oobabooga
|
f035b01823
|
Update README.md
|
2023-04-10 16:20:23 -03:00 |
|
Jeff Lefebvre
|
b7ca89ba3f
|
Mention that build-essential is required (#1013)
|
2023-04-10 16:19:10 -03:00 |
|
loeken
|
52339e9b20
|
add make/g++ to docker (#1015)
|
2023-04-10 16:18:07 -03:00 |
|
oobabooga
|
4961f43702
|
Improve header bar colors
|
2023-04-10 16:15:16 -03:00 |
|
oobabooga
|
617530296e
|
Instruct mode color/style improvements
|
2023-04-10 16:04:21 -03:00 |
|
oobabooga
|
0f1627eff1
|
Don't treat Intruct mode histories as regular histories
* They must now be saved/loaded manually
* Also improved browser caching of pfps
* Also changed the global default preset
|
2023-04-10 15:48:07 -03:00 |
|
oobabooga
|
d679c4be13
|
Change a label
|
2023-04-10 11:44:37 -03:00 |
|
oobabooga
|
45244ed125
|
More descriptive download info
|
2023-04-10 11:42:12 -03:00 |
|
oobabooga
|
7e70741a4e
|
Download models from Model tab (#954 from UsamaKenway/main)
|
2023-04-10 11:38:30 -03:00 |
|
oobabooga
|
11b23db8d4
|
Remove unused imports
|
2023-04-10 11:37:42 -03:00 |
|
oobabooga
|
2c14df81a8
|
Use download-model.py to download the model
|
2023-04-10 11:36:39 -03:00 |
|
oobabooga
|
c6e9ba20a4
|
Merge branch 'main' into UsamaKenway-main
|
2023-04-10 11:14:03 -03:00 |
|
oobabooga
|
843f672227
|
fix random seeds to actually randomize (#1004 from mcmonkey4eva/seed-fix)
|
2023-04-10 10:56:12 -03:00 |
|
oobabooga
|
769aa900ea
|
Print the used seed
|
2023-04-10 10:53:31 -03:00 |
|
jllllll
|
254609daca
|
Update llama-cpp-python link to official wheel (#19)
|
2023-04-10 10:48:56 -03:00 |
|
oobabooga
|
32d078487e
|
Add llama-cpp-python to requirements.txt
|
2023-04-10 10:45:51 -03:00 |
|
Alex "mcmonkey" Goodwin
|
30befe492a
|
fix random seeds to actually randomize
Without this fix, manual seeds get locked in.
|
2023-04-10 06:29:10 -07:00 |
|
jllllll
|
c3e1a58cb3
|
Correct llama-cpp-python wheel link (#17)
|
2023-04-09 23:46:54 -03:00 |
|
oobabooga
|
1911504f82
|
Minor bug fix
|
2023-04-09 23:45:41 -03:00 |
|
BlueprintCoding
|
8178fde2cb
|
Added dropdown to character bias. (#986)
|
2023-04-09 23:44:31 -03:00 |
|
oobabooga
|
dba2000d2b
|
Do things that I am not proud of
|
2023-04-09 23:40:49 -03:00 |
|
oobabooga
|
97840c92f9
|
Add working llamaa-cpp-python install from wheel. (#13 from Loufe/oobabooga-windows)
|
2023-04-09 23:23:27 -03:00 |
|
oobabooga
|
65552d2157
|
Merge branch 'main' of github.com:oobabooga/text-generation-webui
|
2023-04-09 23:19:53 -03:00 |
|
oobabooga
|
8c6155251a
|
More robust 4-bit model loading
|
2023-04-09 23:19:28 -03:00 |
|
MarkovInequality
|
992663fa20
|
Added xformers support to Llama (#950)
|
2023-04-09 23:08:40 -03:00 |
|
Brian O'Connor
|
625d81f495
|
Update character log logic (#977)
* When logs are cleared, save the cleared log over the old log files
* Generate a log file when a character is loaded the first time
|
2023-04-09 22:20:21 -03:00 |
|
oobabooga
|
57f768eaad
|
Better preset in api-example.py
|
2023-04-09 22:18:40 -03:00 |
|
oobabooga
|
a3085dba07
|
Fix LlamaTokenizer eos_token (attempt)
|
2023-04-09 21:19:39 -03:00 |
|
oobabooga
|
120f5662cf
|
Better handle spaces for Continue
|
2023-04-09 20:37:31 -03:00 |
|
oobabooga
|
b27d757fd1
|
Minor change
|
2023-04-09 20:06:20 -03:00 |
|
oobabooga
|
d29f4624e9
|
Add a Continue button to chat mode
|
2023-04-09 20:04:16 -03:00 |
|
oobabooga
|
170e0c05c4
|
Typo
|
2023-04-09 17:00:59 -03:00 |
|
oobabooga
|
34ec02d41d
|
Make download-model.py importable
|
2023-04-09 16:59:59 -03:00 |
|
oobabooga
|
f91d3a3ff4
|
server.py readability
|
2023-04-09 14:46:32 -03:00 |
|
Usama Kenway
|
ebdf4c8c12
|
path fixed
|
2023-04-09 16:53:21 +05:00 |
|
Usama Kenway
|
7436dd5b4a
|
download custom model menu (from hugging face) added in model tab
|
2023-04-09 16:11:43 +05:00 |
|
oobabooga
|
bce1b7fbb2
|
Update README.md
|
2023-04-09 02:19:40 -03:00 |
|
oobabooga
|
f7860ce192
|
Update README.md
|
2023-04-09 02:19:17 -03:00 |
|
oobabooga
|
ece8ed2c84
|
Update README.md
|
2023-04-09 02:18:42 -03:00 |
|
oobabooga
|
cc693a7546
|
Remove obsolete code
|
2023-04-09 00:51:07 -03:00 |
|
Lou Bernardi
|
0818bc93ad
|
Add working llamaa-cpp-python install from wheel.
|
2023-04-08 22:44:55 -04:00 |
|
oobabooga
|
2fde50a800
|
Delete docker.md
|
2023-04-08 22:37:54 -03:00 |
|
loeken
|
acc235aced
|
updated docs for docker, setup video added, removed left over GPTQ_VERSION from docker-compose (#940)
|
2023-04-08 22:35:15 -03:00 |
|
Blake Wyatt
|
df561fd896
|
Fix ggml downloading in download-model.py (#915)
|
2023-04-08 18:52:30 -03:00 |
|
oobabooga
|
d272ac46dd
|
Add Pillow as a requirement
|
2023-04-08 18:48:46 -03:00 |
|
oobabooga
|
cb169d0834
|
Minor formatting changes
|
2023-04-08 17:34:07 -03:00 |
|
oobabooga
|
2f16d0afca
|
Remove redundant events
|
2023-04-08 17:32:36 -03:00 |
|
oobabooga
|
a6a00cb82f
|
Properly concatenate chat events
|
2023-04-08 17:25:21 -03:00 |
|
Φφ
|
c97c270040
|
Send_pictures small fix (#546)
|
2023-04-08 01:55:16 -03:00 |
|
oobabooga
|
0b458bf82d
|
Simplify a function
|
2023-04-07 21:37:41 -03:00 |
|