Lukas
5ad92c940e
lora training fixes: ( #970 )
...
Fix wrong input format being picked
Fix crash when an entry in the dataset has an attribute of value None
2023-04-12 11:38:01 -03:00
oobabooga
80f4eabb2a
Fix send_pictures extension
2023-04-12 10:27:06 -03:00
oobabooga
8265d45db8
Add send dummy message/reply buttons
...
Useful for starting a new reply.
2023-04-11 22:21:41 -03:00
oobabooga
37d52c96bc
Fix Continue in chat mode
2023-04-11 21:46:17 -03:00
oobabooga
cacbcda208
Two new options: truncation length and ban eos token
2023-04-11 18:46:06 -03:00
catalpaaa
78bbc66fc4
allow custom stopping strings in all modes ( #903 )
2023-04-11 12:30:06 -03:00
oobabooga
0f212093a3
Refactor the UI
...
A single dictionary called 'interface_state' is now passed as input to all functions. The values are updated only when necessary.
The goal is to make it easier to add new elements to the UI.
2023-04-11 11:46:30 -03:00
IggoOnCode
09d8119e3c
Add CPU LoRA training ( #938 )
...
(It's very slow)
2023-04-10 17:29:00 -03:00
Alex "mcmonkey" Goodwin
0caf718a21
add on-page documentation to parameters ( #1008 )
2023-04-10 17:19:12 -03:00
oobabooga
bd04ff27ad
Make the bos token optional
2023-04-10 16:44:22 -03:00
oobabooga
0f1627eff1
Don't treat Intruct mode histories as regular histories
...
* They must now be saved/loaded manually
* Also improved browser caching of pfps
* Also changed the global default preset
2023-04-10 15:48:07 -03:00
oobabooga
769aa900ea
Print the used seed
2023-04-10 10:53:31 -03:00
Alex "mcmonkey" Goodwin
30befe492a
fix random seeds to actually randomize
...
Without this fix, manual seeds get locked in.
2023-04-10 06:29:10 -07:00
oobabooga
1911504f82
Minor bug fix
2023-04-09 23:45:41 -03:00
oobabooga
dba2000d2b
Do things that I am not proud of
2023-04-09 23:40:49 -03:00
oobabooga
65552d2157
Merge branch 'main' of github.com:oobabooga/text-generation-webui
2023-04-09 23:19:53 -03:00
oobabooga
8c6155251a
More robust 4-bit model loading
2023-04-09 23:19:28 -03:00
MarkovInequality
992663fa20
Added xformers support to Llama ( #950 )
2023-04-09 23:08:40 -03:00
Brian O'Connor
625d81f495
Update character log logic ( #977 )
...
* When logs are cleared, save the cleared log over the old log files
* Generate a log file when a character is loaded the first time
2023-04-09 22:20:21 -03:00
oobabooga
a3085dba07
Fix LlamaTokenizer eos_token (attempt)
2023-04-09 21:19:39 -03:00
oobabooga
120f5662cf
Better handle spaces for Continue
2023-04-09 20:37:31 -03:00
oobabooga
b27d757fd1
Minor change
2023-04-09 20:06:20 -03:00
oobabooga
d29f4624e9
Add a Continue button to chat mode
2023-04-09 20:04:16 -03:00
oobabooga
cc693a7546
Remove obsolete code
2023-04-09 00:51:07 -03:00
oobabooga
cb169d0834
Minor formatting changes
2023-04-08 17:34:07 -03:00
oobabooga
0b458bf82d
Simplify a function
2023-04-07 21:37:41 -03:00
Φφ
ffd102e5c0
SD Api Pics extension, v.1.1 ( #596 )
2023-04-07 21:36:04 -03:00
oobabooga
1dc464dcb0
Sort imports
2023-04-07 14:42:03 -03:00
oobabooga
42ea6a3fc0
Change the timing for setup() calls
2023-04-07 12:20:57 -03:00
oobabooga
768354239b
Change training file encoding
2023-04-07 11:15:52 -03:00
oobabooga
6762e62a40
Simplifications
2023-04-07 11:14:32 -03:00
oobabooga
a453d4e9c4
Reorganize some chat functions
2023-04-07 11:07:03 -03:00
Maya
8fa182cfa7
Fix regeneration of first message in instruct mode ( #881 )
2023-04-07 10:45:42 -03:00
oobabooga
46c4654226
More PEP8 stuff
2023-04-07 00:52:02 -03:00
oobabooga
ea6e77df72
Make the code more like PEP8 for readability ( #862 )
2023-04-07 00:15:45 -03:00
OWKenobi
310bf46a94
Instruction Character Vicuna, Instruction Mode Bugfix ( #838 )
2023-04-06 17:40:44 -03:00
oobabooga
113f94b61e
Bump transformers (16-bit llama must be reconverted/redownloaded)
2023-04-06 16:04:03 -03:00
oobabooga
03cb44fc8c
Add new llama.cpp library (2048 context, temperature, etc now work)
2023-04-06 13:12:14 -03:00
EyeDeck
39f3fec913
Broaden GPTQ-for-LLaMA branch support ( #820 )
2023-04-06 12:16:48 -03:00
Alex "mcmonkey" Goodwin
0c7ef26981
Lora trainer improvements ( #763 )
2023-04-06 02:04:11 -03:00
oobabooga
e94ab5dac1
Minor fixes
2023-04-06 01:43:10 -03:00
oobabooga
3f3e42e26c
Refactor several function calls and the API
2023-04-06 01:22:15 -03:00
SDS
378d21e80c
Add LLaMA-Precise preset ( #767 )
2023-04-05 18:52:36 -03:00
Forkoz
8203ce0cac
Stop character pic from being cached when changing chars or clearing. ( #798 )
...
Tested on both FF and chromium
2023-04-05 14:25:01 -03:00
oobabooga
7f66421369
Fix loading characters
2023-04-05 14:22:32 -03:00
oobabooga
e722c240af
Add Instruct mode
2023-04-05 13:54:50 -03:00
oobabooga
3d6cb5ed63
Minor rewrite
2023-04-05 01:21:40 -03:00
oobabooga
f3a2e0b8a9
Disable pre_layer when the model type is not llama
2023-04-05 01:19:26 -03:00
catalpaaa
4ab679480e
allow quantized model to be loaded from model dir ( #760 )
2023-04-04 23:19:38 -03:00
oobabooga
ae1fe45bc0
One more cache reset
2023-04-04 23:15:57 -03:00
oobabooga
8ef89730a5
Try to better handle browser image cache
2023-04-04 23:09:28 -03:00
oobabooga
cc6c7a37f3
Add make_thumbnail function
2023-04-04 23:03:58 -03:00
oobabooga
80dfba05f3
Better crop/resize cached images
2023-04-04 22:52:15 -03:00
oobabooga
65d8a24a6d
Show profile pictures in the Character tab
2023-04-04 22:28:49 -03:00
OWKenobi
ee4547cd34
Detect "vicuna" as llama model type ( #772 )
2023-04-04 13:23:27 -03:00
oobabooga
b24147c7ca
Document --pre_layer
2023-04-03 17:34:25 -03:00
oobabooga
4c9ed09270
Update settings template
2023-04-03 14:59:26 -03:00
OWKenobi
dcf61a8897
"character greeting" displayed and editable on the fly ( #743 )
...
* Add greetings field
* add greeting field and make it interactive
* Minor changes
* Fix a bug
* Simplify clear_chat_log
* Change a label
* Minor change
* Simplifications
* Simplification
* Simplify loading the default character history
* Fix regression
---------
Co-authored-by: oobabooga
2023-04-03 12:16:15 -03:00
Alex "mcmonkey" Goodwin
8b1f20aa04
Fix some old JSON characters not loading ( #740 )
2023-04-03 10:49:28 -03:00
oobabooga
8b442305ac
Rename another variable
2023-04-03 01:15:20 -03:00
oobabooga
08448fb637
Rename a variable
2023-04-03 01:02:11 -03:00
oobabooga
2a267011dc
Use Path.stem for simplicity
2023-04-03 00:56:14 -03:00
Alex "mcmonkey" Goodwin
ea97303509
Apply dialogue format in all character fields not just example dialogue ( #650 )
2023-04-02 21:54:29 -03:00
TheTerrasque
2157bb4319
New yaml character format ( #337 from TheTerrasque/feature/yaml-characters)
...
This doesn't break backward compatibility with JSON characters.
2023-04-02 20:34:25 -03:00
oobabooga
5f3f3faa96
Better handle CUDA out of memory errors in chat mode
2023-04-02 17:48:00 -03:00
oobabooga
b0890a7925
Add shared.is_chat() function
2023-04-01 20:15:00 -03:00
oobabooga
b857f4655b
Update shared.py
2023-04-01 13:56:47 -03:00
oobabooga
fcda3f8776
Add also_return_rows to generate_chat_prompt
2023-04-01 01:12:13 -03:00
oobabooga
2c52310642
Add --threads flag for llama.cpp
2023-03-31 21:18:05 -03:00
oobabooga
eeafd60713
Fix streaming
2023-03-31 19:05:38 -03:00
oobabooga
52065ae4cd
Add repetition_penalty
2023-03-31 19:01:34 -03:00
oobabooga
2259143fec
Fix llama.cpp with --no-stream
2023-03-31 18:43:45 -03:00
oobabooga
3a47a602a3
Detect ggml*.bin files automatically
2023-03-31 17:18:21 -03:00
oobabooga
0aee7341d8
Properly count tokens/s for llama.cpp in chat mode
2023-03-31 17:04:32 -03:00
oobabooga
ea3ba6fc73
Merge branch 'feature/llamacpp' of github.com:thomasantony/text-generation-webui into thomasantony-feature/llamacpp
2023-03-31 14:45:53 -03:00
oobabooga
09b0a3aafb
Add repetition_penalty
2023-03-31 14:45:17 -03:00
oobabooga
4d98623041
Merge branch 'main' into feature/llamacpp
2023-03-31 14:37:04 -03:00
oobabooga
4c27562157
Minor changes
2023-03-31 14:33:46 -03:00
oobabooga
9d1dcf880a
General improvements
2023-03-31 14:27:01 -03:00
oobabooga
770ff0efa9
Merge branch 'main' of github.com:oobabooga/text-generation-webui
2023-03-31 12:22:22 -03:00
oobabooga
1d1d9e40cd
Add seed to settings
2023-03-31 12:22:07 -03:00
Maya
b246d17513
Fix type object is not subscriptable
...
Fix `type object is not subscriptable` on python 3.8
2023-03-31 14:20:31 +03:00
oobabooga
d4a9b5ea97
Remove redundant preset (see the plot in #587 )
2023-03-30 17:34:44 -03:00
Thomas Antony
7fa5d96c22
Update to use new llamacpp API
2023-03-30 11:23:05 +01:00
Thomas Antony
79fa2b6d7e
Add support for alpaca
2023-03-30 11:23:04 +01:00
Thomas Antony
a5f5736e74
Add to text_generation.py
2023-03-30 11:22:38 +01:00
Thomas Antony
7745faa7bb
Add llamacpp to models.py
2023-03-30 11:22:37 +01:00
Thomas Antony
7a562481fa
Initial version of llamacpp_model.py
2023-03-30 11:22:07 +01:00
oobabooga
a21e580782
Move an import
2023-03-29 22:50:58 -03:00
oobabooga
55755e27b9
Don't hardcode prompts in the settings dict/json
2023-03-29 22:47:01 -03:00
oobabooga
1cb9246160
Adapt to the new model names
2023-03-29 21:47:36 -03:00
oobabooga
58349f44a0
Handle training exception for unsupported models
2023-03-29 11:55:34 -03:00
oobabooga
a6d0373063
Fix training dataset loading #636
2023-03-29 11:48:17 -03:00
oobabooga
1edfb96778
Fix loading extensions from within the interface
2023-03-28 23:27:02 -03:00
oobabooga
304f812c63
Gracefully handle CUDA out of memory errors with streaming
2023-03-28 19:20:50 -03:00
oobabooga
010b259dde
Update documentation
2023-03-28 17:46:00 -03:00
oobabooga
0bec15ebcd
Reorder imports
2023-03-28 17:34:15 -03:00
Maya Eary
41ec682834
Disable kernel threshold for gpt-j
2023-03-28 22:45:38 +03:00
Maya
1ac003d41c
Merge branch 'oobabooga:main' into feature/gpt-j-4bit-v2
2023-03-28 22:30:39 +03:00
Maya Eary
1c075d8d21
Fix typo
2023-03-28 20:43:50 +03:00
Maya Eary
c8207d474f
Generalized load_quantized
2023-03-28 20:38:55 +03:00
oobabooga
8579fe51dd
Fix new lines in the HTML tab
2023-03-28 12:59:34 -03:00
Alex "mcmonkey" Goodwin
e817fac542
better defaults
2023-03-27 22:29:23 -07:00
Alex "mcmonkey" Goodwin
2e08af4edf
implement initial Raw Text File Input
...
also bump default Rank & Alpha for values that will make sense in testing if you don't know what you're doing and leave the defaults.
2023-03-27 22:15:32 -07:00
Alex "mcmonkey" Goodwin
b749952fe3
change number minimums to 0
...
gradio calculates 'step' relative to the minimum, so at '1' the step values were all offset awkwardly. 0 isn't valid, but, uh, just don't slam the slider to the left.
2023-03-27 21:22:43 -07:00
Alex "mcmonkey" Goodwin
ec6224f556
use new shared.args.lora_dir
2023-03-27 20:04:16 -07:00
Alex "mcmonkey" Goodwin
31f04dc615
Merge branch 'main' into add-train-lora-tab
2023-03-27 20:03:30 -07:00
oobabooga
53da672315
Fix FlexGen
2023-03-27 23:44:21 -03:00
oobabooga
ee95e55df6
Fix RWKV tokenizer
2023-03-27 23:42:29 -03:00
oobabooga
036163a751
Change description
2023-03-27 23:39:26 -03:00
oobabooga
005f552ea3
Some simplifications
2023-03-27 23:29:52 -03:00
oobabooga
fde92048af
Merge branch 'main' into catalpaaa-lora-and-model-dir
2023-03-27 23:16:44 -03:00
Alex "mcmonkey" Goodwin
8a97f6ba29
corrections per the PR comments
2023-03-27 18:39:06 -07:00
Alex "mcmonkey" Goodwin
7fab7ea1b6
couple missed camelCases
2023-03-27 18:19:06 -07:00
Alex "mcmonkey" Goodwin
6368dad7db
Fix camelCase to snake_case to match repo format standard
2023-03-27 18:17:42 -07:00
oobabooga
2f0571bfa4
Small style changes
2023-03-27 21:24:39 -03:00
oobabooga
c2cad30772
Merge branch 'main' into mcmonkey4eva-add-train-lora-tab
2023-03-27 21:05:44 -03:00
Alex "mcmonkey" Goodwin
9ced75746d
add total time estimate
2023-03-27 10:57:27 -07:00
Alex "mcmonkey" Goodwin
16ea4fc36d
interrupt button
2023-03-27 10:43:01 -07:00
Alex "mcmonkey" Goodwin
8fc723fc95
initial progress tracker in UI
2023-03-27 10:25:08 -07:00
oobabooga
48a6c9513e
Merge pull request #572 from clusterfudge/issues/571
...
Potential fix for issues/571
2023-03-27 14:06:38 -03:00
Alex "mcmonkey" Goodwin
c07bcd0850
add some outputs to indicate progress updates (sorta)
...
Actual progressbar still needed. Also minor formatting fixes.
2023-03-27 09:41:06 -07:00
oobabooga
af65c12900
Change Stop button behavior
2023-03-27 13:23:59 -03:00
Alex "mcmonkey" Goodwin
d911c22af9
use shared rows to make the LoRA Trainer interface a bit more compact / clean
2023-03-27 08:31:49 -07:00
Alex "mcmonkey" Goodwin
e439228ed8
Merge branch 'main' into add-train-lora-tab
2023-03-27 08:21:19 -07:00
oobabooga
3dc61284d5
Handle unloading LoRA from dropdown menu icon
2023-03-27 00:04:43 -03:00
oobabooga
1c77fdca4c
Change notebook mode appearance
2023-03-26 22:20:30 -03:00
oobabooga
49c10c5570
Add support for the latest GPTQ models with group-size ( #530 )
...
**Warning: old 4-bit weights will not work anymore!**
See here how to get up to date weights: https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model#step-2-get-the-pre-converted-weights
2023-03-26 00:11:33 -03:00
Sean Fitzgerald
0bac80d9eb
Potential fix for issues/571
2023-03-25 13:08:45 -07:00
Alex "mcmonkey" Goodwin
f1ba2196b1
make 'model' variables less ambiguous
2023-03-25 12:57:36 -07:00
Alex "mcmonkey" Goodwin
8da237223e
document options better
2023-03-25 12:48:35 -07:00
Alex "mcmonkey" Goodwin
5c49a0dcd0
fix error from prepare call running twice in a row
2023-03-25 12:37:32 -07:00
Alex "mcmonkey" Goodwin
7bf601107c
automatically strip empty data entries (for better alpaca dataset compat)
2023-03-25 12:28:46 -07:00
Alex "mcmonkey" Goodwin
566898a79a
initial lora training tab
2023-03-25 12:08:26 -07:00
oobabooga
8c8e8b4450
Fix the early stopping callback #559
2023-03-25 12:35:52 -03:00
oobabooga
a1f12d607f
Merge pull request #538 from Ph0rk0z/display-input-context
...
Add display of context when input was generated
2023-03-25 11:56:18 -03:00
catalpaaa
f740ee558c
Merge branch 'oobabooga:main' into lora-and-model-dir
2023-03-25 01:28:33 -07:00
oobabooga
25be9698c7
Fix LoRA on mps
2023-03-25 01:18:32 -03:00
oobabooga
3da633a497
Merge pull request #529 from EyeDeck/main
...
Allow loading of .safetensors through GPTQ-for-LLaMa
2023-03-24 23:51:01 -03:00
catalpaaa
b37c54edcf
lora-dir, model-dir and login auth
...
Added lora-dir, model-dir, and a login auth arguments that points to a file contains usernames and passwords in the format of "u:pw,u:pw,..."
2023-03-24 17:30:18 -07:00
oobabooga
9fa47c0eed
Revert GPTQ_loader.py (accident)
2023-03-24 19:57:12 -03:00
oobabooga
a6bf54739c
Revert models.py (accident)
2023-03-24 19:56:45 -03:00
oobabooga
0a16224451
Update GPTQ_loader.py
2023-03-24 19:54:36 -03:00
oobabooga
a80aa65986
Update models.py
2023-03-24 19:53:20 -03:00
oobabooga
507db0929d
Do not use empty user messages in chat mode
...
This allows the bot to send messages by clicking on Generate with empty inputs.
2023-03-24 17:22:22 -03:00
oobabooga
6e1b16c2aa
Update html_generator.py
2023-03-24 17:18:27 -03:00
oobabooga
ffb0187e83
Update chat.py
2023-03-24 17:17:29 -03:00
oobabooga
bfe960731f
Merge branch 'main' into fix/api-reload
2023-03-24 16:54:41 -03:00
oobabooga
8fad84abc2
Update extensions.py
2023-03-24 16:51:27 -03:00
Forkoz
b740c5b284
Add display of context when input was generated
...
Not sure if I did this right but it does move with the conversation and seems to match value.
2023-03-24 08:56:07 -05:00