oobabooga
|
511be06dcc
|
Merge branch 'main' of github.com:oobabooga/text-generation-webui
|
2023-03-26 22:21:29 -03:00 |
|
oobabooga
|
1c77fdca4c
|
Change notebook mode appearance
|
2023-03-26 22:20:30 -03:00 |
|
oobabooga
|
9ff6a538b6
|
Bump gradio version
Make sure to upgrade with
`pip install -r requirements.txt --upgrade`
|
2023-03-26 22:11:19 -03:00 |
|
oobabooga
|
a04b7cf264
|
Merge pull request #585 from fkusche/also-download-markdown
Also download Markdown files
|
2023-03-26 14:51:23 -03:00 |
|
Florian Kusche
|
19174842b8
|
Also download Markdown files
|
2023-03-26 19:41:14 +02:00 |
|
oobabooga
|
8222d32240
|
Merge pull request #565 from mcmonkey4eva/improve-gitignore
improve/simplify gitignore
|
2023-03-26 13:31:45 -03:00 |
|
oobabooga
|
49c10c5570
|
Add support for the latest GPTQ models with group-size (#530)
**Warning: old 4-bit weights will not work anymore!**
See here how to get up to date weights: https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model#step-2-get-the-pre-converted-weights
|
2023-03-26 00:11:33 -03:00 |
|
Alex "mcmonkey" Goodwin
|
8134c4b334
|
add training/datsets to gitignore for #570
|
2023-03-25 12:41:18 -07:00 |
|
Alex "mcmonkey" Goodwin
|
9ccf505ccd
|
improve/simplify gitignore
- add repositories
- remove the redundant "/*" on folders
- remove the exclusions for files that already exist
|
2023-03-25 10:04:00 -07:00 |
|
oobabooga
|
8c8e8b4450
|
Fix the early stopping callback #559
|
2023-03-25 12:35:52 -03:00 |
|
oobabooga
|
a1f12d607f
|
Merge pull request #538 from Ph0rk0z/display-input-context
Add display of context when input was generated
|
2023-03-25 11:56:18 -03:00 |
|
oobabooga
|
70f9565f37
|
Update README.md
|
2023-03-25 02:35:30 -03:00 |
|
oobabooga
|
25be9698c7
|
Fix LoRA on mps
|
2023-03-25 01:18:32 -03:00 |
|
oobabooga
|
3da633a497
|
Merge pull request #529 from EyeDeck/main
Allow loading of .safetensors through GPTQ-for-LLaMa
|
2023-03-24 23:51:01 -03:00 |
|
oobabooga
|
9fa47c0eed
|
Revert GPTQ_loader.py (accident)
|
2023-03-24 19:57:12 -03:00 |
|
oobabooga
|
a6bf54739c
|
Revert models.py (accident)
|
2023-03-24 19:56:45 -03:00 |
|
oobabooga
|
0a16224451
|
Update GPTQ_loader.py
|
2023-03-24 19:54:36 -03:00 |
|
oobabooga
|
a80aa65986
|
Update models.py
|
2023-03-24 19:53:20 -03:00 |
|
oobabooga
|
507db0929d
|
Do not use empty user messages in chat mode
This allows the bot to send messages by clicking on Generate with empty inputs.
|
2023-03-24 17:22:22 -03:00 |
|
oobabooga
|
6e1b16c2aa
|
Update html_generator.py
|
2023-03-24 17:18:27 -03:00 |
|
oobabooga
|
ffb0187e83
|
Update chat.py
|
2023-03-24 17:17:29 -03:00 |
|
oobabooga
|
c14e598f14
|
Merge pull request #433 from mayaeary/fix/api-reload
Fix api extension duplicating
|
2023-03-24 16:56:10 -03:00 |
|
oobabooga
|
bfe960731f
|
Merge branch 'main' into fix/api-reload
|
2023-03-24 16:54:41 -03:00 |
|
oobabooga
|
4a724ed22f
|
Reorder imports
|
2023-03-24 16:53:56 -03:00 |
|
oobabooga
|
8fad84abc2
|
Update extensions.py
|
2023-03-24 16:51:27 -03:00 |
|
oobabooga
|
d8e950d6bd
|
Don't load the model twice when using --lora
|
2023-03-24 16:30:32 -03:00 |
|
oobabooga
|
fd99995b01
|
Make the Stop button more consistent in chat mode
|
2023-03-24 15:59:27 -03:00 |
|
Forkoz
|
b740c5b284
|
Add display of context when input was generated
Not sure if I did this right but it does move with the conversation and seems to match value.
|
2023-03-24 08:56:07 -05:00 |
|
oobabooga
|
4f5c2ce785
|
Fix chat_generation_attempts
|
2023-03-24 02:03:30 -03:00 |
|
oobabooga
|
04417b658b
|
Update README.md
|
2023-03-24 01:40:43 -03:00 |
|
oobabooga
|
bb4cb22453
|
Download .pt files using download-model.py (for 4-bit models)
|
2023-03-24 00:49:04 -03:00 |
|
oobabooga
|
143b5b5edf
|
Mention one-click-bandaid in the README
|
2023-03-23 23:28:50 -03:00 |
|
EyeDeck
|
dcfd866402
|
Allow loading of .safetensors through GPTQ-for-LLaMa
|
2023-03-23 21:31:34 -04:00 |
|
oobabooga
|
8747c74339
|
Another missing import
|
2023-03-23 22:19:01 -03:00 |
|
oobabooga
|
7078d168c3
|
Missing import
|
2023-03-23 22:16:08 -03:00 |
|
oobabooga
|
d1327f99f9
|
Fix broken callbacks.py
|
2023-03-23 22:12:24 -03:00 |
|
oobabooga
|
9bdb3c784d
|
Minor fix
|
2023-03-23 22:02:40 -03:00 |
|
oobabooga
|
b0abb327d8
|
Update LoRA.py
|
2023-03-23 22:02:09 -03:00 |
|
oobabooga
|
bf22d16ebc
|
Clear cache while switching LoRAs
|
2023-03-23 21:56:26 -03:00 |
|
oobabooga
|
4578e88ffd
|
Stop the bot from talking for you in chat mode
|
2023-03-23 21:38:20 -03:00 |
|
oobabooga
|
9bf6ecf9e2
|
Fix LoRA device map (attempt)
|
2023-03-23 16:49:41 -03:00 |
|
oobabooga
|
c5ebcc5f7e
|
Change the default names (#518)
* Update shared.py
* Update settings-template.json
|
2023-03-23 13:36:00 -03:00 |
|
oobabooga
|
29bd41d453
|
Fix LoRA in CPU mode
|
2023-03-23 01:05:13 -03:00 |
|
oobabooga
|
eac27f4f55
|
Make LoRAs work in 16-bit mode
|
2023-03-23 00:55:33 -03:00 |
|
oobabooga
|
bfa81e105e
|
Fix FlexGen streaming
|
2023-03-23 00:22:14 -03:00 |
|
oobabooga
|
7b6f85d327
|
Fix markdown headers in light mode
|
2023-03-23 00:13:34 -03:00 |
|
oobabooga
|
de6a09dc7f
|
Properly separate the original prompt from the reply
|
2023-03-23 00:12:40 -03:00 |
|
oobabooga
|
d5fc1bead7
|
Merge pull request #489 from Brawlence/ext-fixes
Extensions performance & memory optimisations
|
2023-03-22 16:10:59 -03:00 |
|
oobabooga
|
bfb1be2820
|
Minor fix
|
2023-03-22 16:09:48 -03:00 |
|
oobabooga
|
0abff499e2
|
Use image.thumbnail
|
2023-03-22 16:03:05 -03:00 |
|