Commit Graph

2770 Commits

Author SHA1 Message Date
jllllll
12baa0e84b
Update for latest GPTQ-for-LLaMa 2023-03-26 00:46:07 -05:00
jllllll
247e8e5b79
Fix for issue in current GPTQ-for-LLaMa. 2023-03-26 00:24:00 -05:00
oobabooga
49c10c5570
Add support for the latest GPTQ models with group-size (#530)
**Warning: old 4-bit weights will not work anymore!**

See here how to get up to date weights: https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model#step-2-get-the-pre-converted-weights
2023-03-26 00:11:33 -03:00
Sean Fitzgerald
0bac80d9eb Potential fix for issues/571 2023-03-25 13:08:45 -07:00
Alex "mcmonkey" Goodwin
f1ba2196b1 make 'model' variables less ambiguous 2023-03-25 12:57:36 -07:00
Alex "mcmonkey" Goodwin
8da237223e document options better 2023-03-25 12:48:35 -07:00
Alex "mcmonkey" Goodwin
8134c4b334 add training/datsets to gitignore for #570 2023-03-25 12:41:18 -07:00
Alex "mcmonkey" Goodwin
5c49a0dcd0 fix error from prepare call running twice in a row 2023-03-25 12:37:32 -07:00
Alex "mcmonkey" Goodwin
7bf601107c automatically strip empty data entries (for better alpaca dataset compat) 2023-03-25 12:28:46 -07:00
Alex "mcmonkey" Goodwin
566898a79a initial lora training tab 2023-03-25 12:08:26 -07:00
Φφ
1a1e420e65 Silero_tts streaming fix
Temporarily suppress the streaming during the audio response as it would interfere with the audio (making it stutter and play anew)
2023-03-25 21:33:30 +03:00
Alex "mcmonkey" Goodwin
9ccf505ccd improve/simplify gitignore
- add repositories
- remove the redundant "/*" on folders
- remove the exclusions for files that already exist
2023-03-25 10:04:00 -07:00
oobabooga
8c8e8b4450
Fix the early stopping callback #559 2023-03-25 12:35:52 -03:00
oobabooga
a1f12d607f
Merge pull request #538 from Ph0rk0z/display-input-context
Add display of context when input was generated
2023-03-25 11:56:18 -03:00
catalpaaa
f740ee558c
Merge branch 'oobabooga:main' into lora-and-model-dir 2023-03-25 01:28:33 -07:00
jllllll
ce9a5e3b53
Update install.bat
Minor fixes
2023-03-25 02:22:02 -05:00
jllllll
2e02d42682 Changed things around to allow Micromamba to work with paths containing spaces. 2023-03-25 01:26:25 -05:00
oobabooga
70f9565f37
Update README.md 2023-03-25 02:35:30 -03:00
oobabooga
25be9698c7
Fix LoRA on mps 2023-03-25 01:18:32 -03:00
oobabooga
3da633a497
Merge pull request #529 from EyeDeck/main
Allow loading of .safetensors through GPTQ-for-LLaMa
2023-03-24 23:51:01 -03:00
jllllll
1e260544cd
Update install.bat
Added C:\Windows\System32 to PATH to avoid issues with broken? Windows installs.
2023-03-24 21:25:14 -05:00
catalpaaa
d51cb8292b Update server.py
yea i should go to bed
2023-03-24 17:36:31 -07:00
catalpaaa
9e2963e0c8 Update server.py 2023-03-24 17:35:45 -07:00
catalpaaa
ec2a1facee Update server.py 2023-03-24 17:34:33 -07:00
catalpaaa
b37c54edcf lora-dir, model-dir and login auth
Added lora-dir, model-dir, and a login auth arguments that points to a file contains usernames and passwords in the format of "u:pw,u:pw,..."
2023-03-24 17:30:18 -07:00
jllllll
fa916aa1de
Update INSTRUCTIONS.txt
Added clarification on new variable added to download-model.bat.
2023-03-24 18:28:46 -05:00
jllllll
586775ad47
Update download-model.bat
Removed redundant %ModelName% variable.
2023-03-24 18:25:49 -05:00
jllllll
bddbc2f898
Update start-webui.bat
Updated virtual environment handling to use Micromamba.
2023-03-24 18:19:23 -05:00
jllllll
2604e3f7ac
Update download-model.bat
Added variables for model selection and text only mode.
Updated virtual environment handling to use Micromamba.
2023-03-24 18:15:24 -05:00
jllllll
24870e51ed
Update micromamba-cmd.bat
Add cd command for admin.
2023-03-24 18:12:02 -05:00
jllllll
f0c82f06c3
Add files via upload
Add script to open cmd within installation environment for easier modification.
2023-03-24 18:09:44 -05:00
oobabooga
9fa47c0eed
Revert GPTQ_loader.py (accident) 2023-03-24 19:57:12 -03:00
oobabooga
a6bf54739c
Revert models.py (accident) 2023-03-24 19:56:45 -03:00
jllllll
eec773b1f4
Update install.bat
Corrected libbitsandbytes_cudaall.dll install.
2023-03-24 17:54:47 -05:00
oobabooga
0a16224451
Update GPTQ_loader.py 2023-03-24 19:54:36 -03:00
oobabooga
a80aa65986
Update models.py 2023-03-24 19:53:20 -03:00
jllllll
817e6c681e
Update install.bat
Added `cd /D "%~dp0"` in case the script is ran as admin.
2023-03-24 17:51:13 -05:00
jllllll
a80a5465f2
Update install.bat
Updated Conda packages and channels to install cuda-toolkit and override 12.0 cuda packages requested by pytorch with their 11.7 equivalent.
Removed Conda installation since we can use the downloaded Micromamba.exe for the same purpose with a smaller footprint.
Removed redundant PATH changes.
Changed %gpuchoice% comparisons to be case-insensitive.
Added additional error handling and removed the use of .tmp files.
Added missing extension requirements.
Added GPTQ installation. Will attempt to compile locally and, if failed, will download and install a precompiled wheel.
Incorporated fixes from one-click-bandaid.
Fixed and expanded first sed command from one-click-bandaid.
libbitsandbytes_cudaall.dll is used here as the cuda116.dll used by one-click-bandaid does not work on my 1080ti. This can be changed if needed.
2023-03-24 17:27:29 -05:00
oobabooga
507db0929d
Do not use empty user messages in chat mode
This allows the bot to send messages by clicking on Generate with empty inputs.
2023-03-24 17:22:22 -03:00
oobabooga
6e1b16c2aa
Update html_generator.py 2023-03-24 17:18:27 -03:00
oobabooga
ffb0187e83
Update chat.py 2023-03-24 17:17:29 -03:00
oobabooga
c14e598f14
Merge pull request #433 from mayaeary/fix/api-reload
Fix api extension duplicating
2023-03-24 16:56:10 -03:00
oobabooga
bfe960731f
Merge branch 'main' into fix/api-reload 2023-03-24 16:54:41 -03:00
oobabooga
4a724ed22f
Reorder imports 2023-03-24 16:53:56 -03:00
oobabooga
8fad84abc2
Update extensions.py 2023-03-24 16:51:27 -03:00
oobabooga
d8e950d6bd
Don't load the model twice when using --lora 2023-03-24 16:30:32 -03:00
oobabooga
fd99995b01
Make the Stop button more consistent in chat mode 2023-03-24 15:59:27 -03:00
Forkoz
b740c5b284
Add display of context when input was generated
Not sure if I did this right but it does move with the conversation and seems to match value.
2023-03-24 08:56:07 -05:00
oobabooga
4f5c2ce785
Fix chat_generation_attempts 2023-03-24 02:03:30 -03:00
oobabooga
04417b658b
Update README.md 2023-03-24 01:40:43 -03:00