Commit Graph

54 Commits

Author SHA1 Message Date
oobabooga
86648d4085 Remove CUDA, keep only pytorch 2023-09-22 08:13:11 -07:00
oobabooga
66363a4d70 Minor changes / reorder some functions 2023-09-22 08:02:21 -07:00
jllllll
69b0aedd95
Fix missing models warning 2023-09-22 01:12:08 -05:00
jllllll
9054c98eca
Use --autostash on git pull 2023-09-21 23:00:33 -05:00
jllllll
498552a92b
More robust installation check for installer 2023-09-21 22:23:23 -05:00
jllllll
6bbfc40d10
Add .git creation to installer 2023-09-21 21:51:58 -05:00
oobabooga
193fe18c8c Resolve conflicts 2023-09-21 17:45:11 -07:00
oobabooga
df39f455ad Merge remote-tracking branch 'second-repo/main' into merge-second-repo 2023-09-21 17:39:54 -07:00
oobabooga
fc2b831692 Basic changes 2023-09-21 15:55:09 -07:00
oobabooga
b04b3957f9 Move one-click-installers into the repository 2023-09-21 15:35:53 -07:00
oobabooga
b74bf5638b Install extensions dependencies before webui dependencies
webui takes precedence over extensions.
2023-08-14 09:15:25 -07:00
jllllll
28e3ce4317
Simplify GPTQ-for-LLaMa installation (#122) 2023-08-10 13:19:47 -03:00
oobabooga
fa4a948b38
Allow users to write one flag per line in CMD_FLAGS.txt 2023-08-09 01:58:23 -03:00
oobabooga
601fc424cd
Several improvements (#117) 2023-08-03 14:39:46 -03:00
jllllll
aca5679968
Properly fix broken gcc_linux-64 package (#115) 2023-08-02 23:39:07 -03:00
jllllll
ecd92d6a4e
Remove unused variable from ROCm GPTQ install (#107) 2023-07-26 22:16:36 -03:00
jllllll
1e3c950c7d
Add AMD GPU support for Linux (#98) 2023-07-26 17:33:02 -03:00
jllllll
52e3b91f5e
Fix broken gxx_linux-64 package. (#106) 2023-07-26 01:55:08 -03:00
oobabooga
cc2ed46d44
Make chat the default again 2023-07-20 18:55:09 -03:00
jllllll
fcb215fed5
Add check for compute support for GPTQ-for-LLaMa (#104)
Installs from main cuda repo if fork not supported
Also removed cuBLAS llama-cpp-python installation in preperation for 4b19b74e6c
2023-07-20 11:11:00 -03:00
jllllll
4df3f72753
Fix GPTQ fail message not being shown on update (#103) 2023-07-19 22:25:09 -03:00
jllllll
11a8fd1eb9
Add cuBLAS llama-cpp-python wheel installation (#102)
Parses requirements.txt using regex to determine required version.
2023-07-16 01:31:33 -03:00
oobabooga
bb79037ebd
Fix wrong pytorch version on Linux+CPU
It was installing nvidia wheels
2023-07-07 20:40:31 -03:00
oobabooga
564a8c507f
Don't launch chat mode by default 2023-07-07 13:32:11 -03:00
jllllll
eac8450ef7
Move special character check to start script (#92)
Also port print_big_message function to batch
2023-06-24 10:06:35 -03:00
jllllll
04cae3e5db
Remove bitsandbytes compatibility workaround (#91)
New bnb does not need it.
Commented out in case it is needed in the futute.
2023-06-21 15:40:41 -03:00
oobabooga
80a615c3ae
Add space 2023-06-20 22:48:45 -03:00
oobabooga
a2116e8b2b
use uninstall -y 2023-06-20 21:24:01 -03:00
oobabooga
c0a1baa46e
Minor changes 2023-06-20 20:23:21 -03:00
jllllll
5cbc0b28f2
Workaround for Peft not updating their package version on the git repo (#88)
* Workaround for Peft not updating their git package version

* Update webui.py

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-06-20 20:21:10 -03:00
jllllll
9bb2fc8cd7
Install Pytorch through pip instead of Conda (#84) 2023-06-20 16:39:23 -03:00
jllllll
b1d05cbbf6
Install exllama (#83)
* Install exllama

* Handle updating exllama
2023-06-17 19:10:36 -03:00
jllllll
b2483e28d1
Check for special characters in path on Windows (#81)
Display warning message if detected
2023-06-17 19:09:22 -03:00
oobabooga
5540335819 Better way to detect if a model has been downloaded 2023-06-01 14:01:19 -03:00
oobabooga
248ef32358 Print a big message for CPU users 2023-06-01 01:40:24 -03:00
oobabooga
290a3374e4 Don't download a model during installation
And some other updates/minor improvements
2023-06-01 01:30:21 -03:00
Sam
dea1bf3d04
Parse g++ version instead of using string matching (#72) 2023-05-31 14:44:36 -03:00
gavin660
97bc7e3fb6
Adds functionality for user to set flags via environment variable (#59) 2023-05-31 14:43:22 -03:00
Sam
5405635305
Install pre-compiled wheels for Linux (#74) 2023-05-31 14:41:54 -03:00
jllllll
be98e74337
Install older bitsandbytes on older gpus + fix llama-cpp-python issue (#75) 2023-05-31 14:41:03 -03:00
oobabooga
c8ce2e777b
Add instructions for CPU mode users 2023-05-25 10:57:52 -03:00
oobabooga
996c49daa7
Remove bitsandbytes installation step
Following 548f05e106
2023-05-25 10:50:20 -03:00
jllllll
4ef2de3486
Fix dependencies downgrading from gptq install (#61) 2023-05-18 12:46:04 -03:00
oobabooga
07510a2414
Change a message 2023-05-18 10:58:37 -03:00
oobabooga
0bcd5b6894
Soothe anxious users 2023-05-18 10:56:49 -03:00
oobabooga
1309cdd257
Add a space 2023-05-10 18:03:12 -03:00
oobabooga
3e19733d35
Remove obsolete comment 2023-05-10 18:01:04 -03:00
oobabooga
d7d3f7f31c
Add a "CMD_FLAGS" variable 2023-05-10 17:54:12 -03:00
oobabooga
b8cfc20e58
Don't install superbooga by default 2023-05-09 14:17:08 -03:00
Semjon Kravtšenko
126d216384
Fix possible crash (#53) 2023-05-06 01:14:09 -03:00