mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-26 17:50:22 +01:00
0b4ee14edc
Have pip directly download and install backup GPTQ wheel instead of first downloading through curl. Install bitsandbytes from wheel compiled for Windows from modified source. Add clarification of minor, intermittent issue to instructions. Add system32 folder to end of PATH rather than beginning. Add warning when installed under a path containing spaces.
19 lines
1.2 KiB
Plaintext
19 lines
1.2 KiB
Plaintext
Thank you for downloading oobabooga/text-generation-webui.
|
|
Here is how to get it up and running:
|
|
|
|
1. Run the "install" script to install the web UI and its requirements in this folder.
|
|
2. Run the "download-model" script to download a model of your choice. Change TextOnly variable at top of script to download only config files.
|
|
3. Run the "start-webui" script to launch the web UI.
|
|
|
|
To add flags like --chat, --notebook, --extensions, etc, edit the
|
|
"start-webui" script using a text editor and add the desired flags
|
|
to the line that says "python server.py...".
|
|
|
|
To get the latest updates in the future, just re-run the "install" script.
|
|
This will only install the updates, so it should be much faster.
|
|
May need to delete '\text-generation-webui\repositories\GPTQ-for-LLaMa' folder if GPTQ-for-LLaMa needs to be updated.
|
|
|
|
You can open a command-line attached to the virtual environment by running the "micromamba-cmd" script.
|
|
|
|
This installer uses a custom-built Windows-compatible version of bitsandbytes. Source: https://github.com/acpopescu/bitsandbytes/tree/cmake_windows
|
|
When starting the webui, you may encounter an error referencing cuda 116. Starting the webui again should allow bitsandbytes to detect the correct version. |