From 36e850fe8923bdd229eff5bdef0019a2d8a01043 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 17:55:41 -0300 Subject: [PATCH 01/20] Update README.md --- README.md | 12 +++++++++--- 1 file changed, 9 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index f9425833..7c49639d 100644 --- a/README.md +++ b/README.md @@ -59,7 +59,7 @@ To define persistent command-line flags like `--listen` or `--api`, edit the `CM #### Other info * There is no need to run any of those scripts as admin/root. -* For additional instructions about AMD setup, WSL setup, and nvcc installation, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). +* For additional instructions about AMD setup and WSL setup, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). * The installer has been tested mostly on NVIDIA GPUs. If you can find a way to improve it for your AMD/Intel Arc/Mac Metal GPU, you are highly encouraged to submit a PR to this repository. The main file to be edited is `one_click.py`. * For automated installation, you can use the `GPU_CHOICE`, `USE_CUDA118`, `LAUNCH_AFTER_INSTALL`, and `INSTALL_EXTENSIONS` environment variables. For instance: `GPU_CHOICE=A USE_CUDA118=FALSE LAUNCH_AFTER_INSTALL=FALSE INSTALL_EXTENSIONS=FALSE ./start_linux.sh`. @@ -98,10 +98,16 @@ conda activate textgen The up-to-date commands can be found here: https://pytorch.org/get-started/locally/. -For NVIDIA, you may also need to manually install the CUDA runtime libraries: +For NVIDIA, you also need to install the CUDA runtime libraries: ``` -conda install -y -c "nvidia/label/cuda-12.1.0" cuda-runtime +conda install -y -c "nvidia/label/cuda-12.1.1" cuda-runtime +``` + +If you need `nvcc` to compile some library manually, replace the command above with + +``` +conda install -y -c "nvidia/label/cuda-12.1.1" cuda ``` #### 3. Install the web UI From f51156705d0b801226b6e431376dea5091870410 Mon Sep 17 00:00:00 2001 From: Bartowski Date: Wed, 13 Dec 2023 16:08:21 -0500 Subject: [PATCH 02/20] Allow symlinked folder within root directory (#4863) --- modules/utils.py | 14 ++++++++------ 1 file changed, 8 insertions(+), 6 deletions(-) diff --git a/modules/utils.py b/modules/utils.py index c1d0fb46..9f9ba0b7 100644 --- a/modules/utils.py +++ b/modules/utils.py @@ -21,16 +21,17 @@ def save_file(fname, contents): return root_folder = Path(__file__).resolve().parent.parent - abs_path = Path(fname).resolve() - rel_path = abs_path.relative_to(root_folder) + abs_path_str = os.path.abspath(fname) + rel_path_str = os.path.relpath(abs_path_str, root_folder) + rel_path = Path(rel_path_str) if rel_path.parts[0] == '..': logger.error(f'Invalid file path: {fname}') return - with open(abs_path, 'w', encoding='utf-8') as f: + with open(abs_path_str, 'w', encoding='utf-8') as f: f.write(contents) - logger.info(f'Saved {abs_path}.') + logger.info(f'Saved {abs_path_str}.') def delete_file(fname): @@ -39,8 +40,9 @@ def delete_file(fname): return root_folder = Path(__file__).resolve().parent.parent - abs_path = Path(fname).resolve() - rel_path = abs_path.relative_to(root_folder) + abs_path_str = os.path.abspath(fname) + rel_path_str = os.path.relpath(abs_path_str, root_folder) + rel_path = Path(rel_path_str) if rel_path.parts[0] == '..': logger.error(f'Invalid file path: {fname}') return From 5754f0c3573b2c310eadb8b26d365d7ceeaadbe1 Mon Sep 17 00:00:00 2001 From: Lounger <4087076+TheLounger@users.noreply.github.com> Date: Thu, 14 Dec 2023 01:54:43 +0100 Subject: [PATCH 03/20] Fix deleting chat logs (#4914) --- modules/utils.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/modules/utils.py b/modules/utils.py index 9f9ba0b7..39d4fb8f 100644 --- a/modules/utils.py +++ b/modules/utils.py @@ -47,8 +47,8 @@ def delete_file(fname): logger.error(f'Invalid file path: {fname}') return - if abs_path.exists(): - abs_path.unlink() + if rel_path.exists(): + rel_path.unlink() logger.info(f'Deleted {fname}.') From d241de86c4157a677339573fa7229e684e7b21a2 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:02:10 -0800 Subject: [PATCH 04/20] Update README --- README.md | 505 ++++++++++++++++++++++++++++-------------------------- 1 file changed, 259 insertions(+), 246 deletions(-) diff --git a/README.md b/README.md index 7c49639d..14845cd3 100644 --- a/README.md +++ b/README.md @@ -10,264 +10,37 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. ## Features -* 3 interface modes: default (two columns), notebook, and chat -* Multiple model backends: [Transformers](https://github.com/huggingface/transformers), [llama.cpp](https://github.com/ggerganov/llama.cpp) (through [llama-cpp-python](https://github.com/abetlen/llama-cpp-python)), [ExLlama](https://github.com/turboderp/exllama), [ExLlamaV2](https://github.com/turboderp/exllamav2), [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ), [AutoAWQ](https://github.com/casper-hansen/AutoAWQ), [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa), [CTransformers](https://github.com/marella/ctransformers), [QuIP#](https://github.com/Cornell-RelaxML/quip-sharp) -* Dropdown menu for quickly switching between different models -* LoRA: load and unload LoRAs on the fly, train a new LoRA using QLoRA -* Precise instruction templates for chat mode, including Llama-2-chat, Alpaca, Vicuna, WizardLM, StableLM, and many others -* 4-bit, 8-bit, and CPU inference through the transformers library -* Use llama.cpp models with transformers samplers (`llamacpp_HF` loader) -* [Multimodal pipelines, including LLaVA and MiniGPT-4](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal) -* [Extensions framework](https://github.com/oobabooga/text-generation-webui/wiki/07-%E2%80%90-Extensions) -* [Custom chat characters](https://github.com/oobabooga/text-generation-webui/wiki/03-%E2%80%90-Parameters-Tab#character) -* Markdown output with LaTeX rendering, to use for instance with [GALACTICA](https://github.com/paperswithcode/galai) -* OpenAI-compatible API server with Chat and Completions endpoints -- see the [examples](https://github.com/oobabooga/text-generation-webui/wiki/12-%E2%80%90-OpenAI-API#examples) +* 3 interface modes: default (two columns), notebook, and chat. +* Multiple model backends: [Transformers](https://github.com/huggingface/transformers), [llama.cpp](https://github.com/ggerganov/llama.cpp) (through [llama-cpp-python](https://github.com/abetlen/llama-cpp-python)), [ExLlama](https://github.com/turboderp/exllama), [ExLlamaV2](https://github.com/turboderp/exllamav2), [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ), [AutoAWQ](https://github.com/casper-hansen/AutoAWQ), [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa), [CTransformers](https://github.com/marella/ctransformers), [QuIP#](https://github.com/Cornell-RelaxML/quip-sharp). +* Dropdown menu for quickly switching between different models. +* Large number of extensions (built-in and user-contributed), including Coqui TTS for voice outputs, Whisper STT for voice inputs, translation, [multimodal pipelines](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal), vector databases, Stable Diffusion integration, and a lot more. See [the wiki](https://github.com/oobabooga/text-generation-webui/wiki/07-%E2%80%90-Extensions) and [the extensions directory](https://github.com/oobabooga/text-generation-webui-extensions) for details. +* Chat with [custom characters](https://github.com/oobabooga/text-generation-webui/wiki/03-%E2%80%90-Parameters-Tab#character). +* Precise templates for instruction-following models, including Llama-2-chat, Alpaca, Vicuna, Mistral, and many others. +* Easy UI for training LoRAs, as well as loading/unloading them on the fly. +* HF transformers integration: load models in 4-bit or 8-bit quantization through bitsandbytes, use llama.cpp with transformers samplers (`llamacpp_HF` loader), CPU inference in 32-bit precision using PyTorch. +* OpenAI-compatible API server with Chat and Completions endpoints -- see the [examples](https://github.com/oobabooga/text-generation-webui/wiki/12-%E2%80%90-OpenAI-API#examples). -## Documentation - -To learn how to use the various features, check out the Documentation: - -https://github.com/oobabooga/text-generation-webui/wiki - -## Installation - -### One-click installers +## How to install 1) Clone or [download](https://github.com/oobabooga/text-generation-webui/archive/refs/heads/main.zip) the repository. 2) Run the `start_linux.sh`, `start_windows.bat`, `start_macos.sh`, or `start_wsl.bat` script depending on your OS. 3) Select your GPU vendor when asked. -4) Have fun! +4) Once the installation ends, browse to `http://localhost:7860/?__theme=dark`. +5) Have fun! -#### How it works - -The script creates a folder called `installer_files` where it sets up a Conda environment using Miniconda. The installation is self-contained: if you want to reinstall, just delete `installer_files` and run the start script again. - -To launch the webui in the future after it is already installed, run the same `start` script. - -#### Getting updates - -Run `update_linux.sh`, `update_windows.bat`, `update_macos.sh`, or `update_wsl.bat`. - -#### Running commands - -If you ever need to install something manually in the `installer_files` environment, you can launch an interactive shell using the cmd script: `cmd_linux.sh`, `cmd_windows.bat`, `cmd_macos.sh`, or `cmd_wsl.bat`. - -#### Defining command-line flags - -To define persistent command-line flags like `--listen` or `--api`, edit the `CMD_FLAGS.txt` file with a text editor and add them there. Flags can also be provided directly to the start scripts, for instance, `./start-linux.sh --listen`. - -#### Other info - -* There is no need to run any of those scripts as admin/root. -* For additional instructions about AMD setup and WSL setup, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). -* The installer has been tested mostly on NVIDIA GPUs. If you can find a way to improve it for your AMD/Intel Arc/Mac Metal GPU, you are highly encouraged to submit a PR to this repository. The main file to be edited is `one_click.py`. -* For automated installation, you can use the `GPU_CHOICE`, `USE_CUDA118`, `LAUNCH_AFTER_INSTALL`, and `INSTALL_EXTENSIONS` environment variables. For instance: `GPU_CHOICE=A USE_CUDA118=FALSE LAUNCH_AFTER_INSTALL=FALSE INSTALL_EXTENSIONS=FALSE ./start_linux.sh`. - -### Manual installation using Conda - -Recommended if you have some experience with the command-line. - -#### 0. Install Conda - -https://docs.conda.io/en/latest/miniconda.html - -On Linux or WSL, it can be automatically installed with these two commands ([source](https://educe-ubc.github.io/conda.html)): - -``` -curl -sL "https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh" > "Miniconda3.sh" -bash Miniconda3.sh -``` - -#### 1. Create a new conda environment - -``` -conda create -n textgen python=3.11 -conda activate textgen -``` - -#### 2. Install Pytorch - -| System | GPU | Command | -|--------|---------|---------| -| Linux/WSL | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121` | -| Linux/WSL | CPU only | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu` | -| Linux | AMD | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.6` | -| MacOS + MPS | Any | `pip3 install torch torchvision torchaudio` | -| Windows | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121` | -| Windows | CPU only | `pip3 install torch torchvision torchaudio` | - -The up-to-date commands can be found here: https://pytorch.org/get-started/locally/. - -For NVIDIA, you also need to install the CUDA runtime libraries: - -``` -conda install -y -c "nvidia/label/cuda-12.1.1" cuda-runtime -``` - -If you need `nvcc` to compile some library manually, replace the command above with - -``` -conda install -y -c "nvidia/label/cuda-12.1.1" cuda -``` - -#### 3. Install the web UI - -``` -git clone https://github.com/oobabooga/text-generation-webui -cd text-generation-webui -pip install -r -``` - -Requirements file to use: - -| GPU | CPU | requirements file to use | -|--------|---------|---------| -| NVIDIA | has AVX2 | `requirements.txt` | -| NVIDIA | no AVX2 | `requirements_noavx2.txt` | -| AMD | has AVX2 | `requirements_amd.txt` | -| AMD | no AVX2 | `requirements_amd_noavx2.txt` | -| CPU only | has AVX2 | `requirements_cpu_only.txt` | -| CPU only | no AVX2 | `requirements_cpu_only_noavx2.txt` | -| Apple | Intel | `requirements_apple_intel.txt` | -| Apple | Apple Silicon | `requirements_apple_silicon.txt` | - -##### AMD GPU on Windows - -1) Use `requirements_cpu_only.txt` or `requirements_cpu_only_noavx2.txt` in the command above. - -2) Manually install llama-cpp-python using the appropriate command for your hardware: [Installation from PyPI](https://github.com/abetlen/llama-cpp-python#installation-with-hardware-acceleration). - * Use the `LLAMA_HIPBLAS=on` toggle. - * Note the [Windows remarks](https://github.com/abetlen/llama-cpp-python#windows-remarks). - -3) Manually install AutoGPTQ: [Installation](https://github.com/PanQiWei/AutoGPTQ#install-from-source). - * Perform the from-source installation - there are no prebuilt ROCm packages for Windows. - -4) Manually install [ExLlama](https://github.com/turboderp/exllama) by simply cloning it into the `repositories` folder (it will be automatically compiled at runtime after that): - -```sh -cd text-generation-webui -git clone https://github.com/turboderp/exllama repositories/exllama -``` - -##### Older NVIDIA GPUs - -1) For Kepler GPUs and older, you will need to install CUDA 11.8 instead of 12: - -``` -pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 -conda install -y -c "nvidia/label/cuda-11.8.0" cuda-runtime -``` - -2) bitsandbytes >= 0.39 may not work. In that case, to use `--load-in-8bit`, you may have to downgrade like this: - * Linux: `pip install bitsandbytes==0.38.1` - * Windows: `pip install https://github.com/jllllll/bitsandbytes-windows-webui/raw/main/bitsandbytes-0.38.1-py3-none-any.whl` - -##### Manual install - -The requirements*.txt above contain various precompiled wheels. If you wish to compile things manually, or if you need to because no suitable wheels are available for your hardware, you can use `requirements_nowheels.txt` and then install your desired loaders manually. - -### Alternative: Docker - -``` -ln -s docker/{nvidia/Dockerfile,docker-compose.yml,.dockerignore} . -cp docker/.env.example .env -# Edit .env and set: -# TORCH_CUDA_ARCH_LIST based on your GPU model -# APP_RUNTIME_GID your host user's group id (run `id -g` in a terminal) -# BUILD_EXTENIONS optionally add comma separated list of extensions to build -docker compose up --build -``` - -* You need to have Docker Compose v2.17 or higher installed. See [this guide](https://github.com/oobabooga/text-generation-webui/wiki/09-%E2%80%90-Docker) for instructions. -* For additional docker files, check out [this repository](https://github.com/Atinoda/text-generation-webui-docker). - -### Updating the requirements - -From time to time, the `requirements*.txt` changes. To update, use these commands: - -``` -conda activate textgen -cd text-generation-webui -pip install -r --upgrade -``` - -## Downloading models - -Models should be placed in the `text-generation-webui/models` folder. They are usually downloaded from [Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads). - -* Transformers or GPTQ models are made of several files and must be placed in a subfolder. Example: - -``` -text-generation-webui -├── models -│   ├── lmsys_vicuna-33b-v1.3 -│   │   ├── config.json -│   │   ├── generation_config.json -│   │   ├── pytorch_model-00001-of-00007.bin -│   │   ├── pytorch_model-00002-of-00007.bin -│   │   ├── pytorch_model-00003-of-00007.bin -│   │   ├── pytorch_model-00004-of-00007.bin -│   │   ├── pytorch_model-00005-of-00007.bin -│   │   ├── pytorch_model-00006-of-00007.bin -│   │   ├── pytorch_model-00007-of-00007.bin -│   │   ├── pytorch_model.bin.index.json -│   │   ├── special_tokens_map.json -│   │   ├── tokenizer_config.json -│   │   └── tokenizer.model -``` - -* GGUF models are a single file and should be placed directly into `models`. Example: - -``` -text-generation-webui -├── models -│   ├── llama-2-13b-chat.Q4_K_M.gguf -``` - -In both cases, you can use the "Model" tab of the UI to download the model from Hugging Face automatically. It is also possible to download via the command-line with `python download-model.py organization/model` (use `--help` to see all the options). - -#### GPT-4chan +To launch the web UI again in the future, run the same `start_` script that you used to install it.
-Instructions +Setup details -[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options: - -* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model) -* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/) - -The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version. - -After downloading the model, follow these steps: - -1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`. -2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json). -3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan): - -``` -python download-model.py EleutherAI/gpt-j-6B --text-only -``` - -When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format: - -![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png) - -
- -## Starting the web UI - - conda activate textgen - cd text-generation-webui - python server.py - -Then browse to - -`http://localhost:7860/?__theme=dark` - -Optionally, you can use the following command-line flags: +Command-line flags can be passed to that script. Alternatively, you can place your flags in the `CMD_FLAGS.txt` file. +
+ +Command-line flags list + #### Basic settings | Flag | Description | @@ -431,6 +204,246 @@ Optionally, you can use the following command-line flags: |---------------------------------------|-------------| | `--multimodal-pipeline PIPELINE` | The multimodal pipeline to use. Examples: `llava-7b`, `llava-13b`. | + +
+ +### One-click-installer + +#### How it works + +The script creates a folder called `installer_files` where it sets up a Conda environment using Miniconda. The installation is self-contained: if you want to reinstall, just delete `installer_files` and run the start script again. + +To launch the webui in the future after it is already installed, run the same `start` script. + +#### Getting updates + +Run `update_linux.sh`, `update_windows.bat`, `update_macos.sh`, or `update_wsl.bat`. + +#### Running commands + +If you ever need to install something manually in the `installer_files` environment, you can launch an interactive shell using the cmd script: `cmd_linux.sh`, `cmd_windows.bat`, `cmd_macos.sh`, or `cmd_wsl.bat`. + +#### Defining command-line flags + +To define persistent command-line flags like `--listen` or `--api`, edit the `CMD_FLAGS.txt` file with a text editor and add them there. Flags can also be provided directly to the start scripts, for instance, `./start-linux.sh --listen`. + +#### Other info + +* There is no need to run any of those scripts as admin/root. +* For additional instructions about AMD setup and WSL setup, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). +* The installer has been tested mostly on NVIDIA GPUs. If you can find a way to improve it for your AMD/Intel Arc/Mac Metal GPU, you are highly encouraged to submit a PR to this repository. The main file to be edited is `one_click.py`. +* For automated installation, you can use the `GPU_CHOICE`, `USE_CUDA118`, `LAUNCH_AFTER_INSTALL`, and `INSTALL_EXTENSIONS` environment variables. For instance: `GPU_CHOICE=A USE_CUDA118=FALSE LAUNCH_AFTER_INSTALL=FALSE INSTALL_EXTENSIONS=FALSE ./start_linux.sh`. + +### Manual installation using Conda + +Recommended if you have some experience with the command-line. + +#### 0. Install Conda + +https://docs.conda.io/en/latest/miniconda.html + +On Linux or WSL, it can be automatically installed with these two commands ([source](https://educe-ubc.github.io/conda.html)): + +``` +curl -sL "https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh" > "Miniconda3.sh" +bash Miniconda3.sh +``` + +#### 1. Create a new conda environment + +``` +conda create -n textgen python=3.11 +conda activate textgen +``` + +#### 2. Install Pytorch + +| System | GPU | Command | +|--------|---------|---------| +| Linux/WSL | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121` | +| Linux/WSL | CPU only | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu` | +| Linux | AMD | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.6` | +| MacOS + MPS | Any | `pip3 install torch torchvision torchaudio` | +| Windows | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121` | +| Windows | CPU only | `pip3 install torch torchvision torchaudio` | + +The up-to-date commands can be found here: https://pytorch.org/get-started/locally/. + +For NVIDIA, you also need to install the CUDA runtime libraries: + +``` +conda install -y -c "nvidia/label/cuda-12.1.1" cuda-runtime +``` + +If you need `nvcc` to compile some library manually, replace the command above with + +``` +conda install -y -c "nvidia/label/cuda-12.1.1" cuda +``` + +#### 3. Install the web UI + +``` +git clone https://github.com/oobabooga/text-generation-webui +cd text-generation-webui +pip install -r +``` + +Requirements file to use: + +| GPU | CPU | requirements file to use | +|--------|---------|---------| +| NVIDIA | has AVX2 | `requirements.txt` | +| NVIDIA | no AVX2 | `requirements_noavx2.txt` | +| AMD | has AVX2 | `requirements_amd.txt` | +| AMD | no AVX2 | `requirements_amd_noavx2.txt` | +| CPU only | has AVX2 | `requirements_cpu_only.txt` | +| CPU only | no AVX2 | `requirements_cpu_only_noavx2.txt` | +| Apple | Intel | `requirements_apple_intel.txt` | +| Apple | Apple Silicon | `requirements_apple_silicon.txt` | + +### Start the web UI + + conda activate textgen + cd text-generation-webui + python server.py + +Then browse to + +`http://localhost:7860/?__theme=dark` + +##### AMD GPU on Windows + +1) Use `requirements_cpu_only.txt` or `requirements_cpu_only_noavx2.txt` in the command above. + +2) Manually install llama-cpp-python using the appropriate command for your hardware: [Installation from PyPI](https://github.com/abetlen/llama-cpp-python#installation-with-hardware-acceleration). + * Use the `LLAMA_HIPBLAS=on` toggle. + * Note the [Windows remarks](https://github.com/abetlen/llama-cpp-python#windows-remarks). + +3) Manually install AutoGPTQ: [Installation](https://github.com/PanQiWei/AutoGPTQ#install-from-source). + * Perform the from-source installation - there are no prebuilt ROCm packages for Windows. + +4) Manually install [ExLlama](https://github.com/turboderp/exllama) by simply cloning it into the `repositories` folder (it will be automatically compiled at runtime after that): + +```sh +cd text-generation-webui +git clone https://github.com/turboderp/exllama repositories/exllama +``` + +##### Older NVIDIA GPUs + +1) For Kepler GPUs and older, you will need to install CUDA 11.8 instead of 12: + +``` +pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 +conda install -y -c "nvidia/label/cuda-11.8.0" cuda-runtime +``` + +2) bitsandbytes >= 0.39 may not work. In that case, to use `--load-in-8bit`, you may have to downgrade like this: + * Linux: `pip install bitsandbytes==0.38.1` + * Windows: `pip install https://github.com/jllllll/bitsandbytes-windows-webui/raw/main/bitsandbytes-0.38.1-py3-none-any.whl` + +##### Manual install + +The requirements*.txt above contain various precompiled wheels. If you wish to compile things manually, or if you need to because no suitable wheels are available for your hardware, you can use `requirements_nowheels.txt` and then install your desired loaders manually. + +### Alternative: Docker + +``` +ln -s docker/{nvidia/Dockerfile,docker-compose.yml,.dockerignore} . +cp docker/.env.example .env +# Edit .env and set: +# TORCH_CUDA_ARCH_LIST based on your GPU model +# APP_RUNTIME_GID your host user's group id (run `id -g` in a terminal) +# BUILD_EXTENIONS optionally add comma separated list of extensions to build +docker compose up --build +``` + +* You need to have Docker Compose v2.17 or higher installed. See [this guide](https://github.com/oobabooga/text-generation-webui/wiki/09-%E2%80%90-Docker) for instructions. +* For additional docker files, check out [this repository](https://github.com/Atinoda/text-generation-webui-docker). + +### Updating the requirements + +From time to time, the `requirements*.txt` changes. To update, use these commands: + +``` +conda activate textgen +cd text-generation-webui +pip install -r --upgrade +``` + + + + +## Documentation + +https://github.com/oobabooga/text-generation-webui/wiki + +## Downloading models + +Models should be placed in the `text-generation-webui/models` folder. They are usually downloaded from [Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads). + +* GGUF models are a single file and should be placed directly into `models`. Example: + +``` +text-generation-webui +├── models +│   ├── llama-2-13b-chat.Q4_K_M.gguf +``` + +* Other models (like 16-bit transformers models and GPTQ models) are made of several files and must be placed in a subfolder. Example: + +``` +text-generation-webui +├── models +│   ├── lmsys_vicuna-33b-v1.3 +│   │   ├── config.json +│   │   ├── generation_config.json +│   │   ├── pytorch_model-00001-of-00007.bin +│   │   ├── pytorch_model-00002-of-00007.bin +│   │   ├── pytorch_model-00003-of-00007.bin +│   │   ├── pytorch_model-00004-of-00007.bin +│   │   ├── pytorch_model-00005-of-00007.bin +│   │   ├── pytorch_model-00006-of-00007.bin +│   │   ├── pytorch_model-00007-of-00007.bin +│   │   ├── pytorch_model.bin.index.json +│   │   ├── special_tokens_map.json +│   │   ├── tokenizer_config.json +│   │   └── tokenizer.model +``` + +In both cases, you can use the "Model" tab of the UI to download the model from Hugging Face automatically. It is also possible to download via the command-line with `python download-model.py organization/model` (use `--help` to see all the options). + +#### GPT-4chan + +
+ +Instructions + + +[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options: + +* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model) +* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/) + +The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version. + +After downloading the model, follow these steps: + +1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`. +2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json). +3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan): + +``` +python download-model.py EleutherAI/gpt-j-6B --text-only +``` + +When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format: + +![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png) + +
+ ## Google Colab notebook https://colab.research.google.com/github/oobabooga/text-generation-webui/blob/main/Colab-TextGen-GPU.ipynb From 0d5ca05ab9e510549ba5f7b8d5fd56f3ec1e3000 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:06:04 -0800 Subject: [PATCH 05/20] Update README --- README.md | 15 ++++++++------- 1 file changed, 8 insertions(+), 7 deletions(-) diff --git a/README.md b/README.md index 14845cd3..53a99d88 100644 --- a/README.md +++ b/README.md @@ -14,7 +14,7 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. * Multiple model backends: [Transformers](https://github.com/huggingface/transformers), [llama.cpp](https://github.com/ggerganov/llama.cpp) (through [llama-cpp-python](https://github.com/abetlen/llama-cpp-python)), [ExLlama](https://github.com/turboderp/exllama), [ExLlamaV2](https://github.com/turboderp/exllamav2), [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ), [AutoAWQ](https://github.com/casper-hansen/AutoAWQ), [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa), [CTransformers](https://github.com/marella/ctransformers), [QuIP#](https://github.com/Cornell-RelaxML/quip-sharp). * Dropdown menu for quickly switching between different models. * Large number of extensions (built-in and user-contributed), including Coqui TTS for voice outputs, Whisper STT for voice inputs, translation, [multimodal pipelines](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal), vector databases, Stable Diffusion integration, and a lot more. See [the wiki](https://github.com/oobabooga/text-generation-webui/wiki/07-%E2%80%90-Extensions) and [the extensions directory](https://github.com/oobabooga/text-generation-webui-extensions) for details. -* Chat with [custom characters](https://github.com/oobabooga/text-generation-webui/wiki/03-%E2%80%90-Parameters-Tab#character). +* [Chat with custom characters](https://github.com/oobabooga/text-generation-webui/wiki/03-%E2%80%90-Parameters-Tab#character). * Precise templates for instruction-following models, including Llama-2-chat, Alpaca, Vicuna, Mistral, and many others. * Easy UI for training LoRAs, as well as loading/unloading them on the fly. * HF transformers integration: load models in 4-bit or 8-bit quantization through bitsandbytes, use llama.cpp with transformers samplers (`llamacpp_HF` loader), CPU inference in 32-bit precision using PyTorch. @@ -35,12 +35,6 @@ To launch the web UI again in the future, run the same `start_` script that you Setup details -Command-line flags can be passed to that script. Alternatively, you can place your flags in the `CMD_FLAGS.txt` file. - -
- -Command-line flags list - #### Basic settings | Flag | Description | @@ -207,6 +201,13 @@ Command-line flags list
+Command-line flags can be passed to that script. Alternatively, you can place your flags in the `CMD_FLAGS.txt` file. + +
+ +Command-line flags list + + ### One-click-installer #### How it works From 634518a4128df1a0028b306ea8b94f29dcdee363 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:08:41 -0800 Subject: [PATCH 06/20] Update README --- README.md | 350 +++++++++++++++++++++++++++--------------------------- 1 file changed, 175 insertions(+), 175 deletions(-) diff --git a/README.md b/README.md index 53a99d88..b26034f9 100644 --- a/README.md +++ b/README.md @@ -32,9 +32,183 @@ To launch the web UI again in the future, run the same `start_` script that you
-Setup details +Setup details and information about installing manually +### One-click-installer + +#### How it works + +The script creates a folder called `installer_files` where it sets up a Conda environment using Miniconda. The installation is self-contained: if you want to reinstall, just delete `installer_files` and run the start script again. + +To launch the webui in the future after it is already installed, run the same `start` script. + +#### Getting updates + +Run `update_linux.sh`, `update_windows.bat`, `update_macos.sh`, or `update_wsl.bat`. + +#### Running commands + +If you ever need to install something manually in the `installer_files` environment, you can launch an interactive shell using the cmd script: `cmd_linux.sh`, `cmd_windows.bat`, `cmd_macos.sh`, or `cmd_wsl.bat`. + +#### Defining command-line flags + +To define persistent command-line flags like `--listen` or `--api`, edit the `CMD_FLAGS.txt` file with a text editor and add them there. Flags can also be provided directly to the start scripts, for instance, `./start-linux.sh --listen`. + +#### Other info + +* There is no need to run any of those scripts as admin/root. +* For additional instructions about AMD setup and WSL setup, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). +* The installer has been tested mostly on NVIDIA GPUs. If you can find a way to improve it for your AMD/Intel Arc/Mac Metal GPU, you are highly encouraged to submit a PR to this repository. The main file to be edited is `one_click.py`. +* For automated installation, you can use the `GPU_CHOICE`, `USE_CUDA118`, `LAUNCH_AFTER_INSTALL`, and `INSTALL_EXTENSIONS` environment variables. For instance: `GPU_CHOICE=A USE_CUDA118=FALSE LAUNCH_AFTER_INSTALL=FALSE INSTALL_EXTENSIONS=FALSE ./start_linux.sh`. + +### Manual installation using Conda + +Recommended if you have some experience with the command-line. + +#### 0. Install Conda + +https://docs.conda.io/en/latest/miniconda.html + +On Linux or WSL, it can be automatically installed with these two commands ([source](https://educe-ubc.github.io/conda.html)): + +``` +curl -sL "https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh" > "Miniconda3.sh" +bash Miniconda3.sh +``` + +#### 1. Create a new conda environment + +``` +conda create -n textgen python=3.11 +conda activate textgen +``` + +#### 2. Install Pytorch + +| System | GPU | Command | +|--------|---------|---------| +| Linux/WSL | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121` | +| Linux/WSL | CPU only | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu` | +| Linux | AMD | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.6` | +| MacOS + MPS | Any | `pip3 install torch torchvision torchaudio` | +| Windows | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121` | +| Windows | CPU only | `pip3 install torch torchvision torchaudio` | + +The up-to-date commands can be found here: https://pytorch.org/get-started/locally/. + +For NVIDIA, you also need to install the CUDA runtime libraries: + +``` +conda install -y -c "nvidia/label/cuda-12.1.1" cuda-runtime +``` + +If you need `nvcc` to compile some library manually, replace the command above with + +``` +conda install -y -c "nvidia/label/cuda-12.1.1" cuda +``` + +#### 3. Install the web UI + +``` +git clone https://github.com/oobabooga/text-generation-webui +cd text-generation-webui +pip install -r +``` + +Requirements file to use: + +| GPU | CPU | requirements file to use | +|--------|---------|---------| +| NVIDIA | has AVX2 | `requirements.txt` | +| NVIDIA | no AVX2 | `requirements_noavx2.txt` | +| AMD | has AVX2 | `requirements_amd.txt` | +| AMD | no AVX2 | `requirements_amd_noavx2.txt` | +| CPU only | has AVX2 | `requirements_cpu_only.txt` | +| CPU only | no AVX2 | `requirements_cpu_only_noavx2.txt` | +| Apple | Intel | `requirements_apple_intel.txt` | +| Apple | Apple Silicon | `requirements_apple_silicon.txt` | + +### Start the web UI + + conda activate textgen + cd text-generation-webui + python server.py + +Then browse to + +`http://localhost:7860/?__theme=dark` + +##### AMD GPU on Windows + +1) Use `requirements_cpu_only.txt` or `requirements_cpu_only_noavx2.txt` in the command above. + +2) Manually install llama-cpp-python using the appropriate command for your hardware: [Installation from PyPI](https://github.com/abetlen/llama-cpp-python#installation-with-hardware-acceleration). + * Use the `LLAMA_HIPBLAS=on` toggle. + * Note the [Windows remarks](https://github.com/abetlen/llama-cpp-python#windows-remarks). + +3) Manually install AutoGPTQ: [Installation](https://github.com/PanQiWei/AutoGPTQ#install-from-source). + * Perform the from-source installation - there are no prebuilt ROCm packages for Windows. + +4) Manually install [ExLlama](https://github.com/turboderp/exllama) by simply cloning it into the `repositories` folder (it will be automatically compiled at runtime after that): + +```sh +cd text-generation-webui +git clone https://github.com/turboderp/exllama repositories/exllama +``` + +##### Older NVIDIA GPUs + +1) For Kepler GPUs and older, you will need to install CUDA 11.8 instead of 12: + +``` +pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 +conda install -y -c "nvidia/label/cuda-11.8.0" cuda-runtime +``` + +2) bitsandbytes >= 0.39 may not work. In that case, to use `--load-in-8bit`, you may have to downgrade like this: + * Linux: `pip install bitsandbytes==0.38.1` + * Windows: `pip install https://github.com/jllllll/bitsandbytes-windows-webui/raw/main/bitsandbytes-0.38.1-py3-none-any.whl` + +##### Manual install + +The requirements*.txt above contain various precompiled wheels. If you wish to compile things manually, or if you need to because no suitable wheels are available for your hardware, you can use `requirements_nowheels.txt` and then install your desired loaders manually. + +### Alternative: Docker + +``` +ln -s docker/{nvidia/Dockerfile,docker-compose.yml,.dockerignore} . +cp docker/.env.example .env +# Edit .env and set: +# TORCH_CUDA_ARCH_LIST based on your GPU model +# APP_RUNTIME_GID your host user's group id (run `id -g` in a terminal) +# BUILD_EXTENIONS optionally add comma separated list of extensions to build +docker compose up --build +``` + +* You need to have Docker Compose v2.17 or higher installed. See [this guide](https://github.com/oobabooga/text-generation-webui/wiki/09-%E2%80%90-Docker) for instructions. +* For additional docker files, check out [this repository](https://github.com/Atinoda/text-generation-webui-docker). + +### Updating the requirements + +From time to time, the `requirements*.txt` changes. To update, use these commands: + +``` +conda activate textgen +cd text-generation-webui +pip install -r --upgrade +``` +
+ +Command-line flags can be passed to the `start_` script. Alternatively, you can open the file `CMD_FLAGS.txt` with a text editor and add your flags there. + +
+ +Command-line flags list + + + #### Basic settings | Flag | Description | @@ -199,180 +373,6 @@ Setup details | `--multimodal-pipeline PIPELINE` | The multimodal pipeline to use. Examples: `llava-7b`, `llava-13b`. | -
- -Command-line flags can be passed to that script. Alternatively, you can place your flags in the `CMD_FLAGS.txt` file. - -
- -Command-line flags list - - -### One-click-installer - -#### How it works - -The script creates a folder called `installer_files` where it sets up a Conda environment using Miniconda. The installation is self-contained: if you want to reinstall, just delete `installer_files` and run the start script again. - -To launch the webui in the future after it is already installed, run the same `start` script. - -#### Getting updates - -Run `update_linux.sh`, `update_windows.bat`, `update_macos.sh`, or `update_wsl.bat`. - -#### Running commands - -If you ever need to install something manually in the `installer_files` environment, you can launch an interactive shell using the cmd script: `cmd_linux.sh`, `cmd_windows.bat`, `cmd_macos.sh`, or `cmd_wsl.bat`. - -#### Defining command-line flags - -To define persistent command-line flags like `--listen` or `--api`, edit the `CMD_FLAGS.txt` file with a text editor and add them there. Flags can also be provided directly to the start scripts, for instance, `./start-linux.sh --listen`. - -#### Other info - -* There is no need to run any of those scripts as admin/root. -* For additional instructions about AMD setup and WSL setup, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). -* The installer has been tested mostly on NVIDIA GPUs. If you can find a way to improve it for your AMD/Intel Arc/Mac Metal GPU, you are highly encouraged to submit a PR to this repository. The main file to be edited is `one_click.py`. -* For automated installation, you can use the `GPU_CHOICE`, `USE_CUDA118`, `LAUNCH_AFTER_INSTALL`, and `INSTALL_EXTENSIONS` environment variables. For instance: `GPU_CHOICE=A USE_CUDA118=FALSE LAUNCH_AFTER_INSTALL=FALSE INSTALL_EXTENSIONS=FALSE ./start_linux.sh`. - -### Manual installation using Conda - -Recommended if you have some experience with the command-line. - -#### 0. Install Conda - -https://docs.conda.io/en/latest/miniconda.html - -On Linux or WSL, it can be automatically installed with these two commands ([source](https://educe-ubc.github.io/conda.html)): - -``` -curl -sL "https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh" > "Miniconda3.sh" -bash Miniconda3.sh -``` - -#### 1. Create a new conda environment - -``` -conda create -n textgen python=3.11 -conda activate textgen -``` - -#### 2. Install Pytorch - -| System | GPU | Command | -|--------|---------|---------| -| Linux/WSL | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121` | -| Linux/WSL | CPU only | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu` | -| Linux | AMD | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.6` | -| MacOS + MPS | Any | `pip3 install torch torchvision torchaudio` | -| Windows | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121` | -| Windows | CPU only | `pip3 install torch torchvision torchaudio` | - -The up-to-date commands can be found here: https://pytorch.org/get-started/locally/. - -For NVIDIA, you also need to install the CUDA runtime libraries: - -``` -conda install -y -c "nvidia/label/cuda-12.1.1" cuda-runtime -``` - -If you need `nvcc` to compile some library manually, replace the command above with - -``` -conda install -y -c "nvidia/label/cuda-12.1.1" cuda -``` - -#### 3. Install the web UI - -``` -git clone https://github.com/oobabooga/text-generation-webui -cd text-generation-webui -pip install -r -``` - -Requirements file to use: - -| GPU | CPU | requirements file to use | -|--------|---------|---------| -| NVIDIA | has AVX2 | `requirements.txt` | -| NVIDIA | no AVX2 | `requirements_noavx2.txt` | -| AMD | has AVX2 | `requirements_amd.txt` | -| AMD | no AVX2 | `requirements_amd_noavx2.txt` | -| CPU only | has AVX2 | `requirements_cpu_only.txt` | -| CPU only | no AVX2 | `requirements_cpu_only_noavx2.txt` | -| Apple | Intel | `requirements_apple_intel.txt` | -| Apple | Apple Silicon | `requirements_apple_silicon.txt` | - -### Start the web UI - - conda activate textgen - cd text-generation-webui - python server.py - -Then browse to - -`http://localhost:7860/?__theme=dark` - -##### AMD GPU on Windows - -1) Use `requirements_cpu_only.txt` or `requirements_cpu_only_noavx2.txt` in the command above. - -2) Manually install llama-cpp-python using the appropriate command for your hardware: [Installation from PyPI](https://github.com/abetlen/llama-cpp-python#installation-with-hardware-acceleration). - * Use the `LLAMA_HIPBLAS=on` toggle. - * Note the [Windows remarks](https://github.com/abetlen/llama-cpp-python#windows-remarks). - -3) Manually install AutoGPTQ: [Installation](https://github.com/PanQiWei/AutoGPTQ#install-from-source). - * Perform the from-source installation - there are no prebuilt ROCm packages for Windows. - -4) Manually install [ExLlama](https://github.com/turboderp/exllama) by simply cloning it into the `repositories` folder (it will be automatically compiled at runtime after that): - -```sh -cd text-generation-webui -git clone https://github.com/turboderp/exllama repositories/exllama -``` - -##### Older NVIDIA GPUs - -1) For Kepler GPUs and older, you will need to install CUDA 11.8 instead of 12: - -``` -pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 -conda install -y -c "nvidia/label/cuda-11.8.0" cuda-runtime -``` - -2) bitsandbytes >= 0.39 may not work. In that case, to use `--load-in-8bit`, you may have to downgrade like this: - * Linux: `pip install bitsandbytes==0.38.1` - * Windows: `pip install https://github.com/jllllll/bitsandbytes-windows-webui/raw/main/bitsandbytes-0.38.1-py3-none-any.whl` - -##### Manual install - -The requirements*.txt above contain various precompiled wheels. If you wish to compile things manually, or if you need to because no suitable wheels are available for your hardware, you can use `requirements_nowheels.txt` and then install your desired loaders manually. - -### Alternative: Docker - -``` -ln -s docker/{nvidia/Dockerfile,docker-compose.yml,.dockerignore} . -cp docker/.env.example .env -# Edit .env and set: -# TORCH_CUDA_ARCH_LIST based on your GPU model -# APP_RUNTIME_GID your host user's group id (run `id -g` in a terminal) -# BUILD_EXTENIONS optionally add comma separated list of extensions to build -docker compose up --build -``` - -* You need to have Docker Compose v2.17 or higher installed. See [this guide](https://github.com/oobabooga/text-generation-webui/wiki/09-%E2%80%90-Docker) for instructions. -* For additional docker files, check out [this repository](https://github.com/Atinoda/text-generation-webui-docker). - -### Updating the requirements - -From time to time, the `requirements*.txt` changes. To update, use these commands: - -``` -conda activate textgen -cd text-generation-webui -pip install -r --upgrade -``` -
From aafd15109d294d6cea5699f61ee00e9933b1ba0b Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:15:58 -0800 Subject: [PATCH 07/20] Update README --- README.md | 20 +++++++++++++------- 1 file changed, 13 insertions(+), 7 deletions(-) diff --git a/README.md b/README.md index b26034f9..c15d25fe 100644 --- a/README.md +++ b/README.md @@ -13,7 +13,7 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. * 3 interface modes: default (two columns), notebook, and chat. * Multiple model backends: [Transformers](https://github.com/huggingface/transformers), [llama.cpp](https://github.com/ggerganov/llama.cpp) (through [llama-cpp-python](https://github.com/abetlen/llama-cpp-python)), [ExLlama](https://github.com/turboderp/exllama), [ExLlamaV2](https://github.com/turboderp/exllamav2), [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ), [AutoAWQ](https://github.com/casper-hansen/AutoAWQ), [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa), [CTransformers](https://github.com/marella/ctransformers), [QuIP#](https://github.com/Cornell-RelaxML/quip-sharp). * Dropdown menu for quickly switching between different models. -* Large number of extensions (built-in and user-contributed), including Coqui TTS for voice outputs, Whisper STT for voice inputs, translation, [multimodal pipelines](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal), vector databases, Stable Diffusion integration, and a lot more. See [the wiki](https://github.com/oobabooga/text-generation-webui/wiki/07-%E2%80%90-Extensions) and [the extensions directory](https://github.com/oobabooga/text-generation-webui-extensions) for details. +* Large number of extensions (built-in and user-contributed), including Coqui TTS for realistic voice outputs, Whisper STT for voice inputs, translation, [multimodal pipelines](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal), vector databases, Stable Diffusion integration, and a lot more. See [the wiki](https://github.com/oobabooga/text-generation-webui/wiki/07-%E2%80%90-Extensions) and [the extensions directory](https://github.com/oobabooga/text-generation-webui-extensions) for details. * [Chat with custom characters](https://github.com/oobabooga/text-generation-webui/wiki/03-%E2%80%90-Parameters-Tab#character). * Precise templates for instruction-following models, including Llama-2-chat, Alpaca, Vicuna, Mistral, and many others. * Easy UI for training LoRAs, as well as loading/unloading them on the fly. @@ -30,6 +30,8 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. To launch the web UI again in the future, run the same `start_` script that you used to install it. +Command-line flags can be passed to the `start_` script. Alternatively, you can open the file `CMD_FLAGS.txt` with a text editor and add your flags there. +
Setup details and information about installing manually @@ -201,8 +203,6 @@ pip install -r --upgrade ```
-Command-line flags can be passed to the `start_` script. Alternatively, you can open the file `CMD_FLAGS.txt` with a text editor and add your flags there. -
Command-line flags list @@ -382,14 +382,14 @@ https://github.com/oobabooga/text-generation-webui/wiki ## Downloading models -Models should be placed in the `text-generation-webui/models` folder. They are usually downloaded from [Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads). +Models should be placed in the folder `text-generation-webui/models`. They are usually downloaded from [Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads). * GGUF models are a single file and should be placed directly into `models`. Example: ``` text-generation-webui -├── models -│   ├── llama-2-13b-chat.Q4_K_M.gguf +└── models + └── llama-2-13b-chat.Q4_K_M.gguf ``` * Other models (like 16-bit transformers models and GPTQ models) are made of several files and must be placed in a subfolder. Example: @@ -413,7 +413,13 @@ text-generation-webui │   │   └── tokenizer.model ``` -In both cases, you can use the "Model" tab of the UI to download the model from Hugging Face automatically. It is also possible to download via the command-line with `python download-model.py organization/model` (use `--help` to see all the options). +In both cases, you can use the "Model" tab of the UI to download the model from Hugging Face automatically. It is also possible to download via the command-line with + +``` +python download-model.py organization/model +``` + +Use `--help` to see all the options. #### GPT-4chan From fade6abfe9dd955476510d4bf6db2ca4d6b9ecb9 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:18:40 -0800 Subject: [PATCH 08/20] Update README --- README.md | 32 +------------------------------- docs/08 - Additional Tips.md | 23 +++++++++++++++++++++++ 2 files changed, 24 insertions(+), 31 deletions(-) diff --git a/README.md b/README.md index c15d25fe..17273093 100644 --- a/README.md +++ b/README.md @@ -205,7 +205,7 @@ pip install -r --upgrade
-Command-line flags list +List of command-line flags @@ -421,36 +421,6 @@ python download-model.py organization/model Use `--help` to see all the options. -#### GPT-4chan - -
- -Instructions - - -[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options: - -* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model) -* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/) - -The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version. - -After downloading the model, follow these steps: - -1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`. -2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json). -3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan): - -``` -python download-model.py EleutherAI/gpt-j-6B --text-only -``` - -When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format: - -![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png) - -
- ## Google Colab notebook https://colab.research.google.com/github/oobabooga/text-generation-webui/blob/main/Colab-TextGen-GPU.ipynb diff --git a/docs/08 - Additional Tips.md b/docs/08 - Additional Tips.md index 89675cca..7ad00ee3 100644 --- a/docs/08 - Additional Tips.md +++ b/docs/08 - Additional Tips.md @@ -13,6 +13,29 @@ Source: https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/1126 This file will be automatically detected the next time you start the web UI. +## GPT-4chan + +[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options: + +* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model) +* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/) + +The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version. + +After downloading the model, follow these steps: + +1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`. +2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json). +3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan): + +``` +python download-model.py EleutherAI/gpt-j-6B --text-only +``` + +When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format: + +![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png) + ## Using LoRAs with GPTQ-for-LLaMa This requires using a monkey patch that is supported by this web UI: https://github.com/johnsmith0031/alpaca_lora_4bit From 0a4fad2d4627bcb3fedd730acab36410404d9eb4 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:20:37 -0800 Subject: [PATCH 09/20] Update README --- README.md | 10 ++-------- 1 file changed, 2 insertions(+), 8 deletions(-) diff --git a/README.md b/README.md index 17273093..205e59b0 100644 --- a/README.md +++ b/README.md @@ -32,6 +32,8 @@ To launch the web UI again in the future, run the same `start_` script that you Command-line flags can be passed to the `start_` script. Alternatively, you can open the file `CMD_FLAGS.txt` with a text editor and add your flags there. +To get updates in the future, run `update_linux.sh`, `update_windows.bat`, `update_macos.sh`, or `update_wsl.bat`. +
Setup details and information about installing manually @@ -45,18 +47,10 @@ The script creates a folder called `installer_files` where it sets up a Conda en To launch the webui in the future after it is already installed, run the same `start` script. -#### Getting updates - -Run `update_linux.sh`, `update_windows.bat`, `update_macos.sh`, or `update_wsl.bat`. - #### Running commands If you ever need to install something manually in the `installer_files` environment, you can launch an interactive shell using the cmd script: `cmd_linux.sh`, `cmd_windows.bat`, `cmd_macos.sh`, or `cmd_wsl.bat`. -#### Defining command-line flags - -To define persistent command-line flags like `--listen` or `--api`, edit the `CMD_FLAGS.txt` file with a text editor and add them there. Flags can also be provided directly to the start scripts, for instance, `./start-linux.sh --listen`. - #### Other info * There is no need to run any of those scripts as admin/root. From d354f5009c044a32137a8dde37f6f00be3207614 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:21:29 -0800 Subject: [PATCH 10/20] Update README --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 205e59b0..c2e8575b 100644 --- a/README.md +++ b/README.md @@ -413,7 +413,7 @@ In both cases, you can use the "Model" tab of the UI to download the model from python download-model.py organization/model ``` -Use `--help` to see all the options. +Run `python download-model.py --help` to see all the options. ## Google Colab notebook From 9695db0ee4022b8496cb2b919eae3b67be9f8aa0 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:30:31 -0800 Subject: [PATCH 11/20] Update README --- README.md | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index c2e8575b..63dfccd7 100644 --- a/README.md +++ b/README.md @@ -28,9 +28,11 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. 4) Once the installation ends, browse to `http://localhost:7860/?__theme=dark`. 5) Have fun! -To launch the web UI again in the future, run the same `start_` script that you used to install it. +To launch the web UI again in the future, run the `start_` script again. -Command-line flags can be passed to the `start_` script. Alternatively, you can open the file `CMD_FLAGS.txt` with a text editor and add your flags there. +This script sets up the web UI environment entirely in the `installer_files` folder -- it doesn't touch your system folders. To reinstall the project requirements, you can simply delete this folder and run the `start_` script again. + +That script accepts command-line flags. Alternatively, you can open the file `CMD_FLAGS.txt` with a text editor and add your flags there. To get updates in the future, run `update_linux.sh`, `update_windows.bat`, `update_macos.sh`, or `update_wsl.bat`. From d36e7f176252ddee43d1be0604f7f0d8197d7b1f Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:35:22 -0800 Subject: [PATCH 12/20] Update README --- README.md | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/README.md b/README.md index 63dfccd7..e3062813 100644 --- a/README.md +++ b/README.md @@ -28,11 +28,9 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. 4) Once the installation ends, browse to `http://localhost:7860/?__theme=dark`. 5) Have fun! -To launch the web UI again in the future, run the `start_` script again. +To restart the web UI later, just run the `start_` script once again. This script will set up the necessary files for the web UI in the `installer_files` folder. In case you need to reinstall the project's requirements, you can easily do so by deleting that folder and then running the `start_` script again. -This script sets up the web UI environment entirely in the `installer_files` folder -- it doesn't touch your system folders. To reinstall the project requirements, you can simply delete this folder and run the `start_` script again. - -That script accepts command-line flags. Alternatively, you can open the file `CMD_FLAGS.txt` with a text editor and add your flags there. +You also have the option of using command-line flags with the script. As an alternative, you can edit the `CMD_FLAGS.txt` file with a text editor and include your flags there. To get updates in the future, run `update_linux.sh`, `update_windows.bat`, `update_macos.sh`, or `update_wsl.bat`. From 7fef16950ff244bb9df48cfaf9635c51bdb40adf Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:42:54 -0800 Subject: [PATCH 13/20] Update README --- README.md | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index e3062813..d144d6df 100644 --- a/README.md +++ b/README.md @@ -428,6 +428,8 @@ If you would like to contribute to the project, check out the [Contributing guid * Subreddit: https://www.reddit.com/r/oobabooga/ * Discord: https://discord.gg/jwZCF2dPQN -## Acknowledgment +## Acknowledgment & support In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition, which will allow me to dedicate more time towards realizing the full potential of text-generation-webui. + +If you find this project useful, I have a [Ko-fi](https://ko-fi.com/oobabooga) where you can make a donation. Your support helps me continue maintaining and improving this project. From 783947a2aaef4fc6a6a48de8e8b182d673fdc22a Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:44:25 -0800 Subject: [PATCH 14/20] Update README --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index d144d6df..673c2903 100644 --- a/README.md +++ b/README.md @@ -432,4 +432,4 @@ If you would like to contribute to the project, check out the [Contributing guid In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition, which will allow me to dedicate more time towards realizing the full potential of text-generation-webui. -If you find this project useful, I have a [Ko-fi](https://ko-fi.com/oobabooga) where you can make a donation. Your support helps me continue maintaining and improving this project. +If you find this project useful, I have a [Ko-fi page](https://ko-fi.com/oobabooga) where you can make a donation. Your support helps me continue maintaining and improving this project. From aad14174e4a95e19f6a785aba75fcc1d5d944cd3 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:46:18 -0800 Subject: [PATCH 15/20] Update README --- README.md | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/README.md b/README.md index 673c2903..31b82249 100644 --- a/README.md +++ b/README.md @@ -43,9 +43,7 @@ Setup details and information about installing manually #### How it works -The script creates a folder called `installer_files` where it sets up a Conda environment using Miniconda. The installation is self-contained: if you want to reinstall, just delete `installer_files` and run the start script again. - -To launch the webui in the future after it is already installed, run the same `start` script. +The script creates a folder called `installer_files` where it sets up a Conda environment using Miniconda. #### Running commands From 03babe7d81833f0874ae88ee267267bcd0b07189 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:47:08 -0800 Subject: [PATCH 16/20] Update README --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 31b82249..1741b5be 100644 --- a/README.md +++ b/README.md @@ -428,6 +428,6 @@ If you would like to contribute to the project, check out the [Contributing guid ## Acknowledgment & support -In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition, which will allow me to dedicate more time towards realizing the full potential of text-generation-webui. +In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition. If you find this project useful, I have a [Ko-fi page](https://ko-fi.com/oobabooga) where you can make a donation. Your support helps me continue maintaining and improving this project. From e7fa17740ab482c7eb3e2a6bf615a385103c939d Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 13 Dec 2023 22:49:42 -0800 Subject: [PATCH 17/20] Update README --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 1741b5be..9a54fd1c 100644 --- a/README.md +++ b/README.md @@ -28,7 +28,7 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. 4) Once the installation ends, browse to `http://localhost:7860/?__theme=dark`. 5) Have fun! -To restart the web UI later, just run the `start_` script once again. This script will set up the necessary files for the web UI in the `installer_files` folder. In case you need to reinstall the project's requirements, you can easily do so by deleting that folder and then running the `start_` script again. +To restart the web UI later, just run the `start_` script once again. This script will set up the project's requirements in the `installer_files` folder. In case you need to reinstall the requirements, you can easily do so by deleting that folder and then running the `start_` script again. You also have the option of using command-line flags with the script. As an alternative, you can edit the `CMD_FLAGS.txt` file with a text editor and include your flags there. From 958799221fd7dc8bef2b7d79cdb0efa9f7ba8824 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Thu, 14 Dec 2023 06:09:03 -0800 Subject: [PATCH 18/20] Update README --- README.md | 30 ++++++++++++------------------ 1 file changed, 12 insertions(+), 18 deletions(-) diff --git a/README.md b/README.md index 9a54fd1c..5df865d5 100644 --- a/README.md +++ b/README.md @@ -15,9 +15,9 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. * Dropdown menu for quickly switching between different models. * Large number of extensions (built-in and user-contributed), including Coqui TTS for realistic voice outputs, Whisper STT for voice inputs, translation, [multimodal pipelines](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal), vector databases, Stable Diffusion integration, and a lot more. See [the wiki](https://github.com/oobabooga/text-generation-webui/wiki/07-%E2%80%90-Extensions) and [the extensions directory](https://github.com/oobabooga/text-generation-webui-extensions) for details. * [Chat with custom characters](https://github.com/oobabooga/text-generation-webui/wiki/03-%E2%80%90-Parameters-Tab#character). -* Precise templates for instruction-following models, including Llama-2-chat, Alpaca, Vicuna, Mistral, and many others. -* Easy UI for training LoRAs, as well as loading/unloading them on the fly. -* HF transformers integration: load models in 4-bit or 8-bit quantization through bitsandbytes, use llama.cpp with transformers samplers (`llamacpp_HF` loader), CPU inference in 32-bit precision using PyTorch. +* Precise chat templates for instruction-following models, including Llama-2-chat, Alpaca, Vicuna, Mistral, and many others. +* LoRA: train new LoRAs with your own data, load/unload LoRAs on the fly for generation. +* Transformers library integration: load models in 4-bit or 8-bit precision through bitsandbytes, use llama.cpp with transformers samplers (`llamacpp_HF` loader), CPU inference in 32-bit precision using PyTorch. * OpenAI-compatible API server with Chat and Completions endpoints -- see the [examples](https://github.com/oobabooga/text-generation-webui/wiki/12-%E2%80%90-OpenAI-API#examples). ## How to install @@ -28,7 +28,7 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github. 4) Once the installation ends, browse to `http://localhost:7860/?__theme=dark`. 5) Have fun! -To restart the web UI later, just run the `start_` script once again. This script will set up the project's requirements in the `installer_files` folder. In case you need to reinstall the requirements, you can easily do so by deleting that folder and then running the `start_` script again. +To restart the web UI in the future, just run the `start_` script again. This script creates a `installer_files` folder where it sets up the project's requirements. In case you need to reinstall the requirements, you can easily do so by deleting that folder and then running the `start_` script again. You also have the option of using command-line flags with the script. As an alternative, you can edit the `CMD_FLAGS.txt` file with a text editor and include your flags there. @@ -41,19 +41,12 @@ Setup details and information about installing manually ### One-click-installer -#### How it works - -The script creates a folder called `installer_files` where it sets up a Conda environment using Miniconda. - -#### Running commands +The script uses Miniconda to set up a Conda environment in the `installer_files` folder. If you ever need to install something manually in the `installer_files` environment, you can launch an interactive shell using the cmd script: `cmd_linux.sh`, `cmd_windows.bat`, `cmd_macos.sh`, or `cmd_wsl.bat`. -#### Other info - * There is no need to run any of those scripts as admin/root. * For additional instructions about AMD setup and WSL setup, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). -* The installer has been tested mostly on NVIDIA GPUs. If you can find a way to improve it for your AMD/Intel Arc/Mac Metal GPU, you are highly encouraged to submit a PR to this repository. The main file to be edited is `one_click.py`. * For automated installation, you can use the `GPU_CHOICE`, `USE_CUDA118`, `LAUNCH_AFTER_INSTALL`, and `INSTALL_EXTENSIONS` environment variables. For instance: `GPU_CHOICE=A USE_CUDA118=FALSE LAUNCH_AFTER_INSTALL=FALSE INSTALL_EXTENSIONS=FALSE ./start_linux.sh`. ### Manual installation using Conda @@ -126,9 +119,11 @@ Requirements file to use: ### Start the web UI - conda activate textgen - cd text-generation-webui - python server.py +``` +conda activate textgen +cd text-generation-webui +python server.py +``` Then browse to @@ -200,7 +195,6 @@ pip install -r --upgrade List of command-line flags - #### Basic settings | Flag | Description | @@ -405,7 +399,7 @@ text-generation-webui │   │   └── tokenizer.model ``` -In both cases, you can use the "Model" tab of the UI to download the model from Hugging Face automatically. It is also possible to download via the command-line with +In both cases, you can use the "Model" tab of the UI to download the model from Hugging Face automatically. It is also possible to download it via the command-line with ``` python download-model.py organization/model @@ -428,6 +422,6 @@ If you would like to contribute to the project, check out the [Contributing guid ## Acknowledgment & support -In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition. +In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project ([link](https://a16z.com/supporting-the-open-source-ai-community/)). I am **extremely** grateful for their trust and recognition. If you find this project useful, I have a [Ko-fi page](https://ko-fi.com/oobabooga) where you can make a donation. Your support helps me continue maintaining and improving this project. From 5b283fff2242d29752bbbecab2ac5146041d2863 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Thu, 14 Dec 2023 06:15:14 -0800 Subject: [PATCH 19/20] Update README --- README.md | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/README.md b/README.md index 5df865d5..8907c4d9 100644 --- a/README.md +++ b/README.md @@ -45,8 +45,8 @@ The script uses Miniconda to set up a Conda environment in the `installer_files` If you ever need to install something manually in the `installer_files` environment, you can launch an interactive shell using the cmd script: `cmd_linux.sh`, `cmd_windows.bat`, `cmd_macos.sh`, or `cmd_wsl.bat`. -* There is no need to run any of those scripts as admin/root. -* For additional instructions about AMD setup and WSL setup, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). +* There is no need to run any of those scripts (`start_`, `update_`, or `cmd_`) as admin/root. +* For additional instructions about AMD and WSL setup, consult [the documentation](https://github.com/oobabooga/text-generation-webui/wiki). * For automated installation, you can use the `GPU_CHOICE`, `USE_CUDA118`, `LAUNCH_AFTER_INSTALL`, and `INSTALL_EXTENSIONS` environment variables. For instance: `GPU_CHOICE=A USE_CUDA118=FALSE LAUNCH_AFTER_INSTALL=FALSE INSTALL_EXTENSIONS=FALSE ./start_linux.sh`. ### Manual installation using Conda @@ -162,7 +162,7 @@ conda install -y -c "nvidia/label/cuda-11.8.0" cuda-runtime ##### Manual install -The requirements*.txt above contain various precompiled wheels. If you wish to compile things manually, or if you need to because no suitable wheels are available for your hardware, you can use `requirements_nowheels.txt` and then install your desired loaders manually. +The `requirements*.txt` above contain various wheels precompiled through GitHub Actions. If you wish to compile things manually, or if you need to because no suitable wheels are available for your hardware, you can use `requirements_nowheels.txt` and then install your desired loaders manually. ### Alternative: Docker @@ -181,7 +181,7 @@ docker compose up --build ### Updating the requirements -From time to time, the `requirements*.txt` changes. To update, use these commands: +From time to time, the `requirements*.txt` change. To update, use these commands: ``` conda activate textgen @@ -422,6 +422,6 @@ If you would like to contribute to the project, check out the [Contributing guid ## Acknowledgment & support -In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project ([link](https://a16z.com/supporting-the-open-source-ai-community/)). I am **extremely** grateful for their trust and recognition. +In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition. If you find this project useful, I have a [Ko-fi page](https://ko-fi.com/oobabooga) where you can make a donation. Your support helps me continue maintaining and improving this project. From d5ec3c34440e8899a6d1cd2db1103668db6045ab Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Thu, 14 Dec 2023 06:20:52 -0800 Subject: [PATCH 20/20] Update README --- README.md | 2 -- 1 file changed, 2 deletions(-) diff --git a/README.md b/README.md index 8907c4d9..4dee47af 100644 --- a/README.md +++ b/README.md @@ -358,8 +358,6 @@ List of command-line flags |---------------------------------------|-------------| | `--multimodal-pipeline PIPELINE` | The multimodal pipeline to use. Examples: `llava-7b`, `llava-13b`. | - -
## Documentation