mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-25 17:29:22 +01:00
179 lines
6.2 KiB
Markdown
179 lines
6.2 KiB
Markdown
## Audio notification
|
|
|
|
If your computer takes a long time to generate each response for the model that you are using, you can enable an audio notification for when the response is completed. This feature was kindly contributed by HappyWorldGames in [#1277](https://github.com/oobabooga/text-generation-webui/pull/1277).
|
|
|
|
### Installation
|
|
|
|
Simply place a file called "notification.mp3" in the same folder as `server.py`. Here you can find some examples:
|
|
|
|
* https://pixabay.com/sound-effects/search/ding/?duration=0-30
|
|
* https://pixabay.com/sound-effects/search/notification/?duration=0-30
|
|
|
|
Source: https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/1126
|
|
|
|
This file will be automatically detected the next time you start the web UI.
|
|
|
|
## GPT-4chan
|
|
|
|
[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options:
|
|
|
|
* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model)
|
|
* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/)
|
|
|
|
The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version.
|
|
|
|
After downloading the model, follow these steps:
|
|
|
|
1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`.
|
|
2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json).
|
|
3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan):
|
|
|
|
```
|
|
python download-model.py EleutherAI/gpt-j-6B --text-only
|
|
```
|
|
|
|
When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format:
|
|
|
|
![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png)
|
|
|
|
## Using LoRAs with GPTQ-for-LLaMa
|
|
|
|
This requires using a monkey patch that is supported by this web UI: https://github.com/johnsmith0031/alpaca_lora_4bit
|
|
|
|
To use it:
|
|
|
|
Install alpaca_lora_4bit using pip
|
|
|
|
```
|
|
git clone https://github.com/johnsmith0031/alpaca_lora_4bit.git
|
|
cd alpaca_lora_4bit
|
|
git fetch origin winglian-setup_pip
|
|
git checkout winglian-setup_pip
|
|
pip install .
|
|
```
|
|
|
|
Start the UI with the --monkey-patch flag:
|
|
|
|
```
|
|
python server.py --model llama-7b-4bit-128g --listen --lora tloen_alpaca-lora-7b --monkey-patch
|
|
```
|
|
|
|
## DeepSpeed
|
|
|
|
`DeepSpeed ZeRO-3` is an alternative offloading strategy for full-precision (16-bit) transformers models.
|
|
|
|
With this, I have been able to load a 6b model (GPT-J 6B) with less than 6GB of VRAM. The speed of text generation is very decent and much better than what would be accomplished with `--auto-devices --gpu-memory 6`.
|
|
|
|
As far as I know, DeepSpeed is only available for Linux at the moment.
|
|
|
|
### How to use it
|
|
|
|
1. Install DeepSpeed:
|
|
|
|
```
|
|
conda install -c conda-forge mpi4py mpich
|
|
pip install -U deepspeed
|
|
```
|
|
|
|
2. Start the web UI replacing `python` with `deepspeed --num_gpus=1` and adding the `--deepspeed` flag. Example:
|
|
|
|
```
|
|
deepspeed --num_gpus=1 server.py --deepspeed --chat --model gpt-j-6B
|
|
```
|
|
|
|
> RWKV: RNN with Transformer-level LLM Performance
|
|
>
|
|
> It combines the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding (using the final hidden state).
|
|
|
|
https://github.com/BlinkDL/RWKV-LM
|
|
|
|
https://github.com/BlinkDL/ChatRWKV
|
|
|
|
## Using RWKV in the web UI
|
|
|
|
### Hugging Face weights
|
|
|
|
Simply download the weights from https://huggingface.co/RWKV and load them as you would for any other model.
|
|
|
|
There is a bug in transformers==4.29.2 that prevents RWKV from being loaded in 8-bit mode. You can install the dev branch to solve this bug: `pip install git+https://github.com/huggingface/transformers`
|
|
|
|
### Original .pth weights
|
|
|
|
The instructions below are from before RWKV was supported in transformers, and they are kept for legacy purposes. The old implementation is possibly faster, but it lacks the full range of samplers that the transformers library offers.
|
|
|
|
#### 0. Install the RWKV library
|
|
|
|
```
|
|
pip install rwkv
|
|
```
|
|
|
|
`0.7.3` was the last version that I tested. If you experience any issues, try ```pip install rwkv==0.7.3```.
|
|
|
|
#### 1. Download the model
|
|
|
|
It is available in different sizes:
|
|
|
|
* https://huggingface.co/BlinkDL/rwkv-4-pile-3b/
|
|
* https://huggingface.co/BlinkDL/rwkv-4-pile-7b/
|
|
* https://huggingface.co/BlinkDL/rwkv-4-pile-14b/
|
|
|
|
There are also older releases with smaller sizes like:
|
|
|
|
* https://huggingface.co/BlinkDL/rwkv-4-pile-169m/resolve/main/RWKV-4-Pile-169M-20220807-8023.pth
|
|
|
|
Download the chosen `.pth` and put it directly in the `models` folder.
|
|
|
|
#### 2. Download the tokenizer
|
|
|
|
[20B_tokenizer.json](https://raw.githubusercontent.com/BlinkDL/ChatRWKV/main/v2/20B_tokenizer.json)
|
|
|
|
Also put it directly in the `models` folder. Make sure to not rename it. It should be called `20B_tokenizer.json`.
|
|
|
|
#### 3. Launch the web UI
|
|
|
|
No additional steps are required. Just launch it as you would with any other model.
|
|
|
|
```
|
|
python server.py --listen --no-stream --model RWKV-4-Pile-169M-20220807-8023.pth
|
|
```
|
|
|
|
#### Setting a custom strategy
|
|
|
|
It is possible to have very fine control over the offloading and precision for the model with the `--rwkv-strategy` flag. Possible values include:
|
|
|
|
```
|
|
"cpu fp32" # CPU mode
|
|
"cuda fp16" # GPU mode with float16 precision
|
|
"cuda fp16 *30 -> cpu fp32" # GPU+CPU offloading. The higher the number after *, the higher the GPU allocation.
|
|
"cuda fp16i8" # GPU mode with 8-bit precision
|
|
```
|
|
|
|
See the README for the PyPl package for more details: https://pypi.org/project/rwkv/
|
|
|
|
#### Compiling the CUDA kernel
|
|
|
|
You can compile the CUDA kernel for the model with `--rwkv-cuda-on`. This should improve the performance a lot but I haven't been able to get it to work yet.
|
|
|
|
## Miscellaneous info
|
|
|
|
### You can train LoRAs in CPU mode
|
|
|
|
Load the web UI with
|
|
|
|
```
|
|
python server.py --cpu
|
|
```
|
|
|
|
and start training the LoRA from the training tab as usual.
|
|
|
|
### You can check the sha256sum of downloaded models with the download script
|
|
|
|
```
|
|
python download-model.py facebook/galactica-125m --check
|
|
```
|
|
|
|
### The download script continues interrupted downloads by default
|
|
|
|
It doesn't start over.
|
|
|