mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-12-26 06:10:39 +01:00
108 lines
3.2 KiB
Markdown
108 lines
3.2 KiB
Markdown
You have now entered a hidden corner of the internet.
|
|
|
|
A confusing yet intriguing realm of paradoxes and contradictions.
|
|
|
|
A place where you will find out that what you thought you knew, you in fact didn't know, and what you didn't know was in front of you all along.
|
|
|
|
![](https://i.pinimg.com/originals/6e/e2/7b/6ee27bad351d3aca470d80f1033ba9c6.jpg)
|
|
|
|
*In other words, here I will document little-known facts about this web UI that I could not find another place for in the wiki.*
|
|
|
|
#### You can train LoRAs in CPU mode
|
|
|
|
Load the web UI with
|
|
|
|
```
|
|
python server.py --cpu
|
|
```
|
|
|
|
and start training the LoRA from the training tab as usual.
|
|
|
|
#### 8-bit mode works with CPU offloading
|
|
|
|
```
|
|
python server.py --load-in-8bit --gpu-memory 4000MiB
|
|
```
|
|
|
|
#### `--pre_layer`, and not `--gpu-memory`, is the right way to do CPU offloading with 4-bit models
|
|
|
|
```
|
|
python server.py --wbits 4 --groupsize 128 --pre_layer 20
|
|
```
|
|
|
|
#### Models can be loaded in 32-bit, 16-bit, 8-bit, and 4-bit modes
|
|
|
|
```
|
|
python server.py --cpu
|
|
python server.py
|
|
python server.py --load-in-8bit
|
|
python server.py --wbits 4
|
|
```
|
|
|
|
#### The web UI works with any version of GPTQ-for-LLaMa
|
|
|
|
Including the up to date triton and cuda branches. But you have to delete the `repositories/GPTQ-for-LLaMa` folder and reinstall the new one every time:
|
|
|
|
```
|
|
cd text-generation-webui/repositories
|
|
rm -r GPTQ-for-LLaMa
|
|
pip uninstall quant-cuda
|
|
git clone https://github.com/oobabooga/GPTQ-for-LLaMa -b cuda # or any other repository and branch
|
|
cd GPTQ-for-LLaMa
|
|
python setup_cuda.py install
|
|
```
|
|
|
|
#### Instruction-following templates are represented as chat characters
|
|
|
|
https://github.com/oobabooga/text-generation-webui/tree/main/characters/instruction-following
|
|
|
|
#### The right way to run Alpaca, Open Assistant, Vicuna, etc is Instruct mode, not normal chat mode
|
|
|
|
Otherwise the prompt will not be formatted correctly.
|
|
|
|
1. Start the web UI with
|
|
|
|
```
|
|
python server.py --chat
|
|
```
|
|
|
|
2. Click on the "instruct" option under "Chat modes"
|
|
|
|
3. Select the correct template in the hidden dropdown menu that will become visible.
|
|
|
|
#### Notebook mode is best mode
|
|
|
|
Ascended individuals have realized that notebook mode is the superset of chat mode and can do chats with ultimate flexibility, including group chats, editing replies, starting a new bot reply in a given way, and impersonating.
|
|
|
|
#### RWKV is a RNN
|
|
|
|
Most models are transformers, but not RWKV, which is a RNN. It's a great model.
|
|
|
|
#### `--gpu-memory` is not a hard limit on the GPU memory
|
|
|
|
It is simply a parameter that is passed to the `accelerate` library while loading the model. More memory will be allocated during generation. That's why this parameter has to be set to less than your total GPU memory.
|
|
|
|
#### Contrastive search perhaps the best preset
|
|
|
|
But it uses a ton of VRAM.
|
|
|
|
#### You can check the sha256sum of downloaded models with the download script
|
|
|
|
```
|
|
python download-model.py facebook/galactica-125m --check
|
|
```
|
|
|
|
#### The download script continues interrupted downloads by default
|
|
|
|
It doesn't start over.
|
|
|
|
#### You can download models with multiple threads
|
|
|
|
```
|
|
python download-model.py facebook/galactica-125m --threads 8
|
|
```
|
|
|
|
#### LoRAs work in 4-bit mode
|
|
|
|
You need to follow [these instructions](GPTQ-models-(4-bit-mode).md#using-loras-in-4-bit-mode) and then start the web UI with the `--monkey-patch` flag.
|