mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 08:07:56 +01:00
Update Using-LoRAs.md
This commit is contained in:
parent
6d4f131d0a
commit
9508f207ba
@ -16,7 +16,7 @@ python server.py --model llama-7b-hf --lora alpaca-lora-7b --load-in-8bit
|
|||||||
python server.py --model llama-7b-hf --lora alpaca-lora-7b --cpu
|
python server.py --model llama-7b-hf --lora alpaca-lora-7b --cpu
|
||||||
```
|
```
|
||||||
|
|
||||||
* For using LoRAs in 4-bit mode, follow these special instructions: https://github.com/oobabooga/text-generation-webui/wiki/GPTQ-models-(4-bit-mode)#using-loras-in-4-bit-mode
|
* For using LoRAs in 4-bit mode, follow [these special instructions](GPTQ-models-(4-bit-mode).md#using-loras-in-4-bit-mode).
|
||||||
|
|
||||||
* Instead of using the `--lora` command-line flag, you can also select the LoRA in the "Parameters" tab of the interface.
|
* Instead of using the `--lora` command-line flag, you can also select the LoRA in the "Parameters" tab of the interface.
|
||||||
|
|
||||||
@ -85,4 +85,4 @@ Run the script with:
|
|||||||
python finetune.py
|
python finetune.py
|
||||||
```
|
```
|
||||||
|
|
||||||
It just works. It runs at 22.32s/it, with 1170 iterations in total, so about 7 hours and a half for training a LoRA. RTX 3090, 18153MiB VRAM used, drawing maximum power (350W, room heater mode).
|
It just works. It runs at 22.32s/it, with 1170 iterations in total, so about 7 hours and a half for training a LoRA. RTX 3090, 18153MiB VRAM used, drawing maximum power (350W, room heater mode).
|
||||||
|
Loading…
Reference in New Issue
Block a user