From 9508f207ba9f30be67a3832843490ef1c87db2fe Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sat, 22 Apr 2023 02:53:01 -0300 Subject: [PATCH] Update Using-LoRAs.md --- docs/Using-LoRAs.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/Using-LoRAs.md b/docs/Using-LoRAs.md index 8049e96b..159345ef 100644 --- a/docs/Using-LoRAs.md +++ b/docs/Using-LoRAs.md @@ -16,7 +16,7 @@ python server.py --model llama-7b-hf --lora alpaca-lora-7b --load-in-8bit python server.py --model llama-7b-hf --lora alpaca-lora-7b --cpu ``` -* For using LoRAs in 4-bit mode, follow these special instructions: https://github.com/oobabooga/text-generation-webui/wiki/GPTQ-models-(4-bit-mode)#using-loras-in-4-bit-mode +* For using LoRAs in 4-bit mode, follow [these special instructions](GPTQ-models-(4-bit-mode).md#using-loras-in-4-bit-mode). * Instead of using the `--lora` command-line flag, you can also select the LoRA in the "Parameters" tab of the interface. @@ -85,4 +85,4 @@ Run the script with: python finetune.py ``` -It just works. It runs at 22.32s/it, with 1170 iterations in total, so about 7 hours and a half for training a LoRA. RTX 3090, 18153MiB VRAM used, drawing maximum power (350W, room heater mode). \ No newline at end of file +It just works. It runs at 22.32s/it, with 1170 iterations in total, so about 7 hours and a half for training a LoRA. RTX 3090, 18153MiB VRAM used, drawing maximum power (350W, room heater mode).