From f5c36cca4076b0b3762ebc7e8c749305dc66db2c Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sat, 22 Apr 2023 02:49:54 -0300 Subject: [PATCH] Update LLaMA-model.md --- docs/LLaMA-model.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/LLaMA-model.md b/docs/LLaMA-model.md index 95e0d35e..299076b2 100644 --- a/docs/LLaMA-model.md +++ b/docs/LLaMA-model.md @@ -3,7 +3,7 @@ LLaMA is a Large Language Model developed by Meta AI. It was trained on more tokens than previous models. The result is that the smallest version with 7 billion parameters has similar performance to GPT-3 with 175 billion parameters. This guide will cover usage through the official `transformers` implementation. For 4-bit mode, head over to [GPTQ models (4 bit mode) -](https://github.com/oobabooga/text-generation-webui/wiki/GPTQ-models-(4-bit-mode)). +](GPTQ-models-(4-bit-mode).md). ## Getting the weights @@ -42,4 +42,4 @@ python convert_llama_weights_to_hf.py --input_dir /path/to/LLaMA --model_size 7B ```python python server.py --model llama-7b -``` \ No newline at end of file +```