From cb9be5db1c1658fca304ae5ea634ebb5f88412df Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Fri, 16 Jun 2023 20:40:12 -0300 Subject: [PATCH] Update ExLlama.md --- docs/ExLlama.md | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/docs/ExLlama.md b/docs/ExLlama.md index a0968927..a9fd016d 100644 --- a/docs/ExLlama.md +++ b/docs/ExLlama.md @@ -4,11 +4,12 @@ ExLlama is an extremely optimized GPTQ backend for LLaMA models. It features much lower VRAM usage and much higher speeds due to not relying on unoptimized transformers code. -# Installation: +## Installation: -1) Clone the ExLlama repository into your `repositories` folder: +1) Clone the ExLlama repository into your `text-generation-webui/repositories` folder: ``` +mkdir repositories cd repositories git clone https://github.com/turboderp/exllama ```