mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 08:07:56 +01:00
Update llama.cpp.md instructions (#3702)
This commit is contained in:
parent
a9a1784420
commit
cdb854db9e
@ -9,7 +9,10 @@ llama.cpp is the best backend in two important scenarios:
|
|||||||
|
|
||||||
#### Pre-converted
|
#### Pre-converted
|
||||||
|
|
||||||
Download the ggml model directly into your `text-generation-webui/models` folder, making sure that its name contains `ggml` somewhere and ends in `.bin`. It's a single file.
|
Download the GGUF or GGML models directly into your `text-generation-webui/models` folder. It will be a single file.
|
||||||
|
|
||||||
|
* For GGUF models, make sure its name contains `.gguf`.
|
||||||
|
* For GGML models, make sure its name contains `ggml` and ends in `.bin`.
|
||||||
|
|
||||||
`q4_K_M` quantization is recommended.
|
`q4_K_M` quantization is recommended.
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user