Update llama.cpp.md instructions (#3702)

This commit is contained in:
q5sys (JT) 2023-08-29 16:56:50 -04:00 committed by GitHub
parent a9a1784420
commit cdb854db9e
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -9,7 +9,10 @@ llama.cpp is the best backend in two important scenarios:
#### Pre-converted #### Pre-converted
Download the ggml model directly into your `text-generation-webui/models` folder, making sure that its name contains `ggml` somewhere and ends in `.bin`. It's a single file. Download the GGUF or GGML models directly into your `text-generation-webui/models` folder. It will be a single file.
* For GGUF models, make sure its name contains `.gguf`.
* For GGML models, make sure its name contains `ggml` and ends in `.bin`.
`q4_K_M` quantization is recommended. `q4_K_M` quantization is recommended.