From d3abe7caa8916c2162d7e4b50fd2de05a026461e Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Tue, 25 Jul 2023 15:33:16 -0700 Subject: [PATCH] Update llama.cpp.md --- docs/llama.cpp.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/llama.cpp.md b/docs/llama.cpp.md index 07d3a1d8..68aa1cfa 100644 --- a/docs/llama.cpp.md +++ b/docs/llama.cpp.md @@ -24,7 +24,7 @@ Enabled with the `--n-gpu-layers` parameter. * If you have enough VRAM, use a high number like `--n-gpu-layers 1000` to offload all layers to the GPU. * Otherwise, start with a low number like `--n-gpu-layers 10` and then gradually increase it until you run out of memory. -This feature works out of the box for NVIDIA GPUs. For other GPUs, you need to uninstall `llama-cpp-python` with +This feature works out of the box for NVIDIA GPUs on Linux (amd64) or Windows. For other GPUs, you need to uninstall `llama-cpp-python` with ``` pip uninstall -y llama-cpp-python