diff --git a/docs/llama.cpp.md b/docs/llama.cpp.md index 07d3a1d8..68aa1cfa 100644 --- a/docs/llama.cpp.md +++ b/docs/llama.cpp.md @@ -24,7 +24,7 @@ Enabled with the `--n-gpu-layers` parameter. * If you have enough VRAM, use a high number like `--n-gpu-layers 1000` to offload all layers to the GPU. * Otherwise, start with a low number like `--n-gpu-layers 10` and then gradually increase it until you run out of memory. -This feature works out of the box for NVIDIA GPUs. For other GPUs, you need to uninstall `llama-cpp-python` with +This feature works out of the box for NVIDIA GPUs on Linux (amd64) or Windows. For other GPUs, you need to uninstall `llama-cpp-python` with ``` pip uninstall -y llama-cpp-python