mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 16:17:57 +01:00
Update llama.cpp-models.md
This commit is contained in:
parent
26cf8c2545
commit
cd9be4c2ba
@ -16,11 +16,22 @@ Enabled with the `--n-gpu-layers` parameter. If you have enough VRAM, use a high
|
|||||||
|
|
||||||
Note that you need to manually install `llama-cpp-python` with GPU support. To do that:
|
Note that you need to manually install `llama-cpp-python` with GPU support. To do that:
|
||||||
|
|
||||||
|
#### Linux
|
||||||
|
|
||||||
```
|
```
|
||||||
pip uninstall -y llama-cpp-python
|
pip uninstall -y llama-cpp-python
|
||||||
CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python --no-cache-dir
|
CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python --no-cache-dir
|
||||||
```
|
```
|
||||||
|
|
||||||
|
#### Windows
|
||||||
|
|
||||||
|
```
|
||||||
|
pip uninstall -y llama-cpp-python
|
||||||
|
set CMAKE_ARGS="-DLLAMA_CUBLAS=on"
|
||||||
|
set FORCE_CMAKE=1
|
||||||
|
pip install llama-cpp-python --no-cache-dir
|
||||||
|
```
|
||||||
|
|
||||||
Here you can find the different compilation options for OpenBLAS / cuBLAS / CLBlast: https://pypi.org/project/llama-cpp-python/
|
Here you can find the different compilation options for OpenBLAS / cuBLAS / CLBlast: https://pypi.org/project/llama-cpp-python/
|
||||||
|
|
||||||
## Performance
|
## Performance
|
||||||
|
Loading…
Reference in New Issue
Block a user