mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-28 18:48:04 +01:00
Update GPTQ-models-(4-bit-mode).md
This commit is contained in:
parent
486ddd62df
commit
419c34eca4
@ -86,7 +86,7 @@ conda install -c conda-forge cudatoolkit-dev
|
|||||||
|
|
||||||
The command above takes some 10 minutes to run and shows no progress bar or updates along the way.
|
The command above takes some 10 minutes to run and shows no progress bar or updates along the way.
|
||||||
|
|
||||||
You are going to need to have a C++ compiler installed. On Linux, `sudo apt install build-essential` or equivalent is enough.
|
You are also going to need to have a C++ compiler installed. On Linux, `sudo apt install build-essential` or equivalent is enough.
|
||||||
|
|
||||||
If you're using an older version of CUDA toolkit (e.g. 11.7) but the latest version of `gcc` and `g++` (12.0+), you should downgrade with: `conda install -c conda-forge gxx==11.3.0`. Kernel compilation will fail otherwise.
|
If you're using an older version of CUDA toolkit (e.g. 11.7) but the latest version of `gcc` and `g++` (12.0+), you should downgrade with: `conda install -c conda-forge gxx==11.3.0`. Kernel compilation will fail otherwise.
|
||||||
|
|
||||||
@ -97,7 +97,9 @@ cd repositories/GPTQ-for-LLaMa
|
|||||||
python setup_cuda.py install
|
python setup_cuda.py install
|
||||||
```
|
```
|
||||||
|
|
||||||
### Pre-converted weights
|
### Getting pre-converted LLaMA weights
|
||||||
|
|
||||||
|
These are models that you can simply download and place in your `models` folder.
|
||||||
|
|
||||||
* Converted without `group-size` (better for the 7b model): https://github.com/oobabooga/text-generation-webui/pull/530#issuecomment-1483891617
|
* Converted without `group-size` (better for the 7b model): https://github.com/oobabooga/text-generation-webui/pull/530#issuecomment-1483891617
|
||||||
* Converted with `group-size` (better from 13b upwards): https://github.com/oobabooga/text-generation-webui/pull/530#issuecomment-1483941105
|
* Converted with `group-size` (better from 13b upwards): https://github.com/oobabooga/text-generation-webui/pull/530#issuecomment-1483941105
|
||||||
|
Loading…
Reference in New Issue
Block a user