mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-24 17:06:53 +01:00
Update README
This commit is contained in:
parent
aafd15109d
commit
fade6abfe9
32
README.md
32
README.md
@ -205,7 +205,7 @@ pip install -r <requirements file that you have used> --upgrade
|
|||||||
|
|
||||||
<details>
|
<details>
|
||||||
<summary>
|
<summary>
|
||||||
Command-line flags list
|
List of command-line flags
|
||||||
</summary>
|
</summary>
|
||||||
|
|
||||||
|
|
||||||
@ -421,36 +421,6 @@ python download-model.py organization/model
|
|||||||
|
|
||||||
Use `--help` to see all the options.
|
Use `--help` to see all the options.
|
||||||
|
|
||||||
#### GPT-4chan
|
|
||||||
|
|
||||||
<details>
|
|
||||||
<summary>
|
|
||||||
Instructions
|
|
||||||
</summary>
|
|
||||||
|
|
||||||
[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options:
|
|
||||||
|
|
||||||
* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model)
|
|
||||||
* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/)
|
|
||||||
|
|
||||||
The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version.
|
|
||||||
|
|
||||||
After downloading the model, follow these steps:
|
|
||||||
|
|
||||||
1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`.
|
|
||||||
2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json).
|
|
||||||
3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan):
|
|
||||||
|
|
||||||
```
|
|
||||||
python download-model.py EleutherAI/gpt-j-6B --text-only
|
|
||||||
```
|
|
||||||
|
|
||||||
When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format:
|
|
||||||
|
|
||||||
![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png)
|
|
||||||
|
|
||||||
</details>
|
|
||||||
|
|
||||||
## Google Colab notebook
|
## Google Colab notebook
|
||||||
|
|
||||||
https://colab.research.google.com/github/oobabooga/text-generation-webui/blob/main/Colab-TextGen-GPU.ipynb
|
https://colab.research.google.com/github/oobabooga/text-generation-webui/blob/main/Colab-TextGen-GPU.ipynb
|
||||||
|
@ -13,6 +13,29 @@ Source: https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/1126
|
|||||||
|
|
||||||
This file will be automatically detected the next time you start the web UI.
|
This file will be automatically detected the next time you start the web UI.
|
||||||
|
|
||||||
|
## GPT-4chan
|
||||||
|
|
||||||
|
[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options:
|
||||||
|
|
||||||
|
* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model)
|
||||||
|
* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/)
|
||||||
|
|
||||||
|
The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version.
|
||||||
|
|
||||||
|
After downloading the model, follow these steps:
|
||||||
|
|
||||||
|
1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`.
|
||||||
|
2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json).
|
||||||
|
3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan):
|
||||||
|
|
||||||
|
```
|
||||||
|
python download-model.py EleutherAI/gpt-j-6B --text-only
|
||||||
|
```
|
||||||
|
|
||||||
|
When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format:
|
||||||
|
|
||||||
|
![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png)
|
||||||
|
|
||||||
## Using LoRAs with GPTQ-for-LLaMa
|
## Using LoRAs with GPTQ-for-LLaMa
|
||||||
|
|
||||||
This requires using a monkey patch that is supported by this web UI: https://github.com/johnsmith0031/alpaca_lora_4bit
|
This requires using a monkey patch that is supported by this web UI: https://github.com/johnsmith0031/alpaca_lora_4bit
|
||||||
|
Loading…
Reference in New Issue
Block a user