mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 16:17:57 +01:00
Revert changes
This commit is contained in:
parent
69f8b35bc9
commit
4a24849715
@ -1,20 +0,0 @@
|
|||||||
## GPT-4chan
|
|
||||||
|
|
||||||
[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options:
|
|
||||||
|
|
||||||
* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model)
|
|
||||||
* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/)
|
|
||||||
|
|
||||||
The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version.
|
|
||||||
|
|
||||||
After downloading the model, follow these steps:
|
|
||||||
|
|
||||||
1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`.
|
|
||||||
2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json).
|
|
||||||
3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan):
|
|
||||||
|
|
||||||
```
|
|
||||||
python download-model.py EleutherAI/gpt-j-6B --text-only
|
|
||||||
```
|
|
||||||
|
|
||||||
When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format.
|
|
@ -9,10 +9,9 @@
|
|||||||
* [ExLlama](ExLlama.md)
|
* [ExLlama](ExLlama.md)
|
||||||
* [Extensions](Extensions.md)
|
* [Extensions](Extensions.md)
|
||||||
* [Generation parameters](Generation-parameters.md)
|
* [Generation parameters](Generation-parameters.md)
|
||||||
* [GGML (llama.cpp) models](GGML-llama.cpp-models.md)
|
|
||||||
* [GPT-4chan model](GPT-4chan-model.md)
|
|
||||||
* [GPTQ models (4 bit mode)](GPTQ-models-(4-bit-mode).md)
|
* [GPTQ models (4 bit mode)](GPTQ-models-(4-bit-mode).md)
|
||||||
* [LLaMA model](LLaMA-model.md)
|
* [LLaMA model](LLaMA-model.md)
|
||||||
|
* [llama.cpp](llama.cpp.md)
|
||||||
* [LoRA](LoRA.md)
|
* [LoRA](LoRA.md)
|
||||||
* [Low VRAM guide](Low-VRAM-guide.md)
|
* [Low VRAM guide](Low-VRAM-guide.md)
|
||||||
* [RWKV model](RWKV-model.md)
|
* [RWKV model](RWKV-model.md)
|
||||||
|
Loading…
Reference in New Issue
Block a user