Alex "mcmonkey" Goodwin
1f50dbe352
Experimental jank multiGPU inference that's 2x faster than native somehow ( #2100 )
2023-05-17 10:41:09 -03:00
Andrei
e657dd342d
Add in-memory cache support for llama.cpp ( #1936 )
2023-05-15 20:19:55 -03:00
AlphaAtlas
071f0776ad
Add llama.cpp GPU offload option ( #2060 )
2023-05-14 22:58:11 -03:00
oobabooga
23d3f6909a
Update README.md
2023-05-11 10:21:20 -03:00
oobabooga
2930e5a895
Update README.md
2023-05-11 10:04:38 -03:00
oobabooga
0ff38c994e
Update README.md
2023-05-11 09:58:58 -03:00
oobabooga
e6959a5d9a
Update README.md
2023-05-11 09:54:22 -03:00
oobabooga
dcfd09b61e
Update README.md
2023-05-11 09:49:57 -03:00
oobabooga
7a49ceab29
Update README.md
2023-05-11 09:42:39 -03:00
oobabooga
57dc44a995
Update README.md
2023-05-10 12:48:25 -03:00
oobabooga
181b102521
Update README.md
2023-05-10 12:09:47 -03:00
Carl Kenner
814f754451
Support for MPT, INCITE, WizardLM, StableLM, Galactica, Vicuna, Guanaco, and Baize instruction following ( #1596 )
2023-05-09 20:37:31 -03:00
Wojtab
e9e75a9ec7
Generalize multimodality (llava/minigpt4 7b and 13b now supported) ( #1741 )
2023-05-09 20:18:02 -03:00
oobabooga
00e333d790
Add MOSS support
2023-05-04 23:20:34 -03:00
oobabooga
b6ff138084
Add --checkpoint argument for GPTQ
2023-05-04 15:17:20 -03:00
Ahmed Said
fbcd32988e
added no_mmap & mlock parameters to llama.cpp and removed llamacpp_model_alternative ( #1649 )
...
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-02 18:25:28 -03:00
oobabooga
f39c99fa14
Load more than one LoRA with --lora, fix a bug
2023-04-25 22:58:48 -03:00
oobabooga
b6af2e56a2
Add --character flag, add character to settings.json
2023-04-24 13:19:42 -03:00
eiery
78d1977ebf
add n_batch support for llama.cpp ( #1115 )
2023-04-24 03:46:18 -03:00
Andy Salerno
654933c634
New universal API with streaming/blocking endpoints ( #990 )
...
Previous title: Add api_streaming extension and update api-example-stream to use it
* Merge with latest main
* Add parameter capturing encoder_repetition_penalty
* Change some defaults, minor fixes
* Add --api, --public-api flags
* remove unneeded/broken comment from blocking API startup. The comment is already correctly emitted in try_start_cloudflared by calling the lambda we pass in.
* Update on_start message for blocking_api, it should say 'non-streaming' and not 'streaming'
* Update the API examples
* Change a comment
* Update README
* Remove the gradio API
* Remove unused import
* Minor change
* Remove unused import
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-23 15:52:43 -03:00
oobabooga
7438f4f6ba
Change GPTQ triton default settings
2023-04-22 12:27:30 -03:00
oobabooga
fe02281477
Update README.md
2023-04-22 03:05:00 -03:00
oobabooga
038fa3eb39
Update README.md
2023-04-22 02:46:07 -03:00
oobabooga
505c2c73e8
Update README.md
2023-04-22 00:11:27 -03:00
oobabooga
f8da9a0424
Update README.md
2023-04-18 20:25:08 -03:00
oobabooga
c3f6e65554
Update README.md
2023-04-18 20:23:31 -03:00
oobabooga
eb15193327
Update README.md
2023-04-18 13:07:08 -03:00
oobabooga
7fbfc489e2
Update README.md
2023-04-18 12:56:37 -03:00
oobabooga
f559f9595b
Update README.md
2023-04-18 12:54:09 -03:00
loeken
89e22d4d6a
added windows/docker docs ( #1027 )
2023-04-18 12:47:43 -03:00
oobabooga
8275989f03
Add new 1-click installers for Linux and MacOS
2023-04-18 02:40:36 -03:00
oobabooga
301c687c64
Update README.md
2023-04-17 11:25:26 -03:00
oobabooga
89bc540557
Update README
2023-04-17 10:55:35 -03:00
practicaldreamer
3961f49524
Add note about --no-fused_mlp ignoring --gpu-memory ( #1301 )
2023-04-17 10:46:37 -03:00
sgsdxzy
b57ffc2ec9
Update to support GPTQ triton commit c90adef ( #1229 )
2023-04-17 01:11:18 -03:00
oobabooga
3e5cdd005f
Update README.md
2023-04-16 23:28:59 -03:00
oobabooga
39099663a0
Add 4-bit LoRA support ( #1200 )
2023-04-16 23:26:52 -03:00
oobabooga
705121161b
Update README.md
2023-04-16 20:03:03 -03:00
oobabooga
50c55a51fc
Update README.md
2023-04-16 19:22:31 -03:00
Forkoz
c6fe1ced01
Add ChatGLM support ( #1256 )
...
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 19:15:03 -03:00
oobabooga
c96529a1b3
Update README.md
2023-04-16 17:00:03 -03:00
oobabooga
004f275efe
Update README.md
2023-04-14 23:36:56 -03:00
oobabooga
83964ed354
Update README.md
2023-04-14 23:33:54 -03:00
oobabooga
c41037db68
Update README.md
2023-04-14 23:32:39 -03:00
v0xie
9d66957207
Add --listen-host launch option ( #1122 )
2023-04-13 21:35:08 -03:00
oobabooga
403be8a27f
Update README.md
2023-04-13 21:23:35 -03:00
Light
97e67d136b
Update README.md
2023-04-13 21:00:58 +08:00
Light
15d5a043f2
Merge remote-tracking branch 'origin/main' into triton
2023-04-13 19:38:51 +08:00
oobabooga
7dfbe54f42
Add --model-menu option
2023-04-12 21:24:26 -03:00
MarlinMr
47daf891fe
Link to developer.nvidia.com ( #1104 )
2023-04-12 15:56:42 -03:00
Light
f3591ccfa1
Keep minimal change.
2023-04-12 23:26:06 +08:00
oobabooga
461ca7faf5
Mention that pull request reviews are welcome
2023-04-11 23:12:48 -03:00
oobabooga
749c08a4ff
Update README.md
2023-04-11 14:42:10 -03:00
IggoOnCode
09d8119e3c
Add CPU LoRA training ( #938 )
...
(It's very slow)
2023-04-10 17:29:00 -03:00
oobabooga
f035b01823
Update README.md
2023-04-10 16:20:23 -03:00
Jeff Lefebvre
b7ca89ba3f
Mention that build-essential is required ( #1013 )
2023-04-10 16:19:10 -03:00
MarkovInequality
992663fa20
Added xformers support to Llama ( #950 )
2023-04-09 23:08:40 -03:00
oobabooga
bce1b7fbb2
Update README.md
2023-04-09 02:19:40 -03:00
oobabooga
f7860ce192
Update README.md
2023-04-09 02:19:17 -03:00
oobabooga
ece8ed2c84
Update README.md
2023-04-09 02:18:42 -03:00
MarlinMr
ec979cd9c4
Use updated docker compose ( #877 )
2023-04-07 10:48:47 -03:00
MarlinMr
2c0018d946
Cosmetic change of README.md ( #878 )
2023-04-07 10:47:10 -03:00
oobabooga
848c4edfd5
Update README.md
2023-04-06 22:52:35 -03:00
oobabooga
e047cd1def
Update README
2023-04-06 22:50:58 -03:00
loeken
08b9d1b23a
creating a layer with Docker/docker-compose ( #633 )
2023-04-06 22:46:04 -03:00
oobabooga
d9e7aba714
Update README.md
2023-04-06 13:42:24 -03:00
oobabooga
eec3665845
Add instructions for updating requirements
2023-04-06 13:24:01 -03:00
oobabooga
4a28f39823
Update README.md
2023-04-06 02:47:27 -03:00
eiery
19b516b11b
fix link to streaming api example ( #803 )
2023-04-05 14:50:23 -03:00
oobabooga
7617ed5bfd
Add AMD instructions
2023-04-05 14:42:58 -03:00
oobabooga
770ef5744f
Update README
2023-04-05 14:38:11 -03:00
oobabooga
65d8a24a6d
Show profile pictures in the Character tab
2023-04-04 22:28:49 -03:00
oobabooga
b24147c7ca
Document --pre_layer
2023-04-03 17:34:25 -03:00
oobabooga
525f729b8e
Update README.md
2023-04-02 21:12:41 -03:00
oobabooga
53084241b4
Update README.md
2023-04-02 20:50:06 -03:00
oobabooga
b6f817be45
Update README.md
2023-04-01 14:54:10 -03:00
oobabooga
88fa38ac01
Update README.md
2023-04-01 14:49:03 -03:00
oobabooga
4b57bd0d99
Update README.md
2023-04-01 14:38:04 -03:00
oobabooga
b53bec5a1f
Update README.md
2023-04-01 14:37:35 -03:00
oobabooga
9160586c04
Update README.md
2023-04-01 14:31:10 -03:00
oobabooga
7ec11ae000
Update README.md
2023-04-01 14:15:19 -03:00
oobabooga
012f4f83b8
Update README.md
2023-04-01 13:55:15 -03:00
oobabooga
2c52310642
Add --threads flag for llama.cpp
2023-03-31 21:18:05 -03:00
oobabooga
cbfe0b944a
Update README.md
2023-03-31 17:49:11 -03:00
oobabooga
5c4e44b452
llama.cpp documentation
2023-03-31 15:20:39 -03:00
oobabooga
d4a9b5ea97
Remove redundant preset (see the plot in #587 )
2023-03-30 17:34:44 -03:00
oobabooga
41b58bc47e
Update README.md
2023-03-29 11:02:29 -03:00
oobabooga
3b4447a4fe
Update README.md
2023-03-29 02:24:11 -03:00
oobabooga
5d0b83c341
Update README.md
2023-03-29 02:22:19 -03:00
oobabooga
c2a863f87d
Mention the updated one-click installer
2023-03-29 02:11:51 -03:00
oobabooga
010b259dde
Update documentation
2023-03-28 17:46:00 -03:00
oobabooga
036163a751
Change description
2023-03-27 23:39:26 -03:00
oobabooga
30585b3e71
Update README
2023-03-27 23:35:01 -03:00
oobabooga
49c10c5570
Add support for the latest GPTQ models with group-size ( #530 )
...
**Warning: old 4-bit weights will not work anymore!**
See here how to get up to date weights: https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model#step-2-get-the-pre-converted-weights
2023-03-26 00:11:33 -03:00
oobabooga
70f9565f37
Update README.md
2023-03-25 02:35:30 -03:00
oobabooga
04417b658b
Update README.md
2023-03-24 01:40:43 -03:00
oobabooga
143b5b5edf
Mention one-click-bandaid in the README
2023-03-23 23:28:50 -03:00
oobabooga
6872ffd976
Update README.md
2023-03-20 16:53:14 -03:00
oobabooga
dd4374edde
Update README
2023-03-19 20:15:15 -03:00
oobabooga
9378754cc7
Update README
2023-03-19 20:14:50 -03:00
oobabooga
7ddf6147ac
Update README.md
2023-03-19 19:25:52 -03:00
oobabooga
ddb62470e9
--no-cache and --gpu-memory in MiB for fine VRAM control
2023-03-19 19:21:41 -03:00
oobabooga
0cbe2dd7e9
Update README.md
2023-03-18 12:24:54 -03:00
oobabooga
d2a7fac8ea
Use pip instead of conda for pytorch
2023-03-18 11:56:04 -03:00
oobabooga
a0b1a30fd5
Specify torchvision/torchaudio versions
2023-03-18 11:23:56 -03:00
oobabooga
a163807f86
Update README.md
2023-03-18 03:07:27 -03:00
oobabooga
a7acfa4893
Update README.md
2023-03-17 22:57:46 -03:00
oobabooga
dc35861184
Update README.md
2023-03-17 21:05:17 -03:00
oobabooga
f2a5ca7d49
Update README.md
2023-03-17 20:50:27 -03:00
oobabooga
8c8286b0e6
Update README.md
2023-03-17 20:49:40 -03:00
oobabooga
0c05e65e5c
Update README.md
2023-03-17 20:25:42 -03:00
oobabooga
66e8d12354
Update README.md
2023-03-17 19:59:37 -03:00
oobabooga
9a871117d7
Update README.md
2023-03-17 19:52:22 -03:00
oobabooga
d4f38b6a1f
Update README.md
2023-03-17 18:57:48 -03:00
oobabooga
ad7c829953
Update README.md
2023-03-17 18:55:01 -03:00
oobabooga
4426f941e0
Update the installation instructions. Tldr use WSL
2023-03-17 18:51:07 -03:00
oobabooga
ebef4a510b
Update README
2023-03-17 11:58:45 -03:00
oobabooga
cdfa787bcb
Update README
2023-03-17 11:53:28 -03:00
oobabooga
dd1c5963da
Update README
2023-03-16 12:45:27 -03:00
oobabooga
445ebf0ba8
Update README.md
2023-03-15 20:06:46 -03:00
oobabooga
09045e4bdb
Add WSL guide
2023-03-15 19:42:06 -03:00
oobabooga
128d18e298
Update README.md
2023-03-14 17:57:25 -03:00
oobabooga
1236c7f971
Update README.md
2023-03-14 17:56:15 -03:00
oobabooga
b419dffba3
Update README.md
2023-03-14 17:55:35 -03:00
oobabooga
87192e2813
Update README
2023-03-14 08:02:21 -03:00
oobabooga
3da73e409f
Merge branch 'main' into Zerogoki00-opt4-bit
2023-03-14 07:50:36 -03:00
Ayanami Rei
b746250b2f
Update README
2023-03-13 20:20:45 +03:00
oobabooga
66b6971b61
Update README
2023-03-13 12:44:18 -03:00
oobabooga
ddea518e0f
Document --auto-launch
2023-03-13 12:43:33 -03:00
oobabooga
d97bfb8713
Update README.md
2023-03-13 12:39:33 -03:00
oobabooga
bdff37f0bb
Update README.md
2023-03-13 11:05:51 -03:00
oobabooga
d168b6e1f7
Update README.md
2023-03-12 17:54:07 -03:00
oobabooga
54e8f0c31f
Update README.md
2023-03-12 16:58:00 -03:00
oobabooga
0a4d8a5cf6
Delete README.md
2023-03-12 16:43:06 -03:00
oobabooga
0b86ac38b1
Initial commit
2023-03-12 16:40:10 -03:00
oobabooga
3375eaece0
Update README
2023-03-12 15:01:32 -03:00
oobabooga
17210ff88f
Update README.md
2023-03-12 14:31:24 -03:00
oobabooga
4dc1d8c091
Update README.md
2023-03-12 12:46:53 -03:00
oobabooga
89e9493509
Update README
2023-03-12 11:23:20 -03:00
draff
28fd4fc970
Change wording to be consistent with other args
2023-03-10 23:34:13 +00:00
draff
804486214b
Re-implement --load-in-4bit and update --llama-bits arg description
2023-03-10 23:21:01 +00:00
draff
e6c631aea4
Replace --load-in-4bit with --llama-bits
...
Replaces --load-in-4bit with a more flexible --llama-bits arg to allow for 2 and 3 bit models as well. This commit also fixes a loading issue with .pt files which are not in the root of the models folder
2023-03-10 21:36:45 +00:00
oobabooga
7c3d1b43c1
Merge pull request #204 from MichealC0/patch-1
...
Update README.md
2023-03-09 23:04:09 -03:00
oobabooga
1a3d25f75d
Merge pull request #206 from oobabooga/llama-4bit
...
Add LLaMA 4-bit support
2023-03-09 21:07:32 -03:00
oobabooga
eb0cb9b6df
Update README
2023-03-09 20:53:52 -03:00
oobabooga
d41e3c233b
Update README.md
2023-03-09 18:02:44 -03:00
oobabooga
33414478bf
Update README
2023-03-09 11:13:03 -03:00
oobabooga
e7adf5fe4e
Add Contrastive Search preset #197
2023-03-09 10:27:11 -03:00
Chimdumebi Nebolisa
4dd14dcab4
Update README.md
2023-03-09 10:22:09 +01:00
oobabooga
b4bfd87319
Update README.md
2023-03-06 20:55:01 -03:00
oobabooga
d0e8780555
Update README.md
2023-03-06 20:17:59 -03:00
oobabooga
18ccfcd7fe
Update README.md
2023-03-06 20:15:55 -03:00
oobabooga
91823e1ed1
Update README.md
2023-03-06 16:48:31 -03:00
oobabooga
aa7ce0665e
Merge branch 'main' of github.com:oobabooga/text-generation-webui
2023-03-06 10:58:41 -03:00
oobabooga
bf56b6c1fb
Load settings.json without the need for --settings settings.json
...
This is for setting UI defaults
2023-03-06 10:57:45 -03:00
oobabooga
2de9f122cd
Update README.md
2023-03-06 09:34:49 -03:00
oobabooga
736f61610b
Update README
2023-03-04 01:33:52 -03:00
oobabooga
76378c6cc2
Update README
2023-03-02 11:27:15 -03:00
oobabooga
f4b130e2bd
Update README.md
2023-02-27 15:15:45 -03:00
oobabooga
c183d2917c
Update README.md
2023-02-26 00:59:07 -03:00
oobabooga
cfe010b244
Update README.md
2023-02-26 00:54:37 -03:00
oobabooga
87d9f3e329
Update README.md
2023-02-26 00:54:19 -03:00
oobabooga
955997a90b
Update README.md
2023-02-26 00:54:07 -03:00
oobabooga
c593dfa827
Update README.md
2023-02-25 18:57:34 -03:00
oobabooga
7872a64f78
Update README.md
2023-02-25 18:56:43 -03:00
oobabooga
88cfc84ddb
Update README
2023-02-25 01:33:26 -03:00
oobabooga
0b90e0b3b6
Update README.md
2023-02-24 12:01:07 -03:00
oobabooga
1a23e6d185
Add Pythia to README
2023-02-24 11:38:01 -03:00
oobabooga
f4f508c8e2
Update README.md
2023-02-24 09:03:09 -03:00
oobabooga
ced5d9ab04
Update README.md
2023-02-23 10:04:07 -03:00
oobabooga
b18071330f
Update README.md
2023-02-23 01:32:05 -03:00
oobabooga
b4a7f5fa70
Update README.md
2023-02-22 01:54:12 -03:00
oobabooga
e195377050
Deprecate torch dumps, move to safetensors (they load even faster)
2023-02-20 15:03:19 -03:00
oobabooga
58520a1f75
Update README.md
2023-02-20 12:44:31 -03:00
oobabooga
05e9da0c12
Update README.md
2023-02-18 22:34:51 -03:00
oobabooga
b1add0e586
Update README.md
2023-02-18 22:32:16 -03:00
oobabooga
348acdf626
Mention deepspeed in the README
2023-02-16 17:29:48 -03:00
oobabooga
05b53e4626
Update README
2023-02-15 14:43:34 -03:00
oobabooga
ed73d00bd5
Update README
2023-02-15 14:43:13 -03:00
oobabooga
30fcb26737
Update README
2023-02-15 14:42:41 -03:00
oobabooga
5ee9283cae
Mention BLIP
2023-02-15 13:53:38 -03:00
oobabooga
8d3b3959e7
Document --picture option
2023-02-15 13:50:18 -03:00
oobabooga
01e5772302
Update README.md
2023-02-14 13:06:26 -03:00
oobabooga
210c918199
Update README.md
2023-02-13 21:49:19 -03:00
oobabooga
b7ddcab53a
Update README.md
2023-02-13 15:52:49 -03:00
oobabooga
939e9d00a2
Update README.md
2023-02-12 00:47:03 -03:00
oobabooga
bf9dd8f8ee
Add --text-only option to the download script
2023-02-12 00:42:56 -03:00
oobabooga
42cc307409
Update README.md
2023-02-12 00:34:55 -03:00
oobabooga
144857acfe
Update README
2023-02-11 14:49:11 -03:00
oobabooga
0dd1409f24
Add penalty_alpha parameter (contrastive search)
2023-02-11 14:48:12 -03:00
oobabooga
8aafb55693
1-click installer now also works for AMD GPUs
...
(I think)
2023-02-11 14:24:47 -03:00
oobabooga
1176d64b13
Update README.md
2023-02-11 07:56:12 -03:00
Spencer-Dawson
c5324d653b
re-added missed README changes
2023-02-11 00:13:06 -07:00
oobabooga
cf89ef1c74
Update README.md
2023-02-10 21:46:29 -03:00
oobabooga
8782ac1911
Update README.md
2023-02-10 17:10:27 -03:00
oobabooga
7d7cc37560
Add Linux 1-click installer
2023-02-10 17:09:53 -03:00
oobabooga
991de5ed40
Update README.md
2023-02-09 14:36:47 -03:00
oobabooga
04d3d0aee6
Add 1-click windows installer (for #45 )
2023-02-09 13:27:30 -03:00
oobabooga
a21620fc59
Update README
2023-02-08 01:17:50 -03:00
oobabooga
fc0493d885
Add credits
2023-02-08 00:09:41 -03:00
oobabooga
53af062fa5
Update README.md
2023-02-05 23:14:25 -03:00
oobabooga
90bb2caffd
Update README.md
2023-02-03 19:45:11 -03:00
oobabooga
93b0d1b1b8
Update README.md
2023-02-03 10:14:52 -03:00
oobabooga
6212b41930
Update README
2023-02-03 09:13:14 -03:00
oobabooga
7f4315b120
Mention 8bit fix for Windows users
...
Closes #44 , #20
2023-02-02 11:00:57 -03:00
oobabooga
7aa3d6583e
Update README.md
2023-01-30 09:45:31 -03:00
oobabooga
239f96a9c5
Add extensions guide
2023-01-30 09:44:57 -03:00
oobabooga
f92996b3c8
Update README.md
2023-01-29 14:37:05 -03:00
oobabooga
9e4db10cd0
Update README.md
2023-01-29 03:13:22 -03:00
oobabooga
1a139664f5
Grammar
2023-01-29 02:54:36 -03:00
oobabooga
89c862b179
Update README
2023-01-28 20:37:43 -03:00
oobabooga
6b5dcd46c5
Add support for extensions
...
This is experimental.
2023-01-27 00:40:39 -03:00
oobabooga
d49710878e
Update README.md
2023-01-26 16:17:15 -03:00
oobabooga
ff180f3e60
Update installation instructions (for #15 )
2023-01-26 12:32:21 -03:00
oobabooga
61611197e0
Add --verbose option (oops)
2023-01-26 02:18:06 -03:00
oobabooga
64f278d248
Update README
2023-01-25 19:43:30 -03:00
oobabooga
b77933d327
File names must be img_me.jpg and img_bot.jpg
2023-01-25 19:40:30 -03:00
oobabooga
fc73188ec7
Allow specifying your own profile picture in chat mode
2023-01-25 19:37:44 -03:00
oobabooga
651eb50dd1
Merge branch 'main' of github.com:oobabooga/text-generation-webui
2023-01-25 10:20:03 -03:00
oobabooga
3b8f0021cc
Stop generating at \nYou: in chat mode
2023-01-25 10:17:55 -03:00
oobabooga
befabbb862
Update README.md
2023-01-24 23:01:43 -03:00
oobabooga
7b98971df1
Update README.md
2023-01-24 18:39:29 -03:00
oobabooga
86b237eb0b
Update README.md
2023-01-24 09:29:34 -03:00
oobabooga
c2d1f04305
Update README.md
2023-01-23 23:34:20 -03:00
oobabooga
6c40f7eeb4
New NovelAI/KoboldAI preset selection
2023-01-23 20:44:27 -03:00
oobabooga
22845ba445
Update README
2023-01-23 13:41:50 -03:00
oobabooga
5b60691367
Update README
2023-01-23 10:05:25 -03:00
oobabooga
085d5cbcb9
Update README
2023-01-23 10:03:19 -03:00
oobabooga
8a68930220
Update README
2023-01-23 10:02:35 -03:00
oobabooga
3c5454c6f8
Update README
2023-01-22 19:38:07 -03:00
oobabooga
8f37383dc6
Update README
2023-01-22 19:29:24 -03:00
oobabooga
9592e7e618
Update README
2023-01-22 19:28:03 -03:00
oobabooga
c410fe5ab8
Add simplified colab notebook
2023-01-22 19:27:30 -03:00
oobabooga
41806fe40d
Update chat screenshot
2023-01-22 17:28:51 -03:00
oobabooga
00f3b0996b
Warn the user that chat mode becomes a lot slower with text streaming
2023-01-22 16:19:11 -03:00
oobabooga
5b50db9dee
Mention pygmalion support
2023-01-22 01:30:55 -03:00
oobabooga
6b06fca2f1
Mention pygmalion support
2023-01-22 01:28:52 -03:00
oobabooga
434d4b128c
Add refresh buttons for the model/preset/character menus
2023-01-22 00:02:46 -03:00
oobabooga
bc664ecf3b
Update the installation instructions for low attention span people
2023-01-21 22:54:35 -03:00
oobabooga
f9dbe7e08e
Update README
2023-01-21 03:05:55 -03:00
oobabooga
990ee54ddd
Move the example dialogue to the chat history, and keep it hidden.
...
This greatly improves the performance of text generation, as
histories can be quite long. It also makes more sense to implement
it this way.
2023-01-21 02:48:06 -03:00
oobabooga
d7299df01f
Rename parameters
2023-01-21 00:33:41 -03:00
oobabooga
5df03bf0fd
Merge branch 'main' into main
2023-01-21 00:25:34 -03:00
oobabooga
faaafe7c0e
Better parameter naming
2023-01-20 23:45:16 -03:00
Silver267
f4634e4c32
Update.
2023-01-20 17:05:43 -05:00
oobabooga
83584ae2d7
Clearer installation instructions
2023-01-20 00:20:35 -03:00
oobabooga
8d4170826f
Update README
2023-01-19 21:08:26 -03:00
oobabooga
cd7b07239f
Add Colab guide
2023-01-19 17:58:04 -03:00
oobabooga
83808171d3
Add --share option for Colab
2023-01-19 17:31:29 -03:00
oobabooga
b054367be2
Update README
2023-01-19 16:54:58 -03:00