Commit Graph

647 Commits

Author SHA1 Message Date
oobabooga
4067cecf67 Bump bitsandbytes version 2023-01-20 12:51:49 -03:00
oobabooga
8f3deec759 Prevent the history from being altered by the html script 2023-01-20 01:59:51 -03:00
oobabooga
78d5a999e6 Improve prompt formatation 2023-01-20 01:54:38 -03:00
oobabooga
70ff685736 Encode the input string correctly 2023-01-20 00:45:02 -03:00
oobabooga
83584ae2d7 Clearer installation instructions 2023-01-20 00:20:35 -03:00
oobabooga
b66d18d5a0 Allow presets/characters with '.' in their names 2023-01-19 21:56:33 -03:00
oobabooga
c4f7a874d5 Fix the regex... 2023-01-19 21:16:11 -03:00
oobabooga
8d4170826f Update README 2023-01-19 21:08:26 -03:00
oobabooga
11c3214981 Fix some regexes 2023-01-19 19:59:34 -03:00
oobabooga
e61138bdad Minor fixes 2023-01-19 19:04:54 -03:00
oobabooga
2181fca709 Better defaults for chat 2023-01-19 18:58:45 -03:00
oobabooga
cd7b07239f Add Colab guide 2023-01-19 17:58:04 -03:00
oobabooga
83808171d3 Add --share option for Colab 2023-01-19 17:31:29 -03:00
oobabooga
b054367be2 Update README 2023-01-19 16:54:58 -03:00
oobabooga
8d788874d7 Add support for characters 2023-01-19 16:46:46 -03:00
oobabooga
3121f4788e Fix uploading chat log in --chat mode 2023-01-19 15:05:42 -03:00
oobabooga
849e4c7f90 Better way of finding the generated reply in the output string 2023-01-19 14:57:01 -03:00
oobabooga
d03b0ad7a8 Implement saving/loading chat logs (#9) 2023-01-19 14:03:47 -03:00
oobabooga
39bfea5a22 Add a progress bar 2023-01-19 12:20:57 -03:00
oobabooga
5390fc87c8 add auto-devices when disk is used 2023-01-19 12:11:44 -03:00
oobabooga
759da435e3 Release 8-bit models memory 2023-01-19 12:03:16 -03:00
oobabooga
f9faad4cfa Add low VRAM guide 2023-01-19 11:25:17 -03:00
oobabooga
7ace04864a Implement sending layers to disk with --disk (#10) 2023-01-19 11:09:24 -03:00
oobabooga
1ce95ee817 Mention text streaming 2023-01-19 10:46:41 -03:00
oobabooga
93fa9bbe01 Clean up the streaming implementation 2023-01-19 10:43:05 -03:00
oobabooga
c90310e40e Small simplification 2023-01-19 00:41:57 -03:00
oobabooga
99536ef5bf Add no-stream option 2023-01-18 23:56:42 -03:00
oobabooga
116299b3ad Manual eos_token implementation 2023-01-18 22:57:39 -03:00
oobabooga
3cb30bed0a Add a "stop" button 2023-01-18 22:44:47 -03:00
oobabooga
8f27d33034 Fix another bug 2023-01-18 22:08:23 -03:00
oobabooga
6c7f187586 Minor change 2023-01-18 21:59:23 -03:00
oobabooga
b3cba0b330 Bug 2023-01-18 21:54:44 -03:00
oobabooga
df2e910421 Stop generating in chat mode when \nYou: is generated 2023-01-18 21:51:18 -03:00
oobabooga
022960a087 This is the correct way of sampling 1 token at a time 2023-01-18 21:37:21 -03:00
oobabooga
0f01a3b1fa Implement text streaming (#10)
Still experimental. There might be bugs.
2023-01-18 19:06:50 -03:00
oobabooga
ca13acdfa0 Ensure that the chat prompt will always contain < 2048 tokens
This way, we can keep the context string at the top of the prompt
even if you keep talking to the bot for hours.

Before this commit, the prompt would be simply truncated and the
context string would eventually be lost.
2023-01-17 20:16:23 -03:00
oobabooga
6456777b09 Clean things up 2023-01-16 16:35:45 -03:00
oobabooga
3a99b2b030 Change a truncation parameter 2023-01-16 13:53:30 -03:00
oobabooga
54bf55372b Truncate prompts to 2048 characters 2023-01-16 13:43:23 -03:00
oobabooga
99d24bdbfe
Update README.md 2023-01-16 11:23:45 -03:00
oobabooga
ed1d2c0d38
Update README.md 2023-01-16 11:19:23 -03:00
oobabooga
82c50a09b2
Update README.md 2023-01-16 10:31:13 -03:00
oobabooga
c7a2818665
Grammar 2023-01-16 10:10:09 -03:00
oobabooga
3a28865eb2
Update README.md 2023-01-16 10:07:00 -03:00
oobabooga
d973897021
Typo 2023-01-16 01:52:28 -03:00
oobabooga
18626c4788
Update README.md 2023-01-16 01:51:59 -03:00
oobabooga
47a20638de Don't need this 2023-01-15 23:15:30 -03:00
oobabooga
b55486fa00 Reorganize things 2023-01-15 23:01:51 -03:00
oobabooga
a3bb18002a Update README 2023-01-15 22:34:56 -03:00
oobabooga
ebf4d5f506 Add --max-gpu-memory parameter for #7 2023-01-15 22:33:35 -03:00