oobabooga
|
0b2a6b2819
|
Add file
|
2023-01-25 10:19:50 -03:00 |
|
oobabooga
|
3b8f0021cc
|
Stop generating at \nYou: in chat mode
|
2023-01-25 10:17:55 -03:00 |
|
oobabooga
|
54e77acac4
|
Rename to "Generation parameters preset" for clarity
|
2023-01-23 20:49:44 -03:00 |
|
oobabooga
|
6c40f7eeb4
|
New NovelAI/KoboldAI preset selection
|
2023-01-23 20:44:27 -03:00 |
|
oobabooga
|
ce4756fb88
|
Allow uploading chat history in official pygmalion web ui format
|
2023-01-23 15:29:01 -03:00 |
|
oobabooga
|
8325e23923
|
Fix bug in loading chat history as text file
|
2023-01-23 14:28:02 -03:00 |
|
oobabooga
|
059d47edb5
|
Submit with enter instead of shift+enter in chat mode
|
2023-01-23 14:04:01 -03:00 |
|
oobabooga
|
22845ba445
|
Update README
|
2023-01-23 13:41:50 -03:00 |
|
oobabooga
|
4820379139
|
Add debug preset (deterministic, should always give the same responses)
|
2023-01-23 13:36:01 -03:00 |
|
oobabooga
|
5b60691367
|
Update README
|
2023-01-23 10:05:25 -03:00 |
|
oobabooga
|
085d5cbcb9
|
Update README
|
2023-01-23 10:03:19 -03:00 |
|
oobabooga
|
8a68930220
|
Update README
|
2023-01-23 10:02:35 -03:00 |
|
oobabooga
|
947b50e8ea
|
Allow uploading chat history as simple text files
|
2023-01-23 09:45:10 -03:00 |
|
oobabooga
|
ebf720585b
|
Mention time and it/s in terminal with streaming off
|
2023-01-22 20:07:19 -03:00 |
|
oobabooga
|
d87310ad61
|
Send last input to the input box when "Remove last" is clicked
|
2023-01-22 19:40:22 -03:00 |
|
oobabooga
|
3c5454c6f8
|
Update README
|
2023-01-22 19:38:07 -03:00 |
|
oobabooga
|
8f37383dc6
|
Update README
|
2023-01-22 19:29:24 -03:00 |
|
oobabooga
|
9592e7e618
|
Update README
|
2023-01-22 19:28:03 -03:00 |
|
oobabooga
|
c410fe5ab8
|
Add simplified colab notebook
|
2023-01-22 19:27:30 -03:00 |
|
oobabooga
|
41806fe40d
|
Update chat screenshot
|
2023-01-22 17:28:51 -03:00 |
|
oobabooga
|
d0ea6d5f86
|
Make the maximum history size in prompt unlimited by default
|
2023-01-22 17:17:35 -03:00 |
|
oobabooga
|
00f3b0996b
|
Warn the user that chat mode becomes a lot slower with text streaming
|
2023-01-22 16:19:11 -03:00 |
|
oobabooga
|
c5cc3a3075
|
Fix bug in "remove last" button
|
2023-01-22 13:10:36 -03:00 |
|
oobabooga
|
a410cf1345
|
Mention that "Chat history size" means "Chat history size in prompt"
|
2023-01-22 03:15:35 -03:00 |
|
oobabooga
|
b3e1a874bc
|
Fix bug in loading history
|
2023-01-22 02:32:54 -03:00 |
|
oobabooga
|
62b533f344
|
Add "regenerate" button to the chat
|
2023-01-22 02:19:58 -03:00 |
|
oobabooga
|
5b50db9dee
|
Mention pygmalion support
|
2023-01-22 01:30:55 -03:00 |
|
oobabooga
|
6b06fca2f1
|
Mention pygmalion support
|
2023-01-22 01:28:52 -03:00 |
|
oobabooga
|
94ecbc6dff
|
Export history as nicely formatted json
|
2023-01-22 01:24:16 -03:00 |
|
oobabooga
|
deacb96c34
|
Change the pygmalion default context
|
2023-01-22 00:49:59 -03:00 |
|
oobabooga
|
23f94f559a
|
Improve the chat prompt design
|
2023-01-22 00:35:42 -03:00 |
|
oobabooga
|
139e2f0ab4
|
Redesign the upload/download chat history buttons
|
2023-01-22 00:22:50 -03:00 |
|
oobabooga
|
434d4b128c
|
Add refresh buttons for the model/preset/character menus
|
2023-01-22 00:02:46 -03:00 |
|
oobabooga
|
bc664ecf3b
|
Update the installation instructions for low attention span people
|
2023-01-21 22:54:35 -03:00 |
|
oobabooga
|
1e5e56fa2e
|
Better recognize the 4chan model (for #19)
|
2023-01-21 22:13:01 -03:00 |
|
oobabooga
|
aadf4e899a
|
Improve example dialogue handling
|
2023-01-21 15:04:13 -03:00 |
|
oobabooga
|
f9dbe7e08e
|
Update README
|
2023-01-21 03:05:55 -03:00 |
|
oobabooga
|
27e2d932b0
|
Don't export include the example dialogue in the export json
|
2023-01-21 02:55:13 -03:00 |
|
oobabooga
|
990ee54ddd
|
Move the example dialogue to the chat history, and keep it hidden.
This greatly improves the performance of text generation, as
histories can be quite long. It also makes more sense to implement
it this way.
|
2023-01-21 02:48:06 -03:00 |
|
oobabooga
|
3f2c1e7170
|
Merge pull request #16 from 81300/model-download
Allow specifying the Hugging Face Git branch when downloading models
|
2023-01-21 00:43:35 -03:00 |
|
oobabooga
|
1e541d4882
|
Update download-model.py
|
2023-01-21 00:43:00 -03:00 |
|
oobabooga
|
18ef72d7c0
|
Update download-model.py
|
2023-01-21 00:38:23 -03:00 |
|
oobabooga
|
d7299df01f
|
Rename parameters
|
2023-01-21 00:33:41 -03:00 |
|
oobabooga
|
86a2832f3b
|
Merge pull request #17 from Silver267/main
Add --disk-cache-dir and --max-cpu-mem parameters (Properly pushed to github)
|
2023-01-21 00:26:19 -03:00 |
|
oobabooga
|
5df03bf0fd
|
Merge branch 'main' into main
|
2023-01-21 00:25:34 -03:00 |
|
oobabooga
|
faaafe7c0e
|
Better parameter naming
|
2023-01-20 23:45:16 -03:00 |
|
Silver267
|
f4634e4c32
|
Update.
|
2023-01-20 17:05:43 -05:00 |
|
81300
|
fffd49e64e
|
Add --branch option to the model download script
|
2023-01-20 22:51:56 +02:00 |
|
oobabooga
|
c0f2367b54
|
Minor fix
|
2023-01-20 17:09:25 -03:00 |
|
oobabooga
|
185587a33e
|
Add a history size parameter to the chat
If too many messages are used in the prompt, the model
gets really slow. It is useful to have the ability to
limit this.
|
2023-01-20 17:03:09 -03:00 |
|