oobabooga
|
72d539dbff
|
Better separate the FlexGen case
|
2023-03-08 02:54:47 -03:00 |
|
oobabooga
|
0e16c0bacb
|
Remove redeclaration of a function
|
2023-03-08 02:50:49 -03:00 |
|
oobabooga
|
ab50f80542
|
New text streaming method (much faster)
|
2023-03-08 02:46:35 -03:00 |
|
oobabooga
|
8e89bc596b
|
Fix encode() for RWKV
|
2023-03-07 23:15:46 -03:00 |
|
oobabooga
|
19a34941ed
|
Add proper streaming to RWKV
|
2023-03-07 18:17:56 -03:00 |
|
oobabooga
|
8660227e1b
|
Add top_k to RWKV
|
2023-03-07 17:24:28 -03:00 |
|
oobabooga
|
153dfeb4dd
|
Add --rwkv-cuda-on parameter, bump rwkv version
|
2023-03-06 20:12:54 -03:00 |
|
oobabooga
|
6904a507c6
|
Change some parameters
|
2023-03-06 16:29:43 -03:00 |
|
oobabooga
|
20bd645f6a
|
Fix bug in multigpu setups (attempt 3)
|
2023-03-06 15:58:18 -03:00 |
|
oobabooga
|
09a7c36e1b
|
Minor improvement while running custom models
|
2023-03-06 15:36:35 -03:00 |
|
oobabooga
|
24c4c20391
|
Fix bug in multigpu setups (attempt #2)
|
2023-03-06 15:23:29 -03:00 |
|
oobabooga
|
d88b7836c6
|
Fix bug in multigpu setups
|
2023-03-06 14:58:30 -03:00 |
|
oobabooga
|
5bed607b77
|
Increase repetition frequency/penalty for RWKV
|
2023-03-06 14:25:48 -03:00 |
|
oobabooga
|
bf56b6c1fb
|
Load settings.json without the need for --settings settings.json
This is for setting UI defaults
|
2023-03-06 10:57:45 -03:00 |
|
oobabooga
|
e91f4bc25a
|
Add RWKV tokenizer
|
2023-03-06 08:45:49 -03:00 |
|
oobabooga
|
c855b828fe
|
Better handle <USER>
|
2023-03-05 17:01:47 -03:00 |
|
oobabooga
|
2af66a4d4c
|
Fix <USER> in pygmalion replies
|
2023-03-05 16:08:50 -03:00 |
|
oobabooga
|
a54b91af77
|
Improve readability
|
2023-03-05 10:21:15 -03:00 |
|
oobabooga
|
8e706df20e
|
Fix a memory leak when text streaming is on
|
2023-03-05 10:12:43 -03:00 |
|
oobabooga
|
c33715ad5b
|
Move towards HF LLaMA implementation
|
2023-03-05 01:20:31 -03:00 |
|
oobabooga
|
bd8aac8fa4
|
Add LLaMA 8-bit support
|
2023-03-04 13:28:42 -03:00 |
|
oobabooga
|
c93f1fa99b
|
Count the tokens more conservatively
|
2023-03-04 03:10:21 -03:00 |
|
oobabooga
|
ed8b35efd2
|
Add --pin-weight parameter for FlexGen
|
2023-03-04 01:04:02 -03:00 |
|
oobabooga
|
05e703b4a4
|
Print the performance information more reliably
|
2023-03-03 21:24:32 -03:00 |
|
oobabooga
|
5a79863df3
|
Increase the sequence length, decrease batch size
I have no idea what I am doing
|
2023-03-03 15:54:13 -03:00 |
|
oobabooga
|
a345a2acd2
|
Add a tokenizer placeholder
|
2023-03-03 15:16:55 -03:00 |
|
oobabooga
|
5b354817f6
|
Make chat minimally work with LLaMA
|
2023-03-03 15:04:41 -03:00 |
|
oobabooga
|
ea5c5eb3da
|
Add LLaMA support
|
2023-03-03 14:39:14 -03:00 |
|
oobabooga
|
2bff646130
|
Stop chat from flashing dark when processing
|
2023-03-03 13:19:13 -03:00 |
|
oobabooga
|
169209805d
|
Model-aware prompts and presets
|
2023-03-02 11:25:04 -03:00 |
|
oobabooga
|
7bbe32f618
|
Don't return a value in an iterator function
|
2023-03-02 00:48:46 -03:00 |
|
oobabooga
|
ff9f649c0c
|
Remove some unused imports
|
2023-03-02 00:36:20 -03:00 |
|
oobabooga
|
1a05860ca3
|
Ensure proper no-streaming with generation_attempts > 1
|
2023-03-02 00:10:10 -03:00 |
|
oobabooga
|
a2a3e8f797
|
Add --rwkv-strategy parameter
|
2023-03-01 20:02:48 -03:00 |
|
oobabooga
|
449116a510
|
Fix RWKV paths on Windows (attempt)
|
2023-03-01 19:17:16 -03:00 |
|
oobabooga
|
955cf431e8
|
Minor consistency fix
|
2023-03-01 19:11:26 -03:00 |
|
oobabooga
|
f3da6dcc8f
|
Merge pull request #149 from oobabooga/RWKV
Add RWKV support
|
2023-03-01 16:57:45 -03:00 |
|
oobabooga
|
831ac7ed3f
|
Add top_p
|
2023-03-01 16:45:48 -03:00 |
|
oobabooga
|
7c4d5ca8cc
|
Improve the text generation call a bit
|
2023-03-01 16:40:25 -03:00 |
|
oobabooga
|
2f16ce309a
|
Rename a variable
|
2023-03-01 12:33:09 -03:00 |
|
oobabooga
|
9e9cfc4b31
|
Parameters
|
2023-03-01 12:19:37 -03:00 |
|
oobabooga
|
0f6708c471
|
Sort the imports
|
2023-03-01 12:18:17 -03:00 |
|
oobabooga
|
e735806c51
|
Add a generate() function for RWKV
|
2023-03-01 12:16:11 -03:00 |
|
oobabooga
|
659bb76722
|
Add RWKVModel class
|
2023-03-01 12:08:55 -03:00 |
|
oobabooga
|
9c86a1cd4a
|
Add RWKV pip package
|
2023-03-01 11:42:49 -03:00 |
|
oobabooga
|
6837d4d72a
|
Load the model by name
|
2023-02-28 02:52:29 -03:00 |
|
oobabooga
|
a1429d1607
|
Add default extensions to the settings
|
2023-02-28 02:20:11 -03:00 |
|
oobabooga
|
19ccb2aaf5
|
Handle <USER> and <BOT>
|
2023-02-28 01:05:43 -03:00 |
|
oobabooga
|
626da6c731
|
Handle {{user}} and {{char}} in example dialogue
|
2023-02-28 00:59:05 -03:00 |
|
oobabooga
|
e861e68e38
|
Move the chat example dialogue to the prompt
|
2023-02-28 00:50:46 -03:00 |
|
oobabooga
|
f871971de1
|
Trying to get the chat to work
|
2023-02-28 00:25:30 -03:00 |
|
oobabooga
|
67ee7bead7
|
Add cpu, bf16 options
|
2023-02-28 00:09:11 -03:00 |
|
oobabooga
|
ebd698905c
|
Add streaming to RWKV
|
2023-02-28 00:04:04 -03:00 |
|
oobabooga
|
70e522732c
|
Move RWKV loader into a separate file
|
2023-02-27 23:50:16 -03:00 |
|
oobabooga
|
ebc64a408c
|
RWKV support prototype
|
2023-02-27 23:03:35 -03:00 |
|
oobabooga
|
021bd55886
|
Better format the prompt when generation attempts > 1
|
2023-02-27 21:37:03 -03:00 |
|
oobabooga
|
43b6ab8673
|
Store thumbnails as files instead of base64 strings
This improves the UI responsiveness for large histories.
|
2023-02-27 13:41:00 -03:00 |
|
oobabooga
|
f24b6e78a3
|
Fix clear history
|
2023-02-26 23:58:04 -03:00 |
|
oobabooga
|
8e3e8a070f
|
Make FlexGen work with the newest API
|
2023-02-26 16:53:41 -03:00 |
|
oobabooga
|
3333f94c30
|
Make the gallery extension work on colab
|
2023-02-26 12:37:26 -03:00 |
|
oobabooga
|
633a2b6be2
|
Don't regenerate/remove last message if the chat is empty
|
2023-02-26 00:43:12 -03:00 |
|
oobabooga
|
6e843a11d6
|
Fix FlexGen in chat mode
|
2023-02-26 00:36:04 -03:00 |
|
oobabooga
|
4548227fb5
|
Downgrade gradio version (file uploads are broken in 3.19.1)
|
2023-02-25 22:59:02 -03:00 |
|
oobabooga
|
9456c1d6ed
|
Prevent streaming with no_stream + generation attempts > 1
|
2023-02-25 17:45:03 -03:00 |
|
oobabooga
|
32f40f3b42
|
Bump gradio version to 3.19.1
|
2023-02-25 17:20:03 -03:00 |
|
oobabooga
|
fa58fd5559
|
Proper way to free the cuda cache
|
2023-02-25 15:50:29 -03:00 |
|
oobabooga
|
b585e382c0
|
Rename the custom prompt generator function
|
2023-02-25 15:13:14 -03:00 |
|
oobabooga
|
700311ce40
|
Empty the cuda cache at model.generate()
|
2023-02-25 14:39:13 -03:00 |
|
oobabooga
|
1878acd9f3
|
Minor bug fix in chat
|
2023-02-25 09:30:59 -03:00 |
|
oobabooga
|
e71ff959f5
|
Clean up some unused code
|
2023-02-25 09:23:02 -03:00 |
|
oobabooga
|
91f5852245
|
Move bot_picture.py inside the extension
|
2023-02-25 03:00:19 -03:00 |
|
oobabooga
|
5ac24b019e
|
Minor fix in the extensions implementation
|
2023-02-25 02:53:18 -03:00 |
|
oobabooga
|
85f914b9b9
|
Disable the hijack after using it
|
2023-02-25 02:36:01 -03:00 |
|
oobabooga
|
7e9f13e29f
|
Rename a variable
|
2023-02-25 01:55:32 -03:00 |
|
oobabooga
|
1741c36092
|
Minor fix
|
2023-02-25 01:47:25 -03:00 |
|
oobabooga
|
7c2babfe39
|
Rename greed to "generation attempts"
|
2023-02-25 01:42:19 -03:00 |
|
oobabooga
|
2dfb999bf1
|
Add greed parameter
|
2023-02-25 01:31:01 -03:00 |
|
oobabooga
|
13f2688134
|
Better way to generate custom prompts
|
2023-02-25 01:08:17 -03:00 |
|
oobabooga
|
67623a52b7
|
Allow for permanent hijacking
|
2023-02-25 00:55:19 -03:00 |
|
oobabooga
|
111b5d42e7
|
Add prompt hijack option for extensions
|
2023-02-25 00:49:18 -03:00 |
|
oobabooga
|
7a527a5581
|
Move "send picture" into an extension
I am not proud of how I did it for now.
|
2023-02-25 00:23:51 -03:00 |
|
oobabooga
|
e51ece21c0
|
Add ui() function to extensions
|
2023-02-24 19:00:11 -03:00 |
|
oobabooga
|
78ad55641b
|
Remove duplicate max_new_tokens parameter
|
2023-02-24 17:19:42 -03:00 |
|
oobabooga
|
65326b545a
|
Move all gradio elements to shared (so that extensions can use them)
|
2023-02-24 16:46:50 -03:00 |
|
oobabooga
|
0817fe1beb
|
Move code back into the chatbot wrapper
|
2023-02-24 14:10:32 -03:00 |
|
oobabooga
|
8a7563ae84
|
Reorder the imports
|
2023-02-24 12:42:43 -03:00 |
|
oobabooga
|
ace74a557a
|
Add some comments
|
2023-02-24 12:41:27 -03:00 |
|
oobabooga
|
fe5057f932
|
Simplify the extensions implementation
|
2023-02-24 10:01:21 -03:00 |
|
oobabooga
|
2fb6ae6970
|
Move chat preprocessing into a separate function
|
2023-02-24 09:40:48 -03:00 |
|
oobabooga
|
f6f792363b
|
Separate command-line params by spaces instead of commas
|
2023-02-24 08:55:09 -03:00 |
|
oobabooga
|
e260e84e5a
|
Merge branch 'max_memory' of https://github.com/elwolf6/text-generation-webui into elwolf6-max_memory
|
2023-02-24 08:47:01 -03:00 |
|
oobabooga
|
146f786c57
|
Reorganize a bit
|
2023-02-24 08:44:54 -03:00 |
|
oobabooga
|
c2f4c395b9
|
Clean up some chat functions
|
2023-02-24 08:31:30 -03:00 |
|
luis
|
5abdc99a7c
|
gpu-memory arg change
|
2023-02-23 18:43:55 -05:00 |
|
oobabooga
|
9ae063e42b
|
Fix softprompts when deepspeed is active (#112)
|
2023-02-23 20:22:47 -03:00 |
|
oobabooga
|
dac6fe0ff4
|
Reset the history if no default history exists on reload
|
2023-02-23 19:53:50 -03:00 |
|
oobabooga
|
3b8cecbab7
|
Reload the default chat on page refresh
|
2023-02-23 19:50:23 -03:00 |
|
oobabooga
|
f1914115d3
|
Fix minor issue with chat logs
|
2023-02-23 16:04:47 -03:00 |
|
oobabooga
|
b78561fba6
|
Minor bug fix
|
2023-02-23 15:26:41 -03:00 |
|
oobabooga
|
2e86a1ec04
|
Move chat history into shared module
|
2023-02-23 15:11:18 -03:00 |
|