oobabooga
|
831ac7ed3f
|
Add top_p
|
2023-03-01 16:45:48 -03:00 |
|
oobabooga
|
7c4d5ca8cc
|
Improve the text generation call a bit
|
2023-03-01 16:40:25 -03:00 |
|
oobabooga
|
0f6708c471
|
Sort the imports
|
2023-03-01 12:18:17 -03:00 |
|
oobabooga
|
e735806c51
|
Add a generate() function for RWKV
|
2023-03-01 12:16:11 -03:00 |
|
oobabooga
|
f871971de1
|
Trying to get the chat to work
|
2023-02-28 00:25:30 -03:00 |
|
oobabooga
|
ebd698905c
|
Add streaming to RWKV
|
2023-02-28 00:04:04 -03:00 |
|
oobabooga
|
70e522732c
|
Move RWKV loader into a separate file
|
2023-02-27 23:50:16 -03:00 |
|
oobabooga
|
ebc64a408c
|
RWKV support prototype
|
2023-02-27 23:03:35 -03:00 |
|
oobabooga
|
6e843a11d6
|
Fix FlexGen in chat mode
|
2023-02-26 00:36:04 -03:00 |
|
oobabooga
|
fa58fd5559
|
Proper way to free the cuda cache
|
2023-02-25 15:50:29 -03:00 |
|
oobabooga
|
700311ce40
|
Empty the cuda cache at model.generate()
|
2023-02-25 14:39:13 -03:00 |
|
oobabooga
|
78ad55641b
|
Remove duplicate max_new_tokens parameter
|
2023-02-24 17:19:42 -03:00 |
|
oobabooga
|
65326b545a
|
Move all gradio elements to shared (so that extensions can use them)
|
2023-02-24 16:46:50 -03:00 |
|
oobabooga
|
9ae063e42b
|
Fix softprompts when deepspeed is active (#112)
|
2023-02-23 20:22:47 -03:00 |
|
oobabooga
|
7224343a70
|
Improve the imports
|
2023-02-23 14:41:42 -03:00 |
|
oobabooga
|
1dacd34165
|
Further refactor
|
2023-02-23 13:28:30 -03:00 |
|
oobabooga
|
ce7feb3641
|
Further refactor
|
2023-02-23 13:03:52 -03:00 |
|