micsthepick
|
3708de2b1f
|
respect model dir for downloads (#3077) (#3079)
|
2023-07-11 18:55:46 -03:00 |
|
matatonic
|
3778816b8d
|
models/config.yaml: +platypus/gplatty, +longchat, +vicuna-33b, +Redmond-Hermes-Coder, +wizardcoder, +more (#2928)
* +platypus/gplatty
* +longchat, +vicuna-33b, +Redmond-Hermes-Coder
* +wizardcoder
* +superplatty
* +Godzilla, +WizardLM-V1.1, +rwkv 8k,
+wizard-mega fix </s>
---------
Co-authored-by: Matthew Ashton <mashton-gitlab@zhero.org>
|
2023-07-11 18:53:48 -03:00 |
|
Ricardo Pinto
|
3e9da5a27c
|
Changed FormComponent to IOComponent (#3017)
Co-authored-by: Ricardo Pinto <1-ricardo.pinto@users.noreply.gitlab.cognitage.com>
|
2023-07-11 18:52:16 -03:00 |
|
matatonic
|
3e7feb699c
|
extensions/openai: Major openai extension updates & fixes (#3049)
* many openai updates
* total reorg & cleanup.
* fixups
* missing import os for images
* +moderations, custom_stopping_strings, more fixes
* fix bugs in completion streaming
* moderation fix (flagged)
* updated moderation categories
---------
Co-authored-by: Matthew Ashton <mashton-gitlab@zhero.org>
|
2023-07-11 18:50:08 -03:00 |
|
Ahmad Fahadh Ilyas
|
8db7e857b1
|
Add token authorization for downloading model (#3067)
|
2023-07-11 18:48:08 -03:00 |
|
FartyPants
|
61102899cd
|
google flan T5 download fix (#3080)
|
2023-07-11 18:46:59 -03:00 |
|
jllllll
|
fdd596f98f
|
Bump bitsandbytes Windows wheel (#3097)
|
2023-07-11 18:41:24 -03:00 |
|
Vadim Peretokin
|
987d522b55
|
Fix API example for loading models (#3101)
|
2023-07-11 18:40:55 -03:00 |
|
Josh XT
|
f4aa11cef6
|
Add default environment variable values to docker compose file (#3102)
Add default environment variable values to docker compose file
|
2023-07-11 18:38:26 -03:00 |
|
ofirkris
|
a81cdd1367
|
Bump cpp llama version (#3081)
Bump cpp llama version to 0.1.70
|
2023-07-10 19:36:15 -03:00 |
|
jllllll
|
f8dbd7519b
|
Bump exllama module version (#3087)
d769533b6f...e61d4d31d4
|
2023-07-10 19:35:59 -03:00 |
|
tianchen zhong
|
c7058afb40
|
Add new possible bin file name regex (#3070)
|
2023-07-09 17:22:56 -03:00 |
|
ofirkris
|
161d984e80
|
Bump llama-cpp-python version (#3072)
Bump llama-cpp-python version to 0.1.69
|
2023-07-09 17:22:24 -03:00 |
|
Salvador E. Tropea
|
463aac2d65
|
[Added] google_translate activate param (#2961)
- So you can quickly enable/disable it, otherwise you must select
English to disable it, and then your language to enable it again.
|
2023-07-09 01:08:20 -03:00 |
|
Forkoz
|
74ea7522a0
|
Lora fixes for AutoGPTQ (#2818)
|
2023-07-09 01:03:43 -03:00 |
|
Chris Rude
|
70b088843d
|
fix for issue #2475: Streaming api deadlock (#3048)
|
2023-07-08 23:21:20 -03:00 |
|
oobabooga
|
5ac4e4da8b
|
Make --model work with argument like models/folder_name
|
2023-07-08 10:22:54 -07:00 |
|
Brandon McClure
|
acf24ebb49
|
Whisper_stt params for model, language, and auto_submit (#3031)
|
2023-07-07 20:54:53 -03:00 |
|
oobabooga
|
79679b3cfd
|
Pin fastapi version (for #3042)
|
2023-07-07 16:40:57 -07:00 |
|
oobabooga
|
bb79037ebd
|
Fix wrong pytorch version on Linux+CPU
It was installing nvidia wheels
|
2023-07-07 20:40:31 -03:00 |
|
oobabooga
|
564a8c507f
|
Don't launch chat mode by default
|
2023-07-07 13:32:11 -03:00 |
|
oobabooga
|
b6643e5039
|
Add decode functions to llama.cpp/exllama
|
2023-07-07 09:11:30 -07:00 |
|
oobabooga
|
1ba2e88551
|
Add truncation to exllama
|
2023-07-07 09:09:23 -07:00 |
|
oobabooga
|
c21b73ff37
|
Minor change to ui.py
|
2023-07-07 09:09:14 -07:00 |
|
oobabooga
|
de994331a4
|
Merge remote-tracking branch 'refs/remotes/origin/main'
|
2023-07-06 22:25:43 -07:00 |
|
oobabooga
|
9aee1064a3
|
Block a cloudfare request
|
2023-07-06 22:24:52 -07:00 |
|
Fernando Tarin Morales
|
d7e14e1f78
|
Fixed the param name when loading a LoRA using a model loaded in 4 or 8 bits (#3036)
|
2023-07-07 02:24:07 -03:00 |
|
Fernando Tarin Morales
|
1f540fa4f8
|
Added the format to be able to finetune Vicuna1.1 models (#3037)
|
2023-07-07 02:22:39 -03:00 |
|
Xiaojian "JJ" Deng
|
ff45317032
|
Update models.py (#3020)
Hopefully fixed error with "ValueError: Tokenizer class GPTNeoXTokenizer does not exist or is not currently
imported."
|
2023-07-05 21:40:43 -03:00 |
|
ofirkris
|
b67c362735
|
Bump llama-cpp-python (#3011)
Bump llama-cpp-python to V0.1.68
|
2023-07-05 11:33:28 -03:00 |
|
jeckyhl
|
88a747b5b9
|
fix: Error when downloading model from UI (#3014)
|
2023-07-05 11:27:29 -03:00 |
|
oobabooga
|
e0a50fb77a
|
Merge pull request #2922 from Honkware/main
Load Salesforce Xgen Models
|
2023-07-04 23:47:21 -03:00 |
|
oobabooga
|
8705eba830
|
Remove universal llama tokenizer support
Instead replace it with a warning if the tokenizer files look off
|
2023-07-04 19:43:19 -07:00 |
|
oobabooga
|
84d6c93d0d
|
Merge branch 'main' into Honkware-main
|
2023-07-04 18:50:07 -07:00 |
|
oobabooga
|
31c297d7e0
|
Various changes
|
2023-07-04 18:50:01 -07:00 |
|
AN Long
|
be4582be40
|
Support specify retry times in download-model.py (#2908)
|
2023-07-04 22:26:30 -03:00 |
|
oobabooga
|
70a4d5dbcf
|
Update chat API (fixes #3006)
|
2023-07-04 17:36:47 -07:00 |
|
oobabooga
|
333075e726
|
Fix #3003
|
2023-07-04 11:38:35 -03:00 |
|
oobabooga
|
40c5722499
|
Fix #2998
|
2023-07-04 11:35:25 -03:00 |
|
oobabooga
|
463ddfffd0
|
Fix start_with
|
2023-07-03 23:32:02 -07:00 |
|
oobabooga
|
55457549cd
|
Add information about presets to the UI
|
2023-07-03 22:39:01 -07:00 |
|
oobabooga
|
373555c4fb
|
Fix loading some histories (thanks kaiokendev)
|
2023-07-03 22:19:28 -07:00 |
|
Panchovix
|
10c8c197bf
|
Add Support for Static NTK RoPE scaling for exllama/exllama_hf (#2955)
|
2023-07-04 01:13:16 -03:00 |
|
jllllll
|
1610d5ffb2
|
Bump exllama module to 0.0.5 (#2993)
|
2023-07-04 00:15:55 -03:00 |
|
FartyPants
|
eb6112d5a2
|
Update server.py - clear LORA after reload (#2952)
|
2023-07-04 00:13:38 -03:00 |
|
oobabooga
|
7e8340b14d
|
Make greetings appear in --multi-user mode
|
2023-07-03 20:08:14 -07:00 |
|
oobabooga
|
4b1804a438
|
Implement sessions + add basic multi-user support (#2991)
|
2023-07-04 00:03:30 -03:00 |
|
FartyPants
|
1f8cae14f9
|
Update training.py - correct use of lora_names (#2988)
|
2023-07-03 17:41:18 -03:00 |
|
FartyPants
|
c23c88ee4c
|
Update LoRA.py - avoid potential error (#2953)
|
2023-07-03 17:40:22 -03:00 |
|
FartyPants
|
33f56fd41d
|
Update models.py to clear LORA names after unload (#2951)
|
2023-07-03 17:39:06 -03:00 |
|