Commit Graph

1479 Commits

Author SHA1 Message Date
nclok1405
b9e2ded6d4
Added UnicodeDecodeError workaround for modules/llamacpp_model.py (#6040)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2025-01-08 21:17:31 -03:00
oobabooga
91a8a87887 Remove obsolete code 2025-01-08 15:07:21 -08:00
oobabooga
7157257c3f
Remove the AutoGPTQ loader (#6641) 2025-01-08 19:28:56 -03:00
oobabooga
c0f600c887 Add a --torch-compile flag for transformers 2025-01-05 05:47:00 -08:00
oobabooga
11af199aff Add a "Static KV cache" option for transformers 2025-01-04 17:52:57 -08:00
oobabooga
3967520e71 Connect XTC, DRY, smoothing_factor, and dynatemp to ExLlamaV2 loader (non-HF) 2025-01-04 16:25:06 -08:00
oobabooga
049297fa66 UI: reduce the size of CSS sent to the UI during streaming 2025-01-04 14:09:36 -08:00
oobabooga
0e673a7a42 UI: reduce the size of HTML sent to the UI during streaming 2025-01-04 11:40:24 -08:00
mamei16
9f24885bd2
Sane handling of markdown lists (#6626) 2025-01-04 15:41:31 -03:00
oobabooga
4b3e1b3757 UI: add a "Search chats" input field 2025-01-02 18:46:40 -08:00
oobabooga
b8fc9010fa UI: fix orjson.JSONDecodeError error on page reload 2025-01-02 16:57:04 -08:00
oobabooga
75f1b5ccde UI: add a "Branch chat" button 2025-01-02 16:24:18 -08:00
Petr Korolev
13c033c745
Fix CUDA error on MPS backend during API request (#6572)
---------

Co-authored-by: oobabooga <oobabooga4@gmail.com>
2025-01-02 00:06:11 -03:00
oobabooga
725639118a UI: Use a tab length of 2 for lists (rather than 4) 2025-01-01 13:53:50 -08:00
oobabooga
7b88724711
Make responses start faster by removing unnecessary cleanup calls (#6625) 2025-01-01 18:33:38 -03:00
oobabooga
64853f8509 Reapply a necessary change that I removed from #6599 (thanks @mamei16!) 2024-12-31 14:43:22 -08:00
mamei16
e953af85cd
Fix newlines in the markdown renderer (#6599)
---------

Co-authored-by: oobabooga <oobabooga4@gmail.com>
2024-12-31 01:04:02 -03:00
oobabooga
39a5c9a49c
UI organization (#6618) 2024-12-29 11:16:17 -03:00
oobabooga
0490ee620a UI: increase the threshold for a <li> to be considered long (some more) 2024-12-19 16:51:34 -08:00
oobabooga
89888bef56 UI: increase the threshold for a <li> to be considered long 2024-12-19 14:38:36 -08:00
oobabooga
2acec386fc UI: improve the streaming cursor 2024-12-19 14:08:56 -08:00
oobabooga
e2fb86e5df UI: further improve the style of lists and headings 2024-12-19 13:59:24 -08:00
oobabooga
c48e4622e8 UI: update a link 2024-12-18 06:28:14 -08:00
oobabooga
b27f6f8915 Lint 2024-12-17 20:13:32 -08:00
oobabooga
b051e2c161 UI: improve a margin for readability 2024-12-17 19:58:21 -08:00
oobabooga
60c93e0c66 UI: Set cache_type to fp16 by default 2024-12-17 19:44:20 -08:00
oobabooga
ddccc0d657 UI: minor change to log messages 2024-12-17 19:39:00 -08:00
oobabooga
3030c79e8c UI: show progress while loading a model 2024-12-17 19:37:43 -08:00
Diner Burger
addad3c63e
Allow more granular KV cache settings (#6561) 2024-12-17 17:43:48 -03:00
oobabooga
c43ee5db11 UI: very minor color change 2024-12-17 07:59:55 -08:00
oobabooga
d769618591
Improved UI (#6575) 2024-12-17 00:47:41 -03:00
oobabooga
350758f81c UI: Fix the history upload event 2024-11-19 20:34:53 -08:00
oobabooga
d01293861b Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-11-18 10:15:36 -08:00
oobabooga
3d19746a5d UI: improve HTML rendering for lists with sub-lists 2024-11-18 10:14:09 -08:00
mefich
1c937dad72
Filter whitespaces in downloader fields in model tab (#6518) 2024-11-18 12:01:40 -03:00
PIRI
e1061ba7e3
Make token bans work again on HF loaders (#6488) 2024-10-24 15:24:02 -03:00
oobabooga
2468cfd8bb Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-10-14 13:25:27 -07:00
oobabooga
bb62e796eb Fix locally compiled llama-cpp-python failing to import 2024-10-14 13:24:13 -07:00
oobabooga
c9a9f63d1b Fix llama.cpp loader not being random (thanks @reydeljuego12345) 2024-10-14 13:07:07 -07:00
PIRI
03a2e70054
Fix temperature_last when temperature not in sampler priority (#6439) 2024-10-09 11:25:14 -03:00
oobabooga
49dfa0adaf Fix the "save preset" event 2024-10-01 11:20:48 -07:00
oobabooga
93c250b9b6 Add a UI element for enable_tp 2024-10-01 11:16:15 -07:00
oobabooga
cca9d6e22d Lint 2024-10-01 10:21:06 -07:00
oobabooga
4d9ce586d3 Update llama_cpp_python_hijack.py, fix llamacpp_hf 2024-09-30 14:49:21 -07:00
oobabooga
bbdeed3cf4 Make sampler priority high if unspecified 2024-09-29 20:45:27 -07:00
Manuel Schmid
0f90a1b50f
Do not set value for histories in chat when --multi-user is used (#6317) 2024-09-29 01:08:55 -03:00
oobabooga
c61b29b9ce Simplify the warning when flash-attn fails to import 2024-09-28 20:33:17 -07:00
oobabooga
b92d7fd43e Add warnings for when AutoGPTQ, TensorRT-LLM, or HQQ are missing 2024-09-28 20:30:24 -07:00
oobabooga
7276dca933 Fix a typo 2024-09-27 20:28:17 -07:00
RandoInternetPreson
46996f6519
ExllamaV2 tensor parallelism to increase multi gpu inference speeds (#6356) 2024-09-28 00:26:03 -03:00