text-generation-webui/modules
2023-05-14 22:58:11 -03:00
..
callbacks.py Remove mutable defaults from function signature. (#1663) 2023-05-08 22:55:41 -03:00
chat.py Add chat-instruct mode (#2049) 2023-05-14 10:43:55 -03:00
deepspeed_parameters.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
evaluate.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
extensions.py Fix custom_generate_chat_prompt (#1965) 2023-05-10 11:29:59 -03:00
GPTQ_loader.py Fix bug in --cpu-memory 2023-05-12 06:17:07 -03:00
html_generator.py Add markdown table rendering 2023-05-10 13:41:23 -03:00
llama_attn_hijack.py Better warning messages 2023-05-03 21:43:17 -03:00
llamacpp_model.py Add llama.cpp GPU offload option (#2060) 2023-05-14 22:58:11 -03:00
logging_colors.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
LoRA.py fixed LoRA loading issue (#1865) 2023-05-08 16:21:55 -03:00
models.py Fix bug in --cpu-memory 2023-05-12 06:17:07 -03:00
monkey_patch_gptq_lora.py Better warning messages 2023-05-03 21:43:17 -03:00
RWKV.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
shared.py Add llama.cpp GPU offload option (#2060) 2023-05-14 22:58:11 -03:00
text_generation.py Better handle spaces in LlamaTokenizer 2023-05-11 17:55:50 -03:00
training.py Sort dropdowns numerically 2023-05-05 23:14:56 -03:00
ui.py Add chat-instruct mode (#2049) 2023-05-14 10:43:55 -03:00
utils.py Fix prompt order in the dropdown 2023-05-10 02:24:09 -03:00