text-generation-webui/modules
2023-05-12 12:50:29 -03:00
..
callbacks.py Remove mutable defaults from function signature. (#1663) 2023-05-08 22:55:41 -03:00
chat.py Change how spaces are handled in continue/generation attempts 2023-05-12 12:50:29 -03:00
deepspeed_parameters.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
evaluate.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
extensions.py Fix custom_generate_chat_prompt (#1965) 2023-05-10 11:29:59 -03:00
GPTQ_loader.py Fix bug in --cpu-memory 2023-05-12 06:17:07 -03:00
html_generator.py Add markdown table rendering 2023-05-10 13:41:23 -03:00
llama_attn_hijack.py Better warning messages 2023-05-03 21:43:17 -03:00
llamacpp_model.py added no_mmap & mlock parameters to llama.cpp and removed llamacpp_model_alternative (#1649) 2023-05-02 18:25:28 -03:00
logging_colors.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
LoRA.py fixed LoRA loading issue (#1865) 2023-05-08 16:21:55 -03:00
models.py Fix bug in --cpu-memory 2023-05-12 06:17:07 -03:00
monkey_patch_gptq_lora.py Better warning messages 2023-05-03 21:43:17 -03:00
RWKV.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
shared.py Change how spaces are handled in continue/generation attempts 2023-05-12 12:50:29 -03:00
text_generation.py Better handle spaces in LlamaTokenizer 2023-05-11 17:55:50 -03:00
training.py Sort dropdowns numerically 2023-05-05 23:14:56 -03:00
ui.py Add support for custom chat styles (#1917) 2023-05-08 12:35:03 -03:00
utils.py Fix prompt order in the dropdown 2023-05-10 02:24:09 -03:00