text-generation-webui/modules
2023-06-21 14:15:08 -05:00
..
AutoGPTQ_loader.py AutoGPTQ: Add UI and command line support for disabling fused attention and fused MLP (#2648) 2023-06-15 23:59:54 -03:00
block_requests.py Add modules/block_requests.py 2023-06-18 16:31:14 -03:00
callbacks.py Add ExLlama support (#2444) 2023-06-16 20:35:38 -03:00
chat.py Add menus for saving presets/characters/instruction templates/prompts (#2621) 2023-06-11 12:19:18 -03:00
deepspeed_parameters.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
evaluate.py Reorganize model loading UI completely (#2720) 2023-06-16 19:00:37 -03:00
exllama_hf.py Implement a demo HF wrapper for exllama to utilize existing HF transformers decoding. (#2777) 2023-06-21 15:31:42 -03:00
exllama.py Update error message 2023-06-20 01:32:12 -05:00
extensions.py Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
github.py Reorganize model loading UI completely (#2720) 2023-06-16 19:00:37 -03:00
GPTQ_loader.py Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
html_generator.py Add markdown table rendering 2023-05-10 13:41:23 -03:00
llama_attn_hijack.py Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
llamacpp_model.py fix usage of self in classmethod (#2781) 2023-06-20 16:18:42 -03:00
loaders.py Implement a demo HF wrapper for exllama to utilize existing HF transformers decoding. (#2777) 2023-06-21 15:31:42 -03:00
logging_colors.py Add menus for saving presets/characters/instruction templates/prompts (#2621) 2023-06-11 12:19:18 -03:00
LoRA.py Add backwards compatibility 2023-06-19 23:38:12 -05:00
models_settings.py Add ExLlama support (#2444) 2023-06-16 20:35:38 -03:00
models.py Implement a demo HF wrapper for exllama to utilize existing HF transformers decoding. (#2777) 2023-06-21 15:31:42 -03:00
monkey_patch_gptq_lora.py Better warning messages 2023-05-03 21:43:17 -03:00
presets.py Reorganize model loading UI completely (#2720) 2023-06-16 19:00:37 -03:00
relative_imports.py Add ExLlama+LoRA support (#2756) 2023-06-19 12:31:24 -03:00
RWKV.py Add ExLlama support (#2444) 2023-06-16 20:35:38 -03:00
sampler_hijack.py Add Mirostat v2 sampling to transformer models (#2571) 2023-06-09 21:26:31 -03:00
shared.py Implement a demo HF wrapper for exllama to utilize existing HF transformers decoding. (#2777) 2023-06-21 15:31:42 -03:00
text_generation.py Implement a demo HF wrapper for exllama to utilize existing HF transformers decoding. (#2777) 2023-06-21 15:31:42 -03:00
training.py Added Falcon LoRA training support (#2684) 2023-06-20 01:03:44 -03:00
ui.py Add gpu_split param to ExLlama 2023-06-16 20:49:36 -03:00
utils.py Allow API requests to use parameter presets 2023-06-14 11:32:20 -03:00