text-generation-webui/modules
2023-05-17 10:41:09 -03:00
..
callbacks.py Remove mutable defaults from function signature. (#1663) 2023-05-08 22:55:41 -03:00
chat.py Add chat-instruct mode (#2049) 2023-05-14 10:43:55 -03:00
deepspeed_parameters.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
evaluate.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
extensions.py Allow extensions to define a new tab 2023-05-17 01:31:56 -03:00
GPTQ_loader.py Experimental jank multiGPU inference that's 2x faster than native somehow (#2100) 2023-05-17 10:41:09 -03:00
html_generator.py Add markdown table rendering 2023-05-10 13:41:23 -03:00
llama_attn_hijack.py Better warning messages 2023-05-03 21:43:17 -03:00
llamacpp_model.py Add in-memory cache support for llama.cpp (#1936) 2023-05-15 20:19:55 -03:00
logging_colors.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
LoRA.py fixed LoRA loading issue (#1865) 2023-05-08 16:21:55 -03:00
models.py Refactor models.py (#2113) 2023-05-16 19:52:22 -03:00
monkey_patch_gptq_lora.py Better warning messages 2023-05-03 21:43:17 -03:00
RWKV.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
shared.py Experimental jank multiGPU inference that's 2x faster than native somehow (#2100) 2023-05-17 10:41:09 -03:00
text_generation.py Better handle spaces in LlamaTokenizer 2023-05-11 17:55:50 -03:00
training.py Fix Training fails when evaluation dataset is selected (#2099) 2023-05-16 13:40:19 -03:00
ui.py Fix no-mmap bug 2023-05-16 17:35:49 -03:00
utils.py Fix prompt order in the dropdown 2023-05-10 02:24:09 -03:00