text-generation-webui/modules
oobabooga b6c407f51d Don't stream at more than 24 fps
This is a performance optimization
2023-05-31 23:41:42 -03:00
..
AutoGPTQ_loader.py Falcon support (trust-remote-code and autogptq checkboxes) (#2367) 2023-05-29 10:20:18 -03:00
callbacks.py Remove mutable defaults from function signature. (#1663) 2023-05-08 22:55:41 -03:00
chat.py Use YAML for presets and settings 2023-05-28 22:34:12 -03:00
deepspeed_parameters.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
evaluate.py Minor fix 2023-05-29 13:31:17 -03:00
extensions.py Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
GPTQ_loader.py Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
html_generator.py Add markdown table rendering 2023-05-10 13:41:23 -03:00
llama_attn_hijack.py Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
llamacpp_model.py Make llama.cpp read prompt size and seed from settings (#2299) 2023-05-25 10:29:31 -03:00
logging_colors.py Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
LoRA.py Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
models.py Change a warning message 2023-05-29 22:40:54 -03:00
monkey_patch_gptq_lora.py Better warning messages 2023-05-03 21:43:17 -03:00
RWKV.py Style improvements (#1957) 2023-05-09 22:49:39 -03:00
sampler_hijack.py Add tail-free and top-a sampling (#2357) 2023-05-29 21:40:01 -03:00
shared.py Falcon support (trust-remote-code and autogptq checkboxes) (#2367) 2023-05-29 10:20:18 -03:00
text_generation.py Don't stream at more than 24 fps 2023-05-31 23:41:42 -03:00
training.py Fix warning for qlora (#2438) 2023-05-30 11:09:18 -03:00
ui.py Add tail-free and top-a sampling (#2357) 2023-05-29 21:40:01 -03:00
utils.py Use YAML for presets and settings 2023-05-28 22:34:12 -03:00