text-generation-webui/modules
2023-12-17 21:05:10 -08:00
..
grammar
AutoGPTQ_loader.py
block_requests.py
callbacks.py
chat.py Instruction templates: better handle unwanted bos tokens 2023-12-17 21:04:30 -08:00
ctransformers_model.py
deepspeed_parameters.py
evaluate.py
exllama_hf.py
exllama.py
exllamav2_hf.py Add --num_experts_per_token parameter (ExLlamav2) (#4955) 2023-12-17 12:08:33 -03:00
exllamav2.py Add --num_experts_per_token parameter (ExLlamav2) (#4955) 2023-12-17 12:08:33 -03:00
extensions.py
github.py
GPTQ_loader.py
html_generator.py
llama_attn_hijack.py
llamacpp_hf.py
llamacpp_model.py
loaders.py Add --num_experts_per_token parameter (ExLlamav2) (#4955) 2023-12-17 12:08:33 -03:00
logging_colors.py
logits.py
LoRA.py
metadata_gguf.py llama.cpp: read instruction template from GGUF metadata (#4975) 2023-12-18 01:51:58 -03:00
models_settings.py llama.cpp: read instruction template from GGUF metadata (#4975) 2023-12-18 01:51:58 -03:00
models.py
monkey_patch_gptq_lora.py
one_click_installer_check.py
presets.py
prompts.py
relative_imports.py
RoPE.py
RWKV.py
sampler_hijack.py
shared.py Add --num_experts_per_token parameter (ExLlamav2) (#4955) 2023-12-17 12:08:33 -03:00
text_generation.py
training.py
ui_chat.py
ui_default.py
ui_file_saving.py
ui_model_menu.py Add --num_experts_per_token parameter (ExLlamav2) (#4955) 2023-12-17 12:08:33 -03:00
ui_notebook.py
ui_parameters.py
ui_session.py
ui.py Add --num_experts_per_token parameter (ExLlamav2) (#4955) 2023-12-17 12:08:33 -03:00
utils.py