diff --git a/modules/models.py b/modules/models.py index 6c38c3c7..5929e868 100644 --- a/modules/models.py +++ b/modules/models.py @@ -100,9 +100,9 @@ def load_model(model_name, loader=None): elif loader in ['llama.cpp', 'llamacpp_HF', 'ctransformers']: shared.settings['truncation_length'] = shared.args.n_ctx - logger.info(f"LOADER: {loader}") + logger.info(f"LOADER: \"{loader}\"") logger.info(f"TRUNCATION LENGTH: {shared.settings['truncation_length']}") - logger.info(f"INSTRUCTION TEMPLATE: {metadata['instruction_template']}") + logger.info(f"INSTRUCTION TEMPLATE: \"{metadata['instruction_template']}\"") logger.info(f"Loaded the model in {(time.time()-t0):.2f} seconds.") return model, tokenizer