Update training.py - correct use of lora_names (#2988)

This commit is contained in:
FartyPants 2023-07-03 16:41:18 -04:00 committed by GitHub
parent c23c88ee4c
commit 1f8cae14f9
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -283,7 +283,7 @@ def do_train(lora_name: str, always_override: bool, save_steps: int, micro_batch
else: else:
model_id = "llama" model_id = "llama"
if model_type == "PeftModelForCausalLM": if model_type == "PeftModelForCausalLM":
if len(shared.args.lora_names) > 0: if len(shared.lora_names) > 0:
yield "You are trying to train a LoRA while you already have another LoRA loaded. This will work, but may have unexpected effects. *(Will continue anyway in 5 seconds, press `Interrupt` to stop.)*" yield "You are trying to train a LoRA while you already have another LoRA loaded. This will work, but may have unexpected effects. *(Will continue anyway in 5 seconds, press `Interrupt` to stop.)*"
logger.warning("Training LoRA over top of another LoRA. May have unexpected effects.") logger.warning("Training LoRA over top of another LoRA. May have unexpected effects.")
else: else: