Skip to content

Commit bb277ad

Browse files
committed
Fix
1 parent f588222 commit bb277ad

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

src/maxdiffusion/configs/base_wan_lora_14b.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -295,7 +295,7 @@ lora_rank: 64
295295
# Values are lists to support multiple LoRA loading during inference in the future.
296296
lora_config: {
297297
lora_model_name_or_path: ["lightx2v/Wan2.1-Distill-Loras"],
298-
weight_name: ["Wan2.1_t2v_14b_lora_rank64_lightx2v_4step.safetensors"],
298+
weight_name: ["wan2.1_t2v_14b_lora_rank64_lightx2v_4step.safetensors"],
299299
adapter_name: ["wan21-distill-lora"],
300300
scale: [1.0],
301301
from_pt: []

0 commit comments

Comments
 (0)