We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 27225eb commit bee4f96Copy full SHA for bee4f96
1 file changed
src/maxdiffusion/models/ltx2/transformer_ltx2.py
@@ -624,9 +624,6 @@ def __init__(
624
qk_norm: str = "rms_norm_across_heads",
625
flash_block_sizes: BlockSizes = None,
626
flash_min_seq_length: int = 4096,
627
- video_gated_attn: bool = False,
628
- audio_gated_attn: bool = False,
629
- cross_attn_mod: bool = False,
630
**kwargs,
631
):
632
self.in_channels = in_channels
@@ -676,7 +673,7 @@ def __init__(
676
673
self.a2v_attention_kernel = a2v_attention_kernel
677
674
self.v2a_attention_kernel = v2a_attention_kernel
678
675
self.flash_min_seq_length = flash_min_seq_length
679
- self.video_gated_attn = video_gated_attn
+ self.video_gated_attn = gated_attn
680
self.audio_gated_attn = audio_gated_attn
681
self.cross_attn_mod = cross_attn_mod
682
0 commit comments