Skip to content

Commit 292fd84

Browse files
committed
Add ulysses to supported attention comment in WAN configs
1 parent faea21a commit 292fd84

5 files changed

Lines changed: 5 additions & 5 deletions

File tree

src/maxdiffusion/configs/base_wan_14b.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ jit_initializers: True
6060
# Set true to load weights from pytorch
6161
from_pt: True
6262
split_head_dim: True
63-
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring
63+
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring, ulysses
6464
flash_min_seq_length: 0
6565

6666
# If mask_padding_tokens is True, we pass in segment ids to splash attention to avoid attending to padding tokens.

src/maxdiffusion/configs/base_wan_1_3b.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ jit_initializers: True
6060
# Set true to load weights from pytorch
6161
from_pt: True
6262
split_head_dim: True
63-
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring
63+
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring, ulysses
6464
flash_min_seq_length: 0
6565

6666
# If mask_padding_tokens is True, we pass in segment ids to splash attention to avoid attending to padding tokens.

src/maxdiffusion/configs/base_wan_27b.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ jit_initializers: True
6060
# Set true to load weights from pytorch
6161
from_pt: True
6262
split_head_dim: True
63-
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring
63+
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring, ulysses
6464
flash_min_seq_length: 4096
6565
# If mask_padding_tokens is True, we pass in segment ids to splash attention to avoid attending to padding tokens.
6666
# Else we do not pass in segment ids and on vpu bound hardware like trillium this is faster.

src/maxdiffusion/configs/base_wan_i2v_14b.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ jit_initializers: True
6060
# Set true to load weights from pytorch
6161
from_pt: True
6262
split_head_dim: True
63-
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring
63+
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring, ulysses
6464
flash_min_seq_length: 4096
6565
dropout: 0.0
6666

src/maxdiffusion/configs/base_wan_i2v_27b.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ jit_initializers: True
6060
# Set true to load weights from pytorch
6161
from_pt: True
6262
split_head_dim: True
63-
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring
63+
attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ring, ulysses
6464
flash_min_seq_length: 4096
6565
dropout: 0.0
6666

0 commit comments

Comments
 (0)