We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 463baaf commit 85cec45Copy full SHA for 85cec45
src/maxdiffusion/configs/base_wan_14b.yml
@@ -63,7 +63,7 @@ attention: 'flash' # Supported attention: dot_product, flash, cudnn_flash_te, ri
63
flash_min_seq_length: 0
64
65
# If mask_padding_tokens is True, we pass in segment ids to splash attention to avoid attending to padding tokens.
66
-# Else we do not pass in segment ids and on vpu bound hardware like (ironwood) this is faster.
+# Else we do not pass in segment ids and on vpu bound hardware like trillium this is faster.
67
# However, when padding tokens are significant, this will lead to worse quality and should be set to True.
68
mask_padding_tokens: True
69
# Maxdiffusion has 2 types of attention sharding strategies:
0 commit comments