We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent be62d37 commit 5498223Copy full SHA for 5498223
src/maxdiffusion/configs/base14.yml
@@ -50,6 +50,8 @@ jit_initializers: True
50
from_pt: False
51
split_head_dim: True
52
attention: 'dot_product' # Supported attention: dot_product, flash
53
+mask_padding_tokens: True # Whether to mask padding tokens in attention computation.
54
+attention_sharding_uniform: True # same sequence sharding rules applied for q in both (self and cross attention)
55
flash_block_sizes: {}
56
# GroupNorm groups
57
norm_num_groups: 32
0 commit comments