Skip to content

Commit

Permalink
Use causal_padding instead of padding
Browse files Browse the repository at this point in the history
Signed-off-by: Reese Wang <[email protected]>
  • Loading branch information
zlsh80826 committed Mar 11, 2024
1 parent e3e785c commit d2ba1bc
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion praxis/contrib/gpu/scripts_gpu/te_helper.py
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ def update_attn_te_tpl(te_tpl, attn_tpl):
assert (transformer_layer_tpl.tr_fflayer_tpl.has_bias ==
transformer_layer_tpl.tr_atten_tpl.use_bias), "TE only allows same bias settings."
te_transformer_tpl.use_bias = transformer_layer_tpl.tr_fflayer_tpl.has_bias
te_transformer_tpl.self_attn_mask_type = 'causal' \
te_transformer_tpl.self_attn_mask_type = 'causal_padding' \
if stacked_transformer_obj.mask_self_attention else 'padding'

te_transformer_tpl.logical_axes_rules = te_flax.extend_logical_axis_rules(tuple())
Expand Down

0 comments on commit d2ba1bc

Please sign in to comment.