Skip to content

Commit

Permalink
hotfix activation typo from #1212
Browse files Browse the repository at this point in the history
  • Loading branch information
Quentin-Anthony committed Sep 9, 2024
1 parent 836aefa commit 3169c84
Showing 1 changed file with 1 addition and 0 deletions.
1 change: 1 addition & 0 deletions megatron/model/transformer.py
Original file line number Diff line number Diff line change
Expand Up @@ -979,6 +979,7 @@ def __init__(
self.gpt_j_tied = neox_args.gpt_j_tied
self.mlp_type = neox_args.mlp_type
self.moe_type = neox_args.moe_type
self.activation = neox_args.activation

if self.gpt_j_residual:
# GPT-J style layers allow us to defer the reduction of results across TP ranks until the end of the two sublayers.
Expand Down

0 comments on commit 3169c84

Please sign in to comment.