Skip to content

Commit

Permalink
cleanup
Browse files Browse the repository at this point in the history
  • Loading branch information
jahatef committed Nov 6, 2024
1 parent 48de682 commit c6fac96
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 2 deletions.
2 changes: 1 addition & 1 deletion configs/rwkv/430M.yml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@
"output_layer_parallelism": "column",
"norm": "rmsnorm",
"rms_norm_epsilon": 1.0e-5,
"train_micro_batch_size_per_gpu": 4,
"train_micro_batch_size_per_gpu": 1,

"attention_config": [[["rwkv"], 24]],

Expand Down
1 change: 0 additions & 1 deletion megatron/model/rwkv/v6/rwkv.py
Original file line number Diff line number Diff line change
Expand Up @@ -258,7 +258,6 @@ def forward(self, x):
B, T, C = x.size()
C_tp = C//mpu.get_model_parallel_world_size()
H = self.neox_args.num_attention_heads//mpu.get_model_parallel_world_size()
H_tp = H//mpu.get_model_parallel_world_size()

r, k, v, g, w = self.jit_func(x)

Expand Down

0 comments on commit c6fac96

Please sign in to comment.