Skip to content

Commit

Permalink
Update NeoXArgs docs automatically
Browse files Browse the repository at this point in the history
  • Loading branch information
github-actions committed May 10, 2024
1 parent ad8417c commit 3dccb2f
Showing 1 changed file with 44 additions and 3 deletions.
47 changes: 44 additions & 3 deletions configs/neox_arguments.md
Original file line number Diff line number Diff line change
Expand Up @@ -111,7 +111,7 @@ Logging Arguments

- **git_hash**: str

Default = 6fb840e
Default = ad8417c

current git hash of repository

Expand Down Expand Up @@ -1201,7 +1201,7 @@ Text Generation arguments
- **num_experts**: int
- **moe_num_experts**: int
Default = 1
Expand Down Expand Up @@ -1243,7 +1243,7 @@ Text Generation arguments
- **moe_token_dropping**: bool
Default = True
Default = False
Whether to drop tokens when exceeding capacity
Expand Down Expand Up @@ -1273,6 +1273,47 @@ Text Generation arguments
- **moe_type**: str
Default = megablocks
Either `deepspeed` or `megablocks`
- **moe_glu**: bool
Default = False
Use gated linear units in MoE
- **moe_lbl_in_fp32**: bool
Default = False
Whether to compute the load balancing loss in fp32.
- **moe_jitter_eps**: float
Default = None
Coefficient for MoE routing jitter. Jitter is
not used if set to None
- **enable_expert_tensor_parallelism**: bool
Default = False
Enable expert tensor parallelism
## NeoXArgsTokenizer
Tokenizer Arguments
Expand Down

0 comments on commit 3dccb2f

Please sign in to comment.