Skip to content

Commit

Permalink
fix config
Browse files Browse the repository at this point in the history
  • Loading branch information
samsja committed Jul 23, 2024
1 parent 19814f6 commit fb60d7a
Show file tree
Hide file tree
Showing 2 changed files with 12 additions and 14 deletions.
21 changes: 9 additions & 12 deletions open_diloco/configs/config_150m.json
Original file line number Diff line number Diff line change
@@ -1,13 +1,10 @@
{
"architectures": [
"LlamaForCausalLM"
],
"model_type": "llama",
"hidden_size": 1024,
"intermediate_size": 2688,
"num_attention_heads": 16,
"num_hidden_layers": 12,
"use_cache": false,
"rms_norm_eps": 1e-05
}

{
"name": "llama150m",
"n_embd": 1024,
"intermediate_size": 4096,
"n_head": 16,
"n_layer": 12,
"vocab_size": 32000,
"block_size": 1024
}
5 changes: 3 additions & 2 deletions open_diloco/configs/config_1b.json
Original file line number Diff line number Diff line change
@@ -1,9 +1,10 @@
{
"name": "llama",
"name": "llama1b",
"n_embd": 2048,
"intermediate_size": 5632,
"n_head": 32,
"n_layer": 22,
"n_query_groups": 4,
"vocab_size": 1024
"vocab_size": 32000,
"block_size": 1024
}

0 comments on commit fb60d7a

Please sign in to comment.