Skip to content

Commit

Permalink
push from ssd-1
Browse files Browse the repository at this point in the history
  • Loading branch information
haileyschoelkopf committed Nov 28, 2023
1 parent 1dd4054 commit 4158a90
Show file tree
Hide file tree
Showing 2 changed files with 18 additions and 8 deletions.
18 changes: 14 additions & 4 deletions configs/pythia/70M.yml
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@
"gpt_j_residual": true,
"output_layer_parallelism": "column",

"attention_config": [[["flash"], 6]],
# "attention_config": [[["flash"], 6]],

"scaled_upper_triang_masked_softmax_fusion": true,
"bias_gelu_fusion": true,
Expand Down Expand Up @@ -42,7 +42,7 @@
"cpu_offload": false
},

"train_micro_batch_size_per_gpu": 32,
"train_micro_batch_size_per_gpu": 4,
"gas": 1,
"data_impl": "mmap",
"num_workers": 1,
Expand Down Expand Up @@ -73,13 +73,23 @@
"lr_decay_style": "cosine",
"warmup": 0.01,
"checkpoint_factor": 1000,
"extra_save_iters": [0,1,2,4,8,16,32,64,128,256,512],
# "extra_save_iters": [0,1,2,4,8,16,32,64,128,256,512],
"eval_interval": 100000,
"eval_iters": 10,

"log_interval": 10,
"steps_per_print": 10,
"wall_clock_breakdown": true,

"tokenizer_type": "HFTokenizer"
"tokenizer_type": "HFTokenizer",
"vocab_file": "/mnt/ssd-1/hailey/pythia/utils/20B_tokenizer.json",

"train_data_paths": ["/mnt/ssd-1/hailey/conditional-training/gpt-neox/enwik8/enwik8/enwik8_text_document"],
"valid_data_paths": ["/mnt/ssd-1/hailey/conditional-training/gpt-neox/enwik8/enwik8/enwik8_text_document"],
"test_data_paths": ["/mnt/ssd-1/hailey/conditional-training/gpt-neox/enwik8/enwik8/enwik8_text_document"],

"force_multi": false,


"conditional_training": false
}
8 changes: 4 additions & 4 deletions eval_tasks/eval_adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,18 +13,18 @@
# limitations under the License.

from megatron.utils import is_local_main, print_rank_0
import best_download
# import best_download

# patch best_download (eval harness downloader) to only happen on the first local rank
fn = best_download.download_file
# fn = best_download.download_file


def _download_file(*args, **kwargs):
if is_local_main():
fn(*args, **kwargs)
pass # fn(*args, **kwargs)


best_download.download_file = _download_file
# best_download.download_file = _download_file

import os
import sys
Expand Down

0 comments on commit 4158a90

Please sign in to comment.