peacock-data-public-datasets-idc-config_toyds
/
bigscience
/experiments
/gpt2-meg
/meg_gpt2_perf_n16.out
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
***************************************** | |
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. | |
***************************************** | |
using world size: 64, data-parallel-size: 1, tensor-model-parallel size: 4, pipeline-model-parallel size: 16 | |
using torch.float16 for parameters ... | |
------------------------ arguments ------------------------ | |
accumulate_allreduce_grads_in_fp32 .............. False | |
adam_beta1 ...................................... 0.9 | |
adam_beta2 ...................................... 0.999 | |
adam_eps ........................................ 1e-08 | |
adlr_autoresume ................................. False | |
adlr_autoresume_interval ........................ 1000 | |
apply_query_key_layer_scaling ................... True | |
apply_residual_connection_post_layernorm ........ False | |
attention_dropout ............................... 0.1 | |
attention_softmax_in_fp32 ....................... False | |
bert_binary_head ................................ True | |
bert_load ....................................... None | |
bf16 ............................................ False | |
bias_dropout_fusion ............................. True | |
bias_gelu_fusion ................................ True | |
biencoder_projection_dim ........................ 0 | |
biencoder_shared_query_context_model ............ False | |
block_data_path ................................. None | |
checkpoint_activations .......................... True | |
checkpoint_num_layers ........................... 1 | |
clip_grad ....................................... 1.0 | |
consumed_train_samples .......................... 0 | |
consumed_valid_samples .......................... 0 | |
data_impl ....................................... mmap | |
data_parallel_size .............................. 1 | |
data_path ....................................... ['/gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document'] | |
dataloader_type ................................. single | |
DDP_impl ........................................ local | |
decoder_seq_length .............................. None | |
distribute_checkpointed_activations ............. False | |
distributed_backend ............................. nccl | |
embedding_path .................................. None | |
encoder_seq_length .............................. 1024 | |
eod_mask_loss ................................... False | |
eval_interval ................................... 100 | |
eval_iters ...................................... 10 | |
evidence_data_path .............................. None | |
exit_duration_in_mins ........................... None | |
exit_interval ................................... None | |
ffn_hidden_size ................................. 32768 | |
finetune ........................................ False | |
fp16 ............................................ True | |
fp16_lm_cross_entropy ........................... False | |
fp32_residual_connection ........................ False | |
global_batch_size ............................... 1024 | |
hidden_dropout .................................. 0.1 | |
hidden_size ..................................... 8192 | |
hysteresis ...................................... 2 | |
ict_head_size ................................... None | |
ict_load ........................................ None | |
img_dim ......................................... 224 | |
indexer_batch_size .............................. 128 | |
indexer_log_interval ............................ 1000 | |
init_method_std ................................. 0.02 | |
init_method_xavier_uniform ...................... False | |
initial_loss_scale .............................. 4294967296 | |
kv_channels ..................................... 256 | |
layernorm_epsilon ............................... 1e-05 | |
lazy_mpu_init ................................... None | |
load ............................................ /gpfsscratch/rech/eha/commun/checkpoints/gpt2-1-node | |
local_rank ...................................... 0 | |
log_batch_size_to_tensorboard ................... False | |
log_interval .................................... 1 | |
log_learning_rate_to_tensorboard ................ True | |
log_loss_scale_to_tensorboard ................... True | |
log_num_zeros_in_grad ........................... False | |
log_params_norm ................................. False | |
log_timers_to_tensorboard ....................... False | |
log_validation_ppl_to_tensorboard ............... False | |
loss_scale ...................................... None | |
loss_scale_window ............................... 1000 | |
lr .............................................. 0.00015 | |
lr_decay_iters .................................. 800 | |
lr_decay_samples ................................ None | |
lr_decay_style .................................. cosine | |
lr_warmup_fraction .............................. 0.01 | |
lr_warmup_iters ................................. 0 | |
lr_warmup_samples ............................... 0 | |
make_vocab_size_divisible_by .................... 128 | |
mask_prob ....................................... 0.15 | |
masked_softmax_fusion ........................... True | |
max_position_embeddings ......................... 1024 | |
merge_file ...................................... /gpfsscratch/rech/eha/commun/models-custom/megatron-gpt2/megatron_lm_345m_v0.0/release/gpt2-merges.txt | |
micro_batch_size ................................ 4 | |
min_loss_scale .................................. 1.0 | |
min_lr .......................................... 1e-05 | |
mmap_warmup ..................................... False | |
no_load_optim ................................... None | |
no_load_rng ..................................... None | |
no_save_optim ................................... None | |
no_save_rng ..................................... None | |
num_attention_heads ............................. 32 | |
num_channels .................................... 3 | |
num_classes ..................................... 1000 | |
num_layers ...................................... 64 | |
num_layers_per_virtual_pipeline_stage ........... None | |
num_workers ..................................... 2 | |
onnx_safe ....................................... None | |
openai_gelu ..................................... False | |
optimizer ....................................... adam | |
override_lr_scheduler ........................... False | |
params_dtype .................................... torch.float16 | |
patch_dim ....................................... 16 | |
pipeline_model_parallel_size .................... 16 | |
query_in_block_prob ............................. 0.1 | |
rampup_batch_size ............................... None | |
rank ............................................ 0 | |
reset_attention_mask ............................ False | |
reset_position_ids .............................. False | |
retriever_report_topk_accuracies ................ [] | |
retriever_score_scaling ......................... False | |
retriever_seq_length ............................ 256 | |
sample_rate ..................................... 1.0 | |
save ............................................ /gpfsscratch/rech/eha/commun/checkpoints/gpt2-1-node | |
save_interval ................................... 500 | |
scatter_gather_tensors_in_pipeline .............. True | |
seed ............................................ 1234 | |
seq_length ...................................... 1024 | |
sgd_momentum .................................... 0.9 | |
short_seq_prob .................................. 0.1 | |
split ........................................... 949,50,1 | |
tensor_model_parallel_size ...................... 4 | |
tensorboard_dir ................................. None | |
tensorboard_log_interval ........................ 1 | |
tensorboard_queue_size .......................... 1000 | |
titles_data_path ................................ None | |
tokenizer_type .................................. GPT2BPETokenizer | |
train_iters ..................................... 1000 | |
train_samples ................................... None | |
use_checkpoint_lr_scheduler ..................... False | |
use_contiguous_buffers_in_ddp ................... False | |
use_cpu_initialization .......................... None | |
use_one_sent_docs ............................... False | |
virtual_pipeline_model_parallel_size ............ None | |
vocab_extra_ids ................................. 0 | |
vocab_file ...................................... /gpfsscratch/rech/eha/commun/models-custom/megatron-gpt2/megatron_lm_345m_v0.0/release/gpt2-vocab.json | |
weight_decay .................................... 0.01 | |
world_size ...................................... 64 | |
-------------------- end of arguments --------------------- | |
setting number of micro-batches to constant 256 | |
> building GPT2BPETokenizer tokenizer ... | |
> padded vocab (size: 50257) with 431 dummy tokens (new size: 50688) | |
> initializing torch distributed ... | |
> initializing tensor model parallel with size 4 | |
> initializing pipeline model parallel with size 16 | |
> setting random seeds to 1234 ... | |
> initializing model parallel cuda seeds on global rank 0, model parallel rank 0, and data parallel rank 0 with model parallel seed: 3952 and data parallel seed: 1234 | |
> compiling dataset index builder ... | |
make: Entering directory '/gpfsdswork/projects/rech/eha/commun/code/megatron-lm/megatron/data' | |
make: Nothing to be done for 'default'. | |
make: Leaving directory '/gpfsdswork/projects/rech/eha/commun/code/megatron-lm/megatron/data' | |
>>> done with dataset index builder. Compilation time: 0.109 seconds | |
> compiling and loading fused kernels ... | |
Detected CUDA files, patching ldflags | |
Emitting ninja build file /gpfsdswork/projects/rech/eha/commun/code/megatron-lm/megatron/fused_kernels/build/build.ninja... | |
Building extension module scaled_upper_triang_masked_softmax_cuda... | |
Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) | |
ninja: no work to do. | |
Loading extension module scaled_upper_triang_masked_softmax_cuda... | |
Detected CUDA files, patching ldflags | |
Emitting ninja build file /gpfsdswork/projects/rech/eha/commun/code/megatron-lm/megatron/fused_kernels/build/build.ninja... | |
Building extension module scaled_masked_softmax_cuda... | |
Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) | |
ninja: no work to do. | |
Loading extension module scaled_masked_softmax_cuda... | |
Detected CUDA files, patching ldflags | |
Emitting ninja build file /gpfsdswork/projects/rech/eha/commun/code/megatron-lm/megatron/fused_kernels/build/build.ninja... | |
Building extension module fused_mix_prec_layer_norm_cuda... | |
Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) | |
ninja: no work to do. | |
Loading extension module fused_mix_prec_layer_norm_cuda... | |
>>> done with compiling and loading fused kernels. Compilation time: 12.395 seconds | |
time to initialize megatron (seconds): -27.152 | |
[after megatron is initialized] datetime: 2021-05-26 03:20:36 | |
building GPT model ... | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 1): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 13): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 2): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 11): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 6): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 1): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 2): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 13): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 6): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 6): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 14): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 10): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 14): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 7): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 5): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 5): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 7): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 4): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 5): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 4): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 4): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 1): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 13): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 13): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 1): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 2): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 2): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 6): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 5): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 12): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 3): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 12): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 3): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 10): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 10): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 4): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 7): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 14): 805560320 > number of parameters on (tensor, pipeline) model parallel rank (3, 14): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 10): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 7): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 11): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 11): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 11): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 3): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 9): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 3): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 12): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 12): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 9): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 9): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 8): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 8): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 8): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 8): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 9): 805560320 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 15): 909385728 | |
> number of parameters on (tensor, pipeline) model parallel rank (3, 0): 917757952 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 15): 909385728 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 15): 909385728 | |
> number of parameters on (tensor, pipeline) model parallel rank (1, 0): 917757952 | |
> number of parameters on (tensor, pipeline) model parallel rank (0, 0): 917757952 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 0): 917757952 | |
> number of parameters on (tensor, pipeline) model parallel rank (2, 15): 909385728 | |
> learning rate decay style: cosine | |
WARNING: could not find the metadata file /gpfsscratch/rech/eha/commun/checkpoints/gpt2-1-node/latest_checkpointed_iteration.txt | |
will not load any checkpoints and will start from random | |
time (ms) | load-checkpoint: 0.47 | |
[after model, optimizer, and learning rate scheduler are built] datetime: 2021-05-26 03:20:37 | |
> building train, validation, and test datasets ... | |
> datasets target sizes (minimum size): | |
train: 1024000 | |
validation: 112640 | |
test: 10240 | |
> building train, validation, and test datasets for GPT ... | |
> building dataset index ... | |
reading sizes... | |
reading pointers... | |
reading document index... | |
creating numpy buffer of mmap... | |
creating memory view of numpy buffer... | |
> finished creating indexed dataset in 0.001563 seconds | |
number of documents: 10000 | |
> dataset split: | |
train: | |
document indices in [0, 9490) total of 9490 documents | |
validation: | |
document indices in [9490, 9990) total of 500 documents | |
test: | |
document indices in [9990, 10000) total of 10 documents | |
> loading doc-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_train_indexmap_1024000ns_1024sl_1234s_doc_idx.npy | |
> loading sample-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_train_indexmap_1024000ns_1024sl_1234s_sample_idx.npy | |
> loading shuffle-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_train_indexmap_1024000ns_1024sl_1234s_shuffle_idx.npy | |
loaded indexed file in 0.013 seconds | |
total number of samples: 1024856 | |
total number of epochs: 99 | |
> loading doc-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_valid_indexmap_112640ns_1024sl_1234s_doc_idx.npy | |
> loading sample-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_valid_indexmap_112640ns_1024sl_1234s_sample_idx.npy | |
> loading shuffle-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_valid_indexmap_112640ns_1024sl_1234s_shuffle_idx.npy | |
loaded indexed file in 0.002 seconds | |
total number of samples: 113200 | |
total number of epochs: 182 | |
> loading doc-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_test_indexmap_10240ns_1024sl_1234s_doc_idx.npy | |
> loading sample-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_test_indexmap_10240ns_1024sl_1234s_sample_idx.npy | |
> loading shuffle-idx mapping from /gpfsscratch/rech/eha/commun/datasets-custom/openwebtext-10k/meg-gpt2_text_document_test_indexmap_10240ns_1024sl_1234s_shuffle_idx.npy | |
loaded indexed file in 0.001 seconds | |
total number of samples: 10255 | |
total number of epochs: 672 | |
> finished creating GPT datasets ... | |
[after dataloaders are built] datetime: 2021-05-26 03:20:38 | |
done with setup ... | |
training ... | |
time (ms) | model-and-optimizer-setup: 336.31 | train/valid/test-data-iterators-setup: 494.63 | |
[before the start of training step] datetime: 2021-05-26 03:20:38 | |
iteration 1/ 1000 | consumed samples: 1024 | elapsed time per iteration (ms): 155483.8 | learning rate: 0.000E+00 | global batch size: 1024 | loss scale: 4294967296.0 | number of skipped iterations: 1 | number of nan iterations: 0 | | |
time (ms) | forward-compute: 32054.16 | forward-recv: 19581.85 | backward-compute: 87380.53 | backward-send: 2.90 | backward-send-forward-recv: 11878.08 | backward-params-all-reduce: 26.06 | backward-embedding-all-reduce: 4181.82 | optimizer-copy-to-main-grad: 8.18 | optimizer-unscale-and-check-inf: 333.72 | optimizer: 342.01 | batch-generator: 182.63 | |
iteration 2/ 1000 | consumed samples: 2048 | elapsed time per iteration (ms): 129028.8 | learning rate: 0.000E+00 | global batch size: 1024 | loss scale: 2147483648.0 | number of skipped iterations: 1 | number of nan iterations: 0 | | |
time (ms) | forward-compute: 31749.74 | forward-recv: 1493.80 | backward-compute: 88911.73 | backward-send: 2.63 | backward-send-forward-recv: 2267.37 | backward-params-all-reduce: 26.04 | backward-embedding-all-reduce: 4492.62 | optimizer-copy-to-main-grad: 8.16 | optimizer-unscale-and-check-inf: 42.71 | optimizer: 51.07 | batch-generator: 184.46 | |
iteration 3/ 1000 | consumed samples: 3072 | elapsed time per iteration (ms): 128891.5 | learning rate: 0.000E+00 | global batch size: 1024 | loss scale: 1073741824.0 | number of skipped iterations: 1 | number of nan iterations: 0 | | |
time (ms) | forward-compute: 31888.01 | forward-recv: 1498.23 | backward-compute: 89444.82 | backward-send: 2.80 | backward-send-forward-recv: 1766.94 | backward-params-all-reduce: 26.45 | backward-embedding-all-reduce: 4179.82 | optimizer-copy-to-main-grad: 8.17 | optimizer-unscale-and-check-inf: 42.35 | optimizer: 50.66 | batch-generator: 187.52 | |