diff --git "a/llama13b_multiling_800M/13-05-2024-11:50:01/log.txt" "b/llama13b_multiling_800M/13-05-2024-11:50:01/log.txt" new file mode 100644--- /dev/null +++ "b/llama13b_multiling_800M/13-05-2024-11:50:01/log.txt" @@ -0,0 +1,3923 @@ +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. + warnings.warn( +[2024-05-13 11:50:03,059] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +[2024-05-13 11:50:04,374] [INFO] [runner.py:574:main] deepspeed_env file = /root/.deepspeed_env +[2024-05-13 11:50:04,374] [INFO] [multinode_runner.py:72:get_cmd] Running on the following workers: 100.83.37.175,100.83.37.188 +[2024-05-13 11:50:04,374] [INFO] [runner.py:585:main] cmd = pdsh -S -f 1024 -w 100.83.37.175,100.83.37.188 export PYTHONPATH=/Model-References/PyTorch/common:/usr/lib/habanalabs/:$PYTHONPATH; export PYTHON=/usr/bin/python3; export PT_HPU_LAZY_ACC_PAR_MODE=0; export PT_HPU_ENABLE_REFINE_DYNAMIC_SHAPES=0; export PT_HPU_ENABLE_WEIGHT_CPU_PERMUTE=0; export MODEL_REFERENCES_ROOT=/Model-References; export ENABLE_CONSOLE=false; export LOG_LEVEL_ALL=4; cd /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed; /usr/bin/python3 -u -m deepspeed.launcher.launch --world_info=eyIxMDAuODMuMzcuMTc1IjogWzAsIDEsIDIsIDMsIDQsIDUsIDYsIDddLCAiMTAwLjgzLjM3LjE4OCI6IFswLCAxLCAyLCAzLCA0LCA1LCA2LCA3XX0= --node_rank=%n --master_addr=100.83.37.175 --master_port=29500 --no_python --no_local_rank /usr/bin/bash -c ' cd /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed && python3 -u ./pretrain_llama.py --deepspeed --tensor-model-parallel-size 1 --pipeline-model-parallel-size 1 --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --layernorm-epsilon 1e-6 --num-layers 24 --hidden-size 2048 --ffn-hidden-size 4096 --num-attention-heads 32 --seq-length 2048 --micro-batch-size 1 --global-batch-size 256 --train-iters 10000 --log-interval 10 --eval-iters 10 --eval-interval 500 --data-path /data/hineng/tokenizer//_raw_content_document --vocab-file /data/hineng/tokenizer//gpt2-vocab.json --merge-file /data/hineng/tokenizer//gpt2-merges.txt --optimizer adamw --adam-beta1 0.9 --adam-beta2 0.95 --adam-eps 1e-6 --lr 3e-4 --lr-decay-style cosine --lr-warmup-iters 2000 --clip-grad 1.0 --weight-decay 0.1 --attention-dropout 0.1 --hidden-dropout 0.1 --tensorboard-dir /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/tensorboard --log-validation-ppl-to-tensorboard --log-batch-size-to-tensorboard --log-timers-to-tensorboard --load /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/checkpoints_zero_stage_2 --deepspeed_config=/data/output/llama13b_multiling_800M/13-05-2024-11:50:01/ds_config.json --zero-stage=0 --exit-interval 0 --no-masked-softmax-fusion --no-bias-gelu-fusion --no-bias-dropout-fusion --use-torch-compile false --use-fused-sdpa true --use-fused-sdpa-with-recompute false --bf16 --use_hpu --distributed-backend=hccl --hpu-deterministic --sequence-parallel --save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/checkpoints_zero_stage_2 --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt --save-interval 500 --verify-checkpoint --verify-checkpoint-model-type LLAMA' +100.83.37.175: Warning: Permanently added '[100.83.37.175]:3122' (ED25519) to the list of known hosts. +100.83.37.188: Warning: Permanently added '[100.83.37.188]:3122' (ED25519) to the list of known hosts. +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: ...done. +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: ...done. +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: [2024-05-13 11:50:06,041] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.188: [2024-05-13 11:50:06,084] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: [2024-05-13 11:50:07,212] [INFO] [launch.py:146:main] WORLD INFO DICT: {'100.83.37.175': [0, 1, 2, 3, 4, 5, 6, 7], '100.83.37.188': [0, 1, 2, 3, 4, 5, 6, 7]} +100.83.37.188: [2024-05-13 11:50:07,212] [INFO] [launch.py:152:main] nnodes=2, num_local_procs=8, node_rank=1 +100.83.37.188: [2024-05-13 11:50:07,212] [INFO] [launch.py:163:main] global_rank_mapping=defaultdict(, {'100.83.37.175': [0, 1, 2, 3, 4, 5, 6, 7], '100.83.37.188': [8, 9, 10, 11, 12, 13, 14, 15]}) +100.83.37.188: [2024-05-13 11:50:07,212] [INFO] [launch.py:164:main] dist_world_size=16 +100.83.37.188: [2024-05-13 11:50:07,212] [INFO] [launch.py:166:main] Setting CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: ...done. +100.83.37.188: ...done. +100.83.37.188: ...done. +100.83.37.188: ...done. +100.83.37.188: ...done. +100.83.37.188: ...done. +100.83.37.188: ...done. +100.83.37.188: * Starting OpenBSD Secure Shell server sshd +100.83.37.188: ...done. +100.83.37.175: [2024-05-13 11:50:07,442] [INFO] [launch.py:146:main] WORLD INFO DICT: {'100.83.37.175': [0, 1, 2, 3, 4, 5, 6, 7], '100.83.37.188': [0, 1, 2, 3, 4, 5, 6, 7]} +100.83.37.175: [2024-05-13 11:50:07,443] [INFO] [launch.py:152:main] nnodes=2, num_local_procs=8, node_rank=0 +100.83.37.175: [2024-05-13 11:50:07,443] [INFO] [launch.py:163:main] global_rank_mapping=defaultdict(, {'100.83.37.175': [0, 1, 2, 3, 4, 5, 6, 7], '100.83.37.188': [8, 9, 10, 11, 12, 13, 14, 15]}) +100.83.37.175: [2024-05-13 11:50:07,443] [INFO] [launch.py:164:main] dist_world_size=16 +100.83.37.175: [2024-05-13 11:50:07,443] [INFO] [launch.py:166:main] Setting CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: ...done. +100.83.37.175: ...done. +100.83.37.175: ...done. +100.83.37.175: ...done. +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: ...done. +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: ...done. +100.83.37.175: * Starting OpenBSD Secure Shell server sshd +100.83.37.175: ...done. +100.83.37.175: ...done. +100.83.37.188: [2024-05-13 11:50:09,078] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.188: [2024-05-13 11:50:09,095] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.188: [2024-05-13 11:50:09,117] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.188: [2024-05-13 11:50:09,117] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: [2024-05-13 11:50:09,117] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: [2024-05-13 11:50:09,136] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.188: [2024-05-13 11:50:09,138] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.175: [2024-05-13 11:50:09,176] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.175: [2024-05-13 11:50:09,176] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: [2024-05-13 11:50:09,178] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: [2024-05-13 11:50:09,180] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: [2024-05-13 11:50:09,192] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: [2024-05-13 11:50:09,194] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: [2024-05-13 11:50:09,725] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.188: warnings.warn( +100.83.37.175: [2024-05-13 11:50:09,863] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: /usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. +100.83.37.175: warnings.warn( +100.83.37.175: [2024-05-13 11:50:09,868] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +100.83.37.188: ---------------------------------------------------------------------------------------------------- +100.83.37.188: +100.83.37.188: DeepSpeed C++/CUDA extension op reportDeepSpeed C++/CUDA extension op report +100.83.37.188: +100.83.37.188: ---------------------------------------------------------------------------------------------------- +100.83.37.188: +100.83.37.188: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.188: runtime if needed. Op compatibility means that your system +100.83.37.188: meet the required dependencies to JIT install the op.NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.188: runtime if needed. Op compatibility means that your system +100.83.37.188: meet the required dependencies to JIT install the op. +100.83.37.188: +100.83.37.188: ---------------------------------------------------------------------------------------------------- +100.83.37.188: +100.83.37.188: JIT compiled ops requires ninjaJIT compiled ops requires ninja +100.83.37.188: +100.83.37.188: ninja .................. [OKAY] +100.83.37.188: -------------------------------------------------- +100.83.37.188: op name ................ installed .. ninjacompatible +100.83.37.188: --------------------------------------------------.................. +100.83.37.188: [OKAY] +100.83.37.188: -------------------------------------------------- +100.83.37.188: op name cpu_adam................ ...............installed [NO].. .......compatible +100.83.37.188: [OKAY] +100.83.37.188: -------------------------------------------------- +100.83.37.188: fused_adam ............. [NO] .......cpu_adam [OKAY]............... +100.83.37.188: [NO]deepspeed_not_implemented ....... [OKAY][NO] +100.83.37.188: ....... [OKAY] +100.83.37.188: fused_adam .............transformer_inference [NO].. .......[NO] [OKAY]....... +100.83.37.188: [OKAY] +100.83.37.188: deepspeed_not_implemented-------------------------------------------------- +100.83.37.188: [NO] ....... [OKAY] +100.83.37.188: transformer_inference .. [NO] ....... [OKAY] +100.83.37.188: -------------------------------------------------- +100.83.37.188: DeepSpeed general environment info: +100.83.37.188: torch install path ............... DeepSpeed general environment info: +100.83.37.188: ['/usr/local/lib/python3.10/dist-packages/torch']torch install path +100.83.37.188: ............... torch version .................... ['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.188: 2.1.1a0+gitb51c9f6 +100.83.37.188: torch versiondeepspeed install path .................... ........... 2.1.1a0+gitb51c9f6 +100.83.37.188: ['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.188: deepspeed install path deepspeed info........... ...................['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.188: 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.188: deepspeed info deepspeed wheel compiled w.................... ......0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.188: deepspeed wheel compiled w.torch 2.1 +100.83.37.188: ......shared memory (/dev/shm) size torch 2.1 .... +100.83.37.188: shared memory (/dev/shm) size503.75 GB +100.83.37.188: .... 503.75 GB +100.83.37.188: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.188: To add an exception for this directory, call: +100.83.37.188: +100.83.37.188: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.188: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.188: To add an exception for this directory, call: +100.83.37.188: +100.83.37.188: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.188: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.188: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.188: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.188: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.188: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.188: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.188: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.188: [--expert-interval EXPERT_INTERVAL] +100.83.37.188: [--hidden-size HIDDEN_SIZE] +100.83.37.188: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.188: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.188: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.188: [--kv-channels KV_CHANNELS] +100.83.37.188: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.188: [--use-rotary-position-embeddings] +100.83.37.188: [--rotary-percent ROTARY_PERCENT] +100.83.37.188: [--no-position-embedding] +100.83.37.188: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.188: [--normalization {layernorm,rmsnorm}] +100.83.37.188: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.188: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.188: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.188: [--apply-residual-connection-post-layernorm] +100.83.37.188: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.188: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.188: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.188: [--untie-embeddings-and-output-weights] +100.83.37.188: [--embedding-weights-in-fp32] +100.83.37.188: [--fix-position-emb-redundant-alloc] +100.83.37.188: [--embed-layernorm] +100.83.37.188: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.188: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.188: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.188: [--weight-decay WEIGHT_DECAY] +100.83.37.188: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.188: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.188: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.188: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.188: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.188: [--sgd-momentum SGD_MOMENTUM] +100.83.37.188: [--do-norm-bias-weight-decay] +100.83.37.188: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.188: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.188: [--batch-size BATCH_SIZE] +100.83.37.188: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.188: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.188: [--recompute-activations] +100.83.37.188: [--recompute-granularity {full,selective}] +100.83.37.188: [--distribute-saved-activations] +100.83.37.188: [--recompute-method {uniform,block}] +100.83.37.188: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.188: [--checkpoint-activations] +100.83.37.188: [--distribute-checkpointed-activations] +100.83.37.188: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.188: [--train-iters TRAIN_ITERS] +100.83.37.188: [--train-samples TRAIN_SAMPLES] +100.83.37.188: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.188: [--log-interval LOG_INTERVAL] +100.83.37.188: [--exit-interval EXIT_INTERVAL] +100.83.37.188: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.188: [--exit-signal-handler] +100.83.37.188: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.188: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.188: [--no-bias-dropout-fusion] +100.83.37.188: [--disable-moe-token-dropping] +100.83.37.188: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.188: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.188: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.188: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.188: [--create-moe-param-group] [--use-flash-attn] +100.83.37.188: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.188: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.188: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.188: [--disable-bias-linear] +100.83.37.188: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.188: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.188: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.188: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.188: [--no-async-tensor-model-parallel-allreduce] +100.83.37.188: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.188: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.188: [--force-ds-sequence-parallel] +100.83.37.188: [--no-gradient-accumulation-fusion] +100.83.37.188: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.188: [--data-parallel-random-init] +100.83.37.188: [--init-method-std INIT_METHOD_STD] +100.83.37.188: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.188: [--lr LR] +100.83.37.188: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.188: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.188: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.188: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.188: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.188: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.188: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.188: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.188: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.188: [--override-opt_param-scheduler] +100.83.37.188: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.188: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.188: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.188: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.188: [--no-initialization] [--use-checkpoint-args] +100.83.37.188: [--exit-on-missing-checkpoint] +100.83.37.188: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.188: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.188: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.188: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.188: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.188: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.188: [--hysteresis HYSTERESIS] +100.83.37.188: [--fp32-residual-connection] +100.83.37.188: [--no-query-key-layer-scaling] +100.83.37.188: [--attention-softmax-in-fp32] +100.83.37.188: [--accumulate-allreduce-grads-in-fp32] +100.83.37.188: [--fp16-lm-cross-entropy] +100.83.37.188: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.188: [--enable-expert-tensor-parallelism] +100.83.37.188: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.188: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.188: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.188: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.188: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.188: [--overlap-p2p-communication] +100.83.37.188: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.188: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.188: [--DDP-impl {local,torch,FSDP}] +100.83.37.188: [--no-contiguous-buffers-in-local-ddp] +100.83.37.188: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.188: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.188: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.188: [--use-cpu-initialization] +100.83.37.188: [--empty-unused-memory-level {0,1,2}] +100.83.37.188: [--standalone-embedding-stage] +100.83.37.188: [--use-distributed-optimizer] +100.83.37.188: [--eval-iters EVAL_ITERS] +100.83.37.188: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.188: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.188: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.188: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.188: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.188: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.188: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.188: [--data-cache-path DATA_CACHE_PATH] +100.83.37.188: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.188: [--merge-file MERGE_FILE] +100.83.37.188: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.188: [--seq-length SEQ_LENGTH] +100.83.37.188: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.188: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.188: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.188: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.188: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.188: [--num-workers NUM_WORKERS] +100.83.37.188: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.188: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.188: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.188: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.188: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.188: [--return-data-index] +100.83.37.188: [--data-efficiency-curriculum-learning] +100.83.37.188: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.188: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.188: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.188: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.188: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.188: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.188: [--adlr-autoresume] +100.83.37.188: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.188: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.188: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.188: [--biencoder-shared-query-context-model] +100.83.37.188: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.188: [--titles-data-path TITLES_DATA_PATH] +100.83.37.188: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.188: [--use-one-sent-docs] +100.83.37.188: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.188: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.188: [--retriever-score-scaling] +100.83.37.188: [--block-data-path BLOCK_DATA_PATH] +100.83.37.188: [--embedding-path EMBEDDING_PATH] +100.83.37.188: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.188: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.188: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.188: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.188: [--patch-dim PATCH_DIM] +100.83.37.188: [--classes-fraction CLASSES_FRACTION] +100.83.37.188: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.188: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.188: [--vision-pretraining] +100.83.37.188: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.188: [--vision-backbone-type {vit,mit,swin}] +100.83.37.188: [--swin-backbone-type {tiny,base,h3}] +100.83.37.188: [--mask-type {random,row}] +100.83.37.188: [--mask-factor MASK_FACTOR] +100.83.37.188: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.188: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.188: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.188: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.188: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.188: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.188: [--dino-norm-last-layer] +100.83.37.188: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.188: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.188: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.188: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.188: [--timing-log-level {0,1,2}] +100.83.37.188: [--no-barrier-with-level-1-timing] +100.83.37.188: [--timing-log-option {max,minmax,all}] +100.83.37.188: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.188: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.188: [--log-timers-to-tensorboard] +100.83.37.188: [--log-batch-size-to-tensorboard] +100.83.37.188: [--no-log-learnig-rate-to-tensorboard] +100.83.37.188: [--no-log-loss-scale-to-tensorboard] +100.83.37.188: [--log-validation-ppl-to-tensorboard] +100.83.37.188: [--log-optimizer-states-to-tensorboard] +100.83.37.188: [--log-memory-to-tensorboard] +100.83.37.188: [--log-world-size-to-tensorboard] +100.83.37.188: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.188: [--zero-contigious-gradients] +100.83.37.188: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.188: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.188: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.188: [--scattered-embeddings] [--split-transformers] +100.83.37.188: [--memory-centric-tiled-linear] +100.83.37.188: [--tile-factor TILE_FACTOR] +100.83.37.188: [--deepspeed-activation-checkpointing] +100.83.37.188: [--partition-activations] +100.83.37.188: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.188: [--synchronize-each-layer] [--profile-backward] +100.83.37.188: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.188: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.188: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.188: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.188: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.188: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.188: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.188: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.188: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.188: [--output-bert-embeddings] +100.83.37.188: [--bert-embedder-type {megatron,huggingface}] +100.83.37.188: [--cache-fp8-weight] +100.83.37.188: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.188: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.188: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.188: [--fp8-interval FP8_INTERVAL] +100.83.37.188: [--transformer-impl {local,transformer_engine}] +100.83.37.188: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.188: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.188: [--retro-workdir RETRO_WORKDIR] +100.83.37.188: [--retro-add-retriever] +100.83.37.188: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.188: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.188: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.188: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.188: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.188: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.188: [--retro-return-doc-ids] +100.83.37.188: [--profile {pt,pt-full,hltv}] +100.83.37.188: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.188: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.188: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.188: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.188: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.188: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.188: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.188: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.188: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.188: [--expert-interval EXPERT_INTERVAL] +100.83.37.188: [--hidden-size HIDDEN_SIZE] +100.83.37.188: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.188: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.188: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.188: [--kv-channels KV_CHANNELS] +100.83.37.188: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.188: [--use-rotary-position-embeddings] +100.83.37.188: [--rotary-percent ROTARY_PERCENT] +100.83.37.188: [--no-position-embedding] +100.83.37.188: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.188: [--normalization {layernorm,rmsnorm}] +100.83.37.188: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.188: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.188: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.188: [--apply-residual-connection-post-layernorm] +100.83.37.188: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.188: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.188: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.188: [--untie-embeddings-and-output-weights] +100.83.37.188: [--embedding-weights-in-fp32] +100.83.37.188: [--fix-position-emb-redundant-alloc] +100.83.37.188: [--embed-layernorm] +100.83.37.188: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.188: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.188: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.188: [--weight-decay WEIGHT_DECAY] +100.83.37.188: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.188: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.188: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.188: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.188: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.188: [--sgd-momentum SGD_MOMENTUM] +100.83.37.188: [--do-norm-bias-weight-decay] +100.83.37.188: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.188: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.188: [--batch-size BATCH_SIZE] +100.83.37.188: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.188: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.188: [--recompute-activations] +100.83.37.188: [--recompute-granularity {full,selective}] +100.83.37.188: [--distribute-saved-activations] +100.83.37.188: [--recompute-method {uniform,block}] +100.83.37.188: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.188: [--checkpoint-activations] +100.83.37.188: [--distribute-checkpointed-activations] +100.83.37.188: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.188: [--train-iters TRAIN_ITERS] +100.83.37.188: [--train-samples TRAIN_SAMPLES] +100.83.37.188: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.188: [--log-interval LOG_INTERVAL] +100.83.37.188: [--exit-interval EXIT_INTERVAL] +100.83.37.188: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.188: [--exit-signal-handler] +100.83.37.188: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.188: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.188: [--no-bias-dropout-fusion] +100.83.37.188: [--disable-moe-token-dropping] +100.83.37.188: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.188: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.188: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.188: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.188: [--create-moe-param-group] [--use-flash-attn] +100.83.37.188: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.188: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.188: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.188: [--disable-bias-linear] +100.83.37.188: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.188: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.188: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.188: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.188: [--no-async-tensor-model-parallel-allreduce] +100.83.37.188: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.188: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.188: [--force-ds-sequence-parallel] +100.83.37.188: [--no-gradient-accumulation-fusion] +100.83.37.188: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.188: [--data-parallel-random-init] +100.83.37.188: [--init-method-std INIT_METHOD_STD] +100.83.37.188: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.188: [--lr LR] +100.83.37.188: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.188: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.188: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.188: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.188: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.188: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.188: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.188: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.188: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.188: [--override-opt_param-scheduler] +100.83.37.188: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.188: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.188: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.188: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.188: [--no-initialization] [--use-checkpoint-args] +100.83.37.188: [--exit-on-missing-checkpoint] +100.83.37.188: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.188: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.188: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.188: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.188: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.188: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.188: [--hysteresis HYSTERESIS] +100.83.37.188: [--fp32-residual-connection] +100.83.37.188: [--no-query-key-layer-scaling] +100.83.37.188: [--attention-softmax-in-fp32] +100.83.37.188: [--accumulate-allreduce-grads-in-fp32] +100.83.37.188: [--fp16-lm-cross-entropy] +100.83.37.188: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.188: [--enable-expert-tensor-parallelism] +100.83.37.188: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.188: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.188: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.188: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.188: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.188: [--overlap-p2p-communication] +100.83.37.188: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.188: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.188: [--DDP-impl {local,torch,FSDP}] +100.83.37.188: [--no-contiguous-buffers-in-local-ddp] +100.83.37.188: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.188: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.188: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.188: [--use-cpu-initialization] +100.83.37.188: [--empty-unused-memory-level {0,1,2}] +100.83.37.188: [--standalone-embedding-stage] +100.83.37.188: [--use-distributed-optimizer] +100.83.37.188: [--eval-iters EVAL_ITERS] +100.83.37.188: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.188: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.188: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.188: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.188: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.188: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.188: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.188: [--data-cache-path DATA_CACHE_PATH] +100.83.37.188: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.188: [--merge-file MERGE_FILE] +100.83.37.188: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.188: [--seq-length SEQ_LENGTH] +100.83.37.188: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.188: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.188: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.188: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.188: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.188: [--num-workers NUM_WORKERS] +100.83.37.188: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.188: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.188: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.188: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.188: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.188: [--return-data-index] +100.83.37.188: [--data-efficiency-curriculum-learning] +100.83.37.188: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.188: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.188: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.188: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.188: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.188: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.188: [--adlr-autoresume] +100.83.37.188: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.188: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.188: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.188: [--biencoder-shared-query-context-model] +100.83.37.188: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.188: [--titles-data-path TITLES_DATA_PATH] +100.83.37.188: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.188: [--use-one-sent-docs] +100.83.37.188: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.188: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.188: [--retriever-score-scaling] +100.83.37.188: [--block-data-path BLOCK_DATA_PATH] +100.83.37.188: [--embedding-path EMBEDDING_PATH] +100.83.37.188: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.188: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.188: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.188: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.188: [--patch-dim PATCH_DIM] +100.83.37.188: [--classes-fraction CLASSES_FRACTION] +100.83.37.188: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.188: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.188: [--vision-pretraining] +100.83.37.188: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.188: [--vision-backbone-type {vit,mit,swin}] +100.83.37.188: [--swin-backbone-type {tiny,base,h3}] +100.83.37.188: [--mask-type {random,row}] +100.83.37.188: [--mask-factor MASK_FACTOR] +100.83.37.188: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.188: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.188: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.188: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.188: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.188: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.188: [--dino-norm-last-layer] +100.83.37.188: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.188: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.188: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.188: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.188: [--timing-log-level {0,1,2}] +100.83.37.188: [--no-barrier-with-level-1-timing] +100.83.37.188: [--timing-log-option {max,minmax,all}] +100.83.37.188: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.188: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.188: [--log-timers-to-tensorboard] +100.83.37.188: [--log-batch-size-to-tensorboard] +100.83.37.188: [--no-log-learnig-rate-to-tensorboard] +100.83.37.188: [--no-log-loss-scale-to-tensorboard] +100.83.37.188: [--log-validation-ppl-to-tensorboard] +100.83.37.188: [--log-optimizer-states-to-tensorboard] +100.83.37.188: [--log-memory-to-tensorboard] +100.83.37.188: [--log-world-size-to-tensorboard] +100.83.37.188: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.188: [--zero-contigious-gradients] +100.83.37.188: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.188: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.188: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.188: [--scattered-embeddings] [--split-transformers] +100.83.37.188: [--memory-centric-tiled-linear] +100.83.37.188: [--tile-factor TILE_FACTOR] +100.83.37.188: [--deepspeed-activation-checkpointing] +100.83.37.188: [--partition-activations] +100.83.37.188: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.188: [--synchronize-each-layer] [--profile-backward] +100.83.37.188: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.188: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.188: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.188: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.188: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.188: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.188: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.188: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.188: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.188: [--output-bert-embeddings] +100.83.37.188: [--bert-embedder-type {megatron,huggingface}] +100.83.37.188: [--cache-fp8-weight] +100.83.37.188: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.188: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.188: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.188: [--fp8-interval FP8_INTERVAL] +100.83.37.188: [--transformer-impl {local,transformer_engine}] +100.83.37.188: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.188: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.188: [--retro-workdir RETRO_WORKDIR] +100.83.37.188: [--retro-add-retriever] +100.83.37.188: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.188: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.188: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.188: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.188: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.188: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.188: [--retro-return-doc-ids] +100.83.37.188: [--profile {pt,pt-full,hltv}] +100.83.37.188: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.188: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.188: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.188: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.188: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.188: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.188: -------------------------------------------------- +100.83.37.188: DeepSpeed C++/CUDA extension op report +100.83.37.188: -------------------------------------------------- +100.83.37.188: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.188: runtime if needed. Op compatibility means that your system +100.83.37.188: meet the required dependencies to JIT install the op. +100.83.37.188: -------------------------------------------------- +100.83.37.188: JIT compiled ops requires ninja +100.83.37.188: -------------------------------------------------- +100.83.37.188: ----------------------------------------------------------------------------------------------------DeepSpeed C++/CUDA extension op report +100.83.37.188: +100.83.37.188: +100.83.37.188: --------------------------------------------------DeepSpeed C++/CUDA extension op report +100.83.37.188: DeepSpeed C++/CUDA extension op report +100.83.37.188: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.188: runtime if needed. Op compatibility means that your system +100.83.37.188: meet the required dependencies to JIT install the op. +100.83.37.188: -------------------------------------------------- +100.83.37.188: +100.83.37.188: ---------------------------------------------------------------------------------------------------- +100.83.37.188: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.188: runtime if needed. Op compatibility means that your system +100.83.37.188: meet the required dependencies to JIT install the op. +100.83.37.188: JIT compiled ops requires ninja +100.83.37.188: +100.83.37.188: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.188: runtime if needed. Op compatibility means that your system +100.83.37.188: meet the required dependencies to JIT install the op.-------------------------------------------------- +100.83.37.188: -------------------------------------------------- +100.83.37.188: +100.83.37.188: JIT compiled ops requires ninja +100.83.37.188: JIT compiled ops requires ninja +100.83.37.188: ninja .................. [OKAY] +100.83.37.188: -------------------------------------------------- +100.83.37.188: op name ................ installed .. compatible +100.83.37.188: -------------------------------------------------- +100.83.37.188: cpu_adam ............... [NO] ....... [OKAY] +100.83.37.188: fused_adam ............. [NO] ....... [OKAY] +100.83.37.188: deepspeed_not_implemented [NO] ....... [OKAY] +100.83.37.188: transformer_inference .. [NO] ....... [OKAY] +100.83.37.188: -------------------------------------------------- +100.83.37.188: DeepSpeed general environment info: +100.83.37.188: torch install path ............... ['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.188: torch version .................... 2.1.1a0+gitb51c9f6 +100.83.37.188: deepspeed install path ........... ['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.188: deepspeed info ................... 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.188: deepspeed wheel compiled w. ...... torch 2.1 +100.83.37.188: shared memory (/dev/shm) size .... 503.75 GB +100.83.37.188: ninja ..................ninja [OKAY]ninja.................. +100.83.37.188: --------------------------------------------------[OKAY].................. +100.83.37.188: +100.83.37.188: [OKAY]op name-------------------------------------------------- +100.83.37.188: +100.83.37.188: ................ --------------------------------------------------op nameinstalled +100.83.37.188: ..................op name installedcompatible................ +100.83.37.188: ..installed -------------------------------------------------- compatible +100.83.37.188: .. +100.83.37.188: compatible-------------------------------------------------- +100.83.37.188: +100.83.37.188: -------------------------------------------------- +100.83.37.188: cpu_adam ............... [NO] cpu_adam....... cpu_adam...............[OKAY] ............... +100.83.37.188: [NO] [NO] fused_adam.............. [OKAY].............[OKAY] +100.83.37.188: +100.83.37.188: [NO] fused_adam.......fused_adam .............[OKAY]............. +100.83.37.188: [NO][NO] ....... deepspeed_not_implemented ....... [OKAY] +100.83.37.188: [OKAY] +100.83.37.188: [NO]deepspeed_not_implemented .......deepspeed_not_implemented [OKAY][NO] +100.83.37.188: [NO]....... transformer_inference ....... [OKAY] .. +100.83.37.188: [OKAY] +100.83.37.188: [NO] transformer_inference....... transformer_inference ..[OKAY].. +100.83.37.188: [NO][NO] -------------------------------------------------- ....... +100.83.37.188: ....... [OKAY][OKAY] +100.83.37.188: +100.83.37.188: ---------------------------------------------------------------------------------------------------- +100.83.37.188: +100.83.37.188: DeepSpeed general environment info: +100.83.37.188: DeepSpeed general environment info:torch install path +100.83.37.188: DeepSpeed general environment info:............... +100.83.37.188: torch install path torch install path............... ['/usr/local/lib/python3.10/dist-packages/torch']............... +100.83.37.188: ['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.188: torch version['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.188: ....................torch version torch version.................... 2.1.1a0+gitb51c9f6 .................... +100.83.37.188: 2.1.1a0+gitb51c9f6 +100.83.37.188: deepspeed install path2.1.1a0+gitb51c9f6deepspeed install path +100.83.37.188: ........... deepspeed install path........... ['/usr/local/lib/python3.10/dist-packages/deepspeed']........... +100.83.37.188: ['/usr/local/lib/python3.10/dist-packages/deepspeed']deepspeed info['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.188: +100.83.37.188: ...................deepspeed infodeepspeed info 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0...................................... +100.83.37.188: 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0deepspeed wheel compiled w.0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.188: +100.83.37.188: deepspeed wheel compiled w.......deepspeed wheel compiled w. ......torch 2.1 ...... +100.83.37.188: torch 2.1 torch 2.1 +100.83.37.188: shared memory (/dev/shm) size +100.83.37.188: shared memory (/dev/shm) size.... shared memory (/dev/shm) size ....503.75 GB +100.83.37.188: ....503.75 GB +100.83.37.188: 503.75 GB +100.83.37.188: -------------------------------------------------- +100.83.37.188: DeepSpeed C++/CUDA extension op report +100.83.37.188: -------------------------------------------------- +100.83.37.188: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.188: runtime if needed. Op compatibility means that your system +100.83.37.188: meet the required dependencies to JIT install the op. +100.83.37.188: -------------------------------------------------- +100.83.37.188: JIT compiled ops requires ninja +100.83.37.188: ninja .................. [OKAY] +100.83.37.188: -------------------------------------------------- +100.83.37.188: op name ................ installed .. compatible +100.83.37.188: -------------------------------------------------- +100.83.37.188: cpu_adam ............... [NO] ....... [OKAY] +100.83.37.188: fused_adam ............. [NO] ....... [OKAY] +100.83.37.188: deepspeed_not_implemented [NO] ....... [OKAY] +100.83.37.188: transformer_inference .. [NO] ....... [OKAY] +100.83.37.188: -------------------------------------------------- +100.83.37.188: DeepSpeed general environment info: +100.83.37.188: torch install path ............... ['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.188: torch version .................... 2.1.1a0+gitb51c9f6 +100.83.37.188: deepspeed install path ........... ['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.188: deepspeed info ................... 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.188: deepspeed wheel compiled w. ...... torch 2.1 +100.83.37.188: shared memory (/dev/shm) size .... 503.75 GB +100.83.37.188: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.188: To add an exception for this directory, call: +100.83.37.188: +100.83.37.188: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.188: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.188: To add an exception for this directory, call: +100.83.37.188: +100.83.37.188: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.188: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.188: To add an exception for this directory, call: +100.83.37.188: +100.83.37.188: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.188: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.188: To add an exception for this directory, call: +100.83.37.188: +100.83.37.188: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.188: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.188: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.188: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.188: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.188: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.188: To add an exception for this directory, call: +100.83.37.188: +100.83.37.188: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.188: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.188: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.188: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.188: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.188: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.188: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.188: [--expert-interval EXPERT_INTERVAL] +100.83.37.188: [--hidden-size HIDDEN_SIZE] +100.83.37.188: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.188: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.188: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.188: [--kv-channels KV_CHANNELS] +100.83.37.188: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.188: [--use-rotary-position-embeddings] +100.83.37.188: [--rotary-percent ROTARY_PERCENT] +100.83.37.188: [--no-position-embedding] +100.83.37.188: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.188: [--normalization {layernorm,rmsnorm}] +100.83.37.188: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.188: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.188: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.188: [--apply-residual-connection-post-layernorm] +100.83.37.188: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.188: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.188: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.188: [--untie-embeddings-and-output-weights] +100.83.37.188: [--embedding-weights-in-fp32] +100.83.37.188: [--fix-position-emb-redundant-alloc] +100.83.37.188: [--embed-layernorm] +100.83.37.188: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.188: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.188: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.188: [--weight-decay WEIGHT_DECAY] +100.83.37.188: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.188: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.188: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.188: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.188: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.188: [--sgd-momentum SGD_MOMENTUM] +100.83.37.188: [--do-norm-bias-weight-decay] +100.83.37.188: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.188: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.188: [--batch-size BATCH_SIZE] +100.83.37.188: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.188: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.188: [--recompute-activations] +100.83.37.188: [--recompute-granularity {full,selective}] +100.83.37.188: [--distribute-saved-activations] +100.83.37.188: [--recompute-method {uniform,block}] +100.83.37.188: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.188: [--checkpoint-activations] +100.83.37.188: [--distribute-checkpointed-activations] +100.83.37.188: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.188: [--train-iters TRAIN_ITERS] +100.83.37.188: [--train-samples TRAIN_SAMPLES] +100.83.37.188: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.188: [--log-interval LOG_INTERVAL] +100.83.37.188: [--exit-interval EXIT_INTERVAL] +100.83.37.188: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.188: [--exit-signal-handler] +100.83.37.188: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.188: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.188: [--no-bias-dropout-fusion] +100.83.37.188: [--disable-moe-token-dropping] +100.83.37.188: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.188: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.188: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.188: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.188: [--create-moe-param-group] [--use-flash-attn] +100.83.37.188: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.188: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.188: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.188: [--disable-bias-linear] +100.83.37.188: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.188: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.188: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.188: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.188: [--no-async-tensor-model-parallel-allreduce] +100.83.37.188: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.188: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.188: [--force-ds-sequence-parallel] +100.83.37.188: [--no-gradient-accumulation-fusion] +100.83.37.188: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.188: [--data-parallel-random-init] +100.83.37.188: [--init-method-std INIT_METHOD_STD] +100.83.37.188: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.188: [--lr LR] +100.83.37.188: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.188: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.188: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.188: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.188: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.188: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.188: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.188: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.188: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.188: [--override-opt_param-scheduler] +100.83.37.188: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.188: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.188: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.188: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.188: [--no-initialization] [--use-checkpoint-args] +100.83.37.188: [--exit-on-missing-checkpoint] +100.83.37.188: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.188: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.188: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.188: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.188: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.188: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.188: [--hysteresis HYSTERESIS] +100.83.37.188: [--fp32-residual-connection] +100.83.37.188: [--no-query-key-layer-scaling] +100.83.37.188: [--attention-softmax-in-fp32] +100.83.37.188: [--accumulate-allreduce-grads-in-fp32] +100.83.37.188: [--fp16-lm-cross-entropy] +100.83.37.188: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.188: [--enable-expert-tensor-parallelism] +100.83.37.188: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.188: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.188: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.188: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.188: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.188: [--overlap-p2p-communication] +100.83.37.188: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.188: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.188: [--DDP-impl {local,torch,FSDP}] +100.83.37.188: [--no-contiguous-buffers-in-local-ddp] +100.83.37.188: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.188: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.188: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.188: [--use-cpu-initialization] +100.83.37.188: [--empty-unused-memory-level {0,1,2}] +100.83.37.188: [--standalone-embedding-stage] +100.83.37.188: [--use-distributed-optimizer] +100.83.37.188: [--eval-iters EVAL_ITERS] +100.83.37.188: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.188: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.188: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.188: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.188: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.188: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.188: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.188: [--data-cache-path DATA_CACHE_PATH] +100.83.37.188: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.188: [--merge-file MERGE_FILE] +100.83.37.188: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.188: [--seq-length SEQ_LENGTH] +100.83.37.188: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.188: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.188: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.188: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.188: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.188: [--num-workers NUM_WORKERS] +100.83.37.188: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.188: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.188: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.188: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.188: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.188: [--return-data-index] +100.83.37.188: [--data-efficiency-curriculum-learning] +100.83.37.188: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.188: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.188: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.188: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.188: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.188: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.188: [--adlr-autoresume] +100.83.37.188: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.188: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.188: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.188: [--biencoder-shared-query-context-model] +100.83.37.188: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.188: [--titles-data-path TITLES_DATA_PATH] +100.83.37.188: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.188: [--use-one-sent-docs] +100.83.37.188: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.188: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.188: [--retriever-score-scaling] +100.83.37.188: [--block-data-path BLOCK_DATA_PATH] +100.83.37.188: [--embedding-path EMBEDDING_PATH] +100.83.37.188: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.188: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.188: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.188: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.188: [--patch-dim PATCH_DIM] +100.83.37.188: [--classes-fraction CLASSES_FRACTION] +100.83.37.188: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.188: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.188: [--vision-pretraining] +100.83.37.188: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.188: [--vision-backbone-type {vit,mit,swin}] +100.83.37.188: [--swin-backbone-type {tiny,base,h3}] +100.83.37.188: [--mask-type {random,row}] +100.83.37.188: [--mask-factor MASK_FACTOR] +100.83.37.188: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.188: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.188: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.188: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.188: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.188: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.188: [--dino-norm-last-layer] +100.83.37.188: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.188: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.188: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.188: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.188: [--timing-log-level {0,1,2}] +100.83.37.188: [--no-barrier-with-level-1-timing] +100.83.37.188: [--timing-log-option {max,minmax,all}] +100.83.37.188: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.188: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.188: [--log-timers-to-tensorboard] +100.83.37.188: [--log-batch-size-to-tensorboard] +100.83.37.188: [--no-log-learnig-rate-to-tensorboard] +100.83.37.188: [--no-log-loss-scale-to-tensorboard] +100.83.37.188: [--log-validation-ppl-to-tensorboard] +100.83.37.188: [--log-optimizer-states-to-tensorboard] +100.83.37.188: [--log-memory-to-tensorboard] +100.83.37.188: [--log-world-size-to-tensorboard] +100.83.37.188: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.188: [--zero-contigious-gradients] +100.83.37.188: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.188: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.188: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.188: [--scattered-embeddings] [--split-transformers] +100.83.37.188: [--memory-centric-tiled-linear] +100.83.37.188: [--tile-factor TILE_FACTOR] +100.83.37.188: [--deepspeed-activation-checkpointing] +100.83.37.188: [--partition-activations] +100.83.37.188: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.188: [--synchronize-each-layer] [--profile-backward] +100.83.37.188: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.188: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.188: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.188: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.188: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.188: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.188: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.188: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.188: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.188: [--output-bert-embeddings] +100.83.37.188: [--bert-embedder-type {megatron,huggingface}] +100.83.37.188: [--cache-fp8-weight] +100.83.37.188: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.188: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.188: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.188: [--fp8-interval FP8_INTERVAL] +100.83.37.188: [--transformer-impl {local,transformer_engine}] +100.83.37.188: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.188: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.188: [--retro-workdir RETRO_WORKDIR] +100.83.37.188: [--retro-add-retriever] +100.83.37.188: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.188: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.188: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.188: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.188: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.188: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.188: [--retro-return-doc-ids] +100.83.37.188: [--profile {pt,pt-full,hltv}] +100.83.37.188: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.188: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.188: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.188: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.188: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.188: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.188: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.188: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.188: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.188: [--expert-interval EXPERT_INTERVAL] +100.83.37.188: [--hidden-size HIDDEN_SIZE] +100.83.37.188: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.188: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.188: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.188: [--kv-channels KV_CHANNELS] +100.83.37.188: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.188: [--use-rotary-position-embeddings] +100.83.37.188: [--rotary-percent ROTARY_PERCENT] +100.83.37.188: [--no-position-embedding] +100.83.37.188: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.188: [--normalization {layernorm,rmsnorm}] +100.83.37.188: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.188: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.188: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.188: [--apply-residual-connection-post-layernorm] +100.83.37.188: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.188: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.188: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.188: [--untie-embeddings-and-output-weights] +100.83.37.188: [--embedding-weights-in-fp32] +100.83.37.188: [--fix-position-emb-redundant-alloc] +100.83.37.188: [--embed-layernorm] +100.83.37.188: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.188: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.188: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.188: [--weight-decay WEIGHT_DECAY] +100.83.37.188: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.188: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.188: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.188: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.188: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.188: [--sgd-momentum SGD_MOMENTUM] +100.83.37.188: [--do-norm-bias-weight-decay] +100.83.37.188: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.188: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.188: [--batch-size BATCH_SIZE] +100.83.37.188: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.188: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.188: [--recompute-activations] +100.83.37.188: [--recompute-granularity {full,selective}] +100.83.37.188: [--distribute-saved-activations] +100.83.37.188: [--recompute-method {uniform,block}] +100.83.37.188: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.188: [--checkpoint-activations] +100.83.37.188: [--distribute-checkpointed-activations] +100.83.37.188: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.188: [--train-iters TRAIN_ITERS] +100.83.37.188: [--train-samples TRAIN_SAMPLES] +100.83.37.188: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.188: [--log-interval LOG_INTERVAL] +100.83.37.188: [--exit-interval EXIT_INTERVAL] +100.83.37.188: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.188: [--exit-signal-handler] +100.83.37.188: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.188: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.188: [--no-bias-dropout-fusion] +100.83.37.188: [--disable-moe-token-dropping] +100.83.37.188: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.188: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.188: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.188: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.188: [--create-moe-param-group] [--use-flash-attn] +100.83.37.188: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.188: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.188: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.188: [--disable-bias-linear] +100.83.37.188: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.188: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.188: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.188: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.188: [--no-async-tensor-model-parallel-allreduce] +100.83.37.188: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.188: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.188: [--force-ds-sequence-parallel] +100.83.37.188: [--no-gradient-accumulation-fusion] +100.83.37.188: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.188: [--data-parallel-random-init] +100.83.37.188: [--init-method-std INIT_METHOD_STD] +100.83.37.188: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.188: [--lr LR] +100.83.37.188: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.188: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.188: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.188: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.188: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.188: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.188: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.188: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.188: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.188: [--override-opt_param-scheduler] +100.83.37.188: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.188: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.188: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.188: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.188: [--no-initialization] [--use-checkpoint-args] +100.83.37.188: [--exit-on-missing-checkpoint] +100.83.37.188: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.188: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.188: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.188: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.188: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.188: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.188: [--hysteresis HYSTERESIS] +100.83.37.188: [--fp32-residual-connection] +100.83.37.188: [--no-query-key-layer-scaling] +100.83.37.188: [--attention-softmax-in-fp32] +100.83.37.188: [--accumulate-allreduce-grads-in-fp32] +100.83.37.188: [--fp16-lm-cross-entropy] +100.83.37.188: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.188: [--enable-expert-tensor-parallelism] +100.83.37.188: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.188: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.188: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.188: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.188: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.188: [--overlap-p2p-communication] +100.83.37.188: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.188: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.188: [--DDP-impl {local,torch,FSDP}] +100.83.37.188: [--no-contiguous-buffers-in-local-ddp] +100.83.37.188: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.188: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.188: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.188: [--use-cpu-initialization] +100.83.37.188: [--empty-unused-memory-level {0,1,2}] +100.83.37.188: [--standalone-embedding-stage] +100.83.37.188: [--use-distributed-optimizer] +100.83.37.188: [--eval-iters EVAL_ITERS] +100.83.37.188: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.188: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.188: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.188: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.188: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.188: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.188: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.188: [--data-cache-path DATA_CACHE_PATH] +100.83.37.188: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.188: [--merge-file MERGE_FILE] +100.83.37.188: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.188: [--seq-length SEQ_LENGTH] +100.83.37.188: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.188: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.188: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.188: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.188: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.188: [--num-workers NUM_WORKERS] +100.83.37.188: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.188: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.188: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.188: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.188: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.188: [--return-data-index] +100.83.37.188: [--data-efficiency-curriculum-learning] +100.83.37.188: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.188: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.188: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.188: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.188: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.188: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.188: [--adlr-autoresume] +100.83.37.188: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.188: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.188: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.188: [--biencoder-shared-query-context-model] +100.83.37.188: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.188: [--titles-data-path TITLES_DATA_PATH] +100.83.37.188: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.188: [--use-one-sent-docs] +100.83.37.188: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.188: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.188: [--retriever-score-scaling] +100.83.37.188: [--block-data-path BLOCK_DATA_PATH] +100.83.37.188: [--embedding-path EMBEDDING_PATH] +100.83.37.188: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.188: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.188: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.188: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.188: [--patch-dim PATCH_DIM] +100.83.37.188: [--classes-fraction CLASSES_FRACTION] +100.83.37.188: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.188: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.188: [--vision-pretraining] +100.83.37.188: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.188: [--vision-backbone-type {vit,mit,swin}] +100.83.37.188: [--swin-backbone-type {tiny,base,h3}] +100.83.37.188: [--mask-type {random,row}] +100.83.37.188: [--mask-factor MASK_FACTOR] +100.83.37.188: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.188: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.188: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.188: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.188: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.188: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.188: [--dino-norm-last-layer] +100.83.37.188: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.188: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.188: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.188: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.188: [--timing-log-level {0,1,2}] +100.83.37.188: [--no-barrier-with-level-1-timing] +100.83.37.188: [--timing-log-option {max,minmax,all}] +100.83.37.188: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.188: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.188: [--log-timers-to-tensorboard] +100.83.37.188: [--log-batch-size-to-tensorboard] +100.83.37.188: [--no-log-learnig-rate-to-tensorboard] +100.83.37.188: [--no-log-loss-scale-to-tensorboard] +100.83.37.188: [--log-validation-ppl-to-tensorboard] +100.83.37.188: [--log-optimizer-states-to-tensorboard] +100.83.37.188: [--log-memory-to-tensorboard] +100.83.37.188: [--log-world-size-to-tensorboard] +100.83.37.188: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.188: [--zero-contigious-gradients] +100.83.37.188: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.188: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.188: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.188: [--scattered-embeddings] [--split-transformers] +100.83.37.188: [--memory-centric-tiled-linear] +100.83.37.188: [--tile-factor TILE_FACTOR] +100.83.37.188: [--deepspeed-activation-checkpointing] +100.83.37.188: [--partition-activations] +100.83.37.188: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.188: [--synchronize-each-layer] [--profile-backward] +100.83.37.188: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.188: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.188: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.188: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.188: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.188: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.188: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.188: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.188: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.188: [--output-bert-embeddings] +100.83.37.188: [--bert-embedder-type {megatron,huggingface}] +100.83.37.188: [--cache-fp8-weight] +100.83.37.188: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.188: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.188: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.188: [--fp8-interval FP8_INTERVAL] +100.83.37.188: [--transformer-impl {local,transformer_engine}] +100.83.37.188: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.188: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.188: [--retro-workdir RETRO_WORKDIR] +100.83.37.188: [--retro-add-retriever] +100.83.37.188: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.188: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.188: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.188: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.188: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.188: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.188: [--retro-return-doc-ids] +100.83.37.188: [--profile {pt,pt-full,hltv}] +100.83.37.188: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.188: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.188: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.188: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.188: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.188: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.188: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.188: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.188: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.188: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.188: [--expert-interval EXPERT_INTERVAL] +100.83.37.188: [--hidden-size HIDDEN_SIZE] +100.83.37.188: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.188: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.188: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.188: [--kv-channels KV_CHANNELS] +100.83.37.188: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.188: [--use-rotary-position-embeddings] +100.83.37.188: [--rotary-percent ROTARY_PERCENT] +100.83.37.188: [--no-position-embedding] +100.83.37.188: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.188: [--normalization {layernorm,rmsnorm}] +100.83.37.188: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.188: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.188: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.188: [--apply-residual-connection-post-layernorm] +100.83.37.188: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.188: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.188: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.188: [--untie-embeddings-and-output-weights] +100.83.37.188: [--embedding-weights-in-fp32] +100.83.37.188: [--fix-position-emb-redundant-alloc] +100.83.37.188: [--embed-layernorm] +100.83.37.188: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.188: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.188: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.188: [--weight-decay WEIGHT_DECAY] +100.83.37.188: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.188: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.188: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.188: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.188: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.188: [--sgd-momentum SGD_MOMENTUM] +100.83.37.188: [--do-norm-bias-weight-decay] +100.83.37.188: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.188: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.188: [--batch-size BATCH_SIZE] +100.83.37.188: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.188: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.188: [--recompute-activations] +100.83.37.188: [--recompute-granularity {full,selective}] +100.83.37.188: [--distribute-saved-activations] +100.83.37.188: [--recompute-method {uniform,block}] +100.83.37.188: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.188: [--checkpoint-activations] +100.83.37.188: [--distribute-checkpointed-activations] +100.83.37.188: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.188: [--train-iters TRAIN_ITERS] +100.83.37.188: [--train-samples TRAIN_SAMPLES] +100.83.37.188: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.188: [--log-interval LOG_INTERVAL] +100.83.37.188: [--exit-interval EXIT_INTERVAL] +100.83.37.188: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.188: [--exit-signal-handler] +100.83.37.188: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.188: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.188: [--no-bias-dropout-fusion] +100.83.37.188: [--disable-moe-token-dropping] +100.83.37.188: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.188: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.188: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.188: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.188: [--create-moe-param-group] [--use-flash-attn] +100.83.37.188: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.188: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.188: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.188: [--disable-bias-linear] +100.83.37.188: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.188: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.188: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.188: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.188: [--no-async-tensor-model-parallel-allreduce] +100.83.37.188: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.188: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.188: [--force-ds-sequence-parallel] +100.83.37.188: [--no-gradient-accumulation-fusion] +100.83.37.188: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.188: [--data-parallel-random-init] +100.83.37.188: [--init-method-std INIT_METHOD_STD] +100.83.37.188: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.188: [--lr LR] +100.83.37.188: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.188: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.188: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.188: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.188: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.188: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.188: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.188: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.188: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.188: [--override-opt_param-scheduler] +100.83.37.188: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.188: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.188: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.188: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.188: [--no-initialization] [--use-checkpoint-args] +100.83.37.188: [--exit-on-missing-checkpoint] +100.83.37.188: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.188: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.188: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.188: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.188: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.188: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.188: [--hysteresis HYSTERESIS] +100.83.37.188: [--fp32-residual-connection] +100.83.37.188: [--no-query-key-layer-scaling] +100.83.37.188: [--attention-softmax-in-fp32] +100.83.37.188: [--accumulate-allreduce-grads-in-fp32] +100.83.37.188: [--fp16-lm-cross-entropy] +100.83.37.188: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.188: [--enable-expert-tensor-parallelism] +100.83.37.188: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.188: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.188: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.188: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.188: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.188: [--overlap-p2p-communication] +100.83.37.188: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.188: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.188: [--DDP-impl {local,torch,FSDP}] +100.83.37.188: [--no-contiguous-buffers-in-local-ddp] +100.83.37.188: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.188: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.188: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.188: [--use-cpu-initialization] +100.83.37.188: [--empty-unused-memory-level {0,1,2}] +100.83.37.188: [--standalone-embedding-stage] +100.83.37.188: [--use-distributed-optimizer] +100.83.37.188: [--eval-iters EVAL_ITERS] +100.83.37.188: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.188: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.188: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.188: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.188: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.188: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.188: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.188: [--data-cache-path DATA_CACHE_PATH] +100.83.37.188: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.188: [--merge-file MERGE_FILE] +100.83.37.188: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.188: [--seq-length SEQ_LENGTH] +100.83.37.188: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.188: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.188: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.188: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.188: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.188: [--num-workers NUM_WORKERS] +100.83.37.188: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.188: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.188: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.188: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.188: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.188: [--return-data-index] +100.83.37.188: [--data-efficiency-curriculum-learning] +100.83.37.188: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.188: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.188: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.188: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.188: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.188: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.188: [--adlr-autoresume] +100.83.37.188: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.188: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.188: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.188: [--biencoder-shared-query-context-model] +100.83.37.188: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.188: [--titles-data-path TITLES_DATA_PATH] +100.83.37.188: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.188: [--use-one-sent-docs] +100.83.37.188: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.188: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.188: [--retriever-score-scaling] +100.83.37.188: [--block-data-path BLOCK_DATA_PATH] +100.83.37.188: [--embedding-path EMBEDDING_PATH] +100.83.37.188: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.188: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.188: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.188: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.188: [--patch-dim PATCH_DIM] +100.83.37.188: [--classes-fraction CLASSES_FRACTION] +100.83.37.188: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.188: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.188: [--vision-pretraining] +100.83.37.188: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.188: [--vision-backbone-type {vit,mit,swin}] +100.83.37.188: [--swin-backbone-type {tiny,base,h3}] +100.83.37.188: [--mask-type {random,row}] +100.83.37.188: [--mask-factor MASK_FACTOR] +100.83.37.188: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.188: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.188: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.188: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.188: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.188: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.188: [--dino-norm-last-layer] +100.83.37.188: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.188: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.188: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.188: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.188: [--timing-log-level {0,1,2}] +100.83.37.188: [--no-barrier-with-level-1-timing] +100.83.37.188: [--timing-log-option {max,minmax,all}] +100.83.37.188: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.188: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.188: [--log-timers-to-tensorboard] +100.83.37.188: [--log-batch-size-to-tensorboard] +100.83.37.188: [--no-log-learnig-rate-to-tensorboard] +100.83.37.188: [--no-log-loss-scale-to-tensorboard] +100.83.37.188: [--log-validation-ppl-to-tensorboard] +100.83.37.188: [--log-optimizer-states-to-tensorboard] +100.83.37.188: [--log-memory-to-tensorboard] +100.83.37.188: [--log-world-size-to-tensorboard] +100.83.37.188: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.188: [--zero-contigious-gradients] +100.83.37.188: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.188: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.188: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.188: [--scattered-embeddings] [--split-transformers] +100.83.37.188: [--memory-centric-tiled-linear] +100.83.37.188: [--tile-factor TILE_FACTOR] +100.83.37.188: [--deepspeed-activation-checkpointing] +100.83.37.188: [--partition-activations] +100.83.37.188: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.188: [--synchronize-each-layer] [--profile-backward] +100.83.37.188: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.188: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.188: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.188: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.188: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.188: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.188: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.188: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.188: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.188: [--output-bert-embeddings] +100.83.37.188: [--bert-embedder-type {megatron,huggingface}] +100.83.37.188: [--cache-fp8-weight] +100.83.37.188: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.188: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.188: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.188: [--fp8-interval FP8_INTERVAL] +100.83.37.188: [--transformer-impl {local,transformer_engine}] +100.83.37.188: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.188: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.188: [--retro-workdir RETRO_WORKDIR] +100.83.37.188: [--retro-add-retriever] +100.83.37.188: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.188: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.188: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.188: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.188: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.188: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.188: [--retro-return-doc-ids] +100.83.37.188: [--profile {pt,pt-full,hltv}] +100.83.37.188: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.188: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.188: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.188: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.188: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.188: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.188: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.188: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.188: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.188: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.188: [--expert-interval EXPERT_INTERVAL] +100.83.37.188: [--hidden-size HIDDEN_SIZE] +100.83.37.188: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.188: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.188: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.188: [--kv-channels KV_CHANNELS] +100.83.37.188: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.188: [--use-rotary-position-embeddings] +100.83.37.188: [--rotary-percent ROTARY_PERCENT] +100.83.37.188: [--no-position-embedding] +100.83.37.188: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.188: [--normalization {layernorm,rmsnorm}] +100.83.37.188: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.188: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.188: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.188: [--apply-residual-connection-post-layernorm] +100.83.37.188: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.188: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.188: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.188: [--untie-embeddings-and-output-weights] +100.83.37.188: [--embedding-weights-in-fp32] +100.83.37.188: [--fix-position-emb-redundant-alloc] +100.83.37.188: [--embed-layernorm] +100.83.37.188: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.188: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.188: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.188: [--weight-decay WEIGHT_DECAY] +100.83.37.188: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.188: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.188: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.188: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.188: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.188: [--sgd-momentum SGD_MOMENTUM] +100.83.37.188: [--do-norm-bias-weight-decay] +100.83.37.188: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.188: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.188: [--batch-size BATCH_SIZE] +100.83.37.188: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.188: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.188: [--recompute-activations] +100.83.37.188: [--recompute-granularity {full,selective}] +100.83.37.188: [--distribute-saved-activations] +100.83.37.188: [--recompute-method {uniform,block}] +100.83.37.188: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.188: [--checkpoint-activations] +100.83.37.188: [--distribute-checkpointed-activations] +100.83.37.188: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.188: [--train-iters TRAIN_ITERS] +100.83.37.188: [--train-samples TRAIN_SAMPLES] +100.83.37.188: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.188: [--log-interval LOG_INTERVAL] +100.83.37.188: [--exit-interval EXIT_INTERVAL] +100.83.37.188: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.188: [--exit-signal-handler] +100.83.37.188: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.188: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.188: [--no-bias-dropout-fusion] +100.83.37.188: [--disable-moe-token-dropping] +100.83.37.188: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.188: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.188: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.188: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.188: [--create-moe-param-group] [--use-flash-attn] +100.83.37.188: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.188: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.188: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.188: [--disable-bias-linear] +100.83.37.188: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.188: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.188: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.188: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.188: [--no-async-tensor-model-parallel-allreduce] +100.83.37.188: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.188: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.188: [--force-ds-sequence-parallel] +100.83.37.188: [--no-gradient-accumulation-fusion] +100.83.37.188: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.188: [--data-parallel-random-init] +100.83.37.188: [--init-method-std INIT_METHOD_STD] +100.83.37.188: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.188: [--lr LR] +100.83.37.188: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.188: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.188: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.188: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.188: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.188: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.188: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.188: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.188: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.188: [--override-opt_param-scheduler] +100.83.37.188: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.188: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.188: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.188: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.188: [--no-initialization] [--use-checkpoint-args] +100.83.37.188: [--exit-on-missing-checkpoint] +100.83.37.188: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.188: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.188: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.188: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.188: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.188: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.188: [--hysteresis HYSTERESIS] +100.83.37.188: [--fp32-residual-connection] +100.83.37.188: [--no-query-key-layer-scaling] +100.83.37.188: [--attention-softmax-in-fp32] +100.83.37.188: [--accumulate-allreduce-grads-in-fp32] +100.83.37.188: [--fp16-lm-cross-entropy] +100.83.37.188: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.188: [--enable-expert-tensor-parallelism] +100.83.37.188: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.188: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.188: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.188: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.188: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.188: [--overlap-p2p-communication] +100.83.37.188: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.188: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.188: [--DDP-impl {local,torch,FSDP}] +100.83.37.188: [--no-contiguous-buffers-in-local-ddp] +100.83.37.188: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.188: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.188: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.188: [--use-cpu-initialization] +100.83.37.188: [--empty-unused-memory-level {0,1,2}] +100.83.37.188: [--standalone-embedding-stage] +100.83.37.188: [--use-distributed-optimizer] +100.83.37.188: [--eval-iters EVAL_ITERS] +100.83.37.188: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.188: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.188: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.188: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.188: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.188: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.188: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.188: [--data-cache-path DATA_CACHE_PATH] +100.83.37.188: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.188: [--merge-file MERGE_FILE] +100.83.37.188: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.188: [--seq-length SEQ_LENGTH] +100.83.37.188: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.188: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.188: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.188: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.188: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.188: [--num-workers NUM_WORKERS] +100.83.37.188: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.188: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.188: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.188: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.188: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.188: [--return-data-index] +100.83.37.188: [--data-efficiency-curriculum-learning] +100.83.37.188: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.188: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.188: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.188: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.188: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.188: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.188: [--adlr-autoresume] +100.83.37.188: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.188: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.188: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.188: [--biencoder-shared-query-context-model] +100.83.37.188: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.188: [--titles-data-path TITLES_DATA_PATH] +100.83.37.188: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.188: [--use-one-sent-docs] +100.83.37.188: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.188: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.188: [--retriever-score-scaling] +100.83.37.188: [--block-data-path BLOCK_DATA_PATH] +100.83.37.188: [--embedding-path EMBEDDING_PATH] +100.83.37.188: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.188: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.188: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.188: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.188: [--patch-dim PATCH_DIM] +100.83.37.188: [--classes-fraction CLASSES_FRACTION] +100.83.37.188: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.188: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.188: [--vision-pretraining] +100.83.37.188: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.188: [--vision-backbone-type {vit,mit,swin}] +100.83.37.188: [--swin-backbone-type {tiny,base,h3}] +100.83.37.188: [--mask-type {random,row}] +100.83.37.188: [--mask-factor MASK_FACTOR] +100.83.37.188: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.188: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.188: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.188: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.188: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.188: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.188: [--dino-norm-last-layer] +100.83.37.188: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.188: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.188: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.188: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.188: [--timing-log-level {0,1,2}] +100.83.37.188: [--no-barrier-with-level-1-timing] +100.83.37.188: [--timing-log-option {max,minmax,all}] +100.83.37.188: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.188: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.188: [--log-timers-to-tensorboard] +100.83.37.188: [--log-batch-size-to-tensorboard] +100.83.37.188: [--no-log-learnig-rate-to-tensorboard] +100.83.37.188: [--no-log-loss-scale-to-tensorboard] +100.83.37.188: [--log-validation-ppl-to-tensorboard] +100.83.37.188: [--log-optimizer-states-to-tensorboard] +100.83.37.188: [--log-memory-to-tensorboard] +100.83.37.188: [--log-world-size-to-tensorboard] +100.83.37.188: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.188: [--zero-contigious-gradients] +100.83.37.188: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.188: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.188: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.188: [--scattered-embeddings] [--split-transformers] +100.83.37.188: [--memory-centric-tiled-linear] +100.83.37.188: [--tile-factor TILE_FACTOR] +100.83.37.188: [--deepspeed-activation-checkpointing] +100.83.37.188: [--partition-activations] +100.83.37.188: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.188: [--synchronize-each-layer] [--profile-backward] +100.83.37.188: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.188: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.188: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.188: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.188: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.188: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.188: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.188: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.188: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.188: [--output-bert-embeddings] +100.83.37.188: [--bert-embedder-type {megatron,huggingface}] +100.83.37.188: [--cache-fp8-weight] +100.83.37.188: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.188: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.188: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.188: [--fp8-interval FP8_INTERVAL] +100.83.37.188: [--transformer-impl {local,transformer_engine}] +100.83.37.188: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.188: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.188: [--retro-workdir RETRO_WORKDIR] +100.83.37.188: [--retro-add-retriever] +100.83.37.188: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.188: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.188: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.188: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.188: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.188: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.188: [--retro-return-doc-ids] +100.83.37.188: [--profile {pt,pt-full,hltv}] +100.83.37.188: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.188: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.188: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.188: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.188: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.188: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.188: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.188: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.188: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.188: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.188: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.188: [--expert-interval EXPERT_INTERVAL] +100.83.37.188: [--hidden-size HIDDEN_SIZE] +100.83.37.188: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.188: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.188: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.188: [--kv-channels KV_CHANNELS] +100.83.37.188: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.188: [--use-rotary-position-embeddings] +100.83.37.188: [--rotary-percent ROTARY_PERCENT] +100.83.37.188: [--no-position-embedding] +100.83.37.188: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.188: [--normalization {layernorm,rmsnorm}] +100.83.37.188: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.188: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.188: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.188: [--apply-residual-connection-post-layernorm] +100.83.37.188: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.188: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.188: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.188: [--untie-embeddings-and-output-weights] +100.83.37.188: [--embedding-weights-in-fp32] +100.83.37.188: [--fix-position-emb-redundant-alloc] +100.83.37.188: [--embed-layernorm] +100.83.37.188: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.188: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.188: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.188: [--weight-decay WEIGHT_DECAY] +100.83.37.188: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.188: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.188: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.188: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.188: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.188: [--sgd-momentum SGD_MOMENTUM] +100.83.37.188: [--do-norm-bias-weight-decay] +100.83.37.188: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.188: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.188: [--batch-size BATCH_SIZE] +100.83.37.188: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.188: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.188: [--recompute-activations] +100.83.37.188: [--recompute-granularity {full,selective}] +100.83.37.188: [--distribute-saved-activations] +100.83.37.188: [--recompute-method {uniform,block}] +100.83.37.188: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.188: [--checkpoint-activations] +100.83.37.188: [--distribute-checkpointed-activations] +100.83.37.188: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.188: [--train-iters TRAIN_ITERS] +100.83.37.188: [--train-samples TRAIN_SAMPLES] +100.83.37.188: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.188: [--log-interval LOG_INTERVAL] +100.83.37.188: [--exit-interval EXIT_INTERVAL] +100.83.37.188: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.188: [--exit-signal-handler] +100.83.37.188: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.188: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.188: [--no-bias-dropout-fusion] +100.83.37.188: [--disable-moe-token-dropping] +100.83.37.188: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.188: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.188: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.188: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.188: [--create-moe-param-group] [--use-flash-attn] +100.83.37.188: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.188: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.188: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.188: [--disable-bias-linear] +100.83.37.188: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.188: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.188: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.188: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.188: [--no-async-tensor-model-parallel-allreduce] +100.83.37.188: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.188: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.188: [--force-ds-sequence-parallel] +100.83.37.188: [--no-gradient-accumulation-fusion] +100.83.37.188: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.188: [--data-parallel-random-init] +100.83.37.188: [--init-method-std INIT_METHOD_STD] +100.83.37.188: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.188: [--lr LR] +100.83.37.188: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.188: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.188: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.188: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.188: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.188: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.188: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.188: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.188: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.188: [--override-opt_param-scheduler] +100.83.37.188: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.188: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.188: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.188: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.188: [--no-initialization] [--use-checkpoint-args] +100.83.37.188: [--exit-on-missing-checkpoint] +100.83.37.188: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.188: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.188: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.188: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.188: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.188: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.188: [--hysteresis HYSTERESIS] +100.83.37.188: [--fp32-residual-connection] +100.83.37.188: [--no-query-key-layer-scaling] +100.83.37.188: [--attention-softmax-in-fp32] +100.83.37.188: [--accumulate-allreduce-grads-in-fp32] +100.83.37.188: [--fp16-lm-cross-entropy] +100.83.37.188: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.188: [--enable-expert-tensor-parallelism] +100.83.37.188: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.188: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.188: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.188: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.188: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.188: [--overlap-p2p-communication] +100.83.37.188: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.188: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.188: [--DDP-impl {local,torch,FSDP}] +100.83.37.188: [--no-contiguous-buffers-in-local-ddp] +100.83.37.188: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.188: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.188: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.188: [--use-cpu-initialization] +100.83.37.188: [--empty-unused-memory-level {0,1,2}] +100.83.37.188: [--standalone-embedding-stage] +100.83.37.188: [--use-distributed-optimizer] +100.83.37.188: [--eval-iters EVAL_ITERS] +100.83.37.188: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.188: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.188: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.188: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.188: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.188: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.188: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.188: [--data-cache-path DATA_CACHE_PATH] +100.83.37.188: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.188: [--merge-file MERGE_FILE] +100.83.37.188: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.188: [--seq-length SEQ_LENGTH] +100.83.37.188: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.188: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.188: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.188: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.188: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.188: [--num-workers NUM_WORKERS] +100.83.37.188: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.188: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.188: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.188: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.188: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.188: [--return-data-index] +100.83.37.188: [--data-efficiency-curriculum-learning] +100.83.37.188: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.188: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.188: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.188: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.188: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.188: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.188: [--adlr-autoresume] +100.83.37.188: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.188: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.188: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.188: [--biencoder-shared-query-context-model] +100.83.37.188: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.188: [--titles-data-path TITLES_DATA_PATH] +100.83.37.188: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.188: [--use-one-sent-docs] +100.83.37.188: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.188: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.188: [--retriever-score-scaling] +100.83.37.188: [--block-data-path BLOCK_DATA_PATH] +100.83.37.188: [--embedding-path EMBEDDING_PATH] +100.83.37.188: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.188: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.188: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.188: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.188: [--patch-dim PATCH_DIM] +100.83.37.188: [--classes-fraction CLASSES_FRACTION] +100.83.37.188: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.188: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.188: [--vision-pretraining] +100.83.37.188: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.188: [--vision-backbone-type {vit,mit,swin}] +100.83.37.188: [--swin-backbone-type {tiny,base,h3}] +100.83.37.188: [--mask-type {random,row}] +100.83.37.188: [--mask-factor MASK_FACTOR] +100.83.37.188: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.188: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.188: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.188: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.188: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.188: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.188: [--dino-norm-last-layer] +100.83.37.188: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.188: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.188: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.188: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.188: [--timing-log-level {0,1,2}] +100.83.37.188: [--no-barrier-with-level-1-timing] +100.83.37.188: [--timing-log-option {max,minmax,all}] +100.83.37.188: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.188: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.188: [--log-timers-to-tensorboard] +100.83.37.188: [--log-batch-size-to-tensorboard] +100.83.37.188: [--no-log-learnig-rate-to-tensorboard] +100.83.37.188: [--no-log-loss-scale-to-tensorboard] +100.83.37.188: [--log-validation-ppl-to-tensorboard] +100.83.37.188: [--log-optimizer-states-to-tensorboard] +100.83.37.188: [--log-memory-to-tensorboard] +100.83.37.188: [--log-world-size-to-tensorboard] +100.83.37.188: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.188: [--zero-contigious-gradients] +100.83.37.188: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.188: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.188: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.188: [--scattered-embeddings] [--split-transformers] +100.83.37.188: [--memory-centric-tiled-linear] +100.83.37.188: [--tile-factor TILE_FACTOR] +100.83.37.188: [--deepspeed-activation-checkpointing] +100.83.37.188: [--partition-activations] +100.83.37.188: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.188: [--synchronize-each-layer] [--profile-backward] +100.83.37.188: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.188: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.188: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.188: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.188: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.188: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.188: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.188: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.188: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.188: [--output-bert-embeddings] +100.83.37.188: [--bert-embedder-type {megatron,huggingface}] +100.83.37.188: [--cache-fp8-weight] +100.83.37.188: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.188: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.188: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.188: [--fp8-interval FP8_INTERVAL] +100.83.37.188: [--transformer-impl {local,transformer_engine}] +100.83.37.188: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.188: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.188: [--retro-workdir RETRO_WORKDIR] +100.83.37.188: [--retro-add-retriever] +100.83.37.188: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.188: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.188: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.188: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.188: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.188: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.188: [--retro-return-doc-ids] +100.83.37.188: [--profile {pt,pt-full,hltv}] +100.83.37.188: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.188: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.188: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.188: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.188: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.175: ---------------------------------------------------------------------------------------------------- +100.83.37.175: +100.83.37.175: DeepSpeed C++/CUDA extension op reportDeepSpeed C++/CUDA extension op report +100.83.37.175: +100.83.37.175: ---------------------------------------------------------------------------------------------------- +100.83.37.175: +100.83.37.175: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.175: runtime if needed. Op compatibility means that your system +100.83.37.175: meet the required dependencies to JIT install the op.NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.175: runtime if needed. Op compatibility means that your system +100.83.37.175: meet the required dependencies to JIT install the op. +100.83.37.175: +100.83.37.175: ---------------------------------------------------------------------------------------------------- +100.83.37.175: +100.83.37.175: JIT compiled ops requires ninjaJIT compiled ops requires ninja +100.83.37.175: +100.83.37.175: ninja .................. [OKAY] +100.83.37.175: ninja-------------------------------------------------- +100.83.37.175: ..................op name [OKAY]................ +100.83.37.175: installed-------------------------------------------------- +100.83.37.175: .. compatibleop name +100.83.37.175: ................-------------------------------------------------- +100.83.37.175: installed .. compatible +100.83.37.175: -------------------------------------------------- +100.83.37.175: cpu_adam ............... [NO] ....... [OKAY]cpu_adam +100.83.37.175: ............... [NO] fused_adam....... .............[OKAY] +100.83.37.175: [NO] ....... fused_adam[OKAY] +100.83.37.175: ............. [NO]deepspeed_not_implemented ....... [OKAY][NO] +100.83.37.175: ....... deepspeed_not_implemented[OKAY] +100.83.37.175: [NO] transformer_inference....... ..[OKAY] +100.83.37.175: [NO] .......transformer_inference [OKAY].. +100.83.37.175: [NO]-------------------------------------------------- +100.83.37.175: ....... [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed general environment info:DeepSpeed general environment info: +100.83.37.175: +100.83.37.175: torch install pathtorch install path .............................. ['/usr/local/lib/python3.10/dist-packages/torch']['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.175: +100.83.37.175: torch versiontorch version ........................................ 2.1.1a0+gitb51c9f62.1.1a0+gitb51c9f6 +100.83.37.175: +100.83.37.175: deepspeed install pathdeepspeed install path ...................... ['/usr/local/lib/python3.10/dist-packages/deepspeed']['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.175: +100.83.37.175: deepspeed infodeepspeed info ...................................... 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.00.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.175: +100.83.37.175: deepspeed wheel compiled w.deepspeed wheel compiled w. ............ torch 2.1 torch 2.1 +100.83.37.175: +100.83.37.175: shared memory (/dev/shm) sizeshared memory (/dev/shm) size ........ 503.75 GB503.75 GB +100.83.37.175: +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed C++/CUDA extension op report +100.83.37.175: -------------------------------------------------- +100.83.37.175: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.175: runtime if needed. Op compatibility means that your system +100.83.37.175: meet the required dependencies to JIT install the op. +100.83.37.175: -------------------------------------------------- +100.83.37.175: JIT compiled ops requires ninja +100.83.37.175: ninja .................. [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: op name ................ installed .. compatible +100.83.37.175: -------------------------------------------------- +100.83.37.175: cpu_adam ............... [NO] ....... [OKAY] +100.83.37.175: fused_adam ............. [NO] ....... [OKAY] +100.83.37.175: deepspeed_not_implemented [NO] ....... [OKAY] +100.83.37.175: transformer_inference .. [NO] ....... [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed general environment info: +100.83.37.175: torch install path ............... ['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.175: torch version .................... 2.1.1a0+gitb51c9f6 +100.83.37.175: deepspeed install path ........... ['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.175: deepspeed info ................... 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.175: deepspeed wheel compiled w. ...... torch 2.1 +100.83.37.175: shared memory (/dev/shm) size .... 503.75 GB +100.83.37.175: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.175: To add an exception for this directory, call: +100.83.37.175: +100.83.37.175: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.175: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.175: To add an exception for this directory, call: +100.83.37.175: +100.83.37.175: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.175: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.175: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.175: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.175: To add an exception for this directory, call: +100.83.37.175: +100.83.37.175: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.175: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed C++/CUDA extension op report +100.83.37.175: -------------------------------------------------- +100.83.37.175: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.175: runtime if needed. Op compatibility means that your system +100.83.37.175: meet the required dependencies to JIT install the op. +100.83.37.175: -------------------------------------------------- +100.83.37.175: JIT compiled ops requires ninja +100.83.37.175: ninja .................. [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: op name ................ installed .. compatible +100.83.37.175: -------------------------------------------------- +100.83.37.175: cpu_adam ............... [NO] ....... [OKAY] +100.83.37.175: fused_adam ............. [NO] ....... [OKAY] +100.83.37.175: deepspeed_not_implemented [NO] ....... [OKAY] +100.83.37.175: transformer_inference .. [NO] ....... [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed general environment info: +100.83.37.175: torch install path ............... ['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.175: torch version .................... 2.1.1a0+gitb51c9f6 +100.83.37.175: deepspeed install path ........... ['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.175: deepspeed info ................... 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.175: deepspeed wheel compiled w. ...... torch 2.1 +100.83.37.175: shared memory (/dev/shm) size .... 503.75 GB +100.83.37.175: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.175: To add an exception for this directory, call: +100.83.37.175: +100.83.37.175: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.175: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.175: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.175: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.175: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.175: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.175: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.175: [--expert-interval EXPERT_INTERVAL] +100.83.37.175: [--hidden-size HIDDEN_SIZE] +100.83.37.175: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.175: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.175: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.175: [--kv-channels KV_CHANNELS] +100.83.37.175: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.175: [--use-rotary-position-embeddings] +100.83.37.175: [--rotary-percent ROTARY_PERCENT] +100.83.37.175: [--no-position-embedding] +100.83.37.175: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.175: [--normalization {layernorm,rmsnorm}] +100.83.37.175: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.175: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.175: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.175: [--apply-residual-connection-post-layernorm] +100.83.37.175: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.175: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.175: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.175: [--untie-embeddings-and-output-weights] +100.83.37.175: [--embedding-weights-in-fp32] +100.83.37.175: [--fix-position-emb-redundant-alloc] +100.83.37.175: [--embed-layernorm] +100.83.37.175: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.175: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.175: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.175: [--weight-decay WEIGHT_DECAY] +100.83.37.175: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.175: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.175: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.175: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.175: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.175: [--sgd-momentum SGD_MOMENTUM] +100.83.37.175: [--do-norm-bias-weight-decay] +100.83.37.175: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.175: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.175: [--batch-size BATCH_SIZE] +100.83.37.175: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.175: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.175: [--recompute-activations] +100.83.37.175: [--recompute-granularity {full,selective}] +100.83.37.175: [--distribute-saved-activations] +100.83.37.175: [--recompute-method {uniform,block}] +100.83.37.175: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.175: [--checkpoint-activations] +100.83.37.175: [--distribute-checkpointed-activations] +100.83.37.175: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.175: [--train-iters TRAIN_ITERS] +100.83.37.175: [--train-samples TRAIN_SAMPLES] +100.83.37.175: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.175: [--log-interval LOG_INTERVAL] +100.83.37.175: [--exit-interval EXIT_INTERVAL] +100.83.37.175: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.175: [--exit-signal-handler] +100.83.37.175: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.175: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.175: [--no-bias-dropout-fusion] +100.83.37.175: [--disable-moe-token-dropping] +100.83.37.175: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.175: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.175: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.175: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.175: [--create-moe-param-group] [--use-flash-attn] +100.83.37.175: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.175: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.175: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.175: [--disable-bias-linear] +100.83.37.175: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.175: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.175: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.175: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.175: [--no-async-tensor-model-parallel-allreduce] +100.83.37.175: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.175: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.175: [--force-ds-sequence-parallel] +100.83.37.175: [--no-gradient-accumulation-fusion] +100.83.37.175: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.175: [--data-parallel-random-init] +100.83.37.175: [--init-method-std INIT_METHOD_STD] +100.83.37.175: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.175: [--lr LR] +100.83.37.175: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.175: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.175: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.175: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.175: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.175: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.175: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.175: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.175: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.175: [--override-opt_param-scheduler] +100.83.37.175: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.175: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.175: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.175: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.175: [--no-initialization] [--use-checkpoint-args] +100.83.37.175: [--exit-on-missing-checkpoint] +100.83.37.175: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.175: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.175: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.175: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.175: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.175: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.175: [--hysteresis HYSTERESIS] +100.83.37.175: [--fp32-residual-connection] +100.83.37.175: [--no-query-key-layer-scaling] +100.83.37.175: [--attention-softmax-in-fp32] +100.83.37.175: [--accumulate-allreduce-grads-in-fp32] +100.83.37.175: [--fp16-lm-cross-entropy] +100.83.37.175: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.175: [--enable-expert-tensor-parallelism] +100.83.37.175: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.175: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.175: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.175: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.175: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.175: [--overlap-p2p-communication] +100.83.37.175: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.175: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.175: [--DDP-impl {local,torch,FSDP}] +100.83.37.175: [--no-contiguous-buffers-in-local-ddp] +100.83.37.175: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.175: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.175: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.175: [--use-cpu-initialization] +100.83.37.175: [--empty-unused-memory-level {0,1,2}] +100.83.37.175: [--standalone-embedding-stage] +100.83.37.175: [--use-distributed-optimizer] +100.83.37.175: [--eval-iters EVAL_ITERS] +100.83.37.175: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.175: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.175: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.175: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.175: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.175: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.175: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.175: [--data-cache-path DATA_CACHE_PATH] +100.83.37.175: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.175: [--merge-file MERGE_FILE] +100.83.37.175: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.175: [--seq-length SEQ_LENGTH] +100.83.37.175: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.175: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.175: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.175: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.175: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.175: [--num-workers NUM_WORKERS] +100.83.37.175: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.175: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.175: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.175: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.175: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.175: [--return-data-index] +100.83.37.175: [--data-efficiency-curriculum-learning] +100.83.37.175: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.175: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.175: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.175: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.175: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.175: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.175: [--adlr-autoresume] +100.83.37.175: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.175: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.175: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.175: [--biencoder-shared-query-context-model] +100.83.37.175: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.175: [--titles-data-path TITLES_DATA_PATH] +100.83.37.175: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.175: [--use-one-sent-docs] +100.83.37.175: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.175: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.175: [--retriever-score-scaling] +100.83.37.175: [--block-data-path BLOCK_DATA_PATH] +100.83.37.175: [--embedding-path EMBEDDING_PATH] +100.83.37.175: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.175: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.175: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.175: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.175: [--patch-dim PATCH_DIM] +100.83.37.175: [--classes-fraction CLASSES_FRACTION] +100.83.37.175: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.175: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.175: [--vision-pretraining] +100.83.37.175: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.175: [--vision-backbone-type {vit,mit,swin}] +100.83.37.175: [--swin-backbone-type {tiny,base,h3}] +100.83.37.175: [--mask-type {random,row}] +100.83.37.175: [--mask-factor MASK_FACTOR] +100.83.37.175: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.175: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.175: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.175: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.175: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.175: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.175: [--dino-norm-last-layer] +100.83.37.175: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.175: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.175: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.175: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.175: [--timing-log-level {0,1,2}] +100.83.37.175: [--no-barrier-with-level-1-timing] +100.83.37.175: [--timing-log-option {max,minmax,all}] +100.83.37.175: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.175: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.175: [--log-timers-to-tensorboard] +100.83.37.175: [--log-batch-size-to-tensorboard] +100.83.37.175: [--no-log-learnig-rate-to-tensorboard] +100.83.37.175: [--no-log-loss-scale-to-tensorboard] +100.83.37.175: [--log-validation-ppl-to-tensorboard] +100.83.37.175: [--log-optimizer-states-to-tensorboard] +100.83.37.175: [--log-memory-to-tensorboard] +100.83.37.175: [--log-world-size-to-tensorboard] +100.83.37.175: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.175: [--zero-contigious-gradients] +100.83.37.175: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.175: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.175: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.175: [--scattered-embeddings] [--split-transformers] +100.83.37.175: [--memory-centric-tiled-linear] +100.83.37.175: [--tile-factor TILE_FACTOR] +100.83.37.175: [--deepspeed-activation-checkpointing] +100.83.37.175: [--partition-activations] +100.83.37.175: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.175: [--synchronize-each-layer] [--profile-backward] +100.83.37.175: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.175: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.175: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.175: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.175: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.175: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.175: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.175: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.175: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.175: [--output-bert-embeddings] +100.83.37.175: [--bert-embedder-type {megatron,huggingface}] +100.83.37.175: [--cache-fp8-weight] +100.83.37.175: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.175: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.175: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.175: [--fp8-interval FP8_INTERVAL] +100.83.37.175: [--transformer-impl {local,transformer_engine}] +100.83.37.175: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.175: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.175: [--retro-workdir RETRO_WORKDIR] +100.83.37.175: [--retro-add-retriever] +100.83.37.175: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.175: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.175: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.175: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.175: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.175: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.175: [--retro-return-doc-ids] +100.83.37.175: [--profile {pt,pt-full,hltv}] +100.83.37.175: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.175: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.175: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.175: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.175: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.175: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.175: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.175: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.175: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.175: [--expert-interval EXPERT_INTERVAL] +100.83.37.175: [--hidden-size HIDDEN_SIZE] +100.83.37.175: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.175: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.175: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.175: [--kv-channels KV_CHANNELS] +100.83.37.175: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.175: [--use-rotary-position-embeddings] +100.83.37.175: [--rotary-percent ROTARY_PERCENT] +100.83.37.175: [--no-position-embedding] +100.83.37.175: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.175: [--normalization {layernorm,rmsnorm}] +100.83.37.175: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.175: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.175: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.175: [--apply-residual-connection-post-layernorm] +100.83.37.175: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.175: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.175: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.175: [--untie-embeddings-and-output-weights] +100.83.37.175: [--embedding-weights-in-fp32] +100.83.37.175: [--fix-position-emb-redundant-alloc] +100.83.37.175: [--embed-layernorm] +100.83.37.175: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.175: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.175: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.175: [--weight-decay WEIGHT_DECAY] +100.83.37.175: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.175: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.175: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.175: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.175: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.175: [--sgd-momentum SGD_MOMENTUM] +100.83.37.175: [--do-norm-bias-weight-decay] +100.83.37.175: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.175: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.175: [--batch-size BATCH_SIZE] +100.83.37.175: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.175: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.175: [--recompute-activations] +100.83.37.175: [--recompute-granularity {full,selective}] +100.83.37.175: [--distribute-saved-activations] +100.83.37.175: [--recompute-method {uniform,block}] +100.83.37.175: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.175: [--checkpoint-activations] +100.83.37.175: [--distribute-checkpointed-activations] +100.83.37.175: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.175: [--train-iters TRAIN_ITERS] +100.83.37.175: [--train-samples TRAIN_SAMPLES] +100.83.37.175: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.175: [--log-interval LOG_INTERVAL] +100.83.37.175: [--exit-interval EXIT_INTERVAL] +100.83.37.175: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.175: [--exit-signal-handler] +100.83.37.175: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.175: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.175: [--no-bias-dropout-fusion] +100.83.37.175: [--disable-moe-token-dropping] +100.83.37.175: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.175: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.175: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.175: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.175: [--create-moe-param-group] [--use-flash-attn] +100.83.37.175: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.175: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.175: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.175: [--disable-bias-linear] +100.83.37.175: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.175: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.175: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.175: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.175: [--no-async-tensor-model-parallel-allreduce] +100.83.37.175: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.175: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.175: [--force-ds-sequence-parallel] +100.83.37.175: [--no-gradient-accumulation-fusion] +100.83.37.175: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.175: [--data-parallel-random-init] +100.83.37.175: [--init-method-std INIT_METHOD_STD] +100.83.37.175: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.175: [--lr LR] +100.83.37.175: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.175: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.175: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.175: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.175: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.175: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.175: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.175: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.175: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.175: [--override-opt_param-scheduler] +100.83.37.175: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.175: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.175: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.175: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.175: [--no-initialization] [--use-checkpoint-args] +100.83.37.175: [--exit-on-missing-checkpoint] +100.83.37.175: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.175: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.175: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.175: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.175: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.175: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.175: [--hysteresis HYSTERESIS] +100.83.37.175: [--fp32-residual-connection] +100.83.37.175: [--no-query-key-layer-scaling] +100.83.37.175: [--attention-softmax-in-fp32] +100.83.37.175: [--accumulate-allreduce-grads-in-fp32] +100.83.37.175: [--fp16-lm-cross-entropy] +100.83.37.175: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.175: [--enable-expert-tensor-parallelism] +100.83.37.175: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.175: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.175: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.175: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.175: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.175: [--overlap-p2p-communication] +100.83.37.175: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.175: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.175: [--DDP-impl {local,torch,FSDP}] +100.83.37.175: [--no-contiguous-buffers-in-local-ddp] +100.83.37.175: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.175: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.175: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.175: [--use-cpu-initialization] +100.83.37.175: [--empty-unused-memory-level {0,1,2}] +100.83.37.175: [--standalone-embedding-stage] +100.83.37.175: [--use-distributed-optimizer] +100.83.37.175: [--eval-iters EVAL_ITERS] +100.83.37.175: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.175: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.175: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.175: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.175: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.175: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.175: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.175: [--data-cache-path DATA_CACHE_PATH] +100.83.37.175: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.175: [--merge-file MERGE_FILE] +100.83.37.175: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.175: [--seq-length SEQ_LENGTH] +100.83.37.175: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.175: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.175: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.175: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.175: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.175: [--num-workers NUM_WORKERS] +100.83.37.175: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.175: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.175: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.175: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.175: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.175: [--return-data-index] +100.83.37.175: [--data-efficiency-curriculum-learning] +100.83.37.175: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.175: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.175: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.175: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.175: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.175: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.175: [--adlr-autoresume] +100.83.37.175: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.175: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.175: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.175: [--biencoder-shared-query-context-model] +100.83.37.175: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.175: [--titles-data-path TITLES_DATA_PATH] +100.83.37.175: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.175: [--use-one-sent-docs] +100.83.37.175: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.175: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.175: [--retriever-score-scaling] +100.83.37.175: [--block-data-path BLOCK_DATA_PATH] +100.83.37.175: [--embedding-path EMBEDDING_PATH] +100.83.37.175: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.175: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.175: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.175: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.175: [--patch-dim PATCH_DIM] +100.83.37.175: [--classes-fraction CLASSES_FRACTION] +100.83.37.175: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.175: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.175: [--vision-pretraining] +100.83.37.175: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.175: [--vision-backbone-type {vit,mit,swin}] +100.83.37.175: [--swin-backbone-type {tiny,base,h3}] +100.83.37.175: [--mask-type {random,row}] +100.83.37.175: [--mask-factor MASK_FACTOR] +100.83.37.175: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.175: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.175: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.175: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.175: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.175: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.175: [--dino-norm-last-layer] +100.83.37.175: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.175: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.175: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.175: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.175: [--timing-log-level {0,1,2}] +100.83.37.175: [--no-barrier-with-level-1-timing] +100.83.37.175: [--timing-log-option {max,minmax,all}] +100.83.37.175: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.175: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.175: [--log-timers-to-tensorboard] +100.83.37.175: [--log-batch-size-to-tensorboard] +100.83.37.175: [--no-log-learnig-rate-to-tensorboard] +100.83.37.175: [--no-log-loss-scale-to-tensorboard] +100.83.37.175: [--log-validation-ppl-to-tensorboard] +100.83.37.175: [--log-optimizer-states-to-tensorboard] +100.83.37.175: [--log-memory-to-tensorboard] +100.83.37.175: [--log-world-size-to-tensorboard] +100.83.37.175: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.175: [--zero-contigious-gradients] +100.83.37.175: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.175: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.175: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.175: [--scattered-embeddings] [--split-transformers] +100.83.37.175: [--memory-centric-tiled-linear] +100.83.37.175: [--tile-factor TILE_FACTOR] +100.83.37.175: [--deepspeed-activation-checkpointing] +100.83.37.175: [--partition-activations] +100.83.37.175: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.175: [--synchronize-each-layer] [--profile-backward] +100.83.37.175: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.175: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.175: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.175: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.175: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.175: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.175: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.175: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.175: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.175: [--output-bert-embeddings] +100.83.37.175: [--bert-embedder-type {megatron,huggingface}] +100.83.37.175: [--cache-fp8-weight] +100.83.37.175: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.175: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.175: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.175: [--fp8-interval FP8_INTERVAL] +100.83.37.175: [--transformer-impl {local,transformer_engine}] +100.83.37.175: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.175: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.175: [--retro-workdir RETRO_WORKDIR] +100.83.37.175: [--retro-add-retriever] +100.83.37.175: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.175: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.175: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.175: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.175: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.175: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.175: [--retro-return-doc-ids] +100.83.37.175: [--profile {pt,pt-full,hltv}] +100.83.37.175: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.175: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.175: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.175: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.175: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.175: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed C++/CUDA extension op report +100.83.37.175: -------------------------------------------------- +100.83.37.175: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.175: runtime if needed. Op compatibility means that your system +100.83.37.175: meet the required dependencies to JIT install the op. +100.83.37.175: -------------------------------------------------- +100.83.37.175: JIT compiled ops requires ninja +100.83.37.175: ninja .................. [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: op name ................ installed .. compatible +100.83.37.175: -------------------------------------------------- +100.83.37.175: cpu_adam ............... [NO] ....... [OKAY] +100.83.37.175: fused_adam ............. [NO] ....... [OKAY] +100.83.37.175: deepspeed_not_implemented [NO] ....... [OKAY] +100.83.37.175: transformer_inference .. [NO] ....... [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed general environment info: +100.83.37.175: torch install path ............... ['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.175: torch version .................... 2.1.1a0+gitb51c9f6 +100.83.37.175: deepspeed install path ........... ['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.175: deepspeed info ................... 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.175: deepspeed wheel compiled w. ...... torch 2.1 +100.83.37.175: shared memory (/dev/shm) size .... 503.75 GB +100.83.37.175: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.175: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.175: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.175: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.175: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.175: [--expert-interval EXPERT_INTERVAL] +100.83.37.175: [--hidden-size HIDDEN_SIZE] +100.83.37.175: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.175: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.175: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.175: [--kv-channels KV_CHANNELS] +100.83.37.175: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.175: [--use-rotary-position-embeddings] +100.83.37.175: [--rotary-percent ROTARY_PERCENT] +100.83.37.175: [--no-position-embedding] +100.83.37.175: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.175: [--normalization {layernorm,rmsnorm}] +100.83.37.175: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.175: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.175: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.175: [--apply-residual-connection-post-layernorm] +100.83.37.175: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.175: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.175: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.175: [--untie-embeddings-and-output-weights] +100.83.37.175: [--embedding-weights-in-fp32] +100.83.37.175: [--fix-position-emb-redundant-alloc] +100.83.37.175: [--embed-layernorm] +100.83.37.175: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.175: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.175: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.175: [--weight-decay WEIGHT_DECAY] +100.83.37.175: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.175: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.175: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.175: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.175: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.175: [--sgd-momentum SGD_MOMENTUM] +100.83.37.175: [--do-norm-bias-weight-decay] +100.83.37.175: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.175: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.175: [--batch-size BATCH_SIZE] +100.83.37.175: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.175: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.175: [--recompute-activations] +100.83.37.175: [--recompute-granularity {full,selective}] +100.83.37.175: [--distribute-saved-activations] +100.83.37.175: [--recompute-method {uniform,block}] +100.83.37.175: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.175: [--checkpoint-activations] +100.83.37.175: [--distribute-checkpointed-activations] +100.83.37.175: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.175: [--train-iters TRAIN_ITERS] +100.83.37.175: [--train-samples TRAIN_SAMPLES] +100.83.37.175: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.175: [--log-interval LOG_INTERVAL] +100.83.37.175: [--exit-interval EXIT_INTERVAL] +100.83.37.175: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.175: [--exit-signal-handler] +100.83.37.175: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.175: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.175: [--no-bias-dropout-fusion] +100.83.37.175: [--disable-moe-token-dropping] +100.83.37.175: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.175: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.175: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.175: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.175: [--create-moe-param-group] [--use-flash-attn] +100.83.37.175: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.175: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.175: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.175: [--disable-bias-linear] +100.83.37.175: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.175: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.175: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.175: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.175: [--no-async-tensor-model-parallel-allreduce] +100.83.37.175: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.175: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.175: [--force-ds-sequence-parallel] +100.83.37.175: [--no-gradient-accumulation-fusion] +100.83.37.175: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.175: [--data-parallel-random-init] +100.83.37.175: [--init-method-std INIT_METHOD_STD] +100.83.37.175: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.175: [--lr LR] +100.83.37.175: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.175: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.175: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.175: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.175: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.175: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.175: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.175: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.175: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.175: [--override-opt_param-scheduler] +100.83.37.175: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.175: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.175: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.175: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.175: [--no-initialization] [--use-checkpoint-args] +100.83.37.175: [--exit-on-missing-checkpoint] +100.83.37.175: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.175: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.175: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.175: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.175: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.175: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.175: [--hysteresis HYSTERESIS] +100.83.37.175: [--fp32-residual-connection] +100.83.37.175: [--no-query-key-layer-scaling] +100.83.37.175: [--attention-softmax-in-fp32] +100.83.37.175: [--accumulate-allreduce-grads-in-fp32] +100.83.37.175: [--fp16-lm-cross-entropy] +100.83.37.175: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.175: [--enable-expert-tensor-parallelism] +100.83.37.175: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.175: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.175: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.175: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.175: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.175: [--overlap-p2p-communication] +100.83.37.175: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.175: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.175: [--DDP-impl {local,torch,FSDP}] +100.83.37.175: [--no-contiguous-buffers-in-local-ddp] +100.83.37.175: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.175: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.175: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.175: [--use-cpu-initialization] +100.83.37.175: [--empty-unused-memory-level {0,1,2}] +100.83.37.175: [--standalone-embedding-stage] +100.83.37.175: [--use-distributed-optimizer] +100.83.37.175: [--eval-iters EVAL_ITERS] +100.83.37.175: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.175: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.175: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.175: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.175: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.175: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.175: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.175: [--data-cache-path DATA_CACHE_PATH] +100.83.37.175: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.175: [--merge-file MERGE_FILE] +100.83.37.175: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.175: [--seq-length SEQ_LENGTH] +100.83.37.175: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.175: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.175: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.175: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.175: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.175: [--num-workers NUM_WORKERS] +100.83.37.175: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.175: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.175: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.175: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.175: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.175: [--return-data-index] +100.83.37.175: [--data-efficiency-curriculum-learning] +100.83.37.175: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.175: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.175: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.175: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.175: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.175: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.175: [--adlr-autoresume] +100.83.37.175: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.175: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.175: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.175: [--biencoder-shared-query-context-model] +100.83.37.175: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.175: [--titles-data-path TITLES_DATA_PATH] +100.83.37.175: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.175: [--use-one-sent-docs] +100.83.37.175: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.175: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.175: [--retriever-score-scaling] +100.83.37.175: [--block-data-path BLOCK_DATA_PATH] +100.83.37.175: [--embedding-path EMBEDDING_PATH] +100.83.37.175: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.175: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.175: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.175: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.175: [--patch-dim PATCH_DIM] +100.83.37.175: [--classes-fraction CLASSES_FRACTION] +100.83.37.175: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.175: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.175: [--vision-pretraining] +100.83.37.175: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.175: [--vision-backbone-type {vit,mit,swin}] +100.83.37.175: [--swin-backbone-type {tiny,base,h3}] +100.83.37.175: [--mask-type {random,row}] +100.83.37.175: [--mask-factor MASK_FACTOR] +100.83.37.175: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.175: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.175: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.175: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.175: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.175: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.175: [--dino-norm-last-layer] +100.83.37.175: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.175: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.175: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.175: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.175: [--timing-log-level {0,1,2}] +100.83.37.175: [--no-barrier-with-level-1-timing] +100.83.37.175: [--timing-log-option {max,minmax,all}] +100.83.37.175: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.175: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.175: [--log-timers-to-tensorboard] +100.83.37.175: [--log-batch-size-to-tensorboard] +100.83.37.175: [--no-log-learnig-rate-to-tensorboard] +100.83.37.175: [--no-log-loss-scale-to-tensorboard] +100.83.37.175: [--log-validation-ppl-to-tensorboard] +100.83.37.175: [--log-optimizer-states-to-tensorboard] +100.83.37.175: [--log-memory-to-tensorboard] +100.83.37.175: [--log-world-size-to-tensorboard] +100.83.37.175: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.175: [--zero-contigious-gradients] +100.83.37.175: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.175: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.175: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.175: [--scattered-embeddings] [--split-transformers] +100.83.37.175: [--memory-centric-tiled-linear] +100.83.37.175: [--tile-factor TILE_FACTOR] +100.83.37.175: [--deepspeed-activation-checkpointing] +100.83.37.175: [--partition-activations] +100.83.37.175: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.175: [--synchronize-each-layer] [--profile-backward] +100.83.37.175: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.175: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.175: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.175: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.175: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.175: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.175: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.175: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.175: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.175: [--output-bert-embeddings] +100.83.37.175: [--bert-embedder-type {megatron,huggingface}] +100.83.37.175: [--cache-fp8-weight] +100.83.37.175: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.175: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.175: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.175: [--fp8-interval FP8_INTERVAL] +100.83.37.175: [--transformer-impl {local,transformer_engine}] +100.83.37.175: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.175: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.175: [--retro-workdir RETRO_WORKDIR] +100.83.37.175: [--retro-add-retriever] +100.83.37.175: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.175: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.175: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.175: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.175: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.175: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.175: [--retro-return-doc-ids] +100.83.37.175: [--profile {pt,pt-full,hltv}] +100.83.37.175: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.175: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.175: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.175: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.175: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.175: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.175: To add an exception for this directory, call: +100.83.37.175: +100.83.37.175: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.175: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.175: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.175: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.175: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.175: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.175: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.175: [--expert-interval EXPERT_INTERVAL] +100.83.37.175: [--hidden-size HIDDEN_SIZE] +100.83.37.175: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.175: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.175: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.175: [--kv-channels KV_CHANNELS] +100.83.37.175: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.175: [--use-rotary-position-embeddings] +100.83.37.175: [--rotary-percent ROTARY_PERCENT] +100.83.37.175: [--no-position-embedding] +100.83.37.175: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.175: [--normalization {layernorm,rmsnorm}] +100.83.37.175: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.175: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.175: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.175: [--apply-residual-connection-post-layernorm] +100.83.37.175: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.175: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.175: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.175: [--untie-embeddings-and-output-weights] +100.83.37.175: [--embedding-weights-in-fp32] +100.83.37.175: [--fix-position-emb-redundant-alloc] +100.83.37.175: [--embed-layernorm] +100.83.37.175: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.175: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.175: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.175: [--weight-decay WEIGHT_DECAY] +100.83.37.175: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.175: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.175: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.175: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.175: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.175: [--sgd-momentum SGD_MOMENTUM] +100.83.37.175: [--do-norm-bias-weight-decay] +100.83.37.175: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.175: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.175: [--batch-size BATCH_SIZE] +100.83.37.175: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.175: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.175: [--recompute-activations] +100.83.37.175: [--recompute-granularity {full,selective}] +100.83.37.175: [--distribute-saved-activations] +100.83.37.175: [--recompute-method {uniform,block}] +100.83.37.175: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.175: [--checkpoint-activations] +100.83.37.175: [--distribute-checkpointed-activations] +100.83.37.175: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.175: [--train-iters TRAIN_ITERS] +100.83.37.175: [--train-samples TRAIN_SAMPLES] +100.83.37.175: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.175: [--log-interval LOG_INTERVAL] +100.83.37.175: [--exit-interval EXIT_INTERVAL] +100.83.37.175: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.175: [--exit-signal-handler] +100.83.37.175: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.175: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.175: [--no-bias-dropout-fusion] +100.83.37.175: [--disable-moe-token-dropping] +100.83.37.175: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.175: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.175: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.175: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.175: [--create-moe-param-group] [--use-flash-attn] +100.83.37.175: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.175: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.175: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.175: [--disable-bias-linear] +100.83.37.175: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.175: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.175: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.175: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.175: [--no-async-tensor-model-parallel-allreduce] +100.83.37.175: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.175: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.175: [--force-ds-sequence-parallel] +100.83.37.175: [--no-gradient-accumulation-fusion] +100.83.37.175: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.175: [--data-parallel-random-init] +100.83.37.175: [--init-method-std INIT_METHOD_STD] +100.83.37.175: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.175: [--lr LR] +100.83.37.175: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.175: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.175: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.175: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.175: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.175: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.175: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.175: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.175: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.175: [--override-opt_param-scheduler] +100.83.37.175: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.175: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.175: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.175: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.175: [--no-initialization] [--use-checkpoint-args] +100.83.37.175: [--exit-on-missing-checkpoint] +100.83.37.175: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.175: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.175: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.175: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.175: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.175: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.175: [--hysteresis HYSTERESIS] +100.83.37.175: [--fp32-residual-connection] +100.83.37.175: [--no-query-key-layer-scaling] +100.83.37.175: [--attention-softmax-in-fp32] +100.83.37.175: [--accumulate-allreduce-grads-in-fp32] +100.83.37.175: [--fp16-lm-cross-entropy] +100.83.37.175: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.175: [--enable-expert-tensor-parallelism] +100.83.37.175: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.175: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.175: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.175: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.175: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.175: [--overlap-p2p-communication] +100.83.37.175: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.175: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.175: [--DDP-impl {local,torch,FSDP}] +100.83.37.175: [--no-contiguous-buffers-in-local-ddp] +100.83.37.175: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.175: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.175: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.175: [--use-cpu-initialization] +100.83.37.175: [--empty-unused-memory-level {0,1,2}] +100.83.37.175: [--standalone-embedding-stage] +100.83.37.175: [--use-distributed-optimizer] +100.83.37.175: [--eval-iters EVAL_ITERS] +100.83.37.175: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.175: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.175: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.175: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.175: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.175: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.175: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.175: [--data-cache-path DATA_CACHE_PATH] +100.83.37.175: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.175: [--merge-file MERGE_FILE] +100.83.37.175: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.175: [--seq-length SEQ_LENGTH] +100.83.37.175: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.175: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.175: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.175: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.175: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.175: [--num-workers NUM_WORKERS] +100.83.37.175: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.175: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.175: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.175: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.175: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.175: [--return-data-index] +100.83.37.175: [--data-efficiency-curriculum-learning] +100.83.37.175: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.175: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.175: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.175: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.175: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.175: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.175: [--adlr-autoresume] +100.83.37.175: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.175: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.175: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.175: [--biencoder-shared-query-context-model] +100.83.37.175: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.175: [--titles-data-path TITLES_DATA_PATH] +100.83.37.175: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.175: [--use-one-sent-docs] +100.83.37.175: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.175: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.175: [--retriever-score-scaling] +100.83.37.175: [--block-data-path BLOCK_DATA_PATH] +100.83.37.175: [--embedding-path EMBEDDING_PATH] +100.83.37.175: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.175: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.175: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.175: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.175: [--patch-dim PATCH_DIM] +100.83.37.175: [--classes-fraction CLASSES_FRACTION] +100.83.37.175: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.175: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.175: [--vision-pretraining] +100.83.37.175: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.175: [--vision-backbone-type {vit,mit,swin}] +100.83.37.175: [--swin-backbone-type {tiny,base,h3}] +100.83.37.175: [--mask-type {random,row}] +100.83.37.175: [--mask-factor MASK_FACTOR] +100.83.37.175: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.175: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.175: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.175: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.175: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.175: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.175: [--dino-norm-last-layer] +100.83.37.175: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.175: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.175: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.175: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.175: [--timing-log-level {0,1,2}] +100.83.37.175: [--no-barrier-with-level-1-timing] +100.83.37.175: [--timing-log-option {max,minmax,all}] +100.83.37.175: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.175: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.175: [--log-timers-to-tensorboard] +100.83.37.175: [--log-batch-size-to-tensorboard] +100.83.37.175: [--no-log-learnig-rate-to-tensorboard] +100.83.37.175: [--no-log-loss-scale-to-tensorboard] +100.83.37.175: [--log-validation-ppl-to-tensorboard] +100.83.37.175: [--log-optimizer-states-to-tensorboard] +100.83.37.175: [--log-memory-to-tensorboard] +100.83.37.175: [--log-world-size-to-tensorboard] +100.83.37.175: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.175: [--zero-contigious-gradients] +100.83.37.175: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.175: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.175: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.175: [--scattered-embeddings] [--split-transformers] +100.83.37.175: [--memory-centric-tiled-linear] +100.83.37.175: [--tile-factor TILE_FACTOR] +100.83.37.175: [--deepspeed-activation-checkpointing] +100.83.37.175: [--partition-activations] +100.83.37.175: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.175: [--synchronize-each-layer] [--profile-backward] +100.83.37.175: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.175: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.175: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.175: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.175: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.175: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.175: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.175: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.175: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.175: [--output-bert-embeddings] +100.83.37.175: [--bert-embedder-type {megatron,huggingface}] +100.83.37.175: [--cache-fp8-weight] +100.83.37.175: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.175: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.175: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.175: [--fp8-interval FP8_INTERVAL] +100.83.37.175: [--transformer-impl {local,transformer_engine}] +100.83.37.175: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.175: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.175: [--retro-workdir RETRO_WORKDIR] +100.83.37.175: [--retro-add-retriever] +100.83.37.175: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.175: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.175: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.175: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.175: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.175: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.175: [--retro-return-doc-ids] +100.83.37.175: [--profile {pt,pt-full,hltv}] +100.83.37.175: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.175: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.175: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.175: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.175: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.175: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.175: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.175: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.175: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.175: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.175: [--expert-interval EXPERT_INTERVAL] +100.83.37.175: [--hidden-size HIDDEN_SIZE] +100.83.37.175: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.175: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.175: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.175: [--kv-channels KV_CHANNELS] +100.83.37.175: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.175: [--use-rotary-position-embeddings] +100.83.37.175: [--rotary-percent ROTARY_PERCENT] +100.83.37.175: [--no-position-embedding] +100.83.37.175: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.175: [--normalization {layernorm,rmsnorm}] +100.83.37.175: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.175: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.175: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.175: [--apply-residual-connection-post-layernorm] +100.83.37.175: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.175: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.175: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.175: [--untie-embeddings-and-output-weights] +100.83.37.175: [--embedding-weights-in-fp32] +100.83.37.175: [--fix-position-emb-redundant-alloc] +100.83.37.175: [--embed-layernorm] +100.83.37.175: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.175: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.175: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.175: [--weight-decay WEIGHT_DECAY] +100.83.37.175: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.175: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.175: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.175: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.175: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.175: [--sgd-momentum SGD_MOMENTUM] +100.83.37.175: [--do-norm-bias-weight-decay] +100.83.37.175: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.175: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.175: [--batch-size BATCH_SIZE] +100.83.37.175: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.175: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.175: [--recompute-activations] +100.83.37.175: [--recompute-granularity {full,selective}] +100.83.37.175: [--distribute-saved-activations] +100.83.37.175: [--recompute-method {uniform,block}] +100.83.37.175: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.175: [--checkpoint-activations] +100.83.37.175: [--distribute-checkpointed-activations] +100.83.37.175: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.175: [--train-iters TRAIN_ITERS] +100.83.37.175: [--train-samples TRAIN_SAMPLES] +100.83.37.175: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.175: [--log-interval LOG_INTERVAL] +100.83.37.175: [--exit-interval EXIT_INTERVAL] +100.83.37.175: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.175: [--exit-signal-handler] +100.83.37.175: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.175: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.175: [--no-bias-dropout-fusion] +100.83.37.175: [--disable-moe-token-dropping] +100.83.37.175: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.175: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.175: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.175: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.175: [--create-moe-param-group] [--use-flash-attn] +100.83.37.175: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.175: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.175: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.175: [--disable-bias-linear] +100.83.37.175: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.175: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.175: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.175: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.175: [--no-async-tensor-model-parallel-allreduce] +100.83.37.175: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.175: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.175: [--force-ds-sequence-parallel] +100.83.37.175: [--no-gradient-accumulation-fusion] +100.83.37.175: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.175: [--data-parallel-random-init] +100.83.37.175: [--init-method-std INIT_METHOD_STD] +100.83.37.175: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.175: [--lr LR] +100.83.37.175: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.175: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.175: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.175: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.175: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.175: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.175: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.175: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.175: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.175: [--override-opt_param-scheduler] +100.83.37.175: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.175: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.175: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.175: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.175: [--no-initialization] [--use-checkpoint-args] +100.83.37.175: [--exit-on-missing-checkpoint] +100.83.37.175: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.175: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.175: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.175: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.175: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.175: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.175: [--hysteresis HYSTERESIS] +100.83.37.175: [--fp32-residual-connection] +100.83.37.175: [--no-query-key-layer-scaling] +100.83.37.175: [--attention-softmax-in-fp32] +100.83.37.175: [--accumulate-allreduce-grads-in-fp32] +100.83.37.175: [--fp16-lm-cross-entropy] +100.83.37.175: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.175: [--enable-expert-tensor-parallelism] +100.83.37.175: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.175: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.175: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.175: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.175: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.175: [--overlap-p2p-communication] +100.83.37.175: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.175: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.175: [--DDP-impl {local,torch,FSDP}] +100.83.37.175: [--no-contiguous-buffers-in-local-ddp] +100.83.37.175: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.175: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.175: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.175: [--use-cpu-initialization] +100.83.37.175: [--empty-unused-memory-level {0,1,2}] +100.83.37.175: [--standalone-embedding-stage] +100.83.37.175: [--use-distributed-optimizer] +100.83.37.175: [--eval-iters EVAL_ITERS] +100.83.37.175: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.175: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.175: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.175: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.175: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.175: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.175: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.175: [--data-cache-path DATA_CACHE_PATH] +100.83.37.175: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.175: [--merge-file MERGE_FILE] +100.83.37.175: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.175: [--seq-length SEQ_LENGTH] +100.83.37.175: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.175: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.175: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.175: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.175: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.175: [--num-workers NUM_WORKERS] +100.83.37.175: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.175: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.175: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.175: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.175: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.175: [--return-data-index] +100.83.37.175: [--data-efficiency-curriculum-learning] +100.83.37.175: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.175: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.175: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.175: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.175: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.175: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.175: [--adlr-autoresume] +100.83.37.175: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.175: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.175: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.175: [--biencoder-shared-query-context-model] +100.83.37.175: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.175: [--titles-data-path TITLES_DATA_PATH] +100.83.37.175: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.175: [--use-one-sent-docs] +100.83.37.175: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.175: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.175: [--retriever-score-scaling] +100.83.37.175: [--block-data-path BLOCK_DATA_PATH] +100.83.37.175: [--embedding-path EMBEDDING_PATH] +100.83.37.175: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.175: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.175: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.175: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.175: [--patch-dim PATCH_DIM] +100.83.37.175: [--classes-fraction CLASSES_FRACTION] +100.83.37.175: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.175: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.175: [--vision-pretraining] +100.83.37.175: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.175: [--vision-backbone-type {vit,mit,swin}] +100.83.37.175: [--swin-backbone-type {tiny,base,h3}] +100.83.37.175: [--mask-type {random,row}] +100.83.37.175: [--mask-factor MASK_FACTOR] +100.83.37.175: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.175: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.175: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.175: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.175: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.175: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.175: [--dino-norm-last-layer] +100.83.37.175: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.175: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.175: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.175: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.175: [--timing-log-level {0,1,2}] +100.83.37.175: [--no-barrier-with-level-1-timing] +100.83.37.175: [--timing-log-option {max,minmax,all}] +100.83.37.175: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.175: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.175: [--log-timers-to-tensorboard] +100.83.37.175: [--log-batch-size-to-tensorboard] +100.83.37.175: [--no-log-learnig-rate-to-tensorboard] +100.83.37.175: [--no-log-loss-scale-to-tensorboard] +100.83.37.175: [--log-validation-ppl-to-tensorboard] +100.83.37.175: [--log-optimizer-states-to-tensorboard] +100.83.37.175: [--log-memory-to-tensorboard] +100.83.37.175: [--log-world-size-to-tensorboard] +100.83.37.175: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.175: [--zero-contigious-gradients] +100.83.37.175: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.175: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.175: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.175: [--scattered-embeddings] [--split-transformers] +100.83.37.175: [--memory-centric-tiled-linear] +100.83.37.175: [--tile-factor TILE_FACTOR] +100.83.37.175: [--deepspeed-activation-checkpointing] +100.83.37.175: [--partition-activations] +100.83.37.175: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.175: [--synchronize-each-layer] [--profile-backward] +100.83.37.175: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.175: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.175: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.175: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.175: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.175: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.175: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.175: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.175: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.175: [--output-bert-embeddings] +100.83.37.175: [--bert-embedder-type {megatron,huggingface}] +100.83.37.175: [--cache-fp8-weight] +100.83.37.175: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.175: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.175: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.175: [--fp8-interval FP8_INTERVAL] +100.83.37.175: [--transformer-impl {local,transformer_engine}] +100.83.37.175: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.175: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.175: [--retro-workdir RETRO_WORKDIR] +100.83.37.175: [--retro-add-retriever] +100.83.37.175: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.175: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.175: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.175: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.175: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.175: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.175: [--retro-return-doc-ids] +100.83.37.175: [--profile {pt,pt-full,hltv}] +100.83.37.175: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.175: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.175: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.175: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.175: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed C++/CUDA extension op report +100.83.37.175: -------------------------------------------------- +100.83.37.175: NOTE: Ops not installed will be just-in-time (JIT) compiled at +100.83.37.175: runtime if needed. Op compatibility means that your system +100.83.37.175: meet the required dependencies to JIT install the op. +100.83.37.175: -------------------------------------------------- +100.83.37.175: JIT compiled ops requires ninja +100.83.37.175: ninja .................. [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: op name ................ installed .. compatible +100.83.37.175: -------------------------------------------------- +100.83.37.175: cpu_adam ............... [NO] ....... [OKAY] +100.83.37.175: fused_adam ............. [NO] ....... [OKAY] +100.83.37.175: deepspeed_not_implemented [NO] ....... [OKAY] +100.83.37.175: transformer_inference .. [NO] ....... [OKAY] +100.83.37.175: -------------------------------------------------- +100.83.37.175: DeepSpeed general environment info: +100.83.37.175: torch install path ............... ['/usr/local/lib/python3.10/dist-packages/torch'] +100.83.37.175: torch version .................... 2.1.1a0+gitb51c9f6 +100.83.37.175: deepspeed install path ........... ['/usr/local/lib/python3.10/dist-packages/deepspeed'] +100.83.37.175: deepspeed info ................... 0.12.4+hpu.synapse.v1.14.0, fad45b2, 1.14.0 +100.83.37.175: deepspeed wheel compiled w. ...... torch 2.1 +100.83.37.175: shared memory (/dev/shm) size .... 503.75 GB +100.83.37.175: fatal: detected dubious ownership in repository at '/Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed' +100.83.37.175: To add an exception for this directory, call: +100.83.37.175: +100.83.37.175: git config --global --add safe.directory /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed +100.83.37.175: **** Git info for Megatron: git_hash=unknown git_branch=unknown **** +100.83.37.175: usage: pretrain_llama.py [-h] [--num-layers NUM_LAYERS] +100.83.37.175: [--encoder-num-layers ENCODER_NUM_LAYERS] +100.83.37.175: [--decoder-num-layers DECODER_NUM_LAYERS] +100.83.37.175: [--num-experts NUM_EXPERTS [NUM_EXPERTS ...]] +100.83.37.175: [--mlp-type MLP_TYPE] [--topk TOPK] +100.83.37.175: [--expert-interval EXPERT_INTERVAL] +100.83.37.175: [--hidden-size HIDDEN_SIZE] +100.83.37.175: [--ffn-hidden-size FFN_HIDDEN_SIZE] +100.83.37.175: [--num-attention-heads NUM_ATTENTION_HEADS] +100.83.37.175: [--num-key-value-heads NUM_KEY_VALUE_HEADS] +100.83.37.175: [--kv-channels KV_CHANNELS] +100.83.37.175: [--max-position-embeddings MAX_POSITION_EMBEDDINGS] +100.83.37.175: [--use-rotary-position-embeddings] +100.83.37.175: [--rotary-percent ROTARY_PERCENT] +100.83.37.175: [--no-position-embedding] +100.83.37.175: [--make-vocab-size-divisible-by MAKE_VOCAB_SIZE_DIVISIBLE_BY] +100.83.37.175: [--normalization {layernorm,rmsnorm}] +100.83.37.175: [--use-fused-rmsnorm USE_FUSED_RMSNORM] +100.83.37.175: [--layernorm-epsilon LAYERNORM_EPSILON] +100.83.37.175: [--apply-layernorm-1p] [--disable-mem-efficient-ln] +100.83.37.175: [--apply-residual-connection-post-layernorm] +100.83.37.175: [--openai-gelu] [--squared-relu] [--swiglu] +100.83.37.175: [--onnx-safe ONNX_SAFE] [--bert-no-binary-head] +100.83.37.175: [--num-experts-switch NUM_EXPERTS_SWITCH] +100.83.37.175: [--untie-embeddings-and-output-weights] +100.83.37.175: [--embedding-weights-in-fp32] +100.83.37.175: [--fix-position-emb-redundant-alloc] +100.83.37.175: [--embed-layernorm] +100.83.37.175: [--kill-switch-path KILL_SWITCH_PATH] +100.83.37.175: [--attention-dropout ATTENTION_DROPOUT] +100.83.37.175: [--hidden-dropout HIDDEN_DROPOUT] +100.83.37.175: [--weight-decay WEIGHT_DECAY] +100.83.37.175: [--start-weight-decay START_WEIGHT_DECAY] +100.83.37.175: [--end-weight-decay END_WEIGHT_DECAY] +100.83.37.175: [--weight-decay-incr-style {constant,linear,cosine}] +100.83.37.175: [--clip-grad CLIP_GRAD] [--adam-beta1 ADAM_BETA1] +100.83.37.175: [--adam-beta2 ADAM_BETA2] [--adam-eps ADAM_EPS] +100.83.37.175: [--sgd-momentum SGD_MOMENTUM] +100.83.37.175: [--do-norm-bias-weight-decay] +100.83.37.175: [--micro-batch-size MICRO_BATCH_SIZE] +100.83.37.175: [--eval-micro-batch-size EVAL_MICRO_BATCH_SIZE] +100.83.37.175: [--batch-size BATCH_SIZE] +100.83.37.175: [--global-batch-size GLOBAL_BATCH_SIZE] +100.83.37.175: [--rampup-batch-size [RAMPUP_BATCH_SIZE ...]] +100.83.37.175: [--recompute-activations] +100.83.37.175: [--recompute-granularity {full,selective}] +100.83.37.175: [--distribute-saved-activations] +100.83.37.175: [--recompute-method {uniform,block}] +100.83.37.175: [--recompute-num-layers RECOMPUTE_NUM_LAYERS] +100.83.37.175: [--checkpoint-activations] +100.83.37.175: [--distribute-checkpointed-activations] +100.83.37.175: [--checkpoint-num-layers CHECKPOINT_NUM_LAYERS] +100.83.37.175: [--train-iters TRAIN_ITERS] +100.83.37.175: [--train-samples TRAIN_SAMPLES] +100.83.37.175: [--train-tokens TRAIN_TOKENS] [--random-ltd] +100.83.37.175: [--log-interval LOG_INTERVAL] +100.83.37.175: [--exit-interval EXIT_INTERVAL] +100.83.37.175: [--exit-duration-in-mins EXIT_DURATION_IN_MINS] +100.83.37.175: [--exit-signal-handler] +100.83.37.175: [--tensorboard-dir TENSORBOARD_DIR] +100.83.37.175: [--no-masked-softmax-fusion] [--no-bias-gelu-fusion] +100.83.37.175: [--no-bias-dropout-fusion] +100.83.37.175: [--disable-moe-token-dropping] +100.83.37.175: [--moe-train-capacity-factor MOE_TRAIN_CAPACITY_FACTOR] +100.83.37.175: [--moe-eval-capacity-factor MOE_EVAL_CAPACITY_FACTOR] +100.83.37.175: [--moe-min-capacity MOE_MIN_CAPACITY] +100.83.37.175: [--moe-loss-coeff MOE_LOSS_COEFF] +100.83.37.175: [--create-moe-param-group] [--use-flash-attn] +100.83.37.175: [--use-flash-attn-v2] [--use-flash-attn-triton] +100.83.37.175: [--use-fused-sdpa USE_FUSED_SDPA] +100.83.37.175: [--use-fused-sdpa-with-recompute USE_FUSED_SDPA_WITH_RECOMPUTE] +100.83.37.175: [--disable-bias-linear] +100.83.37.175: [--optimizer {adam,sgd,adamw,fusedadamw}] +100.83.37.175: [--dataloader-type {single,cyclic}] [--ds-inference] +100.83.37.175: [--cpu-optimizer] [--cpu_torch_adam] +100.83.37.175: [--no-pipeline-parallel] [--use-tutel] [--inference] +100.83.37.175: [--no-async-tensor-model-parallel-allreduce] +100.83.37.175: [--no-persist-layer-norm] [--sequence-parallel] +100.83.37.175: [--ds-sequence-parallel-size DS_SEQUENCE_PARALLEL_SIZE] +100.83.37.175: [--force-ds-sequence-parallel] +100.83.37.175: [--no-gradient-accumulation-fusion] +100.83.37.175: [--use-dataset-only USE_DATASET_ONLY] [--seed SEED] +100.83.37.175: [--data-parallel-random-init] +100.83.37.175: [--init-method-std INIT_METHOD_STD] +100.83.37.175: [--init-method-xavier-uniform] [--no-scaled-init] +100.83.37.175: [--lr LR] +100.83.37.175: [--lr-decay-style {constant,linear,cosine,inverse-square-root}] +100.83.37.175: [--lr-decay-iters LR_DECAY_ITERS] +100.83.37.175: [--lr-decay-samples LR_DECAY_SAMPLES] +100.83.37.175: [--lr-decay-tokens LR_DECAY_TOKENS] +100.83.37.175: [--lr-warmup-fraction LR_WARMUP_FRACTION] +100.83.37.175: [--lr-warmup-iters LR_WARMUP_ITERS] +100.83.37.175: [--lr-warmup-samples LR_WARMUP_SAMPLES] +100.83.37.175: [--lr-warmup-tokens LR_WARMUP_TOKENS] +100.83.37.175: [--warmup WARMUP] [--min-lr MIN_LR] +100.83.37.175: [--override-opt_param-scheduler] +100.83.37.175: [--use-checkpoint-opt_param-scheduler] [--save SAVE] +100.83.37.175: [--save-interval SAVE_INTERVAL] [--no-save-optim] +100.83.37.175: [--no-save-rng] [--load LOAD] [--no-load-optim] +100.83.37.175: [--no-load-rng] [--no-load-lr-state] [--finetune] +100.83.37.175: [--no-initialization] [--use-checkpoint-args] +100.83.37.175: [--exit-on-missing-checkpoint] +100.83.37.175: [--universal-checkpoint] [--verify-checkpoint] +100.83.37.175: [--verify-checkpoint-model-type {GPT,BLOOM,LLAMA}] +100.83.37.175: [--fp16] [--bf16] [--loss-scale LOSS_SCALE] +100.83.37.175: [--initial-loss-scale INITIAL_LOSS_SCALE] +100.83.37.175: [--min-loss-scale MIN_LOSS_SCALE] +100.83.37.175: [--loss-scale-window LOSS_SCALE_WINDOW] +100.83.37.175: [--hysteresis HYSTERESIS] +100.83.37.175: [--fp32-residual-connection] +100.83.37.175: [--no-query-key-layer-scaling] +100.83.37.175: [--attention-softmax-in-fp32] +100.83.37.175: [--accumulate-allreduce-grads-in-fp32] +100.83.37.175: [--fp16-lm-cross-entropy] +100.83.37.175: [--tensor-model-parallel-size TENSOR_MODEL_PARALLEL_SIZE] +100.83.37.175: [--enable-expert-tensor-parallelism] +100.83.37.175: [--pipeline-model-parallel-size PIPELINE_MODEL_PARALLEL_SIZE] +100.83.37.175: [--pipeline-model-parallel-split-rank PIPELINE_MODEL_PARALLEL_SPLIT_RANK] +100.83.37.175: [--moe-expert-parallel-size MOE_EXPERT_PARALLEL_SIZE] +100.83.37.175: [--model-parallel-size MODEL_PARALLEL_SIZE] +100.83.37.175: [--num-layers-per-virtual-pipeline-stage NUM_LAYERS_PER_VIRTUAL_PIPELINE_STAGE] +100.83.37.175: [--overlap-p2p-communication] +100.83.37.175: [--distributed-backend {nccl,gloo,ccl,hccl}] +100.83.37.175: [--distributed-timeout-minutes DISTRIBUTED_TIMEOUT_MINUTES] +100.83.37.175: [--DDP-impl {local,torch,FSDP}] +100.83.37.175: [--no-contiguous-buffers-in-local-ddp] +100.83.37.175: [--no-scatter-gather-tensors-in-pipeline] +100.83.37.175: [--use-ring-exchange-p2p] [--local-rank LOCAL_RANK] +100.83.37.175: [--lazy-mpu-init LAZY_MPU_INIT] +100.83.37.175: [--use-cpu-initialization] +100.83.37.175: [--empty-unused-memory-level {0,1,2}] +100.83.37.175: [--standalone-embedding-stage] +100.83.37.175: [--use-distributed-optimizer] +100.83.37.175: [--eval-iters EVAL_ITERS] +100.83.37.175: [--eval-interval EVAL_INTERVAL] [--skip-train] +100.83.37.175: [--eval-loss-exit-value EVAL_LOSS_EXIT_VALUE] +100.83.37.175: [--aml-data-download-path AML_DATA_DOWNLOAD_PATH] +100.83.37.175: [--data-path [DATA_PATH ...]] [--split SPLIT] +100.83.37.175: [--train-data-path [TRAIN_DATA_PATH ...]] +100.83.37.175: [--valid-data-path [VALID_DATA_PATH ...]] +100.83.37.175: [--test-data-path [TEST_DATA_PATH ...]] +100.83.37.175: [--data-cache-path DATA_CACHE_PATH] +100.83.37.175: [--vocab-size VOCAB_SIZE] [--vocab-file VOCAB_FILE] +100.83.37.175: [--merge-file MERGE_FILE] +100.83.37.175: [--vocab-extra-ids VOCAB_EXTRA_IDS] +100.83.37.175: [--seq-length SEQ_LENGTH] +100.83.37.175: [--encoder-seq-length ENCODER_SEQ_LENGTH] +100.83.37.175: [--decoder-seq-length DECODER_SEQ_LENGTH] +100.83.37.175: [--retriever-seq-length RETRIEVER_SEQ_LENGTH] +100.83.37.175: [--sample-rate SAMPLE_RATE] [--mask-prob MASK_PROB] +100.83.37.175: [--short-seq-prob SHORT_SEQ_PROB] [--mmap-warmup] +100.83.37.175: [--num-workers NUM_WORKERS] +100.83.37.175: [--tokenizer-type {BertWordPieceLowerCase,BertWordPieceCase,GPT2BPETokenizer,SentencePieceTokenizer,GPTSentencePieceTokenizer,NullTokenizer}] +100.83.37.175: [--tokenizer-model TOKENIZER_MODEL] +100.83.37.175: [--data-impl {mmap,infer}] [--reset-position-ids] +100.83.37.175: [--reset-attention-mask] [--eod-mask-loss] +100.83.37.175: [--train-data-exact-num-epochs TRAIN_DATA_EXACT_NUM_EPOCHS] +100.83.37.175: [--return-data-index] +100.83.37.175: [--data-efficiency-curriculum-learning] +100.83.37.175: [--train-idx-path TRAIN_IDX_PATH] +100.83.37.175: [--train-desc-path TRAIN_DESC_PATH] +100.83.37.175: [--train-doc-idx-path TRAIN_DOC_IDX_PATH] +100.83.37.175: [--train-sample-idx-path TRAIN_SAMPLE_IDX_PATH] +100.83.37.175: [--train-shuffle-idx-path TRAIN_SHUFFLE_IDX_PATH] +100.83.37.175: [--mask-tensor-adding] [--no-seq-len-plus-one-tokens] +100.83.37.175: [--adlr-autoresume] +100.83.37.175: [--adlr-autoresume-interval ADLR_AUTORESUME_INTERVAL] +100.83.37.175: [--ict-head-size ICT_HEAD_SIZE] +100.83.37.175: [--biencoder-projection-dim BIENCODER_PROJECTION_DIM] +100.83.37.175: [--biencoder-shared-query-context-model] +100.83.37.175: [--ict-load ICT_LOAD] [--bert-load BERT_LOAD] +100.83.37.175: [--titles-data-path TITLES_DATA_PATH] +100.83.37.175: [--query-in-block-prob QUERY_IN_BLOCK_PROB] +100.83.37.175: [--use-one-sent-docs] +100.83.37.175: [--evidence-data-path EVIDENCE_DATA_PATH] +100.83.37.175: [--retriever-report-topk-accuracies RETRIEVER_REPORT_TOPK_ACCURACIES [RETRIEVER_REPORT_TOPK_ACCURACIES ...]] +100.83.37.175: [--retriever-score-scaling] +100.83.37.175: [--block-data-path BLOCK_DATA_PATH] +100.83.37.175: [--embedding-path EMBEDDING_PATH] +100.83.37.175: [--indexer-batch-size INDEXER_BATCH_SIZE] +100.83.37.175: [--indexer-log-interval INDEXER_LOG_INTERVAL] +100.83.37.175: [--num-classes NUM_CLASSES] [--img-h IMG_H] +100.83.37.175: [--img-w IMG_W] [--num-channels NUM_CHANNELS] +100.83.37.175: [--patch-dim PATCH_DIM] +100.83.37.175: [--classes-fraction CLASSES_FRACTION] +100.83.37.175: [--data-per-class-fraction DATA_PER_CLASS_FRACTION] +100.83.37.175: [--no-data-sharding] [--head-lr-mult HEAD_LR_MULT] +100.83.37.175: [--vision-pretraining] +100.83.37.175: [--vision-pretraining-type {classify,inpaint,dino}] +100.83.37.175: [--vision-backbone-type {vit,mit,swin}] +100.83.37.175: [--swin-backbone-type {tiny,base,h3}] +100.83.37.175: [--mask-type {random,row}] +100.83.37.175: [--mask-factor MASK_FACTOR] +100.83.37.175: [--iter-per-epoch ITER_PER_EPOCH] +100.83.37.175: [--dino-local-img-size DINO_LOCAL_IMG_SIZE] +100.83.37.175: [--dino-local-crops-number DINO_LOCAL_CROPS_NUMBER] +100.83.37.175: [--dino-head-hidden-size DINO_HEAD_HIDDEN_SIZE] +100.83.37.175: [--dino-bottleneck-size DINO_BOTTLENECK_SIZE] +100.83.37.175: [--dino-freeze-last-layer DINO_FREEZE_LAST_LAYER] +100.83.37.175: [--dino-norm-last-layer] +100.83.37.175: [--dino-warmup-teacher-temp DINO_WARMUP_TEACHER_TEMP] +100.83.37.175: [--dino-teacher-temp DINO_TEACHER_TEMP] +100.83.37.175: [--dino-warmup-teacher-temp-epochs DINO_WARMUP_TEACHER_TEMP_EPOCHS] +100.83.37.175: [--log-params-norm] [--log-num-zeros-in-grad] +100.83.37.175: [--timing-log-level {0,1,2}] +100.83.37.175: [--no-barrier-with-level-1-timing] +100.83.37.175: [--timing-log-option {max,minmax,all}] +100.83.37.175: [--tensorboard-log-interval TENSORBOARD_LOG_INTERVAL] +100.83.37.175: [--tensorboard-queue-size TENSORBOARD_QUEUE_SIZE] +100.83.37.175: [--log-timers-to-tensorboard] +100.83.37.175: [--log-batch-size-to-tensorboard] +100.83.37.175: [--no-log-learnig-rate-to-tensorboard] +100.83.37.175: [--no-log-loss-scale-to-tensorboard] +100.83.37.175: [--log-validation-ppl-to-tensorboard] +100.83.37.175: [--log-optimizer-states-to-tensorboard] +100.83.37.175: [--log-memory-to-tensorboard] +100.83.37.175: [--log-world-size-to-tensorboard] +100.83.37.175: [--zero-stage ZERO_STAGE] [--zero-reduce-scatter] +100.83.37.175: [--zero-contigious-gradients] +100.83.37.175: [--zero-reduce-bucket-size ZERO_REDUCE_BUCKET_SIZE] +100.83.37.175: [--zero-allgather-bucket-size ZERO_ALLGATHER_BUCKET_SIZE] +100.83.37.175: [--remote-device {none,cpu,nvme}] [--use-pin-memory] +100.83.37.175: [--scattered-embeddings] [--split-transformers] +100.83.37.175: [--memory-centric-tiled-linear] +100.83.37.175: [--tile-factor TILE_FACTOR] +100.83.37.175: [--deepspeed-activation-checkpointing] +100.83.37.175: [--partition-activations] +100.83.37.175: [--contigious-checkpointing] [--checkpoint-in-cpu] +100.83.37.175: [--synchronize-each-layer] [--profile-backward] +100.83.37.175: [--num-layers-teacher NUM_LAYERS_TEACHER] +100.83.37.175: [--num-experts-teacher NUM_EXPERTS_TEACHER [NUM_EXPERTS_TEACHER ...]] +100.83.37.175: [--hidden-size-teacher HIDDEN_SIZE_TEACHER] +100.83.37.175: [--num-attention-heads-teacher NUM_ATTENTION_HEADS_TEACHER] +100.83.37.175: [--mos] [--kd] [--kd-alpha-ce KD_ALPHA_CE] +100.83.37.175: [--kd-beta-ce KD_BETA_CE] [--kd-temp KD_TEMP] +100.83.37.175: [--reset-iteration] [--load-teacher LOAD_TEACHER] +100.83.37.175: [--inference-batch-times-seqlen-threshold INFERENCE_BATCH_TIMES_SEQLEN_THRESHOLD] +100.83.37.175: [--max-tokens-to-oom MAX_TOKENS_TO_OOM] +100.83.37.175: [--output-bert-embeddings] +100.83.37.175: [--bert-embedder-type {megatron,huggingface}] +100.83.37.175: [--cache-fp8-weight] +100.83.37.175: [--cache-fp8-weight-fwd CACHE_FP8_WEIGHT_FWD] +100.83.37.175: [--fp8-e5m2] [--fp8-e4m3] [--fp8-hybrid] +100.83.37.175: [--no-fp8-wgrad] [--fp8-margin FP8_MARGIN] +100.83.37.175: [--fp8-interval FP8_INTERVAL] +100.83.37.175: [--transformer-impl {local,transformer_engine}] +100.83.37.175: [--fp8-amax-history-len FP8_AMAX_HISTORY_LEN] +100.83.37.175: [--fp8-amax-compute-algo {most_recent,max}] +100.83.37.175: [--retro-workdir RETRO_WORKDIR] +100.83.37.175: [--retro-add-retriever] +100.83.37.175: [--retro-cyclic-train-iters RETRO_CYCLIC_TRAIN_ITERS] +100.83.37.175: [--retro-encoder-layers RETRO_ENCODER_LAYERS] +100.83.37.175: [--retro-encoder-hidden-dropout RETRO_ENCODER_HIDDEN_DROPOUT] +100.83.37.175: [--retro-encoder-attention-dropout RETRO_ENCODER_ATTENTION_DROPOUT] +100.83.37.175: [--retro-num-neighbors RETRO_NUM_NEIGHBORS] +100.83.37.175: [--retro-num-retrieved-chunks RETRO_NUM_RETRIEVED_CHUNKS] +100.83.37.175: [--retro-return-doc-ids] +100.83.37.175: [--profile {pt,pt-full,hltv}] +100.83.37.175: [--profile-steps PROFILE_STEPS] [--deepspeed] +100.83.37.175: [--deepspeed_config DEEPSPEED_CONFIG] [--deepscale] +100.83.37.175: [--deepscale_config DEEPSCALE_CONFIG] +100.83.37.175: [--deepspeed_mpi] [--use_hpu] [--no_cuda] +100.83.37.175: pretrain_llama.py: error: unrecognized arguments: --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --use-torch-compile false --hpu-deterministic --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt +100.83.37.188: [2024-05-13 11:50:12,220] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 127781 +100.83.37.188: [2024-05-13 11:50:12,222] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 127782 +100.83.37.188: [2024-05-13 11:50:12,249] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 127783 +100.83.37.188: [2024-05-13 11:50:12,249] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 127784 +100.83.37.188: [2024-05-13 11:50:12,249] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 127785 +100.83.37.188: [2024-05-13 11:50:12,250] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 127786 +100.83.37.188: [2024-05-13 11:50:12,250] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 127787 +100.83.37.188: [2024-05-13 11:50:12,250] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 127788 +100.83.37.188: [2024-05-13 11:50:12,250] [ERROR] [launch.py:322:sigkill_handler] ['/usr/bin/bash', '-c', ' cd /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed && python3 -u ./pretrain_llama.py --deepspeed --tensor-model-parallel-size 1 --pipeline-model-parallel-size 1 --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --layernorm-epsilon 1e-6 --num-layers 24 --hidden-size 2048 --ffn-hidden-size 4096 --num-attention-heads 32 --seq-length 2048 --micro-batch-size 1 --global-batch-size 256 --train-iters 10000 --log-interval 10 --eval-iters 10 --eval-interval 500 --data-path /data/hineng/tokenizer//_raw_content_document --vocab-file /data/hineng/tokenizer//gpt2-vocab.json --merge-file /data/hineng/tokenizer//gpt2-merges.txt --optimizer adamw --adam-beta1 0.9 --adam-beta2 0.95 --adam-eps 1e-6 --lr 3e-4 --lr-decay-style cosine --lr-warmup-iters 2000 --clip-grad 1.0 --weight-decay 0.1 --attention-dropout 0.1 --hidden-dropout 0.1 --tensorboard-dir /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/tensorboard --log-validation-ppl-to-tensorboard --log-batch-size-to-tensorboard --log-timers-to-tensorboard --load /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/checkpoints_zero_stage_2 --deepspeed_config=/data/output/llama13b_multiling_800M/13-05-2024-11:50:01/ds_config.json --zero-stage=0 --exit-interval 0 --no-masked-softmax-fusion --no-bias-gelu-fusion --no-bias-dropout-fusion --use-torch-compile false --use-fused-sdpa true --use-fused-sdpa-with-recompute false --bf16 --use_hpu --distributed-backend=hccl --hpu-deterministic --sequence-parallel --save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/checkpoints_zero_stage_2 --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt --save-interval 500 --verify-checkpoint --verify-checkpoint-model-type LLAMA'] exits with return code = 2 +100.83.37.175: [2024-05-13 11:50:12,451] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 33636 +100.83.37.175: [2024-05-13 11:50:12,453] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 33637 +100.83.37.175: [2024-05-13 11:50:12,453] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 33638 +100.83.37.175: [2024-05-13 11:50:12,453] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 33639 +100.83.37.175: [2024-05-13 11:50:12,454] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 33641 +100.83.37.175: [2024-05-13 11:50:12,481] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 33644 +100.83.37.175: [2024-05-13 11:50:12,507] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 33647 +100.83.37.175: [2024-05-13 11:50:12,508] [INFO] [launch.py:316:sigkill_handler] Killing subprocess 33649 +100.83.37.175: [2024-05-13 11:50:12,509] [ERROR] [launch.py:322:sigkill_handler] ['/usr/bin/bash', '-c', ' cd /Model-References/PyTorch/nlp/DeepSpeedExamples/Megatron-DeepSpeed && python3 -u ./pretrain_llama.py --deepspeed --tensor-model-parallel-size 1 --pipeline-model-parallel-size 1 --position-embedding-type rotary --no-bias --layernorm-type rmsnorm --activation-func-type swiglu --layernorm-epsilon 1e-6 --num-layers 24 --hidden-size 2048 --ffn-hidden-size 4096 --num-attention-heads 32 --seq-length 2048 --micro-batch-size 1 --global-batch-size 256 --train-iters 10000 --log-interval 10 --eval-iters 10 --eval-interval 500 --data-path /data/hineng/tokenizer//_raw_content_document --vocab-file /data/hineng/tokenizer//gpt2-vocab.json --merge-file /data/hineng/tokenizer//gpt2-merges.txt --optimizer adamw --adam-beta1 0.9 --adam-beta2 0.95 --adam-eps 1e-6 --lr 3e-4 --lr-decay-style cosine --lr-warmup-iters 2000 --clip-grad 1.0 --weight-decay 0.1 --attention-dropout 0.1 --hidden-dropout 0.1 --tensorboard-dir /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/tensorboard --log-validation-ppl-to-tensorboard --log-batch-size-to-tensorboard --log-timers-to-tensorboard --load /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/checkpoints_zero_stage_2 --deepspeed_config=/data/output/llama13b_multiling_800M/13-05-2024-11:50:01/ds_config.json --zero-stage=0 --exit-interval 0 --no-masked-softmax-fusion --no-bias-gelu-fusion --no-bias-dropout-fusion --use-torch-compile false --use-fused-sdpa true --use-fused-sdpa-with-recompute false --bf16 --use_hpu --distributed-backend=hccl --hpu-deterministic --sequence-parallel --save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/checkpoints_zero_stage_2 --hf-save /data/output/llama13b_multiling_800M/13-05-2024-11:50:01/hf_ckpt --save-interval 500 --verify-checkpoint --verify-checkpoint-model-type LLAMA'] exits with return code = 2 +pdsh@vizzhy-150-3: 100.83.37.188: ssh exited with exit code 2