#!/bin/bash | |
# ✅ | |
# Saturate DP for 1B on 1 node | |
# for grad_acc in 2 4 8 16 32 64 128; do | |
# for mbs in 64 32 16 8 4 2 1; do | |
# if [ $((grad_acc * mbs)) -eq 128 ]; then | |
# echo "Running with grad_acc=$grad_acc and mbs=$mbs" | |
# exp_name="dp8_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# python create_config.py --out_dir result_blog_posts --model_name TinyLlama/TinyLlama_v1.1 --exp_name saturate_dp_1b_1_node/$exp_name --mbs=$mbs --grad_acc=$grad_acc --dp=8 --tp=1 --pp=1 --seq_len=4096 --use_wandb | |
# fi | |
# done | |
# done | |
# ✅ | |
# Saturate DP for 1B on 8 node | |
# for grad_acc in 1 2 4 8 16; do | |
# for mbs in 1 2 4 8 16; do | |
# if [ $((grad_acc * mbs)) -eq 16 ]; then | |
# exp_name="dp64_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc and mbs=$mbs" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name TinyLlama/TinyLlama_v1.1 \ | |
# --exp_name saturate_dp_1b_8_node/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=64 \ | |
# --tp=1 \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# done | |
# ✅ | |
# # Saturate DP for 1B on 32 node | |
# for grad_acc in 1 2 4; do | |
# for mbs in 1 2 4; do | |
# if [ $((grad_acc * mbs)) -eq 4 ]; then | |
# exp_name="dp256_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc and mbs=$mbs" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name TinyLlama/TinyLlama_v1.1 \ | |
# --exp_name saturate_dp_1b_32_node/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=256 \ | |
# --tp=1 \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# done | |
# ✅ | |
# # Sature DP for 1b on 1 node (with TP) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256 512; do | |
# for mbs in 1 2 4 8 16 32 64 128 256 512; do | |
# for dp in 1 2 4; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for tp in 1 2 4 8; do | |
# if [ $((dp * tp)) -eq 8 ]; then | |
# exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name TinyLlama/TinyLlama_v1.1 \ | |
# --exp_name saturate_dp_1b_1_node_tp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=$tp \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# ✅ | |
# # Sature DP for 1b on 8 node (with TP) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256 512; do | |
# for mbs in 1 2 4 8 16 32 64 128 256 512; do | |
# for dp in 1 2 4 8 16 32; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for tp in 1 2 4 8; do | |
# if [ $((dp * tp)) -eq 64 ]; then | |
# exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name TinyLlama/TinyLlama_v1.1 \ | |
# --exp_name saturate_dp_1b_8_node_tp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=$tp \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# ✅ | |
# # Saturated DP for 1B on 32 nodes (with TP) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256 512; do | |
# for mbs in 1 2 4 8 16 32 64 128 256 512; do | |
# for dp in 1 2 4 8 16 32 64 128; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for tp in 1 2 4 8; do | |
# if [ $((dp * tp)) -eq 256 ]; then | |
# exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name TinyLlama/TinyLlama_v1.1 \ | |
# --exp_name saturate_dp_1b_32_node_tp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=$tp \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# # ✅ | |
# # Sature DP for 1b on 1 node (with PP 1f1b) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256 512; do | |
# for mbs in 1 2 4 8 16 32 64 128 256 512; do | |
# for dp in 1 2 4; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for pp in 1 2 4 8; do | |
# if [ $((dp * pp)) -eq 8 ]; then | |
# exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=1, pp=$pp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name TinyLlama/TinyLlama_v1.1 \ | |
# --exp_name saturate_dp_1b_1_node_pp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=1 \ | |
# --pp=$pp \ | |
# --pp_engine 1f1b \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# # ✅ | |
# # Sature DP for 1b on 8 node (with PP 1f1b) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256 512; do | |
# for mbs in 1 2 4 8 16 32 64 128 256 512; do | |
# for dp in 1 2 4 8 16 32; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for pp in 1 2 4 8 16 32 64; do | |
# if [ $((dp * pp)) -eq 64 ]; then | |
# exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=1, pp=$pp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name TinyLlama/TinyLlama_v1.1 \ | |
# --exp_name saturate_dp_1b_8_node_pp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=1 \ | |
# --pp=$pp \ | |
# --pp_engine 1f1b \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# # ✅ | |
# # Sature DP for 1b on 32 node (with PP 1f1b) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256 512; do | |
# for mbs in 1 2 4 8 16 32 64 128 256 512; do | |
# for dp in 1 2 4 8 16 32 64 128; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for pp in 1 2 4 8 16 32 64 128 256; do | |
# if [ $((dp * pp)) -eq 256 ]; then | |
# exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=1, pp=$pp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name TinyLlama/TinyLlama_v1.1 \ | |
# --exp_name saturate_dp_1b_32_node_pp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=1 \ | |
# --pp=$pp \ | |
# --pp_engine 1f1b \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# ✅ | |
# # Try to fit 7B on 1 node | |
# for grad_acc in 1 2 4 8 16 32 64 128; do | |
# for mbs in 1 2 4 8 16 32 64 128; do | |
# if [ $((grad_acc * mbs)) -eq 128 ]; then | |
# exp_name="dp8_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc and mbs=$mbs" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name meta-llama/Llama-2-7b-hf \ | |
# --exp_name try_to_fit_7b_1_node/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=8 \ | |
# --tp=1 \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# done | |
# ✅ | |
# # Saturate DP for 7B on 8 nodes (with TP) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256; do | |
# for mbs in 1 2 4 8 16; do | |
# for dp in 1 2 4 8 16 32 64; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for tp in 1 2 4 8; do | |
# if [ $((dp * tp)) -eq 64 ]; then | |
# exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name meta-llama/Llama-2-7b-hf \ | |
# --exp_name saturate_dp_7b_8_node_tp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=$tp \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# ✅ | |
# # Saturate DP for 7B on 8 nodes and TP=16 | |
# for grad_acc in 1 2 4 8 16 32 64 128 256; do | |
# for mbs in 1 2 4 8 16; do | |
# for dp in 1 2 4; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# tp=16 | |
# if [ $((dp * tp)) -eq 64 ]; then | |
# exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name meta-llama/Llama-2-7b-hf \ | |
# --exp_name saturate_dp_7b_8_node_tp16/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=$tp \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# fi | |
# done | |
# done | |
# done | |
# ✅ | |
# # Saturate DP for 7B on 32 nodes | |
# for grad_acc in 1 2 4; do | |
# for mbs in 1 2 4; do | |
# if [ $((grad_acc * mbs)) -eq 4 ]; then | |
# exp_name="dp256_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc and mbs=$mbs" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name meta-llama/Llama-2-7b-hf \ | |
# --exp_name saturate_dp_7b_32_node/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=256 \ | |
# --tp=1 \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# done | |
# ✅ | |
# # Saturate DP for 7B on 32 nodes (with TP) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256; do | |
# for mbs in 1 2 4 8 16; do | |
# for dp in 1 2 4 8 16 32 64 128 256; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for tp in 1 2 4 8; do | |
# if [ $((dp * tp)) -eq 256 ]; then | |
# exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name meta-llama/Llama-2-7b-hf \ | |
# --exp_name saturate_dp_7b_32_node_tp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=$tp \ | |
# --pp=1 \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# # ✅ | |
# # Saturate DP for 7B on 8 nodes (with PP) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256; do | |
# for mbs in 1 2 4 8 16; do | |
# for dp in 1 2 4 8 16 32 64; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for pp in 1 2 4 8 16 32 64; do | |
# if [ $((dp * pp)) -eq 64 ]; then | |
# exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, pp=$pp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name meta-llama/Llama-2-7b-hf \ | |
# --exp_name saturate_dp_7b_8_node_pp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=1 \ | |
# --pp=$pp \ | |
# --pp_engine 1f1b \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# # ✅ | |
# # Saturate DP for 7B on 8 nodes and PP=16 | |
# for grad_acc in 1 2 4 8 16 32 64 128 256; do | |
# for mbs in 1 2 4 8 16; do | |
# for dp in 1 2 4; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# pp=16 | |
# if [ $((dp * pp)) -eq 64 ]; then | |
# exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, pp=$pp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name meta-llama/Llama-2-7b-hf \ | |
# --exp_name saturate_dp_7b_8_node_pp16/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=1 \ | |
# --pp=$pp \ | |
# --pp_engine 1f1b \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# fi | |
# done | |
# done | |
# done | |
# # ✅ | |
# # Saturate DP for 7B on 32 nodes (with PP) | |
# for grad_acc in 1 2 4 8 16 32 64 128 256; do | |
# for mbs in 1 2 4 8 16; do | |
# for dp in 1 2 4 8 16 32 64 128; do | |
# if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then | |
# for pp in 1 2 4 8 16 32 64 128 256; do | |
# if [ $((dp * pp)) -eq 256 ]; then | |
# exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" | |
# echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, pp=$pp" | |
# python create_config.py \ | |
# --out_dir result_blog_posts \ | |
# --model_name meta-llama/Llama-2-7b-hf \ | |
# --exp_name saturate_dp_7b_32_node_pp/$exp_name \ | |
# --mbs=$mbs \ | |
# --grad_acc=$grad_acc \ | |
# --dp=$dp \ | |
# --tp=1 \ | |
# --pp=$pp \ | |
# --pp_engine 1f1b \ | |
# --seq_len=4096 \ | |
# --use_wandb | |
# fi | |
# done | |
# fi | |
# done | |
# done | |
# done | |
# ⌛ | |
# 70B on 8 nodes (with DP + TP + PP) | |
for grad_acc in 1 2 4 8 16 32 64 128 256 512 1024; do | |
for dp in 1 2 4 8 16 32 64; do | |
if [ $((dp * grad_acc)) -eq 1024 ]; then | |
for tp in 2 4 8; do | |
for pp in 2 4 8 16 32 64; do | |
if [ $((dp * tp * pp)) -eq 64 ]; then | |
exp_name="dp${dp}_tp${tp}_pp${pp}_mbs1_ga${grad_acc}_sl4096" | |
echo "Running with grad_acc=$grad_acc, mbs=1, dp=$dp, tp=$tp, pp=$pp" | |
python create_config.py \ | |
--out_dir result_blog_posts \ | |
--model_name meta-llama/Llama-2-70b-hf \ | |
--exp_name 70b_8_node_with_dp_tp_pp/$exp_name \ | |
--mbs=1 \ | |
--grad_acc=$grad_acc \ | |
--dp=$dp \ | |
--tp=$tp \ | |
--pp=$pp \ | |
--pp_engine 1f1b \ | |
--seq_len=4096 \ | |
--use_wandb \ | |
--hf_hub_safetensors_path /fsx/ferdinandmom/hf_model_ckpt/Llama-2-70b-chat-hf | |
fi | |
done | |
done | |
fi | |
done | |
done | |
# ⌛ | |
# 70B on 32 nodes (with DP + TP + PP) | |
for grad_acc in 1 2 4 8 16 32 64 128 256 512 1024; do | |
for dp in 1 2 4 8 16 32 64 128 256; do | |
if [ $((dp * grad_acc)) -eq 1024 ]; then | |
for tp in 2 4 8; do | |
for pp in 2 4 8 16 32 64 128 256; do | |
if [ $((dp * tp * pp)) -eq 256 ]; then | |
exp_name="dp${dp}_tp${tp}_pp${pp}_mbs1_ga${grad_acc}_sl4096" | |
echo "Running with grad_acc=$grad_acc, mbs=1, dp=$dp, tp=$tp, pp=$pp" | |
python create_config.py \ | |
--out_dir result_blog_posts \ | |
--model_name meta-llama/Llama-2-70b-hf \ | |
--exp_name 70b_32_node_with_dp_tp_pp/$exp_name \ | |
--mbs=1 \ | |
--grad_acc=$grad_acc \ | |
--dp=$dp \ | |
--tp=$tp \ | |
--pp=$pp \ | |
--pp_engine 1f1b \ | |
--seq_len=4096 \ | |
--use_wandb \ | |
--hf_hub_safetensors_path /fsx/ferdinandmom/hf_model_ckpt/Llama-2-70b-chat-hf | |
fi | |
done | |
done | |
fi | |
done | |
done |