#!/bin/bash # ✅ # Saturate DP for 1B on 1 node # for grad_acc in 2 4 8 16 32 64 128; do # for mbs in 64 32 16 8 4 2 1; do # if [ $((grad_acc * mbs)) -eq 128 ]; then # echo "Running with grad_acc=$grad_acc and mbs=$mbs" # exp_name="dp8_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # python create_config.py --out_dir result_blog_posts --model_name TinyLlama/TinyLlama_v1.1 --exp_name saturate_dp_1b_1_node/$exp_name --mbs=$mbs --grad_acc=$grad_acc --dp=8 --tp=1 --pp=1 --seq_len=4096 --use_wandb # fi # done # done # ✅ # Saturate DP for 1B on 8 node # for grad_acc in 1 2 4 8 16; do # for mbs in 1 2 4 8 16; do # if [ $((grad_acc * mbs)) -eq 16 ]; then # exp_name="dp64_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc and mbs=$mbs" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name TinyLlama/TinyLlama_v1.1 \ # --exp_name saturate_dp_1b_8_node/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=64 \ # --tp=1 \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # done # ✅ # # Saturate DP for 1B on 32 node # for grad_acc in 1 2 4; do # for mbs in 1 2 4; do # if [ $((grad_acc * mbs)) -eq 4 ]; then # exp_name="dp256_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc and mbs=$mbs" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name TinyLlama/TinyLlama_v1.1 \ # --exp_name saturate_dp_1b_32_node/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=256 \ # --tp=1 \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # done # ✅ # # Sature DP for 1b on 1 node (with TP) # for grad_acc in 1 2 4 8 16 32 64 128 256 512; do # for mbs in 1 2 4 8 16 32 64 128 256 512; do # for dp in 1 2 4; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for tp in 1 2 4 8; do # if [ $((dp * tp)) -eq 8 ]; then # exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name TinyLlama/TinyLlama_v1.1 \ # --exp_name saturate_dp_1b_1_node_tp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=$tp \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # ✅ # # Sature DP for 1b on 8 node (with TP) # for grad_acc in 1 2 4 8 16 32 64 128 256 512; do # for mbs in 1 2 4 8 16 32 64 128 256 512; do # for dp in 1 2 4 8 16 32; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for tp in 1 2 4 8; do # if [ $((dp * tp)) -eq 64 ]; then # exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name TinyLlama/TinyLlama_v1.1 \ # --exp_name saturate_dp_1b_8_node_tp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=$tp \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # ✅ # # Saturated DP for 1B on 32 nodes (with TP) # for grad_acc in 1 2 4 8 16 32 64 128 256 512; do # for mbs in 1 2 4 8 16 32 64 128 256 512; do # for dp in 1 2 4 8 16 32 64 128; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for tp in 1 2 4 8; do # if [ $((dp * tp)) -eq 256 ]; then # exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name TinyLlama/TinyLlama_v1.1 \ # --exp_name saturate_dp_1b_32_node_tp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=$tp \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # # ✅ # # Sature DP for 1b on 1 node (with PP 1f1b) # for grad_acc in 1 2 4 8 16 32 64 128 256 512; do # for mbs in 1 2 4 8 16 32 64 128 256 512; do # for dp in 1 2 4; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for pp in 1 2 4 8; do # if [ $((dp * pp)) -eq 8 ]; then # exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=1, pp=$pp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name TinyLlama/TinyLlama_v1.1 \ # --exp_name saturate_dp_1b_1_node_pp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=1 \ # --pp=$pp \ # --pp_engine 1f1b \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # # ✅ # # Sature DP for 1b on 8 node (with PP 1f1b) # for grad_acc in 1 2 4 8 16 32 64 128 256 512; do # for mbs in 1 2 4 8 16 32 64 128 256 512; do # for dp in 1 2 4 8 16 32; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for pp in 1 2 4 8 16 32 64; do # if [ $((dp * pp)) -eq 64 ]; then # exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=1, pp=$pp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name TinyLlama/TinyLlama_v1.1 \ # --exp_name saturate_dp_1b_8_node_pp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=1 \ # --pp=$pp \ # --pp_engine 1f1b \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # # ✅ # # Sature DP for 1b on 32 node (with PP 1f1b) # for grad_acc in 1 2 4 8 16 32 64 128 256 512; do # for mbs in 1 2 4 8 16 32 64 128 256 512; do # for dp in 1 2 4 8 16 32 64 128; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for pp in 1 2 4 8 16 32 64 128 256; do # if [ $((dp * pp)) -eq 256 ]; then # exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=1, pp=$pp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name TinyLlama/TinyLlama_v1.1 \ # --exp_name saturate_dp_1b_32_node_pp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=1 \ # --pp=$pp \ # --pp_engine 1f1b \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # ✅ # # Try to fit 7B on 1 node # for grad_acc in 1 2 4 8 16 32 64 128; do # for mbs in 1 2 4 8 16 32 64 128; do # if [ $((grad_acc * mbs)) -eq 128 ]; then # exp_name="dp8_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc and mbs=$mbs" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name meta-llama/Llama-2-7b-hf \ # --exp_name try_to_fit_7b_1_node/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=8 \ # --tp=1 \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # done # ✅ # # Saturate DP for 7B on 8 nodes (with TP) # for grad_acc in 1 2 4 8 16 32 64 128 256; do # for mbs in 1 2 4 8 16; do # for dp in 1 2 4 8 16 32 64; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for tp in 1 2 4 8; do # if [ $((dp * tp)) -eq 64 ]; then # exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name meta-llama/Llama-2-7b-hf \ # --exp_name saturate_dp_7b_8_node_tp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=$tp \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # ✅ # # Saturate DP for 7B on 8 nodes and TP=16 # for grad_acc in 1 2 4 8 16 32 64 128 256; do # for mbs in 1 2 4 8 16; do # for dp in 1 2 4; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # tp=16 # if [ $((dp * tp)) -eq 64 ]; then # exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name meta-llama/Llama-2-7b-hf \ # --exp_name saturate_dp_7b_8_node_tp16/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=$tp \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # fi # done # done # done # ✅ # # Saturate DP for 7B on 32 nodes # for grad_acc in 1 2 4; do # for mbs in 1 2 4; do # if [ $((grad_acc * mbs)) -eq 4 ]; then # exp_name="dp256_tp1_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc and mbs=$mbs" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name meta-llama/Llama-2-7b-hf \ # --exp_name saturate_dp_7b_32_node/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=256 \ # --tp=1 \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # done # ✅ # # Saturate DP for 7B on 32 nodes (with TP) # for grad_acc in 1 2 4 8 16 32 64 128 256; do # for mbs in 1 2 4 8 16; do # for dp in 1 2 4 8 16 32 64 128 256; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for tp in 1 2 4 8; do # if [ $((dp * tp)) -eq 256 ]; then # exp_name="dp${dp}_tp${tp}_pp1_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, tp=$tp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name meta-llama/Llama-2-7b-hf \ # --exp_name saturate_dp_7b_32_node_tp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=$tp \ # --pp=1 \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # # ✅ # # Saturate DP for 7B on 8 nodes (with PP) # for grad_acc in 1 2 4 8 16 32 64 128 256; do # for mbs in 1 2 4 8 16; do # for dp in 1 2 4 8 16 32 64; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for pp in 1 2 4 8 16 32 64; do # if [ $((dp * pp)) -eq 64 ]; then # exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, pp=$pp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name meta-llama/Llama-2-7b-hf \ # --exp_name saturate_dp_7b_8_node_pp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=1 \ # --pp=$pp \ # --pp_engine 1f1b \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # # ✅ # # Saturate DP for 7B on 8 nodes and PP=16 # for grad_acc in 1 2 4 8 16 32 64 128 256; do # for mbs in 1 2 4 8 16; do # for dp in 1 2 4; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # pp=16 # if [ $((dp * pp)) -eq 64 ]; then # exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, pp=$pp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name meta-llama/Llama-2-7b-hf \ # --exp_name saturate_dp_7b_8_node_pp16/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=1 \ # --pp=$pp \ # --pp_engine 1f1b \ # --seq_len=4096 \ # --use_wandb # fi # fi # done # done # done # # ✅ # # Saturate DP for 7B on 32 nodes (with PP) # for grad_acc in 1 2 4 8 16 32 64 128 256; do # for mbs in 1 2 4 8 16; do # for dp in 1 2 4 8 16 32 64 128; do # if [ $((dp * mbs * grad_acc)) -eq 1024 ]; then # for pp in 1 2 4 8 16 32 64 128 256; do # if [ $((dp * pp)) -eq 256 ]; then # exp_name="dp${dp}_tp1_pp${pp}_mbs${mbs}_ga${grad_acc}_sl4096" # echo "Running with grad_acc=$grad_acc, mbs=$mbs, dp=$dp, pp=$pp" # python create_config.py \ # --out_dir result_blog_posts \ # --model_name meta-llama/Llama-2-7b-hf \ # --exp_name saturate_dp_7b_32_node_pp/$exp_name \ # --mbs=$mbs \ # --grad_acc=$grad_acc \ # --dp=$dp \ # --tp=1 \ # --pp=$pp \ # --pp_engine 1f1b \ # --seq_len=4096 \ # --use_wandb # fi # done # fi # done # done # done # ⌛ # 70B on 8 nodes (with DP + TP + PP) for grad_acc in 1 2 4 8 16 32 64 128 256 512 1024; do for dp in 1 2 4 8 16 32 64; do if [ $((dp * grad_acc)) -eq 1024 ]; then for tp in 2 4 8; do for pp in 2 4 8 16 32 64; do if [ $((dp * tp * pp)) -eq 64 ]; then exp_name="dp${dp}_tp${tp}_pp${pp}_mbs1_ga${grad_acc}_sl4096" echo "Running with grad_acc=$grad_acc, mbs=1, dp=$dp, tp=$tp, pp=$pp" python create_config.py \ --out_dir result_blog_posts \ --model_name meta-llama/Llama-2-70b-hf \ --exp_name 70b_8_node_with_dp_tp_pp/$exp_name \ --mbs=1 \ --grad_acc=$grad_acc \ --dp=$dp \ --tp=$tp \ --pp=$pp \ --pp_engine 1f1b \ --seq_len=4096 \ --use_wandb \ --hf_hub_safetensors_path /fsx/ferdinandmom/hf_model_ckpt/Llama-2-70b-chat-hf fi done done fi done done # ⌛ # 70B on 32 nodes (with DP + TP + PP) for grad_acc in 1 2 4 8 16 32 64 128 256 512 1024; do for dp in 1 2 4 8 16 32 64 128 256; do if [ $((dp * grad_acc)) -eq 1024 ]; then for tp in 2 4 8; do for pp in 2 4 8 16 32 64 128 256; do if [ $((dp * tp * pp)) -eq 256 ]; then exp_name="dp${dp}_tp${tp}_pp${pp}_mbs1_ga${grad_acc}_sl4096" echo "Running with grad_acc=$grad_acc, mbs=1, dp=$dp, tp=$tp, pp=$pp" python create_config.py \ --out_dir result_blog_posts \ --model_name meta-llama/Llama-2-70b-hf \ --exp_name 70b_32_node_with_dp_tp_pp/$exp_name \ --mbs=1 \ --grad_acc=$grad_acc \ --dp=$dp \ --tp=$tp \ --pp=$pp \ --pp_engine 1f1b \ --seq_len=4096 \ --use_wandb \ --hf_hub_safetensors_path /fsx/ferdinandmom/hf_model_ckpt/Llama-2-70b-chat-hf fi done done fi done done