| #SBATCH --job-name=ofa_large_pretrain_s3 | |
| #SBATCH --nodes=32 | |
| #SBATCH --ntasks=32 | |
| #SBATCH --gpus=256 | |
| #SBATCH --threads-per-core=2 | |
| #SBATCH --gpu-bind=closest | |
| ####SBATCH --nodelist=x1004c7s2b1n0,x1004c7s3b0n0,x1004c7s3b1n0,x1004c7s4b0n0 | |
| #SBATCH --mail-type=END,FAIL | |
| #SBATCH --output=/lus/home/NAT/gda2204/mshukor/logs/slurm/ofa_large_pretrain_s3.out | |
| #SBATCH --exclusive | |
| #SBATCH --time=1:00:00 | |
| #SBATCH -C MI250 | |
| #SBATCH -A gda2204 | |
| #SBATCH --cpus-per-task=128 | |
| #SBATCH [email protected] | |
| cd /lus/home/NAT/gda2204/mshukor/code/ofa_ours/run_scripts | |
| source /lus/home/NAT/gda2204/mshukor/.bashrc | |
| conda activate main | |
| rm core-* | |
| srun -l -N 32 -n 32 -c 128 --gpus=256 --gpu-bind=closest bash pretraining/scaling/ofa_large_pretrain_s3.sh |