diff --git a/.gitattributes b/.gitattributes index fa9d40304e428e053ee3477e0b0b4bee03ccdaaf..ad19ab604240d67bff8d8681fab8562b4f3e4c02 100644 --- a/.gitattributes +++ b/.gitattributes @@ -132,3 +132,4 @@ dravid/tok_dravid/split/finalag_splitaa filter=lfs diff=lfs merge=lfs -text dravid/tok_dravid/split/finalcf_splitab filter=lfs diff=lfs merge=lfs -text dravid/tok_dravid/split/finalap_splitab filter=lfs diff=lfs merge=lfs -text dravid/tok_dravid/split/finalbt_splitaa filter=lfs diff=lfs merge=lfs -text +dravid/tok_dravid/split/finalbp_splitab filter=lfs diff=lfs merge=lfs -text diff --git a/dravid/tok_dravid/split/finalbp_splitab b/dravid/tok_dravid/split/finalbp_splitab new file mode 100644 index 0000000000000000000000000000000000000000..8133822e29b13f95dee46d1d80448e6199431221 --- /dev/null +++ b/dravid/tok_dravid/split/finalbp_splitab @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b37ee28c82aee7cbe04b3e8cf08521ffdf7a494dc1c3c23d29d7be65f74c634 +size 27918701591 diff --git a/english_only/tok_english_only/tok_files/tokenizer_0.sh b/english_only/tok_english_only/tok_files/tokenizer_0.sh new file mode 100644 index 0000000000000000000000000000000000000000..6306abea808020c17e90f2eda7f9bf6346109ac1 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_0.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +#sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split/finalaa +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer0_debug/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split/finalaa \ + --output-prefix $FINAL_DIR/tokenizer0_debug/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_1.sh b/english_only/tok_english_only/tok_files/tokenizer_1.sh new file mode 100644 index 0000000000000000000000000000000000000000..82b9c1442f33defbfb90d284a9c4d8d857360643 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_1.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalab +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer1/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalab \ + --output-prefix $FINAL_DIR/tokenizer1/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_10.sh b/english_only/tok_english_only/tok_files/tokenizer_10.sh new file mode 100644 index 0000000000000000000000000000000000000000..0f3783f32cbbd77850a8fa86091c889105da968d --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_10.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalak +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer10/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalak \ + --output-prefix $FINAL_DIR/tokenizer10/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_11.sh b/english_only/tok_english_only/tok_files/tokenizer_11.sh new file mode 100644 index 0000000000000000000000000000000000000000..ca15b3ddde5155dd9a44187dc722d96b9f7afdf0 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_11.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalal +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer11/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalal \ + --output-prefix $FINAL_DIR/tokenizer11/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_12.sh b/english_only/tok_english_only/tok_files/tokenizer_12.sh new file mode 100644 index 0000000000000000000000000000000000000000..2835538a5f37554a842cc4288ac9100e4ef18759 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_12.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalam +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer12/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalam \ + --output-prefix $FINAL_DIR/tokenizer12/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_13.sh b/english_only/tok_english_only/tok_files/tokenizer_13.sh new file mode 100644 index 0000000000000000000000000000000000000000..e1c50ad214acea171043119f3cd0741fab4a974e --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_13.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalan +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer13/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalan \ + --output-prefix $FINAL_DIR/tokenizer13/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_14.sh b/english_only/tok_english_only/tok_files/tokenizer_14.sh new file mode 100644 index 0000000000000000000000000000000000000000..b43e1256b55f813f600a13fdff42be3d126adbdb --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_14.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalao +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer14/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalao \ + --output-prefix $FINAL_DIR/tokenizer14/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_15.sh b/english_only/tok_english_only/tok_files/tokenizer_15.sh new file mode 100644 index 0000000000000000000000000000000000000000..19912f83b3d45628ce0ae3d579ba181365f307ab --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_15.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalap +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer15/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalap \ + --output-prefix $FINAL_DIR/tokenizer15/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_16.sh b/english_only/tok_english_only/tok_files/tokenizer_16.sh new file mode 100644 index 0000000000000000000000000000000000000000..0c6f3fcc1ba650fc74cdf7177a0c56e8db9a3284 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_16.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaq +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer16/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaq \ + --output-prefix $FINAL_DIR/tokenizer16/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_17.sh b/english_only/tok_english_only/tok_files/tokenizer_17.sh new file mode 100644 index 0000000000000000000000000000000000000000..66d332c1b07a983c1c2f3012c45f8cd457ba2df7 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_17.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalar +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer17/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalar \ + --output-prefix $FINAL_DIR/tokenizer17/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_18.sh b/english_only/tok_english_only/tok_files/tokenizer_18.sh new file mode 100644 index 0000000000000000000000000000000000000000..491c7075d7ddac2f5df27b554d3457c19c2591ed --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_18.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalas +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer18/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalas \ + --output-prefix $FINAL_DIR/tokenizer18/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_19.sh b/english_only/tok_english_only/tok_files/tokenizer_19.sh new file mode 100644 index 0000000000000000000000000000000000000000..9fb4147279f7cf54590616d4cdabc827bb43242a --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_19.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalat +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer19/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalat \ + --output-prefix $FINAL_DIR/tokenizer19/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_2.sh b/english_only/tok_english_only/tok_files/tokenizer_2.sh new file mode 100644 index 0000000000000000000000000000000000000000..aaf7bff9205cd940d37642aa47f1330ac7c08bab --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_2.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalac +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer2/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalac \ + --output-prefix $FINAL_DIR/tokenizer2/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_20.sh b/english_only/tok_english_only/tok_files/tokenizer_20.sh new file mode 100644 index 0000000000000000000000000000000000000000..6b6f40ddc4b111cf116c9b4700e418fff4ae350f --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_20.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalau +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer20/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalau \ + --output-prefix $FINAL_DIR/tokenizer20/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_21.sh b/english_only/tok_english_only/tok_files/tokenizer_21.sh new file mode 100644 index 0000000000000000000000000000000000000000..6a380a34be7f5cd07ea98a06de4d59204db5aa56 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_21.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalav +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer21/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalav \ + --output-prefix $FINAL_DIR/tokenizer21/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_22.sh b/english_only/tok_english_only/tok_files/tokenizer_22.sh new file mode 100644 index 0000000000000000000000000000000000000000..62df80c3cd5c97f2c3444a359fe633c28d0ef86b --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_22.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaw +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer22/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaw \ + --output-prefix $FINAL_DIR/tokenizer22/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_23.sh b/english_only/tok_english_only/tok_files/tokenizer_23.sh new file mode 100644 index 0000000000000000000000000000000000000000..db08b6a7a6eb4773646b4c9d518d0694fb1ab905 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_23.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalax +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer23/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalax \ + --output-prefix $FINAL_DIR/tokenizer23/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_24.sh b/english_only/tok_english_only/tok_files/tokenizer_24.sh new file mode 100644 index 0000000000000000000000000000000000000000..bdaaa6f6039454fa7fe4d797f7ba6023151f153f --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_24.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalay +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer24/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalay \ + --output-prefix $FINAL_DIR/tokenizer24/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_25.sh b/english_only/tok_english_only/tok_files/tokenizer_25.sh new file mode 100644 index 0000000000000000000000000000000000000000..490e2dc4b523129ffee36b0adedb66b7437a65ce --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_25.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaz +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer25/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaz \ + --output-prefix $FINAL_DIR/tokenizer25/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_26.sh b/english_only/tok_english_only/tok_files/tokenizer_26.sh new file mode 100644 index 0000000000000000000000000000000000000000..f5541f657ee8d08436be59ba8527e51fe1a46f4c --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_26.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalba +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer26/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalba \ + --output-prefix $FINAL_DIR/tokenizer26/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_27.sh b/english_only/tok_english_only/tok_files/tokenizer_27.sh new file mode 100644 index 0000000000000000000000000000000000000000..7d11bb9cbbb4985ceea96e948e83c219f4bbea5c --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_27.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbb +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer27/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbb \ + --output-prefix $FINAL_DIR/tokenizer27/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_28.sh b/english_only/tok_english_only/tok_files/tokenizer_28.sh new file mode 100644 index 0000000000000000000000000000000000000000..849721a47aa8144ad6195c7f2fafb41d188395c3 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_28.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbc +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer28/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbc \ + --output-prefix $FINAL_DIR/tokenizer28/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_29.sh b/english_only/tok_english_only/tok_files/tokenizer_29.sh new file mode 100644 index 0000000000000000000000000000000000000000..689a2429565c64d5071481203cfbfd6afb746366 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_29.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbd +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer29/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbd \ + --output-prefix $FINAL_DIR/tokenizer29/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_3.sh b/english_only/tok_english_only/tok_files/tokenizer_3.sh new file mode 100644 index 0000000000000000000000000000000000000000..6d9fc8efa210a048718a1f985db3aaea35ea76ca --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_3.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalad +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer3/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalad \ + --output-prefix $FINAL_DIR/tokenizer3/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_30.sh b/english_only/tok_english_only/tok_files/tokenizer_30.sh new file mode 100644 index 0000000000000000000000000000000000000000..097cc1f8ee6bcc8c2bbae10921ab2bcec946860c --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_30.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbe +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer30/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbe \ + --output-prefix $FINAL_DIR/tokenizer30/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_31.sh b/english_only/tok_english_only/tok_files/tokenizer_31.sh new file mode 100644 index 0000000000000000000000000000000000000000..d7394a4f04b11da05f818a8b2b06ce5a8d209d15 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_31.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbf +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer31/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbf \ + --output-prefix $FINAL_DIR/tokenizer31/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_32.sh b/english_only/tok_english_only/tok_files/tokenizer_32.sh new file mode 100644 index 0000000000000000000000000000000000000000..dd220cabdb42e322583ccfa9e579158076ff22e5 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_32.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbg +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer32/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbg \ + --output-prefix $FINAL_DIR/tokenizer32/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_33.sh b/english_only/tok_english_only/tok_files/tokenizer_33.sh new file mode 100644 index 0000000000000000000000000000000000000000..14eaf28db4a96c5d4aa895a260ad0bfd54b4dbc3 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_33.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbh +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer33/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbh \ + --output-prefix $FINAL_DIR/tokenizer33/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_34.sh b/english_only/tok_english_only/tok_files/tokenizer_34.sh new file mode 100644 index 0000000000000000000000000000000000000000..9f165714c826519bc1c6b844ecd1daccc7e26390 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_34.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbi +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer34/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbi \ + --output-prefix $FINAL_DIR/tokenizer34/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_35.sh b/english_only/tok_english_only/tok_files/tokenizer_35.sh new file mode 100644 index 0000000000000000000000000000000000000000..6dcf9577399604428f91230ad8787c456e010c82 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_35.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbj +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer35/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbj \ + --output-prefix $FINAL_DIR/tokenizer35/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_36.sh b/english_only/tok_english_only/tok_files/tokenizer_36.sh new file mode 100644 index 0000000000000000000000000000000000000000..a26f364ab2039269be3adbfd38d12609995ab954 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_36.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbk +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer36/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbk \ + --output-prefix $FINAL_DIR/tokenizer36/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_37.sh b/english_only/tok_english_only/tok_files/tokenizer_37.sh new file mode 100644 index 0000000000000000000000000000000000000000..5e265ec77b88a01c5efce9534ef68e8e00eb9f41 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_37.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbl +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer37/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbl \ + --output-prefix $FINAL_DIR/tokenizer37/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_38.sh b/english_only/tok_english_only/tok_files/tokenizer_38.sh new file mode 100644 index 0000000000000000000000000000000000000000..fe731219ef87e19b81b2761c47e70b1ca19bb8ae --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_38.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbm +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer38/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbm \ + --output-prefix $FINAL_DIR/tokenizer38/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_39.sh b/english_only/tok_english_only/tok_files/tokenizer_39.sh new file mode 100644 index 0000000000000000000000000000000000000000..40f779a6d9a9e12ddcf0785d4a5fe5ffb44e6f20 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_39.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbn +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer39/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalbn \ + --output-prefix $FINAL_DIR/tokenizer39/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_4.sh b/english_only/tok_english_only/tok_files/tokenizer_4.sh new file mode 100644 index 0000000000000000000000000000000000000000..c7e9f1aa663810ae1395be4589b37ecc7b58c7f0 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_4.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalae +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer4/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalae \ + --output-prefix $FINAL_DIR/tokenizer4/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_5.sh b/english_only/tok_english_only/tok_files/tokenizer_5.sh new file mode 100644 index 0000000000000000000000000000000000000000..58494224d1e125eeaf0d1e8fcfc86315c13ad73a --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_5.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaf +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer5/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaf \ + --output-prefix $FINAL_DIR/tokenizer5/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_6.sh b/english_only/tok_english_only/tok_files/tokenizer_6.sh new file mode 100644 index 0000000000000000000000000000000000000000..0f0f47ed29ae3726c43e401b24b932102ce3634b --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_6.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalag +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer6/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalag \ + --output-prefix $FINAL_DIR/tokenizer6/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_7.sh b/english_only/tok_english_only/tok_files/tokenizer_7.sh new file mode 100644 index 0000000000000000000000000000000000000000..94a4f8f06f57d44c0383d2e39f2cafdd5cccbb8f --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_7.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalah +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer7/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalah \ + --output-prefix $FINAL_DIR/tokenizer7/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_8.sh b/english_only/tok_english_only/tok_files/tokenizer_8.sh new file mode 100644 index 0000000000000000000000000000000000000000..1fbe808a3b3f0fd12df3ecafaf722cd87326d2ab --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_8.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalai +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer8/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalai \ + --output-prefix $FINAL_DIR/tokenizer8/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_files/tokenizer_9.sh b/english_only/tok_english_only/tok_files/tokenizer_9.sh new file mode 100644 index 0000000000000000000000000000000000000000..18d06dbff0e61642f1d35a4dbd853ca7b1a20b70 --- /dev/null +++ b/english_only/tok_english_only/tok_files/tokenizer_9.sh @@ -0,0 +1,25 @@ +#!/bin/bash +set -m +sed -i -e "s/raw_content/text/g" /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaj +echo "above deepspeed$" +FINAL_DIR=/mnt/weka/peacock/idc/hineng/english_only/ +#TOKENIZER=facebook/nllb-200-distilled-600M +#TOKENIZER_TYPE=HuggingFaceTokenizer +TOKENIZER_PATH=/mnt/weka/peacock/tokenization/trained-tokenizer/encode_32k/ +TOKENIZER=$TOKENIZER_PATH/en_code_32kspm_tokenizer.model +VOCAB_FILE=$TOKENIZER_PATH/en_code_32kspm_tokenizer.vocab +TOKENIZER_TYPE=SentencePieceTokenizer +mkdir -p $FINAL_DIR/tokenizer9/ +cd /mnt/weka/peacock/experiments/llama/Megatron-DeepSpeed +echo "inside deepspeed" +pwd +python3 tools/preprocess_data.py \ + --input /mnt/weka/peacock/idc/datasets/english_only//tok_english_only/split//finalaj \ + --output-prefix $FINAL_DIR/tokenizer9/ \ + --tokenizer-model $TOKENIZER \ + --vocab-file $VOCAB_FILE \ + --dataset-impl mmap --tokenizer-type $TOKENIZER_TYPE\ + --append-eod --workers 8 # --partitions 16 #--chunk-size 50 + + # --merge-file $MERGES_FILE \ + diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_0.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_0.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b852e4ba86c812b8bd7df91fe443081b2ac25399 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_0.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-0 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-0 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-0-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_0.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_1.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_1.yaml new file mode 100644 index 0000000000000000000000000000000000000000..718e60ef0ab180cd74bd8d1eb575b5887bdf25d8 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_1.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-1 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-1 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-1-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_1.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_10.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_10.yaml new file mode 100644 index 0000000000000000000000000000000000000000..10afb9eaf4eef7c24b0d3df5b4c287244b2f2618 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_10.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-10 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-10 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-10-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_10.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_11.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_11.yaml new file mode 100644 index 0000000000000000000000000000000000000000..afebaed3e2af197b2c66ad3c0cb4b10f905d303b --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_11.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-11 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-11 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-11-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_11.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_12.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_12.yaml new file mode 100644 index 0000000000000000000000000000000000000000..897b29f4060adc7a6b06a3864279bca429d25795 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_12.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-12 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-12 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-12-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_12.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_13.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_13.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a2abafbb14b3f64ac8c9cbe0d77e99bcbd2c04b8 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_13.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-13 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-13 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-13-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_13.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_16.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_16.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1796ce4fdd191f87a7c5f0e25dd5e13cca3811e9 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_16.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-16 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-16 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-16-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_16.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_17.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_17.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0a2e35eb696ac1c1f5ccd6e2c389ea708bbe2b6e --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_17.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-17 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-17 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-17-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_17.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_18.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_18.yaml new file mode 100644 index 0000000000000000000000000000000000000000..453edb15e8d6f26e59b9e8cdac0be5d3fa0d82d8 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_18.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-18 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-18 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-18-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_18.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_19.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_19.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8401211c4a1795ee9e455b2b2c7bba4d38e78e7a --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_19.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-19 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-19 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-19-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_19.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_2.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_2.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d354a22f14cdad5d2889d7c91cd2fa4f25a23594 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_2.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-2 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-2 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-2-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_2.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_20.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_20.yaml new file mode 100644 index 0000000000000000000000000000000000000000..05507fb850c898e4296d6ec2fddfdbbccea2f88d --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_20.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-20 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-20 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-20-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_20.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_21.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_21.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c5600c37b8c251e0e61614d2f0187f6b494ad236 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_21.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-21 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-21 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-21-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_21.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_24.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_24.yaml new file mode 100644 index 0000000000000000000000000000000000000000..72bb27b158e7594ee653d9e03ea963873b868e9b --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_24.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-24 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-24 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-24-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_24.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_25.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_25.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5e16b3e992e98735d9840c7af40ef0df6e6acfd0 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_25.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-25 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-25 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-25-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_25.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_26.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_26.yaml new file mode 100644 index 0000000000000000000000000000000000000000..406f7ffc9911e577d9aee97b14165bf0e78cd622 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_26.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-26 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-26 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-26-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_26.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_27.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_27.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a5820ea1f0b236ea41d91923e70103479bc529ac --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_27.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-27 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-27 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-27-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_27.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_28.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_28.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2a4b0107f0252dde0e1630981f89b1af8a3e8704 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_28.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-28 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-28 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-28-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_28.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_29.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_29.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cfb37acc7ae1de49bae34eb1e3ac5b5333890fcd --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_29.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-29 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-29 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-29-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_29.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_3.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_3.yaml new file mode 100644 index 0000000000000000000000000000000000000000..425bf6afc4e9de1be01ac4871f997006212a2b9e --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_3.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-3 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-3 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-3-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_3.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_30.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_30.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6fc9aece5b4cc2d730da21a9b81a94c45efda466 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_30.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-30 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-30 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-30-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_30.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_31.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_31.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b607a66ad9df1ecb09534dd2312cb24d1549a189 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_31.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-31 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-31 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-31-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_31.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_32.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_32.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4f2ec7efccc1b890c3ca7544e33533d4c240729d --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_32.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-32 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-32 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-32-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_32.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_33.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_33.yaml new file mode 100644 index 0000000000000000000000000000000000000000..941d9632b8f0596433ef93c967fe5f07a2d26739 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_33.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-33 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-33 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-33-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_33.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_35.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_35.yaml new file mode 100644 index 0000000000000000000000000000000000000000..35950da59834a45b25869cbf3782b70f0ea4981e --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_35.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-35 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-35 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-35-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_35.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_36.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_36.yaml new file mode 100644 index 0000000000000000000000000000000000000000..185fc91e569de8572dde51897e5bf441a42ee8e1 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_36.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-36 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-36 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-36-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_36.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_37.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_37.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f16edfde82d0bdbf3de5de4205dd3cb506e948e0 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_37.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-37 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-37 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-37-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_37.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_38.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_38.yaml new file mode 100644 index 0000000000000000000000000000000000000000..17267580b9b7c16e7cb3083ef43b40c373a6e11f --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_38.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-38 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-38 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-38-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_38.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_4.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_4.yaml new file mode 100644 index 0000000000000000000000000000000000000000..132ea35dd5e15b3d66df9327e2d4f45aab371392 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_4.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-4 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-4 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-4-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_4.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_6.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_6.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2fc012695ada27ffaa2031ace0614ccb475c655e --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_6.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-6 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-6 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-6-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_6.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_7.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_7.yaml new file mode 100644 index 0000000000000000000000000000000000000000..904496ceae9d1befb74b3e80ab3d54840c2b9691 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_7.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-7 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-7 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-7-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_7.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_8.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_8.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5cc2ff1c16835d196e86d362c4fb52ffaf00953f --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_8.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-8 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-8 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-8-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_8.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock diff --git a/english_only/tok_english_only/tok_jobs_check/tokenizer_9.yaml b/english_only/tok_english_only/tok_jobs_check/tokenizer_9.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3fcb21aed17160d4093e7aa5a91500e629c0a940 --- /dev/null +++ b/english_only/tok_english_only/tok_jobs_check/tokenizer_9.yaml @@ -0,0 +1,104 @@ +apiVersion: kubeflow.org/v2beta1 +kind: MPIJob +metadata: + name: peacock-tokenizer-english-only-9 + namespace: peacock + labels: + app: peacock-tokenizer-english-only-9 +spec: + slotsPerWorker: 8 + runPolicy: + backoffLimit: 1 + cleanPodPolicy: Running + mpiReplicaSpecs: + Launcher: + replicas: 1 + template: + spec: + hostIPC: true + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-tokenizer-english-only-9-launcher + imagePullPolicy: IfNotPresent + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock + command: ["/bin/bash", "-c"] + args: + - >- + + export SYNAPSE_VERSION="1.15.1"; + export WORKER_DIR="/mnt/weka/peacock/experiments/llama"; + + export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh" + export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/english_only//tok_english_only/tok_files_check/tokenizer_9.sh + HOSTSFILE=$OMPI_MCA_orte_default_hostfile; + echo "HOSTSFILE=$HOSTSFILE"; + MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)"; + NUM_NODES=$(wc -l < $HOSTSFILE); + CARDS_PER_NODE=8; + N_CARDS=$((NUM_NODES*CARDS_PER_NODE)); + echo "MPI_ROOT=$MPI_ROOT"; + echo "N_CARDS=$N_CARDS"; + echo "MASTER_ADDR=$MASTER_ADDR"; + sleep 20; + + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD; + + mpirun -np $N_CARDS -npernode 8 \ + --tag-output \ + --allow-run-as-root \ + --prefix $MPI_ROOT \ + -x WORKER_DIR=$WORKER_DIR \ + -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD; + + + Worker: + replicas: 1 + template: + spec: + volumes: + - name: work-dir + persistentVolumeClaim: + claimName: peacock-fs-pvc + tolerations: + - key: "habana.ai/gaudi" + operator: "Exists" + effect: "NoSchedule" + - key: "k8s/namespace" + operator: "Equal" + value: "peacock" + effect: "NoSchedule" + hostIPC: true + containers: + - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest + name: peacock-llama-worker + imagePullPolicy: IfNotPresent + securityContext: + capabilities: + add: + - SYSLOG + resources: + limits: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + requests: + habana.ai/gaudi: 8 + hugepages-2Mi: 300Gi + memory: 700Gi + cpu: 150 + volumeMounts: + - name: work-dir + mountPath: /mnt/weka/peacock