applied-ai-018 commited on
Commit
558c60a
·
verified ·
1 Parent(s): db9c742

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_106.yaml +104 -0
  2. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_131.yaml +104 -0
  3. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_154.yaml +104 -0
  4. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_156.yaml +104 -0
  5. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_167.yaml +104 -0
  6. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_177.yaml +104 -0
  7. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_180.yaml +104 -0
  8. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_186.yaml +104 -0
  9. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_187.yaml +104 -0
  10. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_193.yaml +104 -0
  11. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_197.yaml +104 -0
  12. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_198.yaml +104 -0
  13. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_215.yaml +104 -0
  14. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_233.yaml +104 -0
  15. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_279.yaml +104 -0
  16. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_285.yaml +104 -0
  17. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_290.yaml +104 -0
  18. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_293.yaml +104 -0
  19. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_295.yaml +104 -0
  20. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_299.yaml +104 -0
  21. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_309.yaml +104 -0
  22. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_323.yaml +104 -0
  23. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_337.yaml +104 -0
  24. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_341.yaml +104 -0
  25. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_343.yaml +104 -0
  26. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_344.yaml +104 -0
  27. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_353.yaml +104 -0
  28. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_377.yaml +104 -0
  29. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_381.yaml +104 -0
  30. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_384.yaml +104 -0
  31. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_404.yaml +104 -0
  32. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_418.yaml +104 -0
  33. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_425.yaml +104 -0
  34. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_447.yaml +104 -0
  35. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_454.yaml +104 -0
  36. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_455.yaml +104 -0
  37. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_464.yaml +104 -0
  38. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_496.yaml +104 -0
  39. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_5.yaml +104 -0
  40. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_527.yaml +104 -0
  41. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_529.yaml +104 -0
  42. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_531.yaml +104 -0
  43. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_536.yaml +104 -0
  44. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_548.yaml +104 -0
  45. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_55.yaml +104 -0
  46. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_557.yaml +104 -0
  47. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_563.yaml +104 -0
  48. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_572.yaml +104 -0
  49. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_580.yaml +104 -0
  50. hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_581.yaml +104 -0
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_106.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-106
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-106
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-106-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_106.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_131.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-131
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-131
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-131-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_131.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_154.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-154
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-154
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-154-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_154.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_156.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-156
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-156
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-156-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_156.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_167.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-167
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-167
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-167-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_167.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_177.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-177
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-177
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-177-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_177.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_180.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-180
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-180
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-180-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_180.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_186.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-186
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-186
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-186-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_186.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_187.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-187
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-187
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-187-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_187.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_193.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-193
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-193
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-193-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_193.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_197.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-197
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-197
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-197-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_197.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_198.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-198
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-198
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-198-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_198.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_215.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-215
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-215
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-215-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_215.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_233.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-233
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-233
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-233-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_233.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_279.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-279
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-279
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-279-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_279.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_285.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-285
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-285
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-285-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_285.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_290.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-290
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-290
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-290-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_290.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_293.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-293
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-293
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-293-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_293.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_295.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-295
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-295
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-295-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_295.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_299.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-299
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-299
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-299-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_299.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_309.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-309
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-309
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-309-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_309.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_323.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-323
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-323
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-323-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_323.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_337.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-337
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-337
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-337-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_337.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_341.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-341
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-341
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-341-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_341.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_343.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-343
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-343
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-343-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_343.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_344.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-344
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-344
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-344-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_344.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_353.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-353
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-353
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-353-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_353.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_377.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-377
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-377
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-377-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_377.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_381.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-381
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-381
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-381-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_381.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_384.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-384
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-384
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-384-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_384.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_404.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-404
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-404
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-404-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_404.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_418.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-418
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-418
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-418-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_418.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_425.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-425
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-425
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-425-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_425.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_447.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-447
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-447
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-447-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_447.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_454.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-454
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-454
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-454-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_454.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_455.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-455
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-455
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-455-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_455.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_464.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-464
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-464
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-464-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_464.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_496.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-496
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-496
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-496-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_496.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_5.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-5
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-5
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-5-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_5.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_527.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-527
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-527
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-527-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_527.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_529.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-529
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-529
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-529-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_529.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_531.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-531
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-531
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-531-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_531.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_536.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-536
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-536
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-536-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_536.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_548.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-548
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-548
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-548-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_548.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_55.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-55
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-55
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-55-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_55.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_557.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-557
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-557
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-557-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_557.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_563.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-563
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-563
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-563-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_563.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_572.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-572
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-572
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-572-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_572.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_580.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-580
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-580
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-580-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_580.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs/tokenizer_581.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-581
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-581
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-581-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_581.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock