applied-ai-018 commited on
Commit
0274b70
·
verified ·
1 Parent(s): fbdd3dd

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_1.yaml +104 -0
  2. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_13.yaml +104 -0
  3. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_16.yaml +104 -0
  4. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_19.yaml +104 -0
  5. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_22.yaml +104 -0
  6. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_28.yaml +104 -0
  7. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_3.yaml +104 -0
  8. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_33.yaml +104 -0
  9. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_35.yaml +104 -0
  10. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_37.yaml +104 -0
  11. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_39.yaml +104 -0
  12. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_4.yaml +104 -0
  13. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_41.yaml +104 -0
  14. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_43.yaml +104 -0
  15. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_46.yaml +104 -0
  16. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_48.yaml +104 -0
  17. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_49.yaml +104 -0
  18. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_5.yaml +104 -0
  19. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_50.yaml +104 -0
  20. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_51.yaml +104 -0
  21. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_52.yaml +104 -0
  22. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_55.yaml +104 -0
  23. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_58.yaml +104 -0
  24. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_59.yaml +104 -0
  25. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_6.yaml +104 -0
  26. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_60.yaml +104 -0
  27. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_61.yaml +104 -0
  28. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_64.yaml +104 -0
  29. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_65.yaml +104 -0
  30. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_67.yaml +104 -0
  31. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_69.yaml +104 -0
  32. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_7.yaml +104 -0
  33. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_70.yaml +104 -0
  34. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_71.yaml +104 -0
  35. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_72.yaml +104 -0
  36. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_74.yaml +104 -0
  37. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_75.yaml +104 -0
  38. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_76.yaml +104 -0
  39. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_77.yaml +104 -0
  40. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_78.yaml +104 -0
  41. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_79.yaml +104 -0
  42. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_80.yaml +104 -0
  43. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_81.yaml +104 -0
  44. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_82.yaml +104 -0
  45. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_83.yaml +104 -0
  46. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_85.yaml +104 -0
  47. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_86.yaml +104 -0
  48. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_9.yaml +104 -0
  49. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_92.yaml +104 -0
  50. hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_95.yaml +104 -0
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_1.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-1
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-1
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-1-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_1.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_13.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-13
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-13
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-13-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_13.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_16.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-16
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-16
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-16-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_16.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_19.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-19
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-19
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-19-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_19.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_22.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-22
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-22
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-22-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_22.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_28.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-28
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-28
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-28-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_28.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_3.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-3
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-3
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-3-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_3.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_33.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-33
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-33
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-33-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_33.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_35.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-35
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-35
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-35-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_35.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_37.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-37
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-37
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-37-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_37.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_39.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-39
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-39
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-39-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_39.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_4.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-4
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-4
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-4-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_4.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_41.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-41
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-41
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-41-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_41.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_43.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-43
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-43
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-43-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_43.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_46.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-46
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-46
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-46-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_46.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_48.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-48
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-48
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-48-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_48.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_49.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-49
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-49
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-49-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_49.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_5.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-5
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-5
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-5-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_5.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_50.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-50
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-50
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-50-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_50.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_51.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-51
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-51
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-51-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_51.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_52.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-52
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-52
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-52-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_52.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_55.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-55
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-55
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-55-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_55.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_58.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-58
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-58
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-58-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_58.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_59.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-59
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-59
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-59-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_59.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_6.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-6
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-6
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-6-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_6.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_60.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-60
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-60
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-60-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_60.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_61.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-61
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-61
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-61-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_61.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_64.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-64
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-64
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-64-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_64.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_65.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-65
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-65
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-65-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_65.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_67.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-67
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-67
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-67-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_67.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_69.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-69
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-69
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-69-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_69.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_7.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-7
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-7
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-7-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_7.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_70.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-70
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-70
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-70-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_70.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_71.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-71
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-71
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-71-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_71.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_72.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-72
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-72
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-72-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_72.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_74.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-74
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-74
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-74-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_74.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_75.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-75
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-75
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-75-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_75.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_76.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-76
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-76
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-76-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_76.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_77.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-77
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-77
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-77-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_77.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_78.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-78
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-78
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-78-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_78.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_79.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-79
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-79
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-79-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_79.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_80.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-80
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-80
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-80-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_80.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_81.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-81
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-81
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-81-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_81.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_82.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-82
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-82
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-82-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_82.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_83.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-83
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-83
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-83-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_83.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_85.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-85
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-85
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-85-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_85.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_86.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-86
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-86
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-86-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_86.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_9.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-9
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-9
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-9-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_9.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_92.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-92
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-92
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-92-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_92.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_1/tokenizer_95.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-95
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-95
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-95-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_95.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock