applied-ai-018 commited on
Commit
247adee
·
verified ·
1 Parent(s): 3cb4f9c

Add files using upload-large-folder tool

Browse files
Files changed (41) hide show
  1. .gitattributes +5 -0
  2. dravid/tok_dravid/shuffled_files.txt +3 -0
  3. dravid/tok_dravid/unshuffled.txt +3 -0
  4. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_200.yaml +104 -0
  5. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_202.yaml +104 -0
  6. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_203.yaml +104 -0
  7. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_207.yaml +104 -0
  8. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_208.yaml +104 -0
  9. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_209.yaml +104 -0
  10. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_210.yaml +104 -0
  11. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_219.yaml +104 -0
  12. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_225.yaml +104 -0
  13. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_226.yaml +104 -0
  14. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_228.yaml +104 -0
  15. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_230.yaml +104 -0
  16. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_234.yaml +104 -0
  17. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_237.yaml +104 -0
  18. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_238.yaml +104 -0
  19. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_239.yaml +104 -0
  20. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_240.yaml +104 -0
  21. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_241.yaml +104 -0
  22. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_243.yaml +104 -0
  23. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_253.yaml +104 -0
  24. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_255.yaml +104 -0
  25. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_256.yaml +104 -0
  26. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_261.yaml +104 -0
  27. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_263.yaml +104 -0
  28. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_268.yaml +104 -0
  29. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_270.yaml +104 -0
  30. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_276.yaml +104 -0
  31. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_277.yaml +104 -0
  32. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_279.yaml +104 -0
  33. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_281.yaml +104 -0
  34. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_284.yaml +104 -0
  35. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_287.yaml +104 -0
  36. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_288.yaml +104 -0
  37. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_294.yaml +104 -0
  38. hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_296.yaml +104 -0
  39. sangraha/tok_sangraha/split/finalaa_splitaa +3 -0
  40. sangraha/tok_sangraha/split/finalae_splitab +3 -0
  41. sangraha/tok_sangraha/split/finalan_splitaa +3 -0
.gitattributes CHANGED
@@ -91,3 +91,8 @@ sangraha/split_final.json/part-16.json filter=lfs diff=lfs merge=lfs -text
91
  sangraha/tok_sangraha/split/finalaf_splitaa filter=lfs diff=lfs merge=lfs -text
92
  sangraha/tok_sangraha/split/finalad_splitab filter=lfs diff=lfs merge=lfs -text
93
  sangraha/tok_sangraha/split/finalac_splitab filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
91
  sangraha/tok_sangraha/split/finalaf_splitaa filter=lfs diff=lfs merge=lfs -text
92
  sangraha/tok_sangraha/split/finalad_splitab filter=lfs diff=lfs merge=lfs -text
93
  sangraha/tok_sangraha/split/finalac_splitab filter=lfs diff=lfs merge=lfs -text
94
+ sangraha/tok_sangraha/split/finalae_splitab filter=lfs diff=lfs merge=lfs -text
95
+ sangraha/tok_sangraha/split/finalan_splitaa filter=lfs diff=lfs merge=lfs -text
96
+ sangraha/tok_sangraha/split/finalaa_splitaa filter=lfs diff=lfs merge=lfs -text
97
+ dravid/tok_dravid/shuffled_files.txt filter=lfs diff=lfs merge=lfs -text
98
+ dravid/tok_dravid/unshuffled.txt filter=lfs diff=lfs merge=lfs -text
dravid/tok_dravid/shuffled_files.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d896ca5949bf902e9beb35eb7ff926dca1ad17a0598d9845886780a7e8364afc
3
+ size 17592322
dravid/tok_dravid/unshuffled.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf54fdf231c5d4bd1771cdf761dc825a8902d0d3da45d0f52d2a5f4fc4cc18ff
3
+ size 17592322
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_200.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-200
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-200
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-200-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_200.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_202.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-202
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-202
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-202-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_202.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_203.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-203
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-203
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-203-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_203.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_207.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-207
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-207
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-207-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_207.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_208.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-208
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-208
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-208-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_208.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_209.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-209
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-209
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-209-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_209.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_210.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-210
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-210
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-210-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_210.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_219.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-219
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-219
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-219-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_219.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_225.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-225
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-225
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-225-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_225.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_226.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-226
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-226
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-226-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_226.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_228.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-228
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-228
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-228-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_228.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_230.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-230
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-230
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-230-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_230.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_234.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-234
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-234
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-234-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_234.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_237.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-237
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-237
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-237-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_237.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_238.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-238
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-238
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-238-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_238.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_239.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-239
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-239
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-239-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_239.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_240.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-240
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-240
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-240-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_240.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_241.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-241
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-241
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-241-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_241.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_243.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-243
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-243
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-243-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_243.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_253.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-253
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-253
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-253-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_253.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_255.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-255
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-255
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-255-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_255.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_256.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-256
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-256
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-256-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_256.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_261.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-261
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-261
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-261-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_261.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_263.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-263
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-263
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-263-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_263.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_268.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-268
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-268
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-268-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_268.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_270.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-270
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-270
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-270-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_270.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_276.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-276
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-276
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-276-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_276.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_277.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-277
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-277
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-277-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_277.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_279.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-279
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-279
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-279-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_279.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_281.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-281
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-281
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-281-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_281.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_284.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-284
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-284
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-284-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_284.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_287.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-287
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-287
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-287-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_287.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_288.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-288
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-288
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-288-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_288.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_294.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-294
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-294
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-294-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_294.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
hn_eng_bn/tok_en-bn-hn/tok_jobs_3/tokenizer_296.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ apiVersion: kubeflow.org/v2beta1
2
+ kind: MPIJob
3
+ metadata:
4
+ name: peacock-tokenizer-en-bn-hn-296
5
+ namespace: peacock
6
+ labels:
7
+ app: peacock-tokenizer-en-bn-hn-296
8
+ spec:
9
+ slotsPerWorker: 8
10
+ runPolicy:
11
+ backoffLimit: 1
12
+ cleanPodPolicy: Running
13
+ mpiReplicaSpecs:
14
+ Launcher:
15
+ replicas: 1
16
+ template:
17
+ spec:
18
+ hostIPC: true
19
+ volumes:
20
+ - name: work-dir
21
+ persistentVolumeClaim:
22
+ claimName: peacock-fs-pvc
23
+ containers:
24
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
25
+ name: peacock-tokenizer-en-bn-hn-296-launcher
26
+ imagePullPolicy: IfNotPresent
27
+ volumeMounts:
28
+ - name: work-dir
29
+ mountPath: /mnt/weka/peacock
30
+ command: ["/bin/bash", "-c"]
31
+ args:
32
+ - >-
33
+
34
+ export SYNAPSE_VERSION="1.15.1";
35
+ export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
36
+
37
+ export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
38
+ export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/hn_eng_bn/tok_en-bn-hn/tok_files/tokenizer_296.sh
39
+ HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
40
+ echo "HOSTSFILE=$HOSTSFILE";
41
+ MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
42
+ NUM_NODES=$(wc -l < $HOSTSFILE);
43
+ CARDS_PER_NODE=8;
44
+ N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
45
+ echo "MPI_ROOT=$MPI_ROOT";
46
+ echo "N_CARDS=$N_CARDS";
47
+ echo "MASTER_ADDR=$MASTER_ADDR";
48
+ sleep 20;
49
+
50
+
51
+ mpirun -np $N_CARDS -npernode 8 \
52
+ --tag-output \
53
+ --allow-run-as-root \
54
+ --prefix $MPI_ROOT \
55
+ -x WORKER_DIR=$WORKER_DIR \
56
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
57
+
58
+ mpirun -np $N_CARDS -npernode 8 \
59
+ --tag-output \
60
+ --allow-run-as-root \
61
+ --prefix $MPI_ROOT \
62
+ -x WORKER_DIR=$WORKER_DIR \
63
+ -x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
64
+
65
+
66
+ Worker:
67
+ replicas: 1
68
+ template:
69
+ spec:
70
+ volumes:
71
+ - name: work-dir
72
+ persistentVolumeClaim:
73
+ claimName: peacock-fs-pvc
74
+ tolerations:
75
+ - key: "habana.ai/gaudi"
76
+ operator: "Exists"
77
+ effect: "NoSchedule"
78
+ - key: "k8s/namespace"
79
+ operator: "Equal"
80
+ value: "peacock"
81
+ effect: "NoSchedule"
82
+ hostIPC: true
83
+ containers:
84
+ - image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
85
+ name: peacock-llama-worker
86
+ imagePullPolicy: IfNotPresent
87
+ securityContext:
88
+ capabilities:
89
+ add:
90
+ - SYSLOG
91
+ resources:
92
+ limits:
93
+ habana.ai/gaudi: 8
94
+ hugepages-2Mi: 300Gi
95
+ memory: 700Gi
96
+ cpu: 150
97
+ requests:
98
+ habana.ai/gaudi: 8
99
+ hugepages-2Mi: 300Gi
100
+ memory: 700Gi
101
+ cpu: 150
102
+ volumeMounts:
103
+ - name: work-dir
104
+ mountPath: /mnt/weka/peacock
sangraha/tok_sangraha/split/finalaa_splitaa ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e28effa33a7f4be4249ff702e3f9a2080573a807bfcb8c1191aa022b72137b03
3
+ size 12238366095
sangraha/tok_sangraha/split/finalae_splitab ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab08a229a5d342642d47a6b1073e5b2aa4a5c7c486b0575cfd900bd17c7075dc
3
+ size 12254666619
sangraha/tok_sangraha/split/finalan_splitaa ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e087811a706b1f96796994c9c04ecb1ba4739deb02911afc3bcec261979a5ff2
3
+ size 14497495472