applied-ai-018
commited on
Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- dravid/tok_dravid/tok_jobs/tokenizer_0.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_100.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_102.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_103.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_104.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_105.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_106.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_107.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_108.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_109.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_11.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_112.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_114.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_115.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_116.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_117.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_120.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_121.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_123.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_125.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_126.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_14.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_16.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_2.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_21.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_22.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_24.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_26.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_27.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_28.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_29.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_3.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_30.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_32.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_34.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_35.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_36.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_37.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_40.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_41.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_42.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_44.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_45.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_46.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_47.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_48.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_5.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_51.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_53.yaml +104 -0
- dravid/tok_dravid/tok_jobs/tokenizer_54.yaml +104 -0
dravid/tok_dravid/tok_jobs/tokenizer_0.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-0
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-0
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-0-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_0.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_100.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-100
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-100
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-100-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_100.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_102.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-102
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-102
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-102-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_102.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_103.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-103
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-103
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-103-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_103.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_104.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-104
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-104
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-104-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_104.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_105.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-105
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-105
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-105-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_105.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_106.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-106
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-106
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-106-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_106.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_107.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-107
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-107
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-107-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_107.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_108.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-108
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-108
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-108-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_108.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_109.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-109
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-109
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-109-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_109.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_11.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-11
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-11
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-11-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_11.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_112.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-112
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-112
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-112-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_112.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_114.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-114
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-114
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-114-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_114.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_115.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-115
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-115
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-115-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_115.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_116.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-116
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-116
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-116-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_116.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_117.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-117
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-117
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-117-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_117.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_120.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-120
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-120
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-120-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_120.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_121.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-121
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-121
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-121-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_121.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_123.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-123
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-123
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-123-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_123.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_125.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-125
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-125
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-125-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_125.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_126.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-126
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-126
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-126-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_126.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_14.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-14
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-14
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-14-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_14.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_16.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-16
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-16
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-16-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_16.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_2.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-2
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-2
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-2-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_2.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_21.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-21
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-21
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-21-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_21.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_22.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-22
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-22
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-22-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_22.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_24.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-24
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-24
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-24-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_24.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_26.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-26
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-26
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-26-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_26.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_27.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-27
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-27
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-27-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_27.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_28.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-28
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-28
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-28-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_28.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_29.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-29
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-29
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-29-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_29.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_3.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-3
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-3
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-3-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_3.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_30.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-30
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-30
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-30-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_30.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_32.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-32
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-32
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-32-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_32.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_34.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-34
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-34
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-34-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_34.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_35.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-35
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-35
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-35-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_35.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_36.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-36
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-36
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-36-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_36.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_37.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-37
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-37
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-37-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_37.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_40.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-40
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-40
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-40-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_40.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_41.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-41
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-41
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-41-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_41.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_42.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-42
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-42
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-42-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_42.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_44.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-44
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-44
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-44-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_44.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_45.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-45
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-45
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-45-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_45.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_46.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-46
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-46
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-46-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_46.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_47.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-47
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-47
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-47-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_47.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_48.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-48
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-48
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-48-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_48.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_5.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-5
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-5
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-5-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_5.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_51.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-51
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-51
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-51-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_51.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_53.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-53
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-53
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-53-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_53.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|
dravid/tok_dravid/tok_jobs/tokenizer_54.yaml
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
apiVersion: kubeflow.org/v2beta1
|
2 |
+
kind: MPIJob
|
3 |
+
metadata:
|
4 |
+
name: peacock-tokenizer-dravid-54
|
5 |
+
namespace: peacock
|
6 |
+
labels:
|
7 |
+
app: peacock-tokenizer-dravid-54
|
8 |
+
spec:
|
9 |
+
slotsPerWorker: 8
|
10 |
+
runPolicy:
|
11 |
+
backoffLimit: 1
|
12 |
+
cleanPodPolicy: Running
|
13 |
+
mpiReplicaSpecs:
|
14 |
+
Launcher:
|
15 |
+
replicas: 1
|
16 |
+
template:
|
17 |
+
spec:
|
18 |
+
hostIPC: true
|
19 |
+
volumes:
|
20 |
+
- name: work-dir
|
21 |
+
persistentVolumeClaim:
|
22 |
+
claimName: peacock-fs-pvc
|
23 |
+
containers:
|
24 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
25 |
+
name: peacock-tokenizer-dravid-54-launcher
|
26 |
+
imagePullPolicy: IfNotPresent
|
27 |
+
volumeMounts:
|
28 |
+
- name: work-dir
|
29 |
+
mountPath: /mnt/weka/peacock
|
30 |
+
command: ["/bin/bash", "-c"]
|
31 |
+
args:
|
32 |
+
- >-
|
33 |
+
|
34 |
+
export SYNAPSE_VERSION="1.15.1";
|
35 |
+
export WORKER_DIR="/mnt/weka/peacock/experiments/llama";
|
36 |
+
|
37 |
+
export MEGATRON_SETUP_CMD="$WORKER_DIR/launch/setup.sh"
|
38 |
+
export TOKENIZER_CMD=/mnt/weka/peacock/idc/datasets/dravid//tok_dravid/tok_files/tokenizer_54.sh
|
39 |
+
HOSTSFILE=$OMPI_MCA_orte_default_hostfile;
|
40 |
+
echo "HOSTSFILE=$HOSTSFILE";
|
41 |
+
MASTER_ADDR="$(head -n 1 $HOSTSFILE | sed -n s/[[:space:]]slots.*//p)";
|
42 |
+
NUM_NODES=$(wc -l < $HOSTSFILE);
|
43 |
+
CARDS_PER_NODE=8;
|
44 |
+
N_CARDS=$((NUM_NODES*CARDS_PER_NODE));
|
45 |
+
echo "MPI_ROOT=$MPI_ROOT";
|
46 |
+
echo "N_CARDS=$N_CARDS";
|
47 |
+
echo "MASTER_ADDR=$MASTER_ADDR";
|
48 |
+
sleep 20;
|
49 |
+
|
50 |
+
|
51 |
+
mpirun -np $N_CARDS -npernode 8 \
|
52 |
+
--tag-output \
|
53 |
+
--allow-run-as-root \
|
54 |
+
--prefix $MPI_ROOT \
|
55 |
+
-x WORKER_DIR=$WORKER_DIR \
|
56 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $MEGATRON_SETUP_CMD;
|
57 |
+
|
58 |
+
mpirun -np $N_CARDS -npernode 8 \
|
59 |
+
--tag-output \
|
60 |
+
--allow-run-as-root \
|
61 |
+
--prefix $MPI_ROOT \
|
62 |
+
-x WORKER_DIR=$WORKER_DIR \
|
63 |
+
-x SYNAPSE_VERSION=$SYNAPSE_VERSION $TOKENIZER_CMD;
|
64 |
+
|
65 |
+
|
66 |
+
Worker:
|
67 |
+
replicas: 1
|
68 |
+
template:
|
69 |
+
spec:
|
70 |
+
volumes:
|
71 |
+
- name: work-dir
|
72 |
+
persistentVolumeClaim:
|
73 |
+
claimName: peacock-fs-pvc
|
74 |
+
tolerations:
|
75 |
+
- key: "habana.ai/gaudi"
|
76 |
+
operator: "Exists"
|
77 |
+
effect: "NoSchedule"
|
78 |
+
- key: "k8s/namespace"
|
79 |
+
operator: "Equal"
|
80 |
+
value: "peacock"
|
81 |
+
effect: "NoSchedule"
|
82 |
+
hostIPC: true
|
83 |
+
containers:
|
84 |
+
- image: vault.habana.ai/gaudi-docker/1.15.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.0:latest
|
85 |
+
name: peacock-llama-worker
|
86 |
+
imagePullPolicy: IfNotPresent
|
87 |
+
securityContext:
|
88 |
+
capabilities:
|
89 |
+
add:
|
90 |
+
- SYSLOG
|
91 |
+
resources:
|
92 |
+
limits:
|
93 |
+
habana.ai/gaudi: 8
|
94 |
+
hugepages-2Mi: 300Gi
|
95 |
+
memory: 700Gi
|
96 |
+
cpu: 150
|
97 |
+
requests:
|
98 |
+
habana.ai/gaudi: 8
|
99 |
+
hugepages-2Mi: 300Gi
|
100 |
+
memory: 700Gi
|
101 |
+
cpu: 150
|
102 |
+
volumeMounts:
|
103 |
+
- name: work-dir
|
104 |
+
mountPath: /mnt/weka/peacock
|