Saving weights and logs of step 10
Browse files- sinhala-gpt2/config.json → config.json +0 -0
- config.py +1 -1
- sinhala-gpt2/events.out.tfevents.1625744111.t1v-n-0b4afe6c-w-0.150137.3.v2 → events.out.tfevents.1625851395.t1v-n-0b4afe6c-w-0.251724.3.v2 +2 -2
- flax_model.msgpack +3 -0
- run_clm_flax.py +0 -1
- si_tokenizer.py +1 -1
- sinhala-gpt2/flax_model.msgpack +0 -0
- sinhala-gpt2/tokenizer.json → tokenizer.json +0 -0
sinhala-gpt2/config.json → config.json
RENAMED
File without changes
|
config.py
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
from transformers import GPT2Config
|
2 |
|
3 |
-
model_dir = "
|
4 |
|
5 |
config = GPT2Config.from_pretrained("gpt2", resid_pdrop=0.0, embd_pdrop=0.0, attn_pdrop=0.0)
|
6 |
config.save_pretrained(model_dir)
|
|
|
1 |
from transformers import GPT2Config
|
2 |
|
3 |
+
model_dir = "./sihala-gpt2"
|
4 |
|
5 |
config = GPT2Config.from_pretrained("gpt2", resid_pdrop=0.0, embd_pdrop=0.0, attn_pdrop=0.0)
|
6 |
config.save_pretrained(model_dir)
|
sinhala-gpt2/events.out.tfevents.1625744111.t1v-n-0b4afe6c-w-0.150137.3.v2 → events.out.tfevents.1625851395.t1v-n-0b4afe6c-w-0.251724.3.v2
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e497ce08418104de7651e94eb9dffb54daac6d986bc927df823d3b9d8a78ae45
|
3 |
+
size 40
|
flax_model.msgpack
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:694fdb364154b16f66f557226c6f0006187a1fe305f6696be427aa8e7f634748
|
3 |
+
size 497764120
|
run_clm_flax.py
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
/home/Keshan/transformers/examples/flax/language-modeling/run_clm_flax.py
|
|
|
|
si_tokenizer.py
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
from datasets import load_dataset
|
2 |
from tokenizers import trainers, Tokenizer, normalizers, ByteLevelBPETokenizer
|
3 |
|
4 |
-
model_dir = "
|
5 |
|
6 |
# load dataset
|
7 |
dataset = load_dataset("mc4", "si", split="train")
|
|
|
1 |
from datasets import load_dataset
|
2 |
from tokenizers import trainers, Tokenizer, normalizers, ByteLevelBPETokenizer
|
3 |
|
4 |
+
model_dir = "sinhala-gpt2" # ${MODEL_DIR}
|
5 |
|
6 |
# load dataset
|
7 |
dataset = load_dataset("mc4", "si", split="train")
|
sinhala-gpt2/flax_model.msgpack
DELETED
File without changes
|
sinhala-gpt2/tokenizer.json → tokenizer.json
RENAMED
File without changes
|