KoichiYasuoka commited on
Commit
dcdab31
·
1 Parent(s): 84bb563

model improved

Browse files
Files changed (5) hide show
  1. README.md +2 -2
  2. maker.py +1 -1
  3. pytorch_model.bin +1 -1
  4. tokenizer.json +0 -0
  5. tokenizer_config.json +4 -1
README.md CHANGED
@@ -6,7 +6,7 @@ tags:
6
  - "pos"
7
  - "dependency-parsing"
8
  - "modernbert"
9
- base_model: KoichiYasuoka/modernbert-base-japanese-aozora
10
  datasets:
11
  - "universal_dependencies"
12
  license: "apache-2.0"
@@ -19,7 +19,7 @@ widget:
19
 
20
  ## Model Description
21
 
22
- This is a ModernBERT model pretrained for POS-tagging and dependency-parsing (using `goeswith` for subwords), derived from [modernbert-base-japanese-aozora](https://huggingface.co/KoichiYasuoka/modernbert-base-japanese-aozora) and [UD_Japanese-GSDLUW](https://github.com/UniversalDependencies/UD_Japanese-GSDLUW).
23
 
24
  ## How to Use
25
 
 
6
  - "pos"
7
  - "dependency-parsing"
8
  - "modernbert"
9
+ base_model: KoichiYasuoka/modernbert-base-japanese-aozora-upos
10
  datasets:
11
  - "universal_dependencies"
12
  license: "apache-2.0"
 
19
 
20
  ## Model Description
21
 
22
+ This is a ModernBERT model pretrained for POS-tagging and dependency-parsing (using `goeswith` for subwords), derived from [modernbert-base-japanese-aozora-upos](https://huggingface.co/KoichiYasuoka/modernbert-base-japanese-aozora-upos) and [UD_Japanese-GSDLUW](https://github.com/UniversalDependencies/UD_Japanese-GSDLUW).
23
 
24
  ## How to Use
25
 
maker.py CHANGED
@@ -1,5 +1,5 @@
1
  #! /usr/bin/python3
2
- src="KoichiYasuoka/modernbert-base-japanese-aozora"
3
  tgt="KoichiYasuoka/modernbert-base-japanese-aozora-ud-goeswith"
4
  url="https://github.com/UniversalDependencies/UD_Japanese-GSDLUW"
5
  import os
 
1
  #! /usr/bin/python3
2
+ src="KoichiYasuoka/modernbert-base-japanese-aozora-upos"
3
  tgt="KoichiYasuoka/modernbert-base-japanese-aozora-ud-goeswith"
4
  url="https://github.com/UniversalDependencies/UD_Japanese-GSDLUW"
5
  import os
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59986e96524b43b5cf379da759af0304a39d83e53a634d78efed2355ea3b6f96
3
  size 643697266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24f7cf26d7d271e1e48ce615240d1c2edfea7a0d4efb9bedee6f02e00f6d63ae
3
  size 643697266
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -49,7 +49,10 @@
49
  "extra_special_tokens": {},
50
  "keep_accents": true,
51
  "mask_token": "[MASK]",
52
- "model_input_names": ["input_ids", "attention_mask"],
 
 
 
53
  "model_max_length": 1000000000000000019884624838656,
54
  "pad_token": "[PAD]",
55
  "sep_token": "[SEP]",
 
49
  "extra_special_tokens": {},
50
  "keep_accents": true,
51
  "mask_token": "[MASK]",
52
+ "model_input_names": [
53
+ "input_ids",
54
+ "attention_mask"
55
+ ],
56
  "model_max_length": 1000000000000000019884624838656,
57
  "pad_token": "[PAD]",
58
  "sep_token": "[SEP]",