tt1225 commited on
Commit
48454ef
·
1 Parent(s): c967fd0

epoch 220-235

Browse files
Files changed (3) hide show
  1. config.json +16 -17
  2. generation_config.json +1 -1
  3. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,11 +1,11 @@
1
  {
2
- "_commit_hash": "86f68fe78ff0811773d8233a005f1746702d4799",
3
- "_name_or_path": "tt1225/finetuned-trocr-small-vietnamese-nom",
4
  "architectures": [
5
  "VisionEncoderDecoderModel"
6
  ],
7
  "decoder": {
8
- "_name_or_path": "minhtoan/roberta-masked-lm-vietnamese-nom",
9
  "add_cross_attention": true,
10
  "architectures": [
11
  "RobertaForMaskedLM"
@@ -27,10 +27,9 @@
27
  "finetuning_task": null,
28
  "forced_bos_token_id": null,
29
  "forced_eos_token_id": null,
30
- "gradient_checkpointing": false,
31
  "hidden_act": "gelu",
32
  "hidden_dropout_prob": 0.1,
33
- "hidden_size": 768,
34
  "id2label": {
35
  "0": "LABEL_0",
36
  "1": "LABEL_1"
@@ -46,14 +45,14 @@
46
  "layer_norm_eps": 1e-12,
47
  "length_penalty": 1.0,
48
  "max_length": 20,
49
- "max_position_embeddings": 514,
50
  "min_length": 0,
51
  "model_type": "roberta",
52
  "no_repeat_ngram_size": 0,
53
- "num_attention_heads": 12,
54
  "num_beam_groups": 1,
55
  "num_beams": 1,
56
- "num_hidden_layers": 6,
57
  "num_return_sequences": 1,
58
  "output_attentions": false,
59
  "output_hidden_states": false,
@@ -79,7 +78,7 @@
79
  "top_p": 1.0,
80
  "torch_dtype": "float32",
81
  "torchscript": false,
82
- "transformers_version": "4.30.2",
83
  "type_vocab_size": 1,
84
  "typical_p": 1.0,
85
  "use_bfloat16": false,
@@ -89,12 +88,12 @@
89
  "decoder_start_token_id": 5,
90
  "early_stopping": true,
91
  "encoder": {
92
- "_name_or_path": "facebook/deit-tiny-distilled-patch16-224",
93
  "add_cross_attention": false,
94
  "architectures": [
95
- "DeiTForImageClassificationWithTeacher"
96
  ],
97
- "attention_probs_dropout_prob": 0.5,
98
  "bad_words_ids": null,
99
  "begin_suppress_tokens": null,
100
  "bos_token_id": null,
@@ -113,7 +112,7 @@
113
  "forced_eos_token_id": null,
114
  "hidden_act": "gelu",
115
  "hidden_dropout_prob": 0.0,
116
- "hidden_size": 192,
117
  "id2label": {
118
  "0": "tench, Tinca tinca",
119
  "1": "goldfish, Carassius auratus",
@@ -1118,7 +1117,7 @@
1118
  },
1119
  "image_size": 224,
1120
  "initializer_range": 0.02,
1121
- "intermediate_size": 768,
1122
  "is_decoder": false,
1123
  "is_encoder_decoder": false,
1124
  "label2id": {
@@ -2126,9 +2125,9 @@
2126
  "length_penalty": 1.0,
2127
  "max_length": 20,
2128
  "min_length": 0,
2129
- "model_type": "deit",
2130
  "no_repeat_ngram_size": 0,
2131
- "num_attention_heads": 3,
2132
  "num_beam_groups": 1,
2133
  "num_beams": 1,
2134
  "num_channels": 3,
@@ -2159,7 +2158,7 @@
2159
  "top_p": 1.0,
2160
  "torch_dtype": null,
2161
  "torchscript": false,
2162
- "transformers_version": "4.30.2",
2163
  "typical_p": 1.0,
2164
  "use_bfloat16": false
2165
  },
 
1
  {
2
+ "_commit_hash": "620d114582d85d027705435e967dca3667a79825",
3
+ "_name_or_path": "tt1225/finetuned-trocr-tiny-vietnamese-nom",
4
  "architectures": [
5
  "VisionEncoderDecoderModel"
6
  ],
7
  "decoder": {
8
+ "_name_or_path": "nxquang-al/roberta-small-masked-lm-vietnamese-nom",
9
  "add_cross_attention": true,
10
  "architectures": [
11
  "RobertaForMaskedLM"
 
27
  "finetuning_task": null,
28
  "forced_bos_token_id": null,
29
  "forced_eos_token_id": null,
 
30
  "hidden_act": "gelu",
31
  "hidden_dropout_prob": 0.1,
32
+ "hidden_size": 256,
33
  "id2label": {
34
  "0": "LABEL_0",
35
  "1": "LABEL_1"
 
45
  "layer_norm_eps": 1e-12,
46
  "length_penalty": 1.0,
47
  "max_length": 20,
48
+ "max_position_embeddings": 512,
49
  "min_length": 0,
50
  "model_type": "roberta",
51
  "no_repeat_ngram_size": 0,
52
+ "num_attention_heads": 4,
53
  "num_beam_groups": 1,
54
  "num_beams": 1,
55
+ "num_hidden_layers": 12,
56
  "num_return_sequences": 1,
57
  "output_attentions": false,
58
  "output_hidden_states": false,
 
78
  "top_p": 1.0,
79
  "torch_dtype": "float32",
80
  "torchscript": false,
81
+ "transformers_version": "4.30.1",
82
  "type_vocab_size": 1,
83
  "typical_p": 1.0,
84
  "use_bfloat16": false,
 
88
  "decoder_start_token_id": 5,
89
  "early_stopping": true,
90
  "encoder": {
91
+ "_name_or_path": "facebook/deit-small-patch16-224",
92
  "add_cross_attention": false,
93
  "architectures": [
94
+ "ViTForImageClassification"
95
  ],
96
+ "attention_probs_dropout_prob": 0.1,
97
  "bad_words_ids": null,
98
  "begin_suppress_tokens": null,
99
  "bos_token_id": null,
 
112
  "forced_eos_token_id": null,
113
  "hidden_act": "gelu",
114
  "hidden_dropout_prob": 0.0,
115
+ "hidden_size": 384,
116
  "id2label": {
117
  "0": "tench, Tinca tinca",
118
  "1": "goldfish, Carassius auratus",
 
1117
  },
1118
  "image_size": 224,
1119
  "initializer_range": 0.02,
1120
+ "intermediate_size": 1536,
1121
  "is_decoder": false,
1122
  "is_encoder_decoder": false,
1123
  "label2id": {
 
2125
  "length_penalty": 1.0,
2126
  "max_length": 20,
2127
  "min_length": 0,
2128
+ "model_type": "vit",
2129
  "no_repeat_ngram_size": 0,
2130
+ "num_attention_heads": 6,
2131
  "num_beam_groups": 1,
2132
  "num_beams": 1,
2133
  "num_channels": 3,
 
2158
  "top_p": 1.0,
2159
  "torch_dtype": null,
2160
  "torchscript": false,
2161
+ "transformers_version": "4.30.1",
2162
  "typical_p": 1.0,
2163
  "use_bfloat16": false
2164
  },
generation_config.json CHANGED
@@ -9,5 +9,5 @@
9
  "no_repeat_ngram_size": 3,
10
  "num_beams": 4,
11
  "pad_token_id": 6198,
12
- "transformers_version": "4.30.2"
13
  }
 
9
  "no_repeat_ngram_size": 3,
10
  "num_beams": 4,
11
  "pad_token_id": 6198,
12
+ "transformers_version": "4.30.1"
13
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb1591d78d4b0c5b0540b19f6880898f1f2c4cee765ff9818d53b8f86499db37
3
- size 349128597
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a95b2610553c7cffc1356ccd94814585f16dde224ea8b00754bbd2d58dd665df
3
+ size 221492737