epoch 220-235
Browse files- config.json +16 -17
- generation_config.json +1 -1
- pytorch_model.bin +2 -2
config.json
CHANGED
@@ -1,11 +1,11 @@
|
|
1 |
{
|
2 |
-
"_commit_hash": "
|
3 |
-
"_name_or_path": "tt1225/finetuned-trocr-
|
4 |
"architectures": [
|
5 |
"VisionEncoderDecoderModel"
|
6 |
],
|
7 |
"decoder": {
|
8 |
-
"_name_or_path": "
|
9 |
"add_cross_attention": true,
|
10 |
"architectures": [
|
11 |
"RobertaForMaskedLM"
|
@@ -27,10 +27,9 @@
|
|
27 |
"finetuning_task": null,
|
28 |
"forced_bos_token_id": null,
|
29 |
"forced_eos_token_id": null,
|
30 |
-
"gradient_checkpointing": false,
|
31 |
"hidden_act": "gelu",
|
32 |
"hidden_dropout_prob": 0.1,
|
33 |
-
"hidden_size":
|
34 |
"id2label": {
|
35 |
"0": "LABEL_0",
|
36 |
"1": "LABEL_1"
|
@@ -46,14 +45,14 @@
|
|
46 |
"layer_norm_eps": 1e-12,
|
47 |
"length_penalty": 1.0,
|
48 |
"max_length": 20,
|
49 |
-
"max_position_embeddings":
|
50 |
"min_length": 0,
|
51 |
"model_type": "roberta",
|
52 |
"no_repeat_ngram_size": 0,
|
53 |
-
"num_attention_heads":
|
54 |
"num_beam_groups": 1,
|
55 |
"num_beams": 1,
|
56 |
-
"num_hidden_layers":
|
57 |
"num_return_sequences": 1,
|
58 |
"output_attentions": false,
|
59 |
"output_hidden_states": false,
|
@@ -79,7 +78,7 @@
|
|
79 |
"top_p": 1.0,
|
80 |
"torch_dtype": "float32",
|
81 |
"torchscript": false,
|
82 |
-
"transformers_version": "4.30.
|
83 |
"type_vocab_size": 1,
|
84 |
"typical_p": 1.0,
|
85 |
"use_bfloat16": false,
|
@@ -89,12 +88,12 @@
|
|
89 |
"decoder_start_token_id": 5,
|
90 |
"early_stopping": true,
|
91 |
"encoder": {
|
92 |
-
"_name_or_path": "facebook/deit-
|
93 |
"add_cross_attention": false,
|
94 |
"architectures": [
|
95 |
-
"
|
96 |
],
|
97 |
-
"attention_probs_dropout_prob": 0.
|
98 |
"bad_words_ids": null,
|
99 |
"begin_suppress_tokens": null,
|
100 |
"bos_token_id": null,
|
@@ -113,7 +112,7 @@
|
|
113 |
"forced_eos_token_id": null,
|
114 |
"hidden_act": "gelu",
|
115 |
"hidden_dropout_prob": 0.0,
|
116 |
-
"hidden_size":
|
117 |
"id2label": {
|
118 |
"0": "tench, Tinca tinca",
|
119 |
"1": "goldfish, Carassius auratus",
|
@@ -1118,7 +1117,7 @@
|
|
1118 |
},
|
1119 |
"image_size": 224,
|
1120 |
"initializer_range": 0.02,
|
1121 |
-
"intermediate_size":
|
1122 |
"is_decoder": false,
|
1123 |
"is_encoder_decoder": false,
|
1124 |
"label2id": {
|
@@ -2126,9 +2125,9 @@
|
|
2126 |
"length_penalty": 1.0,
|
2127 |
"max_length": 20,
|
2128 |
"min_length": 0,
|
2129 |
-
"model_type": "
|
2130 |
"no_repeat_ngram_size": 0,
|
2131 |
-
"num_attention_heads":
|
2132 |
"num_beam_groups": 1,
|
2133 |
"num_beams": 1,
|
2134 |
"num_channels": 3,
|
@@ -2159,7 +2158,7 @@
|
|
2159 |
"top_p": 1.0,
|
2160 |
"torch_dtype": null,
|
2161 |
"torchscript": false,
|
2162 |
-
"transformers_version": "4.30.
|
2163 |
"typical_p": 1.0,
|
2164 |
"use_bfloat16": false
|
2165 |
},
|
|
|
1 |
{
|
2 |
+
"_commit_hash": "620d114582d85d027705435e967dca3667a79825",
|
3 |
+
"_name_or_path": "tt1225/finetuned-trocr-tiny-vietnamese-nom",
|
4 |
"architectures": [
|
5 |
"VisionEncoderDecoderModel"
|
6 |
],
|
7 |
"decoder": {
|
8 |
+
"_name_or_path": "nxquang-al/roberta-small-masked-lm-vietnamese-nom",
|
9 |
"add_cross_attention": true,
|
10 |
"architectures": [
|
11 |
"RobertaForMaskedLM"
|
|
|
27 |
"finetuning_task": null,
|
28 |
"forced_bos_token_id": null,
|
29 |
"forced_eos_token_id": null,
|
|
|
30 |
"hidden_act": "gelu",
|
31 |
"hidden_dropout_prob": 0.1,
|
32 |
+
"hidden_size": 256,
|
33 |
"id2label": {
|
34 |
"0": "LABEL_0",
|
35 |
"1": "LABEL_1"
|
|
|
45 |
"layer_norm_eps": 1e-12,
|
46 |
"length_penalty": 1.0,
|
47 |
"max_length": 20,
|
48 |
+
"max_position_embeddings": 512,
|
49 |
"min_length": 0,
|
50 |
"model_type": "roberta",
|
51 |
"no_repeat_ngram_size": 0,
|
52 |
+
"num_attention_heads": 4,
|
53 |
"num_beam_groups": 1,
|
54 |
"num_beams": 1,
|
55 |
+
"num_hidden_layers": 12,
|
56 |
"num_return_sequences": 1,
|
57 |
"output_attentions": false,
|
58 |
"output_hidden_states": false,
|
|
|
78 |
"top_p": 1.0,
|
79 |
"torch_dtype": "float32",
|
80 |
"torchscript": false,
|
81 |
+
"transformers_version": "4.30.1",
|
82 |
"type_vocab_size": 1,
|
83 |
"typical_p": 1.0,
|
84 |
"use_bfloat16": false,
|
|
|
88 |
"decoder_start_token_id": 5,
|
89 |
"early_stopping": true,
|
90 |
"encoder": {
|
91 |
+
"_name_or_path": "facebook/deit-small-patch16-224",
|
92 |
"add_cross_attention": false,
|
93 |
"architectures": [
|
94 |
+
"ViTForImageClassification"
|
95 |
],
|
96 |
+
"attention_probs_dropout_prob": 0.1,
|
97 |
"bad_words_ids": null,
|
98 |
"begin_suppress_tokens": null,
|
99 |
"bos_token_id": null,
|
|
|
112 |
"forced_eos_token_id": null,
|
113 |
"hidden_act": "gelu",
|
114 |
"hidden_dropout_prob": 0.0,
|
115 |
+
"hidden_size": 384,
|
116 |
"id2label": {
|
117 |
"0": "tench, Tinca tinca",
|
118 |
"1": "goldfish, Carassius auratus",
|
|
|
1117 |
},
|
1118 |
"image_size": 224,
|
1119 |
"initializer_range": 0.02,
|
1120 |
+
"intermediate_size": 1536,
|
1121 |
"is_decoder": false,
|
1122 |
"is_encoder_decoder": false,
|
1123 |
"label2id": {
|
|
|
2125 |
"length_penalty": 1.0,
|
2126 |
"max_length": 20,
|
2127 |
"min_length": 0,
|
2128 |
+
"model_type": "vit",
|
2129 |
"no_repeat_ngram_size": 0,
|
2130 |
+
"num_attention_heads": 6,
|
2131 |
"num_beam_groups": 1,
|
2132 |
"num_beams": 1,
|
2133 |
"num_channels": 3,
|
|
|
2158 |
"top_p": 1.0,
|
2159 |
"torch_dtype": null,
|
2160 |
"torchscript": false,
|
2161 |
+
"transformers_version": "4.30.1",
|
2162 |
"typical_p": 1.0,
|
2163 |
"use_bfloat16": false
|
2164 |
},
|
generation_config.json
CHANGED
@@ -9,5 +9,5 @@
|
|
9 |
"no_repeat_ngram_size": 3,
|
10 |
"num_beams": 4,
|
11 |
"pad_token_id": 6198,
|
12 |
-
"transformers_version": "4.30.
|
13 |
}
|
|
|
9 |
"no_repeat_ngram_size": 3,
|
10 |
"num_beams": 4,
|
11 |
"pad_token_id": 6198,
|
12 |
+
"transformers_version": "4.30.1"
|
13 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a95b2610553c7cffc1356ccd94814585f16dde224ea8b00754bbd2d58dd665df
|
3 |
+
size 221492737
|