chizhik commited on
Commit
af19d48
·
1 Parent(s): da07a7f

add model weights

Browse files
config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/mnt/beegfs/mc000051/CERPLES/Models/roberta-base-biomedical-clinical-es",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "LABEL_0",
16
+ "1": "LABEL_1",
17
+ "2": "LABEL_2",
18
+ "3": "LABEL_3",
19
+ "4": "LABEL_4",
20
+ "5": "LABEL_5",
21
+ "6": "LABEL_6",
22
+ "7": "LABEL_7",
23
+ "8": "LABEL_8",
24
+ "9": "LABEL_9",
25
+ "10": "LABEL_10",
26
+ "11": "LABEL_11",
27
+ "12": "LABEL_12",
28
+ "13": "LABEL_13",
29
+ "14": "LABEL_14",
30
+ "15": "LABEL_15"
31
+ },
32
+ "initializer_range": 0.02,
33
+ "intermediate_size": 3072,
34
+ "label2id": {
35
+ "LABEL_0": 0,
36
+ "LABEL_1": 1,
37
+ "LABEL_10": 10,
38
+ "LABEL_11": 11,
39
+ "LABEL_12": 12,
40
+ "LABEL_13": 13,
41
+ "LABEL_14": 14,
42
+ "LABEL_15": 15,
43
+ "LABEL_2": 2,
44
+ "LABEL_3": 3,
45
+ "LABEL_4": 4,
46
+ "LABEL_5": 5,
47
+ "LABEL_6": 6,
48
+ "LABEL_7": 7,
49
+ "LABEL_8": 8,
50
+ "LABEL_9": 9
51
+ },
52
+ "layer_norm_eps": 1e-05,
53
+ "max_position_embeddings": 514,
54
+ "model_type": "roberta",
55
+ "num_attention_heads": 12,
56
+ "num_hidden_layers": 12,
57
+ "pad_token_id": 1,
58
+ "position_embedding_type": "absolute",
59
+ "problem_type": "multi_label_classification",
60
+ "torch_dtype": "float32",
61
+ "transformers_version": "4.19.2",
62
+ "type_vocab_size": 1,
63
+ "use_cache": true,
64
+ "vocab_size": 52000
65
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:319bd8b0b3fc3178d2d78724f5baa91dd85918dbcf71a4c47feafcde2ec91c89
3
+ size 1008039837
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09eaddcf87a41588d71397076bdad7e758ea8e4e74269567a6e22ed43d76b437
3
+ size 504033325
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dcd8bd2facbcb38c978d6b2b10cbacd8e9f54834a513467b1196a471554682c
3
+ size 14567
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:064f8a5f5dbc314d6628cb6104463ee1d75e86f55b483f7a798a2a982ed80516
3
+ size 623
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true}}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": true, "trim_offsets": true, "max_len": 512, "special_tokens_map_file": null, "name_or_path": "/mnt/beegfs/mc000051/CERPLES/Models/roberta-base-biomedical-clinical-es", "tokenizer_class": "RobertaTokenizer"}
trainer_state.json ADDED
@@ -0,0 +1,156 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7442957192288286,
3
+ "best_model_checkpoint": "./CARES/checkpoints/roberta-/checkpoint-3500",
4
+ "epoch": 49.29577464788732,
5
+ "global_step": 3500,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 7.04,
12
+ "learning_rate": 4.647887323943662e-05,
13
+ "loss": 0.1802,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 7.04,
18
+ "eval_loss": 0.1463283747434616,
19
+ "eval_macro_f1": 0.38060566625566594,
20
+ "eval_macro_precision": 0.5037224636955915,
21
+ "eval_macro_recall": 0.36676192196038643,
22
+ "eval_micro_f1": 0.7358149620527648,
23
+ "eval_micro_precision": 0.8249594813614263,
24
+ "eval_micro_recall": 0.6640574037834311,
25
+ "eval_runtime": 2.7054,
26
+ "eval_samples_per_second": 357.063,
27
+ "eval_steps_per_second": 11.459,
28
+ "step": 500
29
+ },
30
+ {
31
+ "epoch": 14.08,
32
+ "learning_rate": 4.295774647887324e-05,
33
+ "loss": 0.0484,
34
+ "step": 1000
35
+ },
36
+ {
37
+ "epoch": 14.08,
38
+ "eval_loss": 0.14033463597297668,
39
+ "eval_macro_f1": 0.6067981552982502,
40
+ "eval_macro_precision": 0.6486392263482452,
41
+ "eval_macro_recall": 0.5884363376224964,
42
+ "eval_micro_f1": 0.7908386662175816,
43
+ "eval_micro_precision": 0.8175487465181058,
44
+ "eval_micro_recall": 0.7658186562296151,
45
+ "eval_runtime": 2.6975,
46
+ "eval_samples_per_second": 358.108,
47
+ "eval_steps_per_second": 11.492,
48
+ "step": 1000
49
+ },
50
+ {
51
+ "epoch": 21.13,
52
+ "learning_rate": 3.943661971830986e-05,
53
+ "loss": 0.0179,
54
+ "step": 1500
55
+ },
56
+ {
57
+ "epoch": 21.13,
58
+ "eval_loss": 0.16313208639621735,
59
+ "eval_macro_f1": 0.6876414143457183,
60
+ "eval_macro_precision": 0.7316617517332842,
61
+ "eval_macro_recall": 0.7019290034007357,
62
+ "eval_micro_f1": 0.7892204042348412,
63
+ "eval_micro_precision": 0.7765151515151515,
64
+ "eval_micro_recall": 0.8023483365949119,
65
+ "eval_runtime": 2.6931,
66
+ "eval_samples_per_second": 358.699,
67
+ "eval_steps_per_second": 11.511,
68
+ "step": 1500
69
+ },
70
+ {
71
+ "epoch": 28.17,
72
+ "learning_rate": 3.5915492957746486e-05,
73
+ "loss": 0.0095,
74
+ "step": 2000
75
+ },
76
+ {
77
+ "epoch": 28.17,
78
+ "eval_loss": 0.1628233790397644,
79
+ "eval_macro_f1": 0.7285510286916194,
80
+ "eval_macro_precision": 0.8247103134528249,
81
+ "eval_macro_recall": 0.6865275313487541,
82
+ "eval_micro_f1": 0.8138990978950885,
83
+ "eval_micro_precision": 0.8342465753424657,
84
+ "eval_micro_recall": 0.7945205479452054,
85
+ "eval_runtime": 2.6912,
86
+ "eval_samples_per_second": 358.946,
87
+ "eval_steps_per_second": 11.519,
88
+ "step": 2000
89
+ },
90
+ {
91
+ "epoch": 35.21,
92
+ "learning_rate": 3.23943661971831e-05,
93
+ "loss": 0.0055,
94
+ "step": 2500
95
+ },
96
+ {
97
+ "epoch": 35.21,
98
+ "eval_loss": 0.1734953373670578,
99
+ "eval_macro_f1": 0.7421994453880427,
100
+ "eval_macro_precision": 0.8204432806900727,
101
+ "eval_macro_recall": 0.7088284387134485,
102
+ "eval_micro_f1": 0.8176722716782063,
103
+ "eval_micro_precision": 0.8266666666666667,
104
+ "eval_micro_recall": 0.8088714938030006,
105
+ "eval_runtime": 2.6929,
106
+ "eval_samples_per_second": 358.718,
107
+ "eval_steps_per_second": 11.512,
108
+ "step": 2500
109
+ },
110
+ {
111
+ "epoch": 42.25,
112
+ "learning_rate": 2.887323943661972e-05,
113
+ "loss": 0.0039,
114
+ "step": 3000
115
+ },
116
+ {
117
+ "epoch": 42.25,
118
+ "eval_loss": 0.1867484599351883,
119
+ "eval_macro_f1": 0.7349168384365183,
120
+ "eval_macro_precision": 0.7741277072189265,
121
+ "eval_macro_recall": 0.7306178466824944,
122
+ "eval_micro_f1": 0.8012924071082391,
123
+ "eval_micro_precision": 0.793854033290653,
124
+ "eval_micro_recall": 0.8088714938030006,
125
+ "eval_runtime": 2.6968,
126
+ "eval_samples_per_second": 358.208,
127
+ "eval_steps_per_second": 11.495,
128
+ "step": 3000
129
+ },
130
+ {
131
+ "epoch": 49.3,
132
+ "learning_rate": 2.535211267605634e-05,
133
+ "loss": 0.0032,
134
+ "step": 3500
135
+ },
136
+ {
137
+ "epoch": 49.3,
138
+ "eval_loss": 0.181275874376297,
139
+ "eval_macro_f1": 0.7442957192288286,
140
+ "eval_macro_precision": 0.7848363899123905,
141
+ "eval_macro_recall": 0.7360907675141599,
142
+ "eval_micro_f1": 0.8202794930126746,
143
+ "eval_micro_precision": 0.8173575129533679,
144
+ "eval_micro_recall": 0.8232224396607958,
145
+ "eval_runtime": 2.6934,
146
+ "eval_samples_per_second": 358.652,
147
+ "eval_steps_per_second": 11.51,
148
+ "step": 3500
149
+ }
150
+ ],
151
+ "max_steps": 7100,
152
+ "num_train_epochs": 100,
153
+ "total_flos": 2.877157614419203e+16,
154
+ "trial_name": null,
155
+ "trial_params": null
156
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95ef9e04b1c54450ebb0e5d368ad5f45d662a1cb072d8e0e194111a56c16c7d5
3
+ size 3247
vocab.json ADDED
The diff for this file is too large to render. See raw diff