felixm commited on
Commit
c9242b2
·
verified ·
1 Parent(s): 81f30e3

Upload 11 files

Browse files
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "KBLab/bert-base-swedish-cased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "output_past": true,
20
+ "pad_token_id": 0,
21
+ "position_embedding_type": "absolute",
22
+ "problem_type": "single_label_classification",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.33.0.dev0",
25
+ "type_vocab_size": 2,
26
+ "use_cache": true,
27
+ "vocab_size": 50325
28
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2772e501a6a8f124e73e4ea3d82a2ea8e285aae674d7640fab424dd38708f54
3
+ size 997707525
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55729a3cd73fdeafb1565c3daad3389a6b08a4c846b3d6aeb64df60d2beb138e
3
+ size 498838321
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25741809c006b1e1d45076291787b153d2177d4b2fa6342ff7eec9d0ad7564dc
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70e4f9db91b06117fdbdc942d6fa35caf00f6ff9b123ac028e720b2d7cf74898
3
+ size 627
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_basic_tokenize": true,
5
+ "do_lower_case": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": false,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
trainer_state.json ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7070142891290766,
3
+ "best_model_checkpoint": "/home/felix/kubhist2-testing/SuperLim-2-Testing/results/trainer_output/run-5d9be_00002/checkpoint-8896",
4
+ "epoch": 4.0,
5
+ "eval_steps": 500,
6
+ "global_step": 8896,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.22,
13
+ "learning_rate": 1.1146067415730336e-05,
14
+ "loss": 0.5378,
15
+ "step": 500
16
+ },
17
+ {
18
+ "epoch": 0.45,
19
+ "learning_rate": 2.2382022471910112e-05,
20
+ "loss": 0.3896,
21
+ "step": 1000
22
+ },
23
+ {
24
+ "epoch": 0.67,
25
+ "learning_rate": 2.976895479550347e-05,
26
+ "loss": 0.3679,
27
+ "step": 1500
28
+ },
29
+ {
30
+ "epoch": 0.9,
31
+ "learning_rate": 2.905142310452045e-05,
32
+ "loss": 0.3735,
33
+ "step": 2000
34
+ },
35
+ {
36
+ "epoch": 1.0,
37
+ "eval_accuracy": 0.8432581880051042,
38
+ "eval_krippendorff_nominal": 0.6854738483266531,
39
+ "eval_loss": 0.3858277499675751,
40
+ "eval_runtime": 3.9653,
41
+ "eval_samples_per_second": 1185.781,
42
+ "eval_steps_per_second": 18.662,
43
+ "learning_rate": 2.8729968906960057e-05,
44
+ "step": 2224
45
+ },
46
+ {
47
+ "epoch": 1.12,
48
+ "learning_rate": 2.833389141353743e-05,
49
+ "loss": 0.3276,
50
+ "step": 2500
51
+ },
52
+ {
53
+ "epoch": 1.35,
54
+ "learning_rate": 2.7616359722554415e-05,
55
+ "loss": 0.3014,
56
+ "step": 3000
57
+ },
58
+ {
59
+ "epoch": 1.57,
60
+ "learning_rate": 2.6898828031571395e-05,
61
+ "loss": 0.3091,
62
+ "step": 3500
63
+ },
64
+ {
65
+ "epoch": 1.8,
66
+ "learning_rate": 2.6181296340588376e-05,
67
+ "loss": 0.3179,
68
+ "step": 4000
69
+ },
70
+ {
71
+ "epoch": 2.0,
72
+ "eval_accuracy": 0.8441088898341131,
73
+ "eval_krippendorff_nominal": 0.6860352395475762,
74
+ "eval_loss": 0.3822194039821625,
75
+ "eval_runtime": 3.9064,
76
+ "eval_samples_per_second": 1203.675,
77
+ "eval_steps_per_second": 18.943,
78
+ "learning_rate": 2.5539823008849558e-05,
79
+ "step": 4448
80
+ },
81
+ {
82
+ "epoch": 2.02,
83
+ "learning_rate": 2.5465199712987326e-05,
84
+ "loss": 0.3132,
85
+ "step": 4500
86
+ },
87
+ {
88
+ "epoch": 2.25,
89
+ "learning_rate": 2.4749103085386272e-05,
90
+ "loss": 0.2406,
91
+ "step": 5000
92
+ },
93
+ {
94
+ "epoch": 2.47,
95
+ "learning_rate": 2.4031571394403256e-05,
96
+ "loss": 0.2518,
97
+ "step": 5500
98
+ },
99
+ {
100
+ "epoch": 2.7,
101
+ "learning_rate": 2.3314039703420236e-05,
102
+ "loss": 0.2558,
103
+ "step": 6000
104
+ },
105
+ {
106
+ "epoch": 2.92,
107
+ "learning_rate": 2.2596508012437217e-05,
108
+ "loss": 0.2739,
109
+ "step": 6500
110
+ },
111
+ {
112
+ "epoch": 3.0,
113
+ "eval_accuracy": 0.839430029774564,
114
+ "eval_krippendorff_nominal": 0.6751470306348389,
115
+ "eval_loss": 0.5222128629684448,
116
+ "eval_runtime": 3.8961,
117
+ "eval_samples_per_second": 1206.858,
118
+ "eval_steps_per_second": 18.994,
119
+ "learning_rate": 2.234967711073906e-05,
120
+ "step": 6672
121
+ },
122
+ {
123
+ "epoch": 3.15,
124
+ "learning_rate": 2.1878976321454197e-05,
125
+ "loss": 0.2358,
126
+ "step": 7000
127
+ },
128
+ {
129
+ "epoch": 3.37,
130
+ "learning_rate": 2.116144463047118e-05,
131
+ "loss": 0.2057,
132
+ "step": 7500
133
+ },
134
+ {
135
+ "epoch": 3.6,
136
+ "learning_rate": 2.044391293948816e-05,
137
+ "loss": 0.2157,
138
+ "step": 8000
139
+ },
140
+ {
141
+ "epoch": 3.82,
142
+ "learning_rate": 1.972638124850514e-05,
143
+ "loss": 0.2295,
144
+ "step": 8500
145
+ },
146
+ {
147
+ "epoch": 4.0,
148
+ "eval_accuracy": 0.8543173117822204,
149
+ "eval_krippendorff_nominal": 0.7070142891290766,
150
+ "eval_loss": 0.44575992226600647,
151
+ "eval_runtime": 3.9739,
152
+ "eval_samples_per_second": 1183.225,
153
+ "eval_steps_per_second": 18.622,
154
+ "learning_rate": 1.9158096149246593e-05,
155
+ "step": 8896
156
+ }
157
+ ],
158
+ "logging_steps": 500,
159
+ "max_steps": 22240,
160
+ "num_train_epochs": 10,
161
+ "save_steps": 500,
162
+ "total_flos": 5898489416824320.0,
163
+ "trial_name": null,
164
+ "trial_params": {
165
+ "learning_rate": 3e-05,
166
+ "per_device_train_batch_size": 16
167
+ }
168
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c7d8309c8bf093577cb91837180e01b99f1d6ba51e7684d9ecf39ae93001a6e
3
+ size 4219
vocab.txt ADDED
The diff for this file is too large to render. See raw diff