Rocky807 commited on
Commit
e7f0038
·
verified ·
1 Parent(s): a6e319f

End of training

Browse files
README.md CHANGED
@@ -17,8 +17,8 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [microsoft/beit-base-patch16-224-pt22k-ft22k](https://huggingface.co/microsoft/beit-base-patch16-224-pt22k-ft22k) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.2845
21
- - Accuracy: 0.9058
22
 
23
  ## Model description
24
 
 
17
 
18
  This model is a fine-tuned version of [microsoft/beit-base-patch16-224-pt22k-ft22k](https://huggingface.co/microsoft/beit-base-patch16-224-pt22k-ft22k) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.2796
21
+ - Accuracy: 0.9082
22
 
23
  ## Model description
24
 
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 2.99,
3
- "eval_accuracy": 0.9192,
4
- "eval_loss": 0.23591181635856628,
5
- "eval_runtime": 78.5827,
6
- "eval_samples_per_second": 63.627,
7
- "eval_steps_per_second": 1.998
8
  }
 
1
  {
2
+ "epoch": 1.0,
3
+ "eval_accuracy": 0.9082,
4
+ "eval_loss": 0.2795782685279846,
5
+ "eval_runtime": 78.9983,
6
+ "eval_samples_per_second": 63.293,
7
+ "eval_steps_per_second": 1.987
8
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 2.99,
3
- "eval_accuracy": 0.9192,
4
- "eval_loss": 0.23591181635856628,
5
- "eval_runtime": 78.5827,
6
- "eval_samples_per_second": 63.627,
7
- "eval_steps_per_second": 1.998
8
  }
 
1
  {
2
+ "epoch": 1.0,
3
+ "eval_accuracy": 0.9082,
4
+ "eval_loss": 0.2795782685279846,
5
+ "eval_runtime": 78.9983,
6
+ "eval_samples_per_second": 63.293,
7
+ "eval_steps_per_second": 1.987
8
  }
runs/Apr24_11-55-23_MacBook-Pro.local/events.out.tfevents.1713933335.MacBook-Pro.local.24885.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a0eebd7e731627a42f4dbd395cb5165d97d9aed588729b210142189bbe10031
3
+ size 411
trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.9064,
3
  "best_model_checkpoint": "beit-base-patch16-224-pt22k-ft22k-finetuning/checkpoint-351",
4
  "epoch": 0.997867803837953,
5
  "eval_steps": 500,
@@ -10,265 +10,265 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.03,
13
- "grad_norm": 12.392043113708496,
14
  "learning_rate": 1.388888888888889e-05,
15
- "loss": 2.38,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.06,
20
- "grad_norm": 9.366357803344727,
21
  "learning_rate": 2.777777777777778e-05,
22
- "loss": 2.0468,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.09,
27
- "grad_norm": 10.620584487915039,
28
  "learning_rate": 4.166666666666667e-05,
29
- "loss": 1.5116,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 0.11,
34
- "grad_norm": 9.78640079498291,
35
  "learning_rate": 4.936507936507937e-05,
36
- "loss": 1.0056,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 0.14,
41
- "grad_norm": 7.943943977355957,
42
  "learning_rate": 4.7777777777777784e-05,
43
- "loss": 0.7888,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 0.17,
48
- "grad_norm": 9.149613380432129,
49
  "learning_rate": 4.6190476190476194e-05,
50
- "loss": 0.6486,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 0.2,
55
- "grad_norm": 7.764841079711914,
56
  "learning_rate": 4.460317460317461e-05,
57
- "loss": 0.6232,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 0.23,
62
- "grad_norm": 8.434979438781738,
63
  "learning_rate": 4.301587301587302e-05,
64
- "loss": 0.5555,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 0.26,
69
- "grad_norm": 8.733748435974121,
70
  "learning_rate": 4.1428571428571437e-05,
71
- "loss": 0.4872,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 0.28,
76
- "grad_norm": 7.663728713989258,
77
  "learning_rate": 3.984126984126984e-05,
78
- "loss": 0.5385,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 0.31,
83
- "grad_norm": 6.9408698081970215,
84
  "learning_rate": 3.8253968253968256e-05,
85
- "loss": 0.4591,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 0.34,
90
- "grad_norm": 6.367518424987793,
91
  "learning_rate": 3.6666666666666666e-05,
92
- "loss": 0.4449,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 0.37,
97
- "grad_norm": 6.589195728302002,
98
  "learning_rate": 3.5079365079365075e-05,
99
- "loss": 0.4365,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 0.4,
104
- "grad_norm": 8.5034818649292,
105
  "learning_rate": 3.349206349206349e-05,
106
- "loss": 0.4538,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 0.43,
111
- "grad_norm": 6.432806491851807,
112
  "learning_rate": 3.19047619047619e-05,
113
- "loss": 0.4182,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 0.45,
118
- "grad_norm": 5.926751613616943,
119
  "learning_rate": 3.0317460317460318e-05,
120
- "loss": 0.3683,
121
  "step": 160
122
  },
123
  {
124
  "epoch": 0.48,
125
- "grad_norm": 7.516120433807373,
126
  "learning_rate": 2.8730158730158728e-05,
127
- "loss": 0.4315,
128
  "step": 170
129
  },
130
  {
131
  "epoch": 0.51,
132
- "grad_norm": 7.042306423187256,
133
  "learning_rate": 2.714285714285714e-05,
134
- "loss": 0.3883,
135
  "step": 180
136
  },
137
  {
138
  "epoch": 0.54,
139
- "grad_norm": 6.7353620529174805,
140
  "learning_rate": 2.5555555555555554e-05,
141
- "loss": 0.4158,
142
  "step": 190
143
  },
144
  {
145
  "epoch": 0.57,
146
- "grad_norm": 6.9553914070129395,
147
  "learning_rate": 2.396825396825397e-05,
148
- "loss": 0.3893,
149
  "step": 200
150
  },
151
  {
152
  "epoch": 0.6,
153
- "grad_norm": 6.948991298675537,
154
  "learning_rate": 2.2380952380952384e-05,
155
- "loss": 0.4288,
156
  "step": 210
157
  },
158
  {
159
  "epoch": 0.63,
160
- "grad_norm": 6.90835428237915,
161
  "learning_rate": 2.0793650793650797e-05,
162
- "loss": 0.368,
163
  "step": 220
164
  },
165
  {
166
  "epoch": 0.65,
167
- "grad_norm": 7.47681999206543,
168
  "learning_rate": 1.920634920634921e-05,
169
- "loss": 0.4212,
170
  "step": 230
171
  },
172
  {
173
  "epoch": 0.68,
174
- "grad_norm": 5.4326887130737305,
175
  "learning_rate": 1.761904761904762e-05,
176
- "loss": 0.4101,
177
  "step": 240
178
  },
179
  {
180
  "epoch": 0.71,
181
- "grad_norm": 5.068437576293945,
182
  "learning_rate": 1.6031746031746033e-05,
183
- "loss": 0.3544,
184
  "step": 250
185
  },
186
  {
187
  "epoch": 0.74,
188
- "grad_norm": 5.686431884765625,
189
  "learning_rate": 1.4444444444444444e-05,
190
- "loss": 0.3252,
191
  "step": 260
192
  },
193
  {
194
  "epoch": 0.77,
195
- "grad_norm": 5.882190227508545,
196
  "learning_rate": 1.2857142857142857e-05,
197
- "loss": 0.3475,
198
  "step": 270
199
  },
200
  {
201
  "epoch": 0.8,
202
- "grad_norm": 6.3953447341918945,
203
  "learning_rate": 1.126984126984127e-05,
204
- "loss": 0.3164,
205
  "step": 280
206
  },
207
  {
208
  "epoch": 0.82,
209
- "grad_norm": 7.278485298156738,
210
  "learning_rate": 9.682539682539683e-06,
211
- "loss": 0.328,
212
  "step": 290
213
  },
214
  {
215
  "epoch": 0.85,
216
- "grad_norm": 6.7763166427612305,
217
  "learning_rate": 8.095238095238097e-06,
218
- "loss": 0.3463,
219
  "step": 300
220
  },
221
  {
222
  "epoch": 0.88,
223
- "grad_norm": 7.21019172668457,
224
  "learning_rate": 6.507936507936509e-06,
225
- "loss": 0.3762,
226
  "step": 310
227
  },
228
  {
229
  "epoch": 0.91,
230
- "grad_norm": 8.0834379196167,
231
  "learning_rate": 4.920634920634921e-06,
232
- "loss": 0.3538,
233
  "step": 320
234
  },
235
  {
236
  "epoch": 0.94,
237
- "grad_norm": 5.969318389892578,
238
  "learning_rate": 3.3333333333333333e-06,
239
- "loss": 0.3159,
240
  "step": 330
241
  },
242
  {
243
  "epoch": 0.97,
244
- "grad_norm": 7.134821891784668,
245
  "learning_rate": 1.7460317460317462e-06,
246
- "loss": 0.3528,
247
  "step": 340
248
  },
249
  {
250
  "epoch": 1.0,
251
- "grad_norm": 6.126035213470459,
252
  "learning_rate": 1.5873015873015874e-07,
253
- "loss": 0.2928,
254
  "step": 350
255
  },
256
  {
257
  "epoch": 1.0,
258
- "eval_accuracy": 0.9064,
259
- "eval_loss": 0.27925798296928406,
260
- "eval_runtime": 77.4382,
261
- "eval_samples_per_second": 64.568,
262
- "eval_steps_per_second": 2.027,
263
  "step": 351
264
  },
265
  {
266
  "epoch": 1.0,
267
  "step": 351,
268
  "total_flos": 3.4803216831306793e+18,
269
- "train_loss": 0.5742807560666674,
270
- "train_runtime": 2256.218,
271
- "train_samples_per_second": 19.945,
272
  "train_steps_per_second": 0.156
273
  }
274
  ],
 
1
  {
2
+ "best_metric": 0.9058,
3
  "best_model_checkpoint": "beit-base-patch16-224-pt22k-ft22k-finetuning/checkpoint-351",
4
  "epoch": 0.997867803837953,
5
  "eval_steps": 500,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.03,
13
+ "grad_norm": 11.096900939941406,
14
  "learning_rate": 1.388888888888889e-05,
15
+ "loss": 2.3539,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.06,
20
+ "grad_norm": 8.86681079864502,
21
  "learning_rate": 2.777777777777778e-05,
22
+ "loss": 2.0154,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.09,
27
+ "grad_norm": 9.853775978088379,
28
  "learning_rate": 4.166666666666667e-05,
29
+ "loss": 1.4709,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 0.11,
34
+ "grad_norm": 10.110260963439941,
35
  "learning_rate": 4.936507936507937e-05,
36
+ "loss": 1.0243,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 0.14,
41
+ "grad_norm": 10.437296867370605,
42
  "learning_rate": 4.7777777777777784e-05,
43
+ "loss": 0.7642,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 0.17,
48
+ "grad_norm": 7.33429479598999,
49
  "learning_rate": 4.6190476190476194e-05,
50
+ "loss": 0.6573,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 0.2,
55
+ "grad_norm": 7.737111568450928,
56
  "learning_rate": 4.460317460317461e-05,
57
+ "loss": 0.5998,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 0.23,
62
+ "grad_norm": 7.117513179779053,
63
  "learning_rate": 4.301587301587302e-05,
64
+ "loss": 0.5478,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 0.26,
69
+ "grad_norm": 8.698220252990723,
70
  "learning_rate": 4.1428571428571437e-05,
71
+ "loss": 0.5195,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 0.28,
76
+ "grad_norm": 6.8238677978515625,
77
  "learning_rate": 3.984126984126984e-05,
78
+ "loss": 0.5212,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 0.31,
83
+ "grad_norm": 6.604701042175293,
84
  "learning_rate": 3.8253968253968256e-05,
85
+ "loss": 0.4546,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 0.34,
90
+ "grad_norm": 8.972763061523438,
91
  "learning_rate": 3.6666666666666666e-05,
92
+ "loss": 0.4429,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 0.37,
97
+ "grad_norm": 6.5601959228515625,
98
  "learning_rate": 3.5079365079365075e-05,
99
+ "loss": 0.4326,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 0.4,
104
+ "grad_norm": 7.662568092346191,
105
  "learning_rate": 3.349206349206349e-05,
106
+ "loss": 0.4859,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 0.43,
111
+ "grad_norm": 8.094512939453125,
112
  "learning_rate": 3.19047619047619e-05,
113
+ "loss": 0.4159,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 0.45,
118
+ "grad_norm": 6.0086846351623535,
119
  "learning_rate": 3.0317460317460318e-05,
120
+ "loss": 0.4216,
121
  "step": 160
122
  },
123
  {
124
  "epoch": 0.48,
125
+ "grad_norm": 7.797686576843262,
126
  "learning_rate": 2.8730158730158728e-05,
127
+ "loss": 0.434,
128
  "step": 170
129
  },
130
  {
131
  "epoch": 0.51,
132
+ "grad_norm": 7.188005447387695,
133
  "learning_rate": 2.714285714285714e-05,
134
+ "loss": 0.3736,
135
  "step": 180
136
  },
137
  {
138
  "epoch": 0.54,
139
+ "grad_norm": 5.677178382873535,
140
  "learning_rate": 2.5555555555555554e-05,
141
+ "loss": 0.3726,
142
  "step": 190
143
  },
144
  {
145
  "epoch": 0.57,
146
+ "grad_norm": 8.169827461242676,
147
  "learning_rate": 2.396825396825397e-05,
148
+ "loss": 0.3675,
149
  "step": 200
150
  },
151
  {
152
  "epoch": 0.6,
153
+ "grad_norm": 6.677453994750977,
154
  "learning_rate": 2.2380952380952384e-05,
155
+ "loss": 0.3794,
156
  "step": 210
157
  },
158
  {
159
  "epoch": 0.63,
160
+ "grad_norm": 6.1892876625061035,
161
  "learning_rate": 2.0793650793650797e-05,
162
+ "loss": 0.4131,
163
  "step": 220
164
  },
165
  {
166
  "epoch": 0.65,
167
+ "grad_norm": 7.120225429534912,
168
  "learning_rate": 1.920634920634921e-05,
169
+ "loss": 0.3614,
170
  "step": 230
171
  },
172
  {
173
  "epoch": 0.68,
174
+ "grad_norm": 7.643838405609131,
175
  "learning_rate": 1.761904761904762e-05,
176
+ "loss": 0.3806,
177
  "step": 240
178
  },
179
  {
180
  "epoch": 0.71,
181
+ "grad_norm": 5.949714660644531,
182
  "learning_rate": 1.6031746031746033e-05,
183
+ "loss": 0.33,
184
  "step": 250
185
  },
186
  {
187
  "epoch": 0.74,
188
+ "grad_norm": 7.846484661102295,
189
  "learning_rate": 1.4444444444444444e-05,
190
+ "loss": 0.3952,
191
  "step": 260
192
  },
193
  {
194
  "epoch": 0.77,
195
+ "grad_norm": 8.437196731567383,
196
  "learning_rate": 1.2857142857142857e-05,
197
+ "loss": 0.3512,
198
  "step": 270
199
  },
200
  {
201
  "epoch": 0.8,
202
+ "grad_norm": 6.006718158721924,
203
  "learning_rate": 1.126984126984127e-05,
204
+ "loss": 0.3003,
205
  "step": 280
206
  },
207
  {
208
  "epoch": 0.82,
209
+ "grad_norm": 8.551034927368164,
210
  "learning_rate": 9.682539682539683e-06,
211
+ "loss": 0.3413,
212
  "step": 290
213
  },
214
  {
215
  "epoch": 0.85,
216
+ "grad_norm": 7.815725326538086,
217
  "learning_rate": 8.095238095238097e-06,
218
+ "loss": 0.3174,
219
  "step": 300
220
  },
221
  {
222
  "epoch": 0.88,
223
+ "grad_norm": 7.373149871826172,
224
  "learning_rate": 6.507936507936509e-06,
225
+ "loss": 0.337,
226
  "step": 310
227
  },
228
  {
229
  "epoch": 0.91,
230
+ "grad_norm": 6.722525596618652,
231
  "learning_rate": 4.920634920634921e-06,
232
+ "loss": 0.3622,
233
  "step": 320
234
  },
235
  {
236
  "epoch": 0.94,
237
+ "grad_norm": 8.15896224975586,
238
  "learning_rate": 3.3333333333333333e-06,
239
+ "loss": 0.3043,
240
  "step": 330
241
  },
242
  {
243
  "epoch": 0.97,
244
+ "grad_norm": 6.256545066833496,
245
  "learning_rate": 1.7460317460317462e-06,
246
+ "loss": 0.3147,
247
  "step": 340
248
  },
249
  {
250
  "epoch": 1.0,
251
+ "grad_norm": 6.901019096374512,
252
  "learning_rate": 1.5873015873015874e-07,
253
+ "loss": 0.3075,
254
  "step": 350
255
  },
256
  {
257
  "epoch": 1.0,
258
+ "eval_accuracy": 0.9058,
259
+ "eval_loss": 0.2844555079936981,
260
+ "eval_runtime": 80.5038,
261
+ "eval_samples_per_second": 62.109,
262
+ "eval_steps_per_second": 1.95,
263
  "step": 351
264
  },
265
  {
266
  "epoch": 1.0,
267
  "step": 351,
268
  "total_flos": 3.4803216831306793e+18,
269
+ "train_loss": 0.5666828784253183,
270
+ "train_runtime": 2255.5076,
271
+ "train_samples_per_second": 19.951,
272
  "train_steps_per_second": 0.156
273
  }
274
  ],