anki08 commited on
Commit
3f839dc
·
1 Parent(s): ca3ffac

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +210 -0
README.md ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ metrics:
6
+ - bleu
7
+ model-index:
8
+ - name: t5-small-finetuned-text2log-compute-metrics-v5-400
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # t5-small-finetuned-text2log-compute-metrics-v5-400
16
+
17
+ This model is a fine-tuned version of [mrm8488/t5-small-finetuned-text2log](https://huggingface.co/mrm8488/t5-small-finetuned-text2log) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 0.5820
20
+ - Bleu: 30.1378
21
+ - Gen Len: 18.568
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 5e-05
41
+ - train_batch_size: 8
42
+ - eval_batch_size: 8
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - num_epochs: 150
47
+ - mixed_precision_training: Native AMP
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
52
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
53
+ | No log | 1.0 | 23 | 1.5732 | 8.4784 | 17.3373 |
54
+ | No log | 2.0 | 46 | 1.1400 | 22.8492 | 18.8284 |
55
+ | No log | 3.0 | 69 | 0.9426 | 27.2247 | 18.787 |
56
+ | No log | 4.0 | 92 | 0.8110 | 26.0131 | 18.7101 |
57
+ | No log | 5.0 | 115 | 0.7559 | 24.1316 | 18.7988 |
58
+ | No log | 6.0 | 138 | 0.7102 | 25.4664 | 18.7692 |
59
+ | No log | 7.0 | 161 | 0.6807 | 26.2651 | 18.6923 |
60
+ | No log | 8.0 | 184 | 0.6575 | 27.5298 | 18.6213 |
61
+ | No log | 9.0 | 207 | 0.6369 | 28.8178 | 18.5976 |
62
+ | No log | 10.0 | 230 | 0.6250 | 27.2363 | 18.6864 |
63
+ | No log | 11.0 | 253 | 0.6143 | 28.7917 | 18.6391 |
64
+ | No log | 12.0 | 276 | 0.5973 | 29.6901 | 18.7041 |
65
+ | No log | 13.0 | 299 | 0.5864 | 29.6632 | 18.6568 |
66
+ | No log | 14.0 | 322 | 0.5827 | 29.874 | 18.6391 |
67
+ | No log | 15.0 | 345 | 0.5700 | 29.6188 | 18.6864 |
68
+ | No log | 16.0 | 368 | 0.5620 | 29.9895 | 18.7278 |
69
+ | No log | 17.0 | 391 | 0.5569 | 29.3883 | 18.7278 |
70
+ | No log | 18.0 | 414 | 0.5486 | 29.783 | 18.6864 |
71
+ | No log | 19.0 | 437 | 0.5520 | 29.4199 | 18.6568 |
72
+ | No log | 20.0 | 460 | 0.5496 | 28.8502 | 18.6923 |
73
+ | No log | 21.0 | 483 | 0.5495 | 29.2369 | 18.7337 |
74
+ | 0.8287 | 22.0 | 506 | 0.5411 | 29.1707 | 18.6923 |
75
+ | 0.8287 | 23.0 | 529 | 0.5325 | 28.8466 | 18.6923 |
76
+ | 0.8287 | 24.0 | 552 | 0.5388 | 29.2703 | 18.6746 |
77
+ | 0.8287 | 25.0 | 575 | 0.5288 | 28.7683 | 18.7278 |
78
+ | 0.8287 | 26.0 | 598 | 0.5299 | 28.3489 | 18.7337 |
79
+ | 0.8287 | 27.0 | 621 | 0.5314 | 28.2042 | 18.7278 |
80
+ | 0.8287 | 28.0 | 644 | 0.5159 | 29.5603 | 18.7574 |
81
+ | 0.8287 | 29.0 | 667 | 0.5163 | 28.8959 | 18.6982 |
82
+ | 0.8287 | 30.0 | 690 | 0.5170 | 30.0671 | 18.7219 |
83
+ | 0.8287 | 31.0 | 713 | 0.5202 | 28.9559 | 18.6686 |
84
+ | 0.8287 | 32.0 | 736 | 0.5181 | 28.1951 | 18.7574 |
85
+ | 0.8287 | 33.0 | 759 | 0.5134 | 28.2097 | 18.7041 |
86
+ | 0.8287 | 34.0 | 782 | 0.5134 | 28.0953 | 18.7574 |
87
+ | 0.8287 | 35.0 | 805 | 0.5165 | 28.4171 | 18.7988 |
88
+ | 0.8287 | 36.0 | 828 | 0.5195 | 28.7953 | 18.7633 |
89
+ | 0.8287 | 37.0 | 851 | 0.5185 | 29.1606 | 18.7633 |
90
+ | 0.8287 | 38.0 | 874 | 0.5192 | 29.5226 | 18.6805 |
91
+ | 0.8287 | 39.0 | 897 | 0.5178 | 28.425 | 18.7692 |
92
+ | 0.8287 | 40.0 | 920 | 0.5196 | 29.4795 | 18.7041 |
93
+ | 0.8287 | 41.0 | 943 | 0.5161 | 29.5127 | 18.7041 |
94
+ | 0.8287 | 42.0 | 966 | 0.5164 | 28.6225 | 18.7574 |
95
+ | 0.8287 | 43.0 | 989 | 0.5183 | 29.0629 | 18.6272 |
96
+ | 0.3866 | 44.0 | 1012 | 0.5174 | 28.6628 | 18.6213 |
97
+ | 0.3866 | 45.0 | 1035 | 0.5141 | 28.499 | 18.6213 |
98
+ | 0.3866 | 46.0 | 1058 | 0.5151 | 28.1309 | 18.6272 |
99
+ | 0.3866 | 47.0 | 1081 | 0.5167 | 29.8871 | 18.6391 |
100
+ | 0.3866 | 48.0 | 1104 | 0.5133 | 28.7513 | 18.6331 |
101
+ | 0.3866 | 49.0 | 1127 | 0.5188 | 28.3089 | 18.6213 |
102
+ | 0.3866 | 50.0 | 1150 | 0.5203 | 28.8714 | 18.6331 |
103
+ | 0.3866 | 51.0 | 1173 | 0.5263 | 28.7644 | 18.6331 |
104
+ | 0.3866 | 52.0 | 1196 | 0.5222 | 28.4817 | 18.6331 |
105
+ | 0.3866 | 53.0 | 1219 | 0.5307 | 28.8117 | 18.6272 |
106
+ | 0.3866 | 54.0 | 1242 | 0.5255 | 29.3844 | 18.6213 |
107
+ | 0.3866 | 55.0 | 1265 | 0.5264 | 29.7108 | 18.6213 |
108
+ | 0.3866 | 56.0 | 1288 | 0.5272 | 29.353 | 18.6331 |
109
+ | 0.3866 | 57.0 | 1311 | 0.5283 | 28.5792 | 18.6391 |
110
+ | 0.3866 | 58.0 | 1334 | 0.5301 | 29.9914 | 18.6272 |
111
+ | 0.3866 | 59.0 | 1357 | 0.5320 | 29.3162 | 18.6272 |
112
+ | 0.3866 | 60.0 | 1380 | 0.5380 | 29.2162 | 18.6272 |
113
+ | 0.3866 | 61.0 | 1403 | 0.5349 | 28.5292 | 18.6272 |
114
+ | 0.3866 | 62.0 | 1426 | 0.5313 | 28.7165 | 18.6627 |
115
+ | 0.3866 | 63.0 | 1449 | 0.5335 | 29.3637 | 18.6154 |
116
+ | 0.3866 | 64.0 | 1472 | 0.5350 | 29.3612 | 18.568 |
117
+ | 0.3866 | 65.0 | 1495 | 0.5330 | 29.1338 | 18.5621 |
118
+ | 0.283 | 66.0 | 1518 | 0.5322 | 29.0514 | 18.5562 |
119
+ | 0.283 | 67.0 | 1541 | 0.5362 | 29.2592 | 18.5562 |
120
+ | 0.283 | 68.0 | 1564 | 0.5379 | 29.6757 | 18.568 |
121
+ | 0.283 | 69.0 | 1587 | 0.5386 | 29.5012 | 18.5976 |
122
+ | 0.283 | 70.0 | 1610 | 0.5379 | 29.6616 | 18.5917 |
123
+ | 0.283 | 71.0 | 1633 | 0.5364 | 29.8762 | 18.6154 |
124
+ | 0.283 | 72.0 | 1656 | 0.5392 | 29.7143 | 18.6036 |
125
+ | 0.283 | 73.0 | 1679 | 0.5438 | 29.385 | 18.5976 |
126
+ | 0.283 | 74.0 | 1702 | 0.5386 | 28.3472 | 18.6095 |
127
+ | 0.283 | 75.0 | 1725 | 0.5372 | 29.1045 | 18.574 |
128
+ | 0.283 | 76.0 | 1748 | 0.5406 | 29.0839 | 18.6095 |
129
+ | 0.283 | 77.0 | 1771 | 0.5408 | 29.735 | 18.5799 |
130
+ | 0.283 | 78.0 | 1794 | 0.5406 | 29.5432 | 18.6036 |
131
+ | 0.283 | 79.0 | 1817 | 0.5413 | 29.1501 | 18.5976 |
132
+ | 0.283 | 80.0 | 1840 | 0.5434 | 29.5822 | 18.6095 |
133
+ | 0.283 | 81.0 | 1863 | 0.5491 | 29.1933 | 18.5799 |
134
+ | 0.283 | 82.0 | 1886 | 0.5473 | 28.9065 | 18.5385 |
135
+ | 0.283 | 83.0 | 1909 | 0.5507 | 29.4129 | 18.5385 |
136
+ | 0.283 | 84.0 | 1932 | 0.5534 | 29.2249 | 18.5385 |
137
+ | 0.283 | 85.0 | 1955 | 0.5561 | 29.6955 | 18.5799 |
138
+ | 0.283 | 86.0 | 1978 | 0.5575 | 29.1081 | 18.5385 |
139
+ | 0.2296 | 87.0 | 2001 | 0.5531 | 29.7633 | 18.5385 |
140
+ | 0.2296 | 88.0 | 2024 | 0.5548 | 30.045 | 18.5385 |
141
+ | 0.2296 | 89.0 | 2047 | 0.5567 | 29.9209 | 18.5385 |
142
+ | 0.2296 | 90.0 | 2070 | 0.5577 | 29.1879 | 18.5858 |
143
+ | 0.2296 | 91.0 | 2093 | 0.5602 | 29.1587 | 18.5799 |
144
+ | 0.2296 | 92.0 | 2116 | 0.5595 | 29.5205 | 18.5799 |
145
+ | 0.2296 | 93.0 | 2139 | 0.5605 | 29.3439 | 18.5325 |
146
+ | 0.2296 | 94.0 | 2162 | 0.5583 | 29.4742 | 18.5325 |
147
+ | 0.2296 | 95.0 | 2185 | 0.5576 | 29.132 | 18.5325 |
148
+ | 0.2296 | 96.0 | 2208 | 0.5566 | 29.0861 | 18.5325 |
149
+ | 0.2296 | 97.0 | 2231 | 0.5584 | 29.6618 | 18.5385 |
150
+ | 0.2296 | 98.0 | 2254 | 0.5593 | 29.1068 | 18.5325 |
151
+ | 0.2296 | 99.0 | 2277 | 0.5603 | 29.7081 | 18.5385 |
152
+ | 0.2296 | 100.0 | 2300 | 0.5599 | 29.6368 | 18.5325 |
153
+ | 0.2296 | 101.0 | 2323 | 0.5598 | 29.6263 | 18.5325 |
154
+ | 0.2296 | 102.0 | 2346 | 0.5637 | 29.6321 | 18.5385 |
155
+ | 0.2296 | 103.0 | 2369 | 0.5678 | 29.6306 | 18.5266 |
156
+ | 0.2296 | 104.0 | 2392 | 0.5685 | 29.3279 | 18.5325 |
157
+ | 0.2296 | 105.0 | 2415 | 0.5680 | 29.1363 | 18.5621 |
158
+ | 0.2296 | 106.0 | 2438 | 0.5726 | 29.2666 | 18.5385 |
159
+ | 0.2296 | 107.0 | 2461 | 0.5738 | 29.2981 | 18.5385 |
160
+ | 0.2296 | 108.0 | 2484 | 0.5740 | 29.5752 | 18.5385 |
161
+ | 0.1942 | 109.0 | 2507 | 0.5749 | 29.5596 | 18.5385 |
162
+ | 0.1942 | 110.0 | 2530 | 0.5732 | 29.6728 | 18.574 |
163
+ | 0.1942 | 111.0 | 2553 | 0.5738 | 29.6052 | 18.5325 |
164
+ | 0.1942 | 112.0 | 2576 | 0.5731 | 29.5143 | 18.574 |
165
+ | 0.1942 | 113.0 | 2599 | 0.5744 | 29.8059 | 18.574 |
166
+ | 0.1942 | 114.0 | 2622 | 0.5751 | 29.6796 | 18.574 |
167
+ | 0.1942 | 115.0 | 2645 | 0.5763 | 29.9279 | 18.568 |
168
+ | 0.1942 | 116.0 | 2668 | 0.5746 | 29.892 | 18.568 |
169
+ | 0.1942 | 117.0 | 2691 | 0.5741 | 29.8104 | 18.574 |
170
+ | 0.1942 | 118.0 | 2714 | 0.5759 | 29.6379 | 18.574 |
171
+ | 0.1942 | 119.0 | 2737 | 0.5777 | 29.7949 | 18.574 |
172
+ | 0.1942 | 120.0 | 2760 | 0.5776 | 29.6297 | 18.574 |
173
+ | 0.1942 | 121.0 | 2783 | 0.5789 | 29.5298 | 18.574 |
174
+ | 0.1942 | 122.0 | 2806 | 0.5794 | 29.6102 | 18.574 |
175
+ | 0.1942 | 123.0 | 2829 | 0.5799 | 29.7981 | 18.574 |
176
+ | 0.1942 | 124.0 | 2852 | 0.5811 | 30.0894 | 18.574 |
177
+ | 0.1942 | 125.0 | 2875 | 0.5826 | 29.9849 | 18.574 |
178
+ | 0.1942 | 126.0 | 2898 | 0.5829 | 29.8349 | 18.574 |
179
+ | 0.1942 | 127.0 | 2921 | 0.5817 | 29.6295 | 18.574 |
180
+ | 0.1942 | 128.0 | 2944 | 0.5809 | 29.5264 | 18.568 |
181
+ | 0.1942 | 129.0 | 2967 | 0.5813 | 29.5858 | 18.568 |
182
+ | 0.1942 | 130.0 | 2990 | 0.5843 | 29.6556 | 18.568 |
183
+ | 0.1777 | 131.0 | 3013 | 0.5836 | 30.0165 | 18.568 |
184
+ | 0.1777 | 132.0 | 3036 | 0.5835 | 29.8399 | 18.568 |
185
+ | 0.1777 | 133.0 | 3059 | 0.5824 | 29.8065 | 18.568 |
186
+ | 0.1777 | 134.0 | 3082 | 0.5821 | 29.8948 | 18.574 |
187
+ | 0.1777 | 135.0 | 3105 | 0.5808 | 29.9342 | 18.574 |
188
+ | 0.1777 | 136.0 | 3128 | 0.5810 | 29.7556 | 18.574 |
189
+ | 0.1777 | 137.0 | 3151 | 0.5813 | 30.0425 | 18.568 |
190
+ | 0.1777 | 138.0 | 3174 | 0.5822 | 30.0719 | 18.568 |
191
+ | 0.1777 | 139.0 | 3197 | 0.5823 | 30.0719 | 18.568 |
192
+ | 0.1777 | 140.0 | 3220 | 0.5828 | 30.1124 | 18.568 |
193
+ | 0.1777 | 141.0 | 3243 | 0.5826 | 30.1451 | 18.568 |
194
+ | 0.1777 | 142.0 | 3266 | 0.5828 | 30.1451 | 18.568 |
195
+ | 0.1777 | 143.0 | 3289 | 0.5829 | 30.1451 | 18.568 |
196
+ | 0.1777 | 144.0 | 3312 | 0.5829 | 30.1451 | 18.568 |
197
+ | 0.1777 | 145.0 | 3335 | 0.5825 | 30.1378 | 18.568 |
198
+ | 0.1777 | 146.0 | 3358 | 0.5824 | 30.1378 | 18.568 |
199
+ | 0.1777 | 147.0 | 3381 | 0.5822 | 30.1378 | 18.568 |
200
+ | 0.1777 | 148.0 | 3404 | 0.5820 | 30.1378 | 18.568 |
201
+ | 0.1777 | 149.0 | 3427 | 0.5821 | 30.1378 | 18.568 |
202
+ | 0.1777 | 150.0 | 3450 | 0.5820 | 30.1378 | 18.568 |
203
+
204
+
205
+ ### Framework versions
206
+
207
+ - Transformers 4.21.2
208
+ - Pytorch 1.12.1+cu113
209
+ - Datasets 2.4.0
210
+ - Tokenizers 0.12.1