update model card README.md
Browse files
README.md
ADDED
@@ -0,0 +1,210 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- generated_from_trainer
|
5 |
+
metrics:
|
6 |
+
- bleu
|
7 |
+
model-index:
|
8 |
+
- name: t5-small-finetuned-text2log-compute-metrics-v5-400
|
9 |
+
results: []
|
10 |
+
---
|
11 |
+
|
12 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
+
should probably proofread and complete it, then remove this comment. -->
|
14 |
+
|
15 |
+
# t5-small-finetuned-text2log-compute-metrics-v5-400
|
16 |
+
|
17 |
+
This model is a fine-tuned version of [mrm8488/t5-small-finetuned-text2log](https://huggingface.co/mrm8488/t5-small-finetuned-text2log) on an unknown dataset.
|
18 |
+
It achieves the following results on the evaluation set:
|
19 |
+
- Loss: 0.5820
|
20 |
+
- Bleu: 30.1378
|
21 |
+
- Gen Len: 18.568
|
22 |
+
|
23 |
+
## Model description
|
24 |
+
|
25 |
+
More information needed
|
26 |
+
|
27 |
+
## Intended uses & limitations
|
28 |
+
|
29 |
+
More information needed
|
30 |
+
|
31 |
+
## Training and evaluation data
|
32 |
+
|
33 |
+
More information needed
|
34 |
+
|
35 |
+
## Training procedure
|
36 |
+
|
37 |
+
### Training hyperparameters
|
38 |
+
|
39 |
+
The following hyperparameters were used during training:
|
40 |
+
- learning_rate: 5e-05
|
41 |
+
- train_batch_size: 8
|
42 |
+
- eval_batch_size: 8
|
43 |
+
- seed: 42
|
44 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
45 |
+
- lr_scheduler_type: linear
|
46 |
+
- num_epochs: 150
|
47 |
+
- mixed_precision_training: Native AMP
|
48 |
+
|
49 |
+
### Training results
|
50 |
+
|
51 |
+
| Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
|
52 |
+
|:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
|
53 |
+
| No log | 1.0 | 23 | 1.5732 | 8.4784 | 17.3373 |
|
54 |
+
| No log | 2.0 | 46 | 1.1400 | 22.8492 | 18.8284 |
|
55 |
+
| No log | 3.0 | 69 | 0.9426 | 27.2247 | 18.787 |
|
56 |
+
| No log | 4.0 | 92 | 0.8110 | 26.0131 | 18.7101 |
|
57 |
+
| No log | 5.0 | 115 | 0.7559 | 24.1316 | 18.7988 |
|
58 |
+
| No log | 6.0 | 138 | 0.7102 | 25.4664 | 18.7692 |
|
59 |
+
| No log | 7.0 | 161 | 0.6807 | 26.2651 | 18.6923 |
|
60 |
+
| No log | 8.0 | 184 | 0.6575 | 27.5298 | 18.6213 |
|
61 |
+
| No log | 9.0 | 207 | 0.6369 | 28.8178 | 18.5976 |
|
62 |
+
| No log | 10.0 | 230 | 0.6250 | 27.2363 | 18.6864 |
|
63 |
+
| No log | 11.0 | 253 | 0.6143 | 28.7917 | 18.6391 |
|
64 |
+
| No log | 12.0 | 276 | 0.5973 | 29.6901 | 18.7041 |
|
65 |
+
| No log | 13.0 | 299 | 0.5864 | 29.6632 | 18.6568 |
|
66 |
+
| No log | 14.0 | 322 | 0.5827 | 29.874 | 18.6391 |
|
67 |
+
| No log | 15.0 | 345 | 0.5700 | 29.6188 | 18.6864 |
|
68 |
+
| No log | 16.0 | 368 | 0.5620 | 29.9895 | 18.7278 |
|
69 |
+
| No log | 17.0 | 391 | 0.5569 | 29.3883 | 18.7278 |
|
70 |
+
| No log | 18.0 | 414 | 0.5486 | 29.783 | 18.6864 |
|
71 |
+
| No log | 19.0 | 437 | 0.5520 | 29.4199 | 18.6568 |
|
72 |
+
| No log | 20.0 | 460 | 0.5496 | 28.8502 | 18.6923 |
|
73 |
+
| No log | 21.0 | 483 | 0.5495 | 29.2369 | 18.7337 |
|
74 |
+
| 0.8287 | 22.0 | 506 | 0.5411 | 29.1707 | 18.6923 |
|
75 |
+
| 0.8287 | 23.0 | 529 | 0.5325 | 28.8466 | 18.6923 |
|
76 |
+
| 0.8287 | 24.0 | 552 | 0.5388 | 29.2703 | 18.6746 |
|
77 |
+
| 0.8287 | 25.0 | 575 | 0.5288 | 28.7683 | 18.7278 |
|
78 |
+
| 0.8287 | 26.0 | 598 | 0.5299 | 28.3489 | 18.7337 |
|
79 |
+
| 0.8287 | 27.0 | 621 | 0.5314 | 28.2042 | 18.7278 |
|
80 |
+
| 0.8287 | 28.0 | 644 | 0.5159 | 29.5603 | 18.7574 |
|
81 |
+
| 0.8287 | 29.0 | 667 | 0.5163 | 28.8959 | 18.6982 |
|
82 |
+
| 0.8287 | 30.0 | 690 | 0.5170 | 30.0671 | 18.7219 |
|
83 |
+
| 0.8287 | 31.0 | 713 | 0.5202 | 28.9559 | 18.6686 |
|
84 |
+
| 0.8287 | 32.0 | 736 | 0.5181 | 28.1951 | 18.7574 |
|
85 |
+
| 0.8287 | 33.0 | 759 | 0.5134 | 28.2097 | 18.7041 |
|
86 |
+
| 0.8287 | 34.0 | 782 | 0.5134 | 28.0953 | 18.7574 |
|
87 |
+
| 0.8287 | 35.0 | 805 | 0.5165 | 28.4171 | 18.7988 |
|
88 |
+
| 0.8287 | 36.0 | 828 | 0.5195 | 28.7953 | 18.7633 |
|
89 |
+
| 0.8287 | 37.0 | 851 | 0.5185 | 29.1606 | 18.7633 |
|
90 |
+
| 0.8287 | 38.0 | 874 | 0.5192 | 29.5226 | 18.6805 |
|
91 |
+
| 0.8287 | 39.0 | 897 | 0.5178 | 28.425 | 18.7692 |
|
92 |
+
| 0.8287 | 40.0 | 920 | 0.5196 | 29.4795 | 18.7041 |
|
93 |
+
| 0.8287 | 41.0 | 943 | 0.5161 | 29.5127 | 18.7041 |
|
94 |
+
| 0.8287 | 42.0 | 966 | 0.5164 | 28.6225 | 18.7574 |
|
95 |
+
| 0.8287 | 43.0 | 989 | 0.5183 | 29.0629 | 18.6272 |
|
96 |
+
| 0.3866 | 44.0 | 1012 | 0.5174 | 28.6628 | 18.6213 |
|
97 |
+
| 0.3866 | 45.0 | 1035 | 0.5141 | 28.499 | 18.6213 |
|
98 |
+
| 0.3866 | 46.0 | 1058 | 0.5151 | 28.1309 | 18.6272 |
|
99 |
+
| 0.3866 | 47.0 | 1081 | 0.5167 | 29.8871 | 18.6391 |
|
100 |
+
| 0.3866 | 48.0 | 1104 | 0.5133 | 28.7513 | 18.6331 |
|
101 |
+
| 0.3866 | 49.0 | 1127 | 0.5188 | 28.3089 | 18.6213 |
|
102 |
+
| 0.3866 | 50.0 | 1150 | 0.5203 | 28.8714 | 18.6331 |
|
103 |
+
| 0.3866 | 51.0 | 1173 | 0.5263 | 28.7644 | 18.6331 |
|
104 |
+
| 0.3866 | 52.0 | 1196 | 0.5222 | 28.4817 | 18.6331 |
|
105 |
+
| 0.3866 | 53.0 | 1219 | 0.5307 | 28.8117 | 18.6272 |
|
106 |
+
| 0.3866 | 54.0 | 1242 | 0.5255 | 29.3844 | 18.6213 |
|
107 |
+
| 0.3866 | 55.0 | 1265 | 0.5264 | 29.7108 | 18.6213 |
|
108 |
+
| 0.3866 | 56.0 | 1288 | 0.5272 | 29.353 | 18.6331 |
|
109 |
+
| 0.3866 | 57.0 | 1311 | 0.5283 | 28.5792 | 18.6391 |
|
110 |
+
| 0.3866 | 58.0 | 1334 | 0.5301 | 29.9914 | 18.6272 |
|
111 |
+
| 0.3866 | 59.0 | 1357 | 0.5320 | 29.3162 | 18.6272 |
|
112 |
+
| 0.3866 | 60.0 | 1380 | 0.5380 | 29.2162 | 18.6272 |
|
113 |
+
| 0.3866 | 61.0 | 1403 | 0.5349 | 28.5292 | 18.6272 |
|
114 |
+
| 0.3866 | 62.0 | 1426 | 0.5313 | 28.7165 | 18.6627 |
|
115 |
+
| 0.3866 | 63.0 | 1449 | 0.5335 | 29.3637 | 18.6154 |
|
116 |
+
| 0.3866 | 64.0 | 1472 | 0.5350 | 29.3612 | 18.568 |
|
117 |
+
| 0.3866 | 65.0 | 1495 | 0.5330 | 29.1338 | 18.5621 |
|
118 |
+
| 0.283 | 66.0 | 1518 | 0.5322 | 29.0514 | 18.5562 |
|
119 |
+
| 0.283 | 67.0 | 1541 | 0.5362 | 29.2592 | 18.5562 |
|
120 |
+
| 0.283 | 68.0 | 1564 | 0.5379 | 29.6757 | 18.568 |
|
121 |
+
| 0.283 | 69.0 | 1587 | 0.5386 | 29.5012 | 18.5976 |
|
122 |
+
| 0.283 | 70.0 | 1610 | 0.5379 | 29.6616 | 18.5917 |
|
123 |
+
| 0.283 | 71.0 | 1633 | 0.5364 | 29.8762 | 18.6154 |
|
124 |
+
| 0.283 | 72.0 | 1656 | 0.5392 | 29.7143 | 18.6036 |
|
125 |
+
| 0.283 | 73.0 | 1679 | 0.5438 | 29.385 | 18.5976 |
|
126 |
+
| 0.283 | 74.0 | 1702 | 0.5386 | 28.3472 | 18.6095 |
|
127 |
+
| 0.283 | 75.0 | 1725 | 0.5372 | 29.1045 | 18.574 |
|
128 |
+
| 0.283 | 76.0 | 1748 | 0.5406 | 29.0839 | 18.6095 |
|
129 |
+
| 0.283 | 77.0 | 1771 | 0.5408 | 29.735 | 18.5799 |
|
130 |
+
| 0.283 | 78.0 | 1794 | 0.5406 | 29.5432 | 18.6036 |
|
131 |
+
| 0.283 | 79.0 | 1817 | 0.5413 | 29.1501 | 18.5976 |
|
132 |
+
| 0.283 | 80.0 | 1840 | 0.5434 | 29.5822 | 18.6095 |
|
133 |
+
| 0.283 | 81.0 | 1863 | 0.5491 | 29.1933 | 18.5799 |
|
134 |
+
| 0.283 | 82.0 | 1886 | 0.5473 | 28.9065 | 18.5385 |
|
135 |
+
| 0.283 | 83.0 | 1909 | 0.5507 | 29.4129 | 18.5385 |
|
136 |
+
| 0.283 | 84.0 | 1932 | 0.5534 | 29.2249 | 18.5385 |
|
137 |
+
| 0.283 | 85.0 | 1955 | 0.5561 | 29.6955 | 18.5799 |
|
138 |
+
| 0.283 | 86.0 | 1978 | 0.5575 | 29.1081 | 18.5385 |
|
139 |
+
| 0.2296 | 87.0 | 2001 | 0.5531 | 29.7633 | 18.5385 |
|
140 |
+
| 0.2296 | 88.0 | 2024 | 0.5548 | 30.045 | 18.5385 |
|
141 |
+
| 0.2296 | 89.0 | 2047 | 0.5567 | 29.9209 | 18.5385 |
|
142 |
+
| 0.2296 | 90.0 | 2070 | 0.5577 | 29.1879 | 18.5858 |
|
143 |
+
| 0.2296 | 91.0 | 2093 | 0.5602 | 29.1587 | 18.5799 |
|
144 |
+
| 0.2296 | 92.0 | 2116 | 0.5595 | 29.5205 | 18.5799 |
|
145 |
+
| 0.2296 | 93.0 | 2139 | 0.5605 | 29.3439 | 18.5325 |
|
146 |
+
| 0.2296 | 94.0 | 2162 | 0.5583 | 29.4742 | 18.5325 |
|
147 |
+
| 0.2296 | 95.0 | 2185 | 0.5576 | 29.132 | 18.5325 |
|
148 |
+
| 0.2296 | 96.0 | 2208 | 0.5566 | 29.0861 | 18.5325 |
|
149 |
+
| 0.2296 | 97.0 | 2231 | 0.5584 | 29.6618 | 18.5385 |
|
150 |
+
| 0.2296 | 98.0 | 2254 | 0.5593 | 29.1068 | 18.5325 |
|
151 |
+
| 0.2296 | 99.0 | 2277 | 0.5603 | 29.7081 | 18.5385 |
|
152 |
+
| 0.2296 | 100.0 | 2300 | 0.5599 | 29.6368 | 18.5325 |
|
153 |
+
| 0.2296 | 101.0 | 2323 | 0.5598 | 29.6263 | 18.5325 |
|
154 |
+
| 0.2296 | 102.0 | 2346 | 0.5637 | 29.6321 | 18.5385 |
|
155 |
+
| 0.2296 | 103.0 | 2369 | 0.5678 | 29.6306 | 18.5266 |
|
156 |
+
| 0.2296 | 104.0 | 2392 | 0.5685 | 29.3279 | 18.5325 |
|
157 |
+
| 0.2296 | 105.0 | 2415 | 0.5680 | 29.1363 | 18.5621 |
|
158 |
+
| 0.2296 | 106.0 | 2438 | 0.5726 | 29.2666 | 18.5385 |
|
159 |
+
| 0.2296 | 107.0 | 2461 | 0.5738 | 29.2981 | 18.5385 |
|
160 |
+
| 0.2296 | 108.0 | 2484 | 0.5740 | 29.5752 | 18.5385 |
|
161 |
+
| 0.1942 | 109.0 | 2507 | 0.5749 | 29.5596 | 18.5385 |
|
162 |
+
| 0.1942 | 110.0 | 2530 | 0.5732 | 29.6728 | 18.574 |
|
163 |
+
| 0.1942 | 111.0 | 2553 | 0.5738 | 29.6052 | 18.5325 |
|
164 |
+
| 0.1942 | 112.0 | 2576 | 0.5731 | 29.5143 | 18.574 |
|
165 |
+
| 0.1942 | 113.0 | 2599 | 0.5744 | 29.8059 | 18.574 |
|
166 |
+
| 0.1942 | 114.0 | 2622 | 0.5751 | 29.6796 | 18.574 |
|
167 |
+
| 0.1942 | 115.0 | 2645 | 0.5763 | 29.9279 | 18.568 |
|
168 |
+
| 0.1942 | 116.0 | 2668 | 0.5746 | 29.892 | 18.568 |
|
169 |
+
| 0.1942 | 117.0 | 2691 | 0.5741 | 29.8104 | 18.574 |
|
170 |
+
| 0.1942 | 118.0 | 2714 | 0.5759 | 29.6379 | 18.574 |
|
171 |
+
| 0.1942 | 119.0 | 2737 | 0.5777 | 29.7949 | 18.574 |
|
172 |
+
| 0.1942 | 120.0 | 2760 | 0.5776 | 29.6297 | 18.574 |
|
173 |
+
| 0.1942 | 121.0 | 2783 | 0.5789 | 29.5298 | 18.574 |
|
174 |
+
| 0.1942 | 122.0 | 2806 | 0.5794 | 29.6102 | 18.574 |
|
175 |
+
| 0.1942 | 123.0 | 2829 | 0.5799 | 29.7981 | 18.574 |
|
176 |
+
| 0.1942 | 124.0 | 2852 | 0.5811 | 30.0894 | 18.574 |
|
177 |
+
| 0.1942 | 125.0 | 2875 | 0.5826 | 29.9849 | 18.574 |
|
178 |
+
| 0.1942 | 126.0 | 2898 | 0.5829 | 29.8349 | 18.574 |
|
179 |
+
| 0.1942 | 127.0 | 2921 | 0.5817 | 29.6295 | 18.574 |
|
180 |
+
| 0.1942 | 128.0 | 2944 | 0.5809 | 29.5264 | 18.568 |
|
181 |
+
| 0.1942 | 129.0 | 2967 | 0.5813 | 29.5858 | 18.568 |
|
182 |
+
| 0.1942 | 130.0 | 2990 | 0.5843 | 29.6556 | 18.568 |
|
183 |
+
| 0.1777 | 131.0 | 3013 | 0.5836 | 30.0165 | 18.568 |
|
184 |
+
| 0.1777 | 132.0 | 3036 | 0.5835 | 29.8399 | 18.568 |
|
185 |
+
| 0.1777 | 133.0 | 3059 | 0.5824 | 29.8065 | 18.568 |
|
186 |
+
| 0.1777 | 134.0 | 3082 | 0.5821 | 29.8948 | 18.574 |
|
187 |
+
| 0.1777 | 135.0 | 3105 | 0.5808 | 29.9342 | 18.574 |
|
188 |
+
| 0.1777 | 136.0 | 3128 | 0.5810 | 29.7556 | 18.574 |
|
189 |
+
| 0.1777 | 137.0 | 3151 | 0.5813 | 30.0425 | 18.568 |
|
190 |
+
| 0.1777 | 138.0 | 3174 | 0.5822 | 30.0719 | 18.568 |
|
191 |
+
| 0.1777 | 139.0 | 3197 | 0.5823 | 30.0719 | 18.568 |
|
192 |
+
| 0.1777 | 140.0 | 3220 | 0.5828 | 30.1124 | 18.568 |
|
193 |
+
| 0.1777 | 141.0 | 3243 | 0.5826 | 30.1451 | 18.568 |
|
194 |
+
| 0.1777 | 142.0 | 3266 | 0.5828 | 30.1451 | 18.568 |
|
195 |
+
| 0.1777 | 143.0 | 3289 | 0.5829 | 30.1451 | 18.568 |
|
196 |
+
| 0.1777 | 144.0 | 3312 | 0.5829 | 30.1451 | 18.568 |
|
197 |
+
| 0.1777 | 145.0 | 3335 | 0.5825 | 30.1378 | 18.568 |
|
198 |
+
| 0.1777 | 146.0 | 3358 | 0.5824 | 30.1378 | 18.568 |
|
199 |
+
| 0.1777 | 147.0 | 3381 | 0.5822 | 30.1378 | 18.568 |
|
200 |
+
| 0.1777 | 148.0 | 3404 | 0.5820 | 30.1378 | 18.568 |
|
201 |
+
| 0.1777 | 149.0 | 3427 | 0.5821 | 30.1378 | 18.568 |
|
202 |
+
| 0.1777 | 150.0 | 3450 | 0.5820 | 30.1378 | 18.568 |
|
203 |
+
|
204 |
+
|
205 |
+
### Framework versions
|
206 |
+
|
207 |
+
- Transformers 4.21.2
|
208 |
+
- Pytorch 1.12.1+cu113
|
209 |
+
- Datasets 2.4.0
|
210 |
+
- Tokenizers 0.12.1
|