mdeberta-semeval25_thresh07_fold5
This model is a fine-tuned version of microsoft/mdeberta-v3-base on the None dataset. It achieves the following results on the evaluation set:
- Loss: 8.8074
- Precision Samples: 0.1339
- Recall Samples: 0.5358
- F1 Samples: 0.1978
- Precision Macro: 0.8220
- Recall Macro: 0.3540
- F1 Macro: 0.2374
- Precision Micro: 0.1213
- Recall Micro: 0.4565
- F1 Micro: 0.1917
- Precision Weighted: 0.5384
- Recall Weighted: 0.4565
- F1 Weighted: 0.1237
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 10
Training results
Training Loss | Epoch | Step | Validation Loss | Precision Samples | Recall Samples | F1 Samples | Precision Macro | Recall Macro | F1 Macro | Precision Micro | Recall Micro | F1 Micro | Precision Weighted | Recall Weighted | F1 Weighted |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
10.6336 | 1.0 | 19 | 9.9911 | 0.3379 | 0.1862 | 0.1862 | 0.9913 | 0.2103 | 0.2039 | 0.2195 | 0.0811 | 0.1184 | 0.9320 | 0.0811 | 0.0309 |
9.4459 | 2.0 | 38 | 9.6833 | 0.1397 | 0.2709 | 0.1733 | 0.9607 | 0.2304 | 0.2085 | 0.1365 | 0.1562 | 0.1457 | 0.8325 | 0.1562 | 0.0460 |
9.4688 | 3.0 | 57 | 9.4995 | 0.1076 | 0.3609 | 0.1535 | 0.9405 | 0.2654 | 0.2126 | 0.1060 | 0.2492 | 0.1487 | 0.7808 | 0.2492 | 0.0574 |
9.0478 | 4.0 | 76 | 9.3397 | 0.1137 | 0.3878 | 0.1664 | 0.9204 | 0.2690 | 0.2156 | 0.1112 | 0.2733 | 0.1581 | 0.7045 | 0.2733 | 0.0727 |
9.0907 | 5.0 | 95 | 9.1679 | 0.1250 | 0.4917 | 0.1880 | 0.8628 | 0.3093 | 0.2304 | 0.1237 | 0.3784 | 0.1864 | 0.5947 | 0.3784 | 0.1100 |
8.688 | 6.0 | 114 | 9.0335 | 0.1401 | 0.5148 | 0.2038 | 0.8527 | 0.3255 | 0.2328 | 0.1291 | 0.4174 | 0.1972 | 0.5824 | 0.4174 | 0.1163 |
9.009 | 7.0 | 133 | 8.9266 | 0.1373 | 0.5275 | 0.1996 | 0.8310 | 0.3441 | 0.2341 | 0.1223 | 0.4444 | 0.1918 | 0.5496 | 0.4444 | 0.1177 |
9.1189 | 8.0 | 152 | 8.8585 | 0.1358 | 0.5207 | 0.1983 | 0.8195 | 0.3317 | 0.2335 | 0.1234 | 0.4264 | 0.1914 | 0.5347 | 0.4264 | 0.1178 |
8.5834 | 9.0 | 171 | 8.8189 | 0.1353 | 0.5529 | 0.2007 | 0.8211 | 0.3597 | 0.2366 | 0.1228 | 0.4685 | 0.1946 | 0.5369 | 0.4685 | 0.1223 |
9.118 | 10.0 | 190 | 8.8074 | 0.1339 | 0.5358 | 0.1978 | 0.8220 | 0.3540 | 0.2374 | 0.1213 | 0.4565 | 0.1917 | 0.5384 | 0.4565 | 0.1237 |
Framework versions
- Transformers 4.46.0
- Pytorch 2.3.1
- Datasets 2.21.0
- Tokenizers 0.20.1
- Downloads last month
- 160
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for g-assismoraes/mdeberta-semeval25_thresh07_fold5
Base model
microsoft/mdeberta-v3-base