wav2vec2-xls-r-1b-scandinavian-E3-100h-30-epochs-20250123

This model is a fine-tuned version of facebook/wav2vec2-xls-r-1b on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2819
  • Wer: 21.6559
  • Cer: 5.6618

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 16
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 32
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 6000
  • num_epochs: 30
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Wer Cer
1.0778 0.6307 1000 0.6403 50.3070 15.8896
0.5579 1.2611 2000 0.3271 33.1503 9.3525
0.4943 1.8918 3000 0.2798 29.3342 8.2452
0.364 2.5222 4000 0.2706 29.1800 8.1400
0.3936 3.1526 5000 0.2704 29.8631 8.2974
0.3917 3.7833 6000 0.2831 30.0634 8.6321
0.401 4.4137 7000 0.2827 30.6442 8.7321
0.3556 5.0442 8000 0.2810 29.5504 8.3789
0.3758 5.6749 9000 0.2649 29.0661 8.2426
0.2998 6.3053 10000 0.2632 29.1123 8.1817
0.309 6.9360 11000 0.2585 28.3773 7.9461
0.327 7.5664 12000 0.2638 29.0359 8.1756
0.3288 8.1968 13000 0.2584 28.9768 8.0346
0.3118 8.8275 14000 0.2536 28.0905 7.8639
0.2333 9.4579 15000 0.2543 27.0745 7.5489
0.2406 10.0883 16000 0.2584 27.0399 7.6065
0.2323 10.7190 17000 0.2525 26.9203 7.4725
0.2726 11.3494 18000 0.2438 26.8165 7.4168
0.2361 11.9801 19000 0.2442 26.4202 7.3170
0.2217 12.6105 20000 0.2416 25.8078 7.0961
0.2006 13.2409 21000 0.2481 25.5728 7.0706
0.1744 13.8716 22000 0.2375 25.5901 6.9779
0.1803 14.5020 23000 0.2407 25.4806 6.9786
0.1835 15.1325 24000 0.2390 25.3941 6.9559
0.1835 15.7632 25000 0.2398 24.9474 6.8212
0.1356 16.3936 26000 0.2371 24.1332 6.5554
0.1275 17.0240 27000 0.2414 24.2600 6.5575
0.1354 17.6547 28000 0.2421 23.9026 6.4959
0.1377 18.2851 29000 0.2541 23.6561 6.3947
0.1245 18.9158 30000 0.2326 23.4256 6.3225
0.1146 19.5462 31000 0.2477 22.9918 6.1989
0.0842 20.1766 32000 0.2480 22.9082 6.1661
0.0822 20.8073 33000 0.2508 22.6992 6.0991
0.0813 21.4377 34000 0.2546 22.6776 6.0490
0.0876 22.0681 35000 0.2610 22.4600 5.9675
0.0906 22.6988 36000 0.2623 22.3274 5.9101
0.0755 23.3292 37000 0.2632 22.3750 5.8953
0.0734 23.9599 38000 0.2637 22.0378 5.8312
0.0665 24.5904 39000 0.2690 21.9455 5.7953
0.0628 25.2208 40000 0.2816 21.8274 5.7466
0.068 25.8515 41000 0.2767 21.8245 5.7178
0.0637 26.4819 42000 0.2801 21.7755 5.7103
0.0535 27.1123 43000 0.2787 21.7423 5.6853
0.0523 27.7430 44000 0.2833 21.6746 5.6771
0.0493 28.3734 45000 0.2822 21.6587 5.6658
0.0576 29.0038 46000 0.2805 21.6414 5.6632
0.0559 29.6345 47000 0.2819 21.6559 5.6618

Framework versions

  • Transformers 4.48.1
  • Pytorch 2.5.1+cu124
  • Datasets 3.2.0
  • Tokenizers 0.21.0
Downloads last month
7
Safetensors
Model size
963M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for davidilag/wav2vec2-xls-r-1b-scandinavian-E3-100h-30-epochs-20250123

Finetuned
(105)
this model
Finetunes
1 model