sajidof commited on
Commit
a3fa0b0
·
verified ·
1 Parent(s): 46bb439

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,4 @@
1
  ---
2
- license: apache-2.0
3
- base_model: AntonyG/fine-tune-wav2vec2-large-xls-r-1b-sw
4
  tags:
5
  - generated_from_trainer
6
  datasets:
@@ -22,7 +20,7 @@ model-index:
22
  metrics:
23
  - name: Wer
24
  type: wer
25
- value: 0.8007542426147077
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -30,10 +28,10 @@ should probably proofread and complete it, then remove this comment. -->
30
 
31
  # fine-tune-wav2vec2-large-xls-r-300m-xty_224s
32
 
33
- This model is a fine-tuned version of [AntonyG/fine-tune-wav2vec2-large-xls-r-1b-sw](https://huggingface.co/AntonyG/fine-tune-wav2vec2-large-xls-r-1b-sw) on the ml-superb-subset dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 2.2894
36
- - Wer: 0.8008
37
 
38
  ## Model description
39
 
@@ -61,16 +59,20 @@ The following hyperparameters were used during training:
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_steps: 500
64
- - num_epochs: 20
65
  - mixed_precision_training: Native AMP
66
 
67
  ### Training results
68
 
69
  | Training Loss | Epoch | Step | Validation Loss | Wer |
70
  |:-------------:|:-------:|:----:|:---------------:|:------:|
71
- | 3.356 | 5.5172 | 400 | 1.8297 | 0.9478 |
72
- | 0.9415 | 11.0345 | 800 | 1.7935 | 0.8485 |
73
- | 0.2658 | 16.5517 | 1200 | 2.2894 | 0.8008 |
 
 
 
 
74
 
75
 
76
  ### Framework versions
 
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
4
  datasets:
 
20
  metrics:
21
  - name: Wer
22
  type: wer
23
+ value: 1.0584538026398491
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
28
 
29
  # fine-tune-wav2vec2-large-xls-r-300m-xty_224s
30
 
31
+ This model was trained from scratch on the ml-superb-subset dataset.
32
  It achieves the following results on the evaluation set:
33
+ - Loss: 0.7658
34
+ - Wer: 1.0585
35
 
36
  ## Model description
37
 
 
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
  - lr_scheduler_warmup_steps: 500
62
+ - num_epochs: 40
63
  - mixed_precision_training: Native AMP
64
 
65
  ### Training results
66
 
67
  | Training Loss | Epoch | Step | Validation Loss | Wer |
68
  |:-------------:|:-------:|:----:|:---------------:|:------:|
69
+ | 1.5419 | 5.5172 | 400 | 0.4368 | 0.9994 |
70
+ | 0.423 | 11.0345 | 800 | 0.4315 | 1.0 |
71
+ | 0.3795 | 16.5517 | 1200 | 0.3892 | 1.0151 |
72
+ | 0.3306 | 22.0690 | 1600 | 0.4055 | 1.0013 |
73
+ | 0.2464 | 27.5862 | 2000 | 0.4672 | 1.0421 |
74
+ | 0.1454 | 33.1034 | 2400 | 0.6656 | 1.0333 |
75
+ | 0.0883 | 38.6207 | 2800 | 0.7658 | 1.0585 |
76
 
77
 
78
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa8fb3486f5a7fc2dab93547e4eeb75e84afb30a747eca019a462c6e6e894d81
3
  size 1262012480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c8500b8a8491e31105b8336c55541ab3e0aa2a33064c5a473e597bc8b2855f2
3
  size 1262012480
runs/May22_09-32-03_a5554c14b454/events.out.tfevents.1716370348.a5554c14b454.80691.4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5cfe73e28e8a4010701da591dbff172116974962f918d96537929ab3e98aaace
3
- size 10048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:802502cf8901bc44ca9c5aea7627ce6a768fc0d691ba4931ae711f21f89b7064
3
+ size 10402