|
--- |
|
license: apache-2.0 |
|
--- |
|
# Whisper Small - Swedish |
|
This is a fine-tuned version of the [openai/whisper-small](https://huggingface.co/openai/whisper-small) model on the Common Voice 11.0 dataset. |
|
|
|
The following results were achieved after training for 4000 optimization steps: |
|
|
|
* Training Loss: 0.003900 |
|
* Validation Loss: 0.326255 |
|
* WER: 19.894598 |
|
|
|
# Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
|
|
* train_batch_size: 16 |
|
* gradient_accumulation_steps: 1 |
|
* learning_rate: 1e-5 |
|
* eval_batch_size: 8 |
|
* max_steps: 4000 |
|
* eval_steps: 1000 |
|
|
|
# Framework Versions |
|
|
|
* Transformers 4.25.0 |
|
* Pytorch 1.12.1 |
|
* Datasets 2.7.1 |
|
|