File size: 1,440 Bytes
2398cbf 138a99c 2398cbf 138a99c 2398cbf |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 |
---
license: apache-2.0
base_model: Helsinki/opus-mt-en-mul
datasets:
- ai4bharat/samanantar
- wmt/wmt19
language:
- en
- hi
- gu
metrics:
- bleu
---
# Finetuning
This model is a fine-tuned version of [Varsha00/finetuned-opusmt-en-to-hi](https://huggingface.co/Varsha00/finetuned-opusmt-en-to-hi) on the samanantar & WMT News dataset.
source group: English
target group: Gujarati
model: transformer
## Model description
This model is a sequentially finetuned version of the Helsinki-NLP/opus-mt-en-mul model, designed for translating between English and Gujarati.
The model was initially finetuned on the Hindi language using a substantial dataset and subsequently finetuned on Gujarati using a smaller dataset.
This approach, known as sequential finetuning or cascaded finetuning, allows the model to leverage the knowledge gained from Hindi to improve its
performance on Gujarati translations, despite the limited data available for the latter.
## Training and evaluation data
ai4bharath/samanantar
WMT-News
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-5
- warmup_steps: 500
- gradient_accumulation_steps: 2
- total_train_batch_size: 16
- num_epochs: 10
### Benchamark Evaluation
- BLEU score on Tatoeba: 27.7761903401179
- BLUE score on IN-22: 16.437183600289
### Framework versions
- Transformers 4.42.3
- Pytorch 2.1.2
- Datasets 2.20.0
- Tokenizers 0.19.1 |