File size: 1,115 Bytes
d20e097 bdd3fe4 d20e097 bdd3fe4 d20e097 98f7282 d20e097 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 |
---
language_details: spa_Latn, quy_Latn, nso_Latn, fra_Latn, bos_Latn
pipeline_tag: translation
tags:
- mt5
license: apache-2.0
inference: false
---
# Description
Finetuned [google/mt5-large](https://huggingface.co/google/mt5-large) model to translate between Spanish ("spa_Latn") and Mapuzungún.
We support the following languages/graphemaries:
- Spanish (**spa_Latn**)
- Azümchefe (**quy_Latn**)
- Ragileo (**nso_Latn**)
- Unificado (**fra_Latn**)
# Example
```python
from transformers import T5TokenizerFast, AutoModelForSeq2SeqLM
tokenizer = T5TokenizerFast.from_pretrained("CenIA/mt5-large-spa-rap")
model = AutoModelForSeq2SeqLM.from_pretrained("CenIA/mt5-large-spa-rap")
def translate(sentence: str, translate_from="spa_Latn", translate_to="mri_Latn") -> str:
inputs = tokenizer(translate_from + sentence, return_tensors="pt")
result = model.generate(**inputs, forced_bos_token_id=tokenizer.convert_tokens_to_ids(translate_to))
decoded = tokenizer.batch_decode(result, skip_special_tokens=True)[0]
return decoded
traduction = translate("Hola, ¿cómo estás?")
print(traduction)
``` |