barbaroo/nllb_200_1.3B_en_fo

Model Description

  • Model Architecture: This model is based on the NLLB 1.3B architecture and weights.
  • Languages: This checkpoint is fine-tuned to translate from English (en) to Faroese (fo).
  • Size: ~1.3B parameters.
  • Finetuning Datasets:
  • License: Inherits the original licenses of the NLLB 1.3B model.

Intended Use

  • Primary Use Case: Translate text from English to Faroese.
  • Audience: Researchers, developers, or anyone interested in Faroese language processing.
  • Usage Scenarios:
    • Building Faroese-English translation tools
    • Language research and corpus analysis
    • Synthetic data creation

Important: While the model can produce fluent translations, it is not guaranteed to be perfectly accurate on all inputs. Users should verify critical or sensitive content through human experts.

Metrics

  • Model performance measures:
    This model was evaluated using BLEU, chrF and BERT-score —metrics widely adopted by the machine translation community. Additionally, human evaluation was performed by two human experts using the ESA framework on a small dataset (about 200 sentences) of English sentences from news outlets (BBC, CNN, Al Jazeera).

Evaluation Data

  • Datasets:
    Flores-200 dataset is described in Section 4 of the NLLB paper/documentation.

  • Motivation:
    Flores-200 is currently the only machine translation benchmark available for Faroese.

    How to Use

Below is a simple usage example in Python with Hugging Face Transformers:

from transformers import pipeline

model_name = "barbaroo/nllb_200_600M_en_fo"

translator = pipeline(
    "translation",
    model=model_name,
    tokenizer=model_name,
    src_lang="eng_Latn",   # Language code for English
    tgt_lang="fao_Latn"    # Language code for Faroese
)

text = "Hello, how are you?"
translation = translator(text)
print(translation)

Citation

If you use this model or find it helpful in your research, please cite: [COMING SOON]

Contact

For questions, feedback, or collaboration inquiries, feel free to reach out:

Downloads last month
28
Safetensors
Model size
1.37B params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for barbaroo/nllb_200_1.3B_en_fo

Finetuned
(6)
this model

Datasets used to train barbaroo/nllb_200_1.3B_en_fo

Space using barbaroo/nllb_200_1.3B_en_fo 1