rifkat's picture
Update README.md
5d7acfa
|
raw
history blame
1.12 kB
metadata
language:
  - uz
tags:
  - transformers
  - mit
  - robert
  - uzrobert
  - uzbek
  - cyrillic
  - latin
license: apache-2.0
widget:
  - text: Kuchli yomg‘irlar tufayli bir qator <mask> kuchli sel oqishi kuzatildi.
    example_title: Latin script
  - text: >-
      Алишер Навоий – улуғ ўзбек ва бошқа туркий халқларнинг <mask>, мутафаккири
      ва давлат арбоби бўлган.
    example_title: Cyrillic script

UzRoBerta model.

Pre-prepared model in Uzbek (Cyrillic and latin script) to model the masked language and predict the next sentences.

How to use. You can use this model directly with a pipeline for masked language modeling: fromtransformersimportpipelineunmasker=pipeline(fillmask,model=coppercitylabs/uzbertbaseuncased)unmasker("АлишерНавоий–улуғўзбеквабошқатуркийхалқларнинг[MASK],мутафаккиривадавлатарбобибўлган.") from transformers import pipeline unmasker = pipeline('fill-mask', model='coppercitylabs/uzbert-base-uncased') unmasker("Алишер Навоий – улуғ ўзбек ва бошқа туркий халқларнинг [MASK], мутафаккири ва давлат арбоби бўлган.")

Training data.

UzBERT model was pretrained on ≈2M news articles (≈3Gb).