|
--- |
|
widget: |
|
- text: Em uma bela manhã de |
|
- text: Em uma cidade tão grande como |
|
- text: Maria e Joana são |
|
license: mit |
|
datasets: |
|
- mc4 |
|
language: |
|
- pt |
|
metrics: |
|
- perplexity |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
--- |
|
# OPT-125M finetuned Portuguese |
|
|
|
Fine-tuning the [OPT-125M](https://huggingface.co/facebook/opt-125m) model on a reduced corpus of mc4-Portuguese with approximately 300M tokens. |
|
|
|
###### Hyper-parameters |
|
|
|
|
|
- learning_rate = 5e-5 |
|
- batch_size = 32 |
|
- warmup = 500 |
|
- seq_length = 512 |
|
- num_train_epochs = 2.0 |
|
|
|
With an A100 with 40GB of RAM, the training took around 3 hours |
|
|
|
**Perplexity:** 9.4 |
|
|
|
## Sample Use |
|
|
|
```python |
|
from transformers import pipeline |
|
generator = pipeline('text-generation', model='Mirelle/opt-125M-pt-br-finetuned', max_length=100, do_sample=True) |
|
generator("Em uma bela manhã de") |
|
``` |