Update README.md
Browse files
README.md
CHANGED
@@ -72,7 +72,7 @@ widget:
|
|
72 |
|
73 |
## Model description
|
74 |
|
75 |
-
The **roberta-large-bne-massive** is a Intent Classification model for the
|
76 |
|
77 |
## Intended uses and limitations
|
78 |
|
@@ -99,7 +99,7 @@ At the time of submission, no measures have been taken to estimate the bias embe
|
|
99 |
## Training
|
100 |
|
101 |
### Training data
|
102 |
-
We used the
|
103 |
|
104 |
### Training procedure
|
105 |
The model was trained with a batch size of 16 and a learning rate of 1e-5 for 20 epochs. We then selected the best checkpoint using the downstream task metric in the corresponding development set and then evaluated it on the test set.
|
|
|
72 |
|
73 |
## Model description
|
74 |
|
75 |
+
The **roberta-large-bne-massive** is a Intent Classification model for the Spanish language fine-tuned from the roberta-large-bne-massive model, a [RoBERTa](https://arxiv.org/abs/1907.11692) based model pre-trained on a medium-size corpus collected from publicly available corpora and crawlers.
|
76 |
|
77 |
## Intended uses and limitations
|
78 |
|
|
|
99 |
## Training
|
100 |
|
101 |
### Training data
|
102 |
+
We used the Spanish split of the [MASSIVE](https://huggingface.co/datasets/AmazonScience/massive) dataset for training and evaluation.
|
103 |
|
104 |
### Training procedure
|
105 |
The model was trained with a batch size of 16 and a learning rate of 1e-5 for 20 epochs. We then selected the best checkpoint using the downstream task metric in the corresponding development set and then evaluated it on the test set.
|