Commit
·
9e267cf
1
Parent(s):
6674f66
Update README.md
Browse files
README.md
CHANGED
@@ -16,19 +16,22 @@ tags:
|
|
16 |
|
17 |
Model Card for Loquace-12B
|
18 |
|
19 |
-
# Loquace
|
20 |
|
21 |
## Model Description
|
22 |
|
23 |
-
Loquace-12B is
|
|
|
24 |
|
25 |
## Usage
|
26 |
|
27 |
-
## Training Data
|
28 |
|
29 |
-
|
30 |
-
|
31 |
-
|
|
|
|
|
|
|
32 |
|
33 |
## Limitations
|
34 |
|
|
|
16 |
|
17 |
Model Card for Loquace-12B
|
18 |
|
19 |
+
# Loquace 🇮🇹 An Italian instruction finetuned LargeLanguage model. 🇮🇹
|
20 |
|
21 |
## Model Description
|
22 |
|
23 |
+
Loquace-12B is the first 12B italian Large Language Model trained using QLoRa on a large dataset of 102k question/answer pairs
|
24 |
+
exclusively in Italian.
|
25 |
|
26 |
## Usage
|
27 |
|
|
|
28 |
|
29 |
+
|
30 |
+
## Training
|
31 |
+
|
32 |
+
Loquace-12B was trained on a conversational dataset comprising 102k question/answer pairs in Italian language.
|
33 |
+
The training data was constructed by putting together translations from the original alpaca Dataset and other sources like the OpenAssistant dataset.
|
34 |
+
The model was trained for only 3000 iterations and took 18 hours on a single RTX 3090, kindly provided by Genesis Cloud.
|
35 |
|
36 |
## Limitations
|
37 |
|