cosimoiaia commited on
Commit
9e267cf
·
1 Parent(s): 6674f66

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -6
README.md CHANGED
@@ -16,19 +16,22 @@ tags:
16
 
17
  Model Card for Loquace-12B
18
 
19
- # Loquace - An Italian instruction finetuned LargeLanguage model.
20
 
21
  ## Model Description
22
 
23
- Loquace-12B is a fine-tuned conversational model for the Italian language. It has been trained on a dataset of 102,000 question/answer examples in the Alpaca style. The model is based on the Falcon-12B architecture and was fine-tuned using the qload framework.
 
24
 
25
  ## Usage
26
 
27
- ## Training Data
28
 
29
- Loquace-12B was trained on a conversational dataset comprising 102,000 question/answer pairs in Italian.
30
- The training data was formatted in the Alpaca style, which emphasizes conversational exchanges.
31
- The specific sources and characteristics of the training data are not disclosed.
 
 
 
32
 
33
  ## Limitations
34
 
 
16
 
17
  Model Card for Loquace-12B
18
 
19
+ # Loquace 🇮🇹 An Italian instruction finetuned LargeLanguage model. 🇮🇹
20
 
21
  ## Model Description
22
 
23
+ Loquace-12B is the first 12B italian Large Language Model trained using QLoRa on a large dataset of 102k question/answer pairs
24
+ exclusively in Italian.
25
 
26
  ## Usage
27
 
 
28
 
29
+
30
+ ## Training
31
+
32
+ Loquace-12B was trained on a conversational dataset comprising 102k question/answer pairs in Italian language.
33
+ The training data was constructed by putting together translations from the original alpaca Dataset and other sources like the OpenAssistant dataset.
34
+ The model was trained for only 3000 iterations and took 18 hours on a single RTX 3090, kindly provided by Genesis Cloud.
35
 
36
  ## Limitations
37