stefanoscotta
commited on
Commit
·
ebd6241
1
Parent(s):
366cefc
Update README.md
Browse files
README.md
CHANGED
@@ -24,7 +24,7 @@ An open-source LLaMa2 language model of 7b parameters fine-tuned (using as base
|
|
24 |
This model is a LLM of 7b parameters based on [NousResearch/Nous-Hermes-llama-2-7b](https://huggingface.co/NousResearch/Nous-Hermes-llama-2-7b), a version of [meta-llama/Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b) fine-tuned to follow instructions.
|
25 |
The model was further fine-tuned in order to follow instructions in italian, using [LoRA](https://arxiv.org/abs/2106.09685) approach and a dataset of 120k random pairs of instruction/answer from [raicrits/Orca_ITA_200k](https://huggingface.co/datasets/raicrits/Orca_ITA_200k).
|
26 |
|
27 |
-
This repository contains the model weights merged with the LoRA adapters obtained in the fine-tuning procedure
|
28 |
|
29 |
|
30 |
- **Developed by:** Stefano Scotta ([email protected])
|
@@ -77,7 +77,7 @@ model_name = "raicrits/Hermes7b_ITA_v1"
|
|
77 |
model = LlamaForCausalLM.from_pretrained(
|
78 |
model_name,
|
79 |
device_map="auto",
|
80 |
-
|
81 |
)
|
82 |
|
83 |
tokenizer = AutoTokenizer.from_pretrained("Hermes_ITA_Lora_merged_V2", add_eos_token=False)
|
@@ -113,6 +113,8 @@ The fine-tuning procedure was done using [LoRA](https://arxiv.org/abs/2106.09685
|
|
113 |
|
114 |
- learning_rate=2e-4,
|
115 |
|
|
|
|
|
116 |
|
117 |
**LoRA configuration:**
|
118 |
- r= 8
|
|
|
24 |
This model is a LLM of 7b parameters based on [NousResearch/Nous-Hermes-llama-2-7b](https://huggingface.co/NousResearch/Nous-Hermes-llama-2-7b), a version of [meta-llama/Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b) fine-tuned to follow instructions.
|
25 |
The model was further fine-tuned in order to follow instructions in italian, using [LoRA](https://arxiv.org/abs/2106.09685) approach and a dataset of 120k random pairs of instruction/answer from [raicrits/Orca_ITA_200k](https://huggingface.co/datasets/raicrits/Orca_ITA_200k).
|
26 |
|
27 |
+
This repository contains the model weights merged with the LoRA adapters obtained in the fine-tuning procedure.
|
28 |
|
29 |
|
30 |
- **Developed by:** Stefano Scotta ([email protected])
|
|
|
77 |
model = LlamaForCausalLM.from_pretrained(
|
78 |
model_name,
|
79 |
device_map="auto",
|
80 |
+
torch_dtype=torch.bfloat16
|
81 |
)
|
82 |
|
83 |
tokenizer = AutoTokenizer.from_pretrained("Hermes_ITA_Lora_merged_V2", add_eos_token=False)
|
|
|
113 |
|
114 |
- learning_rate=2e-4,
|
115 |
|
116 |
+
- mixed precision training: float16
|
117 |
+
|
118 |
|
119 |
**LoRA configuration:**
|
120 |
- r= 8
|