Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,95 @@
|
|
1 |
-
---
|
2 |
-
license: llama3
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: llama3
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
library_name: transformers
|
6 |
+
pipeline_tag: text-generation
|
7 |
+
datasets:
|
8 |
+
- cognitivecomputations/WizardLM_alpaca_evol_instruct_70k_unfiltered
|
9 |
+
tags:
|
10 |
+
- finetune
|
11 |
+
- llama
|
12 |
+
---
|
13 |
+
# WizardHermes-2-Pro-Llama-3-8b
|
14 |
+
Finetuned from Meta-Llama-3-8b, 0.55 epoches
|
15 |
+
|
16 |
+
Merged with Hermes-2-Pro-Llama-3-8b
|
17 |
+
|
18 |
+
Eval (0.0 temp) ([https://inflection.ai/assets/MMLU-Examples.pdf](https://inflection.ai/assets/MMLU-Examples.pdf)):
|
19 |
+
```
|
20 |
+
Hermes 2 Pro LoRA (merged) (checkpoint 1): 7/12
|
21 |
+
Hermes 2 Pro LoRA (merged) (checkpoint 2): 9/12
|
22 |
+
Hermes 2 Pro Llama 3 8b: 8/12
|
23 |
+
```
|
24 |
+
|
25 |
+
Trained using MonsterAPI
|
26 |
+
|
27 |
+
LoRA training parameters:
|
28 |
+
```
|
29 |
+
{
|
30 |
+
"deployment_name": "wizardlm-llama-3-8b",
|
31 |
+
"pretrainedmodel_config": {
|
32 |
+
"model_path": "meta-llama/Meta-Llama-3-8B",
|
33 |
+
"other_model_info": null,
|
34 |
+
"resume_checkpoint_path": "",
|
35 |
+
"use_lora": true,
|
36 |
+
"lora_r": 8,
|
37 |
+
"lora_alpha": 16,
|
38 |
+
"lora_dropout": 0,
|
39 |
+
"lora_bias": "none",
|
40 |
+
"use_quantization": false,
|
41 |
+
"use_gradient_checkpointing": false,
|
42 |
+
"parallelization": "nmp"
|
43 |
+
},
|
44 |
+
"data_config": {
|
45 |
+
"data_path": "cognitivecomputations/WizardLM_alpaca_evol_instruct_70k_unfiltered",
|
46 |
+
"data_subset": "default",
|
47 |
+
"data_source_type": "hub_link",
|
48 |
+
"cutoff_len": 4096,
|
49 |
+
"data_split_config": {
|
50 |
+
"train": 0.9,
|
51 |
+
"validation": 0.1
|
52 |
+
},
|
53 |
+
"prevalidated": true,
|
54 |
+
"concat_config": {
|
55 |
+
"0": {
|
56 |
+
"text": "Below is an instruction that describes a task. Write a response that appropriately completes the request. \n\n\n###Instruction:",
|
57 |
+
"column": false
|
58 |
+
},
|
59 |
+
"1": {
|
60 |
+
"text": "instruction",
|
61 |
+
"column": true
|
62 |
+
},
|
63 |
+
"2": {
|
64 |
+
"text": "\n\n\n###Response:",
|
65 |
+
"column": false
|
66 |
+
},
|
67 |
+
"3": {
|
68 |
+
"text": "output",
|
69 |
+
"column": true
|
70 |
+
}
|
71 |
+
}
|
72 |
+
},
|
73 |
+
"training_config": {
|
74 |
+
"early_stopping_patience": 5,
|
75 |
+
"num_train_epochs": 5,
|
76 |
+
"gradient_accumulation_steps": 1,
|
77 |
+
"warmup_steps": 300,
|
78 |
+
"learning_rate": 0.0005,
|
79 |
+
"lr_scheduler_type": "reduce_lr_on_plateau",
|
80 |
+
"group_by_length": false,
|
81 |
+
"use_hugging_face": false
|
82 |
+
},
|
83 |
+
"logging_config": {
|
84 |
+
"use_wandb": false,
|
85 |
+
"wandb_username": "",
|
86 |
+
"wandb_login_key": "",
|
87 |
+
"wandb_project": "",
|
88 |
+
"wandb_run_name": ""
|
89 |
+
},
|
90 |
+
"accessorytasks_config": {
|
91 |
+
"run_eval_report": false,
|
92 |
+
"run_quantize_merge": false
|
93 |
+
}
|
94 |
+
}
|
95 |
+
```
|