NeoChen1024 commited on
Commit
274d766
·
verified ·
1 Parent(s): a92c0d7

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +289 -0
README.md ADDED
@@ -0,0 +1,289 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ - axolotl
6
+ base_model: cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
7
+ datasets:
8
+ - cognitivecomputations/Dolphin-2.9
9
+ - teknium/OpenHermes-2.5
10
+ - m-a-p/CodeFeedback-Filtered-Instruction
11
+ - cognitivecomputations/dolphin-coder
12
+ - cognitivecomputations/samantha-data
13
+ - microsoft/orca-math-word-problems-200k
14
+ - Locutusque/function-calling-chatml
15
+ - internlm/Agent-FLAN
16
+ model-index:
17
+ - name: dolphin-2.9.3-mistral-7B-32k
18
+ results:
19
+ - task:
20
+ type: text-generation
21
+ name: Text Generation
22
+ dataset:
23
+ name: IFEval (0-Shot)
24
+ type: HuggingFaceH4/ifeval
25
+ args:
26
+ num_few_shot: 0
27
+ metrics:
28
+ - type: inst_level_strict_acc and prompt_level_strict_acc
29
+ value: 41.26
30
+ name: strict accuracy
31
+ source:
32
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
33
+ name: Open LLM Leaderboard
34
+ - task:
35
+ type: text-generation
36
+ name: Text Generation
37
+ dataset:
38
+ name: BBH (3-Shot)
39
+ type: BBH
40
+ args:
41
+ num_few_shot: 3
42
+ metrics:
43
+ - type: acc_norm
44
+ value: 26.91
45
+ name: normalized accuracy
46
+ source:
47
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
48
+ name: Open LLM Leaderboard
49
+ - task:
50
+ type: text-generation
51
+ name: Text Generation
52
+ dataset:
53
+ name: MATH Lvl 5 (4-Shot)
54
+ type: hendrycks/competition_math
55
+ args:
56
+ num_few_shot: 4
57
+ metrics:
58
+ - type: exact_match
59
+ value: 4.83
60
+ name: exact match
61
+ source:
62
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
63
+ name: Open LLM Leaderboard
64
+ - task:
65
+ type: text-generation
66
+ name: Text Generation
67
+ dataset:
68
+ name: GPQA (0-shot)
69
+ type: Idavidrein/gpqa
70
+ args:
71
+ num_few_shot: 0
72
+ metrics:
73
+ - type: acc_norm
74
+ value: 4.7
75
+ name: acc_norm
76
+ source:
77
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
78
+ name: Open LLM Leaderboard
79
+ - task:
80
+ type: text-generation
81
+ name: Text Generation
82
+ dataset:
83
+ name: MuSR (0-shot)
84
+ type: TAUR-Lab/MuSR
85
+ args:
86
+ num_few_shot: 0
87
+ metrics:
88
+ - type: acc_norm
89
+ value: 17.93
90
+ name: acc_norm
91
+ source:
92
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
93
+ name: Open LLM Leaderboard
94
+ - task:
95
+ type: text-generation
96
+ name: Text Generation
97
+ dataset:
98
+ name: MMLU-PRO (5-shot)
99
+ type: TIGER-Lab/MMLU-Pro
100
+ config: main
101
+ split: test
102
+ args:
103
+ num_few_shot: 5
104
+ metrics:
105
+ - type: acc
106
+ value: 20.23
107
+ name: accuracy
108
+ source:
109
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
110
+ name: Open LLM Leaderboard
111
+ ---
112
+
113
+ # W8A8 quant of Dolphin 2.9.3 Mistral 7b v0.3 32k 🐬
114
+ Quantization script: <https://github.com/NeoChen1024/scripts/blob/master/llm-compressor-quantize.py>
115
+
116
+ Curated and trained by Eric Hartford and Cognitive Computations
117
+
118
+ [![Discord](https://img.shields.io/discord/1156064224225808488?logo=Discord&logoColor=%23ffffff&label=Discord&link=https%3A%2F%2Fdiscord.gg%2FtCMkMDDHwm)](https://discord.gg/h3K4XGj2RH)
119
+ Discord: https://discord.gg/h3K4XGj2RH
120
+
121
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/63111b2d88942700629f5771/ldkN1J0WIDQwU4vutGYiD.png" width="600" />
122
+
123
+ Our appreciation for the sponsors of Dolphin 2.9.3:
124
+ - [Crusoe Cloud](https://crusoe.ai/) - provided excellent on-demand 8xH100 node
125
+ - [OnDemand](https://on-demand.io/) - provided inference sponsorship
126
+
127
+ This model is based on mistralai/Mistral-7B-v0.3, and is governed by the apache 2.0 license.
128
+
129
+ The base model has 32k context, and our finetuning took place with 8192 sequence length.
130
+
131
+ Dolphin 2.9.3 uses ChatML prompt template format.
132
+
133
+ example:
134
+
135
+ ```
136
+ <|im_start|>system
137
+ You are Dolphin, a helpful AI assistant.<|im_end|>
138
+ <|im_start|>user
139
+ {prompt}<|im_end|>
140
+ <|im_start|>assistant
141
+
142
+ ```
143
+
144
+ Dolphin-2.9.3 has a variety of instruction following, conversational, and coding skills. It also has initial agentic abilities and supports function calling.
145
+
146
+ Dolphin is uncensored. We have filtered the dataset to remove alignment and bias. This makes the model more compliant. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant with any requests, even unethical ones. Please read my blog post about uncensored models. https://erichartford.com/uncensored-models You are responsible for any content you create using this model. Enjoy responsibly.
147
+
148
+ Dolphin is licensed according to apache 2.0 license. We grant permission for any use, including commercial. Dolphin was trained on data generated from GPT4, among other models.
149
+
150
+ ## Evals
151
+
152
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63111b2d88942700629f5771/5KUgfzJyY1IM4Yg6bg3Dq.png)
153
+
154
+ ## Training
155
+
156
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
157
+ <details><summary>See axolotl config</summary>
158
+
159
+ axolotl version: `0.4.0`
160
+ ```yaml
161
+ base_model: mistralai/Mistral-7B-v0.3
162
+ model_type: AutoModelForCausalLM
163
+ tokenizer_type: AutoTokenizer
164
+
165
+ load_in_8bit: false
166
+ # load_in_4bit: true
167
+ strict: false
168
+
169
+ datasets:
170
+ - path: /workspace/datasets/dolphin-2.9.3/dolphin201-sharegpt2.jsonl
171
+ type: sharegpt
172
+ conversation: chatml
173
+ - path: /workspace/datasets/dolphin-2.9.3/SystemChat_filtered_sharegpt.jsonl
174
+ type: sharegpt
175
+ conversation: chatml
176
+ - path: /workspace/datasets/dolphin-2.9.3/SystemChat_multilingual_sharegpt.jsonl
177
+ type: sharegpt
178
+ conversation: chatml
179
+ - path: /workspace/datasets/dolphin-2.9.3/dolphin-coder-translate-sharegpt2.jsonl
180
+ type: sharegpt
181
+ conversation: chatml
182
+ - path: /workspace/datasets/dolphin-2.9.3/dolphin-coder-codegen-sharegpt2.jsonl
183
+ type: sharegpt
184
+ conversation: chatml
185
+ - path: /workspace/datasets/dolphin-2.9.3/m-a-p_Code-Feedback-sharegpt-unfiltered.jsonl
186
+ type: sharegpt
187
+ conversation: chatml
188
+ - path: /workspace/datasets/dolphin-2.9.3/m-a-p_CodeFeedback-Filtered-Instruction-sharegpt-unfiltered.jsonl
189
+ type: sharegpt
190
+ conversation: chatml
191
+ - path: /workspace/datasets/dolphin-2.9.3/not_samantha_norefusals.jsonl
192
+ type: sharegpt
193
+ conversation: chatml
194
+ - path: /workspace/datasets/dolphin-2.9.3/Orca-Math-resort-unfiltered.jsonl
195
+ type: sharegpt
196
+ conversation: chatml
197
+ - path: /workspace/datasets/dolphin-2.9.3/agent_instruct_react_unfiltered.jsonl
198
+ type: sharegpt
199
+ conversation: chatml
200
+ - path: /workspace/datasets/dolphin-2.9.3/toolbench_instruct_j1s1_3k_unfiltered.jsonl
201
+ type: sharegpt
202
+ conversation: chatml
203
+ - path: /workspace/datasets/dolphin-2.9.3/toolbench_negative_unfiltered.jsonl
204
+ type: sharegpt
205
+ conversation: chatml
206
+ - path: /workspace/datasets/dolphin-2.9.3/toolbench_react_10p_unfiltered.jsonl
207
+ type: sharegpt
208
+ conversation: chatml
209
+ - path: /workspace/datasets/dolphin-2.9.3/toolbench_tflan_cot_30p_unfiltered.jsonl
210
+ type: sharegpt
211
+ conversation: chatml
212
+ - path: /workspace/datasets/dolphin-2.9.3/openhermes200k_unfiltered.jsonl
213
+ type: sharegpt
214
+ conversation: chatml
215
+
216
+ chat_template: chatml
217
+ # adapter: qlora
218
+ # lora_r: 128
219
+ # lora_alpha: 16
220
+ # lora_modules_to_save: [embed_tokens, lm_head]
221
+ # lora_dropout: 0.05
222
+ # lora_target_linear: true
223
+
224
+
225
+ dataset_prepared_path: /workspace/axolotl/dolph-2.9.3-prepared
226
+ val_set_size: 0.01
227
+ output_dir: /workspace/axolotl/dolphin-2.9.3-mistral-7B
228
+
229
+ sequence_len: 8192
230
+ sample_packing: true
231
+ pad_to_sequence_len: true
232
+
233
+ wandb_project: dolphin-2.9.3-Mistral-7B
234
+ wandb_watch:
235
+ wandb_run_id:
236
+ wandb_log_model:
237
+
238
+ gradient_accumulation_steps: 16
239
+ micro_batch_size: 1
240
+ num_epochs: 3
241
+ optimizer: adamw_8bit
242
+ lr_scheduler: cosine
243
+ learning_rate: 5e-6
244
+ train_on_inputs: false
245
+ group_by_length: false
246
+ bf16: auto
247
+ fp16:
248
+ tf32:
249
+
250
+ gradient_checkpointing: true
251
+ gradient_checkpointing_kwargs:
252
+ use_reentrant: false
253
+ early_stopping_patience:
254
+ resume_from_checkpoint:
255
+ logging_steps: 1
256
+ xformers_attention:
257
+ flash_attention: true
258
+
259
+ warmup_steps: 100
260
+ # evals_per_epoch: 4
261
+ eval_table_size:
262
+ saves_per_epoch: 1
263
+ save_total_limit: 2
264
+ save_steps:
265
+ debug:
266
+ deepspeed: /workspace/axolotl/deepspeed_configs/zero3_bf16.json
267
+ weight_decay: 0.1
268
+ fsdp:
269
+ fsdp_config:
270
+ special_tokens:
271
+ eos_token: "<|im_end|>"
272
+ tokens:
273
+ - "<|im_start|>"
274
+
275
+ ```
276
+
277
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
278
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_cognitivecomputations__dolphin-2.9.3-mistral-7B-32k)
279
+
280
+ | Metric |Value|
281
+ |-------------------|----:|
282
+ |Avg. |19.31|
283
+ |IFEval (0-Shot) |41.26|
284
+ |BBH (3-Shot) |26.91|
285
+ |MATH Lvl 5 (4-Shot)| 4.83|
286
+ |GPQA (0-shot) | 4.70|
287
+ |MuSR (0-shot) |17.93|
288
+ |MMLU-PRO (5-shot) |20.23|
289
+