Weyaxi commited on
Commit
64021ad
·
verified ·
1 Parent(s): 78fa413

model card

Browse files
Files changed (1) hide show
  1. README.md +106 -47
README.md CHANGED
@@ -1,18 +1,65 @@
1
  ---
2
- license: mit
3
- base_model: microsoft/phi-2
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
7
- model-index:
8
- - name: Einstein-v4-phi2
9
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
14
 
15
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
16
  <details><summary>See axolotl config</summary>
17
 
18
  axolotl version: `0.4.0`
@@ -112,62 +159,74 @@ special_tokens:
112
  pad_token: "<|endoftext|>"
113
  tokens:
114
  - "<|im_start|>"
 
 
 
 
115
 
 
116
 
 
 
 
 
 
 
 
 
 
117
  ```
118
 
119
- </details><br>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
120
 
121
- # Einstein-v4-phi2
122
 
123
- This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on the None dataset.
124
- It achieves the following results on the evaluation set:
125
- - Loss: 0.7362
126
 
127
- ## Model description
128
 
129
- More information needed
130
 
131
- ## Intended uses & limitations
132
 
133
- More information needed
134
 
135
- ## Training and evaluation data
136
 
137
- More information needed
138
 
139
- ## Training procedure
140
 
141
- ### Training hyperparameters
142
 
143
- The following hyperparameters were used during training:
144
- - learning_rate: 5e-06
145
- - train_batch_size: 3
146
- - eval_batch_size: 3
147
- - seed: 42
148
- - distributed_type: multi-GPU
149
- - num_devices: 9
150
- - gradient_accumulation_steps: 4
151
- - total_train_batch_size: 108
152
- - total_eval_batch_size: 27
153
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
154
- - lr_scheduler_type: cosine
155
- - lr_scheduler_warmup_steps: 10
156
- - num_epochs: 2
157
 
158
- ### Training results
159
 
160
- | Training Loss | Epoch | Step | Validation Loss |
161
- |:-------------:|:-----:|:----:|:---------------:|
162
- | 1.1232 | 0.0 | 1 | 1.0997 |
163
- | 0.7597 | 0.5 | 545 | 0.7720 |
164
- | 0.7232 | 1.0 | 1090 | 0.7451 |
165
- | 0.7806 | 1.48 | 1635 | 0.7362 |
166
 
 
 
 
167
 
168
- ### Framework versions
169
 
170
- - Transformers 4.38.2
171
- - Pytorch 2.1.2+cu118
172
- - Datasets 2.18.0
173
- - Tokenizers 0.15.0
 
1
  ---
2
+ license: other
 
3
  tags:
4
  - axolotl
5
  - generated_from_trainer
6
+ - phi
7
+ - phi2
8
+ - einstein
9
+ - instruct
10
+ - finetune
11
+ - chatml
12
+ - gpt4
13
+ - synthetic data
14
+ - science
15
+ - physics
16
+ - chemistry
17
+ - biology
18
+ - math
19
+ base_model: mistralai/Mistral-7B-v0.1
20
+ datasets:
21
+ - allenai/ai2_arc
22
+ - camel-ai/physics
23
+ - camel-ai/chemistry
24
+ - camel-ai/biology
25
+ - camel-ai/math
26
+ - metaeval/reclor
27
+ - openbookqa
28
+ - mandyyyyii/scibench
29
+ - derek-thomas/ScienceQA
30
+ - TIGER-Lab/ScienceEval
31
+ - jondurbin/airoboros-3.2
32
+ - LDJnr/Capybara
33
+ - Cot-Alpaca-GPT4-From-OpenHermes-2.5
34
+ - STEM-AI-mtl/Electrical-engineering
35
+ - knowrohit07/saraswati-stem
36
+ - sablo/oasst2_curated
37
+ - glaiveai/glaive-code-assistant
38
+ - lmsys/lmsys-chat-1m
39
+ - TIGER-Lab/MathInstruct
40
+ - bigbio/med_qa
41
+ - meta-math/MetaMathQA-40K
42
+ - openbookqa
43
+ - piqa
44
+ - metaeval/reclor
45
+ - derek-thomas/ScienceQA
46
+ - scibench
47
+ - sciq
48
+ - Open-Orca/SlimOrca
49
+ - migtissera/Synthia-v1.3
50
+ - TIGER-Lab/ScienceEval
51
+ language:
52
+ - en
53
  ---
54
 
55
+ # 🔬 Einstein-v4-phi2
56
+
57
+ This model is a full fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on diverse datasets.
58
+
59
+ This model is finetuned using `7xRTX3090` + `1xRTXA6000` using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl).
60
+
61
+ This model's training was sponsored by [sablo.ai](https://sablo.ai).
62
 
 
63
  <details><summary>See axolotl config</summary>
64
 
65
  axolotl version: `0.4.0`
 
159
  pad_token: "<|endoftext|>"
160
  tokens:
161
  - "<|im_start|>"
162
+ ```
163
+ </details><br>
164
+
165
+ # 💬 Prompt Template
166
 
167
+ You can use this prompt template while using the model:
168
 
169
+ ### ChatML
170
+
171
+ ```
172
+ <|im_start|>system
173
+ {system}<|im_end|>
174
+ <|im_start|>user
175
+ {user}<|im_end|>
176
+ <|im_start|>assistant
177
+ {asistant}<|im_end|>
178
  ```
179
 
180
+ This prompt template is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating), which means you can format messages using the
181
+ `tokenizer.apply_chat_template()` method:
182
+
183
+ ```python
184
+ messages = [
185
+ {"role": "system", "content": "You are helpful AI asistant."},
186
+ {"role": "user", "content": "Hello!"}
187
+ ]
188
+ gen_input = tokenizer.apply_chat_template(message, return_tensors="pt")
189
+ model.generate(**gen_input)
190
+ ```
191
+
192
+ # 🔄 Quantizationed versions
193
+
194
+ Quantizationed versions of this model is available.
195
 
196
+ ## GGUF [@bartowski](https://hf.co/bartowski):
197
 
198
+ - https://huggingface.co/bartowski/Einstein-v4-phi2-GGUF
 
 
199
 
200
+ ## Exl2 [@bartowski](https://hf.co/bartowski):
201
 
202
+ - https://huggingface.co/bartowski/Einstein-v4-phi2-exl2
203
 
204
+ # 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
205
 
206
+ # 🤖 Additional information about training
207
 
208
+ This model is full fine-tuned for 2 epochs.
209
 
210
+ Total number of steps was 2178.
211
 
212
+ <details><summary>Loss graph</summary>
213
 
214
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/qsoXp0z2AooZjij95lpRU.png)
215
 
216
+ </details><br>
217
+
218
+ # 🤝 Acknowledgments
219
+
220
+ Thanks to [sablo.ai](https://sablo.ai) for sponsoring this model.
 
 
 
 
 
 
 
 
 
221
 
222
+ Thanks to all the dataset authors mentioned in the datasets section.
223
 
224
+ Thanks to [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) for making the repository I used to make this model.
 
 
 
 
 
225
 
226
+ Thanks to all open source AI community.
227
+
228
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
229
 
230
+ If you would like to support me:
231
 
232
+ [☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)