RichardErkhov commited on
Commit
98d68e2
·
verified ·
1 Parent(s): 7d24e83

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +211 -0
README.md ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ multimaster-7b - GGUF
11
+ - Model creator: https://huggingface.co/ibivibiv/
12
+ - Original model: https://huggingface.co/ibivibiv/multimaster-7b/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [multimaster-7b.Q2_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q2_K.gguf) | Q2_K | 2.53GB |
18
+ | [multimaster-7b.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.IQ3_XS.gguf) | IQ3_XS | 2.81GB |
19
+ | [multimaster-7b.IQ3_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.IQ3_S.gguf) | IQ3_S | 2.96GB |
20
+ | [multimaster-7b.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q3_K_S.gguf) | Q3_K_S | 2.95GB |
21
+ | [multimaster-7b.IQ3_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.IQ3_M.gguf) | IQ3_M | 3.06GB |
22
+ | [multimaster-7b.Q3_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q3_K.gguf) | Q3_K | 3.28GB |
23
+ | [multimaster-7b.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q3_K_M.gguf) | Q3_K_M | 3.28GB |
24
+ | [multimaster-7b.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q3_K_L.gguf) | Q3_K_L | 3.56GB |
25
+ | [multimaster-7b.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.IQ4_XS.gguf) | IQ4_XS | 3.67GB |
26
+ | [multimaster-7b.Q4_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q4_0.gguf) | Q4_0 | 3.83GB |
27
+ | [multimaster-7b.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.IQ4_NL.gguf) | IQ4_NL | 3.87GB |
28
+ | [multimaster-7b.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q4_K_S.gguf) | Q4_K_S | 3.86GB |
29
+ | [multimaster-7b.Q4_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q4_K.gguf) | Q4_K | 4.07GB |
30
+ | [multimaster-7b.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q4_K_M.gguf) | Q4_K_M | 4.07GB |
31
+ | [multimaster-7b.Q4_1.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q4_1.gguf) | Q4_1 | 4.24GB |
32
+ | [multimaster-7b.Q5_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q5_0.gguf) | Q5_0 | 4.65GB |
33
+ | [multimaster-7b.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q5_K_S.gguf) | Q5_K_S | 4.65GB |
34
+ | [multimaster-7b.Q5_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q5_K.gguf) | Q5_K | 4.78GB |
35
+ | [multimaster-7b.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q5_K_M.gguf) | Q5_K_M | 4.78GB |
36
+ | [multimaster-7b.Q5_1.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q5_1.gguf) | Q5_1 | 5.07GB |
37
+ | [multimaster-7b.Q6_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q6_K.gguf) | Q6_K | 5.53GB |
38
+ | [multimaster-7b.Q8_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_multimaster-7b-gguf/blob/main/multimaster-7b.Q8_0.gguf) | Q8_0 | 7.17GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ license: apache-2.0
46
+ language:
47
+ - en
48
+ library_name: transformers
49
+ tags:
50
+ - moe
51
+ - moerge
52
+ ---
53
+
54
+ # Multi Master 7B
55
+
56
+ ![img](./multimaster.png)
57
+
58
+ A quick multi-disciplinary fine tune of openchat/openchat-3.5-0106 using an alpaca-style dataset across different disciplines. I used LORA adapters that I then merged back into the main model for ease of use.
59
+
60
+ # Prompting
61
+
62
+ ## Prompt Template for alpaca style
63
+
64
+ ```
65
+ ### Instruction:
66
+
67
+ <prompt> (without the <>)
68
+
69
+ ### Response:
70
+ ```
71
+
72
+ ## Sample Code
73
+
74
+ ```python
75
+ import torch
76
+ from transformers import AutoModelForCausalLM, AutoTokenizer
77
+
78
+ torch.set_default_device("cuda")
79
+
80
+ model = AutoModelForCausalLM.from_pretrained("ibivibiv/multimaster-7b", torch_dtype="auto", device_config='auto')
81
+ tokenizer = AutoTokenizer.from_pretrained("ibivibiv/multimaster-7b")
82
+
83
+ inputs = tokenizer("### Instruction: Who would when in an arm wrestling match between Abraham Lincoln and Chuck Norris?\nA. Abraham Lincoln \nB. Chuck Norris\n### Response:\n", return_tensors="pt", return_attention_mask=False)
84
+
85
+ outputs = model.generate(**inputs, max_length=200)
86
+ text = tokenizer.batch_decode(outputs)[0]
87
+ print(text)
88
+ ```
89
+
90
+ # Model Details
91
+ * **Trained by**: [ibivibiv](https://huggingface.co/ibivibiv)
92
+ * **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
93
+ * **Model type:** **multimaster-7b** is a lora tuned version of openchat/openchat-3.5-0106 with the adapter merged back into the main model
94
+ * **Language(s)**: English
95
+ * **Purpose**: This model is a focus on multi-disciplinary model tuning
96
+
97
+ # Benchmark Scores
98
+
99
+ coming soon
100
+
101
+ ## Citations
102
+
103
+ ```
104
+ @misc{open-llm-leaderboard,
105
+ author = {Edward Beeching and Clémentine Fourrier and Nathan Habib and Sheon Han and Nathan Lambert and Nazneen Rajani and Omar Sanseviero and Lewis Tunstall and Thomas Wolf},
106
+ title = {Open LLM Leaderboard},
107
+ year = {2023},
108
+ publisher = {Hugging Face},
109
+ howpublished = "\url{https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard}"
110
+ }
111
+ ```
112
+ ```
113
+ @software{eval-harness,
114
+ author = {Gao, Leo and
115
+ Tow, Jonathan and
116
+ Biderman, Stella and
117
+ Black, Sid and
118
+ DiPofi, Anthony and
119
+ Foster, Charles and
120
+ Golding, Laurence and
121
+ Hsu, Jeffrey and
122
+ McDonell, Kyle and
123
+ Muennighoff, Niklas and
124
+ Phang, Jason and
125
+ Reynolds, Laria and
126
+ Tang, Eric and
127
+ Thite, Anish and
128
+ Wang, Ben and
129
+ Wang, Kevin and
130
+ Zou, Andy},
131
+ title = {A framework for few-shot language model evaluation},
132
+ month = sep,
133
+ year = 2021,
134
+ publisher = {Zenodo},
135
+ version = {v0.0.1},
136
+ doi = {10.5281/zenodo.5371628},
137
+ url = {https://doi.org/10.5281/zenodo.5371628}
138
+ }
139
+ ```
140
+ ```
141
+ @misc{clark2018think,
142
+ title={Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge},
143
+ author={Peter Clark and Isaac Cowhey and Oren Etzioni and Tushar Khot and Ashish Sabharwal and Carissa Schoenick and Oyvind Tafjord},
144
+ year={2018},
145
+ eprint={1803.05457},
146
+ archivePrefix={arXiv},
147
+ primaryClass={cs.AI}
148
+ }
149
+ ```
150
+ ```
151
+ @misc{zellers2019hellaswag,
152
+ title={HellaSwag: Can a Machine Really Finish Your Sentence?},
153
+ author={Rowan Zellers and Ari Holtzman and Yonatan Bisk and Ali Farhadi and Yejin Choi},
154
+ year={2019},
155
+ eprint={1905.07830},
156
+ archivePrefix={arXiv},
157
+ primaryClass={cs.CL}
158
+ }
159
+ ```
160
+ ```
161
+ @misc{hendrycks2021measuring,
162
+ title={Measuring Massive Multitask Language Understanding},
163
+ author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},
164
+ year={2021},
165
+ eprint={2009.03300},
166
+ archivePrefix={arXiv},
167
+ primaryClass={cs.CY}
168
+ }
169
+ ```
170
+ ```
171
+ @misc{lin2022truthfulqa,
172
+ title={TruthfulQA: Measuring How Models Mimic Human Falsehoods},
173
+ author={Stephanie Lin and Jacob Hilton and Owain Evans},
174
+ year={2022},
175
+ eprint={2109.07958},
176
+ archivePrefix={arXiv},
177
+ primaryClass={cs.CL}
178
+ }
179
+ ```
180
+ ```
181
+ @misc{DBLP:journals/corr/abs-1907-10641,
182
+ title={{WINOGRANDE:} An Adversarial Winograd Schema Challenge at Scale},
183
+ author={Keisuke Sakaguchi and Ronan Le Bras and Chandra Bhagavatula and Yejin Choi},
184
+ year={2019},
185
+ eprint={1907.10641},
186
+ archivePrefix={arXiv},
187
+ primaryClass={cs.CL}
188
+ }
189
+ ```
190
+ ```
191
+ @misc{DBLP:journals/corr/abs-2110-14168,
192
+ title={Training Verifiers to Solve Math Word Problems},
193
+ author={Karl Cobbe and
194
+ Vineet Kosaraju and
195
+ Mohammad Bavarian and
196
+ Mark Chen and
197
+ Heewoo Jun and
198
+ Lukasz Kaiser and
199
+ Matthias Plappert and
200
+ Jerry Tworek and
201
+ Jacob Hilton and
202
+ Reiichiro Nakano and
203
+ Christopher Hesse and
204
+ John Schulman},
205
+ year={2021},
206
+ eprint={2110.14168},
207
+ archivePrefix={arXiv},
208
+ primaryClass={cs.CL}
209
+ }
210
+ ```
211
+