rAIfle commited on
Commit
bd22fc1
·
verified ·
1 Parent(s): d332a1c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +79 -0
README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - jondurbin/bagel-dpo-8x7b-v0.2
4
+ - mistralai/Mixtral-8x7B-v0.1
5
+ - Sao10K/Sensualize-Mixtral-bf16
6
+ - mistralai/Mixtral-8x7B-v0.1
7
+ - Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora
8
+ - mistralai/Mixtral-8x7B-Instruct-v0.1
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+
13
+ ---
14
+
15
+ Quantized using 200 samples of 8192 tokens from an RP-oriented [PIPPA](https://huggingface.co/datasets/royallab/PIPPA-cleaned) dataset.
16
+
17
+ Branches:
18
+ - `main` -- `measurement.json`
19
+ - `2.25b6h` -- 2.25bpw, 6bit lm_head
20
+ - `3.5b6h` -- 3.5bpw, 6bit lm_head
21
+ - `3.7b6h` -- 3.7bpw, 6bit lm_head
22
+ - `5b6h` -- 5bpw, 6bit lm_head
23
+ - `6b6h` -- 6bpw, 6bit lm_head
24
+
25
+
26
+ Requires ExllamaV2 version 0.0.12 and up.
27
+
28
+ Original model link: [Envoid/Fish-8x7B](https://huggingface.co/Envoid/Fish-8x7B)
29
+
30
+ Original model README below.
31
+
32
+ ***
33
+ # BagelWorldTour
34
+
35
+ Requested by [kalomaze](https://huggingface.co/kalomaze)
36
+
37
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
38
+
39
+ ## Merge Details
40
+ ### Merge Method
41
+
42
+ This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [mistralai/Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) as a base.
43
+
44
+ ### Models Merged
45
+
46
+ The following models were included in the merge:
47
+ * [jondurbin/bagel-dpo-8x7b-v0.2](https://huggingface.co/jondurbin/bagel-dpo-8x7b-v0.2)
48
+ * [Sao10K/Sensualize-Mixtral-bf16](https://huggingface.co/Sao10K/Sensualize-Mixtral-bf16)
49
+ * [mistralai/Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) + [Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora](https://huggingface.co/Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora)
50
+ * [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1)
51
+
52
+ ### Configuration
53
+
54
+ The following YAML configuration was used to produce this model:
55
+
56
+ ```yaml
57
+ base_model: mistralai/Mixtral-8x7B-v0.1
58
+ models:
59
+ - model: mistralai/Mixtral-8x7B-v0.1+Doctor-Shotgun/limarp-zloss-mixtral-8x7b-qlora
60
+ parameters:
61
+ density: 0.5
62
+ weight: 0.1
63
+ - model: Sao10K/Sensualize-Mixtral-bf16
64
+ parameters:
65
+ density: 0.5
66
+ weight: 0.1
67
+ - model: mistralai/Mixtral-8x7B-Instruct-v0.1
68
+ parameters:
69
+ density: 0.66
70
+ weight: 1.0
71
+ - model: jondurbin/bagel-dpo-8x7b-v0.2
72
+ parameters:
73
+ density: 0.66
74
+ weight: 0.5
75
+ merge_method: dare_ties
76
+ dtype: bfloat16
77
+
78
+ ```
79
+