Statuo commited on
Commit
200775d
·
verified ·
1 Parent(s): 461d89b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -1
README.md CHANGED
@@ -1,4 +1,28 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  base_model:
3
  - SanjiWatsuki/Kunoichi-DPO-v2-7B
4
  - dreamgen/WizardLM-2-7B
@@ -42,4 +66,4 @@ models:
42
  weight: 0.6
43
  merge_method: linear
44
  dtype: float16
45
- ```
 
1
  ---
2
+ {}
3
+ ---
4
+ # Lemon Kunoichi Wizard - 7b
5
+
6
+ ![LemonKunoichiWizard](https://files.catbox.moe/eivabp.png)
7
+
8
+ [Base Model](https://huggingface.co/Statuo/LemonKunoichiWizardV3/), [4bpw](https://huggingface.co/Statuo/LemonKunoichiWizardv3_4bpw), [6bpw](https://huggingface.co/Statuo/LemonKunoichiWizardv3_6bpw), [8bpw](https://huggingface.co/Statuo/LemonKunoichiWizardv3_8bpw)
9
+
10
+ The Quanted versions come with the measurement files in case you want to do your own quants.
11
+
12
+ A merge of three models, LemonadeRP-4.5.3, Kunoichi-DPO-v2, and WizardLM-2. I used Lemonade as a base with Kunoichi being the second biggest influence and WizardLM-2 for logic capabilities.
13
+
14
+ The end result is a Roleplay-focused model with great character card inference. I ran 4 merges at varying values to see which provided the most accurate output to a character cards quirk, with this v3 version being the winner out of the four.
15
+
16
+ ## Context Template - Alpaca
17
+
18
+ Alpaca preset seems to work well with your own System Prompt.
19
+
20
+ ## Context Size - 8192
21
+
22
+ The model loads at 8192 on my end, but theoretically it should be able to go up to 32k. Not that it'll be coherent at 32k. Most models based on Mistral like this end up being - at best - 12k context size for coherent output. I only tested at 8k which is where the base models tend to shine. YMMV otherwise.
23
+
24
+ ---
25
+
26
  base_model:
27
  - SanjiWatsuki/Kunoichi-DPO-v2-7B
28
  - dreamgen/WizardLM-2-7B
 
66
  weight: 0.6
67
  merge_method: linear
68
  dtype: float16
69
+ ```