ToastyPigeon commited on
Commit
b78ff30
·
verified ·
1 Parent(s): 138f5ed

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -44
README.md CHANGED
@@ -1,59 +1,26 @@
1
  ---
2
  base_model:
3
  - unsloth/Mistral-Small-Instruct-2409
4
- - ToastyPigeon/mistral-small-springdragon-qlora
5
- - unsloth/Mistral-Small-Instruct-2409
6
- - unsloth/Mistral-Small-Instruct-2409
7
- - Alfitaria/mistral-small-fujin-qlora
8
- - nbeerbower/Mistral-Small-Gutenberg-Doppel-22B
9
  library_name: transformers
10
  tags:
11
  - mergekit
12
  - merge
13
 
14
  ---
15
- # ms-literarier-creativeflowerventure
16
-
17
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
-
19
- ## Merge Details
20
- ### Merge Method
21
 
22
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [unsloth/Mistral-Small-Instruct-2409](https://huggingface.co/unsloth/Mistral-Small-Instruct-2409) as a base.
23
 
24
- ### Models Merged
 
 
 
25
 
26
- The following models were included in the merge:
27
- * [unsloth/Mistral-Small-Instruct-2409](https://huggingface.co/unsloth/Mistral-Small-Instruct-2409) + [ToastyPigeon/mistral-small-springdragon-qlora](https://huggingface.co/ToastyPigeon/mistral-small-springdragon-qlora)
28
- * [unsloth/Mistral-Small-Instruct-2409](https://huggingface.co/unsloth/Mistral-Small-Instruct-2409) + [Alfitaria/mistral-small-fujin-qlora](https://huggingface.co/Alfitaria/mistral-small-fujin-qlora)
29
- * [nbeerbower/Mistral-Small-Gutenberg-Doppel-22B](https://huggingface.co/nbeerbower/Mistral-Small-Gutenberg-Doppel-22B)
30
- * output/ms-creative
31
 
32
- ### Configuration
33
 
34
- The following YAML configuration was used to produce this model:
 
35
 
36
- ```yaml
37
- base_model: unsloth/Mistral-Small-Instruct-2409
38
- merge_method: task_arithmetic
39
- slices:
40
- - sources:
41
- - layer_range: [0, 56]
42
- model: output/ms-creative
43
- parameters:
44
- weight: 0.3
45
- - layer_range: [0, 56]
46
- model: nbeerbower/Mistral-Small-Gutenberg-Doppel-22B
47
- parameters:
48
- weight: 0.6
49
- - layer_range: [0, 56]
50
- model: unsloth/Mistral-Small-Instruct-2409+Alfitaria/mistral-small-fujin-qlora
51
- parameters:
52
- weight: 0.4
53
- - layer_range: [0, 56]
54
- model: unsloth/Mistral-Small-Instruct-2409+ToastyPigeon/mistral-small-springdragon-qlora
55
- parameters:
56
- weight: 0.1
57
- - layer_range: [0, 56]
58
- model: unsloth/Mistral-Small-Instruct-2409
59
- ```
 
1
  ---
2
  base_model:
3
  - unsloth/Mistral-Small-Instruct-2409
 
 
 
 
 
4
  library_name: transformers
5
  tags:
6
  - mergekit
7
  - merge
8
 
9
  ---
10
+ # MS-Meadowlark-22B
 
 
 
 
 
11
 
12
+ A roleplay and storywriting model based on Mistral Small 22B.
13
 
14
+ Datasets used in this model:
15
+ - [Dampfinchen/Creative_Writing-Multiturn](https://huggingface.co/datasets/Dampfinchen/Creative_Writing_Multiturn) at 16k
16
+ - [Fizzarolli/rosier-dataset](https://huggingface.co/datasets/Fizzarolli/rosier-dataset) + [Alfitaria/body-inflation-org](https://huggingface.co/datasets/Alfitaria/body-inflation-org) at 16k
17
+ - [ToastyPigeon/SpringDragon](https://huggingface.co/datasets/ToastyPigeon/SpringDragon) at 8k
18
 
19
+ Each dataset was trained separately onto Mistral Small Instruct, and then the component models were merged along with [nbeerbower/Mistral-Small-Gutenberg-Doppel-22B](https://huggingface.co/nbeerbower/Mistral-Small-Gutenberg-Doppel-22B) to create Meadowlark.
 
 
 
 
20
 
21
+ I tried different blends of the component models, and this one seems to be the most stable while retaining creativity and unpredictability added by the trained data.
22
 
23
+ # Instruct Format
24
+ Rosier/bodyinf and SpringDragon were trained in completion format. This model should work with [Kobold Lite](https://lite.koboldai.net/) in Adventure Mode and Story Mode.
25
 
26
+ Creative_Writing_Multiturn