File size: 3,940 Bytes
7886d18 419ba3c 679182a 7886d18 5db79d8 7886d18 3a89443 c00b53e c2aaa2d 2470754 b556f64 aa84a25 c00b53e 3a89443 7886d18 aad8879 7886d18 aad8879 7886d18 419ba3c 679182a 419ba3c 7886d18 16d91e6 7886d18 7c2c628 7886d18 419ba3c b723c6f 1c29bfa b723c6f 5db79d8 1c29bfa 5db79d8 b723c6f 1c29bfa 419ba3c b723c6f 7886d18 2a9d0da 7886d18 2ad5639 19abbcf aad8879 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 |
---
base_model:
- v000000/L3.1-8B-RP-Test-003-Task_Arithmetic
- v000000/L3.1-Niitorm-8B-t0.0001
- Sao10K/L3.1-8B-Niitama-v1.1
- arcee-ai/Llama-3.1-SuperNova-Lite
- akjindal53244/Llama-3.1-Storm-8B
- arcee-ai/Llama-Spark
- v000000/L3.1-8B-RP-Test-002-Task_Arithmetic
- grimjim/Llama-3-Instruct-abliteration-LoRA-8B
library_name: transformers
tags:
- mergekit
- merge
- llama
---
# Llama-3.1-Storniitova-8B
Storniitova-8B is a RP/Instruct model built on the foundation of Llama-3.1-SuperNova-Lite, which is distilled from the 405B parameter variant of Llama-3.1
By only changing the vector tasks, I attempt to retain the full distillation while learning roleplaying capabilties.
-----------------------------------------------------------------------------------------------------------
# merge
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit) and other proprietary tools.
## Merge Details
### Merge Method
This model was merged using the <b>SLERP, Task_Arithmetic and NEARSWAP</b> merge method.
### Models Merged
The following models were included in the merge:
* [v000000/L3.1-Niitorm-8B-t0.0001](https://huggingface.co/v000000/L3.1-Niitorm-8B-t0.0001)
* [akjindal53244/Llama-3.1-Storm-8B](https://huggingface.co/akjindal53244/Llama-3.1-Storm-8B)
* [arcee-ai/Llama-Spark](https://huggingface.co/arcee-ai/Llama-Spark)
* [arcee-ai/Llama-3.1-SuperNova-Lite](https://huggingface.co/arcee-ai/Llama-3.1-SuperNova-Lite)
* [v000000/L3.1-8B-RP-Test-003-Task_Arithmetic](https://huggingface.co/v000000/L3.1-8B-RP-Test-003-Task_Arithmetic)
* [Sao10K/L3.1-8B-Niitama-v1.1](https://huggingface.co/Sao10K/L3.1-8B-Niitama-v1.1) + [grimjim/Llama-3-Instruct-abliteration-LoRA-8B](https://huggingface.co/grimjim/Llama-3-Instruct-abliteration-LoRA-8B)
* [v000000/L3.1-8B-RP-Test-002-Task_Arithmetic](https://huggingface.co/v000000/L3.1-8B-RP-Test-002-Task_Arithmetic) + [grimjim/Llama-3-Instruct-abliteration-LoRA-8B](https://huggingface.co/grimjim/Llama-3-Instruct-abliteration-LoRA-8B)
### Recipe
The following YAML configuration was used to produce this model:
```yaml
#Step1 - Add smarts to Niitama with alchemonaut's algorithm.
slices:
- sources:
- model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
layer_range: [0, 32]
- model: akjindal53244/Llama-3.1-Storm-8B
layer_range: [0, 32]
merge_method: nearswap
base_model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
parameters:
t:
- value: 0.0001
dtype: bfloat16
out_type: float16
#Step 2 - Learn vectors onto Supernova 0.4(Niitorm)
models:
- model: arcee-ai/Llama-3.1-SuperNova-Lite
parameters:
weight: 1.0
- model: v000000/L3.1-Niitorm-8B-t0.0001
parameters:
weight: 0.4
merge_method: task_arithmetic
base_model: arcee-ai/Llama-3.1-SuperNova-Lite
parameters:
normalize: false
dtype: float16
#Step 3 - Fully learn vectors onto Supernova 1.25(Niitorm)
models:
- model: arcee-ai/Llama-3.1-SuperNova-Lite
parameters:
weight: 0.0
- model: v000000/L3.1-Niitorm-8B-t0.0001
parameters:
weight: 1.25
merge_method: task_arithmetic
base_model: arcee-ai/Llama-3.1-SuperNova-Lite
parameters:
normalize: false
dtype: float16
#Step 4 - Merge checkpoints and keep output/input Supernova heavy
#Merge with a triangular slerp from sophosympatheia.
models:
- model: v000000/L3.1-8B-RP-Test-003-Task_Arithmetic
merge_method: slerp
base_model: v000000/L3.1-8B-RP-Test-002-Task_Arithmetic+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
# This model needed some abliteration^
parameters:
t:
- value: [0, 0, 0.3, 0.4, 0.5, 0.6, 0.5, 0.4, 0.3, 0, 0]
dtype: float16
```
*SLERP distribution used to smoothly blend the mostly Supernova base with the 100% roleplay vectors:*
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/GP2LMRvMkhVJwNDSEC4oU.png)
|