|
An attempt to make ParasiticRogue's model a tad better on longer contexts. I just ran the script, all credit for the original merge goes to my friend. |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6550b16f7490049d6237f200/1aFzrFX2W4cilMSZnoX2k.png) |
|
|
|
Exl2 quants already being uploaded by him: |
|
https://huggingface.co/ParasiticRogue/RP-Stew-v2.5-34B-exl2-4.65 |
|
|
|
Also, my samplers, instruct and prompt for the model (updated, new format): |
|
|
|
Samplers: https://files.catbox.moe/8ficm1.json |
|
Instruct: https://files.catbox.moe/nlflxw.json |
|
Story String: https://files.catbox.moe/6bk6gj.json |
|
|
|
If you're using Vectors (I use them for summaries of different story arcs): |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6550b16f7490049d6237f200/-Adevxuhmywjrs2WtexZK.png) |
|
|
|
Alternative samplers and an "old" version of the prompt, plus instruct (classic format): |
|
Samplers: https://files.catbox.moe/ef67mj.json |
|
Instruct: https://files.catbox.moe/uba6o1.json |
|
Story String: https://files.catbox.moe/t5gfun.json |
|
|
|
I'm unsure which one works better, so it's up to your tests. |
|
|
|
``` |
|
models: |
|
- model: F:\Merge\ParasiticRogue_Nontoxic-PiVoT-Bagel-RP-34b |
|
parameters: |
|
weight: 0.16 |
|
density: 0.42 |
|
- model: F:\Merge\ParasiticRogue_Nyakura-CausalLM-RP-34B |
|
parameters: |
|
weight: 0.22 |
|
density: 0.54 |
|
- model: F:\Merge\migtissera_Tess-34B-v1.5b |
|
parameters: |
|
weight: 0.28 |
|
density: 0.66 |
|
- model: F:\Merge\brucethemoose_Capybara-Fixed-Temp |
|
parameters: |
|
weight: 0.34 |
|
density: 0.78 |
|
merge_method: dare_ties |
|
base_model: F:\Merge\chargoddard_Yi-34B-200K-Llama |
|
parameters: |
|
int8_mask: true |
|
dtype: bfloat16 |
|
``` |