Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,56 @@
|
|
1 |
-
---
|
2 |
-
license: llama3
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: llama3
|
3 |
+
license_name: llama3
|
4 |
+
license_link: LICENSE
|
5 |
+
library_name: transformers
|
6 |
+
tags:
|
7 |
+
- not-for-all-audiences
|
8 |
+
datasets:
|
9 |
+
- crestf411/LimaRP-DS
|
10 |
+
- Gryphe/Sonnet3.5-Charcard-Roleplay
|
11 |
+
- anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system
|
12 |
+
- anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
|
13 |
+
- anthracite-org/kalo-opus-instruct-3k-filtered-no-system
|
14 |
+
- anthracite-org/nopm_claude_writing_fixed
|
15 |
+
base_model:
|
16 |
+
- nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
|
17 |
+
---
|
18 |
+
|
19 |
+
Sunfall (2024-10-28) v0.7.0 trained directly against, and merged with [Nemotron 70B Instruct](nvidia/Llama-3.1-Nemotron-70B-Instruct-HF).
|
20 |
+
|
21 |
+
It also contains samples from [Antracite.Org](https://huggingface.co/anthracite-org) datasets. See bottom for details.
|
22 |
+
|
23 |
+
Significant revamping of the dataset metadata generation process, resulting in higher quality dataset overall. *The "Diamond Law" experiment has been removed as it didn't seem to affect the model output enough to warrant set up complexity.*
|
24 |
+
|
25 |
+
Recommended starting point:
|
26 |
+
|
27 |
+
* Temperature: **1**
|
28 |
+
* MinP: **0.05~0.1**
|
29 |
+
* DRY: **0.8 1.75 2 0**
|
30 |
+
|
31 |
+
At early context, I recommend keeping XTC disabled. Once you hit higher context sizes (10k+), enabling XTC at 0.1 / 0.5 seems to significantly improve the output, but YMMV. If the output drones on and is uninspiring, XTC can be extremely effective.
|
32 |
+
|
33 |
+
General heuristic:
|
34 |
+
|
35 |
+
* Lots of slop? Temperature is too low. Raise it, or enable XTC. For early context, temp bump is probably preferred.
|
36 |
+
* Is the model making mistakes about subtle or obvious details in the scene? Temperature is too high, OR XTC is enabled and/or XTC settings are too high. Lower temp and/or disable XTC.
|
37 |
+
|
38 |
+
*Mergers/fine-tuners: [there is a LoRA of this model](https://huggingface.co/crestf411/sunfall-peft/tree/main/l3.1-nemotron-70B). Consider merging that instead of merging this model.*
|
39 |
+
|
40 |
+
This model has been trained on context that mimics that of Silly Tavern's "Llama 3 Instruct" preset, with character names added.
|
41 |
+
|
42 |
+
The model has also been trained to do interactive storywriting. You may steer the model towards specific content by "responding" to the model like so:
|
43 |
+
|
44 |
+
```
|
45 |
+
Continue writing adhering to the following scenario: (things you want to happen next)
|
46 |
+
```
|
47 |
+
|
48 |
+
Additional inclusions (random sampled sub-set, cursorily quality-checked) from:
|
49 |
+
|
50 |
+
- [Gryphe/Sonnet3.5-Charcard-Roleplay](https://huggingface.co/datasets/Gryphe/Sonnet3.5-Charcard-Roleplay)
|
51 |
+
- [anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system](https://huggingface.co/datasets/anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system)
|
52 |
+
- [anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system)
|
53 |
+
- [anthracite-org/kalo-opus-instruct-3k-filtered-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-3k-filtered-no-system)
|
54 |
+
- [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
|
55 |
+
|
56 |
+
As such, the dataset is not 100% slop free, but this addition likely helps the model be a better roleplayer. At some point, I intend to clean up and release the samples, deslopped.
|