crestf411 commited on
Commit
ac42c96
·
verified ·
1 Parent(s): 1b2eab7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -3
README.md CHANGED
@@ -1,3 +1,56 @@
1
- ---
2
- license: llama3.1
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3
3
+ license_name: llama3
4
+ license_link: LICENSE
5
+ library_name: transformers
6
+ tags:
7
+ - not-for-all-audiences
8
+ datasets:
9
+ - crestf411/LimaRP-DS
10
+ - Gryphe/Sonnet3.5-Charcard-Roleplay
11
+ - anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system
12
+ - anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
13
+ - anthracite-org/kalo-opus-instruct-3k-filtered-no-system
14
+ - anthracite-org/nopm_claude_writing_fixed
15
+ base_model:
16
+ - nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
17
+ ---
18
+
19
+ Sunfall (2024-10-28) v0.7.0 trained directly against, and merged with [Nemotron 70B Instruct](nvidia/Llama-3.1-Nemotron-70B-Instruct-HF).
20
+
21
+ It also contains samples from [Antracite.Org](https://huggingface.co/anthracite-org) datasets. See bottom for details.
22
+
23
+ Significant revamping of the dataset metadata generation process, resulting in higher quality dataset overall. *The "Diamond Law" experiment has been removed as it didn't seem to affect the model output enough to warrant set up complexity.*
24
+
25
+ Recommended starting point:
26
+
27
+ * Temperature: **1**
28
+ * MinP: **0.05~0.1**
29
+ * DRY: **0.8 1.75 2 0**
30
+
31
+ At early context, I recommend keeping XTC disabled. Once you hit higher context sizes (10k+), enabling XTC at 0.1 / 0.5 seems to significantly improve the output, but YMMV. If the output drones on and is uninspiring, XTC can be extremely effective.
32
+
33
+ General heuristic:
34
+
35
+ * Lots of slop? Temperature is too low. Raise it, or enable XTC. For early context, temp bump is probably preferred.
36
+ * Is the model making mistakes about subtle or obvious details in the scene? Temperature is too high, OR XTC is enabled and/or XTC settings are too high. Lower temp and/or disable XTC.
37
+
38
+ *Mergers/fine-tuners: [there is a LoRA of this model](https://huggingface.co/crestf411/sunfall-peft/tree/main/l3.1-nemotron-70B). Consider merging that instead of merging this model.*
39
+
40
+ This model has been trained on context that mimics that of Silly Tavern's "Llama 3 Instruct" preset, with character names added.
41
+
42
+ The model has also been trained to do interactive storywriting. You may steer the model towards specific content by "responding" to the model like so:
43
+
44
+ ```
45
+ Continue writing adhering to the following scenario: (things you want to happen next)
46
+ ```
47
+
48
+ Additional inclusions (random sampled sub-set, cursorily quality-checked) from:
49
+
50
+ - [Gryphe/Sonnet3.5-Charcard-Roleplay](https://huggingface.co/datasets/Gryphe/Sonnet3.5-Charcard-Roleplay)
51
+ - [anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system](https://huggingface.co/datasets/anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system)
52
+ - [anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system)
53
+ - [anthracite-org/kalo-opus-instruct-3k-filtered-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-3k-filtered-no-system)
54
+ - [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
55
+
56
+ As such, the dataset is not 100% slop free, but this addition likely helps the model be a better roleplayer. At some point, I intend to clean up and release the samples, deslopped.