totally-not-an-llm
commited on
Commit
·
63bef5d
1
Parent(s):
795d32b
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,41 @@
|
|
1 |
---
|
2 |
license: llama2
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: llama2
|
3 |
+
datasets:
|
4 |
+
- totally-not-an-llm/EverythingLM-data-V2
|
5 |
---
|
6 |
+
|
7 |
+
# EverythingLM-13b-16k
|
8 |
+
|
9 |
+
Introducing EverythingLM, a llama-2 based, general-purpose 13b model with 16k context thanks to LlongMa. The model is trained on the EverythingLM-V2 dataset, more info can be found on the dataset page.
|
10 |
+
|
11 |
+
The model is completely uncensored.
|
12 |
+
|
13 |
+
### GGML quants:
|
14 |
+
soon
|
15 |
+
|
16 |
+
Make sure to use correct rope scaling settings:
|
17 |
+
`-c 16384 --rope-freq-base 10000 --rope-freq-scale 0.25`
|
18 |
+
### GPTQ quants:
|
19 |
+
soon
|
20 |
+
|
21 |
+
### Notable features:
|
22 |
+
- Automatically triggered CoT reasoning.
|
23 |
+
- Verbose and detailed replies.
|
24 |
+
- Creative stories.
|
25 |
+
- Better prompt understanding.
|
26 |
+
|
27 |
+
### Prompt format:
|
28 |
+
It is a modified Vicuna format, the same used in many of ehartford's models.
|
29 |
+
```
|
30 |
+
You are a helpful AI assistant.
|
31 |
+
|
32 |
+
USER: <prompt>
|
33 |
+
ASSISTANT:
|
34 |
+
```
|
35 |
+
|
36 |
+
Training took about 2.5 hours using QLoRa on 1xA100, so this model can be recreated for about $4. QLoRa model can be found here: https://huggingface.co/totally-not-an-llm/EverythingLM-13b-V2-peft.
|
37 |
+
|
38 |
+
### Future plans:
|
39 |
+
- Native finetune.
|
40 |
+
- Other model sizes.
|
41 |
+
- Test some model merges using this model. (Specifically OpenOrca and Platypus models)
|