Update README.md
Browse files
README.md
CHANGED
@@ -139,7 +139,7 @@ This example demonstrates how to load the model and tokenizer, prepare input, ge
|
|
139 |
|
140 |
[Teuken-7B-instruct-research-v0.4](https://huggingface.co/openGPT-X/Teuken-7B-instruct-research-v0.4) was pre-trained on 4 trillion tokens of data from publicly available sources.
|
141 |
The pretraining data has a cutoff of September 2023.
|
142 |
-
More information is available in our
|
143 |
|
144 |
|
145 |
### Instruction-Tuning Data
|
|
|
139 |
|
140 |
[Teuken-7B-instruct-research-v0.4](https://huggingface.co/openGPT-X/Teuken-7B-instruct-research-v0.4) was pre-trained on 4 trillion tokens of data from publicly available sources.
|
141 |
The pretraining data has a cutoff of September 2023.
|
142 |
+
More information is available in our preprint ["Data Processing for the OpenGPT-X Model Family"](http://arxiv.org/abs/2410.08800).
|
143 |
|
144 |
|
145 |
### Instruction-Tuning Data
|