Merge branch 'main' of https://huggingface.co/stabilityai/stablelm-2-12b into main
Browse files- README.md +6 -3
- tokenizer_config.json +1 -1
README.md
CHANGED
@@ -24,6 +24,8 @@ datasets:
|
|
24 |
|
25 |
`Stable LM 2 12B` is a 12.1 billion parameter decoder-only language model pre-trained on 2 trillion tokens of diverse multilingual and code datasets for two epochs.
|
26 |
|
|
|
|
|
27 |
## Usage
|
28 |
|
29 |
Get started generating text with `Stable LM 2 12B` by using the following code snippet:
|
@@ -81,7 +83,8 @@ print(tokenizer.decode(tokens[0], skip_special_tokens=True))
|
|
81 |
* **Language(s)**: English
|
82 |
* **Paper**: [Stable LM 2 Technical Report](https://arxiv.org/abs/2402.17834)
|
83 |
* **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
|
84 |
-
* **License**: [Stability AI Non-Commercial Research Community License](https://huggingface.co/stabilityai/stablelm-2-12b/blob/main/LICENSE).
|
|
|
85 |
* **Contact**: For questions and comments about the model, please email `[email protected]`
|
86 |
|
87 |
### Model Architecture
|
@@ -120,7 +123,7 @@ The model is pre-trained on the aforementioned datasets in `bfloat16` precision,
|
|
120 |
|
121 |
### Intended Use
|
122 |
|
123 |
-
The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications.
|
124 |
|
125 |
### Limitations and Bias
|
126 |
|
@@ -128,7 +131,7 @@ As a base model, this model may exhibit unreliable, unsafe, or other undesirable
|
|
128 |
|
129 |
## How to Cite
|
130 |
|
131 |
-
```
|
132 |
@article{bellagente2024stable,
|
133 |
title={Stable LM 2 1.6 B Technical Report},
|
134 |
author={Bellagente, Marco and Tow, Jonathan and Mahan, Dakota and Phung, Duy and Zhuravinskyi, Maksym and Adithyan, Reshinth and Baicoianu, James and Brooks, Ben and Cooper, Nathan and Datta, Ashish and others},
|
|
|
24 |
|
25 |
`Stable LM 2 12B` is a 12.1 billion parameter decoder-only language model pre-trained on 2 trillion tokens of diverse multilingual and code datasets for two epochs.
|
26 |
|
27 |
+
Please note: For commercial use, please refer to https://stability.ai/membership.
|
28 |
+
|
29 |
## Usage
|
30 |
|
31 |
Get started generating text with `Stable LM 2 12B` by using the following code snippet:
|
|
|
83 |
* **Language(s)**: English
|
84 |
* **Paper**: [Stable LM 2 Technical Report](https://arxiv.org/abs/2402.17834)
|
85 |
* **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
|
86 |
+
* **License**: [Stability AI Non-Commercial Research Community License](https://huggingface.co/stabilityai/stablelm-2-12b/blob/main/LICENSE).
|
87 |
+
* **Commercial License**: to use this model commercially, please refer to https://stability.ai/membership
|
88 |
* **Contact**: For questions and comments about the model, please email `[email protected]`
|
89 |
|
90 |
### Model Architecture
|
|
|
123 |
|
124 |
### Intended Use
|
125 |
|
126 |
+
The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications. For commercial use, please refer to https://stability.ai/membership.
|
127 |
|
128 |
### Limitations and Bias
|
129 |
|
|
|
131 |
|
132 |
## How to Cite
|
133 |
|
134 |
+
```bibtex
|
135 |
@article{bellagente2024stable,
|
136 |
title={Stable LM 2 1.6 B Technical Report},
|
137 |
author={Bellagente, Marco and Tow, Jonathan and Mahan, Dakota and Phung, Duy and Zhuravinskyi, Maksym and Adithyan, Reshinth and Baicoianu, James and Brooks, Ben and Cooper, Nathan and Datta, Ashish and others},
|
tokenizer_config.json
CHANGED
@@ -306,6 +306,6 @@
|
|
306 |
"eos_token": "<|endoftext|>",
|
307 |
"model_max_length": 1000000000000000019884624838656,
|
308 |
"pad_token": "<|endoftext|>",
|
309 |
-
"tokenizer_class": "
|
310 |
"unk_token": "<|endoftext|>"
|
311 |
}
|
|
|
306 |
"eos_token": "<|endoftext|>",
|
307 |
"model_max_length": 1000000000000000019884624838656,
|
308 |
"pad_token": "<|endoftext|>",
|
309 |
+
"tokenizer_class": "GPT2TokenizerFast",
|
310 |
"unk_token": "<|endoftext|>"
|
311 |
}
|