jon-tow commited on
Commit
37e38dc
·
2 Parent(s): 2aa2607 c934d66

Merge branch 'main' of https://huggingface.co/stabilityai/stablelm-2-12b into main

Browse files
Files changed (2) hide show
  1. README.md +6 -3
  2. tokenizer_config.json +1 -1
README.md CHANGED
@@ -24,6 +24,8 @@ datasets:
24
 
25
  `Stable LM 2 12B` is a 12.1 billion parameter decoder-only language model pre-trained on 2 trillion tokens of diverse multilingual and code datasets for two epochs.
26
 
 
 
27
  ## Usage
28
 
29
  Get started generating text with `Stable LM 2 12B` by using the following code snippet:
@@ -81,7 +83,8 @@ print(tokenizer.decode(tokens[0], skip_special_tokens=True))
81
  * **Language(s)**: English
82
  * **Paper**: [Stable LM 2 Technical Report](https://arxiv.org/abs/2402.17834)
83
  * **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
84
- * **License**: [Stability AI Non-Commercial Research Community License](https://huggingface.co/stabilityai/stablelm-2-12b/blob/main/LICENSE). If you'd like to use this model for commercial products or purposes, please contact us [here](https://stability.ai/membership) to learn more.
 
85
  * **Contact**: For questions and comments about the model, please email `[email protected]`
86
 
87
  ### Model Architecture
@@ -120,7 +123,7 @@ The model is pre-trained on the aforementioned datasets in `bfloat16` precision,
120
 
121
  ### Intended Use
122
 
123
- The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications.
124
 
125
  ### Limitations and Bias
126
 
@@ -128,7 +131,7 @@ As a base model, this model may exhibit unreliable, unsafe, or other undesirable
128
 
129
  ## How to Cite
130
 
131
- ```
132
  @article{bellagente2024stable,
133
  title={Stable LM 2 1.6 B Technical Report},
134
  author={Bellagente, Marco and Tow, Jonathan and Mahan, Dakota and Phung, Duy and Zhuravinskyi, Maksym and Adithyan, Reshinth and Baicoianu, James and Brooks, Ben and Cooper, Nathan and Datta, Ashish and others},
 
24
 
25
  `Stable LM 2 12B` is a 12.1 billion parameter decoder-only language model pre-trained on 2 trillion tokens of diverse multilingual and code datasets for two epochs.
26
 
27
+ Please note: For commercial use, please refer to https://stability.ai/membership.
28
+
29
  ## Usage
30
 
31
  Get started generating text with `Stable LM 2 12B` by using the following code snippet:
 
83
  * **Language(s)**: English
84
  * **Paper**: [Stable LM 2 Technical Report](https://arxiv.org/abs/2402.17834)
85
  * **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
86
+ * **License**: [Stability AI Non-Commercial Research Community License](https://huggingface.co/stabilityai/stablelm-2-12b/blob/main/LICENSE).
87
+ * **Commercial License**: to use this model commercially, please refer to https://stability.ai/membership
88
  * **Contact**: For questions and comments about the model, please email `[email protected]`
89
 
90
  ### Model Architecture
 
123
 
124
  ### Intended Use
125
 
126
+ The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications. For commercial use, please refer to https://stability.ai/membership.
127
 
128
  ### Limitations and Bias
129
 
 
131
 
132
  ## How to Cite
133
 
134
+ ```bibtex
135
  @article{bellagente2024stable,
136
  title={Stable LM 2 1.6 B Technical Report},
137
  author={Bellagente, Marco and Tow, Jonathan and Mahan, Dakota and Phung, Duy and Zhuravinskyi, Maksym and Adithyan, Reshinth and Baicoianu, James and Brooks, Ben and Cooper, Nathan and Datta, Ashish and others},
tokenizer_config.json CHANGED
@@ -306,6 +306,6 @@
306
  "eos_token": "<|endoftext|>",
307
  "model_max_length": 1000000000000000019884624838656,
308
  "pad_token": "<|endoftext|>",
309
- "tokenizer_class": "GPT2Tokenizer",
310
  "unk_token": "<|endoftext|>"
311
  }
 
306
  "eos_token": "<|endoftext|>",
307
  "model_max_length": 1000000000000000019884624838656,
308
  "pad_token": "<|endoftext|>",
309
+ "tokenizer_class": "GPT2TokenizerFast",
310
  "unk_token": "<|endoftext|>"
311
  }