Philip Blair
commited on
Commit
·
6f545a3
1
Parent(s):
21e411e
Modify tokenizer config to have non-null pad token
Browse files- README.md +4 -1
- tokenizer_config.json +1 -1
README.md
CHANGED
@@ -10,6 +10,9 @@ inference:
|
|
10 |
|
11 |
# Model Card for Mistral-7B-Instruct-v0.1
|
12 |
|
|
|
|
|
|
|
13 |
The Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) generative text model using a variety of publicly available conversation datasets.
|
14 |
|
15 |
For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/announcing-mistral-7b/).
|
@@ -84,4 +87,4 @@ make the model finely respect guardrails, allowing for deployment in environment
|
|
84 |
|
85 |
## The Mistral AI Team
|
86 |
|
87 |
-
Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
|
|
|
10 |
|
11 |
# Model Card for Mistral-7B-Instruct-v0.1
|
12 |
|
13 |
+
**NOTE**: This is a fork of [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) intended to have a non-null pad token. This has been done in order to
|
14 |
+
facilitate usage of this model with off-the-shelf PEFT tuners, such as what is offered by Google Cloud Vertex AI.
|
15 |
+
|
16 |
The Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) generative text model using a variety of publicly available conversation datasets.
|
17 |
|
18 |
For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/announcing-mistral-7b/).
|
|
|
87 |
|
88 |
## The Mistral AI Team
|
89 |
|
90 |
+
Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
|
tokenizer_config.json
CHANGED
@@ -34,7 +34,7 @@
|
|
34 |
"eos_token": "</s>",
|
35 |
"legacy": true,
|
36 |
"model_max_length": 1000000000000000019884624838656,
|
37 |
-
"pad_token":
|
38 |
"sp_model_kwargs": {},
|
39 |
"spaces_between_special_tokens": false,
|
40 |
"tokenizer_class": "LlamaTokenizer",
|
|
|
34 |
"eos_token": "</s>",
|
35 |
"legacy": true,
|
36 |
"model_max_length": 1000000000000000019884624838656,
|
37 |
+
"pad_token": "<unk>",
|
38 |
"sp_model_kwargs": {},
|
39 |
"spaces_between_special_tokens": false,
|
40 |
"tokenizer_class": "LlamaTokenizer",
|