Philip Blair
commited on
Commit
·
7a6678d
1
Parent(s):
e49018b
Update tokenizer config
Browse files- README.md +4 -1
- tokenizer_config.json +1 -1
README.md
CHANGED
|
@@ -12,6 +12,9 @@ inference:
|
|
| 12 |
|
| 13 |
# Model Card for Mistral-7B-v0.1
|
| 14 |
|
|
|
|
|
|
|
|
|
|
| 15 |
The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters.
|
| 16 |
Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks we tested.
|
| 17 |
|
|
@@ -43,4 +46,4 @@ Mistral 7B is a pretrained base model and therefore does not have any moderation
|
|
| 43 |
|
| 44 |
## The Mistral AI Team
|
| 45 |
|
| 46 |
-
Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
|
|
|
|
| 12 |
|
| 13 |
# Model Card for Mistral-7B-v0.1
|
| 14 |
|
| 15 |
+
**NOTE**: This is a fork of [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) intended to have a non-null pad token. This has been done in order to
|
| 16 |
+
facilitate usage of this model with off-the-shelf PEFT tuners, such as what is offered by Google Cloud Vertex AI.
|
| 17 |
+
|
| 18 |
The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters.
|
| 19 |
Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks we tested.
|
| 20 |
|
|
|
|
| 46 |
|
| 47 |
## The Mistral AI Team
|
| 48 |
|
| 49 |
+
Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
|
tokenizer_config.json
CHANGED
|
@@ -33,7 +33,7 @@
|
|
| 33 |
"eos_token": "</s>",
|
| 34 |
"legacy": true,
|
| 35 |
"model_max_length": 1000000000000000019884624838656,
|
| 36 |
-
"pad_token":
|
| 37 |
"sp_model_kwargs": {},
|
| 38 |
"spaces_between_special_tokens": false,
|
| 39 |
"tokenizer_class": "LlamaTokenizer",
|
|
|
|
| 33 |
"eos_token": "</s>",
|
| 34 |
"legacy": true,
|
| 35 |
"model_max_length": 1000000000000000019884624838656,
|
| 36 |
+
"pad_token": "<unk>",
|
| 37 |
"sp_model_kwargs": {},
|
| 38 |
"spaces_between_special_tokens": false,
|
| 39 |
"tokenizer_class": "LlamaTokenizer",
|