Philip Blair commited on
Commit
7a6678d
·
1 Parent(s): e49018b

Update tokenizer config

Browse files
Files changed (2) hide show
  1. README.md +4 -1
  2. tokenizer_config.json +1 -1
README.md CHANGED
@@ -12,6 +12,9 @@ inference:
12
 
13
  # Model Card for Mistral-7B-v0.1
14
 
 
 
 
15
  The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters.
16
  Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks we tested.
17
 
@@ -43,4 +46,4 @@ Mistral 7B is a pretrained base model and therefore does not have any moderation
43
 
44
  ## The Mistral AI Team
45
 
46
- Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
 
12
 
13
  # Model Card for Mistral-7B-v0.1
14
 
15
+ **NOTE**: This is a fork of [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) intended to have a non-null pad token. This has been done in order to
16
+ facilitate usage of this model with off-the-shelf PEFT tuners, such as what is offered by Google Cloud Vertex AI.
17
+
18
  The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters.
19
  Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks we tested.
20
 
 
46
 
47
  ## The Mistral AI Team
48
 
49
+ Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
tokenizer_config.json CHANGED
@@ -33,7 +33,7 @@
33
  "eos_token": "</s>",
34
  "legacy": true,
35
  "model_max_length": 1000000000000000019884624838656,
36
- "pad_token": null,
37
  "sp_model_kwargs": {},
38
  "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",
 
33
  "eos_token": "</s>",
34
  "legacy": true,
35
  "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": "<unk>",
37
  "sp_model_kwargs": {},
38
  "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",