anujsahani01 commited on
Commit
8449650
·
1 Parent(s): 1b9ba37

Training in progress, step 500

Browse files
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac400f910045230fd7f524717715a99be96e0d01edfc0258022bcd31719bde28
3
  size 1188025
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71b45bcd4ec9083a0a27901057b4140664e3a299e5c765eaf29fc5ee24b6d0aa
3
  size 1188025
added_tokens.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "<ASSISTANT_TASK:>": 32771,
3
  "<END_TASK>": 32772,
 
4
  "<SYSTEM_TASK:>": 32769,
5
  "<USER_TASK:>": 32770,
6
  "<|PAD|>": 32768,
 
1
  {
2
  "<ASSISTANT_TASK:>": 32771,
3
  "<END_TASK>": 32772,
4
+ "<PAD>": 32778,
5
  "<SYSTEM_TASK:>": 32769,
6
  "<USER_TASK:>": 32770,
7
  "<|PAD|>": 32768,
tokenizer.json CHANGED
@@ -18,8 +18,8 @@
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
21
- "normalized": false,
22
- "special": true
23
  },
24
  {
25
  "id": 32769,
@@ -101,6 +101,15 @@
101
  "rstrip": false,
102
  "normalized": true,
103
  "special": false
 
 
 
 
 
 
 
 
 
104
  }
105
  ],
106
  "normalizer": null,
 
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
21
+ "normalized": true,
22
+ "special": false
23
  },
24
  {
25
  "id": 32769,
 
101
  "rstrip": false,
102
  "normalized": true,
103
  "special": false
104
+ },
105
+ {
106
+ "id": 32778,
107
+ "content": "<PAD>",
108
+ "single_word": false,
109
+ "lstrip": false,
110
+ "rstrip": false,
111
+ "normalized": false,
112
+ "special": true
113
  }
114
  ],
115
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -9,8 +9,8 @@
9
  "bos_token": "<|endoftext|>",
10
  "clean_up_tokenization_spaces": true,
11
  "eos_token": "<|endoftext|>",
12
- "model_max_length": 5000,
13
- "pad_token": "<|PAD|>",
14
  "tokenizer_class": "GPT2Tokenizer",
15
  "unk_token": "<|endoftext|>"
16
  }
 
9
  "bos_token": "<|endoftext|>",
10
  "clean_up_tokenization_spaces": true,
11
  "eos_token": "<|endoftext|>",
12
+ "model_max_length": 3000,
13
+ "pad_token": "<PAD>",
14
  "tokenizer_class": "GPT2Tokenizer",
15
  "unk_token": "<|endoftext|>"
16
  }