manato003 commited on
Commit
8340b4f
·
verified ·
1 Parent(s): 3d9b41a

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ library_name: transformers
4
+ tags:
5
+ - autotrain
6
+ - text-classification
7
+ base_model: studio-ousia/luke-japanese-base-lite
8
+ widget:
9
+ - text: "I love AutoTrain"
10
+ ---
11
+
12
+ # Model Trained Using AutoTrain
13
+
14
+ - Problem type: Text Classification
15
+
16
+ ## Validation Metrics
17
+ loss: 0.5321555137634277
18
+
19
+ f1: 0.7355371900826446
20
+
21
+ precision: 0.6544117647058824
22
+
23
+ recall: 0.839622641509434
24
+
25
+ auc: 0.8266129032258065
26
+
27
+ accuracy: 0.7217391304347827
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<ent2>": 32771,
3
+ "<ent>": 32770
4
+ }
checkpoint-230/config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "studio-ousia/luke-japanese-base-lite",
3
+ "_num_labels": 2,
4
+ "architectures": [
5
+ "LukeForSequenceClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "bert_model_name": "models/luke-japanese/hf_xlm_roberta",
9
+ "bos_token_id": 0,
10
+ "classifier_dropout": null,
11
+ "cls_entity_prediction": false,
12
+ "entity_emb_size": 256,
13
+ "entity_vocab_size": 4,
14
+ "eos_token_id": 2,
15
+ "hidden_act": "gelu",
16
+ "hidden_dropout_prob": 0.1,
17
+ "hidden_size": 768,
18
+ "id2label": {
19
+ "0": "\u60aa\u3044",
20
+ "1": "\u826f\u3044"
21
+ },
22
+ "initializer_range": 0.02,
23
+ "intermediate_size": 3072,
24
+ "label2id": {
25
+ "\u60aa\u3044": 0,
26
+ "\u826f\u3044": 1
27
+ },
28
+ "layer_norm_eps": 1e-05,
29
+ "max_position_embeddings": 514,
30
+ "model_type": "luke",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 1,
34
+ "position_embedding_type": "absolute",
35
+ "problem_type": "single_label_classification",
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.48.0",
38
+ "type_vocab_size": 1,
39
+ "use_cache": true,
40
+ "use_entity_aware_attention": true,
41
+ "vocab_size": 32772
42
+ }
checkpoint-230/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de13ef21ccae838fb30a5ec29dd128906a5bfc49538967ccc3b269ce2043e573
3
+ size 532307280
checkpoint-230/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c83aee057460784cfcc3a6b63cfbda532203292f4c7a25f41918a4960e3909e9
3
+ size 889863098
checkpoint-230/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a00ae46a38a73f402956850b524194acd54becd011fcc7d9c07b576676cd3a60
3
+ size 13990
checkpoint-230/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99d2857ac1c5e019d8e0780861a630b7d7dd2080a27821bbdf29e9280ee1c0f2
3
+ size 1064
checkpoint-230/trainer_state.json ADDED
@@ -0,0 +1,390 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.5321555137634277,
3
+ "best_model_checkpoint": "2classification/checkpoint-230",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 230,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.043478260869565216,
13
+ "grad_norm": 3.3017044067382812,
14
+ "learning_rate": 7.142857142857143e-06,
15
+ "loss": 0.701,
16
+ "step": 5
17
+ },
18
+ {
19
+ "epoch": 0.08695652173913043,
20
+ "grad_norm": 3.861875295639038,
21
+ "learning_rate": 1.4285714285714285e-05,
22
+ "loss": 0.6724,
23
+ "step": 10
24
+ },
25
+ {
26
+ "epoch": 0.13043478260869565,
27
+ "grad_norm": 5.9016523361206055,
28
+ "learning_rate": 2.1428571428571428e-05,
29
+ "loss": 0.7108,
30
+ "step": 15
31
+ },
32
+ {
33
+ "epoch": 0.17391304347826086,
34
+ "grad_norm": 2.023663282394409,
35
+ "learning_rate": 2.857142857142857e-05,
36
+ "loss": 0.7146,
37
+ "step": 20
38
+ },
39
+ {
40
+ "epoch": 0.21739130434782608,
41
+ "grad_norm": 2.016728401184082,
42
+ "learning_rate": 3.571428571428572e-05,
43
+ "loss": 0.7238,
44
+ "step": 25
45
+ },
46
+ {
47
+ "epoch": 0.2608695652173913,
48
+ "grad_norm": 1.6572303771972656,
49
+ "learning_rate": 4.2857142857142856e-05,
50
+ "loss": 0.6849,
51
+ "step": 30
52
+ },
53
+ {
54
+ "epoch": 0.30434782608695654,
55
+ "grad_norm": 4.739016532897949,
56
+ "learning_rate": 5e-05,
57
+ "loss": 0.7218,
58
+ "step": 35
59
+ },
60
+ {
61
+ "epoch": 0.34782608695652173,
62
+ "grad_norm": 2.0643484592437744,
63
+ "learning_rate": 4.9193548387096775e-05,
64
+ "loss": 0.6941,
65
+ "step": 40
66
+ },
67
+ {
68
+ "epoch": 0.391304347826087,
69
+ "grad_norm": 2.2051498889923096,
70
+ "learning_rate": 4.8387096774193554e-05,
71
+ "loss": 0.6904,
72
+ "step": 45
73
+ },
74
+ {
75
+ "epoch": 0.43478260869565216,
76
+ "grad_norm": 4.221965789794922,
77
+ "learning_rate": 4.7580645161290326e-05,
78
+ "loss": 0.7442,
79
+ "step": 50
80
+ },
81
+ {
82
+ "epoch": 0.4782608695652174,
83
+ "grad_norm": 3.326643466949463,
84
+ "learning_rate": 4.67741935483871e-05,
85
+ "loss": 0.7255,
86
+ "step": 55
87
+ },
88
+ {
89
+ "epoch": 0.5217391304347826,
90
+ "grad_norm": 5.108890533447266,
91
+ "learning_rate": 4.596774193548387e-05,
92
+ "loss": 0.6994,
93
+ "step": 60
94
+ },
95
+ {
96
+ "epoch": 0.5652173913043478,
97
+ "grad_norm": 1.680823802947998,
98
+ "learning_rate": 4.516129032258064e-05,
99
+ "loss": 0.6759,
100
+ "step": 65
101
+ },
102
+ {
103
+ "epoch": 0.6086956521739131,
104
+ "grad_norm": 4.85138463973999,
105
+ "learning_rate": 4.435483870967742e-05,
106
+ "loss": 0.6261,
107
+ "step": 70
108
+ },
109
+ {
110
+ "epoch": 0.6521739130434783,
111
+ "grad_norm": 4.047025680541992,
112
+ "learning_rate": 4.3548387096774194e-05,
113
+ "loss": 0.5949,
114
+ "step": 75
115
+ },
116
+ {
117
+ "epoch": 0.6956521739130435,
118
+ "grad_norm": 3.6327199935913086,
119
+ "learning_rate": 4.2741935483870973e-05,
120
+ "loss": 0.8464,
121
+ "step": 80
122
+ },
123
+ {
124
+ "epoch": 0.7391304347826086,
125
+ "grad_norm": 3.9471940994262695,
126
+ "learning_rate": 4.1935483870967746e-05,
127
+ "loss": 0.7577,
128
+ "step": 85
129
+ },
130
+ {
131
+ "epoch": 0.782608695652174,
132
+ "grad_norm": 4.150089263916016,
133
+ "learning_rate": 4.112903225806452e-05,
134
+ "loss": 0.6673,
135
+ "step": 90
136
+ },
137
+ {
138
+ "epoch": 0.8260869565217391,
139
+ "grad_norm": 2.2822818756103516,
140
+ "learning_rate": 4.032258064516129e-05,
141
+ "loss": 0.7113,
142
+ "step": 95
143
+ },
144
+ {
145
+ "epoch": 0.8695652173913043,
146
+ "grad_norm": 5.6378984451293945,
147
+ "learning_rate": 3.951612903225806e-05,
148
+ "loss": 0.6567,
149
+ "step": 100
150
+ },
151
+ {
152
+ "epoch": 0.9130434782608695,
153
+ "grad_norm": 1.9879374504089355,
154
+ "learning_rate": 3.870967741935484e-05,
155
+ "loss": 0.7019,
156
+ "step": 105
157
+ },
158
+ {
159
+ "epoch": 0.9565217391304348,
160
+ "grad_norm": 6.7003631591796875,
161
+ "learning_rate": 3.7903225806451614e-05,
162
+ "loss": 0.6505,
163
+ "step": 110
164
+ },
165
+ {
166
+ "epoch": 1.0,
167
+ "grad_norm": 9.337882041931152,
168
+ "learning_rate": 3.7096774193548386e-05,
169
+ "loss": 0.6924,
170
+ "step": 115
171
+ },
172
+ {
173
+ "epoch": 1.0,
174
+ "eval_accuracy": 0.7043478260869566,
175
+ "eval_auc": 0.7728241022519781,
176
+ "eval_f1": 0.6991150442477876,
177
+ "eval_loss": 0.607258677482605,
178
+ "eval_precision": 0.6583333333333333,
179
+ "eval_recall": 0.7452830188679245,
180
+ "eval_runtime": 35.0981,
181
+ "eval_samples_per_second": 6.553,
182
+ "eval_steps_per_second": 0.427,
183
+ "step": 115
184
+ },
185
+ {
186
+ "epoch": 1.0434782608695652,
187
+ "grad_norm": 1.8164705038070679,
188
+ "learning_rate": 3.6290322580645165e-05,
189
+ "loss": 0.6439,
190
+ "step": 120
191
+ },
192
+ {
193
+ "epoch": 1.0869565217391304,
194
+ "grad_norm": 2.1826255321502686,
195
+ "learning_rate": 3.548387096774194e-05,
196
+ "loss": 0.647,
197
+ "step": 125
198
+ },
199
+ {
200
+ "epoch": 1.1304347826086956,
201
+ "grad_norm": 2.288874626159668,
202
+ "learning_rate": 3.467741935483872e-05,
203
+ "loss": 0.6064,
204
+ "step": 130
205
+ },
206
+ {
207
+ "epoch": 1.1739130434782608,
208
+ "grad_norm": 2.857973337173462,
209
+ "learning_rate": 3.387096774193548e-05,
210
+ "loss": 0.5087,
211
+ "step": 135
212
+ },
213
+ {
214
+ "epoch": 1.2173913043478262,
215
+ "grad_norm": 4.019787788391113,
216
+ "learning_rate": 3.306451612903226e-05,
217
+ "loss": 0.5771,
218
+ "step": 140
219
+ },
220
+ {
221
+ "epoch": 1.2608695652173914,
222
+ "grad_norm": 5.981603622436523,
223
+ "learning_rate": 3.2258064516129034e-05,
224
+ "loss": 0.5869,
225
+ "step": 145
226
+ },
227
+ {
228
+ "epoch": 1.3043478260869565,
229
+ "grad_norm": 12.450459480285645,
230
+ "learning_rate": 3.1451612903225806e-05,
231
+ "loss": 0.6328,
232
+ "step": 150
233
+ },
234
+ {
235
+ "epoch": 1.3478260869565217,
236
+ "grad_norm": 9.469613075256348,
237
+ "learning_rate": 3.0645161290322585e-05,
238
+ "loss": 0.6025,
239
+ "step": 155
240
+ },
241
+ {
242
+ "epoch": 1.391304347826087,
243
+ "grad_norm": 4.401791095733643,
244
+ "learning_rate": 2.9838709677419357e-05,
245
+ "loss": 0.4786,
246
+ "step": 160
247
+ },
248
+ {
249
+ "epoch": 1.434782608695652,
250
+ "grad_norm": 5.496943950653076,
251
+ "learning_rate": 2.9032258064516133e-05,
252
+ "loss": 0.7632,
253
+ "step": 165
254
+ },
255
+ {
256
+ "epoch": 1.4782608695652173,
257
+ "grad_norm": 2.510056495666504,
258
+ "learning_rate": 2.822580645161291e-05,
259
+ "loss": 0.412,
260
+ "step": 170
261
+ },
262
+ {
263
+ "epoch": 1.5217391304347827,
264
+ "grad_norm": 4.305450439453125,
265
+ "learning_rate": 2.7419354838709678e-05,
266
+ "loss": 0.4254,
267
+ "step": 175
268
+ },
269
+ {
270
+ "epoch": 1.5652173913043477,
271
+ "grad_norm": 4.428637504577637,
272
+ "learning_rate": 2.661290322580645e-05,
273
+ "loss": 0.6168,
274
+ "step": 180
275
+ },
276
+ {
277
+ "epoch": 1.608695652173913,
278
+ "grad_norm": 4.408801555633545,
279
+ "learning_rate": 2.5806451612903226e-05,
280
+ "loss": 0.5885,
281
+ "step": 185
282
+ },
283
+ {
284
+ "epoch": 1.6521739130434783,
285
+ "grad_norm": 5.0507097244262695,
286
+ "learning_rate": 2.5e-05,
287
+ "loss": 0.5953,
288
+ "step": 190
289
+ },
290
+ {
291
+ "epoch": 1.6956521739130435,
292
+ "grad_norm": 7.350249290466309,
293
+ "learning_rate": 2.4193548387096777e-05,
294
+ "loss": 0.5695,
295
+ "step": 195
296
+ },
297
+ {
298
+ "epoch": 1.7391304347826086,
299
+ "grad_norm": 7.184178352355957,
300
+ "learning_rate": 2.338709677419355e-05,
301
+ "loss": 0.6336,
302
+ "step": 200
303
+ },
304
+ {
305
+ "epoch": 1.7826086956521738,
306
+ "grad_norm": 4.202939987182617,
307
+ "learning_rate": 2.258064516129032e-05,
308
+ "loss": 0.7637,
309
+ "step": 205
310
+ },
311
+ {
312
+ "epoch": 1.8260869565217392,
313
+ "grad_norm": 6.9584431648254395,
314
+ "learning_rate": 2.1774193548387097e-05,
315
+ "loss": 0.5885,
316
+ "step": 210
317
+ },
318
+ {
319
+ "epoch": 1.8695652173913042,
320
+ "grad_norm": 3.880800724029541,
321
+ "learning_rate": 2.0967741935483873e-05,
322
+ "loss": 0.5884,
323
+ "step": 215
324
+ },
325
+ {
326
+ "epoch": 1.9130434782608696,
327
+ "grad_norm": 5.578256130218506,
328
+ "learning_rate": 2.0161290322580645e-05,
329
+ "loss": 0.5477,
330
+ "step": 220
331
+ },
332
+ {
333
+ "epoch": 1.9565217391304348,
334
+ "grad_norm": 5.322207450866699,
335
+ "learning_rate": 1.935483870967742e-05,
336
+ "loss": 0.5686,
337
+ "step": 225
338
+ },
339
+ {
340
+ "epoch": 2.0,
341
+ "grad_norm": 5.499088287353516,
342
+ "learning_rate": 1.8548387096774193e-05,
343
+ "loss": 0.4809,
344
+ "step": 230
345
+ },
346
+ {
347
+ "epoch": 2.0,
348
+ "eval_accuracy": 0.7217391304347827,
349
+ "eval_auc": 0.8266129032258065,
350
+ "eval_f1": 0.7355371900826446,
351
+ "eval_loss": 0.5321555137634277,
352
+ "eval_precision": 0.6544117647058824,
353
+ "eval_recall": 0.839622641509434,
354
+ "eval_runtime": 33.1845,
355
+ "eval_samples_per_second": 6.931,
356
+ "eval_steps_per_second": 0.452,
357
+ "step": 230
358
+ }
359
+ ],
360
+ "logging_steps": 5,
361
+ "max_steps": 345,
362
+ "num_input_tokens_seen": 0,
363
+ "num_train_epochs": 3,
364
+ "save_steps": 500,
365
+ "stateful_callbacks": {
366
+ "EarlyStoppingCallback": {
367
+ "args": {
368
+ "early_stopping_patience": 5,
369
+ "early_stopping_threshold": 0.01
370
+ },
371
+ "attributes": {
372
+ "early_stopping_patience_counter": 0
373
+ }
374
+ },
375
+ "TrainerControl": {
376
+ "args": {
377
+ "should_epoch_stop": false,
378
+ "should_evaluate": false,
379
+ "should_log": false,
380
+ "should_save": true,
381
+ "should_training_stop": false
382
+ },
383
+ "attributes": {}
384
+ }
385
+ },
386
+ "total_flos": 150697802919936.0,
387
+ "train_batch_size": 8,
388
+ "trial_name": null,
389
+ "trial_params": null
390
+ }
checkpoint-230/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:554915da28fbd19adcee6c63005ebe3c27c6d5174ad2ddb01c22b93fd5948bfb
3
+ size 5368
config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "studio-ousia/luke-japanese-base-lite",
3
+ "_num_labels": 2,
4
+ "architectures": [
5
+ "LukeForSequenceClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "bert_model_name": "models/luke-japanese/hf_xlm_roberta",
9
+ "bos_token_id": 0,
10
+ "classifier_dropout": null,
11
+ "cls_entity_prediction": false,
12
+ "entity_emb_size": 256,
13
+ "entity_vocab_size": 4,
14
+ "eos_token_id": 2,
15
+ "hidden_act": "gelu",
16
+ "hidden_dropout_prob": 0.1,
17
+ "hidden_size": 768,
18
+ "id2label": {
19
+ "0": "\u60aa\u3044",
20
+ "1": "\u826f\u3044"
21
+ },
22
+ "initializer_range": 0.02,
23
+ "intermediate_size": 3072,
24
+ "label2id": {
25
+ "\u60aa\u3044": 0,
26
+ "\u826f\u3044": 1
27
+ },
28
+ "layer_norm_eps": 1e-05,
29
+ "max_position_embeddings": 514,
30
+ "model_type": "luke",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 1,
34
+ "position_embedding_type": "absolute",
35
+ "problem_type": "single_label_classification",
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.48.0",
38
+ "type_vocab_size": 1,
39
+ "use_cache": true,
40
+ "use_entity_aware_attention": true,
41
+ "vocab_size": 32772
42
+ }
entity_vocab.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "[MASK2]": 3,
3
+ "[MASK]": 0,
4
+ "[PAD]": 2,
5
+ "[UNK]": 1
6
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de13ef21ccae838fb30a5ec29dd128906a5bfc49538967ccc3b269ce2043e573
3
+ size 532307280
runs/Sep22_00-59-01_r-manato003-dlsite-voice-rec-ip32f285-4552b-n9vba/events.out.tfevents.1758502744.r-manato003-dlsite-voice-rec-ip32f285-4552b-n9vba.63.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d8048e4a26995dfd1949376e9ca7713d5dbc936fd649fc1280bc9724ad9bfee
3
- size 17146
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:326cc783775f1c7994c7d3e9a65884671e8d27031e0a3d35d9110857c786a62f
3
+ size 21817
runs/Sep22_00-59-01_r-manato003-dlsite-voice-rec-ip32f285-4552b-n9vba/events.out.tfevents.1758504439.r-manato003-dlsite-voice-rec-ip32f285-4552b-n9vba.63.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8336f04b736207eb0ae8da9e9ba17e2278c206ec0d6dbeb822632388da0d8f69
3
+ size 607
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8b73a5e054936c920cf5b7d1ec21ce9c281977078269963beb821c6c86fbff7
3
+ size 841889
special_tokens_map.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<ent>",
5
+ "lstrip": false,
6
+ "normalized": true,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<ent2>",
12
+ "lstrip": false,
13
+ "normalized": true,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<ent>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "<ent2>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "<ent>",
33
+ "lstrip": false,
34
+ "normalized": true,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "<ent2>",
40
+ "lstrip": false,
41
+ "normalized": true,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ }
45
+ ],
46
+ "bos_token": {
47
+ "content": "<s>",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false
52
+ },
53
+ "cls_token": {
54
+ "content": "<s>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false
59
+ },
60
+ "eos_token": {
61
+ "content": "</s>",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false
66
+ },
67
+ "mask_token": {
68
+ "content": "<mask>",
69
+ "lstrip": true,
70
+ "normalized": true,
71
+ "rstrip": false,
72
+ "single_word": false
73
+ },
74
+ "pad_token": {
75
+ "content": "<pad>",
76
+ "lstrip": false,
77
+ "normalized": false,
78
+ "rstrip": false,
79
+ "single_word": false
80
+ },
81
+ "sep_token": {
82
+ "content": "</s>",
83
+ "lstrip": false,
84
+ "normalized": false,
85
+ "rstrip": false,
86
+ "single_word": false
87
+ },
88
+ "unk_token": {
89
+ "content": "<unk>",
90
+ "lstrip": false,
91
+ "normalized": false,
92
+ "rstrip": false,
93
+ "single_word": false
94
+ }
95
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "32769": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": true,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "32770": {
44
+ "content": "<ent>",
45
+ "lstrip": false,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "32771": {
52
+ "content": "<ent2>",
53
+ "lstrip": false,
54
+ "normalized": true,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ }
59
+ },
60
+ "additional_special_tokens": [
61
+ "<ent>",
62
+ "<ent2>",
63
+ "<ent>",
64
+ "<ent2>",
65
+ "<ent>",
66
+ "<ent2>"
67
+ ],
68
+ "bos_token": "<s>",
69
+ "clean_up_tokenization_spaces": false,
70
+ "cls_token": "<s>",
71
+ "entity_mask2_token": "[MASK2]",
72
+ "entity_mask_token": "[MASK]",
73
+ "entity_pad_token": "[PAD]",
74
+ "entity_token_1": {
75
+ "__type": "AddedToken",
76
+ "content": "<ent>",
77
+ "lstrip": false,
78
+ "normalized": true,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": false
82
+ },
83
+ "entity_token_2": {
84
+ "__type": "AddedToken",
85
+ "content": "<ent2>",
86
+ "lstrip": false,
87
+ "normalized": true,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": false
91
+ },
92
+ "entity_unk_token": "[UNK]",
93
+ "eos_token": "</s>",
94
+ "extra_special_tokens": {},
95
+ "mask_token": "<mask>",
96
+ "max_entity_length": 32,
97
+ "max_mention_length": 30,
98
+ "model_max_length": 512,
99
+ "pad_token": "<pad>",
100
+ "sep_token": "</s>",
101
+ "sp_model_kwargs": {},
102
+ "task": null,
103
+ "tokenizer_class": "MLukeTokenizer",
104
+ "unk_token": "<unk>"
105
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:554915da28fbd19adcee6c63005ebe3c27c6d5174ad2ddb01c22b93fd5948bfb
3
+ size 5368
training_params.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "2classification/autotrain-data",
3
+ "model": "studio-ousia/luke-japanese-base-lite",
4
+ "lr": 5e-05,
5
+ "epochs": 3,
6
+ "max_seq_length": 128,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "text_column": "autotrain_text",
18
+ "target_column": "autotrain_label",
19
+ "logging_steps": -1,
20
+ "project_name": "2classification",
21
+ "auto_find_batch_size": false,
22
+ "mixed_precision": "fp16",
23
+ "save_total_limit": 1,
24
+ "push_to_hub": true,
25
+ "eval_strategy": "epoch",
26
+ "username": "manato003",
27
+ "log": "tensorboard",
28
+ "early_stopping_patience": 5,
29
+ "early_stopping_threshold": 0.01
30
+ }