thanhdath commited on
Commit
fb7ede6
·
verified ·
1 Parent(s): 47156af

Upload folder using huggingface_hub

Browse files
Files changed (41) hide show
  1. args.json +353 -0
  2. checkpoint-2400/README.md +206 -0
  3. checkpoint-2400/adapter_config.json +49 -0
  4. checkpoint-2400/adapter_model.safetensors +3 -0
  5. checkpoint-2400/additional_config.json +1 -0
  6. checkpoint-2400/args.json +353 -0
  7. checkpoint-2400/optimizer.pt +3 -0
  8. checkpoint-2400/rng_state_0.pth +3 -0
  9. checkpoint-2400/rng_state_1.pth +3 -0
  10. checkpoint-2400/rng_state_2.pth +3 -0
  11. checkpoint-2400/rng_state_3.pth +3 -0
  12. checkpoint-2400/rng_state_4.pth +3 -0
  13. checkpoint-2400/rng_state_5.pth +3 -0
  14. checkpoint-2400/rng_state_6.pth +3 -0
  15. checkpoint-2400/rng_state_7.pth +3 -0
  16. checkpoint-2400/scheduler.pt +3 -0
  17. checkpoint-2400/trainer_state.json +3401 -0
  18. checkpoint-2400/training_args.bin +3 -0
  19. checkpoint-2600/README.md +206 -0
  20. checkpoint-2600/adapter_config.json +49 -0
  21. checkpoint-2600/adapter_model.safetensors +3 -0
  22. checkpoint-2600/additional_config.json +1 -0
  23. checkpoint-2600/args.json +353 -0
  24. checkpoint-2600/optimizer.pt +3 -0
  25. checkpoint-2600/rng_state_0.pth +3 -0
  26. checkpoint-2600/rng_state_1.pth +3 -0
  27. checkpoint-2600/rng_state_2.pth +3 -0
  28. checkpoint-2600/rng_state_3.pth +3 -0
  29. checkpoint-2600/rng_state_4.pth +3 -0
  30. checkpoint-2600/rng_state_5.pth +3 -0
  31. checkpoint-2600/rng_state_6.pth +3 -0
  32. checkpoint-2600/rng_state_7.pth +3 -0
  33. checkpoint-2600/scheduler.pt +3 -0
  34. checkpoint-2600/trainer_state.json +0 -0
  35. checkpoint-2600/training_args.bin +3 -0
  36. images/train_epoch.png +0 -0
  37. images/train_grad_norm.png +0 -0
  38. images/train_learning_rate.png +0 -0
  39. images/train_loss.png +0 -0
  40. logging.jsonl +0 -0
  41. runs/events.out.tfevents.1767168374.server-217.1205148.0 +3 -0
args.json ADDED
@@ -0,0 +1,353 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438",
3
+ "overwrite_output_dir": false,
4
+ "do_train": false,
5
+ "do_eval": false,
6
+ "do_predict": false,
7
+ "eval_strategy": "no",
8
+ "prediction_loss_only": false,
9
+ "per_device_train_batch_size": 4,
10
+ "per_device_eval_batch_size": 1,
11
+ "per_gpu_train_batch_size": null,
12
+ "per_gpu_eval_batch_size": null,
13
+ "gradient_accumulation_steps": 2,
14
+ "eval_accumulation_steps": null,
15
+ "eval_delay": 0,
16
+ "torch_empty_cache_steps": null,
17
+ "learning_rate": 6e-06,
18
+ "weight_decay": 0.1,
19
+ "adam_beta1": 0.9,
20
+ "adam_beta2": 0.95,
21
+ "adam_epsilon": 1e-08,
22
+ "max_grad_norm": 1.0,
23
+ "num_train_epochs": 1.0,
24
+ "max_steps": -1,
25
+ "lr_scheduler_type": "cosine",
26
+ "lr_scheduler_kwargs": null,
27
+ "warmup_ratio": 0.1,
28
+ "warmup_steps": 0,
29
+ "log_level": "passive",
30
+ "log_level_replica": "warning",
31
+ "log_on_each_node": true,
32
+ "logging_dir": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438/runs",
33
+ "logging_strategy": "steps",
34
+ "logging_first_step": true,
35
+ "logging_steps": 5,
36
+ "logging_nan_inf_filter": true,
37
+ "save_strategy": "steps",
38
+ "save_steps": 200.0,
39
+ "save_total_limit": 2,
40
+ "save_safetensors": true,
41
+ "save_on_each_node": false,
42
+ "save_only_model": false,
43
+ "restore_callback_states_from_checkpoint": false,
44
+ "no_cuda": false,
45
+ "use_cpu": false,
46
+ "use_mps_device": false,
47
+ "seed": 42,
48
+ "data_seed": 42,
49
+ "jit_mode_eval": false,
50
+ "bf16": true,
51
+ "fp16": false,
52
+ "fp16_opt_level": "O1",
53
+ "half_precision_backend": "auto",
54
+ "bf16_full_eval": false,
55
+ "fp16_full_eval": false,
56
+ "tf32": null,
57
+ "local_rank": 0,
58
+ "ddp_backend": null,
59
+ "tpu_num_cores": null,
60
+ "tpu_metrics_debug": false,
61
+ "debug": null,
62
+ "dataloader_drop_last": true,
63
+ "eval_steps": 200.0,
64
+ "dataloader_num_workers": 64,
65
+ "dataloader_prefetch_factor": null,
66
+ "past_index": -1,
67
+ "run_name": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438",
68
+ "disable_tqdm": null,
69
+ "remove_unused_columns": true,
70
+ "label_names": null,
71
+ "load_best_model_at_end": false,
72
+ "metric_for_best_model": "loss",
73
+ "greater_is_better": false,
74
+ "ignore_data_skip": false,
75
+ "fsdp": [],
76
+ "fsdp_min_num_params": 0,
77
+ "fsdp_config": null,
78
+ "fsdp_transformer_layer_cls_to_wrap": null,
79
+ "accelerator_config": {
80
+ "dispatch_batches": false
81
+ },
82
+ "parallelism_config": null,
83
+ "deepspeed": null,
84
+ "label_smoothing_factor": 0.0,
85
+ "optim": "adamw_torch_fused",
86
+ "optim_args": null,
87
+ "adafactor": false,
88
+ "group_by_length": false,
89
+ "length_column_name": "length",
90
+ "report_to": [
91
+ "tensorboard"
92
+ ],
93
+ "project": "huggingface",
94
+ "trackio_space_id": "trackio",
95
+ "ddp_find_unused_parameters": null,
96
+ "ddp_bucket_cap_mb": null,
97
+ "ddp_broadcast_buffers": null,
98
+ "dataloader_pin_memory": true,
99
+ "dataloader_persistent_workers": false,
100
+ "skip_memory_metrics": true,
101
+ "use_legacy_prediction_loop": false,
102
+ "push_to_hub": false,
103
+ "resume_from_checkpoint": null,
104
+ "hub_model_id": null,
105
+ "hub_strategy": "every_save",
106
+ "hub_token": null,
107
+ "hub_private_repo": null,
108
+ "hub_always_push": false,
109
+ "hub_revision": null,
110
+ "gradient_checkpointing": true,
111
+ "gradient_checkpointing_kwargs": null,
112
+ "include_inputs_for_metrics": false,
113
+ "include_for_metrics": [],
114
+ "eval_do_concat_batches": true,
115
+ "fp16_backend": "auto",
116
+ "push_to_hub_model_id": null,
117
+ "push_to_hub_organization": null,
118
+ "push_to_hub_token": null,
119
+ "mp_parameters": "",
120
+ "auto_find_batch_size": false,
121
+ "full_determinism": false,
122
+ "torchdynamo": null,
123
+ "ray_scope": "last",
124
+ "ddp_timeout": 18000000,
125
+ "torch_compile": false,
126
+ "torch_compile_backend": null,
127
+ "torch_compile_mode": null,
128
+ "include_tokens_per_second": false,
129
+ "include_num_input_tokens_seen": false,
130
+ "neftune_noise_alpha": null,
131
+ "optim_target_modules": null,
132
+ "batch_eval_metrics": false,
133
+ "eval_on_start": false,
134
+ "use_liger_kernel": true,
135
+ "liger_kernel_config": null,
136
+ "eval_use_gather_object": false,
137
+ "average_tokens_across_devices": true,
138
+ "sortish_sampler": false,
139
+ "predict_with_generate": false,
140
+ "generation_max_length": null,
141
+ "generation_num_beams": null,
142
+ "generation_config": null,
143
+ "tuner_backend": "peft",
144
+ "vit_gradient_checkpointing": null,
145
+ "router_aux_loss_coef": 0.0,
146
+ "enable_dft_loss": false,
147
+ "enable_channel_loss": false,
148
+ "check_model": true,
149
+ "acc_strategy": "token",
150
+ "train_dataloader_shuffle": true,
151
+ "max_epochs": null,
152
+ "aligner_lr": null,
153
+ "vit_lr": null,
154
+ "use_logits_to_keep": null,
155
+ "ds3_gather_for_generation": true,
156
+ "resume_only_model": false,
157
+ "optimizer": null,
158
+ "loss_type": "infonce",
159
+ "metric": null,
160
+ "eval_use_evalscope": false,
161
+ "eval_dataset": [],
162
+ "eval_dataset_args": null,
163
+ "eval_limit": null,
164
+ "eval_generation_config": null,
165
+ "extra_eval_args": null,
166
+ "use_flash_ckpt": false,
167
+ "use_ray": false,
168
+ "ray_exp_name": null,
169
+ "device_groups": null,
170
+ "model": "Qwen/Qwen3-Embedding-4B",
171
+ "model_type": "qwen3_emb",
172
+ "model_revision": null,
173
+ "task_type": "embedding",
174
+ "torch_dtype": "bfloat16",
175
+ "attn_impl": "flash_attn",
176
+ "new_special_tokens": [],
177
+ "num_labels": null,
178
+ "problem_type": null,
179
+ "rope_scaling": null,
180
+ "device_map": null,
181
+ "max_memory": {},
182
+ "max_model_len": null,
183
+ "local_repo_path": null,
184
+ "init_strategy": null,
185
+ "template": "qwen3_emb",
186
+ "system": null,
187
+ "max_length": 1024,
188
+ "truncation_strategy": "right",
189
+ "max_pixels": null,
190
+ "agent_template": null,
191
+ "norm_bbox": null,
192
+ "use_chat_template": true,
193
+ "padding_side": "right",
194
+ "padding_free": false,
195
+ "loss_scale": "default",
196
+ "sequence_parallel_size": 1,
197
+ "template_backend": "swift",
198
+ "response_prefix": null,
199
+ "enable_thinking": null,
200
+ "add_non_thinking_prefix": true,
201
+ "dataset": [
202
+ "../output.jsonl"
203
+ ],
204
+ "val_dataset": [],
205
+ "cached_dataset": [],
206
+ "cached_val_dataset": [],
207
+ "split_dataset_ratio": 0.0,
208
+ "dataset_num_proc": 64,
209
+ "load_from_cache_file": true,
210
+ "dataset_shuffle": true,
211
+ "val_dataset_shuffle": false,
212
+ "streaming": false,
213
+ "interleave_prob": null,
214
+ "stopping_strategy": "first_exhausted",
215
+ "shuffle_buffer_size": 1000,
216
+ "download_mode": "reuse_dataset_if_exists",
217
+ "columns": {},
218
+ "strict": false,
219
+ "model_name": null,
220
+ "model_author": null,
221
+ "custom_dataset_info": [],
222
+ "quant_method": null,
223
+ "quant_bits": null,
224
+ "hqq_axis": null,
225
+ "bnb_4bit_compute_dtype": "bfloat16",
226
+ "bnb_4bit_quant_type": "nf4",
227
+ "bnb_4bit_use_double_quant": true,
228
+ "bnb_4bit_quant_storage": null,
229
+ "max_new_tokens": 64,
230
+ "temperature": 0.0,
231
+ "top_k": null,
232
+ "top_p": null,
233
+ "repetition_penalty": null,
234
+ "num_beams": 1,
235
+ "stream": false,
236
+ "stop_words": [],
237
+ "logprobs": false,
238
+ "top_logprobs": null,
239
+ "structured_outputs_regex": null,
240
+ "ckpt_dir": null,
241
+ "lora_modules": [],
242
+ "train_type": "lora",
243
+ "adapters": [],
244
+ "external_plugins": [],
245
+ "model_kwargs": {},
246
+ "load_args": false,
247
+ "load_data_args": false,
248
+ "packing": false,
249
+ "packing_length": null,
250
+ "packing_num_proc": 1,
251
+ "lazy_tokenize": false,
252
+ "custom_register_path": [],
253
+ "use_hf": false,
254
+ "ignore_args_error": false,
255
+ "use_swift_lora": false,
256
+ "freeze_parameters": [],
257
+ "freeze_parameters_regex": null,
258
+ "freeze_parameters_ratio": 0.0,
259
+ "trainable_parameters": [],
260
+ "trainable_parameters_regex": null,
261
+ "freeze_llm": false,
262
+ "freeze_vit": true,
263
+ "freeze_aligner": true,
264
+ "target_modules": [
265
+ "all-linear"
266
+ ],
267
+ "target_regex": null,
268
+ "target_parameters": null,
269
+ "modules_to_save": [],
270
+ "lora_rank": 8,
271
+ "lora_alpha": 32,
272
+ "lora_dropout": 0.05,
273
+ "lora_bias": "none",
274
+ "lora_dtype": null,
275
+ "lorap_lr_ratio": null,
276
+ "use_rslora": false,
277
+ "use_dora": false,
278
+ "lora_ga_batch_size": 2,
279
+ "lora_ga_iters": 2,
280
+ "lora_ga_max_length": 1024,
281
+ "lora_ga_direction": "ArB2r",
282
+ "lora_ga_scale": "stable",
283
+ "lora_ga_stable_gamma": 16,
284
+ "init_weights": true,
285
+ "fourier_n_frequency": 2000,
286
+ "fourier_scaling": 300.0,
287
+ "boft_block_size": 4,
288
+ "boft_block_num": 0,
289
+ "boft_n_butterfly_factor": 1,
290
+ "boft_dropout": 0.0,
291
+ "vera_rank": 256,
292
+ "vera_projection_prng_key": 0,
293
+ "vera_dropout": 0.0,
294
+ "vera_d_initial": 0.1,
295
+ "adapter_act": "gelu",
296
+ "adapter_length": 128,
297
+ "use_galore": false,
298
+ "galore_target_modules": null,
299
+ "galore_rank": 128,
300
+ "galore_update_proj_gap": 50,
301
+ "galore_scale": 1.0,
302
+ "galore_proj_type": "std",
303
+ "galore_optim_per_parameter": false,
304
+ "galore_with_embedding": false,
305
+ "galore_quantization": false,
306
+ "galore_proj_quant": false,
307
+ "galore_proj_bits": 4,
308
+ "galore_proj_group_size": 256,
309
+ "galore_cos_threshold": 0.4,
310
+ "galore_gamma_proj": 2,
311
+ "galore_queue_size": 5,
312
+ "adalora_target_r": 8,
313
+ "adalora_init_r": 12,
314
+ "adalora_tinit": 0,
315
+ "adalora_tfinal": 0,
316
+ "adalora_deltaT": 1,
317
+ "adalora_beta1": 0.85,
318
+ "adalora_beta2": 0.85,
319
+ "adalora_orth_reg_weight": 0.5,
320
+ "llamapro_num_new_blocks": 4,
321
+ "llamapro_num_groups": null,
322
+ "lisa_activated_layers": 0,
323
+ "lisa_step_interval": 20,
324
+ "reft_layer_key": null,
325
+ "reft_layers": null,
326
+ "reft_rank": 4,
327
+ "reft_intervention_type": "LoreftIntervention",
328
+ "reft_args": null,
329
+ "swanlab_token": null,
330
+ "swanlab_project": "ms-swift",
331
+ "swanlab_workspace": null,
332
+ "swanlab_exp_name": null,
333
+ "swanlab_notification_method": null,
334
+ "swanlab_webhook_url": null,
335
+ "swanlab_secret": null,
336
+ "swanlab_mode": "cloud",
337
+ "add_version": true,
338
+ "create_checkpoint_symlink": false,
339
+ "zero_hpz_partition_size": null,
340
+ "deepspeed_autotp_size": null,
341
+ "early_stop_interval": null,
342
+ "rank": 0,
343
+ "global_world_size": 8,
344
+ "local_world_size": 8,
345
+ "model_suffix": "Qwen3-Embedding-4B",
346
+ "model_info": "ModelInfo(model_type='qwen3_emb', model_dir='/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B', torch_dtype=torch.bfloat16, max_model_len=40960, quant_method=None, quant_bits=None, rope_scaling=None, is_moe_model=False, is_multimodal=False, config=None, task_type='embedding', num_labels=None)",
347
+ "model_meta": "ModelMeta(model_type='qwen3_emb', model_groups=[ModelGroup(models=[Model(ms_model_id='Qwen/Qwen3-Embedding-0.6B', hf_model_id='Qwen/Qwen3-Embedding-0.6B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-Embedding-4B', hf_model_id='Qwen/Qwen3-Embedding-4B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-Embedding-8B', hf_model_id='Qwen/Qwen3-Embedding-8B', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[])], template='qwen3_emb', get_function=<function get_model_tokenizer_with_flash_attn at 0x7f415201c5e0>, model_arch=None, architectures=['Qwen3ForCausalLM'], additional_saved_files=['config_sentence_transformers.json', '1_Pooling', 'modules.json'], torch_dtype=None, is_multimodal=False, is_reward=False, is_reranker=False, task_type=None, ignore_patterns=None, requires=[], tags=[])",
348
+ "model_dir": "/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B",
349
+ "_val_dataset_exists": [],
350
+ "hub": "<class 'swift.hub.hub.MSHub'>",
351
+ "evaluation_strategy": "steps",
352
+ "training_args": "TrainingArguments(output_dir='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438', overwrite_output_dir=False, do_train=False, do_eval=False, do_predict=False, eval_strategy=<IntervalStrategy.NO: 'no'>, prediction_loss_only=False, per_device_train_batch_size=4, per_device_eval_batch_size=1, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=2, eval_accumulation_steps=None, eval_delay=0, torch_empty_cache_steps=None, learning_rate=6e-06, weight_decay=0.1, adam_beta1=0.9, adam_beta2=0.95, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=1.0, max_steps=-1, lr_scheduler_type=<SchedulerType.COSINE: 'cosine'>, lr_scheduler_kwargs=None, warmup_ratio=0.1, warmup_steps=0, log_level='passive', log_level_replica='warning', log_on_each_node=True, logging_dir='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438/runs', logging_strategy=<IntervalStrategy.STEPS: 'steps'>, logging_first_step=True, logging_steps=5, logging_nan_inf_filter=True, save_strategy=<SaveStrategy.STEPS: 'steps'>, save_steps=200, save_total_limit=2, save_safetensors=True, save_on_each_node=False, save_only_model=False, restore_callback_states_from_checkpoint=False, no_cuda=False, use_cpu=False, use_mps_device=False, seed=42, data_seed=42, jit_mode_eval=False, bf16=True, fp16=False, fp16_opt_level='O1', half_precision_backend='auto', bf16_full_eval=False, fp16_full_eval=False, tf32=None, local_rank=0, ddp_backend=None, tpu_num_cores=None, tpu_metrics_debug=False, debug=[], dataloader_drop_last=True, eval_steps=200.0, dataloader_num_workers=64, dataloader_prefetch_factor=2, past_index=-1, run_name='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438', disable_tqdm=False, remove_unused_columns=False, label_names=None, load_best_model_at_end=False, metric_for_best_model='loss', greater_is_better=False, ignore_data_skip=False, fsdp=[], fsdp_min_num_params=0, fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, fsdp_transformer_layer_cls_to_wrap=None, accelerator_config=AcceleratorConfig(split_batches=False, dispatch_batches=False, even_batches=True, use_seedable_sampler=True, non_blocking=False, gradient_accumulation_kwargs=None, use_configured_state=False), parallelism_config=None, deepspeed=None, label_smoothing_factor=0.0, optim=<OptimizerNames.ADAMW_TORCH_FUSED: 'adamw_torch_fused'>, optim_args=None, adafactor=False, group_by_length=False, length_column_name='length', report_to=['tensorboard'], project='huggingface', trackio_space_id='trackio', ddp_find_unused_parameters=None, ddp_bucket_cap_mb=None, ddp_broadcast_buffers=None, dataloader_pin_memory=True, dataloader_persistent_workers=False, skip_memory_metrics=True, use_legacy_prediction_loop=False, push_to_hub=False, resume_from_checkpoint=None, hub_model_id=None, hub_strategy=<HubStrategy.EVERY_SAVE: 'every_save'>, hub_token=None, hub_private_repo=None, hub_always_push=False, hub_revision=None, gradient_checkpointing=True, gradient_checkpointing_kwargs=None, include_inputs_for_metrics=False, include_for_metrics=[], eval_do_concat_batches=True, fp16_backend='auto', push_to_hub_model_id=None, push_to_hub_organization=None, push_to_hub_token=None, mp_parameters='', auto_find_batch_size=False, full_determinism=False, torchdynamo=None, ray_scope='last', ddp_timeout=18000000, torch_compile=False, torch_compile_backend=None, torch_compile_mode=None, include_tokens_per_second=None, include_num_input_tokens_seen=None, neftune_noise_alpha=None, optim_target_modules=None, batch_eval_metrics=False, eval_on_start=False, use_liger_kernel=True, liger_kernel_config=None, eval_use_gather_object=False, average_tokens_across_devices=None, tuner_backend='peft', vit_gradient_checkpointing=True, router_aux_loss_coef=0.0, enable_dft_loss=False, enable_channel_loss=False, check_model=True, acc_strategy='token', train_dataloader_shuffle=True, max_epochs=None, aligner_lr=None, vit_lr=None, use_logits_to_keep=None, ds3_gather_for_generation=True, resume_only_model=False, optimizer=None, loss_type='infonce', metric=None, eval_use_evalscope=False, eval_dataset=[], eval_dataset_args=None, eval_limit=None, eval_generation_config=None, extra_eval_args=None, use_flash_ckpt=False, sft_alpha=0, chord_sft_dataset=[], chord_sft_per_device_train_batch_size=None, chord_enable_phi_function=False, chord_mu_warmup_steps=None, chord_mu_decay_steps=None, chord_mu_peak=None, chord_mu_valley=None, train_type='lora', local_repo_path=None, galore_config=None, task_type='embedding', problem_type=None)"
353
+ }
checkpoint-2400/README.md ADDED
@@ -0,0 +1,206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B
6
+ - lora
7
+ - transformers
8
+ ---
9
+
10
+ # Model Card for Model ID
11
+
12
+ <!-- Provide a quick summary of what the model is/does. -->
13
+
14
+
15
+
16
+ ## Model Details
17
+
18
+ ### Model Description
19
+
20
+ <!-- Provide a longer summary of what this model is. -->
21
+
22
+
23
+
24
+ - **Developed by:** [More Information Needed]
25
+ - **Funded by [optional]:** [More Information Needed]
26
+ - **Shared by [optional]:** [More Information Needed]
27
+ - **Model type:** [More Information Needed]
28
+ - **Language(s) (NLP):** [More Information Needed]
29
+ - **License:** [More Information Needed]
30
+ - **Finetuned from model [optional]:** [More Information Needed]
31
+
32
+ ### Model Sources [optional]
33
+
34
+ <!-- Provide the basic links for the model. -->
35
+
36
+ - **Repository:** [More Information Needed]
37
+ - **Paper [optional]:** [More Information Needed]
38
+ - **Demo [optional]:** [More Information Needed]
39
+
40
+ ## Uses
41
+
42
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
43
+
44
+ ### Direct Use
45
+
46
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
47
+
48
+ [More Information Needed]
49
+
50
+ ### Downstream Use [optional]
51
+
52
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
53
+
54
+ [More Information Needed]
55
+
56
+ ### Out-of-Scope Use
57
+
58
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
59
+
60
+ [More Information Needed]
61
+
62
+ ## Bias, Risks, and Limitations
63
+
64
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
65
+
66
+ [More Information Needed]
67
+
68
+ ### Recommendations
69
+
70
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
71
+
72
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
73
+
74
+ ## How to Get Started with the Model
75
+
76
+ Use the code below to get started with the model.
77
+
78
+ [More Information Needed]
79
+
80
+ ## Training Details
81
+
82
+ ### Training Data
83
+
84
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
85
+
86
+ [More Information Needed]
87
+
88
+ ### Training Procedure
89
+
90
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
91
+
92
+ #### Preprocessing [optional]
93
+
94
+ [More Information Needed]
95
+
96
+
97
+ #### Training Hyperparameters
98
+
99
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
100
+
101
+ #### Speeds, Sizes, Times [optional]
102
+
103
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
104
+
105
+ [More Information Needed]
106
+
107
+ ## Evaluation
108
+
109
+ <!-- This section describes the evaluation protocols and provides the results. -->
110
+
111
+ ### Testing Data, Factors & Metrics
112
+
113
+ #### Testing Data
114
+
115
+ <!-- This should link to a Dataset Card if possible. -->
116
+
117
+ [More Information Needed]
118
+
119
+ #### Factors
120
+
121
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
122
+
123
+ [More Information Needed]
124
+
125
+ #### Metrics
126
+
127
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
128
+
129
+ [More Information Needed]
130
+
131
+ ### Results
132
+
133
+ [More Information Needed]
134
+
135
+ #### Summary
136
+
137
+
138
+
139
+ ## Model Examination [optional]
140
+
141
+ <!-- Relevant interpretability work for the model goes here -->
142
+
143
+ [More Information Needed]
144
+
145
+ ## Environmental Impact
146
+
147
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
148
+
149
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
150
+
151
+ - **Hardware Type:** [More Information Needed]
152
+ - **Hours used:** [More Information Needed]
153
+ - **Cloud Provider:** [More Information Needed]
154
+ - **Compute Region:** [More Information Needed]
155
+ - **Carbon Emitted:** [More Information Needed]
156
+
157
+ ## Technical Specifications [optional]
158
+
159
+ ### Model Architecture and Objective
160
+
161
+ [More Information Needed]
162
+
163
+ ### Compute Infrastructure
164
+
165
+ [More Information Needed]
166
+
167
+ #### Hardware
168
+
169
+ [More Information Needed]
170
+
171
+ #### Software
172
+
173
+ [More Information Needed]
174
+
175
+ ## Citation [optional]
176
+
177
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
178
+
179
+ **BibTeX:**
180
+
181
+ [More Information Needed]
182
+
183
+ **APA:**
184
+
185
+ [More Information Needed]
186
+
187
+ ## Glossary [optional]
188
+
189
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
190
+
191
+ [More Information Needed]
192
+
193
+ ## More Information [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Authors [optional]
198
+
199
+ [More Information Needed]
200
+
201
+ ## Model Card Contact
202
+
203
+ [More Information Needed]
204
+ ### Framework versions
205
+
206
+ - PEFT 0.18.0
checkpoint-2400/adapter_config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "Qwen3ForCausalLM",
7
+ "parent_library": "transformers.models.qwen3.modeling_qwen3"
8
+ },
9
+ "base_model_name_or_path": "/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B",
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": true,
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 32,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.05,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": [],
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.0",
30
+ "qalora_group_size": 16,
31
+ "r": 8,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "k_proj",
36
+ "o_proj",
37
+ "down_proj",
38
+ "v_proj",
39
+ "gate_proj",
40
+ "up_proj",
41
+ "q_proj"
42
+ ],
43
+ "target_parameters": null,
44
+ "task_type": null,
45
+ "trainable_token_indices": null,
46
+ "use_dora": false,
47
+ "use_qalora": false,
48
+ "use_rslora": false
49
+ }
checkpoint-2400/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:380f74ab2e524fb27adcef5219c053fc6c42fa87f6be8f89059d2a75f0ee1538
3
+ size 66126768
checkpoint-2400/additional_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lora_dtype": null, "lorap_lr_ratio": null, "lorap_emb_lr": 1e-06}
checkpoint-2400/args.json ADDED
@@ -0,0 +1,353 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438",
3
+ "overwrite_output_dir": false,
4
+ "do_train": false,
5
+ "do_eval": false,
6
+ "do_predict": false,
7
+ "eval_strategy": "no",
8
+ "prediction_loss_only": false,
9
+ "per_device_train_batch_size": 4,
10
+ "per_device_eval_batch_size": 1,
11
+ "per_gpu_train_batch_size": null,
12
+ "per_gpu_eval_batch_size": null,
13
+ "gradient_accumulation_steps": 2,
14
+ "eval_accumulation_steps": null,
15
+ "eval_delay": 0,
16
+ "torch_empty_cache_steps": null,
17
+ "learning_rate": 6e-06,
18
+ "weight_decay": 0.1,
19
+ "adam_beta1": 0.9,
20
+ "adam_beta2": 0.95,
21
+ "adam_epsilon": 1e-08,
22
+ "max_grad_norm": 1.0,
23
+ "num_train_epochs": 1.0,
24
+ "max_steps": -1,
25
+ "lr_scheduler_type": "cosine",
26
+ "lr_scheduler_kwargs": null,
27
+ "warmup_ratio": 0.1,
28
+ "warmup_steps": 0,
29
+ "log_level": "passive",
30
+ "log_level_replica": "warning",
31
+ "log_on_each_node": true,
32
+ "logging_dir": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438/runs",
33
+ "logging_strategy": "steps",
34
+ "logging_first_step": true,
35
+ "logging_steps": 5,
36
+ "logging_nan_inf_filter": true,
37
+ "save_strategy": "steps",
38
+ "save_steps": 200.0,
39
+ "save_total_limit": 2,
40
+ "save_safetensors": true,
41
+ "save_on_each_node": false,
42
+ "save_only_model": false,
43
+ "restore_callback_states_from_checkpoint": false,
44
+ "no_cuda": false,
45
+ "use_cpu": false,
46
+ "use_mps_device": false,
47
+ "seed": 42,
48
+ "data_seed": 42,
49
+ "jit_mode_eval": false,
50
+ "bf16": true,
51
+ "fp16": false,
52
+ "fp16_opt_level": "O1",
53
+ "half_precision_backend": "auto",
54
+ "bf16_full_eval": false,
55
+ "fp16_full_eval": false,
56
+ "tf32": null,
57
+ "local_rank": 0,
58
+ "ddp_backend": null,
59
+ "tpu_num_cores": null,
60
+ "tpu_metrics_debug": false,
61
+ "debug": null,
62
+ "dataloader_drop_last": true,
63
+ "eval_steps": 200.0,
64
+ "dataloader_num_workers": 64,
65
+ "dataloader_prefetch_factor": null,
66
+ "past_index": -1,
67
+ "run_name": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438",
68
+ "disable_tqdm": null,
69
+ "remove_unused_columns": true,
70
+ "label_names": null,
71
+ "load_best_model_at_end": false,
72
+ "metric_for_best_model": "loss",
73
+ "greater_is_better": false,
74
+ "ignore_data_skip": false,
75
+ "fsdp": [],
76
+ "fsdp_min_num_params": 0,
77
+ "fsdp_config": null,
78
+ "fsdp_transformer_layer_cls_to_wrap": null,
79
+ "accelerator_config": {
80
+ "dispatch_batches": false
81
+ },
82
+ "parallelism_config": null,
83
+ "deepspeed": null,
84
+ "label_smoothing_factor": 0.0,
85
+ "optim": "adamw_torch_fused",
86
+ "optim_args": null,
87
+ "adafactor": false,
88
+ "group_by_length": false,
89
+ "length_column_name": "length",
90
+ "report_to": [
91
+ "tensorboard"
92
+ ],
93
+ "project": "huggingface",
94
+ "trackio_space_id": "trackio",
95
+ "ddp_find_unused_parameters": null,
96
+ "ddp_bucket_cap_mb": null,
97
+ "ddp_broadcast_buffers": null,
98
+ "dataloader_pin_memory": true,
99
+ "dataloader_persistent_workers": false,
100
+ "skip_memory_metrics": true,
101
+ "use_legacy_prediction_loop": false,
102
+ "push_to_hub": false,
103
+ "resume_from_checkpoint": null,
104
+ "hub_model_id": null,
105
+ "hub_strategy": "every_save",
106
+ "hub_token": null,
107
+ "hub_private_repo": null,
108
+ "hub_always_push": false,
109
+ "hub_revision": null,
110
+ "gradient_checkpointing": true,
111
+ "gradient_checkpointing_kwargs": null,
112
+ "include_inputs_for_metrics": false,
113
+ "include_for_metrics": [],
114
+ "eval_do_concat_batches": true,
115
+ "fp16_backend": "auto",
116
+ "push_to_hub_model_id": null,
117
+ "push_to_hub_organization": null,
118
+ "push_to_hub_token": null,
119
+ "mp_parameters": "",
120
+ "auto_find_batch_size": false,
121
+ "full_determinism": false,
122
+ "torchdynamo": null,
123
+ "ray_scope": "last",
124
+ "ddp_timeout": 18000000,
125
+ "torch_compile": false,
126
+ "torch_compile_backend": null,
127
+ "torch_compile_mode": null,
128
+ "include_tokens_per_second": false,
129
+ "include_num_input_tokens_seen": false,
130
+ "neftune_noise_alpha": null,
131
+ "optim_target_modules": null,
132
+ "batch_eval_metrics": false,
133
+ "eval_on_start": false,
134
+ "use_liger_kernel": true,
135
+ "liger_kernel_config": null,
136
+ "eval_use_gather_object": false,
137
+ "average_tokens_across_devices": true,
138
+ "sortish_sampler": false,
139
+ "predict_with_generate": false,
140
+ "generation_max_length": null,
141
+ "generation_num_beams": null,
142
+ "generation_config": null,
143
+ "tuner_backend": "peft",
144
+ "vit_gradient_checkpointing": null,
145
+ "router_aux_loss_coef": 0.0,
146
+ "enable_dft_loss": false,
147
+ "enable_channel_loss": false,
148
+ "check_model": true,
149
+ "acc_strategy": "token",
150
+ "train_dataloader_shuffle": true,
151
+ "max_epochs": null,
152
+ "aligner_lr": null,
153
+ "vit_lr": null,
154
+ "use_logits_to_keep": null,
155
+ "ds3_gather_for_generation": true,
156
+ "resume_only_model": false,
157
+ "optimizer": null,
158
+ "loss_type": "infonce",
159
+ "metric": null,
160
+ "eval_use_evalscope": false,
161
+ "eval_dataset": [],
162
+ "eval_dataset_args": null,
163
+ "eval_limit": null,
164
+ "eval_generation_config": null,
165
+ "extra_eval_args": null,
166
+ "use_flash_ckpt": false,
167
+ "use_ray": false,
168
+ "ray_exp_name": null,
169
+ "device_groups": null,
170
+ "model": "Qwen/Qwen3-Embedding-4B",
171
+ "model_type": "qwen3_emb",
172
+ "model_revision": null,
173
+ "task_type": "embedding",
174
+ "torch_dtype": "bfloat16",
175
+ "attn_impl": "flash_attn",
176
+ "new_special_tokens": [],
177
+ "num_labels": null,
178
+ "problem_type": null,
179
+ "rope_scaling": null,
180
+ "device_map": null,
181
+ "max_memory": {},
182
+ "max_model_len": null,
183
+ "local_repo_path": null,
184
+ "init_strategy": null,
185
+ "template": "qwen3_emb",
186
+ "system": null,
187
+ "max_length": 1024,
188
+ "truncation_strategy": "right",
189
+ "max_pixels": null,
190
+ "agent_template": null,
191
+ "norm_bbox": null,
192
+ "use_chat_template": true,
193
+ "padding_side": "right",
194
+ "padding_free": false,
195
+ "loss_scale": "default",
196
+ "sequence_parallel_size": 1,
197
+ "template_backend": "swift",
198
+ "response_prefix": null,
199
+ "enable_thinking": null,
200
+ "add_non_thinking_prefix": true,
201
+ "dataset": [
202
+ "../output.jsonl"
203
+ ],
204
+ "val_dataset": [],
205
+ "cached_dataset": [],
206
+ "cached_val_dataset": [],
207
+ "split_dataset_ratio": 0.0,
208
+ "dataset_num_proc": 64,
209
+ "load_from_cache_file": true,
210
+ "dataset_shuffle": true,
211
+ "val_dataset_shuffle": false,
212
+ "streaming": false,
213
+ "interleave_prob": null,
214
+ "stopping_strategy": "first_exhausted",
215
+ "shuffle_buffer_size": 1000,
216
+ "download_mode": "reuse_dataset_if_exists",
217
+ "columns": {},
218
+ "strict": false,
219
+ "model_name": null,
220
+ "model_author": null,
221
+ "custom_dataset_info": [],
222
+ "quant_method": null,
223
+ "quant_bits": null,
224
+ "hqq_axis": null,
225
+ "bnb_4bit_compute_dtype": "bfloat16",
226
+ "bnb_4bit_quant_type": "nf4",
227
+ "bnb_4bit_use_double_quant": true,
228
+ "bnb_4bit_quant_storage": null,
229
+ "max_new_tokens": 64,
230
+ "temperature": 0.0,
231
+ "top_k": null,
232
+ "top_p": null,
233
+ "repetition_penalty": null,
234
+ "num_beams": 1,
235
+ "stream": false,
236
+ "stop_words": [],
237
+ "logprobs": false,
238
+ "top_logprobs": null,
239
+ "structured_outputs_regex": null,
240
+ "ckpt_dir": null,
241
+ "lora_modules": [],
242
+ "train_type": "lora",
243
+ "adapters": [],
244
+ "external_plugins": [],
245
+ "model_kwargs": {},
246
+ "load_args": false,
247
+ "load_data_args": false,
248
+ "packing": false,
249
+ "packing_length": null,
250
+ "packing_num_proc": 1,
251
+ "lazy_tokenize": false,
252
+ "custom_register_path": [],
253
+ "use_hf": false,
254
+ "ignore_args_error": false,
255
+ "use_swift_lora": false,
256
+ "freeze_parameters": [],
257
+ "freeze_parameters_regex": null,
258
+ "freeze_parameters_ratio": 0.0,
259
+ "trainable_parameters": [],
260
+ "trainable_parameters_regex": null,
261
+ "freeze_llm": false,
262
+ "freeze_vit": true,
263
+ "freeze_aligner": true,
264
+ "target_modules": [
265
+ "all-linear"
266
+ ],
267
+ "target_regex": null,
268
+ "target_parameters": null,
269
+ "modules_to_save": [],
270
+ "lora_rank": 8,
271
+ "lora_alpha": 32,
272
+ "lora_dropout": 0.05,
273
+ "lora_bias": "none",
274
+ "lora_dtype": null,
275
+ "lorap_lr_ratio": null,
276
+ "use_rslora": false,
277
+ "use_dora": false,
278
+ "lora_ga_batch_size": 2,
279
+ "lora_ga_iters": 2,
280
+ "lora_ga_max_length": 1024,
281
+ "lora_ga_direction": "ArB2r",
282
+ "lora_ga_scale": "stable",
283
+ "lora_ga_stable_gamma": 16,
284
+ "init_weights": true,
285
+ "fourier_n_frequency": 2000,
286
+ "fourier_scaling": 300.0,
287
+ "boft_block_size": 4,
288
+ "boft_block_num": 0,
289
+ "boft_n_butterfly_factor": 1,
290
+ "boft_dropout": 0.0,
291
+ "vera_rank": 256,
292
+ "vera_projection_prng_key": 0,
293
+ "vera_dropout": 0.0,
294
+ "vera_d_initial": 0.1,
295
+ "adapter_act": "gelu",
296
+ "adapter_length": 128,
297
+ "use_galore": false,
298
+ "galore_target_modules": null,
299
+ "galore_rank": 128,
300
+ "galore_update_proj_gap": 50,
301
+ "galore_scale": 1.0,
302
+ "galore_proj_type": "std",
303
+ "galore_optim_per_parameter": false,
304
+ "galore_with_embedding": false,
305
+ "galore_quantization": false,
306
+ "galore_proj_quant": false,
307
+ "galore_proj_bits": 4,
308
+ "galore_proj_group_size": 256,
309
+ "galore_cos_threshold": 0.4,
310
+ "galore_gamma_proj": 2,
311
+ "galore_queue_size": 5,
312
+ "adalora_target_r": 8,
313
+ "adalora_init_r": 12,
314
+ "adalora_tinit": 0,
315
+ "adalora_tfinal": 0,
316
+ "adalora_deltaT": 1,
317
+ "adalora_beta1": 0.85,
318
+ "adalora_beta2": 0.85,
319
+ "adalora_orth_reg_weight": 0.5,
320
+ "llamapro_num_new_blocks": 4,
321
+ "llamapro_num_groups": null,
322
+ "lisa_activated_layers": 0,
323
+ "lisa_step_interval": 20,
324
+ "reft_layer_key": null,
325
+ "reft_layers": null,
326
+ "reft_rank": 4,
327
+ "reft_intervention_type": "LoreftIntervention",
328
+ "reft_args": null,
329
+ "swanlab_token": null,
330
+ "swanlab_project": "ms-swift",
331
+ "swanlab_workspace": null,
332
+ "swanlab_exp_name": null,
333
+ "swanlab_notification_method": null,
334
+ "swanlab_webhook_url": null,
335
+ "swanlab_secret": null,
336
+ "swanlab_mode": "cloud",
337
+ "add_version": true,
338
+ "create_checkpoint_symlink": false,
339
+ "zero_hpz_partition_size": null,
340
+ "deepspeed_autotp_size": null,
341
+ "early_stop_interval": null,
342
+ "rank": 0,
343
+ "global_world_size": 8,
344
+ "local_world_size": 8,
345
+ "model_suffix": "Qwen3-Embedding-4B",
346
+ "model_info": "ModelInfo(model_type='qwen3_emb', model_dir='/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B', torch_dtype=torch.bfloat16, max_model_len=40960, quant_method=None, quant_bits=None, rope_scaling=None, is_moe_model=False, is_multimodal=False, config=None, task_type='embedding', num_labels=None)",
347
+ "model_meta": "ModelMeta(model_type='qwen3_emb', model_groups=[ModelGroup(models=[Model(ms_model_id='Qwen/Qwen3-Embedding-0.6B', hf_model_id='Qwen/Qwen3-Embedding-0.6B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-Embedding-4B', hf_model_id='Qwen/Qwen3-Embedding-4B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-Embedding-8B', hf_model_id='Qwen/Qwen3-Embedding-8B', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[])], template='qwen3_emb', get_function=<function get_model_tokenizer_with_flash_attn at 0x7f415201c5e0>, model_arch=None, architectures=['Qwen3ForCausalLM'], additional_saved_files=['config_sentence_transformers.json', '1_Pooling', 'modules.json'], torch_dtype=None, is_multimodal=False, is_reward=False, is_reranker=False, task_type=None, ignore_patterns=None, requires=[], tags=[])",
348
+ "model_dir": "/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B",
349
+ "_val_dataset_exists": [],
350
+ "hub": "<class 'swift.hub.hub.MSHub'>",
351
+ "evaluation_strategy": "steps",
352
+ "training_args": "TrainingArguments(output_dir='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438', overwrite_output_dir=False, do_train=False, do_eval=False, do_predict=False, eval_strategy=<IntervalStrategy.NO: 'no'>, prediction_loss_only=False, per_device_train_batch_size=4, per_device_eval_batch_size=1, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=2, eval_accumulation_steps=None, eval_delay=0, torch_empty_cache_steps=None, learning_rate=6e-06, weight_decay=0.1, adam_beta1=0.9, adam_beta2=0.95, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=1.0, max_steps=-1, lr_scheduler_type=<SchedulerType.COSINE: 'cosine'>, lr_scheduler_kwargs=None, warmup_ratio=0.1, warmup_steps=0, log_level='passive', log_level_replica='warning', log_on_each_node=True, logging_dir='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438/runs', logging_strategy=<IntervalStrategy.STEPS: 'steps'>, logging_first_step=True, logging_steps=5, logging_nan_inf_filter=True, save_strategy=<SaveStrategy.STEPS: 'steps'>, save_steps=200, save_total_limit=2, save_safetensors=True, save_on_each_node=False, save_only_model=False, restore_callback_states_from_checkpoint=False, no_cuda=False, use_cpu=False, use_mps_device=False, seed=42, data_seed=42, jit_mode_eval=False, bf16=True, fp16=False, fp16_opt_level='O1', half_precision_backend='auto', bf16_full_eval=False, fp16_full_eval=False, tf32=None, local_rank=0, ddp_backend=None, tpu_num_cores=None, tpu_metrics_debug=False, debug=[], dataloader_drop_last=True, eval_steps=200.0, dataloader_num_workers=64, dataloader_prefetch_factor=2, past_index=-1, run_name='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438', disable_tqdm=False, remove_unused_columns=False, label_names=None, load_best_model_at_end=False, metric_for_best_model='loss', greater_is_better=False, ignore_data_skip=False, fsdp=[], fsdp_min_num_params=0, fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, fsdp_transformer_layer_cls_to_wrap=None, accelerator_config=AcceleratorConfig(split_batches=False, dispatch_batches=False, even_batches=True, use_seedable_sampler=True, non_blocking=False, gradient_accumulation_kwargs=None, use_configured_state=False), parallelism_config=None, deepspeed=None, label_smoothing_factor=0.0, optim=<OptimizerNames.ADAMW_TORCH_FUSED: 'adamw_torch_fused'>, optim_args=None, adafactor=False, group_by_length=False, length_column_name='length', report_to=['tensorboard'], project='huggingface', trackio_space_id='trackio', ddp_find_unused_parameters=None, ddp_bucket_cap_mb=None, ddp_broadcast_buffers=None, dataloader_pin_memory=True, dataloader_persistent_workers=False, skip_memory_metrics=True, use_legacy_prediction_loop=False, push_to_hub=False, resume_from_checkpoint=None, hub_model_id=None, hub_strategy=<HubStrategy.EVERY_SAVE: 'every_save'>, hub_token=None, hub_private_repo=None, hub_always_push=False, hub_revision=None, gradient_checkpointing=True, gradient_checkpointing_kwargs=None, include_inputs_for_metrics=False, include_for_metrics=[], eval_do_concat_batches=True, fp16_backend='auto', push_to_hub_model_id=None, push_to_hub_organization=None, push_to_hub_token=None, mp_parameters='', auto_find_batch_size=False, full_determinism=False, torchdynamo=None, ray_scope='last', ddp_timeout=18000000, torch_compile=False, torch_compile_backend=None, torch_compile_mode=None, include_tokens_per_second=None, include_num_input_tokens_seen=None, neftune_noise_alpha=None, optim_target_modules=None, batch_eval_metrics=False, eval_on_start=False, use_liger_kernel=True, liger_kernel_config=None, eval_use_gather_object=False, average_tokens_across_devices=None, tuner_backend='peft', vit_gradient_checkpointing=True, router_aux_loss_coef=0.0, enable_dft_loss=False, enable_channel_loss=False, check_model=True, acc_strategy='token', train_dataloader_shuffle=True, max_epochs=None, aligner_lr=None, vit_lr=None, use_logits_to_keep=None, ds3_gather_for_generation=True, resume_only_model=False, optimizer=None, loss_type='infonce', metric=None, eval_use_evalscope=False, eval_dataset=[], eval_dataset_args=None, eval_limit=None, eval_generation_config=None, extra_eval_args=None, use_flash_ckpt=False, sft_alpha=0, chord_sft_dataset=[], chord_sft_per_device_train_batch_size=None, chord_enable_phi_function=False, chord_mu_warmup_steps=None, chord_mu_decay_steps=None, chord_mu_peak=None, chord_mu_valley=None, train_type='lora', local_repo_path=None, galore_config=None, task_type='embedding', problem_type=None)"
353
+ }
checkpoint-2400/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbd68e5836263cd0f99a5e30a6e1317cb0c8535bb60f449f0b85a845f4ed896f
3
+ size 132552395
checkpoint-2400/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c056507717915e1f78e7bd3e55f190d78a2c8489225f02232f1d6afabc8e4c1d
3
+ size 16389
checkpoint-2400/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfa48f5e8c97c87c0aff80aacaeb6d9964d30a0b0af2896d9d9e63f267f38cdb
3
+ size 16389
checkpoint-2400/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c3549eb23ed28fbb1b3b12a8856f926f76b19676bcef7493de3d57f4ab4b95b
3
+ size 16389
checkpoint-2400/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e1c80b99fc024cdca6f289f3c88f8c4a7a03ca5a66da0009599240b5476b73f
3
+ size 16389
checkpoint-2400/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:080668cf85d24168d22641e8f7f693a1ca0dda4743a911bcc648622dc8e28130
3
+ size 16389
checkpoint-2400/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1c7b588b2c4ac7c986fa1bd39f39885fb44237b3be5fcf2e1e60046baeb4567
3
+ size 16389
checkpoint-2400/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a91da1330948506c18f97183064360f49b41afd5c70feb46c6d7223176a3ce03
3
+ size 16389
checkpoint-2400/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3427e0959cab42cbf93a87f2b3b15ffdf454f411f0489ea197c7fbddce80386d
3
+ size 16389
checkpoint-2400/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:981bd58b68048221592edf233ae4ca3c00afdf32ea4950683cfb71df829d643f
3
+ size 1465
checkpoint-2400/trainer_state.json ADDED
@@ -0,0 +1,3401 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.10710221567708682,
6
+ "eval_steps": 200.0,
7
+ "global_step": 2400,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 4.4625923198786177e-05,
14
+ "grad_norm": 0.38751721382141113,
15
+ "learning_rate": 2.677376171352075e-09,
16
+ "loss": 5.556500434875488,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.00022312961599393088,
21
+ "grad_norm": 0.37884023785591125,
22
+ "learning_rate": 1.3386880856760375e-08,
23
+ "loss": 5.5404791831970215,
24
+ "step": 5
25
+ },
26
+ {
27
+ "epoch": 0.00044625923198786177,
28
+ "grad_norm": 0.38282614946365356,
29
+ "learning_rate": 2.677376171352075e-08,
30
+ "loss": 5.552686309814453,
31
+ "step": 10
32
+ },
33
+ {
34
+ "epoch": 0.0006693888479817927,
35
+ "grad_norm": 0.3849342167377472,
36
+ "learning_rate": 4.016064257028113e-08,
37
+ "loss": 5.491200637817383,
38
+ "step": 15
39
+ },
40
+ {
41
+ "epoch": 0.0008925184639757235,
42
+ "grad_norm": 0.3901270925998688,
43
+ "learning_rate": 5.35475234270415e-08,
44
+ "loss": 5.454954147338867,
45
+ "step": 20
46
+ },
47
+ {
48
+ "epoch": 0.0011156480799696543,
49
+ "grad_norm": 0.38169801235198975,
50
+ "learning_rate": 6.693440428380188e-08,
51
+ "loss": 5.510653686523438,
52
+ "step": 25
53
+ },
54
+ {
55
+ "epoch": 0.0013387776959635854,
56
+ "grad_norm": 0.37212520837783813,
57
+ "learning_rate": 8.032128514056226e-08,
58
+ "loss": 5.510681915283203,
59
+ "step": 30
60
+ },
61
+ {
62
+ "epoch": 0.0015619073119575162,
63
+ "grad_norm": 0.37682151794433594,
64
+ "learning_rate": 9.370816599732262e-08,
65
+ "loss": 5.651955413818359,
66
+ "step": 35
67
+ },
68
+ {
69
+ "epoch": 0.001785036927951447,
70
+ "grad_norm": 0.4001241624355316,
71
+ "learning_rate": 1.07095046854083e-07,
72
+ "loss": 5.452754211425781,
73
+ "step": 40
74
+ },
75
+ {
76
+ "epoch": 0.002008166543945378,
77
+ "grad_norm": 0.405489057302475,
78
+ "learning_rate": 1.2048192771084337e-07,
79
+ "loss": 5.520745849609375,
80
+ "step": 45
81
+ },
82
+ {
83
+ "epoch": 0.0022312961599393086,
84
+ "grad_norm": 0.39249682426452637,
85
+ "learning_rate": 1.3386880856760375e-07,
86
+ "loss": 5.529154205322266,
87
+ "step": 50
88
+ },
89
+ {
90
+ "epoch": 0.0024544257759332396,
91
+ "grad_norm": 0.3897538483142853,
92
+ "learning_rate": 1.4725568942436413e-07,
93
+ "loss": 5.533833312988281,
94
+ "step": 55
95
+ },
96
+ {
97
+ "epoch": 0.0026775553919271707,
98
+ "grad_norm": 0.3601232171058655,
99
+ "learning_rate": 1.606425702811245e-07,
100
+ "loss": 5.433098220825196,
101
+ "step": 60
102
+ },
103
+ {
104
+ "epoch": 0.0029006850079211013,
105
+ "grad_norm": 0.3851630985736847,
106
+ "learning_rate": 1.7402945113788486e-07,
107
+ "loss": 5.608365631103515,
108
+ "step": 65
109
+ },
110
+ {
111
+ "epoch": 0.0031238146239150324,
112
+ "grad_norm": 0.37249451875686646,
113
+ "learning_rate": 1.8741633199464525e-07,
114
+ "loss": 5.5082344055175785,
115
+ "step": 70
116
+ },
117
+ {
118
+ "epoch": 0.003346944239908963,
119
+ "grad_norm": 0.37606358528137207,
120
+ "learning_rate": 2.0080321285140563e-07,
121
+ "loss": 5.465315246582032,
122
+ "step": 75
123
+ },
124
+ {
125
+ "epoch": 0.003570073855902894,
126
+ "grad_norm": 0.3777235448360443,
127
+ "learning_rate": 2.14190093708166e-07,
128
+ "loss": 5.494733428955078,
129
+ "step": 80
130
+ },
131
+ {
132
+ "epoch": 0.0037932034718968248,
133
+ "grad_norm": 0.3441930115222931,
134
+ "learning_rate": 2.2757697456492636e-07,
135
+ "loss": 5.509701538085937,
136
+ "step": 85
137
+ },
138
+ {
139
+ "epoch": 0.004016333087890756,
140
+ "grad_norm": 0.41749539971351624,
141
+ "learning_rate": 2.4096385542168674e-07,
142
+ "loss": 5.484718704223633,
143
+ "step": 90
144
+ },
145
+ {
146
+ "epoch": 0.004239462703884687,
147
+ "grad_norm": 0.426409512758255,
148
+ "learning_rate": 2.543507362784471e-07,
149
+ "loss": 5.485423278808594,
150
+ "step": 95
151
+ },
152
+ {
153
+ "epoch": 0.004462592319878617,
154
+ "grad_norm": 0.3890352249145508,
155
+ "learning_rate": 2.677376171352075e-07,
156
+ "loss": 5.415613174438477,
157
+ "step": 100
158
+ },
159
+ {
160
+ "epoch": 0.004685721935872548,
161
+ "grad_norm": 0.3934517502784729,
162
+ "learning_rate": 2.811244979919679e-07,
163
+ "loss": 5.502951049804688,
164
+ "step": 105
165
+ },
166
+ {
167
+ "epoch": 0.004908851551866479,
168
+ "grad_norm": 0.418043315410614,
169
+ "learning_rate": 2.9451137884872826e-07,
170
+ "loss": 5.533256149291992,
171
+ "step": 110
172
+ },
173
+ {
174
+ "epoch": 0.00513198116786041,
175
+ "grad_norm": 0.38715749979019165,
176
+ "learning_rate": 3.078982597054886e-07,
177
+ "loss": 5.538555908203125,
178
+ "step": 115
179
+ },
180
+ {
181
+ "epoch": 0.005355110783854341,
182
+ "grad_norm": 0.38706645369529724,
183
+ "learning_rate": 3.21285140562249e-07,
184
+ "loss": 5.495933151245117,
185
+ "step": 120
186
+ },
187
+ {
188
+ "epoch": 0.005578240399848272,
189
+ "grad_norm": 0.37968552112579346,
190
+ "learning_rate": 3.346720214190094e-07,
191
+ "loss": 5.507209014892578,
192
+ "step": 125
193
+ },
194
+ {
195
+ "epoch": 0.005801370015842203,
196
+ "grad_norm": 0.37994489073753357,
197
+ "learning_rate": 3.4805890227576973e-07,
198
+ "loss": 5.505420684814453,
199
+ "step": 130
200
+ },
201
+ {
202
+ "epoch": 0.006024499631836134,
203
+ "grad_norm": 0.37164536118507385,
204
+ "learning_rate": 3.614457831325301e-07,
205
+ "loss": 5.491921615600586,
206
+ "step": 135
207
+ },
208
+ {
209
+ "epoch": 0.006247629247830065,
210
+ "grad_norm": 0.3722690939903259,
211
+ "learning_rate": 3.748326639892905e-07,
212
+ "loss": 5.4893798828125,
213
+ "step": 140
214
+ },
215
+ {
216
+ "epoch": 0.006470758863823995,
217
+ "grad_norm": 0.37924614548683167,
218
+ "learning_rate": 3.882195448460509e-07,
219
+ "loss": 5.498228073120117,
220
+ "step": 145
221
+ },
222
+ {
223
+ "epoch": 0.006693888479817926,
224
+ "grad_norm": 0.4063311517238617,
225
+ "learning_rate": 4.0160642570281125e-07,
226
+ "loss": 5.492959213256836,
227
+ "step": 150
228
+ },
229
+ {
230
+ "epoch": 0.006917018095811857,
231
+ "grad_norm": 0.40316569805145264,
232
+ "learning_rate": 4.149933065595716e-07,
233
+ "loss": 5.330196762084961,
234
+ "step": 155
235
+ },
236
+ {
237
+ "epoch": 0.007140147711805788,
238
+ "grad_norm": 0.38558679819107056,
239
+ "learning_rate": 4.28380187416332e-07,
240
+ "loss": 5.483318328857422,
241
+ "step": 160
242
+ },
243
+ {
244
+ "epoch": 0.0073632773277997185,
245
+ "grad_norm": 0.3990952968597412,
246
+ "learning_rate": 4.417670682730924e-07,
247
+ "loss": 5.4153087615966795,
248
+ "step": 165
249
+ },
250
+ {
251
+ "epoch": 0.0075864069437936495,
252
+ "grad_norm": 0.41763484477996826,
253
+ "learning_rate": 4.551539491298527e-07,
254
+ "loss": 5.46466178894043,
255
+ "step": 170
256
+ },
257
+ {
258
+ "epoch": 0.007809536559787581,
259
+ "grad_norm": 0.41940048336982727,
260
+ "learning_rate": 4.6854082998661315e-07,
261
+ "loss": 5.480670166015625,
262
+ "step": 175
263
+ },
264
+ {
265
+ "epoch": 0.008032666175781512,
266
+ "grad_norm": 0.4029385447502136,
267
+ "learning_rate": 4.819277108433735e-07,
268
+ "loss": 5.44397201538086,
269
+ "step": 180
270
+ },
271
+ {
272
+ "epoch": 0.008255795791775442,
273
+ "grad_norm": 0.4002819061279297,
274
+ "learning_rate": 4.953145917001339e-07,
275
+ "loss": 5.355976867675781,
276
+ "step": 185
277
+ },
278
+ {
279
+ "epoch": 0.008478925407769374,
280
+ "grad_norm": 0.41026008129119873,
281
+ "learning_rate": 5.087014725568942e-07,
282
+ "loss": 5.4827728271484375,
283
+ "step": 190
284
+ },
285
+ {
286
+ "epoch": 0.008702055023763304,
287
+ "grad_norm": 0.4057953655719757,
288
+ "learning_rate": 5.220883534136546e-07,
289
+ "loss": 5.4723457336425785,
290
+ "step": 195
291
+ },
292
+ {
293
+ "epoch": 0.008925184639757234,
294
+ "grad_norm": 0.39414113759994507,
295
+ "learning_rate": 5.35475234270415e-07,
296
+ "loss": 5.458871841430664,
297
+ "step": 200
298
+ },
299
+ {
300
+ "epoch": 0.009148314255751166,
301
+ "grad_norm": 0.40346527099609375,
302
+ "learning_rate": 5.488621151271754e-07,
303
+ "loss": 5.332700729370117,
304
+ "step": 205
305
+ },
306
+ {
307
+ "epoch": 0.009371443871745096,
308
+ "grad_norm": 0.4204488694667816,
309
+ "learning_rate": 5.622489959839358e-07,
310
+ "loss": 5.45002326965332,
311
+ "step": 210
312
+ },
313
+ {
314
+ "epoch": 0.009594573487739028,
315
+ "grad_norm": 0.4042298495769501,
316
+ "learning_rate": 5.756358768406961e-07,
317
+ "loss": 5.469032287597656,
318
+ "step": 215
319
+ },
320
+ {
321
+ "epoch": 0.009817703103732959,
322
+ "grad_norm": 0.42059728503227234,
323
+ "learning_rate": 5.890227576974565e-07,
324
+ "loss": 5.416038513183594,
325
+ "step": 220
326
+ },
327
+ {
328
+ "epoch": 0.010040832719726889,
329
+ "grad_norm": 0.4257930517196655,
330
+ "learning_rate": 6.024096385542169e-07,
331
+ "loss": 5.365918731689453,
332
+ "step": 225
333
+ },
334
+ {
335
+ "epoch": 0.01026396233572082,
336
+ "grad_norm": 0.4297005534172058,
337
+ "learning_rate": 6.157965194109772e-07,
338
+ "loss": 5.386648941040039,
339
+ "step": 230
340
+ },
341
+ {
342
+ "epoch": 0.010487091951714751,
343
+ "grad_norm": 0.44034355878829956,
344
+ "learning_rate": 6.291834002677377e-07,
345
+ "loss": 5.4175865173339846,
346
+ "step": 235
347
+ },
348
+ {
349
+ "epoch": 0.010710221567708683,
350
+ "grad_norm": 0.4148600995540619,
351
+ "learning_rate": 6.42570281124498e-07,
352
+ "loss": 5.456137084960938,
353
+ "step": 240
354
+ },
355
+ {
356
+ "epoch": 0.010933351183702613,
357
+ "grad_norm": 0.3927938938140869,
358
+ "learning_rate": 6.559571619812583e-07,
359
+ "loss": 5.421617889404297,
360
+ "step": 245
361
+ },
362
+ {
363
+ "epoch": 0.011156480799696543,
364
+ "grad_norm": 0.46364349126815796,
365
+ "learning_rate": 6.693440428380188e-07,
366
+ "loss": 5.263519287109375,
367
+ "step": 250
368
+ },
369
+ {
370
+ "epoch": 0.011379610415690475,
371
+ "grad_norm": 0.4380209147930145,
372
+ "learning_rate": 6.827309236947791e-07,
373
+ "loss": 5.4163257598876955,
374
+ "step": 255
375
+ },
376
+ {
377
+ "epoch": 0.011602740031684405,
378
+ "grad_norm": 0.44802144169807434,
379
+ "learning_rate": 6.961178045515395e-07,
380
+ "loss": 5.307229232788086,
381
+ "step": 260
382
+ },
383
+ {
384
+ "epoch": 0.011825869647678336,
385
+ "grad_norm": 0.48721882700920105,
386
+ "learning_rate": 7.095046854082999e-07,
387
+ "loss": 5.361153030395508,
388
+ "step": 265
389
+ },
390
+ {
391
+ "epoch": 0.012048999263672268,
392
+ "grad_norm": 0.4634721875190735,
393
+ "learning_rate": 7.228915662650602e-07,
394
+ "loss": 5.335968017578125,
395
+ "step": 270
396
+ },
397
+ {
398
+ "epoch": 0.012272128879666198,
399
+ "grad_norm": 0.44632282853126526,
400
+ "learning_rate": 7.362784471218206e-07,
401
+ "loss": 5.377736282348633,
402
+ "step": 275
403
+ },
404
+ {
405
+ "epoch": 0.01249525849566013,
406
+ "grad_norm": 0.46110817790031433,
407
+ "learning_rate": 7.49665327978581e-07,
408
+ "loss": 5.379433059692383,
409
+ "step": 280
410
+ },
411
+ {
412
+ "epoch": 0.01271838811165406,
413
+ "grad_norm": 0.4558963477611542,
414
+ "learning_rate": 7.630522088353414e-07,
415
+ "loss": 5.252085876464844,
416
+ "step": 285
417
+ },
418
+ {
419
+ "epoch": 0.01294151772764799,
420
+ "grad_norm": 0.4590746760368347,
421
+ "learning_rate": 7.764390896921018e-07,
422
+ "loss": 5.320807647705078,
423
+ "step": 290
424
+ },
425
+ {
426
+ "epoch": 0.013164647343641922,
427
+ "grad_norm": 0.458065927028656,
428
+ "learning_rate": 7.898259705488621e-07,
429
+ "loss": 5.342620086669922,
430
+ "step": 295
431
+ },
432
+ {
433
+ "epoch": 0.013387776959635852,
434
+ "grad_norm": 0.47713929414749146,
435
+ "learning_rate": 8.032128514056225e-07,
436
+ "loss": 5.309723281860352,
437
+ "step": 300
438
+ },
439
+ {
440
+ "epoch": 0.013610906575629784,
441
+ "grad_norm": 0.4549182653427124,
442
+ "learning_rate": 8.16599732262383e-07,
443
+ "loss": 5.263495635986328,
444
+ "step": 305
445
+ },
446
+ {
447
+ "epoch": 0.013834036191623714,
448
+ "grad_norm": 0.45380067825317383,
449
+ "learning_rate": 8.299866131191432e-07,
450
+ "loss": 5.270233535766602,
451
+ "step": 310
452
+ },
453
+ {
454
+ "epoch": 0.014057165807617645,
455
+ "grad_norm": 0.4339616894721985,
456
+ "learning_rate": 8.433734939759036e-07,
457
+ "loss": 5.273446655273437,
458
+ "step": 315
459
+ },
460
+ {
461
+ "epoch": 0.014280295423611577,
462
+ "grad_norm": 0.44467630982398987,
463
+ "learning_rate": 8.56760374832664e-07,
464
+ "loss": 5.210543823242188,
465
+ "step": 320
466
+ },
467
+ {
468
+ "epoch": 0.014503425039605507,
469
+ "grad_norm": 0.4964953064918518,
470
+ "learning_rate": 8.701472556894243e-07,
471
+ "loss": 5.216452026367188,
472
+ "step": 325
473
+ },
474
+ {
475
+ "epoch": 0.014726554655599437,
476
+ "grad_norm": 0.4745877683162689,
477
+ "learning_rate": 8.835341365461848e-07,
478
+ "loss": 5.237066650390625,
479
+ "step": 330
480
+ },
481
+ {
482
+ "epoch": 0.014949684271593369,
483
+ "grad_norm": 0.4923485815525055,
484
+ "learning_rate": 8.969210174029452e-07,
485
+ "loss": 5.16167106628418,
486
+ "step": 335
487
+ },
488
+ {
489
+ "epoch": 0.015172813887587299,
490
+ "grad_norm": 0.45548155903816223,
491
+ "learning_rate": 9.103078982597054e-07,
492
+ "loss": 5.184793090820312,
493
+ "step": 340
494
+ },
495
+ {
496
+ "epoch": 0.015395943503581231,
497
+ "grad_norm": 0.5323435068130493,
498
+ "learning_rate": 9.236947791164659e-07,
499
+ "loss": 5.1162841796875,
500
+ "step": 345
501
+ },
502
+ {
503
+ "epoch": 0.015619073119575161,
504
+ "grad_norm": 0.5220831036567688,
505
+ "learning_rate": 9.370816599732263e-07,
506
+ "loss": 5.135643005371094,
507
+ "step": 350
508
+ },
509
+ {
510
+ "epoch": 0.01584220273556909,
511
+ "grad_norm": 0.5357551574707031,
512
+ "learning_rate": 9.504685408299866e-07,
513
+ "loss": 5.001235580444336,
514
+ "step": 355
515
+ },
516
+ {
517
+ "epoch": 0.016065332351563023,
518
+ "grad_norm": 0.49958014488220215,
519
+ "learning_rate": 9.63855421686747e-07,
520
+ "loss": 5.096822357177734,
521
+ "step": 360
522
+ },
523
+ {
524
+ "epoch": 0.016288461967556955,
525
+ "grad_norm": 0.53453129529953,
526
+ "learning_rate": 9.772423025435074e-07,
527
+ "loss": 5.057343673706055,
528
+ "step": 365
529
+ },
530
+ {
531
+ "epoch": 0.016511591583550884,
532
+ "grad_norm": 0.49969372153282166,
533
+ "learning_rate": 9.906291834002677e-07,
534
+ "loss": 5.097013092041015,
535
+ "step": 370
536
+ },
537
+ {
538
+ "epoch": 0.016734721199544816,
539
+ "grad_norm": 0.477762907743454,
540
+ "learning_rate": 1.0040160642570282e-06,
541
+ "loss": 5.05908317565918,
542
+ "step": 375
543
+ },
544
+ {
545
+ "epoch": 0.016957850815538748,
546
+ "grad_norm": 0.5632966756820679,
547
+ "learning_rate": 1.0174029451137885e-06,
548
+ "loss": 4.880051040649414,
549
+ "step": 380
550
+ },
551
+ {
552
+ "epoch": 0.017180980431532676,
553
+ "grad_norm": 0.4806019961833954,
554
+ "learning_rate": 1.0307898259705488e-06,
555
+ "loss": 4.938255310058594,
556
+ "step": 385
557
+ },
558
+ {
559
+ "epoch": 0.017404110047526608,
560
+ "grad_norm": 0.5882935523986816,
561
+ "learning_rate": 1.0441767068273092e-06,
562
+ "loss": 4.972189331054688,
563
+ "step": 390
564
+ },
565
+ {
566
+ "epoch": 0.01762723966352054,
567
+ "grad_norm": 0.583991289138794,
568
+ "learning_rate": 1.0575635876840697e-06,
569
+ "loss": 4.8074909210205075,
570
+ "step": 395
571
+ },
572
+ {
573
+ "epoch": 0.01785036927951447,
574
+ "grad_norm": 0.6130750179290771,
575
+ "learning_rate": 1.07095046854083e-06,
576
+ "loss": 4.851003265380859,
577
+ "step": 400
578
+ },
579
+ {
580
+ "epoch": 0.0180734988955084,
581
+ "grad_norm": 0.630670428276062,
582
+ "learning_rate": 1.0843373493975905e-06,
583
+ "loss": 4.742978286743164,
584
+ "step": 405
585
+ },
586
+ {
587
+ "epoch": 0.018296628511502332,
588
+ "grad_norm": 0.6021527647972107,
589
+ "learning_rate": 1.0977242302543508e-06,
590
+ "loss": 4.795417022705078,
591
+ "step": 410
592
+ },
593
+ {
594
+ "epoch": 0.018519758127496264,
595
+ "grad_norm": 0.6027496457099915,
596
+ "learning_rate": 1.111111111111111e-06,
597
+ "loss": 4.799094009399414,
598
+ "step": 415
599
+ },
600
+ {
601
+ "epoch": 0.018742887743490193,
602
+ "grad_norm": 0.62542325258255,
603
+ "learning_rate": 1.1244979919678715e-06,
604
+ "loss": 4.734822845458984,
605
+ "step": 420
606
+ },
607
+ {
608
+ "epoch": 0.018966017359484125,
609
+ "grad_norm": 0.6744217872619629,
610
+ "learning_rate": 1.137884872824632e-06,
611
+ "loss": 4.470468902587891,
612
+ "step": 425
613
+ },
614
+ {
615
+ "epoch": 0.019189146975478057,
616
+ "grad_norm": 0.6994063854217529,
617
+ "learning_rate": 1.1512717536813923e-06,
618
+ "loss": 4.7334716796875,
619
+ "step": 430
620
+ },
621
+ {
622
+ "epoch": 0.019412276591471985,
623
+ "grad_norm": 0.6043664813041687,
624
+ "learning_rate": 1.1646586345381526e-06,
625
+ "loss": 4.6335186004638675,
626
+ "step": 435
627
+ },
628
+ {
629
+ "epoch": 0.019635406207465917,
630
+ "grad_norm": 0.600983738899231,
631
+ "learning_rate": 1.178045515394913e-06,
632
+ "loss": 4.521899795532226,
633
+ "step": 440
634
+ },
635
+ {
636
+ "epoch": 0.01985853582345985,
637
+ "grad_norm": 0.632734477519989,
638
+ "learning_rate": 1.1914323962516733e-06,
639
+ "loss": 4.455641555786133,
640
+ "step": 445
641
+ },
642
+ {
643
+ "epoch": 0.020081665439453777,
644
+ "grad_norm": 0.6841812133789062,
645
+ "learning_rate": 1.2048192771084338e-06,
646
+ "loss": 4.552269363403321,
647
+ "step": 450
648
+ },
649
+ {
650
+ "epoch": 0.02030479505544771,
651
+ "grad_norm": 0.7014154195785522,
652
+ "learning_rate": 1.2182061579651943e-06,
653
+ "loss": 4.4612682342529295,
654
+ "step": 455
655
+ },
656
+ {
657
+ "epoch": 0.02052792467144164,
658
+ "grad_norm": 0.7957334518432617,
659
+ "learning_rate": 1.2315930388219544e-06,
660
+ "loss": 4.16539421081543,
661
+ "step": 460
662
+ },
663
+ {
664
+ "epoch": 0.02075105428743557,
665
+ "grad_norm": 0.6679075360298157,
666
+ "learning_rate": 1.2449799196787148e-06,
667
+ "loss": 4.380799102783203,
668
+ "step": 465
669
+ },
670
+ {
671
+ "epoch": 0.020974183903429502,
672
+ "grad_norm": 0.6911934614181519,
673
+ "learning_rate": 1.2583668005354753e-06,
674
+ "loss": 4.188001251220703,
675
+ "step": 470
676
+ },
677
+ {
678
+ "epoch": 0.021197313519423434,
679
+ "grad_norm": 0.6941277384757996,
680
+ "learning_rate": 1.2717536813922356e-06,
681
+ "loss": 4.30323600769043,
682
+ "step": 475
683
+ },
684
+ {
685
+ "epoch": 0.021420443135417366,
686
+ "grad_norm": 0.6411834359169006,
687
+ "learning_rate": 1.285140562248996e-06,
688
+ "loss": 4.210566711425781,
689
+ "step": 480
690
+ },
691
+ {
692
+ "epoch": 0.021643572751411294,
693
+ "grad_norm": 0.6447572112083435,
694
+ "learning_rate": 1.2985274431057564e-06,
695
+ "loss": 4.145897674560547,
696
+ "step": 485
697
+ },
698
+ {
699
+ "epoch": 0.021866702367405226,
700
+ "grad_norm": 0.6956306099891663,
701
+ "learning_rate": 1.3119143239625166e-06,
702
+ "loss": 4.1259113311767575,
703
+ "step": 490
704
+ },
705
+ {
706
+ "epoch": 0.022089831983399158,
707
+ "grad_norm": 0.6369218826293945,
708
+ "learning_rate": 1.3253012048192771e-06,
709
+ "loss": 4.050044250488281,
710
+ "step": 495
711
+ },
712
+ {
713
+ "epoch": 0.022312961599393086,
714
+ "grad_norm": 0.6199172735214233,
715
+ "learning_rate": 1.3386880856760376e-06,
716
+ "loss": 3.9885379791259767,
717
+ "step": 500
718
+ },
719
+ {
720
+ "epoch": 0.02253609121538702,
721
+ "grad_norm": 0.6004146933555603,
722
+ "learning_rate": 1.3520749665327979e-06,
723
+ "loss": 3.9876686096191407,
724
+ "step": 505
725
+ },
726
+ {
727
+ "epoch": 0.02275922083138095,
728
+ "grad_norm": 0.6573876142501831,
729
+ "learning_rate": 1.3654618473895582e-06,
730
+ "loss": 3.860630416870117,
731
+ "step": 510
732
+ },
733
+ {
734
+ "epoch": 0.02298235044737488,
735
+ "grad_norm": 0.6570749878883362,
736
+ "learning_rate": 1.3788487282463186e-06,
737
+ "loss": 3.987852096557617,
738
+ "step": 515
739
+ },
740
+ {
741
+ "epoch": 0.02320548006336881,
742
+ "grad_norm": 0.6052077412605286,
743
+ "learning_rate": 1.392235609103079e-06,
744
+ "loss": 3.773631286621094,
745
+ "step": 520
746
+ },
747
+ {
748
+ "epoch": 0.023428609679362743,
749
+ "grad_norm": 0.6024802923202515,
750
+ "learning_rate": 1.4056224899598394e-06,
751
+ "loss": 3.8334468841552733,
752
+ "step": 525
753
+ },
754
+ {
755
+ "epoch": 0.02365173929535667,
756
+ "grad_norm": 0.6448367238044739,
757
+ "learning_rate": 1.4190093708165999e-06,
758
+ "loss": 3.709330749511719,
759
+ "step": 530
760
+ },
761
+ {
762
+ "epoch": 0.023874868911350603,
763
+ "grad_norm": 0.5610556602478027,
764
+ "learning_rate": 1.4323962516733602e-06,
765
+ "loss": 3.608018493652344,
766
+ "step": 535
767
+ },
768
+ {
769
+ "epoch": 0.024097998527344535,
770
+ "grad_norm": 0.5798413157463074,
771
+ "learning_rate": 1.4457831325301204e-06,
772
+ "loss": 3.513960266113281,
773
+ "step": 540
774
+ },
775
+ {
776
+ "epoch": 0.024321128143338467,
777
+ "grad_norm": 0.5836862921714783,
778
+ "learning_rate": 1.459170013386881e-06,
779
+ "loss": 3.5928466796875,
780
+ "step": 545
781
+ },
782
+ {
783
+ "epoch": 0.024544257759332395,
784
+ "grad_norm": 0.5705676078796387,
785
+ "learning_rate": 1.4725568942436412e-06,
786
+ "loss": 3.5890869140625,
787
+ "step": 550
788
+ },
789
+ {
790
+ "epoch": 0.024767387375326327,
791
+ "grad_norm": 0.6377549767494202,
792
+ "learning_rate": 1.4859437751004017e-06,
793
+ "loss": 3.6042083740234374,
794
+ "step": 555
795
+ },
796
+ {
797
+ "epoch": 0.02499051699132026,
798
+ "grad_norm": 0.5612762570381165,
799
+ "learning_rate": 1.499330655957162e-06,
800
+ "loss": 3.511037826538086,
801
+ "step": 560
802
+ },
803
+ {
804
+ "epoch": 0.025213646607314188,
805
+ "grad_norm": 0.5943495035171509,
806
+ "learning_rate": 1.5127175368139222e-06,
807
+ "loss": 3.5007259368896486,
808
+ "step": 565
809
+ },
810
+ {
811
+ "epoch": 0.02543677622330812,
812
+ "grad_norm": 0.5872607827186584,
813
+ "learning_rate": 1.5261044176706827e-06,
814
+ "loss": 3.3742324829101564,
815
+ "step": 570
816
+ },
817
+ {
818
+ "epoch": 0.02565990583930205,
819
+ "grad_norm": 0.5494210124015808,
820
+ "learning_rate": 1.5394912985274432e-06,
821
+ "loss": 3.3373321533203124,
822
+ "step": 575
823
+ },
824
+ {
825
+ "epoch": 0.02588303545529598,
826
+ "grad_norm": 0.5138423442840576,
827
+ "learning_rate": 1.5528781793842037e-06,
828
+ "loss": 3.346783447265625,
829
+ "step": 580
830
+ },
831
+ {
832
+ "epoch": 0.026106165071289912,
833
+ "grad_norm": 0.5539407730102539,
834
+ "learning_rate": 1.566265060240964e-06,
835
+ "loss": 3.183103561401367,
836
+ "step": 585
837
+ },
838
+ {
839
+ "epoch": 0.026329294687283844,
840
+ "grad_norm": 0.5888996720314026,
841
+ "learning_rate": 1.5796519410977242e-06,
842
+ "loss": 3.357350540161133,
843
+ "step": 590
844
+ },
845
+ {
846
+ "epoch": 0.026552424303277773,
847
+ "grad_norm": 0.5445035696029663,
848
+ "learning_rate": 1.5930388219544845e-06,
849
+ "loss": 3.374500274658203,
850
+ "step": 595
851
+ },
852
+ {
853
+ "epoch": 0.026775553919271704,
854
+ "grad_norm": 0.5013962388038635,
855
+ "learning_rate": 1.606425702811245e-06,
856
+ "loss": 3.3687782287597656,
857
+ "step": 600
858
+ },
859
+ {
860
+ "epoch": 0.026998683535265636,
861
+ "grad_norm": 0.5150508284568787,
862
+ "learning_rate": 1.6198125836680055e-06,
863
+ "loss": 3.2590232849121095,
864
+ "step": 605
865
+ },
866
+ {
867
+ "epoch": 0.02722181315125957,
868
+ "grad_norm": 0.603961169719696,
869
+ "learning_rate": 1.633199464524766e-06,
870
+ "loss": 3.1635900497436524,
871
+ "step": 610
872
+ },
873
+ {
874
+ "epoch": 0.027444942767253497,
875
+ "grad_norm": 0.5365360975265503,
876
+ "learning_rate": 1.6465863453815263e-06,
877
+ "loss": 3.3600418090820314,
878
+ "step": 615
879
+ },
880
+ {
881
+ "epoch": 0.02766807238324743,
882
+ "grad_norm": 0.5401721000671387,
883
+ "learning_rate": 1.6599732262382863e-06,
884
+ "loss": 3.183880424499512,
885
+ "step": 620
886
+ },
887
+ {
888
+ "epoch": 0.02789120199924136,
889
+ "grad_norm": 0.5595034956932068,
890
+ "learning_rate": 1.6733601070950468e-06,
891
+ "loss": 3.0762868881225587,
892
+ "step": 625
893
+ },
894
+ {
895
+ "epoch": 0.02811433161523529,
896
+ "grad_norm": 0.4972860813140869,
897
+ "learning_rate": 1.6867469879518073e-06,
898
+ "loss": 3.0724456787109373,
899
+ "step": 630
900
+ },
901
+ {
902
+ "epoch": 0.02833746123122922,
903
+ "grad_norm": 0.44132477045059204,
904
+ "learning_rate": 1.7001338688085678e-06,
905
+ "loss": 3.1175127029418945,
906
+ "step": 635
907
+ },
908
+ {
909
+ "epoch": 0.028560590847223153,
910
+ "grad_norm": 0.5078781843185425,
911
+ "learning_rate": 1.713520749665328e-06,
912
+ "loss": 3.0849445343017576,
913
+ "step": 640
914
+ },
915
+ {
916
+ "epoch": 0.02878372046321708,
917
+ "grad_norm": 0.4757557809352875,
918
+ "learning_rate": 1.7269076305220885e-06,
919
+ "loss": 3.2029571533203125,
920
+ "step": 645
921
+ },
922
+ {
923
+ "epoch": 0.029006850079211013,
924
+ "grad_norm": 0.4932602345943451,
925
+ "learning_rate": 1.7402945113788486e-06,
926
+ "loss": 3.0668067932128906,
927
+ "step": 650
928
+ },
929
+ {
930
+ "epoch": 0.029229979695204945,
931
+ "grad_norm": 0.5687287449836731,
932
+ "learning_rate": 1.753681392235609e-06,
933
+ "loss": 3.113470268249512,
934
+ "step": 655
935
+ },
936
+ {
937
+ "epoch": 0.029453109311198874,
938
+ "grad_norm": 0.4471919536590576,
939
+ "learning_rate": 1.7670682730923696e-06,
940
+ "loss": 2.9105451583862303,
941
+ "step": 660
942
+ },
943
+ {
944
+ "epoch": 0.029676238927192806,
945
+ "grad_norm": 0.5096343159675598,
946
+ "learning_rate": 1.7804551539491298e-06,
947
+ "loss": 2.94341926574707,
948
+ "step": 665
949
+ },
950
+ {
951
+ "epoch": 0.029899368543186738,
952
+ "grad_norm": 0.5599822402000427,
953
+ "learning_rate": 1.7938420348058903e-06,
954
+ "loss": 2.892704391479492,
955
+ "step": 670
956
+ },
957
+ {
958
+ "epoch": 0.03012249815918067,
959
+ "grad_norm": 0.40055856108665466,
960
+ "learning_rate": 1.8072289156626508e-06,
961
+ "loss": 2.896807861328125,
962
+ "step": 675
963
+ },
964
+ {
965
+ "epoch": 0.030345627775174598,
966
+ "grad_norm": 0.47082021832466125,
967
+ "learning_rate": 1.8206157965194109e-06,
968
+ "loss": 2.7966148376464846,
969
+ "step": 680
970
+ },
971
+ {
972
+ "epoch": 0.03056875739116853,
973
+ "grad_norm": 0.42042815685272217,
974
+ "learning_rate": 1.8340026773761714e-06,
975
+ "loss": 2.839730644226074,
976
+ "step": 685
977
+ },
978
+ {
979
+ "epoch": 0.030791887007162462,
980
+ "grad_norm": 0.49198001623153687,
981
+ "learning_rate": 1.8473895582329318e-06,
982
+ "loss": 2.9913705825805663,
983
+ "step": 690
984
+ },
985
+ {
986
+ "epoch": 0.03101501662315639,
987
+ "grad_norm": 0.4656424820423126,
988
+ "learning_rate": 1.8607764390896921e-06,
989
+ "loss": 3.0195621490478515,
990
+ "step": 695
991
+ },
992
+ {
993
+ "epoch": 0.031238146239150322,
994
+ "grad_norm": 0.46259376406669617,
995
+ "learning_rate": 1.8741633199464526e-06,
996
+ "loss": 2.948585319519043,
997
+ "step": 700
998
+ },
999
+ {
1000
+ "epoch": 0.03146127585514425,
1001
+ "grad_norm": 0.4671732783317566,
1002
+ "learning_rate": 1.887550200803213e-06,
1003
+ "loss": 2.7285741806030273,
1004
+ "step": 705
1005
+ },
1006
+ {
1007
+ "epoch": 0.03168440547113818,
1008
+ "grad_norm": 0.6493679285049438,
1009
+ "learning_rate": 1.9009370816599732e-06,
1010
+ "loss": 2.9263893127441407,
1011
+ "step": 710
1012
+ },
1013
+ {
1014
+ "epoch": 0.031907535087132115,
1015
+ "grad_norm": 0.46389710903167725,
1016
+ "learning_rate": 1.9143239625167336e-06,
1017
+ "loss": 2.634868049621582,
1018
+ "step": 715
1019
+ },
1020
+ {
1021
+ "epoch": 0.03213066470312605,
1022
+ "grad_norm": 0.44173645973205566,
1023
+ "learning_rate": 1.927710843373494e-06,
1024
+ "loss": 2.7959117889404297,
1025
+ "step": 720
1026
+ },
1027
+ {
1028
+ "epoch": 0.03235379431911998,
1029
+ "grad_norm": 0.45440003275871277,
1030
+ "learning_rate": 1.9410977242302546e-06,
1031
+ "loss": 2.9056529998779297,
1032
+ "step": 725
1033
+ },
1034
+ {
1035
+ "epoch": 0.03257692393511391,
1036
+ "grad_norm": 0.44541046023368835,
1037
+ "learning_rate": 1.954484605087015e-06,
1038
+ "loss": 2.7214975357055664,
1039
+ "step": 730
1040
+ },
1041
+ {
1042
+ "epoch": 0.032800053551107836,
1043
+ "grad_norm": 0.4217804968357086,
1044
+ "learning_rate": 1.967871485943775e-06,
1045
+ "loss": 2.7198978424072267,
1046
+ "step": 735
1047
+ },
1048
+ {
1049
+ "epoch": 0.03302318316710177,
1050
+ "grad_norm": 0.46086645126342773,
1051
+ "learning_rate": 1.9812583668005354e-06,
1052
+ "loss": 2.7366127014160155,
1053
+ "step": 740
1054
+ },
1055
+ {
1056
+ "epoch": 0.0332463127830957,
1057
+ "grad_norm": 0.41817471385002136,
1058
+ "learning_rate": 1.9946452476572957e-06,
1059
+ "loss": 2.857570457458496,
1060
+ "step": 745
1061
+ },
1062
+ {
1063
+ "epoch": 0.03346944239908963,
1064
+ "grad_norm": 0.44163694977760315,
1065
+ "learning_rate": 2.0080321285140564e-06,
1066
+ "loss": 2.8251710891723634,
1067
+ "step": 750
1068
+ },
1069
+ {
1070
+ "epoch": 0.03369257201508356,
1071
+ "grad_norm": 0.46645739674568176,
1072
+ "learning_rate": 2.0214190093708167e-06,
1073
+ "loss": 2.548258399963379,
1074
+ "step": 755
1075
+ },
1076
+ {
1077
+ "epoch": 0.033915701631077495,
1078
+ "grad_norm": 0.3971084654331207,
1079
+ "learning_rate": 2.034805890227577e-06,
1080
+ "loss": 2.682722473144531,
1081
+ "step": 760
1082
+ },
1083
+ {
1084
+ "epoch": 0.03413883124707143,
1085
+ "grad_norm": 0.5373950004577637,
1086
+ "learning_rate": 2.0481927710843377e-06,
1087
+ "loss": 2.8539506912231447,
1088
+ "step": 765
1089
+ },
1090
+ {
1091
+ "epoch": 0.03436196086306535,
1092
+ "grad_norm": 0.6133913397789001,
1093
+ "learning_rate": 2.0615796519410975e-06,
1094
+ "loss": 2.7016387939453126,
1095
+ "step": 770
1096
+ },
1097
+ {
1098
+ "epoch": 0.034585090479059284,
1099
+ "grad_norm": 0.5247603058815002,
1100
+ "learning_rate": 2.074966532797858e-06,
1101
+ "loss": 2.645602226257324,
1102
+ "step": 775
1103
+ },
1104
+ {
1105
+ "epoch": 0.034808220095053216,
1106
+ "grad_norm": 0.4487764537334442,
1107
+ "learning_rate": 2.0883534136546185e-06,
1108
+ "loss": 2.699404716491699,
1109
+ "step": 780
1110
+ },
1111
+ {
1112
+ "epoch": 0.03503134971104715,
1113
+ "grad_norm": 0.5962596535682678,
1114
+ "learning_rate": 2.1017402945113788e-06,
1115
+ "loss": 2.7891347885131834,
1116
+ "step": 785
1117
+ },
1118
+ {
1119
+ "epoch": 0.03525447932704108,
1120
+ "grad_norm": 0.5410242080688477,
1121
+ "learning_rate": 2.1151271753681395e-06,
1122
+ "loss": 2.675504684448242,
1123
+ "step": 790
1124
+ },
1125
+ {
1126
+ "epoch": 0.03547760894303501,
1127
+ "grad_norm": 0.4576985239982605,
1128
+ "learning_rate": 2.1285140562248997e-06,
1129
+ "loss": 2.642281341552734,
1130
+ "step": 795
1131
+ },
1132
+ {
1133
+ "epoch": 0.03570073855902894,
1134
+ "grad_norm": 0.45449239015579224,
1135
+ "learning_rate": 2.14190093708166e-06,
1136
+ "loss": 2.7088804244995117,
1137
+ "step": 800
1138
+ },
1139
+ {
1140
+ "epoch": 0.03592386817502287,
1141
+ "grad_norm": 0.4782140552997589,
1142
+ "learning_rate": 2.1552878179384203e-06,
1143
+ "loss": 2.698718452453613,
1144
+ "step": 805
1145
+ },
1146
+ {
1147
+ "epoch": 0.0361469977910168,
1148
+ "grad_norm": 0.43745940923690796,
1149
+ "learning_rate": 2.168674698795181e-06,
1150
+ "loss": 2.7042917251586913,
1151
+ "step": 810
1152
+ },
1153
+ {
1154
+ "epoch": 0.03637012740701073,
1155
+ "grad_norm": 0.6170194149017334,
1156
+ "learning_rate": 2.1820615796519413e-06,
1157
+ "loss": 2.6548912048339846,
1158
+ "step": 815
1159
+ },
1160
+ {
1161
+ "epoch": 0.036593257023004665,
1162
+ "grad_norm": 0.5126772522926331,
1163
+ "learning_rate": 2.1954484605087015e-06,
1164
+ "loss": 2.663498306274414,
1165
+ "step": 820
1166
+ },
1167
+ {
1168
+ "epoch": 0.0368163866389986,
1169
+ "grad_norm": 0.42503371834754944,
1170
+ "learning_rate": 2.2088353413654622e-06,
1171
+ "loss": 2.766114616394043,
1172
+ "step": 825
1173
+ },
1174
+ {
1175
+ "epoch": 0.03703951625499253,
1176
+ "grad_norm": 0.5707629323005676,
1177
+ "learning_rate": 2.222222222222222e-06,
1178
+ "loss": 2.6172439575195314,
1179
+ "step": 830
1180
+ },
1181
+ {
1182
+ "epoch": 0.037262645870986454,
1183
+ "grad_norm": 0.45169416069984436,
1184
+ "learning_rate": 2.2356091030789828e-06,
1185
+ "loss": 2.5655603408813477,
1186
+ "step": 835
1187
+ },
1188
+ {
1189
+ "epoch": 0.037485775486980386,
1190
+ "grad_norm": 0.44726496934890747,
1191
+ "learning_rate": 2.248995983935743e-06,
1192
+ "loss": 2.5273290634155274,
1193
+ "step": 840
1194
+ },
1195
+ {
1196
+ "epoch": 0.03770890510297432,
1197
+ "grad_norm": 0.5287439823150635,
1198
+ "learning_rate": 2.2623828647925033e-06,
1199
+ "loss": 2.6541069030761717,
1200
+ "step": 845
1201
+ },
1202
+ {
1203
+ "epoch": 0.03793203471896825,
1204
+ "grad_norm": 0.46616867184638977,
1205
+ "learning_rate": 2.275769745649264e-06,
1206
+ "loss": 2.56517333984375,
1207
+ "step": 850
1208
+ },
1209
+ {
1210
+ "epoch": 0.03815516433496218,
1211
+ "grad_norm": 0.4369581937789917,
1212
+ "learning_rate": 2.2891566265060243e-06,
1213
+ "loss": 2.4222312927246095,
1214
+ "step": 855
1215
+ },
1216
+ {
1217
+ "epoch": 0.03837829395095611,
1218
+ "grad_norm": 0.4172717332839966,
1219
+ "learning_rate": 2.3025435073627846e-06,
1220
+ "loss": 2.549334716796875,
1221
+ "step": 860
1222
+ },
1223
+ {
1224
+ "epoch": 0.03860142356695004,
1225
+ "grad_norm": 0.46124762296676636,
1226
+ "learning_rate": 2.315930388219545e-06,
1227
+ "loss": 2.5675996780395507,
1228
+ "step": 865
1229
+ },
1230
+ {
1231
+ "epoch": 0.03882455318294397,
1232
+ "grad_norm": 0.4226182699203491,
1233
+ "learning_rate": 2.329317269076305e-06,
1234
+ "loss": 2.4144569396972657,
1235
+ "step": 870
1236
+ },
1237
+ {
1238
+ "epoch": 0.0390476827989379,
1239
+ "grad_norm": 0.4673754870891571,
1240
+ "learning_rate": 2.342704149933066e-06,
1241
+ "loss": 2.3829999923706056,
1242
+ "step": 875
1243
+ },
1244
+ {
1245
+ "epoch": 0.039270812414931834,
1246
+ "grad_norm": 0.5152049660682678,
1247
+ "learning_rate": 2.356091030789826e-06,
1248
+ "loss": 2.5030202865600586,
1249
+ "step": 880
1250
+ },
1251
+ {
1252
+ "epoch": 0.039493942030925766,
1253
+ "grad_norm": 0.5377767086029053,
1254
+ "learning_rate": 2.3694779116465864e-06,
1255
+ "loss": 2.7714466094970702,
1256
+ "step": 885
1257
+ },
1258
+ {
1259
+ "epoch": 0.0397170716469197,
1260
+ "grad_norm": 0.45861950516700745,
1261
+ "learning_rate": 2.3828647925033466e-06,
1262
+ "loss": 2.461492729187012,
1263
+ "step": 890
1264
+ },
1265
+ {
1266
+ "epoch": 0.03994020126291363,
1267
+ "grad_norm": 0.4814665615558624,
1268
+ "learning_rate": 2.396251673360107e-06,
1269
+ "loss": 2.4756641387939453,
1270
+ "step": 895
1271
+ },
1272
+ {
1273
+ "epoch": 0.040163330878907555,
1274
+ "grad_norm": 0.5024599432945251,
1275
+ "learning_rate": 2.4096385542168676e-06,
1276
+ "loss": 2.5811389923095702,
1277
+ "step": 900
1278
+ },
1279
+ {
1280
+ "epoch": 0.04038646049490149,
1281
+ "grad_norm": 0.45141535997390747,
1282
+ "learning_rate": 2.423025435073628e-06,
1283
+ "loss": 2.686309814453125,
1284
+ "step": 905
1285
+ },
1286
+ {
1287
+ "epoch": 0.04060959011089542,
1288
+ "grad_norm": 0.4806381165981293,
1289
+ "learning_rate": 2.4364123159303886e-06,
1290
+ "loss": 2.530971717834473,
1291
+ "step": 910
1292
+ },
1293
+ {
1294
+ "epoch": 0.04083271972688935,
1295
+ "grad_norm": 0.7574505805969238,
1296
+ "learning_rate": 2.449799196787149e-06,
1297
+ "loss": 2.5287572860717775,
1298
+ "step": 915
1299
+ },
1300
+ {
1301
+ "epoch": 0.04105584934288328,
1302
+ "grad_norm": 0.39056211709976196,
1303
+ "learning_rate": 2.4631860776439087e-06,
1304
+ "loss": 2.4373926162719726,
1305
+ "step": 920
1306
+ },
1307
+ {
1308
+ "epoch": 0.041278978958877215,
1309
+ "grad_norm": 0.4549945890903473,
1310
+ "learning_rate": 2.4765729585006694e-06,
1311
+ "loss": 2.508747100830078,
1312
+ "step": 925
1313
+ },
1314
+ {
1315
+ "epoch": 0.04150210857487114,
1316
+ "grad_norm": 0.49253398180007935,
1317
+ "learning_rate": 2.4899598393574297e-06,
1318
+ "loss": 2.5109813690185545,
1319
+ "step": 930
1320
+ },
1321
+ {
1322
+ "epoch": 0.04172523819086507,
1323
+ "grad_norm": 0.4185622036457062,
1324
+ "learning_rate": 2.5033467202141904e-06,
1325
+ "loss": 2.5602550506591797,
1326
+ "step": 935
1327
+ },
1328
+ {
1329
+ "epoch": 0.041948367806859004,
1330
+ "grad_norm": 0.5255184769630432,
1331
+ "learning_rate": 2.5167336010709507e-06,
1332
+ "loss": 2.5018032073974608,
1333
+ "step": 940
1334
+ },
1335
+ {
1336
+ "epoch": 0.042171497422852935,
1337
+ "grad_norm": 0.40105360746383667,
1338
+ "learning_rate": 2.530120481927711e-06,
1339
+ "loss": 2.3703737258911133,
1340
+ "step": 945
1341
+ },
1342
+ {
1343
+ "epoch": 0.04239462703884687,
1344
+ "grad_norm": 0.5396221876144409,
1345
+ "learning_rate": 2.543507362784471e-06,
1346
+ "loss": 2.46767520904541,
1347
+ "step": 950
1348
+ },
1349
+ {
1350
+ "epoch": 0.0426177566548408,
1351
+ "grad_norm": 0.4686439633369446,
1352
+ "learning_rate": 2.5568942436412315e-06,
1353
+ "loss": 2.535739517211914,
1354
+ "step": 955
1355
+ },
1356
+ {
1357
+ "epoch": 0.04284088627083473,
1358
+ "grad_norm": 0.5110803246498108,
1359
+ "learning_rate": 2.570281124497992e-06,
1360
+ "loss": 2.5248756408691406,
1361
+ "step": 960
1362
+ },
1363
+ {
1364
+ "epoch": 0.043064015886828656,
1365
+ "grad_norm": 0.5164591670036316,
1366
+ "learning_rate": 2.5836680053547524e-06,
1367
+ "loss": 2.4275962829589846,
1368
+ "step": 965
1369
+ },
1370
+ {
1371
+ "epoch": 0.04328714550282259,
1372
+ "grad_norm": 0.42096105217933655,
1373
+ "learning_rate": 2.5970548862115127e-06,
1374
+ "loss": 2.4086238861083986,
1375
+ "step": 970
1376
+ },
1377
+ {
1378
+ "epoch": 0.04351027511881652,
1379
+ "grad_norm": 0.5388748645782471,
1380
+ "learning_rate": 2.6104417670682734e-06,
1381
+ "loss": 2.340771293640137,
1382
+ "step": 975
1383
+ },
1384
+ {
1385
+ "epoch": 0.04373340473481045,
1386
+ "grad_norm": 0.4667651653289795,
1387
+ "learning_rate": 2.6238286479250333e-06,
1388
+ "loss": 2.506967544555664,
1389
+ "step": 980
1390
+ },
1391
+ {
1392
+ "epoch": 0.043956534350804384,
1393
+ "grad_norm": 0.5090363621711731,
1394
+ "learning_rate": 2.637215528781794e-06,
1395
+ "loss": 2.251785469055176,
1396
+ "step": 985
1397
+ },
1398
+ {
1399
+ "epoch": 0.044179663966798316,
1400
+ "grad_norm": 0.4600473940372467,
1401
+ "learning_rate": 2.6506024096385542e-06,
1402
+ "loss": 2.3269046783447265,
1403
+ "step": 990
1404
+ },
1405
+ {
1406
+ "epoch": 0.04440279358279224,
1407
+ "grad_norm": 0.4453408718109131,
1408
+ "learning_rate": 2.6639892904953145e-06,
1409
+ "loss": 2.4940771102905273,
1410
+ "step": 995
1411
+ },
1412
+ {
1413
+ "epoch": 0.04462592319878617,
1414
+ "grad_norm": 0.4637933671474457,
1415
+ "learning_rate": 2.6773761713520752e-06,
1416
+ "loss": 2.578083801269531,
1417
+ "step": 1000
1418
+ },
1419
+ {
1420
+ "epoch": 0.044849052814780105,
1421
+ "grad_norm": 0.48008954524993896,
1422
+ "learning_rate": 2.6907630522088355e-06,
1423
+ "loss": 2.509678077697754,
1424
+ "step": 1005
1425
+ },
1426
+ {
1427
+ "epoch": 0.04507218243077404,
1428
+ "grad_norm": 0.5186890363693237,
1429
+ "learning_rate": 2.7041499330655958e-06,
1430
+ "loss": 2.420572280883789,
1431
+ "step": 1010
1432
+ },
1433
+ {
1434
+ "epoch": 0.04529531204676797,
1435
+ "grad_norm": 0.5090295076370239,
1436
+ "learning_rate": 2.717536813922356e-06,
1437
+ "loss": 2.42071533203125,
1438
+ "step": 1015
1439
+ },
1440
+ {
1441
+ "epoch": 0.0455184416627619,
1442
+ "grad_norm": 0.5519959926605225,
1443
+ "learning_rate": 2.7309236947791163e-06,
1444
+ "loss": 2.458280563354492,
1445
+ "step": 1020
1446
+ },
1447
+ {
1448
+ "epoch": 0.04574157127875583,
1449
+ "grad_norm": 0.5525882840156555,
1450
+ "learning_rate": 2.744310575635877e-06,
1451
+ "loss": 2.3400331497192384,
1452
+ "step": 1025
1453
+ },
1454
+ {
1455
+ "epoch": 0.04596470089474976,
1456
+ "grad_norm": 0.48508089780807495,
1457
+ "learning_rate": 2.7576974564926373e-06,
1458
+ "loss": 2.4403892517089845,
1459
+ "step": 1030
1460
+ },
1461
+ {
1462
+ "epoch": 0.04618783051074369,
1463
+ "grad_norm": 0.5230780839920044,
1464
+ "learning_rate": 2.771084337349398e-06,
1465
+ "loss": 2.3652227401733397,
1466
+ "step": 1035
1467
+ },
1468
+ {
1469
+ "epoch": 0.04641096012673762,
1470
+ "grad_norm": 0.5132386088371277,
1471
+ "learning_rate": 2.784471218206158e-06,
1472
+ "loss": 2.313191604614258,
1473
+ "step": 1040
1474
+ },
1475
+ {
1476
+ "epoch": 0.04663408974273155,
1477
+ "grad_norm": 0.490017294883728,
1478
+ "learning_rate": 2.7978580990629185e-06,
1479
+ "loss": 2.24790096282959,
1480
+ "step": 1045
1481
+ },
1482
+ {
1483
+ "epoch": 0.046857219358725485,
1484
+ "grad_norm": 0.5250972509384155,
1485
+ "learning_rate": 2.811244979919679e-06,
1486
+ "loss": 2.1867441177368163,
1487
+ "step": 1050
1488
+ },
1489
+ {
1490
+ "epoch": 0.04708034897471942,
1491
+ "grad_norm": 0.45831558108329773,
1492
+ "learning_rate": 2.824631860776439e-06,
1493
+ "loss": 2.379189300537109,
1494
+ "step": 1055
1495
+ },
1496
+ {
1497
+ "epoch": 0.04730347859071334,
1498
+ "grad_norm": 0.5892948508262634,
1499
+ "learning_rate": 2.8380187416331998e-06,
1500
+ "loss": 2.2735191345214845,
1501
+ "step": 1060
1502
+ },
1503
+ {
1504
+ "epoch": 0.047526608206707274,
1505
+ "grad_norm": 0.5409689545631409,
1506
+ "learning_rate": 2.85140562248996e-06,
1507
+ "loss": 2.3979333877563476,
1508
+ "step": 1065
1509
+ },
1510
+ {
1511
+ "epoch": 0.047749737822701206,
1512
+ "grad_norm": 0.5346503257751465,
1513
+ "learning_rate": 2.8647925033467203e-06,
1514
+ "loss": 2.4819768905639648,
1515
+ "step": 1070
1516
+ },
1517
+ {
1518
+ "epoch": 0.04797286743869514,
1519
+ "grad_norm": 0.6150287985801697,
1520
+ "learning_rate": 2.8781793842034806e-06,
1521
+ "loss": 2.5218339920043946,
1522
+ "step": 1075
1523
+ },
1524
+ {
1525
+ "epoch": 0.04819599705468907,
1526
+ "grad_norm": 0.4347957968711853,
1527
+ "learning_rate": 2.891566265060241e-06,
1528
+ "loss": 2.1929386138916014,
1529
+ "step": 1080
1530
+ },
1531
+ {
1532
+ "epoch": 0.048419126670683,
1533
+ "grad_norm": 0.5238802433013916,
1534
+ "learning_rate": 2.9049531459170016e-06,
1535
+ "loss": 2.215384292602539,
1536
+ "step": 1085
1537
+ },
1538
+ {
1539
+ "epoch": 0.048642256286676934,
1540
+ "grad_norm": 0.5371809601783752,
1541
+ "learning_rate": 2.918340026773762e-06,
1542
+ "loss": 2.4019641876220703,
1543
+ "step": 1090
1544
+ },
1545
+ {
1546
+ "epoch": 0.04886538590267086,
1547
+ "grad_norm": 0.5125853419303894,
1548
+ "learning_rate": 2.931726907630522e-06,
1549
+ "loss": 2.3410377502441406,
1550
+ "step": 1095
1551
+ },
1552
+ {
1553
+ "epoch": 0.04908851551866479,
1554
+ "grad_norm": 0.5087509751319885,
1555
+ "learning_rate": 2.9451137884872824e-06,
1556
+ "loss": 2.3634616851806642,
1557
+ "step": 1100
1558
+ },
1559
+ {
1560
+ "epoch": 0.04931164513465872,
1561
+ "grad_norm": 0.5605607628822327,
1562
+ "learning_rate": 2.9585006693440427e-06,
1563
+ "loss": 2.239984130859375,
1564
+ "step": 1105
1565
+ },
1566
+ {
1567
+ "epoch": 0.049534774750652655,
1568
+ "grad_norm": 0.6371492743492126,
1569
+ "learning_rate": 2.9718875502008034e-06,
1570
+ "loss": 2.499210739135742,
1571
+ "step": 1110
1572
+ },
1573
+ {
1574
+ "epoch": 0.04975790436664659,
1575
+ "grad_norm": 0.5526962280273438,
1576
+ "learning_rate": 2.9852744310575636e-06,
1577
+ "loss": 2.433857536315918,
1578
+ "step": 1115
1579
+ },
1580
+ {
1581
+ "epoch": 0.04998103398264052,
1582
+ "grad_norm": 0.5811582803726196,
1583
+ "learning_rate": 2.998661311914324e-06,
1584
+ "loss": 2.372517967224121,
1585
+ "step": 1120
1586
+ },
1587
+ {
1588
+ "epoch": 0.050204163598634444,
1589
+ "grad_norm": 0.47710540890693665,
1590
+ "learning_rate": 3.0120481927710846e-06,
1591
+ "loss": 2.441071891784668,
1592
+ "step": 1125
1593
+ },
1594
+ {
1595
+ "epoch": 0.050427293214628376,
1596
+ "grad_norm": 0.49463921785354614,
1597
+ "learning_rate": 3.0254350736278445e-06,
1598
+ "loss": 2.218737030029297,
1599
+ "step": 1130
1600
+ },
1601
+ {
1602
+ "epoch": 0.05065042283062231,
1603
+ "grad_norm": 0.7574100494384766,
1604
+ "learning_rate": 3.0388219544846056e-06,
1605
+ "loss": 2.1490055084228517,
1606
+ "step": 1135
1607
+ },
1608
+ {
1609
+ "epoch": 0.05087355244661624,
1610
+ "grad_norm": 0.5574637651443481,
1611
+ "learning_rate": 3.0522088353413654e-06,
1612
+ "loss": 2.3484785079956056,
1613
+ "step": 1140
1614
+ },
1615
+ {
1616
+ "epoch": 0.05109668206261017,
1617
+ "grad_norm": 0.5128051042556763,
1618
+ "learning_rate": 3.0655957161981257e-06,
1619
+ "loss": 2.2358184814453126,
1620
+ "step": 1145
1621
+ },
1622
+ {
1623
+ "epoch": 0.0513198116786041,
1624
+ "grad_norm": 0.4747551679611206,
1625
+ "learning_rate": 3.0789825970548864e-06,
1626
+ "loss": 2.2787382125854494,
1627
+ "step": 1150
1628
+ },
1629
+ {
1630
+ "epoch": 0.051542941294598035,
1631
+ "grad_norm": 0.4365915358066559,
1632
+ "learning_rate": 3.0923694779116467e-06,
1633
+ "loss": 2.327534484863281,
1634
+ "step": 1155
1635
+ },
1636
+ {
1637
+ "epoch": 0.05176607091059196,
1638
+ "grad_norm": 0.6433975696563721,
1639
+ "learning_rate": 3.1057563587684074e-06,
1640
+ "loss": 2.288431167602539,
1641
+ "step": 1160
1642
+ },
1643
+ {
1644
+ "epoch": 0.05198920052658589,
1645
+ "grad_norm": 0.6102743744850159,
1646
+ "learning_rate": 3.1191432396251672e-06,
1647
+ "loss": 2.24114933013916,
1648
+ "step": 1165
1649
+ },
1650
+ {
1651
+ "epoch": 0.052212330142579824,
1652
+ "grad_norm": 0.5493948459625244,
1653
+ "learning_rate": 3.132530120481928e-06,
1654
+ "loss": 2.317499542236328,
1655
+ "step": 1170
1656
+ },
1657
+ {
1658
+ "epoch": 0.052435459758573756,
1659
+ "grad_norm": 0.5411773324012756,
1660
+ "learning_rate": 3.145917001338688e-06,
1661
+ "loss": 2.1778676986694334,
1662
+ "step": 1175
1663
+ },
1664
+ {
1665
+ "epoch": 0.05265858937456769,
1666
+ "grad_norm": 0.5491811633110046,
1667
+ "learning_rate": 3.1593038821954485e-06,
1668
+ "loss": 2.260644721984863,
1669
+ "step": 1180
1670
+ },
1671
+ {
1672
+ "epoch": 0.05288171899056162,
1673
+ "grad_norm": 0.5138102769851685,
1674
+ "learning_rate": 3.172690763052209e-06,
1675
+ "loss": 2.2743204116821287,
1676
+ "step": 1185
1677
+ },
1678
+ {
1679
+ "epoch": 0.053104848606555545,
1680
+ "grad_norm": 0.5310954451560974,
1681
+ "learning_rate": 3.186077643908969e-06,
1682
+ "loss": 2.257237434387207,
1683
+ "step": 1190
1684
+ },
1685
+ {
1686
+ "epoch": 0.05332797822254948,
1687
+ "grad_norm": 0.5704927444458008,
1688
+ "learning_rate": 3.1994645247657297e-06,
1689
+ "loss": 2.0418323516845702,
1690
+ "step": 1195
1691
+ },
1692
+ {
1693
+ "epoch": 0.05355110783854341,
1694
+ "grad_norm": 0.5295515656471252,
1695
+ "learning_rate": 3.21285140562249e-06,
1696
+ "loss": 2.282415771484375,
1697
+ "step": 1200
1698
+ },
1699
+ {
1700
+ "epoch": 0.05377423745453734,
1701
+ "grad_norm": 0.6476492881774902,
1702
+ "learning_rate": 3.2262382864792503e-06,
1703
+ "loss": 2.306402015686035,
1704
+ "step": 1205
1705
+ },
1706
+ {
1707
+ "epoch": 0.05399736707053127,
1708
+ "grad_norm": 0.6115656495094299,
1709
+ "learning_rate": 3.239625167336011e-06,
1710
+ "loss": 2.1535377502441406,
1711
+ "step": 1210
1712
+ },
1713
+ {
1714
+ "epoch": 0.054220496686525205,
1715
+ "grad_norm": 0.5802296996116638,
1716
+ "learning_rate": 3.253012048192771e-06,
1717
+ "loss": 2.2679920196533203,
1718
+ "step": 1215
1719
+ },
1720
+ {
1721
+ "epoch": 0.05444362630251914,
1722
+ "grad_norm": 0.5182514190673828,
1723
+ "learning_rate": 3.266398929049532e-06,
1724
+ "loss": 2.3105010986328125,
1725
+ "step": 1220
1726
+ },
1727
+ {
1728
+ "epoch": 0.05466675591851306,
1729
+ "grad_norm": 0.6288827657699585,
1730
+ "learning_rate": 3.279785809906292e-06,
1731
+ "loss": 2.0794536590576174,
1732
+ "step": 1225
1733
+ },
1734
+ {
1735
+ "epoch": 0.054889885534506994,
1736
+ "grad_norm": 0.7542652487754822,
1737
+ "learning_rate": 3.2931726907630525e-06,
1738
+ "loss": 2.3161176681518554,
1739
+ "step": 1230
1740
+ },
1741
+ {
1742
+ "epoch": 0.055113015150500926,
1743
+ "grad_norm": 0.7416955232620239,
1744
+ "learning_rate": 3.3065595716198128e-06,
1745
+ "loss": 2.279362106323242,
1746
+ "step": 1235
1747
+ },
1748
+ {
1749
+ "epoch": 0.05533614476649486,
1750
+ "grad_norm": 0.5682216882705688,
1751
+ "learning_rate": 3.3199464524765726e-06,
1752
+ "loss": 2.2426963806152345,
1753
+ "step": 1240
1754
+ },
1755
+ {
1756
+ "epoch": 0.05555927438248879,
1757
+ "grad_norm": 0.6715278625488281,
1758
+ "learning_rate": 3.3333333333333337e-06,
1759
+ "loss": 2.1546775817871096,
1760
+ "step": 1245
1761
+ },
1762
+ {
1763
+ "epoch": 0.05578240399848272,
1764
+ "grad_norm": 0.5482577085494995,
1765
+ "learning_rate": 3.3467202141900936e-06,
1766
+ "loss": 2.1484096527099608,
1767
+ "step": 1250
1768
+ },
1769
+ {
1770
+ "epoch": 0.056005533614476646,
1771
+ "grad_norm": 0.6009266972541809,
1772
+ "learning_rate": 3.3601070950468543e-06,
1773
+ "loss": 2.221162796020508,
1774
+ "step": 1255
1775
+ },
1776
+ {
1777
+ "epoch": 0.05622866323047058,
1778
+ "grad_norm": 0.6640043258666992,
1779
+ "learning_rate": 3.3734939759036146e-06,
1780
+ "loss": 2.338100242614746,
1781
+ "step": 1260
1782
+ },
1783
+ {
1784
+ "epoch": 0.05645179284646451,
1785
+ "grad_norm": 0.6514598727226257,
1786
+ "learning_rate": 3.3868808567603744e-06,
1787
+ "loss": 2.2825199127197267,
1788
+ "step": 1265
1789
+ },
1790
+ {
1791
+ "epoch": 0.05667492246245844,
1792
+ "grad_norm": 0.6053763031959534,
1793
+ "learning_rate": 3.4002677376171355e-06,
1794
+ "loss": 2.280612754821777,
1795
+ "step": 1270
1796
+ },
1797
+ {
1798
+ "epoch": 0.056898052078452374,
1799
+ "grad_norm": 0.8854132294654846,
1800
+ "learning_rate": 3.4136546184738954e-06,
1801
+ "loss": 2.3481237411499025,
1802
+ "step": 1275
1803
+ },
1804
+ {
1805
+ "epoch": 0.057121181694446306,
1806
+ "grad_norm": 0.6433872580528259,
1807
+ "learning_rate": 3.427041499330656e-06,
1808
+ "loss": 2.2561552047729494,
1809
+ "step": 1280
1810
+ },
1811
+ {
1812
+ "epoch": 0.05734431131044024,
1813
+ "grad_norm": 0.709402859210968,
1814
+ "learning_rate": 3.4404283801874164e-06,
1815
+ "loss": 2.264519691467285,
1816
+ "step": 1285
1817
+ },
1818
+ {
1819
+ "epoch": 0.05756744092643416,
1820
+ "grad_norm": 0.5431029796600342,
1821
+ "learning_rate": 3.453815261044177e-06,
1822
+ "loss": 2.163587760925293,
1823
+ "step": 1290
1824
+ },
1825
+ {
1826
+ "epoch": 0.057790570542428095,
1827
+ "grad_norm": 0.7146894931793213,
1828
+ "learning_rate": 3.4672021419009373e-06,
1829
+ "loss": 2.2624624252319334,
1830
+ "step": 1295
1831
+ },
1832
+ {
1833
+ "epoch": 0.05801370015842203,
1834
+ "grad_norm": 0.6029372215270996,
1835
+ "learning_rate": 3.480589022757697e-06,
1836
+ "loss": 2.0209115982055663,
1837
+ "step": 1300
1838
+ },
1839
+ {
1840
+ "epoch": 0.05823682977441596,
1841
+ "grad_norm": 0.6099830865859985,
1842
+ "learning_rate": 3.493975903614458e-06,
1843
+ "loss": 2.177568054199219,
1844
+ "step": 1305
1845
+ },
1846
+ {
1847
+ "epoch": 0.05845995939040989,
1848
+ "grad_norm": 0.8421804308891296,
1849
+ "learning_rate": 3.507362784471218e-06,
1850
+ "loss": 2.1980712890625,
1851
+ "step": 1310
1852
+ },
1853
+ {
1854
+ "epoch": 0.05868308900640382,
1855
+ "grad_norm": 0.7849622368812561,
1856
+ "learning_rate": 3.520749665327979e-06,
1857
+ "loss": 2.2047344207763673,
1858
+ "step": 1315
1859
+ },
1860
+ {
1861
+ "epoch": 0.05890621862239775,
1862
+ "grad_norm": 0.5906243324279785,
1863
+ "learning_rate": 3.534136546184739e-06,
1864
+ "loss": 2.3746465682983398,
1865
+ "step": 1320
1866
+ },
1867
+ {
1868
+ "epoch": 0.05912934823839168,
1869
+ "grad_norm": 0.9020318388938904,
1870
+ "learning_rate": 3.547523427041499e-06,
1871
+ "loss": 2.2612667083740234,
1872
+ "step": 1325
1873
+ },
1874
+ {
1875
+ "epoch": 0.05935247785438561,
1876
+ "grad_norm": 0.6542187929153442,
1877
+ "learning_rate": 3.5609103078982597e-06,
1878
+ "loss": 2.2060800552368165,
1879
+ "step": 1330
1880
+ },
1881
+ {
1882
+ "epoch": 0.059575607470379544,
1883
+ "grad_norm": 0.7428557872772217,
1884
+ "learning_rate": 3.57429718875502e-06,
1885
+ "loss": 2.0877670288085937,
1886
+ "step": 1335
1887
+ },
1888
+ {
1889
+ "epoch": 0.059798737086373475,
1890
+ "grad_norm": 0.7468170523643494,
1891
+ "learning_rate": 3.5876840696117807e-06,
1892
+ "loss": 2.3160980224609373,
1893
+ "step": 1340
1894
+ },
1895
+ {
1896
+ "epoch": 0.06002186670236741,
1897
+ "grad_norm": 0.8021371960639954,
1898
+ "learning_rate": 3.601070950468541e-06,
1899
+ "loss": 2.1923311233520506,
1900
+ "step": 1345
1901
+ },
1902
+ {
1903
+ "epoch": 0.06024499631836134,
1904
+ "grad_norm": 0.6904175281524658,
1905
+ "learning_rate": 3.6144578313253016e-06,
1906
+ "loss": 2.1911346435546877,
1907
+ "step": 1350
1908
+ },
1909
+ {
1910
+ "epoch": 0.060468125934355264,
1911
+ "grad_norm": 0.7336163520812988,
1912
+ "learning_rate": 3.627844712182062e-06,
1913
+ "loss": 2.2488683700561523,
1914
+ "step": 1355
1915
+ },
1916
+ {
1917
+ "epoch": 0.060691255550349196,
1918
+ "grad_norm": 0.7925560474395752,
1919
+ "learning_rate": 3.6412315930388218e-06,
1920
+ "loss": 2.174961280822754,
1921
+ "step": 1360
1922
+ },
1923
+ {
1924
+ "epoch": 0.06091438516634313,
1925
+ "grad_norm": 0.8825748562812805,
1926
+ "learning_rate": 3.6546184738955825e-06,
1927
+ "loss": 2.012922668457031,
1928
+ "step": 1365
1929
+ },
1930
+ {
1931
+ "epoch": 0.06113751478233706,
1932
+ "grad_norm": 0.6754929423332214,
1933
+ "learning_rate": 3.6680053547523427e-06,
1934
+ "loss": 2.216470146179199,
1935
+ "step": 1370
1936
+ },
1937
+ {
1938
+ "epoch": 0.06136064439833099,
1939
+ "grad_norm": 0.6616571545600891,
1940
+ "learning_rate": 3.6813922356091034e-06,
1941
+ "loss": 2.067160415649414,
1942
+ "step": 1375
1943
+ },
1944
+ {
1945
+ "epoch": 0.061583774014324924,
1946
+ "grad_norm": 0.7033371925354004,
1947
+ "learning_rate": 3.6947791164658637e-06,
1948
+ "loss": 2.2197628021240234,
1949
+ "step": 1380
1950
+ },
1951
+ {
1952
+ "epoch": 0.06180690363031885,
1953
+ "grad_norm": 0.7256376147270203,
1954
+ "learning_rate": 3.7081659973226235e-06,
1955
+ "loss": 2.2435529708862303,
1956
+ "step": 1385
1957
+ },
1958
+ {
1959
+ "epoch": 0.06203003324631278,
1960
+ "grad_norm": 0.6585844159126282,
1961
+ "learning_rate": 3.7215528781793842e-06,
1962
+ "loss": 2.113719177246094,
1963
+ "step": 1390
1964
+ },
1965
+ {
1966
+ "epoch": 0.06225316286230671,
1967
+ "grad_norm": 0.7192341685295105,
1968
+ "learning_rate": 3.7349397590361445e-06,
1969
+ "loss": 2.2271860122680662,
1970
+ "step": 1395
1971
+ },
1972
+ {
1973
+ "epoch": 0.062476292478300645,
1974
+ "grad_norm": 0.7313777804374695,
1975
+ "learning_rate": 3.7483266398929052e-06,
1976
+ "loss": 2.312770462036133,
1977
+ "step": 1400
1978
+ },
1979
+ {
1980
+ "epoch": 0.06269942209429458,
1981
+ "grad_norm": 0.7106947302818298,
1982
+ "learning_rate": 3.7617135207496655e-06,
1983
+ "loss": 2.041255760192871,
1984
+ "step": 1405
1985
+ },
1986
+ {
1987
+ "epoch": 0.0629225517102885,
1988
+ "grad_norm": 0.7265296578407288,
1989
+ "learning_rate": 3.775100401606426e-06,
1990
+ "loss": 2.2530691146850588,
1991
+ "step": 1410
1992
+ },
1993
+ {
1994
+ "epoch": 0.06314568132628244,
1995
+ "grad_norm": 0.7085316777229309,
1996
+ "learning_rate": 3.788487282463186e-06,
1997
+ "loss": 2.03436222076416,
1998
+ "step": 1415
1999
+ },
2000
+ {
2001
+ "epoch": 0.06336881094227637,
2002
+ "grad_norm": 0.7872986793518066,
2003
+ "learning_rate": 3.8018741633199463e-06,
2004
+ "loss": 2.0966476440429687,
2005
+ "step": 1420
2006
+ },
2007
+ {
2008
+ "epoch": 0.0635919405582703,
2009
+ "grad_norm": 0.8961313366889954,
2010
+ "learning_rate": 3.8152610441767074e-06,
2011
+ "loss": 2.1432809829711914,
2012
+ "step": 1425
2013
+ },
2014
+ {
2015
+ "epoch": 0.06381507017426423,
2016
+ "grad_norm": 0.7127321362495422,
2017
+ "learning_rate": 3.828647925033467e-06,
2018
+ "loss": 2.2022947311401366,
2019
+ "step": 1430
2020
+ },
2021
+ {
2022
+ "epoch": 0.06403819979025815,
2023
+ "grad_norm": 0.8228124976158142,
2024
+ "learning_rate": 3.842034805890228e-06,
2025
+ "loss": 2.39965705871582,
2026
+ "step": 1435
2027
+ },
2028
+ {
2029
+ "epoch": 0.0642613294062521,
2030
+ "grad_norm": 0.775909960269928,
2031
+ "learning_rate": 3.855421686746988e-06,
2032
+ "loss": 2.1313138961791993,
2033
+ "step": 1440
2034
+ },
2035
+ {
2036
+ "epoch": 0.06448445902224602,
2037
+ "grad_norm": 0.7945120334625244,
2038
+ "learning_rate": 3.8688085676037485e-06,
2039
+ "loss": 2.074580955505371,
2040
+ "step": 1445
2041
+ },
2042
+ {
2043
+ "epoch": 0.06470758863823996,
2044
+ "grad_norm": 1.014404535293579,
2045
+ "learning_rate": 3.882195448460509e-06,
2046
+ "loss": 2.26824893951416,
2047
+ "step": 1450
2048
+ },
2049
+ {
2050
+ "epoch": 0.06493071825423388,
2051
+ "grad_norm": 0.986544132232666,
2052
+ "learning_rate": 3.895582329317269e-06,
2053
+ "loss": 2.207651138305664,
2054
+ "step": 1455
2055
+ },
2056
+ {
2057
+ "epoch": 0.06515384787022782,
2058
+ "grad_norm": 0.7337960004806519,
2059
+ "learning_rate": 3.90896921017403e-06,
2060
+ "loss": 2.2294179916381838,
2061
+ "step": 1460
2062
+ },
2063
+ {
2064
+ "epoch": 0.06537697748622175,
2065
+ "grad_norm": 1.2697299718856812,
2066
+ "learning_rate": 3.92235609103079e-06,
2067
+ "loss": 2.114596366882324,
2068
+ "step": 1465
2069
+ },
2070
+ {
2071
+ "epoch": 0.06560010710221567,
2072
+ "grad_norm": 0.8556989431381226,
2073
+ "learning_rate": 3.93574297188755e-06,
2074
+ "loss": 2.21254940032959,
2075
+ "step": 1470
2076
+ },
2077
+ {
2078
+ "epoch": 0.06582323671820961,
2079
+ "grad_norm": 0.9477710723876953,
2080
+ "learning_rate": 3.949129852744311e-06,
2081
+ "loss": 1.977022361755371,
2082
+ "step": 1475
2083
+ },
2084
+ {
2085
+ "epoch": 0.06604636633420354,
2086
+ "grad_norm": 1.7189115285873413,
2087
+ "learning_rate": 3.962516733601071e-06,
2088
+ "loss": 2.012314224243164,
2089
+ "step": 1480
2090
+ },
2091
+ {
2092
+ "epoch": 0.06626949595019747,
2093
+ "grad_norm": 0.826131284236908,
2094
+ "learning_rate": 3.975903614457832e-06,
2095
+ "loss": 2.193516731262207,
2096
+ "step": 1485
2097
+ },
2098
+ {
2099
+ "epoch": 0.0664926255661914,
2100
+ "grad_norm": 1.0023162364959717,
2101
+ "learning_rate": 3.9892904953145914e-06,
2102
+ "loss": 2.2499979019165037,
2103
+ "step": 1490
2104
+ },
2105
+ {
2106
+ "epoch": 0.06671575518218534,
2107
+ "grad_norm": 0.7376646399497986,
2108
+ "learning_rate": 4.002677376171352e-06,
2109
+ "loss": 2.1386489868164062,
2110
+ "step": 1495
2111
+ },
2112
+ {
2113
+ "epoch": 0.06693888479817926,
2114
+ "grad_norm": 0.7226489782333374,
2115
+ "learning_rate": 4.016064257028113e-06,
2116
+ "loss": 2.11199836730957,
2117
+ "step": 1500
2118
+ },
2119
+ {
2120
+ "epoch": 0.06716201441417319,
2121
+ "grad_norm": 0.8427999019622803,
2122
+ "learning_rate": 4.029451137884873e-06,
2123
+ "loss": 2.034942626953125,
2124
+ "step": 1505
2125
+ },
2126
+ {
2127
+ "epoch": 0.06738514403016713,
2128
+ "grad_norm": 0.79557204246521,
2129
+ "learning_rate": 4.042838018741633e-06,
2130
+ "loss": 2.145208549499512,
2131
+ "step": 1510
2132
+ },
2133
+ {
2134
+ "epoch": 0.06760827364616105,
2135
+ "grad_norm": 0.7423805594444275,
2136
+ "learning_rate": 4.056224899598393e-06,
2137
+ "loss": 2.2282032012939452,
2138
+ "step": 1515
2139
+ },
2140
+ {
2141
+ "epoch": 0.06783140326215499,
2142
+ "grad_norm": 0.8701015114784241,
2143
+ "learning_rate": 4.069611780455154e-06,
2144
+ "loss": 2.0600002288818358,
2145
+ "step": 1520
2146
+ },
2147
+ {
2148
+ "epoch": 0.06805453287814892,
2149
+ "grad_norm": 0.9967679977416992,
2150
+ "learning_rate": 4.082998661311915e-06,
2151
+ "loss": 2.20194034576416,
2152
+ "step": 1525
2153
+ },
2154
+ {
2155
+ "epoch": 0.06827766249414285,
2156
+ "grad_norm": 1.0514639616012573,
2157
+ "learning_rate": 4.096385542168675e-06,
2158
+ "loss": 2.250165557861328,
2159
+ "step": 1530
2160
+ },
2161
+ {
2162
+ "epoch": 0.06850079211013678,
2163
+ "grad_norm": 0.7402302026748657,
2164
+ "learning_rate": 4.109772423025435e-06,
2165
+ "loss": 2.0741525650024415,
2166
+ "step": 1535
2167
+ },
2168
+ {
2169
+ "epoch": 0.0687239217261307,
2170
+ "grad_norm": 1.0764678716659546,
2171
+ "learning_rate": 4.123159303882195e-06,
2172
+ "loss": 2.068874549865723,
2173
+ "step": 1540
2174
+ },
2175
+ {
2176
+ "epoch": 0.06894705134212464,
2177
+ "grad_norm": 0.9718510508537292,
2178
+ "learning_rate": 4.136546184738956e-06,
2179
+ "loss": 1.9750284194946288,
2180
+ "step": 1545
2181
+ },
2182
+ {
2183
+ "epoch": 0.06917018095811857,
2184
+ "grad_norm": 0.7534152269363403,
2185
+ "learning_rate": 4.149933065595716e-06,
2186
+ "loss": 2.1882530212402345,
2187
+ "step": 1550
2188
+ },
2189
+ {
2190
+ "epoch": 0.06939331057411251,
2191
+ "grad_norm": 0.8103408813476562,
2192
+ "learning_rate": 4.163319946452477e-06,
2193
+ "loss": 2.076101875305176,
2194
+ "step": 1555
2195
+ },
2196
+ {
2197
+ "epoch": 0.06961644019010643,
2198
+ "grad_norm": 0.7058753967285156,
2199
+ "learning_rate": 4.176706827309237e-06,
2200
+ "loss": 2.1349681854248046,
2201
+ "step": 1560
2202
+ },
2203
+ {
2204
+ "epoch": 0.06983956980610036,
2205
+ "grad_norm": 0.664568305015564,
2206
+ "learning_rate": 4.190093708165997e-06,
2207
+ "loss": 2.1840320587158204,
2208
+ "step": 1565
2209
+ },
2210
+ {
2211
+ "epoch": 0.0700626994220943,
2212
+ "grad_norm": 0.9274365901947021,
2213
+ "learning_rate": 4.2034805890227575e-06,
2214
+ "loss": 2.0162349700927735,
2215
+ "step": 1570
2216
+ },
2217
+ {
2218
+ "epoch": 0.07028582903808822,
2219
+ "grad_norm": 0.806480348110199,
2220
+ "learning_rate": 4.216867469879518e-06,
2221
+ "loss": 2.1628509521484376,
2222
+ "step": 1575
2223
+ },
2224
+ {
2225
+ "epoch": 0.07050895865408216,
2226
+ "grad_norm": 0.9540587067604065,
2227
+ "learning_rate": 4.230254350736279e-06,
2228
+ "loss": 2.239594841003418,
2229
+ "step": 1580
2230
+ },
2231
+ {
2232
+ "epoch": 0.07073208827007609,
2233
+ "grad_norm": 0.8032299876213074,
2234
+ "learning_rate": 4.243641231593039e-06,
2235
+ "loss": 2.170275115966797,
2236
+ "step": 1585
2237
+ },
2238
+ {
2239
+ "epoch": 0.07095521788607002,
2240
+ "grad_norm": 0.9516310095787048,
2241
+ "learning_rate": 4.2570281124497995e-06,
2242
+ "loss": 2.035562515258789,
2243
+ "step": 1590
2244
+ },
2245
+ {
2246
+ "epoch": 0.07117834750206395,
2247
+ "grad_norm": 0.8803707361221313,
2248
+ "learning_rate": 4.270414993306559e-06,
2249
+ "loss": 2.1310283660888674,
2250
+ "step": 1595
2251
+ },
2252
+ {
2253
+ "epoch": 0.07140147711805787,
2254
+ "grad_norm": 1.022939682006836,
2255
+ "learning_rate": 4.28380187416332e-06,
2256
+ "loss": 2.1093021392822267,
2257
+ "step": 1600
2258
+ },
2259
+ {
2260
+ "epoch": 0.07162460673405181,
2261
+ "grad_norm": 0.8254362344741821,
2262
+ "learning_rate": 4.297188755020081e-06,
2263
+ "loss": 1.8178255081176757,
2264
+ "step": 1605
2265
+ },
2266
+ {
2267
+ "epoch": 0.07184773635004574,
2268
+ "grad_norm": 0.779062807559967,
2269
+ "learning_rate": 4.3105756358768406e-06,
2270
+ "loss": 2.261147880554199,
2271
+ "step": 1610
2272
+ },
2273
+ {
2274
+ "epoch": 0.07207086596603968,
2275
+ "grad_norm": 0.8894750475883484,
2276
+ "learning_rate": 4.323962516733601e-06,
2277
+ "loss": 2.0283786773681642,
2278
+ "step": 1615
2279
+ },
2280
+ {
2281
+ "epoch": 0.0722939955820336,
2282
+ "grad_norm": 0.9776813387870789,
2283
+ "learning_rate": 4.337349397590362e-06,
2284
+ "loss": 2.177045440673828,
2285
+ "step": 1620
2286
+ },
2287
+ {
2288
+ "epoch": 0.07251712519802754,
2289
+ "grad_norm": 1.2759140729904175,
2290
+ "learning_rate": 4.350736278447122e-06,
2291
+ "loss": 2.0527717590332033,
2292
+ "step": 1625
2293
+ },
2294
+ {
2295
+ "epoch": 0.07274025481402147,
2296
+ "grad_norm": 0.9953216314315796,
2297
+ "learning_rate": 4.3641231593038825e-06,
2298
+ "loss": 2.05969352722168,
2299
+ "step": 1630
2300
+ },
2301
+ {
2302
+ "epoch": 0.07296338443001539,
2303
+ "grad_norm": 0.8319527506828308,
2304
+ "learning_rate": 4.377510040160642e-06,
2305
+ "loss": 2.003727340698242,
2306
+ "step": 1635
2307
+ },
2308
+ {
2309
+ "epoch": 0.07318651404600933,
2310
+ "grad_norm": 0.8335089087486267,
2311
+ "learning_rate": 4.390896921017403e-06,
2312
+ "loss": 2.0173160552978517,
2313
+ "step": 1640
2314
+ },
2315
+ {
2316
+ "epoch": 0.07340964366200325,
2317
+ "grad_norm": 1.1071332693099976,
2318
+ "learning_rate": 4.404283801874164e-06,
2319
+ "loss": 2.1184307098388673,
2320
+ "step": 1645
2321
+ },
2322
+ {
2323
+ "epoch": 0.0736327732779972,
2324
+ "grad_norm": 1.2880672216415405,
2325
+ "learning_rate": 4.4176706827309244e-06,
2326
+ "loss": 2.0427175521850587,
2327
+ "step": 1650
2328
+ },
2329
+ {
2330
+ "epoch": 0.07385590289399112,
2331
+ "grad_norm": 1.1434367895126343,
2332
+ "learning_rate": 4.431057563587684e-06,
2333
+ "loss": 2.0279747009277345,
2334
+ "step": 1655
2335
+ },
2336
+ {
2337
+ "epoch": 0.07407903250998506,
2338
+ "grad_norm": 0.7727090716362,
2339
+ "learning_rate": 4.444444444444444e-06,
2340
+ "loss": 2.1546749114990233,
2341
+ "step": 1660
2342
+ },
2343
+ {
2344
+ "epoch": 0.07430216212597898,
2345
+ "grad_norm": 1.0890729427337646,
2346
+ "learning_rate": 4.457831325301205e-06,
2347
+ "loss": 2.1247718811035154,
2348
+ "step": 1665
2349
+ },
2350
+ {
2351
+ "epoch": 0.07452529174197291,
2352
+ "grad_norm": 1.0179014205932617,
2353
+ "learning_rate": 4.4712182061579655e-06,
2354
+ "loss": 2.146767234802246,
2355
+ "step": 1670
2356
+ },
2357
+ {
2358
+ "epoch": 0.07474842135796685,
2359
+ "grad_norm": 0.745704174041748,
2360
+ "learning_rate": 4.484605087014726e-06,
2361
+ "loss": 2.224173736572266,
2362
+ "step": 1675
2363
+ },
2364
+ {
2365
+ "epoch": 0.07497155097396077,
2366
+ "grad_norm": 0.771865963935852,
2367
+ "learning_rate": 4.497991967871486e-06,
2368
+ "loss": 2.2265625,
2369
+ "step": 1680
2370
+ },
2371
+ {
2372
+ "epoch": 0.07519468058995471,
2373
+ "grad_norm": 1.352388620376587,
2374
+ "learning_rate": 4.511378848728246e-06,
2375
+ "loss": 2.261865425109863,
2376
+ "step": 1685
2377
+ },
2378
+ {
2379
+ "epoch": 0.07541781020594863,
2380
+ "grad_norm": 0.8781198859214783,
2381
+ "learning_rate": 4.524765729585007e-06,
2382
+ "loss": 2.0096199035644533,
2383
+ "step": 1690
2384
+ },
2385
+ {
2386
+ "epoch": 0.07564093982194256,
2387
+ "grad_norm": 1.0159834623336792,
2388
+ "learning_rate": 4.538152610441767e-06,
2389
+ "loss": 2.0243366241455076,
2390
+ "step": 1695
2391
+ },
2392
+ {
2393
+ "epoch": 0.0758640694379365,
2394
+ "grad_norm": 0.8101127743721008,
2395
+ "learning_rate": 4.551539491298528e-06,
2396
+ "loss": 2.119925117492676,
2397
+ "step": 1700
2398
+ },
2399
+ {
2400
+ "epoch": 0.07608719905393042,
2401
+ "grad_norm": 1.0311274528503418,
2402
+ "learning_rate": 4.564926372155288e-06,
2403
+ "loss": 1.9274934768676757,
2404
+ "step": 1705
2405
+ },
2406
+ {
2407
+ "epoch": 0.07631032866992436,
2408
+ "grad_norm": 1.315838098526001,
2409
+ "learning_rate": 4.578313253012049e-06,
2410
+ "loss": 2.058757209777832,
2411
+ "step": 1710
2412
+ },
2413
+ {
2414
+ "epoch": 0.07653345828591829,
2415
+ "grad_norm": 1.016562819480896,
2416
+ "learning_rate": 4.5917001338688084e-06,
2417
+ "loss": 2.0910503387451174,
2418
+ "step": 1715
2419
+ },
2420
+ {
2421
+ "epoch": 0.07675658790191223,
2422
+ "grad_norm": 1.1127480268478394,
2423
+ "learning_rate": 4.605087014725569e-06,
2424
+ "loss": 2.0274993896484377,
2425
+ "step": 1720
2426
+ },
2427
+ {
2428
+ "epoch": 0.07697971751790615,
2429
+ "grad_norm": 0.7386570572853088,
2430
+ "learning_rate": 4.61847389558233e-06,
2431
+ "loss": 2.03863468170166,
2432
+ "step": 1725
2433
+ },
2434
+ {
2435
+ "epoch": 0.07720284713390008,
2436
+ "grad_norm": 0.9372514486312866,
2437
+ "learning_rate": 4.63186077643909e-06,
2438
+ "loss": 1.9647506713867187,
2439
+ "step": 1730
2440
+ },
2441
+ {
2442
+ "epoch": 0.07742597674989402,
2443
+ "grad_norm": 0.866057276725769,
2444
+ "learning_rate": 4.64524765729585e-06,
2445
+ "loss": 2.0213361740112306,
2446
+ "step": 1735
2447
+ },
2448
+ {
2449
+ "epoch": 0.07764910636588794,
2450
+ "grad_norm": 0.8000425100326538,
2451
+ "learning_rate": 4.65863453815261e-06,
2452
+ "loss": 2.009931755065918,
2453
+ "step": 1740
2454
+ },
2455
+ {
2456
+ "epoch": 0.07787223598188188,
2457
+ "grad_norm": 1.4270923137664795,
2458
+ "learning_rate": 4.672021419009371e-06,
2459
+ "loss": 2.0473087310791014,
2460
+ "step": 1745
2461
+ },
2462
+ {
2463
+ "epoch": 0.0780953655978758,
2464
+ "grad_norm": 0.7464487552642822,
2465
+ "learning_rate": 4.685408299866132e-06,
2466
+ "loss": 2.087724304199219,
2467
+ "step": 1750
2468
+ },
2469
+ {
2470
+ "epoch": 0.07831849521386974,
2471
+ "grad_norm": 1.0895863771438599,
2472
+ "learning_rate": 4.6987951807228915e-06,
2473
+ "loss": 2.0931013107299803,
2474
+ "step": 1755
2475
+ },
2476
+ {
2477
+ "epoch": 0.07854162482986367,
2478
+ "grad_norm": 1.2510039806365967,
2479
+ "learning_rate": 4.712182061579652e-06,
2480
+ "loss": 2.0668258666992188,
2481
+ "step": 1760
2482
+ },
2483
+ {
2484
+ "epoch": 0.0787647544458576,
2485
+ "grad_norm": 0.9064348340034485,
2486
+ "learning_rate": 4.725568942436412e-06,
2487
+ "loss": 1.9942918777465821,
2488
+ "step": 1765
2489
+ },
2490
+ {
2491
+ "epoch": 0.07898788406185153,
2492
+ "grad_norm": 1.0130219459533691,
2493
+ "learning_rate": 4.738955823293173e-06,
2494
+ "loss": 2.0953414916992186,
2495
+ "step": 1770
2496
+ },
2497
+ {
2498
+ "epoch": 0.07921101367784546,
2499
+ "grad_norm": 1.0138216018676758,
2500
+ "learning_rate": 4.7523427041499334e-06,
2501
+ "loss": 2.097544479370117,
2502
+ "step": 1775
2503
+ },
2504
+ {
2505
+ "epoch": 0.0794341432938394,
2506
+ "grad_norm": 0.9159922003746033,
2507
+ "learning_rate": 4.765729585006693e-06,
2508
+ "loss": 2.024347686767578,
2509
+ "step": 1780
2510
+ },
2511
+ {
2512
+ "epoch": 0.07965727290983332,
2513
+ "grad_norm": 0.6978763937950134,
2514
+ "learning_rate": 4.779116465863454e-06,
2515
+ "loss": 2.032564163208008,
2516
+ "step": 1785
2517
+ },
2518
+ {
2519
+ "epoch": 0.07988040252582726,
2520
+ "grad_norm": 0.9015496969223022,
2521
+ "learning_rate": 4.792503346720214e-06,
2522
+ "loss": 2.091436767578125,
2523
+ "step": 1790
2524
+ },
2525
+ {
2526
+ "epoch": 0.08010353214182118,
2527
+ "grad_norm": 0.8910471796989441,
2528
+ "learning_rate": 4.8058902275769745e-06,
2529
+ "loss": 2.095396041870117,
2530
+ "step": 1795
2531
+ },
2532
+ {
2533
+ "epoch": 0.08032666175781511,
2534
+ "grad_norm": 0.8550083637237549,
2535
+ "learning_rate": 4.819277108433735e-06,
2536
+ "loss": 2.061022186279297,
2537
+ "step": 1800
2538
+ },
2539
+ {
2540
+ "epoch": 0.08054979137380905,
2541
+ "grad_norm": 0.888156533241272,
2542
+ "learning_rate": 4.832663989290495e-06,
2543
+ "loss": 2.0070331573486326,
2544
+ "step": 1805
2545
+ },
2546
+ {
2547
+ "epoch": 0.08077292098980297,
2548
+ "grad_norm": 1.0453161001205444,
2549
+ "learning_rate": 4.846050870147256e-06,
2550
+ "loss": 2.0485980987548826,
2551
+ "step": 1810
2552
+ },
2553
+ {
2554
+ "epoch": 0.08099605060579691,
2555
+ "grad_norm": 0.835533857345581,
2556
+ "learning_rate": 4.859437751004016e-06,
2557
+ "loss": 2.0737443923950196,
2558
+ "step": 1815
2559
+ },
2560
+ {
2561
+ "epoch": 0.08121918022179084,
2562
+ "grad_norm": 1.3074698448181152,
2563
+ "learning_rate": 4.872824631860777e-06,
2564
+ "loss": 2.204619216918945,
2565
+ "step": 1820
2566
+ },
2567
+ {
2568
+ "epoch": 0.08144230983778476,
2569
+ "grad_norm": 0.952622652053833,
2570
+ "learning_rate": 4.886211512717537e-06,
2571
+ "loss": 1.8864990234375,
2572
+ "step": 1825
2573
+ },
2574
+ {
2575
+ "epoch": 0.0816654394537787,
2576
+ "grad_norm": 1.069469690322876,
2577
+ "learning_rate": 4.899598393574298e-06,
2578
+ "loss": 2.0791854858398438,
2579
+ "step": 1830
2580
+ },
2581
+ {
2582
+ "epoch": 0.08188856906977263,
2583
+ "grad_norm": 1.1299378871917725,
2584
+ "learning_rate": 4.9129852744310576e-06,
2585
+ "loss": 2.0667266845703125,
2586
+ "step": 1835
2587
+ },
2588
+ {
2589
+ "epoch": 0.08211169868576657,
2590
+ "grad_norm": 0.7093334794044495,
2591
+ "learning_rate": 4.926372155287817e-06,
2592
+ "loss": 1.9636056900024415,
2593
+ "step": 1840
2594
+ },
2595
+ {
2596
+ "epoch": 0.08233482830176049,
2597
+ "grad_norm": 1.1765010356903076,
2598
+ "learning_rate": 4.939759036144579e-06,
2599
+ "loss": 2.0537750244140627,
2600
+ "step": 1845
2601
+ },
2602
+ {
2603
+ "epoch": 0.08255795791775443,
2604
+ "grad_norm": 0.8516542315483093,
2605
+ "learning_rate": 4.953145917001339e-06,
2606
+ "loss": 2.1657791137695312,
2607
+ "step": 1850
2608
+ },
2609
+ {
2610
+ "epoch": 0.08278108753374835,
2611
+ "grad_norm": 1.0495437383651733,
2612
+ "learning_rate": 4.9665327978580995e-06,
2613
+ "loss": 2.1656772613525392,
2614
+ "step": 1855
2615
+ },
2616
+ {
2617
+ "epoch": 0.08300421714974228,
2618
+ "grad_norm": 0.8000126481056213,
2619
+ "learning_rate": 4.979919678714859e-06,
2620
+ "loss": 2.020354461669922,
2621
+ "step": 1860
2622
+ },
2623
+ {
2624
+ "epoch": 0.08322734676573622,
2625
+ "grad_norm": 1.2887593507766724,
2626
+ "learning_rate": 4.993306559571619e-06,
2627
+ "loss": 1.9938453674316405,
2628
+ "step": 1865
2629
+ },
2630
+ {
2631
+ "epoch": 0.08345047638173014,
2632
+ "grad_norm": 0.8419036865234375,
2633
+ "learning_rate": 5.006693440428381e-06,
2634
+ "loss": 2.0833532333374025,
2635
+ "step": 1870
2636
+ },
2637
+ {
2638
+ "epoch": 0.08367360599772408,
2639
+ "grad_norm": 0.9748121500015259,
2640
+ "learning_rate": 5.020080321285141e-06,
2641
+ "loss": 2.0065855026245116,
2642
+ "step": 1875
2643
+ },
2644
+ {
2645
+ "epoch": 0.08389673561371801,
2646
+ "grad_norm": 0.8467640280723572,
2647
+ "learning_rate": 5.033467202141901e-06,
2648
+ "loss": 2.2140008926391603,
2649
+ "step": 1880
2650
+ },
2651
+ {
2652
+ "epoch": 0.08411986522971195,
2653
+ "grad_norm": 0.8260283470153809,
2654
+ "learning_rate": 5.046854082998661e-06,
2655
+ "loss": 1.973113441467285,
2656
+ "step": 1885
2657
+ },
2658
+ {
2659
+ "epoch": 0.08434299484570587,
2660
+ "grad_norm": 1.2561461925506592,
2661
+ "learning_rate": 5.060240963855422e-06,
2662
+ "loss": 1.9699825286865233,
2663
+ "step": 1890
2664
+ },
2665
+ {
2666
+ "epoch": 0.0845661244616998,
2667
+ "grad_norm": 0.7021331191062927,
2668
+ "learning_rate": 5.0736278447121826e-06,
2669
+ "loss": 1.9100543975830078,
2670
+ "step": 1895
2671
+ },
2672
+ {
2673
+ "epoch": 0.08478925407769373,
2674
+ "grad_norm": 0.9285181760787964,
2675
+ "learning_rate": 5.087014725568942e-06,
2676
+ "loss": 2.100686454772949,
2677
+ "step": 1900
2678
+ },
2679
+ {
2680
+ "epoch": 0.08501238369368766,
2681
+ "grad_norm": 1.0512628555297852,
2682
+ "learning_rate": 5.100401606425703e-06,
2683
+ "loss": 2.092105484008789,
2684
+ "step": 1905
2685
+ },
2686
+ {
2687
+ "epoch": 0.0852355133096816,
2688
+ "grad_norm": 0.9993311166763306,
2689
+ "learning_rate": 5.113788487282463e-06,
2690
+ "loss": 2.178700256347656,
2691
+ "step": 1910
2692
+ },
2693
+ {
2694
+ "epoch": 0.08545864292567552,
2695
+ "grad_norm": 0.8007418513298035,
2696
+ "learning_rate": 5.127175368139224e-06,
2697
+ "loss": 2.061174774169922,
2698
+ "step": 1915
2699
+ },
2700
+ {
2701
+ "epoch": 0.08568177254166946,
2702
+ "grad_norm": 0.9044376611709595,
2703
+ "learning_rate": 5.140562248995984e-06,
2704
+ "loss": 1.905235481262207,
2705
+ "step": 1920
2706
+ },
2707
+ {
2708
+ "epoch": 0.08590490215766339,
2709
+ "grad_norm": 1.1532313823699951,
2710
+ "learning_rate": 5.153949129852744e-06,
2711
+ "loss": 2.0893394470214846,
2712
+ "step": 1925
2713
+ },
2714
+ {
2715
+ "epoch": 0.08612803177365731,
2716
+ "grad_norm": 1.0485821962356567,
2717
+ "learning_rate": 5.167336010709505e-06,
2718
+ "loss": 1.9315324783325196,
2719
+ "step": 1930
2720
+ },
2721
+ {
2722
+ "epoch": 0.08635116138965125,
2723
+ "grad_norm": 1.1166177988052368,
2724
+ "learning_rate": 5.180722891566265e-06,
2725
+ "loss": 1.9648147583007813,
2726
+ "step": 1935
2727
+ },
2728
+ {
2729
+ "epoch": 0.08657429100564518,
2730
+ "grad_norm": 1.4419307708740234,
2731
+ "learning_rate": 5.1941097724230254e-06,
2732
+ "loss": 2.0460626602172853,
2733
+ "step": 1940
2734
+ },
2735
+ {
2736
+ "epoch": 0.08679742062163912,
2737
+ "grad_norm": 1.0400481224060059,
2738
+ "learning_rate": 5.207496653279786e-06,
2739
+ "loss": 2.034720230102539,
2740
+ "step": 1945
2741
+ },
2742
+ {
2743
+ "epoch": 0.08702055023763304,
2744
+ "grad_norm": 1.041756510734558,
2745
+ "learning_rate": 5.220883534136547e-06,
2746
+ "loss": 1.9372877120971679,
2747
+ "step": 1950
2748
+ },
2749
+ {
2750
+ "epoch": 0.08724367985362697,
2751
+ "grad_norm": 0.9398750066757202,
2752
+ "learning_rate": 5.234270414993307e-06,
2753
+ "loss": 1.9075811386108399,
2754
+ "step": 1955
2755
+ },
2756
+ {
2757
+ "epoch": 0.0874668094696209,
2758
+ "grad_norm": 0.9996817708015442,
2759
+ "learning_rate": 5.2476572958500665e-06,
2760
+ "loss": 2.04040412902832,
2761
+ "step": 1960
2762
+ },
2763
+ {
2764
+ "epoch": 0.08768993908561483,
2765
+ "grad_norm": 1.1434073448181152,
2766
+ "learning_rate": 5.261044176706827e-06,
2767
+ "loss": 2.05865535736084,
2768
+ "step": 1965
2769
+ },
2770
+ {
2771
+ "epoch": 0.08791306870160877,
2772
+ "grad_norm": 1.1526947021484375,
2773
+ "learning_rate": 5.274431057563588e-06,
2774
+ "loss": 2.0584930419921874,
2775
+ "step": 1970
2776
+ },
2777
+ {
2778
+ "epoch": 0.0881361983176027,
2779
+ "grad_norm": 1.1225361824035645,
2780
+ "learning_rate": 5.287817938420349e-06,
2781
+ "loss": 1.869145965576172,
2782
+ "step": 1975
2783
+ },
2784
+ {
2785
+ "epoch": 0.08835932793359663,
2786
+ "grad_norm": 1.0828170776367188,
2787
+ "learning_rate": 5.3012048192771085e-06,
2788
+ "loss": 2.0549264907836915,
2789
+ "step": 1980
2790
+ },
2791
+ {
2792
+ "epoch": 0.08858245754959056,
2793
+ "grad_norm": 0.9480776786804199,
2794
+ "learning_rate": 5.314591700133868e-06,
2795
+ "loss": 1.987673568725586,
2796
+ "step": 1985
2797
+ },
2798
+ {
2799
+ "epoch": 0.08880558716558448,
2800
+ "grad_norm": 0.9799356460571289,
2801
+ "learning_rate": 5.327978580990629e-06,
2802
+ "loss": 1.9657482147216796,
2803
+ "step": 1990
2804
+ },
2805
+ {
2806
+ "epoch": 0.08902871678157842,
2807
+ "grad_norm": 1.1955480575561523,
2808
+ "learning_rate": 5.34136546184739e-06,
2809
+ "loss": 1.9616188049316405,
2810
+ "step": 1995
2811
+ },
2812
+ {
2813
+ "epoch": 0.08925184639757235,
2814
+ "grad_norm": 0.9528502225875854,
2815
+ "learning_rate": 5.3547523427041504e-06,
2816
+ "loss": 2.1342418670654295,
2817
+ "step": 2000
2818
+ },
2819
+ {
2820
+ "epoch": 0.08947497601356628,
2821
+ "grad_norm": 1.1190009117126465,
2822
+ "learning_rate": 5.36813922356091e-06,
2823
+ "loss": 2.1064796447753906,
2824
+ "step": 2005
2825
+ },
2826
+ {
2827
+ "epoch": 0.08969810562956021,
2828
+ "grad_norm": 1.0113600492477417,
2829
+ "learning_rate": 5.381526104417671e-06,
2830
+ "loss": 1.9123802185058594,
2831
+ "step": 2010
2832
+ },
2833
+ {
2834
+ "epoch": 0.08992123524555415,
2835
+ "grad_norm": 0.9729359745979309,
2836
+ "learning_rate": 5.394912985274431e-06,
2837
+ "loss": 1.9950223922729493,
2838
+ "step": 2015
2839
+ },
2840
+ {
2841
+ "epoch": 0.09014436486154807,
2842
+ "grad_norm": 1.035287618637085,
2843
+ "learning_rate": 5.4082998661311915e-06,
2844
+ "loss": 1.968986701965332,
2845
+ "step": 2020
2846
+ },
2847
+ {
2848
+ "epoch": 0.090367494477542,
2849
+ "grad_norm": 0.7954384684562683,
2850
+ "learning_rate": 5.421686746987952e-06,
2851
+ "loss": 2.075417327880859,
2852
+ "step": 2025
2853
+ },
2854
+ {
2855
+ "epoch": 0.09059062409353594,
2856
+ "grad_norm": 1.2406619787216187,
2857
+ "learning_rate": 5.435073627844712e-06,
2858
+ "loss": 2.097009468078613,
2859
+ "step": 2030
2860
+ },
2861
+ {
2862
+ "epoch": 0.09081375370952986,
2863
+ "grad_norm": 0.8746753931045532,
2864
+ "learning_rate": 5.448460508701473e-06,
2865
+ "loss": 2.04738826751709,
2866
+ "step": 2035
2867
+ },
2868
+ {
2869
+ "epoch": 0.0910368833255238,
2870
+ "grad_norm": 1.137431025505066,
2871
+ "learning_rate": 5.461847389558233e-06,
2872
+ "loss": 1.9636581420898438,
2873
+ "step": 2040
2874
+ },
2875
+ {
2876
+ "epoch": 0.09126001294151773,
2877
+ "grad_norm": 1.4577034711837769,
2878
+ "learning_rate": 5.475234270414993e-06,
2879
+ "loss": 1.8939685821533203,
2880
+ "step": 2045
2881
+ },
2882
+ {
2883
+ "epoch": 0.09148314255751167,
2884
+ "grad_norm": 1.3480346202850342,
2885
+ "learning_rate": 5.488621151271754e-06,
2886
+ "loss": 1.9858078002929687,
2887
+ "step": 2050
2888
+ },
2889
+ {
2890
+ "epoch": 0.09170627217350559,
2891
+ "grad_norm": 0.8342795372009277,
2892
+ "learning_rate": 5.502008032128514e-06,
2893
+ "loss": 1.7992578506469727,
2894
+ "step": 2055
2895
+ },
2896
+ {
2897
+ "epoch": 0.09192940178949952,
2898
+ "grad_norm": 1.1130149364471436,
2899
+ "learning_rate": 5.5153949129852746e-06,
2900
+ "loss": 2.0617441177368163,
2901
+ "step": 2060
2902
+ },
2903
+ {
2904
+ "epoch": 0.09215253140549345,
2905
+ "grad_norm": 1.0582903623580933,
2906
+ "learning_rate": 5.528781793842035e-06,
2907
+ "loss": 2.081201934814453,
2908
+ "step": 2065
2909
+ },
2910
+ {
2911
+ "epoch": 0.09237566102148738,
2912
+ "grad_norm": 0.8622370958328247,
2913
+ "learning_rate": 5.542168674698796e-06,
2914
+ "loss": 2.184007263183594,
2915
+ "step": 2070
2916
+ },
2917
+ {
2918
+ "epoch": 0.09259879063748132,
2919
+ "grad_norm": 1.247605800628662,
2920
+ "learning_rate": 5.555555555555556e-06,
2921
+ "loss": 2.2668291091918946,
2922
+ "step": 2075
2923
+ },
2924
+ {
2925
+ "epoch": 0.09282192025347524,
2926
+ "grad_norm": 0.9048439860343933,
2927
+ "learning_rate": 5.568942436412316e-06,
2928
+ "loss": 2.0036815643310546,
2929
+ "step": 2080
2930
+ },
2931
+ {
2932
+ "epoch": 0.09304504986946917,
2933
+ "grad_norm": 1.5477705001831055,
2934
+ "learning_rate": 5.582329317269076e-06,
2935
+ "loss": 2.1254167556762695,
2936
+ "step": 2085
2937
+ },
2938
+ {
2939
+ "epoch": 0.0932681794854631,
2940
+ "grad_norm": 0.898073673248291,
2941
+ "learning_rate": 5.595716198125837e-06,
2942
+ "loss": 2.0748970031738283,
2943
+ "step": 2090
2944
+ },
2945
+ {
2946
+ "epoch": 0.09349130910145703,
2947
+ "grad_norm": 0.976369321346283,
2948
+ "learning_rate": 5.609103078982598e-06,
2949
+ "loss": 1.9387639999389648,
2950
+ "step": 2095
2951
+ },
2952
+ {
2953
+ "epoch": 0.09371443871745097,
2954
+ "grad_norm": 1.229954719543457,
2955
+ "learning_rate": 5.622489959839358e-06,
2956
+ "loss": 1.9117536544799805,
2957
+ "step": 2100
2958
+ },
2959
+ {
2960
+ "epoch": 0.0939375683334449,
2961
+ "grad_norm": 1.0723400115966797,
2962
+ "learning_rate": 5.6358768406961175e-06,
2963
+ "loss": 1.935175323486328,
2964
+ "step": 2105
2965
+ },
2966
+ {
2967
+ "epoch": 0.09416069794943883,
2968
+ "grad_norm": 1.1824326515197754,
2969
+ "learning_rate": 5.649263721552878e-06,
2970
+ "loss": 2.126773452758789,
2971
+ "step": 2110
2972
+ },
2973
+ {
2974
+ "epoch": 0.09438382756543276,
2975
+ "grad_norm": 0.8257484436035156,
2976
+ "learning_rate": 5.662650602409639e-06,
2977
+ "loss": 2.01248836517334,
2978
+ "step": 2115
2979
+ },
2980
+ {
2981
+ "epoch": 0.09460695718142668,
2982
+ "grad_norm": 1.11427903175354,
2983
+ "learning_rate": 5.6760374832663996e-06,
2984
+ "loss": 2.1134754180908204,
2985
+ "step": 2120
2986
+ },
2987
+ {
2988
+ "epoch": 0.09483008679742062,
2989
+ "grad_norm": 1.5741039514541626,
2990
+ "learning_rate": 5.689424364123159e-06,
2991
+ "loss": 2.0824302673339843,
2992
+ "step": 2125
2993
+ },
2994
+ {
2995
+ "epoch": 0.09505321641341455,
2996
+ "grad_norm": 0.9258787631988525,
2997
+ "learning_rate": 5.70281124497992e-06,
2998
+ "loss": 1.9319812774658203,
2999
+ "step": 2130
3000
+ },
3001
+ {
3002
+ "epoch": 0.09527634602940849,
3003
+ "grad_norm": 1.09774911403656,
3004
+ "learning_rate": 5.71619812583668e-06,
3005
+ "loss": 1.9680803298950196,
3006
+ "step": 2135
3007
+ },
3008
+ {
3009
+ "epoch": 0.09549947564540241,
3010
+ "grad_norm": 0.9626898765563965,
3011
+ "learning_rate": 5.729585006693441e-06,
3012
+ "loss": 1.925799560546875,
3013
+ "step": 2140
3014
+ },
3015
+ {
3016
+ "epoch": 0.09572260526139635,
3017
+ "grad_norm": 0.9046334624290466,
3018
+ "learning_rate": 5.742971887550201e-06,
3019
+ "loss": 1.9070989608764648,
3020
+ "step": 2145
3021
+ },
3022
+ {
3023
+ "epoch": 0.09594573487739028,
3024
+ "grad_norm": 1.0075236558914185,
3025
+ "learning_rate": 5.756358768406961e-06,
3026
+ "loss": 1.9298274993896485,
3027
+ "step": 2150
3028
+ },
3029
+ {
3030
+ "epoch": 0.0961688644933842,
3031
+ "grad_norm": 0.8876293301582336,
3032
+ "learning_rate": 5.769745649263722e-06,
3033
+ "loss": 1.995884323120117,
3034
+ "step": 2155
3035
+ },
3036
+ {
3037
+ "epoch": 0.09639199410937814,
3038
+ "grad_norm": 0.9397478699684143,
3039
+ "learning_rate": 5.783132530120482e-06,
3040
+ "loss": 2.1033966064453127,
3041
+ "step": 2160
3042
+ },
3043
+ {
3044
+ "epoch": 0.09661512372537207,
3045
+ "grad_norm": 1.1887969970703125,
3046
+ "learning_rate": 5.7965194109772425e-06,
3047
+ "loss": 2.2321205139160156,
3048
+ "step": 2165
3049
+ },
3050
+ {
3051
+ "epoch": 0.096838253341366,
3052
+ "grad_norm": 0.9165719151496887,
3053
+ "learning_rate": 5.809906291834003e-06,
3054
+ "loss": 1.9332220077514648,
3055
+ "step": 2170
3056
+ },
3057
+ {
3058
+ "epoch": 0.09706138295735993,
3059
+ "grad_norm": 1.0620633363723755,
3060
+ "learning_rate": 5.823293172690763e-06,
3061
+ "loss": 1.942098617553711,
3062
+ "step": 2175
3063
+ },
3064
+ {
3065
+ "epoch": 0.09728451257335387,
3066
+ "grad_norm": 1.1249985694885254,
3067
+ "learning_rate": 5.836680053547524e-06,
3068
+ "loss": 2.004232406616211,
3069
+ "step": 2180
3070
+ },
3071
+ {
3072
+ "epoch": 0.09750764218934779,
3073
+ "grad_norm": 1.009824514389038,
3074
+ "learning_rate": 5.8500669344042836e-06,
3075
+ "loss": 2.060811424255371,
3076
+ "step": 2185
3077
+ },
3078
+ {
3079
+ "epoch": 0.09773077180534172,
3080
+ "grad_norm": 1.1177529096603394,
3081
+ "learning_rate": 5.863453815261044e-06,
3082
+ "loss": 1.9775056838989258,
3083
+ "step": 2190
3084
+ },
3085
+ {
3086
+ "epoch": 0.09795390142133566,
3087
+ "grad_norm": 0.8883247375488281,
3088
+ "learning_rate": 5.876840696117805e-06,
3089
+ "loss": 2.0491350173950194,
3090
+ "step": 2195
3091
+ },
3092
+ {
3093
+ "epoch": 0.09817703103732958,
3094
+ "grad_norm": 1.0781581401824951,
3095
+ "learning_rate": 5.890227576974565e-06,
3096
+ "loss": 2.130450439453125,
3097
+ "step": 2200
3098
+ },
3099
+ {
3100
+ "epoch": 0.09840016065332352,
3101
+ "grad_norm": 1.022125244140625,
3102
+ "learning_rate": 5.9036144578313255e-06,
3103
+ "loss": 1.9054424285888671,
3104
+ "step": 2205
3105
+ },
3106
+ {
3107
+ "epoch": 0.09862329026931745,
3108
+ "grad_norm": 1.5722389221191406,
3109
+ "learning_rate": 5.917001338688085e-06,
3110
+ "loss": 1.9462669372558594,
3111
+ "step": 2210
3112
+ },
3113
+ {
3114
+ "epoch": 0.09884641988531137,
3115
+ "grad_norm": 0.8767709732055664,
3116
+ "learning_rate": 5.930388219544846e-06,
3117
+ "loss": 1.9716548919677734,
3118
+ "step": 2215
3119
+ },
3120
+ {
3121
+ "epoch": 0.09906954950130531,
3122
+ "grad_norm": 0.8978337049484253,
3123
+ "learning_rate": 5.943775100401607e-06,
3124
+ "loss": 1.932182502746582,
3125
+ "step": 2220
3126
+ },
3127
+ {
3128
+ "epoch": 0.09929267911729923,
3129
+ "grad_norm": 1.1649725437164307,
3130
+ "learning_rate": 5.957161981258367e-06,
3131
+ "loss": 1.949960708618164,
3132
+ "step": 2225
3133
+ },
3134
+ {
3135
+ "epoch": 0.09951580873329317,
3136
+ "grad_norm": 0.9707964062690735,
3137
+ "learning_rate": 5.970548862115127e-06,
3138
+ "loss": 1.9415960311889648,
3139
+ "step": 2230
3140
+ },
3141
+ {
3142
+ "epoch": 0.0997389383492871,
3143
+ "grad_norm": 0.8075286149978638,
3144
+ "learning_rate": 5.983935742971887e-06,
3145
+ "loss": 1.9884584426879883,
3146
+ "step": 2235
3147
+ },
3148
+ {
3149
+ "epoch": 0.09996206796528104,
3150
+ "grad_norm": 0.919001579284668,
3151
+ "learning_rate": 5.997322623828648e-06,
3152
+ "loss": 1.9047340393066405,
3153
+ "step": 2240
3154
+ },
3155
+ {
3156
+ "epoch": 0.10018519758127496,
3157
+ "grad_norm": 1.1452890634536743,
3158
+ "learning_rate": 5.999999417648353e-06,
3159
+ "loss": 1.8911170959472656,
3160
+ "step": 2245
3161
+ },
3162
+ {
3163
+ "epoch": 0.10040832719726889,
3164
+ "grad_norm": 0.9075614213943481,
3165
+ "learning_rate": 5.999997051845177e-06,
3166
+ "loss": 1.9537294387817383,
3167
+ "step": 2250
3168
+ },
3169
+ {
3170
+ "epoch": 0.10063145681326283,
3171
+ "grad_norm": 0.9411978721618652,
3172
+ "learning_rate": 5.999992866194927e-06,
3173
+ "loss": 2.0128080368041994,
3174
+ "step": 2255
3175
+ },
3176
+ {
3177
+ "epoch": 0.10085458642925675,
3178
+ "grad_norm": 1.7404388189315796,
3179
+ "learning_rate": 5.999986860700142e-06,
3180
+ "loss": 1.9837495803833007,
3181
+ "step": 2260
3182
+ },
3183
+ {
3184
+ "epoch": 0.10107771604525069,
3185
+ "grad_norm": 0.990151047706604,
3186
+ "learning_rate": 5.999979035364464e-06,
3187
+ "loss": 1.9975740432739257,
3188
+ "step": 2265
3189
+ },
3190
+ {
3191
+ "epoch": 0.10130084566124462,
3192
+ "grad_norm": 1.042856216430664,
3193
+ "learning_rate": 5.9999693901926435e-06,
3194
+ "loss": 2.016056251525879,
3195
+ "step": 2270
3196
+ },
3197
+ {
3198
+ "epoch": 0.10152397527723855,
3199
+ "grad_norm": 0.8757387399673462,
3200
+ "learning_rate": 5.999957925190527e-06,
3201
+ "loss": 1.8990396499633788,
3202
+ "step": 2275
3203
+ },
3204
+ {
3205
+ "epoch": 0.10174710489323248,
3206
+ "grad_norm": 1.4858282804489136,
3207
+ "learning_rate": 5.999944640365074e-06,
3208
+ "loss": 2.0470705032348633,
3209
+ "step": 2280
3210
+ },
3211
+ {
3212
+ "epoch": 0.1019702345092264,
3213
+ "grad_norm": 1.0206114053726196,
3214
+ "learning_rate": 5.999929535724339e-06,
3215
+ "loss": 2.063922882080078,
3216
+ "step": 2285
3217
+ },
3218
+ {
3219
+ "epoch": 0.10219336412522034,
3220
+ "grad_norm": 1.2412887811660767,
3221
+ "learning_rate": 5.999912611277487e-06,
3222
+ "loss": 1.9137586593627929,
3223
+ "step": 2290
3224
+ },
3225
+ {
3226
+ "epoch": 0.10241649374121427,
3227
+ "grad_norm": 1.0442379713058472,
3228
+ "learning_rate": 5.999893867034784e-06,
3229
+ "loss": 1.9131647109985352,
3230
+ "step": 2295
3231
+ },
3232
+ {
3233
+ "epoch": 0.1026396233572082,
3234
+ "grad_norm": 1.0354846715927124,
3235
+ "learning_rate": 5.9998733030076015e-06,
3236
+ "loss": 1.943178367614746,
3237
+ "step": 2300
3238
+ },
3239
+ {
3240
+ "epoch": 0.10286275297320213,
3241
+ "grad_norm": 0.9171436429023743,
3242
+ "learning_rate": 5.999850919208413e-06,
3243
+ "loss": 1.9243003845214843,
3244
+ "step": 2305
3245
+ },
3246
+ {
3247
+ "epoch": 0.10308588258919607,
3248
+ "grad_norm": 1.3587474822998047,
3249
+ "learning_rate": 5.999826715650797e-06,
3250
+ "loss": 1.7801437377929688,
3251
+ "step": 2310
3252
+ },
3253
+ {
3254
+ "epoch": 0.10330901220519,
3255
+ "grad_norm": 1.6726102828979492,
3256
+ "learning_rate": 5.999800692349435e-06,
3257
+ "loss": 2.0044769287109374,
3258
+ "step": 2315
3259
+ },
3260
+ {
3261
+ "epoch": 0.10353214182118392,
3262
+ "grad_norm": 0.909557580947876,
3263
+ "learning_rate": 5.999772849320115e-06,
3264
+ "loss": 1.8790843963623047,
3265
+ "step": 2320
3266
+ },
3267
+ {
3268
+ "epoch": 0.10375527143717786,
3269
+ "grad_norm": 1.1021777391433716,
3270
+ "learning_rate": 5.9997431865797255e-06,
3271
+ "loss": 1.9645397186279296,
3272
+ "step": 2325
3273
+ },
3274
+ {
3275
+ "epoch": 0.10397840105317178,
3276
+ "grad_norm": 1.0416226387023926,
3277
+ "learning_rate": 5.999711704146261e-06,
3278
+ "loss": 1.9316511154174805,
3279
+ "step": 2330
3280
+ },
3281
+ {
3282
+ "epoch": 0.10420153066916572,
3283
+ "grad_norm": 1.014394998550415,
3284
+ "learning_rate": 5.99967840203882e-06,
3285
+ "loss": 1.9602357864379882,
3286
+ "step": 2335
3287
+ },
3288
+ {
3289
+ "epoch": 0.10442466028515965,
3290
+ "grad_norm": 1.0431203842163086,
3291
+ "learning_rate": 5.999643280277603e-06,
3292
+ "loss": 1.9662353515625,
3293
+ "step": 2340
3294
+ },
3295
+ {
3296
+ "epoch": 0.10464778990115357,
3297
+ "grad_norm": 0.9976010322570801,
3298
+ "learning_rate": 5.999606338883915e-06,
3299
+ "loss": 1.8156768798828125,
3300
+ "step": 2345
3301
+ },
3302
+ {
3303
+ "epoch": 0.10487091951714751,
3304
+ "grad_norm": 1.3542364835739136,
3305
+ "learning_rate": 5.999567577880167e-06,
3306
+ "loss": 1.9991931915283203,
3307
+ "step": 2350
3308
+ },
3309
+ {
3310
+ "epoch": 0.10509404913314144,
3311
+ "grad_norm": 0.9118264317512512,
3312
+ "learning_rate": 5.99952699728987e-06,
3313
+ "loss": 1.6772186279296875,
3314
+ "step": 2355
3315
+ },
3316
+ {
3317
+ "epoch": 0.10531717874913538,
3318
+ "grad_norm": 1.431563138961792,
3319
+ "learning_rate": 5.9994845971376435e-06,
3320
+ "loss": 2.0468074798583986,
3321
+ "step": 2360
3322
+ },
3323
+ {
3324
+ "epoch": 0.1055403083651293,
3325
+ "grad_norm": 0.8818404078483582,
3326
+ "learning_rate": 5.999440377449205e-06,
3327
+ "loss": 1.9281387329101562,
3328
+ "step": 2365
3329
+ },
3330
+ {
3331
+ "epoch": 0.10576343798112324,
3332
+ "grad_norm": 0.905672013759613,
3333
+ "learning_rate": 5.999394338251381e-06,
3334
+ "loss": 2.0013042449951173,
3335
+ "step": 2370
3336
+ },
3337
+ {
3338
+ "epoch": 0.10598656759711717,
3339
+ "grad_norm": 1.0039812326431274,
3340
+ "learning_rate": 5.999346479572099e-06,
3341
+ "loss": 2.0309635162353517,
3342
+ "step": 2375
3343
+ },
3344
+ {
3345
+ "epoch": 0.10620969721311109,
3346
+ "grad_norm": 1.1300805807113647,
3347
+ "learning_rate": 5.999296801440391e-06,
3348
+ "loss": 2.013144111633301,
3349
+ "step": 2380
3350
+ },
3351
+ {
3352
+ "epoch": 0.10643282682910503,
3353
+ "grad_norm": 1.052232027053833,
3354
+ "learning_rate": 5.999245303886392e-06,
3355
+ "loss": 2.089672660827637,
3356
+ "step": 2385
3357
+ },
3358
+ {
3359
+ "epoch": 0.10665595644509895,
3360
+ "grad_norm": 0.8320952653884888,
3361
+ "learning_rate": 5.999191986941342e-06,
3362
+ "loss": 2.0204065322875975,
3363
+ "step": 2390
3364
+ },
3365
+ {
3366
+ "epoch": 0.10687908606109289,
3367
+ "grad_norm": 0.8853346705436707,
3368
+ "learning_rate": 5.999136850637584e-06,
3369
+ "loss": 1.9383895874023438,
3370
+ "step": 2395
3371
+ },
3372
+ {
3373
+ "epoch": 0.10710221567708682,
3374
+ "grad_norm": 1.3795983791351318,
3375
+ "learning_rate": 5.999079895008564e-06,
3376
+ "loss": 1.808138084411621,
3377
+ "step": 2400
3378
+ }
3379
+ ],
3380
+ "logging_steps": 5,
3381
+ "max_steps": 22409,
3382
+ "num_input_tokens_seen": 0,
3383
+ "num_train_epochs": 1,
3384
+ "save_steps": 200,
3385
+ "stateful_callbacks": {
3386
+ "TrainerControl": {
3387
+ "args": {
3388
+ "should_epoch_stop": false,
3389
+ "should_evaluate": false,
3390
+ "should_log": false,
3391
+ "should_save": true,
3392
+ "should_training_stop": false
3393
+ },
3394
+ "attributes": {}
3395
+ }
3396
+ },
3397
+ "total_flos": 3.2148094686056153e+19,
3398
+ "train_batch_size": 4,
3399
+ "trial_name": null,
3400
+ "trial_params": null
3401
+ }
checkpoint-2400/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3809f9af695594d2a92a4b73b98dd0153f0059de103c19d7a1878e59202bb140
3
+ size 6865
checkpoint-2600/README.md ADDED
@@ -0,0 +1,206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B
6
+ - lora
7
+ - transformers
8
+ ---
9
+
10
+ # Model Card for Model ID
11
+
12
+ <!-- Provide a quick summary of what the model is/does. -->
13
+
14
+
15
+
16
+ ## Model Details
17
+
18
+ ### Model Description
19
+
20
+ <!-- Provide a longer summary of what this model is. -->
21
+
22
+
23
+
24
+ - **Developed by:** [More Information Needed]
25
+ - **Funded by [optional]:** [More Information Needed]
26
+ - **Shared by [optional]:** [More Information Needed]
27
+ - **Model type:** [More Information Needed]
28
+ - **Language(s) (NLP):** [More Information Needed]
29
+ - **License:** [More Information Needed]
30
+ - **Finetuned from model [optional]:** [More Information Needed]
31
+
32
+ ### Model Sources [optional]
33
+
34
+ <!-- Provide the basic links for the model. -->
35
+
36
+ - **Repository:** [More Information Needed]
37
+ - **Paper [optional]:** [More Information Needed]
38
+ - **Demo [optional]:** [More Information Needed]
39
+
40
+ ## Uses
41
+
42
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
43
+
44
+ ### Direct Use
45
+
46
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
47
+
48
+ [More Information Needed]
49
+
50
+ ### Downstream Use [optional]
51
+
52
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
53
+
54
+ [More Information Needed]
55
+
56
+ ### Out-of-Scope Use
57
+
58
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
59
+
60
+ [More Information Needed]
61
+
62
+ ## Bias, Risks, and Limitations
63
+
64
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
65
+
66
+ [More Information Needed]
67
+
68
+ ### Recommendations
69
+
70
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
71
+
72
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
73
+
74
+ ## How to Get Started with the Model
75
+
76
+ Use the code below to get started with the model.
77
+
78
+ [More Information Needed]
79
+
80
+ ## Training Details
81
+
82
+ ### Training Data
83
+
84
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
85
+
86
+ [More Information Needed]
87
+
88
+ ### Training Procedure
89
+
90
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
91
+
92
+ #### Preprocessing [optional]
93
+
94
+ [More Information Needed]
95
+
96
+
97
+ #### Training Hyperparameters
98
+
99
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
100
+
101
+ #### Speeds, Sizes, Times [optional]
102
+
103
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
104
+
105
+ [More Information Needed]
106
+
107
+ ## Evaluation
108
+
109
+ <!-- This section describes the evaluation protocols and provides the results. -->
110
+
111
+ ### Testing Data, Factors & Metrics
112
+
113
+ #### Testing Data
114
+
115
+ <!-- This should link to a Dataset Card if possible. -->
116
+
117
+ [More Information Needed]
118
+
119
+ #### Factors
120
+
121
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
122
+
123
+ [More Information Needed]
124
+
125
+ #### Metrics
126
+
127
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
128
+
129
+ [More Information Needed]
130
+
131
+ ### Results
132
+
133
+ [More Information Needed]
134
+
135
+ #### Summary
136
+
137
+
138
+
139
+ ## Model Examination [optional]
140
+
141
+ <!-- Relevant interpretability work for the model goes here -->
142
+
143
+ [More Information Needed]
144
+
145
+ ## Environmental Impact
146
+
147
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
148
+
149
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
150
+
151
+ - **Hardware Type:** [More Information Needed]
152
+ - **Hours used:** [More Information Needed]
153
+ - **Cloud Provider:** [More Information Needed]
154
+ - **Compute Region:** [More Information Needed]
155
+ - **Carbon Emitted:** [More Information Needed]
156
+
157
+ ## Technical Specifications [optional]
158
+
159
+ ### Model Architecture and Objective
160
+
161
+ [More Information Needed]
162
+
163
+ ### Compute Infrastructure
164
+
165
+ [More Information Needed]
166
+
167
+ #### Hardware
168
+
169
+ [More Information Needed]
170
+
171
+ #### Software
172
+
173
+ [More Information Needed]
174
+
175
+ ## Citation [optional]
176
+
177
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
178
+
179
+ **BibTeX:**
180
+
181
+ [More Information Needed]
182
+
183
+ **APA:**
184
+
185
+ [More Information Needed]
186
+
187
+ ## Glossary [optional]
188
+
189
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
190
+
191
+ [More Information Needed]
192
+
193
+ ## More Information [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Authors [optional]
198
+
199
+ [More Information Needed]
200
+
201
+ ## Model Card Contact
202
+
203
+ [More Information Needed]
204
+ ### Framework versions
205
+
206
+ - PEFT 0.18.0
checkpoint-2600/adapter_config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "Qwen3ForCausalLM",
7
+ "parent_library": "transformers.models.qwen3.modeling_qwen3"
8
+ },
9
+ "base_model_name_or_path": "/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B",
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": true,
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 32,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.05,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": [],
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.0",
30
+ "qalora_group_size": 16,
31
+ "r": 8,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "k_proj",
36
+ "o_proj",
37
+ "down_proj",
38
+ "v_proj",
39
+ "gate_proj",
40
+ "up_proj",
41
+ "q_proj"
42
+ ],
43
+ "target_parameters": null,
44
+ "task_type": null,
45
+ "trainable_token_indices": null,
46
+ "use_dora": false,
47
+ "use_qalora": false,
48
+ "use_rslora": false
49
+ }
checkpoint-2600/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06f8d90fdd3283d802fcaef8d874a19603b6741e9f3dd4dce549ed19e00469c3
3
+ size 66126768
checkpoint-2600/additional_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lora_dtype": null, "lorap_lr_ratio": null, "lorap_emb_lr": 1e-06}
checkpoint-2600/args.json ADDED
@@ -0,0 +1,353 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438",
3
+ "overwrite_output_dir": false,
4
+ "do_train": false,
5
+ "do_eval": false,
6
+ "do_predict": false,
7
+ "eval_strategy": "no",
8
+ "prediction_loss_only": false,
9
+ "per_device_train_batch_size": 4,
10
+ "per_device_eval_batch_size": 1,
11
+ "per_gpu_train_batch_size": null,
12
+ "per_gpu_eval_batch_size": null,
13
+ "gradient_accumulation_steps": 2,
14
+ "eval_accumulation_steps": null,
15
+ "eval_delay": 0,
16
+ "torch_empty_cache_steps": null,
17
+ "learning_rate": 6e-06,
18
+ "weight_decay": 0.1,
19
+ "adam_beta1": 0.9,
20
+ "adam_beta2": 0.95,
21
+ "adam_epsilon": 1e-08,
22
+ "max_grad_norm": 1.0,
23
+ "num_train_epochs": 1.0,
24
+ "max_steps": -1,
25
+ "lr_scheduler_type": "cosine",
26
+ "lr_scheduler_kwargs": null,
27
+ "warmup_ratio": 0.1,
28
+ "warmup_steps": 0,
29
+ "log_level": "passive",
30
+ "log_level_replica": "warning",
31
+ "log_on_each_node": true,
32
+ "logging_dir": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438/runs",
33
+ "logging_strategy": "steps",
34
+ "logging_first_step": true,
35
+ "logging_steps": 5,
36
+ "logging_nan_inf_filter": true,
37
+ "save_strategy": "steps",
38
+ "save_steps": 200.0,
39
+ "save_total_limit": 2,
40
+ "save_safetensors": true,
41
+ "save_on_each_node": false,
42
+ "save_only_model": false,
43
+ "restore_callback_states_from_checkpoint": false,
44
+ "no_cuda": false,
45
+ "use_cpu": false,
46
+ "use_mps_device": false,
47
+ "seed": 42,
48
+ "data_seed": 42,
49
+ "jit_mode_eval": false,
50
+ "bf16": true,
51
+ "fp16": false,
52
+ "fp16_opt_level": "O1",
53
+ "half_precision_backend": "auto",
54
+ "bf16_full_eval": false,
55
+ "fp16_full_eval": false,
56
+ "tf32": null,
57
+ "local_rank": 0,
58
+ "ddp_backend": null,
59
+ "tpu_num_cores": null,
60
+ "tpu_metrics_debug": false,
61
+ "debug": null,
62
+ "dataloader_drop_last": true,
63
+ "eval_steps": 200.0,
64
+ "dataloader_num_workers": 64,
65
+ "dataloader_prefetch_factor": null,
66
+ "past_index": -1,
67
+ "run_name": "/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438",
68
+ "disable_tqdm": null,
69
+ "remove_unused_columns": true,
70
+ "label_names": null,
71
+ "load_best_model_at_end": false,
72
+ "metric_for_best_model": "loss",
73
+ "greater_is_better": false,
74
+ "ignore_data_skip": false,
75
+ "fsdp": [],
76
+ "fsdp_min_num_params": 0,
77
+ "fsdp_config": null,
78
+ "fsdp_transformer_layer_cls_to_wrap": null,
79
+ "accelerator_config": {
80
+ "dispatch_batches": false
81
+ },
82
+ "parallelism_config": null,
83
+ "deepspeed": null,
84
+ "label_smoothing_factor": 0.0,
85
+ "optim": "adamw_torch_fused",
86
+ "optim_args": null,
87
+ "adafactor": false,
88
+ "group_by_length": false,
89
+ "length_column_name": "length",
90
+ "report_to": [
91
+ "tensorboard"
92
+ ],
93
+ "project": "huggingface",
94
+ "trackio_space_id": "trackio",
95
+ "ddp_find_unused_parameters": null,
96
+ "ddp_bucket_cap_mb": null,
97
+ "ddp_broadcast_buffers": null,
98
+ "dataloader_pin_memory": true,
99
+ "dataloader_persistent_workers": false,
100
+ "skip_memory_metrics": true,
101
+ "use_legacy_prediction_loop": false,
102
+ "push_to_hub": false,
103
+ "resume_from_checkpoint": null,
104
+ "hub_model_id": null,
105
+ "hub_strategy": "every_save",
106
+ "hub_token": null,
107
+ "hub_private_repo": null,
108
+ "hub_always_push": false,
109
+ "hub_revision": null,
110
+ "gradient_checkpointing": true,
111
+ "gradient_checkpointing_kwargs": null,
112
+ "include_inputs_for_metrics": false,
113
+ "include_for_metrics": [],
114
+ "eval_do_concat_batches": true,
115
+ "fp16_backend": "auto",
116
+ "push_to_hub_model_id": null,
117
+ "push_to_hub_organization": null,
118
+ "push_to_hub_token": null,
119
+ "mp_parameters": "",
120
+ "auto_find_batch_size": false,
121
+ "full_determinism": false,
122
+ "torchdynamo": null,
123
+ "ray_scope": "last",
124
+ "ddp_timeout": 18000000,
125
+ "torch_compile": false,
126
+ "torch_compile_backend": null,
127
+ "torch_compile_mode": null,
128
+ "include_tokens_per_second": false,
129
+ "include_num_input_tokens_seen": false,
130
+ "neftune_noise_alpha": null,
131
+ "optim_target_modules": null,
132
+ "batch_eval_metrics": false,
133
+ "eval_on_start": false,
134
+ "use_liger_kernel": true,
135
+ "liger_kernel_config": null,
136
+ "eval_use_gather_object": false,
137
+ "average_tokens_across_devices": true,
138
+ "sortish_sampler": false,
139
+ "predict_with_generate": false,
140
+ "generation_max_length": null,
141
+ "generation_num_beams": null,
142
+ "generation_config": null,
143
+ "tuner_backend": "peft",
144
+ "vit_gradient_checkpointing": null,
145
+ "router_aux_loss_coef": 0.0,
146
+ "enable_dft_loss": false,
147
+ "enable_channel_loss": false,
148
+ "check_model": true,
149
+ "acc_strategy": "token",
150
+ "train_dataloader_shuffle": true,
151
+ "max_epochs": null,
152
+ "aligner_lr": null,
153
+ "vit_lr": null,
154
+ "use_logits_to_keep": null,
155
+ "ds3_gather_for_generation": true,
156
+ "resume_only_model": false,
157
+ "optimizer": null,
158
+ "loss_type": "infonce",
159
+ "metric": null,
160
+ "eval_use_evalscope": false,
161
+ "eval_dataset": [],
162
+ "eval_dataset_args": null,
163
+ "eval_limit": null,
164
+ "eval_generation_config": null,
165
+ "extra_eval_args": null,
166
+ "use_flash_ckpt": false,
167
+ "use_ray": false,
168
+ "ray_exp_name": null,
169
+ "device_groups": null,
170
+ "model": "Qwen/Qwen3-Embedding-4B",
171
+ "model_type": "qwen3_emb",
172
+ "model_revision": null,
173
+ "task_type": "embedding",
174
+ "torch_dtype": "bfloat16",
175
+ "attn_impl": "flash_attn",
176
+ "new_special_tokens": [],
177
+ "num_labels": null,
178
+ "problem_type": null,
179
+ "rope_scaling": null,
180
+ "device_map": null,
181
+ "max_memory": {},
182
+ "max_model_len": null,
183
+ "local_repo_path": null,
184
+ "init_strategy": null,
185
+ "template": "qwen3_emb",
186
+ "system": null,
187
+ "max_length": 1024,
188
+ "truncation_strategy": "right",
189
+ "max_pixels": null,
190
+ "agent_template": null,
191
+ "norm_bbox": null,
192
+ "use_chat_template": true,
193
+ "padding_side": "right",
194
+ "padding_free": false,
195
+ "loss_scale": "default",
196
+ "sequence_parallel_size": 1,
197
+ "template_backend": "swift",
198
+ "response_prefix": null,
199
+ "enable_thinking": null,
200
+ "add_non_thinking_prefix": true,
201
+ "dataset": [
202
+ "../output.jsonl"
203
+ ],
204
+ "val_dataset": [],
205
+ "cached_dataset": [],
206
+ "cached_val_dataset": [],
207
+ "split_dataset_ratio": 0.0,
208
+ "dataset_num_proc": 64,
209
+ "load_from_cache_file": true,
210
+ "dataset_shuffle": true,
211
+ "val_dataset_shuffle": false,
212
+ "streaming": false,
213
+ "interleave_prob": null,
214
+ "stopping_strategy": "first_exhausted",
215
+ "shuffle_buffer_size": 1000,
216
+ "download_mode": "reuse_dataset_if_exists",
217
+ "columns": {},
218
+ "strict": false,
219
+ "model_name": null,
220
+ "model_author": null,
221
+ "custom_dataset_info": [],
222
+ "quant_method": null,
223
+ "quant_bits": null,
224
+ "hqq_axis": null,
225
+ "bnb_4bit_compute_dtype": "bfloat16",
226
+ "bnb_4bit_quant_type": "nf4",
227
+ "bnb_4bit_use_double_quant": true,
228
+ "bnb_4bit_quant_storage": null,
229
+ "max_new_tokens": 64,
230
+ "temperature": 0.0,
231
+ "top_k": null,
232
+ "top_p": null,
233
+ "repetition_penalty": null,
234
+ "num_beams": 1,
235
+ "stream": false,
236
+ "stop_words": [],
237
+ "logprobs": false,
238
+ "top_logprobs": null,
239
+ "structured_outputs_regex": null,
240
+ "ckpt_dir": null,
241
+ "lora_modules": [],
242
+ "train_type": "lora",
243
+ "adapters": [],
244
+ "external_plugins": [],
245
+ "model_kwargs": {},
246
+ "load_args": false,
247
+ "load_data_args": false,
248
+ "packing": false,
249
+ "packing_length": null,
250
+ "packing_num_proc": 1,
251
+ "lazy_tokenize": false,
252
+ "custom_register_path": [],
253
+ "use_hf": false,
254
+ "ignore_args_error": false,
255
+ "use_swift_lora": false,
256
+ "freeze_parameters": [],
257
+ "freeze_parameters_regex": null,
258
+ "freeze_parameters_ratio": 0.0,
259
+ "trainable_parameters": [],
260
+ "trainable_parameters_regex": null,
261
+ "freeze_llm": false,
262
+ "freeze_vit": true,
263
+ "freeze_aligner": true,
264
+ "target_modules": [
265
+ "all-linear"
266
+ ],
267
+ "target_regex": null,
268
+ "target_parameters": null,
269
+ "modules_to_save": [],
270
+ "lora_rank": 8,
271
+ "lora_alpha": 32,
272
+ "lora_dropout": 0.05,
273
+ "lora_bias": "none",
274
+ "lora_dtype": null,
275
+ "lorap_lr_ratio": null,
276
+ "use_rslora": false,
277
+ "use_dora": false,
278
+ "lora_ga_batch_size": 2,
279
+ "lora_ga_iters": 2,
280
+ "lora_ga_max_length": 1024,
281
+ "lora_ga_direction": "ArB2r",
282
+ "lora_ga_scale": "stable",
283
+ "lora_ga_stable_gamma": 16,
284
+ "init_weights": true,
285
+ "fourier_n_frequency": 2000,
286
+ "fourier_scaling": 300.0,
287
+ "boft_block_size": 4,
288
+ "boft_block_num": 0,
289
+ "boft_n_butterfly_factor": 1,
290
+ "boft_dropout": 0.0,
291
+ "vera_rank": 256,
292
+ "vera_projection_prng_key": 0,
293
+ "vera_dropout": 0.0,
294
+ "vera_d_initial": 0.1,
295
+ "adapter_act": "gelu",
296
+ "adapter_length": 128,
297
+ "use_galore": false,
298
+ "galore_target_modules": null,
299
+ "galore_rank": 128,
300
+ "galore_update_proj_gap": 50,
301
+ "galore_scale": 1.0,
302
+ "galore_proj_type": "std",
303
+ "galore_optim_per_parameter": false,
304
+ "galore_with_embedding": false,
305
+ "galore_quantization": false,
306
+ "galore_proj_quant": false,
307
+ "galore_proj_bits": 4,
308
+ "galore_proj_group_size": 256,
309
+ "galore_cos_threshold": 0.4,
310
+ "galore_gamma_proj": 2,
311
+ "galore_queue_size": 5,
312
+ "adalora_target_r": 8,
313
+ "adalora_init_r": 12,
314
+ "adalora_tinit": 0,
315
+ "adalora_tfinal": 0,
316
+ "adalora_deltaT": 1,
317
+ "adalora_beta1": 0.85,
318
+ "adalora_beta2": 0.85,
319
+ "adalora_orth_reg_weight": 0.5,
320
+ "llamapro_num_new_blocks": 4,
321
+ "llamapro_num_groups": null,
322
+ "lisa_activated_layers": 0,
323
+ "lisa_step_interval": 20,
324
+ "reft_layer_key": null,
325
+ "reft_layers": null,
326
+ "reft_rank": 4,
327
+ "reft_intervention_type": "LoreftIntervention",
328
+ "reft_args": null,
329
+ "swanlab_token": null,
330
+ "swanlab_project": "ms-swift",
331
+ "swanlab_workspace": null,
332
+ "swanlab_exp_name": null,
333
+ "swanlab_notification_method": null,
334
+ "swanlab_webhook_url": null,
335
+ "swanlab_secret": null,
336
+ "swanlab_mode": "cloud",
337
+ "add_version": true,
338
+ "create_checkpoint_symlink": false,
339
+ "zero_hpz_partition_size": null,
340
+ "deepspeed_autotp_size": null,
341
+ "early_stop_interval": null,
342
+ "rank": 0,
343
+ "global_world_size": 8,
344
+ "local_world_size": 8,
345
+ "model_suffix": "Qwen3-Embedding-4B",
346
+ "model_info": "ModelInfo(model_type='qwen3_emb', model_dir='/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B', torch_dtype=torch.bfloat16, max_model_len=40960, quant_method=None, quant_bits=None, rope_scaling=None, is_moe_model=False, is_multimodal=False, config=None, task_type='embedding', num_labels=None)",
347
+ "model_meta": "ModelMeta(model_type='qwen3_emb', model_groups=[ModelGroup(models=[Model(ms_model_id='Qwen/Qwen3-Embedding-0.6B', hf_model_id='Qwen/Qwen3-Embedding-0.6B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-Embedding-4B', hf_model_id='Qwen/Qwen3-Embedding-4B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-Embedding-8B', hf_model_id='Qwen/Qwen3-Embedding-8B', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[])], template='qwen3_emb', get_function=<function get_model_tokenizer_with_flash_attn at 0x7f415201c5e0>, model_arch=None, architectures=['Qwen3ForCausalLM'], additional_saved_files=['config_sentence_transformers.json', '1_Pooling', 'modules.json'], torch_dtype=None, is_multimodal=False, is_reward=False, is_reranker=False, task_type=None, ignore_patterns=None, requires=[], tags=[])",
348
+ "model_dir": "/home/legal/.cache/modelscope/hub/models/Qwen/Qwen3-Embedding-4B",
349
+ "_val_dataset_exists": [],
350
+ "hub": "<class 'swift.hub.hub.MSHub'>",
351
+ "evaluation_strategy": "steps",
352
+ "training_args": "TrainingArguments(output_dir='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438', overwrite_output_dir=False, do_train=False, do_eval=False, do_predict=False, eval_strategy=<IntervalStrategy.NO: 'no'>, prediction_loss_only=False, per_device_train_batch_size=4, per_device_eval_batch_size=1, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=2, eval_accumulation_steps=None, eval_delay=0, torch_empty_cache_steps=None, learning_rate=6e-06, weight_decay=0.1, adam_beta1=0.9, adam_beta2=0.95, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=1.0, max_steps=-1, lr_scheduler_type=<SchedulerType.COSINE: 'cosine'>, lr_scheduler_kwargs=None, warmup_ratio=0.1, warmup_steps=0, log_level='passive', log_level_replica='warning', log_on_each_node=True, logging_dir='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438/runs', logging_strategy=<IntervalStrategy.STEPS: 'steps'>, logging_first_step=True, logging_steps=5, logging_nan_inf_filter=True, save_strategy=<SaveStrategy.STEPS: 'steps'>, save_steps=200, save_total_limit=2, save_safetensors=True, save_on_each_node=False, save_only_model=False, restore_callback_states_from_checkpoint=False, no_cuda=False, use_cpu=False, use_mps_device=False, seed=42, data_seed=42, jit_mode_eval=False, bf16=True, fp16=False, fp16_opt_level='O1', half_precision_backend='auto', bf16_full_eval=False, fp16_full_eval=False, tf32=None, local_rank=0, ddp_backend=None, tpu_num_cores=None, tpu_metrics_debug=False, debug=[], dataloader_drop_last=True, eval_steps=200.0, dataloader_num_workers=64, dataloader_prefetch_factor=2, past_index=-1, run_name='/mnt-legal2/Qwen3-4B-Embedding-finetuned/v2-20251231-080438', disable_tqdm=False, remove_unused_columns=False, label_names=None, load_best_model_at_end=False, metric_for_best_model='loss', greater_is_better=False, ignore_data_skip=False, fsdp=[], fsdp_min_num_params=0, fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, fsdp_transformer_layer_cls_to_wrap=None, accelerator_config=AcceleratorConfig(split_batches=False, dispatch_batches=False, even_batches=True, use_seedable_sampler=True, non_blocking=False, gradient_accumulation_kwargs=None, use_configured_state=False), parallelism_config=None, deepspeed=None, label_smoothing_factor=0.0, optim=<OptimizerNames.ADAMW_TORCH_FUSED: 'adamw_torch_fused'>, optim_args=None, adafactor=False, group_by_length=False, length_column_name='length', report_to=['tensorboard'], project='huggingface', trackio_space_id='trackio', ddp_find_unused_parameters=None, ddp_bucket_cap_mb=None, ddp_broadcast_buffers=None, dataloader_pin_memory=True, dataloader_persistent_workers=False, skip_memory_metrics=True, use_legacy_prediction_loop=False, push_to_hub=False, resume_from_checkpoint=None, hub_model_id=None, hub_strategy=<HubStrategy.EVERY_SAVE: 'every_save'>, hub_token=None, hub_private_repo=None, hub_always_push=False, hub_revision=None, gradient_checkpointing=True, gradient_checkpointing_kwargs=None, include_inputs_for_metrics=False, include_for_metrics=[], eval_do_concat_batches=True, fp16_backend='auto', push_to_hub_model_id=None, push_to_hub_organization=None, push_to_hub_token=None, mp_parameters='', auto_find_batch_size=False, full_determinism=False, torchdynamo=None, ray_scope='last', ddp_timeout=18000000, torch_compile=False, torch_compile_backend=None, torch_compile_mode=None, include_tokens_per_second=None, include_num_input_tokens_seen=None, neftune_noise_alpha=None, optim_target_modules=None, batch_eval_metrics=False, eval_on_start=False, use_liger_kernel=True, liger_kernel_config=None, eval_use_gather_object=False, average_tokens_across_devices=None, tuner_backend='peft', vit_gradient_checkpointing=True, router_aux_loss_coef=0.0, enable_dft_loss=False, enable_channel_loss=False, check_model=True, acc_strategy='token', train_dataloader_shuffle=True, max_epochs=None, aligner_lr=None, vit_lr=None, use_logits_to_keep=None, ds3_gather_for_generation=True, resume_only_model=False, optimizer=None, loss_type='infonce', metric=None, eval_use_evalscope=False, eval_dataset=[], eval_dataset_args=None, eval_limit=None, eval_generation_config=None, extra_eval_args=None, use_flash_ckpt=False, sft_alpha=0, chord_sft_dataset=[], chord_sft_per_device_train_batch_size=None, chord_enable_phi_function=False, chord_mu_warmup_steps=None, chord_mu_decay_steps=None, chord_mu_peak=None, chord_mu_valley=None, train_type='lora', local_repo_path=None, galore_config=None, task_type='embedding', problem_type=None)"
353
+ }
checkpoint-2600/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6806dab5069550b5a8928660b1c100b497dd311bc7b53113c27ac67493087e81
3
+ size 132552395
checkpoint-2600/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f25c3441390b0700398f2ed804c280fb0b1461e1c99fab2162a80bc1aa47d9c
3
+ size 16389
checkpoint-2600/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d454266d3855e490449132c9eb7eb7eb204c1a8bb40a496ac1473aaa188fff1
3
+ size 16389
checkpoint-2600/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:473c5f5025f746a4588c332a844479b2a21b9553a541a3841725a49bddc33b29
3
+ size 16389
checkpoint-2600/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d21ecc9239fbb2a225aa85a6ad167a2a14491724280cadc7d8645e78246126
3
+ size 16389
checkpoint-2600/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d9b563f3099de705ea59e807f1fcab5a14b25317a80b74ac3da2d983e2ec8da
3
+ size 16389
checkpoint-2600/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:240f8c321c5ee27ed9397f38c8d2d68237c2a423dff21823c2ca72d26f9bfe33
3
+ size 16389
checkpoint-2600/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e170d863522d8196c970ef714083d0aa9694c8b173162c79395d1d5de5fb3a7
3
+ size 16389
checkpoint-2600/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5d5a75857423f1dfbb6b51cedec485c669926dfdb1176ca1561f28a0eb906e8
3
+ size 16389
checkpoint-2600/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:137724e165a61b193b1203431c92949bd815145f76e823f92641fb81fff445bf
3
+ size 1465
checkpoint-2600/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-2600/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3809f9af695594d2a92a4b73b98dd0153f0059de103c19d7a1878e59202bb140
3
+ size 6865
images/train_epoch.png ADDED
images/train_grad_norm.png ADDED
images/train_learning_rate.png ADDED
images/train_loss.png ADDED
logging.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
runs/events.out.tfevents.1767168374.server-217.1205148.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9b92797335881fe97885eb5a679af17711990cd499a99adfd78d18c3ae93e30
3
+ size 119430