mazesmazes commited on
Commit
20b7f8f
·
verified ·
1 Parent(s): 2d7274f

Training in progress, step 1000

Browse files
adapter_config.json CHANGED
@@ -18,7 +18,7 @@
18
  "loftq_config": {},
19
  "lora_alpha": 64,
20
  "lora_bias": false,
21
- "lora_dropout": 0.0,
22
  "megatron_config": null,
23
  "megatron_core": "megatron.core",
24
  "modules_to_save": null,
@@ -29,8 +29,13 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
 
 
32
  "q_proj",
33
- "v_proj"
 
 
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
 
18
  "loftq_config": {},
19
  "lora_alpha": 64,
20
  "lora_bias": false,
21
+ "lora_dropout": 0.1,
22
  "megatron_config": null,
23
  "megatron_core": "megatron.core",
24
  "modules_to_save": null,
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "k_proj",
33
+ "v_proj",
34
+ "up_proj",
35
  "q_proj",
36
+ "gate_proj",
37
+ "down_proj",
38
+ "o_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04b0f13a2c6affd9364bb060c7fefc6371c514605e03ab7edc7f209655507dd3
3
- size 18365064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93b61bcfb96c91772e8e5b41eb41f368dba655dbf54fdde2948457d4f63d42cc
3
+ size 80792456
config.json CHANGED
@@ -283,11 +283,16 @@
283
  "length_penalty": 1.0,
284
  "llm_dim": 1024,
285
  "lora_alpha": 64,
286
- "lora_dropout": 0.0,
287
  "lora_rank": 32,
288
  "lora_target_modules": [
289
  "q_proj",
290
- "v_proj"
 
 
 
 
 
291
  ],
292
  "max_new_tokens": 128,
293
  "min_new_tokens": 0,
@@ -378,7 +383,7 @@
378
  "vocab_size": 151670
379
  },
380
  "text_model_id": "Qwen/Qwen3-0.6B",
381
- "time_mask_length": 50,
382
  "transformers_version": "5.0.0.dev0",
383
  "use_cache": false,
384
  "use_lora": true,
 
283
  "length_penalty": 1.0,
284
  "llm_dim": 1024,
285
  "lora_alpha": 64,
286
+ "lora_dropout": 0.1,
287
  "lora_rank": 32,
288
  "lora_target_modules": [
289
  "q_proj",
290
+ "k_proj",
291
+ "v_proj",
292
+ "o_proj",
293
+ "gate_proj",
294
+ "up_proj",
295
+ "down_proj"
296
  ],
297
  "max_new_tokens": 128,
298
  "min_new_tokens": 0,
 
383
  "vocab_size": 151670
384
  },
385
  "text_model_id": "Qwen/Qwen3-0.6B",
386
+ "time_mask_length": 100,
387
  "transformers_version": "5.0.0.dev0",
388
  "use_cache": false,
389
  "use_lora": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a36186bad47fadc818caba35c6ed996f863f4f17da2a91283af70bf15238f70
3
  size 25172384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72848e0848e214362323629f760134c5450a8ec981d6e63cdde28fe7ab1be4d4
3
  size 25172384
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:330e4c09cde3d257e2265f96e24e2d447234d90af25450241177582c5bef7f8b
3
  size 5201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da126f1151bd54f84aad0da286b4ef3ab3076094f0b22438b6bc8e0ef025a9c6
3
  size 5201