Zaynes commited on
Commit
86f2f8f
·
verified ·
1 Parent(s): 767b6e9

Upload folder using huggingface_hub

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:943098d8c73dfce4549b6f6d316bcbde820346c79c1a7fec0b3481003e157c04
3
  size 988097824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cfc35f90ef6d5aac8099351ed95e482789a3930ab9364744ee737be0e9fe7bd
3
  size 988097824
training_artifacts/README.md CHANGED
@@ -12,5 +12,5 @@ This directory contains the training configuration and logs for this model.
12
  ## Job Information
13
 
14
  - Job Name: lf_torch_test__interactive
15
- - Timestamp: 2025-10-22 20:54:32 UTC
16
  - Execution Mode: Local
 
12
  ## Job Information
13
 
14
  - Job Name: lf_torch_test__interactive
15
+ - Timestamp: 2025-10-22 21:18:35 UTC
16
  - Execution Mode: Local
training_artifacts/hydra_config.yaml CHANGED
@@ -167,13 +167,13 @@ lora:
167
  dropout: 0.05
168
  target: all
169
  dataset:
170
- name: my_custom_sft
171
  dir: null
172
  info_json: null
173
  template: default
174
  cutoff_len: 8096
175
  val_size: 0.1
176
- hf_hub_url: TAUR-dev/D-SFT_C-sft_exp_AT_pvv2__fixed-sft-data
177
  formatting: sharegpt
178
  ranking: false
179
  subset: null
 
167
  dropout: 0.05
168
  target: all
169
  dataset:
170
+ name: my_custom_sft9
171
  dir: null
172
  info_json: null
173
  template: default
174
  cutoff_len: 8096
175
  val_size: 0.1
176
+ hf_hub_url: TAUR-dev/D-SFT_C-BASELINE_r1_distillation-sft-data
177
  formatting: sharegpt
178
  ranking: false
179
  subset: null
training_artifacts/logs/pipeline_cleaned.txt CHANGED
The diff for this file is too large to render. See raw diff
 
training_artifacts/train_config.yaml CHANGED
@@ -13,7 +13,7 @@ gradient_accumulation_steps: 1
13
  gradient_checkpointing: true
14
  model_name_or_path: Qwen/Qwen2.5-0.5B
15
  finetuning_type: lora
16
- dataset: my_custom_sft
17
  dataset_dir: /scratch/zrs2020/LlamaFactoryHelper/LLaMA-Factory/data
18
  template: default
19
  cutoff_len: 8096
@@ -23,3 +23,6 @@ lora_alpha: 16
23
  lora_dropout: 0.05
24
  lora_target: all
25
  output_dir: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/checkpoints
 
 
 
 
13
  gradient_checkpointing: true
14
  model_name_or_path: Qwen/Qwen2.5-0.5B
15
  finetuning_type: lora
16
+ dataset: my_custom_sft9
17
  dataset_dir: /scratch/zrs2020/LlamaFactoryHelper/LLaMA-Factory/data
18
  template: default
19
  cutoff_len: 8096
 
23
  lora_dropout: 0.05
24
  lora_target: all
25
  output_dir: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/checkpoints
26
+ preprocessing_num_workers: 1
27
+ disable_tqdm: false
28
+ overwrite_cache: true