Zaynes commited on
Commit
e718c32
·
verified ·
1 Parent(s): c1da67c

Upload folder using huggingface_hub

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcc99ddb69f6afd56f9b1b95b8e6aec14a5ee954e8ff83c4b2fb885211bebef2
3
  size 988097824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db3ea32b05a90d614dfc081fce01b86b6256ac7d62d799e1d43c410f50ae3e67
3
  size 988097824
training_artifacts/README.md CHANGED
@@ -12,5 +12,5 @@ This directory contains the training configuration and logs for this model.
12
  ## Job Information
13
 
14
  - Job Name: lf_torch_test__interactive
15
- - Timestamp: 2025-10-22 21:26:55 UTC
16
  - Execution Mode: Local
 
12
  ## Job Information
13
 
14
  - Job Name: lf_torch_test__interactive
15
+ - Timestamp: 2025-10-23 00:36:20 UTC
16
  - Execution Mode: Local
training_artifacts/hydra_config.yaml CHANGED
@@ -167,7 +167,7 @@ lora:
167
  dropout: 0.05
168
  target: all
169
  dataset:
170
- name: my_custom_sft
171
  dir: null
172
  info_json: null
173
  template: default
@@ -187,6 +187,8 @@ dataset:
187
  content: content
188
  user: user
189
  assistant: assistant
 
 
190
  output:
191
  experiment_dir: ./experiments
192
  wandb:
@@ -213,4 +215,6 @@ training:
213
  per_device_train_batch_size: 1
214
  gradient_accumulation_steps: 1
215
  gradient_checkpointing: true
 
 
216
  merge: {}
 
167
  dropout: 0.05
168
  target: all
169
  dataset:
170
+ name: my_custom_sft12
171
  dir: null
172
  info_json: null
173
  template: default
 
187
  content: content
188
  user: user
189
  assistant: assistant
190
+ tokenized_path: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/tokenized/my_custom_sft12
191
+ data_shared_file_system: true
192
  output:
193
  experiment_dir: ./experiments
194
  wandb:
 
215
  per_device_train_batch_size: 1
216
  gradient_accumulation_steps: 1
217
  gradient_checkpointing: true
218
+ preprocessing_num_workers: 16
219
+ overwrite_cache: true
220
  merge: {}
training_artifacts/logs/pipeline_cleaned.txt CHANGED
The diff for this file is too large to render. See raw diff
 
training_artifacts/train_config.yaml CHANGED
@@ -11,18 +11,19 @@ overwrite_output_dir: true
11
  per_device_train_batch_size: 1
12
  gradient_accumulation_steps: 1
13
  gradient_checkpointing: true
 
 
14
  model_name_or_path: Qwen/Qwen2.5-0.5B
15
  finetuning_type: lora
16
- dataset: my_custom_sft
17
  dataset_dir: /scratch/zrs2020/LlamaFactoryHelper/LLaMA-Factory/data
18
  template: default
19
  cutoff_len: 8096
20
  val_size: 0.1
 
 
21
  lora_rank: 8
22
  lora_alpha: 16
23
  lora_dropout: 0.05
24
  lora_target: all
25
  output_dir: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/checkpoints
26
- preprocessing_num_workers: 1
27
- disable_tqdm: false
28
- overwrite_cache: true
 
11
  per_device_train_batch_size: 1
12
  gradient_accumulation_steps: 1
13
  gradient_checkpointing: true
14
+ preprocessing_num_workers: 16
15
+ overwrite_cache: true
16
  model_name_or_path: Qwen/Qwen2.5-0.5B
17
  finetuning_type: lora
18
+ dataset: my_custom_sft12
19
  dataset_dir: /scratch/zrs2020/LlamaFactoryHelper/LLaMA-Factory/data
20
  template: default
21
  cutoff_len: 8096
22
  val_size: 0.1
23
+ tokenized_path: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/tokenized/my_custom_sft12
24
+ data_shared_file_system: true
25
  lora_rank: 8
26
  lora_alpha: 16
27
  lora_dropout: 0.05
28
  lora_target: all
29
  output_dir: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/checkpoints