| ? '' | |
| : ? '' | |
| : ? '' | |
| : hydra: | |
| run: | |
| dir: . | |
| output_subdir: null | |
| job: | |
| chdir: false | |
| _target_: null | |
| job: | |
| name: ??? | |
| mode: slurm | |
| work_dir: null | |
| dry_run: false | |
| slurm: | |
| time_limit: ??? | |
| constraint: | |
| - h200 | |
| memory: 200 | |
| cpus_per_task: 16 | |
| partition: null | |
| mail_user: user@example.com | |
| execution: | |
| nodes: null | |
| gpus_per_node: null | |
| num_gpus: null | |
| hostfile: null | |
| secrets_file: null | |
| model: | |
| name_or_path: ??? | |
| finetuning_type: lora | |
| dataset: | |
| name: ??? | |
| dir: null | |
| info_json: null | |
| template: default | |
| cutoff_len: 1024 | |
| val_size: 0.1 | |
| hf_hub_url: null | |
| formatting: alpaca | |
| ranking: false | |
| subset: null | |
| split: train | |
| folder: null | |
| num_samples: null | |
| columns: | |
| prompt: null | |
| query: null | |
| response: null | |
| history: null | |
| messages: null | |
| system: null | |
| tools: null | |
| images: null | |
| videos: null | |
| audios: null | |
| chosen: null | |
| rejected: null | |
| kto_tag: null | |
| tags: | |
| role: null | |
| content: null | |
| user: null | |
| assistant: null | |
| observation: null | |
| function: null | |
| system: null | |
| training: | |
| stage: sft | |
| do_train: true | |
| model_name_or_path: null | |
| finetuning_type: lora | |
| trust_remote_code: true | |
| dataset: null | |
| dataset_dir: null | |
| template: default | |
| cutoff_len: 1024 | |
| val_size: 0.1 | |
| preprocessing_num_workers: 1 | |
| dataset_num_proc: 1 | |
| dataloader_num_workers: 0 | |
| streaming: false | |
| learning_rate: 5.0e-05 | |
| num_train_epochs: 3.0 | |
| per_device_train_batch_size: 1 | |
| per_device_eval_batch_size: 1 | |
| gradient_accumulation_steps: 8 | |
| lr_scheduler_type: cosine | |
| warmup_ratio: 0.1 | |
| warmup_steps: 0 | |
| lora_rank: 8 | |
| lora_alpha: 16 | |
| lora_dropout: 0.05 | |
| lora_target: all | |
| optim: adamw_torch | |
| bf16: true | |
| fp16: false | |
| output_dir: null | |
| save_strategy: epoch | |
| save_steps: 500 | |
| save_total_limit: 3 | |
| save_only_model: false | |
| eval_strategy: steps | |
| eval_steps: 500 | |
| do_eval: true | |
| logging_steps: 10 | |
| plot_loss: true | |
| report_to: none | |
| gradient_checkpointing: true | |
| ddp_timeout: 180000000 | |
| include_num_input_tokens_seen: true | |
| overwrite_output_dir: true | |
| overwrite_cache: false | |
| seed: 42 | |
| lora: | |
| rank: 8 | |
| alpha: 16 | |
| dropout: 0.05 | |
| target: all | |
| output: | |
| experiment_dir: ./experiments | |
| merge: | |
| stage: export | |
| model_name_or_path: null | |
| adapter_name_or_path: null | |
| template: default | |
| export_dir: null | |
| export_size: 2 | |
| export_device: auto | |
| export_legacy_format: false | |
| finetuning_type: lora | |
| wandb: | |
| project: null | |
| run_name: null | |
| entity: null | |
| hf: | |
| repo_id: null | |
| private: false | |
| upload_artifacts: true | |
| cleanup: | |
| checkpoints: false | |
| merged: false | |
| job: | |
| name: lf_torch_test__interactive | |
| mode: local | |
| work_dir: null | |
| dry_run: false | |
| slurm: | |
| time_limit: null | |
| constraint: null | |
| memory: null | |
| partition: null | |
| mail_user: null | |
| execution: | |
| nodes: 2 | |
| gpus_per_node: 2 | |
| num_gpus: null | |
| hostfile: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/hostfile_auto_generated.txt | |
| secrets_file: ./secrets.env | |
| model: | |
| name_or_path: Qwen/Qwen2.5-0.5B | |
| finetuning_type: lora | |
| lora: | |
| rank: 8 | |
| alpha: 16 | |
| dropout: 0.05 | |
| target: all | |
| dataset: | |
| name: my_custom_sft12 | |
| dir: null | |
| info_json: null | |
| template: default | |
| cutoff_len: 8096 | |
| val_size: 0.1 | |
| hf_hub_url: TAUR-dev/D-SFT_C-BASELINE_r1_distillation-sft-data | |
| formatting: sharegpt | |
| ranking: false | |
| subset: null | |
| split: train | |
| folder: null | |
| num_samples: null | |
| columns: | |
| messages: conversations | |
| tags: | |
| role: role | |
| content: content | |
| user: user | |
| assistant: assistant | |
| tokenized_path: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/tokenized/my_custom_sft12 | |
| data_shared_file_system: true | |
| output: | |
| experiment_dir: ./experiments | |
| wandb: | |
| project: null | |
| run_name: interactive_test | |
| entity: null | |
| hf: | |
| repo_id: TAUR-dev/testing_llamafactory_helper_quick_test__interactive | |
| private: false | |
| cleanup: | |
| checkpoints: false | |
| merged: false | |
| training: | |
| stage: sft | |
| do_train: true | |
| max_steps: 100 | |
| do_eval: false | |
| save_strategy: steps | |
| save_steps: 50 | |
| logging_steps: 10 | |
| fp16: true | |
| bf16: false | |
| overwrite_output_dir: true | |
| per_device_train_batch_size: 1 | |
| gradient_accumulation_steps: 1 | |
| gradient_checkpointing: true | |
| preprocessing_num_workers: 16 | |
| overwrite_cache: true | |
| merge: {} | |