| add_think_tokens: true | |
| chat_version: Qwen/Qwen2.5-1.5B-Instruct | |
| custom_name: lightr1_3_EN_1024 | |
| dataset: Light-R1 | |
| dataset_args: {} | |
| hf_username: mveroe | |
| max_length: 1024 | |
| mixup: | |
| FPR: 0.0 | |
| TPR: 1.0 | |
| generator_accuracy: 1.0 | |
| model_name: Qwen/Qwen2.5-1.5B | |
| neptune_project: ethsri/label-noise | |
| pretrained_model: true | |
| seed: 42 | |
| set_seed: false | |
| shuffle: true | |
| training_args: | |
| bf16: true | |
| ddp_find_unused_parameters: false | |
| gradient_accumulation_steps: 4 | |
| hub_strategy: all_checkpoints | |
| learning_rate: 5.0e-05 | |
| logging_first_step: true | |
| logging_steps: 1 | |
| logging_strategy: steps | |
| lr_scheduler_type: cosine | |
| num_train_epochs: 3 | |
| optim: adafactor | |
| overwrite_output_dir: true | |
| per_device_train_batch_size: 4 | |
| push_to_hub: true | |
| report_to: none | |
| save_strategy: epoch | |
| seed: 42 | |
| warmup_ratio: 0.1 | |
| training_type: !!python/object/apply:src.configs.TrainingTypes | |
| - sft | |
| use_neptune: true | |