| { | |
| "stage": "CPT", | |
| "phase": "all", | |
| "model_id": "zai-org/GLM-4-32B-Base-0414", | |
| "num_epochs": 1.0, | |
| "max_steps": 1000, | |
| "batch_size": 2, | |
| "grad_accum": 2, | |
| "effective_batch_size": 4, | |
| "learning_rate": 1e-06, | |
| "weight_decay": 0.01, | |
| "warmup_ratio": 0.0, | |
| "warmup_steps": 100, | |
| "max_grad_norm": 1.0, | |
| "max_seq_length": 8192, | |
| "seed": 42, | |
| "sample_seed": null, | |
| "max_samples_per_dataset": null, | |
| "priority_datasets": null, | |
| "priority_repeat": 1, | |
| "cpt_datasets": null, | |
| "cache_key": "dc7008456018", | |
| "domain_counts": {}, | |
| "domain_eval_domains": [] | |
| } |