{ "best_metric": 10.29, "best_model_checkpoint": "/workspace/models/nllb_fongbe_finetuned/checkpoint-3000", "epoch": 4.229820232640113, "eval_steps": 500, "global_step": 3000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.14099400775467041, "grad_norm": 5.562941551208496, "learning_rate": 2.0000000000000003e-06, "loss": 6.7192, "step": 100 }, { "epoch": 0.28198801550934083, "grad_norm": 1.1616981029510498, "learning_rate": 4.000000000000001e-06, "loss": 5.1508, "step": 200 }, { "epoch": 0.4229820232640113, "grad_norm": 1.268113136291504, "learning_rate": 6e-06, "loss": 4.6273, "step": 300 }, { "epoch": 0.5639760310186817, "grad_norm": 1.2506167888641357, "learning_rate": 8.000000000000001e-06, "loss": 4.4506, "step": 400 }, { "epoch": 0.7049700387733522, "grad_norm": 1.2293723821640015, "learning_rate": 1e-05, "loss": 4.3037, "step": 500 }, { "epoch": 0.7049700387733522, "eval_bleu": 4.12, "eval_loss": 4.100268840789795, "eval_runtime": 2071.2386, "eval_samples_per_second": 2.435, "eval_steps_per_second": 0.153, "step": 500 }, { "epoch": 0.8459640465280226, "grad_norm": 1.3162401914596558, "learning_rate": 1.2e-05, "loss": 4.165, "step": 600 }, { "epoch": 0.986958054282693, "grad_norm": 1.1826772689819336, "learning_rate": 1.4e-05, "loss": 4.0721, "step": 700 }, { "epoch": 1.1279520620373633, "grad_norm": 1.2221076488494873, "learning_rate": 1.6000000000000003e-05, "loss": 3.9853, "step": 800 }, { "epoch": 1.2689460697920338, "grad_norm": 1.2303624153137207, "learning_rate": 1.8e-05, "loss": 3.9144, "step": 900 }, { "epoch": 1.4099400775467044, "grad_norm": 1.1688417196273804, "learning_rate": 2e-05, "loss": 3.8332, "step": 1000 }, { "epoch": 1.4099400775467044, "eval_bleu": 5.31, "eval_loss": 3.682112693786621, "eval_runtime": 2266.4204, "eval_samples_per_second": 2.225, "eval_steps_per_second": 0.139, "step": 1000 }, { "epoch": 1.5509340853013747, "grad_norm": 1.1778169870376587, "learning_rate": 1.9214145383104128e-05, "loss": 3.7497, "step": 1100 }, { "epoch": 1.691928093056045, "grad_norm": 1.1701103448867798, "learning_rate": 1.842829076620825e-05, "loss": 3.7144, "step": 1200 }, { "epoch": 1.8329221008107155, "grad_norm": 1.170053243637085, "learning_rate": 1.764243614931238e-05, "loss": 3.6683, "step": 1300 }, { "epoch": 1.973916108565386, "grad_norm": 1.2748677730560303, "learning_rate": 1.6856581532416506e-05, "loss": 3.6192, "step": 1400 }, { "epoch": 2.1149101163200563, "grad_norm": 1.180209994316101, "learning_rate": 1.607072691552063e-05, "loss": 3.553, "step": 1500 }, { "epoch": 2.1149101163200563, "eval_bleu": 6.86, "eval_loss": 3.45310640335083, "eval_runtime": 2226.0872, "eval_samples_per_second": 2.265, "eval_steps_per_second": 0.142, "step": 1500 }, { "epoch": 2.2559041240747266, "grad_norm": 1.1808658838272095, "learning_rate": 1.5284872298624755e-05, "loss": 3.4858, "step": 1600 }, { "epoch": 2.3968981318293974, "grad_norm": 1.2947388887405396, "learning_rate": 1.4499017681728882e-05, "loss": 3.4897, "step": 1700 }, { "epoch": 2.5378921395840677, "grad_norm": 1.2786301374435425, "learning_rate": 1.3713163064833006e-05, "loss": 3.4922, "step": 1800 }, { "epoch": 2.678886147338738, "grad_norm": 1.2938331365585327, "learning_rate": 1.2927308447937132e-05, "loss": 3.4331, "step": 1900 }, { "epoch": 2.8198801550934087, "grad_norm": 1.216395616531372, "learning_rate": 1.2141453831041259e-05, "loss": 3.4279, "step": 2000 }, { "epoch": 2.8198801550934087, "eval_bleu": 8.51, "eval_loss": 3.327824831008911, "eval_runtime": 2139.0671, "eval_samples_per_second": 2.358, "eval_steps_per_second": 0.148, "step": 2000 }, { "epoch": 2.960874162848079, "grad_norm": 1.1774195432662964, "learning_rate": 1.1355599214145383e-05, "loss": 3.4074, "step": 2100 }, { "epoch": 3.1018681706027493, "grad_norm": 1.1595913171768188, "learning_rate": 1.056974459724951e-05, "loss": 3.3459, "step": 2200 }, { "epoch": 3.2428621783574196, "grad_norm": 1.1857880353927612, "learning_rate": 9.783889980353636e-06, "loss": 3.3206, "step": 2300 }, { "epoch": 3.3838561861120904, "grad_norm": 1.1783032417297363, "learning_rate": 8.998035363457762e-06, "loss": 3.3266, "step": 2400 }, { "epoch": 3.5248501938667607, "grad_norm": 1.2441322803497314, "learning_rate": 8.212180746561886e-06, "loss": 3.299, "step": 2500 }, { "epoch": 3.5248501938667607, "eval_bleu": 9.63, "eval_loss": 3.2531518936157227, "eval_runtime": 2099.6399, "eval_samples_per_second": 2.402, "eval_steps_per_second": 0.151, "step": 2500 }, { "epoch": 3.665844201621431, "grad_norm": 1.1659629344940186, "learning_rate": 7.4263261296660124e-06, "loss": 3.2992, "step": 2600 }, { "epoch": 3.8068382093761013, "grad_norm": 1.234079360961914, "learning_rate": 6.640471512770138e-06, "loss": 3.3154, "step": 2700 }, { "epoch": 3.947832217130772, "grad_norm": 1.262974500656128, "learning_rate": 5.854616895874263e-06, "loss": 3.315, "step": 2800 }, { "epoch": 4.088826224885443, "grad_norm": 1.254392385482788, "learning_rate": 5.068762278978389e-06, "loss": 3.2626, "step": 2900 }, { "epoch": 4.229820232640113, "grad_norm": 1.2075427770614624, "learning_rate": 4.282907662082515e-06, "loss": 3.2603, "step": 3000 }, { "epoch": 4.229820232640113, "eval_bleu": 10.29, "eval_loss": 3.213658094406128, "eval_runtime": 2092.8319, "eval_samples_per_second": 2.41, "eval_steps_per_second": 0.151, "step": 3000 } ], "logging_steps": 100, "max_steps": 3545, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 9262611430375424.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }