| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9086702284295991, | |
| "eval_steps": 500, | |
| "global_step": 2700, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00033654452904799967, | |
| "grad_norm": 4.331892967224121, | |
| "learning_rate": 0.0, | |
| "loss": 5.5158, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0006730890580959993, | |
| "grad_norm": 4.539519786834717, | |
| "learning_rate": 6.711409395973154e-07, | |
| "loss": 5.5718, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0013461781161919987, | |
| "grad_norm": 4.208465576171875, | |
| "learning_rate": 2.013422818791946e-06, | |
| "loss": 5.4905, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.002019267174287998, | |
| "grad_norm": 3.891338586807251, | |
| "learning_rate": 3.3557046979865773e-06, | |
| "loss": 5.4511, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.0026923562323839974, | |
| "grad_norm": 2.979590654373169, | |
| "learning_rate": 4.697986577181209e-06, | |
| "loss": 5.3311, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0033654452904799967, | |
| "grad_norm": 3.1792373657226562, | |
| "learning_rate": 6.04026845637584e-06, | |
| "loss": 5.3782, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.004038534348575996, | |
| "grad_norm": 3.9909653663635254, | |
| "learning_rate": 7.382550335570471e-06, | |
| "loss": 5.2583, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.004711623406671995, | |
| "grad_norm": 3.31044340133667, | |
| "learning_rate": 8.724832214765101e-06, | |
| "loss": 5.3198, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.005384712464767995, | |
| "grad_norm": 3.3201687335968018, | |
| "learning_rate": 1.006711409395973e-05, | |
| "loss": 5.1138, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.006057801522863994, | |
| "grad_norm": 2.8034451007843018, | |
| "learning_rate": 1.1409395973154363e-05, | |
| "loss": 5.0022, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.0067308905809599934, | |
| "grad_norm": 2.044506072998047, | |
| "learning_rate": 1.2751677852348994e-05, | |
| "loss": 4.8879, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.007403979639055993, | |
| "grad_norm": 1.6194826364517212, | |
| "learning_rate": 1.4093959731543624e-05, | |
| "loss": 4.7436, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.008077068697151992, | |
| "grad_norm": 1.4383995532989502, | |
| "learning_rate": 1.5436241610738255e-05, | |
| "loss": 4.5798, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.00875015775524799, | |
| "grad_norm": 1.174633264541626, | |
| "learning_rate": 1.6778523489932888e-05, | |
| "loss": 4.4188, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.00942324681334399, | |
| "grad_norm": 1.1277130842208862, | |
| "learning_rate": 1.8120805369127517e-05, | |
| "loss": 4.4374, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.01009633587143999, | |
| "grad_norm": 1.0426617860794067, | |
| "learning_rate": 1.946308724832215e-05, | |
| "loss": 4.2195, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01076942492953599, | |
| "grad_norm": 0.9180749654769897, | |
| "learning_rate": 2.080536912751678e-05, | |
| "loss": 4.1755, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.011442513987631988, | |
| "grad_norm": 1.0104376077651978, | |
| "learning_rate": 2.2147651006711412e-05, | |
| "loss": 4.0892, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.012115603045727988, | |
| "grad_norm": 0.9427777528762817, | |
| "learning_rate": 2.348993288590604e-05, | |
| "loss": 4.0498, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.012788692103823987, | |
| "grad_norm": 1.0120079517364502, | |
| "learning_rate": 2.4832214765100674e-05, | |
| "loss": 3.9471, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.013461781161919987, | |
| "grad_norm": 1.0842680931091309, | |
| "learning_rate": 2.6174496644295304e-05, | |
| "loss": 3.8882, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.014134870220015985, | |
| "grad_norm": 1.3515102863311768, | |
| "learning_rate": 2.7516778523489933e-05, | |
| "loss": 3.8274, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.014807959278111986, | |
| "grad_norm": 1.1003209352493286, | |
| "learning_rate": 2.885906040268457e-05, | |
| "loss": 3.7735, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.015481048336207984, | |
| "grad_norm": 0.9302487373352051, | |
| "learning_rate": 3.02013422818792e-05, | |
| "loss": 3.731, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.016154137394303984, | |
| "grad_norm": 1.0311543941497803, | |
| "learning_rate": 3.1543624161073825e-05, | |
| "loss": 3.6608, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.016827226452399983, | |
| "grad_norm": 1.0927435159683228, | |
| "learning_rate": 3.288590604026846e-05, | |
| "loss": 3.5962, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.01750031551049598, | |
| "grad_norm": 1.3255321979522705, | |
| "learning_rate": 3.422818791946309e-05, | |
| "loss": 3.5669, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.018173404568591983, | |
| "grad_norm": 0.9304305911064148, | |
| "learning_rate": 3.557046979865772e-05, | |
| "loss": 3.5456, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.01884649362668798, | |
| "grad_norm": 1.159280776977539, | |
| "learning_rate": 3.6912751677852356e-05, | |
| "loss": 3.5189, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.01951958268478398, | |
| "grad_norm": 0.9376134276390076, | |
| "learning_rate": 3.8255033557046985e-05, | |
| "loss": 3.432, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.02019267174287998, | |
| "grad_norm": 1.292802095413208, | |
| "learning_rate": 3.959731543624161e-05, | |
| "loss": 3.4564, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.02086576080097598, | |
| "grad_norm": 1.2383852005004883, | |
| "learning_rate": 4.0939597315436244e-05, | |
| "loss": 3.4194, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.02153884985907198, | |
| "grad_norm": 0.8546445369720459, | |
| "learning_rate": 4.228187919463087e-05, | |
| "loss": 3.3873, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.022211938917167977, | |
| "grad_norm": 1.4104743003845215, | |
| "learning_rate": 4.36241610738255e-05, | |
| "loss": 3.3647, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.022885027975263976, | |
| "grad_norm": 1.3548426628112793, | |
| "learning_rate": 4.496644295302014e-05, | |
| "loss": 3.3213, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.023558117033359978, | |
| "grad_norm": 1.1530455350875854, | |
| "learning_rate": 4.630872483221477e-05, | |
| "loss": 3.2789, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.024231206091455976, | |
| "grad_norm": 1.3092457056045532, | |
| "learning_rate": 4.76510067114094e-05, | |
| "loss": 3.2922, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.024904295149551975, | |
| "grad_norm": 1.577699065208435, | |
| "learning_rate": 4.8993288590604034e-05, | |
| "loss": 3.2308, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.025577384207647973, | |
| "grad_norm": 1.3348486423492432, | |
| "learning_rate": 5.033557046979866e-05, | |
| "loss": 3.2656, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.026250473265743975, | |
| "grad_norm": 1.1966625452041626, | |
| "learning_rate": 5.167785234899329e-05, | |
| "loss": 3.2703, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.026923562323839974, | |
| "grad_norm": 1.3125278949737549, | |
| "learning_rate": 5.302013422818792e-05, | |
| "loss": 3.2034, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.027596651381935972, | |
| "grad_norm": 1.1957862377166748, | |
| "learning_rate": 5.436241610738255e-05, | |
| "loss": 3.2476, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.02826974044003197, | |
| "grad_norm": 1.2177337408065796, | |
| "learning_rate": 5.570469798657718e-05, | |
| "loss": 3.2166, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.028942829498127973, | |
| "grad_norm": 0.8806389570236206, | |
| "learning_rate": 5.704697986577181e-05, | |
| "loss": 3.1722, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.02961591855622397, | |
| "grad_norm": 1.8180561065673828, | |
| "learning_rate": 5.838926174496645e-05, | |
| "loss": 3.1597, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.03028900761431997, | |
| "grad_norm": 1.1676297187805176, | |
| "learning_rate": 5.973154362416108e-05, | |
| "loss": 3.1829, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.030962096672415968, | |
| "grad_norm": 1.0163198709487915, | |
| "learning_rate": 6.107382550335571e-05, | |
| "loss": 3.1643, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.03163518573051197, | |
| "grad_norm": 1.0734015703201294, | |
| "learning_rate": 6.241610738255034e-05, | |
| "loss": 3.1692, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.03230827478860797, | |
| "grad_norm": 1.4144916534423828, | |
| "learning_rate": 6.375838926174497e-05, | |
| "loss": 3.1217, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.03298136384670397, | |
| "grad_norm": 1.5647915601730347, | |
| "learning_rate": 6.51006711409396e-05, | |
| "loss": 3.1324, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.033654452904799965, | |
| "grad_norm": 1.1999105215072632, | |
| "learning_rate": 6.644295302013423e-05, | |
| "loss": 3.1103, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.03432754196289597, | |
| "grad_norm": 1.512838363647461, | |
| "learning_rate": 6.778523489932886e-05, | |
| "loss": 3.1036, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.03500063102099196, | |
| "grad_norm": 1.1092990636825562, | |
| "learning_rate": 6.912751677852349e-05, | |
| "loss": 3.1244, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.035673720079087964, | |
| "grad_norm": 1.2763620615005493, | |
| "learning_rate": 7.046979865771812e-05, | |
| "loss": 3.0989, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.036346809137183966, | |
| "grad_norm": 1.2328648567199707, | |
| "learning_rate": 7.181208053691275e-05, | |
| "loss": 3.0506, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.03701989819527996, | |
| "grad_norm": 1.6039047241210938, | |
| "learning_rate": 7.315436241610739e-05, | |
| "loss": 3.0589, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.03769298725337596, | |
| "grad_norm": 1.7161307334899902, | |
| "learning_rate": 7.449664429530202e-05, | |
| "loss": 3.0296, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.03836607631147196, | |
| "grad_norm": 1.2628991603851318, | |
| "learning_rate": 7.583892617449665e-05, | |
| "loss": 3.063, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.03903916536956796, | |
| "grad_norm": 1.1414180994033813, | |
| "learning_rate": 7.718120805369128e-05, | |
| "loss": 3.0366, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.03971225442766396, | |
| "grad_norm": 1.5152932405471802, | |
| "learning_rate": 7.852348993288591e-05, | |
| "loss": 3.0241, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.04038534348575996, | |
| "grad_norm": 1.6119567155838013, | |
| "learning_rate": 7.986577181208054e-05, | |
| "loss": 3.0139, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.04105843254385596, | |
| "grad_norm": 1.6078091859817505, | |
| "learning_rate": 8.120805369127518e-05, | |
| "loss": 3.0288, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.04173152160195196, | |
| "grad_norm": 1.1092705726623535, | |
| "learning_rate": 8.255033557046981e-05, | |
| "loss": 2.9996, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.042404610660047956, | |
| "grad_norm": 1.2352242469787598, | |
| "learning_rate": 8.389261744966444e-05, | |
| "loss": 2.9872, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.04307769971814396, | |
| "grad_norm": 1.636400580406189, | |
| "learning_rate": 8.523489932885907e-05, | |
| "loss": 2.9814, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.04375078877623995, | |
| "grad_norm": 1.4877128601074219, | |
| "learning_rate": 8.65771812080537e-05, | |
| "loss": 2.9756, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.044423877834335955, | |
| "grad_norm": 1.2983709573745728, | |
| "learning_rate": 8.791946308724833e-05, | |
| "loss": 2.9756, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.04509696689243196, | |
| "grad_norm": 1.7350983619689941, | |
| "learning_rate": 8.926174496644296e-05, | |
| "loss": 2.9579, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.04577005595052795, | |
| "grad_norm": 0.978854775428772, | |
| "learning_rate": 9.060402684563759e-05, | |
| "loss": 2.9269, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.046443145008623954, | |
| "grad_norm": 1.362163782119751, | |
| "learning_rate": 9.194630872483221e-05, | |
| "loss": 2.99, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.047116234066719956, | |
| "grad_norm": 1.328202247619629, | |
| "learning_rate": 9.328859060402684e-05, | |
| "loss": 2.966, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.04778932312481595, | |
| "grad_norm": 1.2497445344924927, | |
| "learning_rate": 9.463087248322147e-05, | |
| "loss": 2.9254, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.04846241218291195, | |
| "grad_norm": 1.071092128753662, | |
| "learning_rate": 9.59731543624161e-05, | |
| "loss": 2.9597, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.04913550124100795, | |
| "grad_norm": 1.451729416847229, | |
| "learning_rate": 9.731543624161075e-05, | |
| "loss": 2.9605, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.04980859029910395, | |
| "grad_norm": 1.328731656074524, | |
| "learning_rate": 9.865771812080538e-05, | |
| "loss": 2.9493, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.05048167935719995, | |
| "grad_norm": 1.4675222635269165, | |
| "learning_rate": 0.0001, | |
| "loss": 2.9298, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.05115476841529595, | |
| "grad_norm": 1.208961009979248, | |
| "learning_rate": 9.99998761551904e-05, | |
| "loss": 2.9382, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.05182785747339195, | |
| "grad_norm": 1.0392056703567505, | |
| "learning_rate": 9.999950462137508e-05, | |
| "loss": 2.8829, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.05250094653148795, | |
| "grad_norm": 1.378711462020874, | |
| "learning_rate": 9.999888540039458e-05, | |
| "loss": 2.9041, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.053174035589583946, | |
| "grad_norm": 1.2687252759933472, | |
| "learning_rate": 9.999801849531635e-05, | |
| "loss": 2.9148, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.05384712464767995, | |
| "grad_norm": 1.2382102012634277, | |
| "learning_rate": 9.999690391043487e-05, | |
| "loss": 2.9107, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.05452021370577594, | |
| "grad_norm": 1.0215928554534912, | |
| "learning_rate": 9.999554165127159e-05, | |
| "loss": 2.9187, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.055193302763871945, | |
| "grad_norm": 1.2638540267944336, | |
| "learning_rate": 9.99939317245748e-05, | |
| "loss": 2.9261, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.055866391821967946, | |
| "grad_norm": 1.3826959133148193, | |
| "learning_rate": 9.999207413831982e-05, | |
| "loss": 2.8944, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.05653948088006394, | |
| "grad_norm": 1.0764875411987305, | |
| "learning_rate": 9.998996890170867e-05, | |
| "loss": 2.9031, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.05721256993815994, | |
| "grad_norm": 1.5163936614990234, | |
| "learning_rate": 9.99876160251703e-05, | |
| "loss": 2.8687, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.057885658996255945, | |
| "grad_norm": 1.5821291208267212, | |
| "learning_rate": 9.998501552036037e-05, | |
| "loss": 2.8828, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.05855874805435194, | |
| "grad_norm": 1.4572120904922485, | |
| "learning_rate": 9.998216740016124e-05, | |
| "loss": 2.8898, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.05923183711244794, | |
| "grad_norm": 1.2185218334197998, | |
| "learning_rate": 9.99790716786819e-05, | |
| "loss": 2.8456, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.05990492617054394, | |
| "grad_norm": 1.106446385383606, | |
| "learning_rate": 9.99757283712579e-05, | |
| "loss": 2.8793, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.06057801522863994, | |
| "grad_norm": 1.1448893547058105, | |
| "learning_rate": 9.997213749445129e-05, | |
| "loss": 2.8579, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.06125110428673594, | |
| "grad_norm": 1.1587834358215332, | |
| "learning_rate": 9.996829906605056e-05, | |
| "loss": 2.8839, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.061924193344831936, | |
| "grad_norm": 1.0969592332839966, | |
| "learning_rate": 9.996421310507046e-05, | |
| "loss": 2.8638, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.06259728240292793, | |
| "grad_norm": 0.9740116000175476, | |
| "learning_rate": 9.9959879631752e-05, | |
| "loss": 2.8455, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.06327037146102393, | |
| "grad_norm": 1.2307910919189453, | |
| "learning_rate": 9.995529866756231e-05, | |
| "loss": 2.8534, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.06394346051911994, | |
| "grad_norm": 1.5021939277648926, | |
| "learning_rate": 9.995047023519452e-05, | |
| "loss": 2.8469, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.06461654957721594, | |
| "grad_norm": 1.1044224500656128, | |
| "learning_rate": 9.994539435856771e-05, | |
| "loss": 2.8429, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.06528963863531194, | |
| "grad_norm": 1.4586883783340454, | |
| "learning_rate": 9.99400710628267e-05, | |
| "loss": 2.836, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.06596272769340794, | |
| "grad_norm": 1.2613426446914673, | |
| "learning_rate": 9.993450037434199e-05, | |
| "loss": 2.8243, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.06663581675150393, | |
| "grad_norm": 1.0347422361373901, | |
| "learning_rate": 9.992868232070963e-05, | |
| "loss": 2.7965, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.06730890580959993, | |
| "grad_norm": 2.1357574462890625, | |
| "learning_rate": 9.992261693075103e-05, | |
| "loss": 2.8486, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.06798199486769593, | |
| "grad_norm": 1.0357908010482788, | |
| "learning_rate": 9.991630423451286e-05, | |
| "loss": 2.8386, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.06865508392579193, | |
| "grad_norm": 1.1383159160614014, | |
| "learning_rate": 9.990974426326696e-05, | |
| "loss": 2.7874, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.06932817298388794, | |
| "grad_norm": 0.8452678322792053, | |
| "learning_rate": 9.990293704951001e-05, | |
| "loss": 2.786, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.07000126204198392, | |
| "grad_norm": 0.9482727646827698, | |
| "learning_rate": 9.989588262696357e-05, | |
| "loss": 2.8156, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.07067435110007993, | |
| "grad_norm": 0.8251766562461853, | |
| "learning_rate": 9.988858103057378e-05, | |
| "loss": 2.7588, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.07134744015817593, | |
| "grad_norm": 1.211065649986267, | |
| "learning_rate": 9.988103229651121e-05, | |
| "loss": 2.7623, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.07202052921627193, | |
| "grad_norm": 0.8990377187728882, | |
| "learning_rate": 9.987323646217075e-05, | |
| "loss": 2.8164, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.07269361827436793, | |
| "grad_norm": 0.9878025054931641, | |
| "learning_rate": 9.986519356617132e-05, | |
| "loss": 2.7847, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.07336670733246392, | |
| "grad_norm": 0.7551445364952087, | |
| "learning_rate": 9.985690364835576e-05, | |
| "loss": 2.8111, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.07403979639055992, | |
| "grad_norm": 0.9582260251045227, | |
| "learning_rate": 9.984836674979062e-05, | |
| "loss": 2.793, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.07471288544865592, | |
| "grad_norm": 0.8087739944458008, | |
| "learning_rate": 9.983958291276591e-05, | |
| "loss": 2.7464, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.07538597450675193, | |
| "grad_norm": 1.2373522520065308, | |
| "learning_rate": 9.983055218079493e-05, | |
| "loss": 2.7656, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.07605906356484793, | |
| "grad_norm": 0.9746289849281311, | |
| "learning_rate": 9.982127459861408e-05, | |
| "loss": 2.7765, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.07673215262294392, | |
| "grad_norm": 0.6946307420730591, | |
| "learning_rate": 9.981175021218255e-05, | |
| "loss": 2.7491, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.07740524168103992, | |
| "grad_norm": 0.8959107398986816, | |
| "learning_rate": 9.980197906868215e-05, | |
| "loss": 2.7565, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.07807833073913592, | |
| "grad_norm": 0.9889335036277771, | |
| "learning_rate": 9.979196121651716e-05, | |
| "loss": 2.7974, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.07875141979723192, | |
| "grad_norm": 0.9802746176719666, | |
| "learning_rate": 9.978169670531388e-05, | |
| "loss": 2.7772, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.07942450885532792, | |
| "grad_norm": 0.6934760808944702, | |
| "learning_rate": 9.977118558592059e-05, | |
| "loss": 2.7602, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.08009759791342393, | |
| "grad_norm": 0.8996357917785645, | |
| "learning_rate": 9.97604279104072e-05, | |
| "loss": 2.7669, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.08077068697151991, | |
| "grad_norm": 0.8844061493873596, | |
| "learning_rate": 9.974942373206499e-05, | |
| "loss": 2.7458, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.08144377602961592, | |
| "grad_norm": 1.023626685142517, | |
| "learning_rate": 9.973817310540638e-05, | |
| "loss": 2.7639, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.08211686508771192, | |
| "grad_norm": 0.8241132497787476, | |
| "learning_rate": 9.972667608616466e-05, | |
| "loss": 2.7457, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.08278995414580792, | |
| "grad_norm": 0.7864794135093689, | |
| "learning_rate": 9.971493273129364e-05, | |
| "loss": 2.763, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.08346304320390392, | |
| "grad_norm": 1.071751356124878, | |
| "learning_rate": 9.970294309896747e-05, | |
| "loss": 2.7347, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.08413613226199991, | |
| "grad_norm": 0.8978875279426575, | |
| "learning_rate": 9.969070724858031e-05, | |
| "loss": 2.7807, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.08480922132009591, | |
| "grad_norm": 0.9984204769134521, | |
| "learning_rate": 9.967822524074602e-05, | |
| "loss": 2.7399, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.08548231037819191, | |
| "grad_norm": 0.8611739873886108, | |
| "learning_rate": 9.966549713729787e-05, | |
| "loss": 2.753, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.08615539943628792, | |
| "grad_norm": 0.8647720217704773, | |
| "learning_rate": 9.965252300128826e-05, | |
| "loss": 2.7224, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.08682848849438392, | |
| "grad_norm": 0.8688477873802185, | |
| "learning_rate": 9.963930289698833e-05, | |
| "loss": 2.6879, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.0875015775524799, | |
| "grad_norm": 1.1445469856262207, | |
| "learning_rate": 9.962583688988778e-05, | |
| "loss": 2.739, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.08817466661057591, | |
| "grad_norm": 0.8668599128723145, | |
| "learning_rate": 9.961212504669437e-05, | |
| "loss": 2.6962, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.08884775566867191, | |
| "grad_norm": 0.905125617980957, | |
| "learning_rate": 9.959816743533375e-05, | |
| "loss": 2.7239, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.08952084472676791, | |
| "grad_norm": 0.8252028822898865, | |
| "learning_rate": 9.958396412494901e-05, | |
| "loss": 2.7381, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.09019393378486391, | |
| "grad_norm": 0.7380514740943909, | |
| "learning_rate": 9.956951518590043e-05, | |
| "loss": 2.7135, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.09086702284295992, | |
| "grad_norm": 0.7395239472389221, | |
| "learning_rate": 9.955482068976502e-05, | |
| "loss": 2.6954, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.0915401119010559, | |
| "grad_norm": 0.6564229726791382, | |
| "learning_rate": 9.953988070933628e-05, | |
| "loss": 2.7145, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.0922132009591519, | |
| "grad_norm": 0.7306910157203674, | |
| "learning_rate": 9.952469531862378e-05, | |
| "loss": 2.6951, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.09288629001724791, | |
| "grad_norm": 0.6810031533241272, | |
| "learning_rate": 9.950926459285277e-05, | |
| "loss": 2.7201, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.09355937907534391, | |
| "grad_norm": 0.6724168658256531, | |
| "learning_rate": 9.949358860846388e-05, | |
| "loss": 2.7112, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.09423246813343991, | |
| "grad_norm": 0.7065703272819519, | |
| "learning_rate": 9.947766744311268e-05, | |
| "loss": 2.6884, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.0949055571915359, | |
| "grad_norm": 0.8231908679008484, | |
| "learning_rate": 9.946150117566931e-05, | |
| "loss": 2.7286, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.0955786462496319, | |
| "grad_norm": 0.9570270776748657, | |
| "learning_rate": 9.944508988621812e-05, | |
| "loss": 2.7166, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.0962517353077279, | |
| "grad_norm": 0.9357023239135742, | |
| "learning_rate": 9.94284336560572e-05, | |
| "loss": 2.6768, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.0969248243658239, | |
| "grad_norm": 0.6350796222686768, | |
| "learning_rate": 9.941153256769809e-05, | |
| "loss": 2.6921, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.09759791342391991, | |
| "grad_norm": 0.6700872778892517, | |
| "learning_rate": 9.939438670486525e-05, | |
| "loss": 2.6847, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.0982710024820159, | |
| "grad_norm": 0.6851752400398254, | |
| "learning_rate": 9.937699615249572e-05, | |
| "loss": 2.6586, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.0989440915401119, | |
| "grad_norm": 0.7098946571350098, | |
| "learning_rate": 9.935936099673871e-05, | |
| "loss": 2.6793, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.0996171805982079, | |
| "grad_norm": 0.680543839931488, | |
| "learning_rate": 9.934148132495511e-05, | |
| "loss": 2.6763, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.1002902696563039, | |
| "grad_norm": 0.6832155585289001, | |
| "learning_rate": 9.932335722571709e-05, | |
| "loss": 2.6768, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.1009633587143999, | |
| "grad_norm": 0.8236553072929382, | |
| "learning_rate": 9.930498878880768e-05, | |
| "loss": 2.6738, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.10163644777249589, | |
| "grad_norm": 1.2765145301818848, | |
| "learning_rate": 9.928637610522032e-05, | |
| "loss": 2.6575, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.1023095368305919, | |
| "grad_norm": 0.8765101432800293, | |
| "learning_rate": 9.926751926715836e-05, | |
| "loss": 2.6521, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.1029826258886879, | |
| "grad_norm": 0.7371405959129333, | |
| "learning_rate": 9.924841836803467e-05, | |
| "loss": 2.6916, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.1036557149467839, | |
| "grad_norm": 0.8433207273483276, | |
| "learning_rate": 9.922907350247115e-05, | |
| "loss": 2.6605, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.1043288040048799, | |
| "grad_norm": 0.7216758131980896, | |
| "learning_rate": 9.920948476629821e-05, | |
| "loss": 2.6686, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.1050018930629759, | |
| "grad_norm": 0.7831395864486694, | |
| "learning_rate": 9.918965225655442e-05, | |
| "loss": 2.6783, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.10567498212107189, | |
| "grad_norm": 0.5831722617149353, | |
| "learning_rate": 9.916957607148591e-05, | |
| "loss": 2.6633, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.10634807117916789, | |
| "grad_norm": 0.6143165826797485, | |
| "learning_rate": 9.91492563105459e-05, | |
| "loss": 2.6785, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.1070211602372639, | |
| "grad_norm": 0.8080072999000549, | |
| "learning_rate": 9.912869307439432e-05, | |
| "loss": 2.672, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.1076942492953599, | |
| "grad_norm": 1.0352627038955688, | |
| "learning_rate": 9.910788646489713e-05, | |
| "loss": 2.6474, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.1083673383534559, | |
| "grad_norm": 0.6871941685676575, | |
| "learning_rate": 9.908683658512597e-05, | |
| "loss": 2.6387, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.10904042741155189, | |
| "grad_norm": 0.7802032828330994, | |
| "learning_rate": 9.906554353935758e-05, | |
| "loss": 2.6584, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.10971351646964789, | |
| "grad_norm": 0.7139161825180054, | |
| "learning_rate": 9.904400743307326e-05, | |
| "loss": 2.6728, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.11038660552774389, | |
| "grad_norm": 1.0033543109893799, | |
| "learning_rate": 9.902222837295844e-05, | |
| "loss": 2.666, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.11105969458583989, | |
| "grad_norm": 1.0726121664047241, | |
| "learning_rate": 9.900020646690205e-05, | |
| "loss": 2.6566, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.11173278364393589, | |
| "grad_norm": 0.8290823698043823, | |
| "learning_rate": 9.897794182399606e-05, | |
| "loss": 2.6592, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.11240587270203188, | |
| "grad_norm": 0.7432393431663513, | |
| "learning_rate": 9.895543455453486e-05, | |
| "loss": 2.6604, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.11307896176012788, | |
| "grad_norm": 0.6097532510757446, | |
| "learning_rate": 9.893268477001479e-05, | |
| "loss": 2.6613, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.11375205081822388, | |
| "grad_norm": 0.6818569898605347, | |
| "learning_rate": 9.890969258313358e-05, | |
| "loss": 2.6238, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.11442513987631989, | |
| "grad_norm": 0.6779484748840332, | |
| "learning_rate": 9.888645810778974e-05, | |
| "loss": 2.6512, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.11509822893441589, | |
| "grad_norm": 0.697517454624176, | |
| "learning_rate": 9.886298145908203e-05, | |
| "loss": 2.6278, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.11577131799251189, | |
| "grad_norm": 0.700210452079773, | |
| "learning_rate": 9.88392627533089e-05, | |
| "loss": 2.6592, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.11644440705060788, | |
| "grad_norm": 0.6507501602172852, | |
| "learning_rate": 9.88153021079679e-05, | |
| "loss": 2.6682, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.11711749610870388, | |
| "grad_norm": 0.7211961150169373, | |
| "learning_rate": 9.87910996417551e-05, | |
| "loss": 2.6919, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.11779058516679988, | |
| "grad_norm": 0.7282651662826538, | |
| "learning_rate": 9.876665547456446e-05, | |
| "loss": 2.6148, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.11846367422489588, | |
| "grad_norm": 0.6177457571029663, | |
| "learning_rate": 9.874196972748735e-05, | |
| "loss": 2.6072, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.11913676328299189, | |
| "grad_norm": 0.5497561693191528, | |
| "learning_rate": 9.871704252281179e-05, | |
| "loss": 2.6493, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.11980985234108787, | |
| "grad_norm": 0.6414313912391663, | |
| "learning_rate": 9.869187398402202e-05, | |
| "loss": 2.6505, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.12048294139918388, | |
| "grad_norm": 0.6123400330543518, | |
| "learning_rate": 9.866646423579773e-05, | |
| "loss": 2.6168, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.12115603045727988, | |
| "grad_norm": 0.5811213850975037, | |
| "learning_rate": 9.864081340401354e-05, | |
| "loss": 2.6367, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.12182911951537588, | |
| "grad_norm": 0.7445274591445923, | |
| "learning_rate": 9.861492161573837e-05, | |
| "loss": 2.6288, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.12250220857347188, | |
| "grad_norm": 0.9515259861946106, | |
| "learning_rate": 9.858878899923473e-05, | |
| "loss": 2.6458, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.12317529763156787, | |
| "grad_norm": 0.7661240696907043, | |
| "learning_rate": 9.856241568395818e-05, | |
| "loss": 2.6289, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.12384838668966387, | |
| "grad_norm": 0.704278826713562, | |
| "learning_rate": 9.853580180055669e-05, | |
| "loss": 2.6043, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.12452147574775987, | |
| "grad_norm": 0.6069918274879456, | |
| "learning_rate": 9.850894748086986e-05, | |
| "loss": 2.6193, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.12519456480585586, | |
| "grad_norm": 0.8171132802963257, | |
| "learning_rate": 9.848185285792844e-05, | |
| "loss": 2.6482, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.12586765386395188, | |
| "grad_norm": 0.7318093776702881, | |
| "learning_rate": 9.845451806595354e-05, | |
| "loss": 2.6285, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.12654074292204787, | |
| "grad_norm": 0.7122377157211304, | |
| "learning_rate": 9.84269432403561e-05, | |
| "loss": 2.6147, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.12721383198014388, | |
| "grad_norm": 0.6400437355041504, | |
| "learning_rate": 9.839912851773602e-05, | |
| "loss": 2.576, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.12788692103823987, | |
| "grad_norm": 0.6208680868148804, | |
| "learning_rate": 9.837107403588171e-05, | |
| "loss": 2.6056, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.12856001009633586, | |
| "grad_norm": 0.6690713167190552, | |
| "learning_rate": 9.834277993376924e-05, | |
| "loss": 2.5884, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.12923309915443187, | |
| "grad_norm": 0.613250732421875, | |
| "learning_rate": 9.831424635156168e-05, | |
| "loss": 2.5824, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.12990618821252786, | |
| "grad_norm": 0.6921446323394775, | |
| "learning_rate": 9.82854734306085e-05, | |
| "loss": 2.6239, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.13057927727062388, | |
| "grad_norm": 0.6627881526947021, | |
| "learning_rate": 9.825646131344478e-05, | |
| "loss": 2.5859, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.13125236632871987, | |
| "grad_norm": 0.7399138808250427, | |
| "learning_rate": 9.822721014379052e-05, | |
| "loss": 2.6492, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.13192545538681588, | |
| "grad_norm": 0.633929967880249, | |
| "learning_rate": 9.819772006654993e-05, | |
| "loss": 2.5935, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.13259854444491187, | |
| "grad_norm": 0.6434243321418762, | |
| "learning_rate": 9.816799122781075e-05, | |
| "loss": 2.6231, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.13327163350300786, | |
| "grad_norm": 0.6136692762374878, | |
| "learning_rate": 9.813802377484346e-05, | |
| "loss": 2.5803, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.13394472256110387, | |
| "grad_norm": 0.6683850288391113, | |
| "learning_rate": 9.810781785610062e-05, | |
| "loss": 2.5854, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.13461781161919986, | |
| "grad_norm": 0.705646276473999, | |
| "learning_rate": 9.807737362121605e-05, | |
| "loss": 2.5954, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.13529090067729588, | |
| "grad_norm": 0.7973448038101196, | |
| "learning_rate": 9.80466912210042e-05, | |
| "loss": 2.5824, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.13596398973539187, | |
| "grad_norm": 0.7568982839584351, | |
| "learning_rate": 9.801577080745929e-05, | |
| "loss": 2.5875, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.13663707879348785, | |
| "grad_norm": 0.6286936402320862, | |
| "learning_rate": 9.798461253375462e-05, | |
| "loss": 2.6047, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.13731016785158387, | |
| "grad_norm": 0.6657843589782715, | |
| "learning_rate": 9.795321655424184e-05, | |
| "loss": 2.6066, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.13798325690967986, | |
| "grad_norm": 0.6954363584518433, | |
| "learning_rate": 9.792158302445009e-05, | |
| "loss": 2.5885, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.13865634596777587, | |
| "grad_norm": 0.9607271552085876, | |
| "learning_rate": 9.788971210108533e-05, | |
| "loss": 2.6212, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.13932943502587186, | |
| "grad_norm": 0.9636927843093872, | |
| "learning_rate": 9.785760394202948e-05, | |
| "loss": 2.5552, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.14000252408396785, | |
| "grad_norm": 0.6435789465904236, | |
| "learning_rate": 9.782525870633967e-05, | |
| "loss": 2.6199, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.14067561314206387, | |
| "grad_norm": 0.7337214350700378, | |
| "learning_rate": 9.779267655424754e-05, | |
| "loss": 2.5683, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.14134870220015985, | |
| "grad_norm": 0.6656831502914429, | |
| "learning_rate": 9.775985764715827e-05, | |
| "loss": 2.5735, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.14202179125825587, | |
| "grad_norm": 0.6302557587623596, | |
| "learning_rate": 9.772680214764992e-05, | |
| "loss": 2.5752, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.14269488031635186, | |
| "grad_norm": 0.7694079279899597, | |
| "learning_rate": 9.769351021947258e-05, | |
| "loss": 2.5742, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.14336796937444785, | |
| "grad_norm": 0.743331253528595, | |
| "learning_rate": 9.765998202754753e-05, | |
| "loss": 2.5968, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.14404105843254386, | |
| "grad_norm": 0.7328934669494629, | |
| "learning_rate": 9.762621773796647e-05, | |
| "loss": 2.595, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.14471414749063985, | |
| "grad_norm": 0.786054253578186, | |
| "learning_rate": 9.759221751799072e-05, | |
| "loss": 2.6048, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.14538723654873587, | |
| "grad_norm": 0.7093940377235413, | |
| "learning_rate": 9.755798153605027e-05, | |
| "loss": 2.5766, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.14606032560683185, | |
| "grad_norm": 0.6153858304023743, | |
| "learning_rate": 9.752350996174308e-05, | |
| "loss": 2.5598, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.14673341466492784, | |
| "grad_norm": 0.7648425102233887, | |
| "learning_rate": 9.748880296583418e-05, | |
| "loss": 2.5768, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.14740650372302386, | |
| "grad_norm": 0.777052640914917, | |
| "learning_rate": 9.745386072025481e-05, | |
| "loss": 2.5571, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.14807959278111985, | |
| "grad_norm": 0.6382849812507629, | |
| "learning_rate": 9.741868339810161e-05, | |
| "loss": 2.564, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.14875268183921586, | |
| "grad_norm": 0.6427432298660278, | |
| "learning_rate": 9.738327117363571e-05, | |
| "loss": 2.5625, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.14942577089731185, | |
| "grad_norm": 0.6348729133605957, | |
| "learning_rate": 9.734762422228194e-05, | |
| "loss": 2.5622, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.15009885995540784, | |
| "grad_norm": 0.7136998176574707, | |
| "learning_rate": 9.731174272062791e-05, | |
| "loss": 2.5386, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.15077194901350385, | |
| "grad_norm": 0.7903639078140259, | |
| "learning_rate": 9.727562684642308e-05, | |
| "loss": 2.5593, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.15144503807159984, | |
| "grad_norm": 0.6537742018699646, | |
| "learning_rate": 9.723927677857805e-05, | |
| "loss": 2.5373, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.15211812712969586, | |
| "grad_norm": 0.6430116295814514, | |
| "learning_rate": 9.720269269716347e-05, | |
| "loss": 2.583, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.15279121618779185, | |
| "grad_norm": 0.5976120829582214, | |
| "learning_rate": 9.716587478340928e-05, | |
| "loss": 2.5729, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.15346430524588783, | |
| "grad_norm": 0.7713409066200256, | |
| "learning_rate": 9.712882321970383e-05, | |
| "loss": 2.5858, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.15413739430398385, | |
| "grad_norm": 0.6205691695213318, | |
| "learning_rate": 9.70915381895928e-05, | |
| "loss": 2.5248, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.15481048336207984, | |
| "grad_norm": 0.5880588293075562, | |
| "learning_rate": 9.705401987777855e-05, | |
| "loss": 2.5811, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.15548357242017585, | |
| "grad_norm": 0.645301342010498, | |
| "learning_rate": 9.701626847011899e-05, | |
| "loss": 2.5677, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.15615666147827184, | |
| "grad_norm": 0.5872762799263, | |
| "learning_rate": 9.697828415362674e-05, | |
| "loss": 2.5777, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.15682975053636786, | |
| "grad_norm": 0.5601847767829895, | |
| "learning_rate": 9.694006711646823e-05, | |
| "loss": 2.5563, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.15750283959446384, | |
| "grad_norm": 0.6171532869338989, | |
| "learning_rate": 9.690161754796274e-05, | |
| "loss": 2.5559, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.15817592865255983, | |
| "grad_norm": 0.5213554501533508, | |
| "learning_rate": 9.686293563858142e-05, | |
| "loss": 2.5543, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.15884901771065585, | |
| "grad_norm": 0.5173961520195007, | |
| "learning_rate": 9.682402157994643e-05, | |
| "loss": 2.5674, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.15952210676875184, | |
| "grad_norm": 0.5519447326660156, | |
| "learning_rate": 9.678487556482996e-05, | |
| "loss": 2.5385, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.16019519582684785, | |
| "grad_norm": 0.6517807841300964, | |
| "learning_rate": 9.674549778715322e-05, | |
| "loss": 2.5712, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.16086828488494384, | |
| "grad_norm": 0.745688796043396, | |
| "learning_rate": 9.670588844198554e-05, | |
| "loss": 2.5743, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.16154137394303983, | |
| "grad_norm": 0.6915059685707092, | |
| "learning_rate": 9.666604772554342e-05, | |
| "loss": 2.5536, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.16221446300113584, | |
| "grad_norm": 0.6770395636558533, | |
| "learning_rate": 9.662597583518946e-05, | |
| "loss": 2.5501, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.16288755205923183, | |
| "grad_norm": 0.7015348076820374, | |
| "learning_rate": 9.658567296943151e-05, | |
| "loss": 2.5488, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.16356064111732785, | |
| "grad_norm": 0.6871614456176758, | |
| "learning_rate": 9.654513932792157e-05, | |
| "loss": 2.55, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.16423373017542384, | |
| "grad_norm": 0.634372353553772, | |
| "learning_rate": 9.650437511145494e-05, | |
| "loss": 2.5231, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.16490681923351982, | |
| "grad_norm": 0.6015982031822205, | |
| "learning_rate": 9.646338052196904e-05, | |
| "loss": 2.5338, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.16557990829161584, | |
| "grad_norm": 0.5995615124702454, | |
| "learning_rate": 9.642215576254256e-05, | |
| "loss": 2.5449, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.16625299734971183, | |
| "grad_norm": 0.5862424373626709, | |
| "learning_rate": 9.638070103739443e-05, | |
| "loss": 2.5564, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.16692608640780784, | |
| "grad_norm": 0.49670928716659546, | |
| "learning_rate": 9.633901655188271e-05, | |
| "loss": 2.5283, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.16759917546590383, | |
| "grad_norm": 0.5854897499084473, | |
| "learning_rate": 9.629710251250372e-05, | |
| "loss": 2.5142, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.16827226452399982, | |
| "grad_norm": 0.7148288488388062, | |
| "learning_rate": 9.625495912689088e-05, | |
| "loss": 2.5528, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.16894535358209584, | |
| "grad_norm": 0.7020834684371948, | |
| "learning_rate": 9.62125866038138e-05, | |
| "loss": 2.5157, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.16961844264019182, | |
| "grad_norm": 0.5895742774009705, | |
| "learning_rate": 9.616998515317714e-05, | |
| "loss": 2.5492, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.17029153169828784, | |
| "grad_norm": 0.5936703681945801, | |
| "learning_rate": 9.612715498601966e-05, | |
| "loss": 2.5172, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.17096462075638383, | |
| "grad_norm": 0.6276780962944031, | |
| "learning_rate": 9.608409631451311e-05, | |
| "loss": 2.5448, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.17163770981447982, | |
| "grad_norm": 0.6327413320541382, | |
| "learning_rate": 9.60408093519612e-05, | |
| "loss": 2.4831, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.17231079887257583, | |
| "grad_norm": 0.7642117142677307, | |
| "learning_rate": 9.599729431279856e-05, | |
| "loss": 2.5209, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.17298388793067182, | |
| "grad_norm": 0.7974164485931396, | |
| "learning_rate": 9.595355141258966e-05, | |
| "loss": 2.5102, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.17365697698876784, | |
| "grad_norm": 0.7302394509315491, | |
| "learning_rate": 9.590958086802776e-05, | |
| "loss": 2.4705, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.17433006604686382, | |
| "grad_norm": 0.5221154093742371, | |
| "learning_rate": 9.586538289693378e-05, | |
| "loss": 2.5355, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.1750031551049598, | |
| "grad_norm": 0.7747745513916016, | |
| "learning_rate": 9.582095771825533e-05, | |
| "loss": 2.5375, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.17567624416305583, | |
| "grad_norm": 0.6088679432868958, | |
| "learning_rate": 9.577630555206549e-05, | |
| "loss": 2.546, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.17634933322115182, | |
| "grad_norm": 0.7077321410179138, | |
| "learning_rate": 9.573142661956183e-05, | |
| "loss": 2.5199, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.17702242227924783, | |
| "grad_norm": 0.7267037630081177, | |
| "learning_rate": 9.568632114306524e-05, | |
| "loss": 2.5125, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.17769551133734382, | |
| "grad_norm": 0.591948926448822, | |
| "learning_rate": 9.564098934601895e-05, | |
| "loss": 2.5411, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.1783686003954398, | |
| "grad_norm": 0.5513604283332825, | |
| "learning_rate": 9.559543145298721e-05, | |
| "loss": 2.5178, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.17904168945353582, | |
| "grad_norm": 0.6161476373672485, | |
| "learning_rate": 9.55496476896544e-05, | |
| "loss": 2.5123, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.1797147785116318, | |
| "grad_norm": 0.5339455604553223, | |
| "learning_rate": 9.550363828282375e-05, | |
| "loss": 2.532, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.18038786756972783, | |
| "grad_norm": 0.5792993307113647, | |
| "learning_rate": 9.545740346041634e-05, | |
| "loss": 2.5163, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.18106095662782382, | |
| "grad_norm": 0.6918086409568787, | |
| "learning_rate": 9.541094345146986e-05, | |
| "loss": 2.516, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.18173404568591983, | |
| "grad_norm": 0.577778160572052, | |
| "learning_rate": 9.536425848613754e-05, | |
| "loss": 2.4866, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.18240713474401582, | |
| "grad_norm": 0.6180362105369568, | |
| "learning_rate": 9.531734879568703e-05, | |
| "loss": 2.5186, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.1830802238021118, | |
| "grad_norm": 0.6912726163864136, | |
| "learning_rate": 9.527021461249919e-05, | |
| "loss": 2.5302, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.18375331286020782, | |
| "grad_norm": 0.6083509922027588, | |
| "learning_rate": 9.522285617006695e-05, | |
| "loss": 2.5226, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.1844264019183038, | |
| "grad_norm": 0.6165098547935486, | |
| "learning_rate": 9.517527370299424e-05, | |
| "loss": 2.5136, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.18509949097639983, | |
| "grad_norm": 0.6208174228668213, | |
| "learning_rate": 9.512746744699473e-05, | |
| "loss": 2.4967, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.18577258003449582, | |
| "grad_norm": 0.5808335542678833, | |
| "learning_rate": 9.507943763889062e-05, | |
| "loss": 2.5075, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.1864456690925918, | |
| "grad_norm": 0.5772544741630554, | |
| "learning_rate": 9.503118451661168e-05, | |
| "loss": 2.517, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.18711875815068782, | |
| "grad_norm": 0.5901827812194824, | |
| "learning_rate": 9.498270831919383e-05, | |
| "loss": 2.5008, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.1877918472087838, | |
| "grad_norm": 0.6181182265281677, | |
| "learning_rate": 9.493400928677809e-05, | |
| "loss": 2.4841, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.18846493626687982, | |
| "grad_norm": 0.6515311002731323, | |
| "learning_rate": 9.488508766060935e-05, | |
| "loss": 2.5213, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.1891380253249758, | |
| "grad_norm": 0.5359539985656738, | |
| "learning_rate": 9.483594368303521e-05, | |
| "loss": 2.5276, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.1898111143830718, | |
| "grad_norm": 0.7273654937744141, | |
| "learning_rate": 9.47865775975047e-05, | |
| "loss": 2.4901, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.19048420344116782, | |
| "grad_norm": 0.642062246799469, | |
| "learning_rate": 9.473698964856719e-05, | |
| "loss": 2.4732, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.1911572924992638, | |
| "grad_norm": 0.6698111891746521, | |
| "learning_rate": 9.468718008187106e-05, | |
| "loss": 2.5044, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.19183038155735982, | |
| "grad_norm": 0.6473941206932068, | |
| "learning_rate": 9.463714914416255e-05, | |
| "loss": 2.5149, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.1925034706154558, | |
| "grad_norm": 0.5659685730934143, | |
| "learning_rate": 9.458689708328459e-05, | |
| "loss": 2.4881, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.1931765596735518, | |
| "grad_norm": 0.736530601978302, | |
| "learning_rate": 9.453642414817541e-05, | |
| "loss": 2.5245, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.1938496487316478, | |
| "grad_norm": 0.6213739514350891, | |
| "learning_rate": 9.448573058886747e-05, | |
| "loss": 2.5026, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.1945227377897438, | |
| "grad_norm": 0.6373304724693298, | |
| "learning_rate": 9.443481665648615e-05, | |
| "loss": 2.5033, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.19519582684783982, | |
| "grad_norm": 0.7117682695388794, | |
| "learning_rate": 9.438368260324846e-05, | |
| "loss": 2.5133, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.1958689159059358, | |
| "grad_norm": 0.566567063331604, | |
| "learning_rate": 9.433232868246192e-05, | |
| "loss": 2.4923, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.1965420049640318, | |
| "grad_norm": 0.6423395872116089, | |
| "learning_rate": 9.428075514852319e-05, | |
| "loss": 2.4717, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.1972150940221278, | |
| "grad_norm": 0.6600716710090637, | |
| "learning_rate": 9.422896225691683e-05, | |
| "loss": 2.5027, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.1978881830802238, | |
| "grad_norm": 0.5523872375488281, | |
| "learning_rate": 9.417695026421409e-05, | |
| "loss": 2.4834, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.1985612721383198, | |
| "grad_norm": 0.6183229088783264, | |
| "learning_rate": 9.412471942807157e-05, | |
| "loss": 2.5416, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.1992343611964158, | |
| "grad_norm": 0.5662925839424133, | |
| "learning_rate": 9.407227000723e-05, | |
| "loss": 2.4994, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.1999074502545118, | |
| "grad_norm": 0.58561110496521, | |
| "learning_rate": 9.401960226151291e-05, | |
| "loss": 2.4957, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.2005805393126078, | |
| "grad_norm": 0.5408748984336853, | |
| "learning_rate": 9.396671645182539e-05, | |
| "loss": 2.4932, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.2012536283707038, | |
| "grad_norm": 0.5413419008255005, | |
| "learning_rate": 9.391361284015274e-05, | |
| "loss": 2.4623, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.2019267174287998, | |
| "grad_norm": 0.5581361651420593, | |
| "learning_rate": 9.386029168955925e-05, | |
| "loss": 2.4636, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.2025998064868958, | |
| "grad_norm": 0.5392501950263977, | |
| "learning_rate": 9.380675326418683e-05, | |
| "loss": 2.5353, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.20327289554499178, | |
| "grad_norm": 0.614374041557312, | |
| "learning_rate": 9.37529978292537e-05, | |
| "loss": 2.4918, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.2039459846030878, | |
| "grad_norm": 0.5186980962753296, | |
| "learning_rate": 9.369902565105315e-05, | |
| "loss": 2.4829, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.2046190736611838, | |
| "grad_norm": 0.6255108118057251, | |
| "learning_rate": 9.364483699695215e-05, | |
| "loss": 2.5047, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.2052921627192798, | |
| "grad_norm": 0.7298846244812012, | |
| "learning_rate": 9.359043213539e-05, | |
| "loss": 2.5422, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.2059652517773758, | |
| "grad_norm": 0.7413848042488098, | |
| "learning_rate": 9.353581133587712e-05, | |
| "loss": 2.4572, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.2066383408354718, | |
| "grad_norm": 0.7455918788909912, | |
| "learning_rate": 9.348097486899362e-05, | |
| "loss": 2.4967, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.2073114298935678, | |
| "grad_norm": 0.5773429870605469, | |
| "learning_rate": 9.342592300638795e-05, | |
| "loss": 2.4956, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.20798451895166378, | |
| "grad_norm": 0.5804847478866577, | |
| "learning_rate": 9.337065602077562e-05, | |
| "loss": 2.5137, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.2086576080097598, | |
| "grad_norm": 0.5692312121391296, | |
| "learning_rate": 9.331517418593778e-05, | |
| "loss": 2.4982, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.2093306970678558, | |
| "grad_norm": 0.5889792442321777, | |
| "learning_rate": 9.325947777671996e-05, | |
| "loss": 2.5212, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.2100037861259518, | |
| "grad_norm": 0.6200716495513916, | |
| "learning_rate": 9.320356706903059e-05, | |
| "loss": 2.4906, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.2106768751840478, | |
| "grad_norm": 0.5470912456512451, | |
| "learning_rate": 9.314744233983969e-05, | |
| "loss": 2.4984, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.21134996424214378, | |
| "grad_norm": 0.5696718692779541, | |
| "learning_rate": 9.309110386717755e-05, | |
| "loss": 2.4749, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.2120230533002398, | |
| "grad_norm": 0.5643149614334106, | |
| "learning_rate": 9.303455193013322e-05, | |
| "loss": 2.465, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.21269614235833578, | |
| "grad_norm": 0.581017017364502, | |
| "learning_rate": 9.29777868088533e-05, | |
| "loss": 2.4705, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.2133692314164318, | |
| "grad_norm": 0.5466561913490295, | |
| "learning_rate": 9.29208087845404e-05, | |
| "loss": 2.4705, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.2140423204745278, | |
| "grad_norm": 0.62199467420578, | |
| "learning_rate": 9.286361813945181e-05, | |
| "loss": 2.468, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.21471540953262377, | |
| "grad_norm": 0.6129001975059509, | |
| "learning_rate": 9.280621515689814e-05, | |
| "loss": 2.4807, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.2153884985907198, | |
| "grad_norm": 0.5707617998123169, | |
| "learning_rate": 9.274860012124182e-05, | |
| "loss": 2.4628, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.21606158764881578, | |
| "grad_norm": 0.5998787879943848, | |
| "learning_rate": 9.269077331789578e-05, | |
| "loss": 2.469, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.2167346767069118, | |
| "grad_norm": 0.4872931241989136, | |
| "learning_rate": 9.263273503332201e-05, | |
| "loss": 2.458, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.21740776576500778, | |
| "grad_norm": 0.48162540793418884, | |
| "learning_rate": 9.257448555503013e-05, | |
| "loss": 2.4706, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.21808085482310377, | |
| "grad_norm": 0.5514042377471924, | |
| "learning_rate": 9.251602517157594e-05, | |
| "loss": 2.4834, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.21875394388119979, | |
| "grad_norm": 0.4928090572357178, | |
| "learning_rate": 9.245735417256004e-05, | |
| "loss": 2.4816, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.21942703293929577, | |
| "grad_norm": 0.5588257908821106, | |
| "learning_rate": 9.239847284862639e-05, | |
| "loss": 2.4726, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.2201001219973918, | |
| "grad_norm": 0.4884628355503082, | |
| "learning_rate": 9.233938149146086e-05, | |
| "loss": 2.4834, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.22077321105548778, | |
| "grad_norm": 0.5175626277923584, | |
| "learning_rate": 9.228008039378972e-05, | |
| "loss": 2.4571, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.22144630011358377, | |
| "grad_norm": 0.48624858260154724, | |
| "learning_rate": 9.222056984937835e-05, | |
| "loss": 2.5005, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.22211938917167978, | |
| "grad_norm": 0.6215223073959351, | |
| "learning_rate": 9.21608501530296e-05, | |
| "loss": 2.4743, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.22279247822977577, | |
| "grad_norm": 0.6394065618515015, | |
| "learning_rate": 9.210092160058245e-05, | |
| "loss": 2.5083, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.22346556728787179, | |
| "grad_norm": 0.6800668835639954, | |
| "learning_rate": 9.20407844889105e-05, | |
| "loss": 2.4848, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.22413865634596777, | |
| "grad_norm": 0.7512221932411194, | |
| "learning_rate": 9.198043911592053e-05, | |
| "loss": 2.4759, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.22481174540406376, | |
| "grad_norm": 0.7012692093849182, | |
| "learning_rate": 9.1919885780551e-05, | |
| "loss": 2.4484, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.22548483446215978, | |
| "grad_norm": 0.7792785167694092, | |
| "learning_rate": 9.185912478277052e-05, | |
| "loss": 2.4628, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.22615792352025577, | |
| "grad_norm": 0.6083918213844299, | |
| "learning_rate": 9.17981564235765e-05, | |
| "loss": 2.4606, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.22683101257835178, | |
| "grad_norm": 0.6361806988716125, | |
| "learning_rate": 9.173698100499351e-05, | |
| "loss": 2.4683, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.22750410163644777, | |
| "grad_norm": 0.572401762008667, | |
| "learning_rate": 9.167559883007188e-05, | |
| "loss": 2.4802, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.22817719069454379, | |
| "grad_norm": 0.6176139712333679, | |
| "learning_rate": 9.161401020288616e-05, | |
| "loss": 2.4904, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.22885027975263977, | |
| "grad_norm": 0.5795890688896179, | |
| "learning_rate": 9.155221542853362e-05, | |
| "loss": 2.489, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.22952336881073576, | |
| "grad_norm": 0.544706404209137, | |
| "learning_rate": 9.149021481313276e-05, | |
| "loss": 2.4342, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 0.23019645786883178, | |
| "grad_norm": 0.5396670699119568, | |
| "learning_rate": 9.142800866382173e-05, | |
| "loss": 2.4824, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.23086954692692777, | |
| "grad_norm": 0.5719538927078247, | |
| "learning_rate": 9.13655972887569e-05, | |
| "loss": 2.4956, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 0.23154263598502378, | |
| "grad_norm": 0.5415273308753967, | |
| "learning_rate": 9.130298099711125e-05, | |
| "loss": 2.4569, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.23221572504311977, | |
| "grad_norm": 0.49953708052635193, | |
| "learning_rate": 9.124016009907287e-05, | |
| "loss": 2.4281, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.23288881410121576, | |
| "grad_norm": 0.5383540987968445, | |
| "learning_rate": 9.11771349058435e-05, | |
| "loss": 2.4527, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.23356190315931177, | |
| "grad_norm": 0.5342861413955688, | |
| "learning_rate": 9.111390572963683e-05, | |
| "loss": 2.4786, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 0.23423499221740776, | |
| "grad_norm": 0.5322859883308411, | |
| "learning_rate": 9.105047288367707e-05, | |
| "loss": 2.4587, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.23490808127550378, | |
| "grad_norm": 0.5226521492004395, | |
| "learning_rate": 9.098683668219737e-05, | |
| "loss": 2.4605, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 0.23558117033359977, | |
| "grad_norm": 0.5837043523788452, | |
| "learning_rate": 9.092299744043826e-05, | |
| "loss": 2.4591, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.23625425939169575, | |
| "grad_norm": 0.5048717260360718, | |
| "learning_rate": 9.08589554746461e-05, | |
| "loss": 2.4809, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 0.23692734844979177, | |
| "grad_norm": 0.6950474381446838, | |
| "learning_rate": 9.079471110207149e-05, | |
| "loss": 2.4261, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.23760043750788776, | |
| "grad_norm": 0.6609525680541992, | |
| "learning_rate": 9.07302646409677e-05, | |
| "loss": 2.4631, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 0.23827352656598377, | |
| "grad_norm": 0.7101286053657532, | |
| "learning_rate": 9.066561641058912e-05, | |
| "loss": 2.4427, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.23894661562407976, | |
| "grad_norm": 0.5411195755004883, | |
| "learning_rate": 9.060076673118967e-05, | |
| "loss": 2.4933, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.23961970468217575, | |
| "grad_norm": 0.5625993609428406, | |
| "learning_rate": 9.05357159240212e-05, | |
| "loss": 2.4357, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.24029279374027177, | |
| "grad_norm": 0.8118378520011902, | |
| "learning_rate": 9.04704643113319e-05, | |
| "loss": 2.4366, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 0.24096588279836775, | |
| "grad_norm": 0.8360730409622192, | |
| "learning_rate": 9.040501221636472e-05, | |
| "loss": 2.4807, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.24163897185646377, | |
| "grad_norm": 0.6942176818847656, | |
| "learning_rate": 9.033935996335573e-05, | |
| "loss": 2.4435, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 0.24231206091455976, | |
| "grad_norm": 0.6470984220504761, | |
| "learning_rate": 9.027350787753257e-05, | |
| "loss": 2.4413, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.24298514997265575, | |
| "grad_norm": 0.6180524826049805, | |
| "learning_rate": 9.020745628511281e-05, | |
| "loss": 2.4708, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 0.24365823903075176, | |
| "grad_norm": 0.5775427222251892, | |
| "learning_rate": 9.014120551330232e-05, | |
| "loss": 2.4242, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.24433132808884775, | |
| "grad_norm": 0.4921644628047943, | |
| "learning_rate": 9.007475589029365e-05, | |
| "loss": 2.4759, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 0.24500441714694376, | |
| "grad_norm": 0.6223775148391724, | |
| "learning_rate": 9.000810774526448e-05, | |
| "loss": 2.4407, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.24567750620503975, | |
| "grad_norm": 0.5503790974617004, | |
| "learning_rate": 8.994126140837585e-05, | |
| "loss": 2.4631, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.24635059526313574, | |
| "grad_norm": 0.5364406704902649, | |
| "learning_rate": 8.987421721077062e-05, | |
| "loss": 2.4698, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.24702368432123176, | |
| "grad_norm": 0.5078843235969543, | |
| "learning_rate": 8.980697548457186e-05, | |
| "loss": 2.4727, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 0.24769677337932774, | |
| "grad_norm": 0.49923375248908997, | |
| "learning_rate": 8.973953656288112e-05, | |
| "loss": 2.4305, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.24836986243742376, | |
| "grad_norm": 0.5955345630645752, | |
| "learning_rate": 8.967190077977681e-05, | |
| "loss": 2.4578, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 0.24904295149551975, | |
| "grad_norm": 0.5475841760635376, | |
| "learning_rate": 8.960406847031253e-05, | |
| "loss": 2.4187, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.24971604055361574, | |
| "grad_norm": 0.5308858156204224, | |
| "learning_rate": 8.953603997051548e-05, | |
| "loss": 2.4664, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 0.2503891296117117, | |
| "grad_norm": 0.5813267827033997, | |
| "learning_rate": 8.946781561738474e-05, | |
| "loss": 2.4627, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.25106221866980777, | |
| "grad_norm": 0.5371124744415283, | |
| "learning_rate": 8.939939574888957e-05, | |
| "loss": 2.4528, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 0.25173530772790376, | |
| "grad_norm": 0.6726812720298767, | |
| "learning_rate": 8.933078070396778e-05, | |
| "loss": 2.4286, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.25240839678599974, | |
| "grad_norm": 0.5831666588783264, | |
| "learning_rate": 8.926197082252411e-05, | |
| "loss": 2.4551, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.25308148584409573, | |
| "grad_norm": 0.5222828388214111, | |
| "learning_rate": 8.919296644542837e-05, | |
| "loss": 2.4725, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.2537545749021917, | |
| "grad_norm": 0.5731188058853149, | |
| "learning_rate": 8.912376791451395e-05, | |
| "loss": 2.4416, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 0.25442766396028776, | |
| "grad_norm": 0.7226285338401794, | |
| "learning_rate": 8.905437557257599e-05, | |
| "loss": 2.4516, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.25510075301838375, | |
| "grad_norm": 0.683601975440979, | |
| "learning_rate": 8.898478976336975e-05, | |
| "loss": 2.4514, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 0.25577384207647974, | |
| "grad_norm": 0.623328685760498, | |
| "learning_rate": 8.89150108316089e-05, | |
| "loss": 2.4264, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.25644693113457573, | |
| "grad_norm": 0.5894498825073242, | |
| "learning_rate": 8.884503912296373e-05, | |
| "loss": 2.4572, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 0.2571200201926717, | |
| "grad_norm": 0.5245216488838196, | |
| "learning_rate": 8.877487498405961e-05, | |
| "loss": 2.4725, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.25779310925076776, | |
| "grad_norm": 0.5440263152122498, | |
| "learning_rate": 8.87045187624751e-05, | |
| "loss": 2.4494, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 0.25846619830886375, | |
| "grad_norm": 0.6193727850914001, | |
| "learning_rate": 8.86339708067403e-05, | |
| "loss": 2.4392, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.25913928736695974, | |
| "grad_norm": 0.5997888445854187, | |
| "learning_rate": 8.856323146633517e-05, | |
| "loss": 2.4388, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.2598123764250557, | |
| "grad_norm": 0.7109591960906982, | |
| "learning_rate": 8.849230109168767e-05, | |
| "loss": 2.409, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.2604854654831517, | |
| "grad_norm": 0.5998343825340271, | |
| "learning_rate": 8.842118003417218e-05, | |
| "loss": 2.4443, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 0.26115855454124776, | |
| "grad_norm": 0.5481546521186829, | |
| "learning_rate": 8.834986864610764e-05, | |
| "loss": 2.4334, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.26183164359934374, | |
| "grad_norm": 0.5310930013656616, | |
| "learning_rate": 8.827836728075588e-05, | |
| "loss": 2.4424, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 0.26250473265743973, | |
| "grad_norm": 0.49727270007133484, | |
| "learning_rate": 8.82066762923198e-05, | |
| "loss": 2.4348, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.2631778217155357, | |
| "grad_norm": 0.6515977382659912, | |
| "learning_rate": 8.813479603594168e-05, | |
| "loss": 2.4454, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 0.26385091077363176, | |
| "grad_norm": 0.7124093174934387, | |
| "learning_rate": 8.806272686770139e-05, | |
| "loss": 2.4265, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.26452399983172775, | |
| "grad_norm": 0.5826483964920044, | |
| "learning_rate": 8.799046914461461e-05, | |
| "loss": 2.4736, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 0.26519708888982374, | |
| "grad_norm": 0.5905554294586182, | |
| "learning_rate": 8.791802322463114e-05, | |
| "loss": 2.4467, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.26587017794791973, | |
| "grad_norm": 0.5929023027420044, | |
| "learning_rate": 8.784538946663297e-05, | |
| "loss": 2.4288, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.2665432670060157, | |
| "grad_norm": 0.5478795170783997, | |
| "learning_rate": 8.777256823043269e-05, | |
| "loss": 2.4387, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.26721635606411176, | |
| "grad_norm": 0.5086268782615662, | |
| "learning_rate": 8.769955987677159e-05, | |
| "loss": 2.4384, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 0.26788944512220775, | |
| "grad_norm": 0.6146414875984192, | |
| "learning_rate": 8.762636476731786e-05, | |
| "loss": 2.432, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 0.26856253418030374, | |
| "grad_norm": 0.5799776315689087, | |
| "learning_rate": 8.755298326466495e-05, | |
| "loss": 2.4305, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 0.2692356232383997, | |
| "grad_norm": 0.47361868619918823, | |
| "learning_rate": 8.747941573232951e-05, | |
| "loss": 2.4397, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.2699087122964957, | |
| "grad_norm": 0.4643039405345917, | |
| "learning_rate": 8.740566253474985e-05, | |
| "loss": 2.4515, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 0.27058180135459176, | |
| "grad_norm": 0.4562658369541168, | |
| "learning_rate": 8.733172403728401e-05, | |
| "loss": 2.4386, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 0.27125489041268774, | |
| "grad_norm": 0.5169376730918884, | |
| "learning_rate": 8.725760060620795e-05, | |
| "loss": 2.4331, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 0.27192797947078373, | |
| "grad_norm": 0.5017179250717163, | |
| "learning_rate": 8.718329260871374e-05, | |
| "loss": 2.4203, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 0.2726010685288797, | |
| "grad_norm": 0.5720140337944031, | |
| "learning_rate": 8.71088004129078e-05, | |
| "loss": 2.4548, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.2732741575869757, | |
| "grad_norm": 0.5228462219238281, | |
| "learning_rate": 8.703412438780898e-05, | |
| "loss": 2.3827, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 0.27394724664507175, | |
| "grad_norm": 0.4759220778942108, | |
| "learning_rate": 8.695926490334682e-05, | |
| "loss": 2.4321, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 0.27462033570316774, | |
| "grad_norm": 0.532883882522583, | |
| "learning_rate": 8.688422233035967e-05, | |
| "loss": 2.3892, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 0.2752934247612637, | |
| "grad_norm": 0.480090469121933, | |
| "learning_rate": 8.680899704059283e-05, | |
| "loss": 2.4134, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 0.2759665138193597, | |
| "grad_norm": 0.5285961627960205, | |
| "learning_rate": 8.673358940669679e-05, | |
| "loss": 2.448, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.2766396028774557, | |
| "grad_norm": 0.49371105432510376, | |
| "learning_rate": 8.665799980222528e-05, | |
| "loss": 2.4465, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 0.27731269193555175, | |
| "grad_norm": 0.4970341920852661, | |
| "learning_rate": 8.658222860163356e-05, | |
| "loss": 2.4171, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 0.27798578099364774, | |
| "grad_norm": 0.5542078614234924, | |
| "learning_rate": 8.650627618027638e-05, | |
| "loss": 2.3954, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 0.2786588700517437, | |
| "grad_norm": 0.506982684135437, | |
| "learning_rate": 8.643014291440629e-05, | |
| "loss": 2.4383, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.2793319591098397, | |
| "grad_norm": 0.5749984979629517, | |
| "learning_rate": 8.635382918117167e-05, | |
| "loss": 2.3878, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.2800050481679357, | |
| "grad_norm": 0.5519755482673645, | |
| "learning_rate": 8.627733535861493e-05, | |
| "loss": 2.4304, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.28067813722603174, | |
| "grad_norm": 0.6592808365821838, | |
| "learning_rate": 8.620066182567057e-05, | |
| "loss": 2.3806, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 0.28135122628412773, | |
| "grad_norm": 0.7210227847099304, | |
| "learning_rate": 8.612380896216336e-05, | |
| "loss": 2.4454, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 0.2820243153422237, | |
| "grad_norm": 0.6003880500793457, | |
| "learning_rate": 8.604677714880642e-05, | |
| "loss": 2.4387, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 0.2826974044003197, | |
| "grad_norm": 0.5979674458503723, | |
| "learning_rate": 8.596956676719936e-05, | |
| "loss": 2.4231, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.2833704934584157, | |
| "grad_norm": 0.6536392569541931, | |
| "learning_rate": 8.58921781998264e-05, | |
| "loss": 2.4038, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 0.28404358251651174, | |
| "grad_norm": 0.7131980657577515, | |
| "learning_rate": 8.581461183005441e-05, | |
| "loss": 2.3951, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 0.2847166715746077, | |
| "grad_norm": 0.7326920032501221, | |
| "learning_rate": 8.57368680421311e-05, | |
| "loss": 2.4102, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 0.2853897606327037, | |
| "grad_norm": 0.6664044260978699, | |
| "learning_rate": 8.565894722118307e-05, | |
| "loss": 2.3962, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 0.2860628496907997, | |
| "grad_norm": 0.5486345291137695, | |
| "learning_rate": 8.558084975321384e-05, | |
| "loss": 2.4255, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.2867359387488957, | |
| "grad_norm": 0.5411694049835205, | |
| "learning_rate": 8.55025760251021e-05, | |
| "loss": 2.4395, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 0.28740902780699173, | |
| "grad_norm": 0.490962415933609, | |
| "learning_rate": 8.542412642459963e-05, | |
| "loss": 2.3971, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 0.2880821168650877, | |
| "grad_norm": 0.5245395302772522, | |
| "learning_rate": 8.534550134032944e-05, | |
| "loss": 2.3977, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 0.2887552059231837, | |
| "grad_norm": 0.5046750903129578, | |
| "learning_rate": 8.526670116178391e-05, | |
| "loss": 2.3787, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 0.2894282949812797, | |
| "grad_norm": 0.5423001646995544, | |
| "learning_rate": 8.518772627932276e-05, | |
| "loss": 2.4143, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.2901013840393757, | |
| "grad_norm": 0.5712574124336243, | |
| "learning_rate": 8.510857708417114e-05, | |
| "loss": 2.4738, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 0.29077447309747173, | |
| "grad_norm": 0.6638187766075134, | |
| "learning_rate": 8.502925396841775e-05, | |
| "loss": 2.3985, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.2914475621555677, | |
| "grad_norm": 0.5826054215431213, | |
| "learning_rate": 8.494975732501282e-05, | |
| "loss": 2.4162, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 0.2921206512136637, | |
| "grad_norm": 0.5681880712509155, | |
| "learning_rate": 8.487008754776622e-05, | |
| "loss": 2.407, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 0.2927937402717597, | |
| "grad_norm": 0.5738682746887207, | |
| "learning_rate": 8.47902450313455e-05, | |
| "loss": 2.4274, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.2934668293298557, | |
| "grad_norm": 0.6157481670379639, | |
| "learning_rate": 8.47102301712739e-05, | |
| "loss": 2.4233, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 0.2941399183879517, | |
| "grad_norm": 0.6142215728759766, | |
| "learning_rate": 8.463004336392842e-05, | |
| "loss": 2.4127, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 0.2948130074460477, | |
| "grad_norm": 0.5124562978744507, | |
| "learning_rate": 8.454968500653787e-05, | |
| "loss": 2.4232, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 0.2954860965041437, | |
| "grad_norm": 0.6266094446182251, | |
| "learning_rate": 8.446915549718085e-05, | |
| "loss": 2.4318, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 0.2961591855622397, | |
| "grad_norm": 0.5722962021827698, | |
| "learning_rate": 8.438845523478385e-05, | |
| "loss": 2.4324, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.2968322746203357, | |
| "grad_norm": 0.5898362994194031, | |
| "learning_rate": 8.430758461911919e-05, | |
| "loss": 2.4183, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 0.2975053636784317, | |
| "grad_norm": 0.5322403907775879, | |
| "learning_rate": 8.422654405080315e-05, | |
| "loss": 2.4215, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 0.2981784527365277, | |
| "grad_norm": 0.5670167207717896, | |
| "learning_rate": 8.414533393129383e-05, | |
| "loss": 2.3962, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 0.2988515417946237, | |
| "grad_norm": 0.5440824031829834, | |
| "learning_rate": 8.406395466288935e-05, | |
| "loss": 2.4166, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 0.2995246308527197, | |
| "grad_norm": 0.534233570098877, | |
| "learning_rate": 8.398240664872566e-05, | |
| "loss": 2.4417, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.3001977199108157, | |
| "grad_norm": 0.5321791172027588, | |
| "learning_rate": 8.390069029277474e-05, | |
| "loss": 2.3973, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 0.3008708089689117, | |
| "grad_norm": 0.526823878288269, | |
| "learning_rate": 8.381880599984242e-05, | |
| "loss": 2.4064, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 0.3015438980270077, | |
| "grad_norm": 0.5071905255317688, | |
| "learning_rate": 8.37367541755665e-05, | |
| "loss": 2.4135, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.3022169870851037, | |
| "grad_norm": 0.5760087966918945, | |
| "learning_rate": 8.365453522641467e-05, | |
| "loss": 2.3872, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 0.3028900761431997, | |
| "grad_norm": 0.5876712799072266, | |
| "learning_rate": 8.357214955968256e-05, | |
| "loss": 2.4155, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.30356316520129567, | |
| "grad_norm": 0.5806450247764587, | |
| "learning_rate": 8.348959758349161e-05, | |
| "loss": 2.3914, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 0.3042362542593917, | |
| "grad_norm": 0.727473258972168, | |
| "learning_rate": 8.340687970678724e-05, | |
| "loss": 2.4184, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 0.3049093433174877, | |
| "grad_norm": 0.6061837077140808, | |
| "learning_rate": 8.332399633933658e-05, | |
| "loss": 2.4554, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 0.3055824323755837, | |
| "grad_norm": 0.5309655070304871, | |
| "learning_rate": 8.324094789172667e-05, | |
| "loss": 2.3899, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 0.3062555214336797, | |
| "grad_norm": 0.5148584246635437, | |
| "learning_rate": 8.315773477536222e-05, | |
| "loss": 2.3643, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.30692861049177567, | |
| "grad_norm": 0.5103538036346436, | |
| "learning_rate": 8.307435740246378e-05, | |
| "loss": 2.4013, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 0.3076016995498717, | |
| "grad_norm": 0.5046418309211731, | |
| "learning_rate": 8.299081618606553e-05, | |
| "loss": 2.3934, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 0.3082747886079677, | |
| "grad_norm": 0.5081542134284973, | |
| "learning_rate": 8.29071115400133e-05, | |
| "loss": 2.3995, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 0.3089478776660637, | |
| "grad_norm": 0.49463972449302673, | |
| "learning_rate": 8.282324387896255e-05, | |
| "loss": 2.3822, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 0.3096209667241597, | |
| "grad_norm": 0.6341468691825867, | |
| "learning_rate": 8.273921361837622e-05, | |
| "loss": 2.3997, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.31029405578225566, | |
| "grad_norm": 0.59495609998703, | |
| "learning_rate": 8.265502117452284e-05, | |
| "loss": 2.4188, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 0.3109671448403517, | |
| "grad_norm": 0.5837931632995605, | |
| "learning_rate": 8.257066696447424e-05, | |
| "loss": 2.4005, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 0.3116402338984477, | |
| "grad_norm": 0.5928687453269958, | |
| "learning_rate": 8.24861514061037e-05, | |
| "loss": 2.3863, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 0.3123133229565437, | |
| "grad_norm": 0.5589261651039124, | |
| "learning_rate": 8.240147491808373e-05, | |
| "loss": 2.3889, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.31298641201463967, | |
| "grad_norm": 0.5119801163673401, | |
| "learning_rate": 8.231663791988406e-05, | |
| "loss": 2.4215, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.3136595010727357, | |
| "grad_norm": 0.5372598767280579, | |
| "learning_rate": 8.22316408317696e-05, | |
| "loss": 2.458, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 0.3143325901308317, | |
| "grad_norm": 0.5366113781929016, | |
| "learning_rate": 8.214648407479824e-05, | |
| "loss": 2.374, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 0.3150056791889277, | |
| "grad_norm": 0.5240657329559326, | |
| "learning_rate": 8.206116807081891e-05, | |
| "loss": 2.405, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 0.3156787682470237, | |
| "grad_norm": 0.5178209543228149, | |
| "learning_rate": 8.197569324246937e-05, | |
| "loss": 2.3846, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 0.31635185730511967, | |
| "grad_norm": 0.5788430571556091, | |
| "learning_rate": 8.189006001317413e-05, | |
| "loss": 2.3799, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.3170249463632157, | |
| "grad_norm": 0.508414089679718, | |
| "learning_rate": 8.180426880714251e-05, | |
| "loss": 2.4125, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 0.3176980354213117, | |
| "grad_norm": 0.47710898518562317, | |
| "learning_rate": 8.171832004936627e-05, | |
| "loss": 2.3953, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 0.3183711244794077, | |
| "grad_norm": 0.5494322776794434, | |
| "learning_rate": 8.163221416561773e-05, | |
| "loss": 2.377, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 0.3190442135375037, | |
| "grad_norm": 0.5088549256324768, | |
| "learning_rate": 8.154595158244757e-05, | |
| "loss": 2.3937, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 0.31971730259559966, | |
| "grad_norm": 0.6372072100639343, | |
| "learning_rate": 8.14595327271827e-05, | |
| "loss": 2.3641, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.3203903916536957, | |
| "grad_norm": 0.5864402055740356, | |
| "learning_rate": 8.13729580279242e-05, | |
| "loss": 2.3743, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 0.3210634807117917, | |
| "grad_norm": 0.569735586643219, | |
| "learning_rate": 8.128622791354515e-05, | |
| "loss": 2.4052, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 0.3217365697698877, | |
| "grad_norm": 0.5160011649131775, | |
| "learning_rate": 8.119934281368855e-05, | |
| "loss": 2.3811, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 0.32240965882798367, | |
| "grad_norm": 0.5426251888275146, | |
| "learning_rate": 8.11123031587651e-05, | |
| "loss": 2.4342, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 0.32308274788607966, | |
| "grad_norm": 0.45735329389572144, | |
| "learning_rate": 8.102510937995124e-05, | |
| "loss": 2.3941, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.3237558369441757, | |
| "grad_norm": 0.5503259897232056, | |
| "learning_rate": 8.09377619091868e-05, | |
| "loss": 2.4441, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 0.3244289260022717, | |
| "grad_norm": 0.46749740839004517, | |
| "learning_rate": 8.085026117917303e-05, | |
| "loss": 2.4252, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 0.3251020150603677, | |
| "grad_norm": 0.42537739872932434, | |
| "learning_rate": 8.076260762337038e-05, | |
| "loss": 2.3907, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 0.32577510411846367, | |
| "grad_norm": 0.4988864064216614, | |
| "learning_rate": 8.067480167599637e-05, | |
| "loss": 2.4038, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 0.32644819317655965, | |
| "grad_norm": 0.5707867741584778, | |
| "learning_rate": 8.058684377202343e-05, | |
| "loss": 2.4083, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.3271212822346557, | |
| "grad_norm": 0.48750847578048706, | |
| "learning_rate": 8.049873434717676e-05, | |
| "loss": 2.4191, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 0.3277943712927517, | |
| "grad_norm": 0.4584735631942749, | |
| "learning_rate": 8.041047383793216e-05, | |
| "loss": 2.391, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 0.3284674603508477, | |
| "grad_norm": 0.501753568649292, | |
| "learning_rate": 8.032206268151385e-05, | |
| "loss": 2.385, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 0.32914054940894366, | |
| "grad_norm": 0.45828142762184143, | |
| "learning_rate": 8.023350131589238e-05, | |
| "loss": 2.3861, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 0.32981363846703965, | |
| "grad_norm": 0.4599551260471344, | |
| "learning_rate": 8.014479017978232e-05, | |
| "loss": 2.3826, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.3304867275251357, | |
| "grad_norm": 0.4873054623603821, | |
| "learning_rate": 8.005592971264027e-05, | |
| "loss": 2.4007, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 0.3311598165832317, | |
| "grad_norm": 0.5136446356773376, | |
| "learning_rate": 7.996692035466249e-05, | |
| "loss": 2.4101, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 0.33183290564132767, | |
| "grad_norm": 0.46801379323005676, | |
| "learning_rate": 7.987776254678287e-05, | |
| "loss": 2.3346, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 0.33250599469942366, | |
| "grad_norm": 0.5094547867774963, | |
| "learning_rate": 7.978845673067073e-05, | |
| "loss": 2.4069, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 0.33317908375751965, | |
| "grad_norm": 0.4927618205547333, | |
| "learning_rate": 7.969900334872846e-05, | |
| "loss": 2.351, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.3338521728156157, | |
| "grad_norm": 0.5567794442176819, | |
| "learning_rate": 7.960940284408962e-05, | |
| "loss": 2.3749, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 0.3345252618737117, | |
| "grad_norm": 0.45481762290000916, | |
| "learning_rate": 7.951965566061645e-05, | |
| "loss": 2.3757, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 0.33519835093180766, | |
| "grad_norm": 0.5935695767402649, | |
| "learning_rate": 7.942976224289788e-05, | |
| "loss": 2.3861, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 0.33587143998990365, | |
| "grad_norm": 0.486787348985672, | |
| "learning_rate": 7.933972303624726e-05, | |
| "loss": 2.386, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 0.33654452904799964, | |
| "grad_norm": 0.5005300045013428, | |
| "learning_rate": 7.924953848670011e-05, | |
| "loss": 2.3674, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.3372176181060957, | |
| "grad_norm": 0.4527299404144287, | |
| "learning_rate": 7.915920904101195e-05, | |
| "loss": 2.3521, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 0.3378907071641917, | |
| "grad_norm": 0.5433601140975952, | |
| "learning_rate": 7.906873514665612e-05, | |
| "loss": 2.3559, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 0.33856379622228766, | |
| "grad_norm": 0.47732681035995483, | |
| "learning_rate": 7.897811725182152e-05, | |
| "loss": 2.3874, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 0.33923688528038365, | |
| "grad_norm": 0.5074180960655212, | |
| "learning_rate": 7.888735580541034e-05, | |
| "loss": 2.3631, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 0.33990997433847964, | |
| "grad_norm": 0.47182127833366394, | |
| "learning_rate": 7.879645125703598e-05, | |
| "loss": 2.4256, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.3405830633965757, | |
| "grad_norm": 0.49758678674697876, | |
| "learning_rate": 7.870540405702071e-05, | |
| "loss": 2.3823, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 0.34125615245467167, | |
| "grad_norm": 0.4703919589519501, | |
| "learning_rate": 7.861421465639344e-05, | |
| "loss": 2.3857, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 0.34192924151276766, | |
| "grad_norm": 0.46003982424736023, | |
| "learning_rate": 7.852288350688753e-05, | |
| "loss": 2.4085, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 0.34260233057086364, | |
| "grad_norm": 0.48561301827430725, | |
| "learning_rate": 7.84314110609385e-05, | |
| "loss": 2.3922, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 0.34327541962895963, | |
| "grad_norm": 0.5172334909439087, | |
| "learning_rate": 7.83397977716819e-05, | |
| "loss": 2.3895, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.3439485086870557, | |
| "grad_norm": 0.5409645438194275, | |
| "learning_rate": 7.824804409295093e-05, | |
| "loss": 2.3892, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 0.34462159774515166, | |
| "grad_norm": 0.6464726328849792, | |
| "learning_rate": 7.815615047927427e-05, | |
| "loss": 2.3765, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 0.34529468680324765, | |
| "grad_norm": 0.58461993932724, | |
| "learning_rate": 7.80641173858738e-05, | |
| "loss": 2.384, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 0.34596777586134364, | |
| "grad_norm": 0.5537058711051941, | |
| "learning_rate": 7.797194526866234e-05, | |
| "loss": 2.3723, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 0.34664086491943963, | |
| "grad_norm": 0.48641249537467957, | |
| "learning_rate": 7.787963458424146e-05, | |
| "loss": 2.4044, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.34731395397753567, | |
| "grad_norm": 0.498935341835022, | |
| "learning_rate": 7.77871857898991e-05, | |
| "loss": 2.3665, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 0.34798704303563166, | |
| "grad_norm": 0.4734966456890106, | |
| "learning_rate": 7.769459934360738e-05, | |
| "loss": 2.3664, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 0.34866013209372765, | |
| "grad_norm": 0.5010454654693604, | |
| "learning_rate": 7.760187570402036e-05, | |
| "loss": 2.3691, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 0.34933322115182364, | |
| "grad_norm": 0.5199081897735596, | |
| "learning_rate": 7.750901533047168e-05, | |
| "loss": 2.3806, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 0.3500063102099196, | |
| "grad_norm": 0.4784662127494812, | |
| "learning_rate": 7.741601868297237e-05, | |
| "loss": 2.3737, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.35067939926801567, | |
| "grad_norm": 0.4382856786251068, | |
| "learning_rate": 7.732288622220851e-05, | |
| "loss": 2.3694, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 0.35135248832611166, | |
| "grad_norm": 0.44757765531539917, | |
| "learning_rate": 7.722961840953897e-05, | |
| "loss": 2.3608, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 0.35202557738420764, | |
| "grad_norm": 0.4724947512149811, | |
| "learning_rate": 7.713621570699311e-05, | |
| "loss": 2.3604, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 0.35269866644230363, | |
| "grad_norm": 0.44639867544174194, | |
| "learning_rate": 7.704267857726855e-05, | |
| "loss": 2.369, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 0.3533717555003996, | |
| "grad_norm": 0.4407545030117035, | |
| "learning_rate": 7.694900748372882e-05, | |
| "loss": 2.3792, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.35404484455849566, | |
| "grad_norm": 0.48225051164627075, | |
| "learning_rate": 7.685520289040105e-05, | |
| "loss": 2.3555, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 0.35471793361659165, | |
| "grad_norm": 0.4579322040081024, | |
| "learning_rate": 7.676126526197374e-05, | |
| "loss": 2.4021, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 0.35539102267468764, | |
| "grad_norm": 0.4263804852962494, | |
| "learning_rate": 7.666719506379436e-05, | |
| "loss": 2.4141, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 0.35606411173278363, | |
| "grad_norm": 0.5071961283683777, | |
| "learning_rate": 7.657299276186718e-05, | |
| "loss": 2.3596, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 0.3567372007908796, | |
| "grad_norm": 0.4433095157146454, | |
| "learning_rate": 7.647865882285083e-05, | |
| "loss": 2.3768, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.35741028984897566, | |
| "grad_norm": 0.5640519261360168, | |
| "learning_rate": 7.638419371405606e-05, | |
| "loss": 2.3612, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 0.35808337890707165, | |
| "grad_norm": 0.47830817103385925, | |
| "learning_rate": 7.62895979034434e-05, | |
| "loss": 2.3591, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 0.35875646796516764, | |
| "grad_norm": 0.577599048614502, | |
| "learning_rate": 7.619487185962087e-05, | |
| "loss": 2.3719, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 0.3594295570232636, | |
| "grad_norm": 0.5717925429344177, | |
| "learning_rate": 7.610001605184163e-05, | |
| "loss": 2.378, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 0.36010264608135967, | |
| "grad_norm": 0.5529263019561768, | |
| "learning_rate": 7.600503095000164e-05, | |
| "loss": 2.3936, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.36077573513945566, | |
| "grad_norm": 0.49978116154670715, | |
| "learning_rate": 7.590991702463737e-05, | |
| "loss": 2.3926, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 0.36144882419755164, | |
| "grad_norm": 0.5822058916091919, | |
| "learning_rate": 7.581467474692347e-05, | |
| "loss": 2.3602, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 0.36212191325564763, | |
| "grad_norm": 0.526158332824707, | |
| "learning_rate": 7.571930458867043e-05, | |
| "loss": 2.3559, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 0.3627950023137436, | |
| "grad_norm": 0.4761981964111328, | |
| "learning_rate": 7.562380702232217e-05, | |
| "loss": 2.3603, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 0.36346809137183966, | |
| "grad_norm": 0.4922941029071808, | |
| "learning_rate": 7.552818252095384e-05, | |
| "loss": 2.3626, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.36414118042993565, | |
| "grad_norm": 0.426907479763031, | |
| "learning_rate": 7.543243155826935e-05, | |
| "loss": 2.4049, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 0.36481426948803164, | |
| "grad_norm": 0.4346926510334015, | |
| "learning_rate": 7.53365546085991e-05, | |
| "loss": 2.344, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 0.3654873585461276, | |
| "grad_norm": 0.46861520409584045, | |
| "learning_rate": 7.524055214689758e-05, | |
| "loss": 2.3639, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 0.3661604476042236, | |
| "grad_norm": 0.46749797463417053, | |
| "learning_rate": 7.514442464874107e-05, | |
| "loss": 2.408, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 0.36683353666231966, | |
| "grad_norm": 0.4398956298828125, | |
| "learning_rate": 7.504817259032524e-05, | |
| "loss": 2.3865, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.36750662572041565, | |
| "grad_norm": 0.5109989047050476, | |
| "learning_rate": 7.495179644846279e-05, | |
| "loss": 2.397, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 0.36817971477851164, | |
| "grad_norm": 0.47633254528045654, | |
| "learning_rate": 7.485529670058112e-05, | |
| "loss": 2.3514, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 0.3688528038366076, | |
| "grad_norm": 0.5054311752319336, | |
| "learning_rate": 7.475867382471994e-05, | |
| "loss": 2.3754, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 0.3695258928947036, | |
| "grad_norm": 0.5074468851089478, | |
| "learning_rate": 7.466192829952894e-05, | |
| "loss": 2.3566, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 0.37019898195279966, | |
| "grad_norm": 0.4602699875831604, | |
| "learning_rate": 7.456506060426534e-05, | |
| "loss": 2.3073, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.37087207101089564, | |
| "grad_norm": 0.5205585360527039, | |
| "learning_rate": 7.44680712187916e-05, | |
| "loss": 2.3686, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 0.37154516006899163, | |
| "grad_norm": 0.4438835084438324, | |
| "learning_rate": 7.437096062357303e-05, | |
| "loss": 2.3355, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 0.3722182491270876, | |
| "grad_norm": 0.5131971836090088, | |
| "learning_rate": 7.42737292996753e-05, | |
| "loss": 2.3709, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 0.3728913381851836, | |
| "grad_norm": 0.4649304449558258, | |
| "learning_rate": 7.417637772876224e-05, | |
| "loss": 2.4148, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 0.37356442724327965, | |
| "grad_norm": 0.4761337637901306, | |
| "learning_rate": 7.407890639309329e-05, | |
| "loss": 2.3689, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.37423751630137564, | |
| "grad_norm": 0.47608640789985657, | |
| "learning_rate": 7.398131577552126e-05, | |
| "loss": 2.3554, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 0.3749106053594716, | |
| "grad_norm": 0.45111626386642456, | |
| "learning_rate": 7.388360635948977e-05, | |
| "loss": 2.3621, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 0.3755836944175676, | |
| "grad_norm": 0.4884491562843323, | |
| "learning_rate": 7.378577862903099e-05, | |
| "loss": 2.3873, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 0.3762567834756636, | |
| "grad_norm": 0.5153593420982361, | |
| "learning_rate": 7.368783306876318e-05, | |
| "loss": 2.3955, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 0.37692987253375965, | |
| "grad_norm": 0.4825320541858673, | |
| "learning_rate": 7.358977016388833e-05, | |
| "loss": 2.3099, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.37760296159185563, | |
| "grad_norm": 0.48556360602378845, | |
| "learning_rate": 7.349159040018968e-05, | |
| "loss": 2.3344, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 0.3782760506499516, | |
| "grad_norm": 0.4832114279270172, | |
| "learning_rate": 7.339329426402944e-05, | |
| "loss": 2.3779, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 0.3789491397080476, | |
| "grad_norm": 0.5078182220458984, | |
| "learning_rate": 7.32948822423462e-05, | |
| "loss": 2.3547, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 0.3796222287661436, | |
| "grad_norm": 0.5177402496337891, | |
| "learning_rate": 7.319635482265274e-05, | |
| "loss": 2.3323, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 0.38029531782423964, | |
| "grad_norm": 0.4716585874557495, | |
| "learning_rate": 7.309771249303343e-05, | |
| "loss": 2.3569, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.38096840688233563, | |
| "grad_norm": 0.49241212010383606, | |
| "learning_rate": 7.299895574214186e-05, | |
| "loss": 2.345, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 0.3816414959404316, | |
| "grad_norm": 0.49419349431991577, | |
| "learning_rate": 7.290008505919849e-05, | |
| "loss": 2.329, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 0.3823145849985276, | |
| "grad_norm": 0.45170944929122925, | |
| "learning_rate": 7.280110093398817e-05, | |
| "loss": 2.3797, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 0.3829876740566236, | |
| "grad_norm": 0.49259769916534424, | |
| "learning_rate": 7.270200385685769e-05, | |
| "loss": 2.4009, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 0.38366076311471964, | |
| "grad_norm": 0.5392531752586365, | |
| "learning_rate": 7.260279431871339e-05, | |
| "loss": 2.3521, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.3843338521728156, | |
| "grad_norm": 0.48526453971862793, | |
| "learning_rate": 7.250347281101875e-05, | |
| "loss": 2.368, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 0.3850069412309116, | |
| "grad_norm": 0.4813965857028961, | |
| "learning_rate": 7.240403982579187e-05, | |
| "loss": 2.3605, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 0.3856800302890076, | |
| "grad_norm": 0.553960382938385, | |
| "learning_rate": 7.230449585560312e-05, | |
| "loss": 2.3471, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 0.3863531193471036, | |
| "grad_norm": 0.46131259202957153, | |
| "learning_rate": 7.220484139357267e-05, | |
| "loss": 2.3225, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 0.38702620840519963, | |
| "grad_norm": 0.4981285631656647, | |
| "learning_rate": 7.210507693336806e-05, | |
| "loss": 2.3307, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.3876992974632956, | |
| "grad_norm": 0.4464927911758423, | |
| "learning_rate": 7.200520296920164e-05, | |
| "loss": 2.3736, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 0.3883723865213916, | |
| "grad_norm": 0.5023694038391113, | |
| "learning_rate": 7.190521999582838e-05, | |
| "loss": 2.3258, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 0.3890454755794876, | |
| "grad_norm": 0.4526605010032654, | |
| "learning_rate": 7.18051285085431e-05, | |
| "loss": 2.387, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 0.3897185646375836, | |
| "grad_norm": 0.5418564677238464, | |
| "learning_rate": 7.170492900317832e-05, | |
| "loss": 2.3953, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 0.39039165369567963, | |
| "grad_norm": 0.440750390291214, | |
| "learning_rate": 7.160462197610151e-05, | |
| "loss": 2.3617, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.3910647427537756, | |
| "grad_norm": 0.48780837655067444, | |
| "learning_rate": 7.150420792421291e-05, | |
| "loss": 2.3312, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 0.3917378318118716, | |
| "grad_norm": 0.6161893606185913, | |
| "learning_rate": 7.140368734494286e-05, | |
| "loss": 2.3664, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 0.3924109208699676, | |
| "grad_norm": 0.6196706295013428, | |
| "learning_rate": 7.130306073624944e-05, | |
| "loss": 2.3282, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 0.3930840099280636, | |
| "grad_norm": 0.510915994644165, | |
| "learning_rate": 7.1202328596616e-05, | |
| "loss": 2.3616, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 0.3937570989861596, | |
| "grad_norm": 0.4860905408859253, | |
| "learning_rate": 7.110149142504863e-05, | |
| "loss": 2.3438, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.3944301880442556, | |
| "grad_norm": 0.5084627866744995, | |
| "learning_rate": 7.100054972107373e-05, | |
| "loss": 2.3328, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 0.3951032771023516, | |
| "grad_norm": 0.5457413792610168, | |
| "learning_rate": 7.089950398473558e-05, | |
| "loss": 2.3528, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 0.3957763661604476, | |
| "grad_norm": 0.5079280734062195, | |
| "learning_rate": 7.079835471659375e-05, | |
| "loss": 2.3451, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 0.3964494552185436, | |
| "grad_norm": 0.5098846554756165, | |
| "learning_rate": 7.06971024177207e-05, | |
| "loss": 2.3651, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 0.3971225442766396, | |
| "grad_norm": 0.4790203869342804, | |
| "learning_rate": 7.059574758969934e-05, | |
| "loss": 2.3741, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.3977956333347356, | |
| "grad_norm": 0.5449979901313782, | |
| "learning_rate": 7.049429073462043e-05, | |
| "loss": 2.3693, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 0.3984687223928316, | |
| "grad_norm": 0.4455925524234772, | |
| "learning_rate": 7.039273235508014e-05, | |
| "loss": 2.3259, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 0.3991418114509276, | |
| "grad_norm": 0.4840555787086487, | |
| "learning_rate": 7.029107295417763e-05, | |
| "loss": 2.3527, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 0.3998149005090236, | |
| "grad_norm": 0.6292969584465027, | |
| "learning_rate": 7.018931303551246e-05, | |
| "loss": 2.3742, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 0.4004879895671196, | |
| "grad_norm": 0.5658779144287109, | |
| "learning_rate": 7.008745310318213e-05, | |
| "loss": 2.3608, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.4011610786252156, | |
| "grad_norm": 0.541577160358429, | |
| "learning_rate": 6.998549366177959e-05, | |
| "loss": 2.3359, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 0.4018341676833116, | |
| "grad_norm": 0.48381462693214417, | |
| "learning_rate": 6.988343521639077e-05, | |
| "loss": 2.3626, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 0.4025072567414076, | |
| "grad_norm": 0.5662329196929932, | |
| "learning_rate": 6.978127827259198e-05, | |
| "loss": 2.3511, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 0.40318034579950357, | |
| "grad_norm": 0.5031713843345642, | |
| "learning_rate": 6.967902333644756e-05, | |
| "loss": 2.3438, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 0.4038534348575996, | |
| "grad_norm": 0.5424301028251648, | |
| "learning_rate": 6.957667091450719e-05, | |
| "loss": 2.3443, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.4045265239156956, | |
| "grad_norm": 0.5094406008720398, | |
| "learning_rate": 6.947422151380354e-05, | |
| "loss": 2.3757, | |
| "step": 1202 | |
| }, | |
| { | |
| "epoch": 0.4051996129737916, | |
| "grad_norm": 0.45820188522338867, | |
| "learning_rate": 6.937167564184968e-05, | |
| "loss": 2.3568, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 0.4058727020318876, | |
| "grad_norm": 0.4518091678619385, | |
| "learning_rate": 6.926903380663656e-05, | |
| "loss": 2.3421, | |
| "step": 1206 | |
| }, | |
| { | |
| "epoch": 0.40654579108998357, | |
| "grad_norm": 0.47818735241889954, | |
| "learning_rate": 6.916629651663048e-05, | |
| "loss": 2.3017, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 0.4072188801480796, | |
| "grad_norm": 0.5458183884620667, | |
| "learning_rate": 6.906346428077073e-05, | |
| "loss": 2.296, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.4078919692061756, | |
| "grad_norm": 0.49146702885627747, | |
| "learning_rate": 6.89605376084668e-05, | |
| "loss": 2.3657, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 0.4085650582642716, | |
| "grad_norm": 0.4122360348701477, | |
| "learning_rate": 6.885751700959605e-05, | |
| "loss": 2.3846, | |
| "step": 1214 | |
| }, | |
| { | |
| "epoch": 0.4092381473223676, | |
| "grad_norm": 0.43666061758995056, | |
| "learning_rate": 6.875440299450116e-05, | |
| "loss": 2.3434, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 0.4099112363804636, | |
| "grad_norm": 0.4253632724285126, | |
| "learning_rate": 6.865119607398755e-05, | |
| "loss": 2.3448, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 0.4105843254385596, | |
| "grad_norm": 0.424385130405426, | |
| "learning_rate": 6.854789675932087e-05, | |
| "loss": 2.3686, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.4112574144966556, | |
| "grad_norm": 0.41038796305656433, | |
| "learning_rate": 6.844450556222447e-05, | |
| "loss": 2.3572, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 0.4119305035547516, | |
| "grad_norm": 0.45698851346969604, | |
| "learning_rate": 6.83410229948769e-05, | |
| "loss": 2.3818, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 0.41260359261284757, | |
| "grad_norm": 0.5241548418998718, | |
| "learning_rate": 6.823744956990931e-05, | |
| "loss": 2.3381, | |
| "step": 1226 | |
| }, | |
| { | |
| "epoch": 0.4132766816709436, | |
| "grad_norm": 0.49647489190101624, | |
| "learning_rate": 6.813378580040292e-05, | |
| "loss": 2.37, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 0.4139497707290396, | |
| "grad_norm": 0.4988088607788086, | |
| "learning_rate": 6.803003219988653e-05, | |
| "loss": 2.3302, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.4146228597871356, | |
| "grad_norm": 0.4698541760444641, | |
| "learning_rate": 6.792618928233396e-05, | |
| "loss": 2.3575, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 0.4152959488452316, | |
| "grad_norm": 0.4854598939418793, | |
| "learning_rate": 6.782225756216143e-05, | |
| "loss": 2.3169, | |
| "step": 1234 | |
| }, | |
| { | |
| "epoch": 0.41596903790332757, | |
| "grad_norm": 0.4949102997779846, | |
| "learning_rate": 6.771823755422513e-05, | |
| "loss": 2.3298, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 0.4166421269614236, | |
| "grad_norm": 0.4542332887649536, | |
| "learning_rate": 6.761412977381857e-05, | |
| "loss": 2.3107, | |
| "step": 1238 | |
| }, | |
| { | |
| "epoch": 0.4173152160195196, | |
| "grad_norm": 0.5075053572654724, | |
| "learning_rate": 6.750993473667007e-05, | |
| "loss": 2.3406, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.4179883050776156, | |
| "grad_norm": 0.44473254680633545, | |
| "learning_rate": 6.740565295894026e-05, | |
| "loss": 2.2938, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 0.4186613941357116, | |
| "grad_norm": 0.48458990454673767, | |
| "learning_rate": 6.730128495721935e-05, | |
| "loss": 2.3394, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 0.41933448319380756, | |
| "grad_norm": 0.5320196747779846, | |
| "learning_rate": 6.719683124852479e-05, | |
| "loss": 2.3688, | |
| "step": 1246 | |
| }, | |
| { | |
| "epoch": 0.4200075722519036, | |
| "grad_norm": 0.47306737303733826, | |
| "learning_rate": 6.709229235029855e-05, | |
| "loss": 2.336, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 0.4206806613099996, | |
| "grad_norm": 0.4362936317920685, | |
| "learning_rate": 6.698766878040467e-05, | |
| "loss": 2.345, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.4213537503680956, | |
| "grad_norm": 0.43612322211265564, | |
| "learning_rate": 6.688296105712654e-05, | |
| "loss": 2.3396, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 0.42202683942619157, | |
| "grad_norm": 0.4405270516872406, | |
| "learning_rate": 6.67781696991645e-05, | |
| "loss": 2.3404, | |
| "step": 1254 | |
| }, | |
| { | |
| "epoch": 0.42269992848428756, | |
| "grad_norm": 0.4700160026550293, | |
| "learning_rate": 6.667329522563318e-05, | |
| "loss": 2.3146, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 0.4233730175423836, | |
| "grad_norm": 0.4510868489742279, | |
| "learning_rate": 6.656833815605896e-05, | |
| "loss": 2.3197, | |
| "step": 1258 | |
| }, | |
| { | |
| "epoch": 0.4240461066004796, | |
| "grad_norm": 0.4313759207725525, | |
| "learning_rate": 6.646329901037736e-05, | |
| "loss": 2.3125, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.4247191956585756, | |
| "grad_norm": 0.4161025881767273, | |
| "learning_rate": 6.635817830893052e-05, | |
| "loss": 2.3822, | |
| "step": 1262 | |
| }, | |
| { | |
| "epoch": 0.42539228471667156, | |
| "grad_norm": 0.4245871901512146, | |
| "learning_rate": 6.625297657246453e-05, | |
| "loss": 2.3407, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 0.42606537377476755, | |
| "grad_norm": 0.5218788981437683, | |
| "learning_rate": 6.614769432212699e-05, | |
| "loss": 2.3919, | |
| "step": 1266 | |
| }, | |
| { | |
| "epoch": 0.4267384628328636, | |
| "grad_norm": 0.4118237793445587, | |
| "learning_rate": 6.60423320794643e-05, | |
| "loss": 2.3141, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 0.4274115518909596, | |
| "grad_norm": 0.3999335765838623, | |
| "learning_rate": 6.593689036641913e-05, | |
| "loss": 2.2975, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.4280846409490556, | |
| "grad_norm": 0.47549012303352356, | |
| "learning_rate": 6.583136970532782e-05, | |
| "loss": 2.3479, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 0.42875773000715156, | |
| "grad_norm": 0.46999719738960266, | |
| "learning_rate": 6.572577061891784e-05, | |
| "loss": 2.3476, | |
| "step": 1274 | |
| }, | |
| { | |
| "epoch": 0.42943081906524755, | |
| "grad_norm": 0.44420936703681946, | |
| "learning_rate": 6.562009363030514e-05, | |
| "loss": 2.3197, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 0.4301039081233436, | |
| "grad_norm": 0.43764299154281616, | |
| "learning_rate": 6.551433926299157e-05, | |
| "loss": 2.3312, | |
| "step": 1278 | |
| }, | |
| { | |
| "epoch": 0.4307769971814396, | |
| "grad_norm": 0.452551007270813, | |
| "learning_rate": 6.540850804086232e-05, | |
| "loss": 2.3493, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.43145008623953557, | |
| "grad_norm": 0.46615126729011536, | |
| "learning_rate": 6.530260048818327e-05, | |
| "loss": 2.3476, | |
| "step": 1282 | |
| }, | |
| { | |
| "epoch": 0.43212317529763156, | |
| "grad_norm": 0.48142266273498535, | |
| "learning_rate": 6.519661712959847e-05, | |
| "loss": 2.3278, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 0.43279626435572754, | |
| "grad_norm": 0.4092482924461365, | |
| "learning_rate": 6.509055849012748e-05, | |
| "loss": 2.3442, | |
| "step": 1286 | |
| }, | |
| { | |
| "epoch": 0.4334693534138236, | |
| "grad_norm": 0.48550838232040405, | |
| "learning_rate": 6.498442509516276e-05, | |
| "loss": 2.3207, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 0.4341424424719196, | |
| "grad_norm": 0.44768187403678894, | |
| "learning_rate": 6.487821747046711e-05, | |
| "loss": 2.3474, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.43481553153001556, | |
| "grad_norm": 0.46920204162597656, | |
| "learning_rate": 6.477193614217109e-05, | |
| "loss": 2.3241, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 0.43548862058811155, | |
| "grad_norm": 0.44246378540992737, | |
| "learning_rate": 6.46655816367703e-05, | |
| "loss": 2.34, | |
| "step": 1294 | |
| }, | |
| { | |
| "epoch": 0.43616170964620754, | |
| "grad_norm": 0.4913026690483093, | |
| "learning_rate": 6.455915448112288e-05, | |
| "loss": 2.3744, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 0.4368347987043036, | |
| "grad_norm": 0.4176254868507385, | |
| "learning_rate": 6.445265520244689e-05, | |
| "loss": 2.3856, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 0.43750788776239957, | |
| "grad_norm": 0.428932785987854, | |
| "learning_rate": 6.434608432831761e-05, | |
| "loss": 2.3701, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.43818097682049556, | |
| "grad_norm": 0.483767569065094, | |
| "learning_rate": 6.423944238666507e-05, | |
| "loss": 2.2977, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 0.43885406587859155, | |
| "grad_norm": 0.42508089542388916, | |
| "learning_rate": 6.413272990577124e-05, | |
| "loss": 2.3336, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 0.43952715493668754, | |
| "grad_norm": 0.47280552983283997, | |
| "learning_rate": 6.402594741426767e-05, | |
| "loss": 2.3412, | |
| "step": 1306 | |
| }, | |
| { | |
| "epoch": 0.4402002439947836, | |
| "grad_norm": 0.4695875346660614, | |
| "learning_rate": 6.391909544113259e-05, | |
| "loss": 2.3435, | |
| "step": 1308 | |
| }, | |
| { | |
| "epoch": 0.44087333305287957, | |
| "grad_norm": 0.44250667095184326, | |
| "learning_rate": 6.381217451568853e-05, | |
| "loss": 2.346, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.44154642211097556, | |
| "grad_norm": 0.42314398288726807, | |
| "learning_rate": 6.370518516759953e-05, | |
| "loss": 2.3021, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 0.44221951116907154, | |
| "grad_norm": 0.45168745517730713, | |
| "learning_rate": 6.359812792686865e-05, | |
| "loss": 2.2962, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 0.44289260022716753, | |
| "grad_norm": 0.3884710967540741, | |
| "learning_rate": 6.34910033238352e-05, | |
| "loss": 2.3283, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 0.4435656892852636, | |
| "grad_norm": 0.4065178632736206, | |
| "learning_rate": 6.338381188917222e-05, | |
| "loss": 2.2931, | |
| "step": 1318 | |
| }, | |
| { | |
| "epoch": 0.44423877834335956, | |
| "grad_norm": 0.4447360038757324, | |
| "learning_rate": 6.327655415388385e-05, | |
| "loss": 2.3522, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.44491186740145555, | |
| "grad_norm": 0.3998311758041382, | |
| "learning_rate": 6.316923064930261e-05, | |
| "loss": 2.3458, | |
| "step": 1322 | |
| }, | |
| { | |
| "epoch": 0.44558495645955154, | |
| "grad_norm": 0.40684449672698975, | |
| "learning_rate": 6.306184190708688e-05, | |
| "loss": 2.3077, | |
| "step": 1324 | |
| }, | |
| { | |
| "epoch": 0.44625804551764753, | |
| "grad_norm": 0.4587690830230713, | |
| "learning_rate": 6.29543884592182e-05, | |
| "loss": 2.3301, | |
| "step": 1326 | |
| }, | |
| { | |
| "epoch": 0.44693113457574357, | |
| "grad_norm": 0.4706231653690338, | |
| "learning_rate": 6.284687083799861e-05, | |
| "loss": 2.3705, | |
| "step": 1328 | |
| }, | |
| { | |
| "epoch": 0.44760422363383956, | |
| "grad_norm": 0.4399780333042145, | |
| "learning_rate": 6.27392895760481e-05, | |
| "loss": 2.3849, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.44827731269193555, | |
| "grad_norm": 0.42533183097839355, | |
| "learning_rate": 6.263164520630192e-05, | |
| "loss": 2.3465, | |
| "step": 1332 | |
| }, | |
| { | |
| "epoch": 0.44895040175003154, | |
| "grad_norm": 0.43969234824180603, | |
| "learning_rate": 6.252393826200791e-05, | |
| "loss": 2.3032, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 0.4496234908081275, | |
| "grad_norm": 0.4467422366142273, | |
| "learning_rate": 6.241616927672394e-05, | |
| "loss": 2.3248, | |
| "step": 1336 | |
| }, | |
| { | |
| "epoch": 0.45029657986622357, | |
| "grad_norm": 0.4212711751461029, | |
| "learning_rate": 6.230833878431514e-05, | |
| "loss": 2.3196, | |
| "step": 1338 | |
| }, | |
| { | |
| "epoch": 0.45096966892431956, | |
| "grad_norm": 0.44400525093078613, | |
| "learning_rate": 6.220044731895143e-05, | |
| "loss": 2.3548, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.45164275798241554, | |
| "grad_norm": 0.45930370688438416, | |
| "learning_rate": 6.209249541510469e-05, | |
| "loss": 2.3202, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 0.45231584704051153, | |
| "grad_norm": 0.4817269742488861, | |
| "learning_rate": 6.198448360754626e-05, | |
| "loss": 2.3446, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 0.4529889360986075, | |
| "grad_norm": 0.4762413799762726, | |
| "learning_rate": 6.18764124313442e-05, | |
| "loss": 2.3331, | |
| "step": 1346 | |
| }, | |
| { | |
| "epoch": 0.45366202515670356, | |
| "grad_norm": 0.48614078760147095, | |
| "learning_rate": 6.17682824218607e-05, | |
| "loss": 2.3208, | |
| "step": 1348 | |
| }, | |
| { | |
| "epoch": 0.45433511421479955, | |
| "grad_norm": 0.45794421434402466, | |
| "learning_rate": 6.166009411474936e-05, | |
| "loss": 2.3076, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.45500820327289554, | |
| "grad_norm": 0.4214746952056885, | |
| "learning_rate": 6.15518480459526e-05, | |
| "loss": 2.3296, | |
| "step": 1352 | |
| }, | |
| { | |
| "epoch": 0.4556812923309915, | |
| "grad_norm": 0.4110795557498932, | |
| "learning_rate": 6.144354475169896e-05, | |
| "loss": 2.3486, | |
| "step": 1354 | |
| }, | |
| { | |
| "epoch": 0.45635438138908757, | |
| "grad_norm": 0.44819360971450806, | |
| "learning_rate": 6.133518476850049e-05, | |
| "loss": 2.3241, | |
| "step": 1356 | |
| }, | |
| { | |
| "epoch": 0.45702747044718356, | |
| "grad_norm": 0.44586122035980225, | |
| "learning_rate": 6.122676863315005e-05, | |
| "loss": 2.3141, | |
| "step": 1358 | |
| }, | |
| { | |
| "epoch": 0.45770055950527955, | |
| "grad_norm": 0.45715874433517456, | |
| "learning_rate": 6.111829688271861e-05, | |
| "loss": 2.3451, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.45837364856337554, | |
| "grad_norm": 0.45967382192611694, | |
| "learning_rate": 6.1009770054552796e-05, | |
| "loss": 2.3115, | |
| "step": 1362 | |
| }, | |
| { | |
| "epoch": 0.4590467376214715, | |
| "grad_norm": 0.47256985306739807, | |
| "learning_rate": 6.0901188686271914e-05, | |
| "loss": 2.3371, | |
| "step": 1364 | |
| }, | |
| { | |
| "epoch": 0.45971982667956757, | |
| "grad_norm": 0.48675423860549927, | |
| "learning_rate": 6.0792553315765544e-05, | |
| "loss": 2.309, | |
| "step": 1366 | |
| }, | |
| { | |
| "epoch": 0.46039291573766356, | |
| "grad_norm": 0.42949745059013367, | |
| "learning_rate": 6.068386448119074e-05, | |
| "loss": 2.3253, | |
| "step": 1368 | |
| }, | |
| { | |
| "epoch": 0.46106600479575954, | |
| "grad_norm": 0.4756808578968048, | |
| "learning_rate": 6.057512272096947e-05, | |
| "loss": 2.3142, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.46173909385385553, | |
| "grad_norm": 0.4509829878807068, | |
| "learning_rate": 6.046632857378578e-05, | |
| "loss": 2.3358, | |
| "step": 1372 | |
| }, | |
| { | |
| "epoch": 0.4624121829119515, | |
| "grad_norm": 0.44080397486686707, | |
| "learning_rate": 6.03574825785833e-05, | |
| "loss": 2.3516, | |
| "step": 1374 | |
| }, | |
| { | |
| "epoch": 0.46308527197004756, | |
| "grad_norm": 0.4561088979244232, | |
| "learning_rate": 6.024858527456253e-05, | |
| "loss": 2.3184, | |
| "step": 1376 | |
| }, | |
| { | |
| "epoch": 0.46375836102814355, | |
| "grad_norm": 0.47704899311065674, | |
| "learning_rate": 6.0139637201178054e-05, | |
| "loss": 2.3782, | |
| "step": 1378 | |
| }, | |
| { | |
| "epoch": 0.46443145008623954, | |
| "grad_norm": 0.4723716080188751, | |
| "learning_rate": 6.003063889813605e-05, | |
| "loss": 2.3244, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.4651045391443355, | |
| "grad_norm": 0.5156892538070679, | |
| "learning_rate": 5.9921590905391456e-05, | |
| "loss": 2.3129, | |
| "step": 1382 | |
| }, | |
| { | |
| "epoch": 0.4657776282024315, | |
| "grad_norm": 0.4867628216743469, | |
| "learning_rate": 5.9812493763145404e-05, | |
| "loss": 2.3425, | |
| "step": 1384 | |
| }, | |
| { | |
| "epoch": 0.46645071726052756, | |
| "grad_norm": 0.4723593592643738, | |
| "learning_rate": 5.970334801184246e-05, | |
| "loss": 2.3104, | |
| "step": 1386 | |
| }, | |
| { | |
| "epoch": 0.46712380631862355, | |
| "grad_norm": 0.48298734426498413, | |
| "learning_rate": 5.959415419216805e-05, | |
| "loss": 2.3595, | |
| "step": 1388 | |
| }, | |
| { | |
| "epoch": 0.46779689537671953, | |
| "grad_norm": 0.5155810713768005, | |
| "learning_rate": 5.948491284504567e-05, | |
| "loss": 2.3057, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.4684699844348155, | |
| "grad_norm": 0.5362403392791748, | |
| "learning_rate": 5.937562451163427e-05, | |
| "loss": 2.3246, | |
| "step": 1392 | |
| }, | |
| { | |
| "epoch": 0.4691430734929115, | |
| "grad_norm": 0.4661843776702881, | |
| "learning_rate": 5.9266289733325566e-05, | |
| "loss": 2.3256, | |
| "step": 1394 | |
| }, | |
| { | |
| "epoch": 0.46981616255100755, | |
| "grad_norm": 0.5100911259651184, | |
| "learning_rate": 5.915690905174136e-05, | |
| "loss": 2.3074, | |
| "step": 1396 | |
| }, | |
| { | |
| "epoch": 0.47048925160910354, | |
| "grad_norm": 0.4927884638309479, | |
| "learning_rate": 5.904748300873082e-05, | |
| "loss": 2.3034, | |
| "step": 1398 | |
| }, | |
| { | |
| "epoch": 0.47116234066719953, | |
| "grad_norm": 0.47624877095222473, | |
| "learning_rate": 5.893801214636786e-05, | |
| "loss": 2.2979, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.4718354297252955, | |
| "grad_norm": 0.4616711735725403, | |
| "learning_rate": 5.8828497006948414e-05, | |
| "loss": 2.3246, | |
| "step": 1402 | |
| }, | |
| { | |
| "epoch": 0.4725085187833915, | |
| "grad_norm": 0.5010025501251221, | |
| "learning_rate": 5.87189381329877e-05, | |
| "loss": 2.3405, | |
| "step": 1404 | |
| }, | |
| { | |
| "epoch": 0.47318160784148755, | |
| "grad_norm": 0.46359044313430786, | |
| "learning_rate": 5.860933606721769e-05, | |
| "loss": 2.3485, | |
| "step": 1406 | |
| }, | |
| { | |
| "epoch": 0.47385469689958354, | |
| "grad_norm": 0.452997624874115, | |
| "learning_rate": 5.849969135258423e-05, | |
| "loss": 2.3215, | |
| "step": 1408 | |
| }, | |
| { | |
| "epoch": 0.4745277859576795, | |
| "grad_norm": 0.4528825879096985, | |
| "learning_rate": 5.839000453224448e-05, | |
| "loss": 2.3188, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.4752008750157755, | |
| "grad_norm": 0.4908580780029297, | |
| "learning_rate": 5.828027614956417e-05, | |
| "loss": 2.3088, | |
| "step": 1412 | |
| }, | |
| { | |
| "epoch": 0.4758739640738715, | |
| "grad_norm": 0.44063472747802734, | |
| "learning_rate": 5.817050674811494e-05, | |
| "loss": 2.3237, | |
| "step": 1414 | |
| }, | |
| { | |
| "epoch": 0.47654705313196755, | |
| "grad_norm": 0.3980746865272522, | |
| "learning_rate": 5.806069687167159e-05, | |
| "loss": 2.3336, | |
| "step": 1416 | |
| }, | |
| { | |
| "epoch": 0.47722014219006353, | |
| "grad_norm": 0.5255020260810852, | |
| "learning_rate": 5.795084706420949e-05, | |
| "loss": 2.3321, | |
| "step": 1418 | |
| }, | |
| { | |
| "epoch": 0.4778932312481595, | |
| "grad_norm": 0.46971768140792847, | |
| "learning_rate": 5.7840957869901746e-05, | |
| "loss": 2.3039, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.4785663203062555, | |
| "grad_norm": 0.687181293964386, | |
| "learning_rate": 5.7731029833116615e-05, | |
| "loss": 2.307, | |
| "step": 1422 | |
| }, | |
| { | |
| "epoch": 0.4792394093643515, | |
| "grad_norm": 0.4752247631549835, | |
| "learning_rate": 5.762106349841479e-05, | |
| "loss": 2.3213, | |
| "step": 1424 | |
| }, | |
| { | |
| "epoch": 0.47991249842244754, | |
| "grad_norm": 0.5026934742927551, | |
| "learning_rate": 5.751105941054663e-05, | |
| "loss": 2.3054, | |
| "step": 1426 | |
| }, | |
| { | |
| "epoch": 0.48058558748054353, | |
| "grad_norm": 0.4666305482387543, | |
| "learning_rate": 5.740101811444959e-05, | |
| "loss": 2.3428, | |
| "step": 1428 | |
| }, | |
| { | |
| "epoch": 0.4812586765386395, | |
| "grad_norm": 0.46005183458328247, | |
| "learning_rate": 5.7290940155245344e-05, | |
| "loss": 2.3215, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.4819317655967355, | |
| "grad_norm": 0.5705943703651428, | |
| "learning_rate": 5.7180826078237314e-05, | |
| "loss": 2.3629, | |
| "step": 1432 | |
| }, | |
| { | |
| "epoch": 0.4826048546548315, | |
| "grad_norm": 0.4443369209766388, | |
| "learning_rate": 5.707067642890773e-05, | |
| "loss": 2.3, | |
| "step": 1434 | |
| }, | |
| { | |
| "epoch": 0.48327794371292754, | |
| "grad_norm": 0.5047985315322876, | |
| "learning_rate": 5.6960491752915105e-05, | |
| "loss": 2.3061, | |
| "step": 1436 | |
| }, | |
| { | |
| "epoch": 0.4839510327710235, | |
| "grad_norm": 0.5213374495506287, | |
| "learning_rate": 5.685027259609144e-05, | |
| "loss": 2.3395, | |
| "step": 1438 | |
| }, | |
| { | |
| "epoch": 0.4846241218291195, | |
| "grad_norm": 0.41633522510528564, | |
| "learning_rate": 5.6740019504439576e-05, | |
| "loss": 2.3149, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.4852972108872155, | |
| "grad_norm": 0.4161958694458008, | |
| "learning_rate": 5.662973302413039e-05, | |
| "loss": 2.2994, | |
| "step": 1442 | |
| }, | |
| { | |
| "epoch": 0.4859702999453115, | |
| "grad_norm": 0.44887277483940125, | |
| "learning_rate": 5.651941370150028e-05, | |
| "loss": 2.3016, | |
| "step": 1444 | |
| }, | |
| { | |
| "epoch": 0.48664338900340753, | |
| "grad_norm": 0.43213191628456116, | |
| "learning_rate": 5.6409062083048204e-05, | |
| "loss": 2.3115, | |
| "step": 1446 | |
| }, | |
| { | |
| "epoch": 0.4873164780615035, | |
| "grad_norm": 0.41568759083747864, | |
| "learning_rate": 5.629867871543318e-05, | |
| "loss": 2.3131, | |
| "step": 1448 | |
| }, | |
| { | |
| "epoch": 0.4879895671195995, | |
| "grad_norm": 0.5004609227180481, | |
| "learning_rate": 5.618826414547154e-05, | |
| "loss": 2.3036, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.4886626561776955, | |
| "grad_norm": 0.4459887742996216, | |
| "learning_rate": 5.6077818920134076e-05, | |
| "loss": 2.2908, | |
| "step": 1452 | |
| }, | |
| { | |
| "epoch": 0.4893357452357915, | |
| "grad_norm": 0.48889175057411194, | |
| "learning_rate": 5.596734358654354e-05, | |
| "loss": 2.3218, | |
| "step": 1454 | |
| }, | |
| { | |
| "epoch": 0.49000883429388753, | |
| "grad_norm": 0.5154656767845154, | |
| "learning_rate": 5.585683869197179e-05, | |
| "loss": 2.3291, | |
| "step": 1456 | |
| }, | |
| { | |
| "epoch": 0.4906819233519835, | |
| "grad_norm": 0.4590589702129364, | |
| "learning_rate": 5.574630478383714e-05, | |
| "loss": 2.276, | |
| "step": 1458 | |
| }, | |
| { | |
| "epoch": 0.4913550124100795, | |
| "grad_norm": 0.4735153317451477, | |
| "learning_rate": 5.56357424097016e-05, | |
| "loss": 2.3469, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.4920281014681755, | |
| "grad_norm": 0.4928821921348572, | |
| "learning_rate": 5.552515211726825e-05, | |
| "loss": 2.311, | |
| "step": 1462 | |
| }, | |
| { | |
| "epoch": 0.4927011905262715, | |
| "grad_norm": 0.4758559465408325, | |
| "learning_rate": 5.5414534454378406e-05, | |
| "loss": 2.3068, | |
| "step": 1464 | |
| }, | |
| { | |
| "epoch": 0.4933742795843675, | |
| "grad_norm": 0.45103973150253296, | |
| "learning_rate": 5.5303889969009026e-05, | |
| "loss": 2.313, | |
| "step": 1466 | |
| }, | |
| { | |
| "epoch": 0.4940473686424635, | |
| "grad_norm": 0.42042675614356995, | |
| "learning_rate": 5.519321920926991e-05, | |
| "loss": 2.2479, | |
| "step": 1468 | |
| }, | |
| { | |
| "epoch": 0.4947204577005595, | |
| "grad_norm": 0.5105794072151184, | |
| "learning_rate": 5.508252272340103e-05, | |
| "loss": 2.3064, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.4953935467586555, | |
| "grad_norm": 0.48093077540397644, | |
| "learning_rate": 5.497180105976978e-05, | |
| "loss": 2.336, | |
| "step": 1472 | |
| }, | |
| { | |
| "epoch": 0.4960666358167515, | |
| "grad_norm": 0.47022518515586853, | |
| "learning_rate": 5.4861054766868316e-05, | |
| "loss": 2.3267, | |
| "step": 1474 | |
| }, | |
| { | |
| "epoch": 0.4967397248748475, | |
| "grad_norm": 0.4636436104774475, | |
| "learning_rate": 5.4750284393310756e-05, | |
| "loss": 2.2619, | |
| "step": 1476 | |
| }, | |
| { | |
| "epoch": 0.4974128139329435, | |
| "grad_norm": 0.4825097918510437, | |
| "learning_rate": 5.463949048783056e-05, | |
| "loss": 2.3256, | |
| "step": 1478 | |
| }, | |
| { | |
| "epoch": 0.4980859029910395, | |
| "grad_norm": 0.47207555174827576, | |
| "learning_rate": 5.452867359927771e-05, | |
| "loss": 2.2974, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.4987589920491355, | |
| "grad_norm": 0.4509056508541107, | |
| "learning_rate": 5.441783427661608e-05, | |
| "loss": 2.3031, | |
| "step": 1482 | |
| }, | |
| { | |
| "epoch": 0.4994320811072315, | |
| "grad_norm": 0.48165446519851685, | |
| "learning_rate": 5.430697306892065e-05, | |
| "loss": 2.3011, | |
| "step": 1484 | |
| }, | |
| { | |
| "epoch": 0.5001051701653275, | |
| "grad_norm": 0.4641640782356262, | |
| "learning_rate": 5.419609052537482e-05, | |
| "loss": 2.322, | |
| "step": 1486 | |
| }, | |
| { | |
| "epoch": 0.5007782592234234, | |
| "grad_norm": 0.43199190497398376, | |
| "learning_rate": 5.408518719526773e-05, | |
| "loss": 2.2992, | |
| "step": 1488 | |
| }, | |
| { | |
| "epoch": 0.5014513482815195, | |
| "grad_norm": 0.45555445551872253, | |
| "learning_rate": 5.397426362799141e-05, | |
| "loss": 2.3357, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.5021244373396155, | |
| "grad_norm": 0.4025207757949829, | |
| "learning_rate": 5.3863320373038193e-05, | |
| "loss": 2.3064, | |
| "step": 1492 | |
| }, | |
| { | |
| "epoch": 0.5027975263977115, | |
| "grad_norm": 0.40771278738975525, | |
| "learning_rate": 5.3752357979997936e-05, | |
| "loss": 2.3096, | |
| "step": 1494 | |
| }, | |
| { | |
| "epoch": 0.5034706154558075, | |
| "grad_norm": 0.4379865825176239, | |
| "learning_rate": 5.36413769985553e-05, | |
| "loss": 2.2943, | |
| "step": 1496 | |
| }, | |
| { | |
| "epoch": 0.5041437045139034, | |
| "grad_norm": 0.44733837246894836, | |
| "learning_rate": 5.353037797848702e-05, | |
| "loss": 2.3163, | |
| "step": 1498 | |
| }, | |
| { | |
| "epoch": 0.5048167935719995, | |
| "grad_norm": 0.4397582411766052, | |
| "learning_rate": 5.34193614696592e-05, | |
| "loss": 2.3154, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.5054898826300955, | |
| "grad_norm": 0.4777575135231018, | |
| "learning_rate": 5.330832802202457e-05, | |
| "loss": 2.31, | |
| "step": 1502 | |
| }, | |
| { | |
| "epoch": 0.5061629716881915, | |
| "grad_norm": 0.41434261202812195, | |
| "learning_rate": 5.319727818561977e-05, | |
| "loss": 2.3624, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 0.5068360607462875, | |
| "grad_norm": 0.47366392612457275, | |
| "learning_rate": 5.308621251056266e-05, | |
| "loss": 2.3271, | |
| "step": 1506 | |
| }, | |
| { | |
| "epoch": 0.5075091498043834, | |
| "grad_norm": 0.44623860716819763, | |
| "learning_rate": 5.2975131547049505e-05, | |
| "loss": 2.2921, | |
| "step": 1508 | |
| }, | |
| { | |
| "epoch": 0.5081822388624795, | |
| "grad_norm": 0.39073917269706726, | |
| "learning_rate": 5.2864035845352356e-05, | |
| "loss": 2.3, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.5088553279205755, | |
| "grad_norm": 0.40907222032546997, | |
| "learning_rate": 5.2752925955816246e-05, | |
| "loss": 2.2946, | |
| "step": 1512 | |
| }, | |
| { | |
| "epoch": 0.5095284169786715, | |
| "grad_norm": 0.3879257142543793, | |
| "learning_rate": 5.264180242885652e-05, | |
| "loss": 2.3198, | |
| "step": 1514 | |
| }, | |
| { | |
| "epoch": 0.5102015060367675, | |
| "grad_norm": 0.4007498025894165, | |
| "learning_rate": 5.253066581495601e-05, | |
| "loss": 2.3117, | |
| "step": 1516 | |
| }, | |
| { | |
| "epoch": 0.5108745950948634, | |
| "grad_norm": 0.4116353392601013, | |
| "learning_rate": 5.241951666466249e-05, | |
| "loss": 2.317, | |
| "step": 1518 | |
| }, | |
| { | |
| "epoch": 0.5115476841529595, | |
| "grad_norm": 0.42198294401168823, | |
| "learning_rate": 5.2308355528585726e-05, | |
| "loss": 2.347, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.5122207732110555, | |
| "grad_norm": 0.40731942653656006, | |
| "learning_rate": 5.219718295739494e-05, | |
| "loss": 2.3152, | |
| "step": 1522 | |
| }, | |
| { | |
| "epoch": 0.5128938622691515, | |
| "grad_norm": 0.3913952708244324, | |
| "learning_rate": 5.208599950181593e-05, | |
| "loss": 2.3153, | |
| "step": 1524 | |
| }, | |
| { | |
| "epoch": 0.5135669513272475, | |
| "grad_norm": 0.41797488927841187, | |
| "learning_rate": 5.197480571262849e-05, | |
| "loss": 2.3226, | |
| "step": 1526 | |
| }, | |
| { | |
| "epoch": 0.5142400403853434, | |
| "grad_norm": 0.39869993925094604, | |
| "learning_rate": 5.1863602140663544e-05, | |
| "loss": 2.3063, | |
| "step": 1528 | |
| }, | |
| { | |
| "epoch": 0.5149131294434395, | |
| "grad_norm": 0.3935421109199524, | |
| "learning_rate": 5.175238933680051e-05, | |
| "loss": 2.2692, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.5155862185015355, | |
| "grad_norm": 0.40877199172973633, | |
| "learning_rate": 5.1641167851964535e-05, | |
| "loss": 2.2589, | |
| "step": 1532 | |
| }, | |
| { | |
| "epoch": 0.5162593075596315, | |
| "grad_norm": 0.4104747772216797, | |
| "learning_rate": 5.152993823712373e-05, | |
| "loss": 2.3302, | |
| "step": 1534 | |
| }, | |
| { | |
| "epoch": 0.5169323966177275, | |
| "grad_norm": 0.4258463382720947, | |
| "learning_rate": 5.1418701043286554e-05, | |
| "loss": 2.2981, | |
| "step": 1536 | |
| }, | |
| { | |
| "epoch": 0.5176054856758234, | |
| "grad_norm": 0.4016733467578888, | |
| "learning_rate": 5.1307456821498945e-05, | |
| "loss": 2.2855, | |
| "step": 1538 | |
| }, | |
| { | |
| "epoch": 0.5182785747339195, | |
| "grad_norm": 0.3749251961708069, | |
| "learning_rate": 5.1196206122841705e-05, | |
| "loss": 2.3097, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.5189516637920155, | |
| "grad_norm": 0.41092702746391296, | |
| "learning_rate": 5.108494949842766e-05, | |
| "loss": 2.3075, | |
| "step": 1542 | |
| }, | |
| { | |
| "epoch": 0.5196247528501114, | |
| "grad_norm": 0.4320509731769562, | |
| "learning_rate": 5.097368749939908e-05, | |
| "loss": 2.3192, | |
| "step": 1544 | |
| }, | |
| { | |
| "epoch": 0.5202978419082075, | |
| "grad_norm": 0.3932344913482666, | |
| "learning_rate": 5.086242067692477e-05, | |
| "loss": 2.3131, | |
| "step": 1546 | |
| }, | |
| { | |
| "epoch": 0.5209709309663034, | |
| "grad_norm": 0.40453147888183594, | |
| "learning_rate": 5.0751149582197475e-05, | |
| "loss": 2.327, | |
| "step": 1548 | |
| }, | |
| { | |
| "epoch": 0.5216440200243995, | |
| "grad_norm": 0.4591049253940582, | |
| "learning_rate": 5.063987476643109e-05, | |
| "loss": 2.2951, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.5223171090824955, | |
| "grad_norm": 0.40897637605667114, | |
| "learning_rate": 5.0528596780857995e-05, | |
| "loss": 2.3086, | |
| "step": 1552 | |
| }, | |
| { | |
| "epoch": 0.5229901981405914, | |
| "grad_norm": 0.4454411268234253, | |
| "learning_rate": 5.0417316176726174e-05, | |
| "loss": 2.2754, | |
| "step": 1554 | |
| }, | |
| { | |
| "epoch": 0.5236632871986875, | |
| "grad_norm": 0.41909876465797424, | |
| "learning_rate": 5.0306033505296644e-05, | |
| "loss": 2.3053, | |
| "step": 1556 | |
| }, | |
| { | |
| "epoch": 0.5243363762567834, | |
| "grad_norm": 0.4356496334075928, | |
| "learning_rate": 5.019474931784069e-05, | |
| "loss": 2.2886, | |
| "step": 1558 | |
| }, | |
| { | |
| "epoch": 0.5250094653148795, | |
| "grad_norm": 0.4291853606700897, | |
| "learning_rate": 5.0083464165637006e-05, | |
| "loss": 2.3279, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.5256825543729755, | |
| "grad_norm": 0.4215032160282135, | |
| "learning_rate": 4.9972178599969214e-05, | |
| "loss": 2.3011, | |
| "step": 1562 | |
| }, | |
| { | |
| "epoch": 0.5263556434310714, | |
| "grad_norm": 0.4067915380001068, | |
| "learning_rate": 4.9860893172122844e-05, | |
| "loss": 2.3287, | |
| "step": 1564 | |
| }, | |
| { | |
| "epoch": 0.5270287324891675, | |
| "grad_norm": 0.4543415307998657, | |
| "learning_rate": 4.974960843338282e-05, | |
| "loss": 2.3082, | |
| "step": 1566 | |
| }, | |
| { | |
| "epoch": 0.5277018215472635, | |
| "grad_norm": 0.44435128569602966, | |
| "learning_rate": 4.9638324935030646e-05, | |
| "loss": 2.3121, | |
| "step": 1568 | |
| }, | |
| { | |
| "epoch": 0.5283749106053595, | |
| "grad_norm": 0.40943652391433716, | |
| "learning_rate": 4.9527043228341644e-05, | |
| "loss": 2.27, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.5290479996634555, | |
| "grad_norm": 0.4154236614704132, | |
| "learning_rate": 4.941576386458228e-05, | |
| "loss": 2.299, | |
| "step": 1572 | |
| }, | |
| { | |
| "epoch": 0.5297210887215514, | |
| "grad_norm": 0.41661956906318665, | |
| "learning_rate": 4.9304487395007465e-05, | |
| "loss": 2.2963, | |
| "step": 1574 | |
| }, | |
| { | |
| "epoch": 0.5303941777796475, | |
| "grad_norm": 0.44022658467292786, | |
| "learning_rate": 4.919321437085766e-05, | |
| "loss": 2.2937, | |
| "step": 1576 | |
| }, | |
| { | |
| "epoch": 0.5310672668377435, | |
| "grad_norm": 0.418099582195282, | |
| "learning_rate": 4.9081945343356384e-05, | |
| "loss": 2.3077, | |
| "step": 1578 | |
| }, | |
| { | |
| "epoch": 0.5317403558958395, | |
| "grad_norm": 0.4155904948711395, | |
| "learning_rate": 4.897068086370728e-05, | |
| "loss": 2.2782, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.5324134449539355, | |
| "grad_norm": 0.5036917328834534, | |
| "learning_rate": 4.885942148309146e-05, | |
| "loss": 2.3165, | |
| "step": 1582 | |
| }, | |
| { | |
| "epoch": 0.5330865340120314, | |
| "grad_norm": 0.44698935747146606, | |
| "learning_rate": 4.8748167752664815e-05, | |
| "loss": 2.3194, | |
| "step": 1584 | |
| }, | |
| { | |
| "epoch": 0.5337596230701275, | |
| "grad_norm": 0.4752766788005829, | |
| "learning_rate": 4.863692022355523e-05, | |
| "loss": 2.3011, | |
| "step": 1586 | |
| }, | |
| { | |
| "epoch": 0.5344327121282235, | |
| "grad_norm": 0.45811524987220764, | |
| "learning_rate": 4.852567944685985e-05, | |
| "loss": 2.2886, | |
| "step": 1588 | |
| }, | |
| { | |
| "epoch": 0.5351058011863195, | |
| "grad_norm": 0.4390583336353302, | |
| "learning_rate": 4.841444597364239e-05, | |
| "loss": 2.2841, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.5357788902444155, | |
| "grad_norm": 0.45630621910095215, | |
| "learning_rate": 4.830322035493042e-05, | |
| "loss": 2.2786, | |
| "step": 1592 | |
| }, | |
| { | |
| "epoch": 0.5364519793025114, | |
| "grad_norm": 0.44058138132095337, | |
| "learning_rate": 4.81920031417125e-05, | |
| "loss": 2.3038, | |
| "step": 1594 | |
| }, | |
| { | |
| "epoch": 0.5371250683606075, | |
| "grad_norm": 0.43043145537376404, | |
| "learning_rate": 4.808079488493566e-05, | |
| "loss": 2.3128, | |
| "step": 1596 | |
| }, | |
| { | |
| "epoch": 0.5377981574187035, | |
| "grad_norm": 0.4366210103034973, | |
| "learning_rate": 4.7969596135502515e-05, | |
| "loss": 2.3094, | |
| "step": 1598 | |
| }, | |
| { | |
| "epoch": 0.5384712464767994, | |
| "grad_norm": 0.3835521340370178, | |
| "learning_rate": 4.785840744426856e-05, | |
| "loss": 2.3403, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.5391443355348955, | |
| "grad_norm": 0.4341786205768585, | |
| "learning_rate": 4.7747229362039495e-05, | |
| "loss": 2.2818, | |
| "step": 1602 | |
| }, | |
| { | |
| "epoch": 0.5398174245929914, | |
| "grad_norm": 0.420887291431427, | |
| "learning_rate": 4.7636062439568465e-05, | |
| "loss": 2.2951, | |
| "step": 1604 | |
| }, | |
| { | |
| "epoch": 0.5404905136510875, | |
| "grad_norm": 0.46410495042800903, | |
| "learning_rate": 4.75249072275533e-05, | |
| "loss": 2.2841, | |
| "step": 1606 | |
| }, | |
| { | |
| "epoch": 0.5411636027091835, | |
| "grad_norm": 0.47272729873657227, | |
| "learning_rate": 4.7413764276633884e-05, | |
| "loss": 2.3213, | |
| "step": 1608 | |
| }, | |
| { | |
| "epoch": 0.5418366917672794, | |
| "grad_norm": 0.5024069547653198, | |
| "learning_rate": 4.73026341373893e-05, | |
| "loss": 2.2855, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.5425097808253755, | |
| "grad_norm": 0.5254536271095276, | |
| "learning_rate": 4.719151736033517e-05, | |
| "loss": 2.2954, | |
| "step": 1612 | |
| }, | |
| { | |
| "epoch": 0.5431828698834714, | |
| "grad_norm": 0.45614388585090637, | |
| "learning_rate": 4.7080414495920954e-05, | |
| "loss": 2.333, | |
| "step": 1614 | |
| }, | |
| { | |
| "epoch": 0.5438559589415675, | |
| "grad_norm": 0.42663490772247314, | |
| "learning_rate": 4.696932609452719e-05, | |
| "loss": 2.2926, | |
| "step": 1616 | |
| }, | |
| { | |
| "epoch": 0.5445290479996635, | |
| "grad_norm": 0.4447422921657562, | |
| "learning_rate": 4.685825270646272e-05, | |
| "loss": 2.3185, | |
| "step": 1618 | |
| }, | |
| { | |
| "epoch": 0.5452021370577594, | |
| "grad_norm": 0.4337231516838074, | |
| "learning_rate": 4.6747194881962054e-05, | |
| "loss": 2.3017, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.5458752261158555, | |
| "grad_norm": 0.42868009209632874, | |
| "learning_rate": 4.663615317118264e-05, | |
| "loss": 2.2865, | |
| "step": 1622 | |
| }, | |
| { | |
| "epoch": 0.5465483151739514, | |
| "grad_norm": 0.41032806038856506, | |
| "learning_rate": 4.6525128124202e-05, | |
| "loss": 2.3103, | |
| "step": 1624 | |
| }, | |
| { | |
| "epoch": 0.5472214042320475, | |
| "grad_norm": 0.42203909158706665, | |
| "learning_rate": 4.6414120291015206e-05, | |
| "loss": 2.29, | |
| "step": 1626 | |
| }, | |
| { | |
| "epoch": 0.5478944932901435, | |
| "grad_norm": 0.4146362841129303, | |
| "learning_rate": 4.630313022153201e-05, | |
| "loss": 2.2996, | |
| "step": 1628 | |
| }, | |
| { | |
| "epoch": 0.5485675823482394, | |
| "grad_norm": 0.4136561155319214, | |
| "learning_rate": 4.6192158465574166e-05, | |
| "loss": 2.261, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.5492406714063355, | |
| "grad_norm": 0.4232659935951233, | |
| "learning_rate": 4.608120557287272e-05, | |
| "loss": 2.2871, | |
| "step": 1632 | |
| }, | |
| { | |
| "epoch": 0.5499137604644314, | |
| "grad_norm": 0.4048542380332947, | |
| "learning_rate": 4.597027209306525e-05, | |
| "loss": 2.2924, | |
| "step": 1634 | |
| }, | |
| { | |
| "epoch": 0.5505868495225275, | |
| "grad_norm": 0.3882562816143036, | |
| "learning_rate": 4.5859358575693195e-05, | |
| "loss": 2.3021, | |
| "step": 1636 | |
| }, | |
| { | |
| "epoch": 0.5512599385806235, | |
| "grad_norm": 0.3903749883174896, | |
| "learning_rate": 4.574846557019911e-05, | |
| "loss": 2.3258, | |
| "step": 1638 | |
| }, | |
| { | |
| "epoch": 0.5519330276387194, | |
| "grad_norm": 0.41353288292884827, | |
| "learning_rate": 4.5637593625923886e-05, | |
| "loss": 2.2983, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.5526061166968155, | |
| "grad_norm": 0.40081363916397095, | |
| "learning_rate": 4.5526743292104126e-05, | |
| "loss": 2.3064, | |
| "step": 1642 | |
| }, | |
| { | |
| "epoch": 0.5532792057549114, | |
| "grad_norm": 0.39215919375419617, | |
| "learning_rate": 4.5415915117869406e-05, | |
| "loss": 2.3512, | |
| "step": 1644 | |
| }, | |
| { | |
| "epoch": 0.5539522948130075, | |
| "grad_norm": 0.45746368169784546, | |
| "learning_rate": 4.5305109652239444e-05, | |
| "loss": 2.3371, | |
| "step": 1646 | |
| }, | |
| { | |
| "epoch": 0.5546253838711035, | |
| "grad_norm": 0.4346954822540283, | |
| "learning_rate": 4.5194327444121535e-05, | |
| "loss": 2.3242, | |
| "step": 1648 | |
| }, | |
| { | |
| "epoch": 0.5552984729291994, | |
| "grad_norm": 0.44898685812950134, | |
| "learning_rate": 4.508356904230775e-05, | |
| "loss": 2.2378, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.5559715619872955, | |
| "grad_norm": 0.4585607945919037, | |
| "learning_rate": 4.497283499547218e-05, | |
| "loss": 2.3095, | |
| "step": 1652 | |
| }, | |
| { | |
| "epoch": 0.5566446510453914, | |
| "grad_norm": 0.4320756494998932, | |
| "learning_rate": 4.486212585216832e-05, | |
| "loss": 2.3374, | |
| "step": 1654 | |
| }, | |
| { | |
| "epoch": 0.5573177401034874, | |
| "grad_norm": 0.4416314363479614, | |
| "learning_rate": 4.475144216082631e-05, | |
| "loss": 2.2831, | |
| "step": 1656 | |
| }, | |
| { | |
| "epoch": 0.5579908291615835, | |
| "grad_norm": 0.4329529106616974, | |
| "learning_rate": 4.4640784469750125e-05, | |
| "loss": 2.3037, | |
| "step": 1658 | |
| }, | |
| { | |
| "epoch": 0.5586639182196794, | |
| "grad_norm": 0.44045084714889526, | |
| "learning_rate": 4.4530153327115024e-05, | |
| "loss": 2.293, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.5593370072777755, | |
| "grad_norm": 0.4578095078468323, | |
| "learning_rate": 4.4419549280964736e-05, | |
| "loss": 2.2546, | |
| "step": 1662 | |
| }, | |
| { | |
| "epoch": 0.5600100963358714, | |
| "grad_norm": 0.4369781017303467, | |
| "learning_rate": 4.43089728792087e-05, | |
| "loss": 2.2808, | |
| "step": 1664 | |
| }, | |
| { | |
| "epoch": 0.5606831853939674, | |
| "grad_norm": 0.39533036947250366, | |
| "learning_rate": 4.419842466961947e-05, | |
| "loss": 2.3121, | |
| "step": 1666 | |
| }, | |
| { | |
| "epoch": 0.5613562744520635, | |
| "grad_norm": 0.4261418581008911, | |
| "learning_rate": 4.408790519982994e-05, | |
| "loss": 2.3006, | |
| "step": 1668 | |
| }, | |
| { | |
| "epoch": 0.5620293635101594, | |
| "grad_norm": 0.43180030584335327, | |
| "learning_rate": 4.3977415017330604e-05, | |
| "loss": 2.287, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.5627024525682555, | |
| "grad_norm": 0.40732190012931824, | |
| "learning_rate": 4.386695466946687e-05, | |
| "loss": 2.2768, | |
| "step": 1672 | |
| }, | |
| { | |
| "epoch": 0.5633755416263514, | |
| "grad_norm": 0.45698872208595276, | |
| "learning_rate": 4.375652470343642e-05, | |
| "loss": 2.297, | |
| "step": 1674 | |
| }, | |
| { | |
| "epoch": 0.5640486306844474, | |
| "grad_norm": 0.4643264412879944, | |
| "learning_rate": 4.3646125666286305e-05, | |
| "loss": 2.2932, | |
| "step": 1676 | |
| }, | |
| { | |
| "epoch": 0.5647217197425435, | |
| "grad_norm": 0.43006443977355957, | |
| "learning_rate": 4.3535758104910495e-05, | |
| "loss": 2.2912, | |
| "step": 1678 | |
| }, | |
| { | |
| "epoch": 0.5653948088006394, | |
| "grad_norm": 0.38097435235977173, | |
| "learning_rate": 4.3425422566046965e-05, | |
| "loss": 2.2753, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.5660678978587355, | |
| "grad_norm": 0.40803152322769165, | |
| "learning_rate": 4.331511959627504e-05, | |
| "loss": 2.2997, | |
| "step": 1682 | |
| }, | |
| { | |
| "epoch": 0.5667409869168314, | |
| "grad_norm": 0.42283037304878235, | |
| "learning_rate": 4.320484974201275e-05, | |
| "loss": 2.3095, | |
| "step": 1684 | |
| }, | |
| { | |
| "epoch": 0.5674140759749274, | |
| "grad_norm": 0.389039546251297, | |
| "learning_rate": 4.309461354951407e-05, | |
| "loss": 2.283, | |
| "step": 1686 | |
| }, | |
| { | |
| "epoch": 0.5680871650330235, | |
| "grad_norm": 0.3792765736579895, | |
| "learning_rate": 4.2984411564866186e-05, | |
| "loss": 2.2463, | |
| "step": 1688 | |
| }, | |
| { | |
| "epoch": 0.5687602540911194, | |
| "grad_norm": 0.421396404504776, | |
| "learning_rate": 4.287424433398684e-05, | |
| "loss": 2.2437, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.5694333431492155, | |
| "grad_norm": 0.5416815876960754, | |
| "learning_rate": 4.276411240262169e-05, | |
| "loss": 2.2914, | |
| "step": 1692 | |
| }, | |
| { | |
| "epoch": 0.5701064322073114, | |
| "grad_norm": 0.4610809087753296, | |
| "learning_rate": 4.265401631634138e-05, | |
| "loss": 2.2899, | |
| "step": 1694 | |
| }, | |
| { | |
| "epoch": 0.5707795212654074, | |
| "grad_norm": 0.43153125047683716, | |
| "learning_rate": 4.2543956620539106e-05, | |
| "loss": 2.2828, | |
| "step": 1696 | |
| }, | |
| { | |
| "epoch": 0.5714526103235035, | |
| "grad_norm": 0.4280839264392853, | |
| "learning_rate": 4.2433933860427754e-05, | |
| "loss": 2.2852, | |
| "step": 1698 | |
| }, | |
| { | |
| "epoch": 0.5721256993815994, | |
| "grad_norm": 0.4372383952140808, | |
| "learning_rate": 4.2323948581037206e-05, | |
| "loss": 2.2886, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.5727987884396954, | |
| "grad_norm": 0.41733816266059875, | |
| "learning_rate": 4.22140013272117e-05, | |
| "loss": 2.2735, | |
| "step": 1702 | |
| }, | |
| { | |
| "epoch": 0.5734718774977914, | |
| "grad_norm": 0.39934617280960083, | |
| "learning_rate": 4.210409264360717e-05, | |
| "loss": 2.3215, | |
| "step": 1704 | |
| }, | |
| { | |
| "epoch": 0.5741449665558874, | |
| "grad_norm": 0.3875669240951538, | |
| "learning_rate": 4.1994223074688324e-05, | |
| "loss": 2.2841, | |
| "step": 1706 | |
| }, | |
| { | |
| "epoch": 0.5748180556139835, | |
| "grad_norm": 0.4122958481311798, | |
| "learning_rate": 4.188439316472626e-05, | |
| "loss": 2.2747, | |
| "step": 1708 | |
| }, | |
| { | |
| "epoch": 0.5754911446720794, | |
| "grad_norm": 0.4905928373336792, | |
| "learning_rate": 4.177460345779554e-05, | |
| "loss": 2.2605, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.5761642337301754, | |
| "grad_norm": 0.4066742956638336, | |
| "learning_rate": 4.166485449777155e-05, | |
| "loss": 2.285, | |
| "step": 1712 | |
| }, | |
| { | |
| "epoch": 0.5768373227882714, | |
| "grad_norm": 0.4053576588630676, | |
| "learning_rate": 4.1555146828327875e-05, | |
| "loss": 2.2672, | |
| "step": 1714 | |
| }, | |
| { | |
| "epoch": 0.5775104118463674, | |
| "grad_norm": 0.39931684732437134, | |
| "learning_rate": 4.1445480992933526e-05, | |
| "loss": 2.297, | |
| "step": 1716 | |
| }, | |
| { | |
| "epoch": 0.5781835009044635, | |
| "grad_norm": 0.44321757555007935, | |
| "learning_rate": 4.133585753485029e-05, | |
| "loss": 2.2268, | |
| "step": 1718 | |
| }, | |
| { | |
| "epoch": 0.5788565899625594, | |
| "grad_norm": 0.5384443402290344, | |
| "learning_rate": 4.1226276997130014e-05, | |
| "loss": 2.3537, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.5795296790206554, | |
| "grad_norm": 0.4857224225997925, | |
| "learning_rate": 4.111673992261193e-05, | |
| "loss": 2.3212, | |
| "step": 1722 | |
| }, | |
| { | |
| "epoch": 0.5802027680787514, | |
| "grad_norm": 0.515303373336792, | |
| "learning_rate": 4.100724685391994e-05, | |
| "loss": 2.2802, | |
| "step": 1724 | |
| }, | |
| { | |
| "epoch": 0.5808758571368474, | |
| "grad_norm": 0.46697720885276794, | |
| "learning_rate": 4.089779833346003e-05, | |
| "loss": 2.275, | |
| "step": 1726 | |
| }, | |
| { | |
| "epoch": 0.5815489461949435, | |
| "grad_norm": 0.43453115224838257, | |
| "learning_rate": 4.078839490341739e-05, | |
| "loss": 2.2577, | |
| "step": 1728 | |
| }, | |
| { | |
| "epoch": 0.5822220352530394, | |
| "grad_norm": 0.3856815695762634, | |
| "learning_rate": 4.067903710575393e-05, | |
| "loss": 2.3017, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.5828951243111354, | |
| "grad_norm": 0.40342196822166443, | |
| "learning_rate": 4.0569725482205476e-05, | |
| "loss": 2.288, | |
| "step": 1732 | |
| }, | |
| { | |
| "epoch": 0.5835682133692314, | |
| "grad_norm": 0.4150996208190918, | |
| "learning_rate": 4.046046057427911e-05, | |
| "loss": 2.2616, | |
| "step": 1734 | |
| }, | |
| { | |
| "epoch": 0.5842413024273274, | |
| "grad_norm": 0.408902645111084, | |
| "learning_rate": 4.0351242923250485e-05, | |
| "loss": 2.2993, | |
| "step": 1736 | |
| }, | |
| { | |
| "epoch": 0.5849143914854235, | |
| "grad_norm": 0.42966151237487793, | |
| "learning_rate": 4.024207307016121e-05, | |
| "loss": 2.3073, | |
| "step": 1738 | |
| }, | |
| { | |
| "epoch": 0.5855874805435194, | |
| "grad_norm": 0.4190477430820465, | |
| "learning_rate": 4.013295155581601e-05, | |
| "loss": 2.3066, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.5862605696016154, | |
| "grad_norm": 0.41285306215286255, | |
| "learning_rate": 4.002387892078026e-05, | |
| "loss": 2.2616, | |
| "step": 1742 | |
| }, | |
| { | |
| "epoch": 0.5869336586597114, | |
| "grad_norm": 0.40417051315307617, | |
| "learning_rate": 3.991485570537715e-05, | |
| "loss": 2.2699, | |
| "step": 1744 | |
| }, | |
| { | |
| "epoch": 0.5876067477178074, | |
| "grad_norm": 0.3870849907398224, | |
| "learning_rate": 3.980588244968503e-05, | |
| "loss": 2.3238, | |
| "step": 1746 | |
| }, | |
| { | |
| "epoch": 0.5882798367759035, | |
| "grad_norm": 0.4099171459674835, | |
| "learning_rate": 3.969695969353479e-05, | |
| "loss": 2.2526, | |
| "step": 1748 | |
| }, | |
| { | |
| "epoch": 0.5889529258339994, | |
| "grad_norm": 0.42374101281166077, | |
| "learning_rate": 3.958808797650717e-05, | |
| "loss": 2.2561, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.5896260148920954, | |
| "grad_norm": 0.40672293305397034, | |
| "learning_rate": 3.947926783793002e-05, | |
| "loss": 2.301, | |
| "step": 1752 | |
| }, | |
| { | |
| "epoch": 0.5902991039501914, | |
| "grad_norm": 0.44719624519348145, | |
| "learning_rate": 3.9370499816875724e-05, | |
| "loss": 2.2865, | |
| "step": 1754 | |
| }, | |
| { | |
| "epoch": 0.5909721930082874, | |
| "grad_norm": 0.43917497992515564, | |
| "learning_rate": 3.9261784452158514e-05, | |
| "loss": 2.2604, | |
| "step": 1756 | |
| }, | |
| { | |
| "epoch": 0.5916452820663834, | |
| "grad_norm": 0.37752288579940796, | |
| "learning_rate": 3.915312228233167e-05, | |
| "loss": 2.2627, | |
| "step": 1758 | |
| }, | |
| { | |
| "epoch": 0.5923183711244794, | |
| "grad_norm": 0.379291832447052, | |
| "learning_rate": 3.9044513845685065e-05, | |
| "loss": 2.308, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.5929914601825754, | |
| "grad_norm": 0.37401866912841797, | |
| "learning_rate": 3.893595968024235e-05, | |
| "loss": 2.2601, | |
| "step": 1762 | |
| }, | |
| { | |
| "epoch": 0.5936645492406714, | |
| "grad_norm": 0.39244258403778076, | |
| "learning_rate": 3.88274603237583e-05, | |
| "loss": 2.2949, | |
| "step": 1764 | |
| }, | |
| { | |
| "epoch": 0.5943376382987674, | |
| "grad_norm": 0.3765527904033661, | |
| "learning_rate": 3.87190163137162e-05, | |
| "loss": 2.2914, | |
| "step": 1766 | |
| }, | |
| { | |
| "epoch": 0.5950107273568634, | |
| "grad_norm": 0.39342641830444336, | |
| "learning_rate": 3.8610628187325195e-05, | |
| "loss": 2.2397, | |
| "step": 1768 | |
| }, | |
| { | |
| "epoch": 0.5956838164149594, | |
| "grad_norm": 0.36948031187057495, | |
| "learning_rate": 3.850229648151752e-05, | |
| "loss": 2.3014, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.5963569054730554, | |
| "grad_norm": 0.3758838176727295, | |
| "learning_rate": 3.8394021732945954e-05, | |
| "loss": 2.2858, | |
| "step": 1772 | |
| }, | |
| { | |
| "epoch": 0.5970299945311514, | |
| "grad_norm": 0.3759140968322754, | |
| "learning_rate": 3.8285804477981176e-05, | |
| "loss": 2.3087, | |
| "step": 1774 | |
| }, | |
| { | |
| "epoch": 0.5977030835892474, | |
| "grad_norm": 0.40178316831588745, | |
| "learning_rate": 3.817764525270893e-05, | |
| "loss": 2.2893, | |
| "step": 1776 | |
| }, | |
| { | |
| "epoch": 0.5983761726473434, | |
| "grad_norm": 0.389216810464859, | |
| "learning_rate": 3.8069544592927595e-05, | |
| "loss": 2.3023, | |
| "step": 1778 | |
| }, | |
| { | |
| "epoch": 0.5990492617054394, | |
| "grad_norm": 0.386371910572052, | |
| "learning_rate": 3.796150303414541e-05, | |
| "loss": 2.2999, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.5997223507635354, | |
| "grad_norm": 0.3694564700126648, | |
| "learning_rate": 3.7853521111577804e-05, | |
| "loss": 2.2895, | |
| "step": 1782 | |
| }, | |
| { | |
| "epoch": 0.6003954398216313, | |
| "grad_norm": 0.36854061484336853, | |
| "learning_rate": 3.77455993601448e-05, | |
| "loss": 2.3124, | |
| "step": 1784 | |
| }, | |
| { | |
| "epoch": 0.6010685288797274, | |
| "grad_norm": 0.3862304389476776, | |
| "learning_rate": 3.763773831446839e-05, | |
| "loss": 2.2679, | |
| "step": 1786 | |
| }, | |
| { | |
| "epoch": 0.6017416179378234, | |
| "grad_norm": 0.3964047133922577, | |
| "learning_rate": 3.752993850886973e-05, | |
| "loss": 2.2775, | |
| "step": 1788 | |
| }, | |
| { | |
| "epoch": 0.6024147069959194, | |
| "grad_norm": 0.42583465576171875, | |
| "learning_rate": 3.742220047736674e-05, | |
| "loss": 2.2609, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.6030877960540154, | |
| "grad_norm": 0.4133291244506836, | |
| "learning_rate": 3.731452475367123e-05, | |
| "loss": 2.268, | |
| "step": 1792 | |
| }, | |
| { | |
| "epoch": 0.6037608851121113, | |
| "grad_norm": 0.4323810935020447, | |
| "learning_rate": 3.720691187118638e-05, | |
| "loss": 2.2652, | |
| "step": 1794 | |
| }, | |
| { | |
| "epoch": 0.6044339741702074, | |
| "grad_norm": 0.4155171811580658, | |
| "learning_rate": 3.7099362363004076e-05, | |
| "loss": 2.3004, | |
| "step": 1796 | |
| }, | |
| { | |
| "epoch": 0.6051070632283034, | |
| "grad_norm": 0.40513670444488525, | |
| "learning_rate": 3.6991876761902255e-05, | |
| "loss": 2.2903, | |
| "step": 1798 | |
| }, | |
| { | |
| "epoch": 0.6057801522863994, | |
| "grad_norm": 0.4335862994194031, | |
| "learning_rate": 3.688445560034225e-05, | |
| "loss": 2.3042, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.6064532413444954, | |
| "grad_norm": 0.4351447820663452, | |
| "learning_rate": 3.67770994104662e-05, | |
| "loss": 2.3007, | |
| "step": 1802 | |
| }, | |
| { | |
| "epoch": 0.6071263304025913, | |
| "grad_norm": 0.4499759376049042, | |
| "learning_rate": 3.6669808724094424e-05, | |
| "loss": 2.2837, | |
| "step": 1804 | |
| }, | |
| { | |
| "epoch": 0.6077994194606874, | |
| "grad_norm": 0.4264863133430481, | |
| "learning_rate": 3.656258407272263e-05, | |
| "loss": 2.2729, | |
| "step": 1806 | |
| }, | |
| { | |
| "epoch": 0.6084725085187834, | |
| "grad_norm": 0.4422357976436615, | |
| "learning_rate": 3.6455425987519553e-05, | |
| "loss": 2.3026, | |
| "step": 1808 | |
| }, | |
| { | |
| "epoch": 0.6091455975768794, | |
| "grad_norm": 0.4904687702655792, | |
| "learning_rate": 3.634833499932403e-05, | |
| "loss": 2.2489, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.6098186866349754, | |
| "grad_norm": 0.4359835088253021, | |
| "learning_rate": 3.6241311638642634e-05, | |
| "loss": 2.263, | |
| "step": 1812 | |
| }, | |
| { | |
| "epoch": 0.6104917756930713, | |
| "grad_norm": 0.4219939112663269, | |
| "learning_rate": 3.6134356435646863e-05, | |
| "loss": 2.2819, | |
| "step": 1814 | |
| }, | |
| { | |
| "epoch": 0.6111648647511674, | |
| "grad_norm": 0.4202636182308197, | |
| "learning_rate": 3.602746992017058e-05, | |
| "loss": 2.264, | |
| "step": 1816 | |
| }, | |
| { | |
| "epoch": 0.6118379538092634, | |
| "grad_norm": 0.40362247824668884, | |
| "learning_rate": 3.592065262170739e-05, | |
| "loss": 2.3175, | |
| "step": 1818 | |
| }, | |
| { | |
| "epoch": 0.6125110428673594, | |
| "grad_norm": 0.3956090807914734, | |
| "learning_rate": 3.581390506940803e-05, | |
| "loss": 2.3121, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.6131841319254554, | |
| "grad_norm": 0.37391406297683716, | |
| "learning_rate": 3.570722779207767e-05, | |
| "loss": 2.2965, | |
| "step": 1822 | |
| }, | |
| { | |
| "epoch": 0.6138572209835513, | |
| "grad_norm": 0.3904911279678345, | |
| "learning_rate": 3.5600621318173436e-05, | |
| "loss": 2.2868, | |
| "step": 1824 | |
| }, | |
| { | |
| "epoch": 0.6145303100416474, | |
| "grad_norm": 0.41000476479530334, | |
| "learning_rate": 3.5494086175801666e-05, | |
| "loss": 2.2531, | |
| "step": 1826 | |
| }, | |
| { | |
| "epoch": 0.6152033990997434, | |
| "grad_norm": 0.4269038140773773, | |
| "learning_rate": 3.538762289271531e-05, | |
| "loss": 2.3034, | |
| "step": 1828 | |
| }, | |
| { | |
| "epoch": 0.6158764881578394, | |
| "grad_norm": 0.46176737546920776, | |
| "learning_rate": 3.5281231996311386e-05, | |
| "loss": 2.2755, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.6165495772159354, | |
| "grad_norm": 0.45788294076919556, | |
| "learning_rate": 3.517491401362831e-05, | |
| "loss": 2.2629, | |
| "step": 1832 | |
| }, | |
| { | |
| "epoch": 0.6172226662740313, | |
| "grad_norm": 0.47465789318084717, | |
| "learning_rate": 3.506866947134328e-05, | |
| "loss": 2.2769, | |
| "step": 1834 | |
| }, | |
| { | |
| "epoch": 0.6178957553321274, | |
| "grad_norm": 0.494191974401474, | |
| "learning_rate": 3.496249889576969e-05, | |
| "loss": 2.2551, | |
| "step": 1836 | |
| }, | |
| { | |
| "epoch": 0.6185688443902234, | |
| "grad_norm": 0.415902704000473, | |
| "learning_rate": 3.485640281285459e-05, | |
| "loss": 2.3252, | |
| "step": 1838 | |
| }, | |
| { | |
| "epoch": 0.6192419334483193, | |
| "grad_norm": 0.4104960262775421, | |
| "learning_rate": 3.475038174817587e-05, | |
| "loss": 2.2857, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.6199150225064154, | |
| "grad_norm": 0.421204149723053, | |
| "learning_rate": 3.4644436226939915e-05, | |
| "loss": 2.2967, | |
| "step": 1842 | |
| }, | |
| { | |
| "epoch": 0.6205881115645113, | |
| "grad_norm": 0.4209441542625427, | |
| "learning_rate": 3.4538566773978844e-05, | |
| "loss": 2.3072, | |
| "step": 1844 | |
| }, | |
| { | |
| "epoch": 0.6212612006226074, | |
| "grad_norm": 0.4308975338935852, | |
| "learning_rate": 3.443277391374792e-05, | |
| "loss": 2.2736, | |
| "step": 1846 | |
| }, | |
| { | |
| "epoch": 0.6219342896807034, | |
| "grad_norm": 0.4085467457771301, | |
| "learning_rate": 3.4327058170323036e-05, | |
| "loss": 2.2921, | |
| "step": 1848 | |
| }, | |
| { | |
| "epoch": 0.6226073787387993, | |
| "grad_norm": 0.4512941837310791, | |
| "learning_rate": 3.4221420067398024e-05, | |
| "loss": 2.2672, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.6232804677968954, | |
| "grad_norm": 0.37834271788597107, | |
| "learning_rate": 3.4115860128282103e-05, | |
| "loss": 2.2466, | |
| "step": 1852 | |
| }, | |
| { | |
| "epoch": 0.6239535568549914, | |
| "grad_norm": 0.3885023295879364, | |
| "learning_rate": 3.401037887589731e-05, | |
| "loss": 2.2632, | |
| "step": 1854 | |
| }, | |
| { | |
| "epoch": 0.6246266459130874, | |
| "grad_norm": 0.41913384199142456, | |
| "learning_rate": 3.3904976832775894e-05, | |
| "loss": 2.2625, | |
| "step": 1856 | |
| }, | |
| { | |
| "epoch": 0.6252997349711834, | |
| "grad_norm": 0.3997589647769928, | |
| "learning_rate": 3.379965452105763e-05, | |
| "loss": 2.2701, | |
| "step": 1858 | |
| }, | |
| { | |
| "epoch": 0.6259728240292793, | |
| "grad_norm": 0.3963086009025574, | |
| "learning_rate": 3.3694412462487436e-05, | |
| "loss": 2.2635, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.6266459130873754, | |
| "grad_norm": 0.39688992500305176, | |
| "learning_rate": 3.3589251178412616e-05, | |
| "loss": 2.2917, | |
| "step": 1862 | |
| }, | |
| { | |
| "epoch": 0.6273190021454714, | |
| "grad_norm": 0.38178861141204834, | |
| "learning_rate": 3.348417118978031e-05, | |
| "loss": 2.2731, | |
| "step": 1864 | |
| }, | |
| { | |
| "epoch": 0.6279920912035674, | |
| "grad_norm": 0.39461877942085266, | |
| "learning_rate": 3.337917301713499e-05, | |
| "loss": 2.293, | |
| "step": 1866 | |
| }, | |
| { | |
| "epoch": 0.6286651802616634, | |
| "grad_norm": 0.38475388288497925, | |
| "learning_rate": 3.32742571806158e-05, | |
| "loss": 2.2885, | |
| "step": 1868 | |
| }, | |
| { | |
| "epoch": 0.6293382693197593, | |
| "grad_norm": 0.4194738566875458, | |
| "learning_rate": 3.316942419995399e-05, | |
| "loss": 2.2762, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.6300113583778554, | |
| "grad_norm": 0.421724796295166, | |
| "learning_rate": 3.306467459447041e-05, | |
| "loss": 2.2536, | |
| "step": 1872 | |
| }, | |
| { | |
| "epoch": 0.6306844474359514, | |
| "grad_norm": 0.3988567590713501, | |
| "learning_rate": 3.2960008883072856e-05, | |
| "loss": 2.2885, | |
| "step": 1874 | |
| }, | |
| { | |
| "epoch": 0.6313575364940474, | |
| "grad_norm": 0.3877731263637543, | |
| "learning_rate": 3.285542758425351e-05, | |
| "loss": 2.2979, | |
| "step": 1876 | |
| }, | |
| { | |
| "epoch": 0.6320306255521434, | |
| "grad_norm": 0.36148571968078613, | |
| "learning_rate": 3.2750931216086436e-05, | |
| "loss": 2.2717, | |
| "step": 1878 | |
| }, | |
| { | |
| "epoch": 0.6327037146102393, | |
| "grad_norm": 0.40438732504844666, | |
| "learning_rate": 3.264652029622493e-05, | |
| "loss": 2.2656, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.6333768036683354, | |
| "grad_norm": 0.3884246051311493, | |
| "learning_rate": 3.254219534189903e-05, | |
| "loss": 2.2506, | |
| "step": 1882 | |
| }, | |
| { | |
| "epoch": 0.6340498927264314, | |
| "grad_norm": 0.412960022687912, | |
| "learning_rate": 3.243795686991287e-05, | |
| "loss": 2.3049, | |
| "step": 1884 | |
| }, | |
| { | |
| "epoch": 0.6347229817845274, | |
| "grad_norm": 0.4141545295715332, | |
| "learning_rate": 3.2333805396642245e-05, | |
| "loss": 2.2606, | |
| "step": 1886 | |
| }, | |
| { | |
| "epoch": 0.6353960708426234, | |
| "grad_norm": 0.39594951272010803, | |
| "learning_rate": 3.222974143803188e-05, | |
| "loss": 2.3044, | |
| "step": 1888 | |
| }, | |
| { | |
| "epoch": 0.6360691599007193, | |
| "grad_norm": 0.372408002614975, | |
| "learning_rate": 3.212576550959304e-05, | |
| "loss": 2.272, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.6367422489588154, | |
| "grad_norm": 0.3790285289287567, | |
| "learning_rate": 3.2021878126400916e-05, | |
| "loss": 2.2689, | |
| "step": 1892 | |
| }, | |
| { | |
| "epoch": 0.6374153380169114, | |
| "grad_norm": 0.3891488313674927, | |
| "learning_rate": 3.191807980309199e-05, | |
| "loss": 2.2703, | |
| "step": 1894 | |
| }, | |
| { | |
| "epoch": 0.6380884270750073, | |
| "grad_norm": 0.3948821425437927, | |
| "learning_rate": 3.181437105386164e-05, | |
| "loss": 2.2719, | |
| "step": 1896 | |
| }, | |
| { | |
| "epoch": 0.6387615161331034, | |
| "grad_norm": 0.4036186635494232, | |
| "learning_rate": 3.1710752392461464e-05, | |
| "loss": 2.2787, | |
| "step": 1898 | |
| }, | |
| { | |
| "epoch": 0.6394346051911993, | |
| "grad_norm": 0.3524232506752014, | |
| "learning_rate": 3.160722433219679e-05, | |
| "loss": 2.2778, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.6401076942492954, | |
| "grad_norm": 0.39453738927841187, | |
| "learning_rate": 3.1503787385924155e-05, | |
| "loss": 2.2893, | |
| "step": 1902 | |
| }, | |
| { | |
| "epoch": 0.6407807833073914, | |
| "grad_norm": 0.37081316113471985, | |
| "learning_rate": 3.1400442066048695e-05, | |
| "loss": 2.3099, | |
| "step": 1904 | |
| }, | |
| { | |
| "epoch": 0.6414538723654873, | |
| "grad_norm": 0.3991096317768097, | |
| "learning_rate": 3.129718888452166e-05, | |
| "loss": 2.2747, | |
| "step": 1906 | |
| }, | |
| { | |
| "epoch": 0.6421269614235834, | |
| "grad_norm": 0.3787843585014343, | |
| "learning_rate": 3.119402835283792e-05, | |
| "loss": 2.2386, | |
| "step": 1908 | |
| }, | |
| { | |
| "epoch": 0.6428000504816793, | |
| "grad_norm": 0.3759045898914337, | |
| "learning_rate": 3.1090960982033284e-05, | |
| "loss": 2.25, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.6434731395397754, | |
| "grad_norm": 0.3775229752063751, | |
| "learning_rate": 3.098798728268213e-05, | |
| "loss": 2.2664, | |
| "step": 1912 | |
| }, | |
| { | |
| "epoch": 0.6441462285978714, | |
| "grad_norm": 0.37777119874954224, | |
| "learning_rate": 3.088510776489479e-05, | |
| "loss": 2.2917, | |
| "step": 1914 | |
| }, | |
| { | |
| "epoch": 0.6448193176559673, | |
| "grad_norm": 0.3809587061405182, | |
| "learning_rate": 3.0782322938315023e-05, | |
| "loss": 2.2637, | |
| "step": 1916 | |
| }, | |
| { | |
| "epoch": 0.6454924067140634, | |
| "grad_norm": 0.3507981598377228, | |
| "learning_rate": 3.067963331211752e-05, | |
| "loss": 2.2381, | |
| "step": 1918 | |
| }, | |
| { | |
| "epoch": 0.6461654957721593, | |
| "grad_norm": 0.3526877462863922, | |
| "learning_rate": 3.0577039395005394e-05, | |
| "loss": 2.2455, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.6468385848302554, | |
| "grad_norm": 0.37659838795661926, | |
| "learning_rate": 3.047454169520756e-05, | |
| "loss": 2.2444, | |
| "step": 1922 | |
| }, | |
| { | |
| "epoch": 0.6475116738883514, | |
| "grad_norm": 0.35282251238822937, | |
| "learning_rate": 3.0372140720476396e-05, | |
| "loss": 2.2563, | |
| "step": 1924 | |
| }, | |
| { | |
| "epoch": 0.6481847629464473, | |
| "grad_norm": 0.37368044257164, | |
| "learning_rate": 3.0269836978085042e-05, | |
| "loss": 2.2949, | |
| "step": 1926 | |
| }, | |
| { | |
| "epoch": 0.6488578520045434, | |
| "grad_norm": 0.401157021522522, | |
| "learning_rate": 3.0167630974825002e-05, | |
| "loss": 2.2978, | |
| "step": 1928 | |
| }, | |
| { | |
| "epoch": 0.6495309410626393, | |
| "grad_norm": 0.3910830318927765, | |
| "learning_rate": 3.00655232170036e-05, | |
| "loss": 2.2699, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.6502040301207354, | |
| "grad_norm": 0.3779541254043579, | |
| "learning_rate": 2.996351421044148e-05, | |
| "loss": 2.2641, | |
| "step": 1932 | |
| }, | |
| { | |
| "epoch": 0.6508771191788314, | |
| "grad_norm": 0.36932674050331116, | |
| "learning_rate": 2.9861604460470054e-05, | |
| "loss": 2.3094, | |
| "step": 1934 | |
| }, | |
| { | |
| "epoch": 0.6515502082369273, | |
| "grad_norm": 0.3909669816493988, | |
| "learning_rate": 2.975979447192907e-05, | |
| "loss": 2.2742, | |
| "step": 1936 | |
| }, | |
| { | |
| "epoch": 0.6522232972950234, | |
| "grad_norm": 0.3832036256790161, | |
| "learning_rate": 2.965808474916412e-05, | |
| "loss": 2.281, | |
| "step": 1938 | |
| }, | |
| { | |
| "epoch": 0.6528963863531193, | |
| "grad_norm": 0.38232412934303284, | |
| "learning_rate": 2.9556475796023985e-05, | |
| "loss": 2.2634, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.6535694754112154, | |
| "grad_norm": 0.38755351305007935, | |
| "learning_rate": 2.9454968115858374e-05, | |
| "loss": 2.2544, | |
| "step": 1942 | |
| }, | |
| { | |
| "epoch": 0.6542425644693114, | |
| "grad_norm": 0.4095642566680908, | |
| "learning_rate": 2.9353562211515246e-05, | |
| "loss": 2.2812, | |
| "step": 1944 | |
| }, | |
| { | |
| "epoch": 0.6549156535274073, | |
| "grad_norm": 0.3892836272716522, | |
| "learning_rate": 2.9252258585338387e-05, | |
| "loss": 2.2671, | |
| "step": 1946 | |
| }, | |
| { | |
| "epoch": 0.6555887425855034, | |
| "grad_norm": 0.41320663690567017, | |
| "learning_rate": 2.9151057739164933e-05, | |
| "loss": 2.2601, | |
| "step": 1948 | |
| }, | |
| { | |
| "epoch": 0.6562618316435993, | |
| "grad_norm": 0.3720477223396301, | |
| "learning_rate": 2.9049960174322876e-05, | |
| "loss": 2.2912, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.6569349207016953, | |
| "grad_norm": 0.3487861156463623, | |
| "learning_rate": 2.894896639162854e-05, | |
| "loss": 2.2537, | |
| "step": 1952 | |
| }, | |
| { | |
| "epoch": 0.6576080097597914, | |
| "grad_norm": 0.41094574332237244, | |
| "learning_rate": 2.8848076891384196e-05, | |
| "loss": 2.2645, | |
| "step": 1954 | |
| }, | |
| { | |
| "epoch": 0.6582810988178873, | |
| "grad_norm": 0.3451893627643585, | |
| "learning_rate": 2.874729217337545e-05, | |
| "loss": 2.2353, | |
| "step": 1956 | |
| }, | |
| { | |
| "epoch": 0.6589541878759834, | |
| "grad_norm": 0.36058786511421204, | |
| "learning_rate": 2.864661273686886e-05, | |
| "loss": 2.2692, | |
| "step": 1958 | |
| }, | |
| { | |
| "epoch": 0.6596272769340793, | |
| "grad_norm": 0.37918969988822937, | |
| "learning_rate": 2.8546039080609477e-05, | |
| "loss": 2.2471, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.6603003659921753, | |
| "grad_norm": 0.37483328580856323, | |
| "learning_rate": 2.8445571702818342e-05, | |
| "loss": 2.2872, | |
| "step": 1962 | |
| }, | |
| { | |
| "epoch": 0.6609734550502714, | |
| "grad_norm": 0.37265482544898987, | |
| "learning_rate": 2.8345211101189904e-05, | |
| "loss": 2.2493, | |
| "step": 1964 | |
| }, | |
| { | |
| "epoch": 0.6616465441083673, | |
| "grad_norm": 0.3745740056037903, | |
| "learning_rate": 2.8244957772889803e-05, | |
| "loss": 2.2593, | |
| "step": 1966 | |
| }, | |
| { | |
| "epoch": 0.6623196331664634, | |
| "grad_norm": 0.3740994334220886, | |
| "learning_rate": 2.8144812214552217e-05, | |
| "loss": 2.2502, | |
| "step": 1968 | |
| }, | |
| { | |
| "epoch": 0.6629927222245593, | |
| "grad_norm": 0.370441734790802, | |
| "learning_rate": 2.804477492227744e-05, | |
| "loss": 2.2907, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.6636658112826553, | |
| "grad_norm": 0.37232956290245056, | |
| "learning_rate": 2.7944846391629427e-05, | |
| "loss": 2.2359, | |
| "step": 1972 | |
| }, | |
| { | |
| "epoch": 0.6643389003407514, | |
| "grad_norm": 0.3705579340457916, | |
| "learning_rate": 2.7845027117633404e-05, | |
| "loss": 2.2564, | |
| "step": 1974 | |
| }, | |
| { | |
| "epoch": 0.6650119893988473, | |
| "grad_norm": 0.340466171503067, | |
| "learning_rate": 2.77453175947733e-05, | |
| "loss": 2.2411, | |
| "step": 1976 | |
| }, | |
| { | |
| "epoch": 0.6656850784569434, | |
| "grad_norm": 0.367044597864151, | |
| "learning_rate": 2.7645718316989434e-05, | |
| "loss": 2.2738, | |
| "step": 1978 | |
| }, | |
| { | |
| "epoch": 0.6663581675150393, | |
| "grad_norm": 0.3702086806297302, | |
| "learning_rate": 2.7546229777675914e-05, | |
| "loss": 2.2657, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.6670312565731353, | |
| "grad_norm": 0.3724825978279114, | |
| "learning_rate": 2.7446852469678297e-05, | |
| "loss": 2.237, | |
| "step": 1982 | |
| }, | |
| { | |
| "epoch": 0.6677043456312314, | |
| "grad_norm": 0.3776718080043793, | |
| "learning_rate": 2.7347586885291176e-05, | |
| "loss": 2.2748, | |
| "step": 1984 | |
| }, | |
| { | |
| "epoch": 0.6683774346893273, | |
| "grad_norm": 0.3803219795227051, | |
| "learning_rate": 2.7248433516255633e-05, | |
| "loss": 2.2717, | |
| "step": 1986 | |
| }, | |
| { | |
| "epoch": 0.6690505237474234, | |
| "grad_norm": 0.38277125358581543, | |
| "learning_rate": 2.714939285375684e-05, | |
| "loss": 2.2349, | |
| "step": 1988 | |
| }, | |
| { | |
| "epoch": 0.6697236128055193, | |
| "grad_norm": 0.3710474371910095, | |
| "learning_rate": 2.7050465388421716e-05, | |
| "loss": 2.2445, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.6703967018636153, | |
| "grad_norm": 0.36977726221084595, | |
| "learning_rate": 2.6951651610316374e-05, | |
| "loss": 2.2856, | |
| "step": 1992 | |
| }, | |
| { | |
| "epoch": 0.6710697909217114, | |
| "grad_norm": 0.37834489345550537, | |
| "learning_rate": 2.6852952008943733e-05, | |
| "loss": 2.2603, | |
| "step": 1994 | |
| }, | |
| { | |
| "epoch": 0.6717428799798073, | |
| "grad_norm": 0.36093804240226746, | |
| "learning_rate": 2.675436707324116e-05, | |
| "loss": 2.2417, | |
| "step": 1996 | |
| }, | |
| { | |
| "epoch": 0.6724159690379033, | |
| "grad_norm": 0.38005608320236206, | |
| "learning_rate": 2.6655897291577934e-05, | |
| "loss": 2.2638, | |
| "step": 1998 | |
| }, | |
| { | |
| "epoch": 0.6730890580959993, | |
| "grad_norm": 0.3814466595649719, | |
| "learning_rate": 2.6557543151752934e-05, | |
| "loss": 2.2193, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.6737621471540953, | |
| "grad_norm": 0.37796446681022644, | |
| "learning_rate": 2.6459305140992137e-05, | |
| "loss": 2.3008, | |
| "step": 2002 | |
| }, | |
| { | |
| "epoch": 0.6744352362121914, | |
| "grad_norm": 0.3829179108142853, | |
| "learning_rate": 2.6361183745946227e-05, | |
| "loss": 2.2524, | |
| "step": 2004 | |
| }, | |
| { | |
| "epoch": 0.6751083252702873, | |
| "grad_norm": 0.39577358961105347, | |
| "learning_rate": 2.626317945268827e-05, | |
| "loss": 2.2853, | |
| "step": 2006 | |
| }, | |
| { | |
| "epoch": 0.6757814143283833, | |
| "grad_norm": 0.4012986123561859, | |
| "learning_rate": 2.6165292746711168e-05, | |
| "loss": 2.258, | |
| "step": 2008 | |
| }, | |
| { | |
| "epoch": 0.6764545033864793, | |
| "grad_norm": 0.39091506600379944, | |
| "learning_rate": 2.6067524112925307e-05, | |
| "loss": 2.2566, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.6771275924445753, | |
| "grad_norm": 0.3839189112186432, | |
| "learning_rate": 2.5969874035656237e-05, | |
| "loss": 2.2839, | |
| "step": 2012 | |
| }, | |
| { | |
| "epoch": 0.6778006815026714, | |
| "grad_norm": 0.4070791006088257, | |
| "learning_rate": 2.5872342998642173e-05, | |
| "loss": 2.2573, | |
| "step": 2014 | |
| }, | |
| { | |
| "epoch": 0.6784737705607673, | |
| "grad_norm": 0.3998067080974579, | |
| "learning_rate": 2.5774931485031616e-05, | |
| "loss": 2.2583, | |
| "step": 2016 | |
| }, | |
| { | |
| "epoch": 0.6791468596188633, | |
| "grad_norm": 0.4076973795890808, | |
| "learning_rate": 2.5677639977380953e-05, | |
| "loss": 2.2276, | |
| "step": 2018 | |
| }, | |
| { | |
| "epoch": 0.6798199486769593, | |
| "grad_norm": 0.3782246708869934, | |
| "learning_rate": 2.558046895765215e-05, | |
| "loss": 2.2506, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.6804930377350553, | |
| "grad_norm": 0.3559093475341797, | |
| "learning_rate": 2.5483418907210254e-05, | |
| "loss": 2.2864, | |
| "step": 2022 | |
| }, | |
| { | |
| "epoch": 0.6811661267931514, | |
| "grad_norm": 0.3624304234981537, | |
| "learning_rate": 2.5386490306821037e-05, | |
| "loss": 2.306, | |
| "step": 2024 | |
| }, | |
| { | |
| "epoch": 0.6818392158512473, | |
| "grad_norm": 0.3587932884693146, | |
| "learning_rate": 2.528968363664871e-05, | |
| "loss": 2.2891, | |
| "step": 2026 | |
| }, | |
| { | |
| "epoch": 0.6825123049093433, | |
| "grad_norm": 0.3958047330379486, | |
| "learning_rate": 2.5192999376253368e-05, | |
| "loss": 2.2735, | |
| "step": 2028 | |
| }, | |
| { | |
| "epoch": 0.6831853939674393, | |
| "grad_norm": 0.37902015447616577, | |
| "learning_rate": 2.5096438004588818e-05, | |
| "loss": 2.2716, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.6838584830255353, | |
| "grad_norm": 0.3787776231765747, | |
| "learning_rate": 2.500000000000001e-05, | |
| "loss": 2.274, | |
| "step": 2032 | |
| }, | |
| { | |
| "epoch": 0.6845315720836314, | |
| "grad_norm": 0.3875786364078522, | |
| "learning_rate": 2.4903685840220792e-05, | |
| "loss": 2.2205, | |
| "step": 2034 | |
| }, | |
| { | |
| "epoch": 0.6852046611417273, | |
| "grad_norm": 0.39311864972114563, | |
| "learning_rate": 2.4807496002371548e-05, | |
| "loss": 2.303, | |
| "step": 2036 | |
| }, | |
| { | |
| "epoch": 0.6858777501998233, | |
| "grad_norm": 0.3716079294681549, | |
| "learning_rate": 2.4711430962956745e-05, | |
| "loss": 2.2414, | |
| "step": 2038 | |
| }, | |
| { | |
| "epoch": 0.6865508392579193, | |
| "grad_norm": 0.3402949869632721, | |
| "learning_rate": 2.461549119786262e-05, | |
| "loss": 2.2877, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.6872239283160153, | |
| "grad_norm": 0.39148789644241333, | |
| "learning_rate": 2.4519677182354883e-05, | |
| "loss": 2.2458, | |
| "step": 2042 | |
| }, | |
| { | |
| "epoch": 0.6878970173741114, | |
| "grad_norm": 0.40696027874946594, | |
| "learning_rate": 2.4423989391076296e-05, | |
| "loss": 2.2518, | |
| "step": 2044 | |
| }, | |
| { | |
| "epoch": 0.6885701064322073, | |
| "grad_norm": 0.37364059686660767, | |
| "learning_rate": 2.432842829804425e-05, | |
| "loss": 2.2674, | |
| "step": 2046 | |
| }, | |
| { | |
| "epoch": 0.6892431954903033, | |
| "grad_norm": 0.3884735107421875, | |
| "learning_rate": 2.4232994376648582e-05, | |
| "loss": 2.2851, | |
| "step": 2048 | |
| }, | |
| { | |
| "epoch": 0.6899162845483993, | |
| "grad_norm": 0.3796365261077881, | |
| "learning_rate": 2.413768809964916e-05, | |
| "loss": 2.267, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.6905893736064953, | |
| "grad_norm": 0.38148432970046997, | |
| "learning_rate": 2.4042509939173463e-05, | |
| "loss": 2.2518, | |
| "step": 2052 | |
| }, | |
| { | |
| "epoch": 0.6912624626645913, | |
| "grad_norm": 0.3781225085258484, | |
| "learning_rate": 2.3947460366714323e-05, | |
| "loss": 2.241, | |
| "step": 2054 | |
| }, | |
| { | |
| "epoch": 0.6919355517226873, | |
| "grad_norm": 0.40383458137512207, | |
| "learning_rate": 2.3852539853127626e-05, | |
| "loss": 2.2757, | |
| "step": 2056 | |
| }, | |
| { | |
| "epoch": 0.6926086407807833, | |
| "grad_norm": 0.40038955211639404, | |
| "learning_rate": 2.375774886862987e-05, | |
| "loss": 2.2393, | |
| "step": 2058 | |
| }, | |
| { | |
| "epoch": 0.6932817298388793, | |
| "grad_norm": 0.38713014125823975, | |
| "learning_rate": 2.3663087882795886e-05, | |
| "loss": 2.2698, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.6939548188969753, | |
| "grad_norm": 0.37371134757995605, | |
| "learning_rate": 2.35685573645566e-05, | |
| "loss": 2.283, | |
| "step": 2062 | |
| }, | |
| { | |
| "epoch": 0.6946279079550713, | |
| "grad_norm": 0.37300777435302734, | |
| "learning_rate": 2.3474157782196514e-05, | |
| "loss": 2.2318, | |
| "step": 2064 | |
| }, | |
| { | |
| "epoch": 0.6953009970131673, | |
| "grad_norm": 0.3749702572822571, | |
| "learning_rate": 2.3379889603351617e-05, | |
| "loss": 2.2395, | |
| "step": 2066 | |
| }, | |
| { | |
| "epoch": 0.6959740860712633, | |
| "grad_norm": 0.37506356835365295, | |
| "learning_rate": 2.328575329500687e-05, | |
| "loss": 2.2519, | |
| "step": 2068 | |
| }, | |
| { | |
| "epoch": 0.6966471751293593, | |
| "grad_norm": 0.362093061208725, | |
| "learning_rate": 2.3191749323493972e-05, | |
| "loss": 2.2784, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.6973202641874553, | |
| "grad_norm": 0.3932085335254669, | |
| "learning_rate": 2.309787815448911e-05, | |
| "loss": 2.2183, | |
| "step": 2072 | |
| }, | |
| { | |
| "epoch": 0.6979933532455513, | |
| "grad_norm": 0.3700029253959656, | |
| "learning_rate": 2.3004140253010603e-05, | |
| "loss": 2.2329, | |
| "step": 2074 | |
| }, | |
| { | |
| "epoch": 0.6986664423036473, | |
| "grad_norm": 0.3903138041496277, | |
| "learning_rate": 2.2910536083416478e-05, | |
| "loss": 2.2463, | |
| "step": 2076 | |
| }, | |
| { | |
| "epoch": 0.6993395313617433, | |
| "grad_norm": 0.3758189082145691, | |
| "learning_rate": 2.281706610940241e-05, | |
| "loss": 2.2615, | |
| "step": 2078 | |
| }, | |
| { | |
| "epoch": 0.7000126204198392, | |
| "grad_norm": 0.3749825060367584, | |
| "learning_rate": 2.272373079399921e-05, | |
| "loss": 2.282, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.7006857094779353, | |
| "grad_norm": 0.36719000339508057, | |
| "learning_rate": 2.2630530599570695e-05, | |
| "loss": 2.2534, | |
| "step": 2082 | |
| }, | |
| { | |
| "epoch": 0.7013587985360313, | |
| "grad_norm": 0.38023319840431213, | |
| "learning_rate": 2.2537465987811246e-05, | |
| "loss": 2.2859, | |
| "step": 2084 | |
| }, | |
| { | |
| "epoch": 0.7020318875941273, | |
| "grad_norm": 0.37650057673454285, | |
| "learning_rate": 2.2444537419743622e-05, | |
| "loss": 2.3039, | |
| "step": 2086 | |
| }, | |
| { | |
| "epoch": 0.7027049766522233, | |
| "grad_norm": 0.3848130404949188, | |
| "learning_rate": 2.2351745355716685e-05, | |
| "loss": 2.2191, | |
| "step": 2088 | |
| }, | |
| { | |
| "epoch": 0.7033780657103192, | |
| "grad_norm": 0.3667711019515991, | |
| "learning_rate": 2.2259090255403044e-05, | |
| "loss": 2.2914, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.7040511547684153, | |
| "grad_norm": 0.3482600748538971, | |
| "learning_rate": 2.2166572577796802e-05, | |
| "loss": 2.2745, | |
| "step": 2092 | |
| }, | |
| { | |
| "epoch": 0.7047242438265113, | |
| "grad_norm": 0.3584803640842438, | |
| "learning_rate": 2.2074192781211356e-05, | |
| "loss": 2.2799, | |
| "step": 2094 | |
| }, | |
| { | |
| "epoch": 0.7053973328846073, | |
| "grad_norm": 0.3784869611263275, | |
| "learning_rate": 2.198195132327706e-05, | |
| "loss": 2.2464, | |
| "step": 2096 | |
| }, | |
| { | |
| "epoch": 0.7060704219427033, | |
| "grad_norm": 0.35275179147720337, | |
| "learning_rate": 2.188984866093889e-05, | |
| "loss": 2.2826, | |
| "step": 2098 | |
| }, | |
| { | |
| "epoch": 0.7067435110007992, | |
| "grad_norm": 0.3402944803237915, | |
| "learning_rate": 2.1797885250454336e-05, | |
| "loss": 2.2406, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.7074166000588953, | |
| "grad_norm": 0.3486901819705963, | |
| "learning_rate": 2.1706061547391067e-05, | |
| "loss": 2.2424, | |
| "step": 2102 | |
| }, | |
| { | |
| "epoch": 0.7080896891169913, | |
| "grad_norm": 0.35420867800712585, | |
| "learning_rate": 2.161437800662462e-05, | |
| "loss": 2.2597, | |
| "step": 2104 | |
| }, | |
| { | |
| "epoch": 0.7087627781750873, | |
| "grad_norm": 0.35805556178092957, | |
| "learning_rate": 2.1522835082336203e-05, | |
| "loss": 2.2356, | |
| "step": 2106 | |
| }, | |
| { | |
| "epoch": 0.7094358672331833, | |
| "grad_norm": 0.37050601840019226, | |
| "learning_rate": 2.1431433228010495e-05, | |
| "loss": 2.2569, | |
| "step": 2108 | |
| }, | |
| { | |
| "epoch": 0.7101089562912792, | |
| "grad_norm": 0.37298470735549927, | |
| "learning_rate": 2.1340172896433274e-05, | |
| "loss": 2.2993, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.7107820453493753, | |
| "grad_norm": 0.3525806963443756, | |
| "learning_rate": 2.1249054539689306e-05, | |
| "loss": 2.2822, | |
| "step": 2112 | |
| }, | |
| { | |
| "epoch": 0.7114551344074713, | |
| "grad_norm": 0.3484457731246948, | |
| "learning_rate": 2.1158078609159997e-05, | |
| "loss": 2.2202, | |
| "step": 2114 | |
| }, | |
| { | |
| "epoch": 0.7121282234655673, | |
| "grad_norm": 0.3556087911128998, | |
| "learning_rate": 2.1067245555521202e-05, | |
| "loss": 2.2332, | |
| "step": 2116 | |
| }, | |
| { | |
| "epoch": 0.7128013125236633, | |
| "grad_norm": 0.3609943091869354, | |
| "learning_rate": 2.097655582874104e-05, | |
| "loss": 2.2598, | |
| "step": 2118 | |
| }, | |
| { | |
| "epoch": 0.7134744015817592, | |
| "grad_norm": 0.3892589807510376, | |
| "learning_rate": 2.0886009878077583e-05, | |
| "loss": 2.2588, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.7141474906398553, | |
| "grad_norm": 0.34478136897087097, | |
| "learning_rate": 2.079560815207664e-05, | |
| "loss": 2.2518, | |
| "step": 2122 | |
| }, | |
| { | |
| "epoch": 0.7148205796979513, | |
| "grad_norm": 0.3567972481250763, | |
| "learning_rate": 2.070535109856962e-05, | |
| "loss": 2.2532, | |
| "step": 2124 | |
| }, | |
| { | |
| "epoch": 0.7154936687560473, | |
| "grad_norm": 0.35243430733680725, | |
| "learning_rate": 2.0615239164671257e-05, | |
| "loss": 2.2246, | |
| "step": 2126 | |
| }, | |
| { | |
| "epoch": 0.7161667578141433, | |
| "grad_norm": 0.34242430329322815, | |
| "learning_rate": 2.05252727967773e-05, | |
| "loss": 2.2331, | |
| "step": 2128 | |
| }, | |
| { | |
| "epoch": 0.7168398468722392, | |
| "grad_norm": 0.3531624674797058, | |
| "learning_rate": 2.0435452440562486e-05, | |
| "loss": 2.2581, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.7175129359303353, | |
| "grad_norm": 0.39661160111427307, | |
| "learning_rate": 2.0345778540978226e-05, | |
| "loss": 2.2561, | |
| "step": 2132 | |
| }, | |
| { | |
| "epoch": 0.7181860249884313, | |
| "grad_norm": 0.3662177622318268, | |
| "learning_rate": 2.025625154225039e-05, | |
| "loss": 2.2464, | |
| "step": 2134 | |
| }, | |
| { | |
| "epoch": 0.7188591140465272, | |
| "grad_norm": 0.37029245495796204, | |
| "learning_rate": 2.0166871887877126e-05, | |
| "loss": 2.2844, | |
| "step": 2136 | |
| }, | |
| { | |
| "epoch": 0.7195322031046233, | |
| "grad_norm": 0.35741955041885376, | |
| "learning_rate": 2.0077640020626714e-05, | |
| "loss": 2.2493, | |
| "step": 2138 | |
| }, | |
| { | |
| "epoch": 0.7202052921627193, | |
| "grad_norm": 0.3547651767730713, | |
| "learning_rate": 1.9988556382535285e-05, | |
| "loss": 2.2592, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.7208783812208153, | |
| "grad_norm": 0.3595333695411682, | |
| "learning_rate": 1.989962141490467e-05, | |
| "loss": 2.2684, | |
| "step": 2142 | |
| }, | |
| { | |
| "epoch": 0.7215514702789113, | |
| "grad_norm": 0.365326464176178, | |
| "learning_rate": 1.9810835558300262e-05, | |
| "loss": 2.2691, | |
| "step": 2144 | |
| }, | |
| { | |
| "epoch": 0.7222245593370072, | |
| "grad_norm": 0.3665175437927246, | |
| "learning_rate": 1.9722199252548734e-05, | |
| "loss": 2.2787, | |
| "step": 2146 | |
| }, | |
| { | |
| "epoch": 0.7228976483951033, | |
| "grad_norm": 0.360385537147522, | |
| "learning_rate": 1.9633712936735977e-05, | |
| "loss": 2.2558, | |
| "step": 2148 | |
| }, | |
| { | |
| "epoch": 0.7235707374531993, | |
| "grad_norm": 0.359365850687027, | |
| "learning_rate": 1.954537704920481e-05, | |
| "loss": 2.2286, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.7242438265112953, | |
| "grad_norm": 0.3630097508430481, | |
| "learning_rate": 1.9457192027552856e-05, | |
| "loss": 2.2216, | |
| "step": 2152 | |
| }, | |
| { | |
| "epoch": 0.7249169155693913, | |
| "grad_norm": 0.3533220887184143, | |
| "learning_rate": 1.936915830863043e-05, | |
| "loss": 2.2625, | |
| "step": 2154 | |
| }, | |
| { | |
| "epoch": 0.7255900046274872, | |
| "grad_norm": 0.3711426854133606, | |
| "learning_rate": 1.9281276328538333e-05, | |
| "loss": 2.2508, | |
| "step": 2156 | |
| }, | |
| { | |
| "epoch": 0.7262630936855833, | |
| "grad_norm": 0.3657391667366028, | |
| "learning_rate": 1.919354652262557e-05, | |
| "loss": 2.2723, | |
| "step": 2158 | |
| }, | |
| { | |
| "epoch": 0.7269361827436793, | |
| "grad_norm": 0.3925938010215759, | |
| "learning_rate": 1.910596932548743e-05, | |
| "loss": 2.2504, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.7276092718017753, | |
| "grad_norm": 0.37426644563674927, | |
| "learning_rate": 1.9018545170963183e-05, | |
| "loss": 2.2724, | |
| "step": 2162 | |
| }, | |
| { | |
| "epoch": 0.7282823608598713, | |
| "grad_norm": 0.3587762415409088, | |
| "learning_rate": 1.893127449213391e-05, | |
| "loss": 2.2583, | |
| "step": 2164 | |
| }, | |
| { | |
| "epoch": 0.7289554499179672, | |
| "grad_norm": 0.36685168743133545, | |
| "learning_rate": 1.8844157721320428e-05, | |
| "loss": 2.2852, | |
| "step": 2166 | |
| }, | |
| { | |
| "epoch": 0.7296285389760633, | |
| "grad_norm": 0.3637944459915161, | |
| "learning_rate": 1.8757195290081165e-05, | |
| "loss": 2.2548, | |
| "step": 2168 | |
| }, | |
| { | |
| "epoch": 0.7303016280341593, | |
| "grad_norm": 0.3491983115673065, | |
| "learning_rate": 1.8670387629209935e-05, | |
| "loss": 2.261, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.7309747170922553, | |
| "grad_norm": 0.3791867792606354, | |
| "learning_rate": 1.8583735168733868e-05, | |
| "loss": 2.2376, | |
| "step": 2172 | |
| }, | |
| { | |
| "epoch": 0.7316478061503513, | |
| "grad_norm": 0.34628093242645264, | |
| "learning_rate": 1.8497238337911238e-05, | |
| "loss": 2.2784, | |
| "step": 2174 | |
| }, | |
| { | |
| "epoch": 0.7323208952084472, | |
| "grad_norm": 0.3489970266819, | |
| "learning_rate": 1.841089756522941e-05, | |
| "loss": 2.2598, | |
| "step": 2176 | |
| }, | |
| { | |
| "epoch": 0.7329939842665433, | |
| "grad_norm": 0.3580794334411621, | |
| "learning_rate": 1.8324713278402687e-05, | |
| "loss": 2.2464, | |
| "step": 2178 | |
| }, | |
| { | |
| "epoch": 0.7336670733246393, | |
| "grad_norm": 0.3512752652168274, | |
| "learning_rate": 1.8238685904370052e-05, | |
| "loss": 2.2622, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.7343401623827353, | |
| "grad_norm": 0.3596512973308563, | |
| "learning_rate": 1.81528158692933e-05, | |
| "loss": 2.2323, | |
| "step": 2182 | |
| }, | |
| { | |
| "epoch": 0.7350132514408313, | |
| "grad_norm": 0.35648801922798157, | |
| "learning_rate": 1.806710359855476e-05, | |
| "loss": 2.2611, | |
| "step": 2184 | |
| }, | |
| { | |
| "epoch": 0.7356863404989272, | |
| "grad_norm": 0.35069939494132996, | |
| "learning_rate": 1.7981549516755226e-05, | |
| "loss": 2.2418, | |
| "step": 2186 | |
| }, | |
| { | |
| "epoch": 0.7363594295570233, | |
| "grad_norm": 0.3489362895488739, | |
| "learning_rate": 1.789615404771182e-05, | |
| "loss": 2.2628, | |
| "step": 2188 | |
| }, | |
| { | |
| "epoch": 0.7370325186151193, | |
| "grad_norm": 0.33355581760406494, | |
| "learning_rate": 1.7810917614456025e-05, | |
| "loss": 2.2441, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.7377056076732152, | |
| "grad_norm": 0.3464316427707672, | |
| "learning_rate": 1.772584063923138e-05, | |
| "loss": 2.2644, | |
| "step": 2192 | |
| }, | |
| { | |
| "epoch": 0.7383786967313113, | |
| "grad_norm": 0.34768715500831604, | |
| "learning_rate": 1.76409235434916e-05, | |
| "loss": 2.2672, | |
| "step": 2194 | |
| }, | |
| { | |
| "epoch": 0.7390517857894072, | |
| "grad_norm": 0.3431531488895416, | |
| "learning_rate": 1.755616674789833e-05, | |
| "loss": 2.2555, | |
| "step": 2196 | |
| }, | |
| { | |
| "epoch": 0.7397248748475033, | |
| "grad_norm": 0.35259053111076355, | |
| "learning_rate": 1.7471570672319115e-05, | |
| "loss": 2.254, | |
| "step": 2198 | |
| }, | |
| { | |
| "epoch": 0.7403979639055993, | |
| "grad_norm": 0.37549468874931335, | |
| "learning_rate": 1.738713573582539e-05, | |
| "loss": 2.2276, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.7410710529636952, | |
| "grad_norm": 0.3733035922050476, | |
| "learning_rate": 1.7302862356690286e-05, | |
| "loss": 2.2591, | |
| "step": 2202 | |
| }, | |
| { | |
| "epoch": 0.7417441420217913, | |
| "grad_norm": 0.36966368556022644, | |
| "learning_rate": 1.7218750952386602e-05, | |
| "loss": 2.2375, | |
| "step": 2204 | |
| }, | |
| { | |
| "epoch": 0.7424172310798872, | |
| "grad_norm": 0.3784911334514618, | |
| "learning_rate": 1.71348019395848e-05, | |
| "loss": 2.2737, | |
| "step": 2206 | |
| }, | |
| { | |
| "epoch": 0.7430903201379833, | |
| "grad_norm": 0.36344560980796814, | |
| "learning_rate": 1.7051015734150883e-05, | |
| "loss": 2.2257, | |
| "step": 2208 | |
| }, | |
| { | |
| "epoch": 0.7437634091960793, | |
| "grad_norm": 0.35233139991760254, | |
| "learning_rate": 1.696739275114425e-05, | |
| "loss": 2.2505, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.7444364982541752, | |
| "grad_norm": 0.34449270367622375, | |
| "learning_rate": 1.6883933404815843e-05, | |
| "loss": 2.2428, | |
| "step": 2212 | |
| }, | |
| { | |
| "epoch": 0.7451095873122713, | |
| "grad_norm": 0.34188973903656006, | |
| "learning_rate": 1.6800638108605938e-05, | |
| "loss": 2.2414, | |
| "step": 2214 | |
| }, | |
| { | |
| "epoch": 0.7457826763703672, | |
| "grad_norm": 0.35242316126823425, | |
| "learning_rate": 1.6717507275142135e-05, | |
| "loss": 2.247, | |
| "step": 2216 | |
| }, | |
| { | |
| "epoch": 0.7464557654284633, | |
| "grad_norm": 0.33550506830215454, | |
| "learning_rate": 1.66345413162373e-05, | |
| "loss": 2.2533, | |
| "step": 2218 | |
| }, | |
| { | |
| "epoch": 0.7471288544865593, | |
| "grad_norm": 0.34991684556007385, | |
| "learning_rate": 1.6551740642887604e-05, | |
| "loss": 2.2671, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.7478019435446552, | |
| "grad_norm": 0.3459914028644562, | |
| "learning_rate": 1.6469105665270368e-05, | |
| "loss": 2.2143, | |
| "step": 2222 | |
| }, | |
| { | |
| "epoch": 0.7484750326027513, | |
| "grad_norm": 0.3622877597808838, | |
| "learning_rate": 1.638663679274211e-05, | |
| "loss": 2.2711, | |
| "step": 2224 | |
| }, | |
| { | |
| "epoch": 0.7491481216608472, | |
| "grad_norm": 0.33951377868652344, | |
| "learning_rate": 1.6304334433836532e-05, | |
| "loss": 2.2364, | |
| "step": 2226 | |
| }, | |
| { | |
| "epoch": 0.7498212107189433, | |
| "grad_norm": 0.36192387342453003, | |
| "learning_rate": 1.6222198996262395e-05, | |
| "loss": 2.2949, | |
| "step": 2228 | |
| }, | |
| { | |
| "epoch": 0.7504942997770393, | |
| "grad_norm": 0.3236417770385742, | |
| "learning_rate": 1.614023088690164e-05, | |
| "loss": 2.2348, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.7511673888351352, | |
| "grad_norm": 0.33134379982948303, | |
| "learning_rate": 1.6058430511807255e-05, | |
| "loss": 2.2529, | |
| "step": 2232 | |
| }, | |
| { | |
| "epoch": 0.7518404778932313, | |
| "grad_norm": 0.35659706592559814, | |
| "learning_rate": 1.5976798276201283e-05, | |
| "loss": 2.2506, | |
| "step": 2234 | |
| }, | |
| { | |
| "epoch": 0.7525135669513272, | |
| "grad_norm": 0.3546935021877289, | |
| "learning_rate": 1.589533458447291e-05, | |
| "loss": 2.2268, | |
| "step": 2236 | |
| }, | |
| { | |
| "epoch": 0.7531866560094232, | |
| "grad_norm": 0.33796676993370056, | |
| "learning_rate": 1.5814039840176336e-05, | |
| "loss": 2.2696, | |
| "step": 2238 | |
| }, | |
| { | |
| "epoch": 0.7538597450675193, | |
| "grad_norm": 0.32641658186912537, | |
| "learning_rate": 1.5732914446028835e-05, | |
| "loss": 2.2491, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.7545328341256152, | |
| "grad_norm": 0.33251863718032837, | |
| "learning_rate": 1.565195880390878e-05, | |
| "loss": 2.246, | |
| "step": 2242 | |
| }, | |
| { | |
| "epoch": 0.7552059231837113, | |
| "grad_norm": 0.32375192642211914, | |
| "learning_rate": 1.5571173314853623e-05, | |
| "loss": 2.2281, | |
| "step": 2244 | |
| }, | |
| { | |
| "epoch": 0.7558790122418072, | |
| "grad_norm": 0.3465546667575836, | |
| "learning_rate": 1.5490558379057895e-05, | |
| "loss": 2.2456, | |
| "step": 2246 | |
| }, | |
| { | |
| "epoch": 0.7565521012999032, | |
| "grad_norm": 0.3475060760974884, | |
| "learning_rate": 1.5410114395871238e-05, | |
| "loss": 2.2547, | |
| "step": 2248 | |
| }, | |
| { | |
| "epoch": 0.7572251903579993, | |
| "grad_norm": 0.33916547894477844, | |
| "learning_rate": 1.5329841763796472e-05, | |
| "loss": 2.2757, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.7578982794160952, | |
| "grad_norm": 0.33837759494781494, | |
| "learning_rate": 1.5249740880487528e-05, | |
| "loss": 2.2266, | |
| "step": 2252 | |
| }, | |
| { | |
| "epoch": 0.7585713684741913, | |
| "grad_norm": 0.34779128432273865, | |
| "learning_rate": 1.5169812142747547e-05, | |
| "loss": 2.2443, | |
| "step": 2254 | |
| }, | |
| { | |
| "epoch": 0.7592444575322872, | |
| "grad_norm": 0.3505866229534149, | |
| "learning_rate": 1.5090055946526926e-05, | |
| "loss": 2.207, | |
| "step": 2256 | |
| }, | |
| { | |
| "epoch": 0.7599175465903832, | |
| "grad_norm": 0.3309464454650879, | |
| "learning_rate": 1.5010472686921274e-05, | |
| "loss": 2.2534, | |
| "step": 2258 | |
| }, | |
| { | |
| "epoch": 0.7605906356484793, | |
| "grad_norm": 0.3288039267063141, | |
| "learning_rate": 1.4931062758169583e-05, | |
| "loss": 2.2751, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.7612637247065752, | |
| "grad_norm": 0.3784511089324951, | |
| "learning_rate": 1.4851826553652087e-05, | |
| "loss": 2.2731, | |
| "step": 2262 | |
| }, | |
| { | |
| "epoch": 0.7619368137646713, | |
| "grad_norm": 0.3320266008377075, | |
| "learning_rate": 1.4772764465888528e-05, | |
| "loss": 2.263, | |
| "step": 2264 | |
| }, | |
| { | |
| "epoch": 0.7626099028227672, | |
| "grad_norm": 0.3421779274940491, | |
| "learning_rate": 1.469387688653609e-05, | |
| "loss": 2.2833, | |
| "step": 2266 | |
| }, | |
| { | |
| "epoch": 0.7632829918808632, | |
| "grad_norm": 0.33158448338508606, | |
| "learning_rate": 1.4615164206387438e-05, | |
| "loss": 2.2546, | |
| "step": 2268 | |
| }, | |
| { | |
| "epoch": 0.7639560809389593, | |
| "grad_norm": 0.34830743074417114, | |
| "learning_rate": 1.4536626815368848e-05, | |
| "loss": 2.2465, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.7646291699970552, | |
| "grad_norm": 0.32620352506637573, | |
| "learning_rate": 1.4458265102538266e-05, | |
| "loss": 2.2211, | |
| "step": 2272 | |
| }, | |
| { | |
| "epoch": 0.7653022590551513, | |
| "grad_norm": 0.35142725706100464, | |
| "learning_rate": 1.4380079456083344e-05, | |
| "loss": 2.2335, | |
| "step": 2274 | |
| }, | |
| { | |
| "epoch": 0.7659753481132472, | |
| "grad_norm": 0.3353235721588135, | |
| "learning_rate": 1.4302070263319517e-05, | |
| "loss": 2.2205, | |
| "step": 2276 | |
| }, | |
| { | |
| "epoch": 0.7666484371713432, | |
| "grad_norm": 0.34044894576072693, | |
| "learning_rate": 1.4224237910688166e-05, | |
| "loss": 2.2499, | |
| "step": 2278 | |
| }, | |
| { | |
| "epoch": 0.7673215262294393, | |
| "grad_norm": 0.3547942638397217, | |
| "learning_rate": 1.4146582783754575e-05, | |
| "loss": 2.2422, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.7679946152875352, | |
| "grad_norm": 0.336256742477417, | |
| "learning_rate": 1.4069105267206156e-05, | |
| "loss": 2.2378, | |
| "step": 2282 | |
| }, | |
| { | |
| "epoch": 0.7686677043456313, | |
| "grad_norm": 0.3607028126716614, | |
| "learning_rate": 1.3991805744850417e-05, | |
| "loss": 2.2332, | |
| "step": 2284 | |
| }, | |
| { | |
| "epoch": 0.7693407934037272, | |
| "grad_norm": 0.34364938735961914, | |
| "learning_rate": 1.3914684599613126e-05, | |
| "loss": 2.2148, | |
| "step": 2286 | |
| }, | |
| { | |
| "epoch": 0.7700138824618232, | |
| "grad_norm": 0.35112154483795166, | |
| "learning_rate": 1.3837742213536453e-05, | |
| "loss": 2.29, | |
| "step": 2288 | |
| }, | |
| { | |
| "epoch": 0.7706869715199193, | |
| "grad_norm": 0.3588349521160126, | |
| "learning_rate": 1.3760978967777022e-05, | |
| "loss": 2.263, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.7713600605780152, | |
| "grad_norm": 0.3553431034088135, | |
| "learning_rate": 1.3684395242603958e-05, | |
| "loss": 2.2772, | |
| "step": 2292 | |
| }, | |
| { | |
| "epoch": 0.7720331496361112, | |
| "grad_norm": 0.3674144148826599, | |
| "learning_rate": 1.360799141739717e-05, | |
| "loss": 2.2812, | |
| "step": 2294 | |
| }, | |
| { | |
| "epoch": 0.7727062386942072, | |
| "grad_norm": 0.34299132227897644, | |
| "learning_rate": 1.3531767870645362e-05, | |
| "loss": 2.222, | |
| "step": 2296 | |
| }, | |
| { | |
| "epoch": 0.7733793277523032, | |
| "grad_norm": 0.3339577615261078, | |
| "learning_rate": 1.3455724979944134e-05, | |
| "loss": 2.2301, | |
| "step": 2298 | |
| }, | |
| { | |
| "epoch": 0.7740524168103993, | |
| "grad_norm": 0.3338162899017334, | |
| "learning_rate": 1.3379863121994168e-05, | |
| "loss": 2.2704, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.7747255058684952, | |
| "grad_norm": 0.3388438820838928, | |
| "learning_rate": 1.3304182672599396e-05, | |
| "loss": 2.2879, | |
| "step": 2302 | |
| }, | |
| { | |
| "epoch": 0.7753985949265912, | |
| "grad_norm": 0.3300301134586334, | |
| "learning_rate": 1.3228684006665027e-05, | |
| "loss": 2.232, | |
| "step": 2304 | |
| }, | |
| { | |
| "epoch": 0.7760716839846872, | |
| "grad_norm": 0.3355436325073242, | |
| "learning_rate": 1.3153367498195769e-05, | |
| "loss": 2.2572, | |
| "step": 2306 | |
| }, | |
| { | |
| "epoch": 0.7767447730427832, | |
| "grad_norm": 0.35798126459121704, | |
| "learning_rate": 1.3078233520293992e-05, | |
| "loss": 2.266, | |
| "step": 2308 | |
| }, | |
| { | |
| "epoch": 0.7774178621008793, | |
| "grad_norm": 0.3336060345172882, | |
| "learning_rate": 1.30032824451578e-05, | |
| "loss": 2.2375, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.7780909511589752, | |
| "grad_norm": 0.3310167193412781, | |
| "learning_rate": 1.292851464407928e-05, | |
| "loss": 2.2435, | |
| "step": 2312 | |
| }, | |
| { | |
| "epoch": 0.7787640402170712, | |
| "grad_norm": 0.349125474691391, | |
| "learning_rate": 1.285393048744259e-05, | |
| "loss": 2.2972, | |
| "step": 2314 | |
| }, | |
| { | |
| "epoch": 0.7794371292751672, | |
| "grad_norm": 0.34504398703575134, | |
| "learning_rate": 1.277953034472214e-05, | |
| "loss": 2.2502, | |
| "step": 2316 | |
| }, | |
| { | |
| "epoch": 0.7801102183332632, | |
| "grad_norm": 0.34388425946235657, | |
| "learning_rate": 1.2705314584480815e-05, | |
| "loss": 2.2226, | |
| "step": 2318 | |
| }, | |
| { | |
| "epoch": 0.7807833073913593, | |
| "grad_norm": 0.3372696340084076, | |
| "learning_rate": 1.2631283574368068e-05, | |
| "loss": 2.2817, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.7814563964494552, | |
| "grad_norm": 0.35478445887565613, | |
| "learning_rate": 1.2557437681118145e-05, | |
| "loss": 2.2547, | |
| "step": 2322 | |
| }, | |
| { | |
| "epoch": 0.7821294855075512, | |
| "grad_norm": 0.3373311460018158, | |
| "learning_rate": 1.2483777270548274e-05, | |
| "loss": 2.2589, | |
| "step": 2324 | |
| }, | |
| { | |
| "epoch": 0.7828025745656472, | |
| "grad_norm": 0.3496008515357971, | |
| "learning_rate": 1.2410302707556864e-05, | |
| "loss": 2.2352, | |
| "step": 2326 | |
| }, | |
| { | |
| "epoch": 0.7834756636237432, | |
| "grad_norm": 0.3309776186943054, | |
| "learning_rate": 1.2337014356121624e-05, | |
| "loss": 2.2912, | |
| "step": 2328 | |
| }, | |
| { | |
| "epoch": 0.7841487526818393, | |
| "grad_norm": 0.33283454179763794, | |
| "learning_rate": 1.2263912579297816e-05, | |
| "loss": 2.2531, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.7848218417399352, | |
| "grad_norm": 0.3534455895423889, | |
| "learning_rate": 1.2190997739216493e-05, | |
| "loss": 2.2375, | |
| "step": 2332 | |
| }, | |
| { | |
| "epoch": 0.7854949307980312, | |
| "grad_norm": 0.3482494056224823, | |
| "learning_rate": 1.211827019708262e-05, | |
| "loss": 2.2516, | |
| "step": 2334 | |
| }, | |
| { | |
| "epoch": 0.7861680198561272, | |
| "grad_norm": 0.33788228034973145, | |
| "learning_rate": 1.2045730313173332e-05, | |
| "loss": 2.2135, | |
| "step": 2336 | |
| }, | |
| { | |
| "epoch": 0.7868411089142232, | |
| "grad_norm": 0.3371339440345764, | |
| "learning_rate": 1.1973378446836181e-05, | |
| "loss": 2.2686, | |
| "step": 2338 | |
| }, | |
| { | |
| "epoch": 0.7875141979723193, | |
| "grad_norm": 0.3414793908596039, | |
| "learning_rate": 1.1901214956487255e-05, | |
| "loss": 2.251, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.7881872870304152, | |
| "grad_norm": 0.35052791237831116, | |
| "learning_rate": 1.1829240199609543e-05, | |
| "loss": 2.239, | |
| "step": 2342 | |
| }, | |
| { | |
| "epoch": 0.7888603760885112, | |
| "grad_norm": 0.32874590158462524, | |
| "learning_rate": 1.1757454532751028e-05, | |
| "loss": 2.2594, | |
| "step": 2344 | |
| }, | |
| { | |
| "epoch": 0.7895334651466072, | |
| "grad_norm": 0.33765682578086853, | |
| "learning_rate": 1.168585831152298e-05, | |
| "loss": 2.281, | |
| "step": 2346 | |
| }, | |
| { | |
| "epoch": 0.7902065542047032, | |
| "grad_norm": 0.34867605566978455, | |
| "learning_rate": 1.1614451890598255e-05, | |
| "loss": 2.2004, | |
| "step": 2348 | |
| }, | |
| { | |
| "epoch": 0.7908796432627992, | |
| "grad_norm": 0.33778446912765503, | |
| "learning_rate": 1.1543235623709408e-05, | |
| "loss": 2.2259, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.7915527323208952, | |
| "grad_norm": 0.3365173637866974, | |
| "learning_rate": 1.1472209863647037e-05, | |
| "loss": 2.2524, | |
| "step": 2352 | |
| }, | |
| { | |
| "epoch": 0.7922258213789912, | |
| "grad_norm": 0.3220577836036682, | |
| "learning_rate": 1.1401374962258039e-05, | |
| "loss": 2.2449, | |
| "step": 2354 | |
| }, | |
| { | |
| "epoch": 0.7928989104370872, | |
| "grad_norm": 0.3405933976173401, | |
| "learning_rate": 1.133073127044379e-05, | |
| "loss": 2.2398, | |
| "step": 2356 | |
| }, | |
| { | |
| "epoch": 0.7935719994951832, | |
| "grad_norm": 0.32475680112838745, | |
| "learning_rate": 1.1260279138158459e-05, | |
| "loss": 2.2633, | |
| "step": 2358 | |
| }, | |
| { | |
| "epoch": 0.7942450885532792, | |
| "grad_norm": 0.32878637313842773, | |
| "learning_rate": 1.1190018914407314e-05, | |
| "loss": 2.2229, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.7949181776113752, | |
| "grad_norm": 0.35024937987327576, | |
| "learning_rate": 1.1119950947244878e-05, | |
| "loss": 2.2647, | |
| "step": 2362 | |
| }, | |
| { | |
| "epoch": 0.7955912666694712, | |
| "grad_norm": 0.333114892244339, | |
| "learning_rate": 1.1050075583773357e-05, | |
| "loss": 2.2656, | |
| "step": 2364 | |
| }, | |
| { | |
| "epoch": 0.7962643557275672, | |
| "grad_norm": 0.32833924889564514, | |
| "learning_rate": 1.0980393170140774e-05, | |
| "loss": 2.2241, | |
| "step": 2366 | |
| }, | |
| { | |
| "epoch": 0.7969374447856632, | |
| "grad_norm": 0.328016996383667, | |
| "learning_rate": 1.0910904051539317e-05, | |
| "loss": 2.2428, | |
| "step": 2368 | |
| }, | |
| { | |
| "epoch": 0.7976105338437592, | |
| "grad_norm": 0.342161625623703, | |
| "learning_rate": 1.0841608572203677e-05, | |
| "loss": 2.2641, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.7982836229018552, | |
| "grad_norm": 0.3461630940437317, | |
| "learning_rate": 1.0772507075409293e-05, | |
| "loss": 2.2051, | |
| "step": 2372 | |
| }, | |
| { | |
| "epoch": 0.7989567119599512, | |
| "grad_norm": 0.3533310294151306, | |
| "learning_rate": 1.0703599903470584e-05, | |
| "loss": 2.2123, | |
| "step": 2374 | |
| }, | |
| { | |
| "epoch": 0.7996298010180471, | |
| "grad_norm": 0.3372839689254761, | |
| "learning_rate": 1.0634887397739396e-05, | |
| "loss": 2.2846, | |
| "step": 2376 | |
| }, | |
| { | |
| "epoch": 0.8003028900761432, | |
| "grad_norm": 0.3277933597564697, | |
| "learning_rate": 1.0566369898603235e-05, | |
| "loss": 2.218, | |
| "step": 2378 | |
| }, | |
| { | |
| "epoch": 0.8009759791342392, | |
| "grad_norm": 0.3312545716762543, | |
| "learning_rate": 1.0498047745483547e-05, | |
| "loss": 2.2481, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.8016490681923352, | |
| "grad_norm": 0.32342013716697693, | |
| "learning_rate": 1.0429921276834081e-05, | |
| "loss": 2.2122, | |
| "step": 2382 | |
| }, | |
| { | |
| "epoch": 0.8023221572504312, | |
| "grad_norm": 0.3250105381011963, | |
| "learning_rate": 1.036199083013925e-05, | |
| "loss": 2.2396, | |
| "step": 2384 | |
| }, | |
| { | |
| "epoch": 0.8029952463085271, | |
| "grad_norm": 0.32683631777763367, | |
| "learning_rate": 1.0294256741912373e-05, | |
| "loss": 2.2575, | |
| "step": 2386 | |
| }, | |
| { | |
| "epoch": 0.8036683353666232, | |
| "grad_norm": 0.3309519588947296, | |
| "learning_rate": 1.0226719347694037e-05, | |
| "loss": 2.2532, | |
| "step": 2388 | |
| }, | |
| { | |
| "epoch": 0.8043414244247192, | |
| "grad_norm": 0.33092552423477173, | |
| "learning_rate": 1.01593789820505e-05, | |
| "loss": 2.2294, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.8050145134828152, | |
| "grad_norm": 0.33747655153274536, | |
| "learning_rate": 1.009223597857193e-05, | |
| "loss": 2.2489, | |
| "step": 2392 | |
| }, | |
| { | |
| "epoch": 0.8056876025409112, | |
| "grad_norm": 0.3354330360889435, | |
| "learning_rate": 1.0025290669870851e-05, | |
| "loss": 2.2171, | |
| "step": 2394 | |
| }, | |
| { | |
| "epoch": 0.8063606915990071, | |
| "grad_norm": 0.3374640941619873, | |
| "learning_rate": 9.958543387580415e-06, | |
| "loss": 2.2491, | |
| "step": 2396 | |
| }, | |
| { | |
| "epoch": 0.8070337806571032, | |
| "grad_norm": 0.3381037414073944, | |
| "learning_rate": 9.891994462352776e-06, | |
| "loss": 2.1906, | |
| "step": 2398 | |
| }, | |
| { | |
| "epoch": 0.8077068697151992, | |
| "grad_norm": 0.32526618242263794, | |
| "learning_rate": 9.825644223857533e-06, | |
| "loss": 2.2378, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.8083799587732952, | |
| "grad_norm": 0.3527051508426666, | |
| "learning_rate": 9.759493000779968e-06, | |
| "loss": 2.2474, | |
| "step": 2402 | |
| }, | |
| { | |
| "epoch": 0.8090530478313912, | |
| "grad_norm": 0.322172075510025, | |
| "learning_rate": 9.693541120819499e-06, | |
| "loss": 2.2806, | |
| "step": 2404 | |
| }, | |
| { | |
| "epoch": 0.8097261368894871, | |
| "grad_norm": 0.33850088715553284, | |
| "learning_rate": 9.627788910688058e-06, | |
| "loss": 2.2205, | |
| "step": 2406 | |
| }, | |
| { | |
| "epoch": 0.8103992259475832, | |
| "grad_norm": 0.32618337869644165, | |
| "learning_rate": 9.562236696108456e-06, | |
| "loss": 2.2138, | |
| "step": 2408 | |
| }, | |
| { | |
| "epoch": 0.8110723150056792, | |
| "grad_norm": 0.3170488774776459, | |
| "learning_rate": 9.496884801812738e-06, | |
| "loss": 2.2171, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.8117454040637752, | |
| "grad_norm": 0.3533998131752014, | |
| "learning_rate": 9.431733551540611e-06, | |
| "loss": 2.2691, | |
| "step": 2412 | |
| }, | |
| { | |
| "epoch": 0.8124184931218712, | |
| "grad_norm": 0.3621939420700073, | |
| "learning_rate": 9.36678326803787e-06, | |
| "loss": 2.2157, | |
| "step": 2414 | |
| }, | |
| { | |
| "epoch": 0.8130915821799671, | |
| "grad_norm": 0.3280872106552124, | |
| "learning_rate": 9.302034273054721e-06, | |
| "loss": 2.2696, | |
| "step": 2416 | |
| }, | |
| { | |
| "epoch": 0.8137646712380632, | |
| "grad_norm": 0.35104599595069885, | |
| "learning_rate": 9.237486887344227e-06, | |
| "loss": 2.2475, | |
| "step": 2418 | |
| }, | |
| { | |
| "epoch": 0.8144377602961592, | |
| "grad_norm": 0.329738974571228, | |
| "learning_rate": 9.173141430660759e-06, | |
| "loss": 2.2811, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.8151108493542552, | |
| "grad_norm": 0.3505691885948181, | |
| "learning_rate": 9.10899822175833e-06, | |
| "loss": 2.2466, | |
| "step": 2422 | |
| }, | |
| { | |
| "epoch": 0.8157839384123512, | |
| "grad_norm": 0.3490196764469147, | |
| "learning_rate": 9.045057578389104e-06, | |
| "loss": 2.2562, | |
| "step": 2424 | |
| }, | |
| { | |
| "epoch": 0.8164570274704472, | |
| "grad_norm": 0.332364946603775, | |
| "learning_rate": 8.981319817301737e-06, | |
| "loss": 2.2096, | |
| "step": 2426 | |
| }, | |
| { | |
| "epoch": 0.8171301165285432, | |
| "grad_norm": 0.35335782170295715, | |
| "learning_rate": 8.917785254239858e-06, | |
| "loss": 2.2168, | |
| "step": 2428 | |
| }, | |
| { | |
| "epoch": 0.8178032055866392, | |
| "grad_norm": 0.3289039433002472, | |
| "learning_rate": 8.854454203940521e-06, | |
| "loss": 2.2601, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.8184762946447351, | |
| "grad_norm": 0.3358098268508911, | |
| "learning_rate": 8.791326980132592e-06, | |
| "loss": 2.2859, | |
| "step": 2432 | |
| }, | |
| { | |
| "epoch": 0.8191493837028312, | |
| "grad_norm": 0.34611964225769043, | |
| "learning_rate": 8.728403895535215e-06, | |
| "loss": 2.2581, | |
| "step": 2434 | |
| }, | |
| { | |
| "epoch": 0.8198224727609272, | |
| "grad_norm": 0.3253307342529297, | |
| "learning_rate": 8.665685261856315e-06, | |
| "loss": 2.2434, | |
| "step": 2436 | |
| }, | |
| { | |
| "epoch": 0.8204955618190232, | |
| "grad_norm": 0.3149079978466034, | |
| "learning_rate": 8.603171389790976e-06, | |
| "loss": 2.222, | |
| "step": 2438 | |
| }, | |
| { | |
| "epoch": 0.8211686508771192, | |
| "grad_norm": 0.33320352435112, | |
| "learning_rate": 8.540862589019916e-06, | |
| "loss": 2.2344, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.8218417399352151, | |
| "grad_norm": 0.3100329637527466, | |
| "learning_rate": 8.478759168208028e-06, | |
| "loss": 2.2471, | |
| "step": 2442 | |
| }, | |
| { | |
| "epoch": 0.8225148289933112, | |
| "grad_norm": 0.32634490728378296, | |
| "learning_rate": 8.416861435002742e-06, | |
| "loss": 2.2528, | |
| "step": 2444 | |
| }, | |
| { | |
| "epoch": 0.8231879180514072, | |
| "grad_norm": 0.3271341323852539, | |
| "learning_rate": 8.3551696960326e-06, | |
| "loss": 2.2795, | |
| "step": 2446 | |
| }, | |
| { | |
| "epoch": 0.8238610071095032, | |
| "grad_norm": 0.3098745048046112, | |
| "learning_rate": 8.29368425690566e-06, | |
| "loss": 2.2643, | |
| "step": 2448 | |
| }, | |
| { | |
| "epoch": 0.8245340961675992, | |
| "grad_norm": 0.32310840487480164, | |
| "learning_rate": 8.232405422208006e-06, | |
| "loss": 2.2447, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.8252071852256951, | |
| "grad_norm": 0.32423245906829834, | |
| "learning_rate": 8.171333495502286e-06, | |
| "loss": 2.2321, | |
| "step": 2452 | |
| }, | |
| { | |
| "epoch": 0.8258802742837912, | |
| "grad_norm": 0.31200647354125977, | |
| "learning_rate": 8.110468779326137e-06, | |
| "loss": 2.244, | |
| "step": 2454 | |
| }, | |
| { | |
| "epoch": 0.8265533633418872, | |
| "grad_norm": 0.3206006586551666, | |
| "learning_rate": 8.049811575190713e-06, | |
| "loss": 2.2656, | |
| "step": 2456 | |
| }, | |
| { | |
| "epoch": 0.8272264523999832, | |
| "grad_norm": 0.3135566711425781, | |
| "learning_rate": 7.989362183579219e-06, | |
| "loss": 2.2556, | |
| "step": 2458 | |
| }, | |
| { | |
| "epoch": 0.8278995414580792, | |
| "grad_norm": 0.3741950988769531, | |
| "learning_rate": 7.92912090394541e-06, | |
| "loss": 2.2159, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.8285726305161751, | |
| "grad_norm": 0.3329770565032959, | |
| "learning_rate": 7.869088034712058e-06, | |
| "loss": 2.2522, | |
| "step": 2462 | |
| }, | |
| { | |
| "epoch": 0.8292457195742712, | |
| "grad_norm": 0.34624531865119934, | |
| "learning_rate": 7.809263873269529e-06, | |
| "loss": 2.2416, | |
| "step": 2464 | |
| }, | |
| { | |
| "epoch": 0.8299188086323672, | |
| "grad_norm": 0.32020536065101624, | |
| "learning_rate": 7.749648715974312e-06, | |
| "loss": 2.2282, | |
| "step": 2466 | |
| }, | |
| { | |
| "epoch": 0.8305918976904632, | |
| "grad_norm": 0.33093929290771484, | |
| "learning_rate": 7.690242858147517e-06, | |
| "loss": 2.2272, | |
| "step": 2468 | |
| }, | |
| { | |
| "epoch": 0.8312649867485592, | |
| "grad_norm": 0.32398727536201477, | |
| "learning_rate": 7.631046594073421e-06, | |
| "loss": 2.2332, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.8319380758066551, | |
| "grad_norm": 0.32068392634391785, | |
| "learning_rate": 7.572060216998034e-06, | |
| "loss": 2.1858, | |
| "step": 2472 | |
| }, | |
| { | |
| "epoch": 0.8326111648647512, | |
| "grad_norm": 0.3306787610054016, | |
| "learning_rate": 7.513284019127614e-06, | |
| "loss": 2.2492, | |
| "step": 2474 | |
| }, | |
| { | |
| "epoch": 0.8332842539228472, | |
| "grad_norm": 0.32998260855674744, | |
| "learning_rate": 7.454718291627255e-06, | |
| "loss": 2.2442, | |
| "step": 2476 | |
| }, | |
| { | |
| "epoch": 0.8339573429809432, | |
| "grad_norm": 0.3302423357963562, | |
| "learning_rate": 7.396363324619409e-06, | |
| "loss": 2.221, | |
| "step": 2478 | |
| }, | |
| { | |
| "epoch": 0.8346304320390392, | |
| "grad_norm": 0.3253351151943207, | |
| "learning_rate": 7.338219407182451e-06, | |
| "loss": 2.2184, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.8353035210971351, | |
| "grad_norm": 0.31889471411705017, | |
| "learning_rate": 7.280286827349297e-06, | |
| "loss": 2.284, | |
| "step": 2482 | |
| }, | |
| { | |
| "epoch": 0.8359766101552312, | |
| "grad_norm": 0.31694450974464417, | |
| "learning_rate": 7.222565872105913e-06, | |
| "loss": 2.2276, | |
| "step": 2484 | |
| }, | |
| { | |
| "epoch": 0.8366496992133272, | |
| "grad_norm": 0.32638663053512573, | |
| "learning_rate": 7.165056827389915e-06, | |
| "loss": 2.2378, | |
| "step": 2486 | |
| }, | |
| { | |
| "epoch": 0.8373227882714231, | |
| "grad_norm": 0.31119346618652344, | |
| "learning_rate": 7.107759978089179e-06, | |
| "loss": 2.2545, | |
| "step": 2488 | |
| }, | |
| { | |
| "epoch": 0.8379958773295192, | |
| "grad_norm": 0.3240341544151306, | |
| "learning_rate": 7.050675608040424e-06, | |
| "loss": 2.2581, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.8386689663876151, | |
| "grad_norm": 0.33010658621788025, | |
| "learning_rate": 6.993804000027726e-06, | |
| "loss": 2.2487, | |
| "step": 2492 | |
| }, | |
| { | |
| "epoch": 0.8393420554457112, | |
| "grad_norm": 0.3287750482559204, | |
| "learning_rate": 6.9371454357812445e-06, | |
| "loss": 2.2651, | |
| "step": 2494 | |
| }, | |
| { | |
| "epoch": 0.8400151445038072, | |
| "grad_norm": 0.33341431617736816, | |
| "learning_rate": 6.880700195975753e-06, | |
| "loss": 2.2358, | |
| "step": 2496 | |
| }, | |
| { | |
| "epoch": 0.8406882335619031, | |
| "grad_norm": 0.3379230201244354, | |
| "learning_rate": 6.824468560229247e-06, | |
| "loss": 2.2363, | |
| "step": 2498 | |
| }, | |
| { | |
| "epoch": 0.8413613226199992, | |
| "grad_norm": 0.366382360458374, | |
| "learning_rate": 6.76845080710155e-06, | |
| "loss": 2.2304, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.8420344116780951, | |
| "grad_norm": 0.32602646946907043, | |
| "learning_rate": 6.7126472140930155e-06, | |
| "loss": 2.2457, | |
| "step": 2502 | |
| }, | |
| { | |
| "epoch": 0.8427075007361912, | |
| "grad_norm": 0.3166140019893646, | |
| "learning_rate": 6.657058057643028e-06, | |
| "loss": 2.215, | |
| "step": 2504 | |
| }, | |
| { | |
| "epoch": 0.8433805897942872, | |
| "grad_norm": 0.3184511065483093, | |
| "learning_rate": 6.601683613128756e-06, | |
| "loss": 2.2653, | |
| "step": 2506 | |
| }, | |
| { | |
| "epoch": 0.8440536788523831, | |
| "grad_norm": 0.3298148512840271, | |
| "learning_rate": 6.5465241548636865e-06, | |
| "loss": 2.1946, | |
| "step": 2508 | |
| }, | |
| { | |
| "epoch": 0.8447267679104792, | |
| "grad_norm": 0.320137619972229, | |
| "learning_rate": 6.4915799560963106e-06, | |
| "loss": 2.2354, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.8453998569685751, | |
| "grad_norm": 0.3132067918777466, | |
| "learning_rate": 6.436851289008805e-06, | |
| "loss": 2.2291, | |
| "step": 2512 | |
| }, | |
| { | |
| "epoch": 0.8460729460266712, | |
| "grad_norm": 0.3094591498374939, | |
| "learning_rate": 6.382338424715612e-06, | |
| "loss": 2.2712, | |
| "step": 2514 | |
| }, | |
| { | |
| "epoch": 0.8467460350847672, | |
| "grad_norm": 0.3072458505630493, | |
| "learning_rate": 6.328041633262138e-06, | |
| "loss": 2.2521, | |
| "step": 2516 | |
| }, | |
| { | |
| "epoch": 0.8474191241428631, | |
| "grad_norm": 0.3033926486968994, | |
| "learning_rate": 6.273961183623422e-06, | |
| "loss": 2.2445, | |
| "step": 2518 | |
| }, | |
| { | |
| "epoch": 0.8480922132009592, | |
| "grad_norm": 0.32361242175102234, | |
| "learning_rate": 6.220097343702808e-06, | |
| "loss": 2.2359, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.8487653022590551, | |
| "grad_norm": 0.3142426311969757, | |
| "learning_rate": 6.166450380330524e-06, | |
| "loss": 2.2657, | |
| "step": 2522 | |
| }, | |
| { | |
| "epoch": 0.8494383913171512, | |
| "grad_norm": 0.3298629820346832, | |
| "learning_rate": 6.113020559262522e-06, | |
| "loss": 2.2335, | |
| "step": 2524 | |
| }, | |
| { | |
| "epoch": 0.8501114803752472, | |
| "grad_norm": 0.3207451105117798, | |
| "learning_rate": 6.059808145179047e-06, | |
| "loss": 2.2406, | |
| "step": 2526 | |
| }, | |
| { | |
| "epoch": 0.8507845694333431, | |
| "grad_norm": 0.31240135431289673, | |
| "learning_rate": 6.006813401683337e-06, | |
| "loss": 2.2312, | |
| "step": 2528 | |
| }, | |
| { | |
| "epoch": 0.8514576584914392, | |
| "grad_norm": 0.3212936222553253, | |
| "learning_rate": 5.954036591300355e-06, | |
| "loss": 2.2459, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.8521307475495351, | |
| "grad_norm": 0.3074382543563843, | |
| "learning_rate": 5.90147797547545e-06, | |
| "loss": 2.2376, | |
| "step": 2532 | |
| }, | |
| { | |
| "epoch": 0.8528038366076311, | |
| "grad_norm": 0.3181383013725281, | |
| "learning_rate": 5.849137814573113e-06, | |
| "loss": 2.2709, | |
| "step": 2534 | |
| }, | |
| { | |
| "epoch": 0.8534769256657272, | |
| "grad_norm": 0.30766862630844116, | |
| "learning_rate": 5.797016367875618e-06, | |
| "loss": 2.2719, | |
| "step": 2536 | |
| }, | |
| { | |
| "epoch": 0.8541500147238231, | |
| "grad_norm": 0.31964921951293945, | |
| "learning_rate": 5.745113893581794e-06, | |
| "loss": 2.2327, | |
| "step": 2538 | |
| }, | |
| { | |
| "epoch": 0.8548231037819192, | |
| "grad_norm": 0.3160279095172882, | |
| "learning_rate": 5.6934306488057255e-06, | |
| "loss": 2.2326, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.8554961928400151, | |
| "grad_norm": 0.32105132937431335, | |
| "learning_rate": 5.641966889575484e-06, | |
| "loss": 2.2572, | |
| "step": 2542 | |
| }, | |
| { | |
| "epoch": 0.8561692818981111, | |
| "grad_norm": 0.30872154235839844, | |
| "learning_rate": 5.590722870831844e-06, | |
| "loss": 2.2385, | |
| "step": 2544 | |
| }, | |
| { | |
| "epoch": 0.8568423709562072, | |
| "grad_norm": 0.3115732669830322, | |
| "learning_rate": 5.539698846427016e-06, | |
| "loss": 2.2456, | |
| "step": 2546 | |
| }, | |
| { | |
| "epoch": 0.8575154600143031, | |
| "grad_norm": 0.303165078163147, | |
| "learning_rate": 5.488895069123457e-06, | |
| "loss": 2.2708, | |
| "step": 2548 | |
| }, | |
| { | |
| "epoch": 0.8581885490723992, | |
| "grad_norm": 0.3096655011177063, | |
| "learning_rate": 5.43831179059251e-06, | |
| "loss": 2.2359, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.8588616381304951, | |
| "grad_norm": 0.3103810250759125, | |
| "learning_rate": 5.387949261413233e-06, | |
| "loss": 2.2021, | |
| "step": 2552 | |
| }, | |
| { | |
| "epoch": 0.8595347271885911, | |
| "grad_norm": 0.31433239579200745, | |
| "learning_rate": 5.3378077310711535e-06, | |
| "loss": 2.2285, | |
| "step": 2554 | |
| }, | |
| { | |
| "epoch": 0.8602078162466872, | |
| "grad_norm": 0.31318360567092896, | |
| "learning_rate": 5.287887447956991e-06, | |
| "loss": 2.2288, | |
| "step": 2556 | |
| }, | |
| { | |
| "epoch": 0.8608809053047831, | |
| "grad_norm": 0.3114445209503174, | |
| "learning_rate": 5.238188659365472e-06, | |
| "loss": 2.2684, | |
| "step": 2558 | |
| }, | |
| { | |
| "epoch": 0.8615539943628792, | |
| "grad_norm": 0.31647783517837524, | |
| "learning_rate": 5.188711611494079e-06, | |
| "loss": 2.2554, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.8622270834209751, | |
| "grad_norm": 0.32441526651382446, | |
| "learning_rate": 5.139456549441813e-06, | |
| "loss": 2.2185, | |
| "step": 2562 | |
| }, | |
| { | |
| "epoch": 0.8629001724790711, | |
| "grad_norm": 0.3170822560787201, | |
| "learning_rate": 5.090423717208054e-06, | |
| "loss": 2.2305, | |
| "step": 2564 | |
| }, | |
| { | |
| "epoch": 0.8635732615371672, | |
| "grad_norm": 0.3120192885398865, | |
| "learning_rate": 5.04161335769126e-06, | |
| "loss": 2.2524, | |
| "step": 2566 | |
| }, | |
| { | |
| "epoch": 0.8642463505952631, | |
| "grad_norm": 0.3131890296936035, | |
| "learning_rate": 4.99302571268781e-06, | |
| "loss": 2.2579, | |
| "step": 2568 | |
| }, | |
| { | |
| "epoch": 0.8649194396533592, | |
| "grad_norm": 0.3128916323184967, | |
| "learning_rate": 4.9446610228908175e-06, | |
| "loss": 2.2561, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.8655925287114551, | |
| "grad_norm": 0.3291536271572113, | |
| "learning_rate": 4.896519527888937e-06, | |
| "loss": 2.2246, | |
| "step": 2572 | |
| }, | |
| { | |
| "epoch": 0.8662656177695511, | |
| "grad_norm": 0.3148040175437927, | |
| "learning_rate": 4.848601466165098e-06, | |
| "loss": 2.2804, | |
| "step": 2574 | |
| }, | |
| { | |
| "epoch": 0.8669387068276472, | |
| "grad_norm": 0.320907860994339, | |
| "learning_rate": 4.800907075095446e-06, | |
| "loss": 2.2305, | |
| "step": 2576 | |
| }, | |
| { | |
| "epoch": 0.8676117958857431, | |
| "grad_norm": 0.3029519021511078, | |
| "learning_rate": 4.753436590948113e-06, | |
| "loss": 2.2462, | |
| "step": 2578 | |
| }, | |
| { | |
| "epoch": 0.8682848849438392, | |
| "grad_norm": 0.3168580234050751, | |
| "learning_rate": 4.706190248881998e-06, | |
| "loss": 2.2336, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.8689579740019351, | |
| "grad_norm": 0.3087359368801117, | |
| "learning_rate": 4.659168282945669e-06, | |
| "loss": 2.242, | |
| "step": 2582 | |
| }, | |
| { | |
| "epoch": 0.8696310630600311, | |
| "grad_norm": 0.3013114035129547, | |
| "learning_rate": 4.612370926076204e-06, | |
| "loss": 2.2436, | |
| "step": 2584 | |
| }, | |
| { | |
| "epoch": 0.8703041521181272, | |
| "grad_norm": 0.3221624195575714, | |
| "learning_rate": 4.5657984100979716e-06, | |
| "loss": 2.2483, | |
| "step": 2586 | |
| }, | |
| { | |
| "epoch": 0.8709772411762231, | |
| "grad_norm": 0.30573809146881104, | |
| "learning_rate": 4.5194509657215615e-06, | |
| "loss": 2.2518, | |
| "step": 2588 | |
| }, | |
| { | |
| "epoch": 0.8716503302343191, | |
| "grad_norm": 0.3036879897117615, | |
| "learning_rate": 4.473328822542583e-06, | |
| "loss": 2.2708, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.8723234192924151, | |
| "grad_norm": 0.3120042383670807, | |
| "learning_rate": 4.4274322090405555e-06, | |
| "loss": 2.2511, | |
| "step": 2592 | |
| }, | |
| { | |
| "epoch": 0.8729965083505111, | |
| "grad_norm": 0.31255391240119934, | |
| "learning_rate": 4.381761352577779e-06, | |
| "loss": 2.1972, | |
| "step": 2594 | |
| }, | |
| { | |
| "epoch": 0.8736695974086072, | |
| "grad_norm": 0.31823691725730896, | |
| "learning_rate": 4.3363164793982e-06, | |
| "loss": 2.2381, | |
| "step": 2596 | |
| }, | |
| { | |
| "epoch": 0.8743426864667031, | |
| "grad_norm": 0.3054344356060028, | |
| "learning_rate": 4.291097814626266e-06, | |
| "loss": 2.2272, | |
| "step": 2598 | |
| }, | |
| { | |
| "epoch": 0.8750157755247991, | |
| "grad_norm": 0.3115767538547516, | |
| "learning_rate": 4.246105582265863e-06, | |
| "loss": 2.2353, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.8756888645828951, | |
| "grad_norm": 0.3003162145614624, | |
| "learning_rate": 4.201340005199194e-06, | |
| "loss": 2.2184, | |
| "step": 2602 | |
| }, | |
| { | |
| "epoch": 0.8763619536409911, | |
| "grad_norm": 0.30938950181007385, | |
| "learning_rate": 4.156801305185592e-06, | |
| "loss": 2.2482, | |
| "step": 2604 | |
| }, | |
| { | |
| "epoch": 0.8770350426990872, | |
| "grad_norm": 0.30393266677856445, | |
| "learning_rate": 4.11248970286055e-06, | |
| "loss": 2.2564, | |
| "step": 2606 | |
| }, | |
| { | |
| "epoch": 0.8777081317571831, | |
| "grad_norm": 0.31042414903640747, | |
| "learning_rate": 4.068405417734556e-06, | |
| "loss": 2.2386, | |
| "step": 2608 | |
| }, | |
| { | |
| "epoch": 0.8783812208152791, | |
| "grad_norm": 0.3065999746322632, | |
| "learning_rate": 4.024548668191996e-06, | |
| "loss": 2.1974, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.8790543098733751, | |
| "grad_norm": 0.30271115899086, | |
| "learning_rate": 3.980919671490091e-06, | |
| "loss": 2.2398, | |
| "step": 2612 | |
| }, | |
| { | |
| "epoch": 0.8797273989314711, | |
| "grad_norm": 0.30883553624153137, | |
| "learning_rate": 3.937518643757854e-06, | |
| "loss": 2.2771, | |
| "step": 2614 | |
| }, | |
| { | |
| "epoch": 0.8804004879895672, | |
| "grad_norm": 0.3028290271759033, | |
| "learning_rate": 3.894345799994953e-06, | |
| "loss": 2.2441, | |
| "step": 2616 | |
| }, | |
| { | |
| "epoch": 0.8810735770476631, | |
| "grad_norm": 0.315090537071228, | |
| "learning_rate": 3.851401354070694e-06, | |
| "loss": 2.2283, | |
| "step": 2618 | |
| }, | |
| { | |
| "epoch": 0.8817466661057591, | |
| "grad_norm": 0.31428200006484985, | |
| "learning_rate": 3.8086855187229433e-06, | |
| "loss": 2.2358, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.8824197551638551, | |
| "grad_norm": 0.2981523275375366, | |
| "learning_rate": 3.766198505557078e-06, | |
| "loss": 2.2666, | |
| "step": 2622 | |
| }, | |
| { | |
| "epoch": 0.8830928442219511, | |
| "grad_norm": 0.32344120740890503, | |
| "learning_rate": 3.7239405250449554e-06, | |
| "loss": 2.2656, | |
| "step": 2624 | |
| }, | |
| { | |
| "epoch": 0.8837659332800472, | |
| "grad_norm": 0.30745527148246765, | |
| "learning_rate": 3.681911786523834e-06, | |
| "loss": 2.2662, | |
| "step": 2626 | |
| }, | |
| { | |
| "epoch": 0.8844390223381431, | |
| "grad_norm": 0.328919380903244, | |
| "learning_rate": 3.640112498195336e-06, | |
| "loss": 2.2531, | |
| "step": 2628 | |
| }, | |
| { | |
| "epoch": 0.8851121113962391, | |
| "grad_norm": 0.3164350986480713, | |
| "learning_rate": 3.598542867124488e-06, | |
| "loss": 2.2241, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.8857852004543351, | |
| "grad_norm": 0.30678606033325195, | |
| "learning_rate": 3.5572030992385985e-06, | |
| "loss": 2.2276, | |
| "step": 2632 | |
| }, | |
| { | |
| "epoch": 0.8864582895124311, | |
| "grad_norm": 0.30014273524284363, | |
| "learning_rate": 3.516093399326287e-06, | |
| "loss": 2.2144, | |
| "step": 2634 | |
| }, | |
| { | |
| "epoch": 0.8871313785705272, | |
| "grad_norm": 0.3091329038143158, | |
| "learning_rate": 3.4752139710364895e-06, | |
| "loss": 2.2914, | |
| "step": 2636 | |
| }, | |
| { | |
| "epoch": 0.8878044676286231, | |
| "grad_norm": 0.3015601933002472, | |
| "learning_rate": 3.434565016877389e-06, | |
| "loss": 2.2749, | |
| "step": 2638 | |
| }, | |
| { | |
| "epoch": 0.8884775566867191, | |
| "grad_norm": 0.3072759211063385, | |
| "learning_rate": 3.3941467382154756e-06, | |
| "loss": 2.2574, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.8891506457448151, | |
| "grad_norm": 0.3078758120536804, | |
| "learning_rate": 3.353959335274509e-06, | |
| "loss": 2.2382, | |
| "step": 2642 | |
| }, | |
| { | |
| "epoch": 0.8898237348029111, | |
| "grad_norm": 0.31162092089653015, | |
| "learning_rate": 3.3140030071345363e-06, | |
| "loss": 2.2175, | |
| "step": 2644 | |
| }, | |
| { | |
| "epoch": 0.8904968238610071, | |
| "grad_norm": 0.3078956604003906, | |
| "learning_rate": 3.274277951730914e-06, | |
| "loss": 2.2284, | |
| "step": 2646 | |
| }, | |
| { | |
| "epoch": 0.8911699129191031, | |
| "grad_norm": 0.30499154329299927, | |
| "learning_rate": 3.2347843658533216e-06, | |
| "loss": 2.2452, | |
| "step": 2648 | |
| }, | |
| { | |
| "epoch": 0.8918430019771991, | |
| "grad_norm": 0.29736632108688354, | |
| "learning_rate": 3.1955224451447716e-06, | |
| "loss": 2.2429, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.8925160910352951, | |
| "grad_norm": 0.30445072054862976, | |
| "learning_rate": 3.156492384100679e-06, | |
| "loss": 2.2254, | |
| "step": 2652 | |
| }, | |
| { | |
| "epoch": 0.8931891800933911, | |
| "grad_norm": 0.3087076246738434, | |
| "learning_rate": 3.1176943760678755e-06, | |
| "loss": 2.2565, | |
| "step": 2654 | |
| }, | |
| { | |
| "epoch": 0.8938622691514871, | |
| "grad_norm": 0.3023519515991211, | |
| "learning_rate": 3.0791286132436094e-06, | |
| "loss": 2.2612, | |
| "step": 2656 | |
| }, | |
| { | |
| "epoch": 0.8945353582095831, | |
| "grad_norm": 0.3006771504878998, | |
| "learning_rate": 3.04079528667468e-06, | |
| "loss": 2.2838, | |
| "step": 2658 | |
| }, | |
| { | |
| "epoch": 0.8952084472676791, | |
| "grad_norm": 0.3032989799976349, | |
| "learning_rate": 3.002694586256427e-06, | |
| "loss": 2.2154, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.895881536325775, | |
| "grad_norm": 0.31874269247055054, | |
| "learning_rate": 2.9648267007318096e-06, | |
| "loss": 2.2215, | |
| "step": 2662 | |
| }, | |
| { | |
| "epoch": 0.8965546253838711, | |
| "grad_norm": 0.29716014862060547, | |
| "learning_rate": 2.9271918176904665e-06, | |
| "loss": 2.2523, | |
| "step": 2664 | |
| }, | |
| { | |
| "epoch": 0.8972277144419671, | |
| "grad_norm": 0.29171815514564514, | |
| "learning_rate": 2.8897901235678036e-06, | |
| "loss": 2.2343, | |
| "step": 2666 | |
| }, | |
| { | |
| "epoch": 0.8979008035000631, | |
| "grad_norm": 0.30495819449424744, | |
| "learning_rate": 2.85262180364404e-06, | |
| "loss": 2.265, | |
| "step": 2668 | |
| }, | |
| { | |
| "epoch": 0.8985738925581591, | |
| "grad_norm": 0.29562342166900635, | |
| "learning_rate": 2.8156870420433137e-06, | |
| "loss": 2.2464, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.899246981616255, | |
| "grad_norm": 0.3044832646846771, | |
| "learning_rate": 2.778986021732771e-06, | |
| "loss": 2.2473, | |
| "step": 2672 | |
| }, | |
| { | |
| "epoch": 0.8999200706743511, | |
| "grad_norm": 0.31051772832870483, | |
| "learning_rate": 2.7425189245216474e-06, | |
| "loss": 2.2233, | |
| "step": 2674 | |
| }, | |
| { | |
| "epoch": 0.9005931597324471, | |
| "grad_norm": 0.3047817647457123, | |
| "learning_rate": 2.7062859310603707e-06, | |
| "loss": 2.3102, | |
| "step": 2676 | |
| }, | |
| { | |
| "epoch": 0.9012662487905431, | |
| "grad_norm": 0.29298290610313416, | |
| "learning_rate": 2.670287220839668e-06, | |
| "loss": 2.2302, | |
| "step": 2678 | |
| }, | |
| { | |
| "epoch": 0.9019393378486391, | |
| "grad_norm": 0.3079676330089569, | |
| "learning_rate": 2.6345229721896734e-06, | |
| "loss": 2.2445, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.902612426906735, | |
| "grad_norm": 0.30154451727867126, | |
| "learning_rate": 2.598993362279045e-06, | |
| "loss": 2.2674, | |
| "step": 2682 | |
| }, | |
| { | |
| "epoch": 0.9032855159648311, | |
| "grad_norm": 0.29776298999786377, | |
| "learning_rate": 2.5636985671141145e-06, | |
| "loss": 2.2305, | |
| "step": 2684 | |
| }, | |
| { | |
| "epoch": 0.9039586050229271, | |
| "grad_norm": 0.3009389340877533, | |
| "learning_rate": 2.5286387615379436e-06, | |
| "loss": 2.234, | |
| "step": 2686 | |
| }, | |
| { | |
| "epoch": 0.9046316940810231, | |
| "grad_norm": 0.3106025457382202, | |
| "learning_rate": 2.4938141192295428e-06, | |
| "loss": 2.2528, | |
| "step": 2688 | |
| }, | |
| { | |
| "epoch": 0.9053047831391191, | |
| "grad_norm": 0.3065546751022339, | |
| "learning_rate": 2.4592248127029593e-06, | |
| "loss": 2.2227, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.905977872197215, | |
| "grad_norm": 0.3005795478820801, | |
| "learning_rate": 2.42487101330644e-06, | |
| "loss": 2.2276, | |
| "step": 2692 | |
| }, | |
| { | |
| "epoch": 0.9066509612553111, | |
| "grad_norm": 0.3081531226634979, | |
| "learning_rate": 2.3907528912215647e-06, | |
| "loss": 2.269, | |
| "step": 2694 | |
| }, | |
| { | |
| "epoch": 0.9073240503134071, | |
| "grad_norm": 0.30866527557373047, | |
| "learning_rate": 2.3568706154624364e-06, | |
| "loss": 2.2805, | |
| "step": 2696 | |
| }, | |
| { | |
| "epoch": 0.9079971393715031, | |
| "grad_norm": 0.3044450879096985, | |
| "learning_rate": 2.323224353874809e-06, | |
| "loss": 2.2795, | |
| "step": 2698 | |
| }, | |
| { | |
| "epoch": 0.9086702284295991, | |
| "grad_norm": 0.3019695580005646, | |
| "learning_rate": 2.2898142731352712e-06, | |
| "loss": 2.181, | |
| "step": 2700 | |
| } | |
| ], | |
| "logging_steps": 2, | |
| "max_steps": 2972, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 300, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.039795302576947e+18, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |