{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.2019267174287998, "eval_steps": 500, "global_step": 600, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00033654452904799967, "grad_norm": 4.331892967224121, "learning_rate": 0.0, "loss": 5.5158, "step": 1 }, { "epoch": 0.0006730890580959993, "grad_norm": 4.539519786834717, "learning_rate": 6.711409395973154e-07, "loss": 5.5718, "step": 2 }, { "epoch": 0.0013461781161919987, "grad_norm": 4.208465576171875, "learning_rate": 2.013422818791946e-06, "loss": 5.4905, "step": 4 }, { "epoch": 0.002019267174287998, "grad_norm": 3.891338586807251, "learning_rate": 3.3557046979865773e-06, "loss": 5.4511, "step": 6 }, { "epoch": 0.0026923562323839974, "grad_norm": 2.979590654373169, "learning_rate": 4.697986577181209e-06, "loss": 5.3311, "step": 8 }, { "epoch": 0.0033654452904799967, "grad_norm": 3.1792373657226562, "learning_rate": 6.04026845637584e-06, "loss": 5.3782, "step": 10 }, { "epoch": 0.004038534348575996, "grad_norm": 3.9909653663635254, "learning_rate": 7.382550335570471e-06, "loss": 5.2583, "step": 12 }, { "epoch": 0.004711623406671995, "grad_norm": 3.31044340133667, "learning_rate": 8.724832214765101e-06, "loss": 5.3198, "step": 14 }, { "epoch": 0.005384712464767995, "grad_norm": 3.3201687335968018, "learning_rate": 1.006711409395973e-05, "loss": 5.1138, "step": 16 }, { "epoch": 0.006057801522863994, "grad_norm": 2.8034451007843018, "learning_rate": 1.1409395973154363e-05, "loss": 5.0022, "step": 18 }, { "epoch": 0.0067308905809599934, "grad_norm": 2.044506072998047, "learning_rate": 1.2751677852348994e-05, "loss": 4.8879, "step": 20 }, { "epoch": 0.007403979639055993, "grad_norm": 1.6194826364517212, "learning_rate": 1.4093959731543624e-05, "loss": 4.7436, "step": 22 }, { "epoch": 0.008077068697151992, "grad_norm": 1.4383995532989502, "learning_rate": 1.5436241610738255e-05, "loss": 4.5798, "step": 24 }, { "epoch": 0.00875015775524799, "grad_norm": 1.174633264541626, "learning_rate": 1.6778523489932888e-05, "loss": 4.4188, "step": 26 }, { "epoch": 0.00942324681334399, "grad_norm": 1.1277130842208862, "learning_rate": 1.8120805369127517e-05, "loss": 4.4374, "step": 28 }, { "epoch": 0.01009633587143999, "grad_norm": 1.0426617860794067, "learning_rate": 1.946308724832215e-05, "loss": 4.2195, "step": 30 }, { "epoch": 0.01076942492953599, "grad_norm": 0.9180749654769897, "learning_rate": 2.080536912751678e-05, "loss": 4.1755, "step": 32 }, { "epoch": 0.011442513987631988, "grad_norm": 1.0104376077651978, "learning_rate": 2.2147651006711412e-05, "loss": 4.0892, "step": 34 }, { "epoch": 0.012115603045727988, "grad_norm": 0.9427777528762817, "learning_rate": 2.348993288590604e-05, "loss": 4.0498, "step": 36 }, { "epoch": 0.012788692103823987, "grad_norm": 1.0120079517364502, "learning_rate": 2.4832214765100674e-05, "loss": 3.9471, "step": 38 }, { "epoch": 0.013461781161919987, "grad_norm": 1.0842680931091309, "learning_rate": 2.6174496644295304e-05, "loss": 3.8882, "step": 40 }, { "epoch": 0.014134870220015985, "grad_norm": 1.3515102863311768, "learning_rate": 2.7516778523489933e-05, "loss": 3.8274, "step": 42 }, { "epoch": 0.014807959278111986, "grad_norm": 1.1003209352493286, "learning_rate": 2.885906040268457e-05, "loss": 3.7735, "step": 44 }, { "epoch": 0.015481048336207984, "grad_norm": 0.9302487373352051, "learning_rate": 3.02013422818792e-05, "loss": 3.731, "step": 46 }, { "epoch": 0.016154137394303984, "grad_norm": 1.0311543941497803, "learning_rate": 3.1543624161073825e-05, "loss": 3.6608, "step": 48 }, { "epoch": 0.016827226452399983, "grad_norm": 1.0927435159683228, "learning_rate": 3.288590604026846e-05, "loss": 3.5962, "step": 50 }, { "epoch": 0.01750031551049598, "grad_norm": 1.3255321979522705, "learning_rate": 3.422818791946309e-05, "loss": 3.5669, "step": 52 }, { "epoch": 0.018173404568591983, "grad_norm": 0.9304305911064148, "learning_rate": 3.557046979865772e-05, "loss": 3.5456, "step": 54 }, { "epoch": 0.01884649362668798, "grad_norm": 1.159280776977539, "learning_rate": 3.6912751677852356e-05, "loss": 3.5189, "step": 56 }, { "epoch": 0.01951958268478398, "grad_norm": 0.9376134276390076, "learning_rate": 3.8255033557046985e-05, "loss": 3.432, "step": 58 }, { "epoch": 0.02019267174287998, "grad_norm": 1.292802095413208, "learning_rate": 3.959731543624161e-05, "loss": 3.4564, "step": 60 }, { "epoch": 0.02086576080097598, "grad_norm": 1.2383852005004883, "learning_rate": 4.0939597315436244e-05, "loss": 3.4194, "step": 62 }, { "epoch": 0.02153884985907198, "grad_norm": 0.8546445369720459, "learning_rate": 4.228187919463087e-05, "loss": 3.3873, "step": 64 }, { "epoch": 0.022211938917167977, "grad_norm": 1.4104743003845215, "learning_rate": 4.36241610738255e-05, "loss": 3.3647, "step": 66 }, { "epoch": 0.022885027975263976, "grad_norm": 1.3548426628112793, "learning_rate": 4.496644295302014e-05, "loss": 3.3213, "step": 68 }, { "epoch": 0.023558117033359978, "grad_norm": 1.1530455350875854, "learning_rate": 4.630872483221477e-05, "loss": 3.2789, "step": 70 }, { "epoch": 0.024231206091455976, "grad_norm": 1.3092457056045532, "learning_rate": 4.76510067114094e-05, "loss": 3.2922, "step": 72 }, { "epoch": 0.024904295149551975, "grad_norm": 1.577699065208435, "learning_rate": 4.8993288590604034e-05, "loss": 3.2308, "step": 74 }, { "epoch": 0.025577384207647973, "grad_norm": 1.3348486423492432, "learning_rate": 5.033557046979866e-05, "loss": 3.2656, "step": 76 }, { "epoch": 0.026250473265743975, "grad_norm": 1.1966625452041626, "learning_rate": 5.167785234899329e-05, "loss": 3.2703, "step": 78 }, { "epoch": 0.026923562323839974, "grad_norm": 1.3125278949737549, "learning_rate": 5.302013422818792e-05, "loss": 3.2034, "step": 80 }, { "epoch": 0.027596651381935972, "grad_norm": 1.1957862377166748, "learning_rate": 5.436241610738255e-05, "loss": 3.2476, "step": 82 }, { "epoch": 0.02826974044003197, "grad_norm": 1.2177337408065796, "learning_rate": 5.570469798657718e-05, "loss": 3.2166, "step": 84 }, { "epoch": 0.028942829498127973, "grad_norm": 0.8806389570236206, "learning_rate": 5.704697986577181e-05, "loss": 3.1722, "step": 86 }, { "epoch": 0.02961591855622397, "grad_norm": 1.8180561065673828, "learning_rate": 5.838926174496645e-05, "loss": 3.1597, "step": 88 }, { "epoch": 0.03028900761431997, "grad_norm": 1.1676297187805176, "learning_rate": 5.973154362416108e-05, "loss": 3.1829, "step": 90 }, { "epoch": 0.030962096672415968, "grad_norm": 1.0163198709487915, "learning_rate": 6.107382550335571e-05, "loss": 3.1643, "step": 92 }, { "epoch": 0.03163518573051197, "grad_norm": 1.0734015703201294, "learning_rate": 6.241610738255034e-05, "loss": 3.1692, "step": 94 }, { "epoch": 0.03230827478860797, "grad_norm": 1.4144916534423828, "learning_rate": 6.375838926174497e-05, "loss": 3.1217, "step": 96 }, { "epoch": 0.03298136384670397, "grad_norm": 1.5647915601730347, "learning_rate": 6.51006711409396e-05, "loss": 3.1324, "step": 98 }, { "epoch": 0.033654452904799965, "grad_norm": 1.1999105215072632, "learning_rate": 6.644295302013423e-05, "loss": 3.1103, "step": 100 }, { "epoch": 0.03432754196289597, "grad_norm": 1.512838363647461, "learning_rate": 6.778523489932886e-05, "loss": 3.1036, "step": 102 }, { "epoch": 0.03500063102099196, "grad_norm": 1.1092990636825562, "learning_rate": 6.912751677852349e-05, "loss": 3.1244, "step": 104 }, { "epoch": 0.035673720079087964, "grad_norm": 1.2763620615005493, "learning_rate": 7.046979865771812e-05, "loss": 3.0989, "step": 106 }, { "epoch": 0.036346809137183966, "grad_norm": 1.2328648567199707, "learning_rate": 7.181208053691275e-05, "loss": 3.0506, "step": 108 }, { "epoch": 0.03701989819527996, "grad_norm": 1.6039047241210938, "learning_rate": 7.315436241610739e-05, "loss": 3.0589, "step": 110 }, { "epoch": 0.03769298725337596, "grad_norm": 1.7161307334899902, "learning_rate": 7.449664429530202e-05, "loss": 3.0296, "step": 112 }, { "epoch": 0.03836607631147196, "grad_norm": 1.2628991603851318, "learning_rate": 7.583892617449665e-05, "loss": 3.063, "step": 114 }, { "epoch": 0.03903916536956796, "grad_norm": 1.1414180994033813, "learning_rate": 7.718120805369128e-05, "loss": 3.0366, "step": 116 }, { "epoch": 0.03971225442766396, "grad_norm": 1.5152932405471802, "learning_rate": 7.852348993288591e-05, "loss": 3.0241, "step": 118 }, { "epoch": 0.04038534348575996, "grad_norm": 1.6119567155838013, "learning_rate": 7.986577181208054e-05, "loss": 3.0139, "step": 120 }, { "epoch": 0.04105843254385596, "grad_norm": 1.6078091859817505, "learning_rate": 8.120805369127518e-05, "loss": 3.0288, "step": 122 }, { "epoch": 0.04173152160195196, "grad_norm": 1.1092705726623535, "learning_rate": 8.255033557046981e-05, "loss": 2.9996, "step": 124 }, { "epoch": 0.042404610660047956, "grad_norm": 1.2352242469787598, "learning_rate": 8.389261744966444e-05, "loss": 2.9872, "step": 126 }, { "epoch": 0.04307769971814396, "grad_norm": 1.636400580406189, "learning_rate": 8.523489932885907e-05, "loss": 2.9814, "step": 128 }, { "epoch": 0.04375078877623995, "grad_norm": 1.4877128601074219, "learning_rate": 8.65771812080537e-05, "loss": 2.9756, "step": 130 }, { "epoch": 0.044423877834335955, "grad_norm": 1.2983709573745728, "learning_rate": 8.791946308724833e-05, "loss": 2.9756, "step": 132 }, { "epoch": 0.04509696689243196, "grad_norm": 1.7350983619689941, "learning_rate": 8.926174496644296e-05, "loss": 2.9579, "step": 134 }, { "epoch": 0.04577005595052795, "grad_norm": 0.978854775428772, "learning_rate": 9.060402684563759e-05, "loss": 2.9269, "step": 136 }, { "epoch": 0.046443145008623954, "grad_norm": 1.362163782119751, "learning_rate": 9.194630872483221e-05, "loss": 2.99, "step": 138 }, { "epoch": 0.047116234066719956, "grad_norm": 1.328202247619629, "learning_rate": 9.328859060402684e-05, "loss": 2.966, "step": 140 }, { "epoch": 0.04778932312481595, "grad_norm": 1.2497445344924927, "learning_rate": 9.463087248322147e-05, "loss": 2.9254, "step": 142 }, { "epoch": 0.04846241218291195, "grad_norm": 1.071092128753662, "learning_rate": 9.59731543624161e-05, "loss": 2.9597, "step": 144 }, { "epoch": 0.04913550124100795, "grad_norm": 1.451729416847229, "learning_rate": 9.731543624161075e-05, "loss": 2.9605, "step": 146 }, { "epoch": 0.04980859029910395, "grad_norm": 1.328731656074524, "learning_rate": 9.865771812080538e-05, "loss": 2.9493, "step": 148 }, { "epoch": 0.05048167935719995, "grad_norm": 1.4675222635269165, "learning_rate": 0.0001, "loss": 2.9298, "step": 150 }, { "epoch": 0.05115476841529595, "grad_norm": 1.208961009979248, "learning_rate": 9.99998761551904e-05, "loss": 2.9382, "step": 152 }, { "epoch": 0.05182785747339195, "grad_norm": 1.0392056703567505, "learning_rate": 9.999950462137508e-05, "loss": 2.8829, "step": 154 }, { "epoch": 0.05250094653148795, "grad_norm": 1.378711462020874, "learning_rate": 9.999888540039458e-05, "loss": 2.9041, "step": 156 }, { "epoch": 0.053174035589583946, "grad_norm": 1.2687252759933472, "learning_rate": 9.999801849531635e-05, "loss": 2.9148, "step": 158 }, { "epoch": 0.05384712464767995, "grad_norm": 1.2382102012634277, "learning_rate": 9.999690391043487e-05, "loss": 2.9107, "step": 160 }, { "epoch": 0.05452021370577594, "grad_norm": 1.0215928554534912, "learning_rate": 9.999554165127159e-05, "loss": 2.9187, "step": 162 }, { "epoch": 0.055193302763871945, "grad_norm": 1.2638540267944336, "learning_rate": 9.99939317245748e-05, "loss": 2.9261, "step": 164 }, { "epoch": 0.055866391821967946, "grad_norm": 1.3826959133148193, "learning_rate": 9.999207413831982e-05, "loss": 2.8944, "step": 166 }, { "epoch": 0.05653948088006394, "grad_norm": 1.0764875411987305, "learning_rate": 9.998996890170867e-05, "loss": 2.9031, "step": 168 }, { "epoch": 0.05721256993815994, "grad_norm": 1.5163936614990234, "learning_rate": 9.99876160251703e-05, "loss": 2.8687, "step": 170 }, { "epoch": 0.057885658996255945, "grad_norm": 1.5821291208267212, "learning_rate": 9.998501552036037e-05, "loss": 2.8828, "step": 172 }, { "epoch": 0.05855874805435194, "grad_norm": 1.4572120904922485, "learning_rate": 9.998216740016124e-05, "loss": 2.8898, "step": 174 }, { "epoch": 0.05923183711244794, "grad_norm": 1.2185218334197998, "learning_rate": 9.99790716786819e-05, "loss": 2.8456, "step": 176 }, { "epoch": 0.05990492617054394, "grad_norm": 1.106446385383606, "learning_rate": 9.99757283712579e-05, "loss": 2.8793, "step": 178 }, { "epoch": 0.06057801522863994, "grad_norm": 1.1448893547058105, "learning_rate": 9.997213749445129e-05, "loss": 2.8579, "step": 180 }, { "epoch": 0.06125110428673594, "grad_norm": 1.1587834358215332, "learning_rate": 9.996829906605056e-05, "loss": 2.8839, "step": 182 }, { "epoch": 0.061924193344831936, "grad_norm": 1.0969592332839966, "learning_rate": 9.996421310507046e-05, "loss": 2.8638, "step": 184 }, { "epoch": 0.06259728240292793, "grad_norm": 0.9740116000175476, "learning_rate": 9.9959879631752e-05, "loss": 2.8455, "step": 186 }, { "epoch": 0.06327037146102393, "grad_norm": 1.2307910919189453, "learning_rate": 9.995529866756231e-05, "loss": 2.8534, "step": 188 }, { "epoch": 0.06394346051911994, "grad_norm": 1.5021939277648926, "learning_rate": 9.995047023519452e-05, "loss": 2.8469, "step": 190 }, { "epoch": 0.06461654957721594, "grad_norm": 1.1044224500656128, "learning_rate": 9.994539435856771e-05, "loss": 2.8429, "step": 192 }, { "epoch": 0.06528963863531194, "grad_norm": 1.4586883783340454, "learning_rate": 9.99400710628267e-05, "loss": 2.836, "step": 194 }, { "epoch": 0.06596272769340794, "grad_norm": 1.2613426446914673, "learning_rate": 9.993450037434199e-05, "loss": 2.8243, "step": 196 }, { "epoch": 0.06663581675150393, "grad_norm": 1.0347422361373901, "learning_rate": 9.992868232070963e-05, "loss": 2.7965, "step": 198 }, { "epoch": 0.06730890580959993, "grad_norm": 2.1357574462890625, "learning_rate": 9.992261693075103e-05, "loss": 2.8486, "step": 200 }, { "epoch": 0.06798199486769593, "grad_norm": 1.0357908010482788, "learning_rate": 9.991630423451286e-05, "loss": 2.8386, "step": 202 }, { "epoch": 0.06865508392579193, "grad_norm": 1.1383159160614014, "learning_rate": 9.990974426326696e-05, "loss": 2.7874, "step": 204 }, { "epoch": 0.06932817298388794, "grad_norm": 0.8452678322792053, "learning_rate": 9.990293704951001e-05, "loss": 2.786, "step": 206 }, { "epoch": 0.07000126204198392, "grad_norm": 0.9482727646827698, "learning_rate": 9.989588262696357e-05, "loss": 2.8156, "step": 208 }, { "epoch": 0.07067435110007993, "grad_norm": 0.8251766562461853, "learning_rate": 9.988858103057378e-05, "loss": 2.7588, "step": 210 }, { "epoch": 0.07134744015817593, "grad_norm": 1.211065649986267, "learning_rate": 9.988103229651121e-05, "loss": 2.7623, "step": 212 }, { "epoch": 0.07202052921627193, "grad_norm": 0.8990377187728882, "learning_rate": 9.987323646217075e-05, "loss": 2.8164, "step": 214 }, { "epoch": 0.07269361827436793, "grad_norm": 0.9878025054931641, "learning_rate": 9.986519356617132e-05, "loss": 2.7847, "step": 216 }, { "epoch": 0.07336670733246392, "grad_norm": 0.7551445364952087, "learning_rate": 9.985690364835576e-05, "loss": 2.8111, "step": 218 }, { "epoch": 0.07403979639055992, "grad_norm": 0.9582260251045227, "learning_rate": 9.984836674979062e-05, "loss": 2.793, "step": 220 }, { "epoch": 0.07471288544865592, "grad_norm": 0.8087739944458008, "learning_rate": 9.983958291276591e-05, "loss": 2.7464, "step": 222 }, { "epoch": 0.07538597450675193, "grad_norm": 1.2373522520065308, "learning_rate": 9.983055218079493e-05, "loss": 2.7656, "step": 224 }, { "epoch": 0.07605906356484793, "grad_norm": 0.9746289849281311, "learning_rate": 9.982127459861408e-05, "loss": 2.7765, "step": 226 }, { "epoch": 0.07673215262294392, "grad_norm": 0.6946307420730591, "learning_rate": 9.981175021218255e-05, "loss": 2.7491, "step": 228 }, { "epoch": 0.07740524168103992, "grad_norm": 0.8959107398986816, "learning_rate": 9.980197906868215e-05, "loss": 2.7565, "step": 230 }, { "epoch": 0.07807833073913592, "grad_norm": 0.9889335036277771, "learning_rate": 9.979196121651716e-05, "loss": 2.7974, "step": 232 }, { "epoch": 0.07875141979723192, "grad_norm": 0.9802746176719666, "learning_rate": 9.978169670531388e-05, "loss": 2.7772, "step": 234 }, { "epoch": 0.07942450885532792, "grad_norm": 0.6934760808944702, "learning_rate": 9.977118558592059e-05, "loss": 2.7602, "step": 236 }, { "epoch": 0.08009759791342393, "grad_norm": 0.8996357917785645, "learning_rate": 9.97604279104072e-05, "loss": 2.7669, "step": 238 }, { "epoch": 0.08077068697151991, "grad_norm": 0.8844061493873596, "learning_rate": 9.974942373206499e-05, "loss": 2.7458, "step": 240 }, { "epoch": 0.08144377602961592, "grad_norm": 1.023626685142517, "learning_rate": 9.973817310540638e-05, "loss": 2.7639, "step": 242 }, { "epoch": 0.08211686508771192, "grad_norm": 0.8241132497787476, "learning_rate": 9.972667608616466e-05, "loss": 2.7457, "step": 244 }, { "epoch": 0.08278995414580792, "grad_norm": 0.7864794135093689, "learning_rate": 9.971493273129364e-05, "loss": 2.763, "step": 246 }, { "epoch": 0.08346304320390392, "grad_norm": 1.071751356124878, "learning_rate": 9.970294309896747e-05, "loss": 2.7347, "step": 248 }, { "epoch": 0.08413613226199991, "grad_norm": 0.8978875279426575, "learning_rate": 9.969070724858031e-05, "loss": 2.7807, "step": 250 }, { "epoch": 0.08480922132009591, "grad_norm": 0.9984204769134521, "learning_rate": 9.967822524074602e-05, "loss": 2.7399, "step": 252 }, { "epoch": 0.08548231037819191, "grad_norm": 0.8611739873886108, "learning_rate": 9.966549713729787e-05, "loss": 2.753, "step": 254 }, { "epoch": 0.08615539943628792, "grad_norm": 0.8647720217704773, "learning_rate": 9.965252300128826e-05, "loss": 2.7224, "step": 256 }, { "epoch": 0.08682848849438392, "grad_norm": 0.8688477873802185, "learning_rate": 9.963930289698833e-05, "loss": 2.6879, "step": 258 }, { "epoch": 0.0875015775524799, "grad_norm": 1.1445469856262207, "learning_rate": 9.962583688988778e-05, "loss": 2.739, "step": 260 }, { "epoch": 0.08817466661057591, "grad_norm": 0.8668599128723145, "learning_rate": 9.961212504669437e-05, "loss": 2.6962, "step": 262 }, { "epoch": 0.08884775566867191, "grad_norm": 0.905125617980957, "learning_rate": 9.959816743533375e-05, "loss": 2.7239, "step": 264 }, { "epoch": 0.08952084472676791, "grad_norm": 0.8252028822898865, "learning_rate": 9.958396412494901e-05, "loss": 2.7381, "step": 266 }, { "epoch": 0.09019393378486391, "grad_norm": 0.7380514740943909, "learning_rate": 9.956951518590043e-05, "loss": 2.7135, "step": 268 }, { "epoch": 0.09086702284295992, "grad_norm": 0.7395239472389221, "learning_rate": 9.955482068976502e-05, "loss": 2.6954, "step": 270 }, { "epoch": 0.0915401119010559, "grad_norm": 0.6564229726791382, "learning_rate": 9.953988070933628e-05, "loss": 2.7145, "step": 272 }, { "epoch": 0.0922132009591519, "grad_norm": 0.7306910157203674, "learning_rate": 9.952469531862378e-05, "loss": 2.6951, "step": 274 }, { "epoch": 0.09288629001724791, "grad_norm": 0.6810031533241272, "learning_rate": 9.950926459285277e-05, "loss": 2.7201, "step": 276 }, { "epoch": 0.09355937907534391, "grad_norm": 0.6724168658256531, "learning_rate": 9.949358860846388e-05, "loss": 2.7112, "step": 278 }, { "epoch": 0.09423246813343991, "grad_norm": 0.7065703272819519, "learning_rate": 9.947766744311268e-05, "loss": 2.6884, "step": 280 }, { "epoch": 0.0949055571915359, "grad_norm": 0.8231908679008484, "learning_rate": 9.946150117566931e-05, "loss": 2.7286, "step": 282 }, { "epoch": 0.0955786462496319, "grad_norm": 0.9570270776748657, "learning_rate": 9.944508988621812e-05, "loss": 2.7166, "step": 284 }, { "epoch": 0.0962517353077279, "grad_norm": 0.9357023239135742, "learning_rate": 9.94284336560572e-05, "loss": 2.6768, "step": 286 }, { "epoch": 0.0969248243658239, "grad_norm": 0.6350796222686768, "learning_rate": 9.941153256769809e-05, "loss": 2.6921, "step": 288 }, { "epoch": 0.09759791342391991, "grad_norm": 0.6700872778892517, "learning_rate": 9.939438670486525e-05, "loss": 2.6847, "step": 290 }, { "epoch": 0.0982710024820159, "grad_norm": 0.6851752400398254, "learning_rate": 9.937699615249572e-05, "loss": 2.6586, "step": 292 }, { "epoch": 0.0989440915401119, "grad_norm": 0.7098946571350098, "learning_rate": 9.935936099673871e-05, "loss": 2.6793, "step": 294 }, { "epoch": 0.0996171805982079, "grad_norm": 0.680543839931488, "learning_rate": 9.934148132495511e-05, "loss": 2.6763, "step": 296 }, { "epoch": 0.1002902696563039, "grad_norm": 0.6832155585289001, "learning_rate": 9.932335722571709e-05, "loss": 2.6768, "step": 298 }, { "epoch": 0.1009633587143999, "grad_norm": 0.8236553072929382, "learning_rate": 9.930498878880768e-05, "loss": 2.6738, "step": 300 }, { "epoch": 0.10163644777249589, "grad_norm": 1.2765145301818848, "learning_rate": 9.928637610522032e-05, "loss": 2.6575, "step": 302 }, { "epoch": 0.1023095368305919, "grad_norm": 0.8765101432800293, "learning_rate": 9.926751926715836e-05, "loss": 2.6521, "step": 304 }, { "epoch": 0.1029826258886879, "grad_norm": 0.7371405959129333, "learning_rate": 9.924841836803467e-05, "loss": 2.6916, "step": 306 }, { "epoch": 0.1036557149467839, "grad_norm": 0.8433207273483276, "learning_rate": 9.922907350247115e-05, "loss": 2.6605, "step": 308 }, { "epoch": 0.1043288040048799, "grad_norm": 0.7216758131980896, "learning_rate": 9.920948476629821e-05, "loss": 2.6686, "step": 310 }, { "epoch": 0.1050018930629759, "grad_norm": 0.7831395864486694, "learning_rate": 9.918965225655442e-05, "loss": 2.6783, "step": 312 }, { "epoch": 0.10567498212107189, "grad_norm": 0.5831722617149353, "learning_rate": 9.916957607148591e-05, "loss": 2.6633, "step": 314 }, { "epoch": 0.10634807117916789, "grad_norm": 0.6143165826797485, "learning_rate": 9.91492563105459e-05, "loss": 2.6785, "step": 316 }, { "epoch": 0.1070211602372639, "grad_norm": 0.8080072999000549, "learning_rate": 9.912869307439432e-05, "loss": 2.672, "step": 318 }, { "epoch": 0.1076942492953599, "grad_norm": 1.0352627038955688, "learning_rate": 9.910788646489713e-05, "loss": 2.6474, "step": 320 }, { "epoch": 0.1083673383534559, "grad_norm": 0.6871941685676575, "learning_rate": 9.908683658512597e-05, "loss": 2.6387, "step": 322 }, { "epoch": 0.10904042741155189, "grad_norm": 0.7802032828330994, "learning_rate": 9.906554353935758e-05, "loss": 2.6584, "step": 324 }, { "epoch": 0.10971351646964789, "grad_norm": 0.7139161825180054, "learning_rate": 9.904400743307326e-05, "loss": 2.6728, "step": 326 }, { "epoch": 0.11038660552774389, "grad_norm": 1.0033543109893799, "learning_rate": 9.902222837295844e-05, "loss": 2.666, "step": 328 }, { "epoch": 0.11105969458583989, "grad_norm": 1.0726121664047241, "learning_rate": 9.900020646690205e-05, "loss": 2.6566, "step": 330 }, { "epoch": 0.11173278364393589, "grad_norm": 0.8290823698043823, "learning_rate": 9.897794182399606e-05, "loss": 2.6592, "step": 332 }, { "epoch": 0.11240587270203188, "grad_norm": 0.7432393431663513, "learning_rate": 9.895543455453486e-05, "loss": 2.6604, "step": 334 }, { "epoch": 0.11307896176012788, "grad_norm": 0.6097532510757446, "learning_rate": 9.893268477001479e-05, "loss": 2.6613, "step": 336 }, { "epoch": 0.11375205081822388, "grad_norm": 0.6818569898605347, "learning_rate": 9.890969258313358e-05, "loss": 2.6238, "step": 338 }, { "epoch": 0.11442513987631989, "grad_norm": 0.6779484748840332, "learning_rate": 9.888645810778974e-05, "loss": 2.6512, "step": 340 }, { "epoch": 0.11509822893441589, "grad_norm": 0.697517454624176, "learning_rate": 9.886298145908203e-05, "loss": 2.6278, "step": 342 }, { "epoch": 0.11577131799251189, "grad_norm": 0.700210452079773, "learning_rate": 9.88392627533089e-05, "loss": 2.6592, "step": 344 }, { "epoch": 0.11644440705060788, "grad_norm": 0.6507501602172852, "learning_rate": 9.88153021079679e-05, "loss": 2.6682, "step": 346 }, { "epoch": 0.11711749610870388, "grad_norm": 0.7211961150169373, "learning_rate": 9.87910996417551e-05, "loss": 2.6919, "step": 348 }, { "epoch": 0.11779058516679988, "grad_norm": 0.7282651662826538, "learning_rate": 9.876665547456446e-05, "loss": 2.6148, "step": 350 }, { "epoch": 0.11846367422489588, "grad_norm": 0.6177457571029663, "learning_rate": 9.874196972748735e-05, "loss": 2.6072, "step": 352 }, { "epoch": 0.11913676328299189, "grad_norm": 0.5497561693191528, "learning_rate": 9.871704252281179e-05, "loss": 2.6493, "step": 354 }, { "epoch": 0.11980985234108787, "grad_norm": 0.6414313912391663, "learning_rate": 9.869187398402202e-05, "loss": 2.6505, "step": 356 }, { "epoch": 0.12048294139918388, "grad_norm": 0.6123400330543518, "learning_rate": 9.866646423579773e-05, "loss": 2.6168, "step": 358 }, { "epoch": 0.12115603045727988, "grad_norm": 0.5811213850975037, "learning_rate": 9.864081340401354e-05, "loss": 2.6367, "step": 360 }, { "epoch": 0.12182911951537588, "grad_norm": 0.7445274591445923, "learning_rate": 9.861492161573837e-05, "loss": 2.6288, "step": 362 }, { "epoch": 0.12250220857347188, "grad_norm": 0.9515259861946106, "learning_rate": 9.858878899923473e-05, "loss": 2.6458, "step": 364 }, { "epoch": 0.12317529763156787, "grad_norm": 0.7661240696907043, "learning_rate": 9.856241568395818e-05, "loss": 2.6289, "step": 366 }, { "epoch": 0.12384838668966387, "grad_norm": 0.704278826713562, "learning_rate": 9.853580180055669e-05, "loss": 2.6043, "step": 368 }, { "epoch": 0.12452147574775987, "grad_norm": 0.6069918274879456, "learning_rate": 9.850894748086986e-05, "loss": 2.6193, "step": 370 }, { "epoch": 0.12519456480585586, "grad_norm": 0.8171132802963257, "learning_rate": 9.848185285792844e-05, "loss": 2.6482, "step": 372 }, { "epoch": 0.12586765386395188, "grad_norm": 0.7318093776702881, "learning_rate": 9.845451806595354e-05, "loss": 2.6285, "step": 374 }, { "epoch": 0.12654074292204787, "grad_norm": 0.7122377157211304, "learning_rate": 9.84269432403561e-05, "loss": 2.6147, "step": 376 }, { "epoch": 0.12721383198014388, "grad_norm": 0.6400437355041504, "learning_rate": 9.839912851773602e-05, "loss": 2.576, "step": 378 }, { "epoch": 0.12788692103823987, "grad_norm": 0.6208680868148804, "learning_rate": 9.837107403588171e-05, "loss": 2.6056, "step": 380 }, { "epoch": 0.12856001009633586, "grad_norm": 0.6690713167190552, "learning_rate": 9.834277993376924e-05, "loss": 2.5884, "step": 382 }, { "epoch": 0.12923309915443187, "grad_norm": 0.613250732421875, "learning_rate": 9.831424635156168e-05, "loss": 2.5824, "step": 384 }, { "epoch": 0.12990618821252786, "grad_norm": 0.6921446323394775, "learning_rate": 9.82854734306085e-05, "loss": 2.6239, "step": 386 }, { "epoch": 0.13057927727062388, "grad_norm": 0.6627881526947021, "learning_rate": 9.825646131344478e-05, "loss": 2.5859, "step": 388 }, { "epoch": 0.13125236632871987, "grad_norm": 0.7399138808250427, "learning_rate": 9.822721014379052e-05, "loss": 2.6492, "step": 390 }, { "epoch": 0.13192545538681588, "grad_norm": 0.633929967880249, "learning_rate": 9.819772006654993e-05, "loss": 2.5935, "step": 392 }, { "epoch": 0.13259854444491187, "grad_norm": 0.6434243321418762, "learning_rate": 9.816799122781075e-05, "loss": 2.6231, "step": 394 }, { "epoch": 0.13327163350300786, "grad_norm": 0.6136692762374878, "learning_rate": 9.813802377484346e-05, "loss": 2.5803, "step": 396 }, { "epoch": 0.13394472256110387, "grad_norm": 0.6683850288391113, "learning_rate": 9.810781785610062e-05, "loss": 2.5854, "step": 398 }, { "epoch": 0.13461781161919986, "grad_norm": 0.705646276473999, "learning_rate": 9.807737362121605e-05, "loss": 2.5954, "step": 400 }, { "epoch": 0.13529090067729588, "grad_norm": 0.7973448038101196, "learning_rate": 9.80466912210042e-05, "loss": 2.5824, "step": 402 }, { "epoch": 0.13596398973539187, "grad_norm": 0.7568982839584351, "learning_rate": 9.801577080745929e-05, "loss": 2.5875, "step": 404 }, { "epoch": 0.13663707879348785, "grad_norm": 0.6286936402320862, "learning_rate": 9.798461253375462e-05, "loss": 2.6047, "step": 406 }, { "epoch": 0.13731016785158387, "grad_norm": 0.6657843589782715, "learning_rate": 9.795321655424184e-05, "loss": 2.6066, "step": 408 }, { "epoch": 0.13798325690967986, "grad_norm": 0.6954363584518433, "learning_rate": 9.792158302445009e-05, "loss": 2.5885, "step": 410 }, { "epoch": 0.13865634596777587, "grad_norm": 0.9607271552085876, "learning_rate": 9.788971210108533e-05, "loss": 2.6212, "step": 412 }, { "epoch": 0.13932943502587186, "grad_norm": 0.9636927843093872, "learning_rate": 9.785760394202948e-05, "loss": 2.5552, "step": 414 }, { "epoch": 0.14000252408396785, "grad_norm": 0.6435789465904236, "learning_rate": 9.782525870633967e-05, "loss": 2.6199, "step": 416 }, { "epoch": 0.14067561314206387, "grad_norm": 0.7337214350700378, "learning_rate": 9.779267655424754e-05, "loss": 2.5683, "step": 418 }, { "epoch": 0.14134870220015985, "grad_norm": 0.6656831502914429, "learning_rate": 9.775985764715827e-05, "loss": 2.5735, "step": 420 }, { "epoch": 0.14202179125825587, "grad_norm": 0.6302557587623596, "learning_rate": 9.772680214764992e-05, "loss": 2.5752, "step": 422 }, { "epoch": 0.14269488031635186, "grad_norm": 0.7694079279899597, "learning_rate": 9.769351021947258e-05, "loss": 2.5742, "step": 424 }, { "epoch": 0.14336796937444785, "grad_norm": 0.743331253528595, "learning_rate": 9.765998202754753e-05, "loss": 2.5968, "step": 426 }, { "epoch": 0.14404105843254386, "grad_norm": 0.7328934669494629, "learning_rate": 9.762621773796647e-05, "loss": 2.595, "step": 428 }, { "epoch": 0.14471414749063985, "grad_norm": 0.786054253578186, "learning_rate": 9.759221751799072e-05, "loss": 2.6048, "step": 430 }, { "epoch": 0.14538723654873587, "grad_norm": 0.7093940377235413, "learning_rate": 9.755798153605027e-05, "loss": 2.5766, "step": 432 }, { "epoch": 0.14606032560683185, "grad_norm": 0.6153858304023743, "learning_rate": 9.752350996174308e-05, "loss": 2.5598, "step": 434 }, { "epoch": 0.14673341466492784, "grad_norm": 0.7648425102233887, "learning_rate": 9.748880296583418e-05, "loss": 2.5768, "step": 436 }, { "epoch": 0.14740650372302386, "grad_norm": 0.777052640914917, "learning_rate": 9.745386072025481e-05, "loss": 2.5571, "step": 438 }, { "epoch": 0.14807959278111985, "grad_norm": 0.6382849812507629, "learning_rate": 9.741868339810161e-05, "loss": 2.564, "step": 440 }, { "epoch": 0.14875268183921586, "grad_norm": 0.6427432298660278, "learning_rate": 9.738327117363571e-05, "loss": 2.5625, "step": 442 }, { "epoch": 0.14942577089731185, "grad_norm": 0.6348729133605957, "learning_rate": 9.734762422228194e-05, "loss": 2.5622, "step": 444 }, { "epoch": 0.15009885995540784, "grad_norm": 0.7136998176574707, "learning_rate": 9.731174272062791e-05, "loss": 2.5386, "step": 446 }, { "epoch": 0.15077194901350385, "grad_norm": 0.7903639078140259, "learning_rate": 9.727562684642308e-05, "loss": 2.5593, "step": 448 }, { "epoch": 0.15144503807159984, "grad_norm": 0.6537742018699646, "learning_rate": 9.723927677857805e-05, "loss": 2.5373, "step": 450 }, { "epoch": 0.15211812712969586, "grad_norm": 0.6430116295814514, "learning_rate": 9.720269269716347e-05, "loss": 2.583, "step": 452 }, { "epoch": 0.15279121618779185, "grad_norm": 0.5976120829582214, "learning_rate": 9.716587478340928e-05, "loss": 2.5729, "step": 454 }, { "epoch": 0.15346430524588783, "grad_norm": 0.7713409066200256, "learning_rate": 9.712882321970383e-05, "loss": 2.5858, "step": 456 }, { "epoch": 0.15413739430398385, "grad_norm": 0.6205691695213318, "learning_rate": 9.70915381895928e-05, "loss": 2.5248, "step": 458 }, { "epoch": 0.15481048336207984, "grad_norm": 0.5880588293075562, "learning_rate": 9.705401987777855e-05, "loss": 2.5811, "step": 460 }, { "epoch": 0.15548357242017585, "grad_norm": 0.645301342010498, "learning_rate": 9.701626847011899e-05, "loss": 2.5677, "step": 462 }, { "epoch": 0.15615666147827184, "grad_norm": 0.5872762799263, "learning_rate": 9.697828415362674e-05, "loss": 2.5777, "step": 464 }, { "epoch": 0.15682975053636786, "grad_norm": 0.5601847767829895, "learning_rate": 9.694006711646823e-05, "loss": 2.5563, "step": 466 }, { "epoch": 0.15750283959446384, "grad_norm": 0.6171532869338989, "learning_rate": 9.690161754796274e-05, "loss": 2.5559, "step": 468 }, { "epoch": 0.15817592865255983, "grad_norm": 0.5213554501533508, "learning_rate": 9.686293563858142e-05, "loss": 2.5543, "step": 470 }, { "epoch": 0.15884901771065585, "grad_norm": 0.5173961520195007, "learning_rate": 9.682402157994643e-05, "loss": 2.5674, "step": 472 }, { "epoch": 0.15952210676875184, "grad_norm": 0.5519447326660156, "learning_rate": 9.678487556482996e-05, "loss": 2.5385, "step": 474 }, { "epoch": 0.16019519582684785, "grad_norm": 0.6517807841300964, "learning_rate": 9.674549778715322e-05, "loss": 2.5712, "step": 476 }, { "epoch": 0.16086828488494384, "grad_norm": 0.745688796043396, "learning_rate": 9.670588844198554e-05, "loss": 2.5743, "step": 478 }, { "epoch": 0.16154137394303983, "grad_norm": 0.6915059685707092, "learning_rate": 9.666604772554342e-05, "loss": 2.5536, "step": 480 }, { "epoch": 0.16221446300113584, "grad_norm": 0.6770395636558533, "learning_rate": 9.662597583518946e-05, "loss": 2.5501, "step": 482 }, { "epoch": 0.16288755205923183, "grad_norm": 0.7015348076820374, "learning_rate": 9.658567296943151e-05, "loss": 2.5488, "step": 484 }, { "epoch": 0.16356064111732785, "grad_norm": 0.6871614456176758, "learning_rate": 9.654513932792157e-05, "loss": 2.55, "step": 486 }, { "epoch": 0.16423373017542384, "grad_norm": 0.634372353553772, "learning_rate": 9.650437511145494e-05, "loss": 2.5231, "step": 488 }, { "epoch": 0.16490681923351982, "grad_norm": 0.6015982031822205, "learning_rate": 9.646338052196904e-05, "loss": 2.5338, "step": 490 }, { "epoch": 0.16557990829161584, "grad_norm": 0.5995615124702454, "learning_rate": 9.642215576254256e-05, "loss": 2.5449, "step": 492 }, { "epoch": 0.16625299734971183, "grad_norm": 0.5862424373626709, "learning_rate": 9.638070103739443e-05, "loss": 2.5564, "step": 494 }, { "epoch": 0.16692608640780784, "grad_norm": 0.49670928716659546, "learning_rate": 9.633901655188271e-05, "loss": 2.5283, "step": 496 }, { "epoch": 0.16759917546590383, "grad_norm": 0.5854897499084473, "learning_rate": 9.629710251250372e-05, "loss": 2.5142, "step": 498 }, { "epoch": 0.16827226452399982, "grad_norm": 0.7148288488388062, "learning_rate": 9.625495912689088e-05, "loss": 2.5528, "step": 500 }, { "epoch": 0.16894535358209584, "grad_norm": 0.7020834684371948, "learning_rate": 9.62125866038138e-05, "loss": 2.5157, "step": 502 }, { "epoch": 0.16961844264019182, "grad_norm": 0.5895742774009705, "learning_rate": 9.616998515317714e-05, "loss": 2.5492, "step": 504 }, { "epoch": 0.17029153169828784, "grad_norm": 0.5936703681945801, "learning_rate": 9.612715498601966e-05, "loss": 2.5172, "step": 506 }, { "epoch": 0.17096462075638383, "grad_norm": 0.6276780962944031, "learning_rate": 9.608409631451311e-05, "loss": 2.5448, "step": 508 }, { "epoch": 0.17163770981447982, "grad_norm": 0.6327413320541382, "learning_rate": 9.60408093519612e-05, "loss": 2.4831, "step": 510 }, { "epoch": 0.17231079887257583, "grad_norm": 0.7642117142677307, "learning_rate": 9.599729431279856e-05, "loss": 2.5209, "step": 512 }, { "epoch": 0.17298388793067182, "grad_norm": 0.7974164485931396, "learning_rate": 9.595355141258966e-05, "loss": 2.5102, "step": 514 }, { "epoch": 0.17365697698876784, "grad_norm": 0.7302394509315491, "learning_rate": 9.590958086802776e-05, "loss": 2.4705, "step": 516 }, { "epoch": 0.17433006604686382, "grad_norm": 0.5221154093742371, "learning_rate": 9.586538289693378e-05, "loss": 2.5355, "step": 518 }, { "epoch": 0.1750031551049598, "grad_norm": 0.7747745513916016, "learning_rate": 9.582095771825533e-05, "loss": 2.5375, "step": 520 }, { "epoch": 0.17567624416305583, "grad_norm": 0.6088679432868958, "learning_rate": 9.577630555206549e-05, "loss": 2.546, "step": 522 }, { "epoch": 0.17634933322115182, "grad_norm": 0.7077321410179138, "learning_rate": 9.573142661956183e-05, "loss": 2.5199, "step": 524 }, { "epoch": 0.17702242227924783, "grad_norm": 0.7267037630081177, "learning_rate": 9.568632114306524e-05, "loss": 2.5125, "step": 526 }, { "epoch": 0.17769551133734382, "grad_norm": 0.591948926448822, "learning_rate": 9.564098934601895e-05, "loss": 2.5411, "step": 528 }, { "epoch": 0.1783686003954398, "grad_norm": 0.5513604283332825, "learning_rate": 9.559543145298721e-05, "loss": 2.5178, "step": 530 }, { "epoch": 0.17904168945353582, "grad_norm": 0.6161476373672485, "learning_rate": 9.55496476896544e-05, "loss": 2.5123, "step": 532 }, { "epoch": 0.1797147785116318, "grad_norm": 0.5339455604553223, "learning_rate": 9.550363828282375e-05, "loss": 2.532, "step": 534 }, { "epoch": 0.18038786756972783, "grad_norm": 0.5792993307113647, "learning_rate": 9.545740346041634e-05, "loss": 2.5163, "step": 536 }, { "epoch": 0.18106095662782382, "grad_norm": 0.6918086409568787, "learning_rate": 9.541094345146986e-05, "loss": 2.516, "step": 538 }, { "epoch": 0.18173404568591983, "grad_norm": 0.577778160572052, "learning_rate": 9.536425848613754e-05, "loss": 2.4866, "step": 540 }, { "epoch": 0.18240713474401582, "grad_norm": 0.6180362105369568, "learning_rate": 9.531734879568703e-05, "loss": 2.5186, "step": 542 }, { "epoch": 0.1830802238021118, "grad_norm": 0.6912726163864136, "learning_rate": 9.527021461249919e-05, "loss": 2.5302, "step": 544 }, { "epoch": 0.18375331286020782, "grad_norm": 0.6083509922027588, "learning_rate": 9.522285617006695e-05, "loss": 2.5226, "step": 546 }, { "epoch": 0.1844264019183038, "grad_norm": 0.6165098547935486, "learning_rate": 9.517527370299424e-05, "loss": 2.5136, "step": 548 }, { "epoch": 0.18509949097639983, "grad_norm": 0.6208174228668213, "learning_rate": 9.512746744699473e-05, "loss": 2.4967, "step": 550 }, { "epoch": 0.18577258003449582, "grad_norm": 0.5808335542678833, "learning_rate": 9.507943763889062e-05, "loss": 2.5075, "step": 552 }, { "epoch": 0.1864456690925918, "grad_norm": 0.5772544741630554, "learning_rate": 9.503118451661168e-05, "loss": 2.517, "step": 554 }, { "epoch": 0.18711875815068782, "grad_norm": 0.5901827812194824, "learning_rate": 9.498270831919383e-05, "loss": 2.5008, "step": 556 }, { "epoch": 0.1877918472087838, "grad_norm": 0.6181182265281677, "learning_rate": 9.493400928677809e-05, "loss": 2.4841, "step": 558 }, { "epoch": 0.18846493626687982, "grad_norm": 0.6515311002731323, "learning_rate": 9.488508766060935e-05, "loss": 2.5213, "step": 560 }, { "epoch": 0.1891380253249758, "grad_norm": 0.5359539985656738, "learning_rate": 9.483594368303521e-05, "loss": 2.5276, "step": 562 }, { "epoch": 0.1898111143830718, "grad_norm": 0.7273654937744141, "learning_rate": 9.47865775975047e-05, "loss": 2.4901, "step": 564 }, { "epoch": 0.19048420344116782, "grad_norm": 0.642062246799469, "learning_rate": 9.473698964856719e-05, "loss": 2.4732, "step": 566 }, { "epoch": 0.1911572924992638, "grad_norm": 0.6698111891746521, "learning_rate": 9.468718008187106e-05, "loss": 2.5044, "step": 568 }, { "epoch": 0.19183038155735982, "grad_norm": 0.6473941206932068, "learning_rate": 9.463714914416255e-05, "loss": 2.5149, "step": 570 }, { "epoch": 0.1925034706154558, "grad_norm": 0.5659685730934143, "learning_rate": 9.458689708328459e-05, "loss": 2.4881, "step": 572 }, { "epoch": 0.1931765596735518, "grad_norm": 0.736530601978302, "learning_rate": 9.453642414817541e-05, "loss": 2.5245, "step": 574 }, { "epoch": 0.1938496487316478, "grad_norm": 0.6213739514350891, "learning_rate": 9.448573058886747e-05, "loss": 2.5026, "step": 576 }, { "epoch": 0.1945227377897438, "grad_norm": 0.6373304724693298, "learning_rate": 9.443481665648615e-05, "loss": 2.5033, "step": 578 }, { "epoch": 0.19519582684783982, "grad_norm": 0.7117682695388794, "learning_rate": 9.438368260324846e-05, "loss": 2.5133, "step": 580 }, { "epoch": 0.1958689159059358, "grad_norm": 0.566567063331604, "learning_rate": 9.433232868246192e-05, "loss": 2.4923, "step": 582 }, { "epoch": 0.1965420049640318, "grad_norm": 0.6423395872116089, "learning_rate": 9.428075514852319e-05, "loss": 2.4717, "step": 584 }, { "epoch": 0.1972150940221278, "grad_norm": 0.6600716710090637, "learning_rate": 9.422896225691683e-05, "loss": 2.5027, "step": 586 }, { "epoch": 0.1978881830802238, "grad_norm": 0.5523872375488281, "learning_rate": 9.417695026421409e-05, "loss": 2.4834, "step": 588 }, { "epoch": 0.1985612721383198, "grad_norm": 0.6183229088783264, "learning_rate": 9.412471942807157e-05, "loss": 2.5416, "step": 590 }, { "epoch": 0.1992343611964158, "grad_norm": 0.5662925839424133, "learning_rate": 9.407227000723e-05, "loss": 2.4994, "step": 592 }, { "epoch": 0.1999074502545118, "grad_norm": 0.58561110496521, "learning_rate": 9.401960226151291e-05, "loss": 2.4957, "step": 594 }, { "epoch": 0.2005805393126078, "grad_norm": 0.5408748984336853, "learning_rate": 9.396671645182539e-05, "loss": 2.4932, "step": 596 }, { "epoch": 0.2012536283707038, "grad_norm": 0.5413419008255005, "learning_rate": 9.391361284015274e-05, "loss": 2.4623, "step": 598 }, { "epoch": 0.2019267174287998, "grad_norm": 0.5581361651420593, "learning_rate": 9.386029168955925e-05, "loss": 2.4636, "step": 600 } ], "logging_steps": 2, "max_steps": 2972, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 300, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.755100672393216e+17, "train_batch_size": 16, "trial_name": null, "trial_params": null }