| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.9994856848962796, | |
| "eval_steps": 500, | |
| "global_step": 4374, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0006857534716269502, | |
| "grad_norm": 5.500612367122994e+18, | |
| "learning_rate": 2.2831050228310502e-08, | |
| "loss": 1.1976, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0034287673581347507, | |
| "grad_norm": 324.38059468358585, | |
| "learning_rate": 1.1415525114155251e-07, | |
| "loss": 1.175, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.006857534716269501, | |
| "grad_norm": 265.1750103761022, | |
| "learning_rate": 2.2831050228310502e-07, | |
| "loss": 1.1601, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.010286302074404252, | |
| "grad_norm": 88.79558048967972, | |
| "learning_rate": 3.4246575342465755e-07, | |
| "loss": 1.1645, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.013715069432539003, | |
| "grad_norm": 997.1728503325202, | |
| "learning_rate": 4.5662100456621004e-07, | |
| "loss": 1.1547, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01714383679067375, | |
| "grad_norm": 138.0901381156419, | |
| "learning_rate": 5.707762557077626e-07, | |
| "loss": 1.1146, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.020572604148808505, | |
| "grad_norm": 3.6970703101396913, | |
| "learning_rate": 6.849315068493151e-07, | |
| "loss": 1.0849, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.024001371506943255, | |
| "grad_norm": 2.837595372313584, | |
| "learning_rate": 7.990867579908676e-07, | |
| "loss": 1.0569, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.027430138865078006, | |
| "grad_norm": 6.412430536312489, | |
| "learning_rate": 9.132420091324201e-07, | |
| "loss": 1.0229, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.030858906223212756, | |
| "grad_norm": 1.429707929152891, | |
| "learning_rate": 1.0273972602739727e-06, | |
| "loss": 0.9816, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.0342876735813475, | |
| "grad_norm": 1.1893517204804187, | |
| "learning_rate": 1.1415525114155251e-06, | |
| "loss": 0.9501, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03771644093948225, | |
| "grad_norm": 1.316763932645065, | |
| "learning_rate": 1.2557077625570776e-06, | |
| "loss": 0.9212, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.04114520829761701, | |
| "grad_norm": 1.2662857030430075, | |
| "learning_rate": 1.3698630136986302e-06, | |
| "loss": 0.9013, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.04457397565575176, | |
| "grad_norm": 1.1097798063025857, | |
| "learning_rate": 1.4840182648401826e-06, | |
| "loss": 0.8951, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.04800274301388651, | |
| "grad_norm": 1.1990919640253435, | |
| "learning_rate": 1.5981735159817353e-06, | |
| "loss": 0.8941, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.05143151037202126, | |
| "grad_norm": 1.154454828366115, | |
| "learning_rate": 1.7123287671232877e-06, | |
| "loss": 0.8742, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.05486027773015601, | |
| "grad_norm": 1.1710287337749217, | |
| "learning_rate": 1.8264840182648401e-06, | |
| "loss": 0.8719, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.05828904508829076, | |
| "grad_norm": 1.1235385190661975, | |
| "learning_rate": 1.9406392694063926e-06, | |
| "loss": 0.8612, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.06171781244642551, | |
| "grad_norm": 1.3849808686622438, | |
| "learning_rate": 2.0547945205479454e-06, | |
| "loss": 0.8373, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.06514657980456026, | |
| "grad_norm": 1.2705169052299314, | |
| "learning_rate": 2.168949771689498e-06, | |
| "loss": 0.846, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.068575347162695, | |
| "grad_norm": 1.2257177989089036, | |
| "learning_rate": 2.2831050228310503e-06, | |
| "loss": 0.8327, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.07200411452082976, | |
| "grad_norm": 1.3499311190223524, | |
| "learning_rate": 2.3972602739726027e-06, | |
| "loss": 0.8223, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.0754328818789645, | |
| "grad_norm": 1.1551811574767, | |
| "learning_rate": 2.511415525114155e-06, | |
| "loss": 0.8075, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.07886164923709926, | |
| "grad_norm": 1.1152359835046965, | |
| "learning_rate": 2.625570776255708e-06, | |
| "loss": 0.8072, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.08229041659523402, | |
| "grad_norm": 1.128721063171254, | |
| "learning_rate": 2.7397260273972604e-06, | |
| "loss": 0.799, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.08571918395336876, | |
| "grad_norm": 1.0606734214287294, | |
| "learning_rate": 2.8538812785388133e-06, | |
| "loss": 0.8028, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.08914795131150352, | |
| "grad_norm": 1.0207053836629467, | |
| "learning_rate": 2.9680365296803653e-06, | |
| "loss": 0.7793, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.09257671866963826, | |
| "grad_norm": 0.9678836166894507, | |
| "learning_rate": 3.082191780821918e-06, | |
| "loss": 0.7818, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.09600548602777302, | |
| "grad_norm": 0.958558687009426, | |
| "learning_rate": 3.1963470319634706e-06, | |
| "loss": 0.78, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.09943425338590776, | |
| "grad_norm": 0.9103829103840835, | |
| "learning_rate": 3.310502283105023e-06, | |
| "loss": 0.7745, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.10286302074404252, | |
| "grad_norm": 1.072709628680254, | |
| "learning_rate": 3.4246575342465754e-06, | |
| "loss": 0.76, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.10629178810217726, | |
| "grad_norm": 0.87005113272475, | |
| "learning_rate": 3.5388127853881283e-06, | |
| "loss": 0.7583, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.10972055546031202, | |
| "grad_norm": 0.7285413995681538, | |
| "learning_rate": 3.6529680365296803e-06, | |
| "loss": 0.7624, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.11314932281844677, | |
| "grad_norm": 0.8949714023624847, | |
| "learning_rate": 3.767123287671233e-06, | |
| "loss": 0.7475, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.11657809017658152, | |
| "grad_norm": 0.721003961805146, | |
| "learning_rate": 3.881278538812785e-06, | |
| "loss": 0.7564, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.12000685753471627, | |
| "grad_norm": 0.7605946035325595, | |
| "learning_rate": 3.995433789954338e-06, | |
| "loss": 0.7383, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.12343562489285102, | |
| "grad_norm": 1.0401397098538707, | |
| "learning_rate": 4.109589041095891e-06, | |
| "loss": 0.7488, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.12686439225098578, | |
| "grad_norm": 0.9619516801856884, | |
| "learning_rate": 4.223744292237444e-06, | |
| "loss": 0.7361, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.13029315960912052, | |
| "grad_norm": 0.7114806408995153, | |
| "learning_rate": 4.337899543378996e-06, | |
| "loss": 0.7413, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.13372192696725527, | |
| "grad_norm": 0.7788966170765217, | |
| "learning_rate": 4.4520547945205486e-06, | |
| "loss": 0.7288, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.13715069432539, | |
| "grad_norm": 0.7788077365287874, | |
| "learning_rate": 4.566210045662101e-06, | |
| "loss": 0.7241, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.14057946168352478, | |
| "grad_norm": 0.713495426998104, | |
| "learning_rate": 4.6803652968036534e-06, | |
| "loss": 0.7335, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.14400822904165952, | |
| "grad_norm": 0.7672558645856348, | |
| "learning_rate": 4.7945205479452054e-06, | |
| "loss": 0.7262, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.14743699639979427, | |
| "grad_norm": 0.7374473797133105, | |
| "learning_rate": 4.908675799086758e-06, | |
| "loss": 0.721, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.150865763757929, | |
| "grad_norm": 0.7501307226424111, | |
| "learning_rate": 5.02283105022831e-06, | |
| "loss": 0.7186, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.15429453111606378, | |
| "grad_norm": 0.6971360520708836, | |
| "learning_rate": 5.136986301369864e-06, | |
| "loss": 0.7194, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.15772329847419853, | |
| "grad_norm": 0.6367123001071819, | |
| "learning_rate": 5.251141552511416e-06, | |
| "loss": 0.719, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.16115206583233327, | |
| "grad_norm": 1.6312076754734648, | |
| "learning_rate": 5.365296803652969e-06, | |
| "loss": 0.7139, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.16458083319046804, | |
| "grad_norm": 0.8262096198704073, | |
| "learning_rate": 5.479452054794521e-06, | |
| "loss": 0.7089, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.16800960054860278, | |
| "grad_norm": 0.9448491152723366, | |
| "learning_rate": 5.593607305936074e-06, | |
| "loss": 0.7103, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.17143836790673753, | |
| "grad_norm": 0.8493857772402849, | |
| "learning_rate": 5.7077625570776266e-06, | |
| "loss": 0.7011, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.17486713526487227, | |
| "grad_norm": 0.7340850975823853, | |
| "learning_rate": 5.821917808219179e-06, | |
| "loss": 0.704, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.17829590262300704, | |
| "grad_norm": 0.713122055338853, | |
| "learning_rate": 5.936073059360731e-06, | |
| "loss": 0.7025, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.18172466998114178, | |
| "grad_norm": 0.7862099070131026, | |
| "learning_rate": 6.050228310502284e-06, | |
| "loss": 0.7023, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.18515343733927653, | |
| "grad_norm": 0.8003249221177992, | |
| "learning_rate": 6.164383561643836e-06, | |
| "loss": 0.6985, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.18858220469741127, | |
| "grad_norm": 0.72870713699412, | |
| "learning_rate": 6.278538812785388e-06, | |
| "loss": 0.6896, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.19201097205554604, | |
| "grad_norm": 0.68892271391433, | |
| "learning_rate": 6.392694063926941e-06, | |
| "loss": 0.6937, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.19543973941368079, | |
| "grad_norm": 0.8188477395535939, | |
| "learning_rate": 6.506849315068494e-06, | |
| "loss": 0.6927, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.19886850677181553, | |
| "grad_norm": 0.7692820637205647, | |
| "learning_rate": 6.621004566210046e-06, | |
| "loss": 0.7009, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.20229727412995027, | |
| "grad_norm": 0.6627944117356517, | |
| "learning_rate": 6.735159817351599e-06, | |
| "loss": 0.6876, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.20572604148808504, | |
| "grad_norm": 0.7266007369078403, | |
| "learning_rate": 6.849315068493151e-06, | |
| "loss": 0.6853, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.2091548088462198, | |
| "grad_norm": 0.6424680301484417, | |
| "learning_rate": 6.9634703196347046e-06, | |
| "loss": 0.6938, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.21258357620435453, | |
| "grad_norm": 0.7129680875859563, | |
| "learning_rate": 7.077625570776257e-06, | |
| "loss": 0.6909, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.21601234356248927, | |
| "grad_norm": 0.9106015540335944, | |
| "learning_rate": 7.191780821917809e-06, | |
| "loss": 0.6916, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.21944111092062404, | |
| "grad_norm": 0.6760706144438374, | |
| "learning_rate": 7.305936073059361e-06, | |
| "loss": 0.6851, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.2228698782787588, | |
| "grad_norm": 0.6512553623989397, | |
| "learning_rate": 7.420091324200914e-06, | |
| "loss": 0.6897, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.22629864563689353, | |
| "grad_norm": 0.7850669023798967, | |
| "learning_rate": 7.534246575342466e-06, | |
| "loss": 0.6835, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.22972741299502827, | |
| "grad_norm": 1.3429850514334625, | |
| "learning_rate": 7.648401826484018e-06, | |
| "loss": 0.6862, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.23315618035316305, | |
| "grad_norm": 1.0601962354732333, | |
| "learning_rate": 7.76255707762557e-06, | |
| "loss": 0.6865, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.2365849477112978, | |
| "grad_norm": 0.7310633594389766, | |
| "learning_rate": 7.876712328767124e-06, | |
| "loss": 0.6796, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.24001371506943253, | |
| "grad_norm": 1.2791540029610797, | |
| "learning_rate": 7.990867579908676e-06, | |
| "loss": 0.678, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.2434424824275673, | |
| "grad_norm": 1.0514021597488898, | |
| "learning_rate": 8.105022831050228e-06, | |
| "loss": 0.6773, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.24687124978570205, | |
| "grad_norm": 0.6590297903145794, | |
| "learning_rate": 8.219178082191782e-06, | |
| "loss": 0.6787, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.2503000171438368, | |
| "grad_norm": 1.4407668581292676, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 0.6747, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.25372878450197156, | |
| "grad_norm": 0.5783708488581324, | |
| "learning_rate": 8.447488584474887e-06, | |
| "loss": 0.6761, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.2571575518601063, | |
| "grad_norm": 0.7196205508845748, | |
| "learning_rate": 8.56164383561644e-06, | |
| "loss": 0.6741, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.26058631921824105, | |
| "grad_norm": 0.8426560938281862, | |
| "learning_rate": 8.675799086757991e-06, | |
| "loss": 0.6783, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.2640150865763758, | |
| "grad_norm": 0.6248407108995583, | |
| "learning_rate": 8.789954337899545e-06, | |
| "loss": 0.6812, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.26744385393451053, | |
| "grad_norm": 0.625929290973614, | |
| "learning_rate": 8.904109589041097e-06, | |
| "loss": 0.6722, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.2708726212926453, | |
| "grad_norm": 0.8071769890986592, | |
| "learning_rate": 9.01826484018265e-06, | |
| "loss": 0.6777, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.27430138865078, | |
| "grad_norm": 0.9295641422813844, | |
| "learning_rate": 9.132420091324201e-06, | |
| "loss": 0.6744, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.2777301560089148, | |
| "grad_norm": 0.7634843175176039, | |
| "learning_rate": 9.246575342465755e-06, | |
| "loss": 0.6718, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.28115892336704956, | |
| "grad_norm": 0.8056800028036205, | |
| "learning_rate": 9.360730593607307e-06, | |
| "loss": 0.6685, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.2845876907251843, | |
| "grad_norm": 0.7359796110052708, | |
| "learning_rate": 9.474885844748859e-06, | |
| "loss": 0.6686, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.28801645808331905, | |
| "grad_norm": 0.8275083411284923, | |
| "learning_rate": 9.589041095890411e-06, | |
| "loss": 0.6713, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.2914452254414538, | |
| "grad_norm": 0.9767666036041072, | |
| "learning_rate": 9.703196347031965e-06, | |
| "loss": 0.6644, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.29487399279958854, | |
| "grad_norm": 0.6970783265751548, | |
| "learning_rate": 9.817351598173517e-06, | |
| "loss": 0.6692, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.2983027601577233, | |
| "grad_norm": 0.584165558382207, | |
| "learning_rate": 9.931506849315069e-06, | |
| "loss": 0.6659, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.301731527515858, | |
| "grad_norm": 0.6245810626396433, | |
| "learning_rate": 9.999993629265979e-06, | |
| "loss": 0.6647, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.3051602948739928, | |
| "grad_norm": 0.7335612100903357, | |
| "learning_rate": 9.999921958694681e-06, | |
| "loss": 0.667, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.30858906223212756, | |
| "grad_norm": 0.6843530498214124, | |
| "learning_rate": 9.999770655279843e-06, | |
| "loss": 0.6683, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.3120178295902623, | |
| "grad_norm": 0.6124751321352965, | |
| "learning_rate": 9.999539721431253e-06, | |
| "loss": 0.6619, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.31544659694839705, | |
| "grad_norm": 0.8614237317784202, | |
| "learning_rate": 9.999229160826947e-06, | |
| "loss": 0.6637, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.3188753643065318, | |
| "grad_norm": 0.6836210220640591, | |
| "learning_rate": 9.998838978413167e-06, | |
| "loss": 0.664, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.32230413166466654, | |
| "grad_norm": 0.7142979223973136, | |
| "learning_rate": 9.998369180404283e-06, | |
| "loss": 0.6682, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.3257328990228013, | |
| "grad_norm": 0.6970601000506839, | |
| "learning_rate": 9.997819774282676e-06, | |
| "loss": 0.6654, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.3291616663809361, | |
| "grad_norm": 0.7855321640617767, | |
| "learning_rate": 9.997190768798639e-06, | |
| "loss": 0.6628, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.3325904337390708, | |
| "grad_norm": 0.8068114924796952, | |
| "learning_rate": 9.996482173970227e-06, | |
| "loss": 0.6532, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.33601920109720557, | |
| "grad_norm": 0.6190829521241801, | |
| "learning_rate": 9.995694001083103e-06, | |
| "loss": 0.6625, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.3394479684553403, | |
| "grad_norm": 0.6559745100197821, | |
| "learning_rate": 9.994826262690347e-06, | |
| "loss": 0.6601, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.34287673581347505, | |
| "grad_norm": 0.6960576238421567, | |
| "learning_rate": 9.993878972612276e-06, | |
| "loss": 0.6575, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.3463055031716098, | |
| "grad_norm": 0.8092838621716005, | |
| "learning_rate": 9.992852145936202e-06, | |
| "loss": 0.6562, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.34973427052974454, | |
| "grad_norm": 0.7436128976237549, | |
| "learning_rate": 9.991745799016206e-06, | |
| "loss": 0.6514, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.3531630378878793, | |
| "grad_norm": 0.754160742003698, | |
| "learning_rate": 9.990559949472875e-06, | |
| "loss": 0.6579, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.3565918052460141, | |
| "grad_norm": 0.6596249438947815, | |
| "learning_rate": 9.989294616193018e-06, | |
| "loss": 0.6619, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.3600205726041488, | |
| "grad_norm": 0.7268392727602604, | |
| "learning_rate": 9.987949819329366e-06, | |
| "loss": 0.6569, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.36344933996228357, | |
| "grad_norm": 0.7908918422796353, | |
| "learning_rate": 9.986525580300253e-06, | |
| "loss": 0.6602, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.3668781073204183, | |
| "grad_norm": 0.731825676141184, | |
| "learning_rate": 9.985021921789274e-06, | |
| "loss": 0.659, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.37030687467855306, | |
| "grad_norm": 0.6826536905758885, | |
| "learning_rate": 9.983438867744923e-06, | |
| "loss": 0.6501, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.3737356420366878, | |
| "grad_norm": 0.6873425512137389, | |
| "learning_rate": 9.981776443380214e-06, | |
| "loss": 0.653, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.37716440939482254, | |
| "grad_norm": 0.6087207671115116, | |
| "learning_rate": 9.980034675172274e-06, | |
| "loss": 0.6501, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.3805931767529573, | |
| "grad_norm": 0.6758600678083865, | |
| "learning_rate": 9.97821359086193e-06, | |
| "loss": 0.6457, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.3840219441110921, | |
| "grad_norm": 0.6996994398315339, | |
| "learning_rate": 9.976313219453255e-06, | |
| "loss": 0.6516, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3874507114692268, | |
| "grad_norm": 0.5956097888192097, | |
| "learning_rate": 9.97433359121312e-06, | |
| "loss": 0.6561, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.39087947882736157, | |
| "grad_norm": 0.5809586858657639, | |
| "learning_rate": 9.972274737670702e-06, | |
| "loss": 0.6515, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.3943082461854963, | |
| "grad_norm": 0.6234744336805654, | |
| "learning_rate": 9.970136691616985e-06, | |
| "loss": 0.6567, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.39773701354363106, | |
| "grad_norm": 0.7019194825681323, | |
| "learning_rate": 9.967919487104237e-06, | |
| "loss": 0.6407, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.4011657809017658, | |
| "grad_norm": 0.6464500473501631, | |
| "learning_rate": 9.965623159445471e-06, | |
| "loss": 0.6478, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.40459454825990054, | |
| "grad_norm": 0.6397619177468029, | |
| "learning_rate": 9.963247745213876e-06, | |
| "loss": 0.6438, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.40802331561803534, | |
| "grad_norm": 0.6910890699251881, | |
| "learning_rate": 9.960793282242243e-06, | |
| "loss": 0.649, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.4114520829761701, | |
| "grad_norm": 0.7163950332830918, | |
| "learning_rate": 9.958259809622353e-06, | |
| "loss": 0.6455, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.41488085033430483, | |
| "grad_norm": 0.8206374824286882, | |
| "learning_rate": 9.955647367704362e-06, | |
| "loss": 0.6569, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.4183096176924396, | |
| "grad_norm": 0.7207309045108575, | |
| "learning_rate": 9.952955998096155e-06, | |
| "loss": 0.6446, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.4217383850505743, | |
| "grad_norm": 0.5651258904128029, | |
| "learning_rate": 9.950185743662685e-06, | |
| "loss": 0.6444, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.42516715240870906, | |
| "grad_norm": 0.5443692863606336, | |
| "learning_rate": 9.94733664852529e-06, | |
| "loss": 0.6418, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.4285959197668438, | |
| "grad_norm": 0.6042279487491544, | |
| "learning_rate": 9.944408758060982e-06, | |
| "loss": 0.6463, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.43202468712497855, | |
| "grad_norm": 0.5656088824328167, | |
| "learning_rate": 9.941402118901743e-06, | |
| "loss": 0.6407, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.43545345448311334, | |
| "grad_norm": 0.658014001778976, | |
| "learning_rate": 9.938316778933763e-06, | |
| "loss": 0.6468, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.4388822218412481, | |
| "grad_norm": 0.6639637664489146, | |
| "learning_rate": 9.935152787296689e-06, | |
| "loss": 0.6494, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.44231098919938283, | |
| "grad_norm": 0.5783566760560125, | |
| "learning_rate": 9.931910194382836e-06, | |
| "loss": 0.6463, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.4457397565575176, | |
| "grad_norm": 0.6490847428513987, | |
| "learning_rate": 9.928589051836392e-06, | |
| "loss": 0.6438, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.4491685239156523, | |
| "grad_norm": 0.5672865850934122, | |
| "learning_rate": 9.925189412552585e-06, | |
| "loss": 0.6333, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.45259729127378706, | |
| "grad_norm": 0.569087352718797, | |
| "learning_rate": 9.921711330676848e-06, | |
| "loss": 0.6478, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.4560260586319218, | |
| "grad_norm": 0.6207854393319554, | |
| "learning_rate": 9.918154861603959e-06, | |
| "loss": 0.6431, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.45945482599005655, | |
| "grad_norm": 0.586346598273814, | |
| "learning_rate": 9.91452006197715e-06, | |
| "loss": 0.6383, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.46288359334819135, | |
| "grad_norm": 0.5045583675197564, | |
| "learning_rate": 9.910806989687207e-06, | |
| "loss": 0.6365, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.4663123607063261, | |
| "grad_norm": 0.6674046560237406, | |
| "learning_rate": 9.907015703871558e-06, | |
| "loss": 0.6446, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.46974112806446083, | |
| "grad_norm": 0.6966429050410609, | |
| "learning_rate": 9.903146264913318e-06, | |
| "loss": 0.6416, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.4731698954225956, | |
| "grad_norm": 0.6302591461200223, | |
| "learning_rate": 9.899198734440335e-06, | |
| "loss": 0.6371, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.4765986627807303, | |
| "grad_norm": 0.8144573920281895, | |
| "learning_rate": 9.895173175324205e-06, | |
| "loss": 0.6342, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.48002743013886506, | |
| "grad_norm": 0.7886436043083611, | |
| "learning_rate": 9.891069651679273e-06, | |
| "loss": 0.6355, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.4834561974969998, | |
| "grad_norm": 0.8615238941842659, | |
| "learning_rate": 9.886888228861608e-06, | |
| "loss": 0.6364, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.4868849648551346, | |
| "grad_norm": 0.6441681216184715, | |
| "learning_rate": 9.882628973467972e-06, | |
| "loss": 0.6365, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.49031373221326935, | |
| "grad_norm": 0.539258830386699, | |
| "learning_rate": 9.878291953334744e-06, | |
| "loss": 0.6409, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.4937424995714041, | |
| "grad_norm": 0.5498193436814545, | |
| "learning_rate": 9.873877237536854e-06, | |
| "loss": 0.6394, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.49717126692953884, | |
| "grad_norm": 0.574880600021989, | |
| "learning_rate": 9.869384896386669e-06, | |
| "loss": 0.6413, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.5006000342876736, | |
| "grad_norm": 0.618195186077868, | |
| "learning_rate": 9.86481500143289e-06, | |
| "loss": 0.6416, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.5040288016458083, | |
| "grad_norm": 0.5418237252054477, | |
| "learning_rate": 9.860167625459398e-06, | |
| "loss": 0.6372, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.5074575690039431, | |
| "grad_norm": 0.5863545275273223, | |
| "learning_rate": 9.855442842484101e-06, | |
| "loss": 0.6345, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.5108863363620778, | |
| "grad_norm": 0.553229939288754, | |
| "learning_rate": 9.850640727757753e-06, | |
| "loss": 0.6347, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.5143151037202126, | |
| "grad_norm": 0.6152186401867638, | |
| "learning_rate": 9.84576135776276e-06, | |
| "loss": 0.6355, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.5177438710783473, | |
| "grad_norm": 0.6357690608918045, | |
| "learning_rate": 9.840804810211954e-06, | |
| "loss": 0.6347, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.5211726384364821, | |
| "grad_norm": 0.5531025379600504, | |
| "learning_rate": 9.835771164047365e-06, | |
| "loss": 0.6367, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.5246014057946168, | |
| "grad_norm": 0.5598603097452307, | |
| "learning_rate": 9.830660499438955e-06, | |
| "loss": 0.6322, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.5280301731527516, | |
| "grad_norm": 0.5973164400797644, | |
| "learning_rate": 9.825472897783344e-06, | |
| "loss": 0.6363, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.5314589405108864, | |
| "grad_norm": 0.6319771585799853, | |
| "learning_rate": 9.820208441702516e-06, | |
| "loss": 0.632, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.5348877078690211, | |
| "grad_norm": 0.6648742107102181, | |
| "learning_rate": 9.814867215042503e-06, | |
| "loss": 0.6379, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.5383164752271559, | |
| "grad_norm": 0.7342225799559353, | |
| "learning_rate": 9.809449302872045e-06, | |
| "loss": 0.6405, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.5417452425852906, | |
| "grad_norm": 0.7688881238342142, | |
| "learning_rate": 9.803954791481239e-06, | |
| "loss": 0.634, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.5451740099434254, | |
| "grad_norm": 0.639917210365233, | |
| "learning_rate": 9.798383768380164e-06, | |
| "loss": 0.6348, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.54860277730156, | |
| "grad_norm": 0.5045187940762229, | |
| "learning_rate": 9.792736322297489e-06, | |
| "loss": 0.6362, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.5520315446596948, | |
| "grad_norm": 0.5381061183966045, | |
| "learning_rate": 9.787012543179053e-06, | |
| "loss": 0.6391, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.5554603120178296, | |
| "grad_norm": 0.5429280857380482, | |
| "learning_rate": 9.781212522186442e-06, | |
| "loss": 0.6307, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.5588890793759643, | |
| "grad_norm": 0.594664535184356, | |
| "learning_rate": 9.77533635169553e-06, | |
| "loss": 0.6298, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.5623178467340991, | |
| "grad_norm": 0.6702565938406244, | |
| "learning_rate": 9.769384125295012e-06, | |
| "loss": 0.6342, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.5657466140922338, | |
| "grad_norm": 0.7024332099849335, | |
| "learning_rate": 9.763355937784908e-06, | |
| "loss": 0.6261, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.5691753814503686, | |
| "grad_norm": 0.6596631748942144, | |
| "learning_rate": 9.757251885175063e-06, | |
| "loss": 0.6377, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.5726041488085033, | |
| "grad_norm": 0.580750320713107, | |
| "learning_rate": 9.751072064683604e-06, | |
| "loss": 0.638, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.5760329161666381, | |
| "grad_norm": 0.5865851818669229, | |
| "learning_rate": 9.744816574735405e-06, | |
| "loss": 0.6257, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.5794616835247729, | |
| "grad_norm": 0.6490256085654721, | |
| "learning_rate": 9.73848551496051e-06, | |
| "loss": 0.6268, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.5828904508829076, | |
| "grad_norm": 0.5699242035532047, | |
| "learning_rate": 9.732078986192552e-06, | |
| "loss": 0.6343, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.5863192182410424, | |
| "grad_norm": 0.5773888152890198, | |
| "learning_rate": 9.725597090467145e-06, | |
| "loss": 0.6335, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.5897479855991771, | |
| "grad_norm": 0.5030181039807039, | |
| "learning_rate": 9.719039931020258e-06, | |
| "loss": 0.6297, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.5931767529573119, | |
| "grad_norm": 0.6927945011578892, | |
| "learning_rate": 9.712407612286568e-06, | |
| "loss": 0.6317, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.5966055203154466, | |
| "grad_norm": 0.6863261538586617, | |
| "learning_rate": 9.705700239897809e-06, | |
| "loss": 0.6349, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.6000342876735814, | |
| "grad_norm": 0.5863644745576692, | |
| "learning_rate": 9.698917920681072e-06, | |
| "loss": 0.6336, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.603463055031716, | |
| "grad_norm": 0.569507321040072, | |
| "learning_rate": 9.692060762657118e-06, | |
| "loss": 0.6311, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.6068918223898508, | |
| "grad_norm": 0.5320598410866568, | |
| "learning_rate": 9.685128875038648e-06, | |
| "loss": 0.628, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.6103205897479856, | |
| "grad_norm": 0.6614795760343849, | |
| "learning_rate": 9.678122368228571e-06, | |
| "loss": 0.6273, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.6137493571061203, | |
| "grad_norm": 0.5899705788099368, | |
| "learning_rate": 9.67104135381824e-06, | |
| "loss": 0.626, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.6171781244642551, | |
| "grad_norm": 0.6244814955846902, | |
| "learning_rate": 9.66388594458568e-06, | |
| "loss": 0.6202, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.6206068918223898, | |
| "grad_norm": 0.5468113327023256, | |
| "learning_rate": 9.656656254493783e-06, | |
| "loss": 0.6247, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.6240356591805246, | |
| "grad_norm": 0.5197583062967639, | |
| "learning_rate": 9.649352398688506e-06, | |
| "loss": 0.6341, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.6274644265386593, | |
| "grad_norm": 0.578711261845795, | |
| "learning_rate": 9.641974493497024e-06, | |
| "loss": 0.6274, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.6308931938967941, | |
| "grad_norm": 0.536259271020302, | |
| "learning_rate": 9.634522656425885e-06, | |
| "loss": 0.638, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.6343219612549289, | |
| "grad_norm": 0.5108361690063496, | |
| "learning_rate": 9.626997006159135e-06, | |
| "loss": 0.6268, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.6377507286130636, | |
| "grad_norm": 0.5258629436603048, | |
| "learning_rate": 9.619397662556434e-06, | |
| "loss": 0.6306, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.6411794959711984, | |
| "grad_norm": 0.5457258768955178, | |
| "learning_rate": 9.61172474665114e-06, | |
| "loss": 0.6286, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.6446082633293331, | |
| "grad_norm": 0.6608031766605356, | |
| "learning_rate": 9.603978380648375e-06, | |
| "loss": 0.6295, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.6480370306874679, | |
| "grad_norm": 0.5276661545510922, | |
| "learning_rate": 9.596158687923105e-06, | |
| "loss": 0.6263, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.6514657980456026, | |
| "grad_norm": 0.530790337719698, | |
| "learning_rate": 9.588265793018141e-06, | |
| "loss": 0.625, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.6548945654037374, | |
| "grad_norm": 0.562840339494226, | |
| "learning_rate": 9.58029982164218e-06, | |
| "loss": 0.6286, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.6583233327618722, | |
| "grad_norm": 0.4947380099055408, | |
| "learning_rate": 9.572260900667794e-06, | |
| "loss": 0.6236, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.6617521001200068, | |
| "grad_norm": 0.47265335918997425, | |
| "learning_rate": 9.564149158129406e-06, | |
| "loss": 0.6296, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.6651808674781416, | |
| "grad_norm": 0.5775895832290389, | |
| "learning_rate": 9.555964723221258e-06, | |
| "loss": 0.6235, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.6686096348362763, | |
| "grad_norm": 0.5561593222147068, | |
| "learning_rate": 9.54770772629535e-06, | |
| "loss": 0.6203, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.6720384021944111, | |
| "grad_norm": 0.5564531018664398, | |
| "learning_rate": 9.539378298859365e-06, | |
| "loss": 0.6131, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.6754671695525458, | |
| "grad_norm": 0.6036545672788025, | |
| "learning_rate": 9.53097657357457e-06, | |
| "loss": 0.6166, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.6788959369106806, | |
| "grad_norm": 0.5409723710972858, | |
| "learning_rate": 9.522502684253709e-06, | |
| "loss": 0.6229, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.6823247042688153, | |
| "grad_norm": 0.509669382435153, | |
| "learning_rate": 9.51395676585887e-06, | |
| "loss": 0.6222, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.6857534716269501, | |
| "grad_norm": 0.6082473746180043, | |
| "learning_rate": 9.505338954499332e-06, | |
| "loss": 0.6148, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.6891822389850849, | |
| "grad_norm": 0.5381865337988982, | |
| "learning_rate": 9.496649387429405e-06, | |
| "loss": 0.6253, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.6926110063432196, | |
| "grad_norm": 0.5519342026599645, | |
| "learning_rate": 9.487888203046232e-06, | |
| "loss": 0.6272, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.6960397737013544, | |
| "grad_norm": 0.5016332652349211, | |
| "learning_rate": 9.479055540887599e-06, | |
| "loss": 0.6228, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.6994685410594891, | |
| "grad_norm": 0.6079273353999184, | |
| "learning_rate": 9.4701515416297e-06, | |
| "loss": 0.6208, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.7028973084176239, | |
| "grad_norm": 0.700304976409857, | |
| "learning_rate": 9.461176347084909e-06, | |
| "loss": 0.6214, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.7063260757757586, | |
| "grad_norm": 0.5534413781804028, | |
| "learning_rate": 9.452130100199504e-06, | |
| "loss": 0.6191, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.7097548431338934, | |
| "grad_norm": 0.5985964197962274, | |
| "learning_rate": 9.44301294505141e-06, | |
| "loss": 0.6232, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.7131836104920282, | |
| "grad_norm": 0.5866063492223905, | |
| "learning_rate": 9.433825026847891e-06, | |
| "loss": 0.6233, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.7166123778501629, | |
| "grad_norm": 0.6647059350986494, | |
| "learning_rate": 9.42456649192324e-06, | |
| "loss": 0.6197, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.7200411452082977, | |
| "grad_norm": 0.6902179520870335, | |
| "learning_rate": 9.415237487736452e-06, | |
| "loss": 0.6162, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.7234699125664323, | |
| "grad_norm": 0.5946749998378368, | |
| "learning_rate": 9.405838162868874e-06, | |
| "loss": 0.6169, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.7268986799245671, | |
| "grad_norm": 0.5785836386179494, | |
| "learning_rate": 9.396368667021835e-06, | |
| "loss": 0.6204, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.7303274472827018, | |
| "grad_norm": 0.5715009908016999, | |
| "learning_rate": 9.386829151014262e-06, | |
| "loss": 0.6161, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.7337562146408366, | |
| "grad_norm": 0.5493040557418012, | |
| "learning_rate": 9.377219766780288e-06, | |
| "loss": 0.6177, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.7371849819989714, | |
| "grad_norm": 0.5614424899020751, | |
| "learning_rate": 9.367540667366816e-06, | |
| "loss": 0.6213, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.7406137493571061, | |
| "grad_norm": 0.5955493656424166, | |
| "learning_rate": 9.3577920069311e-06, | |
| "loss": 0.6115, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.7440425167152409, | |
| "grad_norm": 0.6005939054835772, | |
| "learning_rate": 9.347973940738266e-06, | |
| "loss": 0.6146, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.7474712840733756, | |
| "grad_norm": 0.5179120177804368, | |
| "learning_rate": 9.338086625158867e-06, | |
| "loss": 0.6095, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.7509000514315104, | |
| "grad_norm": 0.5736625803401397, | |
| "learning_rate": 9.328130217666366e-06, | |
| "loss": 0.6184, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.7543288187896451, | |
| "grad_norm": 0.48238347864003633, | |
| "learning_rate": 9.318104876834652e-06, | |
| "loss": 0.6147, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.7577575861477799, | |
| "grad_norm": 0.5972927862243746, | |
| "learning_rate": 9.308010762335492e-06, | |
| "loss": 0.6239, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.7611863535059146, | |
| "grad_norm": 0.7711586637339868, | |
| "learning_rate": 9.297848034936007e-06, | |
| "loss": 0.622, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.7646151208640494, | |
| "grad_norm": 0.617930863519868, | |
| "learning_rate": 9.287616856496097e-06, | |
| "loss": 0.6049, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.7680438882221842, | |
| "grad_norm": 0.562401219116577, | |
| "learning_rate": 9.277317389965871e-06, | |
| "loss": 0.6174, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.7714726555803189, | |
| "grad_norm": 0.6288444408114011, | |
| "learning_rate": 9.266949799383053e-06, | |
| "loss": 0.6132, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.7749014229384537, | |
| "grad_norm": 0.5138761921522274, | |
| "learning_rate": 9.256514249870366e-06, | |
| "loss": 0.6144, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.7783301902965883, | |
| "grad_norm": 0.5557394107732245, | |
| "learning_rate": 9.246010907632894e-06, | |
| "loss": 0.625, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.7817589576547231, | |
| "grad_norm": 0.4785123138439094, | |
| "learning_rate": 9.235439939955458e-06, | |
| "loss": 0.6129, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.7851877250128578, | |
| "grad_norm": 0.4845330276832426, | |
| "learning_rate": 9.22480151519992e-06, | |
| "loss": 0.6249, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.7886164923709926, | |
| "grad_norm": 0.5198630830752909, | |
| "learning_rate": 9.214095802802533e-06, | |
| "loss": 0.6181, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.7920452597291274, | |
| "grad_norm": 0.5032353282124736, | |
| "learning_rate": 9.203322973271223e-06, | |
| "loss": 0.6208, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.7954740270872621, | |
| "grad_norm": 0.710468651962232, | |
| "learning_rate": 9.192483198182876e-06, | |
| "loss": 0.6148, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.7989027944453969, | |
| "grad_norm": 0.5960153692293567, | |
| "learning_rate": 9.181576650180606e-06, | |
| "loss": 0.6216, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.8023315618035316, | |
| "grad_norm": 0.5541889269998591, | |
| "learning_rate": 9.170603502971017e-06, | |
| "loss": 0.6181, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.8057603291616664, | |
| "grad_norm": 0.4842737593004163, | |
| "learning_rate": 9.159563931321416e-06, | |
| "loss": 0.6118, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.8091890965198011, | |
| "grad_norm": 0.5527421600120501, | |
| "learning_rate": 9.148458111057043e-06, | |
| "loss": 0.617, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.8126178638779359, | |
| "grad_norm": 0.517599668993277, | |
| "learning_rate": 9.13728621905827e-06, | |
| "loss": 0.6081, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.8160466312360707, | |
| "grad_norm": 0.5442628844464218, | |
| "learning_rate": 9.12604843325778e-06, | |
| "loss": 0.6211, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.8194753985942054, | |
| "grad_norm": 0.5027283276915621, | |
| "learning_rate": 9.114744932637734e-06, | |
| "loss": 0.6127, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.8229041659523402, | |
| "grad_norm": 0.5296307123252137, | |
| "learning_rate": 9.103375897226919e-06, | |
| "loss": 0.613, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.8263329333104749, | |
| "grad_norm": 0.5490731527916699, | |
| "learning_rate": 9.091941508097886e-06, | |
| "loss": 0.6141, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.8297617006686097, | |
| "grad_norm": 0.48494095011709015, | |
| "learning_rate": 9.080441947364065e-06, | |
| "loss": 0.6137, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.8331904680267443, | |
| "grad_norm": 0.5200558540203623, | |
| "learning_rate": 9.068877398176854e-06, | |
| "loss": 0.6086, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.8366192353848791, | |
| "grad_norm": 0.5307240302552572, | |
| "learning_rate": 9.057248044722718e-06, | |
| "loss": 0.6186, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.8400480027430138, | |
| "grad_norm": 0.6928206617626188, | |
| "learning_rate": 9.045554072220241e-06, | |
| "loss": 0.616, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.8434767701011486, | |
| "grad_norm": 0.6694845253440784, | |
| "learning_rate": 9.033795666917191e-06, | |
| "loss": 0.6105, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.8469055374592834, | |
| "grad_norm": 0.48405743005066365, | |
| "learning_rate": 9.02197301608754e-06, | |
| "loss": 0.606, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.8503343048174181, | |
| "grad_norm": 0.5821371319328923, | |
| "learning_rate": 9.010086308028487e-06, | |
| "loss": 0.6129, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.8537630721755529, | |
| "grad_norm": 0.6308021249120006, | |
| "learning_rate": 8.998135732057458e-06, | |
| "loss": 0.6101, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.8571918395336876, | |
| "grad_norm": 0.6393999209710391, | |
| "learning_rate": 8.986121478509096e-06, | |
| "loss": 0.6099, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.8606206068918224, | |
| "grad_norm": 0.5210274848669988, | |
| "learning_rate": 8.97404373873222e-06, | |
| "loss": 0.6136, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.8640493742499571, | |
| "grad_norm": 0.5493385621402502, | |
| "learning_rate": 8.961902705086785e-06, | |
| "loss": 0.6114, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.8674781416080919, | |
| "grad_norm": 0.565383947144019, | |
| "learning_rate": 8.949698570940816e-06, | |
| "loss": 0.6117, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.8709069089662267, | |
| "grad_norm": 0.4809292251179887, | |
| "learning_rate": 8.937431530667329e-06, | |
| "loss": 0.6163, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.8743356763243614, | |
| "grad_norm": 0.6569955954243558, | |
| "learning_rate": 8.925101779641232e-06, | |
| "loss": 0.6136, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.8777644436824962, | |
| "grad_norm": 0.7586122995884393, | |
| "learning_rate": 8.912709514236218e-06, | |
| "loss": 0.6152, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.8811932110406309, | |
| "grad_norm": 0.5863591455341749, | |
| "learning_rate": 8.900254931821632e-06, | |
| "loss": 0.6105, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.8846219783987657, | |
| "grad_norm": 0.5808595100846887, | |
| "learning_rate": 8.887738230759334e-06, | |
| "loss": 0.6103, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.8880507457569004, | |
| "grad_norm": 0.6637416525469315, | |
| "learning_rate": 8.875159610400535e-06, | |
| "loss": 0.6151, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.8914795131150351, | |
| "grad_norm": 0.635843590965775, | |
| "learning_rate": 8.862519271082624e-06, | |
| "loss": 0.6122, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.89490828047317, | |
| "grad_norm": 0.48644669408598906, | |
| "learning_rate": 8.849817414125973e-06, | |
| "loss": 0.6107, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.8983370478313046, | |
| "grad_norm": 0.7387834738899194, | |
| "learning_rate": 8.83705424183074e-06, | |
| "loss": 0.6144, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.9017658151894394, | |
| "grad_norm": 0.47318607768496285, | |
| "learning_rate": 8.824229957473638e-06, | |
| "loss": 0.615, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.9051945825475741, | |
| "grad_norm": 0.6099273713462994, | |
| "learning_rate": 8.811344765304698e-06, | |
| "loss": 0.6126, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.9086233499057089, | |
| "grad_norm": 0.5213938292230176, | |
| "learning_rate": 8.798398870544027e-06, | |
| "loss": 0.6034, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.9120521172638436, | |
| "grad_norm": 0.5394625940524758, | |
| "learning_rate": 8.785392479378522e-06, | |
| "loss": 0.6039, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.9154808846219784, | |
| "grad_norm": 0.48958440285150634, | |
| "learning_rate": 8.772325798958597e-06, | |
| "loss": 0.6029, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.9189096519801131, | |
| "grad_norm": 0.49051315574848703, | |
| "learning_rate": 8.759199037394888e-06, | |
| "loss": 0.6007, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.9223384193382479, | |
| "grad_norm": 0.5004329030521917, | |
| "learning_rate": 8.746012403754924e-06, | |
| "loss": 0.6079, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.9257671866963827, | |
| "grad_norm": 0.5018875978088632, | |
| "learning_rate": 8.732766108059814e-06, | |
| "loss": 0.614, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.9291959540545174, | |
| "grad_norm": 0.5162939197438497, | |
| "learning_rate": 8.719460361280888e-06, | |
| "loss": 0.6161, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.9326247214126522, | |
| "grad_norm": 0.4971369395606836, | |
| "learning_rate": 8.70609537533634e-06, | |
| "loss": 0.605, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.9360534887707869, | |
| "grad_norm": 0.5016579726115805, | |
| "learning_rate": 8.692671363087864e-06, | |
| "loss": 0.6173, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.9394822561289217, | |
| "grad_norm": 0.5580053234884178, | |
| "learning_rate": 8.679188538337248e-06, | |
| "loss": 0.6087, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.9429110234870564, | |
| "grad_norm": 0.4577425988677661, | |
| "learning_rate": 8.665647115822981e-06, | |
| "loss": 0.5988, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.9463397908451912, | |
| "grad_norm": 0.45683981074908053, | |
| "learning_rate": 8.652047311216823e-06, | |
| "loss": 0.6064, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.949768558203326, | |
| "grad_norm": 0.5288094567173458, | |
| "learning_rate": 8.638389341120378e-06, | |
| "loss": 0.6089, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.9531973255614606, | |
| "grad_norm": 0.5037863596617661, | |
| "learning_rate": 8.62467342306164e-06, | |
| "loss": 0.6085, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.9566260929195954, | |
| "grad_norm": 0.4706120011800052, | |
| "learning_rate": 8.610899775491531e-06, | |
| "loss": 0.6023, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.9600548602777301, | |
| "grad_norm": 0.5245905876570266, | |
| "learning_rate": 8.597068617780419e-06, | |
| "loss": 0.6152, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.9634836276358649, | |
| "grad_norm": 0.5294292049294829, | |
| "learning_rate": 8.583180170214625e-06, | |
| "loss": 0.6087, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.9669123949939996, | |
| "grad_norm": 0.5012747145599434, | |
| "learning_rate": 8.569234653992916e-06, | |
| "loss": 0.6074, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.9703411623521344, | |
| "grad_norm": 0.5519787441985884, | |
| "learning_rate": 8.555232291222982e-06, | |
| "loss": 0.608, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.9737699297102692, | |
| "grad_norm": 0.566179243123642, | |
| "learning_rate": 8.541173304917895e-06, | |
| "loss": 0.6057, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.9771986970684039, | |
| "grad_norm": 0.4833724946956298, | |
| "learning_rate": 8.527057918992564e-06, | |
| "loss": 0.6048, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.9806274644265387, | |
| "grad_norm": 0.5444527229286149, | |
| "learning_rate": 8.512886358260162e-06, | |
| "loss": 0.6074, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.9840562317846734, | |
| "grad_norm": 0.6265580889574098, | |
| "learning_rate": 8.498658848428541e-06, | |
| "loss": 0.6104, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.9874849991428082, | |
| "grad_norm": 0.511037692990781, | |
| "learning_rate": 8.484375616096658e-06, | |
| "loss": 0.6111, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.9909137665009429, | |
| "grad_norm": 0.618617261146988, | |
| "learning_rate": 8.470036888750941e-06, | |
| "loss": 0.6059, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.9943425338590777, | |
| "grad_norm": 0.6912314169042373, | |
| "learning_rate": 8.455642894761684e-06, | |
| "loss": 0.6109, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.9977713012172124, | |
| "grad_norm": 0.6536131654185897, | |
| "learning_rate": 8.441193863379396e-06, | |
| "loss": 0.6117, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.9998285616320932, | |
| "eval_loss": 0.6076797842979431, | |
| "eval_runtime": 204.4178, | |
| "eval_samples_per_second": 48.073, | |
| "eval_steps_per_second": 1.507, | |
| "step": 1458 | |
| }, | |
| { | |
| "epoch": 1.0012000685753473, | |
| "grad_norm": 0.5744253109468758, | |
| "learning_rate": 8.426690024731161e-06, | |
| "loss": 0.5965, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.0046288359334818, | |
| "grad_norm": 0.6126234725401744, | |
| "learning_rate": 8.412131609816968e-06, | |
| "loss": 0.5426, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 1.0080576032916166, | |
| "grad_norm": 0.5502946067152649, | |
| "learning_rate": 8.39751885050603e-06, | |
| "loss": 0.5483, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.0114863706497514, | |
| "grad_norm": 0.5203632756557842, | |
| "learning_rate": 8.382851979533095e-06, | |
| "loss": 0.5429, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 1.0149151380078862, | |
| "grad_norm": 0.5256969786092444, | |
| "learning_rate": 8.36813123049474e-06, | |
| "loss": 0.5416, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.0183439053660208, | |
| "grad_norm": 0.5199622468399526, | |
| "learning_rate": 8.353356837845642e-06, | |
| "loss": 0.5497, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 1.0217726727241556, | |
| "grad_norm": 0.5409282245995303, | |
| "learning_rate": 8.338529036894855e-06, | |
| "loss": 0.549, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.0252014400822904, | |
| "grad_norm": 0.4884152712675349, | |
| "learning_rate": 8.323648063802056e-06, | |
| "loss": 0.5444, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 1.0286302074404252, | |
| "grad_norm": 0.5014268314366117, | |
| "learning_rate": 8.308714155573785e-06, | |
| "loss": 0.5495, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.03205897479856, | |
| "grad_norm": 0.4846536766697211, | |
| "learning_rate": 8.293727550059668e-06, | |
| "loss": 0.5457, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 1.0354877421566946, | |
| "grad_norm": 0.5390111306520632, | |
| "learning_rate": 8.278688485948634e-06, | |
| "loss": 0.5479, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.0389165095148294, | |
| "grad_norm": 0.5668492664534543, | |
| "learning_rate": 8.263597202765109e-06, | |
| "loss": 0.5412, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 1.0423452768729642, | |
| "grad_norm": 0.49509927353990013, | |
| "learning_rate": 8.248453940865204e-06, | |
| "loss": 0.5509, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.045774044231099, | |
| "grad_norm": 0.5658638682509576, | |
| "learning_rate": 8.233258941432883e-06, | |
| "loss": 0.5466, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 1.0492028115892336, | |
| "grad_norm": 0.599624817317566, | |
| "learning_rate": 8.218012446476128e-06, | |
| "loss": 0.5456, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.0526315789473684, | |
| "grad_norm": 0.5811724248156606, | |
| "learning_rate": 8.202714698823078e-06, | |
| "loss": 0.5478, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 1.0560603463055032, | |
| "grad_norm": 0.5829817578564964, | |
| "learning_rate": 8.187365942118162e-06, | |
| "loss": 0.5447, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.059489113663638, | |
| "grad_norm": 0.6195689312853232, | |
| "learning_rate": 8.171966420818227e-06, | |
| "loss": 0.5462, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 1.0629178810217728, | |
| "grad_norm": 0.5753478735487544, | |
| "learning_rate": 8.156516380188635e-06, | |
| "loss": 0.5461, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.0663466483799073, | |
| "grad_norm": 0.7022154988882346, | |
| "learning_rate": 8.14101606629936e-06, | |
| "loss": 0.5499, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 1.0697754157380421, | |
| "grad_norm": 0.6777327419801819, | |
| "learning_rate": 8.125465726021068e-06, | |
| "loss": 0.5499, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.073204183096177, | |
| "grad_norm": 0.5453482327699332, | |
| "learning_rate": 8.10986560702119e-06, | |
| "loss": 0.5503, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 1.0766329504543117, | |
| "grad_norm": 0.5825073699218344, | |
| "learning_rate": 8.09421595775997e-06, | |
| "loss": 0.5451, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.0800617178124465, | |
| "grad_norm": 0.5964267696290744, | |
| "learning_rate": 8.07851702748651e-06, | |
| "loss": 0.5497, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 1.083490485170581, | |
| "grad_norm": 0.5945145884647497, | |
| "learning_rate": 8.062769066234807e-06, | |
| "loss": 0.5467, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.086919252528716, | |
| "grad_norm": 0.5596939286945352, | |
| "learning_rate": 8.046972324819762e-06, | |
| "loss": 0.5469, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 1.0903480198868507, | |
| "grad_norm": 0.5511374508562613, | |
| "learning_rate": 8.031127054833192e-06, | |
| "loss": 0.5473, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.0937767872449855, | |
| "grad_norm": 0.5908893744280362, | |
| "learning_rate": 8.015233508639811e-06, | |
| "loss": 0.5504, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 1.09720555460312, | |
| "grad_norm": 0.6257657254754231, | |
| "learning_rate": 7.999291939373232e-06, | |
| "loss": 0.5473, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.1006343219612549, | |
| "grad_norm": 0.6187428942884239, | |
| "learning_rate": 7.983302600931911e-06, | |
| "loss": 0.5509, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 1.1040630893193897, | |
| "grad_norm": 0.5783570762769669, | |
| "learning_rate": 7.967265747975124e-06, | |
| "loss": 0.5507, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.1074918566775245, | |
| "grad_norm": 0.6002185456345444, | |
| "learning_rate": 7.951181635918895e-06, | |
| "loss": 0.5462, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 1.1109206240356593, | |
| "grad_norm": 0.5515256307910629, | |
| "learning_rate": 7.93505052093194e-06, | |
| "loss": 0.5533, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.1143493913937939, | |
| "grad_norm": 0.6078936624704993, | |
| "learning_rate": 7.91887265993158e-06, | |
| "loss": 0.5472, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 1.1177781587519287, | |
| "grad_norm": 0.5442571913199759, | |
| "learning_rate": 7.90264831057965e-06, | |
| "loss": 0.5486, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.1212069261100635, | |
| "grad_norm": 0.5524504533522675, | |
| "learning_rate": 7.8863777312784e-06, | |
| "loss": 0.5456, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 1.1246356934681982, | |
| "grad_norm": 0.543132690088679, | |
| "learning_rate": 7.870061181166372e-06, | |
| "loss": 0.5541, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.1280644608263328, | |
| "grad_norm": 0.5329328262846712, | |
| "learning_rate": 7.853698920114281e-06, | |
| "loss": 0.5528, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 1.1314932281844676, | |
| "grad_norm": 0.5139681859606589, | |
| "learning_rate": 7.837291208720867e-06, | |
| "loss": 0.5539, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.1349219955426024, | |
| "grad_norm": 0.5328884447403797, | |
| "learning_rate": 7.820838308308751e-06, | |
| "loss": 0.5489, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 1.1383507629007372, | |
| "grad_norm": 0.6114939943677603, | |
| "learning_rate": 7.804340480920274e-06, | |
| "loss": 0.5487, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.141779530258872, | |
| "grad_norm": 0.6051189586236244, | |
| "learning_rate": 7.787797989313317e-06, | |
| "loss": 0.5488, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 1.1452082976170066, | |
| "grad_norm": 0.5408630245450696, | |
| "learning_rate": 7.771211096957125e-06, | |
| "loss": 0.552, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.1486370649751414, | |
| "grad_norm": 0.5456867423021694, | |
| "learning_rate": 7.754580068028096e-06, | |
| "loss": 0.5469, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 1.1520658323332762, | |
| "grad_norm": 0.5474469138093225, | |
| "learning_rate": 7.737905167405596e-06, | |
| "loss": 0.5527, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.155494599691411, | |
| "grad_norm": 0.521490633503822, | |
| "learning_rate": 7.721186660667715e-06, | |
| "loss": 0.5476, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 1.1589233670495456, | |
| "grad_norm": 0.6191340910966903, | |
| "learning_rate": 7.704424814087056e-06, | |
| "loss": 0.5466, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.1623521344076804, | |
| "grad_norm": 0.5052183730102225, | |
| "learning_rate": 7.687619894626493e-06, | |
| "loss": 0.555, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 1.1657809017658152, | |
| "grad_norm": 0.5722854768006315, | |
| "learning_rate": 7.670772169934902e-06, | |
| "loss": 0.5513, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.16920966912395, | |
| "grad_norm": 0.5067922499590597, | |
| "learning_rate": 7.653881908342916e-06, | |
| "loss": 0.5489, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 1.1726384364820848, | |
| "grad_norm": 0.5604900851332252, | |
| "learning_rate": 7.636949378858647e-06, | |
| "loss": 0.5389, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.1760672038402193, | |
| "grad_norm": 0.5018446215200031, | |
| "learning_rate": 7.6199748511634005e-06, | |
| "loss": 0.5463, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 1.1794959711983541, | |
| "grad_norm": 0.48196349070673195, | |
| "learning_rate": 7.602958595607375e-06, | |
| "loss": 0.549, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.182924738556489, | |
| "grad_norm": 0.5951894113068615, | |
| "learning_rate": 7.5859008832053636e-06, | |
| "loss": 0.5484, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 1.1863535059146237, | |
| "grad_norm": 0.6230046151022545, | |
| "learning_rate": 7.568801985632439e-06, | |
| "loss": 0.5495, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.1897822732727585, | |
| "grad_norm": 0.5119551401124481, | |
| "learning_rate": 7.551662175219614e-06, | |
| "loss": 0.5466, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 1.1932110406308931, | |
| "grad_norm": 0.5302998077477141, | |
| "learning_rate": 7.5344817249495195e-06, | |
| "loss": 0.539, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.196639807989028, | |
| "grad_norm": 0.5804990070353486, | |
| "learning_rate": 7.51726090845205e-06, | |
| "loss": 0.5503, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 1.2000685753471627, | |
| "grad_norm": 0.5861147333730155, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 0.5481, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.2034973427052975, | |
| "grad_norm": 0.5966836470331061, | |
| "learning_rate": 7.482699274504708e-06, | |
| "loss": 0.5438, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 1.2069261100634323, | |
| "grad_norm": 0.5203444560983739, | |
| "learning_rate": 7.465359007511667e-06, | |
| "loss": 0.5476, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.2103548774215669, | |
| "grad_norm": 0.4994889019945669, | |
| "learning_rate": 7.44797947519614e-06, | |
| "loss": 0.5526, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 1.2137836447797017, | |
| "grad_norm": 0.5743096754445195, | |
| "learning_rate": 7.430560954358764e-06, | |
| "loss": 0.5484, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.2172124121378365, | |
| "grad_norm": 0.5032391615298711, | |
| "learning_rate": 7.413103722421139e-06, | |
| "loss": 0.5431, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 1.2206411794959713, | |
| "grad_norm": 0.5072263050309479, | |
| "learning_rate": 7.395608057421406e-06, | |
| "loss": 0.5415, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.2240699468541059, | |
| "grad_norm": 0.5302212227201574, | |
| "learning_rate": 7.378074238009827e-06, | |
| "loss": 0.5375, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 1.2274987142122407, | |
| "grad_norm": 0.5042827018544307, | |
| "learning_rate": 7.360502543444339e-06, | |
| "loss": 0.5493, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.2309274815703755, | |
| "grad_norm": 0.5067327603110284, | |
| "learning_rate": 7.342893253586112e-06, | |
| "loss": 0.5547, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 1.2343562489285103, | |
| "grad_norm": 0.5008933072194933, | |
| "learning_rate": 7.325246648895089e-06, | |
| "loss": 0.541, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.237785016286645, | |
| "grad_norm": 0.5065899468061212, | |
| "learning_rate": 7.307563010425517e-06, | |
| "loss": 0.5497, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 1.2412137836447796, | |
| "grad_norm": 0.5456582166819376, | |
| "learning_rate": 7.289842619821475e-06, | |
| "loss": 0.5501, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.2446425510029144, | |
| "grad_norm": 0.625178112160367, | |
| "learning_rate": 7.272085759312389e-06, | |
| "loss": 0.5478, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 1.2480713183610492, | |
| "grad_norm": 0.5527992246013412, | |
| "learning_rate": 7.254292711708529e-06, | |
| "loss": 0.5527, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.251500085719184, | |
| "grad_norm": 0.6681908362750082, | |
| "learning_rate": 7.236463760396516e-06, | |
| "loss": 0.5627, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 1.2549288530773186, | |
| "grad_norm": 0.5414761906585449, | |
| "learning_rate": 7.218599189334799e-06, | |
| "loss": 0.5437, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.2583576204354534, | |
| "grad_norm": 0.545519378308768, | |
| "learning_rate": 7.200699283049138e-06, | |
| "loss": 0.5458, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 1.2617863877935882, | |
| "grad_norm": 0.5175128490097378, | |
| "learning_rate": 7.182764326628068e-06, | |
| "loss": 0.5471, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.265215155151723, | |
| "grad_norm": 0.5436775659977866, | |
| "learning_rate": 7.164794605718366e-06, | |
| "loss": 0.5515, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 1.2686439225098578, | |
| "grad_norm": 0.6110723983754518, | |
| "learning_rate": 7.146790406520491e-06, | |
| "loss": 0.5456, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.2720726898679924, | |
| "grad_norm": 0.5770751655663271, | |
| "learning_rate": 7.128752015784036e-06, | |
| "loss": 0.5443, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 1.2755014572261272, | |
| "grad_norm": 0.4962264740154398, | |
| "learning_rate": 7.1106797208031554e-06, | |
| "loss": 0.5455, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.278930224584262, | |
| "grad_norm": 0.5486471659395239, | |
| "learning_rate": 7.0925738094119865e-06, | |
| "loss": 0.5519, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 1.2823589919423968, | |
| "grad_norm": 0.5503873404673431, | |
| "learning_rate": 7.0744345699800755e-06, | |
| "loss": 0.5478, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.2857877593005314, | |
| "grad_norm": 0.5042441453375152, | |
| "learning_rate": 7.056262291407773e-06, | |
| "loss": 0.5495, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 1.2892165266586662, | |
| "grad_norm": 0.5245616326038465, | |
| "learning_rate": 7.038057263121639e-06, | |
| "loss": 0.5397, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.292645294016801, | |
| "grad_norm": 0.5294940877618698, | |
| "learning_rate": 7.019819775069834e-06, | |
| "loss": 0.5403, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 1.2960740613749357, | |
| "grad_norm": 0.6147918213943395, | |
| "learning_rate": 7.001550117717499e-06, | |
| "loss": 0.5405, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.2995028287330705, | |
| "grad_norm": 0.5184070224956804, | |
| "learning_rate": 6.983248582042126e-06, | |
| "loss": 0.5481, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 1.3029315960912053, | |
| "grad_norm": 0.5295411569445538, | |
| "learning_rate": 6.9649154595289326e-06, | |
| "loss": 0.5466, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.30636036344934, | |
| "grad_norm": 0.49956769450406724, | |
| "learning_rate": 6.94655104216621e-06, | |
| "loss": 0.5461, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 1.3097891308074747, | |
| "grad_norm": 0.5529813856201232, | |
| "learning_rate": 6.92815562244068e-06, | |
| "loss": 0.5456, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.3132178981656095, | |
| "grad_norm": 0.48747897180543187, | |
| "learning_rate": 6.90972949333283e-06, | |
| "loss": 0.5506, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 1.316646665523744, | |
| "grad_norm": 0.5423084499887999, | |
| "learning_rate": 6.891272948312251e-06, | |
| "loss": 0.5497, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.320075432881879, | |
| "grad_norm": 0.4982712510800406, | |
| "learning_rate": 6.872786281332965e-06, | |
| "loss": 0.5493, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 1.3235042002400137, | |
| "grad_norm": 0.47720857931691807, | |
| "learning_rate": 6.854269786828741e-06, | |
| "loss": 0.5471, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.3269329675981485, | |
| "grad_norm": 0.5568806463136045, | |
| "learning_rate": 6.8357237597084015e-06, | |
| "loss": 0.5527, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 1.3303617349562833, | |
| "grad_norm": 0.510634543124876, | |
| "learning_rate": 6.817148495351131e-06, | |
| "loss": 0.5495, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.333790502314418, | |
| "grad_norm": 0.5164556362855032, | |
| "learning_rate": 6.7985442896017765e-06, | |
| "loss": 0.5463, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 1.3372192696725527, | |
| "grad_norm": 0.5254656454551451, | |
| "learning_rate": 6.779911438766117e-06, | |
| "loss": 0.5426, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.3406480370306875, | |
| "grad_norm": 0.49029587603019725, | |
| "learning_rate": 6.7612502396061685e-06, | |
| "loss": 0.5429, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 1.3440768043888223, | |
| "grad_norm": 0.4789585618635431, | |
| "learning_rate": 6.742560989335438e-06, | |
| "loss": 0.543, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.3475055717469568, | |
| "grad_norm": 0.5351724574021284, | |
| "learning_rate": 6.723843985614201e-06, | |
| "loss": 0.549, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 1.3509343391050916, | |
| "grad_norm": 0.5285309854529648, | |
| "learning_rate": 6.705099526544757e-06, | |
| "loss": 0.5491, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.3543631064632264, | |
| "grad_norm": 0.5744541149885962, | |
| "learning_rate": 6.686327910666679e-06, | |
| "loss": 0.5469, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 1.3577918738213612, | |
| "grad_norm": 0.5276476840483824, | |
| "learning_rate": 6.667529436952064e-06, | |
| "loss": 0.5501, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.361220641179496, | |
| "grad_norm": 0.5136271937577194, | |
| "learning_rate": 6.6487044048007696e-06, | |
| "loss": 0.5457, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 1.3646494085376308, | |
| "grad_norm": 0.46138577500457834, | |
| "learning_rate": 6.629853114035643e-06, | |
| "loss": 0.5461, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.3680781758957654, | |
| "grad_norm": 0.4895982866432622, | |
| "learning_rate": 6.610975864897746e-06, | |
| "loss": 0.5392, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 1.3715069432539002, | |
| "grad_norm": 0.5042414486514584, | |
| "learning_rate": 6.5920729580415795e-06, | |
| "loss": 0.5463, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.374935710612035, | |
| "grad_norm": 0.507137387593949, | |
| "learning_rate": 6.573144694530287e-06, | |
| "loss": 0.5486, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 1.3783644779701696, | |
| "grad_norm": 0.5035112896994087, | |
| "learning_rate": 6.554191375830861e-06, | |
| "loss": 0.5399, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.3817932453283044, | |
| "grad_norm": 0.5191434631626274, | |
| "learning_rate": 6.535213303809349e-06, | |
| "loss": 0.5462, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 1.3852220126864392, | |
| "grad_norm": 0.5215645909213554, | |
| "learning_rate": 6.516210780726032e-06, | |
| "loss": 0.5492, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.388650780044574, | |
| "grad_norm": 0.4955063426711122, | |
| "learning_rate": 6.497184109230628e-06, | |
| "loss": 0.547, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 1.3920795474027088, | |
| "grad_norm": 0.5083904799973671, | |
| "learning_rate": 6.478133592357455e-06, | |
| "loss": 0.5495, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.3955083147608436, | |
| "grad_norm": 0.5413459772025052, | |
| "learning_rate": 6.4590595335206154e-06, | |
| "loss": 0.5471, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 1.3989370821189782, | |
| "grad_norm": 0.5159096563706932, | |
| "learning_rate": 6.43996223650916e-06, | |
| "loss": 0.5475, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.402365849477113, | |
| "grad_norm": 0.5232964962054851, | |
| "learning_rate": 6.420842005482248e-06, | |
| "loss": 0.5526, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 1.4057946168352478, | |
| "grad_norm": 0.4879763438255916, | |
| "learning_rate": 6.401699144964306e-06, | |
| "loss": 0.5563, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.4092233841933826, | |
| "grad_norm": 0.49746751110206966, | |
| "learning_rate": 6.382533959840178e-06, | |
| "loss": 0.5401, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 1.4126521515515171, | |
| "grad_norm": 0.551660200659416, | |
| "learning_rate": 6.3633467553502625e-06, | |
| "loss": 0.5464, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.416080918909652, | |
| "grad_norm": 0.5127978902578759, | |
| "learning_rate": 6.344137837085662e-06, | |
| "loss": 0.5436, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 1.4195096862677867, | |
| "grad_norm": 0.528021714081405, | |
| "learning_rate": 6.32490751098331e-06, | |
| "loss": 0.5475, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.4229384536259215, | |
| "grad_norm": 0.5235426406662949, | |
| "learning_rate": 6.305656083321092e-06, | |
| "loss": 0.5495, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 1.4263672209840563, | |
| "grad_norm": 0.5932716644747056, | |
| "learning_rate": 6.286383860712982e-06, | |
| "loss": 0.5437, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.429795988342191, | |
| "grad_norm": 0.5329514416163247, | |
| "learning_rate": 6.2670911501041495e-06, | |
| "loss": 0.5537, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 1.4332247557003257, | |
| "grad_norm": 0.5203312437593606, | |
| "learning_rate": 6.247778258766069e-06, | |
| "loss": 0.5453, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.4366535230584605, | |
| "grad_norm": 0.5530831064402718, | |
| "learning_rate": 6.228445494291635e-06, | |
| "loss": 0.5441, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 1.4400822904165953, | |
| "grad_norm": 0.5356248838373525, | |
| "learning_rate": 6.209093164590253e-06, | |
| "loss": 0.5467, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.4435110577747299, | |
| "grad_norm": 0.507854588281778, | |
| "learning_rate": 6.189721577882942e-06, | |
| "loss": 0.5473, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 1.4469398251328647, | |
| "grad_norm": 0.5020349063977403, | |
| "learning_rate": 6.170331042697425e-06, | |
| "loss": 0.5439, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 1.4503685924909995, | |
| "grad_norm": 0.5299904655120313, | |
| "learning_rate": 6.150921867863215e-06, | |
| "loss": 0.5429, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 1.4537973598491343, | |
| "grad_norm": 0.5599366493504166, | |
| "learning_rate": 6.131494362506693e-06, | |
| "loss": 0.5479, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.457226127207269, | |
| "grad_norm": 0.5419693437182285, | |
| "learning_rate": 6.112048836046185e-06, | |
| "loss": 0.5515, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 1.4606548945654039, | |
| "grad_norm": 0.5619194156719064, | |
| "learning_rate": 6.09258559818704e-06, | |
| "loss": 0.5493, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.4640836619235384, | |
| "grad_norm": 0.5180975989810822, | |
| "learning_rate": 6.0731049589166895e-06, | |
| "loss": 0.5447, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 1.4675124292816732, | |
| "grad_norm": 0.47862105476236594, | |
| "learning_rate": 6.053607228499719e-06, | |
| "loss": 0.5411, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.470941196639808, | |
| "grad_norm": 0.5019570636347335, | |
| "learning_rate": 6.034092717472917e-06, | |
| "loss": 0.5479, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 1.4743699639979426, | |
| "grad_norm": 0.5005596379353178, | |
| "learning_rate": 6.014561736640334e-06, | |
| "loss": 0.5435, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.4777987313560774, | |
| "grad_norm": 0.4759889724446824, | |
| "learning_rate": 5.9950145970683375e-06, | |
| "loss": 0.5434, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 1.4812274987142122, | |
| "grad_norm": 0.47708542568255435, | |
| "learning_rate": 5.975451610080643e-06, | |
| "loss": 0.5382, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.484656266072347, | |
| "grad_norm": 0.507797233049438, | |
| "learning_rate": 5.955873087253371e-06, | |
| "loss": 0.5488, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 1.4880850334304818, | |
| "grad_norm": 0.48901155035523547, | |
| "learning_rate": 5.936279340410082e-06, | |
| "loss": 0.5379, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 1.4915138007886166, | |
| "grad_norm": 0.5482350970529439, | |
| "learning_rate": 5.916670681616797e-06, | |
| "loss": 0.5444, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 1.4949425681467512, | |
| "grad_norm": 0.48909311036019854, | |
| "learning_rate": 5.8970474231770445e-06, | |
| "loss": 0.5438, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.498371335504886, | |
| "grad_norm": 0.5029129982744719, | |
| "learning_rate": 5.877409877626876e-06, | |
| "loss": 0.543, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 1.5018001028630208, | |
| "grad_norm": 0.49508908810255237, | |
| "learning_rate": 5.857758357729892e-06, | |
| "loss": 0.5401, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 1.5052288702211554, | |
| "grad_norm": 0.5186990415007219, | |
| "learning_rate": 5.838093176472259e-06, | |
| "loss": 0.544, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 1.5086576375792902, | |
| "grad_norm": 0.5236949094459556, | |
| "learning_rate": 5.8184146470577265e-06, | |
| "loss": 0.5334, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.512086404937425, | |
| "grad_norm": 0.4893285394654287, | |
| "learning_rate": 5.798723082902636e-06, | |
| "loss": 0.5433, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 1.5155151722955598, | |
| "grad_norm": 0.4907763180175273, | |
| "learning_rate": 5.779018797630934e-06, | |
| "loss": 0.543, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 1.5189439396536946, | |
| "grad_norm": 0.5637821736714953, | |
| "learning_rate": 5.75930210506917e-06, | |
| "loss": 0.5406, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 1.5223727070118294, | |
| "grad_norm": 0.5100910825032349, | |
| "learning_rate": 5.739573319241505e-06, | |
| "loss": 0.5397, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.5258014743699642, | |
| "grad_norm": 0.45580474474329813, | |
| "learning_rate": 5.719832754364707e-06, | |
| "loss": 0.5468, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 1.5292302417280987, | |
| "grad_norm": 0.5045515176992598, | |
| "learning_rate": 5.7000807248431466e-06, | |
| "loss": 0.5367, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 1.5326590090862335, | |
| "grad_norm": 0.4933966502093522, | |
| "learning_rate": 5.680317545263786e-06, | |
| "loss": 0.5461, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 1.5360877764443681, | |
| "grad_norm": 0.5029727396999837, | |
| "learning_rate": 5.66054353039118e-06, | |
| "loss": 0.5419, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.539516543802503, | |
| "grad_norm": 0.5034621640442088, | |
| "learning_rate": 5.640758995162446e-06, | |
| "loss": 0.5425, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 1.5429453111606377, | |
| "grad_norm": 0.5084425747704971, | |
| "learning_rate": 5.620964254682267e-06, | |
| "loss": 0.5414, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.5463740785187725, | |
| "grad_norm": 0.505196202702042, | |
| "learning_rate": 5.601159624217854e-06, | |
| "loss": 0.5474, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 1.5498028458769073, | |
| "grad_norm": 0.5308371720246889, | |
| "learning_rate": 5.58134541919394e-06, | |
| "loss": 0.5433, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 1.553231613235042, | |
| "grad_norm": 0.4934713800815739, | |
| "learning_rate": 5.5615219551877476e-06, | |
| "loss": 0.5504, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 1.556660380593177, | |
| "grad_norm": 0.46471141316877956, | |
| "learning_rate": 5.5416895479239665e-06, | |
| "loss": 0.5413, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 1.5600891479513115, | |
| "grad_norm": 0.47931849486997197, | |
| "learning_rate": 5.521848513269723e-06, | |
| "loss": 0.5373, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 1.5635179153094463, | |
| "grad_norm": 0.4684090998446449, | |
| "learning_rate": 5.501999167229554e-06, | |
| "loss": 0.5419, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 1.5669466826675809, | |
| "grad_norm": 0.5051008793716584, | |
| "learning_rate": 5.482141825940363e-06, | |
| "loss": 0.539, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 1.5703754500257157, | |
| "grad_norm": 0.5270612668963798, | |
| "learning_rate": 5.4622768056664e-06, | |
| "loss": 0.552, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 1.5738042173838505, | |
| "grad_norm": 0.5102700734031906, | |
| "learning_rate": 5.442404422794212e-06, | |
| "loss": 0.5413, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 1.5772329847419853, | |
| "grad_norm": 0.4651045699311542, | |
| "learning_rate": 5.42252499382761e-06, | |
| "loss": 0.5457, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.58066175210012, | |
| "grad_norm": 0.48408561746393897, | |
| "learning_rate": 5.402638835382628e-06, | |
| "loss": 0.5454, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 1.5840905194582549, | |
| "grad_norm": 0.4886707605239818, | |
| "learning_rate": 5.38274626418248e-06, | |
| "loss": 0.5404, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 1.5875192868163897, | |
| "grad_norm": 0.5358058957587754, | |
| "learning_rate": 5.362847597052509e-06, | |
| "loss": 0.5407, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 1.5909480541745242, | |
| "grad_norm": 0.4969840985996426, | |
| "learning_rate": 5.3429431509151515e-06, | |
| "loss": 0.5408, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 1.594376821532659, | |
| "grad_norm": 0.4903079059609584, | |
| "learning_rate": 5.323033242784889e-06, | |
| "loss": 0.5368, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 1.5978055888907936, | |
| "grad_norm": 0.4738390208380356, | |
| "learning_rate": 5.303118189763187e-06, | |
| "loss": 0.5482, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 1.6012343562489284, | |
| "grad_norm": 0.5096387167156792, | |
| "learning_rate": 5.283198309033457e-06, | |
| "loss": 0.5476, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 1.6046631236070632, | |
| "grad_norm": 0.5265182211890799, | |
| "learning_rate": 5.263273917856e-06, | |
| "loss": 0.5363, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 1.608091890965198, | |
| "grad_norm": 0.5159782309776882, | |
| "learning_rate": 5.243345333562954e-06, | |
| "loss": 0.5453, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 1.6115206583233328, | |
| "grad_norm": 0.4866586786251063, | |
| "learning_rate": 5.22341287355324e-06, | |
| "loss": 0.5366, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 1.6149494256814676, | |
| "grad_norm": 0.5069139373526034, | |
| "learning_rate": 5.203476855287507e-06, | |
| "loss": 0.5424, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 1.6183781930396024, | |
| "grad_norm": 0.47559808993135405, | |
| "learning_rate": 5.183537596283075e-06, | |
| "loss": 0.5437, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 1.621806960397737, | |
| "grad_norm": 0.4777303119256066, | |
| "learning_rate": 5.1635954141088815e-06, | |
| "loss": 0.5432, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 1.6252357277558718, | |
| "grad_norm": 0.46998224709647424, | |
| "learning_rate": 5.143650626380417e-06, | |
| "loss": 0.5487, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 1.6286644951140063, | |
| "grad_norm": 0.5020515855916554, | |
| "learning_rate": 5.123703550754669e-06, | |
| "loss": 0.5437, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 1.6320932624721411, | |
| "grad_norm": 0.4986048548791825, | |
| "learning_rate": 5.103754504925071e-06, | |
| "loss": 0.548, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 1.635522029830276, | |
| "grad_norm": 0.49818017928688363, | |
| "learning_rate": 5.083803806616428e-06, | |
| "loss": 0.5407, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 1.6389507971884107, | |
| "grad_norm": 0.4687036183917943, | |
| "learning_rate": 5.06385177357987e-06, | |
| "loss": 0.5418, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 1.6423795645465455, | |
| "grad_norm": 0.5165271708659772, | |
| "learning_rate": 5.043898723587779e-06, | |
| "loss": 0.5473, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 1.6458083319046803, | |
| "grad_norm": 0.583421869027607, | |
| "learning_rate": 5.023944974428739e-06, | |
| "loss": 0.5383, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.6492370992628151, | |
| "grad_norm": 0.508317228928292, | |
| "learning_rate": 5.003990843902463e-06, | |
| "loss": 0.5386, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 1.6526658666209497, | |
| "grad_norm": 0.5359492047497594, | |
| "learning_rate": 4.9840366498147495e-06, | |
| "loss": 0.5425, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 1.6560946339790845, | |
| "grad_norm": 0.5183122191069008, | |
| "learning_rate": 4.964082709972394e-06, | |
| "loss": 0.5502, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 1.6595234013372193, | |
| "grad_norm": 0.5124930538710774, | |
| "learning_rate": 4.944129342178156e-06, | |
| "loss": 0.54, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 1.662952168695354, | |
| "grad_norm": 0.5037854005505844, | |
| "learning_rate": 4.924176864225678e-06, | |
| "loss": 0.538, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 1.6663809360534887, | |
| "grad_norm": 0.5034524367425028, | |
| "learning_rate": 4.90422559389443e-06, | |
| "loss": 0.5342, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 1.6698097034116235, | |
| "grad_norm": 0.48847928157114906, | |
| "learning_rate": 4.88427584894465e-06, | |
| "loss": 0.5414, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 1.6732384707697583, | |
| "grad_norm": 0.5197993207547689, | |
| "learning_rate": 4.864327947112281e-06, | |
| "loss": 0.5437, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 1.676667238127893, | |
| "grad_norm": 0.4979107929661232, | |
| "learning_rate": 4.84438220610391e-06, | |
| "loss": 0.541, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 1.6800960054860279, | |
| "grad_norm": 0.47357100832452387, | |
| "learning_rate": 4.82443894359171e-06, | |
| "loss": 0.5375, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 1.6835247728441627, | |
| "grad_norm": 0.47499915823273264, | |
| "learning_rate": 4.804498477208382e-06, | |
| "loss": 0.5435, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 1.6869535402022973, | |
| "grad_norm": 0.5246936291111205, | |
| "learning_rate": 4.784561124542088e-06, | |
| "loss": 0.5399, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 1.690382307560432, | |
| "grad_norm": 0.5396195065092629, | |
| "learning_rate": 4.764627203131401e-06, | |
| "loss": 0.5323, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 1.6938110749185666, | |
| "grad_norm": 0.49550562745484766, | |
| "learning_rate": 4.744697030460248e-06, | |
| "loss": 0.54, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 1.6972398422767014, | |
| "grad_norm": 0.4834798755139004, | |
| "learning_rate": 4.724770923952844e-06, | |
| "loss": 0.5441, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 1.7006686096348362, | |
| "grad_norm": 0.5285955350607044, | |
| "learning_rate": 4.7048492009686525e-06, | |
| "loss": 0.5391, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 1.704097376992971, | |
| "grad_norm": 0.5121886893362284, | |
| "learning_rate": 4.6849321787973116e-06, | |
| "loss": 0.5406, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 1.7075261443511058, | |
| "grad_norm": 0.5563183347246416, | |
| "learning_rate": 4.6650201746535926e-06, | |
| "loss": 0.5381, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 1.7109549117092406, | |
| "grad_norm": 0.49534162399833953, | |
| "learning_rate": 4.645113505672353e-06, | |
| "loss": 0.5359, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 1.7143836790673754, | |
| "grad_norm": 0.4754638729290019, | |
| "learning_rate": 4.625212488903467e-06, | |
| "loss": 0.5411, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.71781244642551, | |
| "grad_norm": 0.4701836831162324, | |
| "learning_rate": 4.605317441306798e-06, | |
| "loss": 0.5362, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 1.7212412137836448, | |
| "grad_norm": 0.48575768441353867, | |
| "learning_rate": 4.585428679747133e-06, | |
| "loss": 0.5419, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 1.7246699811417794, | |
| "grad_norm": 0.561176281859862, | |
| "learning_rate": 4.565546520989139e-06, | |
| "loss": 0.5361, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 1.7280987484999142, | |
| "grad_norm": 0.4991619958671278, | |
| "learning_rate": 4.545671281692331e-06, | |
| "loss": 0.5379, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.731527515858049, | |
| "grad_norm": 0.4894062203682473, | |
| "learning_rate": 4.525803278406011e-06, | |
| "loss": 0.5359, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 1.7349562832161838, | |
| "grad_norm": 0.4811044764793809, | |
| "learning_rate": 4.505942827564242e-06, | |
| "loss": 0.5426, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 1.7383850505743186, | |
| "grad_norm": 0.48626341909810544, | |
| "learning_rate": 4.4860902454807905e-06, | |
| "loss": 0.5336, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 1.7418138179324534, | |
| "grad_norm": 0.4677680916526279, | |
| "learning_rate": 4.466245848344106e-06, | |
| "loss": 0.5408, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.7452425852905882, | |
| "grad_norm": 0.49984747393620377, | |
| "learning_rate": 4.446409952212278e-06, | |
| "loss": 0.5347, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 1.7486713526487228, | |
| "grad_norm": 0.4932054681232252, | |
| "learning_rate": 4.426582873007999e-06, | |
| "loss": 0.5451, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 1.7521001200068576, | |
| "grad_norm": 0.4794644104456506, | |
| "learning_rate": 4.406764926513536e-06, | |
| "loss": 0.5403, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 1.7555288873649921, | |
| "grad_norm": 0.45717917978063133, | |
| "learning_rate": 4.386956428365701e-06, | |
| "loss": 0.5379, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.758957654723127, | |
| "grad_norm": 0.46935105917323766, | |
| "learning_rate": 4.36715769405083e-06, | |
| "loss": 0.5366, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 1.7623864220812617, | |
| "grad_norm": 0.4839602278611378, | |
| "learning_rate": 4.347369038899744e-06, | |
| "loss": 0.5403, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 1.7658151894393965, | |
| "grad_norm": 0.4996673443974444, | |
| "learning_rate": 4.327590778082734e-06, | |
| "loss": 0.5413, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 1.7692439567975313, | |
| "grad_norm": 0.47519860849274637, | |
| "learning_rate": 4.307823226604555e-06, | |
| "loss": 0.5402, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.7726727241556661, | |
| "grad_norm": 0.47545446063997776, | |
| "learning_rate": 4.28806669929938e-06, | |
| "loss": 0.5374, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 1.776101491513801, | |
| "grad_norm": 0.5148035850870979, | |
| "learning_rate": 4.2683215108258145e-06, | |
| "loss": 0.5392, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 1.7795302588719355, | |
| "grad_norm": 0.49138867529600283, | |
| "learning_rate": 4.2485879756618685e-06, | |
| "loss": 0.5377, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 1.7829590262300703, | |
| "grad_norm": 0.4888120553433302, | |
| "learning_rate": 4.228866408099945e-06, | |
| "loss": 0.5415, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.7863877935882049, | |
| "grad_norm": 0.46500050166415424, | |
| "learning_rate": 4.2091571222418546e-06, | |
| "loss": 0.536, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 1.7898165609463397, | |
| "grad_norm": 0.4829554221776432, | |
| "learning_rate": 4.189460431993788e-06, | |
| "loss": 0.5451, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 1.7932453283044745, | |
| "grad_norm": 0.4917384629382616, | |
| "learning_rate": 4.169776651061334e-06, | |
| "loss": 0.5409, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 1.7966740956626093, | |
| "grad_norm": 0.49620621093693373, | |
| "learning_rate": 4.150106092944475e-06, | |
| "loss": 0.541, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.800102863020744, | |
| "grad_norm": 0.467340353977474, | |
| "learning_rate": 4.130449070932594e-06, | |
| "loss": 0.5407, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 1.8035316303788789, | |
| "grad_norm": 0.49470765077176454, | |
| "learning_rate": 4.110805898099492e-06, | |
| "loss": 0.5416, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 1.8069603977370137, | |
| "grad_norm": 0.4666535437410526, | |
| "learning_rate": 4.091176887298392e-06, | |
| "loss": 0.5406, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 1.8103891650951482, | |
| "grad_norm": 0.46143547716397726, | |
| "learning_rate": 4.071562351156966e-06, | |
| "loss": 0.5344, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.813817932453283, | |
| "grad_norm": 0.46910481572596163, | |
| "learning_rate": 4.051962602072343e-06, | |
| "loss": 0.5364, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 1.8172466998114178, | |
| "grad_norm": 0.4694928382743332, | |
| "learning_rate": 4.032377952206148e-06, | |
| "loss": 0.5413, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 1.8206754671695524, | |
| "grad_norm": 0.47422238985591536, | |
| "learning_rate": 4.012808713479523e-06, | |
| "loss": 0.5435, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 1.8241042345276872, | |
| "grad_norm": 0.48396135634665843, | |
| "learning_rate": 3.993255197568154e-06, | |
| "loss": 0.5336, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.827533001885822, | |
| "grad_norm": 0.4873226150393351, | |
| "learning_rate": 3.9737177158973185e-06, | |
| "loss": 0.5309, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 1.8309617692439568, | |
| "grad_norm": 0.4576384143762062, | |
| "learning_rate": 3.954196579636918e-06, | |
| "loss": 0.5444, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 1.8343905366020916, | |
| "grad_norm": 0.49454868144127, | |
| "learning_rate": 3.93469209969652e-06, | |
| "loss": 0.5353, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 1.8378193039602264, | |
| "grad_norm": 0.4490405182223933, | |
| "learning_rate": 3.91520458672042e-06, | |
| "loss": 0.5352, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.8412480713183612, | |
| "grad_norm": 0.4892412353541936, | |
| "learning_rate": 3.895734351082668e-06, | |
| "loss": 0.5364, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 1.8446768386764958, | |
| "grad_norm": 0.44948826305297584, | |
| "learning_rate": 3.876281702882156e-06, | |
| "loss": 0.5382, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 1.8481056060346306, | |
| "grad_norm": 0.46974192317523716, | |
| "learning_rate": 3.8568469519376585e-06, | |
| "loss": 0.534, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 1.8515343733927652, | |
| "grad_norm": 0.4541143574340376, | |
| "learning_rate": 3.837430407782896e-06, | |
| "loss": 0.5426, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.8549631407509, | |
| "grad_norm": 0.4749792696125447, | |
| "learning_rate": 3.818032379661626e-06, | |
| "loss": 0.5396, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 1.8583919081090348, | |
| "grad_norm": 0.46058829711339094, | |
| "learning_rate": 3.7986531765226965e-06, | |
| "loss": 0.5385, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 1.8618206754671696, | |
| "grad_norm": 0.45724757617193085, | |
| "learning_rate": 3.779293107015136e-06, | |
| "loss": 0.5325, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 1.8652494428253044, | |
| "grad_norm": 0.4613637136202585, | |
| "learning_rate": 3.759952479483232e-06, | |
| "loss": 0.5332, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.8686782101834392, | |
| "grad_norm": 0.45902944748623037, | |
| "learning_rate": 3.7406316019616297e-06, | |
| "loss": 0.5402, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 1.872106977541574, | |
| "grad_norm": 0.4810427911391462, | |
| "learning_rate": 3.7213307821704115e-06, | |
| "loss": 0.5397, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 1.8755357448997085, | |
| "grad_norm": 0.5021212230417156, | |
| "learning_rate": 3.7020503275102095e-06, | |
| "loss": 0.5478, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 1.8789645122578433, | |
| "grad_norm": 0.4836776357761823, | |
| "learning_rate": 3.6827905450573022e-06, | |
| "loss": 0.5358, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.882393279615978, | |
| "grad_norm": 0.5101566373577191, | |
| "learning_rate": 3.6635517415587264e-06, | |
| "loss": 0.5299, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 1.8858220469741127, | |
| "grad_norm": 0.4734869935919425, | |
| "learning_rate": 3.6443342234273905e-06, | |
| "loss": 0.5364, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 1.8892508143322475, | |
| "grad_norm": 0.4977350264147008, | |
| "learning_rate": 3.6251382967371938e-06, | |
| "loss": 0.5344, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 1.8926795816903823, | |
| "grad_norm": 0.45745712112103115, | |
| "learning_rate": 3.6059642672181537e-06, | |
| "loss": 0.5338, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.896108349048517, | |
| "grad_norm": 0.5104324084550051, | |
| "learning_rate": 3.586812440251537e-06, | |
| "loss": 0.5347, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 1.899537116406652, | |
| "grad_norm": 0.5335392069435976, | |
| "learning_rate": 3.5676831208649887e-06, | |
| "loss": 0.5346, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 1.9029658837647867, | |
| "grad_norm": 0.4895513127909529, | |
| "learning_rate": 3.5485766137276894e-06, | |
| "loss": 0.5418, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 1.9063946511229213, | |
| "grad_norm": 0.4542167715320498, | |
| "learning_rate": 3.5294932231454838e-06, | |
| "loss": 0.5385, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.909823418481056, | |
| "grad_norm": 0.4744313937102546, | |
| "learning_rate": 3.510433253056045e-06, | |
| "loss": 0.5391, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 1.9132521858391907, | |
| "grad_norm": 0.4427663775130247, | |
| "learning_rate": 3.4913970070240388e-06, | |
| "loss": 0.5444, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 1.9166809531973255, | |
| "grad_norm": 0.47529978421135, | |
| "learning_rate": 3.4723847882362755e-06, | |
| "loss": 0.5373, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 1.9201097205554603, | |
| "grad_norm": 0.4561099449022098, | |
| "learning_rate": 3.4533968994968913e-06, | |
| "loss": 0.5332, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.923538487913595, | |
| "grad_norm": 0.47185420315317955, | |
| "learning_rate": 3.4344336432225205e-06, | |
| "loss": 0.5277, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 1.9269672552717299, | |
| "grad_norm": 0.4730035039099479, | |
| "learning_rate": 3.41549532143748e-06, | |
| "loss": 0.5377, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 1.9303960226298646, | |
| "grad_norm": 0.48079447901439726, | |
| "learning_rate": 3.3965822357689626e-06, | |
| "loss": 0.5349, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 1.9338247899879994, | |
| "grad_norm": 0.458616349152276, | |
| "learning_rate": 3.3776946874422268e-06, | |
| "loss": 0.5312, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.937253557346134, | |
| "grad_norm": 0.4555844676968641, | |
| "learning_rate": 3.3588329772758054e-06, | |
| "loss": 0.5359, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 1.9406823247042688, | |
| "grad_norm": 0.4916927479459658, | |
| "learning_rate": 3.3399974056767095e-06, | |
| "loss": 0.5343, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 1.9441110920624034, | |
| "grad_norm": 0.47636277234450725, | |
| "learning_rate": 3.3211882726356447e-06, | |
| "loss": 0.5363, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 1.9475398594205382, | |
| "grad_norm": 0.46891572696488587, | |
| "learning_rate": 3.30240587772224e-06, | |
| "loss": 0.537, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.950968626778673, | |
| "grad_norm": 0.4845143780415882, | |
| "learning_rate": 3.283650520080265e-06, | |
| "loss": 0.5367, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 1.9543973941368078, | |
| "grad_norm": 0.4714569513171776, | |
| "learning_rate": 3.2649224984228756e-06, | |
| "loss": 0.5394, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 1.9578261614949426, | |
| "grad_norm": 0.46470157896584324, | |
| "learning_rate": 3.246222111027849e-06, | |
| "loss": 0.5404, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 1.9612549288530774, | |
| "grad_norm": 0.46189223769008775, | |
| "learning_rate": 3.227549655732843e-06, | |
| "loss": 0.5394, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.9646836962112122, | |
| "grad_norm": 0.459933595776418, | |
| "learning_rate": 3.2089054299306376e-06, | |
| "loss": 0.5351, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 1.9681124635693468, | |
| "grad_norm": 0.4881070886064201, | |
| "learning_rate": 3.19028973056441e-06, | |
| "loss": 0.5326, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 1.9715412309274816, | |
| "grad_norm": 0.4518011695478976, | |
| "learning_rate": 3.171702854123008e-06, | |
| "loss": 0.5423, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 1.9749699982856164, | |
| "grad_norm": 0.45756039660819475, | |
| "learning_rate": 3.153145096636211e-06, | |
| "loss": 0.5339, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.978398765643751, | |
| "grad_norm": 0.4544722411338075, | |
| "learning_rate": 3.134616753670036e-06, | |
| "loss": 0.5308, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 1.9818275330018857, | |
| "grad_norm": 0.4592115025179121, | |
| "learning_rate": 3.1161181203220146e-06, | |
| "loss": 0.5353, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 1.9852563003600205, | |
| "grad_norm": 0.46992937248193806, | |
| "learning_rate": 3.0976494912165e-06, | |
| "loss": 0.5369, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 1.9886850677181553, | |
| "grad_norm": 0.4675123294101481, | |
| "learning_rate": 3.079211160499975e-06, | |
| "loss": 0.5326, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.9921138350762901, | |
| "grad_norm": 0.480213421331211, | |
| "learning_rate": 3.060803421836363e-06, | |
| "loss": 0.5392, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 1.995542602434425, | |
| "grad_norm": 0.4628568848487098, | |
| "learning_rate": 3.0424265684023556e-06, | |
| "loss": 0.5321, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 1.9989713697925597, | |
| "grad_norm": 0.49388511981349187, | |
| "learning_rate": 3.0240808928827397e-06, | |
| "loss": 0.5287, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 1.9996571232641864, | |
| "eval_loss": 0.5881877541542053, | |
| "eval_runtime": 204.4729, | |
| "eval_samples_per_second": 48.06, | |
| "eval_steps_per_second": 1.506, | |
| "step": 2916 | |
| }, | |
| { | |
| "epoch": 2.0024001371506945, | |
| "grad_norm": 0.874753190390515, | |
| "learning_rate": 3.0057666874657365e-06, | |
| "loss": 0.4981, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 2.005828904508829, | |
| "grad_norm": 0.5805650180046337, | |
| "learning_rate": 2.9874842438383508e-06, | |
| "loss": 0.4781, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 2.0092576718669637, | |
| "grad_norm": 0.5290132742038788, | |
| "learning_rate": 2.9692338531817205e-06, | |
| "loss": 0.473, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 2.0126864392250985, | |
| "grad_norm": 0.5751295049236377, | |
| "learning_rate": 2.951015806166485e-06, | |
| "loss": 0.4788, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 2.0161152065832333, | |
| "grad_norm": 0.5442073882893506, | |
| "learning_rate": 2.9328303929481507e-06, | |
| "loss": 0.4756, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 2.019543973941368, | |
| "grad_norm": 0.5046978835481467, | |
| "learning_rate": 2.9146779031624683e-06, | |
| "loss": 0.4747, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 2.022972741299503, | |
| "grad_norm": 0.4694017759591277, | |
| "learning_rate": 2.8965586259208295e-06, | |
| "loss": 0.4787, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.0264015086576377, | |
| "grad_norm": 0.49182764229033304, | |
| "learning_rate": 2.878472849805651e-06, | |
| "loss": 0.4817, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 2.0298302760157725, | |
| "grad_norm": 0.5353082540824908, | |
| "learning_rate": 2.860420862865787e-06, | |
| "loss": 0.4814, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 2.0332590433739073, | |
| "grad_norm": 0.4849568396076438, | |
| "learning_rate": 2.842402952611931e-06, | |
| "loss": 0.4774, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 2.0366878107320416, | |
| "grad_norm": 0.455032309231727, | |
| "learning_rate": 2.82441940601205e-06, | |
| "loss": 0.488, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 2.0401165780901764, | |
| "grad_norm": 0.47238583732747885, | |
| "learning_rate": 2.8064705094868074e-06, | |
| "loss": 0.4847, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 2.0435453454483112, | |
| "grad_norm": 0.46589829010871375, | |
| "learning_rate": 2.7885565489049948e-06, | |
| "loss": 0.4745, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 2.046974112806446, | |
| "grad_norm": 0.5097253834071658, | |
| "learning_rate": 2.7706778095789905e-06, | |
| "loss": 0.4822, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 2.050402880164581, | |
| "grad_norm": 0.4744363699843961, | |
| "learning_rate": 2.7528345762602125e-06, | |
| "loss": 0.4837, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 2.0538316475227156, | |
| "grad_norm": 0.48096853101811016, | |
| "learning_rate": 2.7350271331345733e-06, | |
| "loss": 0.4689, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 2.0572604148808504, | |
| "grad_norm": 0.45888264058390604, | |
| "learning_rate": 2.7172557638179674e-06, | |
| "loss": 0.4814, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.0606891822389852, | |
| "grad_norm": 0.5074519719478089, | |
| "learning_rate": 2.6995207513517484e-06, | |
| "loss": 0.4835, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 2.06411794959712, | |
| "grad_norm": 0.46736698036035224, | |
| "learning_rate": 2.681822378198221e-06, | |
| "loss": 0.4821, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 2.0675467169552544, | |
| "grad_norm": 0.4677318714005963, | |
| "learning_rate": 2.6641609262361343e-06, | |
| "loss": 0.4825, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 2.070975484313389, | |
| "grad_norm": 0.47774880777961237, | |
| "learning_rate": 2.6465366767562162e-06, | |
| "loss": 0.4748, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 2.074404251671524, | |
| "grad_norm": 0.45924654536691306, | |
| "learning_rate": 2.628949910456663e-06, | |
| "loss": 0.4767, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 2.077833019029659, | |
| "grad_norm": 0.47391360563192103, | |
| "learning_rate": 2.611400907438685e-06, | |
| "loss": 0.4775, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 2.0812617863877936, | |
| "grad_norm": 0.49156177694024655, | |
| "learning_rate": 2.5938899472020545e-06, | |
| "loss": 0.4789, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 2.0846905537459284, | |
| "grad_norm": 0.4723037751330721, | |
| "learning_rate": 2.5764173086406306e-06, | |
| "loss": 0.4829, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 2.088119321104063, | |
| "grad_norm": 0.46114598626056696, | |
| "learning_rate": 2.558983270037937e-06, | |
| "loss": 0.4743, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 2.091548088462198, | |
| "grad_norm": 0.4718425543262081, | |
| "learning_rate": 2.5415881090627227e-06, | |
| "loss": 0.4785, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 2.0949768558203328, | |
| "grad_norm": 0.48700567548085505, | |
| "learning_rate": 2.5242321027645354e-06, | |
| "loss": 0.4731, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 2.098405623178467, | |
| "grad_norm": 0.47428262784263286, | |
| "learning_rate": 2.506915527569318e-06, | |
| "loss": 0.4805, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 2.101834390536602, | |
| "grad_norm": 0.482699000253055, | |
| "learning_rate": 2.4896386592749994e-06, | |
| "loss": 0.4809, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 2.1052631578947367, | |
| "grad_norm": 0.48854325554987943, | |
| "learning_rate": 2.472401773047107e-06, | |
| "loss": 0.4753, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 2.1086919252528715, | |
| "grad_norm": 0.5091234833700028, | |
| "learning_rate": 2.455205143414375e-06, | |
| "loss": 0.477, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 2.1121206926110063, | |
| "grad_norm": 0.474218431216323, | |
| "learning_rate": 2.438049044264382e-06, | |
| "loss": 0.4779, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 2.115549459969141, | |
| "grad_norm": 0.4762105524360361, | |
| "learning_rate": 2.4209337488391886e-06, | |
| "loss": 0.4763, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 2.118978227327276, | |
| "grad_norm": 0.4708922326438511, | |
| "learning_rate": 2.4038595297309712e-06, | |
| "loss": 0.4803, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 2.1224069946854107, | |
| "grad_norm": 0.48521692596512245, | |
| "learning_rate": 2.386826658877706e-06, | |
| "loss": 0.482, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 2.1258357620435455, | |
| "grad_norm": 0.494787221830382, | |
| "learning_rate": 2.3698354075588105e-06, | |
| "loss": 0.4803, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 2.12926452940168, | |
| "grad_norm": 0.48222263673783083, | |
| "learning_rate": 2.3528860463908375e-06, | |
| "loss": 0.472, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 2.1326932967598147, | |
| "grad_norm": 0.4824015117741457, | |
| "learning_rate": 2.3359788453231723e-06, | |
| "loss": 0.4797, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 2.1361220641179495, | |
| "grad_norm": 0.49605929290504214, | |
| "learning_rate": 2.319114073633712e-06, | |
| "loss": 0.4788, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 2.1395508314760843, | |
| "grad_norm": 0.474543321413953, | |
| "learning_rate": 2.3022919999245964e-06, | |
| "loss": 0.4773, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 2.142979598834219, | |
| "grad_norm": 0.501540938539211, | |
| "learning_rate": 2.285512892117921e-06, | |
| "loss": 0.4822, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 2.146408366192354, | |
| "grad_norm": 0.5038191337466582, | |
| "learning_rate": 2.2687770174514674e-06, | |
| "loss": 0.4791, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 2.1498371335504887, | |
| "grad_norm": 0.4648773455909595, | |
| "learning_rate": 2.2520846424744545e-06, | |
| "loss": 0.4828, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 2.1532659009086235, | |
| "grad_norm": 0.4620270257517909, | |
| "learning_rate": 2.23543603304329e-06, | |
| "loss": 0.479, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 2.1566946682667583, | |
| "grad_norm": 0.4950092604387871, | |
| "learning_rate": 2.218831454317336e-06, | |
| "loss": 0.4746, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 2.160123435624893, | |
| "grad_norm": 0.46350733624535256, | |
| "learning_rate": 2.20227117075468e-06, | |
| "loss": 0.4817, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 2.1635522029830274, | |
| "grad_norm": 0.4842429161440888, | |
| "learning_rate": 2.1857554461079356e-06, | |
| "loss": 0.4829, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 2.166980970341162, | |
| "grad_norm": 0.48989158037943475, | |
| "learning_rate": 2.1692845434200323e-06, | |
| "loss": 0.477, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 2.170409737699297, | |
| "grad_norm": 0.444636329908828, | |
| "learning_rate": 2.152858725020025e-06, | |
| "loss": 0.4762, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 2.173838505057432, | |
| "grad_norm": 0.46845184246079136, | |
| "learning_rate": 2.136478252518924e-06, | |
| "loss": 0.4867, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 2.1772672724155666, | |
| "grad_norm": 0.5158451108463605, | |
| "learning_rate": 2.1201433868055223e-06, | |
| "loss": 0.4773, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 2.1806960397737014, | |
| "grad_norm": 0.484433856048475, | |
| "learning_rate": 2.103854388042243e-06, | |
| "loss": 0.4808, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 2.184124807131836, | |
| "grad_norm": 0.4689424096794033, | |
| "learning_rate": 2.08761151566099e-06, | |
| "loss": 0.4851, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 2.187553574489971, | |
| "grad_norm": 0.46853301342354137, | |
| "learning_rate": 2.071415028359026e-06, | |
| "loss": 0.4854, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 2.190982341848106, | |
| "grad_norm": 0.46534481693626095, | |
| "learning_rate": 2.0552651840948473e-06, | |
| "loss": 0.4789, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 2.19441110920624, | |
| "grad_norm": 0.47458944353516924, | |
| "learning_rate": 2.0391622400840665e-06, | |
| "loss": 0.4801, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 2.197839876564375, | |
| "grad_norm": 0.46593527956244135, | |
| "learning_rate": 2.023106452795339e-06, | |
| "loss": 0.4726, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 2.2012686439225098, | |
| "grad_norm": 0.4773652150154532, | |
| "learning_rate": 2.0070980779462513e-06, | |
| "loss": 0.4806, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 2.2046974112806446, | |
| "grad_norm": 0.4647727192391881, | |
| "learning_rate": 1.9911373704992617e-06, | |
| "loss": 0.4806, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 2.2081261786387794, | |
| "grad_norm": 0.4794479506202543, | |
| "learning_rate": 1.975224584657648e-06, | |
| "loss": 0.4791, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 2.211554945996914, | |
| "grad_norm": 0.49928694289640674, | |
| "learning_rate": 1.9593599738614382e-06, | |
| "loss": 0.4812, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 2.214983713355049, | |
| "grad_norm": 0.489572736256811, | |
| "learning_rate": 1.943543790783392e-06, | |
| "loss": 0.4822, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 2.2184124807131838, | |
| "grad_norm": 0.460165152565946, | |
| "learning_rate": 1.927776287324969e-06, | |
| "loss": 0.4845, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 2.2218412480713186, | |
| "grad_norm": 0.4741945403355791, | |
| "learning_rate": 1.9120577146123125e-06, | |
| "loss": 0.4753, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 2.225270015429453, | |
| "grad_norm": 0.4816051218900053, | |
| "learning_rate": 1.8963883229922603e-06, | |
| "loss": 0.4759, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 2.2286987827875877, | |
| "grad_norm": 0.4865842140542505, | |
| "learning_rate": 1.8807683620283496e-06, | |
| "loss": 0.4887, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 2.2321275501457225, | |
| "grad_norm": 0.49327525598298544, | |
| "learning_rate": 1.8651980804968466e-06, | |
| "loss": 0.4828, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 2.2355563175038573, | |
| "grad_norm": 0.4710718086720219, | |
| "learning_rate": 1.8496777263827775e-06, | |
| "loss": 0.4805, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 2.238985084861992, | |
| "grad_norm": 0.4790195567473799, | |
| "learning_rate": 1.834207546875988e-06, | |
| "loss": 0.4761, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 2.242413852220127, | |
| "grad_norm": 0.47138983012088903, | |
| "learning_rate": 1.8187877883672024e-06, | |
| "loss": 0.4809, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 2.2458426195782617, | |
| "grad_norm": 0.48900189445082154, | |
| "learning_rate": 1.8034186964440937e-06, | |
| "loss": 0.482, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 2.2492713869363965, | |
| "grad_norm": 0.4723460008561715, | |
| "learning_rate": 1.7881005158873826e-06, | |
| "loss": 0.478, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 2.2527001542945313, | |
| "grad_norm": 0.45431679994556473, | |
| "learning_rate": 1.7728334906669342e-06, | |
| "loss": 0.4828, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 2.2561289216526657, | |
| "grad_norm": 0.45761440455845676, | |
| "learning_rate": 1.757617863937865e-06, | |
| "loss": 0.4707, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 2.2595576890108005, | |
| "grad_norm": 0.48498068790945004, | |
| "learning_rate": 1.7424538780366884e-06, | |
| "loss": 0.4724, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 2.2629864563689353, | |
| "grad_norm": 0.45783948313254996, | |
| "learning_rate": 1.7273417744774323e-06, | |
| "loss": 0.4774, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 2.26641522372707, | |
| "grad_norm": 0.46625983570633794, | |
| "learning_rate": 1.7122817939478104e-06, | |
| "loss": 0.4781, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 2.269843991085205, | |
| "grad_norm": 0.46981244219286544, | |
| "learning_rate": 1.6972741763053835e-06, | |
| "loss": 0.4771, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 2.2732727584433396, | |
| "grad_norm": 0.47749785750136037, | |
| "learning_rate": 1.6823191605737316e-06, | |
| "loss": 0.4812, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 2.2767015258014744, | |
| "grad_norm": 0.4781030507544295, | |
| "learning_rate": 1.6674169849386606e-06, | |
| "loss": 0.4745, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 2.2801302931596092, | |
| "grad_norm": 0.47642004107759217, | |
| "learning_rate": 1.6525678867443989e-06, | |
| "loss": 0.4782, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 2.283559060517744, | |
| "grad_norm": 0.4595236798861582, | |
| "learning_rate": 1.6377721024898214e-06, | |
| "loss": 0.4846, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 2.286987827875879, | |
| "grad_norm": 0.4607279352731881, | |
| "learning_rate": 1.623029867824678e-06, | |
| "loss": 0.4783, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 2.290416595234013, | |
| "grad_norm": 0.4857822760894794, | |
| "learning_rate": 1.608341417545849e-06, | |
| "loss": 0.4774, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 2.293845362592148, | |
| "grad_norm": 0.4577566161740047, | |
| "learning_rate": 1.593706985593599e-06, | |
| "loss": 0.4727, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 2.297274129950283, | |
| "grad_norm": 0.4573512037078888, | |
| "learning_rate": 1.5791268050478487e-06, | |
| "loss": 0.4746, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 2.3007028973084176, | |
| "grad_norm": 0.47439501092754294, | |
| "learning_rate": 1.5646011081244717e-06, | |
| "loss": 0.4865, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 2.3041316646665524, | |
| "grad_norm": 0.49446637460132087, | |
| "learning_rate": 1.5501301261715896e-06, | |
| "loss": 0.4751, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 2.307560432024687, | |
| "grad_norm": 0.4952360092448791, | |
| "learning_rate": 1.535714089665889e-06, | |
| "loss": 0.4815, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 2.310989199382822, | |
| "grad_norm": 0.4731157255344, | |
| "learning_rate": 1.5213532282089466e-06, | |
| "loss": 0.4792, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 2.314417966740957, | |
| "grad_norm": 0.4538368226200657, | |
| "learning_rate": 1.5070477705235803e-06, | |
| "loss": 0.4818, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 2.317846734099091, | |
| "grad_norm": 0.45299648856103863, | |
| "learning_rate": 1.4927979444502028e-06, | |
| "loss": 0.4761, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 2.321275501457226, | |
| "grad_norm": 0.48133972008005105, | |
| "learning_rate": 1.478603976943186e-06, | |
| "loss": 0.4796, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 2.3247042688153607, | |
| "grad_norm": 0.46559397796529, | |
| "learning_rate": 1.4644660940672628e-06, | |
| "loss": 0.4833, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 2.3281330361734955, | |
| "grad_norm": 0.46261472772428197, | |
| "learning_rate": 1.4503845209939083e-06, | |
| "loss": 0.4866, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 2.3315618035316303, | |
| "grad_norm": 0.48901577194657203, | |
| "learning_rate": 1.4363594819977606e-06, | |
| "loss": 0.4802, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 2.334990570889765, | |
| "grad_norm": 0.47026562258797505, | |
| "learning_rate": 1.4223912004530593e-06, | |
| "loss": 0.4842, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 2.3384193382479, | |
| "grad_norm": 0.46476370601090344, | |
| "learning_rate": 1.4084798988300684e-06, | |
| "loss": 0.4746, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 2.3418481056060347, | |
| "grad_norm": 0.4736014629302977, | |
| "learning_rate": 1.3946257986915486e-06, | |
| "loss": 0.4811, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 2.3452768729641695, | |
| "grad_norm": 0.4734489622060502, | |
| "learning_rate": 1.3808291206892232e-06, | |
| "loss": 0.4773, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 2.3487056403223043, | |
| "grad_norm": 0.4629066750832987, | |
| "learning_rate": 1.3670900845602585e-06, | |
| "loss": 0.4829, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 2.3521344076804387, | |
| "grad_norm": 0.4681227467302218, | |
| "learning_rate": 1.3534089091237757e-06, | |
| "loss": 0.476, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 2.3555631750385735, | |
| "grad_norm": 0.454191418129244, | |
| "learning_rate": 1.3397858122773544e-06, | |
| "loss": 0.4747, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 2.3589919423967083, | |
| "grad_norm": 0.4579452618693956, | |
| "learning_rate": 1.3262210109935719e-06, | |
| "loss": 0.4832, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 2.362420709754843, | |
| "grad_norm": 0.45074466888367914, | |
| "learning_rate": 1.312714721316537e-06, | |
| "loss": 0.4755, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 2.365849477112978, | |
| "grad_norm": 0.4723349908020595, | |
| "learning_rate": 1.2992671583584587e-06, | |
| "loss": 0.4862, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 2.3692782444711127, | |
| "grad_norm": 0.4708208262941508, | |
| "learning_rate": 1.2858785362962166e-06, | |
| "loss": 0.4777, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 2.3727070118292475, | |
| "grad_norm": 0.47375162486732203, | |
| "learning_rate": 1.2725490683679458e-06, | |
| "loss": 0.4729, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 2.3761357791873823, | |
| "grad_norm": 0.4678384719988464, | |
| "learning_rate": 1.2592789668696482e-06, | |
| "loss": 0.4787, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 2.379564546545517, | |
| "grad_norm": 0.4702212675672681, | |
| "learning_rate": 1.2460684431518055e-06, | |
| "loss": 0.4813, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 2.3829933139036514, | |
| "grad_norm": 0.4532398056344406, | |
| "learning_rate": 1.232917707616017e-06, | |
| "loss": 0.4735, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 2.3864220812617862, | |
| "grad_norm": 0.45797439884914765, | |
| "learning_rate": 1.2198269697116416e-06, | |
| "loss": 0.4778, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 2.389850848619921, | |
| "grad_norm": 0.4723176003130105, | |
| "learning_rate": 1.206796437932472e-06, | |
| "loss": 0.4774, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 2.393279615978056, | |
| "grad_norm": 0.46906434827216664, | |
| "learning_rate": 1.1938263198134087e-06, | |
| "loss": 0.4808, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 2.3967083833361906, | |
| "grad_norm": 0.4678613569139536, | |
| "learning_rate": 1.1809168219271488e-06, | |
| "loss": 0.4795, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 2.4001371506943254, | |
| "grad_norm": 0.47248188702171634, | |
| "learning_rate": 1.168068149880912e-06, | |
| "loss": 0.4781, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 2.4035659180524602, | |
| "grad_norm": 0.46791595503949973, | |
| "learning_rate": 1.1552805083131468e-06, | |
| "loss": 0.4786, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 2.406994685410595, | |
| "grad_norm": 0.45897414399547337, | |
| "learning_rate": 1.1425541008902852e-06, | |
| "loss": 0.4775, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 2.41042345276873, | |
| "grad_norm": 0.4610154781111884, | |
| "learning_rate": 1.1298891303034942e-06, | |
| "loss": 0.4823, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 2.4138522201268646, | |
| "grad_norm": 0.46406483424598494, | |
| "learning_rate": 1.1172857982654445e-06, | |
| "loss": 0.4761, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 2.417280987484999, | |
| "grad_norm": 0.4475457001447567, | |
| "learning_rate": 1.1047443055071033e-06, | |
| "loss": 0.4784, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 2.4207097548431338, | |
| "grad_norm": 0.46031864575357934, | |
| "learning_rate": 1.092264851774536e-06, | |
| "loss": 0.4741, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 2.4241385222012686, | |
| "grad_norm": 0.4642032024999473, | |
| "learning_rate": 1.0798476358257198e-06, | |
| "loss": 0.4763, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 2.4275672895594034, | |
| "grad_norm": 0.46462241022456946, | |
| "learning_rate": 1.067492855427385e-06, | |
| "loss": 0.4787, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 2.430996056917538, | |
| "grad_norm": 0.4769725916465071, | |
| "learning_rate": 1.0552007073518628e-06, | |
| "loss": 0.4826, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 2.434424824275673, | |
| "grad_norm": 0.46174302133938927, | |
| "learning_rate": 1.0429713873739505e-06, | |
| "loss": 0.4847, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 2.4378535916338078, | |
| "grad_norm": 0.45270102445912286, | |
| "learning_rate": 1.0308050902677907e-06, | |
| "loss": 0.4793, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 2.4412823589919426, | |
| "grad_norm": 0.46466350347691254, | |
| "learning_rate": 1.0187020098037759e-06, | |
| "loss": 0.4807, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 2.444711126350077, | |
| "grad_norm": 0.4615768268062226, | |
| "learning_rate": 1.0066623387454582e-06, | |
| "loss": 0.4811, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 2.4481398937082117, | |
| "grad_norm": 0.4544451248675266, | |
| "learning_rate": 9.946862688464753e-07, | |
| "loss": 0.4767, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 2.4515686610663465, | |
| "grad_norm": 0.4742062204892906, | |
| "learning_rate": 9.827739908475099e-07, | |
| "loss": 0.4776, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 2.4549974284244813, | |
| "grad_norm": 0.4672145082984867, | |
| "learning_rate": 9.709256944732343e-07, | |
| "loss": 0.4804, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 2.458426195782616, | |
| "grad_norm": 0.44899579410081497, | |
| "learning_rate": 9.591415684292975e-07, | |
| "loss": 0.4778, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 2.461854963140751, | |
| "grad_norm": 0.46844243005424735, | |
| "learning_rate": 9.474218003993275e-07, | |
| "loss": 0.4753, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 2.4652837304988857, | |
| "grad_norm": 0.4654191148385016, | |
| "learning_rate": 9.357665770419244e-07, | |
| "loss": 0.482, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 2.4687124978570205, | |
| "grad_norm": 0.45942731935750447, | |
| "learning_rate": 9.241760839877023e-07, | |
| "loss": 0.478, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 2.4721412652151553, | |
| "grad_norm": 0.4666227072724768, | |
| "learning_rate": 9.126505058363283e-07, | |
| "loss": 0.4753, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 2.47557003257329, | |
| "grad_norm": 0.4519254146818986, | |
| "learning_rate": 9.011900261535767e-07, | |
| "loss": 0.4783, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 2.4789987999314245, | |
| "grad_norm": 0.463358478501908, | |
| "learning_rate": 8.897948274684154e-07, | |
| "loss": 0.4739, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 2.4824275672895593, | |
| "grad_norm": 0.45511211330917706, | |
| "learning_rate": 8.784650912700909e-07, | |
| "loss": 0.4729, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 2.485856334647694, | |
| "grad_norm": 0.47432321264965754, | |
| "learning_rate": 8.672009980052442e-07, | |
| "loss": 0.4811, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 2.489285102005829, | |
| "grad_norm": 0.4673877749252423, | |
| "learning_rate": 8.560027270750276e-07, | |
| "loss": 0.4814, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 2.4927138693639637, | |
| "grad_norm": 0.4579464435065071, | |
| "learning_rate": 8.44870456832258e-07, | |
| "loss": 0.4806, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 2.4961426367220985, | |
| "grad_norm": 0.476376159360163, | |
| "learning_rate": 8.338043645785698e-07, | |
| "loss": 0.4823, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 2.4995714040802333, | |
| "grad_norm": 0.46673661354408297, | |
| "learning_rate": 8.228046265615908e-07, | |
| "loss": 0.4759, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 2.503000171438368, | |
| "grad_norm": 0.4689234252839364, | |
| "learning_rate": 8.118714179721404e-07, | |
| "loss": 0.4781, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 2.5064289387965024, | |
| "grad_norm": 0.457577939504501, | |
| "learning_rate": 8.010049129414333e-07, | |
| "loss": 0.4754, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 2.509857706154637, | |
| "grad_norm": 0.46767489107398075, | |
| "learning_rate": 7.902052845383112e-07, | |
| "loss": 0.4832, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 2.513286473512772, | |
| "grad_norm": 0.47806526391479376, | |
| "learning_rate": 7.794727047664807e-07, | |
| "loss": 0.4777, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 2.516715240870907, | |
| "grad_norm": 0.44998017534315693, | |
| "learning_rate": 7.6880734456178e-07, | |
| "loss": 0.4728, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 2.5201440082290416, | |
| "grad_norm": 0.4477060965035502, | |
| "learning_rate": 7.58209373789453e-07, | |
| "loss": 0.4734, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 2.5235727755871764, | |
| "grad_norm": 0.46217386088270807, | |
| "learning_rate": 7.476789612414414e-07, | |
| "loss": 0.4846, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 2.527001542945311, | |
| "grad_norm": 0.4563904666440247, | |
| "learning_rate": 7.372162746337059e-07, | |
| "loss": 0.4816, | |
| "step": 3685 | |
| }, | |
| { | |
| "epoch": 2.530430310303446, | |
| "grad_norm": 0.44520759689704725, | |
| "learning_rate": 7.268214806035423e-07, | |
| "loss": 0.4768, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 2.533859077661581, | |
| "grad_norm": 0.4542091051330333, | |
| "learning_rate": 7.164947447069343e-07, | |
| "loss": 0.4798, | |
| "step": 3695 | |
| }, | |
| { | |
| "epoch": 2.5372878450197156, | |
| "grad_norm": 0.45146271333603777, | |
| "learning_rate": 7.062362314159211e-07, | |
| "loss": 0.474, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 2.5407166123778504, | |
| "grad_norm": 0.4843676817399376, | |
| "learning_rate": 6.960461041159677e-07, | |
| "loss": 0.4826, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 2.5441453797359848, | |
| "grad_norm": 0.450384975906084, | |
| "learning_rate": 6.859245251033697e-07, | |
| "loss": 0.4768, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 2.5475741470941196, | |
| "grad_norm": 0.4460709873735881, | |
| "learning_rate": 6.758716555826683e-07, | |
| "loss": 0.4733, | |
| "step": 3715 | |
| }, | |
| { | |
| "epoch": 2.5510029144522544, | |
| "grad_norm": 0.4521937918515753, | |
| "learning_rate": 6.658876556640781e-07, | |
| "loss": 0.4769, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 2.554431681810389, | |
| "grad_norm": 0.4490013681026186, | |
| "learning_rate": 6.559726843609421e-07, | |
| "loss": 0.4805, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 2.557860449168524, | |
| "grad_norm": 0.45604478361558604, | |
| "learning_rate": 6.461268995871967e-07, | |
| "loss": 0.472, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 2.5612892165266588, | |
| "grad_norm": 0.4684730083561354, | |
| "learning_rate": 6.36350458154859e-07, | |
| "loss": 0.4823, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 2.5647179838847936, | |
| "grad_norm": 0.44901484120067714, | |
| "learning_rate": 6.266435157715222e-07, | |
| "loss": 0.4799, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 2.568146751242928, | |
| "grad_norm": 0.45631182390582853, | |
| "learning_rate": 6.170062270378862e-07, | |
| "loss": 0.4762, | |
| "step": 3745 | |
| }, | |
| { | |
| "epoch": 2.5715755186010627, | |
| "grad_norm": 0.4602387391697606, | |
| "learning_rate": 6.074387454452891e-07, | |
| "loss": 0.4713, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 2.5750042859591975, | |
| "grad_norm": 0.4491465188344547, | |
| "learning_rate": 5.979412233732578e-07, | |
| "loss": 0.4751, | |
| "step": 3755 | |
| }, | |
| { | |
| "epoch": 2.5784330533173323, | |
| "grad_norm": 0.46805364957632967, | |
| "learning_rate": 5.885138120870965e-07, | |
| "loss": 0.4789, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 2.581861820675467, | |
| "grad_norm": 0.4531331451215189, | |
| "learning_rate": 5.791566617354599e-07, | |
| "loss": 0.4754, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 2.585290588033602, | |
| "grad_norm": 0.4652091913696696, | |
| "learning_rate": 5.698699213479697e-07, | |
| "loss": 0.4763, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 2.5887193553917367, | |
| "grad_norm": 0.4542601341667562, | |
| "learning_rate": 5.606537388328481e-07, | |
| "loss": 0.4789, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 2.5921481227498715, | |
| "grad_norm": 0.4447156024624712, | |
| "learning_rate": 5.515082609745465e-07, | |
| "loss": 0.476, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 2.5955768901080063, | |
| "grad_norm": 0.45794390952267283, | |
| "learning_rate": 5.424336334314212e-07, | |
| "loss": 0.4781, | |
| "step": 3785 | |
| }, | |
| { | |
| "epoch": 2.599005657466141, | |
| "grad_norm": 0.45160699771932494, | |
| "learning_rate": 5.334300007334065e-07, | |
| "loss": 0.4771, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 2.602434424824276, | |
| "grad_norm": 0.4618426839968327, | |
| "learning_rate": 5.244975062797176e-07, | |
| "loss": 0.4728, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 2.6058631921824107, | |
| "grad_norm": 0.4486150363583273, | |
| "learning_rate": 5.156362923365587e-07, | |
| "loss": 0.4748, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 2.609291959540545, | |
| "grad_norm": 0.4545115387932617, | |
| "learning_rate": 5.068465000348666e-07, | |
| "loss": 0.4739, | |
| "step": 3805 | |
| }, | |
| { | |
| "epoch": 2.61272072689868, | |
| "grad_norm": 0.4623008940487107, | |
| "learning_rate": 4.981282693680584e-07, | |
| "loss": 0.4783, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 2.6161494942568146, | |
| "grad_norm": 0.4653422446290839, | |
| "learning_rate": 4.894817391897988e-07, | |
| "loss": 0.4854, | |
| "step": 3815 | |
| }, | |
| { | |
| "epoch": 2.6195782616149494, | |
| "grad_norm": 0.4515048388429379, | |
| "learning_rate": 4.80907047211796e-07, | |
| "loss": 0.4831, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 2.6230070289730842, | |
| "grad_norm": 0.4639284889413238, | |
| "learning_rate": 4.7240433000160233e-07, | |
| "loss": 0.4836, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 2.626435796331219, | |
| "grad_norm": 0.4465998513923373, | |
| "learning_rate": 4.639737229804403e-07, | |
| "loss": 0.4789, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 2.629864563689354, | |
| "grad_norm": 0.46955783511456173, | |
| "learning_rate": 4.556153604210484e-07, | |
| "loss": 0.4823, | |
| "step": 3835 | |
| }, | |
| { | |
| "epoch": 2.633293331047488, | |
| "grad_norm": 0.4509791198423544, | |
| "learning_rate": 4.473293754455399e-07, | |
| "loss": 0.477, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 2.636722098405623, | |
| "grad_norm": 0.46642581073567424, | |
| "learning_rate": 4.3911590002328476e-07, | |
| "loss": 0.4787, | |
| "step": 3845 | |
| }, | |
| { | |
| "epoch": 2.640150865763758, | |
| "grad_norm": 0.45567923757743767, | |
| "learning_rate": 4.3097506496880325e-07, | |
| "loss": 0.478, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 2.6435796331218926, | |
| "grad_norm": 0.46511398506896673, | |
| "learning_rate": 4.2290699993968897e-07, | |
| "loss": 0.4796, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 2.6470084004800274, | |
| "grad_norm": 0.4739348227964508, | |
| "learning_rate": 4.149118334345403e-07, | |
| "loss": 0.474, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 2.650437167838162, | |
| "grad_norm": 0.46126850845642486, | |
| "learning_rate": 4.0698969279091115e-07, | |
| "loss": 0.4786, | |
| "step": 3865 | |
| }, | |
| { | |
| "epoch": 2.653865935196297, | |
| "grad_norm": 0.4536599491788961, | |
| "learning_rate": 3.9914070418329123e-07, | |
| "loss": 0.477, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 2.657294702554432, | |
| "grad_norm": 0.4620488468718019, | |
| "learning_rate": 3.9136499262108716e-07, | |
| "loss": 0.4808, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 2.6607234699125666, | |
| "grad_norm": 0.4655535577302228, | |
| "learning_rate": 3.836626819466338e-07, | |
| "loss": 0.4809, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 2.6641522372707014, | |
| "grad_norm": 0.5048993299270467, | |
| "learning_rate": 3.760338948332293e-07, | |
| "loss": 0.4731, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 2.667581004628836, | |
| "grad_norm": 0.47154869975353264, | |
| "learning_rate": 3.684787527831707e-07, | |
| "loss": 0.477, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 2.6710097719869705, | |
| "grad_norm": 0.4642713924482236, | |
| "learning_rate": 3.6099737612582455e-07, | |
| "loss": 0.475, | |
| "step": 3895 | |
| }, | |
| { | |
| "epoch": 2.6744385393451053, | |
| "grad_norm": 0.4483908444332171, | |
| "learning_rate": 3.53589884015712e-07, | |
| "loss": 0.4753, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 2.67786730670324, | |
| "grad_norm": 0.4652782765678345, | |
| "learning_rate": 3.462563944306047e-07, | |
| "loss": 0.4816, | |
| "step": 3905 | |
| }, | |
| { | |
| "epoch": 2.681296074061375, | |
| "grad_norm": 0.45331932775420813, | |
| "learning_rate": 3.3899702416965166e-07, | |
| "loss": 0.4777, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 2.6847248414195097, | |
| "grad_norm": 0.4412624964963365, | |
| "learning_rate": 3.3181188885151706e-07, | |
| "loss": 0.4763, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 2.6881536087776445, | |
| "grad_norm": 0.45014302824757474, | |
| "learning_rate": 3.247011029125391e-07, | |
| "loss": 0.4786, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 2.6915823761357793, | |
| "grad_norm": 0.4631051719702464, | |
| "learning_rate": 3.1766477960490485e-07, | |
| "loss": 0.4846, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 2.6950111434939137, | |
| "grad_norm": 0.4581555189827555, | |
| "learning_rate": 3.1070303099485055e-07, | |
| "loss": 0.484, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 2.6984399108520485, | |
| "grad_norm": 0.45350853465967644, | |
| "learning_rate": 3.038159679608754e-07, | |
| "loss": 0.4789, | |
| "step": 3935 | |
| }, | |
| { | |
| "epoch": 2.7018686782101833, | |
| "grad_norm": 0.4546945454352565, | |
| "learning_rate": 2.9700370019197287e-07, | |
| "loss": 0.4763, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 2.705297445568318, | |
| "grad_norm": 0.449234094604194, | |
| "learning_rate": 2.9026633618588706e-07, | |
| "loss": 0.4759, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 2.708726212926453, | |
| "grad_norm": 0.45160809662086393, | |
| "learning_rate": 2.8360398324738415e-07, | |
| "loss": 0.4716, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 2.7121549802845877, | |
| "grad_norm": 0.4482757694655283, | |
| "learning_rate": 2.7701674748654206e-07, | |
| "loss": 0.4774, | |
| "step": 3955 | |
| }, | |
| { | |
| "epoch": 2.7155837476427225, | |
| "grad_norm": 0.4605988211239266, | |
| "learning_rate": 2.7050473381706186e-07, | |
| "loss": 0.4796, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 2.7190125150008573, | |
| "grad_norm": 0.4542079439617735, | |
| "learning_rate": 2.640680459545941e-07, | |
| "loss": 0.4814, | |
| "step": 3965 | |
| }, | |
| { | |
| "epoch": 2.722441282358992, | |
| "grad_norm": 0.47170501988492425, | |
| "learning_rate": 2.577067864150906e-07, | |
| "loss": 0.477, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 2.725870049717127, | |
| "grad_norm": 0.45766219103133904, | |
| "learning_rate": 2.514210565131708e-07, | |
| "loss": 0.479, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 2.7292988170752617, | |
| "grad_norm": 0.46091999743971984, | |
| "learning_rate": 2.452109563605065e-07, | |
| "loss": 0.4795, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 2.732727584433396, | |
| "grad_norm": 0.46053964874575465, | |
| "learning_rate": 2.3907658486422734e-07, | |
| "loss": 0.4816, | |
| "step": 3985 | |
| }, | |
| { | |
| "epoch": 2.736156351791531, | |
| "grad_norm": 0.44585333829135504, | |
| "learning_rate": 2.330180397253473e-07, | |
| "loss": 0.4712, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 2.7395851191496656, | |
| "grad_norm": 0.45616604492168555, | |
| "learning_rate": 2.2703541743720913e-07, | |
| "loss": 0.4765, | |
| "step": 3995 | |
| }, | |
| { | |
| "epoch": 2.7430138865078004, | |
| "grad_norm": 0.458862963479115, | |
| "learning_rate": 2.2112881328394287e-07, | |
| "loss": 0.4734, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 2.7464426538659352, | |
| "grad_norm": 0.4564004005454542, | |
| "learning_rate": 2.152983213389559e-07, | |
| "loss": 0.4817, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 2.74987142122407, | |
| "grad_norm": 0.4527978346632276, | |
| "learning_rate": 2.0954403446342753e-07, | |
| "loss": 0.4778, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 2.753300188582205, | |
| "grad_norm": 0.44674470885527423, | |
| "learning_rate": 2.0386604430483237e-07, | |
| "loss": 0.4753, | |
| "step": 4015 | |
| }, | |
| { | |
| "epoch": 2.756728955940339, | |
| "grad_norm": 0.4573693361590737, | |
| "learning_rate": 1.9826444129548317e-07, | |
| "loss": 0.4829, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 2.760157723298474, | |
| "grad_norm": 0.46202280759285275, | |
| "learning_rate": 1.927393146510881e-07, | |
| "loss": 0.4794, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 2.7635864906566088, | |
| "grad_norm": 0.44872435639531505, | |
| "learning_rate": 1.8729075236932903e-07, | |
| "loss": 0.4712, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 2.7670152580147436, | |
| "grad_norm": 0.4562789796076479, | |
| "learning_rate": 1.8191884122846226e-07, | |
| "loss": 0.4775, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 2.7704440253728784, | |
| "grad_norm": 0.4585752798754004, | |
| "learning_rate": 1.7662366678593502e-07, | |
| "loss": 0.4803, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 2.773872792731013, | |
| "grad_norm": 0.44460297881292643, | |
| "learning_rate": 1.7140531337702383e-07, | |
| "loss": 0.4747, | |
| "step": 4045 | |
| }, | |
| { | |
| "epoch": 2.777301560089148, | |
| "grad_norm": 0.4425263511378165, | |
| "learning_rate": 1.6626386411348783e-07, | |
| "loss": 0.4715, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 2.7807303274472828, | |
| "grad_norm": 0.448778429615437, | |
| "learning_rate": 1.6119940088225317e-07, | |
| "loss": 0.4821, | |
| "step": 4055 | |
| }, | |
| { | |
| "epoch": 2.7841590948054176, | |
| "grad_norm": 0.4454280991695363, | |
| "learning_rate": 1.56212004344099e-07, | |
| "loss": 0.4725, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 2.7875878621635524, | |
| "grad_norm": 0.45389256280014806, | |
| "learning_rate": 1.5130175393237744e-07, | |
| "loss": 0.482, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 2.791016629521687, | |
| "grad_norm": 0.44797422922275487, | |
| "learning_rate": 1.4646872785175182e-07, | |
| "loss": 0.4736, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 2.794445396879822, | |
| "grad_norm": 0.4500167947023535, | |
| "learning_rate": 1.417130030769448e-07, | |
| "loss": 0.4748, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 2.7978741642379563, | |
| "grad_norm": 0.4384121757610316, | |
| "learning_rate": 1.3703465535151505e-07, | |
| "loss": 0.4787, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 2.801302931596091, | |
| "grad_norm": 0.4542996969309403, | |
| "learning_rate": 1.3243375918665314e-07, | |
| "loss": 0.4757, | |
| "step": 4085 | |
| }, | |
| { | |
| "epoch": 2.804731698954226, | |
| "grad_norm": 0.44313787892697587, | |
| "learning_rate": 1.2791038785999243e-07, | |
| "loss": 0.4723, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 2.8081604663123607, | |
| "grad_norm": 0.45221326925330546, | |
| "learning_rate": 1.2346461341443962e-07, | |
| "loss": 0.4739, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 2.8115892336704955, | |
| "grad_norm": 0.4493204806606999, | |
| "learning_rate": 1.1909650665703265e-07, | |
| "loss": 0.476, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 2.8150180010286303, | |
| "grad_norm": 0.4511251142367096, | |
| "learning_rate": 1.1480613715781074e-07, | |
| "loss": 0.4712, | |
| "step": 4105 | |
| }, | |
| { | |
| "epoch": 2.818446768386765, | |
| "grad_norm": 0.44576223137927595, | |
| "learning_rate": 1.1059357324870456e-07, | |
| "loss": 0.4804, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 2.8218755357448995, | |
| "grad_norm": 0.45162764316225673, | |
| "learning_rate": 1.064588820224488e-07, | |
| "loss": 0.4756, | |
| "step": 4115 | |
| }, | |
| { | |
| "epoch": 2.8253043031030343, | |
| "grad_norm": 0.45919178354925205, | |
| "learning_rate": 1.024021293315175e-07, | |
| "loss": 0.4789, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 2.828733070461169, | |
| "grad_norm": 0.45178753105354696, | |
| "learning_rate": 9.842337978706707e-08, | |
| "loss": 0.4712, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 2.832161837819304, | |
| "grad_norm": 0.44369641899607265, | |
| "learning_rate": 9.452269675791603e-08, | |
| "loss": 0.4693, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 2.8355906051774387, | |
| "grad_norm": 0.4397395162890691, | |
| "learning_rate": 9.070014236952973e-08, | |
| "loss": 0.4725, | |
| "step": 4135 | |
| }, | |
| { | |
| "epoch": 2.8390193725355735, | |
| "grad_norm": 0.4515355683540246, | |
| "learning_rate": 8.69557775030344e-08, | |
| "loss": 0.4773, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 2.8424481398937083, | |
| "grad_norm": 0.4403255731977004, | |
| "learning_rate": 8.328966179424413e-08, | |
| "loss": 0.4733, | |
| "step": 4145 | |
| }, | |
| { | |
| "epoch": 2.845876907251843, | |
| "grad_norm": 0.4543200394332973, | |
| "learning_rate": 7.970185363271432e-08, | |
| "loss": 0.481, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 2.849305674609978, | |
| "grad_norm": 0.4564283136704236, | |
| "learning_rate": 7.619241016081081e-08, | |
| "loss": 0.4761, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 2.8527344419681127, | |
| "grad_norm": 0.44588624487284356, | |
| "learning_rate": 7.276138727279669e-08, | |
| "loss": 0.4797, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 2.8561632093262475, | |
| "grad_norm": 0.45198814901899365, | |
| "learning_rate": 6.940883961394917e-08, | |
| "loss": 0.4798, | |
| "step": 4165 | |
| }, | |
| { | |
| "epoch": 2.859591976684382, | |
| "grad_norm": 0.4501783006120383, | |
| "learning_rate": 6.613482057968023e-08, | |
| "loss": 0.4732, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 2.8630207440425166, | |
| "grad_norm": 0.4373982769658886, | |
| "learning_rate": 6.293938231469398e-08, | |
| "loss": 0.4736, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 2.8664495114006514, | |
| "grad_norm": 0.4468552193790084, | |
| "learning_rate": 5.982257571215178e-08, | |
| "loss": 0.4775, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 2.869878278758786, | |
| "grad_norm": 0.44830321173439996, | |
| "learning_rate": 5.6784450412862293e-08, | |
| "loss": 0.4721, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 2.873307046116921, | |
| "grad_norm": 0.44597562076766367, | |
| "learning_rate": 5.382505480449274e-08, | |
| "loss": 0.4757, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 2.876735813475056, | |
| "grad_norm": 0.46198888054755305, | |
| "learning_rate": 5.094443602079613e-08, | |
| "loss": 0.4769, | |
| "step": 4195 | |
| }, | |
| { | |
| "epoch": 2.8801645808331906, | |
| "grad_norm": 0.4481590814268269, | |
| "learning_rate": 4.814263994086077e-08, | |
| "loss": 0.4745, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 2.883593348191325, | |
| "grad_norm": 0.44268948043488404, | |
| "learning_rate": 4.541971118838196e-08, | |
| "loss": 0.4802, | |
| "step": 4205 | |
| }, | |
| { | |
| "epoch": 2.8870221155494598, | |
| "grad_norm": 0.4450782983770214, | |
| "learning_rate": 4.2775693130948094e-08, | |
| "loss": 0.4786, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 2.8904508829075946, | |
| "grad_norm": 0.4605099335234408, | |
| "learning_rate": 4.021062787935181e-08, | |
| "loss": 0.4784, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 2.8938796502657294, | |
| "grad_norm": 0.4481715420899884, | |
| "learning_rate": 3.772455628691829e-08, | |
| "loss": 0.4704, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 2.897308417623864, | |
| "grad_norm": 0.4646866287413105, | |
| "learning_rate": 3.531751794885574e-08, | |
| "loss": 0.4799, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 2.900737184981999, | |
| "grad_norm": 0.4565890986941479, | |
| "learning_rate": 3.2989551201624836e-08, | |
| "loss": 0.4825, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 2.9041659523401337, | |
| "grad_norm": 0.44645402505000176, | |
| "learning_rate": 3.0740693122325304e-08, | |
| "loss": 0.4735, | |
| "step": 4235 | |
| }, | |
| { | |
| "epoch": 2.9075947196982685, | |
| "grad_norm": 0.4507869739723372, | |
| "learning_rate": 2.857097952810972e-08, | |
| "loss": 0.4862, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 2.9110234870564033, | |
| "grad_norm": 0.4501912072102517, | |
| "learning_rate": 2.6480444975610086e-08, | |
| "loss": 0.4789, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 2.914452254414538, | |
| "grad_norm": 0.447987603121036, | |
| "learning_rate": 2.4469122760388264e-08, | |
| "loss": 0.4778, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 2.917881021772673, | |
| "grad_norm": 0.44971834009936357, | |
| "learning_rate": 2.2537044916405292e-08, | |
| "loss": 0.4772, | |
| "step": 4255 | |
| }, | |
| { | |
| "epoch": 2.9213097891308077, | |
| "grad_norm": 0.4495077806550815, | |
| "learning_rate": 2.0684242215511797e-08, | |
| "loss": 0.4749, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 2.924738556488942, | |
| "grad_norm": 0.44694368287876113, | |
| "learning_rate": 1.8910744166958926e-08, | |
| "loss": 0.4801, | |
| "step": 4265 | |
| }, | |
| { | |
| "epoch": 2.928167323847077, | |
| "grad_norm": 0.4497877428185916, | |
| "learning_rate": 1.7216579016925415e-08, | |
| "loss": 0.484, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 2.9315960912052117, | |
| "grad_norm": 0.44920658823268794, | |
| "learning_rate": 1.5601773748070147e-08, | |
| "loss": 0.4814, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 2.9350248585633465, | |
| "grad_norm": 0.4438689776804719, | |
| "learning_rate": 1.4066354079101396e-08, | |
| "loss": 0.4792, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 2.9384536259214813, | |
| "grad_norm": 0.446097901399973, | |
| "learning_rate": 1.2610344464367707e-08, | |
| "loss": 0.478, | |
| "step": 4285 | |
| }, | |
| { | |
| "epoch": 2.941882393279616, | |
| "grad_norm": 0.447115729483812, | |
| "learning_rate": 1.1233768093468766e-08, | |
| "loss": 0.4757, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 2.945311160637751, | |
| "grad_norm": 0.4481899023438092, | |
| "learning_rate": 9.936646890884582e-09, | |
| "loss": 0.4739, | |
| "step": 4295 | |
| }, | |
| { | |
| "epoch": 2.9487399279958852, | |
| "grad_norm": 0.4506621904879036, | |
| "learning_rate": 8.719001515627434e-09, | |
| "loss": 0.479, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 2.95216869535402, | |
| "grad_norm": 0.4533955216869492, | |
| "learning_rate": 7.58085136091269e-09, | |
| "loss": 0.4817, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 2.955597462712155, | |
| "grad_norm": 0.43437042453552366, | |
| "learning_rate": 6.5222145538501595e-09, | |
| "loss": 0.4821, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 2.9590262300702896, | |
| "grad_norm": 0.45340023997429035, | |
| "learning_rate": 5.543107955154892e-09, | |
| "loss": 0.4722, | |
| "step": 4315 | |
| }, | |
| { | |
| "epoch": 2.9624549974284244, | |
| "grad_norm": 0.4547732433450872, | |
| "learning_rate": 4.643547158878492e-09, | |
| "loss": 0.4783, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 2.9658837647865592, | |
| "grad_norm": 0.4459015086328452, | |
| "learning_rate": 3.8235464921621e-09, | |
| "loss": 0.4759, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 2.969312532144694, | |
| "grad_norm": 0.43573098371648133, | |
| "learning_rate": 3.0831190150054646e-09, | |
| "loss": 0.4793, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 2.972741299502829, | |
| "grad_norm": 0.4558549928949452, | |
| "learning_rate": 2.422276520062661e-09, | |
| "loss": 0.4754, | |
| "step": 4335 | |
| }, | |
| { | |
| "epoch": 2.9761700668609636, | |
| "grad_norm": 0.45059494788354154, | |
| "learning_rate": 1.8410295324505778e-09, | |
| "loss": 0.4808, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 2.9795988342190984, | |
| "grad_norm": 0.4527487921892312, | |
| "learning_rate": 1.3393873095846055e-09, | |
| "loss": 0.4771, | |
| "step": 4345 | |
| }, | |
| { | |
| "epoch": 2.9830276015772332, | |
| "grad_norm": 0.44727140352439004, | |
| "learning_rate": 9.173578410281992e-10, | |
| "loss": 0.4793, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 2.9864563689353676, | |
| "grad_norm": 0.44819706968866224, | |
| "learning_rate": 5.749478483679794e-10, | |
| "loss": 0.4785, | |
| "step": 4355 | |
| }, | |
| { | |
| "epoch": 2.9898851362935024, | |
| "grad_norm": 0.4540122997268082, | |
| "learning_rate": 3.1216278510493027e-10, | |
| "loss": 0.4768, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 2.993313903651637, | |
| "grad_norm": 0.44168345357587324, | |
| "learning_rate": 1.2900683656891232e-10, | |
| "loss": 0.4688, | |
| "step": 4365 | |
| }, | |
| { | |
| "epoch": 2.996742671009772, | |
| "grad_norm": 0.45853813294171786, | |
| "learning_rate": 2.548291985149387e-11, | |
| "loss": 0.4775, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 2.9994856848962796, | |
| "eval_loss": 0.5953422784805298, | |
| "eval_runtime": 204.761, | |
| "eval_samples_per_second": 47.993, | |
| "eval_steps_per_second": 1.504, | |
| "step": 4374 | |
| }, | |
| { | |
| "epoch": 2.9994856848962796, | |
| "step": 4374, | |
| "total_flos": 3663097076121600.0, | |
| "train_loss": 0.5645236694000604, | |
| "train_runtime": 46689.0597, | |
| "train_samples_per_second": 11.993, | |
| "train_steps_per_second": 0.094 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 4374, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": false, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3663097076121600.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |