{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9215686274509802, "eval_steps": 500, "global_step": 98000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.000196078431372549, "grad_norm": 2.6763179302215576, "learning_rate": 1.7647058823529412e-09, "loss": 0.1724, "step": 10 }, { "epoch": 0.000392156862745098, "grad_norm": 4.8292555809021, "learning_rate": 3.725490196078431e-09, "loss": 0.1903, "step": 20 }, { "epoch": 0.000588235294117647, "grad_norm": 1.628282070159912, "learning_rate": 5.686274509803921e-09, "loss": 0.1971, "step": 30 }, { "epoch": 0.000784313725490196, "grad_norm": 1.3956987857818604, "learning_rate": 7.647058823529412e-09, "loss": 0.2181, "step": 40 }, { "epoch": 0.000980392156862745, "grad_norm": 2.308335542678833, "learning_rate": 9.607843137254902e-09, "loss": 0.2213, "step": 50 }, { "epoch": 0.001176470588235294, "grad_norm": 1.961266040802002, "learning_rate": 1.1568627450980391e-08, "loss": 0.2309, "step": 60 }, { "epoch": 0.0013725490196078432, "grad_norm": 4.421730041503906, "learning_rate": 1.3529411764705881e-08, "loss": 0.1989, "step": 70 }, { "epoch": 0.001568627450980392, "grad_norm": 2.598287582397461, "learning_rate": 1.549019607843137e-08, "loss": 0.1962, "step": 80 }, { "epoch": 0.0017647058823529412, "grad_norm": 2.9455976486206055, "learning_rate": 1.7450980392156864e-08, "loss": 0.2122, "step": 90 }, { "epoch": 0.00196078431372549, "grad_norm": 2.911093235015869, "learning_rate": 1.9411764705882354e-08, "loss": 0.1884, "step": 100 }, { "epoch": 0.002156862745098039, "grad_norm": 2.1638474464416504, "learning_rate": 2.1372549019607844e-08, "loss": 0.2043, "step": 110 }, { "epoch": 0.002352941176470588, "grad_norm": 3.4772584438323975, "learning_rate": 2.3333333333333334e-08, "loss": 0.2043, "step": 120 }, { "epoch": 0.0025490196078431374, "grad_norm": 1.9596714973449707, "learning_rate": 2.5294117647058824e-08, "loss": 0.1957, "step": 130 }, { "epoch": 0.0027450980392156863, "grad_norm": 4.894094467163086, "learning_rate": 2.7254901960784314e-08, "loss": 0.2196, "step": 140 }, { "epoch": 0.0029411764705882353, "grad_norm": 2.2870545387268066, "learning_rate": 2.9215686274509803e-08, "loss": 0.1756, "step": 150 }, { "epoch": 0.003137254901960784, "grad_norm": 2.5873420238494873, "learning_rate": 3.1176470588235297e-08, "loss": 0.1925, "step": 160 }, { "epoch": 0.0033333333333333335, "grad_norm": 1.4312834739685059, "learning_rate": 3.313725490196078e-08, "loss": 0.171, "step": 170 }, { "epoch": 0.0035294117647058825, "grad_norm": 2.34684419631958, "learning_rate": 3.509803921568627e-08, "loss": 0.2081, "step": 180 }, { "epoch": 0.0037254901960784314, "grad_norm": 2.4413387775421143, "learning_rate": 3.705882352941176e-08, "loss": 0.1805, "step": 190 }, { "epoch": 0.00392156862745098, "grad_norm": 2.489954710006714, "learning_rate": 3.901960784313725e-08, "loss": 0.2292, "step": 200 }, { "epoch": 0.00411764705882353, "grad_norm": 2.6573326587677, "learning_rate": 4.098039215686274e-08, "loss": 0.2044, "step": 210 }, { "epoch": 0.004313725490196078, "grad_norm": 3.6163153648376465, "learning_rate": 4.294117647058823e-08, "loss": 0.238, "step": 220 }, { "epoch": 0.0045098039215686276, "grad_norm": 2.298286199569702, "learning_rate": 4.490196078431372e-08, "loss": 0.1936, "step": 230 }, { "epoch": 0.004705882352941176, "grad_norm": 2.6092772483825684, "learning_rate": 4.686274509803921e-08, "loss": 0.1798, "step": 240 }, { "epoch": 0.004901960784313725, "grad_norm": 2.010282278060913, "learning_rate": 4.88235294117647e-08, "loss": 0.2137, "step": 250 }, { "epoch": 0.005098039215686275, "grad_norm": 4.156600475311279, "learning_rate": 5.078431372549019e-08, "loss": 0.1921, "step": 260 }, { "epoch": 0.005294117647058823, "grad_norm": 2.0165882110595703, "learning_rate": 5.274509803921568e-08, "loss": 0.1746, "step": 270 }, { "epoch": 0.005490196078431373, "grad_norm": 3.4648563861846924, "learning_rate": 5.470588235294117e-08, "loss": 0.176, "step": 280 }, { "epoch": 0.005686274509803921, "grad_norm": 3.827937126159668, "learning_rate": 5.666666666666666e-08, "loss": 0.2012, "step": 290 }, { "epoch": 0.0058823529411764705, "grad_norm": 1.5170515775680542, "learning_rate": 5.8627450980392154e-08, "loss": 0.1574, "step": 300 }, { "epoch": 0.00607843137254902, "grad_norm": 1.257917881011963, "learning_rate": 6.058823529411764e-08, "loss": 0.2041, "step": 310 }, { "epoch": 0.006274509803921568, "grad_norm": 5.107327938079834, "learning_rate": 6.254901960784313e-08, "loss": 0.2278, "step": 320 }, { "epoch": 0.006470588235294118, "grad_norm": 3.3016722202301025, "learning_rate": 6.450980392156863e-08, "loss": 0.2102, "step": 330 }, { "epoch": 0.006666666666666667, "grad_norm": 2.578871726989746, "learning_rate": 6.647058823529411e-08, "loss": 0.2121, "step": 340 }, { "epoch": 0.006862745098039216, "grad_norm": 2.580610990524292, "learning_rate": 6.843137254901961e-08, "loss": 0.2076, "step": 350 }, { "epoch": 0.007058823529411765, "grad_norm": 3.645918846130371, "learning_rate": 7.039215686274509e-08, "loss": 0.186, "step": 360 }, { "epoch": 0.0072549019607843135, "grad_norm": 1.7273955345153809, "learning_rate": 7.235294117647059e-08, "loss": 0.1976, "step": 370 }, { "epoch": 0.007450980392156863, "grad_norm": 2.6824705600738525, "learning_rate": 7.431372549019607e-08, "loss": 0.1736, "step": 380 }, { "epoch": 0.007647058823529412, "grad_norm": 3.5790247917175293, "learning_rate": 7.627450980392157e-08, "loss": 0.218, "step": 390 }, { "epoch": 0.00784313725490196, "grad_norm": 7.130075931549072, "learning_rate": 7.823529411764705e-08, "loss": 0.1859, "step": 400 }, { "epoch": 0.00803921568627451, "grad_norm": 2.84069561958313, "learning_rate": 8.019607843137255e-08, "loss": 0.2269, "step": 410 }, { "epoch": 0.00823529411764706, "grad_norm": 1.5767112970352173, "learning_rate": 8.215686274509803e-08, "loss": 0.1749, "step": 420 }, { "epoch": 0.008431372549019607, "grad_norm": 3.0915749073028564, "learning_rate": 8.411764705882353e-08, "loss": 0.1757, "step": 430 }, { "epoch": 0.008627450980392156, "grad_norm": 2.701918601989746, "learning_rate": 8.607843137254902e-08, "loss": 0.1785, "step": 440 }, { "epoch": 0.008823529411764706, "grad_norm": 2.840264081954956, "learning_rate": 8.803921568627451e-08, "loss": 0.2224, "step": 450 }, { "epoch": 0.009019607843137255, "grad_norm": 2.7778820991516113, "learning_rate": 9e-08, "loss": 0.1857, "step": 460 }, { "epoch": 0.009215686274509804, "grad_norm": 2.0893542766571045, "learning_rate": 9.196078431372549e-08, "loss": 0.2149, "step": 470 }, { "epoch": 0.009411764705882352, "grad_norm": 3.8348495960235596, "learning_rate": 9.392156862745098e-08, "loss": 0.2246, "step": 480 }, { "epoch": 0.009607843137254901, "grad_norm": 2.520922899246216, "learning_rate": 9.588235294117647e-08, "loss": 0.1918, "step": 490 }, { "epoch": 0.00980392156862745, "grad_norm": 2.9252495765686035, "learning_rate": 9.784313725490196e-08, "loss": 0.2553, "step": 500 }, { "epoch": 0.01, "grad_norm": 3.035047769546509, "learning_rate": 9.980392156862745e-08, "loss": 0.1643, "step": 510 }, { "epoch": 0.01019607843137255, "grad_norm": 1.9366408586502075, "learning_rate": 1.0176470588235294e-07, "loss": 0.2233, "step": 520 }, { "epoch": 0.010392156862745097, "grad_norm": 3.2076892852783203, "learning_rate": 1.0372549019607843e-07, "loss": 0.2097, "step": 530 }, { "epoch": 0.010588235294117647, "grad_norm": 2.9406332969665527, "learning_rate": 1.0568627450980392e-07, "loss": 0.2207, "step": 540 }, { "epoch": 0.010784313725490196, "grad_norm": 4.685433387756348, "learning_rate": 1.0764705882352941e-07, "loss": 0.2226, "step": 550 }, { "epoch": 0.010980392156862745, "grad_norm": 3.0491065979003906, "learning_rate": 1.096078431372549e-07, "loss": 0.2191, "step": 560 }, { "epoch": 0.011176470588235295, "grad_norm": 1.4955699443817139, "learning_rate": 1.1156862745098039e-07, "loss": 0.2045, "step": 570 }, { "epoch": 0.011372549019607842, "grad_norm": 2.8227932453155518, "learning_rate": 1.1352941176470588e-07, "loss": 0.2304, "step": 580 }, { "epoch": 0.011568627450980392, "grad_norm": 2.0291354656219482, "learning_rate": 1.1549019607843137e-07, "loss": 0.2323, "step": 590 }, { "epoch": 0.011764705882352941, "grad_norm": 2.433321952819824, "learning_rate": 1.1745098039215686e-07, "loss": 0.2098, "step": 600 }, { "epoch": 0.01196078431372549, "grad_norm": 1.0978672504425049, "learning_rate": 1.1941176470588235e-07, "loss": 0.2222, "step": 610 }, { "epoch": 0.01215686274509804, "grad_norm": 2.8603363037109375, "learning_rate": 1.2137254901960782e-07, "loss": 0.1828, "step": 620 }, { "epoch": 0.012352941176470587, "grad_norm": 1.4105048179626465, "learning_rate": 1.2333333333333333e-07, "loss": 0.2093, "step": 630 }, { "epoch": 0.012549019607843137, "grad_norm": 3.6006417274475098, "learning_rate": 1.2529411764705883e-07, "loss": 0.1821, "step": 640 }, { "epoch": 0.012745098039215686, "grad_norm": 2.3522982597351074, "learning_rate": 1.272549019607843e-07, "loss": 0.2065, "step": 650 }, { "epoch": 0.012941176470588235, "grad_norm": 4.552095413208008, "learning_rate": 1.2921568627450978e-07, "loss": 0.1963, "step": 660 }, { "epoch": 0.013137254901960785, "grad_norm": 3.105801820755005, "learning_rate": 1.3117647058823529e-07, "loss": 0.2059, "step": 670 }, { "epoch": 0.013333333333333334, "grad_norm": 1.8590614795684814, "learning_rate": 1.331372549019608e-07, "loss": 0.1862, "step": 680 }, { "epoch": 0.013529411764705882, "grad_norm": 2.33857798576355, "learning_rate": 1.3509803921568627e-07, "loss": 0.1937, "step": 690 }, { "epoch": 0.013725490196078431, "grad_norm": 3.4690136909484863, "learning_rate": 1.3705882352941174e-07, "loss": 0.2526, "step": 700 }, { "epoch": 0.01392156862745098, "grad_norm": 4.926452159881592, "learning_rate": 1.3901960784313724e-07, "loss": 0.2148, "step": 710 }, { "epoch": 0.01411764705882353, "grad_norm": 4.828888893127441, "learning_rate": 1.4098039215686275e-07, "loss": 0.1974, "step": 720 }, { "epoch": 0.01431372549019608, "grad_norm": 3.172999858856201, "learning_rate": 1.4294117647058822e-07, "loss": 0.2205, "step": 730 }, { "epoch": 0.014509803921568627, "grad_norm": 2.7665116786956787, "learning_rate": 1.4490196078431373e-07, "loss": 0.1772, "step": 740 }, { "epoch": 0.014705882352941176, "grad_norm": 2.2016220092773438, "learning_rate": 1.468627450980392e-07, "loss": 0.2141, "step": 750 }, { "epoch": 0.014901960784313726, "grad_norm": 7.375670909881592, "learning_rate": 1.488235294117647e-07, "loss": 0.1848, "step": 760 }, { "epoch": 0.015098039215686275, "grad_norm": 4.045589447021484, "learning_rate": 1.5078431372549018e-07, "loss": 0.2149, "step": 770 }, { "epoch": 0.015294117647058824, "grad_norm": 1.4687575101852417, "learning_rate": 1.527450980392157e-07, "loss": 0.1785, "step": 780 }, { "epoch": 0.015490196078431372, "grad_norm": 2.897915840148926, "learning_rate": 1.5470588235294116e-07, "loss": 0.1826, "step": 790 }, { "epoch": 0.01568627450980392, "grad_norm": 6.9971604347229, "learning_rate": 1.5666666666666667e-07, "loss": 0.2462, "step": 800 }, { "epoch": 0.01588235294117647, "grad_norm": 2.6937057971954346, "learning_rate": 1.5862745098039214e-07, "loss": 0.2155, "step": 810 }, { "epoch": 0.01607843137254902, "grad_norm": 2.972672700881958, "learning_rate": 1.6058823529411765e-07, "loss": 0.2301, "step": 820 }, { "epoch": 0.016274509803921568, "grad_norm": 3.5203969478607178, "learning_rate": 1.6254901960784312e-07, "loss": 0.2227, "step": 830 }, { "epoch": 0.01647058823529412, "grad_norm": 2.7854881286621094, "learning_rate": 1.6450980392156863e-07, "loss": 0.1975, "step": 840 }, { "epoch": 0.016666666666666666, "grad_norm": 2.832660675048828, "learning_rate": 1.664705882352941e-07, "loss": 0.2016, "step": 850 }, { "epoch": 0.016862745098039214, "grad_norm": 1.9016097784042358, "learning_rate": 1.684313725490196e-07, "loss": 0.1981, "step": 860 }, { "epoch": 0.017058823529411765, "grad_norm": 3.1742653846740723, "learning_rate": 1.7039215686274508e-07, "loss": 0.2347, "step": 870 }, { "epoch": 0.017254901960784313, "grad_norm": 3.4997215270996094, "learning_rate": 1.7235294117647059e-07, "loss": 0.2374, "step": 880 }, { "epoch": 0.017450980392156864, "grad_norm": 2.4480819702148438, "learning_rate": 1.7431372549019606e-07, "loss": 0.3168, "step": 890 }, { "epoch": 0.01764705882352941, "grad_norm": 1.604883074760437, "learning_rate": 1.7627450980392157e-07, "loss": 0.1925, "step": 900 }, { "epoch": 0.01784313725490196, "grad_norm": 3.549654722213745, "learning_rate": 1.7823529411764704e-07, "loss": 0.2359, "step": 910 }, { "epoch": 0.01803921568627451, "grad_norm": 4.209599494934082, "learning_rate": 1.8019607843137254e-07, "loss": 0.2015, "step": 920 }, { "epoch": 0.018235294117647058, "grad_norm": 1.7110475301742554, "learning_rate": 1.8215686274509802e-07, "loss": 0.1549, "step": 930 }, { "epoch": 0.01843137254901961, "grad_norm": 4.232614994049072, "learning_rate": 1.8411764705882352e-07, "loss": 0.2029, "step": 940 }, { "epoch": 0.018627450980392157, "grad_norm": 2.2283616065979004, "learning_rate": 1.86078431372549e-07, "loss": 0.1918, "step": 950 }, { "epoch": 0.018823529411764704, "grad_norm": 5.205087661743164, "learning_rate": 1.880392156862745e-07, "loss": 0.214, "step": 960 }, { "epoch": 0.019019607843137255, "grad_norm": 3.1778151988983154, "learning_rate": 1.8999999999999998e-07, "loss": 0.2395, "step": 970 }, { "epoch": 0.019215686274509803, "grad_norm": 2.5020692348480225, "learning_rate": 1.9196078431372548e-07, "loss": 0.2099, "step": 980 }, { "epoch": 0.019411764705882354, "grad_norm": 2.8813419342041016, "learning_rate": 1.9392156862745096e-07, "loss": 0.2175, "step": 990 }, { "epoch": 0.0196078431372549, "grad_norm": 5.034890174865723, "learning_rate": 1.9588235294117646e-07, "loss": 0.2195, "step": 1000 }, { "epoch": 0.01980392156862745, "grad_norm": 2.199882745742798, "learning_rate": 1.9784313725490197e-07, "loss": 0.2329, "step": 1010 }, { "epoch": 0.02, "grad_norm": 3.7657508850097656, "learning_rate": 1.9980392156862744e-07, "loss": 0.1985, "step": 1020 }, { "epoch": 0.020196078431372548, "grad_norm": 1.738678216934204, "learning_rate": 2.0176470588235292e-07, "loss": 0.1931, "step": 1030 }, { "epoch": 0.0203921568627451, "grad_norm": 2.506545066833496, "learning_rate": 2.0372549019607842e-07, "loss": 0.242, "step": 1040 }, { "epoch": 0.020588235294117647, "grad_norm": 3.4589574337005615, "learning_rate": 2.0568627450980393e-07, "loss": 0.2077, "step": 1050 }, { "epoch": 0.020784313725490194, "grad_norm": 3.32539701461792, "learning_rate": 2.076470588235294e-07, "loss": 0.1978, "step": 1060 }, { "epoch": 0.020980392156862746, "grad_norm": 3.646425724029541, "learning_rate": 2.0960784313725488e-07, "loss": 0.1957, "step": 1070 }, { "epoch": 0.021176470588235293, "grad_norm": 13.688029289245605, "learning_rate": 2.1156862745098038e-07, "loss": 0.1858, "step": 1080 }, { "epoch": 0.021372549019607844, "grad_norm": 1.7299249172210693, "learning_rate": 2.1352941176470589e-07, "loss": 0.1873, "step": 1090 }, { "epoch": 0.021568627450980392, "grad_norm": 2.190490245819092, "learning_rate": 2.1549019607843136e-07, "loss": 0.2431, "step": 1100 }, { "epoch": 0.02176470588235294, "grad_norm": 2.4254512786865234, "learning_rate": 2.1745098039215684e-07, "loss": 0.1748, "step": 1110 }, { "epoch": 0.02196078431372549, "grad_norm": 2.423584222793579, "learning_rate": 2.1941176470588234e-07, "loss": 0.2213, "step": 1120 }, { "epoch": 0.022156862745098038, "grad_norm": 2.6117727756500244, "learning_rate": 2.2137254901960784e-07, "loss": 0.1804, "step": 1130 }, { "epoch": 0.02235294117647059, "grad_norm": 1.8259599208831787, "learning_rate": 2.2333333333333332e-07, "loss": 0.1953, "step": 1140 }, { "epoch": 0.022549019607843137, "grad_norm": 2.6838877201080322, "learning_rate": 2.252941176470588e-07, "loss": 0.1846, "step": 1150 }, { "epoch": 0.022745098039215685, "grad_norm": 2.9006917476654053, "learning_rate": 2.272549019607843e-07, "loss": 0.2096, "step": 1160 }, { "epoch": 0.022941176470588236, "grad_norm": 2.9995810985565186, "learning_rate": 2.292156862745098e-07, "loss": 0.2259, "step": 1170 }, { "epoch": 0.023137254901960783, "grad_norm": 2.303877353668213, "learning_rate": 2.3117647058823528e-07, "loss": 0.2007, "step": 1180 }, { "epoch": 0.023333333333333334, "grad_norm": 1.9802242517471313, "learning_rate": 2.3313725490196076e-07, "loss": 0.2264, "step": 1190 }, { "epoch": 0.023529411764705882, "grad_norm": 2.456413745880127, "learning_rate": 2.3509803921568626e-07, "loss": 0.2135, "step": 1200 }, { "epoch": 0.02372549019607843, "grad_norm": 4.042865753173828, "learning_rate": 2.3705882352941176e-07, "loss": 0.1991, "step": 1210 }, { "epoch": 0.02392156862745098, "grad_norm": 3.221712589263916, "learning_rate": 2.390196078431372e-07, "loss": 0.217, "step": 1220 }, { "epoch": 0.02411764705882353, "grad_norm": 2.4783685207366943, "learning_rate": 2.4098039215686274e-07, "loss": 0.2181, "step": 1230 }, { "epoch": 0.02431372549019608, "grad_norm": 2.7155158519744873, "learning_rate": 2.429411764705882e-07, "loss": 0.192, "step": 1240 }, { "epoch": 0.024509803921568627, "grad_norm": 3.0510873794555664, "learning_rate": 2.449019607843137e-07, "loss": 0.2012, "step": 1250 }, { "epoch": 0.024705882352941175, "grad_norm": 2.6850430965423584, "learning_rate": 2.4686274509803917e-07, "loss": 0.171, "step": 1260 }, { "epoch": 0.024901960784313726, "grad_norm": 3.3726608753204346, "learning_rate": 2.488235294117647e-07, "loss": 0.2082, "step": 1270 }, { "epoch": 0.025098039215686273, "grad_norm": 3.0024449825286865, "learning_rate": 2.507843137254902e-07, "loss": 0.1784, "step": 1280 }, { "epoch": 0.025294117647058825, "grad_norm": 4.871731758117676, "learning_rate": 2.5274509803921566e-07, "loss": 0.2298, "step": 1290 }, { "epoch": 0.025490196078431372, "grad_norm": 4.273900985717773, "learning_rate": 2.547058823529412e-07, "loss": 0.2315, "step": 1300 }, { "epoch": 0.025686274509803923, "grad_norm": 3.133518695831299, "learning_rate": 2.5666666666666666e-07, "loss": 0.1884, "step": 1310 }, { "epoch": 0.02588235294117647, "grad_norm": 2.8455162048339844, "learning_rate": 2.5862745098039214e-07, "loss": 0.1859, "step": 1320 }, { "epoch": 0.02607843137254902, "grad_norm": 2.273074150085449, "learning_rate": 2.605882352941176e-07, "loss": 0.2354, "step": 1330 }, { "epoch": 0.02627450980392157, "grad_norm": 3.2788944244384766, "learning_rate": 2.6254901960784314e-07, "loss": 0.1952, "step": 1340 }, { "epoch": 0.026470588235294117, "grad_norm": 4.694311618804932, "learning_rate": 2.645098039215686e-07, "loss": 0.2053, "step": 1350 }, { "epoch": 0.02666666666666667, "grad_norm": 1.7802472114562988, "learning_rate": 2.664705882352941e-07, "loss": 0.2437, "step": 1360 }, { "epoch": 0.026862745098039216, "grad_norm": 1.496583104133606, "learning_rate": 2.684313725490196e-07, "loss": 0.2048, "step": 1370 }, { "epoch": 0.027058823529411764, "grad_norm": 3.0002849102020264, "learning_rate": 2.703921568627451e-07, "loss": 0.2102, "step": 1380 }, { "epoch": 0.027254901960784315, "grad_norm": 2.2182552814483643, "learning_rate": 2.723529411764706e-07, "loss": 0.2118, "step": 1390 }, { "epoch": 0.027450980392156862, "grad_norm": 3.4683403968811035, "learning_rate": 2.7431372549019606e-07, "loss": 0.2236, "step": 1400 }, { "epoch": 0.027647058823529413, "grad_norm": 2.334892988204956, "learning_rate": 2.7627450980392153e-07, "loss": 0.1828, "step": 1410 }, { "epoch": 0.02784313725490196, "grad_norm": 2.039668560028076, "learning_rate": 2.7823529411764706e-07, "loss": 0.2101, "step": 1420 }, { "epoch": 0.02803921568627451, "grad_norm": 2.187269687652588, "learning_rate": 2.8019607843137254e-07, "loss": 0.199, "step": 1430 }, { "epoch": 0.02823529411764706, "grad_norm": 2.6782257556915283, "learning_rate": 2.82156862745098e-07, "loss": 0.2087, "step": 1440 }, { "epoch": 0.028431372549019607, "grad_norm": 2.8355424404144287, "learning_rate": 2.841176470588235e-07, "loss": 0.1897, "step": 1450 }, { "epoch": 0.02862745098039216, "grad_norm": 2.3069262504577637, "learning_rate": 2.86078431372549e-07, "loss": 0.2054, "step": 1460 }, { "epoch": 0.028823529411764706, "grad_norm": 5.331542015075684, "learning_rate": 2.880392156862745e-07, "loss": 0.2014, "step": 1470 }, { "epoch": 0.029019607843137254, "grad_norm": 3.2946064472198486, "learning_rate": 2.9e-07, "loss": 0.2316, "step": 1480 }, { "epoch": 0.029215686274509805, "grad_norm": 2.943204164505005, "learning_rate": 2.9196078431372545e-07, "loss": 0.2139, "step": 1490 }, { "epoch": 0.029411764705882353, "grad_norm": 3.040644407272339, "learning_rate": 2.93921568627451e-07, "loss": 0.1999, "step": 1500 }, { "epoch": 0.029607843137254904, "grad_norm": 3.9258229732513428, "learning_rate": 2.9588235294117646e-07, "loss": 0.2017, "step": 1510 }, { "epoch": 0.02980392156862745, "grad_norm": 2.669543504714966, "learning_rate": 2.9784313725490194e-07, "loss": 0.1803, "step": 1520 }, { "epoch": 0.03, "grad_norm": 1.9782130718231201, "learning_rate": 2.998039215686274e-07, "loss": 0.1824, "step": 1530 }, { "epoch": 0.03019607843137255, "grad_norm": 4.341161727905273, "learning_rate": 3.0176470588235294e-07, "loss": 0.1644, "step": 1540 }, { "epoch": 0.030392156862745098, "grad_norm": 1.8393055200576782, "learning_rate": 3.037254901960784e-07, "loss": 0.1834, "step": 1550 }, { "epoch": 0.03058823529411765, "grad_norm": 3.5511372089385986, "learning_rate": 3.056862745098039e-07, "loss": 0.2161, "step": 1560 }, { "epoch": 0.030784313725490196, "grad_norm": 2.466632843017578, "learning_rate": 3.0764705882352937e-07, "loss": 0.2005, "step": 1570 }, { "epoch": 0.030980392156862744, "grad_norm": 2.075239419937134, "learning_rate": 3.096078431372549e-07, "loss": 0.2118, "step": 1580 }, { "epoch": 0.031176470588235295, "grad_norm": 3.3345069885253906, "learning_rate": 3.115686274509804e-07, "loss": 0.1865, "step": 1590 }, { "epoch": 0.03137254901960784, "grad_norm": 3.064138650894165, "learning_rate": 3.1352941176470585e-07, "loss": 0.2133, "step": 1600 }, { "epoch": 0.031568627450980394, "grad_norm": 2.285356044769287, "learning_rate": 3.154901960784314e-07, "loss": 0.1649, "step": 1610 }, { "epoch": 0.03176470588235294, "grad_norm": 2.0132076740264893, "learning_rate": 3.1745098039215686e-07, "loss": 0.1998, "step": 1620 }, { "epoch": 0.03196078431372549, "grad_norm": 2.7852389812469482, "learning_rate": 3.1941176470588234e-07, "loss": 0.159, "step": 1630 }, { "epoch": 0.03215686274509804, "grad_norm": 1.691112995147705, "learning_rate": 3.213725490196078e-07, "loss": 0.2099, "step": 1640 }, { "epoch": 0.03235294117647059, "grad_norm": 2.3763959407806396, "learning_rate": 3.233333333333333e-07, "loss": 0.2079, "step": 1650 }, { "epoch": 0.032549019607843135, "grad_norm": 4.417029857635498, "learning_rate": 3.252941176470588e-07, "loss": 0.2082, "step": 1660 }, { "epoch": 0.032745098039215687, "grad_norm": 2.4511497020721436, "learning_rate": 3.272549019607843e-07, "loss": 0.2042, "step": 1670 }, { "epoch": 0.03294117647058824, "grad_norm": 3.1663379669189453, "learning_rate": 3.2921568627450977e-07, "loss": 0.2033, "step": 1680 }, { "epoch": 0.03313725490196078, "grad_norm": 2.6681387424468994, "learning_rate": 3.311764705882353e-07, "loss": 0.2202, "step": 1690 }, { "epoch": 0.03333333333333333, "grad_norm": 3.301950216293335, "learning_rate": 3.331372549019608e-07, "loss": 0.1811, "step": 1700 }, { "epoch": 0.033529411764705884, "grad_norm": 3.376903772354126, "learning_rate": 3.3509803921568626e-07, "loss": 0.1844, "step": 1710 }, { "epoch": 0.03372549019607843, "grad_norm": 3.5456833839416504, "learning_rate": 3.3705882352941173e-07, "loss": 0.2026, "step": 1720 }, { "epoch": 0.03392156862745098, "grad_norm": 3.4473533630371094, "learning_rate": 3.390196078431372e-07, "loss": 0.1791, "step": 1730 }, { "epoch": 0.03411764705882353, "grad_norm": 2.914616823196411, "learning_rate": 3.4098039215686274e-07, "loss": 0.218, "step": 1740 }, { "epoch": 0.03431372549019608, "grad_norm": 4.259304523468018, "learning_rate": 3.429411764705882e-07, "loss": 0.1876, "step": 1750 }, { "epoch": 0.034509803921568626, "grad_norm": 2.343111515045166, "learning_rate": 3.449019607843137e-07, "loss": 0.2214, "step": 1760 }, { "epoch": 0.03470588235294118, "grad_norm": 4.733630657196045, "learning_rate": 3.468627450980392e-07, "loss": 0.2044, "step": 1770 }, { "epoch": 0.03490196078431373, "grad_norm": 3.5949387550354004, "learning_rate": 3.488235294117647e-07, "loss": 0.2412, "step": 1780 }, { "epoch": 0.03509803921568627, "grad_norm": 2.319788932800293, "learning_rate": 3.507843137254902e-07, "loss": 0.1899, "step": 1790 }, { "epoch": 0.03529411764705882, "grad_norm": 3.8382904529571533, "learning_rate": 3.5274509803921565e-07, "loss": 0.1851, "step": 1800 }, { "epoch": 0.035490196078431374, "grad_norm": 3.1545403003692627, "learning_rate": 3.547058823529412e-07, "loss": 0.1974, "step": 1810 }, { "epoch": 0.03568627450980392, "grad_norm": 2.272691011428833, "learning_rate": 3.5666666666666666e-07, "loss": 0.1837, "step": 1820 }, { "epoch": 0.03588235294117647, "grad_norm": 2.2936556339263916, "learning_rate": 3.5862745098039213e-07, "loss": 0.1754, "step": 1830 }, { "epoch": 0.03607843137254902, "grad_norm": 2.031348466873169, "learning_rate": 3.6058823529411766e-07, "loss": 0.2087, "step": 1840 }, { "epoch": 0.03627450980392157, "grad_norm": 2.2454469203948975, "learning_rate": 3.6254901960784314e-07, "loss": 0.1861, "step": 1850 }, { "epoch": 0.036470588235294116, "grad_norm": 2.2150423526763916, "learning_rate": 3.645098039215686e-07, "loss": 0.1627, "step": 1860 }, { "epoch": 0.03666666666666667, "grad_norm": 1.7423158884048462, "learning_rate": 3.664705882352941e-07, "loss": 0.1896, "step": 1870 }, { "epoch": 0.03686274509803922, "grad_norm": 3.8175764083862305, "learning_rate": 3.6843137254901957e-07, "loss": 0.2128, "step": 1880 }, { "epoch": 0.03705882352941176, "grad_norm": 3.566584348678589, "learning_rate": 3.703921568627451e-07, "loss": 0.2475, "step": 1890 }, { "epoch": 0.03725490196078431, "grad_norm": 2.674799680709839, "learning_rate": 3.723529411764706e-07, "loss": 0.2118, "step": 1900 }, { "epoch": 0.037450980392156864, "grad_norm": 3.6460888385772705, "learning_rate": 3.7431372549019605e-07, "loss": 0.184, "step": 1910 }, { "epoch": 0.03764705882352941, "grad_norm": 2.144700050354004, "learning_rate": 3.762745098039216e-07, "loss": 0.1831, "step": 1920 }, { "epoch": 0.03784313725490196, "grad_norm": 2.871405601501465, "learning_rate": 3.7823529411764706e-07, "loss": 0.1932, "step": 1930 }, { "epoch": 0.03803921568627451, "grad_norm": 3.03185772895813, "learning_rate": 3.8019607843137254e-07, "loss": 0.2097, "step": 1940 }, { "epoch": 0.03823529411764706, "grad_norm": 3.685610294342041, "learning_rate": 3.82156862745098e-07, "loss": 0.2012, "step": 1950 }, { "epoch": 0.038431372549019606, "grad_norm": 2.7546932697296143, "learning_rate": 3.841176470588235e-07, "loss": 0.2024, "step": 1960 }, { "epoch": 0.03862745098039216, "grad_norm": 3.8974952697753906, "learning_rate": 3.86078431372549e-07, "loss": 0.2419, "step": 1970 }, { "epoch": 0.03882352941176471, "grad_norm": 3.8983378410339355, "learning_rate": 3.880392156862745e-07, "loss": 0.1958, "step": 1980 }, { "epoch": 0.03901960784313725, "grad_norm": 2.5328612327575684, "learning_rate": 3.8999999999999997e-07, "loss": 0.1959, "step": 1990 }, { "epoch": 0.0392156862745098, "grad_norm": 3.017338991165161, "learning_rate": 3.919607843137255e-07, "loss": 0.1952, "step": 2000 }, { "epoch": 0.039411764705882354, "grad_norm": 2.2104644775390625, "learning_rate": 3.93921568627451e-07, "loss": 0.1858, "step": 2010 }, { "epoch": 0.0396078431372549, "grad_norm": 6.310041904449463, "learning_rate": 3.9588235294117645e-07, "loss": 0.2262, "step": 2020 }, { "epoch": 0.03980392156862745, "grad_norm": 3.676623582839966, "learning_rate": 3.9784313725490193e-07, "loss": 0.1724, "step": 2030 }, { "epoch": 0.04, "grad_norm": 2.1587581634521484, "learning_rate": 3.998039215686274e-07, "loss": 0.2247, "step": 2040 }, { "epoch": 0.04019607843137255, "grad_norm": 3.18282413482666, "learning_rate": 4.0176470588235294e-07, "loss": 0.2349, "step": 2050 }, { "epoch": 0.040392156862745096, "grad_norm": 3.8631222248077393, "learning_rate": 4.037254901960784e-07, "loss": 0.1539, "step": 2060 }, { "epoch": 0.04058823529411765, "grad_norm": 3.0927159786224365, "learning_rate": 4.056862745098039e-07, "loss": 0.203, "step": 2070 }, { "epoch": 0.0407843137254902, "grad_norm": 2.273651599884033, "learning_rate": 4.076470588235294e-07, "loss": 0.1933, "step": 2080 }, { "epoch": 0.04098039215686274, "grad_norm": 2.820582389831543, "learning_rate": 4.096078431372549e-07, "loss": 0.1763, "step": 2090 }, { "epoch": 0.041176470588235294, "grad_norm": 2.6153640747070312, "learning_rate": 4.1156862745098037e-07, "loss": 0.242, "step": 2100 }, { "epoch": 0.041372549019607845, "grad_norm": 1.7456018924713135, "learning_rate": 4.1352941176470585e-07, "loss": 0.1559, "step": 2110 }, { "epoch": 0.04156862745098039, "grad_norm": 2.6023061275482178, "learning_rate": 4.1549019607843133e-07, "loss": 0.1946, "step": 2120 }, { "epoch": 0.04176470588235294, "grad_norm": 2.7573719024658203, "learning_rate": 4.1745098039215686e-07, "loss": 0.1846, "step": 2130 }, { "epoch": 0.04196078431372549, "grad_norm": 3.8828041553497314, "learning_rate": 4.1941176470588233e-07, "loss": 0.1947, "step": 2140 }, { "epoch": 0.04215686274509804, "grad_norm": 3.04514479637146, "learning_rate": 4.2137254901960786e-07, "loss": 0.2046, "step": 2150 }, { "epoch": 0.042352941176470586, "grad_norm": 4.05771017074585, "learning_rate": 4.2333333333333334e-07, "loss": 0.185, "step": 2160 }, { "epoch": 0.04254901960784314, "grad_norm": 12.687215805053711, "learning_rate": 4.252941176470588e-07, "loss": 0.2021, "step": 2170 }, { "epoch": 0.04274509803921569, "grad_norm": 2.9035496711730957, "learning_rate": 4.272549019607843e-07, "loss": 0.1999, "step": 2180 }, { "epoch": 0.04294117647058823, "grad_norm": 4.291037082672119, "learning_rate": 4.2921568627450977e-07, "loss": 0.2487, "step": 2190 }, { "epoch": 0.043137254901960784, "grad_norm": 2.9060707092285156, "learning_rate": 4.3117647058823525e-07, "loss": 0.2424, "step": 2200 }, { "epoch": 0.043333333333333335, "grad_norm": 2.7251791954040527, "learning_rate": 4.331372549019608e-07, "loss": 0.1818, "step": 2210 }, { "epoch": 0.04352941176470588, "grad_norm": 1.8093441724777222, "learning_rate": 4.3509803921568625e-07, "loss": 0.1905, "step": 2220 }, { "epoch": 0.04372549019607843, "grad_norm": 2.042854070663452, "learning_rate": 4.370588235294118e-07, "loss": 0.2111, "step": 2230 }, { "epoch": 0.04392156862745098, "grad_norm": 1.566724181175232, "learning_rate": 4.3901960784313726e-07, "loss": 0.175, "step": 2240 }, { "epoch": 0.04411764705882353, "grad_norm": 6.421918869018555, "learning_rate": 4.4098039215686273e-07, "loss": 0.1873, "step": 2250 }, { "epoch": 0.044313725490196076, "grad_norm": 2.8671746253967285, "learning_rate": 4.429411764705882e-07, "loss": 0.1953, "step": 2260 }, { "epoch": 0.04450980392156863, "grad_norm": 1.647371768951416, "learning_rate": 4.449019607843137e-07, "loss": 0.2201, "step": 2270 }, { "epoch": 0.04470588235294118, "grad_norm": 3.8033668994903564, "learning_rate": 4.4686274509803916e-07, "loss": 0.2137, "step": 2280 }, { "epoch": 0.04490196078431372, "grad_norm": 2.233917713165283, "learning_rate": 4.488235294117647e-07, "loss": 0.1738, "step": 2290 }, { "epoch": 0.045098039215686274, "grad_norm": 1.9051834344863892, "learning_rate": 4.5078431372549017e-07, "loss": 0.1848, "step": 2300 }, { "epoch": 0.045294117647058825, "grad_norm": 1.389155387878418, "learning_rate": 4.527450980392157e-07, "loss": 0.1948, "step": 2310 }, { "epoch": 0.04549019607843137, "grad_norm": 1.6292574405670166, "learning_rate": 4.547058823529412e-07, "loss": 0.1779, "step": 2320 }, { "epoch": 0.04568627450980392, "grad_norm": 2.842968225479126, "learning_rate": 4.5666666666666665e-07, "loss": 0.1669, "step": 2330 }, { "epoch": 0.04588235294117647, "grad_norm": 3.004565954208374, "learning_rate": 4.5862745098039213e-07, "loss": 0.2407, "step": 2340 }, { "epoch": 0.04607843137254902, "grad_norm": 2.346376419067383, "learning_rate": 4.605882352941176e-07, "loss": 0.2226, "step": 2350 }, { "epoch": 0.04627450980392157, "grad_norm": 4.173306941986084, "learning_rate": 4.625490196078431e-07, "loss": 0.2728, "step": 2360 }, { "epoch": 0.04647058823529412, "grad_norm": 4.030762672424316, "learning_rate": 4.645098039215686e-07, "loss": 0.2027, "step": 2370 }, { "epoch": 0.04666666666666667, "grad_norm": 3.40122127532959, "learning_rate": 4.6647058823529414e-07, "loss": 0.1926, "step": 2380 }, { "epoch": 0.04686274509803921, "grad_norm": 2.1672606468200684, "learning_rate": 4.684313725490196e-07, "loss": 0.1884, "step": 2390 }, { "epoch": 0.047058823529411764, "grad_norm": 2.663543701171875, "learning_rate": 4.703921568627451e-07, "loss": 0.1799, "step": 2400 }, { "epoch": 0.047254901960784315, "grad_norm": 2.500825881958008, "learning_rate": 4.7235294117647057e-07, "loss": 0.2025, "step": 2410 }, { "epoch": 0.04745098039215686, "grad_norm": 3.3113651275634766, "learning_rate": 4.7431372549019605e-07, "loss": 0.1941, "step": 2420 }, { "epoch": 0.04764705882352941, "grad_norm": 5.039560794830322, "learning_rate": 4.762745098039215e-07, "loss": 0.2214, "step": 2430 }, { "epoch": 0.04784313725490196, "grad_norm": 1.9254804849624634, "learning_rate": 4.78235294117647e-07, "loss": 0.1868, "step": 2440 }, { "epoch": 0.04803921568627451, "grad_norm": 2.1486563682556152, "learning_rate": 4.801960784313725e-07, "loss": 0.2184, "step": 2450 }, { "epoch": 0.04823529411764706, "grad_norm": 1.050247311592102, "learning_rate": 4.821568627450981e-07, "loss": 0.2273, "step": 2460 }, { "epoch": 0.04843137254901961, "grad_norm": 1.8149820566177368, "learning_rate": 4.841176470588235e-07, "loss": 0.1813, "step": 2470 }, { "epoch": 0.04862745098039216, "grad_norm": 1.038099765777588, "learning_rate": 4.86078431372549e-07, "loss": 0.2182, "step": 2480 }, { "epoch": 0.0488235294117647, "grad_norm": 2.827169418334961, "learning_rate": 4.880392156862744e-07, "loss": 0.1753, "step": 2490 }, { "epoch": 0.049019607843137254, "grad_norm": 2.1705260276794434, "learning_rate": 4.9e-07, "loss": 0.1738, "step": 2500 }, { "epoch": 0.049215686274509805, "grad_norm": 2.702012777328491, "learning_rate": 4.919607843137255e-07, "loss": 0.254, "step": 2510 }, { "epoch": 0.04941176470588235, "grad_norm": 3.804274082183838, "learning_rate": 4.939215686274509e-07, "loss": 0.1718, "step": 2520 }, { "epoch": 0.0496078431372549, "grad_norm": 2.47497296333313, "learning_rate": 4.958823529411765e-07, "loss": 0.2081, "step": 2530 }, { "epoch": 0.04980392156862745, "grad_norm": 1.561738133430481, "learning_rate": 4.97843137254902e-07, "loss": 0.1621, "step": 2540 }, { "epoch": 0.05, "grad_norm": 1.7904165983200073, "learning_rate": 4.998039215686274e-07, "loss": 0.178, "step": 2550 }, { "epoch": 0.05019607843137255, "grad_norm": 3.2636663913726807, "learning_rate": 5.017647058823528e-07, "loss": 0.1953, "step": 2560 }, { "epoch": 0.0503921568627451, "grad_norm": 4.642648220062256, "learning_rate": 5.037254901960785e-07, "loss": 0.1845, "step": 2570 }, { "epoch": 0.05058823529411765, "grad_norm": 2.524141788482666, "learning_rate": 5.05686274509804e-07, "loss": 0.1648, "step": 2580 }, { "epoch": 0.05078431372549019, "grad_norm": 1.575900912284851, "learning_rate": 5.076470588235294e-07, "loss": 0.1821, "step": 2590 }, { "epoch": 0.050980392156862744, "grad_norm": 2.6308038234710693, "learning_rate": 5.096078431372549e-07, "loss": 0.1433, "step": 2600 }, { "epoch": 0.051176470588235295, "grad_norm": 1.6228832006454468, "learning_rate": 5.115686274509804e-07, "loss": 0.164, "step": 2610 }, { "epoch": 0.05137254901960785, "grad_norm": 1.9744882583618164, "learning_rate": 5.135294117647059e-07, "loss": 0.1543, "step": 2620 }, { "epoch": 0.05156862745098039, "grad_norm": 2.5851263999938965, "learning_rate": 5.154901960784313e-07, "loss": 0.1916, "step": 2630 }, { "epoch": 0.05176470588235294, "grad_norm": 2.893012046813965, "learning_rate": 5.174509803921569e-07, "loss": 0.2447, "step": 2640 }, { "epoch": 0.05196078431372549, "grad_norm": 2.2568612098693848, "learning_rate": 5.194117647058823e-07, "loss": 0.1971, "step": 2650 }, { "epoch": 0.05215686274509804, "grad_norm": 4.419097423553467, "learning_rate": 5.213725490196078e-07, "loss": 0.2084, "step": 2660 }, { "epoch": 0.05235294117647059, "grad_norm": 2.660299301147461, "learning_rate": 5.233333333333333e-07, "loss": 0.2307, "step": 2670 }, { "epoch": 0.05254901960784314, "grad_norm": 2.794532060623169, "learning_rate": 5.252941176470588e-07, "loss": 0.2197, "step": 2680 }, { "epoch": 0.05274509803921568, "grad_norm": 2.391566038131714, "learning_rate": 5.272549019607843e-07, "loss": 0.2226, "step": 2690 }, { "epoch": 0.052941176470588235, "grad_norm": 3.0155463218688965, "learning_rate": 5.292156862745097e-07, "loss": 0.1548, "step": 2700 }, { "epoch": 0.053137254901960786, "grad_norm": 2.9043197631835938, "learning_rate": 5.311764705882352e-07, "loss": 0.1734, "step": 2710 }, { "epoch": 0.05333333333333334, "grad_norm": 4.487123012542725, "learning_rate": 5.331372549019607e-07, "loss": 0.1965, "step": 2720 }, { "epoch": 0.05352941176470588, "grad_norm": 1.998080849647522, "learning_rate": 5.350980392156863e-07, "loss": 0.1634, "step": 2730 }, { "epoch": 0.05372549019607843, "grad_norm": 2.375079393386841, "learning_rate": 5.370588235294118e-07, "loss": 0.1709, "step": 2740 }, { "epoch": 0.05392156862745098, "grad_norm": 2.17889142036438, "learning_rate": 5.390196078431373e-07, "loss": 0.1579, "step": 2750 }, { "epoch": 0.05411764705882353, "grad_norm": 4.838517665863037, "learning_rate": 5.409803921568628e-07, "loss": 0.205, "step": 2760 }, { "epoch": 0.05431372549019608, "grad_norm": 2.5293469429016113, "learning_rate": 5.429411764705882e-07, "loss": 0.1877, "step": 2770 }, { "epoch": 0.05450980392156863, "grad_norm": 3.7085185050964355, "learning_rate": 5.449019607843137e-07, "loss": 0.1898, "step": 2780 }, { "epoch": 0.054705882352941174, "grad_norm": 2.4288878440856934, "learning_rate": 5.468627450980392e-07, "loss": 0.212, "step": 2790 }, { "epoch": 0.054901960784313725, "grad_norm": 2.252490997314453, "learning_rate": 5.488235294117647e-07, "loss": 0.1984, "step": 2800 }, { "epoch": 0.055098039215686276, "grad_norm": 2.979335069656372, "learning_rate": 5.507843137254902e-07, "loss": 0.1872, "step": 2810 }, { "epoch": 0.05529411764705883, "grad_norm": 2.502167224884033, "learning_rate": 5.527450980392156e-07, "loss": 0.2036, "step": 2820 }, { "epoch": 0.05549019607843137, "grad_norm": 8.726903915405273, "learning_rate": 5.547058823529412e-07, "loss": 0.1932, "step": 2830 }, { "epoch": 0.05568627450980392, "grad_norm": 2.4564144611358643, "learning_rate": 5.566666666666666e-07, "loss": 0.1946, "step": 2840 }, { "epoch": 0.05588235294117647, "grad_norm": 2.4475178718566895, "learning_rate": 5.586274509803921e-07, "loss": 0.1954, "step": 2850 }, { "epoch": 0.05607843137254902, "grad_norm": 3.373850107192993, "learning_rate": 5.605882352941175e-07, "loss": 0.2254, "step": 2860 }, { "epoch": 0.05627450980392157, "grad_norm": 3.1207592487335205, "learning_rate": 5.625490196078431e-07, "loss": 0.2199, "step": 2870 }, { "epoch": 0.05647058823529412, "grad_norm": 2.057288885116577, "learning_rate": 5.645098039215685e-07, "loss": 0.201, "step": 2880 }, { "epoch": 0.056666666666666664, "grad_norm": 3.683852434158325, "learning_rate": 5.664705882352941e-07, "loss": 0.2022, "step": 2890 }, { "epoch": 0.056862745098039215, "grad_norm": 1.9630478620529175, "learning_rate": 5.684313725490197e-07, "loss": 0.177, "step": 2900 }, { "epoch": 0.057058823529411766, "grad_norm": 2.999586343765259, "learning_rate": 5.703921568627451e-07, "loss": 0.1932, "step": 2910 }, { "epoch": 0.05725490196078432, "grad_norm": 2.0837106704711914, "learning_rate": 5.723529411764706e-07, "loss": 0.2385, "step": 2920 }, { "epoch": 0.05745098039215686, "grad_norm": 1.7918663024902344, "learning_rate": 5.74313725490196e-07, "loss": 0.2148, "step": 2930 }, { "epoch": 0.05764705882352941, "grad_norm": 2.4507970809936523, "learning_rate": 5.762745098039216e-07, "loss": 0.1967, "step": 2940 }, { "epoch": 0.05784313725490196, "grad_norm": 3.001332998275757, "learning_rate": 5.78235294117647e-07, "loss": 0.2005, "step": 2950 }, { "epoch": 0.05803921568627451, "grad_norm": 2.1292388439178467, "learning_rate": 5.801960784313725e-07, "loss": 0.2457, "step": 2960 }, { "epoch": 0.05823529411764706, "grad_norm": 3.217041254043579, "learning_rate": 5.821568627450981e-07, "loss": 0.2224, "step": 2970 }, { "epoch": 0.05843137254901961, "grad_norm": 3.1178836822509766, "learning_rate": 5.841176470588235e-07, "loss": 0.214, "step": 2980 }, { "epoch": 0.058627450980392154, "grad_norm": 2.770151138305664, "learning_rate": 5.86078431372549e-07, "loss": 0.187, "step": 2990 }, { "epoch": 0.058823529411764705, "grad_norm": 4.772091865539551, "learning_rate": 5.880392156862744e-07, "loss": 0.1903, "step": 3000 }, { "epoch": 0.059019607843137256, "grad_norm": 2.3392953872680664, "learning_rate": 5.9e-07, "loss": 0.1785, "step": 3010 }, { "epoch": 0.05921568627450981, "grad_norm": 4.751784801483154, "learning_rate": 5.919607843137254e-07, "loss": 0.1923, "step": 3020 }, { "epoch": 0.05941176470588235, "grad_norm": 3.2652649879455566, "learning_rate": 5.939215686274509e-07, "loss": 0.2335, "step": 3030 }, { "epoch": 0.0596078431372549, "grad_norm": 2.091738700866699, "learning_rate": 5.958823529411763e-07, "loss": 0.1993, "step": 3040 }, { "epoch": 0.059803921568627454, "grad_norm": 2.979250907897949, "learning_rate": 5.97843137254902e-07, "loss": 0.2204, "step": 3050 }, { "epoch": 0.06, "grad_norm": 3.0943524837493896, "learning_rate": 5.998039215686275e-07, "loss": 0.2033, "step": 3060 }, { "epoch": 0.06019607843137255, "grad_norm": 3.3576834201812744, "learning_rate": 6.017647058823529e-07, "loss": 0.209, "step": 3070 }, { "epoch": 0.0603921568627451, "grad_norm": 3.4279446601867676, "learning_rate": 6.037254901960785e-07, "loss": 0.2087, "step": 3080 }, { "epoch": 0.060588235294117644, "grad_norm": 3.759692907333374, "learning_rate": 6.056862745098039e-07, "loss": 0.1564, "step": 3090 }, { "epoch": 0.060784313725490195, "grad_norm": 2.622617721557617, "learning_rate": 6.076470588235294e-07, "loss": 0.1945, "step": 3100 }, { "epoch": 0.060980392156862746, "grad_norm": 2.1470530033111572, "learning_rate": 6.096078431372548e-07, "loss": 0.1869, "step": 3110 }, { "epoch": 0.0611764705882353, "grad_norm": 2.4790213108062744, "learning_rate": 6.115686274509804e-07, "loss": 0.2214, "step": 3120 }, { "epoch": 0.06137254901960784, "grad_norm": 2.5357706546783447, "learning_rate": 6.135294117647059e-07, "loss": 0.1775, "step": 3130 }, { "epoch": 0.06156862745098039, "grad_norm": 3.3663432598114014, "learning_rate": 6.154901960784313e-07, "loss": 0.2017, "step": 3140 }, { "epoch": 0.061764705882352944, "grad_norm": 2.5787577629089355, "learning_rate": 6.174509803921568e-07, "loss": 0.2116, "step": 3150 }, { "epoch": 0.06196078431372549, "grad_norm": 2.607107639312744, "learning_rate": 6.194117647058823e-07, "loss": 0.2419, "step": 3160 }, { "epoch": 0.06215686274509804, "grad_norm": 2.427081823348999, "learning_rate": 6.213725490196078e-07, "loss": 0.1968, "step": 3170 }, { "epoch": 0.06235294117647059, "grad_norm": 3.8898117542266846, "learning_rate": 6.233333333333332e-07, "loss": 0.1848, "step": 3180 }, { "epoch": 0.06254901960784313, "grad_norm": 2.3590660095214844, "learning_rate": 6.252941176470588e-07, "loss": 0.2223, "step": 3190 }, { "epoch": 0.06274509803921569, "grad_norm": 5.244894981384277, "learning_rate": 6.272549019607843e-07, "loss": 0.1884, "step": 3200 }, { "epoch": 0.06294117647058824, "grad_norm": 5.8015947341918945, "learning_rate": 6.292156862745098e-07, "loss": 0.2022, "step": 3210 }, { "epoch": 0.06313725490196079, "grad_norm": 5.453890800476074, "learning_rate": 6.311764705882353e-07, "loss": 0.2252, "step": 3220 }, { "epoch": 0.06333333333333334, "grad_norm": 2.379526376724243, "learning_rate": 6.331372549019608e-07, "loss": 0.171, "step": 3230 }, { "epoch": 0.06352941176470588, "grad_norm": 2.8273093700408936, "learning_rate": 6.350980392156863e-07, "loss": 0.1923, "step": 3240 }, { "epoch": 0.06372549019607843, "grad_norm": 1.8225847482681274, "learning_rate": 6.370588235294117e-07, "loss": 0.1476, "step": 3250 }, { "epoch": 0.06392156862745098, "grad_norm": 2.6780216693878174, "learning_rate": 6.390196078431372e-07, "loss": 0.1809, "step": 3260 }, { "epoch": 0.06411764705882353, "grad_norm": 4.624337196350098, "learning_rate": 6.409803921568628e-07, "loss": 0.1803, "step": 3270 }, { "epoch": 0.06431372549019608, "grad_norm": 2.4611167907714844, "learning_rate": 6.429411764705882e-07, "loss": 0.1994, "step": 3280 }, { "epoch": 0.06450980392156863, "grad_norm": 2.640899181365967, "learning_rate": 6.449019607843137e-07, "loss": 0.2013, "step": 3290 }, { "epoch": 0.06470588235294118, "grad_norm": 1.5827934741973877, "learning_rate": 6.468627450980392e-07, "loss": 0.2328, "step": 3300 }, { "epoch": 0.06490196078431372, "grad_norm": 3.1288492679595947, "learning_rate": 6.488235294117647e-07, "loss": 0.2036, "step": 3310 }, { "epoch": 0.06509803921568627, "grad_norm": 3.9613397121429443, "learning_rate": 6.507843137254901e-07, "loss": 0.1996, "step": 3320 }, { "epoch": 0.06529411764705882, "grad_norm": 2.1854116916656494, "learning_rate": 6.527450980392156e-07, "loss": 0.1885, "step": 3330 }, { "epoch": 0.06549019607843137, "grad_norm": 2.2846286296844482, "learning_rate": 6.547058823529411e-07, "loss": 0.1735, "step": 3340 }, { "epoch": 0.06568627450980392, "grad_norm": 2.6201133728027344, "learning_rate": 6.566666666666666e-07, "loss": 0.1762, "step": 3350 }, { "epoch": 0.06588235294117648, "grad_norm": 2.4327175617218018, "learning_rate": 6.586274509803921e-07, "loss": 0.1936, "step": 3360 }, { "epoch": 0.06607843137254903, "grad_norm": 3.2735559940338135, "learning_rate": 6.605882352941177e-07, "loss": 0.17, "step": 3370 }, { "epoch": 0.06627450980392156, "grad_norm": 7.041193008422852, "learning_rate": 6.625490196078432e-07, "loss": 0.2694, "step": 3380 }, { "epoch": 0.06647058823529411, "grad_norm": 3.8683226108551025, "learning_rate": 6.645098039215686e-07, "loss": 0.2015, "step": 3390 }, { "epoch": 0.06666666666666667, "grad_norm": 1.9060108661651611, "learning_rate": 6.664705882352941e-07, "loss": 0.1762, "step": 3400 }, { "epoch": 0.06686274509803922, "grad_norm": 3.5043435096740723, "learning_rate": 6.684313725490196e-07, "loss": 0.2202, "step": 3410 }, { "epoch": 0.06705882352941177, "grad_norm": 3.7755308151245117, "learning_rate": 6.703921568627451e-07, "loss": 0.2012, "step": 3420 }, { "epoch": 0.06725490196078432, "grad_norm": 4.484972953796387, "learning_rate": 6.723529411764706e-07, "loss": 0.2258, "step": 3430 }, { "epoch": 0.06745098039215686, "grad_norm": 2.8117787837982178, "learning_rate": 6.74313725490196e-07, "loss": 0.1835, "step": 3440 }, { "epoch": 0.06764705882352941, "grad_norm": 1.9734784364700317, "learning_rate": 6.762745098039216e-07, "loss": 0.2034, "step": 3450 }, { "epoch": 0.06784313725490196, "grad_norm": 2.5098302364349365, "learning_rate": 6.78235294117647e-07, "loss": 0.1781, "step": 3460 }, { "epoch": 0.06803921568627451, "grad_norm": 2.2713403701782227, "learning_rate": 6.801960784313725e-07, "loss": 0.2301, "step": 3470 }, { "epoch": 0.06823529411764706, "grad_norm": 2.221928834915161, "learning_rate": 6.821568627450979e-07, "loss": 0.2117, "step": 3480 }, { "epoch": 0.06843137254901961, "grad_norm": 3.215176820755005, "learning_rate": 6.841176470588235e-07, "loss": 0.1904, "step": 3490 }, { "epoch": 0.06862745098039216, "grad_norm": 2.774852991104126, "learning_rate": 6.860784313725489e-07, "loss": 0.1833, "step": 3500 }, { "epoch": 0.0688235294117647, "grad_norm": 3.603203058242798, "learning_rate": 6.880392156862744e-07, "loss": 0.1954, "step": 3510 }, { "epoch": 0.06901960784313725, "grad_norm": 2.9706196784973145, "learning_rate": 6.9e-07, "loss": 0.1716, "step": 3520 }, { "epoch": 0.0692156862745098, "grad_norm": 2.3297643661499023, "learning_rate": 6.919607843137255e-07, "loss": 0.2131, "step": 3530 }, { "epoch": 0.06941176470588235, "grad_norm": 3.163532257080078, "learning_rate": 6.93921568627451e-07, "loss": 0.162, "step": 3540 }, { "epoch": 0.0696078431372549, "grad_norm": 3.2276830673217773, "learning_rate": 6.958823529411764e-07, "loss": 0.2223, "step": 3550 }, { "epoch": 0.06980392156862746, "grad_norm": 2.3361899852752686, "learning_rate": 6.97843137254902e-07, "loss": 0.1836, "step": 3560 }, { "epoch": 0.07, "grad_norm": 3.4899144172668457, "learning_rate": 6.998039215686274e-07, "loss": 0.1938, "step": 3570 }, { "epoch": 0.07019607843137254, "grad_norm": 3.9240634441375732, "learning_rate": 7.017647058823529e-07, "loss": 0.1958, "step": 3580 }, { "epoch": 0.0703921568627451, "grad_norm": 1.53433358669281, "learning_rate": 7.037254901960785e-07, "loss": 0.1441, "step": 3590 }, { "epoch": 0.07058823529411765, "grad_norm": 2.5721330642700195, "learning_rate": 7.056862745098039e-07, "loss": 0.202, "step": 3600 }, { "epoch": 0.0707843137254902, "grad_norm": 0.9262362122535706, "learning_rate": 7.076470588235294e-07, "loss": 0.1822, "step": 3610 }, { "epoch": 0.07098039215686275, "grad_norm": 3.6266536712646484, "learning_rate": 7.096078431372548e-07, "loss": 0.1724, "step": 3620 }, { "epoch": 0.0711764705882353, "grad_norm": 4.926189422607422, "learning_rate": 7.115686274509804e-07, "loss": 0.18, "step": 3630 }, { "epoch": 0.07137254901960784, "grad_norm": 2.8171019554138184, "learning_rate": 7.135294117647058e-07, "loss": 0.1889, "step": 3640 }, { "epoch": 0.07156862745098039, "grad_norm": 3.5256845951080322, "learning_rate": 7.154901960784313e-07, "loss": 0.1796, "step": 3650 }, { "epoch": 0.07176470588235294, "grad_norm": 1.223862886428833, "learning_rate": 7.174509803921568e-07, "loss": 0.1768, "step": 3660 }, { "epoch": 0.07196078431372549, "grad_norm": 3.4330637454986572, "learning_rate": 7.194117647058823e-07, "loss": 0.2006, "step": 3670 }, { "epoch": 0.07215686274509804, "grad_norm": 3.3045032024383545, "learning_rate": 7.213725490196078e-07, "loss": 0.2032, "step": 3680 }, { "epoch": 0.07235294117647059, "grad_norm": 4.276102542877197, "learning_rate": 7.233333333333333e-07, "loss": 0.2072, "step": 3690 }, { "epoch": 0.07254901960784314, "grad_norm": 2.233032464981079, "learning_rate": 7.252941176470589e-07, "loss": 0.2012, "step": 3700 }, { "epoch": 0.07274509803921568, "grad_norm": 2.553008794784546, "learning_rate": 7.272549019607843e-07, "loss": 0.1998, "step": 3710 }, { "epoch": 0.07294117647058823, "grad_norm": 2.5650973320007324, "learning_rate": 7.292156862745098e-07, "loss": 0.2159, "step": 3720 }, { "epoch": 0.07313725490196078, "grad_norm": 2.661440849304199, "learning_rate": 7.311764705882352e-07, "loss": 0.1894, "step": 3730 }, { "epoch": 0.07333333333333333, "grad_norm": 3.3190484046936035, "learning_rate": 7.331372549019608e-07, "loss": 0.2045, "step": 3740 }, { "epoch": 0.07352941176470588, "grad_norm": 3.263941526412964, "learning_rate": 7.350980392156863e-07, "loss": 0.2056, "step": 3750 }, { "epoch": 0.07372549019607844, "grad_norm": 4.04332971572876, "learning_rate": 7.370588235294117e-07, "loss": 0.1816, "step": 3760 }, { "epoch": 0.07392156862745099, "grad_norm": 3.6129958629608154, "learning_rate": 7.390196078431372e-07, "loss": 0.1585, "step": 3770 }, { "epoch": 0.07411764705882352, "grad_norm": 3.8102896213531494, "learning_rate": 7.409803921568627e-07, "loss": 0.19, "step": 3780 }, { "epoch": 0.07431372549019608, "grad_norm": 2.430119037628174, "learning_rate": 7.429411764705882e-07, "loss": 0.1814, "step": 3790 }, { "epoch": 0.07450980392156863, "grad_norm": 0.6968156099319458, "learning_rate": 7.449019607843136e-07, "loss": 0.217, "step": 3800 }, { "epoch": 0.07470588235294118, "grad_norm": 1.9729390144348145, "learning_rate": 7.468627450980392e-07, "loss": 0.1906, "step": 3810 }, { "epoch": 0.07490196078431373, "grad_norm": 2.616109848022461, "learning_rate": 7.488235294117647e-07, "loss": 0.2282, "step": 3820 }, { "epoch": 0.07509803921568628, "grad_norm": 2.958254337310791, "learning_rate": 7.507843137254901e-07, "loss": 0.1825, "step": 3830 }, { "epoch": 0.07529411764705882, "grad_norm": 1.705356478691101, "learning_rate": 7.527450980392156e-07, "loss": 0.2098, "step": 3840 }, { "epoch": 0.07549019607843137, "grad_norm": 4.266743183135986, "learning_rate": 7.547058823529412e-07, "loss": 0.2063, "step": 3850 }, { "epoch": 0.07568627450980392, "grad_norm": 2.1829326152801514, "learning_rate": 7.566666666666667e-07, "loss": 0.1562, "step": 3860 }, { "epoch": 0.07588235294117647, "grad_norm": 2.2065250873565674, "learning_rate": 7.586274509803921e-07, "loss": 0.1861, "step": 3870 }, { "epoch": 0.07607843137254902, "grad_norm": 1.6659119129180908, "learning_rate": 7.605882352941176e-07, "loss": 0.1954, "step": 3880 }, { "epoch": 0.07627450980392157, "grad_norm": 3.2298014163970947, "learning_rate": 7.625490196078432e-07, "loss": 0.1613, "step": 3890 }, { "epoch": 0.07647058823529412, "grad_norm": 3.0127077102661133, "learning_rate": 7.645098039215686e-07, "loss": 0.2124, "step": 3900 }, { "epoch": 0.07666666666666666, "grad_norm": 2.557715654373169, "learning_rate": 7.664705882352941e-07, "loss": 0.2099, "step": 3910 }, { "epoch": 0.07686274509803921, "grad_norm": 2.4278197288513184, "learning_rate": 7.684313725490196e-07, "loss": 0.1977, "step": 3920 }, { "epoch": 0.07705882352941176, "grad_norm": 1.5481489896774292, "learning_rate": 7.703921568627451e-07, "loss": 0.2051, "step": 3930 }, { "epoch": 0.07725490196078431, "grad_norm": 3.577603340148926, "learning_rate": 7.723529411764705e-07, "loss": 0.1999, "step": 3940 }, { "epoch": 0.07745098039215687, "grad_norm": 3.82441782951355, "learning_rate": 7.74313725490196e-07, "loss": 0.2388, "step": 3950 }, { "epoch": 0.07764705882352942, "grad_norm": 2.902674913406372, "learning_rate": 7.762745098039215e-07, "loss": 0.2174, "step": 3960 }, { "epoch": 0.07784313725490197, "grad_norm": 1.820460319519043, "learning_rate": 7.78235294117647e-07, "loss": 0.1865, "step": 3970 }, { "epoch": 0.0780392156862745, "grad_norm": 4.0797529220581055, "learning_rate": 7.801960784313725e-07, "loss": 0.2463, "step": 3980 }, { "epoch": 0.07823529411764706, "grad_norm": 4.631940841674805, "learning_rate": 7.821568627450979e-07, "loss": 0.1847, "step": 3990 }, { "epoch": 0.0784313725490196, "grad_norm": 3.4990079402923584, "learning_rate": 7.841176470588235e-07, "loss": 0.2029, "step": 4000 }, { "epoch": 0.07862745098039216, "grad_norm": 2.4234843254089355, "learning_rate": 7.86078431372549e-07, "loss": 0.1586, "step": 4010 }, { "epoch": 0.07882352941176471, "grad_norm": 2.4945619106292725, "learning_rate": 7.880392156862745e-07, "loss": 0.2208, "step": 4020 }, { "epoch": 0.07901960784313726, "grad_norm": 11.397479057312012, "learning_rate": 7.9e-07, "loss": 0.2112, "step": 4030 }, { "epoch": 0.0792156862745098, "grad_norm": 3.4011662006378174, "learning_rate": 7.919607843137255e-07, "loss": 0.2131, "step": 4040 }, { "epoch": 0.07941176470588235, "grad_norm": 2.7123970985412598, "learning_rate": 7.93921568627451e-07, "loss": 0.1414, "step": 4050 }, { "epoch": 0.0796078431372549, "grad_norm": 1.3853148221969604, "learning_rate": 7.958823529411764e-07, "loss": 0.17, "step": 4060 }, { "epoch": 0.07980392156862745, "grad_norm": 2.4156007766723633, "learning_rate": 7.97843137254902e-07, "loss": 0.2102, "step": 4070 }, { "epoch": 0.08, "grad_norm": 1.761945128440857, "learning_rate": 7.998039215686274e-07, "loss": 0.1848, "step": 4080 }, { "epoch": 0.08019607843137255, "grad_norm": 1.758612036705017, "learning_rate": 8.017647058823529e-07, "loss": 0.2183, "step": 4090 }, { "epoch": 0.0803921568627451, "grad_norm": 3.965364694595337, "learning_rate": 8.037254901960783e-07, "loss": 0.1887, "step": 4100 }, { "epoch": 0.08058823529411764, "grad_norm": 3.4807605743408203, "learning_rate": 8.056862745098039e-07, "loss": 0.1667, "step": 4110 }, { "epoch": 0.08078431372549019, "grad_norm": 2.3339781761169434, "learning_rate": 8.076470588235293e-07, "loss": 0.1803, "step": 4120 }, { "epoch": 0.08098039215686274, "grad_norm": 5.0291218757629395, "learning_rate": 8.096078431372548e-07, "loss": 0.1884, "step": 4130 }, { "epoch": 0.0811764705882353, "grad_norm": 2.3535830974578857, "learning_rate": 8.115686274509804e-07, "loss": 0.1839, "step": 4140 }, { "epoch": 0.08137254901960785, "grad_norm": 3.1544597148895264, "learning_rate": 8.135294117647058e-07, "loss": 0.2072, "step": 4150 }, { "epoch": 0.0815686274509804, "grad_norm": 1.3231886625289917, "learning_rate": 8.154901960784313e-07, "loss": 0.15, "step": 4160 }, { "epoch": 0.08176470588235295, "grad_norm": 2.2060279846191406, "learning_rate": 8.174509803921568e-07, "loss": 0.1796, "step": 4170 }, { "epoch": 0.08196078431372548, "grad_norm": 2.981175422668457, "learning_rate": 8.194117647058824e-07, "loss": 0.2093, "step": 4180 }, { "epoch": 0.08215686274509804, "grad_norm": 2.78774094581604, "learning_rate": 8.213725490196078e-07, "loss": 0.2178, "step": 4190 }, { "epoch": 0.08235294117647059, "grad_norm": 2.590579032897949, "learning_rate": 8.233333333333333e-07, "loss": 0.1941, "step": 4200 }, { "epoch": 0.08254901960784314, "grad_norm": 2.0308637619018555, "learning_rate": 8.252941176470589e-07, "loss": 0.1581, "step": 4210 }, { "epoch": 0.08274509803921569, "grad_norm": 2.994051694869995, "learning_rate": 8.272549019607843e-07, "loss": 0.2133, "step": 4220 }, { "epoch": 0.08294117647058824, "grad_norm": 1.8066377639770508, "learning_rate": 8.292156862745098e-07, "loss": 0.1584, "step": 4230 }, { "epoch": 0.08313725490196078, "grad_norm": 3.273902654647827, "learning_rate": 8.311764705882352e-07, "loss": 0.1749, "step": 4240 }, { "epoch": 0.08333333333333333, "grad_norm": 2.153937339782715, "learning_rate": 8.331372549019608e-07, "loss": 0.2252, "step": 4250 }, { "epoch": 0.08352941176470588, "grad_norm": 3.8431806564331055, "learning_rate": 8.350980392156862e-07, "loss": 0.1895, "step": 4260 }, { "epoch": 0.08372549019607843, "grad_norm": 2.6749844551086426, "learning_rate": 8.370588235294117e-07, "loss": 0.164, "step": 4270 }, { "epoch": 0.08392156862745098, "grad_norm": 2.8832712173461914, "learning_rate": 8.390196078431372e-07, "loss": 0.2099, "step": 4280 }, { "epoch": 0.08411764705882353, "grad_norm": 3.3181400299072266, "learning_rate": 8.409803921568627e-07, "loss": 0.2023, "step": 4290 }, { "epoch": 0.08431372549019608, "grad_norm": 3.0225913524627686, "learning_rate": 8.429411764705882e-07, "loss": 0.1834, "step": 4300 }, { "epoch": 0.08450980392156862, "grad_norm": 3.4151864051818848, "learning_rate": 8.449019607843136e-07, "loss": 0.1567, "step": 4310 }, { "epoch": 0.08470588235294117, "grad_norm": 2.2450852394104004, "learning_rate": 8.468627450980391e-07, "loss": 0.18, "step": 4320 }, { "epoch": 0.08490196078431372, "grad_norm": 3.1525487899780273, "learning_rate": 8.488235294117647e-07, "loss": 0.2024, "step": 4330 }, { "epoch": 0.08509803921568627, "grad_norm": 2.351762294769287, "learning_rate": 8.507843137254902e-07, "loss": 0.2028, "step": 4340 }, { "epoch": 0.08529411764705883, "grad_norm": 2.759901285171509, "learning_rate": 8.527450980392157e-07, "loss": 0.2074, "step": 4350 }, { "epoch": 0.08549019607843138, "grad_norm": 1.8655637502670288, "learning_rate": 8.547058823529412e-07, "loss": 0.2386, "step": 4360 }, { "epoch": 0.08568627450980393, "grad_norm": 2.5192418098449707, "learning_rate": 8.566666666666667e-07, "loss": 0.2068, "step": 4370 }, { "epoch": 0.08588235294117647, "grad_norm": 2.775587797164917, "learning_rate": 8.586274509803921e-07, "loss": 0.1844, "step": 4380 }, { "epoch": 0.08607843137254902, "grad_norm": 1.915915846824646, "learning_rate": 8.605882352941176e-07, "loss": 0.2149, "step": 4390 }, { "epoch": 0.08627450980392157, "grad_norm": 2.014655590057373, "learning_rate": 8.625490196078431e-07, "loss": 0.1745, "step": 4400 }, { "epoch": 0.08647058823529412, "grad_norm": 2.3732750415802, "learning_rate": 8.645098039215686e-07, "loss": 0.1771, "step": 4410 }, { "epoch": 0.08666666666666667, "grad_norm": 2.5433192253112793, "learning_rate": 8.66470588235294e-07, "loss": 0.2197, "step": 4420 }, { "epoch": 0.08686274509803922, "grad_norm": 2.1101813316345215, "learning_rate": 8.684313725490195e-07, "loss": 0.179, "step": 4430 }, { "epoch": 0.08705882352941176, "grad_norm": 4.146600246429443, "learning_rate": 8.703921568627451e-07, "loss": 0.216, "step": 4440 }, { "epoch": 0.08725490196078431, "grad_norm": 2.5137178897857666, "learning_rate": 8.723529411764705e-07, "loss": 0.2079, "step": 4450 }, { "epoch": 0.08745098039215686, "grad_norm": 4.309109210968018, "learning_rate": 8.74313725490196e-07, "loss": 0.1847, "step": 4460 }, { "epoch": 0.08764705882352941, "grad_norm": 1.5906715393066406, "learning_rate": 8.762745098039215e-07, "loss": 0.1818, "step": 4470 }, { "epoch": 0.08784313725490196, "grad_norm": 2.1933562755584717, "learning_rate": 8.782352941176471e-07, "loss": 0.1929, "step": 4480 }, { "epoch": 0.08803921568627451, "grad_norm": 2.9720571041107178, "learning_rate": 8.801960784313725e-07, "loss": 0.1685, "step": 4490 }, { "epoch": 0.08823529411764706, "grad_norm": 2.51102614402771, "learning_rate": 8.82156862745098e-07, "loss": 0.1826, "step": 4500 }, { "epoch": 0.0884313725490196, "grad_norm": 3.0620720386505127, "learning_rate": 8.841176470588236e-07, "loss": 0.1904, "step": 4510 }, { "epoch": 0.08862745098039215, "grad_norm": 3.29909086227417, "learning_rate": 8.86078431372549e-07, "loss": 0.2093, "step": 4520 }, { "epoch": 0.0888235294117647, "grad_norm": 2.660931348800659, "learning_rate": 8.880392156862745e-07, "loss": 0.2238, "step": 4530 }, { "epoch": 0.08901960784313726, "grad_norm": 2.226191282272339, "learning_rate": 8.9e-07, "loss": 0.1929, "step": 4540 }, { "epoch": 0.0892156862745098, "grad_norm": 2.8439109325408936, "learning_rate": 8.919607843137255e-07, "loss": 0.1878, "step": 4550 }, { "epoch": 0.08941176470588236, "grad_norm": 5.07834005355835, "learning_rate": 8.939215686274509e-07, "loss": 0.1885, "step": 4560 }, { "epoch": 0.08960784313725491, "grad_norm": 3.4411540031433105, "learning_rate": 8.958823529411764e-07, "loss": 0.1924, "step": 4570 }, { "epoch": 0.08980392156862745, "grad_norm": 2.620816469192505, "learning_rate": 8.978431372549019e-07, "loss": 0.1865, "step": 4580 }, { "epoch": 0.09, "grad_norm": 1.4460548162460327, "learning_rate": 8.998039215686274e-07, "loss": 0.176, "step": 4590 }, { "epoch": 0.09019607843137255, "grad_norm": 4.028953552246094, "learning_rate": 9.017647058823529e-07, "loss": 0.1828, "step": 4600 }, { "epoch": 0.0903921568627451, "grad_norm": 3.388392686843872, "learning_rate": 9.037254901960783e-07, "loss": 0.1951, "step": 4610 }, { "epoch": 0.09058823529411765, "grad_norm": 3.1092123985290527, "learning_rate": 9.056862745098039e-07, "loss": 0.1913, "step": 4620 }, { "epoch": 0.0907843137254902, "grad_norm": 2.6942882537841797, "learning_rate": 9.076470588235293e-07, "loss": 0.1856, "step": 4630 }, { "epoch": 0.09098039215686274, "grad_norm": 1.666735053062439, "learning_rate": 9.096078431372549e-07, "loss": 0.2372, "step": 4640 }, { "epoch": 0.09117647058823529, "grad_norm": 3.609112024307251, "learning_rate": 9.115686274509804e-07, "loss": 0.2062, "step": 4650 }, { "epoch": 0.09137254901960784, "grad_norm": 1.8214279413223267, "learning_rate": 9.135294117647059e-07, "loss": 0.1797, "step": 4660 }, { "epoch": 0.09156862745098039, "grad_norm": 1.6812957525253296, "learning_rate": 9.154901960784314e-07, "loss": 0.1967, "step": 4670 }, { "epoch": 0.09176470588235294, "grad_norm": 2.225823163986206, "learning_rate": 9.174509803921568e-07, "loss": 0.1743, "step": 4680 }, { "epoch": 0.0919607843137255, "grad_norm": 3.5805299282073975, "learning_rate": 9.194117647058824e-07, "loss": 0.181, "step": 4690 }, { "epoch": 0.09215686274509804, "grad_norm": 3.1303577423095703, "learning_rate": 9.213725490196078e-07, "loss": 0.2235, "step": 4700 }, { "epoch": 0.09235294117647058, "grad_norm": 2.173021078109741, "learning_rate": 9.233333333333333e-07, "loss": 0.1848, "step": 4710 }, { "epoch": 0.09254901960784313, "grad_norm": 2.808845281600952, "learning_rate": 9.252941176470587e-07, "loss": 0.1515, "step": 4720 }, { "epoch": 0.09274509803921568, "grad_norm": 2.0805037021636963, "learning_rate": 9.272549019607843e-07, "loss": 0.1852, "step": 4730 }, { "epoch": 0.09294117647058824, "grad_norm": 3.027742862701416, "learning_rate": 9.292156862745098e-07, "loss": 0.1802, "step": 4740 }, { "epoch": 0.09313725490196079, "grad_norm": 2.9263851642608643, "learning_rate": 9.311764705882352e-07, "loss": 0.1803, "step": 4750 }, { "epoch": 0.09333333333333334, "grad_norm": 4.127167224884033, "learning_rate": 9.331372549019608e-07, "loss": 0.1552, "step": 4760 }, { "epoch": 0.09352941176470589, "grad_norm": 2.805945634841919, "learning_rate": 9.350980392156862e-07, "loss": 0.1746, "step": 4770 }, { "epoch": 0.09372549019607843, "grad_norm": 2.7245898246765137, "learning_rate": 9.370588235294117e-07, "loss": 0.2204, "step": 4780 }, { "epoch": 0.09392156862745098, "grad_norm": 2.033198356628418, "learning_rate": 9.390196078431371e-07, "loss": 0.2087, "step": 4790 }, { "epoch": 0.09411764705882353, "grad_norm": 1.9819586277008057, "learning_rate": 9.409803921568628e-07, "loss": 0.2013, "step": 4800 }, { "epoch": 0.09431372549019608, "grad_norm": 2.4850988388061523, "learning_rate": 9.429411764705882e-07, "loss": 0.1785, "step": 4810 }, { "epoch": 0.09450980392156863, "grad_norm": 2.142763137817383, "learning_rate": 9.449019607843137e-07, "loss": 0.2141, "step": 4820 }, { "epoch": 0.09470588235294118, "grad_norm": 3.120683431625366, "learning_rate": 9.468627450980392e-07, "loss": 0.216, "step": 4830 }, { "epoch": 0.09490196078431372, "grad_norm": 4.3091959953308105, "learning_rate": 9.488235294117647e-07, "loss": 0.2032, "step": 4840 }, { "epoch": 0.09509803921568627, "grad_norm": 2.032409191131592, "learning_rate": 9.507843137254902e-07, "loss": 0.1857, "step": 4850 }, { "epoch": 0.09529411764705882, "grad_norm": 1.520799160003662, "learning_rate": 9.527450980392156e-07, "loss": 0.1781, "step": 4860 }, { "epoch": 0.09549019607843137, "grad_norm": 2.805920362472534, "learning_rate": 9.547058823529412e-07, "loss": 0.1915, "step": 4870 }, { "epoch": 0.09568627450980392, "grad_norm": 2.4971210956573486, "learning_rate": 9.566666666666667e-07, "loss": 0.2094, "step": 4880 }, { "epoch": 0.09588235294117647, "grad_norm": 4.3101806640625, "learning_rate": 9.58627450980392e-07, "loss": 0.176, "step": 4890 }, { "epoch": 0.09607843137254903, "grad_norm": 3.221756935119629, "learning_rate": 9.605882352941175e-07, "loss": 0.1867, "step": 4900 }, { "epoch": 0.09627450980392156, "grad_norm": 2.1800642013549805, "learning_rate": 9.62549019607843e-07, "loss": 0.2026, "step": 4910 }, { "epoch": 0.09647058823529411, "grad_norm": 3.8331825733184814, "learning_rate": 9.645098039215686e-07, "loss": 0.186, "step": 4920 }, { "epoch": 0.09666666666666666, "grad_norm": 3.116273880004883, "learning_rate": 9.664705882352941e-07, "loss": 0.1848, "step": 4930 }, { "epoch": 0.09686274509803922, "grad_norm": 2.3887856006622314, "learning_rate": 9.684313725490194e-07, "loss": 0.1978, "step": 4940 }, { "epoch": 0.09705882352941177, "grad_norm": 3.013380765914917, "learning_rate": 9.70392156862745e-07, "loss": 0.1805, "step": 4950 }, { "epoch": 0.09725490196078432, "grad_norm": 3.318415641784668, "learning_rate": 9.723529411764705e-07, "loss": 0.1613, "step": 4960 }, { "epoch": 0.09745098039215687, "grad_norm": 3.6313235759735107, "learning_rate": 9.74313725490196e-07, "loss": 0.1718, "step": 4970 }, { "epoch": 0.0976470588235294, "grad_norm": 2.0793163776397705, "learning_rate": 9.762745098039216e-07, "loss": 0.212, "step": 4980 }, { "epoch": 0.09784313725490196, "grad_norm": 4.086233139038086, "learning_rate": 9.78235294117647e-07, "loss": 0.2098, "step": 4990 }, { "epoch": 0.09803921568627451, "grad_norm": 3.3871865272521973, "learning_rate": 9.801960784313726e-07, "loss": 0.1482, "step": 5000 }, { "epoch": 0.09823529411764706, "grad_norm": 3.2505223751068115, "learning_rate": 9.82156862745098e-07, "loss": 0.2034, "step": 5010 }, { "epoch": 0.09843137254901961, "grad_norm": 2.6460282802581787, "learning_rate": 9.841176470588235e-07, "loss": 0.1877, "step": 5020 }, { "epoch": 0.09862745098039216, "grad_norm": 4.048020839691162, "learning_rate": 9.86078431372549e-07, "loss": 0.1924, "step": 5030 }, { "epoch": 0.0988235294117647, "grad_norm": 2.5681862831115723, "learning_rate": 9.880392156862745e-07, "loss": 0.1859, "step": 5040 }, { "epoch": 0.09901960784313725, "grad_norm": 2.6397156715393066, "learning_rate": 9.9e-07, "loss": 0.2139, "step": 5050 }, { "epoch": 0.0992156862745098, "grad_norm": 2.250730276107788, "learning_rate": 9.919607843137254e-07, "loss": 0.1976, "step": 5060 }, { "epoch": 0.09941176470588235, "grad_norm": 2.408060073852539, "learning_rate": 9.93921568627451e-07, "loss": 0.2074, "step": 5070 }, { "epoch": 0.0996078431372549, "grad_norm": 3.9261326789855957, "learning_rate": 9.958823529411764e-07, "loss": 0.1884, "step": 5080 }, { "epoch": 0.09980392156862745, "grad_norm": 2.345799684524536, "learning_rate": 9.97843137254902e-07, "loss": 0.2225, "step": 5090 }, { "epoch": 0.1, "grad_norm": 1.9670310020446777, "learning_rate": 9.998039215686273e-07, "loss": 0.1799, "step": 5100 }, { "epoch": 0.10019607843137254, "grad_norm": 4.0252556800842285, "learning_rate": 9.999999787148254e-07, "loss": 0.1925, "step": 5110 }, { "epoch": 0.1003921568627451, "grad_norm": 7.67744255065918, "learning_rate": 9.999999051364466e-07, "loss": 0.2365, "step": 5120 }, { "epoch": 0.10058823529411764, "grad_norm": 3.654590368270874, "learning_rate": 9.999997790020912e-07, "loss": 0.1871, "step": 5130 }, { "epoch": 0.1007843137254902, "grad_norm": 1.170494794845581, "learning_rate": 9.99999600311773e-07, "loss": 0.1884, "step": 5140 }, { "epoch": 0.10098039215686275, "grad_norm": 5.212301254272461, "learning_rate": 9.9999936906551e-07, "loss": 0.2192, "step": 5150 }, { "epoch": 0.1011764705882353, "grad_norm": 3.0794801712036133, "learning_rate": 9.99999085263327e-07, "loss": 0.1764, "step": 5160 }, { "epoch": 0.10137254901960785, "grad_norm": 4.881508827209473, "learning_rate": 9.999987489052538e-07, "loss": 0.1827, "step": 5170 }, { "epoch": 0.10156862745098039, "grad_norm": 1.7163325548171997, "learning_rate": 9.999983599913257e-07, "loss": 0.1753, "step": 5180 }, { "epoch": 0.10176470588235294, "grad_norm": 2.107046365737915, "learning_rate": 9.999979185215834e-07, "loss": 0.1959, "step": 5190 }, { "epoch": 0.10196078431372549, "grad_norm": 4.8714470863342285, "learning_rate": 9.999974244960737e-07, "loss": 0.2226, "step": 5200 }, { "epoch": 0.10215686274509804, "grad_norm": 4.2069573402404785, "learning_rate": 9.99996877914848e-07, "loss": 0.2033, "step": 5210 }, { "epoch": 0.10235294117647059, "grad_norm": 2.3682830333709717, "learning_rate": 9.999962787779642e-07, "loss": 0.1871, "step": 5220 }, { "epoch": 0.10254901960784314, "grad_norm": 2.677098035812378, "learning_rate": 9.99995627085485e-07, "loss": 0.1725, "step": 5230 }, { "epoch": 0.1027450980392157, "grad_norm": 2.5124852657318115, "learning_rate": 9.99994922837479e-07, "loss": 0.1579, "step": 5240 }, { "epoch": 0.10294117647058823, "grad_norm": 4.2292914390563965, "learning_rate": 9.999941660340204e-07, "loss": 0.1666, "step": 5250 }, { "epoch": 0.10313725490196078, "grad_norm": 3.0313010215759277, "learning_rate": 9.999933566751886e-07, "loss": 0.1715, "step": 5260 }, { "epoch": 0.10333333333333333, "grad_norm": 7.285390377044678, "learning_rate": 9.999924947610685e-07, "loss": 0.1911, "step": 5270 }, { "epoch": 0.10352941176470588, "grad_norm": 3.2727975845336914, "learning_rate": 9.999915802917509e-07, "loss": 0.1866, "step": 5280 }, { "epoch": 0.10372549019607843, "grad_norm": 2.407641887664795, "learning_rate": 9.99990613267332e-07, "loss": 0.1985, "step": 5290 }, { "epoch": 0.10392156862745099, "grad_norm": 2.6784958839416504, "learning_rate": 9.999895936879132e-07, "loss": 0.2213, "step": 5300 }, { "epoch": 0.10411764705882352, "grad_norm": 2.677206039428711, "learning_rate": 9.999885215536018e-07, "loss": 0.1971, "step": 5310 }, { "epoch": 0.10431372549019607, "grad_norm": 4.839699745178223, "learning_rate": 9.999873968645108e-07, "loss": 0.1929, "step": 5320 }, { "epoch": 0.10450980392156863, "grad_norm": 3.209017276763916, "learning_rate": 9.999862196207579e-07, "loss": 0.1904, "step": 5330 }, { "epoch": 0.10470588235294118, "grad_norm": 1.8858166933059692, "learning_rate": 9.99984989822467e-07, "loss": 0.1909, "step": 5340 }, { "epoch": 0.10490196078431373, "grad_norm": 10.797245979309082, "learning_rate": 9.999837074697673e-07, "loss": 0.1721, "step": 5350 }, { "epoch": 0.10509803921568628, "grad_norm": 3.5454201698303223, "learning_rate": 9.999823725627938e-07, "loss": 0.1354, "step": 5360 }, { "epoch": 0.10529411764705883, "grad_norm": 2.7670021057128906, "learning_rate": 9.999809851016868e-07, "loss": 0.1855, "step": 5370 }, { "epoch": 0.10549019607843137, "grad_norm": 3.280217409133911, "learning_rate": 9.999795450865922e-07, "loss": 0.175, "step": 5380 }, { "epoch": 0.10568627450980392, "grad_norm": 2.429615020751953, "learning_rate": 9.99978052517661e-07, "loss": 0.1996, "step": 5390 }, { "epoch": 0.10588235294117647, "grad_norm": 2.2265820503234863, "learning_rate": 9.999765073950504e-07, "loss": 0.2143, "step": 5400 }, { "epoch": 0.10607843137254902, "grad_norm": 2.4983999729156494, "learning_rate": 9.999749097189227e-07, "loss": 0.1779, "step": 5410 }, { "epoch": 0.10627450980392157, "grad_norm": 2.057594060897827, "learning_rate": 9.99973259489446e-07, "loss": 0.1861, "step": 5420 }, { "epoch": 0.10647058823529412, "grad_norm": 2.370163917541504, "learning_rate": 9.999715567067934e-07, "loss": 0.1876, "step": 5430 }, { "epoch": 0.10666666666666667, "grad_norm": 3.22011137008667, "learning_rate": 9.999698013711442e-07, "loss": 0.1693, "step": 5440 }, { "epoch": 0.10686274509803921, "grad_norm": 2.6384687423706055, "learning_rate": 9.999679934826828e-07, "loss": 0.1969, "step": 5450 }, { "epoch": 0.10705882352941176, "grad_norm": 2.72566294670105, "learning_rate": 9.99966133041599e-07, "loss": 0.2289, "step": 5460 }, { "epoch": 0.10725490196078431, "grad_norm": 3.7704033851623535, "learning_rate": 9.99964220048089e-07, "loss": 0.205, "step": 5470 }, { "epoch": 0.10745098039215686, "grad_norm": 3.276524543762207, "learning_rate": 9.999622545023535e-07, "loss": 0.1783, "step": 5480 }, { "epoch": 0.10764705882352942, "grad_norm": 3.3527612686157227, "learning_rate": 9.999602364045988e-07, "loss": 0.2002, "step": 5490 }, { "epoch": 0.10784313725490197, "grad_norm": 1.9396904706954956, "learning_rate": 9.999581657550373e-07, "loss": 0.1689, "step": 5500 }, { "epoch": 0.1080392156862745, "grad_norm": 2.1544265747070312, "learning_rate": 9.999560425538867e-07, "loss": 0.202, "step": 5510 }, { "epoch": 0.10823529411764705, "grad_norm": 3.1238925457000732, "learning_rate": 9.999538668013701e-07, "loss": 0.1747, "step": 5520 }, { "epoch": 0.1084313725490196, "grad_norm": 2.3855512142181396, "learning_rate": 9.999516384977164e-07, "loss": 0.19, "step": 5530 }, { "epoch": 0.10862745098039216, "grad_norm": 2.756009817123413, "learning_rate": 9.999493576431595e-07, "loss": 0.1822, "step": 5540 }, { "epoch": 0.10882352941176471, "grad_norm": 3.241616725921631, "learning_rate": 9.999470242379392e-07, "loss": 0.1627, "step": 5550 }, { "epoch": 0.10901960784313726, "grad_norm": 1.8717703819274902, "learning_rate": 9.999446382823011e-07, "loss": 0.1594, "step": 5560 }, { "epoch": 0.10921568627450981, "grad_norm": 2.459388256072998, "learning_rate": 9.999421997764957e-07, "loss": 0.2359, "step": 5570 }, { "epoch": 0.10941176470588235, "grad_norm": 3.2836453914642334, "learning_rate": 9.99939708720779e-07, "loss": 0.1842, "step": 5580 }, { "epoch": 0.1096078431372549, "grad_norm": 17.066057205200195, "learning_rate": 9.999371651154135e-07, "loss": 0.189, "step": 5590 }, { "epoch": 0.10980392156862745, "grad_norm": 3.982179641723633, "learning_rate": 9.999345689606664e-07, "loss": 0.216, "step": 5600 }, { "epoch": 0.11, "grad_norm": 2.2547080516815186, "learning_rate": 9.999319202568103e-07, "loss": 0.1982, "step": 5610 }, { "epoch": 0.11019607843137255, "grad_norm": 2.9204049110412598, "learning_rate": 9.99929219004124e-07, "loss": 0.1809, "step": 5620 }, { "epoch": 0.1103921568627451, "grad_norm": 3.166696548461914, "learning_rate": 9.99926465202891e-07, "loss": 0.1685, "step": 5630 }, { "epoch": 0.11058823529411765, "grad_norm": 3.168828248977661, "learning_rate": 9.999236588534007e-07, "loss": 0.2015, "step": 5640 }, { "epoch": 0.11078431372549019, "grad_norm": 5.352186679840088, "learning_rate": 9.999207999559488e-07, "loss": 0.1914, "step": 5650 }, { "epoch": 0.11098039215686274, "grad_norm": 2.4359519481658936, "learning_rate": 9.99917888510835e-07, "loss": 0.238, "step": 5660 }, { "epoch": 0.1111764705882353, "grad_norm": 2.5077052116394043, "learning_rate": 9.999149245183658e-07, "loss": 0.161, "step": 5670 }, { "epoch": 0.11137254901960784, "grad_norm": 3.2428884506225586, "learning_rate": 9.999119079788527e-07, "loss": 0.2009, "step": 5680 }, { "epoch": 0.1115686274509804, "grad_norm": 2.6359174251556396, "learning_rate": 9.999088388926127e-07, "loss": 0.209, "step": 5690 }, { "epoch": 0.11176470588235295, "grad_norm": 2.603724479675293, "learning_rate": 9.999057172599682e-07, "loss": 0.1938, "step": 5700 }, { "epoch": 0.11196078431372548, "grad_norm": 5.197614669799805, "learning_rate": 9.999025430812476e-07, "loss": 0.1357, "step": 5710 }, { "epoch": 0.11215686274509803, "grad_norm": 1.9098247289657593, "learning_rate": 9.998993163567847e-07, "loss": 0.2094, "step": 5720 }, { "epoch": 0.11235294117647059, "grad_norm": 2.076608896255493, "learning_rate": 9.998960370869182e-07, "loss": 0.2001, "step": 5730 }, { "epoch": 0.11254901960784314, "grad_norm": 1.96406888961792, "learning_rate": 9.99892705271993e-07, "loss": 0.1837, "step": 5740 }, { "epoch": 0.11274509803921569, "grad_norm": 1.7358982563018799, "learning_rate": 9.998893209123593e-07, "loss": 0.1925, "step": 5750 }, { "epoch": 0.11294117647058824, "grad_norm": 2.5608649253845215, "learning_rate": 9.99885884008373e-07, "loss": 0.183, "step": 5760 }, { "epoch": 0.11313725490196079, "grad_norm": 1.3457411527633667, "learning_rate": 9.998823945603951e-07, "loss": 0.1735, "step": 5770 }, { "epoch": 0.11333333333333333, "grad_norm": 2.721052646636963, "learning_rate": 9.998788525687925e-07, "loss": 0.175, "step": 5780 }, { "epoch": 0.11352941176470588, "grad_norm": 3.5463204383850098, "learning_rate": 9.998752580339376e-07, "loss": 0.1535, "step": 5790 }, { "epoch": 0.11372549019607843, "grad_norm": 2.487586736679077, "learning_rate": 9.998716109562082e-07, "loss": 0.1761, "step": 5800 }, { "epoch": 0.11392156862745098, "grad_norm": 3.439589500427246, "learning_rate": 9.998679113359876e-07, "loss": 0.1827, "step": 5810 }, { "epoch": 0.11411764705882353, "grad_norm": 2.739125967025757, "learning_rate": 9.998641591736645e-07, "loss": 0.2203, "step": 5820 }, { "epoch": 0.11431372549019608, "grad_norm": 2.656378984451294, "learning_rate": 9.998603544696337e-07, "loss": 0.2073, "step": 5830 }, { "epoch": 0.11450980392156863, "grad_norm": 3.778212547302246, "learning_rate": 9.998564972242948e-07, "loss": 0.2275, "step": 5840 }, { "epoch": 0.11470588235294117, "grad_norm": 2.36609148979187, "learning_rate": 9.998525874380533e-07, "loss": 0.1744, "step": 5850 }, { "epoch": 0.11490196078431372, "grad_norm": 2.581596612930298, "learning_rate": 9.9984862511132e-07, "loss": 0.1943, "step": 5860 }, { "epoch": 0.11509803921568627, "grad_norm": 2.443136692047119, "learning_rate": 9.99844610244512e-07, "loss": 0.1813, "step": 5870 }, { "epoch": 0.11529411764705882, "grad_norm": 1.8129066228866577, "learning_rate": 9.998405428380506e-07, "loss": 0.2535, "step": 5880 }, { "epoch": 0.11549019607843138, "grad_norm": 2.275730609893799, "learning_rate": 9.998364228923637e-07, "loss": 0.1768, "step": 5890 }, { "epoch": 0.11568627450980393, "grad_norm": 3.0744359493255615, "learning_rate": 9.998322504078843e-07, "loss": 0.1899, "step": 5900 }, { "epoch": 0.11588235294117646, "grad_norm": 3.1373114585876465, "learning_rate": 9.998280253850509e-07, "loss": 0.2072, "step": 5910 }, { "epoch": 0.11607843137254902, "grad_norm": 2.627175807952881, "learning_rate": 9.998237478243077e-07, "loss": 0.2151, "step": 5920 }, { "epoch": 0.11627450980392157, "grad_norm": 3.6560373306274414, "learning_rate": 9.998194177261043e-07, "loss": 0.2169, "step": 5930 }, { "epoch": 0.11647058823529412, "grad_norm": 2.4112884998321533, "learning_rate": 9.99815035090896e-07, "loss": 0.1851, "step": 5940 }, { "epoch": 0.11666666666666667, "grad_norm": 1.601889729499817, "learning_rate": 9.99810599919143e-07, "loss": 0.1928, "step": 5950 }, { "epoch": 0.11686274509803922, "grad_norm": 2.5471951961517334, "learning_rate": 9.99806112211312e-07, "loss": 0.2134, "step": 5960 }, { "epoch": 0.11705882352941177, "grad_norm": 3.0749363899230957, "learning_rate": 9.998015719678743e-07, "loss": 0.2075, "step": 5970 }, { "epoch": 0.11725490196078431, "grad_norm": 2.0946547985076904, "learning_rate": 9.997969791893075e-07, "loss": 0.1914, "step": 5980 }, { "epoch": 0.11745098039215686, "grad_norm": 4.00174617767334, "learning_rate": 9.99792333876094e-07, "loss": 0.1988, "step": 5990 }, { "epoch": 0.11764705882352941, "grad_norm": 6.68895959854126, "learning_rate": 9.997876360287224e-07, "loss": 0.189, "step": 6000 }, { "epoch": 0.11784313725490196, "grad_norm": 2.791578531265259, "learning_rate": 9.997828856476865e-07, "loss": 0.148, "step": 6010 }, { "epoch": 0.11803921568627451, "grad_norm": 2.195007085800171, "learning_rate": 9.997780827334853e-07, "loss": 0.1517, "step": 6020 }, { "epoch": 0.11823529411764706, "grad_norm": 3.5669362545013428, "learning_rate": 9.99773227286624e-07, "loss": 0.181, "step": 6030 }, { "epoch": 0.11843137254901961, "grad_norm": 2.9197566509246826, "learning_rate": 9.997683193076126e-07, "loss": 0.2128, "step": 6040 }, { "epoch": 0.11862745098039215, "grad_norm": 2.623122453689575, "learning_rate": 9.997633587969674e-07, "loss": 0.1522, "step": 6050 }, { "epoch": 0.1188235294117647, "grad_norm": 2.0663273334503174, "learning_rate": 9.997583457552094e-07, "loss": 0.224, "step": 6060 }, { "epoch": 0.11901960784313725, "grad_norm": 2.2926177978515625, "learning_rate": 9.997532801828658e-07, "loss": 0.1746, "step": 6070 }, { "epoch": 0.1192156862745098, "grad_norm": 2.706284523010254, "learning_rate": 9.99748162080469e-07, "loss": 0.1938, "step": 6080 }, { "epoch": 0.11941176470588236, "grad_norm": 1.8832333087921143, "learning_rate": 9.997429914485568e-07, "loss": 0.2059, "step": 6090 }, { "epoch": 0.11960784313725491, "grad_norm": 1.9907909631729126, "learning_rate": 9.99737768287673e-07, "loss": 0.2018, "step": 6100 }, { "epoch": 0.11980392156862744, "grad_norm": 3.505500078201294, "learning_rate": 9.997324925983664e-07, "loss": 0.1677, "step": 6110 }, { "epoch": 0.12, "grad_norm": 3.0165932178497314, "learning_rate": 9.997271643811917e-07, "loss": 0.1638, "step": 6120 }, { "epoch": 0.12019607843137255, "grad_norm": 8.133214950561523, "learning_rate": 9.997217836367088e-07, "loss": 0.208, "step": 6130 }, { "epoch": 0.1203921568627451, "grad_norm": 3.9127750396728516, "learning_rate": 9.997163503654832e-07, "loss": 0.1838, "step": 6140 }, { "epoch": 0.12058823529411765, "grad_norm": 2.5933632850646973, "learning_rate": 9.997108645680863e-07, "loss": 0.179, "step": 6150 }, { "epoch": 0.1207843137254902, "grad_norm": 3.2774078845977783, "learning_rate": 9.997053262450947e-07, "loss": 0.1752, "step": 6160 }, { "epoch": 0.12098039215686275, "grad_norm": 2.042994737625122, "learning_rate": 9.996997353970902e-07, "loss": 0.1816, "step": 6170 }, { "epoch": 0.12117647058823529, "grad_norm": 3.729684829711914, "learning_rate": 9.996940920246608e-07, "loss": 0.191, "step": 6180 }, { "epoch": 0.12137254901960784, "grad_norm": 1.710037112236023, "learning_rate": 9.996883961283996e-07, "loss": 0.2025, "step": 6190 }, { "epoch": 0.12156862745098039, "grad_norm": 2.008704662322998, "learning_rate": 9.996826477089051e-07, "loss": 0.1798, "step": 6200 }, { "epoch": 0.12176470588235294, "grad_norm": 2.296579360961914, "learning_rate": 9.996768467667817e-07, "loss": 0.148, "step": 6210 }, { "epoch": 0.12196078431372549, "grad_norm": 2.73836350440979, "learning_rate": 9.996709933026392e-07, "loss": 0.191, "step": 6220 }, { "epoch": 0.12215686274509804, "grad_norm": 2.8447446823120117, "learning_rate": 9.99665087317093e-07, "loss": 0.187, "step": 6230 }, { "epoch": 0.1223529411764706, "grad_norm": 2.05029296875, "learning_rate": 9.996591288107633e-07, "loss": 0.1628, "step": 6240 }, { "epoch": 0.12254901960784313, "grad_norm": 2.637033462524414, "learning_rate": 9.996531177842771e-07, "loss": 0.1884, "step": 6250 }, { "epoch": 0.12274509803921568, "grad_norm": 2.9382121562957764, "learning_rate": 9.996470542382659e-07, "loss": 0.149, "step": 6260 }, { "epoch": 0.12294117647058823, "grad_norm": 2.4139244556427, "learning_rate": 9.99640938173367e-07, "loss": 0.2003, "step": 6270 }, { "epoch": 0.12313725490196079, "grad_norm": 1.6154303550720215, "learning_rate": 9.996347695902235e-07, "loss": 0.1676, "step": 6280 }, { "epoch": 0.12333333333333334, "grad_norm": 1.6540695428848267, "learning_rate": 9.996285484894835e-07, "loss": 0.1807, "step": 6290 }, { "epoch": 0.12352941176470589, "grad_norm": 2.7168636322021484, "learning_rate": 9.996222748718013e-07, "loss": 0.2167, "step": 6300 }, { "epoch": 0.12372549019607842, "grad_norm": 2.7988102436065674, "learning_rate": 9.99615948737836e-07, "loss": 0.1473, "step": 6310 }, { "epoch": 0.12392156862745098, "grad_norm": 2.748189926147461, "learning_rate": 9.996095700882527e-07, "loss": 0.1756, "step": 6320 }, { "epoch": 0.12411764705882353, "grad_norm": 2.8368234634399414, "learning_rate": 9.996031389237217e-07, "loss": 0.2184, "step": 6330 }, { "epoch": 0.12431372549019608, "grad_norm": 3.815906524658203, "learning_rate": 9.995966552449191e-07, "loss": 0.2015, "step": 6340 }, { "epoch": 0.12450980392156863, "grad_norm": 2.807727336883545, "learning_rate": 9.995901190525266e-07, "loss": 0.2152, "step": 6350 }, { "epoch": 0.12470588235294118, "grad_norm": 2.672886848449707, "learning_rate": 9.99583530347231e-07, "loss": 0.2344, "step": 6360 }, { "epoch": 0.12490196078431373, "grad_norm": 2.4272866249084473, "learning_rate": 9.995768891297248e-07, "loss": 0.1853, "step": 6370 }, { "epoch": 0.12509803921568627, "grad_norm": 1.7276209592819214, "learning_rate": 9.995701954007066e-07, "loss": 0.1844, "step": 6380 }, { "epoch": 0.12529411764705883, "grad_norm": 2.391085147857666, "learning_rate": 9.995634491608793e-07, "loss": 0.1703, "step": 6390 }, { "epoch": 0.12549019607843137, "grad_norm": 3.7352516651153564, "learning_rate": 9.995566504109522e-07, "loss": 0.1917, "step": 6400 }, { "epoch": 0.1256862745098039, "grad_norm": 1.8789658546447754, "learning_rate": 9.995497991516401e-07, "loss": 0.1709, "step": 6410 }, { "epoch": 0.12588235294117647, "grad_norm": 2.8376383781433105, "learning_rate": 9.995428953836631e-07, "loss": 0.1537, "step": 6420 }, { "epoch": 0.126078431372549, "grad_norm": 6.954867839813232, "learning_rate": 9.995359391077468e-07, "loss": 0.2292, "step": 6430 }, { "epoch": 0.12627450980392158, "grad_norm": 1.761800765991211, "learning_rate": 9.995289303246225e-07, "loss": 0.1643, "step": 6440 }, { "epoch": 0.1264705882352941, "grad_norm": 2.659116506576538, "learning_rate": 9.995218690350267e-07, "loss": 0.1713, "step": 6450 }, { "epoch": 0.12666666666666668, "grad_norm": 2.720003128051758, "learning_rate": 9.995147552397018e-07, "loss": 0.1715, "step": 6460 }, { "epoch": 0.12686274509803921, "grad_norm": 3.891976833343506, "learning_rate": 9.995075889393957e-07, "loss": 0.2102, "step": 6470 }, { "epoch": 0.12705882352941175, "grad_norm": 2.4839861392974854, "learning_rate": 9.995003701348612e-07, "loss": 0.1553, "step": 6480 }, { "epoch": 0.12725490196078432, "grad_norm": 3.2632768154144287, "learning_rate": 9.994930988268573e-07, "loss": 0.1872, "step": 6490 }, { "epoch": 0.12745098039215685, "grad_norm": 3.377126455307007, "learning_rate": 9.994857750161486e-07, "loss": 0.1862, "step": 6500 }, { "epoch": 0.12764705882352942, "grad_norm": 2.855311870574951, "learning_rate": 9.994783987035046e-07, "loss": 0.1816, "step": 6510 }, { "epoch": 0.12784313725490196, "grad_norm": 2.6466355323791504, "learning_rate": 9.994709698897005e-07, "loss": 0.1949, "step": 6520 }, { "epoch": 0.12803921568627452, "grad_norm": 2.480194330215454, "learning_rate": 9.994634885755178e-07, "loss": 0.1938, "step": 6530 }, { "epoch": 0.12823529411764706, "grad_norm": 3.3616795539855957, "learning_rate": 9.994559547617423e-07, "loss": 0.1941, "step": 6540 }, { "epoch": 0.1284313725490196, "grad_norm": 4.044515132904053, "learning_rate": 9.994483684491657e-07, "loss": 0.1556, "step": 6550 }, { "epoch": 0.12862745098039216, "grad_norm": 3.320668935775757, "learning_rate": 9.994407296385861e-07, "loss": 0.1631, "step": 6560 }, { "epoch": 0.1288235294117647, "grad_norm": 3.048491954803467, "learning_rate": 9.99433038330806e-07, "loss": 0.1887, "step": 6570 }, { "epoch": 0.12901960784313726, "grad_norm": 2.7971298694610596, "learning_rate": 9.99425294526634e-07, "loss": 0.1904, "step": 6580 }, { "epoch": 0.1292156862745098, "grad_norm": 1.534031867980957, "learning_rate": 9.994174982268838e-07, "loss": 0.1908, "step": 6590 }, { "epoch": 0.12941176470588237, "grad_norm": 2.5483474731445312, "learning_rate": 9.994096494323752e-07, "loss": 0.1787, "step": 6600 }, { "epoch": 0.1296078431372549, "grad_norm": 2.7553694248199463, "learning_rate": 9.99401748143933e-07, "loss": 0.2046, "step": 6610 }, { "epoch": 0.12980392156862744, "grad_norm": 2.212742567062378, "learning_rate": 9.99393794362388e-07, "loss": 0.1852, "step": 6620 }, { "epoch": 0.13, "grad_norm": 2.0198347568511963, "learning_rate": 9.99385788088576e-07, "loss": 0.152, "step": 6630 }, { "epoch": 0.13019607843137254, "grad_norm": 2.486053466796875, "learning_rate": 9.993777293233384e-07, "loss": 0.1825, "step": 6640 }, { "epoch": 0.1303921568627451, "grad_norm": 3.332677125930786, "learning_rate": 9.993696180675225e-07, "loss": 0.2055, "step": 6650 }, { "epoch": 0.13058823529411764, "grad_norm": 2.6523139476776123, "learning_rate": 9.99361454321981e-07, "loss": 0.1722, "step": 6660 }, { "epoch": 0.1307843137254902, "grad_norm": 2.233004331588745, "learning_rate": 9.993532380875719e-07, "loss": 0.208, "step": 6670 }, { "epoch": 0.13098039215686275, "grad_norm": 1.9198969602584839, "learning_rate": 9.993449693651587e-07, "loss": 0.1717, "step": 6680 }, { "epoch": 0.13117647058823528, "grad_norm": 3.143139123916626, "learning_rate": 9.993366481556106e-07, "loss": 0.1874, "step": 6690 }, { "epoch": 0.13137254901960785, "grad_norm": 3.3917179107666016, "learning_rate": 9.993282744598025e-07, "loss": 0.2106, "step": 6700 }, { "epoch": 0.13156862745098039, "grad_norm": 2.773022413253784, "learning_rate": 9.99319848278614e-07, "loss": 0.1709, "step": 6710 }, { "epoch": 0.13176470588235295, "grad_norm": 2.146509885787964, "learning_rate": 9.993113696129314e-07, "loss": 0.1913, "step": 6720 }, { "epoch": 0.1319607843137255, "grad_norm": 3.1679816246032715, "learning_rate": 9.993028384636458e-07, "loss": 0.168, "step": 6730 }, { "epoch": 0.13215686274509805, "grad_norm": 2.7378077507019043, "learning_rate": 9.992942548316534e-07, "loss": 0.1822, "step": 6740 }, { "epoch": 0.1323529411764706, "grad_norm": 1.9040966033935547, "learning_rate": 9.99285618717857e-07, "loss": 0.1511, "step": 6750 }, { "epoch": 0.13254901960784313, "grad_norm": 2.9378502368927, "learning_rate": 9.992769301231646e-07, "loss": 0.2005, "step": 6760 }, { "epoch": 0.1327450980392157, "grad_norm": 4.758029460906982, "learning_rate": 9.992681890484886e-07, "loss": 0.1506, "step": 6770 }, { "epoch": 0.13294117647058823, "grad_norm": 3.9254584312438965, "learning_rate": 9.992593954947484e-07, "loss": 0.1758, "step": 6780 }, { "epoch": 0.1331372549019608, "grad_norm": 1.9239280223846436, "learning_rate": 9.992505494628681e-07, "loss": 0.1657, "step": 6790 }, { "epoch": 0.13333333333333333, "grad_norm": 2.404731512069702, "learning_rate": 9.992416509537777e-07, "loss": 0.1794, "step": 6800 }, { "epoch": 0.13352941176470587, "grad_norm": 6.111047267913818, "learning_rate": 9.992326999684124e-07, "loss": 0.237, "step": 6810 }, { "epoch": 0.13372549019607843, "grad_norm": 1.4080718755722046, "learning_rate": 9.992236965077131e-07, "loss": 0.1786, "step": 6820 }, { "epoch": 0.13392156862745097, "grad_norm": 1.6912685632705688, "learning_rate": 9.99214640572626e-07, "loss": 0.1787, "step": 6830 }, { "epoch": 0.13411764705882354, "grad_norm": 2.6911873817443848, "learning_rate": 9.992055321641036e-07, "loss": 0.1465, "step": 6840 }, { "epoch": 0.13431372549019607, "grad_norm": 2.4005093574523926, "learning_rate": 9.991963712831025e-07, "loss": 0.1951, "step": 6850 }, { "epoch": 0.13450980392156864, "grad_norm": 2.9884634017944336, "learning_rate": 9.99187157930586e-07, "loss": 0.1663, "step": 6860 }, { "epoch": 0.13470588235294118, "grad_norm": 5.373988151550293, "learning_rate": 9.991778921075225e-07, "loss": 0.1545, "step": 6870 }, { "epoch": 0.1349019607843137, "grad_norm": 3.90604829788208, "learning_rate": 9.99168573814886e-07, "loss": 0.2092, "step": 6880 }, { "epoch": 0.13509803921568628, "grad_norm": 4.548996925354004, "learning_rate": 9.99159203053656e-07, "loss": 0.1818, "step": 6890 }, { "epoch": 0.13529411764705881, "grad_norm": 2.8396291732788086, "learning_rate": 9.991497798248171e-07, "loss": 0.1761, "step": 6900 }, { "epoch": 0.13549019607843138, "grad_norm": 4.806790351867676, "learning_rate": 9.991403041293605e-07, "loss": 0.163, "step": 6910 }, { "epoch": 0.13568627450980392, "grad_norm": 2.44119930267334, "learning_rate": 9.991307759682815e-07, "loss": 0.1652, "step": 6920 }, { "epoch": 0.13588235294117648, "grad_norm": 4.105988025665283, "learning_rate": 9.99121195342582e-07, "loss": 0.1979, "step": 6930 }, { "epoch": 0.13607843137254902, "grad_norm": 2.5205495357513428, "learning_rate": 9.99111562253269e-07, "loss": 0.218, "step": 6940 }, { "epoch": 0.13627450980392156, "grad_norm": 2.4460084438323975, "learning_rate": 9.99101876701355e-07, "loss": 0.1749, "step": 6950 }, { "epoch": 0.13647058823529412, "grad_norm": 1.1024105548858643, "learning_rate": 9.99092138687858e-07, "loss": 0.1769, "step": 6960 }, { "epoch": 0.13666666666666666, "grad_norm": 4.507024765014648, "learning_rate": 9.990823482138017e-07, "loss": 0.1905, "step": 6970 }, { "epoch": 0.13686274509803922, "grad_norm": 2.8278448581695557, "learning_rate": 9.990725052802154e-07, "loss": 0.2096, "step": 6980 }, { "epoch": 0.13705882352941176, "grad_norm": 2.873161792755127, "learning_rate": 9.990626098881332e-07, "loss": 0.1923, "step": 6990 }, { "epoch": 0.13725490196078433, "grad_norm": 2.5680088996887207, "learning_rate": 9.990526620385956e-07, "loss": 0.1878, "step": 7000 }, { "epoch": 0.13745098039215686, "grad_norm": 2.5988173484802246, "learning_rate": 9.99042661732648e-07, "loss": 0.1773, "step": 7010 }, { "epoch": 0.1376470588235294, "grad_norm": 4.173630714416504, "learning_rate": 9.990326089713418e-07, "loss": 0.173, "step": 7020 }, { "epoch": 0.13784313725490197, "grad_norm": 1.7947198152542114, "learning_rate": 9.990225037557337e-07, "loss": 0.1728, "step": 7030 }, { "epoch": 0.1380392156862745, "grad_norm": 2.607656955718994, "learning_rate": 9.990123460868857e-07, "loss": 0.1904, "step": 7040 }, { "epoch": 0.13823529411764707, "grad_norm": 5.1830525398254395, "learning_rate": 9.990021359658652e-07, "loss": 0.1664, "step": 7050 }, { "epoch": 0.1384313725490196, "grad_norm": 3.257505178451538, "learning_rate": 9.98991873393746e-07, "loss": 0.1675, "step": 7060 }, { "epoch": 0.13862745098039217, "grad_norm": 2.4895081520080566, "learning_rate": 9.989815583716064e-07, "loss": 0.1819, "step": 7070 }, { "epoch": 0.1388235294117647, "grad_norm": 2.728515625, "learning_rate": 9.989711909005309e-07, "loss": 0.168, "step": 7080 }, { "epoch": 0.13901960784313724, "grad_norm": 2.4217169284820557, "learning_rate": 9.98960770981609e-07, "loss": 0.1851, "step": 7090 }, { "epoch": 0.1392156862745098, "grad_norm": 2.9849510192871094, "learning_rate": 9.98950298615936e-07, "loss": 0.1766, "step": 7100 }, { "epoch": 0.13941176470588235, "grad_norm": 2.6366424560546875, "learning_rate": 9.98939773804613e-07, "loss": 0.2113, "step": 7110 }, { "epoch": 0.1396078431372549, "grad_norm": 2.5402214527130127, "learning_rate": 9.989291965487462e-07, "loss": 0.1516, "step": 7120 }, { "epoch": 0.13980392156862745, "grad_norm": 2.4800984859466553, "learning_rate": 9.989185668494469e-07, "loss": 0.1727, "step": 7130 }, { "epoch": 0.14, "grad_norm": 2.438647508621216, "learning_rate": 9.989078847078327e-07, "loss": 0.203, "step": 7140 }, { "epoch": 0.14019607843137255, "grad_norm": 3.5197503566741943, "learning_rate": 9.988971501250266e-07, "loss": 0.2048, "step": 7150 }, { "epoch": 0.1403921568627451, "grad_norm": 4.170396327972412, "learning_rate": 9.988863631021569e-07, "loss": 0.2145, "step": 7160 }, { "epoch": 0.14058823529411765, "grad_norm": 3.649611473083496, "learning_rate": 9.988755236403572e-07, "loss": 0.1943, "step": 7170 }, { "epoch": 0.1407843137254902, "grad_norm": 2.0277953147888184, "learning_rate": 9.98864631740767e-07, "loss": 0.1916, "step": 7180 }, { "epoch": 0.14098039215686275, "grad_norm": 2.1280651092529297, "learning_rate": 9.988536874045312e-07, "loss": 0.1912, "step": 7190 }, { "epoch": 0.1411764705882353, "grad_norm": 2.1693246364593506, "learning_rate": 9.988426906328002e-07, "loss": 0.1786, "step": 7200 }, { "epoch": 0.14137254901960783, "grad_norm": 3.465407609939575, "learning_rate": 9.988316414267297e-07, "loss": 0.1877, "step": 7210 }, { "epoch": 0.1415686274509804, "grad_norm": 1.3366395235061646, "learning_rate": 9.988205397874814e-07, "loss": 0.1915, "step": 7220 }, { "epoch": 0.14176470588235293, "grad_norm": 2.909632921218872, "learning_rate": 9.988093857162222e-07, "loss": 0.1951, "step": 7230 }, { "epoch": 0.1419607843137255, "grad_norm": 6.22321081161499, "learning_rate": 9.98798179214124e-07, "loss": 0.1881, "step": 7240 }, { "epoch": 0.14215686274509803, "grad_norm": 3.3325953483581543, "learning_rate": 9.987869202823653e-07, "loss": 0.1814, "step": 7250 }, { "epoch": 0.1423529411764706, "grad_norm": 5.281031608581543, "learning_rate": 9.987756089221295e-07, "loss": 0.1757, "step": 7260 }, { "epoch": 0.14254901960784314, "grad_norm": 3.662926197052002, "learning_rate": 9.987642451346053e-07, "loss": 0.1885, "step": 7270 }, { "epoch": 0.14274509803921567, "grad_norm": 2.7030675411224365, "learning_rate": 9.987528289209872e-07, "loss": 0.1914, "step": 7280 }, { "epoch": 0.14294117647058824, "grad_norm": 1.405630111694336, "learning_rate": 9.987413602824755e-07, "loss": 0.1553, "step": 7290 }, { "epoch": 0.14313725490196078, "grad_norm": 2.2022881507873535, "learning_rate": 9.987298392202754e-07, "loss": 0.1998, "step": 7300 }, { "epoch": 0.14333333333333334, "grad_norm": 1.5452877283096313, "learning_rate": 9.987182657355979e-07, "loss": 0.1865, "step": 7310 }, { "epoch": 0.14352941176470588, "grad_norm": 1.6513594388961792, "learning_rate": 9.987066398296596e-07, "loss": 0.1532, "step": 7320 }, { "epoch": 0.14372549019607844, "grad_norm": 2.299046516418457, "learning_rate": 9.986949615036825e-07, "loss": 0.2042, "step": 7330 }, { "epoch": 0.14392156862745098, "grad_norm": 2.03257155418396, "learning_rate": 9.986832307588943e-07, "loss": 0.1279, "step": 7340 }, { "epoch": 0.14411764705882352, "grad_norm": 3.126216173171997, "learning_rate": 9.98671447596528e-07, "loss": 0.1622, "step": 7350 }, { "epoch": 0.14431372549019608, "grad_norm": 2.38985276222229, "learning_rate": 9.986596120178217e-07, "loss": 0.2003, "step": 7360 }, { "epoch": 0.14450980392156862, "grad_norm": 3.334390163421631, "learning_rate": 9.9864772402402e-07, "loss": 0.2028, "step": 7370 }, { "epoch": 0.14470588235294118, "grad_norm": 2.3894333839416504, "learning_rate": 9.98635783616372e-07, "loss": 0.1549, "step": 7380 }, { "epoch": 0.14490196078431372, "grad_norm": 5.379003047943115, "learning_rate": 9.986237907961332e-07, "loss": 0.185, "step": 7390 }, { "epoch": 0.1450980392156863, "grad_norm": 3.7518906593322754, "learning_rate": 9.986117455645642e-07, "loss": 0.1781, "step": 7400 }, { "epoch": 0.14529411764705882, "grad_norm": 5.223618984222412, "learning_rate": 9.985996479229306e-07, "loss": 0.1787, "step": 7410 }, { "epoch": 0.14549019607843136, "grad_norm": 2.6500532627105713, "learning_rate": 9.985874978725046e-07, "loss": 0.1769, "step": 7420 }, { "epoch": 0.14568627450980393, "grad_norm": 4.085933685302734, "learning_rate": 9.98575295414563e-07, "loss": 0.1805, "step": 7430 }, { "epoch": 0.14588235294117646, "grad_norm": 3.0027854442596436, "learning_rate": 9.985630405503885e-07, "loss": 0.176, "step": 7440 }, { "epoch": 0.14607843137254903, "grad_norm": 3.0603368282318115, "learning_rate": 9.985507332812693e-07, "loss": 0.1544, "step": 7450 }, { "epoch": 0.14627450980392157, "grad_norm": 2.876213550567627, "learning_rate": 9.985383736084987e-07, "loss": 0.2075, "step": 7460 }, { "epoch": 0.14647058823529413, "grad_norm": 2.816080331802368, "learning_rate": 9.985259615333763e-07, "loss": 0.1839, "step": 7470 }, { "epoch": 0.14666666666666667, "grad_norm": 8.99306869506836, "learning_rate": 9.985134970572065e-07, "loss": 0.1799, "step": 7480 }, { "epoch": 0.1468627450980392, "grad_norm": 1.7647682428359985, "learning_rate": 9.985009801812996e-07, "loss": 0.178, "step": 7490 }, { "epoch": 0.14705882352941177, "grad_norm": 5.5277299880981445, "learning_rate": 9.98488410906971e-07, "loss": 0.2084, "step": 7500 }, { "epoch": 0.1472549019607843, "grad_norm": 4.2457966804504395, "learning_rate": 9.984757892355423e-07, "loss": 0.1881, "step": 7510 }, { "epoch": 0.14745098039215687, "grad_norm": 2.200181484222412, "learning_rate": 9.9846311516834e-07, "loss": 0.1887, "step": 7520 }, { "epoch": 0.1476470588235294, "grad_norm": 2.2309112548828125, "learning_rate": 9.98450388706696e-07, "loss": 0.1745, "step": 7530 }, { "epoch": 0.14784313725490197, "grad_norm": 5.130068302154541, "learning_rate": 9.984376098519485e-07, "loss": 0.1737, "step": 7540 }, { "epoch": 0.1480392156862745, "grad_norm": 2.6900362968444824, "learning_rate": 9.984247786054404e-07, "loss": 0.1617, "step": 7550 }, { "epoch": 0.14823529411764705, "grad_norm": 3.111534357070923, "learning_rate": 9.984118949685205e-07, "loss": 0.1821, "step": 7560 }, { "epoch": 0.1484313725490196, "grad_norm": 2.72414231300354, "learning_rate": 9.983989589425432e-07, "loss": 0.1902, "step": 7570 }, { "epoch": 0.14862745098039215, "grad_norm": 2.1952896118164062, "learning_rate": 9.983859705288677e-07, "loss": 0.1649, "step": 7580 }, { "epoch": 0.14882352941176472, "grad_norm": 1.672301173210144, "learning_rate": 9.9837292972886e-07, "loss": 0.171, "step": 7590 }, { "epoch": 0.14901960784313725, "grad_norm": 2.096061944961548, "learning_rate": 9.9835983654389e-07, "loss": 0.2065, "step": 7600 }, { "epoch": 0.1492156862745098, "grad_norm": 3.919215679168701, "learning_rate": 9.983466909753347e-07, "loss": 0.1775, "step": 7610 }, { "epoch": 0.14941176470588236, "grad_norm": 1.5032695531845093, "learning_rate": 9.983334930245755e-07, "loss": 0.1944, "step": 7620 }, { "epoch": 0.1496078431372549, "grad_norm": 2.4801414012908936, "learning_rate": 9.983202426929998e-07, "loss": 0.1733, "step": 7630 }, { "epoch": 0.14980392156862746, "grad_norm": 4.645618438720703, "learning_rate": 9.98306939982e-07, "loss": 0.1739, "step": 7640 }, { "epoch": 0.15, "grad_norm": 1.7635382413864136, "learning_rate": 9.982935848929748e-07, "loss": 0.1797, "step": 7650 }, { "epoch": 0.15019607843137256, "grad_norm": 4.675463676452637, "learning_rate": 9.98280177427328e-07, "loss": 0.1869, "step": 7660 }, { "epoch": 0.1503921568627451, "grad_norm": 4.448642730712891, "learning_rate": 9.982667175864684e-07, "loss": 0.1982, "step": 7670 }, { "epoch": 0.15058823529411763, "grad_norm": 1.1663918495178223, "learning_rate": 9.982532053718113e-07, "loss": 0.1521, "step": 7680 }, { "epoch": 0.1507843137254902, "grad_norm": 3.833740234375, "learning_rate": 9.982396407847767e-07, "loss": 0.2312, "step": 7690 }, { "epoch": 0.15098039215686274, "grad_norm": 1.841633915901184, "learning_rate": 9.982260238267906e-07, "loss": 0.2002, "step": 7700 }, { "epoch": 0.1511764705882353, "grad_norm": 2.555298089981079, "learning_rate": 9.98212354499284e-07, "loss": 0.1964, "step": 7710 }, { "epoch": 0.15137254901960784, "grad_norm": 2.5998036861419678, "learning_rate": 9.98198632803694e-07, "loss": 0.2099, "step": 7720 }, { "epoch": 0.1515686274509804, "grad_norm": 2.9881253242492676, "learning_rate": 9.98184858741463e-07, "loss": 0.2274, "step": 7730 }, { "epoch": 0.15176470588235294, "grad_norm": 2.0980870723724365, "learning_rate": 9.981710323140385e-07, "loss": 0.1474, "step": 7740 }, { "epoch": 0.15196078431372548, "grad_norm": 2.509999990463257, "learning_rate": 9.98157153522874e-07, "loss": 0.2122, "step": 7750 }, { "epoch": 0.15215686274509804, "grad_norm": 1.6198532581329346, "learning_rate": 9.981432223694284e-07, "loss": 0.2147, "step": 7760 }, { "epoch": 0.15235294117647058, "grad_norm": 4.021377086639404, "learning_rate": 9.98129238855166e-07, "loss": 0.1848, "step": 7770 }, { "epoch": 0.15254901960784314, "grad_norm": 3.9742379188537598, "learning_rate": 9.981152029815563e-07, "loss": 0.1889, "step": 7780 }, { "epoch": 0.15274509803921568, "grad_norm": 1.367796778678894, "learning_rate": 9.98101114750075e-07, "loss": 0.1739, "step": 7790 }, { "epoch": 0.15294117647058825, "grad_norm": 1.753460168838501, "learning_rate": 9.98086974162203e-07, "loss": 0.1757, "step": 7800 }, { "epoch": 0.15313725490196078, "grad_norm": 2.0941693782806396, "learning_rate": 9.980727812194263e-07, "loss": 0.1892, "step": 7810 }, { "epoch": 0.15333333333333332, "grad_norm": 4.647737503051758, "learning_rate": 9.98058535923237e-07, "loss": 0.2382, "step": 7820 }, { "epoch": 0.1535294117647059, "grad_norm": 1.9468588829040527, "learning_rate": 9.980442382751324e-07, "loss": 0.2065, "step": 7830 }, { "epoch": 0.15372549019607842, "grad_norm": 2.548431396484375, "learning_rate": 9.980298882766154e-07, "loss": 0.1838, "step": 7840 }, { "epoch": 0.153921568627451, "grad_norm": 3.8116936683654785, "learning_rate": 9.98015485929194e-07, "loss": 0.171, "step": 7850 }, { "epoch": 0.15411764705882353, "grad_norm": 2.350362539291382, "learning_rate": 9.980010312343827e-07, "loss": 0.1677, "step": 7860 }, { "epoch": 0.1543137254901961, "grad_norm": 2.1329479217529297, "learning_rate": 9.979865241937004e-07, "loss": 0.1774, "step": 7870 }, { "epoch": 0.15450980392156863, "grad_norm": 2.463714599609375, "learning_rate": 9.979719648086723e-07, "loss": 0.2152, "step": 7880 }, { "epoch": 0.15470588235294117, "grad_norm": 2.0687999725341797, "learning_rate": 9.979573530808281e-07, "loss": 0.2015, "step": 7890 }, { "epoch": 0.15490196078431373, "grad_norm": 1.9685782194137573, "learning_rate": 9.979426890117045e-07, "loss": 0.1419, "step": 7900 }, { "epoch": 0.15509803921568627, "grad_norm": 6.730506420135498, "learning_rate": 9.979279726028423e-07, "loss": 0.246, "step": 7910 }, { "epoch": 0.15529411764705883, "grad_norm": 3.966226100921631, "learning_rate": 9.979132038557887e-07, "loss": 0.1961, "step": 7920 }, { "epoch": 0.15549019607843137, "grad_norm": 2.842186212539673, "learning_rate": 9.978983827720958e-07, "loss": 0.1874, "step": 7930 }, { "epoch": 0.15568627450980393, "grad_norm": 2.8871521949768066, "learning_rate": 9.978835093533216e-07, "loss": 0.167, "step": 7940 }, { "epoch": 0.15588235294117647, "grad_norm": 2.457091808319092, "learning_rate": 9.978685836010295e-07, "loss": 0.1825, "step": 7950 }, { "epoch": 0.156078431372549, "grad_norm": 4.292448043823242, "learning_rate": 9.978536055167885e-07, "loss": 0.1824, "step": 7960 }, { "epoch": 0.15627450980392157, "grad_norm": 2.301600933074951, "learning_rate": 9.978385751021728e-07, "loss": 0.16, "step": 7970 }, { "epoch": 0.1564705882352941, "grad_norm": 2.7621572017669678, "learning_rate": 9.978234923587622e-07, "loss": 0.1627, "step": 7980 }, { "epoch": 0.15666666666666668, "grad_norm": 2.0186409950256348, "learning_rate": 9.978083572881421e-07, "loss": 0.1646, "step": 7990 }, { "epoch": 0.1568627450980392, "grad_norm": 4.50340461730957, "learning_rate": 9.977931698919036e-07, "loss": 0.1879, "step": 8000 }, { "epoch": 0.15705882352941178, "grad_norm": 3.8736331462860107, "learning_rate": 9.97777930171643e-07, "loss": 0.1898, "step": 8010 }, { "epoch": 0.15725490196078432, "grad_norm": 3.184471368789673, "learning_rate": 9.977626381289618e-07, "loss": 0.1589, "step": 8020 }, { "epoch": 0.15745098039215685, "grad_norm": 2.6676266193389893, "learning_rate": 9.97747293765468e-07, "loss": 0.1496, "step": 8030 }, { "epoch": 0.15764705882352942, "grad_norm": 2.462852716445923, "learning_rate": 9.977318970827742e-07, "loss": 0.1868, "step": 8040 }, { "epoch": 0.15784313725490196, "grad_norm": 4.197325229644775, "learning_rate": 9.977164480824984e-07, "loss": 0.1848, "step": 8050 }, { "epoch": 0.15803921568627452, "grad_norm": 1.8114478588104248, "learning_rate": 9.977009467662652e-07, "loss": 0.1701, "step": 8060 }, { "epoch": 0.15823529411764706, "grad_norm": 6.293909072875977, "learning_rate": 9.976853931357034e-07, "loss": 0.2228, "step": 8070 }, { "epoch": 0.1584313725490196, "grad_norm": 2.901611089706421, "learning_rate": 9.97669787192448e-07, "loss": 0.1888, "step": 8080 }, { "epoch": 0.15862745098039216, "grad_norm": 1.9995803833007812, "learning_rate": 9.976541289381396e-07, "loss": 0.15, "step": 8090 }, { "epoch": 0.1588235294117647, "grad_norm": 3.0782406330108643, "learning_rate": 9.97638418374424e-07, "loss": 0.2021, "step": 8100 }, { "epoch": 0.15901960784313726, "grad_norm": 2.4309229850769043, "learning_rate": 9.976226555029522e-07, "loss": 0.1718, "step": 8110 }, { "epoch": 0.1592156862745098, "grad_norm": 2.4714653491973877, "learning_rate": 9.976068403253814e-07, "loss": 0.2118, "step": 8120 }, { "epoch": 0.15941176470588236, "grad_norm": 4.1763386726379395, "learning_rate": 9.97590972843374e-07, "loss": 0.2194, "step": 8130 }, { "epoch": 0.1596078431372549, "grad_norm": 2.832195997238159, "learning_rate": 9.975750530585978e-07, "loss": 0.2179, "step": 8140 }, { "epoch": 0.15980392156862744, "grad_norm": 3.4007511138916016, "learning_rate": 9.975590809727259e-07, "loss": 0.1611, "step": 8150 }, { "epoch": 0.16, "grad_norm": 2.916689395904541, "learning_rate": 9.975430565874376e-07, "loss": 0.2002, "step": 8160 }, { "epoch": 0.16019607843137254, "grad_norm": 2.270042896270752, "learning_rate": 9.97526979904417e-07, "loss": 0.2027, "step": 8170 }, { "epoch": 0.1603921568627451, "grad_norm": 1.7552350759506226, "learning_rate": 9.97510850925354e-07, "loss": 0.207, "step": 8180 }, { "epoch": 0.16058823529411764, "grad_norm": 2.7135753631591797, "learning_rate": 9.97494669651944e-07, "loss": 0.2101, "step": 8190 }, { "epoch": 0.1607843137254902, "grad_norm": 2.0745344161987305, "learning_rate": 9.974784360858878e-07, "loss": 0.2022, "step": 8200 }, { "epoch": 0.16098039215686274, "grad_norm": 2.374152183532715, "learning_rate": 9.974621502288913e-07, "loss": 0.1988, "step": 8210 }, { "epoch": 0.16117647058823528, "grad_norm": 2.908977508544922, "learning_rate": 9.974458120826672e-07, "loss": 0.1755, "step": 8220 }, { "epoch": 0.16137254901960785, "grad_norm": 3.6278388500213623, "learning_rate": 9.974294216489323e-07, "loss": 0.1677, "step": 8230 }, { "epoch": 0.16156862745098038, "grad_norm": 2.6080222129821777, "learning_rate": 9.974129789294095e-07, "loss": 0.1916, "step": 8240 }, { "epoch": 0.16176470588235295, "grad_norm": 1.7692177295684814, "learning_rate": 9.973964839258274e-07, "loss": 0.1936, "step": 8250 }, { "epoch": 0.1619607843137255, "grad_norm": 4.910453796386719, "learning_rate": 9.973799366399192e-07, "loss": 0.1904, "step": 8260 }, { "epoch": 0.16215686274509805, "grad_norm": 1.648457646369934, "learning_rate": 9.97363337073425e-07, "loss": 0.1965, "step": 8270 }, { "epoch": 0.1623529411764706, "grad_norm": 2.6836204528808594, "learning_rate": 9.973466852280888e-07, "loss": 0.184, "step": 8280 }, { "epoch": 0.16254901960784313, "grad_norm": 2.9240357875823975, "learning_rate": 9.973299811056614e-07, "loss": 0.1708, "step": 8290 }, { "epoch": 0.1627450980392157, "grad_norm": 7.7264604568481445, "learning_rate": 9.973132247078988e-07, "loss": 0.165, "step": 8300 }, { "epoch": 0.16294117647058823, "grad_norm": 2.296403408050537, "learning_rate": 9.972964160365618e-07, "loss": 0.1664, "step": 8310 }, { "epoch": 0.1631372549019608, "grad_norm": 3.1774544715881348, "learning_rate": 9.972795550934176e-07, "loss": 0.2167, "step": 8320 }, { "epoch": 0.16333333333333333, "grad_norm": 2.3982973098754883, "learning_rate": 9.972626418802382e-07, "loss": 0.1801, "step": 8330 }, { "epoch": 0.1635294117647059, "grad_norm": 3.576829433441162, "learning_rate": 9.972456763988014e-07, "loss": 0.1733, "step": 8340 }, { "epoch": 0.16372549019607843, "grad_norm": 2.1517603397369385, "learning_rate": 9.972286586508905e-07, "loss": 0.1615, "step": 8350 }, { "epoch": 0.16392156862745097, "grad_norm": 2.2313339710235596, "learning_rate": 9.972115886382947e-07, "loss": 0.1907, "step": 8360 }, { "epoch": 0.16411764705882353, "grad_norm": 3.31294584274292, "learning_rate": 9.971944663628075e-07, "loss": 0.1728, "step": 8370 }, { "epoch": 0.16431372549019607, "grad_norm": 3.3163905143737793, "learning_rate": 9.97177291826229e-07, "loss": 0.1622, "step": 8380 }, { "epoch": 0.16450980392156864, "grad_norm": 2.258397102355957, "learning_rate": 9.97160065030365e-07, "loss": 0.1934, "step": 8390 }, { "epoch": 0.16470588235294117, "grad_norm": 2.182321786880493, "learning_rate": 9.971427859770252e-07, "loss": 0.2008, "step": 8400 }, { "epoch": 0.16490196078431374, "grad_norm": 2.6920697689056396, "learning_rate": 9.971254546680267e-07, "loss": 0.1771, "step": 8410 }, { "epoch": 0.16509803921568628, "grad_norm": 2.584045648574829, "learning_rate": 9.971080711051906e-07, "loss": 0.1641, "step": 8420 }, { "epoch": 0.1652941176470588, "grad_norm": 2.7721147537231445, "learning_rate": 9.970906352903449e-07, "loss": 0.1788, "step": 8430 }, { "epoch": 0.16549019607843138, "grad_norm": 3.32953143119812, "learning_rate": 9.970731472253214e-07, "loss": 0.167, "step": 8440 }, { "epoch": 0.16568627450980392, "grad_norm": 4.396429538726807, "learning_rate": 9.97055606911959e-07, "loss": 0.182, "step": 8450 }, { "epoch": 0.16588235294117648, "grad_norm": 2.5037972927093506, "learning_rate": 9.97038014352101e-07, "loss": 0.152, "step": 8460 }, { "epoch": 0.16607843137254902, "grad_norm": 2.1387128829956055, "learning_rate": 9.970203695475967e-07, "loss": 0.1583, "step": 8470 }, { "epoch": 0.16627450980392156, "grad_norm": 2.845815420150757, "learning_rate": 9.97002672500301e-07, "loss": 0.1949, "step": 8480 }, { "epoch": 0.16647058823529412, "grad_norm": 4.892502784729004, "learning_rate": 9.969849232120735e-07, "loss": 0.1544, "step": 8490 }, { "epoch": 0.16666666666666666, "grad_norm": 2.701620101928711, "learning_rate": 9.969671216847806e-07, "loss": 0.1963, "step": 8500 }, { "epoch": 0.16686274509803922, "grad_norm": 3.3733198642730713, "learning_rate": 9.96949267920293e-07, "loss": 0.1792, "step": 8510 }, { "epoch": 0.16705882352941176, "grad_norm": 2.8766958713531494, "learning_rate": 9.969313619204875e-07, "loss": 0.1972, "step": 8520 }, { "epoch": 0.16725490196078432, "grad_norm": 2.609391212463379, "learning_rate": 9.96913403687246e-07, "loss": 0.1823, "step": 8530 }, { "epoch": 0.16745098039215686, "grad_norm": 3.149052143096924, "learning_rate": 9.968953932224563e-07, "loss": 0.1893, "step": 8540 }, { "epoch": 0.1676470588235294, "grad_norm": 6.030562877655029, "learning_rate": 9.968773305280116e-07, "loss": 0.1965, "step": 8550 }, { "epoch": 0.16784313725490196, "grad_norm": 2.383648157119751, "learning_rate": 9.968592156058106e-07, "loss": 0.1705, "step": 8560 }, { "epoch": 0.1680392156862745, "grad_norm": 3.191723346710205, "learning_rate": 9.96841048457757e-07, "loss": 0.186, "step": 8570 }, { "epoch": 0.16823529411764707, "grad_norm": 1.7033705711364746, "learning_rate": 9.968228290857604e-07, "loss": 0.1838, "step": 8580 }, { "epoch": 0.1684313725490196, "grad_norm": 2.996513605117798, "learning_rate": 9.968045574917366e-07, "loss": 0.1691, "step": 8590 }, { "epoch": 0.16862745098039217, "grad_norm": 19.30753517150879, "learning_rate": 9.967862336776052e-07, "loss": 0.1997, "step": 8600 }, { "epoch": 0.1688235294117647, "grad_norm": 2.50734543800354, "learning_rate": 9.967678576452929e-07, "loss": 0.148, "step": 8610 }, { "epoch": 0.16901960784313724, "grad_norm": 1.5622843503952026, "learning_rate": 9.96749429396731e-07, "loss": 0.191, "step": 8620 }, { "epoch": 0.1692156862745098, "grad_norm": 3.0996222496032715, "learning_rate": 9.967309489338565e-07, "loss": 0.2032, "step": 8630 }, { "epoch": 0.16941176470588235, "grad_norm": 2.726145029067993, "learning_rate": 9.96712416258612e-07, "loss": 0.1651, "step": 8640 }, { "epoch": 0.1696078431372549, "grad_norm": 3.0598158836364746, "learning_rate": 9.966938313729455e-07, "loss": 0.1816, "step": 8650 }, { "epoch": 0.16980392156862745, "grad_norm": 2.5404350757598877, "learning_rate": 9.966751942788106e-07, "loss": 0.2146, "step": 8660 }, { "epoch": 0.17, "grad_norm": 2.457442283630371, "learning_rate": 9.966565049781658e-07, "loss": 0.1435, "step": 8670 }, { "epoch": 0.17019607843137255, "grad_norm": 4.071306228637695, "learning_rate": 9.966377634729764e-07, "loss": 0.1953, "step": 8680 }, { "epoch": 0.1703921568627451, "grad_norm": 1.0025033950805664, "learning_rate": 9.966189697652115e-07, "loss": 0.1606, "step": 8690 }, { "epoch": 0.17058823529411765, "grad_norm": 3.969038724899292, "learning_rate": 9.966001238568471e-07, "loss": 0.1859, "step": 8700 }, { "epoch": 0.1707843137254902, "grad_norm": 2.1944820880889893, "learning_rate": 9.96581225749864e-07, "loss": 0.1771, "step": 8710 }, { "epoch": 0.17098039215686275, "grad_norm": 3.378026008605957, "learning_rate": 9.965622754462485e-07, "loss": 0.2, "step": 8720 }, { "epoch": 0.1711764705882353, "grad_norm": 3.4759912490844727, "learning_rate": 9.965432729479925e-07, "loss": 0.1827, "step": 8730 }, { "epoch": 0.17137254901960786, "grad_norm": 3.663853645324707, "learning_rate": 9.965242182570936e-07, "loss": 0.2099, "step": 8740 }, { "epoch": 0.1715686274509804, "grad_norm": 2.0682942867279053, "learning_rate": 9.965051113755544e-07, "loss": 0.1788, "step": 8750 }, { "epoch": 0.17176470588235293, "grad_norm": 4.060184478759766, "learning_rate": 9.964859523053837e-07, "loss": 0.2271, "step": 8760 }, { "epoch": 0.1719607843137255, "grad_norm": 4.005316734313965, "learning_rate": 9.964667410485948e-07, "loss": 0.194, "step": 8770 }, { "epoch": 0.17215686274509803, "grad_norm": 3.096754789352417, "learning_rate": 9.964474776072073e-07, "loss": 0.1866, "step": 8780 }, { "epoch": 0.1723529411764706, "grad_norm": 2.0129971504211426, "learning_rate": 9.96428161983246e-07, "loss": 0.1582, "step": 8790 }, { "epoch": 0.17254901960784313, "grad_norm": 2.6152336597442627, "learning_rate": 9.964087941787416e-07, "loss": 0.1868, "step": 8800 }, { "epoch": 0.1727450980392157, "grad_norm": 2.9085135459899902, "learning_rate": 9.96389374195729e-07, "loss": 0.2245, "step": 8810 }, { "epoch": 0.17294117647058824, "grad_norm": 3.146440029144287, "learning_rate": 9.963699020362502e-07, "loss": 0.2113, "step": 8820 }, { "epoch": 0.17313725490196077, "grad_norm": 2.4185245037078857, "learning_rate": 9.963503777023517e-07, "loss": 0.1828, "step": 8830 }, { "epoch": 0.17333333333333334, "grad_norm": 4.350964069366455, "learning_rate": 9.963308011960857e-07, "loss": 0.1704, "step": 8840 }, { "epoch": 0.17352941176470588, "grad_norm": 3.1013126373291016, "learning_rate": 9.9631117251951e-07, "loss": 0.1791, "step": 8850 }, { "epoch": 0.17372549019607844, "grad_norm": 1.8065590858459473, "learning_rate": 9.96291491674688e-07, "loss": 0.2005, "step": 8860 }, { "epoch": 0.17392156862745098, "grad_norm": 5.214871883392334, "learning_rate": 9.962717586636882e-07, "loss": 0.2055, "step": 8870 }, { "epoch": 0.17411764705882352, "grad_norm": 6.631687641143799, "learning_rate": 9.962519734885845e-07, "loss": 0.1693, "step": 8880 }, { "epoch": 0.17431372549019608, "grad_norm": 1.4681464433670044, "learning_rate": 9.962321361514569e-07, "loss": 0.2167, "step": 8890 }, { "epoch": 0.17450980392156862, "grad_norm": 2.68235445022583, "learning_rate": 9.962122466543908e-07, "loss": 0.1826, "step": 8900 }, { "epoch": 0.17470588235294118, "grad_norm": 3.1456315517425537, "learning_rate": 9.961923049994761e-07, "loss": 0.1799, "step": 8910 }, { "epoch": 0.17490196078431372, "grad_norm": 2.404709577560425, "learning_rate": 9.961723111888095e-07, "loss": 0.192, "step": 8920 }, { "epoch": 0.17509803921568629, "grad_norm": 2.2832093238830566, "learning_rate": 9.961522652244922e-07, "loss": 0.1466, "step": 8930 }, { "epoch": 0.17529411764705882, "grad_norm": 2.9227120876312256, "learning_rate": 9.961321671086317e-07, "loss": 0.2253, "step": 8940 }, { "epoch": 0.17549019607843136, "grad_norm": 3.528646469116211, "learning_rate": 9.961120168433401e-07, "loss": 0.1777, "step": 8950 }, { "epoch": 0.17568627450980392, "grad_norm": 1.7941030263900757, "learning_rate": 9.960918144307358e-07, "loss": 0.1756, "step": 8960 }, { "epoch": 0.17588235294117646, "grad_norm": 3.2534339427948, "learning_rate": 9.960715598729421e-07, "loss": 0.1642, "step": 8970 }, { "epoch": 0.17607843137254903, "grad_norm": 2.802990674972534, "learning_rate": 9.96051253172088e-07, "loss": 0.1762, "step": 8980 }, { "epoch": 0.17627450980392156, "grad_norm": 1.9862173795700073, "learning_rate": 9.960308943303085e-07, "loss": 0.1805, "step": 8990 }, { "epoch": 0.17647058823529413, "grad_norm": 2.0407562255859375, "learning_rate": 9.960104833497428e-07, "loss": 0.1483, "step": 9000 }, { "epoch": 0.17666666666666667, "grad_norm": 1.7913830280303955, "learning_rate": 9.959900202325365e-07, "loss": 0.2005, "step": 9010 }, { "epoch": 0.1768627450980392, "grad_norm": 2.4688940048217773, "learning_rate": 9.959695049808406e-07, "loss": 0.2019, "step": 9020 }, { "epoch": 0.17705882352941177, "grad_norm": 3.1725406646728516, "learning_rate": 9.959489375968118e-07, "loss": 0.1665, "step": 9030 }, { "epoch": 0.1772549019607843, "grad_norm": 1.4455329179763794, "learning_rate": 9.959283180826114e-07, "loss": 0.1476, "step": 9040 }, { "epoch": 0.17745098039215687, "grad_norm": 3.784137010574341, "learning_rate": 9.959076464404074e-07, "loss": 0.1388, "step": 9050 }, { "epoch": 0.1776470588235294, "grad_norm": 2.597874164581299, "learning_rate": 9.958869226723723e-07, "loss": 0.186, "step": 9060 }, { "epoch": 0.17784313725490197, "grad_norm": 2.1318883895874023, "learning_rate": 9.958661467806843e-07, "loss": 0.1585, "step": 9070 }, { "epoch": 0.1780392156862745, "grad_norm": 2.3908629417419434, "learning_rate": 9.958453187675276e-07, "loss": 0.1685, "step": 9080 }, { "epoch": 0.17823529411764705, "grad_norm": 3.445103406906128, "learning_rate": 9.958244386350909e-07, "loss": 0.1973, "step": 9090 }, { "epoch": 0.1784313725490196, "grad_norm": 2.613475799560547, "learning_rate": 9.958035063855695e-07, "loss": 0.1791, "step": 9100 }, { "epoch": 0.17862745098039215, "grad_norm": 2.213376998901367, "learning_rate": 9.95782522021163e-07, "loss": 0.1631, "step": 9110 }, { "epoch": 0.17882352941176471, "grad_norm": 3.698637008666992, "learning_rate": 9.95761485544078e-07, "loss": 0.1726, "step": 9120 }, { "epoch": 0.17901960784313725, "grad_norm": 3.7360074520111084, "learning_rate": 9.95740396956525e-07, "loss": 0.1611, "step": 9130 }, { "epoch": 0.17921568627450982, "grad_norm": 2.4989724159240723, "learning_rate": 9.957192562607208e-07, "loss": 0.1757, "step": 9140 }, { "epoch": 0.17941176470588235, "grad_norm": 6.221375942230225, "learning_rate": 9.956980634588876e-07, "loss": 0.1841, "step": 9150 }, { "epoch": 0.1796078431372549, "grad_norm": 2.68314528465271, "learning_rate": 9.956768185532532e-07, "loss": 0.1542, "step": 9160 }, { "epoch": 0.17980392156862746, "grad_norm": 1.5143659114837646, "learning_rate": 9.956555215460503e-07, "loss": 0.1832, "step": 9170 }, { "epoch": 0.18, "grad_norm": 3.9720399379730225, "learning_rate": 9.956341724395177e-07, "loss": 0.2033, "step": 9180 }, { "epoch": 0.18019607843137256, "grad_norm": 1.3034254312515259, "learning_rate": 9.956127712358992e-07, "loss": 0.208, "step": 9190 }, { "epoch": 0.1803921568627451, "grad_norm": 2.0341978073120117, "learning_rate": 9.955913179374448e-07, "loss": 0.1566, "step": 9200 }, { "epoch": 0.18058823529411766, "grad_norm": 2.892878770828247, "learning_rate": 9.955698125464093e-07, "loss": 0.2087, "step": 9210 }, { "epoch": 0.1807843137254902, "grad_norm": 4.039526462554932, "learning_rate": 9.95548255065053e-07, "loss": 0.1465, "step": 9220 }, { "epoch": 0.18098039215686273, "grad_norm": 2.449415922164917, "learning_rate": 9.95526645495642e-07, "loss": 0.1621, "step": 9230 }, { "epoch": 0.1811764705882353, "grad_norm": 2.1200995445251465, "learning_rate": 9.955049838404477e-07, "loss": 0.1473, "step": 9240 }, { "epoch": 0.18137254901960784, "grad_norm": 3.976106643676758, "learning_rate": 9.95483270101747e-07, "loss": 0.2106, "step": 9250 }, { "epoch": 0.1815686274509804, "grad_norm": 2.3969168663024902, "learning_rate": 9.954615042818222e-07, "loss": 0.1444, "step": 9260 }, { "epoch": 0.18176470588235294, "grad_norm": 2.1083595752716064, "learning_rate": 9.954396863829612e-07, "loss": 0.1674, "step": 9270 }, { "epoch": 0.18196078431372548, "grad_norm": 3.020693063735962, "learning_rate": 9.954178164074574e-07, "loss": 0.2011, "step": 9280 }, { "epoch": 0.18215686274509804, "grad_norm": 2.439800977706909, "learning_rate": 9.953958943576094e-07, "loss": 0.1497, "step": 9290 }, { "epoch": 0.18235294117647058, "grad_norm": 2.7243897914886475, "learning_rate": 9.953739202357217e-07, "loss": 0.165, "step": 9300 }, { "epoch": 0.18254901960784314, "grad_norm": 1.907538890838623, "learning_rate": 9.95351894044104e-07, "loss": 0.195, "step": 9310 }, { "epoch": 0.18274509803921568, "grad_norm": 2.8317954540252686, "learning_rate": 9.953298157850713e-07, "loss": 0.1651, "step": 9320 }, { "epoch": 0.18294117647058825, "grad_norm": 2.669828176498413, "learning_rate": 9.953076854609446e-07, "loss": 0.1529, "step": 9330 }, { "epoch": 0.18313725490196078, "grad_norm": 3.831561803817749, "learning_rate": 9.9528550307405e-07, "loss": 0.1818, "step": 9340 }, { "epoch": 0.18333333333333332, "grad_norm": 3.107996940612793, "learning_rate": 9.952632686267188e-07, "loss": 0.1944, "step": 9350 }, { "epoch": 0.18352941176470589, "grad_norm": 3.092655897140503, "learning_rate": 9.952409821212884e-07, "loss": 0.1712, "step": 9360 }, { "epoch": 0.18372549019607842, "grad_norm": 1.8516172170639038, "learning_rate": 9.952186435601014e-07, "loss": 0.1836, "step": 9370 }, { "epoch": 0.183921568627451, "grad_norm": 3.442471504211426, "learning_rate": 9.951962529455058e-07, "loss": 0.1695, "step": 9380 }, { "epoch": 0.18411764705882352, "grad_norm": 3.378180503845215, "learning_rate": 9.951738102798551e-07, "loss": 0.1909, "step": 9390 }, { "epoch": 0.1843137254901961, "grad_norm": 1.9433608055114746, "learning_rate": 9.951513155655081e-07, "loss": 0.1422, "step": 9400 }, { "epoch": 0.18450980392156863, "grad_norm": 1.8897285461425781, "learning_rate": 9.951287688048297e-07, "loss": 0.181, "step": 9410 }, { "epoch": 0.18470588235294116, "grad_norm": 3.2185990810394287, "learning_rate": 9.951061700001895e-07, "loss": 0.1699, "step": 9420 }, { "epoch": 0.18490196078431373, "grad_norm": 1.9820467233657837, "learning_rate": 9.95083519153963e-07, "loss": 0.1662, "step": 9430 }, { "epoch": 0.18509803921568627, "grad_norm": 2.425863265991211, "learning_rate": 9.95060816268531e-07, "loss": 0.183, "step": 9440 }, { "epoch": 0.18529411764705883, "grad_norm": 2.413529634475708, "learning_rate": 9.9503806134628e-07, "loss": 0.1715, "step": 9450 }, { "epoch": 0.18549019607843137, "grad_norm": 2.860034942626953, "learning_rate": 9.950152543896018e-07, "loss": 0.1695, "step": 9460 }, { "epoch": 0.18568627450980393, "grad_norm": 2.7054827213287354, "learning_rate": 9.949923954008933e-07, "loss": 0.1801, "step": 9470 }, { "epoch": 0.18588235294117647, "grad_norm": 3.9521799087524414, "learning_rate": 9.94969484382558e-07, "loss": 0.2084, "step": 9480 }, { "epoch": 0.186078431372549, "grad_norm": 3.3850016593933105, "learning_rate": 9.949465213370034e-07, "loss": 0.1608, "step": 9490 }, { "epoch": 0.18627450980392157, "grad_norm": 1.337053894996643, "learning_rate": 9.949235062666438e-07, "loss": 0.208, "step": 9500 }, { "epoch": 0.1864705882352941, "grad_norm": 2.6107280254364014, "learning_rate": 9.949004391738976e-07, "loss": 0.1816, "step": 9510 }, { "epoch": 0.18666666666666668, "grad_norm": 3.5164549350738525, "learning_rate": 9.948773200611902e-07, "loss": 0.1912, "step": 9520 }, { "epoch": 0.1868627450980392, "grad_norm": 3.8253538608551025, "learning_rate": 9.948541489309512e-07, "loss": 0.1812, "step": 9530 }, { "epoch": 0.18705882352941178, "grad_norm": 7.946608066558838, "learning_rate": 9.948309257856165e-07, "loss": 0.1783, "step": 9540 }, { "epoch": 0.18725490196078431, "grad_norm": 3.806922435760498, "learning_rate": 9.948076506276266e-07, "loss": 0.1806, "step": 9550 }, { "epoch": 0.18745098039215685, "grad_norm": 3.384982109069824, "learning_rate": 9.947843234594287e-07, "loss": 0.1845, "step": 9560 }, { "epoch": 0.18764705882352942, "grad_norm": 3.971163749694824, "learning_rate": 9.947609442834744e-07, "loss": 0.1846, "step": 9570 }, { "epoch": 0.18784313725490195, "grad_norm": 3.0595908164978027, "learning_rate": 9.94737513102221e-07, "loss": 0.1819, "step": 9580 }, { "epoch": 0.18803921568627452, "grad_norm": 3.9814367294311523, "learning_rate": 9.947140299181318e-07, "loss": 0.159, "step": 9590 }, { "epoch": 0.18823529411764706, "grad_norm": 1.8617308139801025, "learning_rate": 9.946904947336748e-07, "loss": 0.1734, "step": 9600 }, { "epoch": 0.18843137254901962, "grad_norm": 2.1451480388641357, "learning_rate": 9.946669075513237e-07, "loss": 0.1903, "step": 9610 }, { "epoch": 0.18862745098039216, "grad_norm": 2.218167781829834, "learning_rate": 9.946432683735583e-07, "loss": 0.1865, "step": 9620 }, { "epoch": 0.1888235294117647, "grad_norm": 2.010166883468628, "learning_rate": 9.946195772028629e-07, "loss": 0.1801, "step": 9630 }, { "epoch": 0.18901960784313726, "grad_norm": 2.298654794692993, "learning_rate": 9.945958340417281e-07, "loss": 0.1878, "step": 9640 }, { "epoch": 0.1892156862745098, "grad_norm": 1.9531927108764648, "learning_rate": 9.945720388926495e-07, "loss": 0.1734, "step": 9650 }, { "epoch": 0.18941176470588236, "grad_norm": 2.0184319019317627, "learning_rate": 9.94548191758128e-07, "loss": 0.1836, "step": 9660 }, { "epoch": 0.1896078431372549, "grad_norm": 2.274027109146118, "learning_rate": 9.945242926406703e-07, "loss": 0.1807, "step": 9670 }, { "epoch": 0.18980392156862744, "grad_norm": 1.8537201881408691, "learning_rate": 9.94500341542789e-07, "loss": 0.1921, "step": 9680 }, { "epoch": 0.19, "grad_norm": 5.245425224304199, "learning_rate": 9.944763384670007e-07, "loss": 0.1675, "step": 9690 }, { "epoch": 0.19019607843137254, "grad_norm": 3.275794506072998, "learning_rate": 9.944522834158292e-07, "loss": 0.1928, "step": 9700 }, { "epoch": 0.1903921568627451, "grad_norm": 3.0515005588531494, "learning_rate": 9.944281763918028e-07, "loss": 0.2215, "step": 9710 }, { "epoch": 0.19058823529411764, "grad_norm": 4.298598766326904, "learning_rate": 9.94404017397455e-07, "loss": 0.2092, "step": 9720 }, { "epoch": 0.1907843137254902, "grad_norm": 2.7523105144500732, "learning_rate": 9.94379806435326e-07, "loss": 0.1783, "step": 9730 }, { "epoch": 0.19098039215686274, "grad_norm": 4.949766159057617, "learning_rate": 9.943555435079599e-07, "loss": 0.1919, "step": 9740 }, { "epoch": 0.19117647058823528, "grad_norm": 2.006464958190918, "learning_rate": 9.943312286179073e-07, "loss": 0.2083, "step": 9750 }, { "epoch": 0.19137254901960785, "grad_norm": 2.9829325675964355, "learning_rate": 9.943068617677242e-07, "loss": 0.179, "step": 9760 }, { "epoch": 0.19156862745098038, "grad_norm": 3.231487989425659, "learning_rate": 9.942824429599714e-07, "loss": 0.1413, "step": 9770 }, { "epoch": 0.19176470588235295, "grad_norm": 3.086510419845581, "learning_rate": 9.942579721972161e-07, "loss": 0.176, "step": 9780 }, { "epoch": 0.19196078431372549, "grad_norm": 3.436910629272461, "learning_rate": 9.942334494820302e-07, "loss": 0.1665, "step": 9790 }, { "epoch": 0.19215686274509805, "grad_norm": 3.191215991973877, "learning_rate": 9.942088748169914e-07, "loss": 0.1897, "step": 9800 }, { "epoch": 0.1923529411764706, "grad_norm": 2.7022736072540283, "learning_rate": 9.941842482046827e-07, "loss": 0.1618, "step": 9810 }, { "epoch": 0.19254901960784312, "grad_norm": 2.48551869392395, "learning_rate": 9.941595696476928e-07, "loss": 0.1659, "step": 9820 }, { "epoch": 0.1927450980392157, "grad_norm": 4.949916839599609, "learning_rate": 9.941348391486155e-07, "loss": 0.189, "step": 9830 }, { "epoch": 0.19294117647058823, "grad_norm": 2.0677602291107178, "learning_rate": 9.941100567100503e-07, "loss": 0.1582, "step": 9840 }, { "epoch": 0.1931372549019608, "grad_norm": 4.488876819610596, "learning_rate": 9.940852223346025e-07, "loss": 0.1703, "step": 9850 }, { "epoch": 0.19333333333333333, "grad_norm": 3.978987455368042, "learning_rate": 9.94060336024882e-07, "loss": 0.155, "step": 9860 }, { "epoch": 0.1935294117647059, "grad_norm": 2.1414432525634766, "learning_rate": 9.940353977835052e-07, "loss": 0.1714, "step": 9870 }, { "epoch": 0.19372549019607843, "grad_norm": 2.921506404876709, "learning_rate": 9.94010407613093e-07, "loss": 0.2053, "step": 9880 }, { "epoch": 0.19392156862745097, "grad_norm": 3.0950207710266113, "learning_rate": 9.93985365516272e-07, "loss": 0.1599, "step": 9890 }, { "epoch": 0.19411764705882353, "grad_norm": 3.3762035369873047, "learning_rate": 9.939602714956747e-07, "loss": 0.1509, "step": 9900 }, { "epoch": 0.19431372549019607, "grad_norm": 4.223522186279297, "learning_rate": 9.939351255539388e-07, "loss": 0.1886, "step": 9910 }, { "epoch": 0.19450980392156864, "grad_norm": 3.465184450149536, "learning_rate": 9.939099276937073e-07, "loss": 0.183, "step": 9920 }, { "epoch": 0.19470588235294117, "grad_norm": 2.50283145904541, "learning_rate": 9.93884677917629e-07, "loss": 0.1702, "step": 9930 }, { "epoch": 0.19490196078431374, "grad_norm": 4.419827938079834, "learning_rate": 9.938593762283581e-07, "loss": 0.1869, "step": 9940 }, { "epoch": 0.19509803921568628, "grad_norm": 2.5644493103027344, "learning_rate": 9.938340226285534e-07, "loss": 0.1874, "step": 9950 }, { "epoch": 0.1952941176470588, "grad_norm": 3.1269032955169678, "learning_rate": 9.938086171208806e-07, "loss": 0.1559, "step": 9960 }, { "epoch": 0.19549019607843138, "grad_norm": 2.8141098022460938, "learning_rate": 9.937831597080098e-07, "loss": 0.1835, "step": 9970 }, { "epoch": 0.19568627450980391, "grad_norm": 2.812638282775879, "learning_rate": 9.937576503926167e-07, "loss": 0.1419, "step": 9980 }, { "epoch": 0.19588235294117648, "grad_norm": 2.607740640640259, "learning_rate": 9.93732089177383e-07, "loss": 0.2046, "step": 9990 }, { "epoch": 0.19607843137254902, "grad_norm": 2.2673757076263428, "learning_rate": 9.937064760649955e-07, "loss": 0.2188, "step": 10000 }, { "epoch": 0.19627450980392158, "grad_norm": 2.0601460933685303, "learning_rate": 9.936808110581461e-07, "loss": 0.1858, "step": 10010 }, { "epoch": 0.19647058823529412, "grad_norm": 2.9653773307800293, "learning_rate": 9.936550941595328e-07, "loss": 0.1667, "step": 10020 }, { "epoch": 0.19666666666666666, "grad_norm": 2.6720733642578125, "learning_rate": 9.936293253718586e-07, "loss": 0.1603, "step": 10030 }, { "epoch": 0.19686274509803922, "grad_norm": 3.0169005393981934, "learning_rate": 9.93603504697832e-07, "loss": 0.1837, "step": 10040 }, { "epoch": 0.19705882352941176, "grad_norm": 1.5626157522201538, "learning_rate": 9.935776321401675e-07, "loss": 0.1721, "step": 10050 }, { "epoch": 0.19725490196078432, "grad_norm": 2.5274500846862793, "learning_rate": 9.93551707701584e-07, "loss": 0.1778, "step": 10060 }, { "epoch": 0.19745098039215686, "grad_norm": 2.608076572418213, "learning_rate": 9.93525731384807e-07, "loss": 0.1846, "step": 10070 }, { "epoch": 0.1976470588235294, "grad_norm": 2.789475202560425, "learning_rate": 9.934997031925668e-07, "loss": 0.1922, "step": 10080 }, { "epoch": 0.19784313725490196, "grad_norm": 1.8384568691253662, "learning_rate": 9.93473623127599e-07, "loss": 0.1739, "step": 10090 }, { "epoch": 0.1980392156862745, "grad_norm": 3.232091188430786, "learning_rate": 9.934474911926453e-07, "loss": 0.1786, "step": 10100 }, { "epoch": 0.19823529411764707, "grad_norm": 2.406766176223755, "learning_rate": 9.934213073904521e-07, "loss": 0.1651, "step": 10110 }, { "epoch": 0.1984313725490196, "grad_norm": 2.457000732421875, "learning_rate": 9.93395071723772e-07, "loss": 0.1665, "step": 10120 }, { "epoch": 0.19862745098039217, "grad_norm": 3.034681797027588, "learning_rate": 9.933687841953624e-07, "loss": 0.1986, "step": 10130 }, { "epoch": 0.1988235294117647, "grad_norm": 4.424872875213623, "learning_rate": 9.933424448079865e-07, "loss": 0.1581, "step": 10140 }, { "epoch": 0.19901960784313724, "grad_norm": 1.9509388208389282, "learning_rate": 9.93316053564413e-07, "loss": 0.1809, "step": 10150 }, { "epoch": 0.1992156862745098, "grad_norm": 5.058659553527832, "learning_rate": 9.93289610467416e-07, "loss": 0.2066, "step": 10160 }, { "epoch": 0.19941176470588234, "grad_norm": 3.5896153450012207, "learning_rate": 9.932631155197746e-07, "loss": 0.1905, "step": 10170 }, { "epoch": 0.1996078431372549, "grad_norm": 2.0011239051818848, "learning_rate": 9.932365687242741e-07, "loss": 0.2048, "step": 10180 }, { "epoch": 0.19980392156862745, "grad_norm": 2.147430658340454, "learning_rate": 9.932099700837049e-07, "loss": 0.1702, "step": 10190 }, { "epoch": 0.2, "grad_norm": 2.405978202819824, "learning_rate": 9.931833196008625e-07, "loss": 0.1591, "step": 10200 }, { "epoch": 0.20019607843137255, "grad_norm": 2.6638479232788086, "learning_rate": 9.931566172785486e-07, "loss": 0.1571, "step": 10210 }, { "epoch": 0.20039215686274509, "grad_norm": 4.3879780769348145, "learning_rate": 9.931298631195698e-07, "loss": 0.2157, "step": 10220 }, { "epoch": 0.20058823529411765, "grad_norm": 3.2602813243865967, "learning_rate": 9.931030571267378e-07, "loss": 0.2219, "step": 10230 }, { "epoch": 0.2007843137254902, "grad_norm": 2.5136005878448486, "learning_rate": 9.93076199302871e-07, "loss": 0.1643, "step": 10240 }, { "epoch": 0.20098039215686275, "grad_norm": 8.51932430267334, "learning_rate": 9.93049289650792e-07, "loss": 0.1958, "step": 10250 }, { "epoch": 0.2011764705882353, "grad_norm": 1.8447037935256958, "learning_rate": 9.930223281733294e-07, "loss": 0.1909, "step": 10260 }, { "epoch": 0.20137254901960785, "grad_norm": 1.5373139381408691, "learning_rate": 9.929953148733172e-07, "loss": 0.1489, "step": 10270 }, { "epoch": 0.2015686274509804, "grad_norm": 1.7257084846496582, "learning_rate": 9.929682497535952e-07, "loss": 0.1476, "step": 10280 }, { "epoch": 0.20176470588235293, "grad_norm": 2.6621978282928467, "learning_rate": 9.929411328170075e-07, "loss": 0.1951, "step": 10290 }, { "epoch": 0.2019607843137255, "grad_norm": 3.2021191120147705, "learning_rate": 9.92913964066405e-07, "loss": 0.2073, "step": 10300 }, { "epoch": 0.20215686274509803, "grad_norm": 3.3065617084503174, "learning_rate": 9.928867435046433e-07, "loss": 0.1512, "step": 10310 }, { "epoch": 0.2023529411764706, "grad_norm": 2.0243186950683594, "learning_rate": 9.928594711345835e-07, "loss": 0.1618, "step": 10320 }, { "epoch": 0.20254901960784313, "grad_norm": 2.873044490814209, "learning_rate": 9.928321469590924e-07, "loss": 0.1867, "step": 10330 }, { "epoch": 0.2027450980392157, "grad_norm": 2.407919406890869, "learning_rate": 9.92804770981042e-07, "loss": 0.1878, "step": 10340 }, { "epoch": 0.20294117647058824, "grad_norm": 3.5230329036712646, "learning_rate": 9.9277734320331e-07, "loss": 0.1914, "step": 10350 }, { "epoch": 0.20313725490196077, "grad_norm": 3.382927179336548, "learning_rate": 9.927498636287792e-07, "loss": 0.1507, "step": 10360 }, { "epoch": 0.20333333333333334, "grad_norm": 2.605539083480835, "learning_rate": 9.927223322603383e-07, "loss": 0.1893, "step": 10370 }, { "epoch": 0.20352941176470588, "grad_norm": 2.8128983974456787, "learning_rate": 9.926947491008805e-07, "loss": 0.1734, "step": 10380 }, { "epoch": 0.20372549019607844, "grad_norm": 1.9513368606567383, "learning_rate": 9.926671141533062e-07, "loss": 0.1543, "step": 10390 }, { "epoch": 0.20392156862745098, "grad_norm": 2.170217752456665, "learning_rate": 9.926394274205193e-07, "loss": 0.1824, "step": 10400 }, { "epoch": 0.20411764705882354, "grad_norm": 3.384787082672119, "learning_rate": 9.9261168890543e-07, "loss": 0.1715, "step": 10410 }, { "epoch": 0.20431372549019608, "grad_norm": 7.126421928405762, "learning_rate": 9.925838986109546e-07, "loss": 0.2027, "step": 10420 }, { "epoch": 0.20450980392156862, "grad_norm": 3.5645618438720703, "learning_rate": 9.925560565400137e-07, "loss": 0.1882, "step": 10430 }, { "epoch": 0.20470588235294118, "grad_norm": 3.4626739025115967, "learning_rate": 9.92528162695534e-07, "loss": 0.2175, "step": 10440 }, { "epoch": 0.20490196078431372, "grad_norm": 14.35198974609375, "learning_rate": 9.925002170804472e-07, "loss": 0.1959, "step": 10450 }, { "epoch": 0.20509803921568628, "grad_norm": 4.715295314788818, "learning_rate": 9.924722196976915e-07, "loss": 0.2327, "step": 10460 }, { "epoch": 0.20529411764705882, "grad_norm": 3.1066551208496094, "learning_rate": 9.924441705502086e-07, "loss": 0.1804, "step": 10470 }, { "epoch": 0.2054901960784314, "grad_norm": 1.5931637287139893, "learning_rate": 9.924160696409477e-07, "loss": 0.1767, "step": 10480 }, { "epoch": 0.20568627450980392, "grad_norm": 3.2130885124206543, "learning_rate": 9.923879169728623e-07, "loss": 0.1734, "step": 10490 }, { "epoch": 0.20588235294117646, "grad_norm": 1.92341947555542, "learning_rate": 9.923597125489114e-07, "loss": 0.1742, "step": 10500 }, { "epoch": 0.20607843137254903, "grad_norm": 2.82692551612854, "learning_rate": 9.923314563720596e-07, "loss": 0.1609, "step": 10510 }, { "epoch": 0.20627450980392156, "grad_norm": 2.5197830200195312, "learning_rate": 9.923031484452776e-07, "loss": 0.1614, "step": 10520 }, { "epoch": 0.20647058823529413, "grad_norm": 2.875676393508911, "learning_rate": 9.922747887715401e-07, "loss": 0.1808, "step": 10530 }, { "epoch": 0.20666666666666667, "grad_norm": 3.3176443576812744, "learning_rate": 9.922463773538283e-07, "loss": 0.1682, "step": 10540 }, { "epoch": 0.2068627450980392, "grad_norm": 1.7781569957733154, "learning_rate": 9.922179141951288e-07, "loss": 0.1977, "step": 10550 }, { "epoch": 0.20705882352941177, "grad_norm": 3.648460626602173, "learning_rate": 9.921893992984332e-07, "loss": 0.1928, "step": 10560 }, { "epoch": 0.2072549019607843, "grad_norm": 3.7801687717437744, "learning_rate": 9.921608326667388e-07, "loss": 0.194, "step": 10570 }, { "epoch": 0.20745098039215687, "grad_norm": 3.6482207775115967, "learning_rate": 9.921322143030484e-07, "loss": 0.1749, "step": 10580 }, { "epoch": 0.2076470588235294, "grad_norm": 3.4673917293548584, "learning_rate": 9.9210354421037e-07, "loss": 0.1759, "step": 10590 }, { "epoch": 0.20784313725490197, "grad_norm": 1.8823972940444946, "learning_rate": 9.920748223917173e-07, "loss": 0.1708, "step": 10600 }, { "epoch": 0.2080392156862745, "grad_norm": 1.9498069286346436, "learning_rate": 9.920460488501091e-07, "loss": 0.1781, "step": 10610 }, { "epoch": 0.20823529411764705, "grad_norm": 3.3815603256225586, "learning_rate": 9.9201722358857e-07, "loss": 0.1925, "step": 10620 }, { "epoch": 0.2084313725490196, "grad_norm": 4.099130153656006, "learning_rate": 9.9198834661013e-07, "loss": 0.1921, "step": 10630 }, { "epoch": 0.20862745098039215, "grad_norm": 2.0449235439300537, "learning_rate": 9.919594179178242e-07, "loss": 0.1742, "step": 10640 }, { "epoch": 0.2088235294117647, "grad_norm": 1.8203942775726318, "learning_rate": 9.919304375146932e-07, "loss": 0.224, "step": 10650 }, { "epoch": 0.20901960784313725, "grad_norm": 3.417109251022339, "learning_rate": 9.919014054037835e-07, "loss": 0.1728, "step": 10660 }, { "epoch": 0.20921568627450982, "grad_norm": 6.422234058380127, "learning_rate": 9.918723215881467e-07, "loss": 0.1754, "step": 10670 }, { "epoch": 0.20941176470588235, "grad_norm": 4.659250259399414, "learning_rate": 9.918431860708398e-07, "loss": 0.1753, "step": 10680 }, { "epoch": 0.2096078431372549, "grad_norm": 3.748878240585327, "learning_rate": 9.918139988549253e-07, "loss": 0.1832, "step": 10690 }, { "epoch": 0.20980392156862746, "grad_norm": 2.0956571102142334, "learning_rate": 9.91784759943471e-07, "loss": 0.1631, "step": 10700 }, { "epoch": 0.21, "grad_norm": 1.295262098312378, "learning_rate": 9.917554693395504e-07, "loss": 0.1494, "step": 10710 }, { "epoch": 0.21019607843137256, "grad_norm": 3.3781840801239014, "learning_rate": 9.917261270462423e-07, "loss": 0.169, "step": 10720 }, { "epoch": 0.2103921568627451, "grad_norm": 4.571416854858398, "learning_rate": 9.916967330666309e-07, "loss": 0.1744, "step": 10730 }, { "epoch": 0.21058823529411766, "grad_norm": 3.108875274658203, "learning_rate": 9.916672874038056e-07, "loss": 0.1791, "step": 10740 }, { "epoch": 0.2107843137254902, "grad_norm": 2.4681007862091064, "learning_rate": 9.916377900608619e-07, "loss": 0.191, "step": 10750 }, { "epoch": 0.21098039215686273, "grad_norm": 1.939278244972229, "learning_rate": 9.916082410409001e-07, "loss": 0.1818, "step": 10760 }, { "epoch": 0.2111764705882353, "grad_norm": 3.8861424922943115, "learning_rate": 9.915786403470264e-07, "loss": 0.1655, "step": 10770 }, { "epoch": 0.21137254901960784, "grad_norm": 3.244882106781006, "learning_rate": 9.915489879823518e-07, "loss": 0.1867, "step": 10780 }, { "epoch": 0.2115686274509804, "grad_norm": 3.5621280670166016, "learning_rate": 9.915192839499933e-07, "loss": 0.1912, "step": 10790 }, { "epoch": 0.21176470588235294, "grad_norm": 3.099177598953247, "learning_rate": 9.914895282530734e-07, "loss": 0.1736, "step": 10800 }, { "epoch": 0.2119607843137255, "grad_norm": 2.463710069656372, "learning_rate": 9.914597208947194e-07, "loss": 0.2189, "step": 10810 }, { "epoch": 0.21215686274509804, "grad_norm": 2.7170937061309814, "learning_rate": 9.914298618780646e-07, "loss": 0.2018, "step": 10820 }, { "epoch": 0.21235294117647058, "grad_norm": 2.7253429889678955, "learning_rate": 9.913999512062475e-07, "loss": 0.1773, "step": 10830 }, { "epoch": 0.21254901960784314, "grad_norm": 2.8462541103363037, "learning_rate": 9.91369988882412e-07, "loss": 0.1692, "step": 10840 }, { "epoch": 0.21274509803921568, "grad_norm": 3.543452501296997, "learning_rate": 9.913399749097076e-07, "loss": 0.1832, "step": 10850 }, { "epoch": 0.21294117647058824, "grad_norm": 2.972637176513672, "learning_rate": 9.91309909291289e-07, "loss": 0.1642, "step": 10860 }, { "epoch": 0.21313725490196078, "grad_norm": 3.2268848419189453, "learning_rate": 9.912797920303167e-07, "loss": 0.1601, "step": 10870 }, { "epoch": 0.21333333333333335, "grad_norm": 2.125889301300049, "learning_rate": 9.91249623129956e-07, "loss": 0.1838, "step": 10880 }, { "epoch": 0.21352941176470588, "grad_norm": 2.3399617671966553, "learning_rate": 9.912194025933783e-07, "loss": 0.1718, "step": 10890 }, { "epoch": 0.21372549019607842, "grad_norm": 1.9145036935806274, "learning_rate": 9.911891304237603e-07, "loss": 0.1843, "step": 10900 }, { "epoch": 0.213921568627451, "grad_norm": 2.7048158645629883, "learning_rate": 9.911588066242838e-07, "loss": 0.2133, "step": 10910 }, { "epoch": 0.21411764705882352, "grad_norm": 5.95388126373291, "learning_rate": 9.911284311981359e-07, "loss": 0.1945, "step": 10920 }, { "epoch": 0.2143137254901961, "grad_norm": 3.385345935821533, "learning_rate": 9.910980041485097e-07, "loss": 0.1766, "step": 10930 }, { "epoch": 0.21450980392156863, "grad_norm": 3.6890037059783936, "learning_rate": 9.910675254786036e-07, "loss": 0.2551, "step": 10940 }, { "epoch": 0.21470588235294116, "grad_norm": 2.2170655727386475, "learning_rate": 9.91036995191621e-07, "loss": 0.1587, "step": 10950 }, { "epoch": 0.21490196078431373, "grad_norm": 2.97515606880188, "learning_rate": 9.910064132907712e-07, "loss": 0.1869, "step": 10960 }, { "epoch": 0.21509803921568627, "grad_norm": 3.3899025917053223, "learning_rate": 9.909757797792684e-07, "loss": 0.1486, "step": 10970 }, { "epoch": 0.21529411764705883, "grad_norm": 5.0938401222229, "learning_rate": 9.90945094660333e-07, "loss": 0.1709, "step": 10980 }, { "epoch": 0.21549019607843137, "grad_norm": 2.5038702487945557, "learning_rate": 9.909143579371897e-07, "loss": 0.1687, "step": 10990 }, { "epoch": 0.21568627450980393, "grad_norm": 2.768780469894409, "learning_rate": 9.9088356961307e-07, "loss": 0.1968, "step": 11000 }, { "epoch": 0.21588235294117647, "grad_norm": 2.0253849029541016, "learning_rate": 9.908527296912099e-07, "loss": 0.17, "step": 11010 }, { "epoch": 0.216078431372549, "grad_norm": 2.6945888996124268, "learning_rate": 9.90821838174851e-07, "loss": 0.1735, "step": 11020 }, { "epoch": 0.21627450980392157, "grad_norm": 2.8105242252349854, "learning_rate": 9.907908950672402e-07, "loss": 0.1772, "step": 11030 }, { "epoch": 0.2164705882352941, "grad_norm": 3.0674164295196533, "learning_rate": 9.907599003716305e-07, "loss": 0.2053, "step": 11040 }, { "epoch": 0.21666666666666667, "grad_norm": 2.825551748275757, "learning_rate": 9.907288540912792e-07, "loss": 0.1737, "step": 11050 }, { "epoch": 0.2168627450980392, "grad_norm": 2.9103057384490967, "learning_rate": 9.9069775622945e-07, "loss": 0.1878, "step": 11060 }, { "epoch": 0.21705882352941178, "grad_norm": 2.6731786727905273, "learning_rate": 9.906666067894115e-07, "loss": 0.1749, "step": 11070 }, { "epoch": 0.2172549019607843, "grad_norm": 2.3222906589508057, "learning_rate": 9.906354057744379e-07, "loss": 0.1418, "step": 11080 }, { "epoch": 0.21745098039215685, "grad_norm": 2.473581314086914, "learning_rate": 9.90604153187809e-07, "loss": 0.2073, "step": 11090 }, { "epoch": 0.21764705882352942, "grad_norm": 2.0490713119506836, "learning_rate": 9.905728490328096e-07, "loss": 0.1765, "step": 11100 }, { "epoch": 0.21784313725490195, "grad_norm": 2.8773348331451416, "learning_rate": 9.905414933127302e-07, "loss": 0.1945, "step": 11110 }, { "epoch": 0.21803921568627452, "grad_norm": 1.7416884899139404, "learning_rate": 9.905100860308668e-07, "loss": 0.1571, "step": 11120 }, { "epoch": 0.21823529411764706, "grad_norm": 2.591438055038452, "learning_rate": 9.904786271905203e-07, "loss": 0.1798, "step": 11130 }, { "epoch": 0.21843137254901962, "grad_norm": 3.799985885620117, "learning_rate": 9.90447116794998e-07, "loss": 0.2302, "step": 11140 }, { "epoch": 0.21862745098039216, "grad_norm": 2.833477258682251, "learning_rate": 9.904155548476112e-07, "loss": 0.1654, "step": 11150 }, { "epoch": 0.2188235294117647, "grad_norm": 3.561474084854126, "learning_rate": 9.903839413516783e-07, "loss": 0.168, "step": 11160 }, { "epoch": 0.21901960784313726, "grad_norm": 2.5616235733032227, "learning_rate": 9.903522763105216e-07, "loss": 0.1883, "step": 11170 }, { "epoch": 0.2192156862745098, "grad_norm": 2.2266509532928467, "learning_rate": 9.9032055972747e-07, "loss": 0.1361, "step": 11180 }, { "epoch": 0.21941176470588236, "grad_norm": 2.0477688312530518, "learning_rate": 9.902887916058568e-07, "loss": 0.1537, "step": 11190 }, { "epoch": 0.2196078431372549, "grad_norm": 2.6420669555664062, "learning_rate": 9.902569719490216e-07, "loss": 0.1862, "step": 11200 }, { "epoch": 0.21980392156862746, "grad_norm": 1.3863639831542969, "learning_rate": 9.90225100760309e-07, "loss": 0.1606, "step": 11210 }, { "epoch": 0.22, "grad_norm": 1.796696424484253, "learning_rate": 9.901931780430685e-07, "loss": 0.1164, "step": 11220 }, { "epoch": 0.22019607843137254, "grad_norm": 2.71370005607605, "learning_rate": 9.901612038006562e-07, "loss": 0.1594, "step": 11230 }, { "epoch": 0.2203921568627451, "grad_norm": 2.356344223022461, "learning_rate": 9.901291780364328e-07, "loss": 0.1849, "step": 11240 }, { "epoch": 0.22058823529411764, "grad_norm": 5.46912956237793, "learning_rate": 9.900971007537646e-07, "loss": 0.1881, "step": 11250 }, { "epoch": 0.2207843137254902, "grad_norm": 4.415577411651611, "learning_rate": 9.90064971956023e-07, "loss": 0.1795, "step": 11260 }, { "epoch": 0.22098039215686274, "grad_norm": 2.5699450969696045, "learning_rate": 9.900327916465857e-07, "loss": 0.1699, "step": 11270 }, { "epoch": 0.2211764705882353, "grad_norm": 4.4226813316345215, "learning_rate": 9.900005598288345e-07, "loss": 0.1397, "step": 11280 }, { "epoch": 0.22137254901960784, "grad_norm": 2.336641550064087, "learning_rate": 9.89968276506158e-07, "loss": 0.1657, "step": 11290 }, { "epoch": 0.22156862745098038, "grad_norm": 2.177450656890869, "learning_rate": 9.899359416819493e-07, "loss": 0.178, "step": 11300 }, { "epoch": 0.22176470588235295, "grad_norm": 2.3312788009643555, "learning_rate": 9.899035553596073e-07, "loss": 0.1748, "step": 11310 }, { "epoch": 0.22196078431372548, "grad_norm": 2.7555551528930664, "learning_rate": 9.89871117542536e-07, "loss": 0.1689, "step": 11320 }, { "epoch": 0.22215686274509805, "grad_norm": 3.5507025718688965, "learning_rate": 9.89838628234145e-07, "loss": 0.209, "step": 11330 }, { "epoch": 0.2223529411764706, "grad_norm": 2.7689125537872314, "learning_rate": 9.898060874378495e-07, "loss": 0.1723, "step": 11340 }, { "epoch": 0.22254901960784312, "grad_norm": 3.189014434814453, "learning_rate": 9.897734951570697e-07, "loss": 0.2, "step": 11350 }, { "epoch": 0.2227450980392157, "grad_norm": 2.4776713848114014, "learning_rate": 9.897408513952316e-07, "loss": 0.1498, "step": 11360 }, { "epoch": 0.22294117647058823, "grad_norm": 6.618722915649414, "learning_rate": 9.897081561557666e-07, "loss": 0.1786, "step": 11370 }, { "epoch": 0.2231372549019608, "grad_norm": 2.044400691986084, "learning_rate": 9.89675409442111e-07, "loss": 0.1997, "step": 11380 }, { "epoch": 0.22333333333333333, "grad_norm": 3.4189045429229736, "learning_rate": 9.896426112577071e-07, "loss": 0.1879, "step": 11390 }, { "epoch": 0.2235294117647059, "grad_norm": 1.3487788438796997, "learning_rate": 9.896097616060025e-07, "loss": 0.1427, "step": 11400 }, { "epoch": 0.22372549019607843, "grad_norm": 1.4452611207962036, "learning_rate": 9.895768604904498e-07, "loss": 0.1831, "step": 11410 }, { "epoch": 0.22392156862745097, "grad_norm": 1.935536503791809, "learning_rate": 9.895439079145073e-07, "loss": 0.1884, "step": 11420 }, { "epoch": 0.22411764705882353, "grad_norm": 2.423469066619873, "learning_rate": 9.89510903881639e-07, "loss": 0.2094, "step": 11430 }, { "epoch": 0.22431372549019607, "grad_norm": 6.385718822479248, "learning_rate": 9.89477848395314e-07, "loss": 0.1767, "step": 11440 }, { "epoch": 0.22450980392156863, "grad_norm": 1.6027650833129883, "learning_rate": 9.894447414590063e-07, "loss": 0.1942, "step": 11450 }, { "epoch": 0.22470588235294117, "grad_norm": 4.201247692108154, "learning_rate": 9.894115830761965e-07, "loss": 0.1953, "step": 11460 }, { "epoch": 0.22490196078431374, "grad_norm": 1.4320788383483887, "learning_rate": 9.893783732503696e-07, "loss": 0.1671, "step": 11470 }, { "epoch": 0.22509803921568627, "grad_norm": 2.789837598800659, "learning_rate": 9.893451119850164e-07, "loss": 0.1849, "step": 11480 }, { "epoch": 0.2252941176470588, "grad_norm": 2.76719069480896, "learning_rate": 9.893117992836331e-07, "loss": 0.1809, "step": 11490 }, { "epoch": 0.22549019607843138, "grad_norm": 5.126953125, "learning_rate": 9.892784351497211e-07, "loss": 0.1805, "step": 11500 }, { "epoch": 0.2256862745098039, "grad_norm": 2.308375597000122, "learning_rate": 9.892450195867877e-07, "loss": 0.178, "step": 11510 }, { "epoch": 0.22588235294117648, "grad_norm": 2.7283546924591064, "learning_rate": 9.892115525983449e-07, "loss": 0.1787, "step": 11520 }, { "epoch": 0.22607843137254902, "grad_norm": 2.227639675140381, "learning_rate": 9.891780341879108e-07, "loss": 0.2004, "step": 11530 }, { "epoch": 0.22627450980392158, "grad_norm": 3.0813345909118652, "learning_rate": 9.891444643590085e-07, "loss": 0.1802, "step": 11540 }, { "epoch": 0.22647058823529412, "grad_norm": 5.161715030670166, "learning_rate": 9.891108431151664e-07, "loss": 0.1809, "step": 11550 }, { "epoch": 0.22666666666666666, "grad_norm": 1.9214069843292236, "learning_rate": 9.890771704599185e-07, "loss": 0.1974, "step": 11560 }, { "epoch": 0.22686274509803922, "grad_norm": 2.170832872390747, "learning_rate": 9.890434463968047e-07, "loss": 0.1943, "step": 11570 }, { "epoch": 0.22705882352941176, "grad_norm": 2.154695987701416, "learning_rate": 9.890096709293691e-07, "loss": 0.1546, "step": 11580 }, { "epoch": 0.22725490196078432, "grad_norm": 1.4538726806640625, "learning_rate": 9.889758440611624e-07, "loss": 0.1702, "step": 11590 }, { "epoch": 0.22745098039215686, "grad_norm": 4.004199028015137, "learning_rate": 9.8894196579574e-07, "loss": 0.1824, "step": 11600 }, { "epoch": 0.22764705882352942, "grad_norm": 2.9116928577423096, "learning_rate": 9.889080361366632e-07, "loss": 0.1669, "step": 11610 }, { "epoch": 0.22784313725490196, "grad_norm": 2.529592752456665, "learning_rate": 9.888740550874978e-07, "loss": 0.1588, "step": 11620 }, { "epoch": 0.2280392156862745, "grad_norm": 2.9911611080169678, "learning_rate": 9.888400226518163e-07, "loss": 0.2134, "step": 11630 }, { "epoch": 0.22823529411764706, "grad_norm": 2.378365993499756, "learning_rate": 9.888059388331955e-07, "loss": 0.1529, "step": 11640 }, { "epoch": 0.2284313725490196, "grad_norm": 3.6135828495025635, "learning_rate": 9.88771803635218e-07, "loss": 0.1819, "step": 11650 }, { "epoch": 0.22862745098039217, "grad_norm": 2.0288000106811523, "learning_rate": 9.88737617061472e-07, "loss": 0.1844, "step": 11660 }, { "epoch": 0.2288235294117647, "grad_norm": 1.5987428426742554, "learning_rate": 9.88703379115551e-07, "loss": 0.1475, "step": 11670 }, { "epoch": 0.22901960784313727, "grad_norm": 2.1229517459869385, "learning_rate": 9.886690898010534e-07, "loss": 0.1471, "step": 11680 }, { "epoch": 0.2292156862745098, "grad_norm": 1.066391944885254, "learning_rate": 9.88634749121584e-07, "loss": 0.1713, "step": 11690 }, { "epoch": 0.22941176470588234, "grad_norm": 5.059505462646484, "learning_rate": 9.88600357080752e-07, "loss": 0.1852, "step": 11700 }, { "epoch": 0.2296078431372549, "grad_norm": 2.931218385696411, "learning_rate": 9.885659136821725e-07, "loss": 0.1979, "step": 11710 }, { "epoch": 0.22980392156862745, "grad_norm": 1.5540964603424072, "learning_rate": 9.88531418929466e-07, "loss": 0.1732, "step": 11720 }, { "epoch": 0.23, "grad_norm": 8.86218547821045, "learning_rate": 9.884968728262582e-07, "loss": 0.1883, "step": 11730 }, { "epoch": 0.23019607843137255, "grad_norm": 2.6698617935180664, "learning_rate": 9.884622753761805e-07, "loss": 0.1952, "step": 11740 }, { "epoch": 0.23039215686274508, "grad_norm": 3.920872926712036, "learning_rate": 9.88427626582869e-07, "loss": 0.1819, "step": 11750 }, { "epoch": 0.23058823529411765, "grad_norm": 3.243504285812378, "learning_rate": 9.883929264499665e-07, "loss": 0.1681, "step": 11760 }, { "epoch": 0.2307843137254902, "grad_norm": 2.5490894317626953, "learning_rate": 9.883581749811198e-07, "loss": 0.1475, "step": 11770 }, { "epoch": 0.23098039215686275, "grad_norm": 4.179109573364258, "learning_rate": 9.883233721799818e-07, "loss": 0.1761, "step": 11780 }, { "epoch": 0.2311764705882353, "grad_norm": 2.57543683052063, "learning_rate": 9.882885180502108e-07, "loss": 0.1581, "step": 11790 }, { "epoch": 0.23137254901960785, "grad_norm": 2.385228395462036, "learning_rate": 9.882536125954703e-07, "loss": 0.164, "step": 11800 }, { "epoch": 0.2315686274509804, "grad_norm": 2.581437826156616, "learning_rate": 9.882186558194293e-07, "loss": 0.1712, "step": 11810 }, { "epoch": 0.23176470588235293, "grad_norm": 2.421449899673462, "learning_rate": 9.881836477257624e-07, "loss": 0.1865, "step": 11820 }, { "epoch": 0.2319607843137255, "grad_norm": 4.001059532165527, "learning_rate": 9.881485883181492e-07, "loss": 0.1706, "step": 11830 }, { "epoch": 0.23215686274509803, "grad_norm": 1.9798890352249146, "learning_rate": 9.881134776002744e-07, "loss": 0.1865, "step": 11840 }, { "epoch": 0.2323529411764706, "grad_norm": 1.4766466617584229, "learning_rate": 9.880783155758294e-07, "loss": 0.1638, "step": 11850 }, { "epoch": 0.23254901960784313, "grad_norm": 3.9924750328063965, "learning_rate": 9.880431022485097e-07, "loss": 0.2059, "step": 11860 }, { "epoch": 0.2327450980392157, "grad_norm": 2.5536441802978516, "learning_rate": 9.880078376220164e-07, "loss": 0.1733, "step": 11870 }, { "epoch": 0.23294117647058823, "grad_norm": 5.877531051635742, "learning_rate": 9.87972521700057e-07, "loss": 0.1894, "step": 11880 }, { "epoch": 0.23313725490196077, "grad_norm": 3.32609224319458, "learning_rate": 9.879371544863426e-07, "loss": 0.1844, "step": 11890 }, { "epoch": 0.23333333333333334, "grad_norm": 3.952564001083374, "learning_rate": 9.879017359845917e-07, "loss": 0.2019, "step": 11900 }, { "epoch": 0.23352941176470587, "grad_norm": 2.7884485721588135, "learning_rate": 9.878662661985264e-07, "loss": 0.2096, "step": 11910 }, { "epoch": 0.23372549019607844, "grad_norm": 2.027512311935425, "learning_rate": 9.878307451318758e-07, "loss": 0.1773, "step": 11920 }, { "epoch": 0.23392156862745098, "grad_norm": 3.3238208293914795, "learning_rate": 9.87795172788373e-07, "loss": 0.1933, "step": 11930 }, { "epoch": 0.23411764705882354, "grad_norm": 1.6844696998596191, "learning_rate": 9.87759549171757e-07, "loss": 0.1297, "step": 11940 }, { "epoch": 0.23431372549019608, "grad_norm": 2.0296874046325684, "learning_rate": 9.877238742857727e-07, "loss": 0.2104, "step": 11950 }, { "epoch": 0.23450980392156862, "grad_norm": 3.103407382965088, "learning_rate": 9.876881481341697e-07, "loss": 0.1745, "step": 11960 }, { "epoch": 0.23470588235294118, "grad_norm": 3.339665174484253, "learning_rate": 9.876523707207034e-07, "loss": 0.1737, "step": 11970 }, { "epoch": 0.23490196078431372, "grad_norm": 3.7519161701202393, "learning_rate": 9.876165420491342e-07, "loss": 0.1927, "step": 11980 }, { "epoch": 0.23509803921568628, "grad_norm": 2.0055015087127686, "learning_rate": 9.875806621232283e-07, "loss": 0.1908, "step": 11990 }, { "epoch": 0.23529411764705882, "grad_norm": 1.7609869241714478, "learning_rate": 9.875447309467572e-07, "loss": 0.1732, "step": 12000 }, { "epoch": 0.23549019607843139, "grad_norm": 2.90618896484375, "learning_rate": 9.875087485234973e-07, "loss": 0.1745, "step": 12010 }, { "epoch": 0.23568627450980392, "grad_norm": 2.748868227005005, "learning_rate": 9.874727148572314e-07, "loss": 0.1833, "step": 12020 }, { "epoch": 0.23588235294117646, "grad_norm": 1.8543007373809814, "learning_rate": 9.874366299517466e-07, "loss": 0.21, "step": 12030 }, { "epoch": 0.23607843137254902, "grad_norm": 2.1918978691101074, "learning_rate": 9.874004938108361e-07, "loss": 0.1816, "step": 12040 }, { "epoch": 0.23627450980392156, "grad_norm": 2.3377084732055664, "learning_rate": 9.87364306438298e-07, "loss": 0.1753, "step": 12050 }, { "epoch": 0.23647058823529413, "grad_norm": 1.73680579662323, "learning_rate": 9.87328067837936e-07, "loss": 0.182, "step": 12060 }, { "epoch": 0.23666666666666666, "grad_norm": 2.7381200790405273, "learning_rate": 9.872917780135594e-07, "loss": 0.1915, "step": 12070 }, { "epoch": 0.23686274509803923, "grad_norm": 3.094179630279541, "learning_rate": 9.87255436968983e-07, "loss": 0.1803, "step": 12080 }, { "epoch": 0.23705882352941177, "grad_norm": 2.88330078125, "learning_rate": 9.87219044708026e-07, "loss": 0.2127, "step": 12090 }, { "epoch": 0.2372549019607843, "grad_norm": 2.361603260040283, "learning_rate": 9.871826012345141e-07, "loss": 0.2141, "step": 12100 }, { "epoch": 0.23745098039215687, "grad_norm": 3.416677236557007, "learning_rate": 9.871461065522778e-07, "loss": 0.1806, "step": 12110 }, { "epoch": 0.2376470588235294, "grad_norm": 4.9980292320251465, "learning_rate": 9.871095606651532e-07, "loss": 0.1967, "step": 12120 }, { "epoch": 0.23784313725490197, "grad_norm": 3.6967828273773193, "learning_rate": 9.870729635769816e-07, "loss": 0.1923, "step": 12130 }, { "epoch": 0.2380392156862745, "grad_norm": 2.694230556488037, "learning_rate": 9.8703631529161e-07, "loss": 0.1472, "step": 12140 }, { "epoch": 0.23823529411764705, "grad_norm": 1.2287651300430298, "learning_rate": 9.869996158128903e-07, "loss": 0.184, "step": 12150 }, { "epoch": 0.2384313725490196, "grad_norm": 3.5003581047058105, "learning_rate": 9.869628651446803e-07, "loss": 0.1854, "step": 12160 }, { "epoch": 0.23862745098039215, "grad_norm": 2.677340269088745, "learning_rate": 9.86926063290843e-07, "loss": 0.1475, "step": 12170 }, { "epoch": 0.2388235294117647, "grad_norm": 2.3696844577789307, "learning_rate": 9.868892102552465e-07, "loss": 0.1775, "step": 12180 }, { "epoch": 0.23901960784313725, "grad_norm": 2.3659119606018066, "learning_rate": 9.868523060417645e-07, "loss": 0.1988, "step": 12190 }, { "epoch": 0.23921568627450981, "grad_norm": 3.3760998249053955, "learning_rate": 9.86815350654276e-07, "loss": 0.2008, "step": 12200 }, { "epoch": 0.23941176470588235, "grad_norm": 1.6655700206756592, "learning_rate": 9.867783440966655e-07, "loss": 0.1452, "step": 12210 }, { "epoch": 0.2396078431372549, "grad_norm": 2.4615936279296875, "learning_rate": 9.86741286372823e-07, "loss": 0.1835, "step": 12220 }, { "epoch": 0.23980392156862745, "grad_norm": 3.392470121383667, "learning_rate": 9.867041774866437e-07, "loss": 0.1927, "step": 12230 }, { "epoch": 0.24, "grad_norm": 3.989393472671509, "learning_rate": 9.86667017442028e-07, "loss": 0.1456, "step": 12240 }, { "epoch": 0.24019607843137256, "grad_norm": 1.9006162881851196, "learning_rate": 9.866298062428822e-07, "loss": 0.1615, "step": 12250 }, { "epoch": 0.2403921568627451, "grad_norm": 2.492732524871826, "learning_rate": 9.865925438931172e-07, "loss": 0.1836, "step": 12260 }, { "epoch": 0.24058823529411766, "grad_norm": 2.7476296424865723, "learning_rate": 9.8655523039665e-07, "loss": 0.1637, "step": 12270 }, { "epoch": 0.2407843137254902, "grad_norm": 4.355039119720459, "learning_rate": 9.865178657574026e-07, "loss": 0.1783, "step": 12280 }, { "epoch": 0.24098039215686273, "grad_norm": 2.899683713912964, "learning_rate": 9.864804499793026e-07, "loss": 0.1579, "step": 12290 }, { "epoch": 0.2411764705882353, "grad_norm": 5.22709846496582, "learning_rate": 9.864429830662826e-07, "loss": 0.1643, "step": 12300 }, { "epoch": 0.24137254901960783, "grad_norm": 3.0263278484344482, "learning_rate": 9.86405465022281e-07, "loss": 0.2233, "step": 12310 }, { "epoch": 0.2415686274509804, "grad_norm": 2.8173418045043945, "learning_rate": 9.863678958512413e-07, "loss": 0.1861, "step": 12320 }, { "epoch": 0.24176470588235294, "grad_norm": 3.194180727005005, "learning_rate": 9.863302755571128e-07, "loss": 0.139, "step": 12330 }, { "epoch": 0.2419607843137255, "grad_norm": 2.5850346088409424, "learning_rate": 9.862926041438493e-07, "loss": 0.163, "step": 12340 }, { "epoch": 0.24215686274509804, "grad_norm": 3.365431547164917, "learning_rate": 9.862548816154108e-07, "loss": 0.2021, "step": 12350 }, { "epoch": 0.24235294117647058, "grad_norm": 2.946910858154297, "learning_rate": 9.862171079757626e-07, "loss": 0.1604, "step": 12360 }, { "epoch": 0.24254901960784314, "grad_norm": 2.3259336948394775, "learning_rate": 9.86179283228875e-07, "loss": 0.1689, "step": 12370 }, { "epoch": 0.24274509803921568, "grad_norm": 2.4714818000793457, "learning_rate": 9.861414073787234e-07, "loss": 0.1461, "step": 12380 }, { "epoch": 0.24294117647058824, "grad_norm": 2.037689685821533, "learning_rate": 9.861034804292895e-07, "loss": 0.1778, "step": 12390 }, { "epoch": 0.24313725490196078, "grad_norm": 4.6607666015625, "learning_rate": 9.8606550238456e-07, "loss": 0.1874, "step": 12400 }, { "epoch": 0.24333333333333335, "grad_norm": 3.1703684329986572, "learning_rate": 9.860274732485262e-07, "loss": 0.164, "step": 12410 }, { "epoch": 0.24352941176470588, "grad_norm": 1.863977313041687, "learning_rate": 9.859893930251862e-07, "loss": 0.1619, "step": 12420 }, { "epoch": 0.24372549019607842, "grad_norm": 1.6084284782409668, "learning_rate": 9.859512617185422e-07, "loss": 0.1774, "step": 12430 }, { "epoch": 0.24392156862745099, "grad_norm": 2.2002758979797363, "learning_rate": 9.859130793326024e-07, "loss": 0.15, "step": 12440 }, { "epoch": 0.24411764705882352, "grad_norm": 2.4956421852111816, "learning_rate": 9.858748458713801e-07, "loss": 0.1573, "step": 12450 }, { "epoch": 0.2443137254901961, "grad_norm": 2.074476480484009, "learning_rate": 9.858365613388942e-07, "loss": 0.2044, "step": 12460 }, { "epoch": 0.24450980392156862, "grad_norm": 2.3653147220611572, "learning_rate": 9.85798225739169e-07, "loss": 0.217, "step": 12470 }, { "epoch": 0.2447058823529412, "grad_norm": 2.777721881866455, "learning_rate": 9.857598390762337e-07, "loss": 0.2124, "step": 12480 }, { "epoch": 0.24490196078431373, "grad_norm": 2.094691753387451, "learning_rate": 9.857214013541235e-07, "loss": 0.1652, "step": 12490 }, { "epoch": 0.24509803921568626, "grad_norm": 0.7753708362579346, "learning_rate": 9.856829125768784e-07, "loss": 0.1681, "step": 12500 }, { "epoch": 0.24529411764705883, "grad_norm": 2.3328115940093994, "learning_rate": 9.856443727485444e-07, "loss": 0.1773, "step": 12510 }, { "epoch": 0.24549019607843137, "grad_norm": 2.5875799655914307, "learning_rate": 9.85605781873172e-07, "loss": 0.1593, "step": 12520 }, { "epoch": 0.24568627450980393, "grad_norm": 3.7355921268463135, "learning_rate": 9.85567139954818e-07, "loss": 0.1822, "step": 12530 }, { "epoch": 0.24588235294117647, "grad_norm": 2.140399932861328, "learning_rate": 9.85528446997544e-07, "loss": 0.1729, "step": 12540 }, { "epoch": 0.246078431372549, "grad_norm": 1.051795482635498, "learning_rate": 9.85489703005417e-07, "loss": 0.1703, "step": 12550 }, { "epoch": 0.24627450980392157, "grad_norm": 3.310673952102661, "learning_rate": 9.854509079825093e-07, "loss": 0.1911, "step": 12560 }, { "epoch": 0.2464705882352941, "grad_norm": 2.819679021835327, "learning_rate": 9.854120619328994e-07, "loss": 0.1698, "step": 12570 }, { "epoch": 0.24666666666666667, "grad_norm": 2.379108190536499, "learning_rate": 9.853731648606695e-07, "loss": 0.1703, "step": 12580 }, { "epoch": 0.2468627450980392, "grad_norm": 4.100790023803711, "learning_rate": 9.85334216769909e-07, "loss": 0.1952, "step": 12590 }, { "epoch": 0.24705882352941178, "grad_norm": 1.4799448251724243, "learning_rate": 9.852952176647115e-07, "loss": 0.1696, "step": 12600 }, { "epoch": 0.2472549019607843, "grad_norm": 2.024698495864868, "learning_rate": 9.852561675491762e-07, "loss": 0.1815, "step": 12610 }, { "epoch": 0.24745098039215685, "grad_norm": 2.7906017303466797, "learning_rate": 9.852170664274077e-07, "loss": 0.1628, "step": 12620 }, { "epoch": 0.24764705882352941, "grad_norm": 2.938981771469116, "learning_rate": 9.85177914303516e-07, "loss": 0.1827, "step": 12630 }, { "epoch": 0.24784313725490195, "grad_norm": 3.034855604171753, "learning_rate": 9.851387111816168e-07, "loss": 0.1865, "step": 12640 }, { "epoch": 0.24803921568627452, "grad_norm": 2.040660858154297, "learning_rate": 9.850994570658305e-07, "loss": 0.1721, "step": 12650 }, { "epoch": 0.24823529411764705, "grad_norm": 5.801543235778809, "learning_rate": 9.850601519602832e-07, "loss": 0.1903, "step": 12660 }, { "epoch": 0.24843137254901962, "grad_norm": 2.728334903717041, "learning_rate": 9.850207958691065e-07, "loss": 0.1557, "step": 12670 }, { "epoch": 0.24862745098039216, "grad_norm": 9.42239761352539, "learning_rate": 9.849813887964366e-07, "loss": 0.1973, "step": 12680 }, { "epoch": 0.2488235294117647, "grad_norm": 1.9510993957519531, "learning_rate": 9.849419307464166e-07, "loss": 0.1632, "step": 12690 }, { "epoch": 0.24901960784313726, "grad_norm": 2.6374051570892334, "learning_rate": 9.849024217231934e-07, "loss": 0.1776, "step": 12700 }, { "epoch": 0.2492156862745098, "grad_norm": 1.517199158668518, "learning_rate": 9.8486286173092e-07, "loss": 0.154, "step": 12710 }, { "epoch": 0.24941176470588236, "grad_norm": 2.8284287452697754, "learning_rate": 9.848232507737546e-07, "loss": 0.1988, "step": 12720 }, { "epoch": 0.2496078431372549, "grad_norm": 3.2362380027770996, "learning_rate": 9.847835888558608e-07, "loss": 0.151, "step": 12730 }, { "epoch": 0.24980392156862746, "grad_norm": 1.3082737922668457, "learning_rate": 9.847438759814076e-07, "loss": 0.1793, "step": 12740 }, { "epoch": 0.25, "grad_norm": 2.7534148693084717, "learning_rate": 9.847041121545695e-07, "loss": 0.1656, "step": 12750 }, { "epoch": 0.25019607843137254, "grad_norm": 4.072992324829102, "learning_rate": 9.846642973795254e-07, "loss": 0.1967, "step": 12760 }, { "epoch": 0.2503921568627451, "grad_norm": 3.799250602722168, "learning_rate": 9.846244316604613e-07, "loss": 0.2116, "step": 12770 }, { "epoch": 0.25058823529411767, "grad_norm": 19.661481857299805, "learning_rate": 9.845845150015667e-07, "loss": 0.1509, "step": 12780 }, { "epoch": 0.2507843137254902, "grad_norm": 3.6094212532043457, "learning_rate": 9.84544547407038e-07, "loss": 0.1908, "step": 12790 }, { "epoch": 0.25098039215686274, "grad_norm": 3.040278196334839, "learning_rate": 9.84504528881076e-07, "loss": 0.1993, "step": 12800 }, { "epoch": 0.2511764705882353, "grad_norm": 3.0486888885498047, "learning_rate": 9.84464459427887e-07, "loss": 0.1609, "step": 12810 }, { "epoch": 0.2513725490196078, "grad_norm": 4.762712001800537, "learning_rate": 9.84424339051683e-07, "loss": 0.1783, "step": 12820 }, { "epoch": 0.2515686274509804, "grad_norm": 1.4785561561584473, "learning_rate": 9.84384167756681e-07, "loss": 0.1462, "step": 12830 }, { "epoch": 0.25176470588235295, "grad_norm": 2.0533952713012695, "learning_rate": 9.843439455471033e-07, "loss": 0.1428, "step": 12840 }, { "epoch": 0.2519607843137255, "grad_norm": 2.6006414890289307, "learning_rate": 9.843036724271782e-07, "loss": 0.1824, "step": 12850 }, { "epoch": 0.252156862745098, "grad_norm": 5.703633785247803, "learning_rate": 9.842633484011384e-07, "loss": 0.2139, "step": 12860 }, { "epoch": 0.2523529411764706, "grad_norm": 2.6953938007354736, "learning_rate": 9.842229734732227e-07, "loss": 0.1859, "step": 12870 }, { "epoch": 0.25254901960784315, "grad_norm": 3.1226556301116943, "learning_rate": 9.841825476476752e-07, "loss": 0.1791, "step": 12880 }, { "epoch": 0.2527450980392157, "grad_norm": 5.607292175292969, "learning_rate": 9.841420709287447e-07, "loss": 0.1563, "step": 12890 }, { "epoch": 0.2529411764705882, "grad_norm": 2.316096544265747, "learning_rate": 9.84101543320686e-07, "loss": 0.1445, "step": 12900 }, { "epoch": 0.25313725490196076, "grad_norm": 7.302420139312744, "learning_rate": 9.840609648277592e-07, "loss": 0.1929, "step": 12910 }, { "epoch": 0.25333333333333335, "grad_norm": 3.1818723678588867, "learning_rate": 9.84020335454229e-07, "loss": 0.18, "step": 12920 }, { "epoch": 0.2535294117647059, "grad_norm": 4.5286173820495605, "learning_rate": 9.839796552043668e-07, "loss": 0.1773, "step": 12930 }, { "epoch": 0.25372549019607843, "grad_norm": 2.348741292953491, "learning_rate": 9.83938924082448e-07, "loss": 0.1364, "step": 12940 }, { "epoch": 0.25392156862745097, "grad_norm": 5.979569435119629, "learning_rate": 9.838981420927544e-07, "loss": 0.1567, "step": 12950 }, { "epoch": 0.2541176470588235, "grad_norm": 3.8631153106689453, "learning_rate": 9.838573092395723e-07, "loss": 0.1863, "step": 12960 }, { "epoch": 0.2543137254901961, "grad_norm": 1.7644429206848145, "learning_rate": 9.838164255271937e-07, "loss": 0.2083, "step": 12970 }, { "epoch": 0.25450980392156863, "grad_norm": 1.637229084968567, "learning_rate": 9.837754909599162e-07, "loss": 0.1361, "step": 12980 }, { "epoch": 0.25470588235294117, "grad_norm": 2.5735373497009277, "learning_rate": 9.837345055420426e-07, "loss": 0.1851, "step": 12990 }, { "epoch": 0.2549019607843137, "grad_norm": 2.2110304832458496, "learning_rate": 9.836934692778805e-07, "loss": 0.149, "step": 13000 }, { "epoch": 0.2550980392156863, "grad_norm": 1.674881100654602, "learning_rate": 9.836523821717438e-07, "loss": 0.1337, "step": 13010 }, { "epoch": 0.25529411764705884, "grad_norm": 1.942336082458496, "learning_rate": 9.836112442279509e-07, "loss": 0.1926, "step": 13020 }, { "epoch": 0.2554901960784314, "grad_norm": 2.234812021255493, "learning_rate": 9.83570055450826e-07, "loss": 0.1478, "step": 13030 }, { "epoch": 0.2556862745098039, "grad_norm": 5.039580345153809, "learning_rate": 9.835288158446985e-07, "loss": 0.1756, "step": 13040 }, { "epoch": 0.25588235294117645, "grad_norm": 5.8890767097473145, "learning_rate": 9.834875254139034e-07, "loss": 0.2107, "step": 13050 }, { "epoch": 0.25607843137254904, "grad_norm": 3.274477005004883, "learning_rate": 9.83446184162781e-07, "loss": 0.1722, "step": 13060 }, { "epoch": 0.2562745098039216, "grad_norm": 1.8198460340499878, "learning_rate": 9.834047920956758e-07, "loss": 0.1591, "step": 13070 }, { "epoch": 0.2564705882352941, "grad_norm": 4.063549518585205, "learning_rate": 9.833633492169392e-07, "loss": 0.1774, "step": 13080 }, { "epoch": 0.25666666666666665, "grad_norm": 2.6599574089050293, "learning_rate": 9.833218555309278e-07, "loss": 0.1484, "step": 13090 }, { "epoch": 0.2568627450980392, "grad_norm": 3.8232555389404297, "learning_rate": 9.832803110420025e-07, "loss": 0.1971, "step": 13100 }, { "epoch": 0.2570588235294118, "grad_norm": 2.2347021102905273, "learning_rate": 9.8323871575453e-07, "loss": 0.1635, "step": 13110 }, { "epoch": 0.2572549019607843, "grad_norm": 2.0123448371887207, "learning_rate": 9.83197069672883e-07, "loss": 0.1577, "step": 13120 }, { "epoch": 0.25745098039215686, "grad_norm": 2.0358381271362305, "learning_rate": 9.831553728014387e-07, "loss": 0.1898, "step": 13130 }, { "epoch": 0.2576470588235294, "grad_norm": 2.3452372550964355, "learning_rate": 9.8311362514458e-07, "loss": 0.1725, "step": 13140 }, { "epoch": 0.257843137254902, "grad_norm": 2.257474184036255, "learning_rate": 9.830718267066949e-07, "loss": 0.1667, "step": 13150 }, { "epoch": 0.2580392156862745, "grad_norm": 2.6325128078460693, "learning_rate": 9.830299774921771e-07, "loss": 0.2059, "step": 13160 }, { "epoch": 0.25823529411764706, "grad_norm": 3.210723400115967, "learning_rate": 9.829880775054254e-07, "loss": 0.2291, "step": 13170 }, { "epoch": 0.2584313725490196, "grad_norm": 4.390870094299316, "learning_rate": 9.82946126750844e-07, "loss": 0.1737, "step": 13180 }, { "epoch": 0.25862745098039214, "grad_norm": 2.880934715270996, "learning_rate": 9.829041252328426e-07, "loss": 0.1945, "step": 13190 }, { "epoch": 0.25882352941176473, "grad_norm": 2.579094171524048, "learning_rate": 9.828620729558356e-07, "loss": 0.1705, "step": 13200 }, { "epoch": 0.25901960784313727, "grad_norm": 3.576809883117676, "learning_rate": 9.828199699242439e-07, "loss": 0.1571, "step": 13210 }, { "epoch": 0.2592156862745098, "grad_norm": 2.722618341445923, "learning_rate": 9.827778161424923e-07, "loss": 0.1423, "step": 13220 }, { "epoch": 0.25941176470588234, "grad_norm": 4.90212869644165, "learning_rate": 9.82735611615012e-07, "loss": 0.1825, "step": 13230 }, { "epoch": 0.2596078431372549, "grad_norm": 2.3107476234436035, "learning_rate": 9.826933563462395e-07, "loss": 0.1831, "step": 13240 }, { "epoch": 0.25980392156862747, "grad_norm": 2.3922533988952637, "learning_rate": 9.826510503406157e-07, "loss": 0.1789, "step": 13250 }, { "epoch": 0.26, "grad_norm": 3.018284559249878, "learning_rate": 9.82608693602588e-07, "loss": 0.1639, "step": 13260 }, { "epoch": 0.26019607843137255, "grad_norm": 13.464309692382812, "learning_rate": 9.825662861366082e-07, "loss": 0.1853, "step": 13270 }, { "epoch": 0.2603921568627451, "grad_norm": 2.038442850112915, "learning_rate": 9.82523827947134e-07, "loss": 0.1709, "step": 13280 }, { "epoch": 0.2605882352941176, "grad_norm": 2.170511245727539, "learning_rate": 9.824813190386285e-07, "loss": 0.1858, "step": 13290 }, { "epoch": 0.2607843137254902, "grad_norm": 1.8209043741226196, "learning_rate": 9.824387594155597e-07, "loss": 0.1738, "step": 13300 }, { "epoch": 0.26098039215686275, "grad_norm": 2.133368968963623, "learning_rate": 9.82396149082401e-07, "loss": 0.1628, "step": 13310 }, { "epoch": 0.2611764705882353, "grad_norm": 2.3813486099243164, "learning_rate": 9.823534880436315e-07, "loss": 0.1509, "step": 13320 }, { "epoch": 0.2613725490196078, "grad_norm": 2.7740695476531982, "learning_rate": 9.823107763037352e-07, "loss": 0.1942, "step": 13330 }, { "epoch": 0.2615686274509804, "grad_norm": 3.065174102783203, "learning_rate": 9.822680138672017e-07, "loss": 0.1748, "step": 13340 }, { "epoch": 0.26176470588235295, "grad_norm": 4.218511581420898, "learning_rate": 9.822252007385257e-07, "loss": 0.1556, "step": 13350 }, { "epoch": 0.2619607843137255, "grad_norm": 2.448058605194092, "learning_rate": 9.821823369222075e-07, "loss": 0.1812, "step": 13360 }, { "epoch": 0.26215686274509803, "grad_norm": 3.7697222232818604, "learning_rate": 9.821394224227526e-07, "loss": 0.2016, "step": 13370 }, { "epoch": 0.26235294117647057, "grad_norm": 2.5719804763793945, "learning_rate": 9.820964572446719e-07, "loss": 0.1492, "step": 13380 }, { "epoch": 0.26254901960784316, "grad_norm": 3.0224807262420654, "learning_rate": 9.820534413924813e-07, "loss": 0.1534, "step": 13390 }, { "epoch": 0.2627450980392157, "grad_norm": 2.6966440677642822, "learning_rate": 9.820103748707027e-07, "loss": 0.1642, "step": 13400 }, { "epoch": 0.26294117647058823, "grad_norm": 4.422606945037842, "learning_rate": 9.819672576838624e-07, "loss": 0.1972, "step": 13410 }, { "epoch": 0.26313725490196077, "grad_norm": 2.7314746379852295, "learning_rate": 9.81924089836493e-07, "loss": 0.147, "step": 13420 }, { "epoch": 0.2633333333333333, "grad_norm": 2.0766518115997314, "learning_rate": 9.818808713331317e-07, "loss": 0.1647, "step": 13430 }, { "epoch": 0.2635294117647059, "grad_norm": 1.8193608522415161, "learning_rate": 9.818376021783213e-07, "loss": 0.15, "step": 13440 }, { "epoch": 0.26372549019607844, "grad_norm": 3.5558059215545654, "learning_rate": 9.8179428237661e-07, "loss": 0.1831, "step": 13450 }, { "epoch": 0.263921568627451, "grad_norm": 2.1404106616973877, "learning_rate": 9.81750911932551e-07, "loss": 0.1692, "step": 13460 }, { "epoch": 0.2641176470588235, "grad_norm": 3.4041988849639893, "learning_rate": 9.817074908507033e-07, "loss": 0.1566, "step": 13470 }, { "epoch": 0.2643137254901961, "grad_norm": 2.4117038249969482, "learning_rate": 9.81664019135631e-07, "loss": 0.2073, "step": 13480 }, { "epoch": 0.26450980392156864, "grad_norm": 3.3128085136413574, "learning_rate": 9.816204967919032e-07, "loss": 0.1798, "step": 13490 }, { "epoch": 0.2647058823529412, "grad_norm": 5.374880790710449, "learning_rate": 9.815769238240952e-07, "loss": 0.155, "step": 13500 }, { "epoch": 0.2649019607843137, "grad_norm": 2.583709955215454, "learning_rate": 9.815333002367863e-07, "loss": 0.1901, "step": 13510 }, { "epoch": 0.26509803921568625, "grad_norm": 4.531609058380127, "learning_rate": 9.814896260345622e-07, "loss": 0.1987, "step": 13520 }, { "epoch": 0.26529411764705885, "grad_norm": 2.977025032043457, "learning_rate": 9.814459012220137e-07, "loss": 0.15, "step": 13530 }, { "epoch": 0.2654901960784314, "grad_norm": 3.0425844192504883, "learning_rate": 9.814021258037369e-07, "loss": 0.1755, "step": 13540 }, { "epoch": 0.2656862745098039, "grad_norm": 2.097895860671997, "learning_rate": 9.813582997843327e-07, "loss": 0.1405, "step": 13550 }, { "epoch": 0.26588235294117646, "grad_norm": 2.427708148956299, "learning_rate": 9.813144231684079e-07, "loss": 0.2185, "step": 13560 }, { "epoch": 0.266078431372549, "grad_norm": 6.079519271850586, "learning_rate": 9.812704959605746e-07, "loss": 0.1981, "step": 13570 }, { "epoch": 0.2662745098039216, "grad_norm": 1.84262216091156, "learning_rate": 9.8122651816545e-07, "loss": 0.1601, "step": 13580 }, { "epoch": 0.2664705882352941, "grad_norm": 5.3261494636535645, "learning_rate": 9.811824897876567e-07, "loss": 0.2089, "step": 13590 }, { "epoch": 0.26666666666666666, "grad_norm": 3.0309367179870605, "learning_rate": 9.811384108318226e-07, "loss": 0.201, "step": 13600 }, { "epoch": 0.2668627450980392, "grad_norm": 3.72986102104187, "learning_rate": 9.810942813025808e-07, "loss": 0.1615, "step": 13610 }, { "epoch": 0.26705882352941174, "grad_norm": 2.9481489658355713, "learning_rate": 9.8105010120457e-07, "loss": 0.1898, "step": 13620 }, { "epoch": 0.26725490196078433, "grad_norm": 2.240187883377075, "learning_rate": 9.81005870542434e-07, "loss": 0.1429, "step": 13630 }, { "epoch": 0.26745098039215687, "grad_norm": 3.511165142059326, "learning_rate": 9.80961589320822e-07, "loss": 0.1793, "step": 13640 }, { "epoch": 0.2676470588235294, "grad_norm": 3.131261110305786, "learning_rate": 9.809172575443887e-07, "loss": 0.1819, "step": 13650 }, { "epoch": 0.26784313725490194, "grad_norm": 2.363475799560547, "learning_rate": 9.808728752177935e-07, "loss": 0.1663, "step": 13660 }, { "epoch": 0.26803921568627453, "grad_norm": 4.721470355987549, "learning_rate": 9.808284423457017e-07, "loss": 0.1886, "step": 13670 }, { "epoch": 0.26823529411764707, "grad_norm": 2.6897647380828857, "learning_rate": 9.807839589327838e-07, "loss": 0.1647, "step": 13680 }, { "epoch": 0.2684313725490196, "grad_norm": 3.6896252632141113, "learning_rate": 9.807394249837153e-07, "loss": 0.1833, "step": 13690 }, { "epoch": 0.26862745098039215, "grad_norm": 3.3750736713409424, "learning_rate": 9.806948405031775e-07, "loss": 0.1807, "step": 13700 }, { "epoch": 0.2688235294117647, "grad_norm": 3.763897180557251, "learning_rate": 9.806502054958566e-07, "loss": 0.1975, "step": 13710 }, { "epoch": 0.2690196078431373, "grad_norm": 3.8310437202453613, "learning_rate": 9.806055199664445e-07, "loss": 0.1709, "step": 13720 }, { "epoch": 0.2692156862745098, "grad_norm": 1.315673589706421, "learning_rate": 9.80560783919638e-07, "loss": 0.1358, "step": 13730 }, { "epoch": 0.26941176470588235, "grad_norm": 2.323986291885376, "learning_rate": 9.805159973601392e-07, "loss": 0.1462, "step": 13740 }, { "epoch": 0.2696078431372549, "grad_norm": 4.321135520935059, "learning_rate": 9.80471160292656e-07, "loss": 0.1486, "step": 13750 }, { "epoch": 0.2698039215686274, "grad_norm": 1.5979951620101929, "learning_rate": 9.804262727219014e-07, "loss": 0.163, "step": 13760 }, { "epoch": 0.27, "grad_norm": 3.835998058319092, "learning_rate": 9.803813346525936e-07, "loss": 0.1977, "step": 13770 }, { "epoch": 0.27019607843137255, "grad_norm": 1.5440388917922974, "learning_rate": 9.803363460894557e-07, "loss": 0.2152, "step": 13780 }, { "epoch": 0.2703921568627451, "grad_norm": 2.536715507507324, "learning_rate": 9.802913070372168e-07, "loss": 0.1902, "step": 13790 }, { "epoch": 0.27058823529411763, "grad_norm": 2.508431911468506, "learning_rate": 9.802462175006113e-07, "loss": 0.1926, "step": 13800 }, { "epoch": 0.2707843137254902, "grad_norm": 3.4774527549743652, "learning_rate": 9.802010774843783e-07, "loss": 0.1639, "step": 13810 }, { "epoch": 0.27098039215686276, "grad_norm": 5.849239349365234, "learning_rate": 9.801558869932626e-07, "loss": 0.1784, "step": 13820 }, { "epoch": 0.2711764705882353, "grad_norm": 2.6271779537200928, "learning_rate": 9.801106460320145e-07, "loss": 0.2194, "step": 13830 }, { "epoch": 0.27137254901960783, "grad_norm": 3.6946256160736084, "learning_rate": 9.800653546053891e-07, "loss": 0.2068, "step": 13840 }, { "epoch": 0.27156862745098037, "grad_norm": 4.334146499633789, "learning_rate": 9.800200127181473e-07, "loss": 0.176, "step": 13850 }, { "epoch": 0.27176470588235296, "grad_norm": 2.9751474857330322, "learning_rate": 9.799746203750547e-07, "loss": 0.1794, "step": 13860 }, { "epoch": 0.2719607843137255, "grad_norm": 2.1796085834503174, "learning_rate": 9.79929177580883e-07, "loss": 0.1722, "step": 13870 }, { "epoch": 0.27215686274509804, "grad_norm": 5.238569259643555, "learning_rate": 9.798836843404086e-07, "loss": 0.1859, "step": 13880 }, { "epoch": 0.2723529411764706, "grad_norm": 1.2736085653305054, "learning_rate": 9.798381406584133e-07, "loss": 0.1555, "step": 13890 }, { "epoch": 0.2725490196078431, "grad_norm": 2.350111722946167, "learning_rate": 9.797925465396845e-07, "loss": 0.1953, "step": 13900 }, { "epoch": 0.2727450980392157, "grad_norm": 2.603228807449341, "learning_rate": 9.797469019890146e-07, "loss": 0.1968, "step": 13910 }, { "epoch": 0.27294117647058824, "grad_norm": 4.948174953460693, "learning_rate": 9.797012070112013e-07, "loss": 0.1628, "step": 13920 }, { "epoch": 0.2731372549019608, "grad_norm": 2.4516797065734863, "learning_rate": 9.796554616110477e-07, "loss": 0.1912, "step": 13930 }, { "epoch": 0.2733333333333333, "grad_norm": 3.7827625274658203, "learning_rate": 9.796096657933623e-07, "loss": 0.1658, "step": 13940 }, { "epoch": 0.2735294117647059, "grad_norm": 3.1509745121002197, "learning_rate": 9.795638195629587e-07, "loss": 0.1781, "step": 13950 }, { "epoch": 0.27372549019607845, "grad_norm": 2.885683059692383, "learning_rate": 9.79517922924656e-07, "loss": 0.1352, "step": 13960 }, { "epoch": 0.273921568627451, "grad_norm": 2.5326199531555176, "learning_rate": 9.794719758832783e-07, "loss": 0.186, "step": 13970 }, { "epoch": 0.2741176470588235, "grad_norm": 3.0795249938964844, "learning_rate": 9.794259784436554e-07, "loss": 0.1676, "step": 13980 }, { "epoch": 0.27431372549019606, "grad_norm": 3.196307897567749, "learning_rate": 9.79379930610622e-07, "loss": 0.1854, "step": 13990 }, { "epoch": 0.27450980392156865, "grad_norm": 1.4346392154693604, "learning_rate": 9.793338323890183e-07, "loss": 0.2039, "step": 14000 }, { "epoch": 0.2747058823529412, "grad_norm": 2.6618452072143555, "learning_rate": 9.7928768378369e-07, "loss": 0.1698, "step": 14010 }, { "epoch": 0.2749019607843137, "grad_norm": 2.2464311122894287, "learning_rate": 9.792414847994875e-07, "loss": 0.1838, "step": 14020 }, { "epoch": 0.27509803921568626, "grad_norm": 2.3884339332580566, "learning_rate": 9.79195235441267e-07, "loss": 0.1951, "step": 14030 }, { "epoch": 0.2752941176470588, "grad_norm": 5.9033894538879395, "learning_rate": 9.7914893571389e-07, "loss": 0.181, "step": 14040 }, { "epoch": 0.2754901960784314, "grad_norm": 3.0442514419555664, "learning_rate": 9.791025856222232e-07, "loss": 0.158, "step": 14050 }, { "epoch": 0.27568627450980393, "grad_norm": 1.9047820568084717, "learning_rate": 9.790561851711382e-07, "loss": 0.1483, "step": 14060 }, { "epoch": 0.27588235294117647, "grad_norm": 3.862842321395874, "learning_rate": 9.790097343655125e-07, "loss": 0.1821, "step": 14070 }, { "epoch": 0.276078431372549, "grad_norm": 2.409548759460449, "learning_rate": 9.78963233210229e-07, "loss": 0.1604, "step": 14080 }, { "epoch": 0.27627450980392154, "grad_norm": 3.0714926719665527, "learning_rate": 9.789166817101746e-07, "loss": 0.1715, "step": 14090 }, { "epoch": 0.27647058823529413, "grad_norm": 2.5201330184936523, "learning_rate": 9.788700798702432e-07, "loss": 0.1759, "step": 14100 }, { "epoch": 0.27666666666666667, "grad_norm": 2.6174938678741455, "learning_rate": 9.78823427695333e-07, "loss": 0.1423, "step": 14110 }, { "epoch": 0.2768627450980392, "grad_norm": 3.839832305908203, "learning_rate": 9.787767251903476e-07, "loss": 0.178, "step": 14120 }, { "epoch": 0.27705882352941175, "grad_norm": 4.7401957511901855, "learning_rate": 9.78729972360196e-07, "loss": 0.1587, "step": 14130 }, { "epoch": 0.27725490196078434, "grad_norm": 2.4868903160095215, "learning_rate": 9.786831692097926e-07, "loss": 0.1537, "step": 14140 }, { "epoch": 0.2774509803921569, "grad_norm": 2.1199073791503906, "learning_rate": 9.78636315744057e-07, "loss": 0.1844, "step": 14150 }, { "epoch": 0.2776470588235294, "grad_norm": 3.2760212421417236, "learning_rate": 9.785894119679137e-07, "loss": 0.163, "step": 14160 }, { "epoch": 0.27784313725490195, "grad_norm": 2.6633710861206055, "learning_rate": 9.785424578862933e-07, "loss": 0.1973, "step": 14170 }, { "epoch": 0.2780392156862745, "grad_norm": 2.562700033187866, "learning_rate": 9.78495453504131e-07, "loss": 0.1588, "step": 14180 }, { "epoch": 0.2782352941176471, "grad_norm": 2.000216245651245, "learning_rate": 9.784483988263677e-07, "loss": 0.1957, "step": 14190 }, { "epoch": 0.2784313725490196, "grad_norm": 3.051067590713501, "learning_rate": 9.784012938579492e-07, "loss": 0.1722, "step": 14200 }, { "epoch": 0.27862745098039216, "grad_norm": 3.8240299224853516, "learning_rate": 9.78354138603827e-07, "loss": 0.1679, "step": 14210 }, { "epoch": 0.2788235294117647, "grad_norm": 2.6512668132781982, "learning_rate": 9.783069330689575e-07, "loss": 0.1742, "step": 14220 }, { "epoch": 0.27901960784313723, "grad_norm": 3.761075735092163, "learning_rate": 9.782596772583025e-07, "loss": 0.2012, "step": 14230 }, { "epoch": 0.2792156862745098, "grad_norm": 1.968023657798767, "learning_rate": 9.782123711768295e-07, "loss": 0.1601, "step": 14240 }, { "epoch": 0.27941176470588236, "grad_norm": 2.5836727619171143, "learning_rate": 9.781650148295105e-07, "loss": 0.2065, "step": 14250 }, { "epoch": 0.2796078431372549, "grad_norm": 3.106264114379883, "learning_rate": 9.781176082213237e-07, "loss": 0.207, "step": 14260 }, { "epoch": 0.27980392156862743, "grad_norm": 4.7932353019714355, "learning_rate": 9.780701513572516e-07, "loss": 0.1597, "step": 14270 }, { "epoch": 0.28, "grad_norm": 2.6639297008514404, "learning_rate": 9.780226442422827e-07, "loss": 0.1669, "step": 14280 }, { "epoch": 0.28019607843137256, "grad_norm": 2.1701149940490723, "learning_rate": 9.779750868814107e-07, "loss": 0.2103, "step": 14290 }, { "epoch": 0.2803921568627451, "grad_norm": 3.9998080730438232, "learning_rate": 9.779274792796344e-07, "loss": 0.1833, "step": 14300 }, { "epoch": 0.28058823529411764, "grad_norm": 2.734114170074463, "learning_rate": 9.778798214419578e-07, "loss": 0.1719, "step": 14310 }, { "epoch": 0.2807843137254902, "grad_norm": 1.6632474660873413, "learning_rate": 9.778321133733904e-07, "loss": 0.1398, "step": 14320 }, { "epoch": 0.28098039215686277, "grad_norm": 2.948023557662964, "learning_rate": 9.77784355078947e-07, "loss": 0.1673, "step": 14330 }, { "epoch": 0.2811764705882353, "grad_norm": 3.545137405395508, "learning_rate": 9.777365465636473e-07, "loss": 0.1565, "step": 14340 }, { "epoch": 0.28137254901960784, "grad_norm": 5.0868988037109375, "learning_rate": 9.776886878325166e-07, "loss": 0.1806, "step": 14350 }, { "epoch": 0.2815686274509804, "grad_norm": 2.6549882888793945, "learning_rate": 9.776407788905856e-07, "loss": 0.1639, "step": 14360 }, { "epoch": 0.2817647058823529, "grad_norm": 2.224914789199829, "learning_rate": 9.775928197428899e-07, "loss": 0.1726, "step": 14370 }, { "epoch": 0.2819607843137255, "grad_norm": 3.2208549976348877, "learning_rate": 9.77544810394471e-07, "loss": 0.1732, "step": 14380 }, { "epoch": 0.28215686274509805, "grad_norm": 2.4533393383026123, "learning_rate": 9.774967508503747e-07, "loss": 0.193, "step": 14390 }, { "epoch": 0.2823529411764706, "grad_norm": 4.112372875213623, "learning_rate": 9.77448641115653e-07, "loss": 0.2099, "step": 14400 }, { "epoch": 0.2825490196078431, "grad_norm": 3.535951614379883, "learning_rate": 9.774004811953628e-07, "loss": 0.1639, "step": 14410 }, { "epoch": 0.28274509803921566, "grad_norm": 3.981532335281372, "learning_rate": 9.773522710945659e-07, "loss": 0.1673, "step": 14420 }, { "epoch": 0.28294117647058825, "grad_norm": 2.16951847076416, "learning_rate": 9.7730401081833e-07, "loss": 0.1517, "step": 14430 }, { "epoch": 0.2831372549019608, "grad_norm": 2.328422784805298, "learning_rate": 9.772557003717282e-07, "loss": 0.1451, "step": 14440 }, { "epoch": 0.2833333333333333, "grad_norm": 2.3962557315826416, "learning_rate": 9.772073397598379e-07, "loss": 0.1674, "step": 14450 }, { "epoch": 0.28352941176470586, "grad_norm": 4.888566493988037, "learning_rate": 9.771589289877427e-07, "loss": 0.1638, "step": 14460 }, { "epoch": 0.28372549019607846, "grad_norm": 3.5664377212524414, "learning_rate": 9.771104680605313e-07, "loss": 0.1818, "step": 14470 }, { "epoch": 0.283921568627451, "grad_norm": 2.7267706394195557, "learning_rate": 9.77061956983297e-07, "loss": 0.1888, "step": 14480 }, { "epoch": 0.28411764705882353, "grad_norm": 3.0010297298431396, "learning_rate": 9.770133957611395e-07, "loss": 0.1537, "step": 14490 }, { "epoch": 0.28431372549019607, "grad_norm": 3.047715187072754, "learning_rate": 9.769647843991629e-07, "loss": 0.1703, "step": 14500 }, { "epoch": 0.2845098039215686, "grad_norm": 2.0460660457611084, "learning_rate": 9.769161229024767e-07, "loss": 0.1663, "step": 14510 }, { "epoch": 0.2847058823529412, "grad_norm": 3.033848762512207, "learning_rate": 9.76867411276196e-07, "loss": 0.185, "step": 14520 }, { "epoch": 0.28490196078431373, "grad_norm": 2.5242490768432617, "learning_rate": 9.76818649525441e-07, "loss": 0.1638, "step": 14530 }, { "epoch": 0.28509803921568627, "grad_norm": 2.8229427337646484, "learning_rate": 9.767698376553368e-07, "loss": 0.1596, "step": 14540 }, { "epoch": 0.2852941176470588, "grad_norm": 4.206874370574951, "learning_rate": 9.767209756710145e-07, "loss": 0.1725, "step": 14550 }, { "epoch": 0.28549019607843135, "grad_norm": 3.1815261840820312, "learning_rate": 9.7667206357761e-07, "loss": 0.1921, "step": 14560 }, { "epoch": 0.28568627450980394, "grad_norm": 3.8132472038269043, "learning_rate": 9.766231013802644e-07, "loss": 0.1626, "step": 14570 }, { "epoch": 0.2858823529411765, "grad_norm": 3.1191446781158447, "learning_rate": 9.765740890841244e-07, "loss": 0.1487, "step": 14580 }, { "epoch": 0.286078431372549, "grad_norm": 2.6841585636138916, "learning_rate": 9.765250266943415e-07, "loss": 0.1855, "step": 14590 }, { "epoch": 0.28627450980392155, "grad_norm": 1.3246322870254517, "learning_rate": 9.764759142160732e-07, "loss": 0.1589, "step": 14600 }, { "epoch": 0.28647058823529414, "grad_norm": 3.3477284908294678, "learning_rate": 9.764267516544812e-07, "loss": 0.236, "step": 14610 }, { "epoch": 0.2866666666666667, "grad_norm": 2.412524461746216, "learning_rate": 9.763775390147337e-07, "loss": 0.2092, "step": 14620 }, { "epoch": 0.2868627450980392, "grad_norm": 2.7598698139190674, "learning_rate": 9.763282763020032e-07, "loss": 0.1908, "step": 14630 }, { "epoch": 0.28705882352941176, "grad_norm": 3.28755521774292, "learning_rate": 9.762789635214677e-07, "loss": 0.1581, "step": 14640 }, { "epoch": 0.2872549019607843, "grad_norm": 3.8816146850585938, "learning_rate": 9.762296006783108e-07, "loss": 0.1583, "step": 14650 }, { "epoch": 0.2874509803921569, "grad_norm": 2.521023988723755, "learning_rate": 9.761801877777212e-07, "loss": 0.174, "step": 14660 }, { "epoch": 0.2876470588235294, "grad_norm": 4.2556915283203125, "learning_rate": 9.761307248248925e-07, "loss": 0.1787, "step": 14670 }, { "epoch": 0.28784313725490196, "grad_norm": 3.6596155166625977, "learning_rate": 9.760812118250238e-07, "loss": 0.173, "step": 14680 }, { "epoch": 0.2880392156862745, "grad_norm": 4.218498229980469, "learning_rate": 9.760316487833199e-07, "loss": 0.154, "step": 14690 }, { "epoch": 0.28823529411764703, "grad_norm": 1.6003023386001587, "learning_rate": 9.759820357049902e-07, "loss": 0.1723, "step": 14700 }, { "epoch": 0.2884313725490196, "grad_norm": 1.4351754188537598, "learning_rate": 9.759323725952498e-07, "loss": 0.1748, "step": 14710 }, { "epoch": 0.28862745098039216, "grad_norm": 3.342681884765625, "learning_rate": 9.758826594593186e-07, "loss": 0.1626, "step": 14720 }, { "epoch": 0.2888235294117647, "grad_norm": 2.8295748233795166, "learning_rate": 9.758328963024223e-07, "loss": 0.1974, "step": 14730 }, { "epoch": 0.28901960784313724, "grad_norm": 2.1441454887390137, "learning_rate": 9.757830831297913e-07, "loss": 0.179, "step": 14740 }, { "epoch": 0.28921568627450983, "grad_norm": 3.0074307918548584, "learning_rate": 9.75733219946662e-07, "loss": 0.1741, "step": 14750 }, { "epoch": 0.28941176470588237, "grad_norm": 3.550647735595703, "learning_rate": 9.756833067582754e-07, "loss": 0.1687, "step": 14760 }, { "epoch": 0.2896078431372549, "grad_norm": 1.5227304697036743, "learning_rate": 9.75633343569878e-07, "loss": 0.1694, "step": 14770 }, { "epoch": 0.28980392156862744, "grad_norm": 2.037102699279785, "learning_rate": 9.755833303867213e-07, "loss": 0.1732, "step": 14780 }, { "epoch": 0.29, "grad_norm": 2.2708797454833984, "learning_rate": 9.755332672140627e-07, "loss": 0.1642, "step": 14790 }, { "epoch": 0.2901960784313726, "grad_norm": 3.1441519260406494, "learning_rate": 9.754831540571642e-07, "loss": 0.1943, "step": 14800 }, { "epoch": 0.2903921568627451, "grad_norm": 1.7201482057571411, "learning_rate": 9.754329909212932e-07, "loss": 0.1599, "step": 14810 }, { "epoch": 0.29058823529411765, "grad_norm": 2.345517635345459, "learning_rate": 9.753827778117226e-07, "loss": 0.1767, "step": 14820 }, { "epoch": 0.2907843137254902, "grad_norm": 4.445740222930908, "learning_rate": 9.753325147337304e-07, "loss": 0.1883, "step": 14830 }, { "epoch": 0.2909803921568627, "grad_norm": 2.986908435821533, "learning_rate": 9.752822016925997e-07, "loss": 0.1559, "step": 14840 }, { "epoch": 0.2911764705882353, "grad_norm": 3.782353639602661, "learning_rate": 9.752318386936193e-07, "loss": 0.17, "step": 14850 }, { "epoch": 0.29137254901960785, "grad_norm": 2.494483470916748, "learning_rate": 9.751814257420827e-07, "loss": 0.1838, "step": 14860 }, { "epoch": 0.2915686274509804, "grad_norm": 5.139228343963623, "learning_rate": 9.75130962843289e-07, "loss": 0.1961, "step": 14870 }, { "epoch": 0.2917647058823529, "grad_norm": 3.8150856494903564, "learning_rate": 9.750804500025422e-07, "loss": 0.1667, "step": 14880 }, { "epoch": 0.29196078431372546, "grad_norm": 4.9914631843566895, "learning_rate": 9.750298872251522e-07, "loss": 0.2201, "step": 14890 }, { "epoch": 0.29215686274509806, "grad_norm": 2.5390541553497314, "learning_rate": 9.749792745164338e-07, "loss": 0.1735, "step": 14900 }, { "epoch": 0.2923529411764706, "grad_norm": 2.3658523559570312, "learning_rate": 9.749286118817066e-07, "loss": 0.1583, "step": 14910 }, { "epoch": 0.29254901960784313, "grad_norm": 6.638339519500732, "learning_rate": 9.74877899326296e-07, "loss": 0.1744, "step": 14920 }, { "epoch": 0.29274509803921567, "grad_norm": 2.7779855728149414, "learning_rate": 9.748271368555324e-07, "loss": 0.1415, "step": 14930 }, { "epoch": 0.29294117647058826, "grad_norm": 1.8053475618362427, "learning_rate": 9.74776324474752e-07, "loss": 0.1625, "step": 14940 }, { "epoch": 0.2931372549019608, "grad_norm": 2.541982889175415, "learning_rate": 9.747254621892953e-07, "loss": 0.1662, "step": 14950 }, { "epoch": 0.29333333333333333, "grad_norm": 3.453139066696167, "learning_rate": 9.746745500045087e-07, "loss": 0.1862, "step": 14960 }, { "epoch": 0.29352941176470587, "grad_norm": 2.802656888961792, "learning_rate": 9.746235879257438e-07, "loss": 0.1939, "step": 14970 }, { "epoch": 0.2937254901960784, "grad_norm": 2.0815329551696777, "learning_rate": 9.74572575958357e-07, "loss": 0.1908, "step": 14980 }, { "epoch": 0.293921568627451, "grad_norm": 4.187165260314941, "learning_rate": 9.745215141077105e-07, "loss": 0.1648, "step": 14990 }, { "epoch": 0.29411764705882354, "grad_norm": 1.863895297050476, "learning_rate": 9.744704023791717e-07, "loss": 0.2109, "step": 15000 }, { "epoch": 0.2943137254901961, "grad_norm": 2.325272798538208, "learning_rate": 9.744192407781127e-07, "loss": 0.1779, "step": 15010 }, { "epoch": 0.2945098039215686, "grad_norm": 2.599215030670166, "learning_rate": 9.743680293099115e-07, "loss": 0.1862, "step": 15020 }, { "epoch": 0.29470588235294115, "grad_norm": 2.2575979232788086, "learning_rate": 9.743167679799507e-07, "loss": 0.1571, "step": 15030 }, { "epoch": 0.29490196078431374, "grad_norm": 2.449100971221924, "learning_rate": 9.742654567936186e-07, "loss": 0.1697, "step": 15040 }, { "epoch": 0.2950980392156863, "grad_norm": 2.622889995574951, "learning_rate": 9.742140957563089e-07, "loss": 0.1819, "step": 15050 }, { "epoch": 0.2952941176470588, "grad_norm": 1.8866087198257446, "learning_rate": 9.741626848734198e-07, "loss": 0.1461, "step": 15060 }, { "epoch": 0.29549019607843136, "grad_norm": 2.729205846786499, "learning_rate": 9.741112241503555e-07, "loss": 0.145, "step": 15070 }, { "epoch": 0.29568627450980395, "grad_norm": 3.690246105194092, "learning_rate": 9.740597135925252e-07, "loss": 0.197, "step": 15080 }, { "epoch": 0.2958823529411765, "grad_norm": 1.4176781177520752, "learning_rate": 9.74008153205343e-07, "loss": 0.1668, "step": 15090 }, { "epoch": 0.296078431372549, "grad_norm": 3.785255193710327, "learning_rate": 9.739565429942289e-07, "loss": 0.1585, "step": 15100 }, { "epoch": 0.29627450980392156, "grad_norm": 1.8336912393569946, "learning_rate": 9.739048829646073e-07, "loss": 0.1784, "step": 15110 }, { "epoch": 0.2964705882352941, "grad_norm": 2.4806017875671387, "learning_rate": 9.738531731219086e-07, "loss": 0.1561, "step": 15120 }, { "epoch": 0.2966666666666667, "grad_norm": 3.4405901432037354, "learning_rate": 9.73801413471568e-07, "loss": 0.1974, "step": 15130 }, { "epoch": 0.2968627450980392, "grad_norm": 3.995011568069458, "learning_rate": 9.73749604019026e-07, "loss": 0.1761, "step": 15140 }, { "epoch": 0.29705882352941176, "grad_norm": 3.0959861278533936, "learning_rate": 9.736977447697284e-07, "loss": 0.1698, "step": 15150 }, { "epoch": 0.2972549019607843, "grad_norm": 1.9571332931518555, "learning_rate": 9.736458357291266e-07, "loss": 0.1717, "step": 15160 }, { "epoch": 0.29745098039215684, "grad_norm": 3.27424693107605, "learning_rate": 9.735938769026766e-07, "loss": 0.1879, "step": 15170 }, { "epoch": 0.29764705882352943, "grad_norm": 2.3015553951263428, "learning_rate": 9.735418682958397e-07, "loss": 0.1649, "step": 15180 }, { "epoch": 0.29784313725490197, "grad_norm": 2.00634503364563, "learning_rate": 9.734898099140827e-07, "loss": 0.1593, "step": 15190 }, { "epoch": 0.2980392156862745, "grad_norm": 2.7280964851379395, "learning_rate": 9.734377017628778e-07, "loss": 0.1868, "step": 15200 }, { "epoch": 0.29823529411764704, "grad_norm": 1.566468596458435, "learning_rate": 9.73385543847702e-07, "loss": 0.1891, "step": 15210 }, { "epoch": 0.2984313725490196, "grad_norm": 1.1674988269805908, "learning_rate": 9.733333361740376e-07, "loss": 0.1714, "step": 15220 }, { "epoch": 0.2986274509803922, "grad_norm": 1.4727542400360107, "learning_rate": 9.732810787473728e-07, "loss": 0.1658, "step": 15230 }, { "epoch": 0.2988235294117647, "grad_norm": 2.6211254596710205, "learning_rate": 9.732287715731998e-07, "loss": 0.1922, "step": 15240 }, { "epoch": 0.29901960784313725, "grad_norm": 2.9030988216400146, "learning_rate": 9.731764146570173e-07, "loss": 0.2167, "step": 15250 }, { "epoch": 0.2992156862745098, "grad_norm": 1.8939210176467896, "learning_rate": 9.73124008004328e-07, "loss": 0.1781, "step": 15260 }, { "epoch": 0.2994117647058824, "grad_norm": 2.863764762878418, "learning_rate": 9.730715516206408e-07, "loss": 0.1999, "step": 15270 }, { "epoch": 0.2996078431372549, "grad_norm": 3.7644240856170654, "learning_rate": 9.730190455114698e-07, "loss": 0.1934, "step": 15280 }, { "epoch": 0.29980392156862745, "grad_norm": 2.1877920627593994, "learning_rate": 9.729664896823335e-07, "loss": 0.1614, "step": 15290 }, { "epoch": 0.3, "grad_norm": 2.0247979164123535, "learning_rate": 9.729138841387565e-07, "loss": 0.1497, "step": 15300 }, { "epoch": 0.3001960784313725, "grad_norm": 2.9772884845733643, "learning_rate": 9.72861228886268e-07, "loss": 0.1721, "step": 15310 }, { "epoch": 0.3003921568627451, "grad_norm": 2.390174627304077, "learning_rate": 9.72808523930403e-07, "loss": 0.1661, "step": 15320 }, { "epoch": 0.30058823529411766, "grad_norm": 5.88565731048584, "learning_rate": 9.72755769276701e-07, "loss": 0.1837, "step": 15330 }, { "epoch": 0.3007843137254902, "grad_norm": 2.1644110679626465, "learning_rate": 9.727029649307077e-07, "loss": 0.1475, "step": 15340 }, { "epoch": 0.30098039215686273, "grad_norm": 1.709626317024231, "learning_rate": 9.726501108979732e-07, "loss": 0.1676, "step": 15350 }, { "epoch": 0.30117647058823527, "grad_norm": 3.2600913047790527, "learning_rate": 9.72597207184053e-07, "loss": 0.2194, "step": 15360 }, { "epoch": 0.30137254901960786, "grad_norm": 3.2202887535095215, "learning_rate": 9.72544253794508e-07, "loss": 0.135, "step": 15370 }, { "epoch": 0.3015686274509804, "grad_norm": 3.7573184967041016, "learning_rate": 9.72491250734904e-07, "loss": 0.1799, "step": 15380 }, { "epoch": 0.30176470588235293, "grad_norm": 3.2780556678771973, "learning_rate": 9.724381980108127e-07, "loss": 0.1992, "step": 15390 }, { "epoch": 0.30196078431372547, "grad_norm": 2.1081323623657227, "learning_rate": 9.723850956278103e-07, "loss": 0.184, "step": 15400 }, { "epoch": 0.30215686274509806, "grad_norm": 3.0930240154266357, "learning_rate": 9.723319435914786e-07, "loss": 0.1636, "step": 15410 }, { "epoch": 0.3023529411764706, "grad_norm": 3.3803396224975586, "learning_rate": 9.722787419074044e-07, "loss": 0.1731, "step": 15420 }, { "epoch": 0.30254901960784314, "grad_norm": 2.0954439640045166, "learning_rate": 9.7222549058118e-07, "loss": 0.1715, "step": 15430 }, { "epoch": 0.3027450980392157, "grad_norm": 3.4940543174743652, "learning_rate": 9.721721896184025e-07, "loss": 0.168, "step": 15440 }, { "epoch": 0.3029411764705882, "grad_norm": 4.970141410827637, "learning_rate": 9.721188390246745e-07, "loss": 0.1902, "step": 15450 }, { "epoch": 0.3031372549019608, "grad_norm": 3.199249505996704, "learning_rate": 9.720654388056041e-07, "loss": 0.1998, "step": 15460 }, { "epoch": 0.30333333333333334, "grad_norm": 1.62307608127594, "learning_rate": 9.720119889668043e-07, "loss": 0.1845, "step": 15470 }, { "epoch": 0.3035294117647059, "grad_norm": 2.234714984893799, "learning_rate": 9.719584895138928e-07, "loss": 0.1737, "step": 15480 }, { "epoch": 0.3037254901960784, "grad_norm": 3.424557685852051, "learning_rate": 9.719049404524936e-07, "loss": 0.1718, "step": 15490 }, { "epoch": 0.30392156862745096, "grad_norm": 2.2434005737304688, "learning_rate": 9.71851341788235e-07, "loss": 0.1243, "step": 15500 }, { "epoch": 0.30411764705882355, "grad_norm": 3.8377861976623535, "learning_rate": 9.717976935267509e-07, "loss": 0.1877, "step": 15510 }, { "epoch": 0.3043137254901961, "grad_norm": 1.7140600681304932, "learning_rate": 9.717439956736806e-07, "loss": 0.165, "step": 15520 }, { "epoch": 0.3045098039215686, "grad_norm": 2.6248812675476074, "learning_rate": 9.71690248234668e-07, "loss": 0.177, "step": 15530 }, { "epoch": 0.30470588235294116, "grad_norm": 2.471184015274048, "learning_rate": 9.716364512153631e-07, "loss": 0.1857, "step": 15540 }, { "epoch": 0.30490196078431375, "grad_norm": 3.9616429805755615, "learning_rate": 9.715826046214202e-07, "loss": 0.1545, "step": 15550 }, { "epoch": 0.3050980392156863, "grad_norm": 2.6601078510284424, "learning_rate": 9.715287084584992e-07, "loss": 0.1772, "step": 15560 }, { "epoch": 0.3052941176470588, "grad_norm": 9.998884201049805, "learning_rate": 9.714747627322658e-07, "loss": 0.1882, "step": 15570 }, { "epoch": 0.30549019607843136, "grad_norm": 2.7756142616271973, "learning_rate": 9.714207674483896e-07, "loss": 0.1386, "step": 15580 }, { "epoch": 0.3056862745098039, "grad_norm": 2.22334361076355, "learning_rate": 9.713667226125466e-07, "loss": 0.1904, "step": 15590 }, { "epoch": 0.3058823529411765, "grad_norm": 2.270505428314209, "learning_rate": 9.713126282304175e-07, "loss": 0.1976, "step": 15600 }, { "epoch": 0.30607843137254903, "grad_norm": 2.7368695735931396, "learning_rate": 9.71258484307688e-07, "loss": 0.1715, "step": 15610 }, { "epoch": 0.30627450980392157, "grad_norm": 2.9312403202056885, "learning_rate": 9.712042908500497e-07, "loss": 0.1693, "step": 15620 }, { "epoch": 0.3064705882352941, "grad_norm": 2.7026278972625732, "learning_rate": 9.711500478631989e-07, "loss": 0.1627, "step": 15630 }, { "epoch": 0.30666666666666664, "grad_norm": 2.1616835594177246, "learning_rate": 9.71095755352837e-07, "loss": 0.1889, "step": 15640 }, { "epoch": 0.30686274509803924, "grad_norm": 3.3344740867614746, "learning_rate": 9.710414133246706e-07, "loss": 0.1969, "step": 15650 }, { "epoch": 0.3070588235294118, "grad_norm": 1.528059720993042, "learning_rate": 9.709870217844122e-07, "loss": 0.1555, "step": 15660 }, { "epoch": 0.3072549019607843, "grad_norm": 4.1594719886779785, "learning_rate": 9.709325807377783e-07, "loss": 0.1828, "step": 15670 }, { "epoch": 0.30745098039215685, "grad_norm": 1.736340880393982, "learning_rate": 9.708780901904922e-07, "loss": 0.1762, "step": 15680 }, { "epoch": 0.3076470588235294, "grad_norm": 2.6724908351898193, "learning_rate": 9.70823550148281e-07, "loss": 0.1598, "step": 15690 }, { "epoch": 0.307843137254902, "grad_norm": 1.9092471599578857, "learning_rate": 9.707689606168773e-07, "loss": 0.163, "step": 15700 }, { "epoch": 0.3080392156862745, "grad_norm": 2.763746500015259, "learning_rate": 9.707143216020198e-07, "loss": 0.1667, "step": 15710 }, { "epoch": 0.30823529411764705, "grad_norm": 6.17631196975708, "learning_rate": 9.706596331094511e-07, "loss": 0.1677, "step": 15720 }, { "epoch": 0.3084313725490196, "grad_norm": 1.2946248054504395, "learning_rate": 9.706048951449199e-07, "loss": 0.1644, "step": 15730 }, { "epoch": 0.3086274509803922, "grad_norm": 2.6555378437042236, "learning_rate": 9.705501077141795e-07, "loss": 0.1811, "step": 15740 }, { "epoch": 0.3088235294117647, "grad_norm": 1.9927791357040405, "learning_rate": 9.704952708229891e-07, "loss": 0.1813, "step": 15750 }, { "epoch": 0.30901960784313726, "grad_norm": 1.594557762145996, "learning_rate": 9.704403844771127e-07, "loss": 0.1883, "step": 15760 }, { "epoch": 0.3092156862745098, "grad_norm": 1.7914798259735107, "learning_rate": 9.703854486823193e-07, "loss": 0.1506, "step": 15770 }, { "epoch": 0.30941176470588233, "grad_norm": 2.9308969974517822, "learning_rate": 9.703304634443833e-07, "loss": 0.1721, "step": 15780 }, { "epoch": 0.3096078431372549, "grad_norm": 3.3547794818878174, "learning_rate": 9.702754287690844e-07, "loss": 0.1713, "step": 15790 }, { "epoch": 0.30980392156862746, "grad_norm": 2.842036247253418, "learning_rate": 9.702203446622075e-07, "loss": 0.1793, "step": 15800 }, { "epoch": 0.31, "grad_norm": 2.041027069091797, "learning_rate": 9.701652111295424e-07, "loss": 0.1772, "step": 15810 }, { "epoch": 0.31019607843137254, "grad_norm": 2.0521512031555176, "learning_rate": 9.701100281768844e-07, "loss": 0.1508, "step": 15820 }, { "epoch": 0.31039215686274507, "grad_norm": 1.3364057540893555, "learning_rate": 9.70054795810034e-07, "loss": 0.1519, "step": 15830 }, { "epoch": 0.31058823529411766, "grad_norm": 2.8177082538604736, "learning_rate": 9.699995140347965e-07, "loss": 0.1398, "step": 15840 }, { "epoch": 0.3107843137254902, "grad_norm": 2.4271724224090576, "learning_rate": 9.699441828569828e-07, "loss": 0.1824, "step": 15850 }, { "epoch": 0.31098039215686274, "grad_norm": 1.9436302185058594, "learning_rate": 9.698888022824089e-07, "loss": 0.1578, "step": 15860 }, { "epoch": 0.3111764705882353, "grad_norm": 5.900735378265381, "learning_rate": 9.698333723168961e-07, "loss": 0.1738, "step": 15870 }, { "epoch": 0.31137254901960787, "grad_norm": 2.582244634628296, "learning_rate": 9.697778929662707e-07, "loss": 0.1758, "step": 15880 }, { "epoch": 0.3115686274509804, "grad_norm": 2.6577675342559814, "learning_rate": 9.697223642363641e-07, "loss": 0.2063, "step": 15890 }, { "epoch": 0.31176470588235294, "grad_norm": 3.0808284282684326, "learning_rate": 9.69666786133013e-07, "loss": 0.2026, "step": 15900 }, { "epoch": 0.3119607843137255, "grad_norm": 3.6916871070861816, "learning_rate": 9.696111586620594e-07, "loss": 0.1659, "step": 15910 }, { "epoch": 0.312156862745098, "grad_norm": 1.340400218963623, "learning_rate": 9.695554818293505e-07, "loss": 0.1341, "step": 15920 }, { "epoch": 0.3123529411764706, "grad_norm": 1.2755281925201416, "learning_rate": 9.694997556407386e-07, "loss": 0.1592, "step": 15930 }, { "epoch": 0.31254901960784315, "grad_norm": 3.3621714115142822, "learning_rate": 9.69443980102081e-07, "loss": 0.1406, "step": 15940 }, { "epoch": 0.3127450980392157, "grad_norm": 2.2103500366210938, "learning_rate": 9.693881552192406e-07, "loss": 0.2018, "step": 15950 }, { "epoch": 0.3129411764705882, "grad_norm": 1.6549783945083618, "learning_rate": 9.693322809980854e-07, "loss": 0.1549, "step": 15960 }, { "epoch": 0.31313725490196076, "grad_norm": 5.434072494506836, "learning_rate": 9.692763574444878e-07, "loss": 0.2382, "step": 15970 }, { "epoch": 0.31333333333333335, "grad_norm": 2.8112425804138184, "learning_rate": 9.692203845643266e-07, "loss": 0.1885, "step": 15980 }, { "epoch": 0.3135294117647059, "grad_norm": 1.4537203311920166, "learning_rate": 9.69164362363485e-07, "loss": 0.1787, "step": 15990 }, { "epoch": 0.3137254901960784, "grad_norm": 7.1828179359436035, "learning_rate": 9.691082908478518e-07, "loss": 0.2046, "step": 16000 }, { "epoch": 0.31392156862745096, "grad_norm": 2.9740233421325684, "learning_rate": 9.690521700233208e-07, "loss": 0.1539, "step": 16010 }, { "epoch": 0.31411764705882356, "grad_norm": 2.0929126739501953, "learning_rate": 9.689959998957907e-07, "loss": 0.1894, "step": 16020 }, { "epoch": 0.3143137254901961, "grad_norm": 3.323915719985962, "learning_rate": 9.689397804711657e-07, "loss": 0.1746, "step": 16030 }, { "epoch": 0.31450980392156863, "grad_norm": 4.237054824829102, "learning_rate": 9.688835117553554e-07, "loss": 0.1713, "step": 16040 }, { "epoch": 0.31470588235294117, "grad_norm": 4.558916091918945, "learning_rate": 9.68827193754274e-07, "loss": 0.1822, "step": 16050 }, { "epoch": 0.3149019607843137, "grad_norm": 4.795844078063965, "learning_rate": 9.687708264738412e-07, "loss": 0.1712, "step": 16060 }, { "epoch": 0.3150980392156863, "grad_norm": 2.7933452129364014, "learning_rate": 9.687144099199822e-07, "loss": 0.2028, "step": 16070 }, { "epoch": 0.31529411764705884, "grad_norm": 3.7849769592285156, "learning_rate": 9.686579440986269e-07, "loss": 0.2124, "step": 16080 }, { "epoch": 0.3154901960784314, "grad_norm": 3.6496753692626953, "learning_rate": 9.686014290157105e-07, "loss": 0.1779, "step": 16090 }, { "epoch": 0.3156862745098039, "grad_norm": 4.254093170166016, "learning_rate": 9.685448646771734e-07, "loss": 0.1782, "step": 16100 }, { "epoch": 0.31588235294117645, "grad_norm": 4.081014156341553, "learning_rate": 9.68488251088961e-07, "loss": 0.1686, "step": 16110 }, { "epoch": 0.31607843137254904, "grad_norm": 6.200359344482422, "learning_rate": 9.684315882570245e-07, "loss": 0.1771, "step": 16120 }, { "epoch": 0.3162745098039216, "grad_norm": 3.7284371852874756, "learning_rate": 9.683748761873197e-07, "loss": 0.1625, "step": 16130 }, { "epoch": 0.3164705882352941, "grad_norm": 4.313432693481445, "learning_rate": 9.683181148858073e-07, "loss": 0.1922, "step": 16140 }, { "epoch": 0.31666666666666665, "grad_norm": 1.8996423482894897, "learning_rate": 9.682613043584541e-07, "loss": 0.163, "step": 16150 }, { "epoch": 0.3168627450980392, "grad_norm": 2.0974621772766113, "learning_rate": 9.682044446112314e-07, "loss": 0.1533, "step": 16160 }, { "epoch": 0.3170588235294118, "grad_norm": 1.93149995803833, "learning_rate": 9.681475356501158e-07, "loss": 0.1675, "step": 16170 }, { "epoch": 0.3172549019607843, "grad_norm": 2.607273578643799, "learning_rate": 9.68090577481089e-07, "loss": 0.1751, "step": 16180 }, { "epoch": 0.31745098039215686, "grad_norm": 2.6452157497406006, "learning_rate": 9.680335701101383e-07, "loss": 0.1599, "step": 16190 }, { "epoch": 0.3176470588235294, "grad_norm": 1.635303020477295, "learning_rate": 9.679765135432555e-07, "loss": 0.1542, "step": 16200 }, { "epoch": 0.317843137254902, "grad_norm": 4.2130608558654785, "learning_rate": 9.679194077864384e-07, "loss": 0.1759, "step": 16210 }, { "epoch": 0.3180392156862745, "grad_norm": 3.2958474159240723, "learning_rate": 9.67862252845689e-07, "loss": 0.1789, "step": 16220 }, { "epoch": 0.31823529411764706, "grad_norm": 2.9924979209899902, "learning_rate": 9.678050487270153e-07, "loss": 0.1375, "step": 16230 }, { "epoch": 0.3184313725490196, "grad_norm": 3.333651065826416, "learning_rate": 9.677477954364298e-07, "loss": 0.1845, "step": 16240 }, { "epoch": 0.31862745098039214, "grad_norm": 3.5930235385894775, "learning_rate": 9.67690492979951e-07, "loss": 0.1637, "step": 16250 }, { "epoch": 0.31882352941176473, "grad_norm": 2.989150285720825, "learning_rate": 9.676331413636017e-07, "loss": 0.1787, "step": 16260 }, { "epoch": 0.31901960784313727, "grad_norm": 2.3188259601593018, "learning_rate": 9.675757405934103e-07, "loss": 0.1683, "step": 16270 }, { "epoch": 0.3192156862745098, "grad_norm": 1.4342639446258545, "learning_rate": 9.675182906754104e-07, "loss": 0.1866, "step": 16280 }, { "epoch": 0.31941176470588234, "grad_norm": 2.1078944206237793, "learning_rate": 9.674607916156405e-07, "loss": 0.1575, "step": 16290 }, { "epoch": 0.3196078431372549, "grad_norm": 7.413783073425293, "learning_rate": 9.674032434201447e-07, "loss": 0.1798, "step": 16300 }, { "epoch": 0.31980392156862747, "grad_norm": 3.5730817317962646, "learning_rate": 9.67345646094972e-07, "loss": 0.1913, "step": 16310 }, { "epoch": 0.32, "grad_norm": 2.190549373626709, "learning_rate": 9.67287999646176e-07, "loss": 0.1477, "step": 16320 }, { "epoch": 0.32019607843137254, "grad_norm": 2.859093427658081, "learning_rate": 9.672303040798166e-07, "loss": 0.1966, "step": 16330 }, { "epoch": 0.3203921568627451, "grad_norm": 2.950479507446289, "learning_rate": 9.671725594019583e-07, "loss": 0.1701, "step": 16340 }, { "epoch": 0.3205882352941177, "grad_norm": 4.067565441131592, "learning_rate": 9.671147656186707e-07, "loss": 0.1674, "step": 16350 }, { "epoch": 0.3207843137254902, "grad_norm": 2.807668924331665, "learning_rate": 9.670569227360283e-07, "loss": 0.1924, "step": 16360 }, { "epoch": 0.32098039215686275, "grad_norm": 2.0443484783172607, "learning_rate": 9.669990307601114e-07, "loss": 0.1651, "step": 16370 }, { "epoch": 0.3211764705882353, "grad_norm": 2.0659289360046387, "learning_rate": 9.66941089697005e-07, "loss": 0.1191, "step": 16380 }, { "epoch": 0.3213725490196078, "grad_norm": 2.924168348312378, "learning_rate": 9.668830995527998e-07, "loss": 0.1504, "step": 16390 }, { "epoch": 0.3215686274509804, "grad_norm": 3.595038414001465, "learning_rate": 9.668250603335904e-07, "loss": 0.163, "step": 16400 }, { "epoch": 0.32176470588235295, "grad_norm": 2.2711713314056396, "learning_rate": 9.667669720454783e-07, "loss": 0.1634, "step": 16410 }, { "epoch": 0.3219607843137255, "grad_norm": 1.6738516092300415, "learning_rate": 9.667088346945686e-07, "loss": 0.1663, "step": 16420 }, { "epoch": 0.322156862745098, "grad_norm": 2.528287410736084, "learning_rate": 9.666506482869728e-07, "loss": 0.1864, "step": 16430 }, { "epoch": 0.32235294117647056, "grad_norm": 3.1893532276153564, "learning_rate": 9.665924128288067e-07, "loss": 0.1836, "step": 16440 }, { "epoch": 0.32254901960784316, "grad_norm": 2.376847267150879, "learning_rate": 9.665341283261914e-07, "loss": 0.1635, "step": 16450 }, { "epoch": 0.3227450980392157, "grad_norm": 2.6940557956695557, "learning_rate": 9.664757947852536e-07, "loss": 0.1796, "step": 16460 }, { "epoch": 0.32294117647058823, "grad_norm": 3.67919659614563, "learning_rate": 9.664174122121246e-07, "loss": 0.1975, "step": 16470 }, { "epoch": 0.32313725490196077, "grad_norm": 2.0399630069732666, "learning_rate": 9.663589806129415e-07, "loss": 0.1504, "step": 16480 }, { "epoch": 0.3233333333333333, "grad_norm": 4.625296115875244, "learning_rate": 9.663004999938455e-07, "loss": 0.1613, "step": 16490 }, { "epoch": 0.3235294117647059, "grad_norm": 1.976119041442871, "learning_rate": 9.662419703609842e-07, "loss": 0.1872, "step": 16500 }, { "epoch": 0.32372549019607844, "grad_norm": 3.3390111923217773, "learning_rate": 9.661833917205097e-07, "loss": 0.175, "step": 16510 }, { "epoch": 0.323921568627451, "grad_norm": 2.0878076553344727, "learning_rate": 9.66124764078579e-07, "loss": 0.1991, "step": 16520 }, { "epoch": 0.3241176470588235, "grad_norm": 3.5707638263702393, "learning_rate": 9.660660874413548e-07, "loss": 0.1658, "step": 16530 }, { "epoch": 0.3243137254901961, "grad_norm": 4.372936248779297, "learning_rate": 9.660073618150046e-07, "loss": 0.2083, "step": 16540 }, { "epoch": 0.32450980392156864, "grad_norm": 2.6390838623046875, "learning_rate": 9.659485872057014e-07, "loss": 0.1264, "step": 16550 }, { "epoch": 0.3247058823529412, "grad_norm": 2.6910572052001953, "learning_rate": 9.658897636196228e-07, "loss": 0.1846, "step": 16560 }, { "epoch": 0.3249019607843137, "grad_norm": 3.403135299682617, "learning_rate": 9.65830891062952e-07, "loss": 0.182, "step": 16570 }, { "epoch": 0.32509803921568625, "grad_norm": 2.845431089401245, "learning_rate": 9.657719695418773e-07, "loss": 0.2126, "step": 16580 }, { "epoch": 0.32529411764705884, "grad_norm": 2.601463794708252, "learning_rate": 9.65712999062592e-07, "loss": 0.1538, "step": 16590 }, { "epoch": 0.3254901960784314, "grad_norm": 3.008991003036499, "learning_rate": 9.656539796312947e-07, "loss": 0.1855, "step": 16600 }, { "epoch": 0.3256862745098039, "grad_norm": 2.2144734859466553, "learning_rate": 9.655949112541886e-07, "loss": 0.1973, "step": 16610 }, { "epoch": 0.32588235294117646, "grad_norm": 2.4124038219451904, "learning_rate": 9.65535793937483e-07, "loss": 0.1543, "step": 16620 }, { "epoch": 0.326078431372549, "grad_norm": 2.4536197185516357, "learning_rate": 9.654766276873917e-07, "loss": 0.1946, "step": 16630 }, { "epoch": 0.3262745098039216, "grad_norm": 4.5563883781433105, "learning_rate": 9.654174125101338e-07, "loss": 0.1804, "step": 16640 }, { "epoch": 0.3264705882352941, "grad_norm": 1.432192325592041, "learning_rate": 9.653581484119332e-07, "loss": 0.1694, "step": 16650 }, { "epoch": 0.32666666666666666, "grad_norm": 2.1169052124023438, "learning_rate": 9.652988353990199e-07, "loss": 0.1509, "step": 16660 }, { "epoch": 0.3268627450980392, "grad_norm": 2.84993839263916, "learning_rate": 9.652394734776277e-07, "loss": 0.1389, "step": 16670 }, { "epoch": 0.3270588235294118, "grad_norm": 1.976631999015808, "learning_rate": 9.651800626539967e-07, "loss": 0.1811, "step": 16680 }, { "epoch": 0.32725490196078433, "grad_norm": 3.764317750930786, "learning_rate": 9.651206029343716e-07, "loss": 0.142, "step": 16690 }, { "epoch": 0.32745098039215687, "grad_norm": 3.473397970199585, "learning_rate": 9.650610943250023e-07, "loss": 0.2026, "step": 16700 }, { "epoch": 0.3276470588235294, "grad_norm": 2.058077096939087, "learning_rate": 9.65001536832144e-07, "loss": 0.1647, "step": 16710 }, { "epoch": 0.32784313725490194, "grad_norm": 2.269468069076538, "learning_rate": 9.649419304620568e-07, "loss": 0.1671, "step": 16720 }, { "epoch": 0.32803921568627453, "grad_norm": 3.31630539894104, "learning_rate": 9.64882275221006e-07, "loss": 0.1835, "step": 16730 }, { "epoch": 0.32823529411764707, "grad_norm": 3.630028009414673, "learning_rate": 9.648225711152618e-07, "loss": 0.1679, "step": 16740 }, { "epoch": 0.3284313725490196, "grad_norm": 1.6994318962097168, "learning_rate": 9.647628181511006e-07, "loss": 0.1426, "step": 16750 }, { "epoch": 0.32862745098039214, "grad_norm": 2.095123291015625, "learning_rate": 9.647030163348023e-07, "loss": 0.1597, "step": 16760 }, { "epoch": 0.3288235294117647, "grad_norm": 2.975661516189575, "learning_rate": 9.646431656726532e-07, "loss": 0.1926, "step": 16770 }, { "epoch": 0.3290196078431373, "grad_norm": 2.7162680625915527, "learning_rate": 9.645832661709443e-07, "loss": 0.1571, "step": 16780 }, { "epoch": 0.3292156862745098, "grad_norm": 2.9517412185668945, "learning_rate": 9.64523317835972e-07, "loss": 0.1709, "step": 16790 }, { "epoch": 0.32941176470588235, "grad_norm": 2.7956788539886475, "learning_rate": 9.64463320674037e-07, "loss": 0.1875, "step": 16800 }, { "epoch": 0.3296078431372549, "grad_norm": 2.1692047119140625, "learning_rate": 9.644032746914462e-07, "loss": 0.2001, "step": 16810 }, { "epoch": 0.3298039215686275, "grad_norm": 2.384352207183838, "learning_rate": 9.64343179894511e-07, "loss": 0.1707, "step": 16820 }, { "epoch": 0.33, "grad_norm": 3.985081195831299, "learning_rate": 9.64283036289548e-07, "loss": 0.1554, "step": 16830 }, { "epoch": 0.33019607843137255, "grad_norm": 5.287329196929932, "learning_rate": 9.642228438828791e-07, "loss": 0.1796, "step": 16840 }, { "epoch": 0.3303921568627451, "grad_norm": 2.745349407196045, "learning_rate": 9.641626026808312e-07, "loss": 0.2088, "step": 16850 }, { "epoch": 0.3305882352941176, "grad_norm": 4.532341957092285, "learning_rate": 9.641023126897365e-07, "loss": 0.1956, "step": 16860 }, { "epoch": 0.3307843137254902, "grad_norm": 2.8519043922424316, "learning_rate": 9.640419739159322e-07, "loss": 0.1618, "step": 16870 }, { "epoch": 0.33098039215686276, "grad_norm": 2.3892078399658203, "learning_rate": 9.639815863657604e-07, "loss": 0.1925, "step": 16880 }, { "epoch": 0.3311764705882353, "grad_norm": 2.6431150436401367, "learning_rate": 9.639211500455685e-07, "loss": 0.1884, "step": 16890 }, { "epoch": 0.33137254901960783, "grad_norm": 1.8350988626480103, "learning_rate": 9.638606649617096e-07, "loss": 0.1649, "step": 16900 }, { "epoch": 0.33156862745098037, "grad_norm": 2.420539140701294, "learning_rate": 9.63800131120541e-07, "loss": 0.1879, "step": 16910 }, { "epoch": 0.33176470588235296, "grad_norm": 1.6756631135940552, "learning_rate": 9.637395485284254e-07, "loss": 0.135, "step": 16920 }, { "epoch": 0.3319607843137255, "grad_norm": 1.8401501178741455, "learning_rate": 9.636789171917313e-07, "loss": 0.1648, "step": 16930 }, { "epoch": 0.33215686274509804, "grad_norm": 3.236325979232788, "learning_rate": 9.63618237116831e-07, "loss": 0.179, "step": 16940 }, { "epoch": 0.3323529411764706, "grad_norm": 3.1073789596557617, "learning_rate": 9.635575083101034e-07, "loss": 0.1973, "step": 16950 }, { "epoch": 0.3325490196078431, "grad_norm": 3.2894482612609863, "learning_rate": 9.634967307779317e-07, "loss": 0.1873, "step": 16960 }, { "epoch": 0.3327450980392157, "grad_norm": 3.6459760665893555, "learning_rate": 9.63435904526704e-07, "loss": 0.1681, "step": 16970 }, { "epoch": 0.33294117647058824, "grad_norm": 3.243448495864868, "learning_rate": 9.633750295628143e-07, "loss": 0.1982, "step": 16980 }, { "epoch": 0.3331372549019608, "grad_norm": 2.7217910289764404, "learning_rate": 9.63314105892661e-07, "loss": 0.1604, "step": 16990 }, { "epoch": 0.3333333333333333, "grad_norm": 2.5451059341430664, "learning_rate": 9.632531335226478e-07, "loss": 0.1651, "step": 17000 }, { "epoch": 0.3335294117647059, "grad_norm": 1.9048495292663574, "learning_rate": 9.631921124591839e-07, "loss": 0.1617, "step": 17010 }, { "epoch": 0.33372549019607844, "grad_norm": 2.2334444522857666, "learning_rate": 9.631310427086831e-07, "loss": 0.1964, "step": 17020 }, { "epoch": 0.333921568627451, "grad_norm": 2.5918009281158447, "learning_rate": 9.63069924277565e-07, "loss": 0.168, "step": 17030 }, { "epoch": 0.3341176470588235, "grad_norm": 3.3118515014648438, "learning_rate": 9.630087571722533e-07, "loss": 0.1835, "step": 17040 }, { "epoch": 0.33431372549019606, "grad_norm": 4.589146137237549, "learning_rate": 9.62947541399178e-07, "loss": 0.1751, "step": 17050 }, { "epoch": 0.33450980392156865, "grad_norm": 2.3321354389190674, "learning_rate": 9.628862769647732e-07, "loss": 0.1746, "step": 17060 }, { "epoch": 0.3347058823529412, "grad_norm": 4.32031774520874, "learning_rate": 9.628249638754784e-07, "loss": 0.1827, "step": 17070 }, { "epoch": 0.3349019607843137, "grad_norm": 2.933992385864258, "learning_rate": 9.627636021377388e-07, "loss": 0.1584, "step": 17080 }, { "epoch": 0.33509803921568626, "grad_norm": 1.981451153755188, "learning_rate": 9.627021917580038e-07, "loss": 0.1394, "step": 17090 }, { "epoch": 0.3352941176470588, "grad_norm": 4.202577114105225, "learning_rate": 9.626407327427287e-07, "loss": 0.1564, "step": 17100 }, { "epoch": 0.3354901960784314, "grad_norm": 2.1704180240631104, "learning_rate": 9.625792250983736e-07, "loss": 0.1461, "step": 17110 }, { "epoch": 0.33568627450980393, "grad_norm": 64.31185150146484, "learning_rate": 9.625176688314033e-07, "loss": 0.182, "step": 17120 }, { "epoch": 0.33588235294117647, "grad_norm": 3.973344326019287, "learning_rate": 9.624560639482884e-07, "loss": 0.1774, "step": 17130 }, { "epoch": 0.336078431372549, "grad_norm": 4.342866897583008, "learning_rate": 9.623944104555044e-07, "loss": 0.1598, "step": 17140 }, { "epoch": 0.3362745098039216, "grad_norm": 2.452596426010132, "learning_rate": 9.623327083595315e-07, "loss": 0.1883, "step": 17150 }, { "epoch": 0.33647058823529413, "grad_norm": 2.7377305030822754, "learning_rate": 9.622709576668558e-07, "loss": 0.1977, "step": 17160 }, { "epoch": 0.33666666666666667, "grad_norm": 3.5561625957489014, "learning_rate": 9.622091583839673e-07, "loss": 0.1487, "step": 17170 }, { "epoch": 0.3368627450980392, "grad_norm": 1.60821533203125, "learning_rate": 9.621473105173626e-07, "loss": 0.146, "step": 17180 }, { "epoch": 0.33705882352941174, "grad_norm": 2.5558743476867676, "learning_rate": 9.620854140735423e-07, "loss": 0.1787, "step": 17190 }, { "epoch": 0.33725490196078434, "grad_norm": 1.4481199979782104, "learning_rate": 9.620234690590125e-07, "loss": 0.1326, "step": 17200 }, { "epoch": 0.3374509803921569, "grad_norm": 2.141871213912964, "learning_rate": 9.619614754802844e-07, "loss": 0.1881, "step": 17210 }, { "epoch": 0.3376470588235294, "grad_norm": 4.862191677093506, "learning_rate": 9.61899433343874e-07, "loss": 0.1689, "step": 17220 }, { "epoch": 0.33784313725490195, "grad_norm": 3.447308301925659, "learning_rate": 9.618373426563031e-07, "loss": 0.1614, "step": 17230 }, { "epoch": 0.3380392156862745, "grad_norm": 1.4837919473648071, "learning_rate": 9.61775203424098e-07, "loss": 0.1484, "step": 17240 }, { "epoch": 0.3382352941176471, "grad_norm": 1.8071621656417847, "learning_rate": 9.617130156537901e-07, "loss": 0.1768, "step": 17250 }, { "epoch": 0.3384313725490196, "grad_norm": 1.722058653831482, "learning_rate": 9.616507793519165e-07, "loss": 0.1612, "step": 17260 }, { "epoch": 0.33862745098039215, "grad_norm": 3.7126636505126953, "learning_rate": 9.615884945250184e-07, "loss": 0.1629, "step": 17270 }, { "epoch": 0.3388235294117647, "grad_norm": 2.5114071369171143, "learning_rate": 9.61526161179643e-07, "loss": 0.187, "step": 17280 }, { "epoch": 0.3390196078431372, "grad_norm": 1.506424903869629, "learning_rate": 9.614637793223424e-07, "loss": 0.1756, "step": 17290 }, { "epoch": 0.3392156862745098, "grad_norm": 1.4202415943145752, "learning_rate": 9.614013489596735e-07, "loss": 0.1582, "step": 17300 }, { "epoch": 0.33941176470588236, "grad_norm": 2.9779772758483887, "learning_rate": 9.613388700981986e-07, "loss": 0.1601, "step": 17310 }, { "epoch": 0.3396078431372549, "grad_norm": 1.3647172451019287, "learning_rate": 9.61276342744485e-07, "loss": 0.1373, "step": 17320 }, { "epoch": 0.33980392156862743, "grad_norm": 2.555346965789795, "learning_rate": 9.612137669051049e-07, "loss": 0.1514, "step": 17330 }, { "epoch": 0.34, "grad_norm": 5.430665969848633, "learning_rate": 9.611511425866356e-07, "loss": 0.1542, "step": 17340 }, { "epoch": 0.34019607843137256, "grad_norm": 5.161334037780762, "learning_rate": 9.610884697956602e-07, "loss": 0.163, "step": 17350 }, { "epoch": 0.3403921568627451, "grad_norm": 2.5546138286590576, "learning_rate": 9.610257485387661e-07, "loss": 0.1656, "step": 17360 }, { "epoch": 0.34058823529411764, "grad_norm": 1.965795636177063, "learning_rate": 9.60962978822546e-07, "loss": 0.1762, "step": 17370 }, { "epoch": 0.3407843137254902, "grad_norm": 4.149137496948242, "learning_rate": 9.609001606535977e-07, "loss": 0.1451, "step": 17380 }, { "epoch": 0.34098039215686277, "grad_norm": 2.671832799911499, "learning_rate": 9.608372940385243e-07, "loss": 0.1755, "step": 17390 }, { "epoch": 0.3411764705882353, "grad_norm": 2.0048906803131104, "learning_rate": 9.607743789839336e-07, "loss": 0.1768, "step": 17400 }, { "epoch": 0.34137254901960784, "grad_norm": 3.448824644088745, "learning_rate": 9.60711415496439e-07, "loss": 0.183, "step": 17410 }, { "epoch": 0.3415686274509804, "grad_norm": 1.8874578475952148, "learning_rate": 9.606484035826586e-07, "loss": 0.1777, "step": 17420 }, { "epoch": 0.3417647058823529, "grad_norm": 3.814511775970459, "learning_rate": 9.605853432492158e-07, "loss": 0.1851, "step": 17430 }, { "epoch": 0.3419607843137255, "grad_norm": 2.0988168716430664, "learning_rate": 9.605222345027385e-07, "loss": 0.1911, "step": 17440 }, { "epoch": 0.34215686274509804, "grad_norm": 2.383424758911133, "learning_rate": 9.60459077349861e-07, "loss": 0.1756, "step": 17450 }, { "epoch": 0.3423529411764706, "grad_norm": 3.5447003841400146, "learning_rate": 9.603958717972214e-07, "loss": 0.1883, "step": 17460 }, { "epoch": 0.3425490196078431, "grad_norm": 3.4452764987945557, "learning_rate": 9.603326178514633e-07, "loss": 0.1734, "step": 17470 }, { "epoch": 0.3427450980392157, "grad_norm": 1.5616416931152344, "learning_rate": 9.602693155192355e-07, "loss": 0.1917, "step": 17480 }, { "epoch": 0.34294117647058825, "grad_norm": 2.160130023956299, "learning_rate": 9.60205964807192e-07, "loss": 0.1366, "step": 17490 }, { "epoch": 0.3431372549019608, "grad_norm": 3.4950015544891357, "learning_rate": 9.601425657219919e-07, "loss": 0.202, "step": 17500 }, { "epoch": 0.3433333333333333, "grad_norm": 2.8430752754211426, "learning_rate": 9.600791182702986e-07, "loss": 0.1839, "step": 17510 }, { "epoch": 0.34352941176470586, "grad_norm": 3.202892541885376, "learning_rate": 9.600156224587815e-07, "loss": 0.1599, "step": 17520 }, { "epoch": 0.34372549019607845, "grad_norm": 2.3155770301818848, "learning_rate": 9.599520782941148e-07, "loss": 0.1831, "step": 17530 }, { "epoch": 0.343921568627451, "grad_norm": 4.380270004272461, "learning_rate": 9.59888485782978e-07, "loss": 0.2265, "step": 17540 }, { "epoch": 0.34411764705882353, "grad_norm": 2.5794942378997803, "learning_rate": 9.598248449320547e-07, "loss": 0.162, "step": 17550 }, { "epoch": 0.34431372549019607, "grad_norm": 3.7602858543395996, "learning_rate": 9.597611557480353e-07, "loss": 0.1976, "step": 17560 }, { "epoch": 0.3445098039215686, "grad_norm": 5.351653575897217, "learning_rate": 9.596974182376133e-07, "loss": 0.1601, "step": 17570 }, { "epoch": 0.3447058823529412, "grad_norm": 6.495424270629883, "learning_rate": 9.59633632407489e-07, "loss": 0.1857, "step": 17580 }, { "epoch": 0.34490196078431373, "grad_norm": 1.8539018630981445, "learning_rate": 9.595697982643668e-07, "loss": 0.179, "step": 17590 }, { "epoch": 0.34509803921568627, "grad_norm": 2.635315179824829, "learning_rate": 9.595059158149563e-07, "loss": 0.1897, "step": 17600 }, { "epoch": 0.3452941176470588, "grad_norm": 3.039846897125244, "learning_rate": 9.594419850659727e-07, "loss": 0.1723, "step": 17610 }, { "epoch": 0.3454901960784314, "grad_norm": 4.5740814208984375, "learning_rate": 9.593780060241355e-07, "loss": 0.1684, "step": 17620 }, { "epoch": 0.34568627450980394, "grad_norm": 4.348100185394287, "learning_rate": 9.593139786961698e-07, "loss": 0.1807, "step": 17630 }, { "epoch": 0.3458823529411765, "grad_norm": 2.012362480163574, "learning_rate": 9.592499030888054e-07, "loss": 0.159, "step": 17640 }, { "epoch": 0.346078431372549, "grad_norm": 3.2964327335357666, "learning_rate": 9.59185779208778e-07, "loss": 0.1676, "step": 17650 }, { "epoch": 0.34627450980392155, "grad_norm": 2.443364143371582, "learning_rate": 9.591216070628274e-07, "loss": 0.1482, "step": 17660 }, { "epoch": 0.34647058823529414, "grad_norm": 1.504224419593811, "learning_rate": 9.590573866576987e-07, "loss": 0.1386, "step": 17670 }, { "epoch": 0.3466666666666667, "grad_norm": 13.873739242553711, "learning_rate": 9.589931180001426e-07, "loss": 0.1709, "step": 17680 }, { "epoch": 0.3468627450980392, "grad_norm": 1.4385544061660767, "learning_rate": 9.589288010969143e-07, "loss": 0.1568, "step": 17690 }, { "epoch": 0.34705882352941175, "grad_norm": 2.552950859069824, "learning_rate": 9.588644359547744e-07, "loss": 0.1757, "step": 17700 }, { "epoch": 0.3472549019607843, "grad_norm": 1.1205060482025146, "learning_rate": 9.58800022580488e-07, "loss": 0.1641, "step": 17710 }, { "epoch": 0.3474509803921569, "grad_norm": 7.426922798156738, "learning_rate": 9.587355609808265e-07, "loss": 0.2029, "step": 17720 }, { "epoch": 0.3476470588235294, "grad_norm": 1.9759958982467651, "learning_rate": 9.58671051162565e-07, "loss": 0.1549, "step": 17730 }, { "epoch": 0.34784313725490196, "grad_norm": 1.8053147792816162, "learning_rate": 9.586064931324844e-07, "loss": 0.1594, "step": 17740 }, { "epoch": 0.3480392156862745, "grad_norm": 2.4013452529907227, "learning_rate": 9.585418868973705e-07, "loss": 0.1353, "step": 17750 }, { "epoch": 0.34823529411764703, "grad_norm": 3.2591183185577393, "learning_rate": 9.584772324640144e-07, "loss": 0.1827, "step": 17760 }, { "epoch": 0.3484313725490196, "grad_norm": 2.990060567855835, "learning_rate": 9.58412529839212e-07, "loss": 0.1793, "step": 17770 }, { "epoch": 0.34862745098039216, "grad_norm": 3.1493470668792725, "learning_rate": 9.583477790297639e-07, "loss": 0.1348, "step": 17780 }, { "epoch": 0.3488235294117647, "grad_norm": 1.485344409942627, "learning_rate": 9.582829800424765e-07, "loss": 0.1533, "step": 17790 }, { "epoch": 0.34901960784313724, "grad_norm": 2.2150285243988037, "learning_rate": 9.58218132884161e-07, "loss": 0.1809, "step": 17800 }, { "epoch": 0.34921568627450983, "grad_norm": 2.932947874069214, "learning_rate": 9.581532375616335e-07, "loss": 0.2074, "step": 17810 }, { "epoch": 0.34941176470588237, "grad_norm": 2.9187405109405518, "learning_rate": 9.580882940817155e-07, "loss": 0.1932, "step": 17820 }, { "epoch": 0.3496078431372549, "grad_norm": 2.3328659534454346, "learning_rate": 9.580233024512329e-07, "loss": 0.1551, "step": 17830 }, { "epoch": 0.34980392156862744, "grad_norm": 1.911021113395691, "learning_rate": 9.579582626770175e-07, "loss": 0.1323, "step": 17840 }, { "epoch": 0.35, "grad_norm": 3.527097702026367, "learning_rate": 9.578931747659058e-07, "loss": 0.1668, "step": 17850 }, { "epoch": 0.35019607843137257, "grad_norm": 3.1569247245788574, "learning_rate": 9.57828038724739e-07, "loss": 0.1981, "step": 17860 }, { "epoch": 0.3503921568627451, "grad_norm": 3.469979763031006, "learning_rate": 9.577628545603639e-07, "loss": 0.1661, "step": 17870 }, { "epoch": 0.35058823529411764, "grad_norm": 4.284945487976074, "learning_rate": 9.576976222796318e-07, "loss": 0.2031, "step": 17880 }, { "epoch": 0.3507843137254902, "grad_norm": 2.740546703338623, "learning_rate": 9.576323418894e-07, "loss": 0.172, "step": 17890 }, { "epoch": 0.3509803921568627, "grad_norm": 3.1533172130584717, "learning_rate": 9.575670133965296e-07, "loss": 0.1733, "step": 17900 }, { "epoch": 0.3511764705882353, "grad_norm": 3.4303243160247803, "learning_rate": 9.57501636807888e-07, "loss": 0.1793, "step": 17910 }, { "epoch": 0.35137254901960785, "grad_norm": 8.692713737487793, "learning_rate": 9.574362121303465e-07, "loss": 0.1789, "step": 17920 }, { "epoch": 0.3515686274509804, "grad_norm": 4.54182767868042, "learning_rate": 9.573707393707823e-07, "loss": 0.1805, "step": 17930 }, { "epoch": 0.3517647058823529, "grad_norm": 5.6574482917785645, "learning_rate": 9.573052185360774e-07, "loss": 0.1658, "step": 17940 }, { "epoch": 0.3519607843137255, "grad_norm": 2.5202760696411133, "learning_rate": 9.57239649633119e-07, "loss": 0.1554, "step": 17950 }, { "epoch": 0.35215686274509805, "grad_norm": 3.6397440433502197, "learning_rate": 9.571740326687987e-07, "loss": 0.1695, "step": 17960 }, { "epoch": 0.3523529411764706, "grad_norm": 1.7541970014572144, "learning_rate": 9.57108367650014e-07, "loss": 0.1712, "step": 17970 }, { "epoch": 0.35254901960784313, "grad_norm": 1.9935483932495117, "learning_rate": 9.57042654583667e-07, "loss": 0.1504, "step": 17980 }, { "epoch": 0.35274509803921567, "grad_norm": 3.119462490081787, "learning_rate": 9.569768934766649e-07, "loss": 0.1769, "step": 17990 }, { "epoch": 0.35294117647058826, "grad_norm": 1.9114563465118408, "learning_rate": 9.569110843359198e-07, "loss": 0.1545, "step": 18000 }, { "epoch": 0.3531372549019608, "grad_norm": 5.776495933532715, "learning_rate": 9.568452271683494e-07, "loss": 0.1588, "step": 18010 }, { "epoch": 0.35333333333333333, "grad_norm": 1.4888848066329956, "learning_rate": 9.567793219808757e-07, "loss": 0.1632, "step": 18020 }, { "epoch": 0.35352941176470587, "grad_norm": 1.939670205116272, "learning_rate": 9.567133687804266e-07, "loss": 0.1352, "step": 18030 }, { "epoch": 0.3537254901960784, "grad_norm": 2.8305771350860596, "learning_rate": 9.56647367573934e-07, "loss": 0.1541, "step": 18040 }, { "epoch": 0.353921568627451, "grad_norm": 2.2130517959594727, "learning_rate": 9.565813183683357e-07, "loss": 0.1828, "step": 18050 }, { "epoch": 0.35411764705882354, "grad_norm": 2.27154803276062, "learning_rate": 9.565152211705744e-07, "loss": 0.1383, "step": 18060 }, { "epoch": 0.3543137254901961, "grad_norm": 2.2165043354034424, "learning_rate": 9.564490759875975e-07, "loss": 0.1558, "step": 18070 }, { "epoch": 0.3545098039215686, "grad_norm": 2.8652381896972656, "learning_rate": 9.563828828263577e-07, "loss": 0.201, "step": 18080 }, { "epoch": 0.3547058823529412, "grad_norm": 2.014169216156006, "learning_rate": 9.563166416938127e-07, "loss": 0.1877, "step": 18090 }, { "epoch": 0.35490196078431374, "grad_norm": 1.4503389596939087, "learning_rate": 9.562503525969253e-07, "loss": 0.1663, "step": 18100 }, { "epoch": 0.3550980392156863, "grad_norm": 1.849640965461731, "learning_rate": 9.56184015542663e-07, "loss": 0.1781, "step": 18110 }, { "epoch": 0.3552941176470588, "grad_norm": 2.5195183753967285, "learning_rate": 9.561176305379992e-07, "loss": 0.1608, "step": 18120 }, { "epoch": 0.35549019607843135, "grad_norm": 2.73960280418396, "learning_rate": 9.560511975899112e-07, "loss": 0.1921, "step": 18130 }, { "epoch": 0.35568627450980395, "grad_norm": 3.8419253826141357, "learning_rate": 9.559847167053822e-07, "loss": 0.1778, "step": 18140 }, { "epoch": 0.3558823529411765, "grad_norm": 3.1208715438842773, "learning_rate": 9.559181878914e-07, "loss": 0.1856, "step": 18150 }, { "epoch": 0.356078431372549, "grad_norm": 3.2293310165405273, "learning_rate": 9.558516111549574e-07, "loss": 0.1988, "step": 18160 }, { "epoch": 0.35627450980392156, "grad_norm": 3.2213456630706787, "learning_rate": 9.557849865030528e-07, "loss": 0.1719, "step": 18170 }, { "epoch": 0.3564705882352941, "grad_norm": 2.779463052749634, "learning_rate": 9.55718313942689e-07, "loss": 0.17, "step": 18180 }, { "epoch": 0.3566666666666667, "grad_norm": 3.5201687812805176, "learning_rate": 9.556515934808742e-07, "loss": 0.1921, "step": 18190 }, { "epoch": 0.3568627450980392, "grad_norm": 2.1550443172454834, "learning_rate": 9.555848251246218e-07, "loss": 0.2012, "step": 18200 }, { "epoch": 0.35705882352941176, "grad_norm": 3.7320218086242676, "learning_rate": 9.55518008880949e-07, "loss": 0.1773, "step": 18210 }, { "epoch": 0.3572549019607843, "grad_norm": 3.181173801422119, "learning_rate": 9.5545114475688e-07, "loss": 0.1748, "step": 18220 }, { "epoch": 0.35745098039215684, "grad_norm": 3.2071471214294434, "learning_rate": 9.553842327594425e-07, "loss": 0.1631, "step": 18230 }, { "epoch": 0.35764705882352943, "grad_norm": 3.6278064250946045, "learning_rate": 9.5531727289567e-07, "loss": 0.16, "step": 18240 }, { "epoch": 0.35784313725490197, "grad_norm": 2.8055176734924316, "learning_rate": 9.552502651726006e-07, "loss": 0.1678, "step": 18250 }, { "epoch": 0.3580392156862745, "grad_norm": 2.108327865600586, "learning_rate": 9.551832095972778e-07, "loss": 0.155, "step": 18260 }, { "epoch": 0.35823529411764704, "grad_norm": 3.483288288116455, "learning_rate": 9.551161061767497e-07, "loss": 0.1827, "step": 18270 }, { "epoch": 0.35843137254901963, "grad_norm": 3.2572290897369385, "learning_rate": 9.550489549180699e-07, "loss": 0.1804, "step": 18280 }, { "epoch": 0.35862745098039217, "grad_norm": 3.281959295272827, "learning_rate": 9.549817558282965e-07, "loss": 0.1485, "step": 18290 }, { "epoch": 0.3588235294117647, "grad_norm": 1.6275568008422852, "learning_rate": 9.549145089144933e-07, "loss": 0.1643, "step": 18300 }, { "epoch": 0.35901960784313725, "grad_norm": 3.4711036682128906, "learning_rate": 9.548472141837285e-07, "loss": 0.1813, "step": 18310 }, { "epoch": 0.3592156862745098, "grad_norm": 2.8882434368133545, "learning_rate": 9.547798716430758e-07, "loss": 0.1711, "step": 18320 }, { "epoch": 0.3594117647058824, "grad_norm": 3.7509007453918457, "learning_rate": 9.547124812996136e-07, "loss": 0.1704, "step": 18330 }, { "epoch": 0.3596078431372549, "grad_norm": 2.323331117630005, "learning_rate": 9.546450431604252e-07, "loss": 0.149, "step": 18340 }, { "epoch": 0.35980392156862745, "grad_norm": 3.7932064533233643, "learning_rate": 9.545775572325994e-07, "loss": 0.185, "step": 18350 }, { "epoch": 0.36, "grad_norm": 2.896311044692993, "learning_rate": 9.545100235232298e-07, "loss": 0.1343, "step": 18360 }, { "epoch": 0.3601960784313725, "grad_norm": 2.519399881362915, "learning_rate": 9.544424420394151e-07, "loss": 0.163, "step": 18370 }, { "epoch": 0.3603921568627451, "grad_norm": 2.472062587738037, "learning_rate": 9.543748127882584e-07, "loss": 0.1754, "step": 18380 }, { "epoch": 0.36058823529411765, "grad_norm": 1.8200421333312988, "learning_rate": 9.54307135776869e-07, "loss": 0.1749, "step": 18390 }, { "epoch": 0.3607843137254902, "grad_norm": 2.2597239017486572, "learning_rate": 9.542394110123602e-07, "loss": 0.1639, "step": 18400 }, { "epoch": 0.36098039215686273, "grad_norm": 1.7466936111450195, "learning_rate": 9.541716385018509e-07, "loss": 0.1626, "step": 18410 }, { "epoch": 0.3611764705882353, "grad_norm": 1.6306592226028442, "learning_rate": 9.541038182524643e-07, "loss": 0.1494, "step": 18420 }, { "epoch": 0.36137254901960786, "grad_norm": 2.1518735885620117, "learning_rate": 9.540359502713297e-07, "loss": 0.1801, "step": 18430 }, { "epoch": 0.3615686274509804, "grad_norm": 1.9311226606369019, "learning_rate": 9.539680345655808e-07, "loss": 0.1621, "step": 18440 }, { "epoch": 0.36176470588235293, "grad_norm": 3.0651469230651855, "learning_rate": 9.53900071142356e-07, "loss": 0.1727, "step": 18450 }, { "epoch": 0.36196078431372547, "grad_norm": 3.0096328258514404, "learning_rate": 9.53832060008799e-07, "loss": 0.1323, "step": 18460 }, { "epoch": 0.36215686274509806, "grad_norm": 8.17846393585205, "learning_rate": 9.537640011720593e-07, "loss": 0.1511, "step": 18470 }, { "epoch": 0.3623529411764706, "grad_norm": 2.68349027633667, "learning_rate": 9.536958946392899e-07, "loss": 0.1977, "step": 18480 }, { "epoch": 0.36254901960784314, "grad_norm": 3.1851038932800293, "learning_rate": 9.5362774041765e-07, "loss": 0.1508, "step": 18490 }, { "epoch": 0.3627450980392157, "grad_norm": 4.075669288635254, "learning_rate": 9.535595385143034e-07, "loss": 0.1342, "step": 18500 }, { "epoch": 0.3629411764705882, "grad_norm": 2.7475461959838867, "learning_rate": 9.534912889364189e-07, "loss": 0.155, "step": 18510 }, { "epoch": 0.3631372549019608, "grad_norm": 1.8580695390701294, "learning_rate": 9.534229916911703e-07, "loss": 0.1689, "step": 18520 }, { "epoch": 0.36333333333333334, "grad_norm": 1.6574654579162598, "learning_rate": 9.533546467857365e-07, "loss": 0.1777, "step": 18530 }, { "epoch": 0.3635294117647059, "grad_norm": 2.7139549255371094, "learning_rate": 9.532862542273014e-07, "loss": 0.1674, "step": 18540 }, { "epoch": 0.3637254901960784, "grad_norm": 3.58244252204895, "learning_rate": 9.53217814023054e-07, "loss": 0.1373, "step": 18550 }, { "epoch": 0.36392156862745095, "grad_norm": 2.8236725330352783, "learning_rate": 9.531493261801879e-07, "loss": 0.2023, "step": 18560 }, { "epoch": 0.36411764705882355, "grad_norm": 3.153860569000244, "learning_rate": 9.530807907059019e-07, "loss": 0.2089, "step": 18570 }, { "epoch": 0.3643137254901961, "grad_norm": 2.122807025909424, "learning_rate": 9.530122076074005e-07, "loss": 0.1653, "step": 18580 }, { "epoch": 0.3645098039215686, "grad_norm": 2.075221061706543, "learning_rate": 9.52943576891892e-07, "loss": 0.1839, "step": 18590 }, { "epoch": 0.36470588235294116, "grad_norm": 2.0271830558776855, "learning_rate": 9.528748985665906e-07, "loss": 0.1774, "step": 18600 }, { "epoch": 0.36490196078431375, "grad_norm": 2.2892229557037354, "learning_rate": 9.528061726387152e-07, "loss": 0.1606, "step": 18610 }, { "epoch": 0.3650980392156863, "grad_norm": 2.8887221813201904, "learning_rate": 9.527373991154897e-07, "loss": 0.196, "step": 18620 }, { "epoch": 0.3652941176470588, "grad_norm": 2.1374964714050293, "learning_rate": 9.526685780041429e-07, "loss": 0.1423, "step": 18630 }, { "epoch": 0.36549019607843136, "grad_norm": 1.8805214166641235, "learning_rate": 9.525997093119088e-07, "loss": 0.1373, "step": 18640 }, { "epoch": 0.3656862745098039, "grad_norm": 2.9640777111053467, "learning_rate": 9.525307930460265e-07, "loss": 0.1938, "step": 18650 }, { "epoch": 0.3658823529411765, "grad_norm": 2.84464168548584, "learning_rate": 9.524618292137397e-07, "loss": 0.1665, "step": 18660 }, { "epoch": 0.36607843137254903, "grad_norm": 2.498542547225952, "learning_rate": 9.523928178222972e-07, "loss": 0.1841, "step": 18670 }, { "epoch": 0.36627450980392157, "grad_norm": 4.211982250213623, "learning_rate": 9.523237588789532e-07, "loss": 0.1779, "step": 18680 }, { "epoch": 0.3664705882352941, "grad_norm": 2.2312989234924316, "learning_rate": 9.522546523909667e-07, "loss": 0.1557, "step": 18690 }, { "epoch": 0.36666666666666664, "grad_norm": 5.596581935882568, "learning_rate": 9.521854983656012e-07, "loss": 0.1868, "step": 18700 }, { "epoch": 0.36686274509803923, "grad_norm": 3.1209254264831543, "learning_rate": 9.52116296810126e-07, "loss": 0.1351, "step": 18710 }, { "epoch": 0.36705882352941177, "grad_norm": 3.066851854324341, "learning_rate": 9.52047047731815e-07, "loss": 0.1709, "step": 18720 }, { "epoch": 0.3672549019607843, "grad_norm": 3.5349130630493164, "learning_rate": 9.519777511379468e-07, "loss": 0.1796, "step": 18730 }, { "epoch": 0.36745098039215685, "grad_norm": 3.3594744205474854, "learning_rate": 9.519084070358056e-07, "loss": 0.1649, "step": 18740 }, { "epoch": 0.36764705882352944, "grad_norm": 2.459867477416992, "learning_rate": 9.518390154326801e-07, "loss": 0.158, "step": 18750 }, { "epoch": 0.367843137254902, "grad_norm": 2.196578025817871, "learning_rate": 9.517695763358644e-07, "loss": 0.2011, "step": 18760 }, { "epoch": 0.3680392156862745, "grad_norm": 7.797895908355713, "learning_rate": 9.517000897526571e-07, "loss": 0.1502, "step": 18770 }, { "epoch": 0.36823529411764705, "grad_norm": 2.019598960876465, "learning_rate": 9.516305556903624e-07, "loss": 0.1587, "step": 18780 }, { "epoch": 0.3684313725490196, "grad_norm": 3.0581727027893066, "learning_rate": 9.515609741562889e-07, "loss": 0.1618, "step": 18790 }, { "epoch": 0.3686274509803922, "grad_norm": 3.2742528915405273, "learning_rate": 9.514913451577504e-07, "loss": 0.1706, "step": 18800 }, { "epoch": 0.3688235294117647, "grad_norm": 3.0979321002960205, "learning_rate": 9.514216687020661e-07, "loss": 0.1841, "step": 18810 }, { "epoch": 0.36901960784313725, "grad_norm": 2.409700632095337, "learning_rate": 9.513519447965595e-07, "loss": 0.1996, "step": 18820 }, { "epoch": 0.3692156862745098, "grad_norm": 1.4475224018096924, "learning_rate": 9.512821734485597e-07, "loss": 0.1818, "step": 18830 }, { "epoch": 0.36941176470588233, "grad_norm": 5.072595119476318, "learning_rate": 9.512123546654001e-07, "loss": 0.145, "step": 18840 }, { "epoch": 0.3696078431372549, "grad_norm": 3.00062894821167, "learning_rate": 9.511424884544197e-07, "loss": 0.1624, "step": 18850 }, { "epoch": 0.36980392156862746, "grad_norm": 3.9431731700897217, "learning_rate": 9.510725748229624e-07, "loss": 0.1904, "step": 18860 }, { "epoch": 0.37, "grad_norm": 2.647209644317627, "learning_rate": 9.510026137783769e-07, "loss": 0.1757, "step": 18870 }, { "epoch": 0.37019607843137253, "grad_norm": 2.361011505126953, "learning_rate": 9.509326053280169e-07, "loss": 0.1594, "step": 18880 }, { "epoch": 0.3703921568627451, "grad_norm": 3.321416139602661, "learning_rate": 9.50862549479241e-07, "loss": 0.187, "step": 18890 }, { "epoch": 0.37058823529411766, "grad_norm": 2.845273017883301, "learning_rate": 9.507924462394132e-07, "loss": 0.1915, "step": 18900 }, { "epoch": 0.3707843137254902, "grad_norm": 3.400665521621704, "learning_rate": 9.507222956159021e-07, "loss": 0.1624, "step": 18910 }, { "epoch": 0.37098039215686274, "grad_norm": 3.154435634613037, "learning_rate": 9.506520976160811e-07, "loss": 0.1548, "step": 18920 }, { "epoch": 0.3711764705882353, "grad_norm": 2.483719825744629, "learning_rate": 9.505818522473291e-07, "loss": 0.2118, "step": 18930 }, { "epoch": 0.37137254901960787, "grad_norm": 2.8475022315979004, "learning_rate": 9.505115595170297e-07, "loss": 0.1979, "step": 18940 }, { "epoch": 0.3715686274509804, "grad_norm": 2.7406930923461914, "learning_rate": 9.504412194325715e-07, "loss": 0.1817, "step": 18950 }, { "epoch": 0.37176470588235294, "grad_norm": 3.1972029209136963, "learning_rate": 9.50370832001348e-07, "loss": 0.162, "step": 18960 }, { "epoch": 0.3719607843137255, "grad_norm": 2.568873643875122, "learning_rate": 9.503003972307579e-07, "loss": 0.1685, "step": 18970 }, { "epoch": 0.372156862745098, "grad_norm": 5.571108341217041, "learning_rate": 9.502299151282046e-07, "loss": 0.1987, "step": 18980 }, { "epoch": 0.3723529411764706, "grad_norm": 2.877887010574341, "learning_rate": 9.501593857010968e-07, "loss": 0.1914, "step": 18990 }, { "epoch": 0.37254901960784315, "grad_norm": 3.846593141555786, "learning_rate": 9.500888089568478e-07, "loss": 0.1655, "step": 19000 }, { "epoch": 0.3727450980392157, "grad_norm": 2.7042734622955322, "learning_rate": 9.500181849028761e-07, "loss": 0.1757, "step": 19010 }, { "epoch": 0.3729411764705882, "grad_norm": 2.6091151237487793, "learning_rate": 9.499475135466051e-07, "loss": 0.1842, "step": 19020 }, { "epoch": 0.37313725490196076, "grad_norm": 5.8022966384887695, "learning_rate": 9.498767948954633e-07, "loss": 0.1604, "step": 19030 }, { "epoch": 0.37333333333333335, "grad_norm": 3.3402891159057617, "learning_rate": 9.498060289568841e-07, "loss": 0.1835, "step": 19040 }, { "epoch": 0.3735294117647059, "grad_norm": 2.7009992599487305, "learning_rate": 9.497352157383058e-07, "loss": 0.1944, "step": 19050 }, { "epoch": 0.3737254901960784, "grad_norm": 2.689500331878662, "learning_rate": 9.496643552471717e-07, "loss": 0.212, "step": 19060 }, { "epoch": 0.37392156862745096, "grad_norm": 3.201547145843506, "learning_rate": 9.495934474909302e-07, "loss": 0.202, "step": 19070 }, { "epoch": 0.37411764705882355, "grad_norm": 3.2583975791931152, "learning_rate": 9.495224924770342e-07, "loss": 0.1791, "step": 19080 }, { "epoch": 0.3743137254901961, "grad_norm": 2.0278091430664062, "learning_rate": 9.494514902129424e-07, "loss": 0.1736, "step": 19090 }, { "epoch": 0.37450980392156863, "grad_norm": 2.3864245414733887, "learning_rate": 9.493804407061176e-07, "loss": 0.1231, "step": 19100 }, { "epoch": 0.37470588235294117, "grad_norm": 2.620558738708496, "learning_rate": 9.493093439640283e-07, "loss": 0.1758, "step": 19110 }, { "epoch": 0.3749019607843137, "grad_norm": 1.8873084783554077, "learning_rate": 9.492381999941472e-07, "loss": 0.1494, "step": 19120 }, { "epoch": 0.3750980392156863, "grad_norm": 12.297877311706543, "learning_rate": 9.491670088039528e-07, "loss": 0.1768, "step": 19130 }, { "epoch": 0.37529411764705883, "grad_norm": 3.567490816116333, "learning_rate": 9.49095770400928e-07, "loss": 0.1506, "step": 19140 }, { "epoch": 0.37549019607843137, "grad_norm": 1.4598451852798462, "learning_rate": 9.490244847925606e-07, "loss": 0.1549, "step": 19150 }, { "epoch": 0.3756862745098039, "grad_norm": 2.0076887607574463, "learning_rate": 9.489531519863439e-07, "loss": 0.1859, "step": 19160 }, { "epoch": 0.37588235294117645, "grad_norm": 3.5803372859954834, "learning_rate": 9.488817719897754e-07, "loss": 0.1712, "step": 19170 }, { "epoch": 0.37607843137254904, "grad_norm": 2.2117629051208496, "learning_rate": 9.488103448103586e-07, "loss": 0.1564, "step": 19180 }, { "epoch": 0.3762745098039216, "grad_norm": 2.8999199867248535, "learning_rate": 9.487388704556008e-07, "loss": 0.1586, "step": 19190 }, { "epoch": 0.3764705882352941, "grad_norm": 2.064548969268799, "learning_rate": 9.48667348933015e-07, "loss": 0.1386, "step": 19200 }, { "epoch": 0.37666666666666665, "grad_norm": 2.2592849731445312, "learning_rate": 9.485957802501192e-07, "loss": 0.1884, "step": 19210 }, { "epoch": 0.37686274509803924, "grad_norm": 4.35361385345459, "learning_rate": 9.485241644144357e-07, "loss": 0.1546, "step": 19220 }, { "epoch": 0.3770588235294118, "grad_norm": 2.5871171951293945, "learning_rate": 9.484525014334927e-07, "loss": 0.1714, "step": 19230 }, { "epoch": 0.3772549019607843, "grad_norm": 2.851006031036377, "learning_rate": 9.483807913148223e-07, "loss": 0.1588, "step": 19240 }, { "epoch": 0.37745098039215685, "grad_norm": 2.271023750305176, "learning_rate": 9.483090340659623e-07, "loss": 0.1492, "step": 19250 }, { "epoch": 0.3776470588235294, "grad_norm": 2.107614755630493, "learning_rate": 9.482372296944555e-07, "loss": 0.1361, "step": 19260 }, { "epoch": 0.377843137254902, "grad_norm": 3.2300453186035156, "learning_rate": 9.48165378207849e-07, "loss": 0.2029, "step": 19270 }, { "epoch": 0.3780392156862745, "grad_norm": 2.4898619651794434, "learning_rate": 9.480934796136954e-07, "loss": 0.1908, "step": 19280 }, { "epoch": 0.37823529411764706, "grad_norm": 1.8646881580352783, "learning_rate": 9.480215339195524e-07, "loss": 0.1842, "step": 19290 }, { "epoch": 0.3784313725490196, "grad_norm": 3.9274253845214844, "learning_rate": 9.479495411329818e-07, "loss": 0.1981, "step": 19300 }, { "epoch": 0.37862745098039213, "grad_norm": 3.2133543491363525, "learning_rate": 9.478775012615513e-07, "loss": 0.1659, "step": 19310 }, { "epoch": 0.3788235294117647, "grad_norm": 3.3596415519714355, "learning_rate": 9.478054143128331e-07, "loss": 0.1702, "step": 19320 }, { "epoch": 0.37901960784313726, "grad_norm": 3.294736862182617, "learning_rate": 9.477332802944043e-07, "loss": 0.1886, "step": 19330 }, { "epoch": 0.3792156862745098, "grad_norm": 2.590306520462036, "learning_rate": 9.47661099213847e-07, "loss": 0.1534, "step": 19340 }, { "epoch": 0.37941176470588234, "grad_norm": 1.8753730058670044, "learning_rate": 9.475888710787486e-07, "loss": 0.1491, "step": 19350 }, { "epoch": 0.3796078431372549, "grad_norm": 1.24417245388031, "learning_rate": 9.475165958967006e-07, "loss": 0.1818, "step": 19360 }, { "epoch": 0.37980392156862747, "grad_norm": 2.2477638721466064, "learning_rate": 9.474442736753005e-07, "loss": 0.1717, "step": 19370 }, { "epoch": 0.38, "grad_norm": 4.1015706062316895, "learning_rate": 9.473719044221502e-07, "loss": 0.2226, "step": 19380 }, { "epoch": 0.38019607843137254, "grad_norm": 2.2248833179473877, "learning_rate": 9.472994881448563e-07, "loss": 0.18, "step": 19390 }, { "epoch": 0.3803921568627451, "grad_norm": 2.433987617492676, "learning_rate": 9.472270248510308e-07, "loss": 0.1445, "step": 19400 }, { "epoch": 0.38058823529411767, "grad_norm": 2.8225841522216797, "learning_rate": 9.471545145482902e-07, "loss": 0.1608, "step": 19410 }, { "epoch": 0.3807843137254902, "grad_norm": 0.9210551381111145, "learning_rate": 9.470819572442565e-07, "loss": 0.1517, "step": 19420 }, { "epoch": 0.38098039215686275, "grad_norm": 2.6172032356262207, "learning_rate": 9.470093529465564e-07, "loss": 0.1443, "step": 19430 }, { "epoch": 0.3811764705882353, "grad_norm": 3.3238914012908936, "learning_rate": 9.469367016628211e-07, "loss": 0.1729, "step": 19440 }, { "epoch": 0.3813725490196078, "grad_norm": 3.1196606159210205, "learning_rate": 9.468640034006876e-07, "loss": 0.1519, "step": 19450 }, { "epoch": 0.3815686274509804, "grad_norm": 2.74922513961792, "learning_rate": 9.467912581677969e-07, "loss": 0.1482, "step": 19460 }, { "epoch": 0.38176470588235295, "grad_norm": 12.110001564025879, "learning_rate": 9.467184659717958e-07, "loss": 0.1699, "step": 19470 }, { "epoch": 0.3819607843137255, "grad_norm": 5.485336780548096, "learning_rate": 9.466456268203352e-07, "loss": 0.1955, "step": 19480 }, { "epoch": 0.382156862745098, "grad_norm": 7.759063243865967, "learning_rate": 9.465727407210717e-07, "loss": 0.2078, "step": 19490 }, { "epoch": 0.38235294117647056, "grad_norm": 2.168095588684082, "learning_rate": 9.464998076816664e-07, "loss": 0.1758, "step": 19500 }, { "epoch": 0.38254901960784315, "grad_norm": 2.2985799312591553, "learning_rate": 9.464268277097853e-07, "loss": 0.2012, "step": 19510 }, { "epoch": 0.3827450980392157, "grad_norm": 3.78286075592041, "learning_rate": 9.463538008130998e-07, "loss": 0.1978, "step": 19520 }, { "epoch": 0.38294117647058823, "grad_norm": 4.356653690338135, "learning_rate": 9.462807269992857e-07, "loss": 0.1809, "step": 19530 }, { "epoch": 0.38313725490196077, "grad_norm": 3.8525094985961914, "learning_rate": 9.462076062760238e-07, "loss": 0.2015, "step": 19540 }, { "epoch": 0.38333333333333336, "grad_norm": 2.420833110809326, "learning_rate": 9.461344386510002e-07, "loss": 0.1964, "step": 19550 }, { "epoch": 0.3835294117647059, "grad_norm": 2.9396731853485107, "learning_rate": 9.460612241319055e-07, "loss": 0.1744, "step": 19560 }, { "epoch": 0.38372549019607843, "grad_norm": 4.084555149078369, "learning_rate": 9.459879627264356e-07, "loss": 0.1818, "step": 19570 }, { "epoch": 0.38392156862745097, "grad_norm": 3.149167060852051, "learning_rate": 9.459146544422911e-07, "loss": 0.1609, "step": 19580 }, { "epoch": 0.3841176470588235, "grad_norm": 1.806666612625122, "learning_rate": 9.458412992871774e-07, "loss": 0.1687, "step": 19590 }, { "epoch": 0.3843137254901961, "grad_norm": 2.221745491027832, "learning_rate": 9.457678972688052e-07, "loss": 0.1877, "step": 19600 }, { "epoch": 0.38450980392156864, "grad_norm": 3.902146339416504, "learning_rate": 9.4569444839489e-07, "loss": 0.1683, "step": 19610 }, { "epoch": 0.3847058823529412, "grad_norm": 1.8777323961257935, "learning_rate": 9.45620952673152e-07, "loss": 0.1786, "step": 19620 }, { "epoch": 0.3849019607843137, "grad_norm": 2.210796356201172, "learning_rate": 9.455474101113165e-07, "loss": 0.163, "step": 19630 }, { "epoch": 0.38509803921568625, "grad_norm": 3.2800419330596924, "learning_rate": 9.454738207171135e-07, "loss": 0.1474, "step": 19640 }, { "epoch": 0.38529411764705884, "grad_norm": 3.2330422401428223, "learning_rate": 9.454001844982786e-07, "loss": 0.1672, "step": 19650 }, { "epoch": 0.3854901960784314, "grad_norm": 2.503862142562866, "learning_rate": 9.453265014625515e-07, "loss": 0.1929, "step": 19660 }, { "epoch": 0.3856862745098039, "grad_norm": 1.9227880239486694, "learning_rate": 9.452527716176775e-07, "loss": 0.1693, "step": 19670 }, { "epoch": 0.38588235294117645, "grad_norm": 1.5647629499435425, "learning_rate": 9.45178994971406e-07, "loss": 0.1667, "step": 19680 }, { "epoch": 0.38607843137254905, "grad_norm": 3.567216157913208, "learning_rate": 9.451051715314922e-07, "loss": 0.2067, "step": 19690 }, { "epoch": 0.3862745098039216, "grad_norm": 2.306844711303711, "learning_rate": 9.450313013056956e-07, "loss": 0.1799, "step": 19700 }, { "epoch": 0.3864705882352941, "grad_norm": 2.1852385997772217, "learning_rate": 9.449573843017808e-07, "loss": 0.1867, "step": 19710 }, { "epoch": 0.38666666666666666, "grad_norm": 2.844312906265259, "learning_rate": 9.448834205275178e-07, "loss": 0.1284, "step": 19720 }, { "epoch": 0.3868627450980392, "grad_norm": 3.141352891921997, "learning_rate": 9.448094099906806e-07, "loss": 0.1679, "step": 19730 }, { "epoch": 0.3870588235294118, "grad_norm": 2.283867835998535, "learning_rate": 9.447353526990487e-07, "loss": 0.16, "step": 19740 }, { "epoch": 0.3872549019607843, "grad_norm": 2.3539044857025146, "learning_rate": 9.446612486604067e-07, "loss": 0.232, "step": 19750 }, { "epoch": 0.38745098039215686, "grad_norm": 6.21832799911499, "learning_rate": 9.445870978825434e-07, "loss": 0.176, "step": 19760 }, { "epoch": 0.3876470588235294, "grad_norm": 2.2737369537353516, "learning_rate": 9.445129003732532e-07, "loss": 0.1847, "step": 19770 }, { "epoch": 0.38784313725490194, "grad_norm": 2.846290111541748, "learning_rate": 9.444386561403351e-07, "loss": 0.1852, "step": 19780 }, { "epoch": 0.38803921568627453, "grad_norm": 6.41343879699707, "learning_rate": 9.443643651915932e-07, "loss": 0.1875, "step": 19790 }, { "epoch": 0.38823529411764707, "grad_norm": 3.1807806491851807, "learning_rate": 9.442900275348359e-07, "loss": 0.1788, "step": 19800 }, { "epoch": 0.3884313725490196, "grad_norm": 3.3777530193328857, "learning_rate": 9.442156431778773e-07, "loss": 0.1788, "step": 19810 }, { "epoch": 0.38862745098039214, "grad_norm": 4.957129001617432, "learning_rate": 9.441412121285364e-07, "loss": 0.1507, "step": 19820 }, { "epoch": 0.3888235294117647, "grad_norm": 3.647386312484741, "learning_rate": 9.440667343946363e-07, "loss": 0.1806, "step": 19830 }, { "epoch": 0.38901960784313727, "grad_norm": 3.3603625297546387, "learning_rate": 9.439922099840054e-07, "loss": 0.1608, "step": 19840 }, { "epoch": 0.3892156862745098, "grad_norm": 3.4730348587036133, "learning_rate": 9.439176389044778e-07, "loss": 0.1583, "step": 19850 }, { "epoch": 0.38941176470588235, "grad_norm": 7.585697650909424, "learning_rate": 9.43843021163891e-07, "loss": 0.1695, "step": 19860 }, { "epoch": 0.3896078431372549, "grad_norm": 3.017805576324463, "learning_rate": 9.43768356770089e-07, "loss": 0.17, "step": 19870 }, { "epoch": 0.3898039215686275, "grad_norm": 2.717661142349243, "learning_rate": 9.436936457309192e-07, "loss": 0.1765, "step": 19880 }, { "epoch": 0.39, "grad_norm": 5.889300346374512, "learning_rate": 9.436188880542351e-07, "loss": 0.1787, "step": 19890 }, { "epoch": 0.39019607843137255, "grad_norm": 4.0266289710998535, "learning_rate": 9.435440837478945e-07, "loss": 0.1423, "step": 19900 }, { "epoch": 0.3903921568627451, "grad_norm": 2.7589356899261475, "learning_rate": 9.434692328197601e-07, "loss": 0.158, "step": 19910 }, { "epoch": 0.3905882352941176, "grad_norm": 2.0604732036590576, "learning_rate": 9.433943352776999e-07, "loss": 0.1667, "step": 19920 }, { "epoch": 0.3907843137254902, "grad_norm": 2.149616241455078, "learning_rate": 9.433193911295863e-07, "loss": 0.1859, "step": 19930 }, { "epoch": 0.39098039215686275, "grad_norm": 2.4754393100738525, "learning_rate": 9.432444003832966e-07, "loss": 0.1678, "step": 19940 }, { "epoch": 0.3911764705882353, "grad_norm": 3.3602399826049805, "learning_rate": 9.431693630467139e-07, "loss": 0.1886, "step": 19950 }, { "epoch": 0.39137254901960783, "grad_norm": 5.0498948097229, "learning_rate": 9.43094279127725e-07, "loss": 0.1974, "step": 19960 }, { "epoch": 0.39156862745098037, "grad_norm": 1.699371099472046, "learning_rate": 9.430191486342223e-07, "loss": 0.1998, "step": 19970 }, { "epoch": 0.39176470588235296, "grad_norm": 1.30162513256073, "learning_rate": 9.429439715741029e-07, "loss": 0.1884, "step": 19980 }, { "epoch": 0.3919607843137255, "grad_norm": 4.704903602600098, "learning_rate": 9.428687479552688e-07, "loss": 0.146, "step": 19990 }, { "epoch": 0.39215686274509803, "grad_norm": 1.8472563028335571, "learning_rate": 9.427934777856268e-07, "loss": 0.1603, "step": 20000 }, { "epoch": 0.39235294117647057, "grad_norm": 2.5458967685699463, "learning_rate": 9.427181610730888e-07, "loss": 0.1631, "step": 20010 }, { "epoch": 0.39254901960784316, "grad_norm": 3.7229037284851074, "learning_rate": 9.426427978255714e-07, "loss": 0.18, "step": 20020 }, { "epoch": 0.3927450980392157, "grad_norm": 1.9043515920639038, "learning_rate": 9.425673880509963e-07, "loss": 0.1865, "step": 20030 }, { "epoch": 0.39294117647058824, "grad_norm": 10.953206062316895, "learning_rate": 9.424919317572899e-07, "loss": 0.2115, "step": 20040 }, { "epoch": 0.3931372549019608, "grad_norm": 3.1615025997161865, "learning_rate": 9.424164289523837e-07, "loss": 0.1666, "step": 20050 }, { "epoch": 0.3933333333333333, "grad_norm": 2.170717239379883, "learning_rate": 9.423408796442137e-07, "loss": 0.179, "step": 20060 }, { "epoch": 0.3935294117647059, "grad_norm": 1.0188337564468384, "learning_rate": 9.422652838407212e-07, "loss": 0.1207, "step": 20070 }, { "epoch": 0.39372549019607844, "grad_norm": 11.508759498596191, "learning_rate": 9.421896415498522e-07, "loss": 0.167, "step": 20080 }, { "epoch": 0.393921568627451, "grad_norm": 4.081296920776367, "learning_rate": 9.421139527795576e-07, "loss": 0.1642, "step": 20090 }, { "epoch": 0.3941176470588235, "grad_norm": 2.4292311668395996, "learning_rate": 9.420382175377933e-07, "loss": 0.2083, "step": 20100 }, { "epoch": 0.39431372549019605, "grad_norm": 2.071728467941284, "learning_rate": 9.419624358325197e-07, "loss": 0.1629, "step": 20110 }, { "epoch": 0.39450980392156865, "grad_norm": 2.5166823863983154, "learning_rate": 9.418866076717026e-07, "loss": 0.1826, "step": 20120 }, { "epoch": 0.3947058823529412, "grad_norm": 1.6519423723220825, "learning_rate": 9.418107330633124e-07, "loss": 0.1651, "step": 20130 }, { "epoch": 0.3949019607843137, "grad_norm": 2.147045850753784, "learning_rate": 9.417348120153243e-07, "loss": 0.1836, "step": 20140 }, { "epoch": 0.39509803921568626, "grad_norm": 2.3976149559020996, "learning_rate": 9.416588445357188e-07, "loss": 0.1675, "step": 20150 }, { "epoch": 0.3952941176470588, "grad_norm": 2.217919111251831, "learning_rate": 9.415828306324809e-07, "loss": 0.1666, "step": 20160 }, { "epoch": 0.3954901960784314, "grad_norm": 2.5541563034057617, "learning_rate": 9.415067703136004e-07, "loss": 0.1825, "step": 20170 }, { "epoch": 0.3956862745098039, "grad_norm": 2.494649887084961, "learning_rate": 9.414306635870721e-07, "loss": 0.159, "step": 20180 }, { "epoch": 0.39588235294117646, "grad_norm": 3.4442696571350098, "learning_rate": 9.41354510460896e-07, "loss": 0.185, "step": 20190 }, { "epoch": 0.396078431372549, "grad_norm": 10.209486961364746, "learning_rate": 9.412783109430765e-07, "loss": 0.1722, "step": 20200 }, { "epoch": 0.3962745098039216, "grad_norm": 2.2608067989349365, "learning_rate": 9.412020650416233e-07, "loss": 0.1906, "step": 20210 }, { "epoch": 0.39647058823529413, "grad_norm": 1.7208473682403564, "learning_rate": 9.411257727645505e-07, "loss": 0.1475, "step": 20220 }, { "epoch": 0.39666666666666667, "grad_norm": 3.693610191345215, "learning_rate": 9.410494341198773e-07, "loss": 0.1438, "step": 20230 }, { "epoch": 0.3968627450980392, "grad_norm": 3.3024516105651855, "learning_rate": 9.409730491156283e-07, "loss": 0.1683, "step": 20240 }, { "epoch": 0.39705882352941174, "grad_norm": 2.4333462715148926, "learning_rate": 9.408966177598318e-07, "loss": 0.1706, "step": 20250 }, { "epoch": 0.39725490196078433, "grad_norm": 3.4808433055877686, "learning_rate": 9.408201400605223e-07, "loss": 0.1656, "step": 20260 }, { "epoch": 0.39745098039215687, "grad_norm": 3.0565764904022217, "learning_rate": 9.407436160257379e-07, "loss": 0.1625, "step": 20270 }, { "epoch": 0.3976470588235294, "grad_norm": 2.02392578125, "learning_rate": 9.406670456635227e-07, "loss": 0.1679, "step": 20280 }, { "epoch": 0.39784313725490195, "grad_norm": 3.7655911445617676, "learning_rate": 9.405904289819249e-07, "loss": 0.1784, "step": 20290 }, { "epoch": 0.3980392156862745, "grad_norm": 2.601513147354126, "learning_rate": 9.405137659889979e-07, "loss": 0.1449, "step": 20300 }, { "epoch": 0.3982352941176471, "grad_norm": 2.6397552490234375, "learning_rate": 9.404370566927997e-07, "loss": 0.194, "step": 20310 }, { "epoch": 0.3984313725490196, "grad_norm": 2.5825068950653076, "learning_rate": 9.403603011013939e-07, "loss": 0.1652, "step": 20320 }, { "epoch": 0.39862745098039215, "grad_norm": 5.348821640014648, "learning_rate": 9.402834992228478e-07, "loss": 0.1757, "step": 20330 }, { "epoch": 0.3988235294117647, "grad_norm": 2.6837198734283447, "learning_rate": 9.402066510652346e-07, "loss": 0.1699, "step": 20340 }, { "epoch": 0.3990196078431373, "grad_norm": 4.446052551269531, "learning_rate": 9.401297566366317e-07, "loss": 0.1519, "step": 20350 }, { "epoch": 0.3992156862745098, "grad_norm": 2.3395025730133057, "learning_rate": 9.40052815945122e-07, "loss": 0.1796, "step": 20360 }, { "epoch": 0.39941176470588236, "grad_norm": 1.5359055995941162, "learning_rate": 9.399758289987924e-07, "loss": 0.1698, "step": 20370 }, { "epoch": 0.3996078431372549, "grad_norm": 2.453338146209717, "learning_rate": 9.398987958057356e-07, "loss": 0.144, "step": 20380 }, { "epoch": 0.39980392156862743, "grad_norm": 1.8340368270874023, "learning_rate": 9.398217163740485e-07, "loss": 0.1453, "step": 20390 }, { "epoch": 0.4, "grad_norm": 1.605747938156128, "learning_rate": 9.397445907118331e-07, "loss": 0.1466, "step": 20400 }, { "epoch": 0.40019607843137256, "grad_norm": 2.810354709625244, "learning_rate": 9.39667418827196e-07, "loss": 0.1449, "step": 20410 }, { "epoch": 0.4003921568627451, "grad_norm": 2.8670828342437744, "learning_rate": 9.395902007282494e-07, "loss": 0.172, "step": 20420 }, { "epoch": 0.40058823529411763, "grad_norm": 3.747004508972168, "learning_rate": 9.395129364231094e-07, "loss": 0.1798, "step": 20430 }, { "epoch": 0.40078431372549017, "grad_norm": 1.9854226112365723, "learning_rate": 9.394356259198976e-07, "loss": 0.1599, "step": 20440 }, { "epoch": 0.40098039215686276, "grad_norm": 2.149052858352661, "learning_rate": 9.393582692267401e-07, "loss": 0.1578, "step": 20450 }, { "epoch": 0.4011764705882353, "grad_norm": 3.3811655044555664, "learning_rate": 9.392808663517682e-07, "loss": 0.1318, "step": 20460 }, { "epoch": 0.40137254901960784, "grad_norm": 2.86049222946167, "learning_rate": 9.392034173031178e-07, "loss": 0.1826, "step": 20470 }, { "epoch": 0.4015686274509804, "grad_norm": 4.804776191711426, "learning_rate": 9.391259220889298e-07, "loss": 0.1682, "step": 20480 }, { "epoch": 0.40176470588235297, "grad_norm": 2.6175425052642822, "learning_rate": 9.390483807173497e-07, "loss": 0.171, "step": 20490 }, { "epoch": 0.4019607843137255, "grad_norm": 3.183725595474243, "learning_rate": 9.389707931965281e-07, "loss": 0.1786, "step": 20500 }, { "epoch": 0.40215686274509804, "grad_norm": 2.440281391143799, "learning_rate": 9.388931595346206e-07, "loss": 0.1777, "step": 20510 }, { "epoch": 0.4023529411764706, "grad_norm": 2.870927095413208, "learning_rate": 9.38815479739787e-07, "loss": 0.1829, "step": 20520 }, { "epoch": 0.4025490196078431, "grad_norm": 2.0948281288146973, "learning_rate": 9.387377538201929e-07, "loss": 0.1784, "step": 20530 }, { "epoch": 0.4027450980392157, "grad_norm": 5.189283847808838, "learning_rate": 9.386599817840076e-07, "loss": 0.1983, "step": 20540 }, { "epoch": 0.40294117647058825, "grad_norm": 1.7753307819366455, "learning_rate": 9.385821636394062e-07, "loss": 0.1537, "step": 20550 }, { "epoch": 0.4031372549019608, "grad_norm": 2.7938148975372314, "learning_rate": 9.385042993945684e-07, "loss": 0.1596, "step": 20560 }, { "epoch": 0.4033333333333333, "grad_norm": 3.8227860927581787, "learning_rate": 9.384263890576787e-07, "loss": 0.1905, "step": 20570 }, { "epoch": 0.40352941176470586, "grad_norm": 2.4829607009887695, "learning_rate": 9.383484326369262e-07, "loss": 0.1504, "step": 20580 }, { "epoch": 0.40372549019607845, "grad_norm": 2.2026824951171875, "learning_rate": 9.382704301405053e-07, "loss": 0.1573, "step": 20590 }, { "epoch": 0.403921568627451, "grad_norm": 3.4033544063568115, "learning_rate": 9.381923815766146e-07, "loss": 0.1804, "step": 20600 }, { "epoch": 0.4041176470588235, "grad_norm": 3.0987977981567383, "learning_rate": 9.381142869534584e-07, "loss": 0.1466, "step": 20610 }, { "epoch": 0.40431372549019606, "grad_norm": 2.9706788063049316, "learning_rate": 9.38036146279245e-07, "loss": 0.2016, "step": 20620 }, { "epoch": 0.4045098039215686, "grad_norm": 1.485044240951538, "learning_rate": 9.379579595621881e-07, "loss": 0.1924, "step": 20630 }, { "epoch": 0.4047058823529412, "grad_norm": 1.0902645587921143, "learning_rate": 9.378797268105061e-07, "loss": 0.1916, "step": 20640 }, { "epoch": 0.40490196078431373, "grad_norm": 3.5557429790496826, "learning_rate": 9.378014480324221e-07, "loss": 0.1578, "step": 20650 }, { "epoch": 0.40509803921568627, "grad_norm": 5.533471584320068, "learning_rate": 9.377231232361642e-07, "loss": 0.2028, "step": 20660 }, { "epoch": 0.4052941176470588, "grad_norm": 2.6551294326782227, "learning_rate": 9.376447524299652e-07, "loss": 0.1783, "step": 20670 }, { "epoch": 0.4054901960784314, "grad_norm": 2.9838616847991943, "learning_rate": 9.375663356220629e-07, "loss": 0.1644, "step": 20680 }, { "epoch": 0.40568627450980393, "grad_norm": 8.982877731323242, "learning_rate": 9.374878728206998e-07, "loss": 0.1713, "step": 20690 }, { "epoch": 0.40588235294117647, "grad_norm": 2.054863452911377, "learning_rate": 9.374093640341232e-07, "loss": 0.1455, "step": 20700 }, { "epoch": 0.406078431372549, "grad_norm": 1.34307062625885, "learning_rate": 9.373308092705856e-07, "loss": 0.1382, "step": 20710 }, { "epoch": 0.40627450980392155, "grad_norm": 1.5947449207305908, "learning_rate": 9.372522085383437e-07, "loss": 0.1459, "step": 20720 }, { "epoch": 0.40647058823529414, "grad_norm": 2.3110599517822266, "learning_rate": 9.371735618456595e-07, "loss": 0.1728, "step": 20730 }, { "epoch": 0.4066666666666667, "grad_norm": 2.2489447593688965, "learning_rate": 9.370948692007999e-07, "loss": 0.1613, "step": 20740 }, { "epoch": 0.4068627450980392, "grad_norm": 2.6151411533355713, "learning_rate": 9.370161306120362e-07, "loss": 0.182, "step": 20750 }, { "epoch": 0.40705882352941175, "grad_norm": 1.4880660772323608, "learning_rate": 9.369373460876449e-07, "loss": 0.1677, "step": 20760 }, { "epoch": 0.4072549019607843, "grad_norm": 2.30322265625, "learning_rate": 9.36858515635907e-07, "loss": 0.1434, "step": 20770 }, { "epoch": 0.4074509803921569, "grad_norm": 3.2742605209350586, "learning_rate": 9.367796392651086e-07, "loss": 0.1439, "step": 20780 }, { "epoch": 0.4076470588235294, "grad_norm": 1.8707994222640991, "learning_rate": 9.367007169835408e-07, "loss": 0.1508, "step": 20790 }, { "epoch": 0.40784313725490196, "grad_norm": 2.267015218734741, "learning_rate": 9.366217487994989e-07, "loss": 0.2018, "step": 20800 }, { "epoch": 0.4080392156862745, "grad_norm": 11.521477699279785, "learning_rate": 9.365427347212838e-07, "loss": 0.1826, "step": 20810 }, { "epoch": 0.4082352941176471, "grad_norm": 4.392251968383789, "learning_rate": 9.364636747572005e-07, "loss": 0.157, "step": 20820 }, { "epoch": 0.4084313725490196, "grad_norm": 1.6892868280410767, "learning_rate": 9.363845689155594e-07, "loss": 0.1623, "step": 20830 }, { "epoch": 0.40862745098039216, "grad_norm": 3.9831888675689697, "learning_rate": 9.363054172046751e-07, "loss": 0.1697, "step": 20840 }, { "epoch": 0.4088235294117647, "grad_norm": 2.8625476360321045, "learning_rate": 9.362262196328678e-07, "loss": 0.1848, "step": 20850 }, { "epoch": 0.40901960784313723, "grad_norm": 1.6932222843170166, "learning_rate": 9.36146976208462e-07, "loss": 0.1345, "step": 20860 }, { "epoch": 0.4092156862745098, "grad_norm": 2.6892430782318115, "learning_rate": 9.360676869397869e-07, "loss": 0.1434, "step": 20870 }, { "epoch": 0.40941176470588236, "grad_norm": 3.4918556213378906, "learning_rate": 9.35988351835177e-07, "loss": 0.1463, "step": 20880 }, { "epoch": 0.4096078431372549, "grad_norm": 3.1388537883758545, "learning_rate": 9.359089709029713e-07, "loss": 0.1455, "step": 20890 }, { "epoch": 0.40980392156862744, "grad_norm": 3.46244740486145, "learning_rate": 9.358295441515136e-07, "loss": 0.1633, "step": 20900 }, { "epoch": 0.41, "grad_norm": 2.145573139190674, "learning_rate": 9.357500715891529e-07, "loss": 0.1751, "step": 20910 }, { "epoch": 0.41019607843137257, "grad_norm": 3.776747226715088, "learning_rate": 9.356705532242422e-07, "loss": 0.1721, "step": 20920 }, { "epoch": 0.4103921568627451, "grad_norm": 3.174232244491577, "learning_rate": 9.355909890651403e-07, "loss": 0.1942, "step": 20930 }, { "epoch": 0.41058823529411764, "grad_norm": 1.8857003450393677, "learning_rate": 9.355113791202102e-07, "loss": 0.1672, "step": 20940 }, { "epoch": 0.4107843137254902, "grad_norm": 3.5731029510498047, "learning_rate": 9.354317233978199e-07, "loss": 0.1533, "step": 20950 }, { "epoch": 0.4109803921568628, "grad_norm": 5.642405033111572, "learning_rate": 9.35352021906342e-07, "loss": 0.1733, "step": 20960 }, { "epoch": 0.4111764705882353, "grad_norm": 4.111735820770264, "learning_rate": 9.352722746541541e-07, "loss": 0.1688, "step": 20970 }, { "epoch": 0.41137254901960785, "grad_norm": 1.8824149370193481, "learning_rate": 9.351924816496388e-07, "loss": 0.163, "step": 20980 }, { "epoch": 0.4115686274509804, "grad_norm": 2.3566360473632812, "learning_rate": 9.351126429011833e-07, "loss": 0.1794, "step": 20990 }, { "epoch": 0.4117647058823529, "grad_norm": 1.5881422758102417, "learning_rate": 9.350327584171794e-07, "loss": 0.1561, "step": 21000 }, { "epoch": 0.4119607843137255, "grad_norm": 1.9135240316390991, "learning_rate": 9.34952828206024e-07, "loss": 0.1816, "step": 21010 }, { "epoch": 0.41215686274509805, "grad_norm": 2.3458430767059326, "learning_rate": 9.348728522761186e-07, "loss": 0.1668, "step": 21020 }, { "epoch": 0.4123529411764706, "grad_norm": 4.808922290802002, "learning_rate": 9.347928306358699e-07, "loss": 0.1717, "step": 21030 }, { "epoch": 0.4125490196078431, "grad_norm": 3.906130075454712, "learning_rate": 9.347127632936888e-07, "loss": 0.1532, "step": 21040 }, { "epoch": 0.41274509803921566, "grad_norm": 2.8587069511413574, "learning_rate": 9.346326502579918e-07, "loss": 0.2102, "step": 21050 }, { "epoch": 0.41294117647058826, "grad_norm": 3.4995334148406982, "learning_rate": 9.345524915371993e-07, "loss": 0.1623, "step": 21060 }, { "epoch": 0.4131372549019608, "grad_norm": 2.526463508605957, "learning_rate": 9.344722871397371e-07, "loss": 0.1701, "step": 21070 }, { "epoch": 0.41333333333333333, "grad_norm": 2.745635986328125, "learning_rate": 9.343920370740357e-07, "loss": 0.1788, "step": 21080 }, { "epoch": 0.41352941176470587, "grad_norm": 3.377352476119995, "learning_rate": 9.343117413485302e-07, "loss": 0.1598, "step": 21090 }, { "epoch": 0.4137254901960784, "grad_norm": 2.742332696914673, "learning_rate": 9.342313999716609e-07, "loss": 0.1618, "step": 21100 }, { "epoch": 0.413921568627451, "grad_norm": 2.878074884414673, "learning_rate": 9.341510129518724e-07, "loss": 0.1816, "step": 21110 }, { "epoch": 0.41411764705882353, "grad_norm": 2.9448888301849365, "learning_rate": 9.340705802976143e-07, "loss": 0.1784, "step": 21120 }, { "epoch": 0.41431372549019607, "grad_norm": 2.7295618057250977, "learning_rate": 9.339901020173412e-07, "loss": 0.1783, "step": 21130 }, { "epoch": 0.4145098039215686, "grad_norm": 2.604792356491089, "learning_rate": 9.339095781195123e-07, "loss": 0.1807, "step": 21140 }, { "epoch": 0.4147058823529412, "grad_norm": 1.659531831741333, "learning_rate": 9.338290086125916e-07, "loss": 0.1646, "step": 21150 }, { "epoch": 0.41490196078431374, "grad_norm": 2.9718194007873535, "learning_rate": 9.33748393505048e-07, "loss": 0.1524, "step": 21160 }, { "epoch": 0.4150980392156863, "grad_norm": 1.5957567691802979, "learning_rate": 9.336677328053548e-07, "loss": 0.1796, "step": 21170 }, { "epoch": 0.4152941176470588, "grad_norm": 3.412355422973633, "learning_rate": 9.335870265219909e-07, "loss": 0.1909, "step": 21180 }, { "epoch": 0.41549019607843135, "grad_norm": 1.696632981300354, "learning_rate": 9.335062746634392e-07, "loss": 0.1806, "step": 21190 }, { "epoch": 0.41568627450980394, "grad_norm": 2.638367176055908, "learning_rate": 9.334254772381876e-07, "loss": 0.1443, "step": 21200 }, { "epoch": 0.4158823529411765, "grad_norm": 2.0758750438690186, "learning_rate": 9.333446342547291e-07, "loss": 0.159, "step": 21210 }, { "epoch": 0.416078431372549, "grad_norm": 2.9868528842926025, "learning_rate": 9.332637457215611e-07, "loss": 0.1625, "step": 21220 }, { "epoch": 0.41627450980392156, "grad_norm": 1.165798306465149, "learning_rate": 9.33182811647186e-07, "loss": 0.1751, "step": 21230 }, { "epoch": 0.4164705882352941, "grad_norm": 3.3799948692321777, "learning_rate": 9.33101832040111e-07, "loss": 0.1649, "step": 21240 }, { "epoch": 0.4166666666666667, "grad_norm": 2.064636468887329, "learning_rate": 9.330208069088481e-07, "loss": 0.1509, "step": 21250 }, { "epoch": 0.4168627450980392, "grad_norm": 2.3495872020721436, "learning_rate": 9.329397362619138e-07, "loss": 0.1934, "step": 21260 }, { "epoch": 0.41705882352941176, "grad_norm": 2.901149034500122, "learning_rate": 9.328586201078297e-07, "loss": 0.1728, "step": 21270 }, { "epoch": 0.4172549019607843, "grad_norm": 2.6672582626342773, "learning_rate": 9.32777458455122e-07, "loss": 0.1983, "step": 21280 }, { "epoch": 0.4174509803921569, "grad_norm": 1.734527349472046, "learning_rate": 9.326962513123218e-07, "loss": 0.1489, "step": 21290 }, { "epoch": 0.4176470588235294, "grad_norm": 1.8268369436264038, "learning_rate": 9.326149986879652e-07, "loss": 0.172, "step": 21300 }, { "epoch": 0.41784313725490196, "grad_norm": 2.132667064666748, "learning_rate": 9.325337005905925e-07, "loss": 0.1586, "step": 21310 }, { "epoch": 0.4180392156862745, "grad_norm": 2.588036060333252, "learning_rate": 9.32452357028749e-07, "loss": 0.1466, "step": 21320 }, { "epoch": 0.41823529411764704, "grad_norm": 2.0273373126983643, "learning_rate": 9.323709680109854e-07, "loss": 0.1964, "step": 21330 }, { "epoch": 0.41843137254901963, "grad_norm": 2.5874545574188232, "learning_rate": 9.322895335458563e-07, "loss": 0.1445, "step": 21340 }, { "epoch": 0.41862745098039217, "grad_norm": 2.297943115234375, "learning_rate": 9.322080536419212e-07, "loss": 0.1339, "step": 21350 }, { "epoch": 0.4188235294117647, "grad_norm": 1.578871250152588, "learning_rate": 9.321265283077451e-07, "loss": 0.1539, "step": 21360 }, { "epoch": 0.41901960784313724, "grad_norm": 1.7751213312149048, "learning_rate": 9.320449575518972e-07, "loss": 0.1546, "step": 21370 }, { "epoch": 0.4192156862745098, "grad_norm": 3.6914446353912354, "learning_rate": 9.319633413829513e-07, "loss": 0.1558, "step": 21380 }, { "epoch": 0.4194117647058824, "grad_norm": 2.763582229614258, "learning_rate": 9.318816798094864e-07, "loss": 0.154, "step": 21390 }, { "epoch": 0.4196078431372549, "grad_norm": 3.2682957649230957, "learning_rate": 9.31799972840086e-07, "loss": 0.1574, "step": 21400 }, { "epoch": 0.41980392156862745, "grad_norm": 3.363485336303711, "learning_rate": 9.317182204833387e-07, "loss": 0.1598, "step": 21410 }, { "epoch": 0.42, "grad_norm": 3.381268262863159, "learning_rate": 9.316364227478375e-07, "loss": 0.1481, "step": 21420 }, { "epoch": 0.4201960784313725, "grad_norm": 2.449537754058838, "learning_rate": 9.315545796421803e-07, "loss": 0.1921, "step": 21430 }, { "epoch": 0.4203921568627451, "grad_norm": 2.262160539627075, "learning_rate": 9.314726911749699e-07, "loss": 0.1532, "step": 21440 }, { "epoch": 0.42058823529411765, "grad_norm": 1.5897060632705688, "learning_rate": 9.313907573548137e-07, "loss": 0.1436, "step": 21450 }, { "epoch": 0.4207843137254902, "grad_norm": 3.036552667617798, "learning_rate": 9.313087781903239e-07, "loss": 0.1838, "step": 21460 }, { "epoch": 0.4209803921568627, "grad_norm": 2.9093472957611084, "learning_rate": 9.312267536901175e-07, "loss": 0.1909, "step": 21470 }, { "epoch": 0.4211764705882353, "grad_norm": 2.3647687435150146, "learning_rate": 9.311446838628163e-07, "loss": 0.1674, "step": 21480 }, { "epoch": 0.42137254901960786, "grad_norm": 1.8285959959030151, "learning_rate": 9.310625687170468e-07, "loss": 0.1803, "step": 21490 }, { "epoch": 0.4215686274509804, "grad_norm": 1.8851724863052368, "learning_rate": 9.309804082614404e-07, "loss": 0.1613, "step": 21500 }, { "epoch": 0.42176470588235293, "grad_norm": 5.110695838928223, "learning_rate": 9.308982025046327e-07, "loss": 0.1563, "step": 21510 }, { "epoch": 0.42196078431372547, "grad_norm": 3.397195339202881, "learning_rate": 9.308159514552651e-07, "loss": 0.1894, "step": 21520 }, { "epoch": 0.42215686274509806, "grad_norm": 6.766827583312988, "learning_rate": 9.307336551219827e-07, "loss": 0.175, "step": 21530 }, { "epoch": 0.4223529411764706, "grad_norm": 2.643519401550293, "learning_rate": 9.306513135134362e-07, "loss": 0.1723, "step": 21540 }, { "epoch": 0.42254901960784313, "grad_norm": 3.189241647720337, "learning_rate": 9.305689266382805e-07, "loss": 0.1796, "step": 21550 }, { "epoch": 0.42274509803921567, "grad_norm": 2.214768409729004, "learning_rate": 9.304864945051754e-07, "loss": 0.1652, "step": 21560 }, { "epoch": 0.4229411764705882, "grad_norm": 3.3105456829071045, "learning_rate": 9.304040171227856e-07, "loss": 0.1687, "step": 21570 }, { "epoch": 0.4231372549019608, "grad_norm": 2.1851212978363037, "learning_rate": 9.303214944997805e-07, "loss": 0.1589, "step": 21580 }, { "epoch": 0.42333333333333334, "grad_norm": 2.870166540145874, "learning_rate": 9.302389266448341e-07, "loss": 0.179, "step": 21590 }, { "epoch": 0.4235294117647059, "grad_norm": 2.8401875495910645, "learning_rate": 9.301563135666253e-07, "loss": 0.1536, "step": 21600 }, { "epoch": 0.4237254901960784, "grad_norm": 3.1666927337646484, "learning_rate": 9.300736552738377e-07, "loss": 0.1597, "step": 21610 }, { "epoch": 0.423921568627451, "grad_norm": 2.9062981605529785, "learning_rate": 9.299909517751599e-07, "loss": 0.178, "step": 21620 }, { "epoch": 0.42411764705882354, "grad_norm": 1.6487419605255127, "learning_rate": 9.299082030792846e-07, "loss": 0.1611, "step": 21630 }, { "epoch": 0.4243137254901961, "grad_norm": 2.179795742034912, "learning_rate": 9.298254091949101e-07, "loss": 0.147, "step": 21640 }, { "epoch": 0.4245098039215686, "grad_norm": 3.3724968433380127, "learning_rate": 9.297425701307389e-07, "loss": 0.1612, "step": 21650 }, { "epoch": 0.42470588235294116, "grad_norm": 1.9622832536697388, "learning_rate": 9.296596858954782e-07, "loss": 0.1511, "step": 21660 }, { "epoch": 0.42490196078431375, "grad_norm": 2.872915029525757, "learning_rate": 9.295767564978403e-07, "loss": 0.1702, "step": 21670 }, { "epoch": 0.4250980392156863, "grad_norm": 1.9804270267486572, "learning_rate": 9.29493781946542e-07, "loss": 0.1584, "step": 21680 }, { "epoch": 0.4252941176470588, "grad_norm": 2.2880635261535645, "learning_rate": 9.294107622503051e-07, "loss": 0.1725, "step": 21690 }, { "epoch": 0.42549019607843136, "grad_norm": 3.7739148139953613, "learning_rate": 9.293276974178557e-07, "loss": 0.1654, "step": 21700 }, { "epoch": 0.4256862745098039, "grad_norm": 2.0780155658721924, "learning_rate": 9.29244587457925e-07, "loss": 0.1618, "step": 21710 }, { "epoch": 0.4258823529411765, "grad_norm": 2.484346866607666, "learning_rate": 9.291614323792489e-07, "loss": 0.1729, "step": 21720 }, { "epoch": 0.426078431372549, "grad_norm": 2.781574249267578, "learning_rate": 9.29078232190568e-07, "loss": 0.1553, "step": 21730 }, { "epoch": 0.42627450980392156, "grad_norm": 3.8960673809051514, "learning_rate": 9.289949869006275e-07, "loss": 0.1778, "step": 21740 }, { "epoch": 0.4264705882352941, "grad_norm": 2.388864278793335, "learning_rate": 9.289116965181777e-07, "loss": 0.188, "step": 21750 }, { "epoch": 0.4266666666666667, "grad_norm": 3.181873083114624, "learning_rate": 9.288283610519732e-07, "loss": 0.1767, "step": 21760 }, { "epoch": 0.42686274509803923, "grad_norm": 3.173170328140259, "learning_rate": 9.287449805107737e-07, "loss": 0.2187, "step": 21770 }, { "epoch": 0.42705882352941177, "grad_norm": 2.4756710529327393, "learning_rate": 9.286615549033435e-07, "loss": 0.1741, "step": 21780 }, { "epoch": 0.4272549019607843, "grad_norm": 3.678098201751709, "learning_rate": 9.285780842384515e-07, "loss": 0.1708, "step": 21790 }, { "epoch": 0.42745098039215684, "grad_norm": 3.307929515838623, "learning_rate": 9.284945685248716e-07, "loss": 0.1761, "step": 21800 }, { "epoch": 0.42764705882352944, "grad_norm": 1.4677321910858154, "learning_rate": 9.284110077713822e-07, "loss": 0.1688, "step": 21810 }, { "epoch": 0.427843137254902, "grad_norm": 2.6666419506073, "learning_rate": 9.283274019867666e-07, "loss": 0.1653, "step": 21820 }, { "epoch": 0.4280392156862745, "grad_norm": 3.7756168842315674, "learning_rate": 9.282437511798127e-07, "loss": 0.1664, "step": 21830 }, { "epoch": 0.42823529411764705, "grad_norm": 3.0504207611083984, "learning_rate": 9.281600553593132e-07, "loss": 0.1852, "step": 21840 }, { "epoch": 0.4284313725490196, "grad_norm": 1.6800938844680786, "learning_rate": 9.280763145340657e-07, "loss": 0.1344, "step": 21850 }, { "epoch": 0.4286274509803922, "grad_norm": 2.538555383682251, "learning_rate": 9.279925287128723e-07, "loss": 0.1483, "step": 21860 }, { "epoch": 0.4288235294117647, "grad_norm": 2.693547010421753, "learning_rate": 9.279086979045396e-07, "loss": 0.1956, "step": 21870 }, { "epoch": 0.42901960784313725, "grad_norm": 1.5507091283798218, "learning_rate": 9.278248221178797e-07, "loss": 0.1726, "step": 21880 }, { "epoch": 0.4292156862745098, "grad_norm": 2.2167513370513916, "learning_rate": 9.277409013617085e-07, "loss": 0.1767, "step": 21890 }, { "epoch": 0.4294117647058823, "grad_norm": 2.445352554321289, "learning_rate": 9.276569356448475e-07, "loss": 0.1707, "step": 21900 }, { "epoch": 0.4296078431372549, "grad_norm": 2.5628950595855713, "learning_rate": 9.275729249761222e-07, "loss": 0.1925, "step": 21910 }, { "epoch": 0.42980392156862746, "grad_norm": 1.8141348361968994, "learning_rate": 9.274888693643632e-07, "loss": 0.169, "step": 21920 }, { "epoch": 0.43, "grad_norm": 2.9966185092926025, "learning_rate": 9.274047688184059e-07, "loss": 0.194, "step": 21930 }, { "epoch": 0.43019607843137253, "grad_norm": 4.301174640655518, "learning_rate": 9.273206233470899e-07, "loss": 0.1548, "step": 21940 }, { "epoch": 0.4303921568627451, "grad_norm": 1.8273216485977173, "learning_rate": 9.272364329592604e-07, "loss": 0.1785, "step": 21950 }, { "epoch": 0.43058823529411766, "grad_norm": 3.6464686393737793, "learning_rate": 9.271521976637664e-07, "loss": 0.1866, "step": 21960 }, { "epoch": 0.4307843137254902, "grad_norm": 2.0255236625671387, "learning_rate": 9.270679174694622e-07, "loss": 0.1769, "step": 21970 }, { "epoch": 0.43098039215686273, "grad_norm": 1.8653080463409424, "learning_rate": 9.269835923852065e-07, "loss": 0.166, "step": 21980 }, { "epoch": 0.43117647058823527, "grad_norm": 3.855792999267578, "learning_rate": 9.268992224198633e-07, "loss": 0.164, "step": 21990 }, { "epoch": 0.43137254901960786, "grad_norm": 4.508403778076172, "learning_rate": 9.268148075823004e-07, "loss": 0.1481, "step": 22000 }, { "epoch": 0.4315686274509804, "grad_norm": 4.076746463775635, "learning_rate": 9.267303478813912e-07, "loss": 0.155, "step": 22010 }, { "epoch": 0.43176470588235294, "grad_norm": 1.4850783348083496, "learning_rate": 9.266458433260131e-07, "loss": 0.1399, "step": 22020 }, { "epoch": 0.4319607843137255, "grad_norm": 2.6378915309906006, "learning_rate": 9.265612939250488e-07, "loss": 0.1722, "step": 22030 }, { "epoch": 0.432156862745098, "grad_norm": 2.5938143730163574, "learning_rate": 9.264766996873854e-07, "loss": 0.183, "step": 22040 }, { "epoch": 0.4323529411764706, "grad_norm": 2.695058822631836, "learning_rate": 9.263920606219147e-07, "loss": 0.1412, "step": 22050 }, { "epoch": 0.43254901960784314, "grad_norm": 1.8530274629592896, "learning_rate": 9.263073767375331e-07, "loss": 0.1841, "step": 22060 }, { "epoch": 0.4327450980392157, "grad_norm": 2.3604934215545654, "learning_rate": 9.262226480431424e-07, "loss": 0.1691, "step": 22070 }, { "epoch": 0.4329411764705882, "grad_norm": 2.7879557609558105, "learning_rate": 9.261378745476481e-07, "loss": 0.1604, "step": 22080 }, { "epoch": 0.4331372549019608, "grad_norm": 2.279524564743042, "learning_rate": 9.260530562599611e-07, "loss": 0.177, "step": 22090 }, { "epoch": 0.43333333333333335, "grad_norm": 2.8312814235687256, "learning_rate": 9.259681931889968e-07, "loss": 0.2048, "step": 22100 }, { "epoch": 0.4335294117647059, "grad_norm": 2.650791645050049, "learning_rate": 9.258832853436753e-07, "loss": 0.1589, "step": 22110 }, { "epoch": 0.4337254901960784, "grad_norm": 4.411615371704102, "learning_rate": 9.257983327329216e-07, "loss": 0.1733, "step": 22120 }, { "epoch": 0.43392156862745096, "grad_norm": 1.4521484375, "learning_rate": 9.257133353656649e-07, "loss": 0.1754, "step": 22130 }, { "epoch": 0.43411764705882355, "grad_norm": 2.296584129333496, "learning_rate": 9.256282932508399e-07, "loss": 0.1288, "step": 22140 }, { "epoch": 0.4343137254901961, "grad_norm": 2.3468430042266846, "learning_rate": 9.255432063973853e-07, "loss": 0.1313, "step": 22150 }, { "epoch": 0.4345098039215686, "grad_norm": 2.889188289642334, "learning_rate": 9.254580748142447e-07, "loss": 0.178, "step": 22160 }, { "epoch": 0.43470588235294116, "grad_norm": 3.0211944580078125, "learning_rate": 9.253728985103664e-07, "loss": 0.1545, "step": 22170 }, { "epoch": 0.4349019607843137, "grad_norm": 1.7788277864456177, "learning_rate": 9.252876774947035e-07, "loss": 0.1576, "step": 22180 }, { "epoch": 0.4350980392156863, "grad_norm": 3.8082799911499023, "learning_rate": 9.252024117762138e-07, "loss": 0.1834, "step": 22190 }, { "epoch": 0.43529411764705883, "grad_norm": 3.5232555866241455, "learning_rate": 9.251171013638598e-07, "loss": 0.1392, "step": 22200 }, { "epoch": 0.43549019607843137, "grad_norm": 2.342480421066284, "learning_rate": 9.250317462666085e-07, "loss": 0.1692, "step": 22210 }, { "epoch": 0.4356862745098039, "grad_norm": 2.302021026611328, "learning_rate": 9.24946346493432e-07, "loss": 0.1253, "step": 22220 }, { "epoch": 0.43588235294117644, "grad_norm": 2.7750003337860107, "learning_rate": 9.248609020533065e-07, "loss": 0.1587, "step": 22230 }, { "epoch": 0.43607843137254904, "grad_norm": 6.826268196105957, "learning_rate": 9.247754129552134e-07, "loss": 0.1734, "step": 22240 }, { "epoch": 0.4362745098039216, "grad_norm": 2.448964834213257, "learning_rate": 9.246898792081386e-07, "loss": 0.179, "step": 22250 }, { "epoch": 0.4364705882352941, "grad_norm": 2.5147695541381836, "learning_rate": 9.246043008210729e-07, "loss": 0.1433, "step": 22260 }, { "epoch": 0.43666666666666665, "grad_norm": 16.055950164794922, "learning_rate": 9.245186778030114e-07, "loss": 0.1538, "step": 22270 }, { "epoch": 0.43686274509803924, "grad_norm": 2.1411736011505127, "learning_rate": 9.244330101629541e-07, "loss": 0.1661, "step": 22280 }, { "epoch": 0.4370588235294118, "grad_norm": 2.2429590225219727, "learning_rate": 9.243472979099057e-07, "loss": 0.1708, "step": 22290 }, { "epoch": 0.4372549019607843, "grad_norm": 3.9071826934814453, "learning_rate": 9.242615410528758e-07, "loss": 0.175, "step": 22300 }, { "epoch": 0.43745098039215685, "grad_norm": 1.7342981100082397, "learning_rate": 9.241757396008782e-07, "loss": 0.1526, "step": 22310 }, { "epoch": 0.4376470588235294, "grad_norm": 2.5802195072174072, "learning_rate": 9.240898935629318e-07, "loss": 0.159, "step": 22320 }, { "epoch": 0.437843137254902, "grad_norm": 7.408921241760254, "learning_rate": 9.2400400294806e-07, "loss": 0.166, "step": 22330 }, { "epoch": 0.4380392156862745, "grad_norm": 3.0129830837249756, "learning_rate": 9.239180677652909e-07, "loss": 0.1826, "step": 22340 }, { "epoch": 0.43823529411764706, "grad_norm": 3.8063807487487793, "learning_rate": 9.238320880236575e-07, "loss": 0.1491, "step": 22350 }, { "epoch": 0.4384313725490196, "grad_norm": 3.772390842437744, "learning_rate": 9.237460637321972e-07, "loss": 0.169, "step": 22360 }, { "epoch": 0.43862745098039213, "grad_norm": 2.1623587608337402, "learning_rate": 9.236599948999519e-07, "loss": 0.1608, "step": 22370 }, { "epoch": 0.4388235294117647, "grad_norm": 3.141451358795166, "learning_rate": 9.235738815359689e-07, "loss": 0.207, "step": 22380 }, { "epoch": 0.43901960784313726, "grad_norm": 2.599355936050415, "learning_rate": 9.234877236492996e-07, "loss": 0.179, "step": 22390 }, { "epoch": 0.4392156862745098, "grad_norm": 2.3716607093811035, "learning_rate": 9.234015212490001e-07, "loss": 0.1722, "step": 22400 }, { "epoch": 0.43941176470588234, "grad_norm": 3.280579090118408, "learning_rate": 9.233152743441314e-07, "loss": 0.1559, "step": 22410 }, { "epoch": 0.43960784313725493, "grad_norm": 1.317641258239746, "learning_rate": 9.232289829437591e-07, "loss": 0.1752, "step": 22420 }, { "epoch": 0.43980392156862746, "grad_norm": 2.25495982170105, "learning_rate": 9.231426470569534e-07, "loss": 0.1646, "step": 22430 }, { "epoch": 0.44, "grad_norm": 1.4769701957702637, "learning_rate": 9.230562666927893e-07, "loss": 0.1536, "step": 22440 }, { "epoch": 0.44019607843137254, "grad_norm": 1.3845436573028564, "learning_rate": 9.229698418603465e-07, "loss": 0.1914, "step": 22450 }, { "epoch": 0.4403921568627451, "grad_norm": 2.1117966175079346, "learning_rate": 9.228833725687088e-07, "loss": 0.1751, "step": 22460 }, { "epoch": 0.44058823529411767, "grad_norm": 2.3616724014282227, "learning_rate": 9.227968588269659e-07, "loss": 0.1688, "step": 22470 }, { "epoch": 0.4407843137254902, "grad_norm": 2.7182705402374268, "learning_rate": 9.227103006442108e-07, "loss": 0.153, "step": 22480 }, { "epoch": 0.44098039215686274, "grad_norm": 4.925293922424316, "learning_rate": 9.226236980295421e-07, "loss": 0.2009, "step": 22490 }, { "epoch": 0.4411764705882353, "grad_norm": 2.38887357711792, "learning_rate": 9.225370509920629e-07, "loss": 0.1739, "step": 22500 }, { "epoch": 0.4413725490196078, "grad_norm": 2.5383551120758057, "learning_rate": 9.224503595408805e-07, "loss": 0.1967, "step": 22510 }, { "epoch": 0.4415686274509804, "grad_norm": 2.4769177436828613, "learning_rate": 9.223636236851074e-07, "loss": 0.1701, "step": 22520 }, { "epoch": 0.44176470588235295, "grad_norm": 2.543577194213867, "learning_rate": 9.222768434338606e-07, "loss": 0.2045, "step": 22530 }, { "epoch": 0.4419607843137255, "grad_norm": 2.568244695663452, "learning_rate": 9.221900187962616e-07, "loss": 0.1693, "step": 22540 }, { "epoch": 0.442156862745098, "grad_norm": 2.8657383918762207, "learning_rate": 9.221031497814369e-07, "loss": 0.1991, "step": 22550 }, { "epoch": 0.4423529411764706, "grad_norm": 3.7963974475860596, "learning_rate": 9.220162363985173e-07, "loss": 0.1819, "step": 22560 }, { "epoch": 0.44254901960784315, "grad_norm": 2.310288429260254, "learning_rate": 9.219292786566385e-07, "loss": 0.1609, "step": 22570 }, { "epoch": 0.4427450980392157, "grad_norm": 2.5056796073913574, "learning_rate": 9.21842276564941e-07, "loss": 0.1755, "step": 22580 }, { "epoch": 0.4429411764705882, "grad_norm": 2.3645849227905273, "learning_rate": 9.217552301325693e-07, "loss": 0.1631, "step": 22590 }, { "epoch": 0.44313725490196076, "grad_norm": 2.496903896331787, "learning_rate": 9.216681393686734e-07, "loss": 0.2052, "step": 22600 }, { "epoch": 0.44333333333333336, "grad_norm": 3.7321629524230957, "learning_rate": 9.215810042824076e-07, "loss": 0.1393, "step": 22610 }, { "epoch": 0.4435294117647059, "grad_norm": 3.09440016746521, "learning_rate": 9.214938248829307e-07, "loss": 0.148, "step": 22620 }, { "epoch": 0.44372549019607843, "grad_norm": 3.3803319931030273, "learning_rate": 9.214066011794061e-07, "loss": 0.1735, "step": 22630 }, { "epoch": 0.44392156862745097, "grad_norm": 3.535640001296997, "learning_rate": 9.213193331810025e-07, "loss": 0.1767, "step": 22640 }, { "epoch": 0.4441176470588235, "grad_norm": 2.448988437652588, "learning_rate": 9.212320208968924e-07, "loss": 0.2296, "step": 22650 }, { "epoch": 0.4443137254901961, "grad_norm": 2.98779296875, "learning_rate": 9.211446643362538e-07, "loss": 0.1857, "step": 22660 }, { "epoch": 0.44450980392156864, "grad_norm": 2.9839189052581787, "learning_rate": 9.210572635082685e-07, "loss": 0.1461, "step": 22670 }, { "epoch": 0.4447058823529412, "grad_norm": 2.533306360244751, "learning_rate": 9.209698184221236e-07, "loss": 0.1831, "step": 22680 }, { "epoch": 0.4449019607843137, "grad_norm": 2.5007386207580566, "learning_rate": 9.208823290870106e-07, "loss": 0.1785, "step": 22690 }, { "epoch": 0.44509803921568625, "grad_norm": 2.258082151412964, "learning_rate": 9.207947955121255e-07, "loss": 0.1537, "step": 22700 }, { "epoch": 0.44529411764705884, "grad_norm": 2.8153929710388184, "learning_rate": 9.207072177066695e-07, "loss": 0.1663, "step": 22710 }, { "epoch": 0.4454901960784314, "grad_norm": 2.812551259994507, "learning_rate": 9.206195956798477e-07, "loss": 0.1537, "step": 22720 }, { "epoch": 0.4456862745098039, "grad_norm": 1.5918962955474854, "learning_rate": 9.205319294408704e-07, "loss": 0.1457, "step": 22730 }, { "epoch": 0.44588235294117645, "grad_norm": 2.4046318531036377, "learning_rate": 9.204442189989524e-07, "loss": 0.142, "step": 22740 }, { "epoch": 0.44607843137254904, "grad_norm": 3.1064672470092773, "learning_rate": 9.20356464363313e-07, "loss": 0.1513, "step": 22750 }, { "epoch": 0.4462745098039216, "grad_norm": 1.0023839473724365, "learning_rate": 9.202686655431764e-07, "loss": 0.1625, "step": 22760 }, { "epoch": 0.4464705882352941, "grad_norm": 1.6906144618988037, "learning_rate": 9.201808225477712e-07, "loss": 0.1446, "step": 22770 }, { "epoch": 0.44666666666666666, "grad_norm": 2.982560396194458, "learning_rate": 9.200929353863308e-07, "loss": 0.1783, "step": 22780 }, { "epoch": 0.4468627450980392, "grad_norm": 2.142914056777954, "learning_rate": 9.200050040680931e-07, "loss": 0.1713, "step": 22790 }, { "epoch": 0.4470588235294118, "grad_norm": 2.43640398979187, "learning_rate": 9.199170286023008e-07, "loss": 0.1954, "step": 22800 }, { "epoch": 0.4472549019607843, "grad_norm": 3.5785884857177734, "learning_rate": 9.198290089982013e-07, "loss": 0.1758, "step": 22810 }, { "epoch": 0.44745098039215686, "grad_norm": 1.7915781736373901, "learning_rate": 9.197409452650464e-07, "loss": 0.1661, "step": 22820 }, { "epoch": 0.4476470588235294, "grad_norm": 2.327780246734619, "learning_rate": 9.196528374120925e-07, "loss": 0.1796, "step": 22830 }, { "epoch": 0.44784313725490194, "grad_norm": 3.194242477416992, "learning_rate": 9.195646854486011e-07, "loss": 0.1817, "step": 22840 }, { "epoch": 0.44803921568627453, "grad_norm": 1.5915757417678833, "learning_rate": 9.194764893838378e-07, "loss": 0.1414, "step": 22850 }, { "epoch": 0.44823529411764707, "grad_norm": 3.8822312355041504, "learning_rate": 9.193882492270731e-07, "loss": 0.1851, "step": 22860 }, { "epoch": 0.4484313725490196, "grad_norm": 2.006378173828125, "learning_rate": 9.192999649875822e-07, "loss": 0.136, "step": 22870 }, { "epoch": 0.44862745098039214, "grad_norm": 2.7618508338928223, "learning_rate": 9.192116366746448e-07, "loss": 0.1767, "step": 22880 }, { "epoch": 0.44882352941176473, "grad_norm": 1.69658625125885, "learning_rate": 9.191232642975451e-07, "loss": 0.1309, "step": 22890 }, { "epoch": 0.44901960784313727, "grad_norm": 2.497910499572754, "learning_rate": 9.190348478655723e-07, "loss": 0.1836, "step": 22900 }, { "epoch": 0.4492156862745098, "grad_norm": 2.8470065593719482, "learning_rate": 9.1894638738802e-07, "loss": 0.1778, "step": 22910 }, { "epoch": 0.44941176470588234, "grad_norm": 2.3400049209594727, "learning_rate": 9.188578828741863e-07, "loss": 0.1543, "step": 22920 }, { "epoch": 0.4496078431372549, "grad_norm": 2.7571728229522705, "learning_rate": 9.187693343333742e-07, "loss": 0.1861, "step": 22930 }, { "epoch": 0.4498039215686275, "grad_norm": 1.4554328918457031, "learning_rate": 9.186807417748912e-07, "loss": 0.1693, "step": 22940 }, { "epoch": 0.45, "grad_norm": 2.0646443367004395, "learning_rate": 9.185921052080495e-07, "loss": 0.1461, "step": 22950 }, { "epoch": 0.45019607843137255, "grad_norm": 2.2673027515411377, "learning_rate": 9.185034246421656e-07, "loss": 0.177, "step": 22960 }, { "epoch": 0.4503921568627451, "grad_norm": 5.7994585037231445, "learning_rate": 9.184147000865614e-07, "loss": 0.1764, "step": 22970 }, { "epoch": 0.4505882352941176, "grad_norm": 1.8624571561813354, "learning_rate": 9.183259315505625e-07, "loss": 0.1806, "step": 22980 }, { "epoch": 0.4507843137254902, "grad_norm": 2.2104134559631348, "learning_rate": 9.182371190434994e-07, "loss": 0.2043, "step": 22990 }, { "epoch": 0.45098039215686275, "grad_norm": 1.9225444793701172, "learning_rate": 9.181482625747079e-07, "loss": 0.1783, "step": 23000 }, { "epoch": 0.4511764705882353, "grad_norm": 2.3427839279174805, "learning_rate": 9.180593621535274e-07, "loss": 0.1516, "step": 23010 }, { "epoch": 0.4513725490196078, "grad_norm": 2.8337244987487793, "learning_rate": 9.179704177893027e-07, "loss": 0.1678, "step": 23020 }, { "epoch": 0.4515686274509804, "grad_norm": 3.515887498855591, "learning_rate": 9.178814294913826e-07, "loss": 0.166, "step": 23030 }, { "epoch": 0.45176470588235296, "grad_norm": 2.753659248352051, "learning_rate": 9.177923972691212e-07, "loss": 0.1644, "step": 23040 }, { "epoch": 0.4519607843137255, "grad_norm": 3.176373243331909, "learning_rate": 9.177033211318766e-07, "loss": 0.1441, "step": 23050 }, { "epoch": 0.45215686274509803, "grad_norm": 3.0322883129119873, "learning_rate": 9.176142010890118e-07, "loss": 0.1762, "step": 23060 }, { "epoch": 0.45235294117647057, "grad_norm": 1.6154632568359375, "learning_rate": 9.175250371498945e-07, "loss": 0.1475, "step": 23070 }, { "epoch": 0.45254901960784316, "grad_norm": 2.0903704166412354, "learning_rate": 9.174358293238969e-07, "loss": 0.1472, "step": 23080 }, { "epoch": 0.4527450980392157, "grad_norm": 3.2624711990356445, "learning_rate": 9.173465776203956e-07, "loss": 0.1724, "step": 23090 }, { "epoch": 0.45294117647058824, "grad_norm": 3.1536929607391357, "learning_rate": 9.172572820487722e-07, "loss": 0.1541, "step": 23100 }, { "epoch": 0.4531372549019608, "grad_norm": 1.9734611511230469, "learning_rate": 9.171679426184127e-07, "loss": 0.1673, "step": 23110 }, { "epoch": 0.4533333333333333, "grad_norm": 6.592508792877197, "learning_rate": 9.170785593387078e-07, "loss": 0.1899, "step": 23120 }, { "epoch": 0.4535294117647059, "grad_norm": 2.2028982639312744, "learning_rate": 9.169891322190526e-07, "loss": 0.1436, "step": 23130 }, { "epoch": 0.45372549019607844, "grad_norm": 2.658607244491577, "learning_rate": 9.168996612688471e-07, "loss": 0.1653, "step": 23140 }, { "epoch": 0.453921568627451, "grad_norm": 4.538134574890137, "learning_rate": 9.168101464974957e-07, "loss": 0.1755, "step": 23150 }, { "epoch": 0.4541176470588235, "grad_norm": 2.5554187297821045, "learning_rate": 9.167205879144075e-07, "loss": 0.1525, "step": 23160 }, { "epoch": 0.45431372549019605, "grad_norm": 3.2520899772644043, "learning_rate": 9.166309855289962e-07, "loss": 0.1638, "step": 23170 }, { "epoch": 0.45450980392156864, "grad_norm": 5.713288307189941, "learning_rate": 9.1654133935068e-07, "loss": 0.1483, "step": 23180 }, { "epoch": 0.4547058823529412, "grad_norm": 2.575212001800537, "learning_rate": 9.164516493888819e-07, "loss": 0.1729, "step": 23190 }, { "epoch": 0.4549019607843137, "grad_norm": 2.521440267562866, "learning_rate": 9.163619156530293e-07, "loss": 0.167, "step": 23200 }, { "epoch": 0.45509803921568626, "grad_norm": 3.694328546524048, "learning_rate": 9.162721381525542e-07, "loss": 0.1671, "step": 23210 }, { "epoch": 0.45529411764705885, "grad_norm": 3.4887356758117676, "learning_rate": 9.161823168968937e-07, "loss": 0.1371, "step": 23220 }, { "epoch": 0.4554901960784314, "grad_norm": 1.5723356008529663, "learning_rate": 9.160924518954886e-07, "loss": 0.1523, "step": 23230 }, { "epoch": 0.4556862745098039, "grad_norm": 2.0711848735809326, "learning_rate": 9.16002543157785e-07, "loss": 0.1733, "step": 23240 }, { "epoch": 0.45588235294117646, "grad_norm": 2.5758180618286133, "learning_rate": 9.159125906932335e-07, "loss": 0.1694, "step": 23250 }, { "epoch": 0.456078431372549, "grad_norm": 2.183934450149536, "learning_rate": 9.15822594511289e-07, "loss": 0.1431, "step": 23260 }, { "epoch": 0.4562745098039216, "grad_norm": 1.4307106733322144, "learning_rate": 9.157325546214112e-07, "loss": 0.1878, "step": 23270 }, { "epoch": 0.45647058823529413, "grad_norm": 2.157191753387451, "learning_rate": 9.156424710330643e-07, "loss": 0.1167, "step": 23280 }, { "epoch": 0.45666666666666667, "grad_norm": 1.3755992650985718, "learning_rate": 9.155523437557173e-07, "loss": 0.1706, "step": 23290 }, { "epoch": 0.4568627450980392, "grad_norm": 3.2840511798858643, "learning_rate": 9.154621727988438e-07, "loss": 0.1601, "step": 23300 }, { "epoch": 0.45705882352941174, "grad_norm": 2.2534239292144775, "learning_rate": 9.153719581719216e-07, "loss": 0.1566, "step": 23310 }, { "epoch": 0.45725490196078433, "grad_norm": 2.3752129077911377, "learning_rate": 9.152816998844333e-07, "loss": 0.1369, "step": 23320 }, { "epoch": 0.45745098039215687, "grad_norm": 3.1411099433898926, "learning_rate": 9.151913979458663e-07, "loss": 0.1775, "step": 23330 }, { "epoch": 0.4576470588235294, "grad_norm": 3.689465284347534, "learning_rate": 9.151010523657123e-07, "loss": 0.1681, "step": 23340 }, { "epoch": 0.45784313725490194, "grad_norm": 2.0997629165649414, "learning_rate": 9.150106631534679e-07, "loss": 0.2147, "step": 23350 }, { "epoch": 0.45803921568627454, "grad_norm": 3.924826145172119, "learning_rate": 9.149202303186338e-07, "loss": 0.151, "step": 23360 }, { "epoch": 0.4582352941176471, "grad_norm": 3.4135448932647705, "learning_rate": 9.148297538707157e-07, "loss": 0.1747, "step": 23370 }, { "epoch": 0.4584313725490196, "grad_norm": 2.8467626571655273, "learning_rate": 9.147392338192239e-07, "loss": 0.2039, "step": 23380 }, { "epoch": 0.45862745098039215, "grad_norm": 1.9932317733764648, "learning_rate": 9.146486701736729e-07, "loss": 0.1985, "step": 23390 }, { "epoch": 0.4588235294117647, "grad_norm": 2.2840616703033447, "learning_rate": 9.145580629435823e-07, "loss": 0.1522, "step": 23400 }, { "epoch": 0.4590196078431373, "grad_norm": 2.3753535747528076, "learning_rate": 9.144674121384756e-07, "loss": 0.1757, "step": 23410 }, { "epoch": 0.4592156862745098, "grad_norm": 1.1443525552749634, "learning_rate": 9.143767177678817e-07, "loss": 0.1249, "step": 23420 }, { "epoch": 0.45941176470588235, "grad_norm": 1.5480291843414307, "learning_rate": 9.142859798413334e-07, "loss": 0.1451, "step": 23430 }, { "epoch": 0.4596078431372549, "grad_norm": 1.6122437715530396, "learning_rate": 9.141951983683684e-07, "loss": 0.1475, "step": 23440 }, { "epoch": 0.4598039215686274, "grad_norm": 2.833127975463867, "learning_rate": 9.14104373358529e-07, "loss": 0.1609, "step": 23450 }, { "epoch": 0.46, "grad_norm": 5.034513473510742, "learning_rate": 9.140135048213619e-07, "loss": 0.1432, "step": 23460 }, { "epoch": 0.46019607843137256, "grad_norm": 2.3132028579711914, "learning_rate": 9.139225927664187e-07, "loss": 0.1758, "step": 23470 }, { "epoch": 0.4603921568627451, "grad_norm": 2.3213906288146973, "learning_rate": 9.138316372032551e-07, "loss": 0.1546, "step": 23480 }, { "epoch": 0.46058823529411763, "grad_norm": 2.5739359855651855, "learning_rate": 9.137406381414316e-07, "loss": 0.1691, "step": 23490 }, { "epoch": 0.46078431372549017, "grad_norm": 1.8318675756454468, "learning_rate": 9.136495955905134e-07, "loss": 0.1671, "step": 23500 }, { "epoch": 0.46098039215686276, "grad_norm": 3.6552398204803467, "learning_rate": 9.135585095600701e-07, "loss": 0.1601, "step": 23510 }, { "epoch": 0.4611764705882353, "grad_norm": 2.957923173904419, "learning_rate": 9.13467380059676e-07, "loss": 0.1639, "step": 23520 }, { "epoch": 0.46137254901960784, "grad_norm": 3.9084019660949707, "learning_rate": 9.133762070989098e-07, "loss": 0.1523, "step": 23530 }, { "epoch": 0.4615686274509804, "grad_norm": 2.4471490383148193, "learning_rate": 9.132849906873549e-07, "loss": 0.1382, "step": 23540 }, { "epoch": 0.46176470588235297, "grad_norm": 2.8357648849487305, "learning_rate": 9.131937308345996e-07, "loss": 0.1695, "step": 23550 }, { "epoch": 0.4619607843137255, "grad_norm": 1.7780876159667969, "learning_rate": 9.131024275502358e-07, "loss": 0.1396, "step": 23560 }, { "epoch": 0.46215686274509804, "grad_norm": 4.326201438903809, "learning_rate": 9.13011080843861e-07, "loss": 0.172, "step": 23570 }, { "epoch": 0.4623529411764706, "grad_norm": 1.5729035139083862, "learning_rate": 9.129196907250764e-07, "loss": 0.1677, "step": 23580 }, { "epoch": 0.4625490196078431, "grad_norm": 2.069939613342285, "learning_rate": 9.128282572034887e-07, "loss": 0.1502, "step": 23590 }, { "epoch": 0.4627450980392157, "grad_norm": 2.8928356170654297, "learning_rate": 9.127367802887085e-07, "loss": 0.1913, "step": 23600 }, { "epoch": 0.46294117647058824, "grad_norm": 2.0025951862335205, "learning_rate": 9.126452599903509e-07, "loss": 0.144, "step": 23610 }, { "epoch": 0.4631372549019608, "grad_norm": 2.075535774230957, "learning_rate": 9.12553696318036e-07, "loss": 0.1805, "step": 23620 }, { "epoch": 0.4633333333333333, "grad_norm": 3.555325984954834, "learning_rate": 9.124620892813881e-07, "loss": 0.1691, "step": 23630 }, { "epoch": 0.46352941176470586, "grad_norm": 2.254519462585449, "learning_rate": 9.123704388900364e-07, "loss": 0.137, "step": 23640 }, { "epoch": 0.46372549019607845, "grad_norm": 4.9631123542785645, "learning_rate": 9.122787451536142e-07, "loss": 0.1702, "step": 23650 }, { "epoch": 0.463921568627451, "grad_norm": 0.9497820734977722, "learning_rate": 9.121870080817599e-07, "loss": 0.1406, "step": 23660 }, { "epoch": 0.4641176470588235, "grad_norm": 4.727434158325195, "learning_rate": 9.120952276841159e-07, "loss": 0.1621, "step": 23670 }, { "epoch": 0.46431372549019606, "grad_norm": 2.592251777648926, "learning_rate": 9.120034039703294e-07, "loss": 0.1405, "step": 23680 }, { "epoch": 0.46450980392156865, "grad_norm": 2.4626708030700684, "learning_rate": 9.119115369500523e-07, "loss": 0.185, "step": 23690 }, { "epoch": 0.4647058823529412, "grad_norm": 1.94853675365448, "learning_rate": 9.11819626632941e-07, "loss": 0.1675, "step": 23700 }, { "epoch": 0.46490196078431373, "grad_norm": 3.8486573696136475, "learning_rate": 9.117276730286563e-07, "loss": 0.1648, "step": 23710 }, { "epoch": 0.46509803921568627, "grad_norm": 5.0813798904418945, "learning_rate": 9.116356761468636e-07, "loss": 0.1692, "step": 23720 }, { "epoch": 0.4652941176470588, "grad_norm": 2.328591823577881, "learning_rate": 9.115436359972328e-07, "loss": 0.1725, "step": 23730 }, { "epoch": 0.4654901960784314, "grad_norm": 2.716559648513794, "learning_rate": 9.114515525894386e-07, "loss": 0.1599, "step": 23740 }, { "epoch": 0.46568627450980393, "grad_norm": 3.0804641246795654, "learning_rate": 9.113594259331598e-07, "loss": 0.1473, "step": 23750 }, { "epoch": 0.46588235294117647, "grad_norm": 3.021937847137451, "learning_rate": 9.112672560380803e-07, "loss": 0.1597, "step": 23760 }, { "epoch": 0.466078431372549, "grad_norm": 1.8742561340332031, "learning_rate": 9.111750429138883e-07, "loss": 0.1744, "step": 23770 }, { "epoch": 0.46627450980392154, "grad_norm": 4.683237075805664, "learning_rate": 9.110827865702761e-07, "loss": 0.1787, "step": 23780 }, { "epoch": 0.46647058823529414, "grad_norm": 2.1854636669158936, "learning_rate": 9.109904870169413e-07, "loss": 0.1779, "step": 23790 }, { "epoch": 0.4666666666666667, "grad_norm": 3.8022139072418213, "learning_rate": 9.108981442635857e-07, "loss": 0.1576, "step": 23800 }, { "epoch": 0.4668627450980392, "grad_norm": 2.2104408740997314, "learning_rate": 9.108057583199154e-07, "loss": 0.1636, "step": 23810 }, { "epoch": 0.46705882352941175, "grad_norm": 2.6690070629119873, "learning_rate": 9.107133291956413e-07, "loss": 0.1519, "step": 23820 }, { "epoch": 0.46725490196078434, "grad_norm": 2.801361560821533, "learning_rate": 9.10620856900479e-07, "loss": 0.1698, "step": 23830 }, { "epoch": 0.4674509803921569, "grad_norm": 2.343665599822998, "learning_rate": 9.105283414441483e-07, "loss": 0.1685, "step": 23840 }, { "epoch": 0.4676470588235294, "grad_norm": 2.3081154823303223, "learning_rate": 9.104357828363737e-07, "loss": 0.1789, "step": 23850 }, { "epoch": 0.46784313725490195, "grad_norm": 4.028531074523926, "learning_rate": 9.103431810868842e-07, "loss": 0.1525, "step": 23860 }, { "epoch": 0.4680392156862745, "grad_norm": 3.2975423336029053, "learning_rate": 9.102505362054135e-07, "loss": 0.1639, "step": 23870 }, { "epoch": 0.4682352941176471, "grad_norm": 2.844179630279541, "learning_rate": 9.101578482016995e-07, "loss": 0.1784, "step": 23880 }, { "epoch": 0.4684313725490196, "grad_norm": 3.0112762451171875, "learning_rate": 9.10065117085485e-07, "loss": 0.171, "step": 23890 }, { "epoch": 0.46862745098039216, "grad_norm": 3.0858445167541504, "learning_rate": 9.099723428665169e-07, "loss": 0.182, "step": 23900 }, { "epoch": 0.4688235294117647, "grad_norm": 2.380362033843994, "learning_rate": 9.098795255545471e-07, "loss": 0.1418, "step": 23910 }, { "epoch": 0.46901960784313723, "grad_norm": 1.4888864755630493, "learning_rate": 9.097866651593315e-07, "loss": 0.1668, "step": 23920 }, { "epoch": 0.4692156862745098, "grad_norm": 2.8000316619873047, "learning_rate": 9.096937616906315e-07, "loss": 0.175, "step": 23930 }, { "epoch": 0.46941176470588236, "grad_norm": 4.751358509063721, "learning_rate": 9.096008151582117e-07, "loss": 0.1775, "step": 23940 }, { "epoch": 0.4696078431372549, "grad_norm": 1.6092724800109863, "learning_rate": 9.09507825571842e-07, "loss": 0.1333, "step": 23950 }, { "epoch": 0.46980392156862744, "grad_norm": 2.2173235416412354, "learning_rate": 9.094147929412972e-07, "loss": 0.165, "step": 23960 }, { "epoch": 0.47, "grad_norm": 4.120408535003662, "learning_rate": 9.093217172763555e-07, "loss": 0.1446, "step": 23970 }, { "epoch": 0.47019607843137257, "grad_norm": 2.176358938217163, "learning_rate": 9.092285985868007e-07, "loss": 0.1626, "step": 23980 }, { "epoch": 0.4703921568627451, "grad_norm": 2.2862868309020996, "learning_rate": 9.091354368824204e-07, "loss": 0.15, "step": 23990 }, { "epoch": 0.47058823529411764, "grad_norm": 5.623892307281494, "learning_rate": 9.090422321730073e-07, "loss": 0.1802, "step": 24000 }, { "epoch": 0.4707843137254902, "grad_norm": 1.8884265422821045, "learning_rate": 9.089489844683583e-07, "loss": 0.1439, "step": 24010 }, { "epoch": 0.47098039215686277, "grad_norm": 0.957724928855896, "learning_rate": 9.088556937782746e-07, "loss": 0.1951, "step": 24020 }, { "epoch": 0.4711764705882353, "grad_norm": 2.720651865005493, "learning_rate": 9.087623601125623e-07, "loss": 0.1582, "step": 24030 }, { "epoch": 0.47137254901960784, "grad_norm": 3.0476651191711426, "learning_rate": 9.086689834810319e-07, "loss": 0.1661, "step": 24040 }, { "epoch": 0.4715686274509804, "grad_norm": 1.6762953996658325, "learning_rate": 9.085755638934986e-07, "loss": 0.154, "step": 24050 }, { "epoch": 0.4717647058823529, "grad_norm": 3.7230682373046875, "learning_rate": 9.084821013597815e-07, "loss": 0.1959, "step": 24060 }, { "epoch": 0.4719607843137255, "grad_norm": 4.060484409332275, "learning_rate": 9.083885958897049e-07, "loss": 0.158, "step": 24070 }, { "epoch": 0.47215686274509805, "grad_norm": 2.423158884048462, "learning_rate": 9.082950474930972e-07, "loss": 0.1321, "step": 24080 }, { "epoch": 0.4723529411764706, "grad_norm": 1.6420104503631592, "learning_rate": 9.082014561797917e-07, "loss": 0.1724, "step": 24090 }, { "epoch": 0.4725490196078431, "grad_norm": 2.889910936355591, "learning_rate": 9.081078219596259e-07, "loss": 0.1588, "step": 24100 }, { "epoch": 0.47274509803921566, "grad_norm": 2.226160764694214, "learning_rate": 9.080141448424417e-07, "loss": 0.1404, "step": 24110 }, { "epoch": 0.47294117647058825, "grad_norm": 1.6058361530303955, "learning_rate": 9.079204248380856e-07, "loss": 0.1555, "step": 24120 }, { "epoch": 0.4731372549019608, "grad_norm": 3.7315642833709717, "learning_rate": 9.07826661956409e-07, "loss": 0.1685, "step": 24130 }, { "epoch": 0.47333333333333333, "grad_norm": 2.4122471809387207, "learning_rate": 9.077328562072675e-07, "loss": 0.1676, "step": 24140 }, { "epoch": 0.47352941176470587, "grad_norm": 3.473085641860962, "learning_rate": 9.076390076005209e-07, "loss": 0.1948, "step": 24150 }, { "epoch": 0.47372549019607846, "grad_norm": 4.241771697998047, "learning_rate": 9.075451161460342e-07, "loss": 0.1923, "step": 24160 }, { "epoch": 0.473921568627451, "grad_norm": 2.3498566150665283, "learning_rate": 9.074511818536761e-07, "loss": 0.1634, "step": 24170 }, { "epoch": 0.47411764705882353, "grad_norm": 3.01627254486084, "learning_rate": 9.073572047333205e-07, "loss": 0.1904, "step": 24180 }, { "epoch": 0.47431372549019607, "grad_norm": 3.2591781616210938, "learning_rate": 9.072631847948455e-07, "loss": 0.1848, "step": 24190 }, { "epoch": 0.4745098039215686, "grad_norm": 2.454066038131714, "learning_rate": 9.071691220481337e-07, "loss": 0.1721, "step": 24200 }, { "epoch": 0.4747058823529412, "grad_norm": 1.2405847311019897, "learning_rate": 9.070750165030722e-07, "loss": 0.1879, "step": 24210 }, { "epoch": 0.47490196078431374, "grad_norm": 5.299139976501465, "learning_rate": 9.069808681695525e-07, "loss": 0.1732, "step": 24220 }, { "epoch": 0.4750980392156863, "grad_norm": 2.2013163566589355, "learning_rate": 9.068866770574709e-07, "loss": 0.178, "step": 24230 }, { "epoch": 0.4752941176470588, "grad_norm": 3.0320825576782227, "learning_rate": 9.06792443176728e-07, "loss": 0.1734, "step": 24240 }, { "epoch": 0.47549019607843135, "grad_norm": 3.744232177734375, "learning_rate": 9.066981665372288e-07, "loss": 0.2006, "step": 24250 }, { "epoch": 0.47568627450980394, "grad_norm": 1.6255619525909424, "learning_rate": 9.066038471488828e-07, "loss": 0.14, "step": 24260 }, { "epoch": 0.4758823529411765, "grad_norm": 2.2227110862731934, "learning_rate": 9.065094850216045e-07, "loss": 0.1557, "step": 24270 }, { "epoch": 0.476078431372549, "grad_norm": 1.8542286157608032, "learning_rate": 9.064150801653121e-07, "loss": 0.1595, "step": 24280 }, { "epoch": 0.47627450980392155, "grad_norm": 2.312717914581299, "learning_rate": 9.06320632589929e-07, "loss": 0.1768, "step": 24290 }, { "epoch": 0.4764705882352941, "grad_norm": 4.612179756164551, "learning_rate": 9.062261423053823e-07, "loss": 0.1567, "step": 24300 }, { "epoch": 0.4766666666666667, "grad_norm": 4.386453628540039, "learning_rate": 9.061316093216046e-07, "loss": 0.1466, "step": 24310 }, { "epoch": 0.4768627450980392, "grad_norm": 3.843641996383667, "learning_rate": 9.060370336485321e-07, "loss": 0.1494, "step": 24320 }, { "epoch": 0.47705882352941176, "grad_norm": 1.4089046716690063, "learning_rate": 9.059424152961061e-07, "loss": 0.1798, "step": 24330 }, { "epoch": 0.4772549019607843, "grad_norm": 3.7043139934539795, "learning_rate": 9.058477542742717e-07, "loss": 0.1823, "step": 24340 }, { "epoch": 0.4774509803921569, "grad_norm": 1.3964428901672363, "learning_rate": 9.057530505929793e-07, "loss": 0.1468, "step": 24350 }, { "epoch": 0.4776470588235294, "grad_norm": 1.999626874923706, "learning_rate": 9.056583042621833e-07, "loss": 0.1623, "step": 24360 }, { "epoch": 0.47784313725490196, "grad_norm": 7.409075736999512, "learning_rate": 9.055635152918426e-07, "loss": 0.1894, "step": 24370 }, { "epoch": 0.4780392156862745, "grad_norm": 2.3672220706939697, "learning_rate": 9.054686836919207e-07, "loss": 0.1634, "step": 24380 }, { "epoch": 0.47823529411764704, "grad_norm": 1.947332501411438, "learning_rate": 9.053738094723855e-07, "loss": 0.1836, "step": 24390 }, { "epoch": 0.47843137254901963, "grad_norm": 3.899850606918335, "learning_rate": 9.052788926432094e-07, "loss": 0.1538, "step": 24400 }, { "epoch": 0.47862745098039217, "grad_norm": 3.4537274837493896, "learning_rate": 9.051839332143695e-07, "loss": 0.1802, "step": 24410 }, { "epoch": 0.4788235294117647, "grad_norm": 2.885040521621704, "learning_rate": 9.050889311958468e-07, "loss": 0.1501, "step": 24420 }, { "epoch": 0.47901960784313724, "grad_norm": 3.0507149696350098, "learning_rate": 9.049938865976275e-07, "loss": 0.1593, "step": 24430 }, { "epoch": 0.4792156862745098, "grad_norm": 3.2847282886505127, "learning_rate": 9.048987994297017e-07, "loss": 0.1519, "step": 24440 }, { "epoch": 0.47941176470588237, "grad_norm": 2.802032470703125, "learning_rate": 9.048036697020644e-07, "loss": 0.1494, "step": 24450 }, { "epoch": 0.4796078431372549, "grad_norm": 2.941970109939575, "learning_rate": 9.047084974247146e-07, "loss": 0.1543, "step": 24460 }, { "epoch": 0.47980392156862745, "grad_norm": 2.4389562606811523, "learning_rate": 9.046132826076563e-07, "loss": 0.192, "step": 24470 }, { "epoch": 0.48, "grad_norm": 2.625566005706787, "learning_rate": 9.045180252608976e-07, "loss": 0.1394, "step": 24480 }, { "epoch": 0.4801960784313726, "grad_norm": 2.6525511741638184, "learning_rate": 9.044227253944512e-07, "loss": 0.1577, "step": 24490 }, { "epoch": 0.4803921568627451, "grad_norm": 5.398136138916016, "learning_rate": 9.043273830183342e-07, "loss": 0.1424, "step": 24500 }, { "epoch": 0.48058823529411765, "grad_norm": 2.5044467449188232, "learning_rate": 9.042319981425685e-07, "loss": 0.1563, "step": 24510 }, { "epoch": 0.4807843137254902, "grad_norm": 8.885017395019531, "learning_rate": 9.041365707771799e-07, "loss": 0.1639, "step": 24520 }, { "epoch": 0.4809803921568627, "grad_norm": 2.438899278640747, "learning_rate": 9.040411009321989e-07, "loss": 0.1711, "step": 24530 }, { "epoch": 0.4811764705882353, "grad_norm": 2.322005033493042, "learning_rate": 9.039455886176608e-07, "loss": 0.1483, "step": 24540 }, { "epoch": 0.48137254901960785, "grad_norm": 1.8334647417068481, "learning_rate": 9.03850033843605e-07, "loss": 0.1754, "step": 24550 }, { "epoch": 0.4815686274509804, "grad_norm": 4.417409896850586, "learning_rate": 9.037544366200753e-07, "loss": 0.184, "step": 24560 }, { "epoch": 0.48176470588235293, "grad_norm": 2.7178986072540283, "learning_rate": 9.036587969571201e-07, "loss": 0.1415, "step": 24570 }, { "epoch": 0.48196078431372547, "grad_norm": 3.5415523052215576, "learning_rate": 9.035631148647925e-07, "loss": 0.1718, "step": 24580 }, { "epoch": 0.48215686274509806, "grad_norm": 3.166921615600586, "learning_rate": 9.034673903531498e-07, "loss": 0.1825, "step": 24590 }, { "epoch": 0.4823529411764706, "grad_norm": 3.8621463775634766, "learning_rate": 9.033716234322537e-07, "loss": 0.1584, "step": 24600 }, { "epoch": 0.48254901960784313, "grad_norm": 4.49177885055542, "learning_rate": 9.032758141121704e-07, "loss": 0.183, "step": 24610 }, { "epoch": 0.48274509803921567, "grad_norm": 2.0303854942321777, "learning_rate": 9.031799624029707e-07, "loss": 0.2159, "step": 24620 }, { "epoch": 0.48294117647058826, "grad_norm": 1.8122600317001343, "learning_rate": 9.030840683147298e-07, "loss": 0.1615, "step": 24630 }, { "epoch": 0.4831372549019608, "grad_norm": 2.3625760078430176, "learning_rate": 9.029881318575272e-07, "loss": 0.1301, "step": 24640 }, { "epoch": 0.48333333333333334, "grad_norm": 3.074671983718872, "learning_rate": 9.02892153041447e-07, "loss": 0.196, "step": 24650 }, { "epoch": 0.4835294117647059, "grad_norm": 4.043420791625977, "learning_rate": 9.027961318765778e-07, "loss": 0.1634, "step": 24660 }, { "epoch": 0.4837254901960784, "grad_norm": 5.222151279449463, "learning_rate": 9.027000683730123e-07, "loss": 0.1363, "step": 24670 }, { "epoch": 0.483921568627451, "grad_norm": 2.5906999111175537, "learning_rate": 9.026039625408483e-07, "loss": 0.1752, "step": 24680 }, { "epoch": 0.48411764705882354, "grad_norm": 2.3125174045562744, "learning_rate": 9.025078143901876e-07, "loss": 0.1384, "step": 24690 }, { "epoch": 0.4843137254901961, "grad_norm": 3.1569430828094482, "learning_rate": 9.024116239311365e-07, "loss": 0.1217, "step": 24700 }, { "epoch": 0.4845098039215686, "grad_norm": 3.2144675254821777, "learning_rate": 9.023153911738055e-07, "loss": 0.1775, "step": 24710 }, { "epoch": 0.48470588235294115, "grad_norm": 4.369509696960449, "learning_rate": 9.022191161283102e-07, "loss": 0.177, "step": 24720 }, { "epoch": 0.48490196078431375, "grad_norm": 1.24225652217865, "learning_rate": 9.021227988047699e-07, "loss": 0.1533, "step": 24730 }, { "epoch": 0.4850980392156863, "grad_norm": 2.1702823638916016, "learning_rate": 9.02026439213309e-07, "loss": 0.2142, "step": 24740 }, { "epoch": 0.4852941176470588, "grad_norm": 5.039830684661865, "learning_rate": 9.019300373640559e-07, "loss": 0.1648, "step": 24750 }, { "epoch": 0.48549019607843136, "grad_norm": 2.8065221309661865, "learning_rate": 9.018335932671437e-07, "loss": 0.1749, "step": 24760 }, { "epoch": 0.4856862745098039, "grad_norm": 2.09317684173584, "learning_rate": 9.017371069327095e-07, "loss": 0.1712, "step": 24770 }, { "epoch": 0.4858823529411765, "grad_norm": 1.8034565448760986, "learning_rate": 9.016405783708957e-07, "loss": 0.1539, "step": 24780 }, { "epoch": 0.486078431372549, "grad_norm": 2.0748770236968994, "learning_rate": 9.015440075918482e-07, "loss": 0.1498, "step": 24790 }, { "epoch": 0.48627450980392156, "grad_norm": 2.3411972522735596, "learning_rate": 9.01447394605718e-07, "loss": 0.1368, "step": 24800 }, { "epoch": 0.4864705882352941, "grad_norm": 2.4566750526428223, "learning_rate": 9.0135073942266e-07, "loss": 0.1598, "step": 24810 }, { "epoch": 0.4866666666666667, "grad_norm": 1.2859565019607544, "learning_rate": 9.012540420528341e-07, "loss": 0.1199, "step": 24820 }, { "epoch": 0.48686274509803923, "grad_norm": 1.8370988368988037, "learning_rate": 9.011573025064041e-07, "loss": 0.1529, "step": 24830 }, { "epoch": 0.48705882352941177, "grad_norm": 3.665910243988037, "learning_rate": 9.010605207935387e-07, "loss": 0.1562, "step": 24840 }, { "epoch": 0.4872549019607843, "grad_norm": 1.6088730096817017, "learning_rate": 9.009636969244105e-07, "loss": 0.1738, "step": 24850 }, { "epoch": 0.48745098039215684, "grad_norm": 2.885850667953491, "learning_rate": 9.008668309091973e-07, "loss": 0.1678, "step": 24860 }, { "epoch": 0.48764705882352943, "grad_norm": 2.412092447280884, "learning_rate": 9.007699227580807e-07, "loss": 0.1541, "step": 24870 }, { "epoch": 0.48784313725490197, "grad_norm": 3.607527017593384, "learning_rate": 9.006729724812468e-07, "loss": 0.1559, "step": 24880 }, { "epoch": 0.4880392156862745, "grad_norm": 1.2256574630737305, "learning_rate": 9.005759800888862e-07, "loss": 0.144, "step": 24890 }, { "epoch": 0.48823529411764705, "grad_norm": 3.18145489692688, "learning_rate": 9.004789455911941e-07, "loss": 0.1516, "step": 24900 }, { "epoch": 0.4884313725490196, "grad_norm": 3.4014768600463867, "learning_rate": 9.0038186899837e-07, "loss": 0.1467, "step": 24910 }, { "epoch": 0.4886274509803922, "grad_norm": 5.392017364501953, "learning_rate": 9.002847503206177e-07, "loss": 0.1803, "step": 24920 }, { "epoch": 0.4888235294117647, "grad_norm": 1.8412593603134155, "learning_rate": 9.001875895681454e-07, "loss": 0.1581, "step": 24930 }, { "epoch": 0.48901960784313725, "grad_norm": 2.4822802543640137, "learning_rate": 9.000903867511664e-07, "loss": 0.1654, "step": 24940 }, { "epoch": 0.4892156862745098, "grad_norm": 2.05365252494812, "learning_rate": 8.999931418798974e-07, "loss": 0.1448, "step": 24950 }, { "epoch": 0.4894117647058824, "grad_norm": 2.4245526790618896, "learning_rate": 8.9989585496456e-07, "loss": 0.1615, "step": 24960 }, { "epoch": 0.4896078431372549, "grad_norm": 2.1090381145477295, "learning_rate": 8.997985260153804e-07, "loss": 0.2137, "step": 24970 }, { "epoch": 0.48980392156862745, "grad_norm": 4.2941389083862305, "learning_rate": 8.99701155042589e-07, "loss": 0.1639, "step": 24980 }, { "epoch": 0.49, "grad_norm": 2.35571551322937, "learning_rate": 8.996037420564207e-07, "loss": 0.1637, "step": 24990 }, { "epoch": 0.49019607843137253, "grad_norm": 2.1038014888763428, "learning_rate": 8.995062870671147e-07, "loss": 0.1792, "step": 25000 }, { "epoch": 0.4903921568627451, "grad_norm": 2.6237165927886963, "learning_rate": 8.994087900849148e-07, "loss": 0.1747, "step": 25010 }, { "epoch": 0.49058823529411766, "grad_norm": 1.8538998365402222, "learning_rate": 8.993112511200688e-07, "loss": 0.1721, "step": 25020 }, { "epoch": 0.4907843137254902, "grad_norm": 4.622958183288574, "learning_rate": 8.992136701828295e-07, "loss": 0.1608, "step": 25030 }, { "epoch": 0.49098039215686273, "grad_norm": 2.7657594680786133, "learning_rate": 8.991160472834536e-07, "loss": 0.1714, "step": 25040 }, { "epoch": 0.49117647058823527, "grad_norm": 3.495009660720825, "learning_rate": 8.990183824322028e-07, "loss": 0.1628, "step": 25050 }, { "epoch": 0.49137254901960786, "grad_norm": 3.044152021408081, "learning_rate": 8.989206756393424e-07, "loss": 0.174, "step": 25060 }, { "epoch": 0.4915686274509804, "grad_norm": 3.183093786239624, "learning_rate": 8.988229269151429e-07, "loss": 0.1748, "step": 25070 }, { "epoch": 0.49176470588235294, "grad_norm": 2.4767255783081055, "learning_rate": 8.987251362698787e-07, "loss": 0.1666, "step": 25080 }, { "epoch": 0.4919607843137255, "grad_norm": 1.8321486711502075, "learning_rate": 8.986273037138288e-07, "loss": 0.1744, "step": 25090 }, { "epoch": 0.492156862745098, "grad_norm": 2.9087460041046143, "learning_rate": 8.985294292572767e-07, "loss": 0.1407, "step": 25100 }, { "epoch": 0.4923529411764706, "grad_norm": 1.1424922943115234, "learning_rate": 8.984315129105098e-07, "loss": 0.1547, "step": 25110 }, { "epoch": 0.49254901960784314, "grad_norm": 2.0906972885131836, "learning_rate": 8.983335546838207e-07, "loss": 0.1904, "step": 25120 }, { "epoch": 0.4927450980392157, "grad_norm": 1.8081951141357422, "learning_rate": 8.982355545875058e-07, "loss": 0.1585, "step": 25130 }, { "epoch": 0.4929411764705882, "grad_norm": 2.768888473510742, "learning_rate": 8.98137512631866e-07, "loss": 0.1649, "step": 25140 }, { "epoch": 0.4931372549019608, "grad_norm": 1.9143553972244263, "learning_rate": 8.980394288272069e-07, "loss": 0.1449, "step": 25150 }, { "epoch": 0.49333333333333335, "grad_norm": 3.818718910217285, "learning_rate": 8.979413031838381e-07, "loss": 0.194, "step": 25160 }, { "epoch": 0.4935294117647059, "grad_norm": 2.8573215007781982, "learning_rate": 8.978431357120737e-07, "loss": 0.1485, "step": 25170 }, { "epoch": 0.4937254901960784, "grad_norm": 1.8309166431427002, "learning_rate": 8.977449264222326e-07, "loss": 0.1251, "step": 25180 }, { "epoch": 0.49392156862745096, "grad_norm": 4.058661460876465, "learning_rate": 8.976466753246376e-07, "loss": 0.1717, "step": 25190 }, { "epoch": 0.49411764705882355, "grad_norm": 2.5623557567596436, "learning_rate": 8.975483824296159e-07, "loss": 0.1531, "step": 25200 }, { "epoch": 0.4943137254901961, "grad_norm": 2.371208667755127, "learning_rate": 8.974500477474995e-07, "loss": 0.1729, "step": 25210 }, { "epoch": 0.4945098039215686, "grad_norm": 2.430969476699829, "learning_rate": 8.973516712886247e-07, "loss": 0.151, "step": 25220 }, { "epoch": 0.49470588235294116, "grad_norm": 2.539085626602173, "learning_rate": 8.972532530633314e-07, "loss": 0.1674, "step": 25230 }, { "epoch": 0.4949019607843137, "grad_norm": 3.260556221008301, "learning_rate": 8.971547930819651e-07, "loss": 0.1904, "step": 25240 }, { "epoch": 0.4950980392156863, "grad_norm": 2.3472506999969482, "learning_rate": 8.970562913548752e-07, "loss": 0.1476, "step": 25250 }, { "epoch": 0.49529411764705883, "grad_norm": 2.360715389251709, "learning_rate": 8.969577478924149e-07, "loss": 0.156, "step": 25260 }, { "epoch": 0.49549019607843137, "grad_norm": 24.9849853515625, "learning_rate": 8.968591627049427e-07, "loss": 0.1489, "step": 25270 }, { "epoch": 0.4956862745098039, "grad_norm": 2.08481764793396, "learning_rate": 8.967605358028209e-07, "loss": 0.1545, "step": 25280 }, { "epoch": 0.4958823529411765, "grad_norm": 2.881788730621338, "learning_rate": 8.966618671964165e-07, "loss": 0.178, "step": 25290 }, { "epoch": 0.49607843137254903, "grad_norm": 2.438833713531494, "learning_rate": 8.965631568961007e-07, "loss": 0.1665, "step": 25300 }, { "epoch": 0.49627450980392157, "grad_norm": 2.2862696647644043, "learning_rate": 8.96464404912249e-07, "loss": 0.1592, "step": 25310 }, { "epoch": 0.4964705882352941, "grad_norm": 2.9391798973083496, "learning_rate": 8.963656112552416e-07, "loss": 0.142, "step": 25320 }, { "epoch": 0.49666666666666665, "grad_norm": 1.389941692352295, "learning_rate": 8.962667759354628e-07, "loss": 0.1676, "step": 25330 }, { "epoch": 0.49686274509803924, "grad_norm": 1.8823703527450562, "learning_rate": 8.961678989633015e-07, "loss": 0.1859, "step": 25340 }, { "epoch": 0.4970588235294118, "grad_norm": 2.494550943374634, "learning_rate": 8.960689803491507e-07, "loss": 0.1501, "step": 25350 }, { "epoch": 0.4972549019607843, "grad_norm": 2.1866025924682617, "learning_rate": 8.95970020103408e-07, "loss": 0.158, "step": 25360 }, { "epoch": 0.49745098039215685, "grad_norm": 14.743940353393555, "learning_rate": 8.958710182364753e-07, "loss": 0.1792, "step": 25370 }, { "epoch": 0.4976470588235294, "grad_norm": 2.1910746097564697, "learning_rate": 8.957719747587587e-07, "loss": 0.1787, "step": 25380 }, { "epoch": 0.497843137254902, "grad_norm": 3.3317513465881348, "learning_rate": 8.956728896806692e-07, "loss": 0.1585, "step": 25390 }, { "epoch": 0.4980392156862745, "grad_norm": 2.904733657836914, "learning_rate": 8.955737630126217e-07, "loss": 0.1638, "step": 25400 }, { "epoch": 0.49823529411764705, "grad_norm": 5.984801292419434, "learning_rate": 8.954745947650355e-07, "loss": 0.1648, "step": 25410 }, { "epoch": 0.4984313725490196, "grad_norm": 2.6967105865478516, "learning_rate": 8.953753849483343e-07, "loss": 0.1795, "step": 25420 }, { "epoch": 0.4986274509803922, "grad_norm": 3.0815131664276123, "learning_rate": 8.952761335729466e-07, "loss": 0.1689, "step": 25430 }, { "epoch": 0.4988235294117647, "grad_norm": 2.92000150680542, "learning_rate": 8.951768406493045e-07, "loss": 0.1798, "step": 25440 }, { "epoch": 0.49901960784313726, "grad_norm": 2.6907994747161865, "learning_rate": 8.950775061878452e-07, "loss": 0.1612, "step": 25450 }, { "epoch": 0.4992156862745098, "grad_norm": 2.459678888320923, "learning_rate": 8.949781301990096e-07, "loss": 0.1673, "step": 25460 }, { "epoch": 0.49941176470588233, "grad_norm": 2.7139971256256104, "learning_rate": 8.948787126932437e-07, "loss": 0.1743, "step": 25470 }, { "epoch": 0.4996078431372549, "grad_norm": 3.455734968185425, "learning_rate": 8.947792536809971e-07, "loss": 0.1769, "step": 25480 }, { "epoch": 0.49980392156862746, "grad_norm": 1.9984755516052246, "learning_rate": 8.946797531727243e-07, "loss": 0.1425, "step": 25490 }, { "epoch": 0.5, "grad_norm": 3.2788777351379395, "learning_rate": 8.945802111788842e-07, "loss": 0.1748, "step": 25500 }, { "epoch": 0.5001960784313726, "grad_norm": 2.94987416267395, "learning_rate": 8.944806277099395e-07, "loss": 0.1731, "step": 25510 }, { "epoch": 0.5003921568627451, "grad_norm": 2.083582878112793, "learning_rate": 8.943810027763579e-07, "loss": 0.162, "step": 25520 }, { "epoch": 0.5005882352941177, "grad_norm": 3.640512704849243, "learning_rate": 8.942813363886109e-07, "loss": 0.1734, "step": 25530 }, { "epoch": 0.5007843137254901, "grad_norm": 2.850459098815918, "learning_rate": 8.941816285571748e-07, "loss": 0.1649, "step": 25540 }, { "epoch": 0.5009803921568627, "grad_norm": 2.9086899757385254, "learning_rate": 8.9408187929253e-07, "loss": 0.1311, "step": 25550 }, { "epoch": 0.5011764705882353, "grad_norm": 4.257396697998047, "learning_rate": 8.939820886051615e-07, "loss": 0.1403, "step": 25560 }, { "epoch": 0.5013725490196078, "grad_norm": 6.8138108253479, "learning_rate": 8.938822565055583e-07, "loss": 0.1659, "step": 25570 }, { "epoch": 0.5015686274509804, "grad_norm": 1.5326273441314697, "learning_rate": 8.937823830042142e-07, "loss": 0.1654, "step": 25580 }, { "epoch": 0.5017647058823529, "grad_norm": 5.51688289642334, "learning_rate": 8.936824681116268e-07, "loss": 0.1442, "step": 25590 }, { "epoch": 0.5019607843137255, "grad_norm": 3.026277542114258, "learning_rate": 8.935825118382985e-07, "loss": 0.2063, "step": 25600 }, { "epoch": 0.5021568627450981, "grad_norm": 3.6119699478149414, "learning_rate": 8.93482514194736e-07, "loss": 0.152, "step": 25610 }, { "epoch": 0.5023529411764706, "grad_norm": 2.8206820487976074, "learning_rate": 8.933824751914501e-07, "loss": 0.1643, "step": 25620 }, { "epoch": 0.5025490196078432, "grad_norm": 2.1781721115112305, "learning_rate": 8.932823948389562e-07, "loss": 0.1701, "step": 25630 }, { "epoch": 0.5027450980392156, "grad_norm": 1.5342005491256714, "learning_rate": 8.931822731477738e-07, "loss": 0.1543, "step": 25640 }, { "epoch": 0.5029411764705882, "grad_norm": 3.0182414054870605, "learning_rate": 8.93082110128427e-07, "loss": 0.1611, "step": 25650 }, { "epoch": 0.5031372549019608, "grad_norm": 1.6072666645050049, "learning_rate": 8.929819057914442e-07, "loss": 0.1328, "step": 25660 }, { "epoch": 0.5033333333333333, "grad_norm": 2.793079137802124, "learning_rate": 8.928816601473579e-07, "loss": 0.1796, "step": 25670 }, { "epoch": 0.5035294117647059, "grad_norm": 2.436704397201538, "learning_rate": 8.927813732067052e-07, "loss": 0.1603, "step": 25680 }, { "epoch": 0.5037254901960785, "grad_norm": 2.0419297218322754, "learning_rate": 8.926810449800275e-07, "loss": 0.149, "step": 25690 }, { "epoch": 0.503921568627451, "grad_norm": 3.0227816104888916, "learning_rate": 8.925806754778704e-07, "loss": 0.1505, "step": 25700 }, { "epoch": 0.5041176470588236, "grad_norm": 4.98631477355957, "learning_rate": 8.92480264710784e-07, "loss": 0.1425, "step": 25710 }, { "epoch": 0.504313725490196, "grad_norm": 4.257578372955322, "learning_rate": 8.923798126893227e-07, "loss": 0.2097, "step": 25720 }, { "epoch": 0.5045098039215686, "grad_norm": 2.911529779434204, "learning_rate": 8.922793194240452e-07, "loss": 0.2187, "step": 25730 }, { "epoch": 0.5047058823529412, "grad_norm": 3.196640729904175, "learning_rate": 8.921787849255145e-07, "loss": 0.1975, "step": 25740 }, { "epoch": 0.5049019607843137, "grad_norm": 2.4060182571411133, "learning_rate": 8.92078209204298e-07, "loss": 0.148, "step": 25750 }, { "epoch": 0.5050980392156863, "grad_norm": 4.824855804443359, "learning_rate": 8.919775922709673e-07, "loss": 0.1822, "step": 25760 }, { "epoch": 0.5052941176470588, "grad_norm": 2.969752550125122, "learning_rate": 8.918769341360988e-07, "loss": 0.1618, "step": 25770 }, { "epoch": 0.5054901960784314, "grad_norm": 4.035074710845947, "learning_rate": 8.917762348102724e-07, "loss": 0.1776, "step": 25780 }, { "epoch": 0.505686274509804, "grad_norm": 2.595327854156494, "learning_rate": 8.916754943040732e-07, "loss": 0.163, "step": 25790 }, { "epoch": 0.5058823529411764, "grad_norm": 2.1777918338775635, "learning_rate": 8.9157471262809e-07, "loss": 0.1618, "step": 25800 }, { "epoch": 0.506078431372549, "grad_norm": 3.423992156982422, "learning_rate": 8.914738897929162e-07, "loss": 0.1444, "step": 25810 }, { "epoch": 0.5062745098039215, "grad_norm": 3.9598894119262695, "learning_rate": 8.913730258091496e-07, "loss": 0.1565, "step": 25820 }, { "epoch": 0.5064705882352941, "grad_norm": 2.689690351486206, "learning_rate": 8.912721206873921e-07, "loss": 0.1489, "step": 25830 }, { "epoch": 0.5066666666666667, "grad_norm": 2.707911252975464, "learning_rate": 8.911711744382502e-07, "loss": 0.1725, "step": 25840 }, { "epoch": 0.5068627450980392, "grad_norm": 3.342078924179077, "learning_rate": 8.910701870723343e-07, "loss": 0.1424, "step": 25850 }, { "epoch": 0.5070588235294118, "grad_norm": 3.0183868408203125, "learning_rate": 8.909691586002595e-07, "loss": 0.1687, "step": 25860 }, { "epoch": 0.5072549019607843, "grad_norm": 2.561461925506592, "learning_rate": 8.908680890326451e-07, "loss": 0.1638, "step": 25870 }, { "epoch": 0.5074509803921569, "grad_norm": 2.182682752609253, "learning_rate": 8.907669783801148e-07, "loss": 0.1676, "step": 25880 }, { "epoch": 0.5076470588235295, "grad_norm": 2.1583807468414307, "learning_rate": 8.906658266532964e-07, "loss": 0.1448, "step": 25890 }, { "epoch": 0.5078431372549019, "grad_norm": 2.251164674758911, "learning_rate": 8.905646338628223e-07, "loss": 0.1645, "step": 25900 }, { "epoch": 0.5080392156862745, "grad_norm": 3.061868906021118, "learning_rate": 8.90463400019329e-07, "loss": 0.1528, "step": 25910 }, { "epoch": 0.508235294117647, "grad_norm": 3.621892213821411, "learning_rate": 8.903621251334574e-07, "loss": 0.1705, "step": 25920 }, { "epoch": 0.5084313725490196, "grad_norm": 3.1058077812194824, "learning_rate": 8.902608092158526e-07, "loss": 0.1441, "step": 25930 }, { "epoch": 0.5086274509803922, "grad_norm": 1.3250032663345337, "learning_rate": 8.901594522771643e-07, "loss": 0.2015, "step": 25940 }, { "epoch": 0.5088235294117647, "grad_norm": 4.550673961639404, "learning_rate": 8.900580543280463e-07, "loss": 0.2028, "step": 25950 }, { "epoch": 0.5090196078431373, "grad_norm": 4.632724761962891, "learning_rate": 8.899566153791565e-07, "loss": 0.1706, "step": 25960 }, { "epoch": 0.5092156862745097, "grad_norm": 1.8889857530593872, "learning_rate": 8.898551354411577e-07, "loss": 0.1693, "step": 25970 }, { "epoch": 0.5094117647058823, "grad_norm": 2.764158010482788, "learning_rate": 8.897536145247164e-07, "loss": 0.1605, "step": 25980 }, { "epoch": 0.5096078431372549, "grad_norm": 2.0191099643707275, "learning_rate": 8.896520526405038e-07, "loss": 0.1464, "step": 25990 }, { "epoch": 0.5098039215686274, "grad_norm": 3.1552376747131348, "learning_rate": 8.895504497991952e-07, "loss": 0.1519, "step": 26000 }, { "epoch": 0.51, "grad_norm": 8.030131340026855, "learning_rate": 8.894488060114702e-07, "loss": 0.2081, "step": 26010 }, { "epoch": 0.5101960784313726, "grad_norm": 2.12567400932312, "learning_rate": 8.893471212880131e-07, "loss": 0.1785, "step": 26020 }, { "epoch": 0.5103921568627451, "grad_norm": 2.9505906105041504, "learning_rate": 8.892453956395118e-07, "loss": 0.1785, "step": 26030 }, { "epoch": 0.5105882352941177, "grad_norm": 2.396299123764038, "learning_rate": 8.89143629076659e-07, "loss": 0.1808, "step": 26040 }, { "epoch": 0.5107843137254902, "grad_norm": 6.345235824584961, "learning_rate": 8.890418216101517e-07, "loss": 0.1415, "step": 26050 }, { "epoch": 0.5109803921568628, "grad_norm": 1.7161972522735596, "learning_rate": 8.889399732506911e-07, "loss": 0.1666, "step": 26060 }, { "epoch": 0.5111764705882353, "grad_norm": 2.5136163234710693, "learning_rate": 8.888380840089824e-07, "loss": 0.1791, "step": 26070 }, { "epoch": 0.5113725490196078, "grad_norm": 1.5666674375534058, "learning_rate": 8.887361538957357e-07, "loss": 0.159, "step": 26080 }, { "epoch": 0.5115686274509804, "grad_norm": 2.190021514892578, "learning_rate": 8.886341829216648e-07, "loss": 0.1862, "step": 26090 }, { "epoch": 0.5117647058823529, "grad_norm": 2.7090468406677246, "learning_rate": 8.885321710974883e-07, "loss": 0.154, "step": 26100 }, { "epoch": 0.5119607843137255, "grad_norm": 2.657778263092041, "learning_rate": 8.884301184339288e-07, "loss": 0.1494, "step": 26110 }, { "epoch": 0.5121568627450981, "grad_norm": 1.690309762954712, "learning_rate": 8.883280249417131e-07, "loss": 0.1764, "step": 26120 }, { "epoch": 0.5123529411764706, "grad_norm": 2.114065170288086, "learning_rate": 8.882258906315727e-07, "loss": 0.1459, "step": 26130 }, { "epoch": 0.5125490196078432, "grad_norm": 2.238528251647949, "learning_rate": 8.881237155142432e-07, "loss": 0.1678, "step": 26140 }, { "epoch": 0.5127450980392156, "grad_norm": 1.8304730653762817, "learning_rate": 8.880214996004641e-07, "loss": 0.1626, "step": 26150 }, { "epoch": 0.5129411764705882, "grad_norm": 2.7795205116271973, "learning_rate": 8.879192429009794e-07, "loss": 0.1456, "step": 26160 }, { "epoch": 0.5131372549019608, "grad_norm": 2.5387825965881348, "learning_rate": 8.878169454265381e-07, "loss": 0.1671, "step": 26170 }, { "epoch": 0.5133333333333333, "grad_norm": 3.363293409347534, "learning_rate": 8.877146071878925e-07, "loss": 0.1517, "step": 26180 }, { "epoch": 0.5135294117647059, "grad_norm": 3.200901746749878, "learning_rate": 8.876122281957995e-07, "loss": 0.1737, "step": 26190 }, { "epoch": 0.5137254901960784, "grad_norm": 1.367678165435791, "learning_rate": 8.875098084610206e-07, "loss": 0.1573, "step": 26200 }, { "epoch": 0.513921568627451, "grad_norm": 5.300647735595703, "learning_rate": 8.874073479943211e-07, "loss": 0.1865, "step": 26210 }, { "epoch": 0.5141176470588236, "grad_norm": 2.4575698375701904, "learning_rate": 8.873048468064713e-07, "loss": 0.1495, "step": 26220 }, { "epoch": 0.514313725490196, "grad_norm": 2.317368745803833, "learning_rate": 8.872023049082446e-07, "loss": 0.1768, "step": 26230 }, { "epoch": 0.5145098039215686, "grad_norm": 4.756613731384277, "learning_rate": 8.870997223104198e-07, "loss": 0.158, "step": 26240 }, { "epoch": 0.5147058823529411, "grad_norm": 1.7269606590270996, "learning_rate": 8.869970990237796e-07, "loss": 0.1567, "step": 26250 }, { "epoch": 0.5149019607843137, "grad_norm": 3.2215259075164795, "learning_rate": 8.868944350591108e-07, "loss": 0.2039, "step": 26260 }, { "epoch": 0.5150980392156863, "grad_norm": 3.5256526470184326, "learning_rate": 8.867917304272046e-07, "loss": 0.1531, "step": 26270 }, { "epoch": 0.5152941176470588, "grad_norm": 2.5800771713256836, "learning_rate": 8.866889851388566e-07, "loss": 0.1612, "step": 26280 }, { "epoch": 0.5154901960784314, "grad_norm": 10.889979362487793, "learning_rate": 8.865861992048665e-07, "loss": 0.1602, "step": 26290 }, { "epoch": 0.515686274509804, "grad_norm": 8.58509349822998, "learning_rate": 8.864833726360381e-07, "loss": 0.1755, "step": 26300 }, { "epoch": 0.5158823529411765, "grad_norm": 2.5510382652282715, "learning_rate": 8.863805054431801e-07, "loss": 0.187, "step": 26310 }, { "epoch": 0.516078431372549, "grad_norm": 2.379842519760132, "learning_rate": 8.862775976371049e-07, "loss": 0.1456, "step": 26320 }, { "epoch": 0.5162745098039215, "grad_norm": 6.114444255828857, "learning_rate": 8.861746492286293e-07, "loss": 0.1711, "step": 26330 }, { "epoch": 0.5164705882352941, "grad_norm": 1.643066644668579, "learning_rate": 8.860716602285745e-07, "loss": 0.1463, "step": 26340 }, { "epoch": 0.5166666666666667, "grad_norm": 1.511488437652588, "learning_rate": 8.859686306477658e-07, "loss": 0.1577, "step": 26350 }, { "epoch": 0.5168627450980392, "grad_norm": 2.9457015991210938, "learning_rate": 8.858655604970329e-07, "loss": 0.1338, "step": 26360 }, { "epoch": 0.5170588235294118, "grad_norm": 3.0512497425079346, "learning_rate": 8.857624497872097e-07, "loss": 0.2021, "step": 26370 }, { "epoch": 0.5172549019607843, "grad_norm": 2.3173859119415283, "learning_rate": 8.856592985291343e-07, "loss": 0.1548, "step": 26380 }, { "epoch": 0.5174509803921569, "grad_norm": 2.154543161392212, "learning_rate": 8.855561067336492e-07, "loss": 0.1608, "step": 26390 }, { "epoch": 0.5176470588235295, "grad_norm": 2.329479455947876, "learning_rate": 8.85452874411601e-07, "loss": 0.1627, "step": 26400 }, { "epoch": 0.5178431372549019, "grad_norm": 3.8357608318328857, "learning_rate": 8.853496015738408e-07, "loss": 0.1887, "step": 26410 }, { "epoch": 0.5180392156862745, "grad_norm": 2.652498960494995, "learning_rate": 8.852462882312237e-07, "loss": 0.1533, "step": 26420 }, { "epoch": 0.518235294117647, "grad_norm": 2.701223134994507, "learning_rate": 8.851429343946092e-07, "loss": 0.1625, "step": 26430 }, { "epoch": 0.5184313725490196, "grad_norm": 2.0112335681915283, "learning_rate": 8.850395400748612e-07, "loss": 0.1279, "step": 26440 }, { "epoch": 0.5186274509803922, "grad_norm": 2.8851001262664795, "learning_rate": 8.849361052828474e-07, "loss": 0.1956, "step": 26450 }, { "epoch": 0.5188235294117647, "grad_norm": 4.483410835266113, "learning_rate": 8.848326300294402e-07, "loss": 0.1985, "step": 26460 }, { "epoch": 0.5190196078431373, "grad_norm": 2.6038084030151367, "learning_rate": 8.847291143255158e-07, "loss": 0.1457, "step": 26470 }, { "epoch": 0.5192156862745098, "grad_norm": 2.9075605869293213, "learning_rate": 8.846255581819555e-07, "loss": 0.1499, "step": 26480 }, { "epoch": 0.5194117647058824, "grad_norm": 1.4977184534072876, "learning_rate": 8.845219616096438e-07, "loss": 0.1511, "step": 26490 }, { "epoch": 0.5196078431372549, "grad_norm": 2.6324312686920166, "learning_rate": 8.844183246194701e-07, "loss": 0.1591, "step": 26500 }, { "epoch": 0.5198039215686274, "grad_norm": 2.5922000408172607, "learning_rate": 8.84314647222328e-07, "loss": 0.1391, "step": 26510 }, { "epoch": 0.52, "grad_norm": 2.2736711502075195, "learning_rate": 8.842109294291149e-07, "loss": 0.1331, "step": 26520 }, { "epoch": 0.5201960784313725, "grad_norm": 3.0898020267486572, "learning_rate": 8.841071712507333e-07, "loss": 0.1364, "step": 26530 }, { "epoch": 0.5203921568627451, "grad_norm": 3.6554410457611084, "learning_rate": 8.840033726980889e-07, "loss": 0.1793, "step": 26540 }, { "epoch": 0.5205882352941177, "grad_norm": 2.3550353050231934, "learning_rate": 8.838995337820925e-07, "loss": 0.1867, "step": 26550 }, { "epoch": 0.5207843137254902, "grad_norm": 4.174932479858398, "learning_rate": 8.837956545136589e-07, "loss": 0.1859, "step": 26560 }, { "epoch": 0.5209803921568628, "grad_norm": 1.8251268863677979, "learning_rate": 8.836917349037067e-07, "loss": 0.1799, "step": 26570 }, { "epoch": 0.5211764705882352, "grad_norm": 2.484668254852295, "learning_rate": 8.835877749631593e-07, "loss": 0.2127, "step": 26580 }, { "epoch": 0.5213725490196078, "grad_norm": 1.8765742778778076, "learning_rate": 8.83483774702944e-07, "loss": 0.1541, "step": 26590 }, { "epoch": 0.5215686274509804, "grad_norm": 1.965352177619934, "learning_rate": 8.833797341339928e-07, "loss": 0.1616, "step": 26600 }, { "epoch": 0.5217647058823529, "grad_norm": 2.158848524093628, "learning_rate": 8.832756532672413e-07, "loss": 0.1859, "step": 26610 }, { "epoch": 0.5219607843137255, "grad_norm": 1.8781852722167969, "learning_rate": 8.831715321136297e-07, "loss": 0.1444, "step": 26620 }, { "epoch": 0.5221568627450981, "grad_norm": 2.273172378540039, "learning_rate": 8.830673706841024e-07, "loss": 0.165, "step": 26630 }, { "epoch": 0.5223529411764706, "grad_norm": 2.6267902851104736, "learning_rate": 8.829631689896081e-07, "loss": 0.1696, "step": 26640 }, { "epoch": 0.5225490196078432, "grad_norm": 2.631007671356201, "learning_rate": 8.828589270410996e-07, "loss": 0.1627, "step": 26650 }, { "epoch": 0.5227450980392156, "grad_norm": 1.5894253253936768, "learning_rate": 8.827546448495338e-07, "loss": 0.1724, "step": 26660 }, { "epoch": 0.5229411764705882, "grad_norm": 2.5199666023254395, "learning_rate": 8.826503224258723e-07, "loss": 0.1219, "step": 26670 }, { "epoch": 0.5231372549019608, "grad_norm": 2.529285192489624, "learning_rate": 8.825459597810805e-07, "loss": 0.1688, "step": 26680 }, { "epoch": 0.5233333333333333, "grad_norm": 2.0041801929473877, "learning_rate": 8.824415569261281e-07, "loss": 0.1357, "step": 26690 }, { "epoch": 0.5235294117647059, "grad_norm": 2.0342631340026855, "learning_rate": 8.82337113871989e-07, "loss": 0.1808, "step": 26700 }, { "epoch": 0.5237254901960784, "grad_norm": 2.903748035430908, "learning_rate": 8.822326306296419e-07, "loss": 0.1739, "step": 26710 }, { "epoch": 0.523921568627451, "grad_norm": 2.744570255279541, "learning_rate": 8.821281072100686e-07, "loss": 0.1782, "step": 26720 }, { "epoch": 0.5241176470588236, "grad_norm": 2.476252794265747, "learning_rate": 8.820235436242562e-07, "loss": 0.1409, "step": 26730 }, { "epoch": 0.5243137254901961, "grad_norm": 2.18163800239563, "learning_rate": 8.819189398831954e-07, "loss": 0.1378, "step": 26740 }, { "epoch": 0.5245098039215687, "grad_norm": 3.118400812149048, "learning_rate": 8.818142959978816e-07, "loss": 0.1603, "step": 26750 }, { "epoch": 0.5247058823529411, "grad_norm": 3.838132858276367, "learning_rate": 8.817096119793135e-07, "loss": 0.1262, "step": 26760 }, { "epoch": 0.5249019607843137, "grad_norm": 3.0116519927978516, "learning_rate": 8.816048878384954e-07, "loss": 0.1566, "step": 26770 }, { "epoch": 0.5250980392156863, "grad_norm": 4.937889575958252, "learning_rate": 8.815001235864343e-07, "loss": 0.2046, "step": 26780 }, { "epoch": 0.5252941176470588, "grad_norm": 3.0143344402313232, "learning_rate": 8.813953192341429e-07, "loss": 0.1786, "step": 26790 }, { "epoch": 0.5254901960784314, "grad_norm": 2.76338267326355, "learning_rate": 8.812904747926368e-07, "loss": 0.2022, "step": 26800 }, { "epoch": 0.5256862745098039, "grad_norm": 2.979842185974121, "learning_rate": 8.811855902729368e-07, "loss": 0.1663, "step": 26810 }, { "epoch": 0.5258823529411765, "grad_norm": 1.4616903066635132, "learning_rate": 8.810806656860674e-07, "loss": 0.1582, "step": 26820 }, { "epoch": 0.5260784313725491, "grad_norm": 3.2583796977996826, "learning_rate": 8.809757010430574e-07, "loss": 0.163, "step": 26830 }, { "epoch": 0.5262745098039215, "grad_norm": 1.928894281387329, "learning_rate": 8.8087069635494e-07, "loss": 0.1551, "step": 26840 }, { "epoch": 0.5264705882352941, "grad_norm": 0.8360648155212402, "learning_rate": 8.807656516327519e-07, "loss": 0.17, "step": 26850 }, { "epoch": 0.5266666666666666, "grad_norm": 2.8887407779693604, "learning_rate": 8.806605668875353e-07, "loss": 0.1801, "step": 26860 }, { "epoch": 0.5268627450980392, "grad_norm": 1.8250000476837158, "learning_rate": 8.805554421303354e-07, "loss": 0.1853, "step": 26870 }, { "epoch": 0.5270588235294118, "grad_norm": 1.6162852048873901, "learning_rate": 8.804502773722022e-07, "loss": 0.1524, "step": 26880 }, { "epoch": 0.5272549019607843, "grad_norm": 4.308277130126953, "learning_rate": 8.803450726241899e-07, "loss": 0.1385, "step": 26890 }, { "epoch": 0.5274509803921569, "grad_norm": 2.4269118309020996, "learning_rate": 8.802398278973565e-07, "loss": 0.1564, "step": 26900 }, { "epoch": 0.5276470588235294, "grad_norm": 3.145597457885742, "learning_rate": 8.801345432027645e-07, "loss": 0.2161, "step": 26910 }, { "epoch": 0.527843137254902, "grad_norm": 2.2062509059906006, "learning_rate": 8.800292185514811e-07, "loss": 0.1715, "step": 26920 }, { "epoch": 0.5280392156862745, "grad_norm": 5.228329181671143, "learning_rate": 8.799238539545764e-07, "loss": 0.1477, "step": 26930 }, { "epoch": 0.528235294117647, "grad_norm": 2.139601945877075, "learning_rate": 8.798184494231261e-07, "loss": 0.1728, "step": 26940 }, { "epoch": 0.5284313725490196, "grad_norm": 2.5617167949676514, "learning_rate": 8.797130049682092e-07, "loss": 0.1824, "step": 26950 }, { "epoch": 0.5286274509803922, "grad_norm": 3.529071569442749, "learning_rate": 8.796075206009093e-07, "loss": 0.1738, "step": 26960 }, { "epoch": 0.5288235294117647, "grad_norm": 2.236891746520996, "learning_rate": 8.795019963323138e-07, "loss": 0.1269, "step": 26970 }, { "epoch": 0.5290196078431373, "grad_norm": 2.2613205909729004, "learning_rate": 8.79396432173515e-07, "loss": 0.164, "step": 26980 }, { "epoch": 0.5292156862745098, "grad_norm": 4.034409999847412, "learning_rate": 8.792908281356087e-07, "loss": 0.1727, "step": 26990 }, { "epoch": 0.5294117647058824, "grad_norm": 4.6653337478637695, "learning_rate": 8.791851842296949e-07, "loss": 0.1597, "step": 27000 }, { "epoch": 0.529607843137255, "grad_norm": 2.641500473022461, "learning_rate": 8.790795004668785e-07, "loss": 0.1528, "step": 27010 }, { "epoch": 0.5298039215686274, "grad_norm": 1.6557461023330688, "learning_rate": 8.789737768582679e-07, "loss": 0.1649, "step": 27020 }, { "epoch": 0.53, "grad_norm": 2.2125303745269775, "learning_rate": 8.788680134149759e-07, "loss": 0.1876, "step": 27030 }, { "epoch": 0.5301960784313725, "grad_norm": 3.53863787651062, "learning_rate": 8.787622101481198e-07, "loss": 0.1742, "step": 27040 }, { "epoch": 0.5303921568627451, "grad_norm": 3.038278579711914, "learning_rate": 8.786563670688203e-07, "loss": 0.1406, "step": 27050 }, { "epoch": 0.5305882352941177, "grad_norm": 3.6290485858917236, "learning_rate": 8.785504841882032e-07, "loss": 0.1645, "step": 27060 }, { "epoch": 0.5307843137254902, "grad_norm": 3.601513624191284, "learning_rate": 8.784445615173977e-07, "loss": 0.1452, "step": 27070 }, { "epoch": 0.5309803921568628, "grad_norm": 2.717808723449707, "learning_rate": 8.783385990675378e-07, "loss": 0.1621, "step": 27080 }, { "epoch": 0.5311764705882352, "grad_norm": 1.8598272800445557, "learning_rate": 8.782325968497613e-07, "loss": 0.1893, "step": 27090 }, { "epoch": 0.5313725490196078, "grad_norm": 3.7754170894622803, "learning_rate": 8.781265548752103e-07, "loss": 0.1998, "step": 27100 }, { "epoch": 0.5315686274509804, "grad_norm": 3.148582935333252, "learning_rate": 8.780204731550313e-07, "loss": 0.1666, "step": 27110 }, { "epoch": 0.5317647058823529, "grad_norm": 2.138633966445923, "learning_rate": 8.779143517003747e-07, "loss": 0.17, "step": 27120 }, { "epoch": 0.5319607843137255, "grad_norm": 3.542800188064575, "learning_rate": 8.778081905223946e-07, "loss": 0.1642, "step": 27130 }, { "epoch": 0.532156862745098, "grad_norm": 2.2340967655181885, "learning_rate": 8.777019896322506e-07, "loss": 0.1582, "step": 27140 }, { "epoch": 0.5323529411764706, "grad_norm": 2.3552844524383545, "learning_rate": 8.775957490411052e-07, "loss": 0.1667, "step": 27150 }, { "epoch": 0.5325490196078432, "grad_norm": 2.3776254653930664, "learning_rate": 8.774894687601255e-07, "loss": 0.1728, "step": 27160 }, { "epoch": 0.5327450980392157, "grad_norm": 2.6546878814697266, "learning_rate": 8.773831488004832e-07, "loss": 0.1406, "step": 27170 }, { "epoch": 0.5329411764705883, "grad_norm": 1.6420385837554932, "learning_rate": 8.772767891733537e-07, "loss": 0.1407, "step": 27180 }, { "epoch": 0.5331372549019607, "grad_norm": 2.8440396785736084, "learning_rate": 8.771703898899163e-07, "loss": 0.1777, "step": 27190 }, { "epoch": 0.5333333333333333, "grad_norm": 1.7372618913650513, "learning_rate": 8.770639509613553e-07, "loss": 0.1734, "step": 27200 }, { "epoch": 0.5335294117647059, "grad_norm": 3.8457107543945312, "learning_rate": 8.769574723988584e-07, "loss": 0.1704, "step": 27210 }, { "epoch": 0.5337254901960784, "grad_norm": 3.3520750999450684, "learning_rate": 8.76850954213618e-07, "loss": 0.177, "step": 27220 }, { "epoch": 0.533921568627451, "grad_norm": 2.536594867706299, "learning_rate": 8.767443964168304e-07, "loss": 0.161, "step": 27230 }, { "epoch": 0.5341176470588235, "grad_norm": 2.7124366760253906, "learning_rate": 8.766377990196958e-07, "loss": 0.1485, "step": 27240 }, { "epoch": 0.5343137254901961, "grad_norm": 3.0238289833068848, "learning_rate": 8.765311620334192e-07, "loss": 0.1591, "step": 27250 }, { "epoch": 0.5345098039215687, "grad_norm": 2.9540135860443115, "learning_rate": 8.764244854692094e-07, "loss": 0.1617, "step": 27260 }, { "epoch": 0.5347058823529411, "grad_norm": 2.0712790489196777, "learning_rate": 8.763177693382793e-07, "loss": 0.1782, "step": 27270 }, { "epoch": 0.5349019607843137, "grad_norm": 2.188962697982788, "learning_rate": 8.76211013651846e-07, "loss": 0.1421, "step": 27280 }, { "epoch": 0.5350980392156863, "grad_norm": 2.041367292404175, "learning_rate": 8.761042184211308e-07, "loss": 0.1671, "step": 27290 }, { "epoch": 0.5352941176470588, "grad_norm": 3.020658016204834, "learning_rate": 8.759973836573592e-07, "loss": 0.1658, "step": 27300 }, { "epoch": 0.5354901960784314, "grad_norm": 1.7654461860656738, "learning_rate": 8.758905093717608e-07, "loss": 0.1332, "step": 27310 }, { "epoch": 0.5356862745098039, "grad_norm": 2.048755168914795, "learning_rate": 8.757835955755695e-07, "loss": 0.13, "step": 27320 }, { "epoch": 0.5358823529411765, "grad_norm": 2.1870920658111572, "learning_rate": 8.756766422800231e-07, "loss": 0.1479, "step": 27330 }, { "epoch": 0.5360784313725491, "grad_norm": 3.2392961978912354, "learning_rate": 8.755696494963635e-07, "loss": 0.1654, "step": 27340 }, { "epoch": 0.5362745098039216, "grad_norm": 3.0758464336395264, "learning_rate": 8.754626172358372e-07, "loss": 0.1753, "step": 27350 }, { "epoch": 0.5364705882352941, "grad_norm": 2.3353450298309326, "learning_rate": 8.753555455096944e-07, "loss": 0.1225, "step": 27360 }, { "epoch": 0.5366666666666666, "grad_norm": 2.937659978866577, "learning_rate": 8.752484343291899e-07, "loss": 0.1274, "step": 27370 }, { "epoch": 0.5368627450980392, "grad_norm": 4.1967949867248535, "learning_rate": 8.75141283705582e-07, "loss": 0.1618, "step": 27380 }, { "epoch": 0.5370588235294118, "grad_norm": 3.279972553253174, "learning_rate": 8.750340936501336e-07, "loss": 0.1551, "step": 27390 }, { "epoch": 0.5372549019607843, "grad_norm": 7.0490193367004395, "learning_rate": 8.749268641741118e-07, "loss": 0.1807, "step": 27400 }, { "epoch": 0.5374509803921569, "grad_norm": 2.7205073833465576, "learning_rate": 8.748195952887875e-07, "loss": 0.1927, "step": 27410 }, { "epoch": 0.5376470588235294, "grad_norm": 2.541997194290161, "learning_rate": 8.747122870054364e-07, "loss": 0.1666, "step": 27420 }, { "epoch": 0.537843137254902, "grad_norm": 1.8604172468185425, "learning_rate": 8.746049393353373e-07, "loss": 0.1735, "step": 27430 }, { "epoch": 0.5380392156862746, "grad_norm": 2.6667592525482178, "learning_rate": 8.74497552289774e-07, "loss": 0.1552, "step": 27440 }, { "epoch": 0.538235294117647, "grad_norm": 2.574878454208374, "learning_rate": 8.743901258800342e-07, "loss": 0.1477, "step": 27450 }, { "epoch": 0.5384313725490196, "grad_norm": 4.25118350982666, "learning_rate": 8.742826601174097e-07, "loss": 0.1981, "step": 27460 }, { "epoch": 0.5386274509803921, "grad_norm": 2.4985084533691406, "learning_rate": 8.741751550131964e-07, "loss": 0.1537, "step": 27470 }, { "epoch": 0.5388235294117647, "grad_norm": 2.811323881149292, "learning_rate": 8.740676105786941e-07, "loss": 0.1789, "step": 27480 }, { "epoch": 0.5390196078431373, "grad_norm": 1.5278515815734863, "learning_rate": 8.739600268252076e-07, "loss": 0.1668, "step": 27490 }, { "epoch": 0.5392156862745098, "grad_norm": 3.5678963661193848, "learning_rate": 8.738524037640448e-07, "loss": 0.1455, "step": 27500 }, { "epoch": 0.5394117647058824, "grad_norm": 1.4811186790466309, "learning_rate": 8.737447414065182e-07, "loss": 0.1337, "step": 27510 }, { "epoch": 0.5396078431372549, "grad_norm": 2.857637882232666, "learning_rate": 8.736370397639445e-07, "loss": 0.157, "step": 27520 }, { "epoch": 0.5398039215686274, "grad_norm": 3.9997737407684326, "learning_rate": 8.735292988476446e-07, "loss": 0.1953, "step": 27530 }, { "epoch": 0.54, "grad_norm": 2.671470880508423, "learning_rate": 8.734215186689431e-07, "loss": 0.1656, "step": 27540 }, { "epoch": 0.5401960784313725, "grad_norm": 3.5651917457580566, "learning_rate": 8.73313699239169e-07, "loss": 0.1768, "step": 27550 }, { "epoch": 0.5403921568627451, "grad_norm": 2.508708953857422, "learning_rate": 8.732058405696555e-07, "loss": 0.1779, "step": 27560 }, { "epoch": 0.5405882352941176, "grad_norm": 1.7881042957305908, "learning_rate": 8.730979426717398e-07, "loss": 0.1599, "step": 27570 }, { "epoch": 0.5407843137254902, "grad_norm": 2.78855299949646, "learning_rate": 8.729900055567631e-07, "loss": 0.1684, "step": 27580 }, { "epoch": 0.5409803921568628, "grad_norm": 3.806973457336426, "learning_rate": 8.728820292360712e-07, "loss": 0.1577, "step": 27590 }, { "epoch": 0.5411764705882353, "grad_norm": 4.120628356933594, "learning_rate": 8.727740137210138e-07, "loss": 0.1719, "step": 27600 }, { "epoch": 0.5413725490196079, "grad_norm": 2.3373851776123047, "learning_rate": 8.726659590229441e-07, "loss": 0.1779, "step": 27610 }, { "epoch": 0.5415686274509804, "grad_norm": 1.999319314956665, "learning_rate": 8.725578651532204e-07, "loss": 0.169, "step": 27620 }, { "epoch": 0.5417647058823529, "grad_norm": 5.617596626281738, "learning_rate": 8.724497321232043e-07, "loss": 0.1641, "step": 27630 }, { "epoch": 0.5419607843137255, "grad_norm": 2.3475608825683594, "learning_rate": 8.723415599442623e-07, "loss": 0.147, "step": 27640 }, { "epoch": 0.542156862745098, "grad_norm": 2.909273147583008, "learning_rate": 8.722333486277642e-07, "loss": 0.1411, "step": 27650 }, { "epoch": 0.5423529411764706, "grad_norm": 2.9045569896698, "learning_rate": 8.721250981850845e-07, "loss": 0.151, "step": 27660 }, { "epoch": 0.5425490196078432, "grad_norm": 4.987287998199463, "learning_rate": 8.720168086276015e-07, "loss": 0.1534, "step": 27670 }, { "epoch": 0.5427450980392157, "grad_norm": 2.375915288925171, "learning_rate": 8.71908479966698e-07, "loss": 0.1717, "step": 27680 }, { "epoch": 0.5429411764705883, "grad_norm": 2.6611034870147705, "learning_rate": 8.718001122137602e-07, "loss": 0.1981, "step": 27690 }, { "epoch": 0.5431372549019607, "grad_norm": 2.484426498413086, "learning_rate": 8.716917053801793e-07, "loss": 0.1574, "step": 27700 }, { "epoch": 0.5433333333333333, "grad_norm": 5.695218086242676, "learning_rate": 8.715832594773497e-07, "loss": 0.151, "step": 27710 }, { "epoch": 0.5435294117647059, "grad_norm": 2.7504563331604004, "learning_rate": 8.71474774516671e-07, "loss": 0.1427, "step": 27720 }, { "epoch": 0.5437254901960784, "grad_norm": 1.6659337282180786, "learning_rate": 8.713662505095456e-07, "loss": 0.1498, "step": 27730 }, { "epoch": 0.543921568627451, "grad_norm": 3.993114709854126, "learning_rate": 8.712576874673812e-07, "loss": 0.1624, "step": 27740 }, { "epoch": 0.5441176470588235, "grad_norm": 2.4371538162231445, "learning_rate": 8.711490854015886e-07, "loss": 0.1554, "step": 27750 }, { "epoch": 0.5443137254901961, "grad_norm": 2.9397597312927246, "learning_rate": 8.710404443235835e-07, "loss": 0.1596, "step": 27760 }, { "epoch": 0.5445098039215687, "grad_norm": 5.076921463012695, "learning_rate": 8.709317642447852e-07, "loss": 0.1544, "step": 27770 }, { "epoch": 0.5447058823529412, "grad_norm": 2.066572427749634, "learning_rate": 8.708230451766174e-07, "loss": 0.1887, "step": 27780 }, { "epoch": 0.5449019607843137, "grad_norm": 2.522334575653076, "learning_rate": 8.707142871305078e-07, "loss": 0.1754, "step": 27790 }, { "epoch": 0.5450980392156862, "grad_norm": 1.5990687608718872, "learning_rate": 8.706054901178881e-07, "loss": 0.1674, "step": 27800 }, { "epoch": 0.5452941176470588, "grad_norm": 2.349374294281006, "learning_rate": 8.704966541501941e-07, "loss": 0.1558, "step": 27810 }, { "epoch": 0.5454901960784314, "grad_norm": 2.606538772583008, "learning_rate": 8.703877792388659e-07, "loss": 0.1578, "step": 27820 }, { "epoch": 0.5456862745098039, "grad_norm": 3.3815293312072754, "learning_rate": 8.702788653953476e-07, "loss": 0.1809, "step": 27830 }, { "epoch": 0.5458823529411765, "grad_norm": 2.331704616546631, "learning_rate": 8.70169912631087e-07, "loss": 0.1365, "step": 27840 }, { "epoch": 0.546078431372549, "grad_norm": 3.1907944679260254, "learning_rate": 8.700609209575369e-07, "loss": 0.2046, "step": 27850 }, { "epoch": 0.5462745098039216, "grad_norm": 2.5974299907684326, "learning_rate": 8.699518903861531e-07, "loss": 0.1647, "step": 27860 }, { "epoch": 0.5464705882352942, "grad_norm": 2.0925090312957764, "learning_rate": 8.698428209283963e-07, "loss": 0.1827, "step": 27870 }, { "epoch": 0.5466666666666666, "grad_norm": 1.7472610473632812, "learning_rate": 8.69733712595731e-07, "loss": 0.164, "step": 27880 }, { "epoch": 0.5468627450980392, "grad_norm": 2.4538331031799316, "learning_rate": 8.696245653996257e-07, "loss": 0.1788, "step": 27890 }, { "epoch": 0.5470588235294118, "grad_norm": 4.6579670906066895, "learning_rate": 8.695153793515532e-07, "loss": 0.1766, "step": 27900 }, { "epoch": 0.5472549019607843, "grad_norm": 3.46321177482605, "learning_rate": 8.694061544629901e-07, "loss": 0.1797, "step": 27910 }, { "epoch": 0.5474509803921569, "grad_norm": 2.254951000213623, "learning_rate": 8.692968907454172e-07, "loss": 0.1442, "step": 27920 }, { "epoch": 0.5476470588235294, "grad_norm": 4.426253795623779, "learning_rate": 8.691875882103197e-07, "loss": 0.1721, "step": 27930 }, { "epoch": 0.547843137254902, "grad_norm": 3.620758295059204, "learning_rate": 8.690782468691864e-07, "loss": 0.1451, "step": 27940 }, { "epoch": 0.5480392156862746, "grad_norm": 2.226604461669922, "learning_rate": 8.689688667335104e-07, "loss": 0.1693, "step": 27950 }, { "epoch": 0.548235294117647, "grad_norm": 1.9758752584457397, "learning_rate": 8.688594478147891e-07, "loss": 0.1354, "step": 27960 }, { "epoch": 0.5484313725490196, "grad_norm": 3.083063840866089, "learning_rate": 8.687499901245232e-07, "loss": 0.1882, "step": 27970 }, { "epoch": 0.5486274509803921, "grad_norm": 4.171585559844971, "learning_rate": 8.686404936742184e-07, "loss": 0.1806, "step": 27980 }, { "epoch": 0.5488235294117647, "grad_norm": 3.0237748622894287, "learning_rate": 8.685309584753843e-07, "loss": 0.1309, "step": 27990 }, { "epoch": 0.5490196078431373, "grad_norm": 2.3221280574798584, "learning_rate": 8.684213845395338e-07, "loss": 0.1513, "step": 28000 }, { "epoch": 0.5492156862745098, "grad_norm": 4.328914165496826, "learning_rate": 8.68311771878185e-07, "loss": 0.1554, "step": 28010 }, { "epoch": 0.5494117647058824, "grad_norm": 5.749283313751221, "learning_rate": 8.68202120502859e-07, "loss": 0.1379, "step": 28020 }, { "epoch": 0.5496078431372549, "grad_norm": 2.7608449459075928, "learning_rate": 8.680924304250819e-07, "loss": 0.1657, "step": 28030 }, { "epoch": 0.5498039215686275, "grad_norm": 2.830470085144043, "learning_rate": 8.679827016563831e-07, "loss": 0.1562, "step": 28040 }, { "epoch": 0.55, "grad_norm": 3.0933523178100586, "learning_rate": 8.678729342082967e-07, "loss": 0.149, "step": 28050 }, { "epoch": 0.5501960784313725, "grad_norm": 2.2099318504333496, "learning_rate": 8.677631280923602e-07, "loss": 0.1555, "step": 28060 }, { "epoch": 0.5503921568627451, "grad_norm": 3.032026767730713, "learning_rate": 8.676532833201159e-07, "loss": 0.1518, "step": 28070 }, { "epoch": 0.5505882352941176, "grad_norm": 1.7762701511383057, "learning_rate": 8.675433999031096e-07, "loss": 0.1698, "step": 28080 }, { "epoch": 0.5507843137254902, "grad_norm": 2.3775627613067627, "learning_rate": 8.674334778528916e-07, "loss": 0.1664, "step": 28090 }, { "epoch": 0.5509803921568628, "grad_norm": 2.6225550174713135, "learning_rate": 8.673235171810156e-07, "loss": 0.1634, "step": 28100 }, { "epoch": 0.5511764705882353, "grad_norm": 1.4714066982269287, "learning_rate": 8.672135178990401e-07, "loss": 0.1389, "step": 28110 }, { "epoch": 0.5513725490196079, "grad_norm": 2.3510444164276123, "learning_rate": 8.671034800185274e-07, "loss": 0.1678, "step": 28120 }, { "epoch": 0.5515686274509803, "grad_norm": 4.445239543914795, "learning_rate": 8.669934035510435e-07, "loss": 0.1718, "step": 28130 }, { "epoch": 0.5517647058823529, "grad_norm": 2.9841229915618896, "learning_rate": 8.668832885081592e-07, "loss": 0.1816, "step": 28140 }, { "epoch": 0.5519607843137255, "grad_norm": 3.5224244594573975, "learning_rate": 8.667731349014483e-07, "loss": 0.1656, "step": 28150 }, { "epoch": 0.552156862745098, "grad_norm": 2.7372887134552, "learning_rate": 8.666629427424898e-07, "loss": 0.1444, "step": 28160 }, { "epoch": 0.5523529411764706, "grad_norm": 2.47469162940979, "learning_rate": 8.66552712042866e-07, "loss": 0.1529, "step": 28170 }, { "epoch": 0.5525490196078431, "grad_norm": 8.365970611572266, "learning_rate": 8.664424428141635e-07, "loss": 0.1697, "step": 28180 }, { "epoch": 0.5527450980392157, "grad_norm": 3.3649165630340576, "learning_rate": 8.663321350679728e-07, "loss": 0.1859, "step": 28190 }, { "epoch": 0.5529411764705883, "grad_norm": 1.9823707342147827, "learning_rate": 8.662217888158887e-07, "loss": 0.1865, "step": 28200 }, { "epoch": 0.5531372549019608, "grad_norm": 1.7567764520645142, "learning_rate": 8.661114040695099e-07, "loss": 0.1702, "step": 28210 }, { "epoch": 0.5533333333333333, "grad_norm": 4.412212371826172, "learning_rate": 8.660009808404391e-07, "loss": 0.1586, "step": 28220 }, { "epoch": 0.5535294117647059, "grad_norm": 2.297511100769043, "learning_rate": 8.658905191402832e-07, "loss": 0.1491, "step": 28230 }, { "epoch": 0.5537254901960784, "grad_norm": 3.0087502002716064, "learning_rate": 8.657800189806529e-07, "loss": 0.1679, "step": 28240 }, { "epoch": 0.553921568627451, "grad_norm": 2.791802167892456, "learning_rate": 8.656694803731632e-07, "loss": 0.1622, "step": 28250 }, { "epoch": 0.5541176470588235, "grad_norm": 2.7099428176879883, "learning_rate": 8.65558903329433e-07, "loss": 0.1483, "step": 28260 }, { "epoch": 0.5543137254901961, "grad_norm": 3.132108211517334, "learning_rate": 8.654482878610854e-07, "loss": 0.1825, "step": 28270 }, { "epoch": 0.5545098039215687, "grad_norm": 3.3306021690368652, "learning_rate": 8.653376339797472e-07, "loss": 0.1424, "step": 28280 }, { "epoch": 0.5547058823529412, "grad_norm": 1.897638201713562, "learning_rate": 8.652269416970493e-07, "loss": 0.138, "step": 28290 }, { "epoch": 0.5549019607843138, "grad_norm": 2.975782871246338, "learning_rate": 8.651162110246273e-07, "loss": 0.1499, "step": 28300 }, { "epoch": 0.5550980392156862, "grad_norm": 4.619785785675049, "learning_rate": 8.6500544197412e-07, "loss": 0.1471, "step": 28310 }, { "epoch": 0.5552941176470588, "grad_norm": 2.7451863288879395, "learning_rate": 8.648946345571706e-07, "loss": 0.1454, "step": 28320 }, { "epoch": 0.5554901960784314, "grad_norm": 2.993403434753418, "learning_rate": 8.647837887854261e-07, "loss": 0.133, "step": 28330 }, { "epoch": 0.5556862745098039, "grad_norm": 1.5596566200256348, "learning_rate": 8.646729046705381e-07, "loss": 0.1551, "step": 28340 }, { "epoch": 0.5558823529411765, "grad_norm": 2.3759825229644775, "learning_rate": 8.645619822241615e-07, "loss": 0.1548, "step": 28350 }, { "epoch": 0.556078431372549, "grad_norm": 2.234107494354248, "learning_rate": 8.644510214579559e-07, "loss": 0.1415, "step": 28360 }, { "epoch": 0.5562745098039216, "grad_norm": 2.441697835922241, "learning_rate": 8.643400223835842e-07, "loss": 0.1616, "step": 28370 }, { "epoch": 0.5564705882352942, "grad_norm": 1.8736451864242554, "learning_rate": 8.642289850127141e-07, "loss": 0.1374, "step": 28380 }, { "epoch": 0.5566666666666666, "grad_norm": 2.0731050968170166, "learning_rate": 8.641179093570167e-07, "loss": 0.1507, "step": 28390 }, { "epoch": 0.5568627450980392, "grad_norm": 1.5026859045028687, "learning_rate": 8.640067954281677e-07, "loss": 0.1567, "step": 28400 }, { "epoch": 0.5570588235294117, "grad_norm": 1.961782693862915, "learning_rate": 8.638956432378462e-07, "loss": 0.1432, "step": 28410 }, { "epoch": 0.5572549019607843, "grad_norm": 4.809487819671631, "learning_rate": 8.637844527977355e-07, "loss": 0.1652, "step": 28420 }, { "epoch": 0.5574509803921569, "grad_norm": 2.0447843074798584, "learning_rate": 8.636732241195237e-07, "loss": 0.1516, "step": 28430 }, { "epoch": 0.5576470588235294, "grad_norm": 1.8143125772476196, "learning_rate": 8.635619572149015e-07, "loss": 0.149, "step": 28440 }, { "epoch": 0.557843137254902, "grad_norm": 1.8403927087783813, "learning_rate": 8.634506520955649e-07, "loss": 0.1075, "step": 28450 }, { "epoch": 0.5580392156862745, "grad_norm": 2.0635406970977783, "learning_rate": 8.63339308773213e-07, "loss": 0.1673, "step": 28460 }, { "epoch": 0.558235294117647, "grad_norm": 1.7096974849700928, "learning_rate": 8.632279272595498e-07, "loss": 0.1376, "step": 28470 }, { "epoch": 0.5584313725490196, "grad_norm": 2.2018487453460693, "learning_rate": 8.631165075662825e-07, "loss": 0.1721, "step": 28480 }, { "epoch": 0.5586274509803921, "grad_norm": 4.786693572998047, "learning_rate": 8.630050497051227e-07, "loss": 0.1474, "step": 28490 }, { "epoch": 0.5588235294117647, "grad_norm": 3.096879005432129, "learning_rate": 8.628935536877858e-07, "loss": 0.169, "step": 28500 }, { "epoch": 0.5590196078431372, "grad_norm": 3.251723527908325, "learning_rate": 8.627820195259917e-07, "loss": 0.1779, "step": 28510 }, { "epoch": 0.5592156862745098, "grad_norm": 5.452677249908447, "learning_rate": 8.626704472314639e-07, "loss": 0.1536, "step": 28520 }, { "epoch": 0.5594117647058824, "grad_norm": 2.8416314125061035, "learning_rate": 8.625588368159298e-07, "loss": 0.1518, "step": 28530 }, { "epoch": 0.5596078431372549, "grad_norm": 2.157465934753418, "learning_rate": 8.62447188291121e-07, "loss": 0.166, "step": 28540 }, { "epoch": 0.5598039215686275, "grad_norm": 2.3529000282287598, "learning_rate": 8.623355016687732e-07, "loss": 0.1433, "step": 28550 }, { "epoch": 0.56, "grad_norm": 2.6963343620300293, "learning_rate": 8.622237769606259e-07, "loss": 0.146, "step": 28560 }, { "epoch": 0.5601960784313725, "grad_norm": 2.781414031982422, "learning_rate": 8.621120141784231e-07, "loss": 0.1557, "step": 28570 }, { "epoch": 0.5603921568627451, "grad_norm": 3.687610626220703, "learning_rate": 8.620002133339119e-07, "loss": 0.1719, "step": 28580 }, { "epoch": 0.5605882352941176, "grad_norm": 1.9843610525131226, "learning_rate": 8.618883744388441e-07, "loss": 0.1669, "step": 28590 }, { "epoch": 0.5607843137254902, "grad_norm": 1.8599961996078491, "learning_rate": 8.617764975049752e-07, "loss": 0.154, "step": 28600 }, { "epoch": 0.5609803921568628, "grad_norm": 2.9777090549468994, "learning_rate": 8.616645825440651e-07, "loss": 0.1527, "step": 28610 }, { "epoch": 0.5611764705882353, "grad_norm": 4.071366786956787, "learning_rate": 8.615526295678772e-07, "loss": 0.1426, "step": 28620 }, { "epoch": 0.5613725490196079, "grad_norm": 3.5675251483917236, "learning_rate": 8.614406385881791e-07, "loss": 0.1543, "step": 28630 }, { "epoch": 0.5615686274509804, "grad_norm": 3.9028961658477783, "learning_rate": 8.613286096167423e-07, "loss": 0.1782, "step": 28640 }, { "epoch": 0.5617647058823529, "grad_norm": 2.1214044094085693, "learning_rate": 8.612165426653426e-07, "loss": 0.1452, "step": 28650 }, { "epoch": 0.5619607843137255, "grad_norm": 2.770949363708496, "learning_rate": 8.611044377457595e-07, "loss": 0.1899, "step": 28660 }, { "epoch": 0.562156862745098, "grad_norm": 2.571028232574463, "learning_rate": 8.609922948697763e-07, "loss": 0.1634, "step": 28670 }, { "epoch": 0.5623529411764706, "grad_norm": 2.343104124069214, "learning_rate": 8.60880114049181e-07, "loss": 0.1741, "step": 28680 }, { "epoch": 0.5625490196078431, "grad_norm": 2.938167095184326, "learning_rate": 8.607678952957649e-07, "loss": 0.1615, "step": 28690 }, { "epoch": 0.5627450980392157, "grad_norm": 2.2470591068267822, "learning_rate": 8.606556386213237e-07, "loss": 0.1873, "step": 28700 }, { "epoch": 0.5629411764705883, "grad_norm": 2.7789947986602783, "learning_rate": 8.605433440376567e-07, "loss": 0.1281, "step": 28710 }, { "epoch": 0.5631372549019608, "grad_norm": 2.9975929260253906, "learning_rate": 8.604310115565676e-07, "loss": 0.1496, "step": 28720 }, { "epoch": 0.5633333333333334, "grad_norm": 2.75927996635437, "learning_rate": 8.603186411898636e-07, "loss": 0.191, "step": 28730 }, { "epoch": 0.5635294117647058, "grad_norm": 2.3570969104766846, "learning_rate": 8.602062329493564e-07, "loss": 0.192, "step": 28740 }, { "epoch": 0.5637254901960784, "grad_norm": 3.0211095809936523, "learning_rate": 8.600937868468616e-07, "loss": 0.1584, "step": 28750 }, { "epoch": 0.563921568627451, "grad_norm": 2.603473424911499, "learning_rate": 8.599813028941983e-07, "loss": 0.1821, "step": 28760 }, { "epoch": 0.5641176470588235, "grad_norm": 3.7477614879608154, "learning_rate": 8.598687811031902e-07, "loss": 0.1745, "step": 28770 }, { "epoch": 0.5643137254901961, "grad_norm": 10.811502456665039, "learning_rate": 8.597562214856644e-07, "loss": 0.1793, "step": 28780 }, { "epoch": 0.5645098039215686, "grad_norm": 2.496785879135132, "learning_rate": 8.596436240534525e-07, "loss": 0.1526, "step": 28790 }, { "epoch": 0.5647058823529412, "grad_norm": 2.4662280082702637, "learning_rate": 8.595309888183897e-07, "loss": 0.1524, "step": 28800 }, { "epoch": 0.5649019607843138, "grad_norm": 3.516436815261841, "learning_rate": 8.594183157923154e-07, "loss": 0.1787, "step": 28810 }, { "epoch": 0.5650980392156862, "grad_norm": 5.330949306488037, "learning_rate": 8.593056049870729e-07, "loss": 0.1692, "step": 28820 }, { "epoch": 0.5652941176470588, "grad_norm": 2.344480276107788, "learning_rate": 8.591928564145092e-07, "loss": 0.1507, "step": 28830 }, { "epoch": 0.5654901960784313, "grad_norm": 4.207390308380127, "learning_rate": 8.59080070086476e-07, "loss": 0.1576, "step": 28840 }, { "epoch": 0.5656862745098039, "grad_norm": 1.9014464616775513, "learning_rate": 8.589672460148281e-07, "loss": 0.1459, "step": 28850 }, { "epoch": 0.5658823529411765, "grad_norm": 1.637876272201538, "learning_rate": 8.588543842114247e-07, "loss": 0.1263, "step": 28860 }, { "epoch": 0.566078431372549, "grad_norm": 1.3130571842193604, "learning_rate": 8.587414846881289e-07, "loss": 0.184, "step": 28870 }, { "epoch": 0.5662745098039216, "grad_norm": 2.6453332901000977, "learning_rate": 8.586285474568082e-07, "loss": 0.1694, "step": 28880 }, { "epoch": 0.5664705882352942, "grad_norm": 3.704343795776367, "learning_rate": 8.585155725293332e-07, "loss": 0.1888, "step": 28890 }, { "epoch": 0.5666666666666667, "grad_norm": 2.194333076477051, "learning_rate": 8.58402559917579e-07, "loss": 0.1401, "step": 28900 }, { "epoch": 0.5668627450980392, "grad_norm": 3.2118659019470215, "learning_rate": 8.582895096334249e-07, "loss": 0.2024, "step": 28910 }, { "epoch": 0.5670588235294117, "grad_norm": 3.102886199951172, "learning_rate": 8.581764216887533e-07, "loss": 0.135, "step": 28920 }, { "epoch": 0.5672549019607843, "grad_norm": 2.247089385986328, "learning_rate": 8.580632960954514e-07, "loss": 0.1777, "step": 28930 }, { "epoch": 0.5674509803921569, "grad_norm": 2.0091726779937744, "learning_rate": 8.579501328654102e-07, "loss": 0.1779, "step": 28940 }, { "epoch": 0.5676470588235294, "grad_norm": 3.267669439315796, "learning_rate": 8.578369320105243e-07, "loss": 0.1689, "step": 28950 }, { "epoch": 0.567843137254902, "grad_norm": 2.153446674346924, "learning_rate": 8.577236935426926e-07, "loss": 0.1857, "step": 28960 }, { "epoch": 0.5680392156862745, "grad_norm": 2.6648213863372803, "learning_rate": 8.576104174738177e-07, "loss": 0.1991, "step": 28970 }, { "epoch": 0.5682352941176471, "grad_norm": 1.4322328567504883, "learning_rate": 8.57497103815806e-07, "loss": 0.1691, "step": 28980 }, { "epoch": 0.5684313725490197, "grad_norm": 1.3948578834533691, "learning_rate": 8.57383752580569e-07, "loss": 0.1329, "step": 28990 }, { "epoch": 0.5686274509803921, "grad_norm": 4.084217071533203, "learning_rate": 8.572703637800202e-07, "loss": 0.1826, "step": 29000 }, { "epoch": 0.5688235294117647, "grad_norm": 2.054375171661377, "learning_rate": 8.571569374260787e-07, "loss": 0.149, "step": 29010 }, { "epoch": 0.5690196078431372, "grad_norm": 6.620234489440918, "learning_rate": 8.57043473530667e-07, "loss": 0.2096, "step": 29020 }, { "epoch": 0.5692156862745098, "grad_norm": 1.8208872079849243, "learning_rate": 8.569299721057115e-07, "loss": 0.1424, "step": 29030 }, { "epoch": 0.5694117647058824, "grad_norm": 4.521175384521484, "learning_rate": 8.568164331631424e-07, "loss": 0.1514, "step": 29040 }, { "epoch": 0.5696078431372549, "grad_norm": 2.4284210205078125, "learning_rate": 8.56702856714894e-07, "loss": 0.1644, "step": 29050 }, { "epoch": 0.5698039215686275, "grad_norm": 8.112000465393066, "learning_rate": 8.565892427729046e-07, "loss": 0.1487, "step": 29060 }, { "epoch": 0.57, "grad_norm": 3.224644899368286, "learning_rate": 8.564755913491164e-07, "loss": 0.1834, "step": 29070 }, { "epoch": 0.5701960784313725, "grad_norm": 3.012273073196411, "learning_rate": 8.563619024554756e-07, "loss": 0.1417, "step": 29080 }, { "epoch": 0.5703921568627451, "grad_norm": 1.5506031513214111, "learning_rate": 8.56248176103932e-07, "loss": 0.1727, "step": 29090 }, { "epoch": 0.5705882352941176, "grad_norm": 3.227888822555542, "learning_rate": 8.561344123064398e-07, "loss": 0.1736, "step": 29100 }, { "epoch": 0.5707843137254902, "grad_norm": 4.635955810546875, "learning_rate": 8.560206110749571e-07, "loss": 0.1497, "step": 29110 }, { "epoch": 0.5709803921568627, "grad_norm": 3.529073715209961, "learning_rate": 8.559067724214455e-07, "loss": 0.152, "step": 29120 }, { "epoch": 0.5711764705882353, "grad_norm": 2.3946664333343506, "learning_rate": 8.557928963578709e-07, "loss": 0.1604, "step": 29130 }, { "epoch": 0.5713725490196079, "grad_norm": 2.819348096847534, "learning_rate": 8.55678982896203e-07, "loss": 0.1503, "step": 29140 }, { "epoch": 0.5715686274509804, "grad_norm": 2.3685519695281982, "learning_rate": 8.555650320484156e-07, "loss": 0.1755, "step": 29150 }, { "epoch": 0.571764705882353, "grad_norm": 4.559390544891357, "learning_rate": 8.554510438264861e-07, "loss": 0.1742, "step": 29160 }, { "epoch": 0.5719607843137255, "grad_norm": 2.1890664100646973, "learning_rate": 8.553370182423963e-07, "loss": 0.1451, "step": 29170 }, { "epoch": 0.572156862745098, "grad_norm": 4.653532028198242, "learning_rate": 8.552229553081311e-07, "loss": 0.1666, "step": 29180 }, { "epoch": 0.5723529411764706, "grad_norm": 10.494361877441406, "learning_rate": 8.551088550356806e-07, "loss": 0.1826, "step": 29190 }, { "epoch": 0.5725490196078431, "grad_norm": 3.9566650390625, "learning_rate": 8.549947174370377e-07, "loss": 0.1492, "step": 29200 }, { "epoch": 0.5727450980392157, "grad_norm": 3.047511339187622, "learning_rate": 8.548805425241997e-07, "loss": 0.1543, "step": 29210 }, { "epoch": 0.5729411764705883, "grad_norm": 2.9922850131988525, "learning_rate": 8.547663303091677e-07, "loss": 0.1766, "step": 29220 }, { "epoch": 0.5731372549019608, "grad_norm": 4.698017120361328, "learning_rate": 8.546520808039468e-07, "loss": 0.1735, "step": 29230 }, { "epoch": 0.5733333333333334, "grad_norm": 4.232255458831787, "learning_rate": 8.545377940205462e-07, "loss": 0.1811, "step": 29240 }, { "epoch": 0.5735294117647058, "grad_norm": 2.1715731620788574, "learning_rate": 8.544234699709784e-07, "loss": 0.1687, "step": 29250 }, { "epoch": 0.5737254901960784, "grad_norm": 4.824649333953857, "learning_rate": 8.543091086672605e-07, "loss": 0.1944, "step": 29260 }, { "epoch": 0.573921568627451, "grad_norm": 2.795924425125122, "learning_rate": 8.541947101214132e-07, "loss": 0.1553, "step": 29270 }, { "epoch": 0.5741176470588235, "grad_norm": 2.3069405555725098, "learning_rate": 8.540802743454611e-07, "loss": 0.1579, "step": 29280 }, { "epoch": 0.5743137254901961, "grad_norm": 1.992737054824829, "learning_rate": 8.539658013514328e-07, "loss": 0.1606, "step": 29290 }, { "epoch": 0.5745098039215686, "grad_norm": 4.188408851623535, "learning_rate": 8.538512911513609e-07, "loss": 0.194, "step": 29300 }, { "epoch": 0.5747058823529412, "grad_norm": 4.355842113494873, "learning_rate": 8.537367437572816e-07, "loss": 0.1992, "step": 29310 }, { "epoch": 0.5749019607843138, "grad_norm": 2.210657835006714, "learning_rate": 8.536221591812354e-07, "loss": 0.1545, "step": 29320 }, { "epoch": 0.5750980392156863, "grad_norm": 2.822265625, "learning_rate": 8.535075374352663e-07, "loss": 0.1822, "step": 29330 }, { "epoch": 0.5752941176470588, "grad_norm": 2.068183422088623, "learning_rate": 8.533928785314226e-07, "loss": 0.1437, "step": 29340 }, { "epoch": 0.5754901960784313, "grad_norm": 4.277740001678467, "learning_rate": 8.532781824817563e-07, "loss": 0.136, "step": 29350 }, { "epoch": 0.5756862745098039, "grad_norm": 3.6549232006073, "learning_rate": 8.531634492983231e-07, "loss": 0.2063, "step": 29360 }, { "epoch": 0.5758823529411765, "grad_norm": 3.1171586513519287, "learning_rate": 8.530486789931831e-07, "loss": 0.1501, "step": 29370 }, { "epoch": 0.576078431372549, "grad_norm": 3.621814012527466, "learning_rate": 8.529338715783998e-07, "loss": 0.186, "step": 29380 }, { "epoch": 0.5762745098039216, "grad_norm": 2.855147361755371, "learning_rate": 8.528190270660411e-07, "loss": 0.1639, "step": 29390 }, { "epoch": 0.5764705882352941, "grad_norm": 3.2005326747894287, "learning_rate": 8.527041454681785e-07, "loss": 0.1546, "step": 29400 }, { "epoch": 0.5766666666666667, "grad_norm": 2.3677964210510254, "learning_rate": 8.525892267968872e-07, "loss": 0.1689, "step": 29410 }, { "epoch": 0.5768627450980393, "grad_norm": 2.323842763900757, "learning_rate": 8.524742710642466e-07, "loss": 0.1633, "step": 29420 }, { "epoch": 0.5770588235294117, "grad_norm": 2.7439565658569336, "learning_rate": 8.523592782823401e-07, "loss": 0.1322, "step": 29430 }, { "epoch": 0.5772549019607843, "grad_norm": 2.8609211444854736, "learning_rate": 8.522442484632547e-07, "loss": 0.1587, "step": 29440 }, { "epoch": 0.5774509803921568, "grad_norm": 1.8958240747451782, "learning_rate": 8.521291816190814e-07, "loss": 0.1626, "step": 29450 }, { "epoch": 0.5776470588235294, "grad_norm": 2.3701398372650146, "learning_rate": 8.520140777619151e-07, "loss": 0.1587, "step": 29460 }, { "epoch": 0.577843137254902, "grad_norm": 2.031890630722046, "learning_rate": 8.518989369038545e-07, "loss": 0.1337, "step": 29470 }, { "epoch": 0.5780392156862745, "grad_norm": 3.365729331970215, "learning_rate": 8.517837590570026e-07, "loss": 0.1637, "step": 29480 }, { "epoch": 0.5782352941176471, "grad_norm": 4.531546115875244, "learning_rate": 8.516685442334656e-07, "loss": 0.1507, "step": 29490 }, { "epoch": 0.5784313725490197, "grad_norm": 2.7437312602996826, "learning_rate": 8.515532924453544e-07, "loss": 0.1649, "step": 29500 }, { "epoch": 0.5786274509803921, "grad_norm": 4.190512180328369, "learning_rate": 8.514380037047828e-07, "loss": 0.162, "step": 29510 }, { "epoch": 0.5788235294117647, "grad_norm": 2.76076602935791, "learning_rate": 8.513226780238694e-07, "loss": 0.1723, "step": 29520 }, { "epoch": 0.5790196078431372, "grad_norm": 2.8702638149261475, "learning_rate": 8.512073154147361e-07, "loss": 0.1365, "step": 29530 }, { "epoch": 0.5792156862745098, "grad_norm": 2.936894178390503, "learning_rate": 8.510919158895091e-07, "loss": 0.176, "step": 29540 }, { "epoch": 0.5794117647058824, "grad_norm": 2.371300220489502, "learning_rate": 8.50976479460318e-07, "loss": 0.1575, "step": 29550 }, { "epoch": 0.5796078431372549, "grad_norm": 2.4345366954803467, "learning_rate": 8.508610061392969e-07, "loss": 0.1572, "step": 29560 }, { "epoch": 0.5798039215686275, "grad_norm": 2.06449031829834, "learning_rate": 8.507454959385833e-07, "loss": 0.1284, "step": 29570 }, { "epoch": 0.58, "grad_norm": 3.6100850105285645, "learning_rate": 8.506299488703185e-07, "loss": 0.1579, "step": 29580 }, { "epoch": 0.5801960784313726, "grad_norm": 1.887877106666565, "learning_rate": 8.505143649466481e-07, "loss": 0.1628, "step": 29590 }, { "epoch": 0.5803921568627451, "grad_norm": 1.933394193649292, "learning_rate": 8.503987441797213e-07, "loss": 0.157, "step": 29600 }, { "epoch": 0.5805882352941176, "grad_norm": 1.7899459600448608, "learning_rate": 8.502830865816911e-07, "loss": 0.1464, "step": 29610 }, { "epoch": 0.5807843137254902, "grad_norm": 2.3671348094940186, "learning_rate": 8.501673921647148e-07, "loss": 0.1772, "step": 29620 }, { "epoch": 0.5809803921568627, "grad_norm": 2.933117151260376, "learning_rate": 8.500516609409529e-07, "loss": 0.1452, "step": 29630 }, { "epoch": 0.5811764705882353, "grad_norm": 3.253150224685669, "learning_rate": 8.499358929225704e-07, "loss": 0.1371, "step": 29640 }, { "epoch": 0.5813725490196079, "grad_norm": 3.2720768451690674, "learning_rate": 8.498200881217357e-07, "loss": 0.1574, "step": 29650 }, { "epoch": 0.5815686274509804, "grad_norm": 2.8682451248168945, "learning_rate": 8.497042465506215e-07, "loss": 0.1618, "step": 29660 }, { "epoch": 0.581764705882353, "grad_norm": 1.1741405725479126, "learning_rate": 8.49588368221404e-07, "loss": 0.174, "step": 29670 }, { "epoch": 0.5819607843137254, "grad_norm": 3.3252382278442383, "learning_rate": 8.494724531462636e-07, "loss": 0.1606, "step": 29680 }, { "epoch": 0.582156862745098, "grad_norm": 3.195183277130127, "learning_rate": 8.493565013373841e-07, "loss": 0.1537, "step": 29690 }, { "epoch": 0.5823529411764706, "grad_norm": 0.7661805748939514, "learning_rate": 8.492405128069533e-07, "loss": 0.1665, "step": 29700 }, { "epoch": 0.5825490196078431, "grad_norm": 1.7559314966201782, "learning_rate": 8.491244875671635e-07, "loss": 0.1676, "step": 29710 }, { "epoch": 0.5827450980392157, "grad_norm": 1.9649525880813599, "learning_rate": 8.490084256302098e-07, "loss": 0.1624, "step": 29720 }, { "epoch": 0.5829411764705882, "grad_norm": 3.7265007495880127, "learning_rate": 8.48892327008292e-07, "loss": 0.1423, "step": 29730 }, { "epoch": 0.5831372549019608, "grad_norm": 1.9647427797317505, "learning_rate": 8.487761917136134e-07, "loss": 0.1613, "step": 29740 }, { "epoch": 0.5833333333333334, "grad_norm": 1.9939818382263184, "learning_rate": 8.486600197583813e-07, "loss": 0.1494, "step": 29750 }, { "epoch": 0.5835294117647059, "grad_norm": 2.2384958267211914, "learning_rate": 8.485438111548064e-07, "loss": 0.1424, "step": 29760 }, { "epoch": 0.5837254901960784, "grad_norm": 2.0333619117736816, "learning_rate": 8.484275659151043e-07, "loss": 0.1174, "step": 29770 }, { "epoch": 0.5839215686274509, "grad_norm": 1.261715292930603, "learning_rate": 8.48311284051493e-07, "loss": 0.127, "step": 29780 }, { "epoch": 0.5841176470588235, "grad_norm": 3.6821837425231934, "learning_rate": 8.481949655761958e-07, "loss": 0.1793, "step": 29790 }, { "epoch": 0.5843137254901961, "grad_norm": 1.306893229484558, "learning_rate": 8.480786105014385e-07, "loss": 0.1806, "step": 29800 }, { "epoch": 0.5845098039215686, "grad_norm": 2.908825397491455, "learning_rate": 8.479622188394518e-07, "loss": 0.1527, "step": 29810 }, { "epoch": 0.5847058823529412, "grad_norm": 1.8027883768081665, "learning_rate": 8.478457906024699e-07, "loss": 0.1575, "step": 29820 }, { "epoch": 0.5849019607843138, "grad_norm": 5.166175842285156, "learning_rate": 8.477293258027306e-07, "loss": 0.1559, "step": 29830 }, { "epoch": 0.5850980392156863, "grad_norm": 3.9173288345336914, "learning_rate": 8.476128244524759e-07, "loss": 0.1619, "step": 29840 }, { "epoch": 0.5852941176470589, "grad_norm": 3.658268451690674, "learning_rate": 8.474962865639515e-07, "loss": 0.1686, "step": 29850 }, { "epoch": 0.5854901960784313, "grad_norm": 5.086117744445801, "learning_rate": 8.473797121494067e-07, "loss": 0.2191, "step": 29860 }, { "epoch": 0.5856862745098039, "grad_norm": 4.270161151885986, "learning_rate": 8.472631012210951e-07, "loss": 0.1722, "step": 29870 }, { "epoch": 0.5858823529411765, "grad_norm": 3.265576124191284, "learning_rate": 8.471464537912739e-07, "loss": 0.177, "step": 29880 }, { "epoch": 0.586078431372549, "grad_norm": 2.6589484214782715, "learning_rate": 8.47029769872204e-07, "loss": 0.1381, "step": 29890 }, { "epoch": 0.5862745098039216, "grad_norm": 3.4393863677978516, "learning_rate": 8.469130494761505e-07, "loss": 0.1972, "step": 29900 }, { "epoch": 0.5864705882352941, "grad_norm": 2.6555025577545166, "learning_rate": 8.467962926153818e-07, "loss": 0.1517, "step": 29910 }, { "epoch": 0.5866666666666667, "grad_norm": 9.595547676086426, "learning_rate": 8.466794993021705e-07, "loss": 0.1716, "step": 29920 }, { "epoch": 0.5868627450980393, "grad_norm": 3.497481346130371, "learning_rate": 8.465626695487933e-07, "loss": 0.1334, "step": 29930 }, { "epoch": 0.5870588235294117, "grad_norm": 3.1345484256744385, "learning_rate": 8.4644580336753e-07, "loss": 0.1637, "step": 29940 }, { "epoch": 0.5872549019607843, "grad_norm": 1.5137602090835571, "learning_rate": 8.463289007706648e-07, "loss": 0.1388, "step": 29950 }, { "epoch": 0.5874509803921568, "grad_norm": 2.814181327819824, "learning_rate": 8.462119617704859e-07, "loss": 0.1425, "step": 29960 }, { "epoch": 0.5876470588235294, "grad_norm": 1.6205066442489624, "learning_rate": 8.460949863792845e-07, "loss": 0.1582, "step": 29970 }, { "epoch": 0.587843137254902, "grad_norm": 2.5136454105377197, "learning_rate": 8.459779746093563e-07, "loss": 0.1678, "step": 29980 }, { "epoch": 0.5880392156862745, "grad_norm": 2.9494473934173584, "learning_rate": 8.458609264730004e-07, "loss": 0.1534, "step": 29990 }, { "epoch": 0.5882352941176471, "grad_norm": 1.3418923616409302, "learning_rate": 8.457438419825203e-07, "loss": 0.1456, "step": 30000 }, { "epoch": 0.5884313725490196, "grad_norm": 2.9415855407714844, "learning_rate": 8.45626721150223e-07, "loss": 0.1702, "step": 30010 }, { "epoch": 0.5886274509803922, "grad_norm": 2.120114326477051, "learning_rate": 8.455095639884191e-07, "loss": 0.1506, "step": 30020 }, { "epoch": 0.5888235294117647, "grad_norm": 3.174844264984131, "learning_rate": 8.453923705094232e-07, "loss": 0.1521, "step": 30030 }, { "epoch": 0.5890196078431372, "grad_norm": 1.4205856323242188, "learning_rate": 8.45275140725554e-07, "loss": 0.1719, "step": 30040 }, { "epoch": 0.5892156862745098, "grad_norm": 4.33406925201416, "learning_rate": 8.451578746491335e-07, "loss": 0.1946, "step": 30050 }, { "epoch": 0.5894117647058823, "grad_norm": 2.4931857585906982, "learning_rate": 8.450405722924878e-07, "loss": 0.1466, "step": 30060 }, { "epoch": 0.5896078431372549, "grad_norm": 2.67171573638916, "learning_rate": 8.449232336679468e-07, "loss": 0.1479, "step": 30070 }, { "epoch": 0.5898039215686275, "grad_norm": 2.8160347938537598, "learning_rate": 8.448058587878444e-07, "loss": 0.1301, "step": 30080 }, { "epoch": 0.59, "grad_norm": 4.566839694976807, "learning_rate": 8.446884476645177e-07, "loss": 0.16, "step": 30090 }, { "epoch": 0.5901960784313726, "grad_norm": 1.7278685569763184, "learning_rate": 8.445710003103084e-07, "loss": 0.1264, "step": 30100 }, { "epoch": 0.590392156862745, "grad_norm": 3.4104161262512207, "learning_rate": 8.444535167375615e-07, "loss": 0.1779, "step": 30110 }, { "epoch": 0.5905882352941176, "grad_norm": 2.8565890789031982, "learning_rate": 8.443359969586258e-07, "loss": 0.1544, "step": 30120 }, { "epoch": 0.5907843137254902, "grad_norm": 2.0833017826080322, "learning_rate": 8.442184409858541e-07, "loss": 0.1419, "step": 30130 }, { "epoch": 0.5909803921568627, "grad_norm": 2.9178476333618164, "learning_rate": 8.441008488316032e-07, "loss": 0.1503, "step": 30140 }, { "epoch": 0.5911764705882353, "grad_norm": 2.5613958835601807, "learning_rate": 8.439832205082329e-07, "loss": 0.1392, "step": 30150 }, { "epoch": 0.5913725490196079, "grad_norm": 3.8678712844848633, "learning_rate": 8.438655560281078e-07, "loss": 0.1657, "step": 30160 }, { "epoch": 0.5915686274509804, "grad_norm": 2.243821620941162, "learning_rate": 8.437478554035959e-07, "loss": 0.1366, "step": 30170 }, { "epoch": 0.591764705882353, "grad_norm": 2.826622486114502, "learning_rate": 8.436301186470684e-07, "loss": 0.158, "step": 30180 }, { "epoch": 0.5919607843137255, "grad_norm": 3.77834153175354, "learning_rate": 8.435123457709015e-07, "loss": 0.2049, "step": 30190 }, { "epoch": 0.592156862745098, "grad_norm": 3.6193766593933105, "learning_rate": 8.43394536787474e-07, "loss": 0.1513, "step": 30200 }, { "epoch": 0.5923529411764706, "grad_norm": 3.053257942199707, "learning_rate": 8.432766917091694e-07, "loss": 0.1379, "step": 30210 }, { "epoch": 0.5925490196078431, "grad_norm": 2.2090487480163574, "learning_rate": 8.431588105483744e-07, "loss": 0.1512, "step": 30220 }, { "epoch": 0.5927450980392157, "grad_norm": 3.028451681137085, "learning_rate": 8.430408933174799e-07, "loss": 0.1553, "step": 30230 }, { "epoch": 0.5929411764705882, "grad_norm": 3.6163926124572754, "learning_rate": 8.429229400288804e-07, "loss": 0.1758, "step": 30240 }, { "epoch": 0.5931372549019608, "grad_norm": 4.560175895690918, "learning_rate": 8.428049506949742e-07, "loss": 0.1864, "step": 30250 }, { "epoch": 0.5933333333333334, "grad_norm": 1.9134865999221802, "learning_rate": 8.426869253281631e-07, "loss": 0.1332, "step": 30260 }, { "epoch": 0.5935294117647059, "grad_norm": 1.5415832996368408, "learning_rate": 8.425688639408533e-07, "loss": 0.154, "step": 30270 }, { "epoch": 0.5937254901960785, "grad_norm": 2.85492205619812, "learning_rate": 8.424507665454543e-07, "loss": 0.1627, "step": 30280 }, { "epoch": 0.5939215686274509, "grad_norm": 4.434545040130615, "learning_rate": 8.423326331543798e-07, "loss": 0.1477, "step": 30290 }, { "epoch": 0.5941176470588235, "grad_norm": 6.476885795593262, "learning_rate": 8.422144637800466e-07, "loss": 0.1575, "step": 30300 }, { "epoch": 0.5943137254901961, "grad_norm": 2.328946352005005, "learning_rate": 8.420962584348762e-07, "loss": 0.1584, "step": 30310 }, { "epoch": 0.5945098039215686, "grad_norm": 3.3492379188537598, "learning_rate": 8.419780171312931e-07, "loss": 0.1502, "step": 30320 }, { "epoch": 0.5947058823529412, "grad_norm": 4.076613426208496, "learning_rate": 8.418597398817259e-07, "loss": 0.1688, "step": 30330 }, { "epoch": 0.5949019607843137, "grad_norm": 4.312715530395508, "learning_rate": 8.417414266986071e-07, "loss": 0.1421, "step": 30340 }, { "epoch": 0.5950980392156863, "grad_norm": 2.586627244949341, "learning_rate": 8.416230775943726e-07, "loss": 0.1642, "step": 30350 }, { "epoch": 0.5952941176470589, "grad_norm": 2.0124781131744385, "learning_rate": 8.415046925814625e-07, "loss": 0.1677, "step": 30360 }, { "epoch": 0.5954901960784313, "grad_norm": 4.191586494445801, "learning_rate": 8.413862716723205e-07, "loss": 0.1489, "step": 30370 }, { "epoch": 0.5956862745098039, "grad_norm": 1.980542778968811, "learning_rate": 8.412678148793939e-07, "loss": 0.1859, "step": 30380 }, { "epoch": 0.5958823529411764, "grad_norm": 1.747719407081604, "learning_rate": 8.41149322215134e-07, "loss": 0.1526, "step": 30390 }, { "epoch": 0.596078431372549, "grad_norm": 2.639695405960083, "learning_rate": 8.410307936919958e-07, "loss": 0.1414, "step": 30400 }, { "epoch": 0.5962745098039216, "grad_norm": 2.064072370529175, "learning_rate": 8.409122293224381e-07, "loss": 0.1704, "step": 30410 }, { "epoch": 0.5964705882352941, "grad_norm": 6.326392650604248, "learning_rate": 8.407936291189234e-07, "loss": 0.2174, "step": 30420 }, { "epoch": 0.5966666666666667, "grad_norm": 3.4864583015441895, "learning_rate": 8.406749930939179e-07, "loss": 0.1457, "step": 30430 }, { "epoch": 0.5968627450980392, "grad_norm": 3.8262417316436768, "learning_rate": 8.40556321259892e-07, "loss": 0.1912, "step": 30440 }, { "epoch": 0.5970588235294118, "grad_norm": 2.1717870235443115, "learning_rate": 8.404376136293191e-07, "loss": 0.1607, "step": 30450 }, { "epoch": 0.5972549019607843, "grad_norm": 2.73565411567688, "learning_rate": 8.403188702146771e-07, "loss": 0.1374, "step": 30460 }, { "epoch": 0.5974509803921568, "grad_norm": 2.7233097553253174, "learning_rate": 8.402000910284472e-07, "loss": 0.1835, "step": 30470 }, { "epoch": 0.5976470588235294, "grad_norm": 1.3397231101989746, "learning_rate": 8.400812760831146e-07, "loss": 0.1504, "step": 30480 }, { "epoch": 0.597843137254902, "grad_norm": 3.8393783569335938, "learning_rate": 8.399624253911681e-07, "loss": 0.1836, "step": 30490 }, { "epoch": 0.5980392156862745, "grad_norm": 3.919389247894287, "learning_rate": 8.398435389651004e-07, "loss": 0.1674, "step": 30500 }, { "epoch": 0.5982352941176471, "grad_norm": 3.726600170135498, "learning_rate": 8.397246168174079e-07, "loss": 0.1658, "step": 30510 }, { "epoch": 0.5984313725490196, "grad_norm": 2.3444395065307617, "learning_rate": 8.396056589605907e-07, "loss": 0.1536, "step": 30520 }, { "epoch": 0.5986274509803922, "grad_norm": 3.121483564376831, "learning_rate": 8.394866654071526e-07, "loss": 0.1422, "step": 30530 }, { "epoch": 0.5988235294117648, "grad_norm": 3.0376052856445312, "learning_rate": 8.393676361696013e-07, "loss": 0.1659, "step": 30540 }, { "epoch": 0.5990196078431372, "grad_norm": 2.9749643802642822, "learning_rate": 8.392485712604482e-07, "loss": 0.1487, "step": 30550 }, { "epoch": 0.5992156862745098, "grad_norm": 3.234847068786621, "learning_rate": 8.391294706922088e-07, "loss": 0.1892, "step": 30560 }, { "epoch": 0.5994117647058823, "grad_norm": 2.9720380306243896, "learning_rate": 8.390103344774015e-07, "loss": 0.1369, "step": 30570 }, { "epoch": 0.5996078431372549, "grad_norm": 2.2422919273376465, "learning_rate": 8.388911626285491e-07, "loss": 0.1718, "step": 30580 }, { "epoch": 0.5998039215686275, "grad_norm": 2.870236873626709, "learning_rate": 8.38771955158178e-07, "loss": 0.1843, "step": 30590 }, { "epoch": 0.6, "grad_norm": 2.2385692596435547, "learning_rate": 8.386527120788184e-07, "loss": 0.1316, "step": 30600 }, { "epoch": 0.6001960784313726, "grad_norm": 1.919142484664917, "learning_rate": 8.385334334030041e-07, "loss": 0.1222, "step": 30610 }, { "epoch": 0.600392156862745, "grad_norm": 2.0993287563323975, "learning_rate": 8.384141191432726e-07, "loss": 0.1888, "step": 30620 }, { "epoch": 0.6005882352941176, "grad_norm": 1.680348515510559, "learning_rate": 8.382947693121656e-07, "loss": 0.1757, "step": 30630 }, { "epoch": 0.6007843137254902, "grad_norm": 2.208672285079956, "learning_rate": 8.381753839222279e-07, "loss": 0.1403, "step": 30640 }, { "epoch": 0.6009803921568627, "grad_norm": 2.8549158573150635, "learning_rate": 8.380559629860084e-07, "loss": 0.1612, "step": 30650 }, { "epoch": 0.6011764705882353, "grad_norm": 1.9325541257858276, "learning_rate": 8.379365065160597e-07, "loss": 0.1643, "step": 30660 }, { "epoch": 0.6013725490196078, "grad_norm": 4.555104732513428, "learning_rate": 8.378170145249381e-07, "loss": 0.1565, "step": 30670 }, { "epoch": 0.6015686274509804, "grad_norm": 5.02286434173584, "learning_rate": 8.376974870252037e-07, "loss": 0.1995, "step": 30680 }, { "epoch": 0.601764705882353, "grad_norm": 2.796895742416382, "learning_rate": 8.375779240294202e-07, "loss": 0.1842, "step": 30690 }, { "epoch": 0.6019607843137255, "grad_norm": 2.4103970527648926, "learning_rate": 8.37458325550155e-07, "loss": 0.1623, "step": 30700 }, { "epoch": 0.602156862745098, "grad_norm": 3.173025369644165, "learning_rate": 8.373386915999796e-07, "loss": 0.1988, "step": 30710 }, { "epoch": 0.6023529411764705, "grad_norm": 2.2311484813690186, "learning_rate": 8.372190221914688e-07, "loss": 0.1554, "step": 30720 }, { "epoch": 0.6025490196078431, "grad_norm": 3.1565465927124023, "learning_rate": 8.370993173372012e-07, "loss": 0.1443, "step": 30730 }, { "epoch": 0.6027450980392157, "grad_norm": 3.652069091796875, "learning_rate": 8.369795770497592e-07, "loss": 0.1751, "step": 30740 }, { "epoch": 0.6029411764705882, "grad_norm": 2.1263668537139893, "learning_rate": 8.368598013417292e-07, "loss": 0.1545, "step": 30750 }, { "epoch": 0.6031372549019608, "grad_norm": 2.947643756866455, "learning_rate": 8.36739990225701e-07, "loss": 0.1891, "step": 30760 }, { "epoch": 0.6033333333333334, "grad_norm": 2.0341999530792236, "learning_rate": 8.366201437142679e-07, "loss": 0.1941, "step": 30770 }, { "epoch": 0.6035294117647059, "grad_norm": 3.3037447929382324, "learning_rate": 8.365002618200275e-07, "loss": 0.1819, "step": 30780 }, { "epoch": 0.6037254901960785, "grad_norm": 3.59294056892395, "learning_rate": 8.363803445555806e-07, "loss": 0.1731, "step": 30790 }, { "epoch": 0.6039215686274509, "grad_norm": 2.4536828994750977, "learning_rate": 8.362603919335321e-07, "loss": 0.1308, "step": 30800 }, { "epoch": 0.6041176470588235, "grad_norm": 2.2538976669311523, "learning_rate": 8.361404039664905e-07, "loss": 0.1815, "step": 30810 }, { "epoch": 0.6043137254901961, "grad_norm": 2.332900285720825, "learning_rate": 8.360203806670678e-07, "loss": 0.154, "step": 30820 }, { "epoch": 0.6045098039215686, "grad_norm": 1.5908918380737305, "learning_rate": 8.359003220478799e-07, "loss": 0.1655, "step": 30830 }, { "epoch": 0.6047058823529412, "grad_norm": 1.3066997528076172, "learning_rate": 8.357802281215467e-07, "loss": 0.1768, "step": 30840 }, { "epoch": 0.6049019607843137, "grad_norm": 5.835001468658447, "learning_rate": 8.35660098900691e-07, "loss": 0.1362, "step": 30850 }, { "epoch": 0.6050980392156863, "grad_norm": 3.5334250926971436, "learning_rate": 8.355399343979403e-07, "loss": 0.1596, "step": 30860 }, { "epoch": 0.6052941176470589, "grad_norm": 4.9331440925598145, "learning_rate": 8.35419734625925e-07, "loss": 0.16, "step": 30870 }, { "epoch": 0.6054901960784314, "grad_norm": 2.252708911895752, "learning_rate": 8.352994995972797e-07, "loss": 0.1728, "step": 30880 }, { "epoch": 0.605686274509804, "grad_norm": 2.7682764530181885, "learning_rate": 8.351792293246425e-07, "loss": 0.1464, "step": 30890 }, { "epoch": 0.6058823529411764, "grad_norm": 2.981584072113037, "learning_rate": 8.350589238206552e-07, "loss": 0.1421, "step": 30900 }, { "epoch": 0.606078431372549, "grad_norm": 14.160927772521973, "learning_rate": 8.349385830979635e-07, "loss": 0.2112, "step": 30910 }, { "epoch": 0.6062745098039216, "grad_norm": 1.9863086938858032, "learning_rate": 8.348182071692164e-07, "loss": 0.1535, "step": 30920 }, { "epoch": 0.6064705882352941, "grad_norm": 2.980092763900757, "learning_rate": 8.346977960470672e-07, "loss": 0.1729, "step": 30930 }, { "epoch": 0.6066666666666667, "grad_norm": 2.411128520965576, "learning_rate": 8.345773497441721e-07, "loss": 0.1576, "step": 30940 }, { "epoch": 0.6068627450980392, "grad_norm": 4.814946174621582, "learning_rate": 8.34456868273192e-07, "loss": 0.1446, "step": 30950 }, { "epoch": 0.6070588235294118, "grad_norm": 2.875297784805298, "learning_rate": 8.343363516467904e-07, "loss": 0.1538, "step": 30960 }, { "epoch": 0.6072549019607844, "grad_norm": 2.883985757827759, "learning_rate": 8.342157998776353e-07, "loss": 0.1606, "step": 30970 }, { "epoch": 0.6074509803921568, "grad_norm": 2.864098072052002, "learning_rate": 8.340952129783981e-07, "loss": 0.1839, "step": 30980 }, { "epoch": 0.6076470588235294, "grad_norm": 1.8892935514450073, "learning_rate": 8.339745909617541e-07, "loss": 0.1698, "step": 30990 }, { "epoch": 0.6078431372549019, "grad_norm": 1.9621670246124268, "learning_rate": 8.338539338403819e-07, "loss": 0.1716, "step": 31000 }, { "epoch": 0.6080392156862745, "grad_norm": 1.4961916208267212, "learning_rate": 8.33733241626964e-07, "loss": 0.1317, "step": 31010 }, { "epoch": 0.6082352941176471, "grad_norm": 3.886159658432007, "learning_rate": 8.336125143341867e-07, "loss": 0.1383, "step": 31020 }, { "epoch": 0.6084313725490196, "grad_norm": 4.154498100280762, "learning_rate": 8.334917519747398e-07, "loss": 0.157, "step": 31030 }, { "epoch": 0.6086274509803922, "grad_norm": 2.1185529232025146, "learning_rate": 8.33370954561317e-07, "loss": 0.1282, "step": 31040 }, { "epoch": 0.6088235294117647, "grad_norm": 1.7361822128295898, "learning_rate": 8.332501221066153e-07, "loss": 0.1443, "step": 31050 }, { "epoch": 0.6090196078431372, "grad_norm": 2.8583645820617676, "learning_rate": 8.33129254623336e-07, "loss": 0.1787, "step": 31060 }, { "epoch": 0.6092156862745098, "grad_norm": 1.8126492500305176, "learning_rate": 8.330083521241835e-07, "loss": 0.1335, "step": 31070 }, { "epoch": 0.6094117647058823, "grad_norm": 2.633856773376465, "learning_rate": 8.32887414621866e-07, "loss": 0.167, "step": 31080 }, { "epoch": 0.6096078431372549, "grad_norm": 3.009176731109619, "learning_rate": 8.327664421290958e-07, "loss": 0.1653, "step": 31090 }, { "epoch": 0.6098039215686275, "grad_norm": 3.1991043090820312, "learning_rate": 8.326454346585883e-07, "loss": 0.166, "step": 31100 }, { "epoch": 0.61, "grad_norm": 4.884282112121582, "learning_rate": 8.325243922230629e-07, "loss": 0.1658, "step": 31110 }, { "epoch": 0.6101960784313726, "grad_norm": 2.874140739440918, "learning_rate": 8.324033148352425e-07, "loss": 0.1491, "step": 31120 }, { "epoch": 0.6103921568627451, "grad_norm": 2.817434549331665, "learning_rate": 8.322822025078539e-07, "loss": 0.1599, "step": 31130 }, { "epoch": 0.6105882352941177, "grad_norm": 3.568326950073242, "learning_rate": 8.321610552536276e-07, "loss": 0.1359, "step": 31140 }, { "epoch": 0.6107843137254902, "grad_norm": 2.8213608264923096, "learning_rate": 8.320398730852974e-07, "loss": 0.1712, "step": 31150 }, { "epoch": 0.6109803921568627, "grad_norm": 5.402161121368408, "learning_rate": 8.319186560156011e-07, "loss": 0.1534, "step": 31160 }, { "epoch": 0.6111764705882353, "grad_norm": 2.320161819458008, "learning_rate": 8.317974040572799e-07, "loss": 0.1828, "step": 31170 }, { "epoch": 0.6113725490196078, "grad_norm": 1.4056458473205566, "learning_rate": 8.316761172230791e-07, "loss": 0.1559, "step": 31180 }, { "epoch": 0.6115686274509804, "grad_norm": 2.413593292236328, "learning_rate": 8.315547955257473e-07, "loss": 0.1556, "step": 31190 }, { "epoch": 0.611764705882353, "grad_norm": 3.3662307262420654, "learning_rate": 8.314334389780368e-07, "loss": 0.155, "step": 31200 }, { "epoch": 0.6119607843137255, "grad_norm": 2.5574228763580322, "learning_rate": 8.313120475927036e-07, "loss": 0.1495, "step": 31210 }, { "epoch": 0.6121568627450981, "grad_norm": 2.3832972049713135, "learning_rate": 8.311906213825076e-07, "loss": 0.1856, "step": 31220 }, { "epoch": 0.6123529411764705, "grad_norm": 3.632082462310791, "learning_rate": 8.310691603602118e-07, "loss": 0.1485, "step": 31230 }, { "epoch": 0.6125490196078431, "grad_norm": 2.392766237258911, "learning_rate": 8.309476645385835e-07, "loss": 0.1597, "step": 31240 }, { "epoch": 0.6127450980392157, "grad_norm": 3.145494222640991, "learning_rate": 8.308261339303933e-07, "loss": 0.1513, "step": 31250 }, { "epoch": 0.6129411764705882, "grad_norm": 2.6049320697784424, "learning_rate": 8.307045685484154e-07, "loss": 0.1678, "step": 31260 }, { "epoch": 0.6131372549019608, "grad_norm": 4.267388343811035, "learning_rate": 8.30582968405428e-07, "loss": 0.1778, "step": 31270 }, { "epoch": 0.6133333333333333, "grad_norm": 2.5480992794036865, "learning_rate": 8.304613335142125e-07, "loss": 0.1373, "step": 31280 }, { "epoch": 0.6135294117647059, "grad_norm": 6.084670066833496, "learning_rate": 8.303396638875543e-07, "loss": 0.1636, "step": 31290 }, { "epoch": 0.6137254901960785, "grad_norm": 4.6633100509643555, "learning_rate": 8.302179595382423e-07, "loss": 0.1463, "step": 31300 }, { "epoch": 0.613921568627451, "grad_norm": 3.2186014652252197, "learning_rate": 8.300962204790692e-07, "loss": 0.1618, "step": 31310 }, { "epoch": 0.6141176470588235, "grad_norm": 2.8567652702331543, "learning_rate": 8.299744467228309e-07, "loss": 0.1573, "step": 31320 }, { "epoch": 0.614313725490196, "grad_norm": 1.6751484870910645, "learning_rate": 8.298526382823276e-07, "loss": 0.1223, "step": 31330 }, { "epoch": 0.6145098039215686, "grad_norm": 3.2908222675323486, "learning_rate": 8.297307951703629e-07, "loss": 0.1524, "step": 31340 }, { "epoch": 0.6147058823529412, "grad_norm": 2.109808921813965, "learning_rate": 8.296089173997436e-07, "loss": 0.1915, "step": 31350 }, { "epoch": 0.6149019607843137, "grad_norm": 3.6267387866973877, "learning_rate": 8.294870049832809e-07, "loss": 0.1732, "step": 31360 }, { "epoch": 0.6150980392156863, "grad_norm": 2.7972054481506348, "learning_rate": 8.29365057933789e-07, "loss": 0.1345, "step": 31370 }, { "epoch": 0.6152941176470588, "grad_norm": 1.4337794780731201, "learning_rate": 8.29243076264086e-07, "loss": 0.1469, "step": 31380 }, { "epoch": 0.6154901960784314, "grad_norm": 4.1285600662231445, "learning_rate": 8.291210599869938e-07, "loss": 0.1547, "step": 31390 }, { "epoch": 0.615686274509804, "grad_norm": 2.646465301513672, "learning_rate": 8.289990091153376e-07, "loss": 0.145, "step": 31400 }, { "epoch": 0.6158823529411764, "grad_norm": 3.7806437015533447, "learning_rate": 8.288769236619466e-07, "loss": 0.1458, "step": 31410 }, { "epoch": 0.616078431372549, "grad_norm": 2.811936378479004, "learning_rate": 8.287548036396532e-07, "loss": 0.1577, "step": 31420 }, { "epoch": 0.6162745098039216, "grad_norm": 2.1635351181030273, "learning_rate": 8.286326490612938e-07, "loss": 0.1869, "step": 31430 }, { "epoch": 0.6164705882352941, "grad_norm": 2.798933744430542, "learning_rate": 8.285104599397083e-07, "loss": 0.156, "step": 31440 }, { "epoch": 0.6166666666666667, "grad_norm": 2.8598439693450928, "learning_rate": 8.283882362877403e-07, "loss": 0.1938, "step": 31450 }, { "epoch": 0.6168627450980392, "grad_norm": 1.6409465074539185, "learning_rate": 8.28265978118237e-07, "loss": 0.1456, "step": 31460 }, { "epoch": 0.6170588235294118, "grad_norm": 3.6688830852508545, "learning_rate": 8.281436854440491e-07, "loss": 0.1616, "step": 31470 }, { "epoch": 0.6172549019607844, "grad_norm": 3.375990629196167, "learning_rate": 8.280213582780311e-07, "loss": 0.1453, "step": 31480 }, { "epoch": 0.6174509803921568, "grad_norm": 2.5802080631256104, "learning_rate": 8.278989966330409e-07, "loss": 0.152, "step": 31490 }, { "epoch": 0.6176470588235294, "grad_norm": 2.8555305004119873, "learning_rate": 8.277766005219402e-07, "loss": 0.1543, "step": 31500 }, { "epoch": 0.6178431372549019, "grad_norm": 2.7347445487976074, "learning_rate": 8.276541699575944e-07, "loss": 0.1715, "step": 31510 }, { "epoch": 0.6180392156862745, "grad_norm": 3.2915596961975098, "learning_rate": 8.275317049528723e-07, "loss": 0.1335, "step": 31520 }, { "epoch": 0.6182352941176471, "grad_norm": 3.843658208847046, "learning_rate": 8.274092055206467e-07, "loss": 0.1587, "step": 31530 }, { "epoch": 0.6184313725490196, "grad_norm": 2.542959213256836, "learning_rate": 8.272866716737935e-07, "loss": 0.1483, "step": 31540 }, { "epoch": 0.6186274509803922, "grad_norm": 1.8608698844909668, "learning_rate": 8.271641034251927e-07, "loss": 0.1726, "step": 31550 }, { "epoch": 0.6188235294117647, "grad_norm": 2.4295308589935303, "learning_rate": 8.270415007877272e-07, "loss": 0.16, "step": 31560 }, { "epoch": 0.6190196078431373, "grad_norm": 2.5303845405578613, "learning_rate": 8.269188637742846e-07, "loss": 0.1641, "step": 31570 }, { "epoch": 0.6192156862745098, "grad_norm": 1.4248783588409424, "learning_rate": 8.267961923977552e-07, "loss": 0.1387, "step": 31580 }, { "epoch": 0.6194117647058823, "grad_norm": 2.591214656829834, "learning_rate": 8.266734866710332e-07, "loss": 0.1615, "step": 31590 }, { "epoch": 0.6196078431372549, "grad_norm": 3.963703155517578, "learning_rate": 8.265507466070167e-07, "loss": 0.1468, "step": 31600 }, { "epoch": 0.6198039215686274, "grad_norm": 2.2748310565948486, "learning_rate": 8.264279722186069e-07, "loss": 0.151, "step": 31610 }, { "epoch": 0.62, "grad_norm": 2.182278871536255, "learning_rate": 8.263051635187089e-07, "loss": 0.1512, "step": 31620 }, { "epoch": 0.6201960784313726, "grad_norm": 1.8998960256576538, "learning_rate": 8.261823205202315e-07, "loss": 0.1729, "step": 31630 }, { "epoch": 0.6203921568627451, "grad_norm": 1.8529902696609497, "learning_rate": 8.260594432360868e-07, "loss": 0.1647, "step": 31640 }, { "epoch": 0.6205882352941177, "grad_norm": 2.488551616668701, "learning_rate": 8.259365316791909e-07, "loss": 0.1554, "step": 31650 }, { "epoch": 0.6207843137254901, "grad_norm": 2.2689366340637207, "learning_rate": 8.25813585862463e-07, "loss": 0.1726, "step": 31660 }, { "epoch": 0.6209803921568627, "grad_norm": 1.452176570892334, "learning_rate": 8.256906057988263e-07, "loss": 0.1756, "step": 31670 }, { "epoch": 0.6211764705882353, "grad_norm": 2.3973124027252197, "learning_rate": 8.255675915012075e-07, "loss": 0.1379, "step": 31680 }, { "epoch": 0.6213725490196078, "grad_norm": 2.2224416732788086, "learning_rate": 8.25444542982537e-07, "loss": 0.1754, "step": 31690 }, { "epoch": 0.6215686274509804, "grad_norm": 1.368644118309021, "learning_rate": 8.253214602557483e-07, "loss": 0.1677, "step": 31700 }, { "epoch": 0.6217647058823529, "grad_norm": 1.8553272485733032, "learning_rate": 8.251983433337792e-07, "loss": 0.1562, "step": 31710 }, { "epoch": 0.6219607843137255, "grad_norm": 3.7206904888153076, "learning_rate": 8.250751922295707e-07, "loss": 0.1876, "step": 31720 }, { "epoch": 0.6221568627450981, "grad_norm": 1.5842915773391724, "learning_rate": 8.249520069560674e-07, "loss": 0.1787, "step": 31730 }, { "epoch": 0.6223529411764706, "grad_norm": 1.7839076519012451, "learning_rate": 8.248287875262176e-07, "loss": 0.1697, "step": 31740 }, { "epoch": 0.6225490196078431, "grad_norm": 2.2463598251342773, "learning_rate": 8.247055339529729e-07, "loss": 0.1618, "step": 31750 }, { "epoch": 0.6227450980392157, "grad_norm": 3.2933764457702637, "learning_rate": 8.245822462492892e-07, "loss": 0.1455, "step": 31760 }, { "epoch": 0.6229411764705882, "grad_norm": 1.8187274932861328, "learning_rate": 8.24458924428125e-07, "loss": 0.1824, "step": 31770 }, { "epoch": 0.6231372549019608, "grad_norm": 1.8331793546676636, "learning_rate": 8.243355685024434e-07, "loss": 0.1485, "step": 31780 }, { "epoch": 0.6233333333333333, "grad_norm": 1.591503620147705, "learning_rate": 8.242121784852101e-07, "loss": 0.1491, "step": 31790 }, { "epoch": 0.6235294117647059, "grad_norm": 2.7404162883758545, "learning_rate": 8.240887543893953e-07, "loss": 0.1725, "step": 31800 }, { "epoch": 0.6237254901960785, "grad_norm": 1.9481900930404663, "learning_rate": 8.239652962279719e-07, "loss": 0.1647, "step": 31810 }, { "epoch": 0.623921568627451, "grad_norm": 1.0927921533584595, "learning_rate": 8.238418040139173e-07, "loss": 0.1656, "step": 31820 }, { "epoch": 0.6241176470588236, "grad_norm": 5.875028133392334, "learning_rate": 8.237182777602116e-07, "loss": 0.1798, "step": 31830 }, { "epoch": 0.624313725490196, "grad_norm": 4.832398891448975, "learning_rate": 8.235947174798392e-07, "loss": 0.1544, "step": 31840 }, { "epoch": 0.6245098039215686, "grad_norm": 5.590737819671631, "learning_rate": 8.234711231857876e-07, "loss": 0.1865, "step": 31850 }, { "epoch": 0.6247058823529412, "grad_norm": 12.95787525177002, "learning_rate": 8.233474948910482e-07, "loss": 0.1575, "step": 31860 }, { "epoch": 0.6249019607843137, "grad_norm": 2.6523711681365967, "learning_rate": 8.232238326086156e-07, "loss": 0.1694, "step": 31870 }, { "epoch": 0.6250980392156863, "grad_norm": 2.876493215560913, "learning_rate": 8.231001363514881e-07, "loss": 0.1683, "step": 31880 }, { "epoch": 0.6252941176470588, "grad_norm": 2.0262954235076904, "learning_rate": 8.229764061326681e-07, "loss": 0.1632, "step": 31890 }, { "epoch": 0.6254901960784314, "grad_norm": 4.736073017120361, "learning_rate": 8.228526419651607e-07, "loss": 0.1748, "step": 31900 }, { "epoch": 0.625686274509804, "grad_norm": 6.129160404205322, "learning_rate": 8.227288438619752e-07, "loss": 0.1583, "step": 31910 }, { "epoch": 0.6258823529411764, "grad_norm": 2.313450336456299, "learning_rate": 8.226050118361243e-07, "loss": 0.1707, "step": 31920 }, { "epoch": 0.626078431372549, "grad_norm": 2.2685675621032715, "learning_rate": 8.224811459006242e-07, "loss": 0.1884, "step": 31930 }, { "epoch": 0.6262745098039215, "grad_norm": 3.636209487915039, "learning_rate": 8.223572460684944e-07, "loss": 0.166, "step": 31940 }, { "epoch": 0.6264705882352941, "grad_norm": 3.679319381713867, "learning_rate": 8.222333123527587e-07, "loss": 0.1535, "step": 31950 }, { "epoch": 0.6266666666666667, "grad_norm": 4.08799409866333, "learning_rate": 8.221093447664437e-07, "loss": 0.1677, "step": 31960 }, { "epoch": 0.6268627450980392, "grad_norm": 3.4077517986297607, "learning_rate": 8.219853433225801e-07, "loss": 0.1446, "step": 31970 }, { "epoch": 0.6270588235294118, "grad_norm": 2.2111222743988037, "learning_rate": 8.218613080342019e-07, "loss": 0.2015, "step": 31980 }, { "epoch": 0.6272549019607843, "grad_norm": 2.8292508125305176, "learning_rate": 8.217372389143464e-07, "loss": 0.1761, "step": 31990 }, { "epoch": 0.6274509803921569, "grad_norm": 3.115649700164795, "learning_rate": 8.216131359760552e-07, "loss": 0.1411, "step": 32000 }, { "epoch": 0.6276470588235294, "grad_norm": 3.2227213382720947, "learning_rate": 8.214889992323726e-07, "loss": 0.1912, "step": 32010 }, { "epoch": 0.6278431372549019, "grad_norm": 4.020289897918701, "learning_rate": 8.21364828696347e-07, "loss": 0.1382, "step": 32020 }, { "epoch": 0.6280392156862745, "grad_norm": 2.5384716987609863, "learning_rate": 8.212406243810305e-07, "loss": 0.1444, "step": 32030 }, { "epoch": 0.6282352941176471, "grad_norm": 3.710080623626709, "learning_rate": 8.211163862994781e-07, "loss": 0.17, "step": 32040 }, { "epoch": 0.6284313725490196, "grad_norm": 3.318026542663574, "learning_rate": 8.209921144647487e-07, "loss": 0.1857, "step": 32050 }, { "epoch": 0.6286274509803922, "grad_norm": 3.623417377471924, "learning_rate": 8.20867808889905e-07, "loss": 0.1511, "step": 32060 }, { "epoch": 0.6288235294117647, "grad_norm": 2.7063839435577393, "learning_rate": 8.207434695880129e-07, "loss": 0.1661, "step": 32070 }, { "epoch": 0.6290196078431373, "grad_norm": 5.485341548919678, "learning_rate": 8.206190965721418e-07, "loss": 0.1727, "step": 32080 }, { "epoch": 0.6292156862745099, "grad_norm": 1.8102277517318726, "learning_rate": 8.20494689855365e-07, "loss": 0.1547, "step": 32090 }, { "epoch": 0.6294117647058823, "grad_norm": 1.9462381601333618, "learning_rate": 8.20370249450759e-07, "loss": 0.1342, "step": 32100 }, { "epoch": 0.6296078431372549, "grad_norm": 2.5821638107299805, "learning_rate": 8.202457753714043e-07, "loss": 0.1449, "step": 32110 }, { "epoch": 0.6298039215686274, "grad_norm": 1.6932883262634277, "learning_rate": 8.20121267630384e-07, "loss": 0.1902, "step": 32120 }, { "epoch": 0.63, "grad_norm": 4.594932556152344, "learning_rate": 8.199967262407859e-07, "loss": 0.1808, "step": 32130 }, { "epoch": 0.6301960784313726, "grad_norm": 2.633728504180908, "learning_rate": 8.198721512157005e-07, "loss": 0.189, "step": 32140 }, { "epoch": 0.6303921568627451, "grad_norm": 3.4829792976379395, "learning_rate": 8.197475425682224e-07, "loss": 0.1646, "step": 32150 }, { "epoch": 0.6305882352941177, "grad_norm": 3.2380757331848145, "learning_rate": 8.196229003114489e-07, "loss": 0.1871, "step": 32160 }, { "epoch": 0.6307843137254902, "grad_norm": 1.094014048576355, "learning_rate": 8.194982244584822e-07, "loss": 0.1425, "step": 32170 }, { "epoch": 0.6309803921568627, "grad_norm": 3.6073241233825684, "learning_rate": 8.193735150224266e-07, "loss": 0.1423, "step": 32180 }, { "epoch": 0.6311764705882353, "grad_norm": 3.8180270195007324, "learning_rate": 8.192487720163909e-07, "loss": 0.1816, "step": 32190 }, { "epoch": 0.6313725490196078, "grad_norm": 2.9367642402648926, "learning_rate": 8.191239954534868e-07, "loss": 0.1702, "step": 32200 }, { "epoch": 0.6315686274509804, "grad_norm": 3.7914302349090576, "learning_rate": 8.189991853468299e-07, "loss": 0.1614, "step": 32210 }, { "epoch": 0.6317647058823529, "grad_norm": 3.1638998985290527, "learning_rate": 8.188743417095394e-07, "loss": 0.1757, "step": 32220 }, { "epoch": 0.6319607843137255, "grad_norm": 5.003864288330078, "learning_rate": 8.187494645547377e-07, "loss": 0.1699, "step": 32230 }, { "epoch": 0.6321568627450981, "grad_norm": 1.7197825908660889, "learning_rate": 8.18624553895551e-07, "loss": 0.1669, "step": 32240 }, { "epoch": 0.6323529411764706, "grad_norm": 4.037989616394043, "learning_rate": 8.184996097451088e-07, "loss": 0.1501, "step": 32250 }, { "epoch": 0.6325490196078432, "grad_norm": 3.2381675243377686, "learning_rate": 8.183746321165444e-07, "loss": 0.1442, "step": 32260 }, { "epoch": 0.6327450980392156, "grad_norm": 1.6325981616973877, "learning_rate": 8.182496210229942e-07, "loss": 0.1522, "step": 32270 }, { "epoch": 0.6329411764705882, "grad_norm": 2.093510866165161, "learning_rate": 8.181245764775983e-07, "loss": 0.1642, "step": 32280 }, { "epoch": 0.6331372549019608, "grad_norm": 4.874569892883301, "learning_rate": 8.179994984935007e-07, "loss": 0.1678, "step": 32290 }, { "epoch": 0.6333333333333333, "grad_norm": 1.6309545040130615, "learning_rate": 8.178743870838485e-07, "loss": 0.1598, "step": 32300 }, { "epoch": 0.6335294117647059, "grad_norm": 2.5517795085906982, "learning_rate": 8.177492422617922e-07, "loss": 0.1594, "step": 32310 }, { "epoch": 0.6337254901960784, "grad_norm": 3.0424728393554688, "learning_rate": 8.176240640404863e-07, "loss": 0.1562, "step": 32320 }, { "epoch": 0.633921568627451, "grad_norm": 2.1998705863952637, "learning_rate": 8.174988524330883e-07, "loss": 0.1582, "step": 32330 }, { "epoch": 0.6341176470588236, "grad_norm": 2.3656797409057617, "learning_rate": 8.173736074527597e-07, "loss": 0.1448, "step": 32340 }, { "epoch": 0.634313725490196, "grad_norm": 3.057373046875, "learning_rate": 8.172483291126651e-07, "loss": 0.1748, "step": 32350 }, { "epoch": 0.6345098039215686, "grad_norm": 2.017770528793335, "learning_rate": 8.171230174259725e-07, "loss": 0.1706, "step": 32360 }, { "epoch": 0.6347058823529412, "grad_norm": 3.143432140350342, "learning_rate": 8.16997672405854e-07, "loss": 0.1382, "step": 32370 }, { "epoch": 0.6349019607843137, "grad_norm": 2.9726147651672363, "learning_rate": 8.168722940654849e-07, "loss": 0.1747, "step": 32380 }, { "epoch": 0.6350980392156863, "grad_norm": 1.8231611251831055, "learning_rate": 8.167468824180437e-07, "loss": 0.1777, "step": 32390 }, { "epoch": 0.6352941176470588, "grad_norm": 3.864163875579834, "learning_rate": 8.166214374767129e-07, "loss": 0.1467, "step": 32400 }, { "epoch": 0.6354901960784314, "grad_norm": 2.221904754638672, "learning_rate": 8.164959592546779e-07, "loss": 0.1458, "step": 32410 }, { "epoch": 0.635686274509804, "grad_norm": 1.2027652263641357, "learning_rate": 8.163704477651285e-07, "loss": 0.2128, "step": 32420 }, { "epoch": 0.6358823529411765, "grad_norm": 3.226015090942383, "learning_rate": 8.16244903021257e-07, "loss": 0.1376, "step": 32430 }, { "epoch": 0.636078431372549, "grad_norm": 3.9156785011291504, "learning_rate": 8.161193250362599e-07, "loss": 0.1566, "step": 32440 }, { "epoch": 0.6362745098039215, "grad_norm": 4.113049507141113, "learning_rate": 8.159937138233371e-07, "loss": 0.1778, "step": 32450 }, { "epoch": 0.6364705882352941, "grad_norm": 1.6203886270523071, "learning_rate": 8.158680693956914e-07, "loss": 0.124, "step": 32460 }, { "epoch": 0.6366666666666667, "grad_norm": 2.7361154556274414, "learning_rate": 8.157423917665298e-07, "loss": 0.1735, "step": 32470 }, { "epoch": 0.6368627450980392, "grad_norm": 2.530008554458618, "learning_rate": 8.156166809490624e-07, "loss": 0.1341, "step": 32480 }, { "epoch": 0.6370588235294118, "grad_norm": 1.2484571933746338, "learning_rate": 8.154909369565032e-07, "loss": 0.2178, "step": 32490 }, { "epoch": 0.6372549019607843, "grad_norm": 1.733277440071106, "learning_rate": 8.15365159802069e-07, "loss": 0.1819, "step": 32500 }, { "epoch": 0.6374509803921569, "grad_norm": 2.6379690170288086, "learning_rate": 8.152393494989808e-07, "loss": 0.1475, "step": 32510 }, { "epoch": 0.6376470588235295, "grad_norm": 3.157655715942383, "learning_rate": 8.151135060604626e-07, "loss": 0.154, "step": 32520 }, { "epoch": 0.6378431372549019, "grad_norm": 2.0725481510162354, "learning_rate": 8.149876294997421e-07, "loss": 0.1326, "step": 32530 }, { "epoch": 0.6380392156862745, "grad_norm": 2.6614623069763184, "learning_rate": 8.148617198300504e-07, "loss": 0.1267, "step": 32540 }, { "epoch": 0.638235294117647, "grad_norm": 2.2318873405456543, "learning_rate": 8.147357770646222e-07, "loss": 0.1592, "step": 32550 }, { "epoch": 0.6384313725490196, "grad_norm": 3.0815536975860596, "learning_rate": 8.146098012166957e-07, "loss": 0.1603, "step": 32560 }, { "epoch": 0.6386274509803922, "grad_norm": 1.9587448835372925, "learning_rate": 8.14483792299512e-07, "loss": 0.1364, "step": 32570 }, { "epoch": 0.6388235294117647, "grad_norm": 10.596631050109863, "learning_rate": 8.143577503263167e-07, "loss": 0.1722, "step": 32580 }, { "epoch": 0.6390196078431373, "grad_norm": 6.383870601654053, "learning_rate": 8.14231675310358e-07, "loss": 0.1398, "step": 32590 }, { "epoch": 0.6392156862745098, "grad_norm": 4.3129448890686035, "learning_rate": 8.14105567264888e-07, "loss": 0.1368, "step": 32600 }, { "epoch": 0.6394117647058823, "grad_norm": 1.9716615676879883, "learning_rate": 8.139794262031621e-07, "loss": 0.1482, "step": 32610 }, { "epoch": 0.6396078431372549, "grad_norm": 0.8848729133605957, "learning_rate": 8.138532521384394e-07, "loss": 0.1523, "step": 32620 }, { "epoch": 0.6398039215686274, "grad_norm": 3.6320064067840576, "learning_rate": 8.137270450839819e-07, "loss": 0.1474, "step": 32630 }, { "epoch": 0.64, "grad_norm": 2.360102653503418, "learning_rate": 8.13600805053056e-07, "loss": 0.1505, "step": 32640 }, { "epoch": 0.6401960784313725, "grad_norm": 2.6268582344055176, "learning_rate": 8.134745320589307e-07, "loss": 0.1408, "step": 32650 }, { "epoch": 0.6403921568627451, "grad_norm": 2.2933409214019775, "learning_rate": 8.13348226114879e-07, "loss": 0.1707, "step": 32660 }, { "epoch": 0.6405882352941177, "grad_norm": 2.0116429328918457, "learning_rate": 8.13221887234177e-07, "loss": 0.1562, "step": 32670 }, { "epoch": 0.6407843137254902, "grad_norm": 3.37490177154541, "learning_rate": 8.130955154301043e-07, "loss": 0.1588, "step": 32680 }, { "epoch": 0.6409803921568628, "grad_norm": 4.133961200714111, "learning_rate": 8.129691107159444e-07, "loss": 0.1517, "step": 32690 }, { "epoch": 0.6411764705882353, "grad_norm": 2.704007863998413, "learning_rate": 8.12842673104984e-07, "loss": 0.1603, "step": 32700 }, { "epoch": 0.6413725490196078, "grad_norm": 1.8684850931167603, "learning_rate": 8.127162026105128e-07, "loss": 0.1644, "step": 32710 }, { "epoch": 0.6415686274509804, "grad_norm": 3.3594539165496826, "learning_rate": 8.125896992458248e-07, "loss": 0.1543, "step": 32720 }, { "epoch": 0.6417647058823529, "grad_norm": 1.8726272583007812, "learning_rate": 8.124631630242167e-07, "loss": 0.1119, "step": 32730 }, { "epoch": 0.6419607843137255, "grad_norm": 3.4209132194519043, "learning_rate": 8.123365939589892e-07, "loss": 0.162, "step": 32740 }, { "epoch": 0.6421568627450981, "grad_norm": 4.362227439880371, "learning_rate": 8.12209992063446e-07, "loss": 0.1346, "step": 32750 }, { "epoch": 0.6423529411764706, "grad_norm": 2.634459972381592, "learning_rate": 8.120833573508947e-07, "loss": 0.1244, "step": 32760 }, { "epoch": 0.6425490196078432, "grad_norm": 2.0199952125549316, "learning_rate": 8.119566898346459e-07, "loss": 0.1814, "step": 32770 }, { "epoch": 0.6427450980392156, "grad_norm": 2.222536563873291, "learning_rate": 8.118299895280139e-07, "loss": 0.1505, "step": 32780 }, { "epoch": 0.6429411764705882, "grad_norm": 2.7085468769073486, "learning_rate": 8.117032564443167e-07, "loss": 0.152, "step": 32790 }, { "epoch": 0.6431372549019608, "grad_norm": 2.37864089012146, "learning_rate": 8.115764905968752e-07, "loss": 0.1523, "step": 32800 }, { "epoch": 0.6433333333333333, "grad_norm": 1.662395715713501, "learning_rate": 8.114496919990141e-07, "loss": 0.144, "step": 32810 }, { "epoch": 0.6435294117647059, "grad_norm": 2.357182264328003, "learning_rate": 8.113228606640613e-07, "loss": 0.1414, "step": 32820 }, { "epoch": 0.6437254901960784, "grad_norm": 2.4223663806915283, "learning_rate": 8.111959966053486e-07, "loss": 0.1333, "step": 32830 }, { "epoch": 0.643921568627451, "grad_norm": 3.109678268432617, "learning_rate": 8.110690998362106e-07, "loss": 0.1529, "step": 32840 }, { "epoch": 0.6441176470588236, "grad_norm": 3.118353843688965, "learning_rate": 8.109421703699859e-07, "loss": 0.1444, "step": 32850 }, { "epoch": 0.644313725490196, "grad_norm": 60.79460525512695, "learning_rate": 8.108152082200161e-07, "loss": 0.1523, "step": 32860 }, { "epoch": 0.6445098039215686, "grad_norm": 1.7356334924697876, "learning_rate": 8.106882133996468e-07, "loss": 0.1418, "step": 32870 }, { "epoch": 0.6447058823529411, "grad_norm": 3.8513782024383545, "learning_rate": 8.105611859222263e-07, "loss": 0.1728, "step": 32880 }, { "epoch": 0.6449019607843137, "grad_norm": 2.857294797897339, "learning_rate": 8.104341258011069e-07, "loss": 0.1904, "step": 32890 }, { "epoch": 0.6450980392156863, "grad_norm": 2.6593925952911377, "learning_rate": 8.103070330496441e-07, "loss": 0.1506, "step": 32900 }, { "epoch": 0.6452941176470588, "grad_norm": 4.016684532165527, "learning_rate": 8.101799076811969e-07, "loss": 0.1752, "step": 32910 }, { "epoch": 0.6454901960784314, "grad_norm": 2.0789315700531006, "learning_rate": 8.100527497091276e-07, "loss": 0.111, "step": 32920 }, { "epoch": 0.6456862745098039, "grad_norm": 3.5777316093444824, "learning_rate": 8.099255591468021e-07, "loss": 0.1783, "step": 32930 }, { "epoch": 0.6458823529411765, "grad_norm": 1.3386141061782837, "learning_rate": 8.097983360075896e-07, "loss": 0.1143, "step": 32940 }, { "epoch": 0.6460784313725491, "grad_norm": 3.0494651794433594, "learning_rate": 8.096710803048628e-07, "loss": 0.1546, "step": 32950 }, { "epoch": 0.6462745098039215, "grad_norm": 5.968482971191406, "learning_rate": 8.095437920519978e-07, "loss": 0.178, "step": 32960 }, { "epoch": 0.6464705882352941, "grad_norm": 2.0141897201538086, "learning_rate": 8.094164712623743e-07, "loss": 0.1916, "step": 32970 }, { "epoch": 0.6466666666666666, "grad_norm": 2.3900320529937744, "learning_rate": 8.092891179493749e-07, "loss": 0.1345, "step": 32980 }, { "epoch": 0.6468627450980392, "grad_norm": 1.3541507720947266, "learning_rate": 8.091617321263861e-07, "loss": 0.1757, "step": 32990 }, { "epoch": 0.6470588235294118, "grad_norm": 2.33510160446167, "learning_rate": 8.090343138067979e-07, "loss": 0.1305, "step": 33000 }, { "epoch": 0.6472549019607843, "grad_norm": 4.192862033843994, "learning_rate": 8.089068630040032e-07, "loss": 0.1649, "step": 33010 }, { "epoch": 0.6474509803921569, "grad_norm": 2.873011589050293, "learning_rate": 8.087793797313987e-07, "loss": 0.1521, "step": 33020 }, { "epoch": 0.6476470588235295, "grad_norm": 3.3632824420928955, "learning_rate": 8.086518640023845e-07, "loss": 0.1477, "step": 33030 }, { "epoch": 0.647843137254902, "grad_norm": 3.582653760910034, "learning_rate": 8.085243158303637e-07, "loss": 0.1338, "step": 33040 }, { "epoch": 0.6480392156862745, "grad_norm": 2.1550583839416504, "learning_rate": 8.083967352287437e-07, "loss": 0.1429, "step": 33050 }, { "epoch": 0.648235294117647, "grad_norm": 2.537113904953003, "learning_rate": 8.082691222109341e-07, "loss": 0.1731, "step": 33060 }, { "epoch": 0.6484313725490196, "grad_norm": 2.4416158199310303, "learning_rate": 8.081414767903491e-07, "loss": 0.1775, "step": 33070 }, { "epoch": 0.6486274509803922, "grad_norm": 2.6441051959991455, "learning_rate": 8.080137989804057e-07, "loss": 0.1697, "step": 33080 }, { "epoch": 0.6488235294117647, "grad_norm": 2.9481287002563477, "learning_rate": 8.078860887945241e-07, "loss": 0.1605, "step": 33090 }, { "epoch": 0.6490196078431373, "grad_norm": 3.1445047855377197, "learning_rate": 8.077583462461282e-07, "loss": 0.1664, "step": 33100 }, { "epoch": 0.6492156862745098, "grad_norm": 2.491447925567627, "learning_rate": 8.076305713486453e-07, "loss": 0.1777, "step": 33110 }, { "epoch": 0.6494117647058824, "grad_norm": 2.9907872676849365, "learning_rate": 8.075027641155064e-07, "loss": 0.1653, "step": 33120 }, { "epoch": 0.649607843137255, "grad_norm": 2.1399011611938477, "learning_rate": 8.07374924560145e-07, "loss": 0.1567, "step": 33130 }, { "epoch": 0.6498039215686274, "grad_norm": 3.526960849761963, "learning_rate": 8.07247052695999e-07, "loss": 0.1729, "step": 33140 }, { "epoch": 0.65, "grad_norm": 3.368643283843994, "learning_rate": 8.071191485365092e-07, "loss": 0.1451, "step": 33150 }, { "epoch": 0.6501960784313725, "grad_norm": 3.2483792304992676, "learning_rate": 8.069912120951197e-07, "loss": 0.171, "step": 33160 }, { "epoch": 0.6503921568627451, "grad_norm": 2.4610824584960938, "learning_rate": 8.068632433852784e-07, "loss": 0.1735, "step": 33170 }, { "epoch": 0.6505882352941177, "grad_norm": 4.005029678344727, "learning_rate": 8.067352424204361e-07, "loss": 0.1663, "step": 33180 }, { "epoch": 0.6507843137254902, "grad_norm": 5.252844333648682, "learning_rate": 8.066072092140473e-07, "loss": 0.1682, "step": 33190 }, { "epoch": 0.6509803921568628, "grad_norm": 4.071515083312988, "learning_rate": 8.064791437795698e-07, "loss": 0.1702, "step": 33200 }, { "epoch": 0.6511764705882352, "grad_norm": 1.6985307931900024, "learning_rate": 8.06351046130465e-07, "loss": 0.1403, "step": 33210 }, { "epoch": 0.6513725490196078, "grad_norm": 1.3598688840866089, "learning_rate": 8.062229162801972e-07, "loss": 0.1539, "step": 33220 }, { "epoch": 0.6515686274509804, "grad_norm": 4.079412460327148, "learning_rate": 8.060947542422346e-07, "loss": 0.1544, "step": 33230 }, { "epoch": 0.6517647058823529, "grad_norm": 3.1254584789276123, "learning_rate": 8.059665600300485e-07, "loss": 0.1443, "step": 33240 }, { "epoch": 0.6519607843137255, "grad_norm": 2.3675246238708496, "learning_rate": 8.058383336571138e-07, "loss": 0.155, "step": 33250 }, { "epoch": 0.652156862745098, "grad_norm": 1.1821318864822388, "learning_rate": 8.057100751369081e-07, "loss": 0.1302, "step": 33260 }, { "epoch": 0.6523529411764706, "grad_norm": 2.4572503566741943, "learning_rate": 8.055817844829136e-07, "loss": 0.1483, "step": 33270 }, { "epoch": 0.6525490196078432, "grad_norm": 2.113262414932251, "learning_rate": 8.054534617086149e-07, "loss": 0.1701, "step": 33280 }, { "epoch": 0.6527450980392157, "grad_norm": 2.7937002182006836, "learning_rate": 8.053251068275e-07, "loss": 0.1369, "step": 33290 }, { "epoch": 0.6529411764705882, "grad_norm": 5.394260406494141, "learning_rate": 8.05196719853061e-07, "loss": 0.1725, "step": 33300 }, { "epoch": 0.6531372549019608, "grad_norm": 3.966327428817749, "learning_rate": 8.050683007987924e-07, "loss": 0.165, "step": 33310 }, { "epoch": 0.6533333333333333, "grad_norm": 3.2484138011932373, "learning_rate": 8.04939849678193e-07, "loss": 0.1582, "step": 33320 }, { "epoch": 0.6535294117647059, "grad_norm": 2.8485264778137207, "learning_rate": 8.048113665047642e-07, "loss": 0.1702, "step": 33330 }, { "epoch": 0.6537254901960784, "grad_norm": 1.8129159212112427, "learning_rate": 8.046828512920116e-07, "loss": 0.175, "step": 33340 }, { "epoch": 0.653921568627451, "grad_norm": 2.6199302673339844, "learning_rate": 8.045543040534433e-07, "loss": 0.1563, "step": 33350 }, { "epoch": 0.6541176470588236, "grad_norm": 2.681779384613037, "learning_rate": 8.044257248025712e-07, "loss": 0.1877, "step": 33360 }, { "epoch": 0.6543137254901961, "grad_norm": 3.231611728668213, "learning_rate": 8.042971135529106e-07, "loss": 0.1415, "step": 33370 }, { "epoch": 0.6545098039215687, "grad_norm": 4.715409278869629, "learning_rate": 8.0416847031798e-07, "loss": 0.1836, "step": 33380 }, { "epoch": 0.6547058823529411, "grad_norm": 7.531062126159668, "learning_rate": 8.040397951113016e-07, "loss": 0.1425, "step": 33390 }, { "epoch": 0.6549019607843137, "grad_norm": 1.9405553340911865, "learning_rate": 8.039110879464004e-07, "loss": 0.1377, "step": 33400 }, { "epoch": 0.6550980392156863, "grad_norm": 2.2290971279144287, "learning_rate": 8.037823488368052e-07, "loss": 0.1404, "step": 33410 }, { "epoch": 0.6552941176470588, "grad_norm": 2.1300792694091797, "learning_rate": 8.036535777960479e-07, "loss": 0.166, "step": 33420 }, { "epoch": 0.6554901960784314, "grad_norm": 1.8078540563583374, "learning_rate": 8.035247748376641e-07, "loss": 0.1598, "step": 33430 }, { "epoch": 0.6556862745098039, "grad_norm": 1.5037360191345215, "learning_rate": 8.033959399751923e-07, "loss": 0.1349, "step": 33440 }, { "epoch": 0.6558823529411765, "grad_norm": 1.8865004777908325, "learning_rate": 8.032670732221747e-07, "loss": 0.1562, "step": 33450 }, { "epoch": 0.6560784313725491, "grad_norm": 1.5439610481262207, "learning_rate": 8.031381745921568e-07, "loss": 0.1508, "step": 33460 }, { "epoch": 0.6562745098039215, "grad_norm": 3.0070786476135254, "learning_rate": 8.030092440986873e-07, "loss": 0.1395, "step": 33470 }, { "epoch": 0.6564705882352941, "grad_norm": 1.2946913242340088, "learning_rate": 8.028802817553185e-07, "loss": 0.1435, "step": 33480 }, { "epoch": 0.6566666666666666, "grad_norm": 3.854668378829956, "learning_rate": 8.027512875756054e-07, "loss": 0.1743, "step": 33490 }, { "epoch": 0.6568627450980392, "grad_norm": 2.173403739929199, "learning_rate": 8.026222615731073e-07, "loss": 0.1416, "step": 33500 }, { "epoch": 0.6570588235294118, "grad_norm": 3.000952959060669, "learning_rate": 8.024932037613864e-07, "loss": 0.1568, "step": 33510 }, { "epoch": 0.6572549019607843, "grad_norm": 4.084436416625977, "learning_rate": 8.02364114154008e-07, "loss": 0.1508, "step": 33520 }, { "epoch": 0.6574509803921569, "grad_norm": 1.3054313659667969, "learning_rate": 8.02234992764541e-07, "loss": 0.1507, "step": 33530 }, { "epoch": 0.6576470588235294, "grad_norm": 3.9762325286865234, "learning_rate": 8.021058396065578e-07, "loss": 0.1411, "step": 33540 }, { "epoch": 0.657843137254902, "grad_norm": 2.025737762451172, "learning_rate": 8.019766546936335e-07, "loss": 0.1715, "step": 33550 }, { "epoch": 0.6580392156862745, "grad_norm": 4.459770679473877, "learning_rate": 8.018474380393477e-07, "loss": 0.1542, "step": 33560 }, { "epoch": 0.658235294117647, "grad_norm": 2.6814143657684326, "learning_rate": 8.017181896572818e-07, "loss": 0.1477, "step": 33570 }, { "epoch": 0.6584313725490196, "grad_norm": 2.7693588733673096, "learning_rate": 8.015889095610218e-07, "loss": 0.1461, "step": 33580 }, { "epoch": 0.6586274509803921, "grad_norm": 7.563344955444336, "learning_rate": 8.014595977641566e-07, "loss": 0.1501, "step": 33590 }, { "epoch": 0.6588235294117647, "grad_norm": 5.453262805938721, "learning_rate": 8.013302542802782e-07, "loss": 0.1612, "step": 33600 }, { "epoch": 0.6590196078431373, "grad_norm": 1.815026044845581, "learning_rate": 8.012008791229825e-07, "loss": 0.1736, "step": 33610 }, { "epoch": 0.6592156862745098, "grad_norm": 2.0678327083587646, "learning_rate": 8.010714723058682e-07, "loss": 0.1195, "step": 33620 }, { "epoch": 0.6594117647058824, "grad_norm": 3.0504093170166016, "learning_rate": 8.009420338425373e-07, "loss": 0.1362, "step": 33630 }, { "epoch": 0.659607843137255, "grad_norm": 2.7428126335144043, "learning_rate": 8.008125637465956e-07, "loss": 0.1773, "step": 33640 }, { "epoch": 0.6598039215686274, "grad_norm": 2.8056721687316895, "learning_rate": 8.006830620316518e-07, "loss": 0.1684, "step": 33650 }, { "epoch": 0.66, "grad_norm": 2.0372867584228516, "learning_rate": 8.005535287113181e-07, "loss": 0.1522, "step": 33660 }, { "epoch": 0.6601960784313725, "grad_norm": 3.232808828353882, "learning_rate": 8.004239637992102e-07, "loss": 0.19, "step": 33670 }, { "epoch": 0.6603921568627451, "grad_norm": 1.7562810182571411, "learning_rate": 8.002943673089466e-07, "loss": 0.1421, "step": 33680 }, { "epoch": 0.6605882352941177, "grad_norm": 1.9229047298431396, "learning_rate": 8.001647392541498e-07, "loss": 0.1407, "step": 33690 }, { "epoch": 0.6607843137254902, "grad_norm": 2.4670844078063965, "learning_rate": 8.00035079648445e-07, "loss": 0.1301, "step": 33700 }, { "epoch": 0.6609803921568628, "grad_norm": 2.258941173553467, "learning_rate": 7.99905388505461e-07, "loss": 0.1521, "step": 33710 }, { "epoch": 0.6611764705882353, "grad_norm": 3.303541421890259, "learning_rate": 7.997756658388299e-07, "loss": 0.1544, "step": 33720 }, { "epoch": 0.6613725490196078, "grad_norm": 2.741373300552368, "learning_rate": 7.996459116621874e-07, "loss": 0.1776, "step": 33730 }, { "epoch": 0.6615686274509804, "grad_norm": 1.6846027374267578, "learning_rate": 7.995161259891717e-07, "loss": 0.1688, "step": 33740 }, { "epoch": 0.6617647058823529, "grad_norm": 7.5438103675842285, "learning_rate": 7.993863088334252e-07, "loss": 0.1989, "step": 33750 }, { "epoch": 0.6619607843137255, "grad_norm": 1.6704068183898926, "learning_rate": 7.992564602085931e-07, "loss": 0.1751, "step": 33760 }, { "epoch": 0.662156862745098, "grad_norm": 2.3485193252563477, "learning_rate": 7.991265801283241e-07, "loss": 0.1522, "step": 33770 }, { "epoch": 0.6623529411764706, "grad_norm": 2.254896402359009, "learning_rate": 7.989966686062702e-07, "loss": 0.1297, "step": 33780 }, { "epoch": 0.6625490196078432, "grad_norm": 2.323906183242798, "learning_rate": 7.988667256560864e-07, "loss": 0.1288, "step": 33790 }, { "epoch": 0.6627450980392157, "grad_norm": 3.990797996520996, "learning_rate": 7.987367512914316e-07, "loss": 0.1574, "step": 33800 }, { "epoch": 0.6629411764705883, "grad_norm": 2.886012315750122, "learning_rate": 7.986067455259675e-07, "loss": 0.1305, "step": 33810 }, { "epoch": 0.6631372549019607, "grad_norm": 1.8266488313674927, "learning_rate": 7.984767083733593e-07, "loss": 0.1191, "step": 33820 }, { "epoch": 0.6633333333333333, "grad_norm": 4.1851043701171875, "learning_rate": 7.983466398472753e-07, "loss": 0.1659, "step": 33830 }, { "epoch": 0.6635294117647059, "grad_norm": 1.4893244504928589, "learning_rate": 7.982165399613875e-07, "loss": 0.1317, "step": 33840 }, { "epoch": 0.6637254901960784, "grad_norm": 1.8875293731689453, "learning_rate": 7.980864087293707e-07, "loss": 0.1621, "step": 33850 }, { "epoch": 0.663921568627451, "grad_norm": 2.295947313308716, "learning_rate": 7.979562461649036e-07, "loss": 0.1506, "step": 33860 }, { "epoch": 0.6641176470588235, "grad_norm": 2.94598388671875, "learning_rate": 7.978260522816674e-07, "loss": 0.1667, "step": 33870 }, { "epoch": 0.6643137254901961, "grad_norm": 1.8321856260299683, "learning_rate": 7.976958270933475e-07, "loss": 0.1559, "step": 33880 }, { "epoch": 0.6645098039215687, "grad_norm": 5.993175029754639, "learning_rate": 7.975655706136318e-07, "loss": 0.1807, "step": 33890 }, { "epoch": 0.6647058823529411, "grad_norm": 1.7002314329147339, "learning_rate": 7.974352828562119e-07, "loss": 0.1611, "step": 33900 }, { "epoch": 0.6649019607843137, "grad_norm": 3.5020148754119873, "learning_rate": 7.973049638347827e-07, "loss": 0.1909, "step": 33910 }, { "epoch": 0.6650980392156862, "grad_norm": 2.536724328994751, "learning_rate": 7.971746135630422e-07, "loss": 0.1844, "step": 33920 }, { "epoch": 0.6652941176470588, "grad_norm": 2.499518632888794, "learning_rate": 7.970442320546917e-07, "loss": 0.1655, "step": 33930 }, { "epoch": 0.6654901960784314, "grad_norm": 1.853437900543213, "learning_rate": 7.969138193234359e-07, "loss": 0.1854, "step": 33940 }, { "epoch": 0.6656862745098039, "grad_norm": 3.339930534362793, "learning_rate": 7.967833753829829e-07, "loss": 0.1729, "step": 33950 }, { "epoch": 0.6658823529411765, "grad_norm": 3.1973822116851807, "learning_rate": 7.966529002470439e-07, "loss": 0.1895, "step": 33960 }, { "epoch": 0.6660784313725491, "grad_norm": 1.9758350849151611, "learning_rate": 7.965223939293331e-07, "loss": 0.166, "step": 33970 }, { "epoch": 0.6662745098039216, "grad_norm": 3.564394235610962, "learning_rate": 7.963918564435685e-07, "loss": 0.1694, "step": 33980 }, { "epoch": 0.6664705882352941, "grad_norm": 3.8977818489074707, "learning_rate": 7.962612878034714e-07, "loss": 0.1635, "step": 33990 }, { "epoch": 0.6666666666666666, "grad_norm": 4.0556488037109375, "learning_rate": 7.961306880227656e-07, "loss": 0.1711, "step": 34000 }, { "epoch": 0.6668627450980392, "grad_norm": 2.763594388961792, "learning_rate": 7.96000057115179e-07, "loss": 0.1792, "step": 34010 }, { "epoch": 0.6670588235294118, "grad_norm": 2.6212403774261475, "learning_rate": 7.958693950944424e-07, "loss": 0.1698, "step": 34020 }, { "epoch": 0.6672549019607843, "grad_norm": 2.543008327484131, "learning_rate": 7.9573870197429e-07, "loss": 0.1726, "step": 34030 }, { "epoch": 0.6674509803921569, "grad_norm": 3.519251823425293, "learning_rate": 7.956079777684592e-07, "loss": 0.16, "step": 34040 }, { "epoch": 0.6676470588235294, "grad_norm": 4.626302719116211, "learning_rate": 7.954772224906907e-07, "loss": 0.1944, "step": 34050 }, { "epoch": 0.667843137254902, "grad_norm": 4.749927520751953, "learning_rate": 7.953464361547283e-07, "loss": 0.1431, "step": 34060 }, { "epoch": 0.6680392156862746, "grad_norm": 2.178975820541382, "learning_rate": 7.952156187743194e-07, "loss": 0.1737, "step": 34070 }, { "epoch": 0.668235294117647, "grad_norm": 2.0019125938415527, "learning_rate": 7.950847703632144e-07, "loss": 0.1941, "step": 34080 }, { "epoch": 0.6684313725490196, "grad_norm": 2.000467300415039, "learning_rate": 7.94953890935167e-07, "loss": 0.1267, "step": 34090 }, { "epoch": 0.6686274509803921, "grad_norm": 2.7011771202087402, "learning_rate": 7.948229805039341e-07, "loss": 0.1573, "step": 34100 }, { "epoch": 0.6688235294117647, "grad_norm": 2.215557098388672, "learning_rate": 7.946920390832763e-07, "loss": 0.1528, "step": 34110 }, { "epoch": 0.6690196078431373, "grad_norm": 2.6458561420440674, "learning_rate": 7.945610666869567e-07, "loss": 0.1945, "step": 34120 }, { "epoch": 0.6692156862745098, "grad_norm": 2.623213291168213, "learning_rate": 7.944300633287425e-07, "loss": 0.1888, "step": 34130 }, { "epoch": 0.6694117647058824, "grad_norm": 4.532837867736816, "learning_rate": 7.942990290224032e-07, "loss": 0.1487, "step": 34140 }, { "epoch": 0.6696078431372549, "grad_norm": 2.544497489929199, "learning_rate": 7.941679637817125e-07, "loss": 0.153, "step": 34150 }, { "epoch": 0.6698039215686274, "grad_norm": 2.696030378341675, "learning_rate": 7.940368676204467e-07, "loss": 0.1479, "step": 34160 }, { "epoch": 0.67, "grad_norm": 2.8368072509765625, "learning_rate": 7.939057405523857e-07, "loss": 0.1575, "step": 34170 }, { "epoch": 0.6701960784313725, "grad_norm": 2.9460268020629883, "learning_rate": 7.937745825913125e-07, "loss": 0.1766, "step": 34180 }, { "epoch": 0.6703921568627451, "grad_norm": 3.4237539768218994, "learning_rate": 7.936433937510134e-07, "loss": 0.1553, "step": 34190 }, { "epoch": 0.6705882352941176, "grad_norm": 2.4620072841644287, "learning_rate": 7.935121740452777e-07, "loss": 0.1445, "step": 34200 }, { "epoch": 0.6707843137254902, "grad_norm": 2.0602259635925293, "learning_rate": 7.933809234878984e-07, "loss": 0.167, "step": 34210 }, { "epoch": 0.6709803921568628, "grad_norm": 3.143026351928711, "learning_rate": 7.932496420926715e-07, "loss": 0.155, "step": 34220 }, { "epoch": 0.6711764705882353, "grad_norm": 2.481452703475952, "learning_rate": 7.931183298733961e-07, "loss": 0.1384, "step": 34230 }, { "epoch": 0.6713725490196079, "grad_norm": 3.328357219696045, "learning_rate": 7.929869868438748e-07, "loss": 0.1609, "step": 34240 }, { "epoch": 0.6715686274509803, "grad_norm": 3.328150749206543, "learning_rate": 7.928556130179134e-07, "loss": 0.1517, "step": 34250 }, { "epoch": 0.6717647058823529, "grad_norm": 1.5788630247116089, "learning_rate": 7.927242084093205e-07, "loss": 0.1355, "step": 34260 }, { "epoch": 0.6719607843137255, "grad_norm": 2.145380735397339, "learning_rate": 7.925927730319088e-07, "loss": 0.1782, "step": 34270 }, { "epoch": 0.672156862745098, "grad_norm": 3.1438825130462646, "learning_rate": 7.924613068994935e-07, "loss": 0.1579, "step": 34280 }, { "epoch": 0.6723529411764706, "grad_norm": 2.759648084640503, "learning_rate": 7.923298100258928e-07, "loss": 0.1671, "step": 34290 }, { "epoch": 0.6725490196078432, "grad_norm": 2.5851540565490723, "learning_rate": 7.921982824249294e-07, "loss": 0.1352, "step": 34300 }, { "epoch": 0.6727450980392157, "grad_norm": 1.509209394454956, "learning_rate": 7.920667241104281e-07, "loss": 0.1647, "step": 34310 }, { "epoch": 0.6729411764705883, "grad_norm": 2.0389652252197266, "learning_rate": 7.919351350962171e-07, "loss": 0.1487, "step": 34320 }, { "epoch": 0.6731372549019607, "grad_norm": 3.2090158462524414, "learning_rate": 7.918035153961282e-07, "loss": 0.1628, "step": 34330 }, { "epoch": 0.6733333333333333, "grad_norm": 3.2659761905670166, "learning_rate": 7.916718650239961e-07, "loss": 0.1487, "step": 34340 }, { "epoch": 0.6735294117647059, "grad_norm": 1.9475083351135254, "learning_rate": 7.915401839936588e-07, "loss": 0.1448, "step": 34350 }, { "epoch": 0.6737254901960784, "grad_norm": 2.7614524364471436, "learning_rate": 7.914084723189575e-07, "loss": 0.1903, "step": 34360 }, { "epoch": 0.673921568627451, "grad_norm": 2.0961406230926514, "learning_rate": 7.91276730013737e-07, "loss": 0.1528, "step": 34370 }, { "epoch": 0.6741176470588235, "grad_norm": 1.5547571182250977, "learning_rate": 7.911449570918445e-07, "loss": 0.1293, "step": 34380 }, { "epoch": 0.6743137254901961, "grad_norm": 2.89570951461792, "learning_rate": 7.910131535671314e-07, "loss": 0.1801, "step": 34390 }, { "epoch": 0.6745098039215687, "grad_norm": 3.557788610458374, "learning_rate": 7.908813194534515e-07, "loss": 0.1871, "step": 34400 }, { "epoch": 0.6747058823529412, "grad_norm": 2.9008662700653076, "learning_rate": 7.907494547646622e-07, "loss": 0.1712, "step": 34410 }, { "epoch": 0.6749019607843137, "grad_norm": 5.837369918823242, "learning_rate": 7.906175595146241e-07, "loss": 0.1542, "step": 34420 }, { "epoch": 0.6750980392156862, "grad_norm": 2.5171656608581543, "learning_rate": 7.904856337172011e-07, "loss": 0.1448, "step": 34430 }, { "epoch": 0.6752941176470588, "grad_norm": 2.5916097164154053, "learning_rate": 7.9035367738626e-07, "loss": 0.1698, "step": 34440 }, { "epoch": 0.6754901960784314, "grad_norm": 3.670297384262085, "learning_rate": 7.90221690535671e-07, "loss": 0.1297, "step": 34450 }, { "epoch": 0.6756862745098039, "grad_norm": 3.3115198612213135, "learning_rate": 7.900896731793076e-07, "loss": 0.1302, "step": 34460 }, { "epoch": 0.6758823529411765, "grad_norm": 4.887085437774658, "learning_rate": 7.899576253310463e-07, "loss": 0.183, "step": 34470 }, { "epoch": 0.676078431372549, "grad_norm": 2.935765266418457, "learning_rate": 7.898255470047669e-07, "loss": 0.1713, "step": 34480 }, { "epoch": 0.6762745098039216, "grad_norm": 2.4889886379241943, "learning_rate": 7.896934382143525e-07, "loss": 0.1614, "step": 34490 }, { "epoch": 0.6764705882352942, "grad_norm": 2.945502996444702, "learning_rate": 7.895612989736893e-07, "loss": 0.1798, "step": 34500 }, { "epoch": 0.6766666666666666, "grad_norm": 2.515206813812256, "learning_rate": 7.894291292966668e-07, "loss": 0.1752, "step": 34510 }, { "epoch": 0.6768627450980392, "grad_norm": 2.3483731746673584, "learning_rate": 7.892969291971774e-07, "loss": 0.1552, "step": 34520 }, { "epoch": 0.6770588235294117, "grad_norm": 3.1746103763580322, "learning_rate": 7.891646986891172e-07, "loss": 0.1638, "step": 34530 }, { "epoch": 0.6772549019607843, "grad_norm": 2.912621021270752, "learning_rate": 7.890324377863849e-07, "loss": 0.1669, "step": 34540 }, { "epoch": 0.6774509803921569, "grad_norm": 1.4209076166152954, "learning_rate": 7.889001465028828e-07, "loss": 0.1722, "step": 34550 }, { "epoch": 0.6776470588235294, "grad_norm": 1.8095186948776245, "learning_rate": 7.887678248525163e-07, "loss": 0.1501, "step": 34560 }, { "epoch": 0.677843137254902, "grad_norm": 2.991658926010132, "learning_rate": 7.886354728491941e-07, "loss": 0.1576, "step": 34570 }, { "epoch": 0.6780392156862745, "grad_norm": 3.8393192291259766, "learning_rate": 7.885030905068279e-07, "loss": 0.1293, "step": 34580 }, { "epoch": 0.678235294117647, "grad_norm": 3.066023111343384, "learning_rate": 7.883706778393329e-07, "loss": 0.1258, "step": 34590 }, { "epoch": 0.6784313725490196, "grad_norm": 2.475964069366455, "learning_rate": 7.882382348606267e-07, "loss": 0.1625, "step": 34600 }, { "epoch": 0.6786274509803921, "grad_norm": 1.7047489881515503, "learning_rate": 7.881057615846312e-07, "loss": 0.1662, "step": 34610 }, { "epoch": 0.6788235294117647, "grad_norm": 2.5935311317443848, "learning_rate": 7.879732580252707e-07, "loss": 0.1985, "step": 34620 }, { "epoch": 0.6790196078431373, "grad_norm": 2.26755690574646, "learning_rate": 7.878407241964728e-07, "loss": 0.1635, "step": 34630 }, { "epoch": 0.6792156862745098, "grad_norm": 2.5444388389587402, "learning_rate": 7.877081601121687e-07, "loss": 0.1472, "step": 34640 }, { "epoch": 0.6794117647058824, "grad_norm": 3.7074575424194336, "learning_rate": 7.875755657862921e-07, "loss": 0.1375, "step": 34650 }, { "epoch": 0.6796078431372549, "grad_norm": 3.416860818862915, "learning_rate": 7.874429412327804e-07, "loss": 0.1533, "step": 34660 }, { "epoch": 0.6798039215686275, "grad_norm": 2.8053829669952393, "learning_rate": 7.873102864655742e-07, "loss": 0.1927, "step": 34670 }, { "epoch": 0.68, "grad_norm": 2.4342751502990723, "learning_rate": 7.871776014986169e-07, "loss": 0.1271, "step": 34680 }, { "epoch": 0.6801960784313725, "grad_norm": 2.2876102924346924, "learning_rate": 7.870448863458554e-07, "loss": 0.1434, "step": 34690 }, { "epoch": 0.6803921568627451, "grad_norm": 3.139826774597168, "learning_rate": 7.869121410212396e-07, "loss": 0.1702, "step": 34700 }, { "epoch": 0.6805882352941176, "grad_norm": 1.6548032760620117, "learning_rate": 7.867793655387225e-07, "loss": 0.1461, "step": 34710 }, { "epoch": 0.6807843137254902, "grad_norm": 2.0491139888763428, "learning_rate": 7.866465599122607e-07, "loss": 0.1617, "step": 34720 }, { "epoch": 0.6809803921568628, "grad_norm": 2.9724466800689697, "learning_rate": 7.865137241558133e-07, "loss": 0.1433, "step": 34730 }, { "epoch": 0.6811764705882353, "grad_norm": 2.5815558433532715, "learning_rate": 7.863808582833431e-07, "loss": 0.164, "step": 34740 }, { "epoch": 0.6813725490196079, "grad_norm": 1.9875789880752563, "learning_rate": 7.86247962308816e-07, "loss": 0.1677, "step": 34750 }, { "epoch": 0.6815686274509803, "grad_norm": 4.142509460449219, "learning_rate": 7.861150362462008e-07, "loss": 0.1376, "step": 34760 }, { "epoch": 0.6817647058823529, "grad_norm": 3.395169973373413, "learning_rate": 7.859820801094699e-07, "loss": 0.128, "step": 34770 }, { "epoch": 0.6819607843137255, "grad_norm": 2.571115732192993, "learning_rate": 7.858490939125981e-07, "loss": 0.1682, "step": 34780 }, { "epoch": 0.682156862745098, "grad_norm": 3.394460439682007, "learning_rate": 7.857160776695641e-07, "loss": 0.1628, "step": 34790 }, { "epoch": 0.6823529411764706, "grad_norm": 2.374249219894409, "learning_rate": 7.855830313943496e-07, "loss": 0.1571, "step": 34800 }, { "epoch": 0.6825490196078431, "grad_norm": 3.037092447280884, "learning_rate": 7.854499551009393e-07, "loss": 0.1656, "step": 34810 }, { "epoch": 0.6827450980392157, "grad_norm": 1.4296932220458984, "learning_rate": 7.853168488033209e-07, "loss": 0.1505, "step": 34820 }, { "epoch": 0.6829411764705883, "grad_norm": 2.1727397441864014, "learning_rate": 7.851837125154856e-07, "loss": 0.1502, "step": 34830 }, { "epoch": 0.6831372549019608, "grad_norm": 2.5374209880828857, "learning_rate": 7.850505462514278e-07, "loss": 0.1723, "step": 34840 }, { "epoch": 0.6833333333333333, "grad_norm": 2.449524402618408, "learning_rate": 7.849173500251447e-07, "loss": 0.1603, "step": 34850 }, { "epoch": 0.6835294117647058, "grad_norm": 3.8911867141723633, "learning_rate": 7.84784123850637e-07, "loss": 0.1535, "step": 34860 }, { "epoch": 0.6837254901960784, "grad_norm": 2.7801403999328613, "learning_rate": 7.84650867741908e-07, "loss": 0.1581, "step": 34870 }, { "epoch": 0.683921568627451, "grad_norm": 2.793788433074951, "learning_rate": 7.845175817129647e-07, "loss": 0.1543, "step": 34880 }, { "epoch": 0.6841176470588235, "grad_norm": 2.8165478706359863, "learning_rate": 7.843842657778172e-07, "loss": 0.1565, "step": 34890 }, { "epoch": 0.6843137254901961, "grad_norm": 1.725196123123169, "learning_rate": 7.842509199504783e-07, "loss": 0.1523, "step": 34900 }, { "epoch": 0.6845098039215687, "grad_norm": 2.170567035675049, "learning_rate": 7.841175442449645e-07, "loss": 0.1471, "step": 34910 }, { "epoch": 0.6847058823529412, "grad_norm": 3.5286052227020264, "learning_rate": 7.83984138675295e-07, "loss": 0.1496, "step": 34920 }, { "epoch": 0.6849019607843138, "grad_norm": 2.7783408164978027, "learning_rate": 7.838507032554926e-07, "loss": 0.1766, "step": 34930 }, { "epoch": 0.6850980392156862, "grad_norm": 1.3121180534362793, "learning_rate": 7.837172379995828e-07, "loss": 0.1349, "step": 34940 }, { "epoch": 0.6852941176470588, "grad_norm": 2.5674896240234375, "learning_rate": 7.835837429215942e-07, "loss": 0.1413, "step": 34950 }, { "epoch": 0.6854901960784314, "grad_norm": 1.7862951755523682, "learning_rate": 7.834502180355591e-07, "loss": 0.1547, "step": 34960 }, { "epoch": 0.6856862745098039, "grad_norm": 3.630197525024414, "learning_rate": 7.833166633555123e-07, "loss": 0.1791, "step": 34970 }, { "epoch": 0.6858823529411765, "grad_norm": 1.5900002717971802, "learning_rate": 7.83183078895492e-07, "loss": 0.1297, "step": 34980 }, { "epoch": 0.686078431372549, "grad_norm": 4.175768852233887, "learning_rate": 7.830494646695397e-07, "loss": 0.1522, "step": 34990 }, { "epoch": 0.6862745098039216, "grad_norm": 1.7201248407363892, "learning_rate": 7.829158206916997e-07, "loss": 0.1448, "step": 35000 }, { "epoch": 0.6864705882352942, "grad_norm": 2.2250583171844482, "learning_rate": 7.827821469760196e-07, "loss": 0.1918, "step": 35010 }, { "epoch": 0.6866666666666666, "grad_norm": 2.4854111671447754, "learning_rate": 7.826484435365502e-07, "loss": 0.1679, "step": 35020 }, { "epoch": 0.6868627450980392, "grad_norm": 3.516824960708618, "learning_rate": 7.825147103873453e-07, "loss": 0.1715, "step": 35030 }, { "epoch": 0.6870588235294117, "grad_norm": 4.702867031097412, "learning_rate": 7.823809475424618e-07, "loss": 0.1729, "step": 35040 }, { "epoch": 0.6872549019607843, "grad_norm": 2.4291188716888428, "learning_rate": 7.822471550159598e-07, "loss": 0.1588, "step": 35050 }, { "epoch": 0.6874509803921569, "grad_norm": 2.1892127990722656, "learning_rate": 7.821133328219025e-07, "loss": 0.1531, "step": 35060 }, { "epoch": 0.6876470588235294, "grad_norm": 1.5417503118515015, "learning_rate": 7.819794809743562e-07, "loss": 0.1494, "step": 35070 }, { "epoch": 0.687843137254902, "grad_norm": 3.427177906036377, "learning_rate": 7.818455994873904e-07, "loss": 0.1557, "step": 35080 }, { "epoch": 0.6880392156862745, "grad_norm": 4.03688907623291, "learning_rate": 7.817116883750776e-07, "loss": 0.1718, "step": 35090 }, { "epoch": 0.6882352941176471, "grad_norm": 1.148983359336853, "learning_rate": 7.815777476514933e-07, "loss": 0.173, "step": 35100 }, { "epoch": 0.6884313725490196, "grad_norm": 4.16136360168457, "learning_rate": 7.814437773307165e-07, "loss": 0.1456, "step": 35110 }, { "epoch": 0.6886274509803921, "grad_norm": 2.242173433303833, "learning_rate": 7.813097774268292e-07, "loss": 0.1298, "step": 35120 }, { "epoch": 0.6888235294117647, "grad_norm": 2.4488236904144287, "learning_rate": 7.811757479539161e-07, "loss": 0.1534, "step": 35130 }, { "epoch": 0.6890196078431372, "grad_norm": 3.2981607913970947, "learning_rate": 7.810416889260653e-07, "loss": 0.1788, "step": 35140 }, { "epoch": 0.6892156862745098, "grad_norm": 2.962474822998047, "learning_rate": 7.809076003573682e-07, "loss": 0.1535, "step": 35150 }, { "epoch": 0.6894117647058824, "grad_norm": 2.7576394081115723, "learning_rate": 7.80773482261919e-07, "loss": 0.1468, "step": 35160 }, { "epoch": 0.6896078431372549, "grad_norm": 2.3642451763153076, "learning_rate": 7.806393346538153e-07, "loss": 0.1545, "step": 35170 }, { "epoch": 0.6898039215686275, "grad_norm": 1.9759970903396606, "learning_rate": 7.805051575471573e-07, "loss": 0.1604, "step": 35180 }, { "epoch": 0.69, "grad_norm": 3.243734836578369, "learning_rate": 7.803709509560488e-07, "loss": 0.1294, "step": 35190 }, { "epoch": 0.6901960784313725, "grad_norm": 7.274235725402832, "learning_rate": 7.802367148945966e-07, "loss": 0.1672, "step": 35200 }, { "epoch": 0.6903921568627451, "grad_norm": 3.2042131423950195, "learning_rate": 7.801024493769104e-07, "loss": 0.1527, "step": 35210 }, { "epoch": 0.6905882352941176, "grad_norm": 1.949638843536377, "learning_rate": 7.799681544171032e-07, "loss": 0.1439, "step": 35220 }, { "epoch": 0.6907843137254902, "grad_norm": 5.704008102416992, "learning_rate": 7.79833830029291e-07, "loss": 0.1659, "step": 35230 }, { "epoch": 0.6909803921568628, "grad_norm": 11.340097427368164, "learning_rate": 7.796994762275927e-07, "loss": 0.1526, "step": 35240 }, { "epoch": 0.6911764705882353, "grad_norm": 2.5568628311157227, "learning_rate": 7.795650930261307e-07, "loss": 0.1488, "step": 35250 }, { "epoch": 0.6913725490196079, "grad_norm": 1.7899116277694702, "learning_rate": 7.794306804390302e-07, "loss": 0.1445, "step": 35260 }, { "epoch": 0.6915686274509804, "grad_norm": 4.328381538391113, "learning_rate": 7.792962384804197e-07, "loss": 0.1689, "step": 35270 }, { "epoch": 0.691764705882353, "grad_norm": 3.1192805767059326, "learning_rate": 7.791617671644304e-07, "loss": 0.184, "step": 35280 }, { "epoch": 0.6919607843137255, "grad_norm": 3.0824854373931885, "learning_rate": 7.790272665051971e-07, "loss": 0.1283, "step": 35290 }, { "epoch": 0.692156862745098, "grad_norm": 1.933801293373108, "learning_rate": 7.788927365168574e-07, "loss": 0.144, "step": 35300 }, { "epoch": 0.6923529411764706, "grad_norm": 2.826002836227417, "learning_rate": 7.787581772135519e-07, "loss": 0.1247, "step": 35310 }, { "epoch": 0.6925490196078431, "grad_norm": 1.3877465724945068, "learning_rate": 7.786235886094245e-07, "loss": 0.1377, "step": 35320 }, { "epoch": 0.6927450980392157, "grad_norm": 2.2898640632629395, "learning_rate": 7.78488970718622e-07, "loss": 0.1601, "step": 35330 }, { "epoch": 0.6929411764705883, "grad_norm": 4.499422073364258, "learning_rate": 7.783543235552943e-07, "loss": 0.1742, "step": 35340 }, { "epoch": 0.6931372549019608, "grad_norm": 2.3637101650238037, "learning_rate": 7.782196471335945e-07, "loss": 0.1456, "step": 35350 }, { "epoch": 0.6933333333333334, "grad_norm": 2.3307559490203857, "learning_rate": 7.780849414676787e-07, "loss": 0.1546, "step": 35360 }, { "epoch": 0.6935294117647058, "grad_norm": 6.920168399810791, "learning_rate": 7.779502065717061e-07, "loss": 0.1633, "step": 35370 }, { "epoch": 0.6937254901960784, "grad_norm": 2.668877124786377, "learning_rate": 7.778154424598388e-07, "loss": 0.1337, "step": 35380 }, { "epoch": 0.693921568627451, "grad_norm": 2.7234997749328613, "learning_rate": 7.776806491462424e-07, "loss": 0.1753, "step": 35390 }, { "epoch": 0.6941176470588235, "grad_norm": 1.8098315000534058, "learning_rate": 7.775458266450851e-07, "loss": 0.1557, "step": 35400 }, { "epoch": 0.6943137254901961, "grad_norm": 2.325291156768799, "learning_rate": 7.774109749705386e-07, "loss": 0.1522, "step": 35410 }, { "epoch": 0.6945098039215686, "grad_norm": 5.308231353759766, "learning_rate": 7.772760941367769e-07, "loss": 0.1809, "step": 35420 }, { "epoch": 0.6947058823529412, "grad_norm": 3.3304457664489746, "learning_rate": 7.771411841579783e-07, "loss": 0.159, "step": 35430 }, { "epoch": 0.6949019607843138, "grad_norm": 2.4522335529327393, "learning_rate": 7.770062450483228e-07, "loss": 0.2019, "step": 35440 }, { "epoch": 0.6950980392156862, "grad_norm": 2.2068898677825928, "learning_rate": 7.768712768219942e-07, "loss": 0.1662, "step": 35450 }, { "epoch": 0.6952941176470588, "grad_norm": 1.9685485363006592, "learning_rate": 7.767362794931798e-07, "loss": 0.1585, "step": 35460 }, { "epoch": 0.6954901960784313, "grad_norm": 3.5234463214874268, "learning_rate": 7.766012530760691e-07, "loss": 0.1717, "step": 35470 }, { "epoch": 0.6956862745098039, "grad_norm": 2.7891507148742676, "learning_rate": 7.764661975848548e-07, "loss": 0.1367, "step": 35480 }, { "epoch": 0.6958823529411765, "grad_norm": 3.1921474933624268, "learning_rate": 7.763311130337332e-07, "loss": 0.1979, "step": 35490 }, { "epoch": 0.696078431372549, "grad_norm": 1.6698758602142334, "learning_rate": 7.761959994369032e-07, "loss": 0.1803, "step": 35500 }, { "epoch": 0.6962745098039216, "grad_norm": 4.959758758544922, "learning_rate": 7.760608568085667e-07, "loss": 0.1635, "step": 35510 }, { "epoch": 0.6964705882352941, "grad_norm": 2.155846357345581, "learning_rate": 7.759256851629288e-07, "loss": 0.1526, "step": 35520 }, { "epoch": 0.6966666666666667, "grad_norm": 2.974440336227417, "learning_rate": 7.75790484514198e-07, "loss": 0.1575, "step": 35530 }, { "epoch": 0.6968627450980392, "grad_norm": 2.259807825088501, "learning_rate": 7.75655254876585e-07, "loss": 0.1396, "step": 35540 }, { "epoch": 0.6970588235294117, "grad_norm": 2.0442633628845215, "learning_rate": 7.755199962643044e-07, "loss": 0.1703, "step": 35550 }, { "epoch": 0.6972549019607843, "grad_norm": 2.6197292804718018, "learning_rate": 7.753847086915736e-07, "loss": 0.1692, "step": 35560 }, { "epoch": 0.6974509803921569, "grad_norm": 2.0671496391296387, "learning_rate": 7.752493921726126e-07, "loss": 0.1853, "step": 35570 }, { "epoch": 0.6976470588235294, "grad_norm": 3.7028613090515137, "learning_rate": 7.751140467216453e-07, "loss": 0.1352, "step": 35580 }, { "epoch": 0.697843137254902, "grad_norm": 1.9276487827301025, "learning_rate": 7.749786723528975e-07, "loss": 0.1406, "step": 35590 }, { "epoch": 0.6980392156862745, "grad_norm": 3.813544273376465, "learning_rate": 7.74843269080599e-07, "loss": 0.1741, "step": 35600 }, { "epoch": 0.6982352941176471, "grad_norm": 2.7047696113586426, "learning_rate": 7.747078369189823e-07, "loss": 0.1767, "step": 35610 }, { "epoch": 0.6984313725490197, "grad_norm": 1.31821870803833, "learning_rate": 7.745723758822829e-07, "loss": 0.1598, "step": 35620 }, { "epoch": 0.6986274509803921, "grad_norm": 2.4844954013824463, "learning_rate": 7.744368859847392e-07, "loss": 0.1685, "step": 35630 }, { "epoch": 0.6988235294117647, "grad_norm": 2.4946141242980957, "learning_rate": 7.743013672405932e-07, "loss": 0.1345, "step": 35640 }, { "epoch": 0.6990196078431372, "grad_norm": 3.051353931427002, "learning_rate": 7.741658196640891e-07, "loss": 0.175, "step": 35650 }, { "epoch": 0.6992156862745098, "grad_norm": 1.410818338394165, "learning_rate": 7.74030243269475e-07, "loss": 0.1434, "step": 35660 }, { "epoch": 0.6994117647058824, "grad_norm": 1.6669416427612305, "learning_rate": 7.738946380710014e-07, "loss": 0.1264, "step": 35670 }, { "epoch": 0.6996078431372549, "grad_norm": 3.0973284244537354, "learning_rate": 7.73759004082922e-07, "loss": 0.1638, "step": 35680 }, { "epoch": 0.6998039215686275, "grad_norm": 3.8499972820281982, "learning_rate": 7.736233413194937e-07, "loss": 0.1595, "step": 35690 }, { "epoch": 0.7, "grad_norm": 1.1866750717163086, "learning_rate": 7.734876497949761e-07, "loss": 0.1717, "step": 35700 }, { "epoch": 0.7001960784313725, "grad_norm": 2.925999402999878, "learning_rate": 7.733519295236319e-07, "loss": 0.1732, "step": 35710 }, { "epoch": 0.7003921568627451, "grad_norm": 2.353712558746338, "learning_rate": 7.732161805197272e-07, "loss": 0.1533, "step": 35720 }, { "epoch": 0.7005882352941176, "grad_norm": 2.4177281856536865, "learning_rate": 7.730804027975309e-07, "loss": 0.1781, "step": 35730 }, { "epoch": 0.7007843137254902, "grad_norm": 4.765989780426025, "learning_rate": 7.729445963713146e-07, "loss": 0.1516, "step": 35740 }, { "epoch": 0.7009803921568627, "grad_norm": 2.290468215942383, "learning_rate": 7.728087612553533e-07, "loss": 0.1433, "step": 35750 }, { "epoch": 0.7011764705882353, "grad_norm": 3.1849474906921387, "learning_rate": 7.726728974639249e-07, "loss": 0.1707, "step": 35760 }, { "epoch": 0.7013725490196079, "grad_norm": 3.3407092094421387, "learning_rate": 7.725370050113103e-07, "loss": 0.1559, "step": 35770 }, { "epoch": 0.7015686274509804, "grad_norm": 3.270143985748291, "learning_rate": 7.724010839117936e-07, "loss": 0.1542, "step": 35780 }, { "epoch": 0.701764705882353, "grad_norm": 2.4948933124542236, "learning_rate": 7.722651341796613e-07, "loss": 0.1492, "step": 35790 }, { "epoch": 0.7019607843137254, "grad_norm": 1.8674254417419434, "learning_rate": 7.721291558292038e-07, "loss": 0.1465, "step": 35800 }, { "epoch": 0.702156862745098, "grad_norm": 2.3617725372314453, "learning_rate": 7.719931488747136e-07, "loss": 0.1367, "step": 35810 }, { "epoch": 0.7023529411764706, "grad_norm": 2.1372923851013184, "learning_rate": 7.718571133304871e-07, "loss": 0.14, "step": 35820 }, { "epoch": 0.7025490196078431, "grad_norm": 2.18353009223938, "learning_rate": 7.717210492108231e-07, "loss": 0.186, "step": 35830 }, { "epoch": 0.7027450980392157, "grad_norm": 2.840219259262085, "learning_rate": 7.715849565300235e-07, "loss": 0.1674, "step": 35840 }, { "epoch": 0.7029411764705882, "grad_norm": 1.7617676258087158, "learning_rate": 7.714488353023934e-07, "loss": 0.1365, "step": 35850 }, { "epoch": 0.7031372549019608, "grad_norm": 3.9646799564361572, "learning_rate": 7.713126855422405e-07, "loss": 0.1568, "step": 35860 }, { "epoch": 0.7033333333333334, "grad_norm": 3.2907912731170654, "learning_rate": 7.711765072638762e-07, "loss": 0.1573, "step": 35870 }, { "epoch": 0.7035294117647058, "grad_norm": 1.907439112663269, "learning_rate": 7.71040300481614e-07, "loss": 0.1805, "step": 35880 }, { "epoch": 0.7037254901960784, "grad_norm": 3.3501768112182617, "learning_rate": 7.709040652097713e-07, "loss": 0.1896, "step": 35890 }, { "epoch": 0.703921568627451, "grad_norm": 16.859107971191406, "learning_rate": 7.707678014626677e-07, "loss": 0.1523, "step": 35900 }, { "epoch": 0.7041176470588235, "grad_norm": 1.9350378513336182, "learning_rate": 7.706315092546264e-07, "loss": 0.125, "step": 35910 }, { "epoch": 0.7043137254901961, "grad_norm": 1.5527974367141724, "learning_rate": 7.704951885999731e-07, "loss": 0.1194, "step": 35920 }, { "epoch": 0.7045098039215686, "grad_norm": 7.84025764465332, "learning_rate": 7.70358839513037e-07, "loss": 0.1533, "step": 35930 }, { "epoch": 0.7047058823529412, "grad_norm": 3.2878568172454834, "learning_rate": 7.702224620081498e-07, "loss": 0.1511, "step": 35940 }, { "epoch": 0.7049019607843138, "grad_norm": 2.372312307357788, "learning_rate": 7.700860560996467e-07, "loss": 0.1812, "step": 35950 }, { "epoch": 0.7050980392156863, "grad_norm": 3.9845130443573, "learning_rate": 7.699496218018654e-07, "loss": 0.1717, "step": 35960 }, { "epoch": 0.7052941176470588, "grad_norm": 3.787832498550415, "learning_rate": 7.698131591291466e-07, "loss": 0.1909, "step": 35970 }, { "epoch": 0.7054901960784313, "grad_norm": 3.5920021533966064, "learning_rate": 7.696766680958344e-07, "loss": 0.1344, "step": 35980 }, { "epoch": 0.7056862745098039, "grad_norm": 3.168285608291626, "learning_rate": 7.695401487162756e-07, "loss": 0.1696, "step": 35990 }, { "epoch": 0.7058823529411765, "grad_norm": 1.2279225587844849, "learning_rate": 7.6940360100482e-07, "loss": 0.1471, "step": 36000 }, { "epoch": 0.706078431372549, "grad_norm": 2.9594857692718506, "learning_rate": 7.692670249758204e-07, "loss": 0.1608, "step": 36010 }, { "epoch": 0.7062745098039216, "grad_norm": 2.341395139694214, "learning_rate": 7.691304206436327e-07, "loss": 0.1497, "step": 36020 }, { "epoch": 0.7064705882352941, "grad_norm": 5.795709133148193, "learning_rate": 7.689937880226152e-07, "loss": 0.158, "step": 36030 }, { "epoch": 0.7066666666666667, "grad_norm": 2.1935462951660156, "learning_rate": 7.688571271271302e-07, "loss": 0.1454, "step": 36040 }, { "epoch": 0.7068627450980393, "grad_norm": 2.6453449726104736, "learning_rate": 7.687204379715422e-07, "loss": 0.1591, "step": 36050 }, { "epoch": 0.7070588235294117, "grad_norm": 1.4385321140289307, "learning_rate": 7.685837205702187e-07, "loss": 0.1657, "step": 36060 }, { "epoch": 0.7072549019607843, "grad_norm": 3.3280820846557617, "learning_rate": 7.684469749375306e-07, "loss": 0.1406, "step": 36070 }, { "epoch": 0.7074509803921568, "grad_norm": 3.593942403793335, "learning_rate": 7.68310201087851e-07, "loss": 0.1573, "step": 36080 }, { "epoch": 0.7076470588235294, "grad_norm": 2.776400327682495, "learning_rate": 7.681733990355572e-07, "loss": 0.1724, "step": 36090 }, { "epoch": 0.707843137254902, "grad_norm": 2.8506102561950684, "learning_rate": 7.680365687950283e-07, "loss": 0.1604, "step": 36100 }, { "epoch": 0.7080392156862745, "grad_norm": 3.025982618331909, "learning_rate": 7.678997103806468e-07, "loss": 0.1452, "step": 36110 }, { "epoch": 0.7082352941176471, "grad_norm": 3.631354808807373, "learning_rate": 7.677628238067982e-07, "loss": 0.1652, "step": 36120 }, { "epoch": 0.7084313725490196, "grad_norm": 3.2991843223571777, "learning_rate": 7.676259090878709e-07, "loss": 0.1596, "step": 36130 }, { "epoch": 0.7086274509803921, "grad_norm": 2.6770238876342773, "learning_rate": 7.674889662382563e-07, "loss": 0.1293, "step": 36140 }, { "epoch": 0.7088235294117647, "grad_norm": 3.5428271293640137, "learning_rate": 7.673519952723488e-07, "loss": 0.1311, "step": 36150 }, { "epoch": 0.7090196078431372, "grad_norm": 2.8329267501831055, "learning_rate": 7.672149962045457e-07, "loss": 0.1528, "step": 36160 }, { "epoch": 0.7092156862745098, "grad_norm": 2.631303310394287, "learning_rate": 7.670779690492471e-07, "loss": 0.1632, "step": 36170 }, { "epoch": 0.7094117647058824, "grad_norm": 2.3531384468078613, "learning_rate": 7.669409138208561e-07, "loss": 0.1556, "step": 36180 }, { "epoch": 0.7096078431372549, "grad_norm": 2.654942512512207, "learning_rate": 7.668038305337791e-07, "loss": 0.1647, "step": 36190 }, { "epoch": 0.7098039215686275, "grad_norm": 3.005509376525879, "learning_rate": 7.666667192024253e-07, "loss": 0.1793, "step": 36200 }, { "epoch": 0.71, "grad_norm": 2.5048861503601074, "learning_rate": 7.665295798412063e-07, "loss": 0.1219, "step": 36210 }, { "epoch": 0.7101960784313726, "grad_norm": 3.2520296573638916, "learning_rate": 7.663924124645375e-07, "loss": 0.1474, "step": 36220 }, { "epoch": 0.7103921568627451, "grad_norm": 2.2210276126861572, "learning_rate": 7.662552170868367e-07, "loss": 0.1813, "step": 36230 }, { "epoch": 0.7105882352941176, "grad_norm": 2.2393980026245117, "learning_rate": 7.661179937225246e-07, "loss": 0.1422, "step": 36240 }, { "epoch": 0.7107843137254902, "grad_norm": 4.123987197875977, "learning_rate": 7.659807423860253e-07, "loss": 0.1596, "step": 36250 }, { "epoch": 0.7109803921568627, "grad_norm": 3.5946106910705566, "learning_rate": 7.658434630917654e-07, "loss": 0.1447, "step": 36260 }, { "epoch": 0.7111764705882353, "grad_norm": 2.4533333778381348, "learning_rate": 7.657061558541744e-07, "loss": 0.1818, "step": 36270 }, { "epoch": 0.7113725490196079, "grad_norm": 2.2999019622802734, "learning_rate": 7.655688206876855e-07, "loss": 0.1216, "step": 36280 }, { "epoch": 0.7115686274509804, "grad_norm": 2.896683931350708, "learning_rate": 7.654314576067338e-07, "loss": 0.1495, "step": 36290 }, { "epoch": 0.711764705882353, "grad_norm": 2.6258039474487305, "learning_rate": 7.65294066625758e-07, "loss": 0.1834, "step": 36300 }, { "epoch": 0.7119607843137254, "grad_norm": 3.5355844497680664, "learning_rate": 7.651566477591993e-07, "loss": 0.178, "step": 36310 }, { "epoch": 0.712156862745098, "grad_norm": 2.1882197856903076, "learning_rate": 7.650192010215022e-07, "loss": 0.1692, "step": 36320 }, { "epoch": 0.7123529411764706, "grad_norm": 3.824450731277466, "learning_rate": 7.648817264271141e-07, "loss": 0.1734, "step": 36330 }, { "epoch": 0.7125490196078431, "grad_norm": 4.630905628204346, "learning_rate": 7.647442239904854e-07, "loss": 0.1416, "step": 36340 }, { "epoch": 0.7127450980392157, "grad_norm": 2.2595179080963135, "learning_rate": 7.646066937260687e-07, "loss": 0.1748, "step": 36350 }, { "epoch": 0.7129411764705882, "grad_norm": 3.1443142890930176, "learning_rate": 7.644691356483206e-07, "loss": 0.1467, "step": 36360 }, { "epoch": 0.7131372549019608, "grad_norm": 3.852226495742798, "learning_rate": 7.643315497716998e-07, "loss": 0.2042, "step": 36370 }, { "epoch": 0.7133333333333334, "grad_norm": 1.7042903900146484, "learning_rate": 7.641939361106682e-07, "loss": 0.164, "step": 36380 }, { "epoch": 0.7135294117647059, "grad_norm": 3.5778610706329346, "learning_rate": 7.64056294679691e-07, "loss": 0.1377, "step": 36390 }, { "epoch": 0.7137254901960784, "grad_norm": 4.087423324584961, "learning_rate": 7.639186254932356e-07, "loss": 0.1813, "step": 36400 }, { "epoch": 0.7139215686274509, "grad_norm": 3.5118772983551025, "learning_rate": 7.637809285657729e-07, "loss": 0.1456, "step": 36410 }, { "epoch": 0.7141176470588235, "grad_norm": 4.626486778259277, "learning_rate": 7.636432039117761e-07, "loss": 0.1556, "step": 36420 }, { "epoch": 0.7143137254901961, "grad_norm": 3.4838669300079346, "learning_rate": 7.635054515457223e-07, "loss": 0.1663, "step": 36430 }, { "epoch": 0.7145098039215686, "grad_norm": 5.8800482749938965, "learning_rate": 7.633676714820905e-07, "loss": 0.1235, "step": 36440 }, { "epoch": 0.7147058823529412, "grad_norm": 2.628371477127075, "learning_rate": 7.632298637353632e-07, "loss": 0.1665, "step": 36450 }, { "epoch": 0.7149019607843137, "grad_norm": 1.1912614107131958, "learning_rate": 7.630920283200255e-07, "loss": 0.1269, "step": 36460 }, { "epoch": 0.7150980392156863, "grad_norm": 2.8335928916931152, "learning_rate": 7.62954165250566e-07, "loss": 0.1617, "step": 36470 }, { "epoch": 0.7152941176470589, "grad_norm": 1.980728268623352, "learning_rate": 7.62816274541475e-07, "loss": 0.1769, "step": 36480 }, { "epoch": 0.7154901960784313, "grad_norm": 2.0493335723876953, "learning_rate": 7.626783562072472e-07, "loss": 0.1313, "step": 36490 }, { "epoch": 0.7156862745098039, "grad_norm": 2.8749825954437256, "learning_rate": 7.62540410262379e-07, "loss": 0.1571, "step": 36500 }, { "epoch": 0.7158823529411765, "grad_norm": 4.213172912597656, "learning_rate": 7.624024367213704e-07, "loss": 0.146, "step": 36510 }, { "epoch": 0.716078431372549, "grad_norm": 3.177386999130249, "learning_rate": 7.622644355987238e-07, "loss": 0.1687, "step": 36520 }, { "epoch": 0.7162745098039216, "grad_norm": 1.9450578689575195, "learning_rate": 7.621264069089452e-07, "loss": 0.1263, "step": 36530 }, { "epoch": 0.7164705882352941, "grad_norm": 1.8273820877075195, "learning_rate": 7.619883506665427e-07, "loss": 0.1592, "step": 36540 }, { "epoch": 0.7166666666666667, "grad_norm": 2.316757917404175, "learning_rate": 7.618502668860279e-07, "loss": 0.155, "step": 36550 }, { "epoch": 0.7168627450980393, "grad_norm": 2.208860397338867, "learning_rate": 7.617121555819148e-07, "loss": 0.1623, "step": 36560 }, { "epoch": 0.7170588235294117, "grad_norm": 3.0965468883514404, "learning_rate": 7.615740167687207e-07, "loss": 0.171, "step": 36570 }, { "epoch": 0.7172549019607843, "grad_norm": 1.7401273250579834, "learning_rate": 7.61435850460966e-07, "loss": 0.1864, "step": 36580 }, { "epoch": 0.7174509803921568, "grad_norm": 4.824897289276123, "learning_rate": 7.612976566731728e-07, "loss": 0.1481, "step": 36590 }, { "epoch": 0.7176470588235294, "grad_norm": 4.291260242462158, "learning_rate": 7.611594354198676e-07, "loss": 0.1881, "step": 36600 }, { "epoch": 0.717843137254902, "grad_norm": 4.001865863800049, "learning_rate": 7.610211867155791e-07, "loss": 0.1435, "step": 36610 }, { "epoch": 0.7180392156862745, "grad_norm": 3.070060968399048, "learning_rate": 7.608829105748384e-07, "loss": 0.1568, "step": 36620 }, { "epoch": 0.7182352941176471, "grad_norm": 4.668734073638916, "learning_rate": 7.607446070121805e-07, "loss": 0.1569, "step": 36630 }, { "epoch": 0.7184313725490196, "grad_norm": 2.2339115142822266, "learning_rate": 7.606062760421424e-07, "loss": 0.1511, "step": 36640 }, { "epoch": 0.7186274509803922, "grad_norm": 2.612351655960083, "learning_rate": 7.604679176792646e-07, "loss": 0.1795, "step": 36650 }, { "epoch": 0.7188235294117648, "grad_norm": 3.104426622390747, "learning_rate": 7.603295319380902e-07, "loss": 0.1447, "step": 36660 }, { "epoch": 0.7190196078431372, "grad_norm": 2.0750279426574707, "learning_rate": 7.601911188331649e-07, "loss": 0.1416, "step": 36670 }, { "epoch": 0.7192156862745098, "grad_norm": 2.872664451599121, "learning_rate": 7.600526783790377e-07, "loss": 0.1729, "step": 36680 }, { "epoch": 0.7194117647058823, "grad_norm": 2.398430347442627, "learning_rate": 7.599142105902607e-07, "loss": 0.1909, "step": 36690 }, { "epoch": 0.7196078431372549, "grad_norm": 5.987502574920654, "learning_rate": 7.597757154813882e-07, "loss": 0.1461, "step": 36700 }, { "epoch": 0.7198039215686275, "grad_norm": 3.2790610790252686, "learning_rate": 7.596371930669776e-07, "loss": 0.1624, "step": 36710 }, { "epoch": 0.72, "grad_norm": 1.5857001543045044, "learning_rate": 7.594986433615895e-07, "loss": 0.1721, "step": 36720 }, { "epoch": 0.7201960784313726, "grad_norm": 1.5932706594467163, "learning_rate": 7.59360066379787e-07, "loss": 0.1637, "step": 36730 }, { "epoch": 0.720392156862745, "grad_norm": 3.026304244995117, "learning_rate": 7.592214621361364e-07, "loss": 0.1643, "step": 36740 }, { "epoch": 0.7205882352941176, "grad_norm": 7.80837345123291, "learning_rate": 7.590828306452063e-07, "loss": 0.1348, "step": 36750 }, { "epoch": 0.7207843137254902, "grad_norm": 4.032628536224365, "learning_rate": 7.589441719215688e-07, "loss": 0.1699, "step": 36760 }, { "epoch": 0.7209803921568627, "grad_norm": 3.42488956451416, "learning_rate": 7.588054859797985e-07, "loss": 0.1815, "step": 36770 }, { "epoch": 0.7211764705882353, "grad_norm": 1.8846123218536377, "learning_rate": 7.586667728344731e-07, "loss": 0.1732, "step": 36780 }, { "epoch": 0.7213725490196078, "grad_norm": 89.154541015625, "learning_rate": 7.585280325001728e-07, "loss": 0.1689, "step": 36790 }, { "epoch": 0.7215686274509804, "grad_norm": 2.013214588165283, "learning_rate": 7.583892649914809e-07, "loss": 0.1501, "step": 36800 }, { "epoch": 0.721764705882353, "grad_norm": 1.4177931547164917, "learning_rate": 7.582504703229837e-07, "loss": 0.1512, "step": 36810 }, { "epoch": 0.7219607843137255, "grad_norm": 2.240224838256836, "learning_rate": 7.581116485092701e-07, "loss": 0.1403, "step": 36820 }, { "epoch": 0.722156862745098, "grad_norm": 2.7391483783721924, "learning_rate": 7.579727995649319e-07, "loss": 0.1672, "step": 36830 }, { "epoch": 0.7223529411764706, "grad_norm": 1.7957608699798584, "learning_rate": 7.578339235045637e-07, "loss": 0.1332, "step": 36840 }, { "epoch": 0.7225490196078431, "grad_norm": 5.962609767913818, "learning_rate": 7.576950203427632e-07, "loss": 0.1505, "step": 36850 }, { "epoch": 0.7227450980392157, "grad_norm": 2.5474205017089844, "learning_rate": 7.575560900941307e-07, "loss": 0.1409, "step": 36860 }, { "epoch": 0.7229411764705882, "grad_norm": 1.8822672367095947, "learning_rate": 7.574171327732694e-07, "loss": 0.1778, "step": 36870 }, { "epoch": 0.7231372549019608, "grad_norm": 4.076420783996582, "learning_rate": 7.572781483947854e-07, "loss": 0.1189, "step": 36880 }, { "epoch": 0.7233333333333334, "grad_norm": 2.0226621627807617, "learning_rate": 7.571391369732878e-07, "loss": 0.1507, "step": 36890 }, { "epoch": 0.7235294117647059, "grad_norm": 2.9418601989746094, "learning_rate": 7.570000985233882e-07, "loss": 0.1517, "step": 36900 }, { "epoch": 0.7237254901960785, "grad_norm": 2.5268213748931885, "learning_rate": 7.568610330597012e-07, "loss": 0.1355, "step": 36910 }, { "epoch": 0.7239215686274509, "grad_norm": 2.48193621635437, "learning_rate": 7.567219405968441e-07, "loss": 0.1411, "step": 36920 }, { "epoch": 0.7241176470588235, "grad_norm": 1.9262467622756958, "learning_rate": 7.565828211494376e-07, "loss": 0.1585, "step": 36930 }, { "epoch": 0.7243137254901961, "grad_norm": 3.9470205307006836, "learning_rate": 7.564436747321043e-07, "loss": 0.1528, "step": 36940 }, { "epoch": 0.7245098039215686, "grad_norm": 4.075769901275635, "learning_rate": 7.563045013594707e-07, "loss": 0.1657, "step": 36950 }, { "epoch": 0.7247058823529412, "grad_norm": 3.179896116256714, "learning_rate": 7.561653010461652e-07, "loss": 0.1515, "step": 36960 }, { "epoch": 0.7249019607843137, "grad_norm": 2.4815666675567627, "learning_rate": 7.560260738068194e-07, "loss": 0.1577, "step": 36970 }, { "epoch": 0.7250980392156863, "grad_norm": 2.9903228282928467, "learning_rate": 7.55886819656068e-07, "loss": 0.1612, "step": 36980 }, { "epoch": 0.7252941176470589, "grad_norm": 3.357996702194214, "learning_rate": 7.557475386085482e-07, "loss": 0.168, "step": 36990 }, { "epoch": 0.7254901960784313, "grad_norm": 4.163931369781494, "learning_rate": 7.556082306788998e-07, "loss": 0.1708, "step": 37000 }, { "epoch": 0.7256862745098039, "grad_norm": 3.2822155952453613, "learning_rate": 7.554688958817663e-07, "loss": 0.1504, "step": 37010 }, { "epoch": 0.7258823529411764, "grad_norm": 1.6180088520050049, "learning_rate": 7.55329534231793e-07, "loss": 0.1386, "step": 37020 }, { "epoch": 0.726078431372549, "grad_norm": 2.1833293437957764, "learning_rate": 7.551901457436287e-07, "loss": 0.1422, "step": 37030 }, { "epoch": 0.7262745098039216, "grad_norm": 3.4997925758361816, "learning_rate": 7.550507304319247e-07, "loss": 0.1481, "step": 37040 }, { "epoch": 0.7264705882352941, "grad_norm": 1.7545380592346191, "learning_rate": 7.549112883113354e-07, "loss": 0.1438, "step": 37050 }, { "epoch": 0.7266666666666667, "grad_norm": 2.1164391040802, "learning_rate": 7.547718193965173e-07, "loss": 0.168, "step": 37060 }, { "epoch": 0.7268627450980392, "grad_norm": 2.7023322582244873, "learning_rate": 7.546323237021309e-07, "loss": 0.1533, "step": 37070 }, { "epoch": 0.7270588235294118, "grad_norm": 3.8405938148498535, "learning_rate": 7.544928012428385e-07, "loss": 0.1498, "step": 37080 }, { "epoch": 0.7272549019607844, "grad_norm": 2.4514288902282715, "learning_rate": 7.543532520333058e-07, "loss": 0.1504, "step": 37090 }, { "epoch": 0.7274509803921568, "grad_norm": 3.794536828994751, "learning_rate": 7.54213676088201e-07, "loss": 0.1886, "step": 37100 }, { "epoch": 0.7276470588235294, "grad_norm": 2.398531198501587, "learning_rate": 7.540740734221951e-07, "loss": 0.1373, "step": 37110 }, { "epoch": 0.7278431372549019, "grad_norm": 3.479365348815918, "learning_rate": 7.539344440499621e-07, "loss": 0.1452, "step": 37120 }, { "epoch": 0.7280392156862745, "grad_norm": 3.678950309753418, "learning_rate": 7.537947879861787e-07, "loss": 0.159, "step": 37130 }, { "epoch": 0.7282352941176471, "grad_norm": 2.9611008167266846, "learning_rate": 7.536551052455244e-07, "loss": 0.1608, "step": 37140 }, { "epoch": 0.7284313725490196, "grad_norm": 3.205310106277466, "learning_rate": 7.535153958426816e-07, "loss": 0.1679, "step": 37150 }, { "epoch": 0.7286274509803922, "grad_norm": 7.154580116271973, "learning_rate": 7.533756597923353e-07, "loss": 0.181, "step": 37160 }, { "epoch": 0.7288235294117648, "grad_norm": 3.5881192684173584, "learning_rate": 7.532358971091737e-07, "loss": 0.152, "step": 37170 }, { "epoch": 0.7290196078431372, "grad_norm": 3.0871427059173584, "learning_rate": 7.530961078078872e-07, "loss": 0.1313, "step": 37180 }, { "epoch": 0.7292156862745098, "grad_norm": 1.9696470499038696, "learning_rate": 7.529562919031696e-07, "loss": 0.1725, "step": 37190 }, { "epoch": 0.7294117647058823, "grad_norm": 4.368664264678955, "learning_rate": 7.528164494097171e-07, "loss": 0.145, "step": 37200 }, { "epoch": 0.7296078431372549, "grad_norm": 2.0408859252929688, "learning_rate": 7.526765803422288e-07, "loss": 0.1509, "step": 37210 }, { "epoch": 0.7298039215686275, "grad_norm": 2.5817103385925293, "learning_rate": 7.525366847154067e-07, "loss": 0.1538, "step": 37220 }, { "epoch": 0.73, "grad_norm": 2.6036324501037598, "learning_rate": 7.523967625439553e-07, "loss": 0.1527, "step": 37230 }, { "epoch": 0.7301960784313726, "grad_norm": 4.4170756340026855, "learning_rate": 7.522568138425823e-07, "loss": 0.1796, "step": 37240 }, { "epoch": 0.7303921568627451, "grad_norm": 2.4125759601593018, "learning_rate": 7.521168386259982e-07, "loss": 0.1281, "step": 37250 }, { "epoch": 0.7305882352941176, "grad_norm": 8.142964363098145, "learning_rate": 7.519768369089155e-07, "loss": 0.1982, "step": 37260 }, { "epoch": 0.7307843137254902, "grad_norm": 4.178279399871826, "learning_rate": 7.518368087060504e-07, "loss": 0.1333, "step": 37270 }, { "epoch": 0.7309803921568627, "grad_norm": 2.655980348587036, "learning_rate": 7.516967540321215e-07, "loss": 0.1496, "step": 37280 }, { "epoch": 0.7311764705882353, "grad_norm": 1.5649076700210571, "learning_rate": 7.515566729018504e-07, "loss": 0.1182, "step": 37290 }, { "epoch": 0.7313725490196078, "grad_norm": 2.849724769592285, "learning_rate": 7.514165653299609e-07, "loss": 0.1598, "step": 37300 }, { "epoch": 0.7315686274509804, "grad_norm": 1.2069460153579712, "learning_rate": 7.512764313311803e-07, "loss": 0.1369, "step": 37310 }, { "epoch": 0.731764705882353, "grad_norm": 3.326528787612915, "learning_rate": 7.511362709202383e-07, "loss": 0.2024, "step": 37320 }, { "epoch": 0.7319607843137255, "grad_norm": 4.563427925109863, "learning_rate": 7.509960841118672e-07, "loss": 0.1549, "step": 37330 }, { "epoch": 0.7321568627450981, "grad_norm": 1.9392099380493164, "learning_rate": 7.508558709208027e-07, "loss": 0.1681, "step": 37340 }, { "epoch": 0.7323529411764705, "grad_norm": 2.1846213340759277, "learning_rate": 7.507156313617826e-07, "loss": 0.1295, "step": 37350 }, { "epoch": 0.7325490196078431, "grad_norm": 2.3000118732452393, "learning_rate": 7.505753654495482e-07, "loss": 0.1403, "step": 37360 }, { "epoch": 0.7327450980392157, "grad_norm": 4.144371032714844, "learning_rate": 7.504350731988425e-07, "loss": 0.1607, "step": 37370 }, { "epoch": 0.7329411764705882, "grad_norm": 2.0664165019989014, "learning_rate": 7.502947546244121e-07, "loss": 0.1363, "step": 37380 }, { "epoch": 0.7331372549019608, "grad_norm": 3.7230634689331055, "learning_rate": 7.501544097410064e-07, "loss": 0.1655, "step": 37390 }, { "epoch": 0.7333333333333333, "grad_norm": 5.9188337326049805, "learning_rate": 7.500140385633772e-07, "loss": 0.1667, "step": 37400 }, { "epoch": 0.7335294117647059, "grad_norm": 2.674025058746338, "learning_rate": 7.49873641106279e-07, "loss": 0.1656, "step": 37410 }, { "epoch": 0.7337254901960785, "grad_norm": 1.8098375797271729, "learning_rate": 7.497332173844696e-07, "loss": 0.1704, "step": 37420 }, { "epoch": 0.733921568627451, "grad_norm": 2.3894400596618652, "learning_rate": 7.49592767412709e-07, "loss": 0.1479, "step": 37430 }, { "epoch": 0.7341176470588235, "grad_norm": 2.3278422355651855, "learning_rate": 7.494522912057601e-07, "loss": 0.146, "step": 37440 }, { "epoch": 0.734313725490196, "grad_norm": 3.4288125038146973, "learning_rate": 7.493117887783888e-07, "loss": 0.1609, "step": 37450 }, { "epoch": 0.7345098039215686, "grad_norm": 1.597348928451538, "learning_rate": 7.491712601453634e-07, "loss": 0.1413, "step": 37460 }, { "epoch": 0.7347058823529412, "grad_norm": 1.7936490774154663, "learning_rate": 7.490307053214556e-07, "loss": 0.1521, "step": 37470 }, { "epoch": 0.7349019607843137, "grad_norm": 4.171685218811035, "learning_rate": 7.488901243214388e-07, "loss": 0.1564, "step": 37480 }, { "epoch": 0.7350980392156863, "grad_norm": 2.5063798427581787, "learning_rate": 7.4874951716009e-07, "loss": 0.1643, "step": 37490 }, { "epoch": 0.7352941176470589, "grad_norm": 2.046250581741333, "learning_rate": 7.486088838521889e-07, "loss": 0.1671, "step": 37500 }, { "epoch": 0.7354901960784314, "grad_norm": 3.182373285293579, "learning_rate": 7.484682244125174e-07, "loss": 0.175, "step": 37510 }, { "epoch": 0.735686274509804, "grad_norm": 2.0326733589172363, "learning_rate": 7.483275388558607e-07, "loss": 0.1431, "step": 37520 }, { "epoch": 0.7358823529411764, "grad_norm": 2.197709321975708, "learning_rate": 7.481868271970064e-07, "loss": 0.1464, "step": 37530 }, { "epoch": 0.736078431372549, "grad_norm": 3.1647093296051025, "learning_rate": 7.48046089450745e-07, "loss": 0.1326, "step": 37540 }, { "epoch": 0.7362745098039216, "grad_norm": 5.906543731689453, "learning_rate": 7.4790532563187e-07, "loss": 0.179, "step": 37550 }, { "epoch": 0.7364705882352941, "grad_norm": 2.71759033203125, "learning_rate": 7.477645357551769e-07, "loss": 0.1685, "step": 37560 }, { "epoch": 0.7366666666666667, "grad_norm": 3.9959754943847656, "learning_rate": 7.476237198354647e-07, "loss": 0.1846, "step": 37570 }, { "epoch": 0.7368627450980392, "grad_norm": 3.3376712799072266, "learning_rate": 7.47482877887535e-07, "loss": 0.1576, "step": 37580 }, { "epoch": 0.7370588235294118, "grad_norm": 1.832293152809143, "learning_rate": 7.473420099261914e-07, "loss": 0.1842, "step": 37590 }, { "epoch": 0.7372549019607844, "grad_norm": 1.6963646411895752, "learning_rate": 7.472011159662414e-07, "loss": 0.1707, "step": 37600 }, { "epoch": 0.7374509803921568, "grad_norm": 1.7343865633010864, "learning_rate": 7.470601960224944e-07, "loss": 0.1699, "step": 37610 }, { "epoch": 0.7376470588235294, "grad_norm": 1.2734652757644653, "learning_rate": 7.469192501097626e-07, "loss": 0.1333, "step": 37620 }, { "epoch": 0.7378431372549019, "grad_norm": 2.2845911979675293, "learning_rate": 7.467782782428616e-07, "loss": 0.1418, "step": 37630 }, { "epoch": 0.7380392156862745, "grad_norm": 2.1191253662109375, "learning_rate": 7.466372804366086e-07, "loss": 0.1826, "step": 37640 }, { "epoch": 0.7382352941176471, "grad_norm": 1.6291826963424683, "learning_rate": 7.464962567058247e-07, "loss": 0.1497, "step": 37650 }, { "epoch": 0.7384313725490196, "grad_norm": 1.8584526777267456, "learning_rate": 7.463552070653329e-07, "loss": 0.1357, "step": 37660 }, { "epoch": 0.7386274509803922, "grad_norm": 2.377307415008545, "learning_rate": 7.462141315299593e-07, "loss": 0.1581, "step": 37670 }, { "epoch": 0.7388235294117647, "grad_norm": 4.653329372406006, "learning_rate": 7.460730301145323e-07, "loss": 0.1628, "step": 37680 }, { "epoch": 0.7390196078431372, "grad_norm": 2.1056087017059326, "learning_rate": 7.45931902833884e-07, "loss": 0.1401, "step": 37690 }, { "epoch": 0.7392156862745098, "grad_norm": 3.7689616680145264, "learning_rate": 7.45790749702848e-07, "loss": 0.1666, "step": 37700 }, { "epoch": 0.7394117647058823, "grad_norm": 3.287217140197754, "learning_rate": 7.456495707362615e-07, "loss": 0.153, "step": 37710 }, { "epoch": 0.7396078431372549, "grad_norm": 1.6175228357315063, "learning_rate": 7.455083659489639e-07, "loss": 0.1525, "step": 37720 }, { "epoch": 0.7398039215686274, "grad_norm": 6.065554618835449, "learning_rate": 7.453671353557977e-07, "loss": 0.179, "step": 37730 }, { "epoch": 0.74, "grad_norm": 1.9635440111160278, "learning_rate": 7.452258789716077e-07, "loss": 0.1404, "step": 37740 }, { "epoch": 0.7401960784313726, "grad_norm": 2.4489071369171143, "learning_rate": 7.450845968112419e-07, "loss": 0.1485, "step": 37750 }, { "epoch": 0.7403921568627451, "grad_norm": 2.7730209827423096, "learning_rate": 7.449432888895505e-07, "loss": 0.154, "step": 37760 }, { "epoch": 0.7405882352941177, "grad_norm": 1.6925272941589355, "learning_rate": 7.448019552213869e-07, "loss": 0.1561, "step": 37770 }, { "epoch": 0.7407843137254903, "grad_norm": 1.4935559034347534, "learning_rate": 7.446605958216065e-07, "loss": 0.1563, "step": 37780 }, { "epoch": 0.7409803921568627, "grad_norm": 2.19084095954895, "learning_rate": 7.445192107050686e-07, "loss": 0.153, "step": 37790 }, { "epoch": 0.7411764705882353, "grad_norm": 5.241160869598389, "learning_rate": 7.443777998866338e-07, "loss": 0.1361, "step": 37800 }, { "epoch": 0.7413725490196078, "grad_norm": 2.2840499877929688, "learning_rate": 7.442363633811665e-07, "loss": 0.1565, "step": 37810 }, { "epoch": 0.7415686274509804, "grad_norm": 2.523529291152954, "learning_rate": 7.440949012035332e-07, "loss": 0.1417, "step": 37820 }, { "epoch": 0.741764705882353, "grad_norm": 2.7508318424224854, "learning_rate": 7.439534133686032e-07, "loss": 0.1761, "step": 37830 }, { "epoch": 0.7419607843137255, "grad_norm": 4.833118438720703, "learning_rate": 7.438118998912486e-07, "loss": 0.1742, "step": 37840 }, { "epoch": 0.7421568627450981, "grad_norm": 2.650923013687134, "learning_rate": 7.436703607863442e-07, "loss": 0.1648, "step": 37850 }, { "epoch": 0.7423529411764705, "grad_norm": 1.9997940063476562, "learning_rate": 7.435287960687675e-07, "loss": 0.1394, "step": 37860 }, { "epoch": 0.7425490196078431, "grad_norm": 2.821626663208008, "learning_rate": 7.433872057533987e-07, "loss": 0.1553, "step": 37870 }, { "epoch": 0.7427450980392157, "grad_norm": 2.2841575145721436, "learning_rate": 7.432455898551203e-07, "loss": 0.1643, "step": 37880 }, { "epoch": 0.7429411764705882, "grad_norm": 3.4307005405426025, "learning_rate": 7.431039483888181e-07, "loss": 0.1612, "step": 37890 }, { "epoch": 0.7431372549019608, "grad_norm": 1.2149813175201416, "learning_rate": 7.429622813693805e-07, "loss": 0.1468, "step": 37900 }, { "epoch": 0.7433333333333333, "grad_norm": 2.6344780921936035, "learning_rate": 7.42820588811698e-07, "loss": 0.1404, "step": 37910 }, { "epoch": 0.7435294117647059, "grad_norm": 3.7833175659179688, "learning_rate": 7.426788707306644e-07, "loss": 0.1514, "step": 37920 }, { "epoch": 0.7437254901960785, "grad_norm": 1.6544266939163208, "learning_rate": 7.425371271411759e-07, "loss": 0.1493, "step": 37930 }, { "epoch": 0.743921568627451, "grad_norm": 3.602144479751587, "learning_rate": 7.423953580581314e-07, "loss": 0.1565, "step": 37940 }, { "epoch": 0.7441176470588236, "grad_norm": 1.7342530488967896, "learning_rate": 7.422535634964327e-07, "loss": 0.162, "step": 37950 }, { "epoch": 0.744313725490196, "grad_norm": 2.0733776092529297, "learning_rate": 7.421117434709841e-07, "loss": 0.1249, "step": 37960 }, { "epoch": 0.7445098039215686, "grad_norm": 1.6613866090774536, "learning_rate": 7.419698979966923e-07, "loss": 0.1207, "step": 37970 }, { "epoch": 0.7447058823529412, "grad_norm": 1.7489408254623413, "learning_rate": 7.418280270884673e-07, "loss": 0.1682, "step": 37980 }, { "epoch": 0.7449019607843137, "grad_norm": 2.870532274246216, "learning_rate": 7.416861307612212e-07, "loss": 0.1407, "step": 37990 }, { "epoch": 0.7450980392156863, "grad_norm": 2.8200337886810303, "learning_rate": 7.41544209029869e-07, "loss": 0.1992, "step": 38000 }, { "epoch": 0.7452941176470588, "grad_norm": 2.596118927001953, "learning_rate": 7.414022619093288e-07, "loss": 0.1423, "step": 38010 }, { "epoch": 0.7454901960784314, "grad_norm": 2.7103655338287354, "learning_rate": 7.412602894145202e-07, "loss": 0.1702, "step": 38020 }, { "epoch": 0.745686274509804, "grad_norm": 4.271402835845947, "learning_rate": 7.411182915603668e-07, "loss": 0.1564, "step": 38030 }, { "epoch": 0.7458823529411764, "grad_norm": 3.006077527999878, "learning_rate": 7.409762683617941e-07, "loss": 0.1708, "step": 38040 }, { "epoch": 0.746078431372549, "grad_norm": 2.6304028034210205, "learning_rate": 7.408342198337303e-07, "loss": 0.173, "step": 38050 }, { "epoch": 0.7462745098039215, "grad_norm": 2.249849319458008, "learning_rate": 7.406921459911066e-07, "loss": 0.1363, "step": 38060 }, { "epoch": 0.7464705882352941, "grad_norm": 2.59271502494812, "learning_rate": 7.405500468488566e-07, "loss": 0.1475, "step": 38070 }, { "epoch": 0.7466666666666667, "grad_norm": 2.5706069469451904, "learning_rate": 7.404079224219165e-07, "loss": 0.1868, "step": 38080 }, { "epoch": 0.7468627450980392, "grad_norm": 1.5173587799072266, "learning_rate": 7.402657727252255e-07, "loss": 0.1448, "step": 38090 }, { "epoch": 0.7470588235294118, "grad_norm": 3.3396008014678955, "learning_rate": 7.401235977737252e-07, "loss": 0.1449, "step": 38100 }, { "epoch": 0.7472549019607844, "grad_norm": 2.970785617828369, "learning_rate": 7.399813975823596e-07, "loss": 0.14, "step": 38110 }, { "epoch": 0.7474509803921568, "grad_norm": 1.5309423208236694, "learning_rate": 7.39839172166076e-07, "loss": 0.1701, "step": 38120 }, { "epoch": 0.7476470588235294, "grad_norm": 1.8615208864212036, "learning_rate": 7.396969215398238e-07, "loss": 0.1737, "step": 38130 }, { "epoch": 0.7478431372549019, "grad_norm": 3.7810475826263428, "learning_rate": 7.395546457185552e-07, "loss": 0.1581, "step": 38140 }, { "epoch": 0.7480392156862745, "grad_norm": 2.899784803390503, "learning_rate": 7.394123447172252e-07, "loss": 0.1396, "step": 38150 }, { "epoch": 0.7482352941176471, "grad_norm": 2.1689445972442627, "learning_rate": 7.392700185507914e-07, "loss": 0.1666, "step": 38160 }, { "epoch": 0.7484313725490196, "grad_norm": 2.4229934215545654, "learning_rate": 7.39127667234214e-07, "loss": 0.1642, "step": 38170 }, { "epoch": 0.7486274509803922, "grad_norm": 2.3840630054473877, "learning_rate": 7.389852907824555e-07, "loss": 0.1358, "step": 38180 }, { "epoch": 0.7488235294117647, "grad_norm": 2.3668458461761475, "learning_rate": 7.388428892104816e-07, "loss": 0.1592, "step": 38190 }, { "epoch": 0.7490196078431373, "grad_norm": 3.26999831199646, "learning_rate": 7.387004625332607e-07, "loss": 0.1348, "step": 38200 }, { "epoch": 0.7492156862745099, "grad_norm": 4.912667274475098, "learning_rate": 7.38558010765763e-07, "loss": 0.1521, "step": 38210 }, { "epoch": 0.7494117647058823, "grad_norm": 1.5833795070648193, "learning_rate": 7.384155339229623e-07, "loss": 0.1653, "step": 38220 }, { "epoch": 0.7496078431372549, "grad_norm": 2.303640604019165, "learning_rate": 7.382730320198343e-07, "loss": 0.1584, "step": 38230 }, { "epoch": 0.7498039215686274, "grad_norm": 3.6752424240112305, "learning_rate": 7.38130505071358e-07, "loss": 0.1398, "step": 38240 }, { "epoch": 0.75, "grad_norm": 3.577833890914917, "learning_rate": 7.379879530925145e-07, "loss": 0.181, "step": 38250 }, { "epoch": 0.7501960784313726, "grad_norm": 3.5843045711517334, "learning_rate": 7.378453760982877e-07, "loss": 0.1707, "step": 38260 }, { "epoch": 0.7503921568627451, "grad_norm": 1.7997618913650513, "learning_rate": 7.377027741036641e-07, "loss": 0.1392, "step": 38270 }, { "epoch": 0.7505882352941177, "grad_norm": 2.82871413230896, "learning_rate": 7.37560147123633e-07, "loss": 0.1632, "step": 38280 }, { "epoch": 0.7507843137254901, "grad_norm": 2.402984380722046, "learning_rate": 7.374174951731862e-07, "loss": 0.1614, "step": 38290 }, { "epoch": 0.7509803921568627, "grad_norm": 2.203172206878662, "learning_rate": 7.37274818267318e-07, "loss": 0.1785, "step": 38300 }, { "epoch": 0.7511764705882353, "grad_norm": 2.937589168548584, "learning_rate": 7.371321164210257e-07, "loss": 0.1413, "step": 38310 }, { "epoch": 0.7513725490196078, "grad_norm": 2.528186082839966, "learning_rate": 7.369893896493087e-07, "loss": 0.1591, "step": 38320 }, { "epoch": 0.7515686274509804, "grad_norm": 1.3831477165222168, "learning_rate": 7.368466379671695e-07, "loss": 0.1615, "step": 38330 }, { "epoch": 0.7517647058823529, "grad_norm": 2.4657649993896484, "learning_rate": 7.367038613896128e-07, "loss": 0.1535, "step": 38340 }, { "epoch": 0.7519607843137255, "grad_norm": 1.9454874992370605, "learning_rate": 7.365610599316463e-07, "loss": 0.149, "step": 38350 }, { "epoch": 0.7521568627450981, "grad_norm": 2.008502244949341, "learning_rate": 7.364182336082801e-07, "loss": 0.1757, "step": 38360 }, { "epoch": 0.7523529411764706, "grad_norm": 3.949296236038208, "learning_rate": 7.362753824345271e-07, "loss": 0.1458, "step": 38370 }, { "epoch": 0.7525490196078432, "grad_norm": 3.1732583045959473, "learning_rate": 7.361325064254023e-07, "loss": 0.1563, "step": 38380 }, { "epoch": 0.7527450980392156, "grad_norm": 2.214287042617798, "learning_rate": 7.359896055959239e-07, "loss": 0.1241, "step": 38390 }, { "epoch": 0.7529411764705882, "grad_norm": 2.233293294906616, "learning_rate": 7.358466799611124e-07, "loss": 0.1321, "step": 38400 }, { "epoch": 0.7531372549019608, "grad_norm": 1.632576584815979, "learning_rate": 7.357037295359913e-07, "loss": 0.1526, "step": 38410 }, { "epoch": 0.7533333333333333, "grad_norm": 2.552367925643921, "learning_rate": 7.355607543355861e-07, "loss": 0.1293, "step": 38420 }, { "epoch": 0.7535294117647059, "grad_norm": 2.2424213886260986, "learning_rate": 7.354177543749253e-07, "loss": 0.1467, "step": 38430 }, { "epoch": 0.7537254901960785, "grad_norm": 2.7653868198394775, "learning_rate": 7.352747296690399e-07, "loss": 0.1625, "step": 38440 }, { "epoch": 0.753921568627451, "grad_norm": 2.939702272415161, "learning_rate": 7.351316802329635e-07, "loss": 0.178, "step": 38450 }, { "epoch": 0.7541176470588236, "grad_norm": 2.261078357696533, "learning_rate": 7.349886060817322e-07, "loss": 0.1418, "step": 38460 }, { "epoch": 0.754313725490196, "grad_norm": 1.659242868423462, "learning_rate": 7.348455072303851e-07, "loss": 0.1389, "step": 38470 }, { "epoch": 0.7545098039215686, "grad_norm": 1.8299736976623535, "learning_rate": 7.347023836939632e-07, "loss": 0.1769, "step": 38480 }, { "epoch": 0.7547058823529412, "grad_norm": 1.9010740518569946, "learning_rate": 7.345592354875108e-07, "loss": 0.1742, "step": 38490 }, { "epoch": 0.7549019607843137, "grad_norm": 6.358440399169922, "learning_rate": 7.344160626260745e-07, "loss": 0.1769, "step": 38500 }, { "epoch": 0.7550980392156863, "grad_norm": 2.4224538803100586, "learning_rate": 7.342728651247032e-07, "loss": 0.1306, "step": 38510 }, { "epoch": 0.7552941176470588, "grad_norm": 2.5001795291900635, "learning_rate": 7.341296429984491e-07, "loss": 0.154, "step": 38520 }, { "epoch": 0.7554901960784314, "grad_norm": 1.1046100854873657, "learning_rate": 7.339863962623661e-07, "loss": 0.14, "step": 38530 }, { "epoch": 0.755686274509804, "grad_norm": 1.450455904006958, "learning_rate": 7.338431249315115e-07, "loss": 0.1607, "step": 38540 }, { "epoch": 0.7558823529411764, "grad_norm": 2.9707372188568115, "learning_rate": 7.336998290209448e-07, "loss": 0.1401, "step": 38550 }, { "epoch": 0.756078431372549, "grad_norm": 3.4741737842559814, "learning_rate": 7.335565085457278e-07, "loss": 0.1499, "step": 38560 }, { "epoch": 0.7562745098039215, "grad_norm": 2.7250189781188965, "learning_rate": 7.334131635209255e-07, "loss": 0.1748, "step": 38570 }, { "epoch": 0.7564705882352941, "grad_norm": 1.5285372734069824, "learning_rate": 7.33269793961605e-07, "loss": 0.1469, "step": 38580 }, { "epoch": 0.7566666666666667, "grad_norm": 2.7478113174438477, "learning_rate": 7.331263998828364e-07, "loss": 0.1319, "step": 38590 }, { "epoch": 0.7568627450980392, "grad_norm": 2.194951057434082, "learning_rate": 7.329829812996919e-07, "loss": 0.1186, "step": 38600 }, { "epoch": 0.7570588235294118, "grad_norm": 2.819566488265991, "learning_rate": 7.328395382272465e-07, "loss": 0.169, "step": 38610 }, { "epoch": 0.7572549019607843, "grad_norm": 3.243062973022461, "learning_rate": 7.326960706805781e-07, "loss": 0.1705, "step": 38620 }, { "epoch": 0.7574509803921569, "grad_norm": 2.4014618396759033, "learning_rate": 7.325525786747666e-07, "loss": 0.1296, "step": 38630 }, { "epoch": 0.7576470588235295, "grad_norm": 4.436540126800537, "learning_rate": 7.324090622248947e-07, "loss": 0.1613, "step": 38640 }, { "epoch": 0.7578431372549019, "grad_norm": 2.238060712814331, "learning_rate": 7.322655213460476e-07, "loss": 0.1673, "step": 38650 }, { "epoch": 0.7580392156862745, "grad_norm": 4.892023086547852, "learning_rate": 7.321219560533136e-07, "loss": 0.1659, "step": 38660 }, { "epoch": 0.758235294117647, "grad_norm": 4.682913780212402, "learning_rate": 7.319783663617825e-07, "loss": 0.149, "step": 38670 }, { "epoch": 0.7584313725490196, "grad_norm": 1.6482113599777222, "learning_rate": 7.318347522865481e-07, "loss": 0.1556, "step": 38680 }, { "epoch": 0.7586274509803922, "grad_norm": 2.4957454204559326, "learning_rate": 7.316911138427052e-07, "loss": 0.1666, "step": 38690 }, { "epoch": 0.7588235294117647, "grad_norm": 2.3638808727264404, "learning_rate": 7.315474510453523e-07, "loss": 0.1562, "step": 38700 }, { "epoch": 0.7590196078431373, "grad_norm": 2.3355178833007812, "learning_rate": 7.3140376390959e-07, "loss": 0.1401, "step": 38710 }, { "epoch": 0.7592156862745097, "grad_norm": 2.540174961090088, "learning_rate": 7.312600524505216e-07, "loss": 0.1716, "step": 38720 }, { "epoch": 0.7594117647058823, "grad_norm": 1.5758079290390015, "learning_rate": 7.311163166832527e-07, "loss": 0.148, "step": 38730 }, { "epoch": 0.7596078431372549, "grad_norm": 2.7389330863952637, "learning_rate": 7.309725566228922e-07, "loss": 0.2103, "step": 38740 }, { "epoch": 0.7598039215686274, "grad_norm": 4.510904312133789, "learning_rate": 7.308287722845502e-07, "loss": 0.1608, "step": 38750 }, { "epoch": 0.76, "grad_norm": 2.322659730911255, "learning_rate": 7.306849636833407e-07, "loss": 0.1327, "step": 38760 }, { "epoch": 0.7601960784313726, "grad_norm": 2.945409059524536, "learning_rate": 7.305411308343794e-07, "loss": 0.1485, "step": 38770 }, { "epoch": 0.7603921568627451, "grad_norm": 2.4334073066711426, "learning_rate": 7.303972737527852e-07, "loss": 0.1581, "step": 38780 }, { "epoch": 0.7605882352941177, "grad_norm": 3.017390012741089, "learning_rate": 7.302533924536789e-07, "loss": 0.1876, "step": 38790 }, { "epoch": 0.7607843137254902, "grad_norm": 1.79999577999115, "learning_rate": 7.301094869521844e-07, "loss": 0.1788, "step": 38800 }, { "epoch": 0.7609803921568628, "grad_norm": 2.886770009994507, "learning_rate": 7.299655572634276e-07, "loss": 0.118, "step": 38810 }, { "epoch": 0.7611764705882353, "grad_norm": 3.409411907196045, "learning_rate": 7.298216034025375e-07, "loss": 0.1511, "step": 38820 }, { "epoch": 0.7613725490196078, "grad_norm": 3.2076175212860107, "learning_rate": 7.296776253846452e-07, "loss": 0.1315, "step": 38830 }, { "epoch": 0.7615686274509804, "grad_norm": 2.279311418533325, "learning_rate": 7.295336232248846e-07, "loss": 0.149, "step": 38840 }, { "epoch": 0.7617647058823529, "grad_norm": 3.557779312133789, "learning_rate": 7.293895969383918e-07, "loss": 0.1965, "step": 38850 }, { "epoch": 0.7619607843137255, "grad_norm": 4.173307418823242, "learning_rate": 7.292455465403062e-07, "loss": 0.1802, "step": 38860 }, { "epoch": 0.7621568627450981, "grad_norm": 3.619450569152832, "learning_rate": 7.29101472045769e-07, "loss": 0.1441, "step": 38870 }, { "epoch": 0.7623529411764706, "grad_norm": 2.7458975315093994, "learning_rate": 7.289573734699239e-07, "loss": 0.1778, "step": 38880 }, { "epoch": 0.7625490196078432, "grad_norm": 2.6739964485168457, "learning_rate": 7.288132508279175e-07, "loss": 0.1889, "step": 38890 }, { "epoch": 0.7627450980392156, "grad_norm": 2.4037492275238037, "learning_rate": 7.286691041348992e-07, "loss": 0.1286, "step": 38900 }, { "epoch": 0.7629411764705882, "grad_norm": 3.206428050994873, "learning_rate": 7.2852493340602e-07, "loss": 0.1648, "step": 38910 }, { "epoch": 0.7631372549019608, "grad_norm": 2.2229769229888916, "learning_rate": 7.283807386564342e-07, "loss": 0.1522, "step": 38920 }, { "epoch": 0.7633333333333333, "grad_norm": 2.3268566131591797, "learning_rate": 7.282365199012985e-07, "loss": 0.1622, "step": 38930 }, { "epoch": 0.7635294117647059, "grad_norm": 4.255580425262451, "learning_rate": 7.280922771557718e-07, "loss": 0.149, "step": 38940 }, { "epoch": 0.7637254901960784, "grad_norm": 3.9126477241516113, "learning_rate": 7.279480104350159e-07, "loss": 0.1642, "step": 38950 }, { "epoch": 0.763921568627451, "grad_norm": 3.8356640338897705, "learning_rate": 7.278037197541949e-07, "loss": 0.1566, "step": 38960 }, { "epoch": 0.7641176470588236, "grad_norm": 3.0861732959747314, "learning_rate": 7.276594051284756e-07, "loss": 0.1466, "step": 38970 }, { "epoch": 0.764313725490196, "grad_norm": 2.421844244003296, "learning_rate": 7.275150665730269e-07, "loss": 0.119, "step": 38980 }, { "epoch": 0.7645098039215686, "grad_norm": 2.936640501022339, "learning_rate": 7.27370704103021e-07, "loss": 0.1155, "step": 38990 }, { "epoch": 0.7647058823529411, "grad_norm": 2.3747360706329346, "learning_rate": 7.272263177336315e-07, "loss": 0.161, "step": 39000 }, { "epoch": 0.7649019607843137, "grad_norm": 3.299443006515503, "learning_rate": 7.270819074800356e-07, "loss": 0.1417, "step": 39010 }, { "epoch": 0.7650980392156863, "grad_norm": 2.9097983837127686, "learning_rate": 7.269374733574124e-07, "loss": 0.1566, "step": 39020 }, { "epoch": 0.7652941176470588, "grad_norm": 4.9433274269104, "learning_rate": 7.267930153809436e-07, "loss": 0.1576, "step": 39030 }, { "epoch": 0.7654901960784314, "grad_norm": 2.884984254837036, "learning_rate": 7.266485335658136e-07, "loss": 0.1594, "step": 39040 }, { "epoch": 0.765686274509804, "grad_norm": 4.500613212585449, "learning_rate": 7.26504027927209e-07, "loss": 0.1924, "step": 39050 }, { "epoch": 0.7658823529411765, "grad_norm": 4.28322172164917, "learning_rate": 7.263594984803193e-07, "loss": 0.1872, "step": 39060 }, { "epoch": 0.766078431372549, "grad_norm": 3.7841153144836426, "learning_rate": 7.262149452403361e-07, "loss": 0.1481, "step": 39070 }, { "epoch": 0.7662745098039215, "grad_norm": 3.1451759338378906, "learning_rate": 7.260703682224536e-07, "loss": 0.1486, "step": 39080 }, { "epoch": 0.7664705882352941, "grad_norm": 1.3686063289642334, "learning_rate": 7.259257674418688e-07, "loss": 0.1672, "step": 39090 }, { "epoch": 0.7666666666666667, "grad_norm": 2.8293328285217285, "learning_rate": 7.257811429137809e-07, "loss": 0.1719, "step": 39100 }, { "epoch": 0.7668627450980392, "grad_norm": 1.384963870048523, "learning_rate": 7.256364946533915e-07, "loss": 0.1353, "step": 39110 }, { "epoch": 0.7670588235294118, "grad_norm": 1.6873880624771118, "learning_rate": 7.254918226759051e-07, "loss": 0.1471, "step": 39120 }, { "epoch": 0.7672549019607843, "grad_norm": 1.2900257110595703, "learning_rate": 7.253471269965284e-07, "loss": 0.1277, "step": 39130 }, { "epoch": 0.7674509803921569, "grad_norm": 2.349062204360962, "learning_rate": 7.252024076304705e-07, "loss": 0.1567, "step": 39140 }, { "epoch": 0.7676470588235295, "grad_norm": 4.169671535491943, "learning_rate": 7.250576645929433e-07, "loss": 0.1431, "step": 39150 }, { "epoch": 0.7678431372549019, "grad_norm": 2.3328678607940674, "learning_rate": 7.249128978991609e-07, "loss": 0.1457, "step": 39160 }, { "epoch": 0.7680392156862745, "grad_norm": 2.6637444496154785, "learning_rate": 7.247681075643401e-07, "loss": 0.1447, "step": 39170 }, { "epoch": 0.768235294117647, "grad_norm": 3.1460914611816406, "learning_rate": 7.246232936037001e-07, "loss": 0.1513, "step": 39180 }, { "epoch": 0.7684313725490196, "grad_norm": 1.836303949356079, "learning_rate": 7.244784560324626e-07, "loss": 0.1605, "step": 39190 }, { "epoch": 0.7686274509803922, "grad_norm": 2.949639320373535, "learning_rate": 7.243335948658518e-07, "loss": 0.1519, "step": 39200 }, { "epoch": 0.7688235294117647, "grad_norm": 3.212808609008789, "learning_rate": 7.24188710119094e-07, "loss": 0.1373, "step": 39210 }, { "epoch": 0.7690196078431373, "grad_norm": 2.7821319103240967, "learning_rate": 7.240438018074187e-07, "loss": 0.1584, "step": 39220 }, { "epoch": 0.7692156862745098, "grad_norm": 2.624630928039551, "learning_rate": 7.238988699460574e-07, "loss": 0.1356, "step": 39230 }, { "epoch": 0.7694117647058824, "grad_norm": 2.634474277496338, "learning_rate": 7.237539145502442e-07, "loss": 0.1736, "step": 39240 }, { "epoch": 0.7696078431372549, "grad_norm": 3.4337692260742188, "learning_rate": 7.236089356352158e-07, "loss": 0.2074, "step": 39250 }, { "epoch": 0.7698039215686274, "grad_norm": 3.619054079055786, "learning_rate": 7.234639332162107e-07, "loss": 0.132, "step": 39260 }, { "epoch": 0.77, "grad_norm": 2.878370523452759, "learning_rate": 7.233189073084708e-07, "loss": 0.1615, "step": 39270 }, { "epoch": 0.7701960784313725, "grad_norm": 2.074929714202881, "learning_rate": 7.2317385792724e-07, "loss": 0.1613, "step": 39280 }, { "epoch": 0.7703921568627451, "grad_norm": 3.9711883068084717, "learning_rate": 7.230287850877648e-07, "loss": 0.1535, "step": 39290 }, { "epoch": 0.7705882352941177, "grad_norm": 2.239661931991577, "learning_rate": 7.228836888052938e-07, "loss": 0.1653, "step": 39300 }, { "epoch": 0.7707843137254902, "grad_norm": 1.7115527391433716, "learning_rate": 7.227385690950787e-07, "loss": 0.1425, "step": 39310 }, { "epoch": 0.7709803921568628, "grad_norm": 3.4963126182556152, "learning_rate": 7.22593425972373e-07, "loss": 0.1662, "step": 39320 }, { "epoch": 0.7711764705882352, "grad_norm": 1.653059482574463, "learning_rate": 7.224482594524333e-07, "loss": 0.1464, "step": 39330 }, { "epoch": 0.7713725490196078, "grad_norm": 1.636924386024475, "learning_rate": 7.223030695505182e-07, "loss": 0.1355, "step": 39340 }, { "epoch": 0.7715686274509804, "grad_norm": 2.5925209522247314, "learning_rate": 7.221578562818888e-07, "loss": 0.1715, "step": 39350 }, { "epoch": 0.7717647058823529, "grad_norm": 2.464535713195801, "learning_rate": 7.22012619661809e-07, "loss": 0.158, "step": 39360 }, { "epoch": 0.7719607843137255, "grad_norm": 3.8764569759368896, "learning_rate": 7.218673597055447e-07, "loss": 0.1685, "step": 39370 }, { "epoch": 0.7721568627450981, "grad_norm": 2.748291015625, "learning_rate": 7.217220764283644e-07, "loss": 0.1509, "step": 39380 }, { "epoch": 0.7723529411764706, "grad_norm": 2.138192653656006, "learning_rate": 7.215767698455393e-07, "loss": 0.1579, "step": 39390 }, { "epoch": 0.7725490196078432, "grad_norm": 2.8493430614471436, "learning_rate": 7.214314399723427e-07, "loss": 0.1632, "step": 39400 }, { "epoch": 0.7727450980392156, "grad_norm": 3.400326728820801, "learning_rate": 7.212860868240506e-07, "loss": 0.1431, "step": 39410 }, { "epoch": 0.7729411764705882, "grad_norm": 3.375680446624756, "learning_rate": 7.211407104159415e-07, "loss": 0.1693, "step": 39420 }, { "epoch": 0.7731372549019608, "grad_norm": 4.255329132080078, "learning_rate": 7.209953107632959e-07, "loss": 0.1648, "step": 39430 }, { "epoch": 0.7733333333333333, "grad_norm": 2.8137283325195312, "learning_rate": 7.208498878813974e-07, "loss": 0.1588, "step": 39440 }, { "epoch": 0.7735294117647059, "grad_norm": 2.575673818588257, "learning_rate": 7.207044417855312e-07, "loss": 0.1616, "step": 39450 }, { "epoch": 0.7737254901960784, "grad_norm": 2.1209232807159424, "learning_rate": 7.205589724909858e-07, "loss": 0.1585, "step": 39460 }, { "epoch": 0.773921568627451, "grad_norm": 2.1180644035339355, "learning_rate": 7.204134800130518e-07, "loss": 0.1333, "step": 39470 }, { "epoch": 0.7741176470588236, "grad_norm": 3.59352970123291, "learning_rate": 7.202679643670218e-07, "loss": 0.1411, "step": 39480 }, { "epoch": 0.7743137254901961, "grad_norm": 1.9136241674423218, "learning_rate": 7.201224255681917e-07, "loss": 0.1581, "step": 39490 }, { "epoch": 0.7745098039215687, "grad_norm": 2.1508166790008545, "learning_rate": 7.199768636318592e-07, "loss": 0.1639, "step": 39500 }, { "epoch": 0.7747058823529411, "grad_norm": 3.8821558952331543, "learning_rate": 7.198312785733245e-07, "loss": 0.1566, "step": 39510 }, { "epoch": 0.7749019607843137, "grad_norm": 4.270737648010254, "learning_rate": 7.196856704078905e-07, "loss": 0.1689, "step": 39520 }, { "epoch": 0.7750980392156863, "grad_norm": 2.289494276046753, "learning_rate": 7.195400391508621e-07, "loss": 0.1542, "step": 39530 }, { "epoch": 0.7752941176470588, "grad_norm": 2.087019205093384, "learning_rate": 7.193943848175472e-07, "loss": 0.1705, "step": 39540 }, { "epoch": 0.7754901960784314, "grad_norm": 4.739058017730713, "learning_rate": 7.192487074232558e-07, "loss": 0.1807, "step": 39550 }, { "epoch": 0.7756862745098039, "grad_norm": 2.880617618560791, "learning_rate": 7.191030069833e-07, "loss": 0.1319, "step": 39560 }, { "epoch": 0.7758823529411765, "grad_norm": 4.893503665924072, "learning_rate": 7.18957283512995e-07, "loss": 0.1715, "step": 39570 }, { "epoch": 0.7760784313725491, "grad_norm": 1.5136566162109375, "learning_rate": 7.188115370276579e-07, "loss": 0.1212, "step": 39580 }, { "epoch": 0.7762745098039215, "grad_norm": 2.4648258686065674, "learning_rate": 7.186657675426085e-07, "loss": 0.1726, "step": 39590 }, { "epoch": 0.7764705882352941, "grad_norm": 3.7664921283721924, "learning_rate": 7.185199750731688e-07, "loss": 0.1961, "step": 39600 }, { "epoch": 0.7766666666666666, "grad_norm": 5.630705833435059, "learning_rate": 7.183741596346635e-07, "loss": 0.1724, "step": 39610 }, { "epoch": 0.7768627450980392, "grad_norm": 3.3438258171081543, "learning_rate": 7.182283212424193e-07, "loss": 0.1486, "step": 39620 }, { "epoch": 0.7770588235294118, "grad_norm": 2.803088665008545, "learning_rate": 7.18082459911766e-07, "loss": 0.1618, "step": 39630 }, { "epoch": 0.7772549019607843, "grad_norm": 1.884819507598877, "learning_rate": 7.179365756580349e-07, "loss": 0.1365, "step": 39640 }, { "epoch": 0.7774509803921569, "grad_norm": 4.263738632202148, "learning_rate": 7.177906684965603e-07, "loss": 0.1677, "step": 39650 }, { "epoch": 0.7776470588235294, "grad_norm": 3.319974184036255, "learning_rate": 7.17644738442679e-07, "loss": 0.1606, "step": 39660 }, { "epoch": 0.777843137254902, "grad_norm": 2.7032954692840576, "learning_rate": 7.174987855117298e-07, "loss": 0.1417, "step": 39670 }, { "epoch": 0.7780392156862745, "grad_norm": 3.5795295238494873, "learning_rate": 7.173528097190543e-07, "loss": 0.175, "step": 39680 }, { "epoch": 0.778235294117647, "grad_norm": 2.0836970806121826, "learning_rate": 7.17206811079996e-07, "loss": 0.148, "step": 39690 }, { "epoch": 0.7784313725490196, "grad_norm": 1.9827059507369995, "learning_rate": 7.170607896099014e-07, "loss": 0.1484, "step": 39700 }, { "epoch": 0.7786274509803922, "grad_norm": 2.018660545349121, "learning_rate": 7.169147453241188e-07, "loss": 0.1499, "step": 39710 }, { "epoch": 0.7788235294117647, "grad_norm": 2.2641234397888184, "learning_rate": 7.167686782379996e-07, "loss": 0.1357, "step": 39720 }, { "epoch": 0.7790196078431373, "grad_norm": 3.290618896484375, "learning_rate": 7.166225883668968e-07, "loss": 0.1924, "step": 39730 }, { "epoch": 0.7792156862745098, "grad_norm": 1.724480390548706, "learning_rate": 7.164764757261666e-07, "loss": 0.1778, "step": 39740 }, { "epoch": 0.7794117647058824, "grad_norm": 1.1316754817962646, "learning_rate": 7.163303403311666e-07, "loss": 0.155, "step": 39750 }, { "epoch": 0.779607843137255, "grad_norm": 2.7377512454986572, "learning_rate": 7.161841821972579e-07, "loss": 0.1488, "step": 39760 }, { "epoch": 0.7798039215686274, "grad_norm": 3.248469352722168, "learning_rate": 7.160380013398033e-07, "loss": 0.1778, "step": 39770 }, { "epoch": 0.78, "grad_norm": 3.671583890914917, "learning_rate": 7.158917977741682e-07, "loss": 0.1362, "step": 39780 }, { "epoch": 0.7801960784313725, "grad_norm": 3.577568292617798, "learning_rate": 7.157455715157204e-07, "loss": 0.1993, "step": 39790 }, { "epoch": 0.7803921568627451, "grad_norm": 5.7486467361450195, "learning_rate": 7.155993225798298e-07, "loss": 0.1332, "step": 39800 }, { "epoch": 0.7805882352941177, "grad_norm": 2.5205600261688232, "learning_rate": 7.154530509818692e-07, "loss": 0.1726, "step": 39810 }, { "epoch": 0.7807843137254902, "grad_norm": 2.3190274238586426, "learning_rate": 7.153067567372133e-07, "loss": 0.1618, "step": 39820 }, { "epoch": 0.7809803921568628, "grad_norm": 3.0642995834350586, "learning_rate": 7.151604398612392e-07, "loss": 0.1673, "step": 39830 }, { "epoch": 0.7811764705882352, "grad_norm": 2.390488862991333, "learning_rate": 7.150141003693271e-07, "loss": 0.1664, "step": 39840 }, { "epoch": 0.7813725490196078, "grad_norm": 4.903113842010498, "learning_rate": 7.148677382768585e-07, "loss": 0.1457, "step": 39850 }, { "epoch": 0.7815686274509804, "grad_norm": 3.177661418914795, "learning_rate": 7.147213535992182e-07, "loss": 0.1405, "step": 39860 }, { "epoch": 0.7817647058823529, "grad_norm": 3.2934982776641846, "learning_rate": 7.145749463517928e-07, "loss": 0.1231, "step": 39870 }, { "epoch": 0.7819607843137255, "grad_norm": 3.1458804607391357, "learning_rate": 7.144285165499713e-07, "loss": 0.1567, "step": 39880 }, { "epoch": 0.782156862745098, "grad_norm": 1.8827464580535889, "learning_rate": 7.142820642091455e-07, "loss": 0.1603, "step": 39890 }, { "epoch": 0.7823529411764706, "grad_norm": 2.1740682125091553, "learning_rate": 7.141355893447092e-07, "loss": 0.1476, "step": 39900 }, { "epoch": 0.7825490196078432, "grad_norm": 1.695892095565796, "learning_rate": 7.139890919720586e-07, "loss": 0.1759, "step": 39910 }, { "epoch": 0.7827450980392157, "grad_norm": 2.7023558616638184, "learning_rate": 7.138425721065924e-07, "loss": 0.1307, "step": 39920 }, { "epoch": 0.7829411764705883, "grad_norm": 3.7520394325256348, "learning_rate": 7.136960297637115e-07, "loss": 0.146, "step": 39930 }, { "epoch": 0.7831372549019607, "grad_norm": 2.2257933616638184, "learning_rate": 7.135494649588194e-07, "loss": 0.1596, "step": 39940 }, { "epoch": 0.7833333333333333, "grad_norm": 3.2732889652252197, "learning_rate": 7.134028777073217e-07, "loss": 0.1764, "step": 39950 }, { "epoch": 0.7835294117647059, "grad_norm": 2.842649221420288, "learning_rate": 7.132562680246266e-07, "loss": 0.1479, "step": 39960 }, { "epoch": 0.7837254901960784, "grad_norm": 5.4511213302612305, "learning_rate": 7.131096359261442e-07, "loss": 0.1848, "step": 39970 }, { "epoch": 0.783921568627451, "grad_norm": 3.6169004440307617, "learning_rate": 7.129629814272878e-07, "loss": 0.1946, "step": 39980 }, { "epoch": 0.7841176470588235, "grad_norm": 3.565134286880493, "learning_rate": 7.128163045434722e-07, "loss": 0.1619, "step": 39990 }, { "epoch": 0.7843137254901961, "grad_norm": 1.9688167572021484, "learning_rate": 7.126696052901149e-07, "loss": 0.2014, "step": 40000 }, { "epoch": 0.7845098039215687, "grad_norm": 4.134446620941162, "learning_rate": 7.125228836826358e-07, "loss": 0.1666, "step": 40010 }, { "epoch": 0.7847058823529411, "grad_norm": 2.240720272064209, "learning_rate": 7.123761397364571e-07, "loss": 0.1494, "step": 40020 }, { "epoch": 0.7849019607843137, "grad_norm": 2.5820870399475098, "learning_rate": 7.122293734670036e-07, "loss": 0.1424, "step": 40030 }, { "epoch": 0.7850980392156863, "grad_norm": 4.761219501495361, "learning_rate": 7.120825848897018e-07, "loss": 0.1585, "step": 40040 }, { "epoch": 0.7852941176470588, "grad_norm": 2.9655158519744873, "learning_rate": 7.119357740199811e-07, "loss": 0.1364, "step": 40050 }, { "epoch": 0.7854901960784314, "grad_norm": 2.2872865200042725, "learning_rate": 7.117889408732731e-07, "loss": 0.1697, "step": 40060 }, { "epoch": 0.7856862745098039, "grad_norm": 6.553128719329834, "learning_rate": 7.116420854650115e-07, "loss": 0.1614, "step": 40070 }, { "epoch": 0.7858823529411765, "grad_norm": 2.73024845123291, "learning_rate": 7.114952078106329e-07, "loss": 0.1703, "step": 40080 }, { "epoch": 0.7860784313725491, "grad_norm": 2.0605623722076416, "learning_rate": 7.11348307925576e-07, "loss": 0.1563, "step": 40090 }, { "epoch": 0.7862745098039216, "grad_norm": 1.3444337844848633, "learning_rate": 7.112013858252812e-07, "loss": 0.1572, "step": 40100 }, { "epoch": 0.7864705882352941, "grad_norm": 1.5451405048370361, "learning_rate": 7.110544415251922e-07, "loss": 0.1557, "step": 40110 }, { "epoch": 0.7866666666666666, "grad_norm": 2.6203343868255615, "learning_rate": 7.109074750407543e-07, "loss": 0.1699, "step": 40120 }, { "epoch": 0.7868627450980392, "grad_norm": 3.509540319442749, "learning_rate": 7.107604863874157e-07, "loss": 0.1287, "step": 40130 }, { "epoch": 0.7870588235294118, "grad_norm": 2.0877795219421387, "learning_rate": 7.106134755806267e-07, "loss": 0.1572, "step": 40140 }, { "epoch": 0.7872549019607843, "grad_norm": 2.1389927864074707, "learning_rate": 7.104664426358396e-07, "loss": 0.181, "step": 40150 }, { "epoch": 0.7874509803921569, "grad_norm": 2.832797050476074, "learning_rate": 7.103193875685096e-07, "loss": 0.1645, "step": 40160 }, { "epoch": 0.7876470588235294, "grad_norm": 2.6281042098999023, "learning_rate": 7.101723103940939e-07, "loss": 0.1618, "step": 40170 }, { "epoch": 0.787843137254902, "grad_norm": 2.4615440368652344, "learning_rate": 7.100252111280519e-07, "loss": 0.1391, "step": 40180 }, { "epoch": 0.7880392156862746, "grad_norm": 2.4368271827697754, "learning_rate": 7.098780897858455e-07, "loss": 0.1601, "step": 40190 }, { "epoch": 0.788235294117647, "grad_norm": 1.8677308559417725, "learning_rate": 7.097309463829391e-07, "loss": 0.1609, "step": 40200 }, { "epoch": 0.7884313725490196, "grad_norm": 1.7947825193405151, "learning_rate": 7.095837809347992e-07, "loss": 0.1299, "step": 40210 }, { "epoch": 0.7886274509803921, "grad_norm": 1.603672981262207, "learning_rate": 7.094365934568947e-07, "loss": 0.1604, "step": 40220 }, { "epoch": 0.7888235294117647, "grad_norm": 3.2499003410339355, "learning_rate": 7.092893839646965e-07, "loss": 0.1732, "step": 40230 }, { "epoch": 0.7890196078431373, "grad_norm": 5.179914474487305, "learning_rate": 7.091421524736783e-07, "loss": 0.1461, "step": 40240 }, { "epoch": 0.7892156862745098, "grad_norm": 1.2110742330551147, "learning_rate": 7.08994898999316e-07, "loss": 0.1688, "step": 40250 }, { "epoch": 0.7894117647058824, "grad_norm": 1.5435436964035034, "learning_rate": 7.088476235570875e-07, "loss": 0.139, "step": 40260 }, { "epoch": 0.7896078431372549, "grad_norm": 1.5666979551315308, "learning_rate": 7.087003261624731e-07, "loss": 0.1532, "step": 40270 }, { "epoch": 0.7898039215686274, "grad_norm": 3.0044972896575928, "learning_rate": 7.085530068309557e-07, "loss": 0.1338, "step": 40280 }, { "epoch": 0.79, "grad_norm": 3.3770530223846436, "learning_rate": 7.084056655780204e-07, "loss": 0.1446, "step": 40290 }, { "epoch": 0.7901960784313725, "grad_norm": 3.309309244155884, "learning_rate": 7.082583024191546e-07, "loss": 0.1721, "step": 40300 }, { "epoch": 0.7903921568627451, "grad_norm": 3.49357533454895, "learning_rate": 7.081109173698475e-07, "loss": 0.1472, "step": 40310 }, { "epoch": 0.7905882352941176, "grad_norm": 4.053843021392822, "learning_rate": 7.079635104455913e-07, "loss": 0.1654, "step": 40320 }, { "epoch": 0.7907843137254902, "grad_norm": 2.4542291164398193, "learning_rate": 7.078160816618805e-07, "loss": 0.1603, "step": 40330 }, { "epoch": 0.7909803921568628, "grad_norm": 3.334069013595581, "learning_rate": 7.076686310342111e-07, "loss": 0.1599, "step": 40340 }, { "epoch": 0.7911764705882353, "grad_norm": 2.3886585235595703, "learning_rate": 7.075211585780823e-07, "loss": 0.1572, "step": 40350 }, { "epoch": 0.7913725490196079, "grad_norm": 2.088690757751465, "learning_rate": 7.073736643089951e-07, "loss": 0.1493, "step": 40360 }, { "epoch": 0.7915686274509804, "grad_norm": 3.756101369857788, "learning_rate": 7.072261482424529e-07, "loss": 0.1703, "step": 40370 }, { "epoch": 0.7917647058823529, "grad_norm": 2.1334879398345947, "learning_rate": 7.070786103939615e-07, "loss": 0.1359, "step": 40380 }, { "epoch": 0.7919607843137255, "grad_norm": 5.499513149261475, "learning_rate": 7.069310507790287e-07, "loss": 0.1279, "step": 40390 }, { "epoch": 0.792156862745098, "grad_norm": 1.5469056367874146, "learning_rate": 7.067834694131649e-07, "loss": 0.147, "step": 40400 }, { "epoch": 0.7923529411764706, "grad_norm": 2.4022529125213623, "learning_rate": 7.066358663118828e-07, "loss": 0.1419, "step": 40410 }, { "epoch": 0.7925490196078432, "grad_norm": 4.701388359069824, "learning_rate": 7.06488241490697e-07, "loss": 0.1933, "step": 40420 }, { "epoch": 0.7927450980392157, "grad_norm": 5.520942211151123, "learning_rate": 7.063405949651249e-07, "loss": 0.1565, "step": 40430 }, { "epoch": 0.7929411764705883, "grad_norm": 2.5420753955841064, "learning_rate": 7.061929267506859e-07, "loss": 0.1456, "step": 40440 }, { "epoch": 0.7931372549019607, "grad_norm": 1.7023513317108154, "learning_rate": 7.060452368629013e-07, "loss": 0.1375, "step": 40450 }, { "epoch": 0.7933333333333333, "grad_norm": 2.8976211547851562, "learning_rate": 7.058975253172954e-07, "loss": 0.1474, "step": 40460 }, { "epoch": 0.7935294117647059, "grad_norm": 3.140739917755127, "learning_rate": 7.057497921293944e-07, "loss": 0.1382, "step": 40470 }, { "epoch": 0.7937254901960784, "grad_norm": 1.6891677379608154, "learning_rate": 7.05602037314727e-07, "loss": 0.1679, "step": 40480 }, { "epoch": 0.793921568627451, "grad_norm": 4.524699687957764, "learning_rate": 7.054542608888239e-07, "loss": 0.1641, "step": 40490 }, { "epoch": 0.7941176470588235, "grad_norm": 2.6581335067749023, "learning_rate": 7.053064628672178e-07, "loss": 0.1466, "step": 40500 }, { "epoch": 0.7943137254901961, "grad_norm": 2.0244638919830322, "learning_rate": 7.051586432654445e-07, "loss": 0.1562, "step": 40510 }, { "epoch": 0.7945098039215687, "grad_norm": 1.6682934761047363, "learning_rate": 7.050108020990415e-07, "loss": 0.1644, "step": 40520 }, { "epoch": 0.7947058823529412, "grad_norm": 3.8364033699035645, "learning_rate": 7.048629393835485e-07, "loss": 0.1329, "step": 40530 }, { "epoch": 0.7949019607843137, "grad_norm": 2.606611967086792, "learning_rate": 7.047150551345079e-07, "loss": 0.1517, "step": 40540 }, { "epoch": 0.7950980392156862, "grad_norm": 2.10400652885437, "learning_rate": 7.04567149367464e-07, "loss": 0.1731, "step": 40550 }, { "epoch": 0.7952941176470588, "grad_norm": 2.5740668773651123, "learning_rate": 7.044192220979633e-07, "loss": 0.1694, "step": 40560 }, { "epoch": 0.7954901960784314, "grad_norm": 2.0313503742218018, "learning_rate": 7.04271273341555e-07, "loss": 0.1411, "step": 40570 }, { "epoch": 0.7956862745098039, "grad_norm": 2.56412935256958, "learning_rate": 7.041233031137902e-07, "loss": 0.1432, "step": 40580 }, { "epoch": 0.7958823529411765, "grad_norm": 3.265395402908325, "learning_rate": 7.039753114302222e-07, "loss": 0.1649, "step": 40590 }, { "epoch": 0.796078431372549, "grad_norm": 4.163300037384033, "learning_rate": 7.038272983064068e-07, "loss": 0.1435, "step": 40600 }, { "epoch": 0.7962745098039216, "grad_norm": 2.3250985145568848, "learning_rate": 7.036792637579021e-07, "loss": 0.131, "step": 40610 }, { "epoch": 0.7964705882352942, "grad_norm": 3.990990161895752, "learning_rate": 7.03531207800268e-07, "loss": 0.1646, "step": 40620 }, { "epoch": 0.7966666666666666, "grad_norm": 2.1660304069519043, "learning_rate": 7.033831304490671e-07, "loss": 0.1761, "step": 40630 }, { "epoch": 0.7968627450980392, "grad_norm": 1.75338876247406, "learning_rate": 7.032350317198641e-07, "loss": 0.1684, "step": 40640 }, { "epoch": 0.7970588235294118, "grad_norm": 2.308126926422119, "learning_rate": 7.030869116282261e-07, "loss": 0.1598, "step": 40650 }, { "epoch": 0.7972549019607843, "grad_norm": 2.279343366622925, "learning_rate": 7.02938770189722e-07, "loss": 0.1461, "step": 40660 }, { "epoch": 0.7974509803921569, "grad_norm": 5.6655497550964355, "learning_rate": 7.027906074199236e-07, "loss": 0.1642, "step": 40670 }, { "epoch": 0.7976470588235294, "grad_norm": 3.2196309566497803, "learning_rate": 7.026424233344043e-07, "loss": 0.1447, "step": 40680 }, { "epoch": 0.797843137254902, "grad_norm": 2.0570874214172363, "learning_rate": 7.024942179487399e-07, "loss": 0.1473, "step": 40690 }, { "epoch": 0.7980392156862746, "grad_norm": 1.9985287189483643, "learning_rate": 7.02345991278509e-07, "loss": 0.1505, "step": 40700 }, { "epoch": 0.798235294117647, "grad_norm": 5.921596527099609, "learning_rate": 7.021977433392917e-07, "loss": 0.2027, "step": 40710 }, { "epoch": 0.7984313725490196, "grad_norm": 2.9446144104003906, "learning_rate": 7.020494741466708e-07, "loss": 0.1546, "step": 40720 }, { "epoch": 0.7986274509803921, "grad_norm": 7.019691467285156, "learning_rate": 7.019011837162308e-07, "loss": 0.1332, "step": 40730 }, { "epoch": 0.7988235294117647, "grad_norm": 2.238997220993042, "learning_rate": 7.017528720635592e-07, "loss": 0.164, "step": 40740 }, { "epoch": 0.7990196078431373, "grad_norm": 3.0609796047210693, "learning_rate": 7.016045392042451e-07, "loss": 0.1293, "step": 40750 }, { "epoch": 0.7992156862745098, "grad_norm": 2.248652219772339, "learning_rate": 7.014561851538802e-07, "loss": 0.1192, "step": 40760 }, { "epoch": 0.7994117647058824, "grad_norm": 2.8365488052368164, "learning_rate": 7.013078099280582e-07, "loss": 0.1348, "step": 40770 }, { "epoch": 0.7996078431372549, "grad_norm": 3.408501148223877, "learning_rate": 7.011594135423753e-07, "loss": 0.1722, "step": 40780 }, { "epoch": 0.7998039215686275, "grad_norm": 4.61792516708374, "learning_rate": 7.010109960124295e-07, "loss": 0.1705, "step": 40790 }, { "epoch": 0.8, "grad_norm": 4.104176998138428, "learning_rate": 7.008625573538213e-07, "loss": 0.1518, "step": 40800 }, { "epoch": 0.8001960784313725, "grad_norm": 1.2628631591796875, "learning_rate": 7.007140975821534e-07, "loss": 0.2039, "step": 40810 }, { "epoch": 0.8003921568627451, "grad_norm": 1.5114822387695312, "learning_rate": 7.005656167130308e-07, "loss": 0.143, "step": 40820 }, { "epoch": 0.8005882352941176, "grad_norm": 2.3784751892089844, "learning_rate": 7.004171147620604e-07, "loss": 0.1411, "step": 40830 }, { "epoch": 0.8007843137254902, "grad_norm": 2.3384957313537598, "learning_rate": 7.002685917448518e-07, "loss": 0.1476, "step": 40840 }, { "epoch": 0.8009803921568628, "grad_norm": 3.254685640335083, "learning_rate": 7.001200476770165e-07, "loss": 0.1497, "step": 40850 }, { "epoch": 0.8011764705882353, "grad_norm": 2.990626096725464, "learning_rate": 6.99971482574168e-07, "loss": 0.1633, "step": 40860 }, { "epoch": 0.8013725490196079, "grad_norm": 5.655246734619141, "learning_rate": 6.998228964519226e-07, "loss": 0.1728, "step": 40870 }, { "epoch": 0.8015686274509803, "grad_norm": 2.013291120529175, "learning_rate": 6.996742893258982e-07, "loss": 0.1616, "step": 40880 }, { "epoch": 0.8017647058823529, "grad_norm": 3.401047945022583, "learning_rate": 6.995256612117154e-07, "loss": 0.1447, "step": 40890 }, { "epoch": 0.8019607843137255, "grad_norm": 2.8905694484710693, "learning_rate": 6.993770121249968e-07, "loss": 0.1541, "step": 40900 }, { "epoch": 0.802156862745098, "grad_norm": 5.784445762634277, "learning_rate": 6.99228342081367e-07, "loss": 0.1319, "step": 40910 }, { "epoch": 0.8023529411764706, "grad_norm": 3.1784987449645996, "learning_rate": 6.990796510964531e-07, "loss": 0.1476, "step": 40920 }, { "epoch": 0.8025490196078431, "grad_norm": 2.7917654514312744, "learning_rate": 6.989309391858843e-07, "loss": 0.1448, "step": 40930 }, { "epoch": 0.8027450980392157, "grad_norm": 1.9262584447860718, "learning_rate": 6.987822063652921e-07, "loss": 0.1811, "step": 40940 }, { "epoch": 0.8029411764705883, "grad_norm": 3.2554972171783447, "learning_rate": 6.9863345265031e-07, "loss": 0.1569, "step": 40950 }, { "epoch": 0.8031372549019608, "grad_norm": 3.0176455974578857, "learning_rate": 6.984846780565737e-07, "loss": 0.1476, "step": 40960 }, { "epoch": 0.8033333333333333, "grad_norm": 4.292030334472656, "learning_rate": 6.983358825997214e-07, "loss": 0.2083, "step": 40970 }, { "epoch": 0.8035294117647059, "grad_norm": 2.5281262397766113, "learning_rate": 6.981870662953932e-07, "loss": 0.164, "step": 40980 }, { "epoch": 0.8037254901960784, "grad_norm": 2.7931885719299316, "learning_rate": 6.980382291592315e-07, "loss": 0.1409, "step": 40990 }, { "epoch": 0.803921568627451, "grad_norm": 2.787452459335327, "learning_rate": 6.978893712068808e-07, "loss": 0.125, "step": 41000 }, { "epoch": 0.8041176470588235, "grad_norm": 2.788698434829712, "learning_rate": 6.977404924539879e-07, "loss": 0.1568, "step": 41010 }, { "epoch": 0.8043137254901961, "grad_norm": 3.441291332244873, "learning_rate": 6.975915929162018e-07, "loss": 0.1131, "step": 41020 }, { "epoch": 0.8045098039215687, "grad_norm": 2.240431785583496, "learning_rate": 6.974426726091734e-07, "loss": 0.1557, "step": 41030 }, { "epoch": 0.8047058823529412, "grad_norm": 3.3121607303619385, "learning_rate": 6.972937315485562e-07, "loss": 0.1335, "step": 41040 }, { "epoch": 0.8049019607843138, "grad_norm": 2.2846593856811523, "learning_rate": 6.971447697500056e-07, "loss": 0.1345, "step": 41050 }, { "epoch": 0.8050980392156862, "grad_norm": 1.92795991897583, "learning_rate": 6.969957872291794e-07, "loss": 0.1472, "step": 41060 }, { "epoch": 0.8052941176470588, "grad_norm": 3.0046749114990234, "learning_rate": 6.968467840017374e-07, "loss": 0.1462, "step": 41070 }, { "epoch": 0.8054901960784314, "grad_norm": 1.8013023138046265, "learning_rate": 6.966977600833414e-07, "loss": 0.1567, "step": 41080 }, { "epoch": 0.8056862745098039, "grad_norm": 2.9002537727355957, "learning_rate": 6.965487154896559e-07, "loss": 0.1591, "step": 41090 }, { "epoch": 0.8058823529411765, "grad_norm": 2.590982437133789, "learning_rate": 6.963996502363472e-07, "loss": 0.1335, "step": 41100 }, { "epoch": 0.806078431372549, "grad_norm": 3.466439723968506, "learning_rate": 6.962505643390837e-07, "loss": 0.1466, "step": 41110 }, { "epoch": 0.8062745098039216, "grad_norm": 3.934016227722168, "learning_rate": 6.961014578135363e-07, "loss": 0.1436, "step": 41120 }, { "epoch": 0.8064705882352942, "grad_norm": 4.20259952545166, "learning_rate": 6.959523306753777e-07, "loss": 0.1595, "step": 41130 }, { "epoch": 0.8066666666666666, "grad_norm": 3.5587615966796875, "learning_rate": 6.958031829402831e-07, "loss": 0.1663, "step": 41140 }, { "epoch": 0.8068627450980392, "grad_norm": 1.823899507522583, "learning_rate": 6.956540146239296e-07, "loss": 0.1373, "step": 41150 }, { "epoch": 0.8070588235294117, "grad_norm": 3.616270065307617, "learning_rate": 6.955048257419966e-07, "loss": 0.1292, "step": 41160 }, { "epoch": 0.8072549019607843, "grad_norm": 2.0623221397399902, "learning_rate": 6.953556163101658e-07, "loss": 0.1642, "step": 41170 }, { "epoch": 0.8074509803921569, "grad_norm": 2.2629873752593994, "learning_rate": 6.952063863441206e-07, "loss": 0.161, "step": 41180 }, { "epoch": 0.8076470588235294, "grad_norm": 2.413954496383667, "learning_rate": 6.950571358595472e-07, "loss": 0.1266, "step": 41190 }, { "epoch": 0.807843137254902, "grad_norm": 2.2187986373901367, "learning_rate": 6.949078648721331e-07, "loss": 0.1403, "step": 41200 }, { "epoch": 0.8080392156862745, "grad_norm": 2.7903616428375244, "learning_rate": 6.947585733975692e-07, "loss": 0.139, "step": 41210 }, { "epoch": 0.808235294117647, "grad_norm": 2.3157427310943604, "learning_rate": 6.946092614515471e-07, "loss": 0.164, "step": 41220 }, { "epoch": 0.8084313725490196, "grad_norm": 3.111637592315674, "learning_rate": 6.944599290497618e-07, "loss": 0.1518, "step": 41230 }, { "epoch": 0.8086274509803921, "grad_norm": 3.1054155826568604, "learning_rate": 6.943105762079096e-07, "loss": 0.1548, "step": 41240 }, { "epoch": 0.8088235294117647, "grad_norm": 4.985475063323975, "learning_rate": 6.941612029416896e-07, "loss": 0.168, "step": 41250 }, { "epoch": 0.8090196078431372, "grad_norm": 3.8418374061584473, "learning_rate": 6.940118092668021e-07, "loss": 0.1719, "step": 41260 }, { "epoch": 0.8092156862745098, "grad_norm": 3.5094377994537354, "learning_rate": 6.93862395198951e-07, "loss": 0.1657, "step": 41270 }, { "epoch": 0.8094117647058824, "grad_norm": 2.39504075050354, "learning_rate": 6.937129607538408e-07, "loss": 0.1775, "step": 41280 }, { "epoch": 0.8096078431372549, "grad_norm": 2.905626058578491, "learning_rate": 6.935635059471791e-07, "loss": 0.1832, "step": 41290 }, { "epoch": 0.8098039215686275, "grad_norm": 2.1327853202819824, "learning_rate": 6.934140307946757e-07, "loss": 0.1556, "step": 41300 }, { "epoch": 0.81, "grad_norm": 2.154127836227417, "learning_rate": 6.932645353120417e-07, "loss": 0.1662, "step": 41310 }, { "epoch": 0.8101960784313725, "grad_norm": 1.622618317604065, "learning_rate": 6.931150195149912e-07, "loss": 0.1574, "step": 41320 }, { "epoch": 0.8103921568627451, "grad_norm": 1.6006509065628052, "learning_rate": 6.9296548341924e-07, "loss": 0.1213, "step": 41330 }, { "epoch": 0.8105882352941176, "grad_norm": 5.132781982421875, "learning_rate": 6.928159270405062e-07, "loss": 0.147, "step": 41340 }, { "epoch": 0.8107843137254902, "grad_norm": 2.7487220764160156, "learning_rate": 6.926663503945098e-07, "loss": 0.161, "step": 41350 }, { "epoch": 0.8109803921568628, "grad_norm": 2.43206524848938, "learning_rate": 6.925167534969732e-07, "loss": 0.1506, "step": 41360 }, { "epoch": 0.8111764705882353, "grad_norm": 2.2513043880462646, "learning_rate": 6.923671363636209e-07, "loss": 0.1198, "step": 41370 }, { "epoch": 0.8113725490196079, "grad_norm": 2.2558529376983643, "learning_rate": 6.922174990101795e-07, "loss": 0.15, "step": 41380 }, { "epoch": 0.8115686274509804, "grad_norm": 6.316461086273193, "learning_rate": 6.920678414523772e-07, "loss": 0.1529, "step": 41390 }, { "epoch": 0.8117647058823529, "grad_norm": 2.038435697555542, "learning_rate": 6.919181637059454e-07, "loss": 0.1424, "step": 41400 }, { "epoch": 0.8119607843137255, "grad_norm": 4.581795692443848, "learning_rate": 6.91768465786617e-07, "loss": 0.1641, "step": 41410 }, { "epoch": 0.812156862745098, "grad_norm": 3.7583484649658203, "learning_rate": 6.916187477101265e-07, "loss": 0.1699, "step": 41420 }, { "epoch": 0.8123529411764706, "grad_norm": 2.554607629776001, "learning_rate": 6.914690094922116e-07, "loss": 0.1379, "step": 41430 }, { "epoch": 0.8125490196078431, "grad_norm": 2.8328890800476074, "learning_rate": 6.913192511486111e-07, "loss": 0.1583, "step": 41440 }, { "epoch": 0.8127450980392157, "grad_norm": 1.7599941492080688, "learning_rate": 6.911694726950668e-07, "loss": 0.1657, "step": 41450 }, { "epoch": 0.8129411764705883, "grad_norm": 2.324828863143921, "learning_rate": 6.910196741473223e-07, "loss": 0.1422, "step": 41460 }, { "epoch": 0.8131372549019608, "grad_norm": 3.256701946258545, "learning_rate": 6.908698555211228e-07, "loss": 0.1629, "step": 41470 }, { "epoch": 0.8133333333333334, "grad_norm": 2.166578531265259, "learning_rate": 6.907200168322161e-07, "loss": 0.1856, "step": 41480 }, { "epoch": 0.8135294117647058, "grad_norm": 2.585761070251465, "learning_rate": 6.905701580963524e-07, "loss": 0.1737, "step": 41490 }, { "epoch": 0.8137254901960784, "grad_norm": 1.8700296878814697, "learning_rate": 6.904202793292833e-07, "loss": 0.1613, "step": 41500 }, { "epoch": 0.813921568627451, "grad_norm": 2.2151196002960205, "learning_rate": 6.90270380546763e-07, "loss": 0.1229, "step": 41510 }, { "epoch": 0.8141176470588235, "grad_norm": 1.2169873714447021, "learning_rate": 6.901204617645477e-07, "loss": 0.1532, "step": 41520 }, { "epoch": 0.8143137254901961, "grad_norm": 3.4181716442108154, "learning_rate": 6.899705229983953e-07, "loss": 0.1562, "step": 41530 }, { "epoch": 0.8145098039215686, "grad_norm": 2.3398122787475586, "learning_rate": 6.898205642640668e-07, "loss": 0.137, "step": 41540 }, { "epoch": 0.8147058823529412, "grad_norm": 5.12694787979126, "learning_rate": 6.896705855773241e-07, "loss": 0.1583, "step": 41550 }, { "epoch": 0.8149019607843138, "grad_norm": 0.8438624143600464, "learning_rate": 6.895205869539318e-07, "loss": 0.1538, "step": 41560 }, { "epoch": 0.8150980392156862, "grad_norm": 4.163977146148682, "learning_rate": 6.893705684096571e-07, "loss": 0.1709, "step": 41570 }, { "epoch": 0.8152941176470588, "grad_norm": 2.253838539123535, "learning_rate": 6.892205299602682e-07, "loss": 0.1382, "step": 41580 }, { "epoch": 0.8154901960784313, "grad_norm": 2.2694735527038574, "learning_rate": 6.89070471621536e-07, "loss": 0.1531, "step": 41590 }, { "epoch": 0.8156862745098039, "grad_norm": 2.2464613914489746, "learning_rate": 6.889203934092335e-07, "loss": 0.1451, "step": 41600 }, { "epoch": 0.8158823529411765, "grad_norm": 2.9337260723114014, "learning_rate": 6.887702953391359e-07, "loss": 0.1333, "step": 41610 }, { "epoch": 0.816078431372549, "grad_norm": 1.7986791133880615, "learning_rate": 6.886201774270199e-07, "loss": 0.149, "step": 41620 }, { "epoch": 0.8162745098039216, "grad_norm": 2.6202080249786377, "learning_rate": 6.884700396886651e-07, "loss": 0.1479, "step": 41630 }, { "epoch": 0.8164705882352942, "grad_norm": 6.380600929260254, "learning_rate": 6.883198821398526e-07, "loss": 0.1745, "step": 41640 }, { "epoch": 0.8166666666666667, "grad_norm": 1.9076604843139648, "learning_rate": 6.881697047963658e-07, "loss": 0.1593, "step": 41650 }, { "epoch": 0.8168627450980392, "grad_norm": 2.6210687160491943, "learning_rate": 6.8801950767399e-07, "loss": 0.1359, "step": 41660 }, { "epoch": 0.8170588235294117, "grad_norm": 4.593018531799316, "learning_rate": 6.878692907885128e-07, "loss": 0.1441, "step": 41670 }, { "epoch": 0.8172549019607843, "grad_norm": 4.022291660308838, "learning_rate": 6.877190541557239e-07, "loss": 0.1636, "step": 41680 }, { "epoch": 0.8174509803921569, "grad_norm": 2.9517364501953125, "learning_rate": 6.875687977914148e-07, "loss": 0.1508, "step": 41690 }, { "epoch": 0.8176470588235294, "grad_norm": 2.9617345333099365, "learning_rate": 6.874185217113794e-07, "loss": 0.1726, "step": 41700 }, { "epoch": 0.817843137254902, "grad_norm": 7.023102283477783, "learning_rate": 6.872682259314133e-07, "loss": 0.1531, "step": 41710 }, { "epoch": 0.8180392156862745, "grad_norm": 3.892937660217285, "learning_rate": 6.871179104673146e-07, "loss": 0.1725, "step": 41720 }, { "epoch": 0.8182352941176471, "grad_norm": 4.944360733032227, "learning_rate": 6.869675753348832e-07, "loss": 0.1693, "step": 41730 }, { "epoch": 0.8184313725490197, "grad_norm": 4.708578586578369, "learning_rate": 6.86817220549921e-07, "loss": 0.1772, "step": 41740 }, { "epoch": 0.8186274509803921, "grad_norm": 1.8787298202514648, "learning_rate": 6.866668461282324e-07, "loss": 0.1641, "step": 41750 }, { "epoch": 0.8188235294117647, "grad_norm": 2.228208065032959, "learning_rate": 6.865164520856231e-07, "loss": 0.1501, "step": 41760 }, { "epoch": 0.8190196078431372, "grad_norm": 2.7745320796966553, "learning_rate": 6.863660384379017e-07, "loss": 0.1616, "step": 41770 }, { "epoch": 0.8192156862745098, "grad_norm": 2.5037386417388916, "learning_rate": 6.862156052008782e-07, "loss": 0.1395, "step": 41780 }, { "epoch": 0.8194117647058824, "grad_norm": 2.82781982421875, "learning_rate": 6.860651523903651e-07, "loss": 0.1646, "step": 41790 }, { "epoch": 0.8196078431372549, "grad_norm": 4.306722164154053, "learning_rate": 6.859146800221767e-07, "loss": 0.1505, "step": 41800 }, { "epoch": 0.8198039215686275, "grad_norm": 1.6087533235549927, "learning_rate": 6.857641881121297e-07, "loss": 0.1523, "step": 41810 }, { "epoch": 0.82, "grad_norm": 5.535120487213135, "learning_rate": 6.856136766760422e-07, "loss": 0.1459, "step": 41820 }, { "epoch": 0.8201960784313725, "grad_norm": 2.8188629150390625, "learning_rate": 6.854631457297351e-07, "loss": 0.1631, "step": 41830 }, { "epoch": 0.8203921568627451, "grad_norm": 2.925342559814453, "learning_rate": 6.853125952890308e-07, "loss": 0.1356, "step": 41840 }, { "epoch": 0.8205882352941176, "grad_norm": 2.7408535480499268, "learning_rate": 6.851620253697539e-07, "loss": 0.1566, "step": 41850 }, { "epoch": 0.8207843137254902, "grad_norm": 2.6890830993652344, "learning_rate": 6.850114359877313e-07, "loss": 0.147, "step": 41860 }, { "epoch": 0.8209803921568627, "grad_norm": 3.743675470352173, "learning_rate": 6.848608271587918e-07, "loss": 0.1558, "step": 41870 }, { "epoch": 0.8211764705882353, "grad_norm": 2.0414505004882812, "learning_rate": 6.847101988987659e-07, "loss": 0.1225, "step": 41880 }, { "epoch": 0.8213725490196079, "grad_norm": 2.65966534614563, "learning_rate": 6.845595512234867e-07, "loss": 0.1557, "step": 41890 }, { "epoch": 0.8215686274509804, "grad_norm": 2.1837215423583984, "learning_rate": 6.844088841487887e-07, "loss": 0.1708, "step": 41900 }, { "epoch": 0.821764705882353, "grad_norm": 1.8732857704162598, "learning_rate": 6.842581976905093e-07, "loss": 0.1628, "step": 41910 }, { "epoch": 0.8219607843137255, "grad_norm": 2.8298020362854004, "learning_rate": 6.841074918644872e-07, "loss": 0.1433, "step": 41920 }, { "epoch": 0.822156862745098, "grad_norm": 3.785677909851074, "learning_rate": 6.839567666865634e-07, "loss": 0.1609, "step": 41930 }, { "epoch": 0.8223529411764706, "grad_norm": 2.3685686588287354, "learning_rate": 6.83806022172581e-07, "loss": 0.1392, "step": 41940 }, { "epoch": 0.8225490196078431, "grad_norm": 2.5206172466278076, "learning_rate": 6.83655258338385e-07, "loss": 0.1573, "step": 41950 }, { "epoch": 0.8227450980392157, "grad_norm": 2.1513490676879883, "learning_rate": 6.835044751998222e-07, "loss": 0.1358, "step": 41960 }, { "epoch": 0.8229411764705883, "grad_norm": 1.924207091331482, "learning_rate": 6.833536727727423e-07, "loss": 0.1685, "step": 41970 }, { "epoch": 0.8231372549019608, "grad_norm": 2.591904878616333, "learning_rate": 6.832028510729959e-07, "loss": 0.1189, "step": 41980 }, { "epoch": 0.8233333333333334, "grad_norm": 1.7232853174209595, "learning_rate": 6.830520101164364e-07, "loss": 0.1522, "step": 41990 }, { "epoch": 0.8235294117647058, "grad_norm": 2.1243348121643066, "learning_rate": 6.829011499189192e-07, "loss": 0.1627, "step": 42000 }, { "epoch": 0.8237254901960784, "grad_norm": 3.5580732822418213, "learning_rate": 6.82750270496301e-07, "loss": 0.1609, "step": 42010 }, { "epoch": 0.823921568627451, "grad_norm": 7.333865642547607, "learning_rate": 6.825993718644416e-07, "loss": 0.1509, "step": 42020 }, { "epoch": 0.8241176470588235, "grad_norm": 2.773085355758667, "learning_rate": 6.824484540392019e-07, "loss": 0.1901, "step": 42030 }, { "epoch": 0.8243137254901961, "grad_norm": 1.796238660812378, "learning_rate": 6.822975170364452e-07, "loss": 0.1732, "step": 42040 }, { "epoch": 0.8245098039215686, "grad_norm": 3.0559048652648926, "learning_rate": 6.821465608720368e-07, "loss": 0.1517, "step": 42050 }, { "epoch": 0.8247058823529412, "grad_norm": 3.1977591514587402, "learning_rate": 6.819955855618442e-07, "loss": 0.1446, "step": 42060 }, { "epoch": 0.8249019607843138, "grad_norm": 2.637722969055176, "learning_rate": 6.818445911217363e-07, "loss": 0.1487, "step": 42070 }, { "epoch": 0.8250980392156863, "grad_norm": 2.134411096572876, "learning_rate": 6.81693577567585e-07, "loss": 0.1456, "step": 42080 }, { "epoch": 0.8252941176470588, "grad_norm": 2.0103650093078613, "learning_rate": 6.81542544915263e-07, "loss": 0.1532, "step": 42090 }, { "epoch": 0.8254901960784313, "grad_norm": 3.5180001258850098, "learning_rate": 6.813914931806461e-07, "loss": 0.1215, "step": 42100 }, { "epoch": 0.8256862745098039, "grad_norm": 2.763042449951172, "learning_rate": 6.812404223796115e-07, "loss": 0.1449, "step": 42110 }, { "epoch": 0.8258823529411765, "grad_norm": 2.880789279937744, "learning_rate": 6.810893325280384e-07, "loss": 0.1544, "step": 42120 }, { "epoch": 0.826078431372549, "grad_norm": 3.322506904602051, "learning_rate": 6.809382236418084e-07, "loss": 0.1762, "step": 42130 }, { "epoch": 0.8262745098039216, "grad_norm": 2.5258147716522217, "learning_rate": 6.807870957368049e-07, "loss": 0.1591, "step": 42140 }, { "epoch": 0.8264705882352941, "grad_norm": 1.7132103443145752, "learning_rate": 6.806359488289129e-07, "loss": 0.1591, "step": 42150 }, { "epoch": 0.8266666666666667, "grad_norm": 1.2648861408233643, "learning_rate": 6.804847829340199e-07, "loss": 0.1512, "step": 42160 }, { "epoch": 0.8268627450980393, "grad_norm": 2.3183839321136475, "learning_rate": 6.803335980680153e-07, "loss": 0.117, "step": 42170 }, { "epoch": 0.8270588235294117, "grad_norm": 1.9157624244689941, "learning_rate": 6.801823942467906e-07, "loss": 0.1448, "step": 42180 }, { "epoch": 0.8272549019607843, "grad_norm": 2.890446662902832, "learning_rate": 6.80031171486239e-07, "loss": 0.1395, "step": 42190 }, { "epoch": 0.8274509803921568, "grad_norm": 2.851205587387085, "learning_rate": 6.798799298022556e-07, "loss": 0.1495, "step": 42200 }, { "epoch": 0.8276470588235294, "grad_norm": 1.409411907196045, "learning_rate": 6.79728669210738e-07, "loss": 0.1558, "step": 42210 }, { "epoch": 0.827843137254902, "grad_norm": 2.5391573905944824, "learning_rate": 6.795773897275855e-07, "loss": 0.1745, "step": 42220 }, { "epoch": 0.8280392156862745, "grad_norm": 3.6200129985809326, "learning_rate": 6.794260913686991e-07, "loss": 0.1533, "step": 42230 }, { "epoch": 0.8282352941176471, "grad_norm": 2.485884666442871, "learning_rate": 6.792747741499822e-07, "loss": 0.1961, "step": 42240 }, { "epoch": 0.8284313725490197, "grad_norm": 5.911131381988525, "learning_rate": 6.791234380873403e-07, "loss": 0.1475, "step": 42250 }, { "epoch": 0.8286274509803921, "grad_norm": 2.2915892601013184, "learning_rate": 6.789720831966803e-07, "loss": 0.1597, "step": 42260 }, { "epoch": 0.8288235294117647, "grad_norm": 3.4578683376312256, "learning_rate": 6.788207094939118e-07, "loss": 0.1618, "step": 42270 }, { "epoch": 0.8290196078431372, "grad_norm": 2.4939911365509033, "learning_rate": 6.786693169949455e-07, "loss": 0.1895, "step": 42280 }, { "epoch": 0.8292156862745098, "grad_norm": 4.520593643188477, "learning_rate": 6.785179057156948e-07, "loss": 0.1983, "step": 42290 }, { "epoch": 0.8294117647058824, "grad_norm": 4.178450584411621, "learning_rate": 6.78366475672075e-07, "loss": 0.1818, "step": 42300 }, { "epoch": 0.8296078431372549, "grad_norm": 1.616969347000122, "learning_rate": 6.782150268800029e-07, "loss": 0.1639, "step": 42310 }, { "epoch": 0.8298039215686275, "grad_norm": 1.950828194618225, "learning_rate": 6.780635593553977e-07, "loss": 0.135, "step": 42320 }, { "epoch": 0.83, "grad_norm": 1.8484303951263428, "learning_rate": 6.779120731141806e-07, "loss": 0.1163, "step": 42330 }, { "epoch": 0.8301960784313726, "grad_norm": 1.696394443511963, "learning_rate": 6.777605681722743e-07, "loss": 0.1585, "step": 42340 }, { "epoch": 0.8303921568627451, "grad_norm": 3.4852445125579834, "learning_rate": 6.776090445456043e-07, "loss": 0.1469, "step": 42350 }, { "epoch": 0.8305882352941176, "grad_norm": 2.497340679168701, "learning_rate": 6.77457502250097e-07, "loss": 0.1506, "step": 42360 }, { "epoch": 0.8307843137254902, "grad_norm": 2.082847833633423, "learning_rate": 6.773059413016816e-07, "loss": 0.1865, "step": 42370 }, { "epoch": 0.8309803921568627, "grad_norm": 2.4711663722991943, "learning_rate": 6.771543617162891e-07, "loss": 0.1344, "step": 42380 }, { "epoch": 0.8311764705882353, "grad_norm": 3.289588212966919, "learning_rate": 6.770027635098519e-07, "loss": 0.1876, "step": 42390 }, { "epoch": 0.8313725490196079, "grad_norm": 5.11430549621582, "learning_rate": 6.76851146698305e-07, "loss": 0.1419, "step": 42400 }, { "epoch": 0.8315686274509804, "grad_norm": 4.942531108856201, "learning_rate": 6.766995112975853e-07, "loss": 0.1536, "step": 42410 }, { "epoch": 0.831764705882353, "grad_norm": 3.014802932739258, "learning_rate": 6.765478573236312e-07, "loss": 0.1385, "step": 42420 }, { "epoch": 0.8319607843137254, "grad_norm": 2.504934787750244, "learning_rate": 6.763961847923837e-07, "loss": 0.149, "step": 42430 }, { "epoch": 0.832156862745098, "grad_norm": 2.078662633895874, "learning_rate": 6.762444937197851e-07, "loss": 0.1517, "step": 42440 }, { "epoch": 0.8323529411764706, "grad_norm": 2.297207832336426, "learning_rate": 6.760927841217801e-07, "loss": 0.1434, "step": 42450 }, { "epoch": 0.8325490196078431, "grad_norm": 3.041660785675049, "learning_rate": 6.759410560143153e-07, "loss": 0.155, "step": 42460 }, { "epoch": 0.8327450980392157, "grad_norm": 2.8262581825256348, "learning_rate": 6.757893094133388e-07, "loss": 0.184, "step": 42470 }, { "epoch": 0.8329411764705882, "grad_norm": 3.9460573196411133, "learning_rate": 6.756375443348013e-07, "loss": 0.1442, "step": 42480 }, { "epoch": 0.8331372549019608, "grad_norm": 0.8730796575546265, "learning_rate": 6.754857607946549e-07, "loss": 0.1483, "step": 42490 }, { "epoch": 0.8333333333333334, "grad_norm": 2.490919828414917, "learning_rate": 6.753339588088541e-07, "loss": 0.1388, "step": 42500 }, { "epoch": 0.8335294117647059, "grad_norm": 2.3303215503692627, "learning_rate": 6.751821383933547e-07, "loss": 0.1445, "step": 42510 }, { "epoch": 0.8337254901960784, "grad_norm": 1.8110909461975098, "learning_rate": 6.750302995641153e-07, "loss": 0.1286, "step": 42520 }, { "epoch": 0.8339215686274509, "grad_norm": 2.22003173828125, "learning_rate": 6.748784423370959e-07, "loss": 0.1419, "step": 42530 }, { "epoch": 0.8341176470588235, "grad_norm": 2.740623712539673, "learning_rate": 6.747265667282584e-07, "loss": 0.1374, "step": 42540 }, { "epoch": 0.8343137254901961, "grad_norm": 2.483210563659668, "learning_rate": 6.745746727535666e-07, "loss": 0.1446, "step": 42550 }, { "epoch": 0.8345098039215686, "grad_norm": 2.7660701274871826, "learning_rate": 6.744227604289866e-07, "loss": 0.1911, "step": 42560 }, { "epoch": 0.8347058823529412, "grad_norm": 2.441727638244629, "learning_rate": 6.742708297704864e-07, "loss": 0.1536, "step": 42570 }, { "epoch": 0.8349019607843138, "grad_norm": 1.87816321849823, "learning_rate": 6.741188807940351e-07, "loss": 0.1732, "step": 42580 }, { "epoch": 0.8350980392156863, "grad_norm": 2.3660836219787598, "learning_rate": 6.739669135156049e-07, "loss": 0.13, "step": 42590 }, { "epoch": 0.8352941176470589, "grad_norm": 2.7262954711914062, "learning_rate": 6.738149279511692e-07, "loss": 0.1734, "step": 42600 }, { "epoch": 0.8354901960784313, "grad_norm": 2.348651170730591, "learning_rate": 6.736629241167033e-07, "loss": 0.1454, "step": 42610 }, { "epoch": 0.8356862745098039, "grad_norm": 2.9127659797668457, "learning_rate": 6.735109020281852e-07, "loss": 0.1548, "step": 42620 }, { "epoch": 0.8358823529411765, "grad_norm": 2.979982376098633, "learning_rate": 6.733588617015937e-07, "loss": 0.1467, "step": 42630 }, { "epoch": 0.836078431372549, "grad_norm": 1.747753381729126, "learning_rate": 6.732068031529101e-07, "loss": 0.1556, "step": 42640 }, { "epoch": 0.8362745098039216, "grad_norm": 2.962661027908325, "learning_rate": 6.73054726398118e-07, "loss": 0.1417, "step": 42650 }, { "epoch": 0.8364705882352941, "grad_norm": 2.208902597427368, "learning_rate": 6.729026314532021e-07, "loss": 0.1186, "step": 42660 }, { "epoch": 0.8366666666666667, "grad_norm": 1.2981764078140259, "learning_rate": 6.727505183341493e-07, "loss": 0.1742, "step": 42670 }, { "epoch": 0.8368627450980393, "grad_norm": 1.4516661167144775, "learning_rate": 6.725983870569489e-07, "loss": 0.1397, "step": 42680 }, { "epoch": 0.8370588235294117, "grad_norm": 2.983807325363159, "learning_rate": 6.724462376375915e-07, "loss": 0.1593, "step": 42690 }, { "epoch": 0.8372549019607843, "grad_norm": 2.697134017944336, "learning_rate": 6.722940700920698e-07, "loss": 0.1917, "step": 42700 }, { "epoch": 0.8374509803921568, "grad_norm": 2.606516122817993, "learning_rate": 6.721418844363784e-07, "loss": 0.1545, "step": 42710 }, { "epoch": 0.8376470588235294, "grad_norm": 3.379955530166626, "learning_rate": 6.719896806865141e-07, "loss": 0.1842, "step": 42720 }, { "epoch": 0.837843137254902, "grad_norm": 2.387467861175537, "learning_rate": 6.718374588584751e-07, "loss": 0.1793, "step": 42730 }, { "epoch": 0.8380392156862745, "grad_norm": 2.786369562149048, "learning_rate": 6.716852189682617e-07, "loss": 0.1387, "step": 42740 }, { "epoch": 0.8382352941176471, "grad_norm": 5.190139293670654, "learning_rate": 6.715329610318763e-07, "loss": 0.1515, "step": 42750 }, { "epoch": 0.8384313725490196, "grad_norm": 2.170034885406494, "learning_rate": 6.713806850653229e-07, "loss": 0.1494, "step": 42760 }, { "epoch": 0.8386274509803922, "grad_norm": 2.782399892807007, "learning_rate": 6.712283910846077e-07, "loss": 0.1908, "step": 42770 }, { "epoch": 0.8388235294117647, "grad_norm": 1.5868244171142578, "learning_rate": 6.710760791057384e-07, "loss": 0.1284, "step": 42780 }, { "epoch": 0.8390196078431372, "grad_norm": 3.7094924449920654, "learning_rate": 6.709237491447249e-07, "loss": 0.1542, "step": 42790 }, { "epoch": 0.8392156862745098, "grad_norm": 1.9384428262710571, "learning_rate": 6.70771401217579e-07, "loss": 0.1409, "step": 42800 }, { "epoch": 0.8394117647058823, "grad_norm": 3.0959417819976807, "learning_rate": 6.706190353403141e-07, "loss": 0.1323, "step": 42810 }, { "epoch": 0.8396078431372549, "grad_norm": 3.3959763050079346, "learning_rate": 6.704666515289458e-07, "loss": 0.1536, "step": 42820 }, { "epoch": 0.8398039215686275, "grad_norm": 4.309385776519775, "learning_rate": 6.703142497994916e-07, "loss": 0.1655, "step": 42830 }, { "epoch": 0.84, "grad_norm": 2.876430034637451, "learning_rate": 6.701618301679705e-07, "loss": 0.1733, "step": 42840 }, { "epoch": 0.8401960784313726, "grad_norm": 3.0497167110443115, "learning_rate": 6.700093926504038e-07, "loss": 0.1868, "step": 42850 }, { "epoch": 0.840392156862745, "grad_norm": 3.009114980697632, "learning_rate": 6.698569372628142e-07, "loss": 0.1472, "step": 42860 }, { "epoch": 0.8405882352941176, "grad_norm": 3.918229579925537, "learning_rate": 6.697044640212271e-07, "loss": 0.1532, "step": 42870 }, { "epoch": 0.8407843137254902, "grad_norm": 2.1716814041137695, "learning_rate": 6.695519729416687e-07, "loss": 0.1498, "step": 42880 }, { "epoch": 0.8409803921568627, "grad_norm": 2.378798246383667, "learning_rate": 6.693994640401683e-07, "loss": 0.1378, "step": 42890 }, { "epoch": 0.8411764705882353, "grad_norm": 1.4182765483856201, "learning_rate": 6.692469373327557e-07, "loss": 0.1351, "step": 42900 }, { "epoch": 0.8413725490196079, "grad_norm": 2.4663374423980713, "learning_rate": 6.690943928354637e-07, "loss": 0.1283, "step": 42910 }, { "epoch": 0.8415686274509804, "grad_norm": 3.1481246948242188, "learning_rate": 6.689418305643267e-07, "loss": 0.1416, "step": 42920 }, { "epoch": 0.841764705882353, "grad_norm": 2.343839168548584, "learning_rate": 6.687892505353802e-07, "loss": 0.157, "step": 42930 }, { "epoch": 0.8419607843137255, "grad_norm": 2.786252737045288, "learning_rate": 6.686366527646628e-07, "loss": 0.1737, "step": 42940 }, { "epoch": 0.842156862745098, "grad_norm": 4.589288711547852, "learning_rate": 6.684840372682142e-07, "loss": 0.1446, "step": 42950 }, { "epoch": 0.8423529411764706, "grad_norm": 2.761115312576294, "learning_rate": 6.683314040620759e-07, "loss": 0.1288, "step": 42960 }, { "epoch": 0.8425490196078431, "grad_norm": 1.3245891332626343, "learning_rate": 6.681787531622917e-07, "loss": 0.1612, "step": 42970 }, { "epoch": 0.8427450980392157, "grad_norm": 3.1540305614471436, "learning_rate": 6.680260845849069e-07, "loss": 0.1834, "step": 42980 }, { "epoch": 0.8429411764705882, "grad_norm": 2.4152138233184814, "learning_rate": 6.67873398345969e-07, "loss": 0.1509, "step": 42990 }, { "epoch": 0.8431372549019608, "grad_norm": 2.2227346897125244, "learning_rate": 6.677206944615271e-07, "loss": 0.1528, "step": 43000 }, { "epoch": 0.8433333333333334, "grad_norm": 2.6962921619415283, "learning_rate": 6.675679729476318e-07, "loss": 0.1357, "step": 43010 }, { "epoch": 0.8435294117647059, "grad_norm": 3.6143295764923096, "learning_rate": 6.674152338203364e-07, "loss": 0.1308, "step": 43020 }, { "epoch": 0.8437254901960785, "grad_norm": 2.1860523223876953, "learning_rate": 6.672624770956954e-07, "loss": 0.137, "step": 43030 }, { "epoch": 0.8439215686274509, "grad_norm": 2.9123334884643555, "learning_rate": 6.671097027897657e-07, "loss": 0.1495, "step": 43040 }, { "epoch": 0.8441176470588235, "grad_norm": 2.435288429260254, "learning_rate": 6.669569109186052e-07, "loss": 0.1291, "step": 43050 }, { "epoch": 0.8443137254901961, "grad_norm": 6.301589012145996, "learning_rate": 6.668041014982745e-07, "loss": 0.154, "step": 43060 }, { "epoch": 0.8445098039215686, "grad_norm": 2.6242895126342773, "learning_rate": 6.666512745448356e-07, "loss": 0.1502, "step": 43070 }, { "epoch": 0.8447058823529412, "grad_norm": 2.8119394779205322, "learning_rate": 6.664984300743526e-07, "loss": 0.1523, "step": 43080 }, { "epoch": 0.8449019607843137, "grad_norm": 3.0920393466949463, "learning_rate": 6.663455681028909e-07, "loss": 0.1575, "step": 43090 }, { "epoch": 0.8450980392156863, "grad_norm": 2.2660701274871826, "learning_rate": 6.661926886465185e-07, "loss": 0.1506, "step": 43100 }, { "epoch": 0.8452941176470589, "grad_norm": 2.6764426231384277, "learning_rate": 6.660397917213048e-07, "loss": 0.1235, "step": 43110 }, { "epoch": 0.8454901960784313, "grad_norm": 3.3483848571777344, "learning_rate": 6.658868773433209e-07, "loss": 0.17, "step": 43120 }, { "epoch": 0.8456862745098039, "grad_norm": 1.870694637298584, "learning_rate": 6.6573394552864e-07, "loss": 0.1274, "step": 43130 }, { "epoch": 0.8458823529411764, "grad_norm": 1.8462951183319092, "learning_rate": 6.655809962933371e-07, "loss": 0.153, "step": 43140 }, { "epoch": 0.846078431372549, "grad_norm": 3.4941091537475586, "learning_rate": 6.654280296534891e-07, "loss": 0.193, "step": 43150 }, { "epoch": 0.8462745098039216, "grad_norm": 4.555943012237549, "learning_rate": 6.652750456251746e-07, "loss": 0.142, "step": 43160 }, { "epoch": 0.8464705882352941, "grad_norm": 2.386890172958374, "learning_rate": 6.651220442244738e-07, "loss": 0.139, "step": 43170 }, { "epoch": 0.8466666666666667, "grad_norm": 1.6819607019424438, "learning_rate": 6.649690254674694e-07, "loss": 0.167, "step": 43180 }, { "epoch": 0.8468627450980392, "grad_norm": 2.3569343090057373, "learning_rate": 6.648159893702451e-07, "loss": 0.1611, "step": 43190 }, { "epoch": 0.8470588235294118, "grad_norm": 2.624958038330078, "learning_rate": 6.646629359488872e-07, "loss": 0.1609, "step": 43200 }, { "epoch": 0.8472549019607843, "grad_norm": 1.265662431716919, "learning_rate": 6.645098652194829e-07, "loss": 0.148, "step": 43210 }, { "epoch": 0.8474509803921568, "grad_norm": 2.2694268226623535, "learning_rate": 6.643567771981224e-07, "loss": 0.1411, "step": 43220 }, { "epoch": 0.8476470588235294, "grad_norm": 2.417433738708496, "learning_rate": 6.642036719008966e-07, "loss": 0.1713, "step": 43230 }, { "epoch": 0.847843137254902, "grad_norm": 3.045372247695923, "learning_rate": 6.640505493438992e-07, "loss": 0.1732, "step": 43240 }, { "epoch": 0.8480392156862745, "grad_norm": 2.4813482761383057, "learning_rate": 6.638974095432246e-07, "loss": 0.1381, "step": 43250 }, { "epoch": 0.8482352941176471, "grad_norm": 4.682314395904541, "learning_rate": 6.637442525149701e-07, "loss": 0.1554, "step": 43260 }, { "epoch": 0.8484313725490196, "grad_norm": 1.6457045078277588, "learning_rate": 6.635910782752342e-07, "loss": 0.1681, "step": 43270 }, { "epoch": 0.8486274509803922, "grad_norm": 3.607520580291748, "learning_rate": 6.634378868401171e-07, "loss": 0.1811, "step": 43280 }, { "epoch": 0.8488235294117648, "grad_norm": 5.316263675689697, "learning_rate": 6.632846782257215e-07, "loss": 0.1415, "step": 43290 }, { "epoch": 0.8490196078431372, "grad_norm": 4.720252990722656, "learning_rate": 6.631314524481513e-07, "loss": 0.1779, "step": 43300 }, { "epoch": 0.8492156862745098, "grad_norm": 2.4043116569519043, "learning_rate": 6.629782095235121e-07, "loss": 0.152, "step": 43310 }, { "epoch": 0.8494117647058823, "grad_norm": 2.50849986076355, "learning_rate": 6.628249494679118e-07, "loss": 0.1358, "step": 43320 }, { "epoch": 0.8496078431372549, "grad_norm": 2.813178539276123, "learning_rate": 6.626716722974598e-07, "loss": 0.1626, "step": 43330 }, { "epoch": 0.8498039215686275, "grad_norm": 3.2428250312805176, "learning_rate": 6.625183780282675e-07, "loss": 0.0922, "step": 43340 }, { "epoch": 0.85, "grad_norm": 1.877694845199585, "learning_rate": 6.623650666764477e-07, "loss": 0.1623, "step": 43350 }, { "epoch": 0.8501960784313726, "grad_norm": 2.855729341506958, "learning_rate": 6.622117382581156e-07, "loss": 0.1408, "step": 43360 }, { "epoch": 0.850392156862745, "grad_norm": 4.913961410522461, "learning_rate": 6.620583927893876e-07, "loss": 0.1742, "step": 43370 }, { "epoch": 0.8505882352941176, "grad_norm": 4.177707195281982, "learning_rate": 6.619050302863823e-07, "loss": 0.1766, "step": 43380 }, { "epoch": 0.8507843137254902, "grad_norm": 1.883953332901001, "learning_rate": 6.617516507652196e-07, "loss": 0.1557, "step": 43390 }, { "epoch": 0.8509803921568627, "grad_norm": 3.2329201698303223, "learning_rate": 6.615982542420222e-07, "loss": 0.1612, "step": 43400 }, { "epoch": 0.8511764705882353, "grad_norm": 4.122861385345459, "learning_rate": 6.614448407329131e-07, "loss": 0.1377, "step": 43410 }, { "epoch": 0.8513725490196078, "grad_norm": 2.2943766117095947, "learning_rate": 6.612914102540184e-07, "loss": 0.1557, "step": 43420 }, { "epoch": 0.8515686274509804, "grad_norm": 2.748389720916748, "learning_rate": 6.611379628214654e-07, "loss": 0.1972, "step": 43430 }, { "epoch": 0.851764705882353, "grad_norm": 2.8770196437835693, "learning_rate": 6.609844984513831e-07, "loss": 0.1984, "step": 43440 }, { "epoch": 0.8519607843137255, "grad_norm": 4.056859016418457, "learning_rate": 6.608310171599027e-07, "loss": 0.1585, "step": 43450 }, { "epoch": 0.852156862745098, "grad_norm": 3.635467767715454, "learning_rate": 6.606775189631568e-07, "loss": 0.1877, "step": 43460 }, { "epoch": 0.8523529411764705, "grad_norm": 2.517413854598999, "learning_rate": 6.605240038772799e-07, "loss": 0.156, "step": 43470 }, { "epoch": 0.8525490196078431, "grad_norm": 2.222588300704956, "learning_rate": 6.603704719184081e-07, "loss": 0.1772, "step": 43480 }, { "epoch": 0.8527450980392157, "grad_norm": 2.5785248279571533, "learning_rate": 6.602169231026798e-07, "loss": 0.1598, "step": 43490 }, { "epoch": 0.8529411764705882, "grad_norm": 1.8299329280853271, "learning_rate": 6.600633574462344e-07, "loss": 0.174, "step": 43500 }, { "epoch": 0.8531372549019608, "grad_norm": 4.875173091888428, "learning_rate": 6.59909774965214e-07, "loss": 0.153, "step": 43510 }, { "epoch": 0.8533333333333334, "grad_norm": 2.761164903640747, "learning_rate": 6.597561756757615e-07, "loss": 0.1667, "step": 43520 }, { "epoch": 0.8535294117647059, "grad_norm": 2.517491102218628, "learning_rate": 6.596025595940223e-07, "loss": 0.1353, "step": 43530 }, { "epoch": 0.8537254901960785, "grad_norm": 2.5735585689544678, "learning_rate": 6.59448926736143e-07, "loss": 0.122, "step": 43540 }, { "epoch": 0.8539215686274509, "grad_norm": 1.8160274028778076, "learning_rate": 6.592952771182726e-07, "loss": 0.1423, "step": 43550 }, { "epoch": 0.8541176470588235, "grad_norm": 2.0266647338867188, "learning_rate": 6.591416107565613e-07, "loss": 0.1511, "step": 43560 }, { "epoch": 0.8543137254901961, "grad_norm": 3.343400239944458, "learning_rate": 6.589879276671615e-07, "loss": 0.1299, "step": 43570 }, { "epoch": 0.8545098039215686, "grad_norm": 2.766038656234741, "learning_rate": 6.588342278662266e-07, "loss": 0.1549, "step": 43580 }, { "epoch": 0.8547058823529412, "grad_norm": 2.857142448425293, "learning_rate": 6.58680511369913e-07, "loss": 0.1639, "step": 43590 }, { "epoch": 0.8549019607843137, "grad_norm": 2.1900124549865723, "learning_rate": 6.585267781943775e-07, "loss": 0.1225, "step": 43600 }, { "epoch": 0.8550980392156863, "grad_norm": 3.5723423957824707, "learning_rate": 6.583730283557798e-07, "loss": 0.1922, "step": 43610 }, { "epoch": 0.8552941176470589, "grad_norm": 2.8300602436065674, "learning_rate": 6.582192618702806e-07, "loss": 0.1327, "step": 43620 }, { "epoch": 0.8554901960784314, "grad_norm": 2.1606812477111816, "learning_rate": 6.580654787540427e-07, "loss": 0.1429, "step": 43630 }, { "epoch": 0.855686274509804, "grad_norm": 3.6750237941741943, "learning_rate": 6.579116790232304e-07, "loss": 0.1488, "step": 43640 }, { "epoch": 0.8558823529411764, "grad_norm": 3.836869955062866, "learning_rate": 6.5775786269401e-07, "loss": 0.1317, "step": 43650 }, { "epoch": 0.856078431372549, "grad_norm": 1.4615529775619507, "learning_rate": 6.576040297825494e-07, "loss": 0.1458, "step": 43660 }, { "epoch": 0.8562745098039216, "grad_norm": 3.529414176940918, "learning_rate": 6.574501803050184e-07, "loss": 0.1677, "step": 43670 }, { "epoch": 0.8564705882352941, "grad_norm": 3.9055604934692383, "learning_rate": 6.572963142775882e-07, "loss": 0.1639, "step": 43680 }, { "epoch": 0.8566666666666667, "grad_norm": 2.3720133304595947, "learning_rate": 6.57142431716432e-07, "loss": 0.1493, "step": 43690 }, { "epoch": 0.8568627450980392, "grad_norm": 4.515248775482178, "learning_rate": 6.56988532637725e-07, "loss": 0.1864, "step": 43700 }, { "epoch": 0.8570588235294118, "grad_norm": 2.2411744594573975, "learning_rate": 6.568346170576436e-07, "loss": 0.1749, "step": 43710 }, { "epoch": 0.8572549019607844, "grad_norm": 3.4544191360473633, "learning_rate": 6.566806849923659e-07, "loss": 0.1403, "step": 43720 }, { "epoch": 0.8574509803921568, "grad_norm": 1.9827306270599365, "learning_rate": 6.565267364580726e-07, "loss": 0.1371, "step": 43730 }, { "epoch": 0.8576470588235294, "grad_norm": 1.7256622314453125, "learning_rate": 6.563727714709451e-07, "loss": 0.1592, "step": 43740 }, { "epoch": 0.8578431372549019, "grad_norm": 2.404794931411743, "learning_rate": 6.562187900471673e-07, "loss": 0.143, "step": 43750 }, { "epoch": 0.8580392156862745, "grad_norm": 1.9917232990264893, "learning_rate": 6.560647922029241e-07, "loss": 0.1466, "step": 43760 }, { "epoch": 0.8582352941176471, "grad_norm": 3.5054800510406494, "learning_rate": 6.559107779544025e-07, "loss": 0.156, "step": 43770 }, { "epoch": 0.8584313725490196, "grad_norm": 2.881207227706909, "learning_rate": 6.557567473177918e-07, "loss": 0.1838, "step": 43780 }, { "epoch": 0.8586274509803922, "grad_norm": 1.260076642036438, "learning_rate": 6.556027003092819e-07, "loss": 0.1566, "step": 43790 }, { "epoch": 0.8588235294117647, "grad_norm": 7.020211219787598, "learning_rate": 6.554486369450651e-07, "loss": 0.1381, "step": 43800 }, { "epoch": 0.8590196078431372, "grad_norm": 2.936131238937378, "learning_rate": 6.552945572413356e-07, "loss": 0.1393, "step": 43810 }, { "epoch": 0.8592156862745098, "grad_norm": 1.8252837657928467, "learning_rate": 6.551404612142888e-07, "loss": 0.1358, "step": 43820 }, { "epoch": 0.8594117647058823, "grad_norm": 2.770134210586548, "learning_rate": 6.54986348880122e-07, "loss": 0.1378, "step": 43830 }, { "epoch": 0.8596078431372549, "grad_norm": 6.13914155960083, "learning_rate": 6.548322202550343e-07, "loss": 0.1762, "step": 43840 }, { "epoch": 0.8598039215686275, "grad_norm": 5.507809162139893, "learning_rate": 6.546780753552264e-07, "loss": 0.1569, "step": 43850 }, { "epoch": 0.86, "grad_norm": 3.8834619522094727, "learning_rate": 6.545239141969009e-07, "loss": 0.1657, "step": 43860 }, { "epoch": 0.8601960784313726, "grad_norm": 2.1885204315185547, "learning_rate": 6.543697367962618e-07, "loss": 0.1438, "step": 43870 }, { "epoch": 0.8603921568627451, "grad_norm": 2.3570139408111572, "learning_rate": 6.542155431695153e-07, "loss": 0.1514, "step": 43880 }, { "epoch": 0.8605882352941177, "grad_norm": 3.050293445587158, "learning_rate": 6.540613333328688e-07, "loss": 0.1729, "step": 43890 }, { "epoch": 0.8607843137254902, "grad_norm": 4.027644634246826, "learning_rate": 6.539071073025315e-07, "loss": 0.1617, "step": 43900 }, { "epoch": 0.8609803921568627, "grad_norm": 3.3213791847229004, "learning_rate": 6.537528650947146e-07, "loss": 0.1534, "step": 43910 }, { "epoch": 0.8611764705882353, "grad_norm": 4.383220672607422, "learning_rate": 6.535986067256307e-07, "loss": 0.1829, "step": 43920 }, { "epoch": 0.8613725490196078, "grad_norm": 4.547692775726318, "learning_rate": 6.534443322114943e-07, "loss": 0.1883, "step": 43930 }, { "epoch": 0.8615686274509804, "grad_norm": 3.769996404647827, "learning_rate": 6.532900415685215e-07, "loss": 0.1454, "step": 43940 }, { "epoch": 0.861764705882353, "grad_norm": 2.840496301651001, "learning_rate": 6.531357348129296e-07, "loss": 0.1354, "step": 43950 }, { "epoch": 0.8619607843137255, "grad_norm": 2.235978364944458, "learning_rate": 6.529814119609387e-07, "loss": 0.1344, "step": 43960 }, { "epoch": 0.8621568627450981, "grad_norm": 2.8368630409240723, "learning_rate": 6.5282707302877e-07, "loss": 0.1279, "step": 43970 }, { "epoch": 0.8623529411764705, "grad_norm": 2.301011562347412, "learning_rate": 6.526727180326459e-07, "loss": 0.1596, "step": 43980 }, { "epoch": 0.8625490196078431, "grad_norm": 2.165631055831909, "learning_rate": 6.525183469887912e-07, "loss": 0.1338, "step": 43990 }, { "epoch": 0.8627450980392157, "grad_norm": 5.72589111328125, "learning_rate": 6.523639599134323e-07, "loss": 0.1736, "step": 44000 }, { "epoch": 0.8629411764705882, "grad_norm": 2.247086763381958, "learning_rate": 6.522095568227969e-07, "loss": 0.1331, "step": 44010 }, { "epoch": 0.8631372549019608, "grad_norm": 2.9604909420013428, "learning_rate": 6.520551377331147e-07, "loss": 0.1641, "step": 44020 }, { "epoch": 0.8633333333333333, "grad_norm": 3.6201677322387695, "learning_rate": 6.51900702660617e-07, "loss": 0.1542, "step": 44030 }, { "epoch": 0.8635294117647059, "grad_norm": 2.571051836013794, "learning_rate": 6.517462516215366e-07, "loss": 0.176, "step": 44040 }, { "epoch": 0.8637254901960785, "grad_norm": 2.489753007888794, "learning_rate": 6.515917846321086e-07, "loss": 0.1771, "step": 44050 }, { "epoch": 0.863921568627451, "grad_norm": 3.094998359680176, "learning_rate": 6.514373017085689e-07, "loss": 0.1434, "step": 44060 }, { "epoch": 0.8641176470588235, "grad_norm": 3.4152591228485107, "learning_rate": 6.512828028671557e-07, "loss": 0.1248, "step": 44070 }, { "epoch": 0.864313725490196, "grad_norm": 2.6584081649780273, "learning_rate": 6.511282881241087e-07, "loss": 0.1264, "step": 44080 }, { "epoch": 0.8645098039215686, "grad_norm": 2.2045907974243164, "learning_rate": 6.509737574956689e-07, "loss": 0.1496, "step": 44090 }, { "epoch": 0.8647058823529412, "grad_norm": 2.634483575820923, "learning_rate": 6.508192109980799e-07, "loss": 0.1715, "step": 44100 }, { "epoch": 0.8649019607843137, "grad_norm": 3.9142017364501953, "learning_rate": 6.50664648647586e-07, "loss": 0.1471, "step": 44110 }, { "epoch": 0.8650980392156863, "grad_norm": 1.7698774337768555, "learning_rate": 6.505100704604336e-07, "loss": 0.1414, "step": 44120 }, { "epoch": 0.8652941176470588, "grad_norm": 1.6202900409698486, "learning_rate": 6.503554764528707e-07, "loss": 0.1606, "step": 44130 }, { "epoch": 0.8654901960784314, "grad_norm": 2.1367130279541016, "learning_rate": 6.502008666411472e-07, "loss": 0.1629, "step": 44140 }, { "epoch": 0.865686274509804, "grad_norm": 2.8449442386627197, "learning_rate": 6.500462410415142e-07, "loss": 0.137, "step": 44150 }, { "epoch": 0.8658823529411764, "grad_norm": 2.0653252601623535, "learning_rate": 6.498915996702248e-07, "loss": 0.1596, "step": 44160 }, { "epoch": 0.866078431372549, "grad_norm": 1.782296061515808, "learning_rate": 6.497369425435335e-07, "loss": 0.165, "step": 44170 }, { "epoch": 0.8662745098039216, "grad_norm": 3.2605819702148438, "learning_rate": 6.495822696776969e-07, "loss": 0.1474, "step": 44180 }, { "epoch": 0.8664705882352941, "grad_norm": 2.7361607551574707, "learning_rate": 6.494275810889728e-07, "loss": 0.1285, "step": 44190 }, { "epoch": 0.8666666666666667, "grad_norm": 4.925535202026367, "learning_rate": 6.49272876793621e-07, "loss": 0.1439, "step": 44200 }, { "epoch": 0.8668627450980392, "grad_norm": 1.6056106090545654, "learning_rate": 6.491181568079025e-07, "loss": 0.1484, "step": 44210 }, { "epoch": 0.8670588235294118, "grad_norm": 2.6743509769439697, "learning_rate": 6.489634211480802e-07, "loss": 0.1632, "step": 44220 }, { "epoch": 0.8672549019607844, "grad_norm": 14.670259475708008, "learning_rate": 6.48808669830419e-07, "loss": 0.1597, "step": 44230 }, { "epoch": 0.8674509803921568, "grad_norm": 1.4582128524780273, "learning_rate": 6.486539028711849e-07, "loss": 0.1407, "step": 44240 }, { "epoch": 0.8676470588235294, "grad_norm": 2.5133867263793945, "learning_rate": 6.484991202866459e-07, "loss": 0.1702, "step": 44250 }, { "epoch": 0.8678431372549019, "grad_norm": 3.4178788661956787, "learning_rate": 6.483443220930712e-07, "loss": 0.1342, "step": 44260 }, { "epoch": 0.8680392156862745, "grad_norm": 2.185673713684082, "learning_rate": 6.481895083067324e-07, "loss": 0.1439, "step": 44270 }, { "epoch": 0.8682352941176471, "grad_norm": 1.6730366945266724, "learning_rate": 6.480346789439017e-07, "loss": 0.132, "step": 44280 }, { "epoch": 0.8684313725490196, "grad_norm": 2.381977081298828, "learning_rate": 6.478798340208541e-07, "loss": 0.1506, "step": 44290 }, { "epoch": 0.8686274509803922, "grad_norm": 3.666130781173706, "learning_rate": 6.477249735538652e-07, "loss": 0.1934, "step": 44300 }, { "epoch": 0.8688235294117647, "grad_norm": 2.1804919242858887, "learning_rate": 6.47570097559213e-07, "loss": 0.1862, "step": 44310 }, { "epoch": 0.8690196078431373, "grad_norm": 3.9657535552978516, "learning_rate": 6.474152060531768e-07, "loss": 0.1766, "step": 44320 }, { "epoch": 0.8692156862745098, "grad_norm": 2.155651569366455, "learning_rate": 6.472602990520373e-07, "loss": 0.1621, "step": 44330 }, { "epoch": 0.8694117647058823, "grad_norm": 3.6763710975646973, "learning_rate": 6.471053765720773e-07, "loss": 0.1502, "step": 44340 }, { "epoch": 0.8696078431372549, "grad_norm": 6.065370082855225, "learning_rate": 6.46950438629581e-07, "loss": 0.157, "step": 44350 }, { "epoch": 0.8698039215686274, "grad_norm": 3.180525779724121, "learning_rate": 6.467954852408342e-07, "loss": 0.1664, "step": 44360 }, { "epoch": 0.87, "grad_norm": 1.9417580366134644, "learning_rate": 6.466405164221243e-07, "loss": 0.1674, "step": 44370 }, { "epoch": 0.8701960784313726, "grad_norm": 2.7400341033935547, "learning_rate": 6.464855321897404e-07, "loss": 0.1564, "step": 44380 }, { "epoch": 0.8703921568627451, "grad_norm": 1.997103214263916, "learning_rate": 6.46330532559973e-07, "loss": 0.1578, "step": 44390 }, { "epoch": 0.8705882352941177, "grad_norm": 3.999185085296631, "learning_rate": 6.46175517549115e-07, "loss": 0.2027, "step": 44400 }, { "epoch": 0.8707843137254901, "grad_norm": 4.421076774597168, "learning_rate": 6.460204871734598e-07, "loss": 0.1371, "step": 44410 }, { "epoch": 0.8709803921568627, "grad_norm": 3.0907318592071533, "learning_rate": 6.458654414493032e-07, "loss": 0.1618, "step": 44420 }, { "epoch": 0.8711764705882353, "grad_norm": 3.098400831222534, "learning_rate": 6.457103803929423e-07, "loss": 0.155, "step": 44430 }, { "epoch": 0.8713725490196078, "grad_norm": 3.3305602073669434, "learning_rate": 6.455553040206756e-07, "loss": 0.1346, "step": 44440 }, { "epoch": 0.8715686274509804, "grad_norm": 2.5970990657806396, "learning_rate": 6.454002123488041e-07, "loss": 0.1551, "step": 44450 }, { "epoch": 0.8717647058823529, "grad_norm": 4.479038238525391, "learning_rate": 6.45245105393629e-07, "loss": 0.1604, "step": 44460 }, { "epoch": 0.8719607843137255, "grad_norm": 2.1414735317230225, "learning_rate": 6.450899831714546e-07, "loss": 0.1417, "step": 44470 }, { "epoch": 0.8721568627450981, "grad_norm": 2.4415817260742188, "learning_rate": 6.449348456985857e-07, "loss": 0.1321, "step": 44480 }, { "epoch": 0.8723529411764706, "grad_norm": 3.118971586227417, "learning_rate": 6.447796929913294e-07, "loss": 0.1416, "step": 44490 }, { "epoch": 0.8725490196078431, "grad_norm": 3.656421422958374, "learning_rate": 6.446245250659938e-07, "loss": 0.1477, "step": 44500 }, { "epoch": 0.8727450980392157, "grad_norm": 3.775343418121338, "learning_rate": 6.444693419388891e-07, "loss": 0.1625, "step": 44510 }, { "epoch": 0.8729411764705882, "grad_norm": 2.145615577697754, "learning_rate": 6.443141436263268e-07, "loss": 0.1673, "step": 44520 }, { "epoch": 0.8731372549019608, "grad_norm": 5.22335147857666, "learning_rate": 6.441589301446202e-07, "loss": 0.1889, "step": 44530 }, { "epoch": 0.8733333333333333, "grad_norm": 3.5353844165802, "learning_rate": 6.440037015100841e-07, "loss": 0.1533, "step": 44540 }, { "epoch": 0.8735294117647059, "grad_norm": 4.73114538192749, "learning_rate": 6.438484577390349e-07, "loss": 0.1722, "step": 44550 }, { "epoch": 0.8737254901960785, "grad_norm": 3.6590938568115234, "learning_rate": 6.436931988477904e-07, "loss": 0.1598, "step": 44560 }, { "epoch": 0.873921568627451, "grad_norm": 1.6233782768249512, "learning_rate": 6.435379248526702e-07, "loss": 0.1405, "step": 44570 }, { "epoch": 0.8741176470588236, "grad_norm": 3.0816330909729004, "learning_rate": 6.433826357699955e-07, "loss": 0.142, "step": 44580 }, { "epoch": 0.874313725490196, "grad_norm": 7.657105922698975, "learning_rate": 6.432273316160893e-07, "loss": 0.1921, "step": 44590 }, { "epoch": 0.8745098039215686, "grad_norm": 2.9515018463134766, "learning_rate": 6.430720124072756e-07, "loss": 0.1761, "step": 44600 }, { "epoch": 0.8747058823529412, "grad_norm": 2.655332326889038, "learning_rate": 6.429166781598804e-07, "loss": 0.1455, "step": 44610 }, { "epoch": 0.8749019607843137, "grad_norm": 3.635849714279175, "learning_rate": 6.427613288902312e-07, "loss": 0.1943, "step": 44620 }, { "epoch": 0.8750980392156863, "grad_norm": 3.670443534851074, "learning_rate": 6.426059646146571e-07, "loss": 0.1771, "step": 44630 }, { "epoch": 0.8752941176470588, "grad_norm": 3.1156349182128906, "learning_rate": 6.424505853494889e-07, "loss": 0.1366, "step": 44640 }, { "epoch": 0.8754901960784314, "grad_norm": 2.1482009887695312, "learning_rate": 6.422951911110583e-07, "loss": 0.1537, "step": 44650 }, { "epoch": 0.875686274509804, "grad_norm": 3.409637212753296, "learning_rate": 6.421397819156995e-07, "loss": 0.1669, "step": 44660 }, { "epoch": 0.8758823529411764, "grad_norm": 6.16284704208374, "learning_rate": 6.41984357779748e-07, "loss": 0.1656, "step": 44670 }, { "epoch": 0.876078431372549, "grad_norm": 1.8825180530548096, "learning_rate": 6.418289187195403e-07, "loss": 0.1452, "step": 44680 }, { "epoch": 0.8762745098039215, "grad_norm": 4.391015529632568, "learning_rate": 6.416734647514153e-07, "loss": 0.1427, "step": 44690 }, { "epoch": 0.8764705882352941, "grad_norm": 1.5704234838485718, "learning_rate": 6.41517995891713e-07, "loss": 0.1692, "step": 44700 }, { "epoch": 0.8766666666666667, "grad_norm": 38.177677154541016, "learning_rate": 6.413625121567749e-07, "loss": 0.1526, "step": 44710 }, { "epoch": 0.8768627450980392, "grad_norm": 2.1721231937408447, "learning_rate": 6.412070135629442e-07, "loss": 0.1741, "step": 44720 }, { "epoch": 0.8770588235294118, "grad_norm": 4.530058860778809, "learning_rate": 6.410515001265657e-07, "loss": 0.1423, "step": 44730 }, { "epoch": 0.8772549019607843, "grad_norm": 2.0925326347351074, "learning_rate": 6.408959718639859e-07, "loss": 0.1634, "step": 44740 }, { "epoch": 0.8774509803921569, "grad_norm": 1.81520414352417, "learning_rate": 6.407404287915523e-07, "loss": 0.1431, "step": 44750 }, { "epoch": 0.8776470588235294, "grad_norm": 4.663355350494385, "learning_rate": 6.405848709256146e-07, "loss": 0.1505, "step": 44760 }, { "epoch": 0.8778431372549019, "grad_norm": 2.541722059249878, "learning_rate": 6.404292982825238e-07, "loss": 0.1371, "step": 44770 }, { "epoch": 0.8780392156862745, "grad_norm": 3.1247475147247314, "learning_rate": 6.402737108786325e-07, "loss": 0.1506, "step": 44780 }, { "epoch": 0.8782352941176471, "grad_norm": 2.1449027061462402, "learning_rate": 6.401181087302947e-07, "loss": 0.1436, "step": 44790 }, { "epoch": 0.8784313725490196, "grad_norm": 2.606501340866089, "learning_rate": 6.399624918538661e-07, "loss": 0.1885, "step": 44800 }, { "epoch": 0.8786274509803922, "grad_norm": 2.2223942279815674, "learning_rate": 6.398068602657038e-07, "loss": 0.1735, "step": 44810 }, { "epoch": 0.8788235294117647, "grad_norm": 2.061694622039795, "learning_rate": 6.396512139821665e-07, "loss": 0.1744, "step": 44820 }, { "epoch": 0.8790196078431373, "grad_norm": 2.7366156578063965, "learning_rate": 6.394955530196147e-07, "loss": 0.1279, "step": 44830 }, { "epoch": 0.8792156862745099, "grad_norm": 3.6365456581115723, "learning_rate": 6.3933987739441e-07, "loss": 0.1413, "step": 44840 }, { "epoch": 0.8794117647058823, "grad_norm": 2.1453914642333984, "learning_rate": 6.391841871229159e-07, "loss": 0.1697, "step": 44850 }, { "epoch": 0.8796078431372549, "grad_norm": 2.399836540222168, "learning_rate": 6.390284822214974e-07, "loss": 0.1656, "step": 44860 }, { "epoch": 0.8798039215686274, "grad_norm": 2.14163875579834, "learning_rate": 6.388727627065208e-07, "loss": 0.1473, "step": 44870 }, { "epoch": 0.88, "grad_norm": 2.2817635536193848, "learning_rate": 6.387170285943542e-07, "loss": 0.1625, "step": 44880 }, { "epoch": 0.8801960784313726, "grad_norm": 2.8016178607940674, "learning_rate": 6.385612799013671e-07, "loss": 0.1686, "step": 44890 }, { "epoch": 0.8803921568627451, "grad_norm": 1.9026844501495361, "learning_rate": 6.384055166439304e-07, "loss": 0.1452, "step": 44900 }, { "epoch": 0.8805882352941177, "grad_norm": 4.072925567626953, "learning_rate": 6.382497388384168e-07, "loss": 0.1686, "step": 44910 }, { "epoch": 0.8807843137254902, "grad_norm": 2.324448585510254, "learning_rate": 6.380939465012004e-07, "loss": 0.1375, "step": 44920 }, { "epoch": 0.8809803921568627, "grad_norm": 2.6050362586975098, "learning_rate": 6.379381396486567e-07, "loss": 0.1624, "step": 44930 }, { "epoch": 0.8811764705882353, "grad_norm": 2.959063768386841, "learning_rate": 6.377823182971633e-07, "loss": 0.1524, "step": 44940 }, { "epoch": 0.8813725490196078, "grad_norm": 3.8082098960876465, "learning_rate": 6.376264824630984e-07, "loss": 0.1787, "step": 44950 }, { "epoch": 0.8815686274509804, "grad_norm": 2.3365283012390137, "learning_rate": 6.374706321628424e-07, "loss": 0.1677, "step": 44960 }, { "epoch": 0.8817647058823529, "grad_norm": 2.0571584701538086, "learning_rate": 6.373147674127773e-07, "loss": 0.1377, "step": 44970 }, { "epoch": 0.8819607843137255, "grad_norm": 2.2340645790100098, "learning_rate": 6.371588882292857e-07, "loss": 0.1524, "step": 44980 }, { "epoch": 0.8821568627450981, "grad_norm": 2.155564308166504, "learning_rate": 6.370029946287531e-07, "loss": 0.1642, "step": 44990 }, { "epoch": 0.8823529411764706, "grad_norm": 2.153233528137207, "learning_rate": 6.368470866275653e-07, "loss": 0.13, "step": 45000 }, { "epoch": 0.8825490196078432, "grad_norm": 4.726171493530273, "learning_rate": 6.366911642421103e-07, "loss": 0.1619, "step": 45010 }, { "epoch": 0.8827450980392156, "grad_norm": 1.8306622505187988, "learning_rate": 6.365352274887773e-07, "loss": 0.1503, "step": 45020 }, { "epoch": 0.8829411764705882, "grad_norm": 2.329392671585083, "learning_rate": 6.363792763839571e-07, "loss": 0.141, "step": 45030 }, { "epoch": 0.8831372549019608, "grad_norm": 3.1640586853027344, "learning_rate": 6.362233109440423e-07, "loss": 0.15, "step": 45040 }, { "epoch": 0.8833333333333333, "grad_norm": 3.9256365299224854, "learning_rate": 6.360673311854264e-07, "loss": 0.1575, "step": 45050 }, { "epoch": 0.8835294117647059, "grad_norm": 2.4078257083892822, "learning_rate": 6.359113371245049e-07, "loss": 0.1834, "step": 45060 }, { "epoch": 0.8837254901960784, "grad_norm": 53.8978271484375, "learning_rate": 6.357553287776747e-07, "loss": 0.13, "step": 45070 }, { "epoch": 0.883921568627451, "grad_norm": 3.1967294216156006, "learning_rate": 6.355993061613341e-07, "loss": 0.1438, "step": 45080 }, { "epoch": 0.8841176470588236, "grad_norm": 3.7506229877471924, "learning_rate": 6.354432692918827e-07, "loss": 0.1455, "step": 45090 }, { "epoch": 0.884313725490196, "grad_norm": 2.533104181289673, "learning_rate": 6.352872181857223e-07, "loss": 0.1527, "step": 45100 }, { "epoch": 0.8845098039215686, "grad_norm": 2.3391222953796387, "learning_rate": 6.351311528592551e-07, "loss": 0.1447, "step": 45110 }, { "epoch": 0.8847058823529412, "grad_norm": 3.081103801727295, "learning_rate": 6.349750733288861e-07, "loss": 0.1385, "step": 45120 }, { "epoch": 0.8849019607843137, "grad_norm": 2.807607412338257, "learning_rate": 6.348189796110209e-07, "loss": 0.1522, "step": 45130 }, { "epoch": 0.8850980392156863, "grad_norm": 1.9264566898345947, "learning_rate": 6.346628717220668e-07, "loss": 0.1497, "step": 45140 }, { "epoch": 0.8852941176470588, "grad_norm": 2.8402135372161865, "learning_rate": 6.345067496784326e-07, "loss": 0.1396, "step": 45150 }, { "epoch": 0.8854901960784314, "grad_norm": 1.9780820608139038, "learning_rate": 6.343506134965284e-07, "loss": 0.1477, "step": 45160 }, { "epoch": 0.885686274509804, "grad_norm": 1.5024033784866333, "learning_rate": 6.341944631927661e-07, "loss": 0.1544, "step": 45170 }, { "epoch": 0.8858823529411765, "grad_norm": 2.3844923973083496, "learning_rate": 6.340382987835593e-07, "loss": 0.17, "step": 45180 }, { "epoch": 0.886078431372549, "grad_norm": 4.25961971282959, "learning_rate": 6.338821202853224e-07, "loss": 0.1737, "step": 45190 }, { "epoch": 0.8862745098039215, "grad_norm": 1.3893561363220215, "learning_rate": 6.337259277144714e-07, "loss": 0.1785, "step": 45200 }, { "epoch": 0.8864705882352941, "grad_norm": 2.8568968772888184, "learning_rate": 6.335697210874247e-07, "loss": 0.1618, "step": 45210 }, { "epoch": 0.8866666666666667, "grad_norm": 2.265130043029785, "learning_rate": 6.33413500420601e-07, "loss": 0.147, "step": 45220 }, { "epoch": 0.8868627450980392, "grad_norm": 1.677486538887024, "learning_rate": 6.332572657304209e-07, "loss": 0.1492, "step": 45230 }, { "epoch": 0.8870588235294118, "grad_norm": 2.8699426651000977, "learning_rate": 6.331010170333069e-07, "loss": 0.1566, "step": 45240 }, { "epoch": 0.8872549019607843, "grad_norm": 2.2644283771514893, "learning_rate": 6.329447543456824e-07, "loss": 0.1639, "step": 45250 }, { "epoch": 0.8874509803921569, "grad_norm": 2.323563575744629, "learning_rate": 6.327884776839725e-07, "loss": 0.1488, "step": 45260 }, { "epoch": 0.8876470588235295, "grad_norm": 2.6769747734069824, "learning_rate": 6.326321870646036e-07, "loss": 0.1598, "step": 45270 }, { "epoch": 0.8878431372549019, "grad_norm": 3.774618148803711, "learning_rate": 6.324758825040039e-07, "loss": 0.1559, "step": 45280 }, { "epoch": 0.8880392156862745, "grad_norm": 1.2156548500061035, "learning_rate": 6.323195640186028e-07, "loss": 0.1264, "step": 45290 }, { "epoch": 0.888235294117647, "grad_norm": 2.7336790561676025, "learning_rate": 6.321632316248313e-07, "loss": 0.139, "step": 45300 }, { "epoch": 0.8884313725490196, "grad_norm": 2.292886257171631, "learning_rate": 6.320068853391219e-07, "loss": 0.1399, "step": 45310 }, { "epoch": 0.8886274509803922, "grad_norm": 1.6405913829803467, "learning_rate": 6.318505251779083e-07, "loss": 0.146, "step": 45320 }, { "epoch": 0.8888235294117647, "grad_norm": 2.2593994140625, "learning_rate": 6.316941511576258e-07, "loss": 0.1415, "step": 45330 }, { "epoch": 0.8890196078431373, "grad_norm": 3.082005739212036, "learning_rate": 6.315377632947114e-07, "loss": 0.16, "step": 45340 }, { "epoch": 0.8892156862745098, "grad_norm": 2.882730007171631, "learning_rate": 6.313813616056032e-07, "loss": 0.1373, "step": 45350 }, { "epoch": 0.8894117647058823, "grad_norm": 4.161201477050781, "learning_rate": 6.312249461067407e-07, "loss": 0.1412, "step": 45360 }, { "epoch": 0.8896078431372549, "grad_norm": 3.205078363418579, "learning_rate": 6.310685168145653e-07, "loss": 0.1422, "step": 45370 }, { "epoch": 0.8898039215686274, "grad_norm": 3.209197521209717, "learning_rate": 6.309120737455195e-07, "loss": 0.1384, "step": 45380 }, { "epoch": 0.89, "grad_norm": 1.5614092350006104, "learning_rate": 6.307556169160473e-07, "loss": 0.1525, "step": 45390 }, { "epoch": 0.8901960784313725, "grad_norm": 1.7918412685394287, "learning_rate": 6.305991463425945e-07, "loss": 0.1353, "step": 45400 }, { "epoch": 0.8903921568627451, "grad_norm": 8.05932903289795, "learning_rate": 6.304426620416074e-07, "loss": 0.1583, "step": 45410 }, { "epoch": 0.8905882352941177, "grad_norm": 3.115476369857788, "learning_rate": 6.302861640295351e-07, "loss": 0.1439, "step": 45420 }, { "epoch": 0.8907843137254902, "grad_norm": 4.800323486328125, "learning_rate": 6.30129652322827e-07, "loss": 0.1774, "step": 45430 }, { "epoch": 0.8909803921568628, "grad_norm": 2.686187982559204, "learning_rate": 6.299731269379342e-07, "loss": 0.1431, "step": 45440 }, { "epoch": 0.8911764705882353, "grad_norm": 4.0217156410217285, "learning_rate": 6.298165878913099e-07, "loss": 0.15, "step": 45450 }, { "epoch": 0.8913725490196078, "grad_norm": 3.070047616958618, "learning_rate": 6.296600351994077e-07, "loss": 0.1164, "step": 45460 }, { "epoch": 0.8915686274509804, "grad_norm": 1.4328978061676025, "learning_rate": 6.295034688786833e-07, "loss": 0.1132, "step": 45470 }, { "epoch": 0.8917647058823529, "grad_norm": 1.9517930746078491, "learning_rate": 6.293468889455942e-07, "loss": 0.1603, "step": 45480 }, { "epoch": 0.8919607843137255, "grad_norm": 5.20967960357666, "learning_rate": 6.291902954165981e-07, "loss": 0.1412, "step": 45490 }, { "epoch": 0.8921568627450981, "grad_norm": 1.330086350440979, "learning_rate": 6.290336883081554e-07, "loss": 0.13, "step": 45500 }, { "epoch": 0.8923529411764706, "grad_norm": 2.130800247192383, "learning_rate": 6.28877067636727e-07, "loss": 0.1717, "step": 45510 }, { "epoch": 0.8925490196078432, "grad_norm": 2.312049627304077, "learning_rate": 6.287204334187756e-07, "loss": 0.1567, "step": 45520 }, { "epoch": 0.8927450980392156, "grad_norm": 3.1115617752075195, "learning_rate": 6.28563785670766e-07, "loss": 0.1659, "step": 45530 }, { "epoch": 0.8929411764705882, "grad_norm": 5.0252838134765625, "learning_rate": 6.284071244091628e-07, "loss": 0.1556, "step": 45540 }, { "epoch": 0.8931372549019608, "grad_norm": 2.5099387168884277, "learning_rate": 6.282504496504335e-07, "loss": 0.1412, "step": 45550 }, { "epoch": 0.8933333333333333, "grad_norm": 2.055234670639038, "learning_rate": 6.280937614110464e-07, "loss": 0.1589, "step": 45560 }, { "epoch": 0.8935294117647059, "grad_norm": 1.8585669994354248, "learning_rate": 6.279370597074715e-07, "loss": 0.1648, "step": 45570 }, { "epoch": 0.8937254901960784, "grad_norm": 3.045457601547241, "learning_rate": 6.277803445561796e-07, "loss": 0.1386, "step": 45580 }, { "epoch": 0.893921568627451, "grad_norm": 3.1706900596618652, "learning_rate": 6.276236159736439e-07, "loss": 0.1576, "step": 45590 }, { "epoch": 0.8941176470588236, "grad_norm": 2.085789203643799, "learning_rate": 6.274668739763381e-07, "loss": 0.1682, "step": 45600 }, { "epoch": 0.894313725490196, "grad_norm": 2.068563222885132, "learning_rate": 6.273101185807375e-07, "loss": 0.1867, "step": 45610 }, { "epoch": 0.8945098039215686, "grad_norm": 3.388179063796997, "learning_rate": 6.271533498033192e-07, "loss": 0.1563, "step": 45620 }, { "epoch": 0.8947058823529411, "grad_norm": 2.231868267059326, "learning_rate": 6.269965676605615e-07, "loss": 0.1551, "step": 45630 }, { "epoch": 0.8949019607843137, "grad_norm": 2.313930034637451, "learning_rate": 6.268397721689441e-07, "loss": 0.181, "step": 45640 }, { "epoch": 0.8950980392156863, "grad_norm": 1.4935158491134644, "learning_rate": 6.266829633449479e-07, "loss": 0.132, "step": 45650 }, { "epoch": 0.8952941176470588, "grad_norm": 2.366804838180542, "learning_rate": 6.265261412050556e-07, "loss": 0.1541, "step": 45660 }, { "epoch": 0.8954901960784314, "grad_norm": 2.4331753253936768, "learning_rate": 6.26369305765751e-07, "loss": 0.1804, "step": 45670 }, { "epoch": 0.8956862745098039, "grad_norm": 1.9984793663024902, "learning_rate": 6.262124570435192e-07, "loss": 0.1516, "step": 45680 }, { "epoch": 0.8958823529411765, "grad_norm": 1.6438982486724854, "learning_rate": 6.260555950548472e-07, "loss": 0.1128, "step": 45690 }, { "epoch": 0.8960784313725491, "grad_norm": 1.5586755275726318, "learning_rate": 6.258987198162229e-07, "loss": 0.1545, "step": 45700 }, { "epoch": 0.8962745098039215, "grad_norm": 3.3755016326904297, "learning_rate": 6.257418313441356e-07, "loss": 0.1828, "step": 45710 }, { "epoch": 0.8964705882352941, "grad_norm": 2.8274097442626953, "learning_rate": 6.255849296550766e-07, "loss": 0.1311, "step": 45720 }, { "epoch": 0.8966666666666666, "grad_norm": 4.677453994750977, "learning_rate": 6.254280147655375e-07, "loss": 0.1618, "step": 45730 }, { "epoch": 0.8968627450980392, "grad_norm": 4.781133651733398, "learning_rate": 6.252710866920126e-07, "loss": 0.1281, "step": 45740 }, { "epoch": 0.8970588235294118, "grad_norm": 2.7340011596679688, "learning_rate": 6.251141454509967e-07, "loss": 0.1279, "step": 45750 }, { "epoch": 0.8972549019607843, "grad_norm": 3.7582991123199463, "learning_rate": 6.249571910589861e-07, "loss": 0.1247, "step": 45760 }, { "epoch": 0.8974509803921569, "grad_norm": 4.620944023132324, "learning_rate": 6.248002235324785e-07, "loss": 0.1309, "step": 45770 }, { "epoch": 0.8976470588235295, "grad_norm": 1.1354283094406128, "learning_rate": 6.246432428879733e-07, "loss": 0.1195, "step": 45780 }, { "epoch": 0.897843137254902, "grad_norm": 3.0295310020446777, "learning_rate": 6.244862491419708e-07, "loss": 0.1658, "step": 45790 }, { "epoch": 0.8980392156862745, "grad_norm": 2.060685634613037, "learning_rate": 6.243292423109733e-07, "loss": 0.1631, "step": 45800 }, { "epoch": 0.898235294117647, "grad_norm": 3.1319937705993652, "learning_rate": 6.241722224114835e-07, "loss": 0.1851, "step": 45810 }, { "epoch": 0.8984313725490196, "grad_norm": 2.037675619125366, "learning_rate": 6.240151894600067e-07, "loss": 0.1193, "step": 45820 }, { "epoch": 0.8986274509803922, "grad_norm": 2.98313307762146, "learning_rate": 6.238581434730486e-07, "loss": 0.1593, "step": 45830 }, { "epoch": 0.8988235294117647, "grad_norm": 4.592033863067627, "learning_rate": 6.237010844671166e-07, "loss": 0.128, "step": 45840 }, { "epoch": 0.8990196078431373, "grad_norm": 3.888303279876709, "learning_rate": 6.235440124587198e-07, "loss": 0.1571, "step": 45850 }, { "epoch": 0.8992156862745098, "grad_norm": 2.1003992557525635, "learning_rate": 6.23386927464368e-07, "loss": 0.152, "step": 45860 }, { "epoch": 0.8994117647058824, "grad_norm": 3.9448108673095703, "learning_rate": 6.232298295005727e-07, "loss": 0.165, "step": 45870 }, { "epoch": 0.899607843137255, "grad_norm": 1.5338553190231323, "learning_rate": 6.230727185838472e-07, "loss": 0.1269, "step": 45880 }, { "epoch": 0.8998039215686274, "grad_norm": 1.8457084894180298, "learning_rate": 6.229155947307051e-07, "loss": 0.1345, "step": 45890 }, { "epoch": 0.9, "grad_norm": 2.5618526935577393, "learning_rate": 6.227584579576625e-07, "loss": 0.1407, "step": 45900 }, { "epoch": 0.9001960784313725, "grad_norm": 2.1716508865356445, "learning_rate": 6.226013082812362e-07, "loss": 0.1387, "step": 45910 }, { "epoch": 0.9003921568627451, "grad_norm": 2.8940269947052, "learning_rate": 6.224441457179445e-07, "loss": 0.1352, "step": 45920 }, { "epoch": 0.9005882352941177, "grad_norm": 1.7451114654541016, "learning_rate": 6.222869702843069e-07, "loss": 0.1504, "step": 45930 }, { "epoch": 0.9007843137254902, "grad_norm": 3.7339272499084473, "learning_rate": 6.22129781996845e-07, "loss": 0.1663, "step": 45940 }, { "epoch": 0.9009803921568628, "grad_norm": 3.7397947311401367, "learning_rate": 6.219725808720804e-07, "loss": 0.1933, "step": 45950 }, { "epoch": 0.9011764705882352, "grad_norm": 1.626413106918335, "learning_rate": 6.218153669265374e-07, "loss": 0.1733, "step": 45960 }, { "epoch": 0.9013725490196078, "grad_norm": 3.1889073848724365, "learning_rate": 6.216581401767409e-07, "loss": 0.1328, "step": 45970 }, { "epoch": 0.9015686274509804, "grad_norm": 3.3349876403808594, "learning_rate": 6.215009006392171e-07, "loss": 0.1809, "step": 45980 }, { "epoch": 0.9017647058823529, "grad_norm": 2.4415042400360107, "learning_rate": 6.213436483304942e-07, "loss": 0.1557, "step": 45990 }, { "epoch": 0.9019607843137255, "grad_norm": 2.8056602478027344, "learning_rate": 6.211863832671006e-07, "loss": 0.1547, "step": 46000 }, { "epoch": 0.902156862745098, "grad_norm": 3.704395294189453, "learning_rate": 6.210291054655674e-07, "loss": 0.1805, "step": 46010 }, { "epoch": 0.9023529411764706, "grad_norm": 1.7523823976516724, "learning_rate": 6.208718149424262e-07, "loss": 0.1424, "step": 46020 }, { "epoch": 0.9025490196078432, "grad_norm": 4.24437952041626, "learning_rate": 6.2071451171421e-07, "loss": 0.1541, "step": 46030 }, { "epoch": 0.9027450980392157, "grad_norm": 4.949716567993164, "learning_rate": 6.205571957974533e-07, "loss": 0.1735, "step": 46040 }, { "epoch": 0.9029411764705882, "grad_norm": 2.1839044094085693, "learning_rate": 6.20399867208692e-07, "loss": 0.1758, "step": 46050 }, { "epoch": 0.9031372549019608, "grad_norm": 3.75048565864563, "learning_rate": 6.202425259644631e-07, "loss": 0.1507, "step": 46060 }, { "epoch": 0.9033333333333333, "grad_norm": 2.919677972793579, "learning_rate": 6.200851720813052e-07, "loss": 0.1542, "step": 46070 }, { "epoch": 0.9035294117647059, "grad_norm": 2.081190824508667, "learning_rate": 6.199278055757578e-07, "loss": 0.1783, "step": 46080 }, { "epoch": 0.9037254901960784, "grad_norm": 3.0440144538879395, "learning_rate": 6.197704264643619e-07, "loss": 0.1273, "step": 46090 }, { "epoch": 0.903921568627451, "grad_norm": 2.5059196949005127, "learning_rate": 6.196130347636606e-07, "loss": 0.1325, "step": 46100 }, { "epoch": 0.9041176470588236, "grad_norm": 3.3979077339172363, "learning_rate": 6.19455630490197e-07, "loss": 0.1348, "step": 46110 }, { "epoch": 0.9043137254901961, "grad_norm": 2.991666555404663, "learning_rate": 6.192982136605166e-07, "loss": 0.1686, "step": 46120 }, { "epoch": 0.9045098039215687, "grad_norm": 2.136002779006958, "learning_rate": 6.191407842911654e-07, "loss": 0.1679, "step": 46130 }, { "epoch": 0.9047058823529411, "grad_norm": 2.2073118686676025, "learning_rate": 6.189833423986915e-07, "loss": 0.1522, "step": 46140 }, { "epoch": 0.9049019607843137, "grad_norm": 1.0902671813964844, "learning_rate": 6.188258879996436e-07, "loss": 0.1522, "step": 46150 }, { "epoch": 0.9050980392156863, "grad_norm": 2.9452567100524902, "learning_rate": 6.186684211105722e-07, "loss": 0.1603, "step": 46160 }, { "epoch": 0.9052941176470588, "grad_norm": 1.3821759223937988, "learning_rate": 6.18510941748029e-07, "loss": 0.1451, "step": 46170 }, { "epoch": 0.9054901960784314, "grad_norm": 3.4787468910217285, "learning_rate": 6.183534499285668e-07, "loss": 0.1644, "step": 46180 }, { "epoch": 0.9056862745098039, "grad_norm": 3.656002998352051, "learning_rate": 6.181959456687401e-07, "loss": 0.1537, "step": 46190 }, { "epoch": 0.9058823529411765, "grad_norm": 2.29294490814209, "learning_rate": 6.180384289851044e-07, "loss": 0.1268, "step": 46200 }, { "epoch": 0.9060784313725491, "grad_norm": 2.6137239933013916, "learning_rate": 6.178808998942164e-07, "loss": 0.1585, "step": 46210 }, { "epoch": 0.9062745098039215, "grad_norm": 3.5491902828216553, "learning_rate": 6.177233584126346e-07, "loss": 0.1418, "step": 46220 }, { "epoch": 0.9064705882352941, "grad_norm": 2.8547260761260986, "learning_rate": 6.175658045569183e-07, "loss": 0.1667, "step": 46230 }, { "epoch": 0.9066666666666666, "grad_norm": 1.5656845569610596, "learning_rate": 6.174082383436283e-07, "loss": 0.1559, "step": 46240 }, { "epoch": 0.9068627450980392, "grad_norm": 2.2811279296875, "learning_rate": 6.172506597893268e-07, "loss": 0.1706, "step": 46250 }, { "epoch": 0.9070588235294118, "grad_norm": 4.734246253967285, "learning_rate": 6.170930689105772e-07, "loss": 0.1338, "step": 46260 }, { "epoch": 0.9072549019607843, "grad_norm": 2.719048023223877, "learning_rate": 6.169354657239438e-07, "loss": 0.1301, "step": 46270 }, { "epoch": 0.9074509803921569, "grad_norm": 3.078364610671997, "learning_rate": 6.167778502459932e-07, "loss": 0.152, "step": 46280 }, { "epoch": 0.9076470588235294, "grad_norm": 3.132981061935425, "learning_rate": 6.166202224932923e-07, "loss": 0.133, "step": 46290 }, { "epoch": 0.907843137254902, "grad_norm": 2.0709054470062256, "learning_rate": 6.164625824824098e-07, "loss": 0.1576, "step": 46300 }, { "epoch": 0.9080392156862745, "grad_norm": 1.5704002380371094, "learning_rate": 6.163049302299155e-07, "loss": 0.1805, "step": 46310 }, { "epoch": 0.908235294117647, "grad_norm": 3.8215596675872803, "learning_rate": 6.161472657523805e-07, "loss": 0.1517, "step": 46320 }, { "epoch": 0.9084313725490196, "grad_norm": 5.801513671875, "learning_rate": 6.159895890663771e-07, "loss": 0.1427, "step": 46330 }, { "epoch": 0.9086274509803921, "grad_norm": 4.907294273376465, "learning_rate": 6.158319001884793e-07, "loss": 0.1488, "step": 46340 }, { "epoch": 0.9088235294117647, "grad_norm": 2.5841310024261475, "learning_rate": 6.156741991352618e-07, "loss": 0.1714, "step": 46350 }, { "epoch": 0.9090196078431373, "grad_norm": 3.2718863487243652, "learning_rate": 6.155164859233011e-07, "loss": 0.1468, "step": 46360 }, { "epoch": 0.9092156862745098, "grad_norm": 3.8785853385925293, "learning_rate": 6.153587605691748e-07, "loss": 0.1766, "step": 46370 }, { "epoch": 0.9094117647058824, "grad_norm": 3.6085946559906006, "learning_rate": 6.152010230894616e-07, "loss": 0.137, "step": 46380 }, { "epoch": 0.909607843137255, "grad_norm": 2.0901312828063965, "learning_rate": 6.150432735007415e-07, "loss": 0.1688, "step": 46390 }, { "epoch": 0.9098039215686274, "grad_norm": 2.025486469268799, "learning_rate": 6.14885511819596e-07, "loss": 0.1583, "step": 46400 }, { "epoch": 0.91, "grad_norm": 2.732841730117798, "learning_rate": 6.147277380626074e-07, "loss": 0.1556, "step": 46410 }, { "epoch": 0.9101960784313725, "grad_norm": 2.140625238418579, "learning_rate": 6.145699522463603e-07, "loss": 0.1385, "step": 46420 }, { "epoch": 0.9103921568627451, "grad_norm": 3.178072690963745, "learning_rate": 6.144121543874394e-07, "loss": 0.174, "step": 46430 }, { "epoch": 0.9105882352941177, "grad_norm": 2.5793328285217285, "learning_rate": 6.142543445024311e-07, "loss": 0.1541, "step": 46440 }, { "epoch": 0.9107843137254902, "grad_norm": 4.6725053787231445, "learning_rate": 6.140965226079233e-07, "loss": 0.1615, "step": 46450 }, { "epoch": 0.9109803921568628, "grad_norm": 2.888000726699829, "learning_rate": 6.139386887205049e-07, "loss": 0.1368, "step": 46460 }, { "epoch": 0.9111764705882353, "grad_norm": 1.6990256309509277, "learning_rate": 6.137808428567662e-07, "loss": 0.1706, "step": 46470 }, { "epoch": 0.9113725490196078, "grad_norm": 3.33890438079834, "learning_rate": 6.136229850332987e-07, "loss": 0.1695, "step": 46480 }, { "epoch": 0.9115686274509804, "grad_norm": 2.585167646408081, "learning_rate": 6.134651152666949e-07, "loss": 0.108, "step": 46490 }, { "epoch": 0.9117647058823529, "grad_norm": 2.5121984481811523, "learning_rate": 6.133072335735492e-07, "loss": 0.1652, "step": 46500 }, { "epoch": 0.9119607843137255, "grad_norm": 13.699408531188965, "learning_rate": 6.131493399704565e-07, "loss": 0.1579, "step": 46510 }, { "epoch": 0.912156862745098, "grad_norm": 2.9522931575775146, "learning_rate": 6.129914344740133e-07, "loss": 0.1563, "step": 46520 }, { "epoch": 0.9123529411764706, "grad_norm": 1.6694997549057007, "learning_rate": 6.128335171008179e-07, "loss": 0.1572, "step": 46530 }, { "epoch": 0.9125490196078432, "grad_norm": 2.945707321166992, "learning_rate": 6.126755878674685e-07, "loss": 0.1957, "step": 46540 }, { "epoch": 0.9127450980392157, "grad_norm": 1.992852807044983, "learning_rate": 6.12517646790566e-07, "loss": 0.1319, "step": 46550 }, { "epoch": 0.9129411764705883, "grad_norm": 5.339662075042725, "learning_rate": 6.123596938867116e-07, "loss": 0.1555, "step": 46560 }, { "epoch": 0.9131372549019607, "grad_norm": 3.0201380252838135, "learning_rate": 6.122017291725081e-07, "loss": 0.1344, "step": 46570 }, { "epoch": 0.9133333333333333, "grad_norm": 4.395500659942627, "learning_rate": 6.120437526645596e-07, "loss": 0.166, "step": 46580 }, { "epoch": 0.9135294117647059, "grad_norm": 3.153700351715088, "learning_rate": 6.11885764379471e-07, "loss": 0.1341, "step": 46590 }, { "epoch": 0.9137254901960784, "grad_norm": 2.5725388526916504, "learning_rate": 6.117277643338492e-07, "loss": 0.1606, "step": 46600 }, { "epoch": 0.913921568627451, "grad_norm": 2.3454477787017822, "learning_rate": 6.115697525443015e-07, "loss": 0.1518, "step": 46610 }, { "epoch": 0.9141176470588235, "grad_norm": 2.730820894241333, "learning_rate": 6.114117290274371e-07, "loss": 0.1373, "step": 46620 }, { "epoch": 0.9143137254901961, "grad_norm": 2.4819276332855225, "learning_rate": 6.112536937998659e-07, "loss": 0.1316, "step": 46630 }, { "epoch": 0.9145098039215687, "grad_norm": 3.035228967666626, "learning_rate": 6.110956468781997e-07, "loss": 0.1426, "step": 46640 }, { "epoch": 0.9147058823529411, "grad_norm": 3.4450907707214355, "learning_rate": 6.109375882790507e-07, "loss": 0.1622, "step": 46650 }, { "epoch": 0.9149019607843137, "grad_norm": 3.172006130218506, "learning_rate": 6.10779518019033e-07, "loss": 0.1218, "step": 46660 }, { "epoch": 0.9150980392156862, "grad_norm": 2.845233201980591, "learning_rate": 6.106214361147617e-07, "loss": 0.1628, "step": 46670 }, { "epoch": 0.9152941176470588, "grad_norm": 1.0433045625686646, "learning_rate": 6.104633425828528e-07, "loss": 0.1357, "step": 46680 }, { "epoch": 0.9154901960784314, "grad_norm": 1.7701919078826904, "learning_rate": 6.103052374399242e-07, "loss": 0.173, "step": 46690 }, { "epoch": 0.9156862745098039, "grad_norm": 2.02402925491333, "learning_rate": 6.101471207025945e-07, "loss": 0.1566, "step": 46700 }, { "epoch": 0.9158823529411765, "grad_norm": 4.652158260345459, "learning_rate": 6.099889923874834e-07, "loss": 0.1491, "step": 46710 }, { "epoch": 0.9160784313725491, "grad_norm": 3.3850533962249756, "learning_rate": 6.098308525112126e-07, "loss": 0.1748, "step": 46720 }, { "epoch": 0.9162745098039216, "grad_norm": 3.195317029953003, "learning_rate": 6.096727010904041e-07, "loss": 0.1539, "step": 46730 }, { "epoch": 0.9164705882352941, "grad_norm": 2.7811262607574463, "learning_rate": 6.095145381416816e-07, "loss": 0.1615, "step": 46740 }, { "epoch": 0.9166666666666666, "grad_norm": 2.261892318725586, "learning_rate": 6.093563636816699e-07, "loss": 0.1677, "step": 46750 }, { "epoch": 0.9168627450980392, "grad_norm": 2.545236349105835, "learning_rate": 6.091981777269951e-07, "loss": 0.1556, "step": 46760 }, { "epoch": 0.9170588235294118, "grad_norm": 2.748758316040039, "learning_rate": 6.090399802942845e-07, "loss": 0.147, "step": 46770 }, { "epoch": 0.9172549019607843, "grad_norm": 1.1187293529510498, "learning_rate": 6.088817714001663e-07, "loss": 0.1328, "step": 46780 }, { "epoch": 0.9174509803921569, "grad_norm": 1.7050269842147827, "learning_rate": 6.087235510612704e-07, "loss": 0.1484, "step": 46790 }, { "epoch": 0.9176470588235294, "grad_norm": 2.5081892013549805, "learning_rate": 6.085653192942277e-07, "loss": 0.1414, "step": 46800 }, { "epoch": 0.917843137254902, "grad_norm": 2.4568772315979004, "learning_rate": 6.084070761156697e-07, "loss": 0.1573, "step": 46810 }, { "epoch": 0.9180392156862746, "grad_norm": 2.708697557449341, "learning_rate": 6.082488215422304e-07, "loss": 0.1466, "step": 46820 }, { "epoch": 0.918235294117647, "grad_norm": 2.545856237411499, "learning_rate": 6.080905555905436e-07, "loss": 0.1232, "step": 46830 }, { "epoch": 0.9184313725490196, "grad_norm": 1.5635558366775513, "learning_rate": 6.079322782772454e-07, "loss": 0.1255, "step": 46840 }, { "epoch": 0.9186274509803921, "grad_norm": 3.067666530609131, "learning_rate": 6.077739896189727e-07, "loss": 0.171, "step": 46850 }, { "epoch": 0.9188235294117647, "grad_norm": 2.827547073364258, "learning_rate": 6.076156896323632e-07, "loss": 0.1498, "step": 46860 }, { "epoch": 0.9190196078431373, "grad_norm": 4.682238578796387, "learning_rate": 6.074573783340562e-07, "loss": 0.1781, "step": 46870 }, { "epoch": 0.9192156862745098, "grad_norm": 0.9607089757919312, "learning_rate": 6.072990557406923e-07, "loss": 0.1529, "step": 46880 }, { "epoch": 0.9194117647058824, "grad_norm": 2.3049426078796387, "learning_rate": 6.071407218689128e-07, "loss": 0.1329, "step": 46890 }, { "epoch": 0.9196078431372549, "grad_norm": 2.006289482116699, "learning_rate": 6.069823767353606e-07, "loss": 0.1405, "step": 46900 }, { "epoch": 0.9198039215686274, "grad_norm": 3.630620002746582, "learning_rate": 6.068240203566799e-07, "loss": 0.1641, "step": 46910 }, { "epoch": 0.92, "grad_norm": 2.568366765975952, "learning_rate": 6.066656527495155e-07, "loss": 0.1307, "step": 46920 }, { "epoch": 0.9201960784313725, "grad_norm": 2.3244006633758545, "learning_rate": 6.065072739305141e-07, "loss": 0.147, "step": 46930 }, { "epoch": 0.9203921568627451, "grad_norm": 2.698336124420166, "learning_rate": 6.063488839163229e-07, "loss": 0.154, "step": 46940 }, { "epoch": 0.9205882352941176, "grad_norm": 2.210325241088867, "learning_rate": 6.061904827235906e-07, "loss": 0.1293, "step": 46950 }, { "epoch": 0.9207843137254902, "grad_norm": 2.14821457862854, "learning_rate": 6.060320703689674e-07, "loss": 0.1355, "step": 46960 }, { "epoch": 0.9209803921568628, "grad_norm": 2.9312610626220703, "learning_rate": 6.05873646869104e-07, "loss": 0.1483, "step": 46970 }, { "epoch": 0.9211764705882353, "grad_norm": 2.991647243499756, "learning_rate": 6.057152122406526e-07, "loss": 0.1462, "step": 46980 }, { "epoch": 0.9213725490196079, "grad_norm": 2.774026393890381, "learning_rate": 6.055567665002668e-07, "loss": 0.1965, "step": 46990 }, { "epoch": 0.9215686274509803, "grad_norm": 2.832124710083008, "learning_rate": 6.05398309664601e-07, "loss": 0.1986, "step": 47000 }, { "epoch": 0.9217647058823529, "grad_norm": 1.3018540143966675, "learning_rate": 6.052398417503109e-07, "loss": 0.1343, "step": 47010 }, { "epoch": 0.9219607843137255, "grad_norm": 1.6466822624206543, "learning_rate": 6.050813627740534e-07, "loss": 0.1623, "step": 47020 }, { "epoch": 0.922156862745098, "grad_norm": 2.2034826278686523, "learning_rate": 6.049228727524866e-07, "loss": 0.1887, "step": 47030 }, { "epoch": 0.9223529411764706, "grad_norm": 3.8110063076019287, "learning_rate": 6.047643717022697e-07, "loss": 0.1552, "step": 47040 }, { "epoch": 0.9225490196078432, "grad_norm": 2.072589635848999, "learning_rate": 6.04605859640063e-07, "loss": 0.1556, "step": 47050 }, { "epoch": 0.9227450980392157, "grad_norm": 3.0836093425750732, "learning_rate": 6.044473365825278e-07, "loss": 0.1582, "step": 47060 }, { "epoch": 0.9229411764705883, "grad_norm": 2.926205635070801, "learning_rate": 6.042888025463273e-07, "loss": 0.1241, "step": 47070 }, { "epoch": 0.9231372549019607, "grad_norm": 4.010020732879639, "learning_rate": 6.041302575481247e-07, "loss": 0.1623, "step": 47080 }, { "epoch": 0.9233333333333333, "grad_norm": 2.169362783432007, "learning_rate": 6.039717016045855e-07, "loss": 0.1513, "step": 47090 }, { "epoch": 0.9235294117647059, "grad_norm": 2.0397908687591553, "learning_rate": 6.038131347323756e-07, "loss": 0.1451, "step": 47100 }, { "epoch": 0.9237254901960784, "grad_norm": 4.8714118003845215, "learning_rate": 6.036545569481623e-07, "loss": 0.1771, "step": 47110 }, { "epoch": 0.923921568627451, "grad_norm": 5.099405765533447, "learning_rate": 6.034959682686141e-07, "loss": 0.1509, "step": 47120 }, { "epoch": 0.9241176470588235, "grad_norm": 1.7362970113754272, "learning_rate": 6.033373687104003e-07, "loss": 0.1569, "step": 47130 }, { "epoch": 0.9243137254901961, "grad_norm": 2.8502485752105713, "learning_rate": 6.031787582901918e-07, "loss": 0.1466, "step": 47140 }, { "epoch": 0.9245098039215687, "grad_norm": 4.113798141479492, "learning_rate": 6.030201370246607e-07, "loss": 0.1762, "step": 47150 }, { "epoch": 0.9247058823529412, "grad_norm": 1.6164143085479736, "learning_rate": 6.028615049304794e-07, "loss": 0.1496, "step": 47160 }, { "epoch": 0.9249019607843137, "grad_norm": 5.093061923980713, "learning_rate": 6.027028620243227e-07, "loss": 0.1624, "step": 47170 }, { "epoch": 0.9250980392156862, "grad_norm": 1.2519563436508179, "learning_rate": 6.025442083228655e-07, "loss": 0.1282, "step": 47180 }, { "epoch": 0.9252941176470588, "grad_norm": 3.4239718914031982, "learning_rate": 6.023855438427841e-07, "loss": 0.1843, "step": 47190 }, { "epoch": 0.9254901960784314, "grad_norm": 5.674184322357178, "learning_rate": 6.022268686007564e-07, "loss": 0.1407, "step": 47200 }, { "epoch": 0.9256862745098039, "grad_norm": 2.757152557373047, "learning_rate": 6.020681826134608e-07, "loss": 0.1447, "step": 47210 }, { "epoch": 0.9258823529411765, "grad_norm": 2.8283071517944336, "learning_rate": 6.019094858975771e-07, "loss": 0.1712, "step": 47220 }, { "epoch": 0.926078431372549, "grad_norm": 2.750307083129883, "learning_rate": 6.017507784697864e-07, "loss": 0.1773, "step": 47230 }, { "epoch": 0.9262745098039216, "grad_norm": 3.480027675628662, "learning_rate": 6.015920603467705e-07, "loss": 0.1764, "step": 47240 }, { "epoch": 0.9264705882352942, "grad_norm": 3.493460178375244, "learning_rate": 6.014333315452127e-07, "loss": 0.1594, "step": 47250 }, { "epoch": 0.9266666666666666, "grad_norm": 1.4895840883255005, "learning_rate": 6.012745920817974e-07, "loss": 0.1821, "step": 47260 }, { "epoch": 0.9268627450980392, "grad_norm": 3.426579236984253, "learning_rate": 6.011158419732098e-07, "loss": 0.1473, "step": 47270 }, { "epoch": 0.9270588235294117, "grad_norm": 2.2568211555480957, "learning_rate": 6.009570812361368e-07, "loss": 0.1573, "step": 47280 }, { "epoch": 0.9272549019607843, "grad_norm": 2.5272529125213623, "learning_rate": 6.007983098872656e-07, "loss": 0.1585, "step": 47290 }, { "epoch": 0.9274509803921569, "grad_norm": 2.8192858695983887, "learning_rate": 6.006395279432852e-07, "loss": 0.1925, "step": 47300 }, { "epoch": 0.9276470588235294, "grad_norm": 5.307064533233643, "learning_rate": 6.004807354208856e-07, "loss": 0.1549, "step": 47310 }, { "epoch": 0.927843137254902, "grad_norm": 3.58259654045105, "learning_rate": 6.003219323367576e-07, "loss": 0.1469, "step": 47320 }, { "epoch": 0.9280392156862745, "grad_norm": 2.2473628520965576, "learning_rate": 6.001631187075932e-07, "loss": 0.1522, "step": 47330 }, { "epoch": 0.928235294117647, "grad_norm": 2.549286127090454, "learning_rate": 6.00004294550086e-07, "loss": 0.1325, "step": 47340 }, { "epoch": 0.9284313725490196, "grad_norm": 3.6230175495147705, "learning_rate": 5.9984545988093e-07, "loss": 0.1509, "step": 47350 }, { "epoch": 0.9286274509803921, "grad_norm": 2.463386297225952, "learning_rate": 5.996866147168209e-07, "loss": 0.1416, "step": 47360 }, { "epoch": 0.9288235294117647, "grad_norm": 2.3587896823883057, "learning_rate": 5.99527759074455e-07, "loss": 0.149, "step": 47370 }, { "epoch": 0.9290196078431373, "grad_norm": 2.5491883754730225, "learning_rate": 5.993688929705299e-07, "loss": 0.1372, "step": 47380 }, { "epoch": 0.9292156862745098, "grad_norm": 2.1967759132385254, "learning_rate": 5.992100164217445e-07, "loss": 0.1328, "step": 47390 }, { "epoch": 0.9294117647058824, "grad_norm": 3.7936782836914062, "learning_rate": 5.990511294447984e-07, "loss": 0.1388, "step": 47400 }, { "epoch": 0.9296078431372549, "grad_norm": 2.1162118911743164, "learning_rate": 5.988922320563928e-07, "loss": 0.1407, "step": 47410 }, { "epoch": 0.9298039215686275, "grad_norm": 5.346748352050781, "learning_rate": 5.987333242732297e-07, "loss": 0.1239, "step": 47420 }, { "epoch": 0.93, "grad_norm": 2.642683506011963, "learning_rate": 5.985744061120117e-07, "loss": 0.1479, "step": 47430 }, { "epoch": 0.9301960784313725, "grad_norm": 2.408900737762451, "learning_rate": 5.984154775894438e-07, "loss": 0.144, "step": 47440 }, { "epoch": 0.9303921568627451, "grad_norm": 2.0321171283721924, "learning_rate": 5.982565387222307e-07, "loss": 0.17, "step": 47450 }, { "epoch": 0.9305882352941176, "grad_norm": 2.1983251571655273, "learning_rate": 5.98097589527079e-07, "loss": 0.1142, "step": 47460 }, { "epoch": 0.9307843137254902, "grad_norm": 8.558822631835938, "learning_rate": 5.979386300206963e-07, "loss": 0.1708, "step": 47470 }, { "epoch": 0.9309803921568628, "grad_norm": 4.699198246002197, "learning_rate": 5.977796602197908e-07, "loss": 0.1439, "step": 47480 }, { "epoch": 0.9311764705882353, "grad_norm": 1.607190489768982, "learning_rate": 5.976206801410722e-07, "loss": 0.1689, "step": 47490 }, { "epoch": 0.9313725490196079, "grad_norm": 4.414806365966797, "learning_rate": 5.974616898012517e-07, "loss": 0.1787, "step": 47500 }, { "epoch": 0.9315686274509803, "grad_norm": 3.1588492393493652, "learning_rate": 5.973026892170403e-07, "loss": 0.1506, "step": 47510 }, { "epoch": 0.9317647058823529, "grad_norm": 3.6914613246917725, "learning_rate": 5.971436784051514e-07, "loss": 0.1558, "step": 47520 }, { "epoch": 0.9319607843137255, "grad_norm": 3.1103367805480957, "learning_rate": 5.969846573822989e-07, "loss": 0.1261, "step": 47530 }, { "epoch": 0.932156862745098, "grad_norm": 3.525517463684082, "learning_rate": 5.968256261651976e-07, "loss": 0.1808, "step": 47540 }, { "epoch": 0.9323529411764706, "grad_norm": 2.510528564453125, "learning_rate": 5.966665847705638e-07, "loss": 0.1499, "step": 47550 }, { "epoch": 0.9325490196078431, "grad_norm": 0.8170267939567566, "learning_rate": 5.965075332151146e-07, "loss": 0.1175, "step": 47560 }, { "epoch": 0.9327450980392157, "grad_norm": 1.2876781225204468, "learning_rate": 5.963484715155681e-07, "loss": 0.1305, "step": 47570 }, { "epoch": 0.9329411764705883, "grad_norm": 2.7710609436035156, "learning_rate": 5.961893996886438e-07, "loss": 0.135, "step": 47580 }, { "epoch": 0.9331372549019608, "grad_norm": 2.9495468139648438, "learning_rate": 5.960303177510618e-07, "loss": 0.144, "step": 47590 }, { "epoch": 0.9333333333333333, "grad_norm": 3.9385156631469727, "learning_rate": 5.958712257195437e-07, "loss": 0.1975, "step": 47600 }, { "epoch": 0.9335294117647058, "grad_norm": 2.619349241256714, "learning_rate": 5.957121236108119e-07, "loss": 0.1715, "step": 47610 }, { "epoch": 0.9337254901960784, "grad_norm": 2.9529080390930176, "learning_rate": 5.9555301144159e-07, "loss": 0.1348, "step": 47620 }, { "epoch": 0.933921568627451, "grad_norm": 2.1329684257507324, "learning_rate": 5.953938892286026e-07, "loss": 0.1637, "step": 47630 }, { "epoch": 0.9341176470588235, "grad_norm": 2.655057907104492, "learning_rate": 5.952347569885751e-07, "loss": 0.1772, "step": 47640 }, { "epoch": 0.9343137254901961, "grad_norm": 3.572221279144287, "learning_rate": 5.950756147382345e-07, "loss": 0.1813, "step": 47650 }, { "epoch": 0.9345098039215687, "grad_norm": 1.7781206369400024, "learning_rate": 5.949164624943087e-07, "loss": 0.1257, "step": 47660 }, { "epoch": 0.9347058823529412, "grad_norm": 4.519789695739746, "learning_rate": 5.94757300273526e-07, "loss": 0.1836, "step": 47670 }, { "epoch": 0.9349019607843138, "grad_norm": 2.9078588485717773, "learning_rate": 5.945981280926167e-07, "loss": 0.1507, "step": 47680 }, { "epoch": 0.9350980392156862, "grad_norm": 2.7350125312805176, "learning_rate": 5.944389459683114e-07, "loss": 0.1412, "step": 47690 }, { "epoch": 0.9352941176470588, "grad_norm": 1.0083881616592407, "learning_rate": 5.942797539173421e-07, "loss": 0.1352, "step": 47700 }, { "epoch": 0.9354901960784314, "grad_norm": 2.7614946365356445, "learning_rate": 5.94120551956442e-07, "loss": 0.1959, "step": 47710 }, { "epoch": 0.9356862745098039, "grad_norm": 2.3619816303253174, "learning_rate": 5.93961340102345e-07, "loss": 0.1504, "step": 47720 }, { "epoch": 0.9358823529411765, "grad_norm": 2.2307000160217285, "learning_rate": 5.93802118371786e-07, "loss": 0.1227, "step": 47730 }, { "epoch": 0.936078431372549, "grad_norm": 2.559771776199341, "learning_rate": 5.936428867815017e-07, "loss": 0.1681, "step": 47740 }, { "epoch": 0.9362745098039216, "grad_norm": 4.050104141235352, "learning_rate": 5.934836453482285e-07, "loss": 0.132, "step": 47750 }, { "epoch": 0.9364705882352942, "grad_norm": 1.557645320892334, "learning_rate": 5.933243940887049e-07, "loss": 0.1429, "step": 47760 }, { "epoch": 0.9366666666666666, "grad_norm": 2.5761351585388184, "learning_rate": 5.931651330196702e-07, "loss": 0.1614, "step": 47770 }, { "epoch": 0.9368627450980392, "grad_norm": 4.151908874511719, "learning_rate": 5.930058621578646e-07, "loss": 0.1773, "step": 47780 }, { "epoch": 0.9370588235294117, "grad_norm": 3.407494068145752, "learning_rate": 5.928465815200292e-07, "loss": 0.1523, "step": 47790 }, { "epoch": 0.9372549019607843, "grad_norm": 1.6711822748184204, "learning_rate": 5.926872911229065e-07, "loss": 0.1587, "step": 47800 }, { "epoch": 0.9374509803921569, "grad_norm": 2.7529828548431396, "learning_rate": 5.925279909832398e-07, "loss": 0.1443, "step": 47810 }, { "epoch": 0.9376470588235294, "grad_norm": 2.6655356884002686, "learning_rate": 5.923686811177735e-07, "loss": 0.135, "step": 47820 }, { "epoch": 0.937843137254902, "grad_norm": 2.8956098556518555, "learning_rate": 5.922093615432527e-07, "loss": 0.1062, "step": 47830 }, { "epoch": 0.9380392156862745, "grad_norm": 1.88449227809906, "learning_rate": 5.920500322764241e-07, "loss": 0.1427, "step": 47840 }, { "epoch": 0.9382352941176471, "grad_norm": 1.772053837776184, "learning_rate": 5.918906933340352e-07, "loss": 0.1375, "step": 47850 }, { "epoch": 0.9384313725490196, "grad_norm": 2.5986921787261963, "learning_rate": 5.91731344732834e-07, "loss": 0.1761, "step": 47860 }, { "epoch": 0.9386274509803921, "grad_norm": 1.7274833917617798, "learning_rate": 5.915719864895702e-07, "loss": 0.1588, "step": 47870 }, { "epoch": 0.9388235294117647, "grad_norm": 2.370091438293457, "learning_rate": 5.914126186209943e-07, "loss": 0.1562, "step": 47880 }, { "epoch": 0.9390196078431372, "grad_norm": 2.183976650238037, "learning_rate": 5.912532411438576e-07, "loss": 0.1308, "step": 47890 }, { "epoch": 0.9392156862745098, "grad_norm": 2.9413721561431885, "learning_rate": 5.910938540749128e-07, "loss": 0.1605, "step": 47900 }, { "epoch": 0.9394117647058824, "grad_norm": 1.921373724937439, "learning_rate": 5.909344574309133e-07, "loss": 0.1362, "step": 47910 }, { "epoch": 0.9396078431372549, "grad_norm": 2.1743526458740234, "learning_rate": 5.907750512286135e-07, "loss": 0.1658, "step": 47920 }, { "epoch": 0.9398039215686275, "grad_norm": 2.347046136856079, "learning_rate": 5.906156354847691e-07, "loss": 0.1755, "step": 47930 }, { "epoch": 0.94, "grad_norm": 2.949563503265381, "learning_rate": 5.904562102161364e-07, "loss": 0.1727, "step": 47940 }, { "epoch": 0.9401960784313725, "grad_norm": 3.122746229171753, "learning_rate": 5.90296775439473e-07, "loss": 0.1882, "step": 47950 }, { "epoch": 0.9403921568627451, "grad_norm": 3.0720810890197754, "learning_rate": 5.901373311715374e-07, "loss": 0.1202, "step": 47960 }, { "epoch": 0.9405882352941176, "grad_norm": 2.1252782344818115, "learning_rate": 5.89977877429089e-07, "loss": 0.1303, "step": 47970 }, { "epoch": 0.9407843137254902, "grad_norm": 2.107551336288452, "learning_rate": 5.898184142288886e-07, "loss": 0.1518, "step": 47980 }, { "epoch": 0.9409803921568628, "grad_norm": 1.5665735006332397, "learning_rate": 5.896589415876973e-07, "loss": 0.1406, "step": 47990 }, { "epoch": 0.9411764705882353, "grad_norm": 2.461834669113159, "learning_rate": 5.894994595222777e-07, "loss": 0.181, "step": 48000 }, { "epoch": 0.9413725490196079, "grad_norm": 4.156539440155029, "learning_rate": 5.893399680493936e-07, "loss": 0.1702, "step": 48010 }, { "epoch": 0.9415686274509804, "grad_norm": 1.6028485298156738, "learning_rate": 5.89180467185809e-07, "loss": 0.1551, "step": 48020 }, { "epoch": 0.941764705882353, "grad_norm": 2.4217686653137207, "learning_rate": 5.890209569482896e-07, "loss": 0.1699, "step": 48030 }, { "epoch": 0.9419607843137255, "grad_norm": 2.460144281387329, "learning_rate": 5.888614373536018e-07, "loss": 0.1607, "step": 48040 }, { "epoch": 0.942156862745098, "grad_norm": 2.933839797973633, "learning_rate": 5.887019084185131e-07, "loss": 0.1409, "step": 48050 }, { "epoch": 0.9423529411764706, "grad_norm": 2.02980899810791, "learning_rate": 5.885423701597917e-07, "loss": 0.1178, "step": 48060 }, { "epoch": 0.9425490196078431, "grad_norm": 1.4507942199707031, "learning_rate": 5.883828225942071e-07, "loss": 0.1566, "step": 48070 }, { "epoch": 0.9427450980392157, "grad_norm": 2.837886333465576, "learning_rate": 5.882232657385296e-07, "loss": 0.1657, "step": 48080 }, { "epoch": 0.9429411764705883, "grad_norm": 4.370419025421143, "learning_rate": 5.880636996095307e-07, "loss": 0.1594, "step": 48090 }, { "epoch": 0.9431372549019608, "grad_norm": 1.1256636381149292, "learning_rate": 5.879041242239825e-07, "loss": 0.116, "step": 48100 }, { "epoch": 0.9433333333333334, "grad_norm": 3.4100961685180664, "learning_rate": 5.877445395986584e-07, "loss": 0.1702, "step": 48110 }, { "epoch": 0.9435294117647058, "grad_norm": 1.6713367700576782, "learning_rate": 5.875849457503328e-07, "loss": 0.1336, "step": 48120 }, { "epoch": 0.9437254901960784, "grad_norm": 2.432894229888916, "learning_rate": 5.874253426957806e-07, "loss": 0.1644, "step": 48130 }, { "epoch": 0.943921568627451, "grad_norm": 2.2965917587280273, "learning_rate": 5.872657304517781e-07, "loss": 0.1505, "step": 48140 }, { "epoch": 0.9441176470588235, "grad_norm": 3.511018991470337, "learning_rate": 5.871061090351026e-07, "loss": 0.141, "step": 48150 }, { "epoch": 0.9443137254901961, "grad_norm": 2.748293399810791, "learning_rate": 5.86946478462532e-07, "loss": 0.1348, "step": 48160 }, { "epoch": 0.9445098039215686, "grad_norm": 2.9224419593811035, "learning_rate": 5.867868387508457e-07, "loss": 0.1824, "step": 48170 }, { "epoch": 0.9447058823529412, "grad_norm": 2.855036497116089, "learning_rate": 5.866271899168235e-07, "loss": 0.1394, "step": 48180 }, { "epoch": 0.9449019607843138, "grad_norm": 3.9648282527923584, "learning_rate": 5.864675319772464e-07, "loss": 0.1586, "step": 48190 }, { "epoch": 0.9450980392156862, "grad_norm": 2.1914734840393066, "learning_rate": 5.863078649488966e-07, "loss": 0.1655, "step": 48200 }, { "epoch": 0.9452941176470588, "grad_norm": 3.3479104042053223, "learning_rate": 5.861481888485566e-07, "loss": 0.1304, "step": 48210 }, { "epoch": 0.9454901960784313, "grad_norm": 2.2471084594726562, "learning_rate": 5.859885036930107e-07, "loss": 0.143, "step": 48220 }, { "epoch": 0.9456862745098039, "grad_norm": 2.805995225906372, "learning_rate": 5.858288094990436e-07, "loss": 0.1647, "step": 48230 }, { "epoch": 0.9458823529411765, "grad_norm": 2.9215927124023438, "learning_rate": 5.856691062834406e-07, "loss": 0.1524, "step": 48240 }, { "epoch": 0.946078431372549, "grad_norm": 2.7056679725646973, "learning_rate": 5.855093940629893e-07, "loss": 0.1693, "step": 48250 }, { "epoch": 0.9462745098039216, "grad_norm": 1.5733002424240112, "learning_rate": 5.853496728544767e-07, "loss": 0.1768, "step": 48260 }, { "epoch": 0.9464705882352941, "grad_norm": 3.4756453037261963, "learning_rate": 5.851899426746915e-07, "loss": 0.1443, "step": 48270 }, { "epoch": 0.9466666666666667, "grad_norm": 4.220004081726074, "learning_rate": 5.850302035404235e-07, "loss": 0.1511, "step": 48280 }, { "epoch": 0.9468627450980392, "grad_norm": 3.3801071643829346, "learning_rate": 5.84870455468463e-07, "loss": 0.1528, "step": 48290 }, { "epoch": 0.9470588235294117, "grad_norm": 3.292362928390503, "learning_rate": 5.847106984756015e-07, "loss": 0.124, "step": 48300 }, { "epoch": 0.9472549019607843, "grad_norm": 3.3895368576049805, "learning_rate": 5.845509325786315e-07, "loss": 0.1438, "step": 48310 }, { "epoch": 0.9474509803921569, "grad_norm": 1.756727933883667, "learning_rate": 5.843911577943459e-07, "loss": 0.1763, "step": 48320 }, { "epoch": 0.9476470588235294, "grad_norm": 2.556865930557251, "learning_rate": 5.842313741395393e-07, "loss": 0.1471, "step": 48330 }, { "epoch": 0.947843137254902, "grad_norm": 2.2129690647125244, "learning_rate": 5.840715816310068e-07, "loss": 0.1294, "step": 48340 }, { "epoch": 0.9480392156862745, "grad_norm": 4.707003116607666, "learning_rate": 5.839117802855444e-07, "loss": 0.1668, "step": 48350 }, { "epoch": 0.9482352941176471, "grad_norm": 2.0091230869293213, "learning_rate": 5.837519701199495e-07, "loss": 0.1572, "step": 48360 }, { "epoch": 0.9484313725490197, "grad_norm": 2.950392246246338, "learning_rate": 5.835921511510196e-07, "loss": 0.1673, "step": 48370 }, { "epoch": 0.9486274509803921, "grad_norm": 2.220982789993286, "learning_rate": 5.834323233955538e-07, "loss": 0.141, "step": 48380 }, { "epoch": 0.9488235294117647, "grad_norm": 2.3021976947784424, "learning_rate": 5.83272486870352e-07, "loss": 0.1513, "step": 48390 }, { "epoch": 0.9490196078431372, "grad_norm": 4.729793071746826, "learning_rate": 5.831126415922147e-07, "loss": 0.1457, "step": 48400 }, { "epoch": 0.9492156862745098, "grad_norm": 3.722808837890625, "learning_rate": 5.829527875779436e-07, "loss": 0.1579, "step": 48410 }, { "epoch": 0.9494117647058824, "grad_norm": 1.8089256286621094, "learning_rate": 5.827929248443415e-07, "loss": 0.1642, "step": 48420 }, { "epoch": 0.9496078431372549, "grad_norm": 1.9880505800247192, "learning_rate": 5.826330534082115e-07, "loss": 0.156, "step": 48430 }, { "epoch": 0.9498039215686275, "grad_norm": 3.940850257873535, "learning_rate": 5.824731732863586e-07, "loss": 0.1732, "step": 48440 }, { "epoch": 0.95, "grad_norm": 1.651511788368225, "learning_rate": 5.823132844955875e-07, "loss": 0.1337, "step": 48450 }, { "epoch": 0.9501960784313725, "grad_norm": 4.092323303222656, "learning_rate": 5.821533870527046e-07, "loss": 0.1588, "step": 48460 }, { "epoch": 0.9503921568627451, "grad_norm": 4.556176662445068, "learning_rate": 5.819934809745174e-07, "loss": 0.1264, "step": 48470 }, { "epoch": 0.9505882352941176, "grad_norm": 2.3637726306915283, "learning_rate": 5.818335662778336e-07, "loss": 0.1572, "step": 48480 }, { "epoch": 0.9507843137254902, "grad_norm": 1.2833175659179688, "learning_rate": 5.81673642979462e-07, "loss": 0.1575, "step": 48490 }, { "epoch": 0.9509803921568627, "grad_norm": 3.1361119747161865, "learning_rate": 5.815137110962127e-07, "loss": 0.1395, "step": 48500 }, { "epoch": 0.9511764705882353, "grad_norm": 1.5676275491714478, "learning_rate": 5.813537706448964e-07, "loss": 0.1469, "step": 48510 }, { "epoch": 0.9513725490196079, "grad_norm": 4.449748516082764, "learning_rate": 5.811938216423249e-07, "loss": 0.1468, "step": 48520 }, { "epoch": 0.9515686274509804, "grad_norm": 1.8463941812515259, "learning_rate": 5.810338641053105e-07, "loss": 0.1628, "step": 48530 }, { "epoch": 0.951764705882353, "grad_norm": 3.1234004497528076, "learning_rate": 5.808738980506667e-07, "loss": 0.1383, "step": 48540 }, { "epoch": 0.9519607843137254, "grad_norm": 3.020096778869629, "learning_rate": 5.807139234952082e-07, "loss": 0.1537, "step": 48550 }, { "epoch": 0.952156862745098, "grad_norm": 2.6236424446105957, "learning_rate": 5.805539404557498e-07, "loss": 0.1562, "step": 48560 }, { "epoch": 0.9523529411764706, "grad_norm": 1.6843942403793335, "learning_rate": 5.803939489491079e-07, "loss": 0.1306, "step": 48570 }, { "epoch": 0.9525490196078431, "grad_norm": 2.2044119834899902, "learning_rate": 5.802339489920993e-07, "loss": 0.1726, "step": 48580 }, { "epoch": 0.9527450980392157, "grad_norm": 2.5059099197387695, "learning_rate": 5.800739406015419e-07, "loss": 0.1376, "step": 48590 }, { "epoch": 0.9529411764705882, "grad_norm": 2.3307902812957764, "learning_rate": 5.79913923794255e-07, "loss": 0.1493, "step": 48600 }, { "epoch": 0.9531372549019608, "grad_norm": 2.9398608207702637, "learning_rate": 5.797538985870576e-07, "loss": 0.1683, "step": 48610 }, { "epoch": 0.9533333333333334, "grad_norm": 4.831429481506348, "learning_rate": 5.795938649967707e-07, "loss": 0.1565, "step": 48620 }, { "epoch": 0.9535294117647058, "grad_norm": 2.6019389629364014, "learning_rate": 5.794338230402156e-07, "loss": 0.1633, "step": 48630 }, { "epoch": 0.9537254901960784, "grad_norm": 2.8118932247161865, "learning_rate": 5.792737727342147e-07, "loss": 0.1546, "step": 48640 }, { "epoch": 0.953921568627451, "grad_norm": 1.5505790710449219, "learning_rate": 5.79113714095591e-07, "loss": 0.1181, "step": 48650 }, { "epoch": 0.9541176470588235, "grad_norm": 1.306450366973877, "learning_rate": 5.789536471411689e-07, "loss": 0.153, "step": 48660 }, { "epoch": 0.9543137254901961, "grad_norm": 2.476302146911621, "learning_rate": 5.787935718877732e-07, "loss": 0.1665, "step": 48670 }, { "epoch": 0.9545098039215686, "grad_norm": 2.369828462600708, "learning_rate": 5.786334883522294e-07, "loss": 0.1695, "step": 48680 }, { "epoch": 0.9547058823529412, "grad_norm": 3.0234453678131104, "learning_rate": 5.784733965513649e-07, "loss": 0.1441, "step": 48690 }, { "epoch": 0.9549019607843138, "grad_norm": 1.9724841117858887, "learning_rate": 5.783132965020066e-07, "loss": 0.1662, "step": 48700 }, { "epoch": 0.9550980392156863, "grad_norm": 2.9156734943389893, "learning_rate": 5.781531882209833e-07, "loss": 0.1359, "step": 48710 }, { "epoch": 0.9552941176470588, "grad_norm": 3.1353771686553955, "learning_rate": 5.779930717251242e-07, "loss": 0.1727, "step": 48720 }, { "epoch": 0.9554901960784313, "grad_norm": 2.2859344482421875, "learning_rate": 5.778329470312595e-07, "loss": 0.1266, "step": 48730 }, { "epoch": 0.9556862745098039, "grad_norm": 1.4125910997390747, "learning_rate": 5.776728141562202e-07, "loss": 0.1655, "step": 48740 }, { "epoch": 0.9558823529411765, "grad_norm": 2.2259559631347656, "learning_rate": 5.775126731168381e-07, "loss": 0.1326, "step": 48750 }, { "epoch": 0.956078431372549, "grad_norm": 3.6433029174804688, "learning_rate": 5.77352523929946e-07, "loss": 0.1822, "step": 48760 }, { "epoch": 0.9562745098039216, "grad_norm": 3.920541286468506, "learning_rate": 5.771923666123776e-07, "loss": 0.1877, "step": 48770 }, { "epoch": 0.9564705882352941, "grad_norm": 2.8983571529388428, "learning_rate": 5.770322011809672e-07, "loss": 0.1341, "step": 48780 }, { "epoch": 0.9566666666666667, "grad_norm": 2.264439105987549, "learning_rate": 5.768720276525502e-07, "loss": 0.1721, "step": 48790 }, { "epoch": 0.9568627450980393, "grad_norm": 3.9352970123291016, "learning_rate": 5.767118460439626e-07, "loss": 0.1675, "step": 48800 }, { "epoch": 0.9570588235294117, "grad_norm": 1.8320379257202148, "learning_rate": 5.765516563720416e-07, "loss": 0.1961, "step": 48810 }, { "epoch": 0.9572549019607843, "grad_norm": 4.821685791015625, "learning_rate": 5.763914586536251e-07, "loss": 0.1519, "step": 48820 }, { "epoch": 0.9574509803921568, "grad_norm": 2.2045786380767822, "learning_rate": 5.762312529055514e-07, "loss": 0.1712, "step": 48830 }, { "epoch": 0.9576470588235294, "grad_norm": 2.0654351711273193, "learning_rate": 5.760710391446604e-07, "loss": 0.1523, "step": 48840 }, { "epoch": 0.957843137254902, "grad_norm": 2.559389591217041, "learning_rate": 5.759108173877926e-07, "loss": 0.1413, "step": 48850 }, { "epoch": 0.9580392156862745, "grad_norm": 16.412940979003906, "learning_rate": 5.757505876517886e-07, "loss": 0.1565, "step": 48860 }, { "epoch": 0.9582352941176471, "grad_norm": 1.7237374782562256, "learning_rate": 5.755903499534913e-07, "loss": 0.1375, "step": 48870 }, { "epoch": 0.9584313725490196, "grad_norm": 1.3851453065872192, "learning_rate": 5.754301043097431e-07, "loss": 0.1619, "step": 48880 }, { "epoch": 0.9586274509803921, "grad_norm": 2.398818016052246, "learning_rate": 5.752698507373876e-07, "loss": 0.1413, "step": 48890 }, { "epoch": 0.9588235294117647, "grad_norm": 4.635507106781006, "learning_rate": 5.751095892532699e-07, "loss": 0.159, "step": 48900 }, { "epoch": 0.9590196078431372, "grad_norm": 2.4757301807403564, "learning_rate": 5.749493198742348e-07, "loss": 0.1255, "step": 48910 }, { "epoch": 0.9592156862745098, "grad_norm": 1.6313878297805786, "learning_rate": 5.74789042617129e-07, "loss": 0.1655, "step": 48920 }, { "epoch": 0.9594117647058824, "grad_norm": 2.960873603820801, "learning_rate": 5.746287574987995e-07, "loss": 0.1338, "step": 48930 }, { "epoch": 0.9596078431372549, "grad_norm": 2.301797866821289, "learning_rate": 5.744684645360938e-07, "loss": 0.1543, "step": 48940 }, { "epoch": 0.9598039215686275, "grad_norm": 2.391719341278076, "learning_rate": 5.743081637458611e-07, "loss": 0.1524, "step": 48950 }, { "epoch": 0.96, "grad_norm": 2.802246570587158, "learning_rate": 5.741478551449505e-07, "loss": 0.1509, "step": 48960 }, { "epoch": 0.9601960784313726, "grad_norm": 1.4984517097473145, "learning_rate": 5.739875387502127e-07, "loss": 0.1566, "step": 48970 }, { "epoch": 0.9603921568627451, "grad_norm": 1.6295645236968994, "learning_rate": 5.738272145784988e-07, "loss": 0.1611, "step": 48980 }, { "epoch": 0.9605882352941176, "grad_norm": 4.233633041381836, "learning_rate": 5.736668826466607e-07, "loss": 0.1699, "step": 48990 }, { "epoch": 0.9607843137254902, "grad_norm": 3.068105697631836, "learning_rate": 5.73506542971551e-07, "loss": 0.1687, "step": 49000 }, { "epoch": 0.9609803921568627, "grad_norm": 1.9322081804275513, "learning_rate": 5.733461955700238e-07, "loss": 0.1365, "step": 49010 }, { "epoch": 0.9611764705882353, "grad_norm": 1.7765012979507446, "learning_rate": 5.731858404589331e-07, "loss": 0.1307, "step": 49020 }, { "epoch": 0.9613725490196079, "grad_norm": 1.7361327409744263, "learning_rate": 5.730254776551344e-07, "loss": 0.1538, "step": 49030 }, { "epoch": 0.9615686274509804, "grad_norm": 2.743293285369873, "learning_rate": 5.728651071754837e-07, "loss": 0.1829, "step": 49040 }, { "epoch": 0.961764705882353, "grad_norm": 1.1274222135543823, "learning_rate": 5.727047290368378e-07, "loss": 0.1297, "step": 49050 }, { "epoch": 0.9619607843137254, "grad_norm": 4.544320106506348, "learning_rate": 5.725443432560545e-07, "loss": 0.1749, "step": 49060 }, { "epoch": 0.962156862745098, "grad_norm": 1.9323972463607788, "learning_rate": 5.723839498499921e-07, "loss": 0.1353, "step": 49070 }, { "epoch": 0.9623529411764706, "grad_norm": 1.9381628036499023, "learning_rate": 5.722235488355098e-07, "loss": 0.1629, "step": 49080 }, { "epoch": 0.9625490196078431, "grad_norm": 3.0699210166931152, "learning_rate": 5.72063140229468e-07, "loss": 0.1698, "step": 49090 }, { "epoch": 0.9627450980392157, "grad_norm": 2.449286699295044, "learning_rate": 5.719027240487273e-07, "loss": 0.1455, "step": 49100 }, { "epoch": 0.9629411764705882, "grad_norm": 5.420107364654541, "learning_rate": 5.717423003101494e-07, "loss": 0.1715, "step": 49110 }, { "epoch": 0.9631372549019608, "grad_norm": 2.076784610748291, "learning_rate": 5.715818690305968e-07, "loss": 0.151, "step": 49120 }, { "epoch": 0.9633333333333334, "grad_norm": 1.571752667427063, "learning_rate": 5.714214302269326e-07, "loss": 0.1653, "step": 49130 }, { "epoch": 0.9635294117647059, "grad_norm": 3.901000499725342, "learning_rate": 5.712609839160211e-07, "loss": 0.2034, "step": 49140 }, { "epoch": 0.9637254901960784, "grad_norm": 2.1208908557891846, "learning_rate": 5.71100530114727e-07, "loss": 0.1504, "step": 49150 }, { "epoch": 0.9639215686274509, "grad_norm": 1.3491193056106567, "learning_rate": 5.709400688399157e-07, "loss": 0.1277, "step": 49160 }, { "epoch": 0.9641176470588235, "grad_norm": 3.507784843444824, "learning_rate": 5.70779600108454e-07, "loss": 0.1536, "step": 49170 }, { "epoch": 0.9643137254901961, "grad_norm": 2.711554765701294, "learning_rate": 5.706191239372089e-07, "loss": 0.135, "step": 49180 }, { "epoch": 0.9645098039215686, "grad_norm": 2.7758262157440186, "learning_rate": 5.704586403430482e-07, "loss": 0.1254, "step": 49190 }, { "epoch": 0.9647058823529412, "grad_norm": 4.6616902351379395, "learning_rate": 5.702981493428412e-07, "loss": 0.1436, "step": 49200 }, { "epoch": 0.9649019607843137, "grad_norm": 3.971546173095703, "learning_rate": 5.701376509534566e-07, "loss": 0.1554, "step": 49210 }, { "epoch": 0.9650980392156863, "grad_norm": 2.7782325744628906, "learning_rate": 5.699771451917653e-07, "loss": 0.121, "step": 49220 }, { "epoch": 0.9652941176470589, "grad_norm": 3.1759915351867676, "learning_rate": 5.698166320746382e-07, "loss": 0.1403, "step": 49230 }, { "epoch": 0.9654901960784313, "grad_norm": 2.505966901779175, "learning_rate": 5.696561116189472e-07, "loss": 0.1435, "step": 49240 }, { "epoch": 0.9656862745098039, "grad_norm": 2.2113945484161377, "learning_rate": 5.694955838415649e-07, "loss": 0.1127, "step": 49250 }, { "epoch": 0.9658823529411765, "grad_norm": 2.252492904663086, "learning_rate": 5.693350487593647e-07, "loss": 0.1561, "step": 49260 }, { "epoch": 0.966078431372549, "grad_norm": 3.6019229888916016, "learning_rate": 5.691745063892206e-07, "loss": 0.1475, "step": 49270 }, { "epoch": 0.9662745098039216, "grad_norm": 2.063668966293335, "learning_rate": 5.690139567480078e-07, "loss": 0.1507, "step": 49280 }, { "epoch": 0.9664705882352941, "grad_norm": 13.019786834716797, "learning_rate": 5.688533998526018e-07, "loss": 0.1627, "step": 49290 }, { "epoch": 0.9666666666666667, "grad_norm": 2.004761219024658, "learning_rate": 5.68692835719879e-07, "loss": 0.1339, "step": 49300 }, { "epoch": 0.9668627450980393, "grad_norm": 2.0874786376953125, "learning_rate": 5.685322643667168e-07, "loss": 0.1185, "step": 49310 }, { "epoch": 0.9670588235294117, "grad_norm": 2.177140235900879, "learning_rate": 5.683716858099931e-07, "loss": 0.1556, "step": 49320 }, { "epoch": 0.9672549019607843, "grad_norm": 2.330493211746216, "learning_rate": 5.682111000665867e-07, "loss": 0.1317, "step": 49330 }, { "epoch": 0.9674509803921568, "grad_norm": 2.057570695877075, "learning_rate": 5.680505071533769e-07, "loss": 0.1321, "step": 49340 }, { "epoch": 0.9676470588235294, "grad_norm": 2.4627621173858643, "learning_rate": 5.678899070872441e-07, "loss": 0.1544, "step": 49350 }, { "epoch": 0.967843137254902, "grad_norm": 2.0704050064086914, "learning_rate": 5.677292998850691e-07, "loss": 0.1629, "step": 49360 }, { "epoch": 0.9680392156862745, "grad_norm": 2.439840078353882, "learning_rate": 5.675686855637339e-07, "loss": 0.1312, "step": 49370 }, { "epoch": 0.9682352941176471, "grad_norm": 1.94171142578125, "learning_rate": 5.674080641401207e-07, "loss": 0.1545, "step": 49380 }, { "epoch": 0.9684313725490196, "grad_norm": 2.1234025955200195, "learning_rate": 5.672474356311131e-07, "loss": 0.1253, "step": 49390 }, { "epoch": 0.9686274509803922, "grad_norm": 2.7136106491088867, "learning_rate": 5.670868000535946e-07, "loss": 0.1893, "step": 49400 }, { "epoch": 0.9688235294117648, "grad_norm": 2.9849307537078857, "learning_rate": 5.669261574244505e-07, "loss": 0.1265, "step": 49410 }, { "epoch": 0.9690196078431372, "grad_norm": 2.9338021278381348, "learning_rate": 5.667655077605659e-07, "loss": 0.1894, "step": 49420 }, { "epoch": 0.9692156862745098, "grad_norm": 1.9236648082733154, "learning_rate": 5.666048510788269e-07, "loss": 0.1393, "step": 49430 }, { "epoch": 0.9694117647058823, "grad_norm": 3.229118585586548, "learning_rate": 5.664441873961208e-07, "loss": 0.1417, "step": 49440 }, { "epoch": 0.9696078431372549, "grad_norm": 3.5577914714813232, "learning_rate": 5.66283516729335e-07, "loss": 0.1348, "step": 49450 }, { "epoch": 0.9698039215686275, "grad_norm": 2.1861941814422607, "learning_rate": 5.661228390953578e-07, "loss": 0.1478, "step": 49460 }, { "epoch": 0.97, "grad_norm": 3.3995022773742676, "learning_rate": 5.659621545110787e-07, "loss": 0.1472, "step": 49470 }, { "epoch": 0.9701960784313726, "grad_norm": 1.6540088653564453, "learning_rate": 5.658014629933875e-07, "loss": 0.1432, "step": 49480 }, { "epoch": 0.970392156862745, "grad_norm": 3.4421911239624023, "learning_rate": 5.656407645591745e-07, "loss": 0.1534, "step": 49490 }, { "epoch": 0.9705882352941176, "grad_norm": 5.254343509674072, "learning_rate": 5.654800592253314e-07, "loss": 0.1481, "step": 49500 }, { "epoch": 0.9707843137254902, "grad_norm": 2.9604384899139404, "learning_rate": 5.6531934700875e-07, "loss": 0.1819, "step": 49510 }, { "epoch": 0.9709803921568627, "grad_norm": 3.305797815322876, "learning_rate": 5.651586279263232e-07, "loss": 0.1608, "step": 49520 }, { "epoch": 0.9711764705882353, "grad_norm": 2.7572972774505615, "learning_rate": 5.649979019949445e-07, "loss": 0.1429, "step": 49530 }, { "epoch": 0.9713725490196078, "grad_norm": 3.0445823669433594, "learning_rate": 5.648371692315079e-07, "loss": 0.1235, "step": 49540 }, { "epoch": 0.9715686274509804, "grad_norm": 2.596099615097046, "learning_rate": 5.646764296529089e-07, "loss": 0.1595, "step": 49550 }, { "epoch": 0.971764705882353, "grad_norm": 3.077268600463867, "learning_rate": 5.645156832760425e-07, "loss": 0.143, "step": 49560 }, { "epoch": 0.9719607843137255, "grad_norm": 1.8642206192016602, "learning_rate": 5.643549301178055e-07, "loss": 0.1433, "step": 49570 }, { "epoch": 0.972156862745098, "grad_norm": 2.756087064743042, "learning_rate": 5.641941701950946e-07, "loss": 0.1362, "step": 49580 }, { "epoch": 0.9723529411764706, "grad_norm": 2.6560463905334473, "learning_rate": 5.64033403524808e-07, "loss": 0.1591, "step": 49590 }, { "epoch": 0.9725490196078431, "grad_norm": 2.5111758708953857, "learning_rate": 5.63872630123844e-07, "loss": 0.1679, "step": 49600 }, { "epoch": 0.9727450980392157, "grad_norm": 3.895082712173462, "learning_rate": 5.637118500091018e-07, "loss": 0.1739, "step": 49610 }, { "epoch": 0.9729411764705882, "grad_norm": 1.3962202072143555, "learning_rate": 5.635510631974813e-07, "loss": 0.1525, "step": 49620 }, { "epoch": 0.9731372549019608, "grad_norm": 2.170001745223999, "learning_rate": 5.633902697058834e-07, "loss": 0.1585, "step": 49630 }, { "epoch": 0.9733333333333334, "grad_norm": 1.8943711519241333, "learning_rate": 5.63229469551209e-07, "loss": 0.1374, "step": 49640 }, { "epoch": 0.9735294117647059, "grad_norm": 2.4152915477752686, "learning_rate": 5.630686627503603e-07, "loss": 0.1421, "step": 49650 }, { "epoch": 0.9737254901960785, "grad_norm": 2.05191969871521, "learning_rate": 5.629078493202402e-07, "loss": 0.1507, "step": 49660 }, { "epoch": 0.9739215686274509, "grad_norm": 1.6618961095809937, "learning_rate": 5.627470292777518e-07, "loss": 0.1633, "step": 49670 }, { "epoch": 0.9741176470588235, "grad_norm": 1.4879257678985596, "learning_rate": 5.625862026397995e-07, "loss": 0.1614, "step": 49680 }, { "epoch": 0.9743137254901961, "grad_norm": 3.632765769958496, "learning_rate": 5.624253694232878e-07, "loss": 0.1719, "step": 49690 }, { "epoch": 0.9745098039215686, "grad_norm": 3.0138962268829346, "learning_rate": 5.622645296451224e-07, "loss": 0.1595, "step": 49700 }, { "epoch": 0.9747058823529412, "grad_norm": 2.033719539642334, "learning_rate": 5.621036833222095e-07, "loss": 0.1558, "step": 49710 }, { "epoch": 0.9749019607843137, "grad_norm": 4.134287357330322, "learning_rate": 5.61942830471456e-07, "loss": 0.163, "step": 49720 }, { "epoch": 0.9750980392156863, "grad_norm": 3.6207940578460693, "learning_rate": 5.617819711097692e-07, "loss": 0.1617, "step": 49730 }, { "epoch": 0.9752941176470589, "grad_norm": 2.310589075088501, "learning_rate": 5.616211052540579e-07, "loss": 0.1727, "step": 49740 }, { "epoch": 0.9754901960784313, "grad_norm": 1.9012341499328613, "learning_rate": 5.614602329212303e-07, "loss": 0.1454, "step": 49750 }, { "epoch": 0.9756862745098039, "grad_norm": 1.621402621269226, "learning_rate": 5.612993541281963e-07, "loss": 0.1501, "step": 49760 }, { "epoch": 0.9758823529411764, "grad_norm": 3.8790950775146484, "learning_rate": 5.611384688918664e-07, "loss": 0.158, "step": 49770 }, { "epoch": 0.976078431372549, "grad_norm": 4.1167168617248535, "learning_rate": 5.609775772291514e-07, "loss": 0.1776, "step": 49780 }, { "epoch": 0.9762745098039216, "grad_norm": 3.3696391582489014, "learning_rate": 5.608166791569631e-07, "loss": 0.1684, "step": 49790 }, { "epoch": 0.9764705882352941, "grad_norm": 1.6629531383514404, "learning_rate": 5.606557746922135e-07, "loss": 0.1538, "step": 49800 }, { "epoch": 0.9766666666666667, "grad_norm": 2.4193217754364014, "learning_rate": 5.604948638518158e-07, "loss": 0.1617, "step": 49810 }, { "epoch": 0.9768627450980392, "grad_norm": 2.2635841369628906, "learning_rate": 5.603339466526836e-07, "loss": 0.1576, "step": 49820 }, { "epoch": 0.9770588235294118, "grad_norm": 2.2540738582611084, "learning_rate": 5.601730231117313e-07, "loss": 0.1528, "step": 49830 }, { "epoch": 0.9772549019607844, "grad_norm": 2.8713622093200684, "learning_rate": 5.600120932458737e-07, "loss": 0.1754, "step": 49840 }, { "epoch": 0.9774509803921568, "grad_norm": 3.181671380996704, "learning_rate": 5.598511570720267e-07, "loss": 0.1779, "step": 49850 }, { "epoch": 0.9776470588235294, "grad_norm": 3.2338759899139404, "learning_rate": 5.596902146071064e-07, "loss": 0.176, "step": 49860 }, { "epoch": 0.9778431372549019, "grad_norm": 2.8381905555725098, "learning_rate": 5.595292658680301e-07, "loss": 0.1653, "step": 49870 }, { "epoch": 0.9780392156862745, "grad_norm": 2.085678815841675, "learning_rate": 5.59368310871715e-07, "loss": 0.138, "step": 49880 }, { "epoch": 0.9782352941176471, "grad_norm": 2.188356637954712, "learning_rate": 5.592073496350796e-07, "loss": 0.1898, "step": 49890 }, { "epoch": 0.9784313725490196, "grad_norm": 4.242002010345459, "learning_rate": 5.59046382175043e-07, "loss": 0.2015, "step": 49900 }, { "epoch": 0.9786274509803922, "grad_norm": 4.584620475769043, "learning_rate": 5.588854085085248e-07, "loss": 0.172, "step": 49910 }, { "epoch": 0.9788235294117648, "grad_norm": 2.1931047439575195, "learning_rate": 5.587244286524448e-07, "loss": 0.1378, "step": 49920 }, { "epoch": 0.9790196078431372, "grad_norm": 2.699418544769287, "learning_rate": 5.585634426237246e-07, "loss": 0.1532, "step": 49930 }, { "epoch": 0.9792156862745098, "grad_norm": 1.061018466949463, "learning_rate": 5.584024504392851e-07, "loss": 0.1302, "step": 49940 }, { "epoch": 0.9794117647058823, "grad_norm": 1.448042392730713, "learning_rate": 5.58241452116049e-07, "loss": 0.1334, "step": 49950 }, { "epoch": 0.9796078431372549, "grad_norm": 2.184399127960205, "learning_rate": 5.580804476709388e-07, "loss": 0.1292, "step": 49960 }, { "epoch": 0.9798039215686275, "grad_norm": 3.6531124114990234, "learning_rate": 5.579194371208785e-07, "loss": 0.1376, "step": 49970 }, { "epoch": 0.98, "grad_norm": 2.9622459411621094, "learning_rate": 5.577584204827917e-07, "loss": 0.1564, "step": 49980 }, { "epoch": 0.9801960784313726, "grad_norm": 1.7787383794784546, "learning_rate": 5.575973977736033e-07, "loss": 0.1433, "step": 49990 }, { "epoch": 0.9803921568627451, "grad_norm": 1.3348379135131836, "learning_rate": 5.574363690102387e-07, "loss": 0.1265, "step": 50000 }, { "epoch": 0.9805882352941176, "grad_norm": 4.964189529418945, "learning_rate": 5.572753342096244e-07, "loss": 0.1303, "step": 50010 }, { "epoch": 0.9807843137254902, "grad_norm": 3.0688931941986084, "learning_rate": 5.571142933886865e-07, "loss": 0.1647, "step": 50020 }, { "epoch": 0.9809803921568627, "grad_norm": 3.3852450847625732, "learning_rate": 5.569532465643524e-07, "loss": 0.1416, "step": 50030 }, { "epoch": 0.9811764705882353, "grad_norm": 4.153070449829102, "learning_rate": 5.567921937535503e-07, "loss": 0.1613, "step": 50040 }, { "epoch": 0.9813725490196078, "grad_norm": 2.1189823150634766, "learning_rate": 5.566311349732087e-07, "loss": 0.1669, "step": 50050 }, { "epoch": 0.9815686274509804, "grad_norm": 3.0674591064453125, "learning_rate": 5.564700702402569e-07, "loss": 0.1518, "step": 50060 }, { "epoch": 0.981764705882353, "grad_norm": 2.7266829013824463, "learning_rate": 5.563089995716244e-07, "loss": 0.129, "step": 50070 }, { "epoch": 0.9819607843137255, "grad_norm": 1.85378897190094, "learning_rate": 5.56147922984242e-07, "loss": 0.1499, "step": 50080 }, { "epoch": 0.9821568627450981, "grad_norm": 2.7069034576416016, "learning_rate": 5.559868404950407e-07, "loss": 0.1477, "step": 50090 }, { "epoch": 0.9823529411764705, "grad_norm": 1.9975937604904175, "learning_rate": 5.55825752120952e-07, "loss": 0.133, "step": 50100 }, { "epoch": 0.9825490196078431, "grad_norm": 2.2315146923065186, "learning_rate": 5.556646578789083e-07, "loss": 0.1593, "step": 50110 }, { "epoch": 0.9827450980392157, "grad_norm": 2.155927896499634, "learning_rate": 5.555035577858428e-07, "loss": 0.1596, "step": 50120 }, { "epoch": 0.9829411764705882, "grad_norm": 2.8606276512145996, "learning_rate": 5.553424518586887e-07, "loss": 0.132, "step": 50130 }, { "epoch": 0.9831372549019608, "grad_norm": 2.2169153690338135, "learning_rate": 5.551813401143804e-07, "loss": 0.1403, "step": 50140 }, { "epoch": 0.9833333333333333, "grad_norm": 1.524489164352417, "learning_rate": 5.550202225698525e-07, "loss": 0.1764, "step": 50150 }, { "epoch": 0.9835294117647059, "grad_norm": 5.235471248626709, "learning_rate": 5.548590992420405e-07, "loss": 0.2114, "step": 50160 }, { "epoch": 0.9837254901960785, "grad_norm": 3.3184120655059814, "learning_rate": 5.546979701478803e-07, "loss": 0.1743, "step": 50170 }, { "epoch": 0.983921568627451, "grad_norm": 2.0534048080444336, "learning_rate": 5.545368353043086e-07, "loss": 0.1411, "step": 50180 }, { "epoch": 0.9841176470588235, "grad_norm": 2.367196798324585, "learning_rate": 5.543756947282625e-07, "loss": 0.1377, "step": 50190 }, { "epoch": 0.984313725490196, "grad_norm": 4.95438289642334, "learning_rate": 5.542145484366799e-07, "loss": 0.1781, "step": 50200 }, { "epoch": 0.9845098039215686, "grad_norm": 4.050638198852539, "learning_rate": 5.540533964464991e-07, "loss": 0.149, "step": 50210 }, { "epoch": 0.9847058823529412, "grad_norm": 1.7635890245437622, "learning_rate": 5.538922387746592e-07, "loss": 0.1373, "step": 50220 }, { "epoch": 0.9849019607843137, "grad_norm": 3.0682787895202637, "learning_rate": 5.537310754380997e-07, "loss": 0.1732, "step": 50230 }, { "epoch": 0.9850980392156863, "grad_norm": 2.1872920989990234, "learning_rate": 5.535699064537609e-07, "loss": 0.1154, "step": 50240 }, { "epoch": 0.9852941176470589, "grad_norm": 2.8988797664642334, "learning_rate": 5.534087318385836e-07, "loss": 0.1483, "step": 50250 }, { "epoch": 0.9854901960784314, "grad_norm": 2.1195740699768066, "learning_rate": 5.532475516095091e-07, "loss": 0.1489, "step": 50260 }, { "epoch": 0.985686274509804, "grad_norm": 3.6115190982818604, "learning_rate": 5.530863657834794e-07, "loss": 0.1543, "step": 50270 }, { "epoch": 0.9858823529411764, "grad_norm": 2.96779465675354, "learning_rate": 5.529251743774371e-07, "loss": 0.1538, "step": 50280 }, { "epoch": 0.986078431372549, "grad_norm": 3.0211427211761475, "learning_rate": 5.527639774083252e-07, "loss": 0.1478, "step": 50290 }, { "epoch": 0.9862745098039216, "grad_norm": 3.853816032409668, "learning_rate": 5.526027748930876e-07, "loss": 0.1437, "step": 50300 }, { "epoch": 0.9864705882352941, "grad_norm": 1.6041043996810913, "learning_rate": 5.524415668486685e-07, "loss": 0.1454, "step": 50310 }, { "epoch": 0.9866666666666667, "grad_norm": 3.863291025161743, "learning_rate": 5.522803532920128e-07, "loss": 0.1641, "step": 50320 }, { "epoch": 0.9868627450980392, "grad_norm": 2.2334797382354736, "learning_rate": 5.521191342400663e-07, "loss": 0.1526, "step": 50330 }, { "epoch": 0.9870588235294118, "grad_norm": 3.956347942352295, "learning_rate": 5.519579097097745e-07, "loss": 0.163, "step": 50340 }, { "epoch": 0.9872549019607844, "grad_norm": 3.6362102031707764, "learning_rate": 5.517966797180844e-07, "loss": 0.1436, "step": 50350 }, { "epoch": 0.9874509803921568, "grad_norm": 1.9747225046157837, "learning_rate": 5.516354442819431e-07, "loss": 0.1609, "step": 50360 }, { "epoch": 0.9876470588235294, "grad_norm": 2.539701461791992, "learning_rate": 5.514742034182984e-07, "loss": 0.1793, "step": 50370 }, { "epoch": 0.9878431372549019, "grad_norm": 1.8920650482177734, "learning_rate": 5.513129571440986e-07, "loss": 0.1791, "step": 50380 }, { "epoch": 0.9880392156862745, "grad_norm": 3.176257848739624, "learning_rate": 5.511517054762926e-07, "loss": 0.1502, "step": 50390 }, { "epoch": 0.9882352941176471, "grad_norm": 1.9291555881500244, "learning_rate": 5.5099044843183e-07, "loss": 0.117, "step": 50400 }, { "epoch": 0.9884313725490196, "grad_norm": 3.8406405448913574, "learning_rate": 5.508291860276608e-07, "loss": 0.1298, "step": 50410 }, { "epoch": 0.9886274509803922, "grad_norm": 1.8105745315551758, "learning_rate": 5.506679182807356e-07, "loss": 0.1629, "step": 50420 }, { "epoch": 0.9888235294117647, "grad_norm": 1.5281645059585571, "learning_rate": 5.505066452080054e-07, "loss": 0.1779, "step": 50430 }, { "epoch": 0.9890196078431372, "grad_norm": 4.207880973815918, "learning_rate": 5.503453668264222e-07, "loss": 0.1321, "step": 50440 }, { "epoch": 0.9892156862745098, "grad_norm": 2.633733034133911, "learning_rate": 5.501840831529381e-07, "loss": 0.1491, "step": 50450 }, { "epoch": 0.9894117647058823, "grad_norm": 2.1136159896850586, "learning_rate": 5.500227942045061e-07, "loss": 0.1468, "step": 50460 }, { "epoch": 0.9896078431372549, "grad_norm": 1.813647985458374, "learning_rate": 5.498614999980793e-07, "loss": 0.1463, "step": 50470 }, { "epoch": 0.9898039215686274, "grad_norm": 2.8043859004974365, "learning_rate": 5.497002005506121e-07, "loss": 0.1602, "step": 50480 }, { "epoch": 0.99, "grad_norm": 6.262386322021484, "learning_rate": 5.495388958790587e-07, "loss": 0.177, "step": 50490 }, { "epoch": 0.9901960784313726, "grad_norm": 2.424912214279175, "learning_rate": 5.493775860003742e-07, "loss": 0.1699, "step": 50500 }, { "epoch": 0.9903921568627451, "grad_norm": 1.4187840223312378, "learning_rate": 5.492162709315141e-07, "loss": 0.1327, "step": 50510 }, { "epoch": 0.9905882352941177, "grad_norm": 3.739084243774414, "learning_rate": 5.49054950689435e-07, "loss": 0.1523, "step": 50520 }, { "epoch": 0.9907843137254903, "grad_norm": 2.639892816543579, "learning_rate": 5.488936252910929e-07, "loss": 0.1698, "step": 50530 }, { "epoch": 0.9909803921568627, "grad_norm": 2.712197780609131, "learning_rate": 5.487322947534455e-07, "loss": 0.1476, "step": 50540 }, { "epoch": 0.9911764705882353, "grad_norm": 1.7042537927627563, "learning_rate": 5.485709590934506e-07, "loss": 0.1513, "step": 50550 }, { "epoch": 0.9913725490196078, "grad_norm": 3.0001018047332764, "learning_rate": 5.484096183280661e-07, "loss": 0.1505, "step": 50560 }, { "epoch": 0.9915686274509804, "grad_norm": 2.1296579837799072, "learning_rate": 5.482482724742514e-07, "loss": 0.1508, "step": 50570 }, { "epoch": 0.991764705882353, "grad_norm": 3.5735318660736084, "learning_rate": 5.480869215489654e-07, "loss": 0.1451, "step": 50580 }, { "epoch": 0.9919607843137255, "grad_norm": 3.0711677074432373, "learning_rate": 5.479255655691682e-07, "loss": 0.1692, "step": 50590 }, { "epoch": 0.9921568627450981, "grad_norm": 4.775904655456543, "learning_rate": 5.477642045518205e-07, "loss": 0.1624, "step": 50600 }, { "epoch": 0.9923529411764705, "grad_norm": 3.591322660446167, "learning_rate": 5.476028385138828e-07, "loss": 0.1742, "step": 50610 }, { "epoch": 0.9925490196078431, "grad_norm": 1.8222888708114624, "learning_rate": 5.474414674723168e-07, "loss": 0.2016, "step": 50620 }, { "epoch": 0.9927450980392157, "grad_norm": 2.1388399600982666, "learning_rate": 5.472800914440847e-07, "loss": 0.1609, "step": 50630 }, { "epoch": 0.9929411764705882, "grad_norm": 2.6746692657470703, "learning_rate": 5.471187104461489e-07, "loss": 0.143, "step": 50640 }, { "epoch": 0.9931372549019608, "grad_norm": 5.111700534820557, "learning_rate": 5.469573244954723e-07, "loss": 0.1688, "step": 50650 }, { "epoch": 0.9933333333333333, "grad_norm": 2.4530234336853027, "learning_rate": 5.467959336090189e-07, "loss": 0.1342, "step": 50660 }, { "epoch": 0.9935294117647059, "grad_norm": 4.0935821533203125, "learning_rate": 5.466345378037524e-07, "loss": 0.159, "step": 50670 }, { "epoch": 0.9937254901960785, "grad_norm": 2.2326951026916504, "learning_rate": 5.464731370966379e-07, "loss": 0.1456, "step": 50680 }, { "epoch": 0.993921568627451, "grad_norm": 4.282197952270508, "learning_rate": 5.463117315046401e-07, "loss": 0.162, "step": 50690 }, { "epoch": 0.9941176470588236, "grad_norm": 2.493539571762085, "learning_rate": 5.461503210447248e-07, "loss": 0.1429, "step": 50700 }, { "epoch": 0.994313725490196, "grad_norm": 3.208813428878784, "learning_rate": 5.459889057338583e-07, "loss": 0.1795, "step": 50710 }, { "epoch": 0.9945098039215686, "grad_norm": 2.5697519779205322, "learning_rate": 5.458274855890071e-07, "loss": 0.1465, "step": 50720 }, { "epoch": 0.9947058823529412, "grad_norm": 3.9323740005493164, "learning_rate": 5.456660606271385e-07, "loss": 0.1521, "step": 50730 }, { "epoch": 0.9949019607843137, "grad_norm": 2.597515106201172, "learning_rate": 5.455046308652201e-07, "loss": 0.1639, "step": 50740 }, { "epoch": 0.9950980392156863, "grad_norm": 4.411145210266113, "learning_rate": 5.453431963202203e-07, "loss": 0.1889, "step": 50750 }, { "epoch": 0.9952941176470588, "grad_norm": 4.51035213470459, "learning_rate": 5.451817570091077e-07, "loss": 0.1589, "step": 50760 }, { "epoch": 0.9954901960784314, "grad_norm": 2.2807724475860596, "learning_rate": 5.450203129488516e-07, "loss": 0.1496, "step": 50770 }, { "epoch": 0.995686274509804, "grad_norm": 3.0159175395965576, "learning_rate": 5.448588641564213e-07, "loss": 0.1972, "step": 50780 }, { "epoch": 0.9958823529411764, "grad_norm": 2.0986876487731934, "learning_rate": 5.446974106487875e-07, "loss": 0.1855, "step": 50790 }, { "epoch": 0.996078431372549, "grad_norm": 3.1978185176849365, "learning_rate": 5.445359524429205e-07, "loss": 0.1472, "step": 50800 }, { "epoch": 0.9962745098039215, "grad_norm": 2.7341148853302, "learning_rate": 5.443744895557918e-07, "loss": 0.1424, "step": 50810 }, { "epoch": 0.9964705882352941, "grad_norm": 2.404075860977173, "learning_rate": 5.44213022004373e-07, "loss": 0.1529, "step": 50820 }, { "epoch": 0.9966666666666667, "grad_norm": 3.142573595046997, "learning_rate": 5.44051549805636e-07, "loss": 0.1756, "step": 50830 }, { "epoch": 0.9968627450980392, "grad_norm": 3.6217262744903564, "learning_rate": 5.43890072976554e-07, "loss": 0.1873, "step": 50840 }, { "epoch": 0.9970588235294118, "grad_norm": 1.257285475730896, "learning_rate": 5.437285915340996e-07, "loss": 0.1106, "step": 50850 }, { "epoch": 0.9972549019607844, "grad_norm": 2.9801876544952393, "learning_rate": 5.435671054952468e-07, "loss": 0.1648, "step": 50860 }, { "epoch": 0.9974509803921568, "grad_norm": 4.229548454284668, "learning_rate": 5.434056148769696e-07, "loss": 0.1547, "step": 50870 }, { "epoch": 0.9976470588235294, "grad_norm": 2.242912530899048, "learning_rate": 5.432441196962426e-07, "loss": 0.134, "step": 50880 }, { "epoch": 0.9978431372549019, "grad_norm": 3.3086421489715576, "learning_rate": 5.430826199700406e-07, "loss": 0.1715, "step": 50890 }, { "epoch": 0.9980392156862745, "grad_norm": 2.615856647491455, "learning_rate": 5.429211157153398e-07, "loss": 0.1393, "step": 50900 }, { "epoch": 0.9982352941176471, "grad_norm": 3.0207812786102295, "learning_rate": 5.427596069491157e-07, "loss": 0.1644, "step": 50910 }, { "epoch": 0.9984313725490196, "grad_norm": 1.3115057945251465, "learning_rate": 5.425980936883449e-07, "loss": 0.1272, "step": 50920 }, { "epoch": 0.9986274509803922, "grad_norm": 3.7368040084838867, "learning_rate": 5.424365759500045e-07, "loss": 0.1339, "step": 50930 }, { "epoch": 0.9988235294117647, "grad_norm": 1.6264506578445435, "learning_rate": 5.422750537510717e-07, "loss": 0.1372, "step": 50940 }, { "epoch": 0.9990196078431373, "grad_norm": 3.674301862716675, "learning_rate": 5.421135271085248e-07, "loss": 0.1774, "step": 50950 }, { "epoch": 0.9992156862745099, "grad_norm": 2.0563595294952393, "learning_rate": 5.419519960393419e-07, "loss": 0.1491, "step": 50960 }, { "epoch": 0.9994117647058823, "grad_norm": 2.473778247833252, "learning_rate": 5.41790460560502e-07, "loss": 0.1476, "step": 50970 }, { "epoch": 0.9996078431372549, "grad_norm": 2.6238882541656494, "learning_rate": 5.416289206889843e-07, "loss": 0.1696, "step": 50980 }, { "epoch": 0.9998039215686274, "grad_norm": 2.288132429122925, "learning_rate": 5.414673764417686e-07, "loss": 0.1385, "step": 50990 }, { "epoch": 1.0, "grad_norm": 1.8299777507781982, "learning_rate": 5.413058278358352e-07, "loss": 0.1532, "step": 51000 }, { "epoch": 1.0001960784313726, "grad_norm": 2.2039947509765625, "learning_rate": 5.411442748881647e-07, "loss": 0.132, "step": 51010 }, { "epoch": 1.0003921568627452, "grad_norm": 2.7582764625549316, "learning_rate": 5.409827176157383e-07, "loss": 0.1652, "step": 51020 }, { "epoch": 1.0005882352941176, "grad_norm": 2.284914493560791, "learning_rate": 5.408211560355376e-07, "loss": 0.1293, "step": 51030 }, { "epoch": 1.0007843137254901, "grad_norm": 2.578050136566162, "learning_rate": 5.406595901645447e-07, "loss": 0.1188, "step": 51040 }, { "epoch": 1.0009803921568627, "grad_norm": 1.0844943523406982, "learning_rate": 5.40498020019742e-07, "loss": 0.1074, "step": 51050 }, { "epoch": 1.0011764705882353, "grad_norm": 2.6098408699035645, "learning_rate": 5.403364456181127e-07, "loss": 0.1108, "step": 51060 }, { "epoch": 1.001372549019608, "grad_norm": 2.4768829345703125, "learning_rate": 5.401748669766398e-07, "loss": 0.121, "step": 51070 }, { "epoch": 1.0015686274509803, "grad_norm": 1.8398700952529907, "learning_rate": 5.400132841123074e-07, "loss": 0.1141, "step": 51080 }, { "epoch": 1.001764705882353, "grad_norm": 1.1034631729125977, "learning_rate": 5.398516970420999e-07, "loss": 0.1196, "step": 51090 }, { "epoch": 1.0019607843137255, "grad_norm": 2.8769729137420654, "learning_rate": 5.396901057830017e-07, "loss": 0.1364, "step": 51100 }, { "epoch": 1.002156862745098, "grad_norm": 2.751586437225342, "learning_rate": 5.395285103519985e-07, "loss": 0.1474, "step": 51110 }, { "epoch": 1.0023529411764707, "grad_norm": 2.935525417327881, "learning_rate": 5.393669107660752e-07, "loss": 0.1248, "step": 51120 }, { "epoch": 1.002549019607843, "grad_norm": 1.5261945724487305, "learning_rate": 5.392053070422185e-07, "loss": 0.1301, "step": 51130 }, { "epoch": 1.0027450980392156, "grad_norm": 1.2391197681427002, "learning_rate": 5.390436991974145e-07, "loss": 0.1206, "step": 51140 }, { "epoch": 1.0029411764705882, "grad_norm": 2.522141218185425, "learning_rate": 5.388820872486504e-07, "loss": 0.1412, "step": 51150 }, { "epoch": 1.0031372549019608, "grad_norm": 1.515822410583496, "learning_rate": 5.387204712129132e-07, "loss": 0.1078, "step": 51160 }, { "epoch": 1.0033333333333334, "grad_norm": 1.6107531785964966, "learning_rate": 5.385588511071911e-07, "loss": 0.1161, "step": 51170 }, { "epoch": 1.0035294117647058, "grad_norm": 2.909897804260254, "learning_rate": 5.383972269484719e-07, "loss": 0.1433, "step": 51180 }, { "epoch": 1.0037254901960784, "grad_norm": 3.188661575317383, "learning_rate": 5.382355987537445e-07, "loss": 0.1466, "step": 51190 }, { "epoch": 1.003921568627451, "grad_norm": 3.033334255218506, "learning_rate": 5.380739665399978e-07, "loss": 0.124, "step": 51200 }, { "epoch": 1.0041176470588236, "grad_norm": 1.0874677896499634, "learning_rate": 5.379123303242215e-07, "loss": 0.1223, "step": 51210 }, { "epoch": 1.0043137254901962, "grad_norm": 2.309921979904175, "learning_rate": 5.377506901234053e-07, "loss": 0.12, "step": 51220 }, { "epoch": 1.0045098039215685, "grad_norm": 1.4894932508468628, "learning_rate": 5.375890459545397e-07, "loss": 0.1313, "step": 51230 }, { "epoch": 1.0047058823529411, "grad_norm": 0.990120530128479, "learning_rate": 5.374273978346151e-07, "loss": 0.114, "step": 51240 }, { "epoch": 1.0049019607843137, "grad_norm": 1.946066975593567, "learning_rate": 5.372657457806231e-07, "loss": 0.1195, "step": 51250 }, { "epoch": 1.0050980392156863, "grad_norm": 2.5025362968444824, "learning_rate": 5.371040898095549e-07, "loss": 0.1299, "step": 51260 }, { "epoch": 1.005294117647059, "grad_norm": 2.0536863803863525, "learning_rate": 5.369424299384027e-07, "loss": 0.1079, "step": 51270 }, { "epoch": 1.0054901960784313, "grad_norm": 2.0912086963653564, "learning_rate": 5.367807661841587e-07, "loss": 0.1368, "step": 51280 }, { "epoch": 1.0056862745098039, "grad_norm": 1.3346920013427734, "learning_rate": 5.366190985638158e-07, "loss": 0.1081, "step": 51290 }, { "epoch": 1.0058823529411764, "grad_norm": 2.0391299724578857, "learning_rate": 5.364574270943673e-07, "loss": 0.1118, "step": 51300 }, { "epoch": 1.006078431372549, "grad_norm": 1.9072707891464233, "learning_rate": 5.362957517928067e-07, "loss": 0.1442, "step": 51310 }, { "epoch": 1.0062745098039216, "grad_norm": 2.183769702911377, "learning_rate": 5.361340726761279e-07, "loss": 0.1119, "step": 51320 }, { "epoch": 1.0064705882352942, "grad_norm": 5.928725242614746, "learning_rate": 5.359723897613256e-07, "loss": 0.1186, "step": 51330 }, { "epoch": 1.0066666666666666, "grad_norm": 1.650256872177124, "learning_rate": 5.358107030653944e-07, "loss": 0.1315, "step": 51340 }, { "epoch": 1.0068627450980392, "grad_norm": 1.851478934288025, "learning_rate": 5.356490126053296e-07, "loss": 0.1202, "step": 51350 }, { "epoch": 1.0070588235294118, "grad_norm": 1.490527629852295, "learning_rate": 5.354873183981267e-07, "loss": 0.1083, "step": 51360 }, { "epoch": 1.0072549019607844, "grad_norm": 2.3123116493225098, "learning_rate": 5.353256204607815e-07, "loss": 0.126, "step": 51370 }, { "epoch": 1.007450980392157, "grad_norm": 0.9207360744476318, "learning_rate": 5.351639188102912e-07, "loss": 0.1115, "step": 51380 }, { "epoch": 1.0076470588235293, "grad_norm": 1.4911082983016968, "learning_rate": 5.350022134636515e-07, "loss": 0.1305, "step": 51390 }, { "epoch": 1.007843137254902, "grad_norm": 1.6998498439788818, "learning_rate": 5.348405044378604e-07, "loss": 0.1175, "step": 51400 }, { "epoch": 1.0080392156862745, "grad_norm": 1.8113360404968262, "learning_rate": 5.34678791749915e-07, "loss": 0.154, "step": 51410 }, { "epoch": 1.0082352941176471, "grad_norm": 1.8733060359954834, "learning_rate": 5.345170754168133e-07, "loss": 0.1145, "step": 51420 }, { "epoch": 1.0084313725490197, "grad_norm": 2.9458611011505127, "learning_rate": 5.343553554555539e-07, "loss": 0.123, "step": 51430 }, { "epoch": 1.008627450980392, "grad_norm": 2.935793399810791, "learning_rate": 5.341936318831353e-07, "loss": 0.135, "step": 51440 }, { "epoch": 1.0088235294117647, "grad_norm": 1.9673413038253784, "learning_rate": 5.340319047165564e-07, "loss": 0.1309, "step": 51450 }, { "epoch": 1.0090196078431373, "grad_norm": 3.186373233795166, "learning_rate": 5.33870173972817e-07, "loss": 0.1314, "step": 51460 }, { "epoch": 1.0092156862745099, "grad_norm": 2.0258872509002686, "learning_rate": 5.337084396689167e-07, "loss": 0.0917, "step": 51470 }, { "epoch": 1.0094117647058825, "grad_norm": 2.843658208847046, "learning_rate": 5.335467018218559e-07, "loss": 0.1245, "step": 51480 }, { "epoch": 1.0096078431372548, "grad_norm": 1.750109076499939, "learning_rate": 5.333849604486352e-07, "loss": 0.139, "step": 51490 }, { "epoch": 1.0098039215686274, "grad_norm": 2.4711103439331055, "learning_rate": 5.332232155662553e-07, "loss": 0.1005, "step": 51500 }, { "epoch": 1.01, "grad_norm": 1.3704826831817627, "learning_rate": 5.330614671917177e-07, "loss": 0.1234, "step": 51510 }, { "epoch": 1.0101960784313726, "grad_norm": 1.5139446258544922, "learning_rate": 5.32899715342024e-07, "loss": 0.1102, "step": 51520 }, { "epoch": 1.0103921568627452, "grad_norm": 1.3666597604751587, "learning_rate": 5.327379600341762e-07, "loss": 0.1154, "step": 51530 }, { "epoch": 1.0105882352941176, "grad_norm": 1.578452229499817, "learning_rate": 5.325762012851771e-07, "loss": 0.1229, "step": 51540 }, { "epoch": 1.0107843137254902, "grad_norm": 1.405012845993042, "learning_rate": 5.324144391120289e-07, "loss": 0.1536, "step": 51550 }, { "epoch": 1.0109803921568628, "grad_norm": 0.9541295766830444, "learning_rate": 5.322526735317353e-07, "loss": 0.1333, "step": 51560 }, { "epoch": 1.0111764705882353, "grad_norm": 2.822641134262085, "learning_rate": 5.320909045612997e-07, "loss": 0.1366, "step": 51570 }, { "epoch": 1.011372549019608, "grad_norm": 2.281449317932129, "learning_rate": 5.319291322177255e-07, "loss": 0.1228, "step": 51580 }, { "epoch": 1.0115686274509803, "grad_norm": 1.236219048500061, "learning_rate": 5.317673565180173e-07, "loss": 0.0922, "step": 51590 }, { "epoch": 1.011764705882353, "grad_norm": 2.527529239654541, "learning_rate": 5.316055774791798e-07, "loss": 0.0968, "step": 51600 }, { "epoch": 1.0119607843137255, "grad_norm": 2.4050638675689697, "learning_rate": 5.314437951182175e-07, "loss": 0.114, "step": 51610 }, { "epoch": 1.012156862745098, "grad_norm": 1.8794289827346802, "learning_rate": 5.312820094521358e-07, "loss": 0.1284, "step": 51620 }, { "epoch": 1.0123529411764707, "grad_norm": 2.6399593353271484, "learning_rate": 5.311202204979403e-07, "loss": 0.1414, "step": 51630 }, { "epoch": 1.012549019607843, "grad_norm": 2.395565986633301, "learning_rate": 5.309584282726373e-07, "loss": 0.1142, "step": 51640 }, { "epoch": 1.0127450980392156, "grad_norm": 1.486802101135254, "learning_rate": 5.307966327932327e-07, "loss": 0.1342, "step": 51650 }, { "epoch": 1.0129411764705882, "grad_norm": 1.8527172803878784, "learning_rate": 5.306348340767332e-07, "loss": 0.123, "step": 51660 }, { "epoch": 1.0131372549019608, "grad_norm": 3.916602849960327, "learning_rate": 5.304730321401459e-07, "loss": 0.1573, "step": 51670 }, { "epoch": 1.0133333333333334, "grad_norm": 3.431809663772583, "learning_rate": 5.303112270004782e-07, "loss": 0.1246, "step": 51680 }, { "epoch": 1.0135294117647058, "grad_norm": 0.4953574538230896, "learning_rate": 5.301494186747375e-07, "loss": 0.1187, "step": 51690 }, { "epoch": 1.0137254901960784, "grad_norm": 1.435222864151001, "learning_rate": 5.29987607179932e-07, "loss": 0.0947, "step": 51700 }, { "epoch": 1.013921568627451, "grad_norm": 1.0524333715438843, "learning_rate": 5.298257925330699e-07, "loss": 0.1376, "step": 51710 }, { "epoch": 1.0141176470588236, "grad_norm": 2.6328306198120117, "learning_rate": 5.296639747511599e-07, "loss": 0.123, "step": 51720 }, { "epoch": 1.0143137254901962, "grad_norm": 3.7000346183776855, "learning_rate": 5.29502153851211e-07, "loss": 0.1457, "step": 51730 }, { "epoch": 1.0145098039215685, "grad_norm": 3.458502769470215, "learning_rate": 5.293403298502324e-07, "loss": 0.1303, "step": 51740 }, { "epoch": 1.0147058823529411, "grad_norm": 1.490855097770691, "learning_rate": 5.29178502765234e-07, "loss": 0.0932, "step": 51750 }, { "epoch": 1.0149019607843137, "grad_norm": 1.8403115272521973, "learning_rate": 5.290166726132257e-07, "loss": 0.1106, "step": 51760 }, { "epoch": 1.0150980392156863, "grad_norm": 2.4610326290130615, "learning_rate": 5.288548394112175e-07, "loss": 0.123, "step": 51770 }, { "epoch": 1.015294117647059, "grad_norm": 1.479305386543274, "learning_rate": 5.286930031762203e-07, "loss": 0.1314, "step": 51780 }, { "epoch": 1.0154901960784313, "grad_norm": 2.6846201419830322, "learning_rate": 5.285311639252449e-07, "loss": 0.1278, "step": 51790 }, { "epoch": 1.0156862745098039, "grad_norm": 1.9630227088928223, "learning_rate": 5.283693216753025e-07, "loss": 0.099, "step": 51800 }, { "epoch": 1.0158823529411765, "grad_norm": 2.9315807819366455, "learning_rate": 5.282074764434048e-07, "loss": 0.1418, "step": 51810 }, { "epoch": 1.016078431372549, "grad_norm": 2.568726062774658, "learning_rate": 5.280456282465636e-07, "loss": 0.1193, "step": 51820 }, { "epoch": 1.0162745098039216, "grad_norm": 1.504319667816162, "learning_rate": 5.27883777101791e-07, "loss": 0.1288, "step": 51830 }, { "epoch": 1.016470588235294, "grad_norm": 2.294276714324951, "learning_rate": 5.277219230260997e-07, "loss": 0.1084, "step": 51840 }, { "epoch": 1.0166666666666666, "grad_norm": 2.220782518386841, "learning_rate": 5.275600660365023e-07, "loss": 0.1238, "step": 51850 }, { "epoch": 1.0168627450980392, "grad_norm": 1.3919798135757446, "learning_rate": 5.273982061500119e-07, "loss": 0.1277, "step": 51860 }, { "epoch": 1.0170588235294118, "grad_norm": 2.3191370964050293, "learning_rate": 5.272363433836423e-07, "loss": 0.1312, "step": 51870 }, { "epoch": 1.0172549019607844, "grad_norm": 2.572420120239258, "learning_rate": 5.270744777544067e-07, "loss": 0.1305, "step": 51880 }, { "epoch": 1.0174509803921568, "grad_norm": 1.6584218740463257, "learning_rate": 5.269126092793194e-07, "loss": 0.0886, "step": 51890 }, { "epoch": 1.0176470588235293, "grad_norm": 3.110861301422119, "learning_rate": 5.267507379753944e-07, "loss": 0.1323, "step": 51900 }, { "epoch": 1.017843137254902, "grad_norm": 2.71783447265625, "learning_rate": 5.265888638596469e-07, "loss": 0.1188, "step": 51910 }, { "epoch": 1.0180392156862745, "grad_norm": 2.6527099609375, "learning_rate": 5.264269869490913e-07, "loss": 0.1003, "step": 51920 }, { "epoch": 1.0182352941176471, "grad_norm": 1.8047453165054321, "learning_rate": 5.26265107260743e-07, "loss": 0.1467, "step": 51930 }, { "epoch": 1.0184313725490197, "grad_norm": 2.661431312561035, "learning_rate": 5.261032248116174e-07, "loss": 0.0892, "step": 51940 }, { "epoch": 1.018627450980392, "grad_norm": 2.898329019546509, "learning_rate": 5.259413396187304e-07, "loss": 0.127, "step": 51950 }, { "epoch": 1.0188235294117647, "grad_norm": 3.0909323692321777, "learning_rate": 5.25779451699098e-07, "loss": 0.1299, "step": 51960 }, { "epoch": 1.0190196078431373, "grad_norm": 2.796268939971924, "learning_rate": 5.256175610697365e-07, "loss": 0.11, "step": 51970 }, { "epoch": 1.0192156862745099, "grad_norm": 1.8842030763626099, "learning_rate": 5.254556677476629e-07, "loss": 0.103, "step": 51980 }, { "epoch": 1.0194117647058825, "grad_norm": 2.377671718597412, "learning_rate": 5.252937717498934e-07, "loss": 0.1081, "step": 51990 }, { "epoch": 1.0196078431372548, "grad_norm": 2.9637444019317627, "learning_rate": 5.25131873093446e-07, "loss": 0.1387, "step": 52000 }, { "epoch": 1.0198039215686274, "grad_norm": 3.5275866985321045, "learning_rate": 5.249699717953378e-07, "loss": 0.1309, "step": 52010 }, { "epoch": 1.02, "grad_norm": 3.633039712905884, "learning_rate": 5.248080678725863e-07, "loss": 0.1222, "step": 52020 }, { "epoch": 1.0201960784313726, "grad_norm": 2.8712856769561768, "learning_rate": 5.246461613422103e-07, "loss": 0.1041, "step": 52030 }, { "epoch": 1.0203921568627452, "grad_norm": 2.666276216506958, "learning_rate": 5.244842522212274e-07, "loss": 0.1099, "step": 52040 }, { "epoch": 1.0205882352941176, "grad_norm": 2.3895983695983887, "learning_rate": 5.243223405266563e-07, "loss": 0.1304, "step": 52050 }, { "epoch": 1.0207843137254902, "grad_norm": 1.9140113592147827, "learning_rate": 5.241604262755163e-07, "loss": 0.119, "step": 52060 }, { "epoch": 1.0209803921568628, "grad_norm": 2.508763313293457, "learning_rate": 5.239985094848262e-07, "loss": 0.1357, "step": 52070 }, { "epoch": 1.0211764705882354, "grad_norm": 3.0517826080322266, "learning_rate": 5.238365901716052e-07, "loss": 0.1147, "step": 52080 }, { "epoch": 1.021372549019608, "grad_norm": 3.1887831687927246, "learning_rate": 5.236746683528734e-07, "loss": 0.1346, "step": 52090 }, { "epoch": 1.0215686274509803, "grad_norm": 3.1859958171844482, "learning_rate": 5.235127440456503e-07, "loss": 0.1415, "step": 52100 }, { "epoch": 1.021764705882353, "grad_norm": 1.460944414138794, "learning_rate": 5.233508172669564e-07, "loss": 0.1214, "step": 52110 }, { "epoch": 1.0219607843137255, "grad_norm": 2.3519446849823, "learning_rate": 5.23188888033812e-07, "loss": 0.1342, "step": 52120 }, { "epoch": 1.022156862745098, "grad_norm": 3.126814126968384, "learning_rate": 5.230269563632379e-07, "loss": 0.1263, "step": 52130 }, { "epoch": 1.0223529411764707, "grad_norm": 1.6402215957641602, "learning_rate": 5.22865022272255e-07, "loss": 0.1133, "step": 52140 }, { "epoch": 1.022549019607843, "grad_norm": 1.7437149286270142, "learning_rate": 5.227030857778844e-07, "loss": 0.1148, "step": 52150 }, { "epoch": 1.0227450980392156, "grad_norm": 3.524247884750366, "learning_rate": 5.225411468971477e-07, "loss": 0.1308, "step": 52160 }, { "epoch": 1.0229411764705882, "grad_norm": 2.308882236480713, "learning_rate": 5.223792056470665e-07, "loss": 0.1099, "step": 52170 }, { "epoch": 1.0231372549019608, "grad_norm": 1.6349581480026245, "learning_rate": 5.22217262044663e-07, "loss": 0.0978, "step": 52180 }, { "epoch": 1.0233333333333334, "grad_norm": 3.1178488731384277, "learning_rate": 5.220553161069591e-07, "loss": 0.1284, "step": 52190 }, { "epoch": 1.0235294117647058, "grad_norm": 6.430595874786377, "learning_rate": 5.218933678509776e-07, "loss": 0.1173, "step": 52200 }, { "epoch": 1.0237254901960784, "grad_norm": 6.71305513381958, "learning_rate": 5.217314172937408e-07, "loss": 0.142, "step": 52210 }, { "epoch": 1.023921568627451, "grad_norm": 1.8634535074234009, "learning_rate": 5.215694644522722e-07, "loss": 0.0927, "step": 52220 }, { "epoch": 1.0241176470588236, "grad_norm": 1.0205146074295044, "learning_rate": 5.214075093435943e-07, "loss": 0.1425, "step": 52230 }, { "epoch": 1.0243137254901962, "grad_norm": 2.2494683265686035, "learning_rate": 5.212455519847309e-07, "loss": 0.1266, "step": 52240 }, { "epoch": 1.0245098039215685, "grad_norm": 2.556915521621704, "learning_rate": 5.210835923927057e-07, "loss": 0.1238, "step": 52250 }, { "epoch": 1.0247058823529411, "grad_norm": 1.5617311000823975, "learning_rate": 5.209216305845424e-07, "loss": 0.0947, "step": 52260 }, { "epoch": 1.0249019607843137, "grad_norm": 2.458904266357422, "learning_rate": 5.207596665772654e-07, "loss": 0.1306, "step": 52270 }, { "epoch": 1.0250980392156863, "grad_norm": 2.52626895904541, "learning_rate": 5.205977003878987e-07, "loss": 0.1084, "step": 52280 }, { "epoch": 1.025294117647059, "grad_norm": 2.8857035636901855, "learning_rate": 5.204357320334673e-07, "loss": 0.1245, "step": 52290 }, { "epoch": 1.0254901960784313, "grad_norm": 1.8379942178726196, "learning_rate": 5.202737615309957e-07, "loss": 0.1295, "step": 52300 }, { "epoch": 1.0256862745098039, "grad_norm": 3.1838979721069336, "learning_rate": 5.201117888975089e-07, "loss": 0.1113, "step": 52310 }, { "epoch": 1.0258823529411765, "grad_norm": 4.314934730529785, "learning_rate": 5.199498141500325e-07, "loss": 0.1157, "step": 52320 }, { "epoch": 1.026078431372549, "grad_norm": 1.4970860481262207, "learning_rate": 5.197878373055917e-07, "loss": 0.139, "step": 52330 }, { "epoch": 1.0262745098039217, "grad_norm": 4.2658185958862305, "learning_rate": 5.196258583812122e-07, "loss": 0.1199, "step": 52340 }, { "epoch": 1.026470588235294, "grad_norm": 2.753537178039551, "learning_rate": 5.194638773939201e-07, "loss": 0.1309, "step": 52350 }, { "epoch": 1.0266666666666666, "grad_norm": 1.487036108970642, "learning_rate": 5.193018943607413e-07, "loss": 0.111, "step": 52360 }, { "epoch": 1.0268627450980392, "grad_norm": 2.1634674072265625, "learning_rate": 5.191399092987024e-07, "loss": 0.1219, "step": 52370 }, { "epoch": 1.0270588235294118, "grad_norm": 2.9156572818756104, "learning_rate": 5.1897792222483e-07, "loss": 0.1225, "step": 52380 }, { "epoch": 1.0272549019607844, "grad_norm": 3.6339383125305176, "learning_rate": 5.188159331561505e-07, "loss": 0.114, "step": 52390 }, { "epoch": 1.0274509803921568, "grad_norm": 2.5447499752044678, "learning_rate": 5.186539421096914e-07, "loss": 0.1069, "step": 52400 }, { "epoch": 1.0276470588235294, "grad_norm": 3.056781053543091, "learning_rate": 5.184919491024796e-07, "loss": 0.112, "step": 52410 }, { "epoch": 1.027843137254902, "grad_norm": 1.7548162937164307, "learning_rate": 5.183299541515423e-07, "loss": 0.117, "step": 52420 }, { "epoch": 1.0280392156862745, "grad_norm": 1.2885265350341797, "learning_rate": 5.181679572739076e-07, "loss": 0.1235, "step": 52430 }, { "epoch": 1.0282352941176471, "grad_norm": 3.6603705883026123, "learning_rate": 5.18005958486603e-07, "loss": 0.0857, "step": 52440 }, { "epoch": 1.0284313725490195, "grad_norm": 6.117382049560547, "learning_rate": 5.178439578066564e-07, "loss": 0.117, "step": 52450 }, { "epoch": 1.028627450980392, "grad_norm": 2.1264564990997314, "learning_rate": 5.176819552510966e-07, "loss": 0.1195, "step": 52460 }, { "epoch": 1.0288235294117647, "grad_norm": 3.022953987121582, "learning_rate": 5.175199508369513e-07, "loss": 0.1565, "step": 52470 }, { "epoch": 1.0290196078431373, "grad_norm": 2.924329996109009, "learning_rate": 5.173579445812495e-07, "loss": 0.1376, "step": 52480 }, { "epoch": 1.0292156862745099, "grad_norm": 1.0740549564361572, "learning_rate": 5.171959365010199e-07, "loss": 0.1206, "step": 52490 }, { "epoch": 1.0294117647058822, "grad_norm": 3.0105104446411133, "learning_rate": 5.170339266132915e-07, "loss": 0.1196, "step": 52500 }, { "epoch": 1.0296078431372548, "grad_norm": 1.304787516593933, "learning_rate": 5.168719149350933e-07, "loss": 0.1061, "step": 52510 }, { "epoch": 1.0298039215686274, "grad_norm": 1.7193729877471924, "learning_rate": 5.16709901483455e-07, "loss": 0.1373, "step": 52520 }, { "epoch": 1.03, "grad_norm": 3.5114901065826416, "learning_rate": 5.165478862754059e-07, "loss": 0.1078, "step": 52530 }, { "epoch": 1.0301960784313726, "grad_norm": 2.7927517890930176, "learning_rate": 5.163858693279758e-07, "loss": 0.1112, "step": 52540 }, { "epoch": 1.030392156862745, "grad_norm": 2.2518839836120605, "learning_rate": 5.162238506581948e-07, "loss": 0.1413, "step": 52550 }, { "epoch": 1.0305882352941176, "grad_norm": 2.490297794342041, "learning_rate": 5.160618302830928e-07, "loss": 0.1285, "step": 52560 }, { "epoch": 1.0307843137254902, "grad_norm": 2.430802583694458, "learning_rate": 5.158998082197003e-07, "loss": 0.1593, "step": 52570 }, { "epoch": 1.0309803921568628, "grad_norm": 3.8084876537323, "learning_rate": 5.157377844850472e-07, "loss": 0.126, "step": 52580 }, { "epoch": 1.0311764705882354, "grad_norm": 2.2640888690948486, "learning_rate": 5.155757590961647e-07, "loss": 0.1125, "step": 52590 }, { "epoch": 1.0313725490196077, "grad_norm": 3.6367175579071045, "learning_rate": 5.154137320700836e-07, "loss": 0.1182, "step": 52600 }, { "epoch": 1.0315686274509803, "grad_norm": 5.215725898742676, "learning_rate": 5.152517034238345e-07, "loss": 0.1409, "step": 52610 }, { "epoch": 1.031764705882353, "grad_norm": 1.9436010122299194, "learning_rate": 5.150896731744488e-07, "loss": 0.0941, "step": 52620 }, { "epoch": 1.0319607843137255, "grad_norm": 3.6732358932495117, "learning_rate": 5.149276413389578e-07, "loss": 0.1311, "step": 52630 }, { "epoch": 1.032156862745098, "grad_norm": 2.79223370552063, "learning_rate": 5.147656079343931e-07, "loss": 0.1198, "step": 52640 }, { "epoch": 1.0323529411764707, "grad_norm": 2.1467931270599365, "learning_rate": 5.146035729777861e-07, "loss": 0.1473, "step": 52650 }, { "epoch": 1.032549019607843, "grad_norm": 1.4937515258789062, "learning_rate": 5.144415364861688e-07, "loss": 0.1218, "step": 52660 }, { "epoch": 1.0327450980392157, "grad_norm": 2.0893630981445312, "learning_rate": 5.14279498476573e-07, "loss": 0.1066, "step": 52670 }, { "epoch": 1.0329411764705883, "grad_norm": 1.8628195524215698, "learning_rate": 5.14117458966031e-07, "loss": 0.1308, "step": 52680 }, { "epoch": 1.0331372549019608, "grad_norm": 2.684520721435547, "learning_rate": 5.139554179715751e-07, "loss": 0.0999, "step": 52690 }, { "epoch": 1.0333333333333334, "grad_norm": 1.3978774547576904, "learning_rate": 5.137933755102375e-07, "loss": 0.1149, "step": 52700 }, { "epoch": 1.0335294117647058, "grad_norm": 1.423580288887024, "learning_rate": 5.13631331599051e-07, "loss": 0.1276, "step": 52710 }, { "epoch": 1.0337254901960784, "grad_norm": 1.957344889640808, "learning_rate": 5.134692862550485e-07, "loss": 0.1241, "step": 52720 }, { "epoch": 1.033921568627451, "grad_norm": 1.8452372550964355, "learning_rate": 5.133072394952627e-07, "loss": 0.0976, "step": 52730 }, { "epoch": 1.0341176470588236, "grad_norm": 1.9686404466629028, "learning_rate": 5.131451913367268e-07, "loss": 0.1108, "step": 52740 }, { "epoch": 1.0343137254901962, "grad_norm": 2.8307082653045654, "learning_rate": 5.129831417964737e-07, "loss": 0.1422, "step": 52750 }, { "epoch": 1.0345098039215685, "grad_norm": 1.962538719177246, "learning_rate": 5.128210908915372e-07, "loss": 0.1242, "step": 52760 }, { "epoch": 1.0347058823529411, "grad_norm": 1.2341598272323608, "learning_rate": 5.126590386389503e-07, "loss": 0.1129, "step": 52770 }, { "epoch": 1.0349019607843137, "grad_norm": 2.204951286315918, "learning_rate": 5.124969850557469e-07, "loss": 0.0927, "step": 52780 }, { "epoch": 1.0350980392156863, "grad_norm": 0.9508833289146423, "learning_rate": 5.123349301589609e-07, "loss": 0.1149, "step": 52790 }, { "epoch": 1.035294117647059, "grad_norm": 1.669791579246521, "learning_rate": 5.121728739656258e-07, "loss": 0.1287, "step": 52800 }, { "epoch": 1.0354901960784313, "grad_norm": 2.7394516468048096, "learning_rate": 5.120108164927762e-07, "loss": 0.1347, "step": 52810 }, { "epoch": 1.0356862745098039, "grad_norm": 3.794827461242676, "learning_rate": 5.118487577574458e-07, "loss": 0.1246, "step": 52820 }, { "epoch": 1.0358823529411765, "grad_norm": 3.154978036880493, "learning_rate": 5.116866977766692e-07, "loss": 0.1228, "step": 52830 }, { "epoch": 1.036078431372549, "grad_norm": 3.593611001968384, "learning_rate": 5.115246365674807e-07, "loss": 0.1212, "step": 52840 }, { "epoch": 1.0362745098039217, "grad_norm": 2.4787681102752686, "learning_rate": 5.113625741469149e-07, "loss": 0.1225, "step": 52850 }, { "epoch": 1.036470588235294, "grad_norm": 2.3508851528167725, "learning_rate": 5.112005105320065e-07, "loss": 0.1277, "step": 52860 }, { "epoch": 1.0366666666666666, "grad_norm": 3.4172351360321045, "learning_rate": 5.110384457397905e-07, "loss": 0.1071, "step": 52870 }, { "epoch": 1.0368627450980392, "grad_norm": 1.0851645469665527, "learning_rate": 5.108763797873015e-07, "loss": 0.1356, "step": 52880 }, { "epoch": 1.0370588235294118, "grad_norm": 4.693179130554199, "learning_rate": 5.107143126915747e-07, "loss": 0.1245, "step": 52890 }, { "epoch": 1.0372549019607844, "grad_norm": 2.0366265773773193, "learning_rate": 5.105522444696455e-07, "loss": 0.1126, "step": 52900 }, { "epoch": 1.0374509803921568, "grad_norm": 2.146643877029419, "learning_rate": 5.10390175138549e-07, "loss": 0.1025, "step": 52910 }, { "epoch": 1.0376470588235294, "grad_norm": 1.416122317314148, "learning_rate": 5.102281047153209e-07, "loss": 0.1214, "step": 52920 }, { "epoch": 1.037843137254902, "grad_norm": 2.3701608180999756, "learning_rate": 5.100660332169963e-07, "loss": 0.1216, "step": 52930 }, { "epoch": 1.0380392156862746, "grad_norm": 2.6469483375549316, "learning_rate": 5.09903960660611e-07, "loss": 0.1171, "step": 52940 }, { "epoch": 1.0382352941176471, "grad_norm": 2.388230323791504, "learning_rate": 5.097418870632012e-07, "loss": 0.1122, "step": 52950 }, { "epoch": 1.0384313725490195, "grad_norm": 4.872939109802246, "learning_rate": 5.095798124418021e-07, "loss": 0.1154, "step": 52960 }, { "epoch": 1.038627450980392, "grad_norm": 2.7294185161590576, "learning_rate": 5.094177368134502e-07, "loss": 0.148, "step": 52970 }, { "epoch": 1.0388235294117647, "grad_norm": 2.2563400268554688, "learning_rate": 5.092556601951814e-07, "loss": 0.1368, "step": 52980 }, { "epoch": 1.0390196078431373, "grad_norm": 2.5893049240112305, "learning_rate": 5.090935826040318e-07, "loss": 0.1154, "step": 52990 }, { "epoch": 1.0392156862745099, "grad_norm": 1.71278715133667, "learning_rate": 5.089315040570379e-07, "loss": 0.0905, "step": 53000 }, { "epoch": 1.0394117647058823, "grad_norm": 1.432680368423462, "learning_rate": 5.08769424571236e-07, "loss": 0.097, "step": 53010 }, { "epoch": 1.0396078431372549, "grad_norm": 2.2749624252319336, "learning_rate": 5.086073441636624e-07, "loss": 0.1378, "step": 53020 }, { "epoch": 1.0398039215686274, "grad_norm": 2.7587692737579346, "learning_rate": 5.08445262851354e-07, "loss": 0.123, "step": 53030 }, { "epoch": 1.04, "grad_norm": 3.66667103767395, "learning_rate": 5.082831806513474e-07, "loss": 0.1336, "step": 53040 }, { "epoch": 1.0401960784313726, "grad_norm": 1.7277958393096924, "learning_rate": 5.081210975806792e-07, "loss": 0.1341, "step": 53050 }, { "epoch": 1.040392156862745, "grad_norm": 2.0278608798980713, "learning_rate": 5.079590136563865e-07, "loss": 0.1305, "step": 53060 }, { "epoch": 1.0405882352941176, "grad_norm": 1.372604250907898, "learning_rate": 5.077969288955062e-07, "loss": 0.1052, "step": 53070 }, { "epoch": 1.0407843137254902, "grad_norm": 3.1764702796936035, "learning_rate": 5.076348433150753e-07, "loss": 0.1162, "step": 53080 }, { "epoch": 1.0409803921568628, "grad_norm": 4.2754693031311035, "learning_rate": 5.07472756932131e-07, "loss": 0.1253, "step": 53090 }, { "epoch": 1.0411764705882354, "grad_norm": 1.9200197458267212, "learning_rate": 5.073106697637104e-07, "loss": 0.1222, "step": 53100 }, { "epoch": 1.0413725490196077, "grad_norm": 2.0522255897521973, "learning_rate": 5.07148581826851e-07, "loss": 0.0995, "step": 53110 }, { "epoch": 1.0415686274509803, "grad_norm": 1.6780035495758057, "learning_rate": 5.069864931385899e-07, "loss": 0.1138, "step": 53120 }, { "epoch": 1.041764705882353, "grad_norm": 3.0444204807281494, "learning_rate": 5.068244037159648e-07, "loss": 0.1274, "step": 53130 }, { "epoch": 1.0419607843137255, "grad_norm": 2.822265386581421, "learning_rate": 5.066623135760133e-07, "loss": 0.1412, "step": 53140 }, { "epoch": 1.0421568627450981, "grad_norm": 1.5877699851989746, "learning_rate": 5.065002227357725e-07, "loss": 0.1422, "step": 53150 }, { "epoch": 1.0423529411764705, "grad_norm": 2.41648530960083, "learning_rate": 5.063381312122808e-07, "loss": 0.1071, "step": 53160 }, { "epoch": 1.042549019607843, "grad_norm": 3.764503002166748, "learning_rate": 5.061760390225756e-07, "loss": 0.1194, "step": 53170 }, { "epoch": 1.0427450980392157, "grad_norm": 1.9822940826416016, "learning_rate": 5.060139461836945e-07, "loss": 0.1083, "step": 53180 }, { "epoch": 1.0429411764705883, "grad_norm": 1.5672938823699951, "learning_rate": 5.05851852712676e-07, "loss": 0.1289, "step": 53190 }, { "epoch": 1.0431372549019609, "grad_norm": 1.7510532140731812, "learning_rate": 5.056897586265575e-07, "loss": 0.1086, "step": 53200 }, { "epoch": 1.0433333333333334, "grad_norm": 3.1909713745117188, "learning_rate": 5.055276639423772e-07, "loss": 0.1411, "step": 53210 }, { "epoch": 1.0435294117647058, "grad_norm": 2.3146920204162598, "learning_rate": 5.053655686771735e-07, "loss": 0.1235, "step": 53220 }, { "epoch": 1.0437254901960784, "grad_norm": 2.675774097442627, "learning_rate": 5.05203472847984e-07, "loss": 0.1165, "step": 53230 }, { "epoch": 1.043921568627451, "grad_norm": 1.6503347158432007, "learning_rate": 5.050413764718473e-07, "loss": 0.1155, "step": 53240 }, { "epoch": 1.0441176470588236, "grad_norm": 1.6026123762130737, "learning_rate": 5.048792795658015e-07, "loss": 0.1186, "step": 53250 }, { "epoch": 1.0443137254901962, "grad_norm": 2.0869178771972656, "learning_rate": 5.047171821468852e-07, "loss": 0.1125, "step": 53260 }, { "epoch": 1.0445098039215686, "grad_norm": 2.766934633255005, "learning_rate": 5.045550842321366e-07, "loss": 0.1172, "step": 53270 }, { "epoch": 1.0447058823529412, "grad_norm": 2.834468126296997, "learning_rate": 5.043929858385939e-07, "loss": 0.1168, "step": 53280 }, { "epoch": 1.0449019607843137, "grad_norm": 1.6227750778198242, "learning_rate": 5.042308869832959e-07, "loss": 0.1243, "step": 53290 }, { "epoch": 1.0450980392156863, "grad_norm": 2.276259422302246, "learning_rate": 5.04068787683281e-07, "loss": 0.1173, "step": 53300 }, { "epoch": 1.045294117647059, "grad_norm": 3.041288137435913, "learning_rate": 5.039066879555878e-07, "loss": 0.1192, "step": 53310 }, { "epoch": 1.0454901960784313, "grad_norm": 5.005396366119385, "learning_rate": 5.037445878172549e-07, "loss": 0.1293, "step": 53320 }, { "epoch": 1.045686274509804, "grad_norm": 4.931292533874512, "learning_rate": 5.03582487285321e-07, "loss": 0.1363, "step": 53330 }, { "epoch": 1.0458823529411765, "grad_norm": 2.752192974090576, "learning_rate": 5.034203863768247e-07, "loss": 0.1254, "step": 53340 }, { "epoch": 1.046078431372549, "grad_norm": 2.218362331390381, "learning_rate": 5.03258285108805e-07, "loss": 0.1287, "step": 53350 }, { "epoch": 1.0462745098039217, "grad_norm": 2.9134135246276855, "learning_rate": 5.030961834983004e-07, "loss": 0.1018, "step": 53360 }, { "epoch": 1.046470588235294, "grad_norm": 2.006577253341675, "learning_rate": 5.0293408156235e-07, "loss": 0.1269, "step": 53370 }, { "epoch": 1.0466666666666666, "grad_norm": 4.338216781616211, "learning_rate": 5.027719793179924e-07, "loss": 0.1542, "step": 53380 }, { "epoch": 1.0468627450980392, "grad_norm": 2.5806546211242676, "learning_rate": 5.026098767822665e-07, "loss": 0.121, "step": 53390 }, { "epoch": 1.0470588235294118, "grad_norm": 1.7765642404556274, "learning_rate": 5.024477739722114e-07, "loss": 0.0968, "step": 53400 }, { "epoch": 1.0472549019607844, "grad_norm": 2.674435615539551, "learning_rate": 5.022856709048661e-07, "loss": 0.094, "step": 53410 }, { "epoch": 1.0474509803921568, "grad_norm": 1.7396690845489502, "learning_rate": 5.02123567597269e-07, "loss": 0.1172, "step": 53420 }, { "epoch": 1.0476470588235294, "grad_norm": 2.2034199237823486, "learning_rate": 5.019614640664599e-07, "loss": 0.1381, "step": 53430 }, { "epoch": 1.047843137254902, "grad_norm": 1.656175136566162, "learning_rate": 5.017993603294773e-07, "loss": 0.11, "step": 53440 }, { "epoch": 1.0480392156862746, "grad_norm": 2.3517141342163086, "learning_rate": 5.016372564033604e-07, "loss": 0.1353, "step": 53450 }, { "epoch": 1.0482352941176472, "grad_norm": 3.6779749393463135, "learning_rate": 5.014751523051483e-07, "loss": 0.1484, "step": 53460 }, { "epoch": 1.0484313725490195, "grad_norm": 2.72495174407959, "learning_rate": 5.0131304805188e-07, "loss": 0.1572, "step": 53470 }, { "epoch": 1.0486274509803921, "grad_norm": 3.238506317138672, "learning_rate": 5.011509436605946e-07, "loss": 0.1075, "step": 53480 }, { "epoch": 1.0488235294117647, "grad_norm": 2.95654034614563, "learning_rate": 5.009888391483314e-07, "loss": 0.1176, "step": 53490 }, { "epoch": 1.0490196078431373, "grad_norm": 2.0376462936401367, "learning_rate": 5.008267345321291e-07, "loss": 0.1134, "step": 53500 }, { "epoch": 1.04921568627451, "grad_norm": 1.391294240951538, "learning_rate": 5.006646298290272e-07, "loss": 0.1179, "step": 53510 }, { "epoch": 1.0494117647058823, "grad_norm": 2.8294551372528076, "learning_rate": 5.005025250560649e-07, "loss": 0.1261, "step": 53520 }, { "epoch": 1.0496078431372549, "grad_norm": 3.556178569793701, "learning_rate": 5.003404202302809e-07, "loss": 0.1166, "step": 53530 }, { "epoch": 1.0498039215686275, "grad_norm": 1.17848801612854, "learning_rate": 5.001783153687149e-07, "loss": 0.1121, "step": 53540 }, { "epoch": 1.05, "grad_norm": 4.0703654289245605, "learning_rate": 5.000162104884058e-07, "loss": 0.1025, "step": 53550 }, { "epoch": 1.0501960784313726, "grad_norm": 1.5639246702194214, "learning_rate": 4.998541056063927e-07, "loss": 0.1301, "step": 53560 }, { "epoch": 1.050392156862745, "grad_norm": 1.8187475204467773, "learning_rate": 4.996920007397149e-07, "loss": 0.1056, "step": 53570 }, { "epoch": 1.0505882352941176, "grad_norm": 2.134652853012085, "learning_rate": 4.995298959054114e-07, "loss": 0.1352, "step": 53580 }, { "epoch": 1.0507843137254902, "grad_norm": 1.4692163467407227, "learning_rate": 4.993677911205217e-07, "loss": 0.1214, "step": 53590 }, { "epoch": 1.0509803921568628, "grad_norm": 2.2606022357940674, "learning_rate": 4.992056864020845e-07, "loss": 0.1107, "step": 53600 }, { "epoch": 1.0511764705882354, "grad_norm": 1.1186784505844116, "learning_rate": 4.990435817671392e-07, "loss": 0.1434, "step": 53610 }, { "epoch": 1.0513725490196077, "grad_norm": 1.5811764001846313, "learning_rate": 4.988814772327251e-07, "loss": 0.1218, "step": 53620 }, { "epoch": 1.0515686274509803, "grad_norm": 5.067830562591553, "learning_rate": 4.987193728158809e-07, "loss": 0.1241, "step": 53630 }, { "epoch": 1.051764705882353, "grad_norm": 2.196347951889038, "learning_rate": 4.985572685336461e-07, "loss": 0.1036, "step": 53640 }, { "epoch": 1.0519607843137255, "grad_norm": 1.8917632102966309, "learning_rate": 4.983951644030597e-07, "loss": 0.1064, "step": 53650 }, { "epoch": 1.0521568627450981, "grad_norm": 1.0450140237808228, "learning_rate": 4.982330604411605e-07, "loss": 0.1366, "step": 53660 }, { "epoch": 1.0523529411764705, "grad_norm": 1.19893217086792, "learning_rate": 4.980709566649879e-07, "loss": 0.1103, "step": 53670 }, { "epoch": 1.052549019607843, "grad_norm": 3.1959033012390137, "learning_rate": 4.979088530915811e-07, "loss": 0.1205, "step": 53680 }, { "epoch": 1.0527450980392157, "grad_norm": 4.028618812561035, "learning_rate": 4.977467497379786e-07, "loss": 0.1167, "step": 53690 }, { "epoch": 1.0529411764705883, "grad_norm": 2.069601058959961, "learning_rate": 4.975846466212197e-07, "loss": 0.1131, "step": 53700 }, { "epoch": 1.0531372549019609, "grad_norm": 2.3703393936157227, "learning_rate": 4.974225437583434e-07, "loss": 0.14, "step": 53710 }, { "epoch": 1.0533333333333332, "grad_norm": 1.6142152547836304, "learning_rate": 4.972604411663887e-07, "loss": 0.1085, "step": 53720 }, { "epoch": 1.0535294117647058, "grad_norm": 2.7219841480255127, "learning_rate": 4.970983388623942e-07, "loss": 0.1203, "step": 53730 }, { "epoch": 1.0537254901960784, "grad_norm": 2.0587923526763916, "learning_rate": 4.969362368633991e-07, "loss": 0.1345, "step": 53740 }, { "epoch": 1.053921568627451, "grad_norm": 2.374295473098755, "learning_rate": 4.967741351864424e-07, "loss": 0.1118, "step": 53750 }, { "epoch": 1.0541176470588236, "grad_norm": 2.8391470909118652, "learning_rate": 4.966120338485624e-07, "loss": 0.1107, "step": 53760 }, { "epoch": 1.054313725490196, "grad_norm": 2.7321078777313232, "learning_rate": 4.964499328667985e-07, "loss": 0.1071, "step": 53770 }, { "epoch": 1.0545098039215686, "grad_norm": 3.4814422130584717, "learning_rate": 4.962878322581891e-07, "loss": 0.1097, "step": 53780 }, { "epoch": 1.0547058823529412, "grad_norm": 3.9524595737457275, "learning_rate": 4.961257320397729e-07, "loss": 0.1548, "step": 53790 }, { "epoch": 1.0549019607843138, "grad_norm": 5.57422399520874, "learning_rate": 4.959636322285886e-07, "loss": 0.1526, "step": 53800 }, { "epoch": 1.0550980392156863, "grad_norm": 1.848609209060669, "learning_rate": 4.958015328416751e-07, "loss": 0.1074, "step": 53810 }, { "epoch": 1.0552941176470587, "grad_norm": 2.707106113433838, "learning_rate": 4.956394338960705e-07, "loss": 0.1181, "step": 53820 }, { "epoch": 1.0554901960784313, "grad_norm": 1.79198157787323, "learning_rate": 4.954773354088137e-07, "loss": 0.1597, "step": 53830 }, { "epoch": 1.055686274509804, "grad_norm": 2.7858450412750244, "learning_rate": 4.953152373969432e-07, "loss": 0.1179, "step": 53840 }, { "epoch": 1.0558823529411765, "grad_norm": 3.681410551071167, "learning_rate": 4.951531398774972e-07, "loss": 0.1421, "step": 53850 }, { "epoch": 1.056078431372549, "grad_norm": 3.006446361541748, "learning_rate": 4.949910428675143e-07, "loss": 0.139, "step": 53860 }, { "epoch": 1.0562745098039215, "grad_norm": 2.0294620990753174, "learning_rate": 4.948289463840329e-07, "loss": 0.1049, "step": 53870 }, { "epoch": 1.056470588235294, "grad_norm": 1.3874255418777466, "learning_rate": 4.946668504440909e-07, "loss": 0.1208, "step": 53880 }, { "epoch": 1.0566666666666666, "grad_norm": 1.5321675539016724, "learning_rate": 4.94504755064727e-07, "loss": 0.1122, "step": 53890 }, { "epoch": 1.0568627450980392, "grad_norm": 1.8542183637619019, "learning_rate": 4.94342660262979e-07, "loss": 0.1505, "step": 53900 }, { "epoch": 1.0570588235294118, "grad_norm": 1.9806387424468994, "learning_rate": 4.941805660558849e-07, "loss": 0.148, "step": 53910 }, { "epoch": 1.0572549019607844, "grad_norm": 2.502454996109009, "learning_rate": 4.940184724604833e-07, "loss": 0.0978, "step": 53920 }, { "epoch": 1.0574509803921568, "grad_norm": 2.142781972885132, "learning_rate": 4.938563794938117e-07, "loss": 0.1149, "step": 53930 }, { "epoch": 1.0576470588235294, "grad_norm": 2.408355236053467, "learning_rate": 4.936942871729084e-07, "loss": 0.0998, "step": 53940 }, { "epoch": 1.057843137254902, "grad_norm": 3.429108142852783, "learning_rate": 4.935321955148108e-07, "loss": 0.142, "step": 53950 }, { "epoch": 1.0580392156862746, "grad_norm": 4.088048458099365, "learning_rate": 4.93370104536557e-07, "loss": 0.1235, "step": 53960 }, { "epoch": 1.0582352941176472, "grad_norm": 3.4740591049194336, "learning_rate": 4.932080142551846e-07, "loss": 0.1353, "step": 53970 }, { "epoch": 1.0584313725490195, "grad_norm": 2.3988349437713623, "learning_rate": 4.930459246877312e-07, "loss": 0.1378, "step": 53980 }, { "epoch": 1.0586274509803921, "grad_norm": 2.918897867202759, "learning_rate": 4.928838358512342e-07, "loss": 0.1358, "step": 53990 }, { "epoch": 1.0588235294117647, "grad_norm": 1.1830048561096191, "learning_rate": 4.927217477627315e-07, "loss": 0.1095, "step": 54000 }, { "epoch": 1.0590196078431373, "grad_norm": 0.9334729313850403, "learning_rate": 4.925596604392602e-07, "loss": 0.1184, "step": 54010 }, { "epoch": 1.05921568627451, "grad_norm": 1.1275761127471924, "learning_rate": 4.923975738978576e-07, "loss": 0.122, "step": 54020 }, { "epoch": 1.0594117647058823, "grad_norm": 3.31122088432312, "learning_rate": 4.922354881555611e-07, "loss": 0.1228, "step": 54030 }, { "epoch": 1.0596078431372549, "grad_norm": 2.0454955101013184, "learning_rate": 4.920734032294078e-07, "loss": 0.107, "step": 54040 }, { "epoch": 1.0598039215686275, "grad_norm": 2.026301383972168, "learning_rate": 4.919113191364344e-07, "loss": 0.1229, "step": 54050 }, { "epoch": 1.06, "grad_norm": 2.892190933227539, "learning_rate": 4.917492358936786e-07, "loss": 0.1138, "step": 54060 }, { "epoch": 1.0601960784313726, "grad_norm": 2.1268999576568604, "learning_rate": 4.915871535181767e-07, "loss": 0.125, "step": 54070 }, { "epoch": 1.060392156862745, "grad_norm": 1.4622046947479248, "learning_rate": 4.914250720269658e-07, "loss": 0.1038, "step": 54080 }, { "epoch": 1.0605882352941176, "grad_norm": 1.407515048980713, "learning_rate": 4.912629914370821e-07, "loss": 0.1012, "step": 54090 }, { "epoch": 1.0607843137254902, "grad_norm": 2.404209852218628, "learning_rate": 4.911009117655631e-07, "loss": 0.1262, "step": 54100 }, { "epoch": 1.0609803921568628, "grad_norm": 3.080606698989868, "learning_rate": 4.909388330294445e-07, "loss": 0.1388, "step": 54110 }, { "epoch": 1.0611764705882354, "grad_norm": 1.1143616437911987, "learning_rate": 4.907767552457632e-07, "loss": 0.1191, "step": 54120 }, { "epoch": 1.0613725490196078, "grad_norm": 1.5861132144927979, "learning_rate": 4.906146784315553e-07, "loss": 0.1053, "step": 54130 }, { "epoch": 1.0615686274509804, "grad_norm": 1.4675579071044922, "learning_rate": 4.904526026038571e-07, "loss": 0.1447, "step": 54140 }, { "epoch": 1.061764705882353, "grad_norm": 2.958130121231079, "learning_rate": 4.902905277797047e-07, "loss": 0.1243, "step": 54150 }, { "epoch": 1.0619607843137255, "grad_norm": 3.044445753097534, "learning_rate": 4.901284539761341e-07, "loss": 0.1455, "step": 54160 }, { "epoch": 1.0621568627450981, "grad_norm": 3.774317741394043, "learning_rate": 4.89966381210181e-07, "loss": 0.1313, "step": 54170 }, { "epoch": 1.0623529411764705, "grad_norm": 1.456394076347351, "learning_rate": 4.898043094988816e-07, "loss": 0.1169, "step": 54180 }, { "epoch": 1.062549019607843, "grad_norm": 1.745240330696106, "learning_rate": 4.896422388592713e-07, "loss": 0.1324, "step": 54190 }, { "epoch": 1.0627450980392157, "grad_norm": 2.5457489490509033, "learning_rate": 4.894801693083855e-07, "loss": 0.1213, "step": 54200 }, { "epoch": 1.0629411764705883, "grad_norm": 2.3959367275238037, "learning_rate": 4.8931810086326e-07, "loss": 0.0954, "step": 54210 }, { "epoch": 1.0631372549019609, "grad_norm": 1.9383354187011719, "learning_rate": 4.891560335409301e-07, "loss": 0.1168, "step": 54220 }, { "epoch": 1.0633333333333332, "grad_norm": 2.818329334259033, "learning_rate": 4.889939673584307e-07, "loss": 0.1176, "step": 54230 }, { "epoch": 1.0635294117647058, "grad_norm": 1.4823410511016846, "learning_rate": 4.888319023327973e-07, "loss": 0.1271, "step": 54240 }, { "epoch": 1.0637254901960784, "grad_norm": 1.7318159341812134, "learning_rate": 4.886698384810646e-07, "loss": 0.1408, "step": 54250 }, { "epoch": 1.063921568627451, "grad_norm": 3.2063443660736084, "learning_rate": 4.885077758202674e-07, "loss": 0.1021, "step": 54260 }, { "epoch": 1.0641176470588236, "grad_norm": 2.642070770263672, "learning_rate": 4.883457143674406e-07, "loss": 0.1226, "step": 54270 }, { "epoch": 1.064313725490196, "grad_norm": 2.270629405975342, "learning_rate": 4.881836541396188e-07, "loss": 0.1262, "step": 54280 }, { "epoch": 1.0645098039215686, "grad_norm": 1.8737584352493286, "learning_rate": 4.880215951538365e-07, "loss": 0.1346, "step": 54290 }, { "epoch": 1.0647058823529412, "grad_norm": 2.4285435676574707, "learning_rate": 4.878595374271279e-07, "loss": 0.1005, "step": 54300 }, { "epoch": 1.0649019607843138, "grad_norm": 3.084437608718872, "learning_rate": 4.876974809765272e-07, "loss": 0.1176, "step": 54310 }, { "epoch": 1.0650980392156864, "grad_norm": 2.286518096923828, "learning_rate": 4.875354258190687e-07, "loss": 0.1435, "step": 54320 }, { "epoch": 1.0652941176470587, "grad_norm": 1.9513431787490845, "learning_rate": 4.873733719717861e-07, "loss": 0.1145, "step": 54330 }, { "epoch": 1.0654901960784313, "grad_norm": 1.346934199333191, "learning_rate": 4.872113194517133e-07, "loss": 0.1104, "step": 54340 }, { "epoch": 1.065686274509804, "grad_norm": 1.6057932376861572, "learning_rate": 4.870492682758841e-07, "loss": 0.1393, "step": 54350 }, { "epoch": 1.0658823529411765, "grad_norm": 2.2691121101379395, "learning_rate": 4.868872184613316e-07, "loss": 0.12, "step": 54360 }, { "epoch": 1.066078431372549, "grad_norm": 1.7121299505233765, "learning_rate": 4.867251700250895e-07, "loss": 0.1374, "step": 54370 }, { "epoch": 1.0662745098039215, "grad_norm": 3.804149627685547, "learning_rate": 4.865631229841911e-07, "loss": 0.1445, "step": 54380 }, { "epoch": 1.066470588235294, "grad_norm": 1.4357250928878784, "learning_rate": 4.864010773556692e-07, "loss": 0.1081, "step": 54390 }, { "epoch": 1.0666666666666667, "grad_norm": 2.561204195022583, "learning_rate": 4.862390331565569e-07, "loss": 0.1236, "step": 54400 }, { "epoch": 1.0668627450980392, "grad_norm": 3.0480692386627197, "learning_rate": 4.860769904038871e-07, "loss": 0.1076, "step": 54410 }, { "epoch": 1.0670588235294118, "grad_norm": 1.5131243467330933, "learning_rate": 4.859149491146922e-07, "loss": 0.1206, "step": 54420 }, { "epoch": 1.0672549019607844, "grad_norm": 2.1113765239715576, "learning_rate": 4.85752909306005e-07, "loss": 0.104, "step": 54430 }, { "epoch": 1.0674509803921568, "grad_norm": 1.99959135055542, "learning_rate": 4.855908709948573e-07, "loss": 0.1279, "step": 54440 }, { "epoch": 1.0676470588235294, "grad_norm": 1.5724812746047974, "learning_rate": 4.854288341982816e-07, "loss": 0.1162, "step": 54450 }, { "epoch": 1.067843137254902, "grad_norm": 2.1791417598724365, "learning_rate": 4.852667989333099e-07, "loss": 0.0983, "step": 54460 }, { "epoch": 1.0680392156862746, "grad_norm": 3.023606777191162, "learning_rate": 4.851047652169741e-07, "loss": 0.1243, "step": 54470 }, { "epoch": 1.0682352941176472, "grad_norm": 1.5729615688323975, "learning_rate": 4.849427330663059e-07, "loss": 0.1262, "step": 54480 }, { "epoch": 1.0684313725490195, "grad_norm": 1.9800243377685547, "learning_rate": 4.847807024983367e-07, "loss": 0.1179, "step": 54490 }, { "epoch": 1.0686274509803921, "grad_norm": 2.569528341293335, "learning_rate": 4.846186735300977e-07, "loss": 0.1468, "step": 54500 }, { "epoch": 1.0688235294117647, "grad_norm": 2.2093770503997803, "learning_rate": 4.844566461786206e-07, "loss": 0.1485, "step": 54510 }, { "epoch": 1.0690196078431373, "grad_norm": 2.2336559295654297, "learning_rate": 4.842946204609359e-07, "loss": 0.0995, "step": 54520 }, { "epoch": 1.06921568627451, "grad_norm": 2.1977715492248535, "learning_rate": 4.841325963940746e-07, "loss": 0.113, "step": 54530 }, { "epoch": 1.0694117647058823, "grad_norm": 1.8000160455703735, "learning_rate": 4.839705739950674e-07, "loss": 0.1079, "step": 54540 }, { "epoch": 1.0696078431372549, "grad_norm": 3.176302433013916, "learning_rate": 4.838085532809449e-07, "loss": 0.137, "step": 54550 }, { "epoch": 1.0698039215686275, "grad_norm": 2.473832368850708, "learning_rate": 4.836465342687371e-07, "loss": 0.1186, "step": 54560 }, { "epoch": 1.07, "grad_norm": 1.7419579029083252, "learning_rate": 4.834845169754747e-07, "loss": 0.1121, "step": 54570 }, { "epoch": 1.0701960784313727, "grad_norm": 2.36032772064209, "learning_rate": 4.833225014181869e-07, "loss": 0.1451, "step": 54580 }, { "epoch": 1.070392156862745, "grad_norm": 2.130903959274292, "learning_rate": 4.831604876139042e-07, "loss": 0.119, "step": 54590 }, { "epoch": 1.0705882352941176, "grad_norm": 2.4885332584381104, "learning_rate": 4.829984755796557e-07, "loss": 0.109, "step": 54600 }, { "epoch": 1.0707843137254902, "grad_norm": 3.1277925968170166, "learning_rate": 4.828364653324709e-07, "loss": 0.149, "step": 54610 }, { "epoch": 1.0709803921568628, "grad_norm": 2.165844440460205, "learning_rate": 4.826744568893793e-07, "loss": 0.1302, "step": 54620 }, { "epoch": 1.0711764705882354, "grad_norm": 1.4963021278381348, "learning_rate": 4.825124502674094e-07, "loss": 0.1029, "step": 54630 }, { "epoch": 1.0713725490196078, "grad_norm": 4.468713760375977, "learning_rate": 4.823504454835906e-07, "loss": 0.1376, "step": 54640 }, { "epoch": 1.0715686274509804, "grad_norm": 3.0208117961883545, "learning_rate": 4.821884425549512e-07, "loss": 0.1283, "step": 54650 }, { "epoch": 1.071764705882353, "grad_norm": 9.171114921569824, "learning_rate": 4.820264414985197e-07, "loss": 0.1177, "step": 54660 }, { "epoch": 1.0719607843137255, "grad_norm": 2.786985397338867, "learning_rate": 4.818644423313246e-07, "loss": 0.134, "step": 54670 }, { "epoch": 1.0721568627450981, "grad_norm": 1.214725375175476, "learning_rate": 4.817024450703937e-07, "loss": 0.1452, "step": 54680 }, { "epoch": 1.0723529411764705, "grad_norm": 2.1189215183258057, "learning_rate": 4.815404497327548e-07, "loss": 0.126, "step": 54690 }, { "epoch": 1.072549019607843, "grad_norm": 1.6550803184509277, "learning_rate": 4.813784563354357e-07, "loss": 0.1286, "step": 54700 }, { "epoch": 1.0727450980392157, "grad_norm": 1.5065555572509766, "learning_rate": 4.812164648954637e-07, "loss": 0.1239, "step": 54710 }, { "epoch": 1.0729411764705883, "grad_norm": 4.261048316955566, "learning_rate": 4.810544754298662e-07, "loss": 0.1155, "step": 54720 }, { "epoch": 1.0731372549019609, "grad_norm": 2.685804605484009, "learning_rate": 4.808924879556703e-07, "loss": 0.1075, "step": 54730 }, { "epoch": 1.0733333333333333, "grad_norm": 2.38415789604187, "learning_rate": 4.807305024899025e-07, "loss": 0.1372, "step": 54740 }, { "epoch": 1.0735294117647058, "grad_norm": 1.6489815711975098, "learning_rate": 4.805685190495896e-07, "loss": 0.1313, "step": 54750 }, { "epoch": 1.0737254901960784, "grad_norm": 2.865230083465576, "learning_rate": 4.804065376517582e-07, "loss": 0.1074, "step": 54760 }, { "epoch": 1.073921568627451, "grad_norm": 2.732969284057617, "learning_rate": 4.80244558313434e-07, "loss": 0.1347, "step": 54770 }, { "epoch": 1.0741176470588236, "grad_norm": 2.1791982650756836, "learning_rate": 4.800825810516435e-07, "loss": 0.0989, "step": 54780 }, { "epoch": 1.074313725490196, "grad_norm": 3.4351987838745117, "learning_rate": 4.799206058834121e-07, "loss": 0.1301, "step": 54790 }, { "epoch": 1.0745098039215686, "grad_norm": 4.523527145385742, "learning_rate": 4.797586328257653e-07, "loss": 0.1185, "step": 54800 }, { "epoch": 1.0747058823529412, "grad_norm": 4.390970706939697, "learning_rate": 4.795966618957285e-07, "loss": 0.124, "step": 54810 }, { "epoch": 1.0749019607843138, "grad_norm": 3.714365005493164, "learning_rate": 4.794346931103271e-07, "loss": 0.13, "step": 54820 }, { "epoch": 1.0750980392156864, "grad_norm": 2.0751230716705322, "learning_rate": 4.792727264865856e-07, "loss": 0.1272, "step": 54830 }, { "epoch": 1.0752941176470587, "grad_norm": 3.4222006797790527, "learning_rate": 4.791107620415285e-07, "loss": 0.1226, "step": 54840 }, { "epoch": 1.0754901960784313, "grad_norm": 2.3953263759613037, "learning_rate": 4.789487997921807e-07, "loss": 0.0968, "step": 54850 }, { "epoch": 1.075686274509804, "grad_norm": 1.604166865348816, "learning_rate": 4.78786839755566e-07, "loss": 0.1195, "step": 54860 }, { "epoch": 1.0758823529411765, "grad_norm": 1.879278540611267, "learning_rate": 4.786248819487084e-07, "loss": 0.1211, "step": 54870 }, { "epoch": 1.076078431372549, "grad_norm": 1.8018507957458496, "learning_rate": 4.784629263886315e-07, "loss": 0.097, "step": 54880 }, { "epoch": 1.0762745098039215, "grad_norm": 1.9475312232971191, "learning_rate": 4.783009730923591e-07, "loss": 0.1559, "step": 54890 }, { "epoch": 1.076470588235294, "grad_norm": 1.907630443572998, "learning_rate": 4.781390220769141e-07, "loss": 0.1413, "step": 54900 }, { "epoch": 1.0766666666666667, "grad_norm": 1.5490409135818481, "learning_rate": 4.779770733593196e-07, "loss": 0.1377, "step": 54910 }, { "epoch": 1.0768627450980393, "grad_norm": 3.0417470932006836, "learning_rate": 4.778151269565985e-07, "loss": 0.1232, "step": 54920 }, { "epoch": 1.0770588235294118, "grad_norm": 3.7172646522521973, "learning_rate": 4.77653182885773e-07, "loss": 0.1224, "step": 54930 }, { "epoch": 1.0772549019607842, "grad_norm": 2.212286949157715, "learning_rate": 4.774912411638658e-07, "loss": 0.1384, "step": 54940 }, { "epoch": 1.0774509803921568, "grad_norm": 2.536520004272461, "learning_rate": 4.773293018078984e-07, "loss": 0.1698, "step": 54950 }, { "epoch": 1.0776470588235294, "grad_norm": 2.382826328277588, "learning_rate": 4.771673648348929e-07, "loss": 0.1283, "step": 54960 }, { "epoch": 1.077843137254902, "grad_norm": 2.4781293869018555, "learning_rate": 4.770054302618708e-07, "loss": 0.1057, "step": 54970 }, { "epoch": 1.0780392156862746, "grad_norm": 1.6992335319519043, "learning_rate": 4.768434981058531e-07, "loss": 0.1349, "step": 54980 }, { "epoch": 1.078235294117647, "grad_norm": 1.4711915254592896, "learning_rate": 4.76681568383861e-07, "loss": 0.1391, "step": 54990 }, { "epoch": 1.0784313725490196, "grad_norm": 1.3882561922073364, "learning_rate": 4.7651964111291537e-07, "loss": 0.118, "step": 55000 }, { "epoch": 1.0786274509803921, "grad_norm": 2.196054458618164, "learning_rate": 4.7635771631003654e-07, "loss": 0.1358, "step": 55010 }, { "epoch": 1.0788235294117647, "grad_norm": 2.2058985233306885, "learning_rate": 4.761957939922449e-07, "loss": 0.1227, "step": 55020 }, { "epoch": 1.0790196078431373, "grad_norm": 3.590280771255493, "learning_rate": 4.7603387417656024e-07, "loss": 0.1257, "step": 55030 }, { "epoch": 1.0792156862745097, "grad_norm": 1.293177843093872, "learning_rate": 4.758719568800023e-07, "loss": 0.112, "step": 55040 }, { "epoch": 1.0794117647058823, "grad_norm": 2.8127994537353516, "learning_rate": 4.757100421195907e-07, "loss": 0.1237, "step": 55050 }, { "epoch": 1.0796078431372549, "grad_norm": 3.4253063201904297, "learning_rate": 4.755481299123444e-07, "loss": 0.1212, "step": 55060 }, { "epoch": 1.0798039215686275, "grad_norm": 2.562976121902466, "learning_rate": 4.7538622027528234e-07, "loss": 0.1422, "step": 55070 }, { "epoch": 1.08, "grad_norm": 2.9054970741271973, "learning_rate": 4.7522431322542344e-07, "loss": 0.1349, "step": 55080 }, { "epoch": 1.0801960784313724, "grad_norm": 2.0872089862823486, "learning_rate": 4.7506240877978565e-07, "loss": 0.122, "step": 55090 }, { "epoch": 1.080392156862745, "grad_norm": 2.0464437007904053, "learning_rate": 4.7490050695538734e-07, "loss": 0.1036, "step": 55100 }, { "epoch": 1.0805882352941176, "grad_norm": 3.975245714187622, "learning_rate": 4.7473860776924637e-07, "loss": 0.1244, "step": 55110 }, { "epoch": 1.0807843137254902, "grad_norm": 1.3828896284103394, "learning_rate": 4.7457671123838e-07, "loss": 0.1076, "step": 55120 }, { "epoch": 1.0809803921568628, "grad_norm": 3.2432920932769775, "learning_rate": 4.7441481737980584e-07, "loss": 0.1072, "step": 55130 }, { "epoch": 1.0811764705882352, "grad_norm": 1.570157766342163, "learning_rate": 4.742529262105406e-07, "loss": 0.1251, "step": 55140 }, { "epoch": 1.0813725490196078, "grad_norm": 1.8093702793121338, "learning_rate": 4.74091037747601e-07, "loss": 0.1163, "step": 55150 }, { "epoch": 1.0815686274509804, "grad_norm": 2.8955609798431396, "learning_rate": 4.739291520080037e-07, "loss": 0.1149, "step": 55160 }, { "epoch": 1.081764705882353, "grad_norm": 2.090184211730957, "learning_rate": 4.7376726900876434e-07, "loss": 0.0993, "step": 55170 }, { "epoch": 1.0819607843137256, "grad_norm": 1.6735177040100098, "learning_rate": 4.736053887668994e-07, "loss": 0.1329, "step": 55180 }, { "epoch": 1.082156862745098, "grad_norm": 1.496122121810913, "learning_rate": 4.7344351129942404e-07, "loss": 0.1165, "step": 55190 }, { "epoch": 1.0823529411764705, "grad_norm": 1.6980687379837036, "learning_rate": 4.732816366233536e-07, "loss": 0.1101, "step": 55200 }, { "epoch": 1.082549019607843, "grad_norm": 2.103656053543091, "learning_rate": 4.7311976475570314e-07, "loss": 0.126, "step": 55210 }, { "epoch": 1.0827450980392157, "grad_norm": 1.5623133182525635, "learning_rate": 4.729578957134871e-07, "loss": 0.0905, "step": 55220 }, { "epoch": 1.0829411764705883, "grad_norm": 1.7397857904434204, "learning_rate": 4.7279602951371997e-07, "loss": 0.1145, "step": 55230 }, { "epoch": 1.0831372549019607, "grad_norm": 2.475733757019043, "learning_rate": 4.7263416617341606e-07, "loss": 0.1211, "step": 55240 }, { "epoch": 1.0833333333333333, "grad_norm": 2.967618227005005, "learning_rate": 4.724723057095887e-07, "loss": 0.1123, "step": 55250 }, { "epoch": 1.0835294117647059, "grad_norm": 1.903901219367981, "learning_rate": 4.723104481392517e-07, "loss": 0.1255, "step": 55260 }, { "epoch": 1.0837254901960784, "grad_norm": 2.6816179752349854, "learning_rate": 4.721485934794182e-07, "loss": 0.1148, "step": 55270 }, { "epoch": 1.083921568627451, "grad_norm": 2.408607006072998, "learning_rate": 4.719867417471009e-07, "loss": 0.1398, "step": 55280 }, { "epoch": 1.0841176470588236, "grad_norm": 2.153160810470581, "learning_rate": 4.7182489295931247e-07, "loss": 0.0945, "step": 55290 }, { "epoch": 1.084313725490196, "grad_norm": 1.6730780601501465, "learning_rate": 4.716630471330651e-07, "loss": 0.1169, "step": 55300 }, { "epoch": 1.0845098039215686, "grad_norm": 3.084204912185669, "learning_rate": 4.7150120428537073e-07, "loss": 0.1283, "step": 55310 }, { "epoch": 1.0847058823529412, "grad_norm": 2.2944862842559814, "learning_rate": 4.7133936443324105e-07, "loss": 0.1301, "step": 55320 }, { "epoch": 1.0849019607843138, "grad_norm": 3.845754623413086, "learning_rate": 4.711775275936873e-07, "loss": 0.1292, "step": 55330 }, { "epoch": 1.0850980392156864, "grad_norm": 3.097975254058838, "learning_rate": 4.7101569378372034e-07, "loss": 0.1193, "step": 55340 }, { "epoch": 1.0852941176470587, "grad_norm": 3.610990524291992, "learning_rate": 4.7085386302035107e-07, "loss": 0.1327, "step": 55350 }, { "epoch": 1.0854901960784313, "grad_norm": 2.9906857013702393, "learning_rate": 4.706920353205898e-07, "loss": 0.129, "step": 55360 }, { "epoch": 1.085686274509804, "grad_norm": 2.374187707901001, "learning_rate": 4.705302107014466e-07, "loss": 0.1393, "step": 55370 }, { "epoch": 1.0858823529411765, "grad_norm": 2.1075844764709473, "learning_rate": 4.70368389179931e-07, "loss": 0.1334, "step": 55380 }, { "epoch": 1.0860784313725491, "grad_norm": 1.2629613876342773, "learning_rate": 4.702065707730525e-07, "loss": 0.1313, "step": 55390 }, { "epoch": 1.0862745098039215, "grad_norm": 1.3353941440582275, "learning_rate": 4.700447554978202e-07, "loss": 0.1391, "step": 55400 }, { "epoch": 1.086470588235294, "grad_norm": 1.4468746185302734, "learning_rate": 4.6988294337124263e-07, "loss": 0.1244, "step": 55410 }, { "epoch": 1.0866666666666667, "grad_norm": 2.458148956298828, "learning_rate": 4.697211344103283e-07, "loss": 0.1433, "step": 55420 }, { "epoch": 1.0868627450980393, "grad_norm": 1.881890058517456, "learning_rate": 4.6955932863208534e-07, "loss": 0.1249, "step": 55430 }, { "epoch": 1.0870588235294119, "grad_norm": 2.0547049045562744, "learning_rate": 4.693975260535214e-07, "loss": 0.1148, "step": 55440 }, { "epoch": 1.0872549019607842, "grad_norm": 2.866917371749878, "learning_rate": 4.692357266916438e-07, "loss": 0.1341, "step": 55450 }, { "epoch": 1.0874509803921568, "grad_norm": 2.95558762550354, "learning_rate": 4.690739305634598e-07, "loss": 0.1273, "step": 55460 }, { "epoch": 1.0876470588235294, "grad_norm": 1.9200323820114136, "learning_rate": 4.6891213768597583e-07, "loss": 0.1148, "step": 55470 }, { "epoch": 1.087843137254902, "grad_norm": 3.5042433738708496, "learning_rate": 4.687503480761985e-07, "loss": 0.1532, "step": 55480 }, { "epoch": 1.0880392156862746, "grad_norm": 3.513707160949707, "learning_rate": 4.6858856175113363e-07, "loss": 0.1187, "step": 55490 }, { "epoch": 1.088235294117647, "grad_norm": 1.423162817955017, "learning_rate": 4.6842677872778705e-07, "loss": 0.1332, "step": 55500 }, { "epoch": 1.0884313725490196, "grad_norm": 1.136121153831482, "learning_rate": 4.6826499902316414e-07, "loss": 0.1287, "step": 55510 }, { "epoch": 1.0886274509803922, "grad_norm": 1.9189655780792236, "learning_rate": 4.6810322265426943e-07, "loss": 0.1538, "step": 55520 }, { "epoch": 1.0888235294117647, "grad_norm": 1.7951054573059082, "learning_rate": 4.679414496381082e-07, "loss": 0.1488, "step": 55530 }, { "epoch": 1.0890196078431373, "grad_norm": 3.378967046737671, "learning_rate": 4.6777967999168444e-07, "loss": 0.1603, "step": 55540 }, { "epoch": 1.0892156862745097, "grad_norm": 2.123039722442627, "learning_rate": 4.6761791373200206e-07, "loss": 0.1243, "step": 55550 }, { "epoch": 1.0894117647058823, "grad_norm": 1.623746633529663, "learning_rate": 4.674561508760647e-07, "loss": 0.1063, "step": 55560 }, { "epoch": 1.089607843137255, "grad_norm": 1.4280405044555664, "learning_rate": 4.672943914408755e-07, "loss": 0.1211, "step": 55570 }, { "epoch": 1.0898039215686275, "grad_norm": 1.5782556533813477, "learning_rate": 4.6713263544343726e-07, "loss": 0.122, "step": 55580 }, { "epoch": 1.09, "grad_norm": 1.8284287452697754, "learning_rate": 4.669708829007527e-07, "loss": 0.1622, "step": 55590 }, { "epoch": 1.0901960784313725, "grad_norm": 2.267216682434082, "learning_rate": 4.6680913382982374e-07, "loss": 0.1165, "step": 55600 }, { "epoch": 1.090392156862745, "grad_norm": 3.3386690616607666, "learning_rate": 4.6664738824765217e-07, "loss": 0.1209, "step": 55610 }, { "epoch": 1.0905882352941176, "grad_norm": 4.5551886558532715, "learning_rate": 4.664856461712395e-07, "loss": 0.15, "step": 55620 }, { "epoch": 1.0907843137254902, "grad_norm": 1.5846166610717773, "learning_rate": 4.6632390761758654e-07, "loss": 0.1366, "step": 55630 }, { "epoch": 1.0909803921568628, "grad_norm": 2.2080161571502686, "learning_rate": 4.661621726036943e-07, "loss": 0.1438, "step": 55640 }, { "epoch": 1.0911764705882352, "grad_norm": 2.516018867492676, "learning_rate": 4.6600044114656267e-07, "loss": 0.1025, "step": 55650 }, { "epoch": 1.0913725490196078, "grad_norm": 3.5984244346618652, "learning_rate": 4.6583871326319176e-07, "loss": 0.0984, "step": 55660 }, { "epoch": 1.0915686274509804, "grad_norm": 2.076978921890259, "learning_rate": 4.6567698897058123e-07, "loss": 0.1264, "step": 55670 }, { "epoch": 1.091764705882353, "grad_norm": 3.0888774394989014, "learning_rate": 4.6551526828572993e-07, "loss": 0.1374, "step": 55680 }, { "epoch": 1.0919607843137256, "grad_norm": 1.7463550567626953, "learning_rate": 4.6535355122563686e-07, "loss": 0.1178, "step": 55690 }, { "epoch": 1.0921568627450982, "grad_norm": 2.2164018154144287, "learning_rate": 4.6519183780730026e-07, "loss": 0.1355, "step": 55700 }, { "epoch": 1.0923529411764705, "grad_norm": 2.099203109741211, "learning_rate": 4.650301280477184e-07, "loss": 0.1279, "step": 55710 }, { "epoch": 1.0925490196078431, "grad_norm": 2.187814712524414, "learning_rate": 4.648684219638888e-07, "loss": 0.1071, "step": 55720 }, { "epoch": 1.0927450980392157, "grad_norm": 2.073310613632202, "learning_rate": 4.6470671957280867e-07, "loss": 0.1208, "step": 55730 }, { "epoch": 1.0929411764705883, "grad_norm": 1.9931756258010864, "learning_rate": 4.645450208914748e-07, "loss": 0.1246, "step": 55740 }, { "epoch": 1.093137254901961, "grad_norm": 3.2539875507354736, "learning_rate": 4.643833259368839e-07, "loss": 0.1287, "step": 55750 }, { "epoch": 1.0933333333333333, "grad_norm": 8.68400764465332, "learning_rate": 4.642216347260317e-07, "loss": 0.121, "step": 55760 }, { "epoch": 1.0935294117647059, "grad_norm": 4.275831699371338, "learning_rate": 4.6405994727591405e-07, "loss": 0.1273, "step": 55770 }, { "epoch": 1.0937254901960785, "grad_norm": 2.0491278171539307, "learning_rate": 4.6389826360352645e-07, "loss": 0.1116, "step": 55780 }, { "epoch": 1.093921568627451, "grad_norm": 2.117588996887207, "learning_rate": 4.6373658372586344e-07, "loss": 0.1145, "step": 55790 }, { "epoch": 1.0941176470588236, "grad_norm": 2.4390995502471924, "learning_rate": 4.6357490765991964e-07, "loss": 0.1178, "step": 55800 }, { "epoch": 1.094313725490196, "grad_norm": 2.0463645458221436, "learning_rate": 4.634132354226892e-07, "loss": 0.1435, "step": 55810 }, { "epoch": 1.0945098039215686, "grad_norm": 1.8078913688659668, "learning_rate": 4.6325156703116577e-07, "loss": 0.1232, "step": 55820 }, { "epoch": 1.0947058823529412, "grad_norm": 2.9879307746887207, "learning_rate": 4.630899025023427e-07, "loss": 0.0974, "step": 55830 }, { "epoch": 1.0949019607843138, "grad_norm": 3.7210147380828857, "learning_rate": 4.629282418532127e-07, "loss": 0.1331, "step": 55840 }, { "epoch": 1.0950980392156864, "grad_norm": 4.224200248718262, "learning_rate": 4.627665851007683e-07, "loss": 0.1125, "step": 55850 }, { "epoch": 1.0952941176470588, "grad_norm": 2.4501848220825195, "learning_rate": 4.626049322620017e-07, "loss": 0.1149, "step": 55860 }, { "epoch": 1.0954901960784313, "grad_norm": 3.1633362770080566, "learning_rate": 4.6244328335390434e-07, "loss": 0.1174, "step": 55870 }, { "epoch": 1.095686274509804, "grad_norm": 3.0514423847198486, "learning_rate": 4.622816383934675e-07, "loss": 0.1173, "step": 55880 }, { "epoch": 1.0958823529411765, "grad_norm": 3.466306447982788, "learning_rate": 4.6211999739768213e-07, "loss": 0.1224, "step": 55890 }, { "epoch": 1.0960784313725491, "grad_norm": 2.4407217502593994, "learning_rate": 4.619583603835386e-07, "loss": 0.1144, "step": 55900 }, { "epoch": 1.0962745098039215, "grad_norm": 3.1908750534057617, "learning_rate": 4.6179672736802693e-07, "loss": 0.1149, "step": 55910 }, { "epoch": 1.096470588235294, "grad_norm": 2.344081163406372, "learning_rate": 4.616350983681365e-07, "loss": 0.1378, "step": 55920 }, { "epoch": 1.0966666666666667, "grad_norm": 2.599818229675293, "learning_rate": 4.614734734008566e-07, "loss": 0.1101, "step": 55930 }, { "epoch": 1.0968627450980393, "grad_norm": 5.003781318664551, "learning_rate": 4.6131185248317606e-07, "loss": 0.1405, "step": 55940 }, { "epoch": 1.0970588235294119, "grad_norm": 2.1974380016326904, "learning_rate": 4.6115023563208285e-07, "loss": 0.1394, "step": 55950 }, { "epoch": 1.0972549019607842, "grad_norm": 2.1161673069000244, "learning_rate": 4.60988622864565e-07, "loss": 0.1101, "step": 55960 }, { "epoch": 1.0974509803921568, "grad_norm": 4.062183856964111, "learning_rate": 4.6082701419761014e-07, "loss": 0.1326, "step": 55970 }, { "epoch": 1.0976470588235294, "grad_norm": 1.3564800024032593, "learning_rate": 4.60665409648205e-07, "loss": 0.1096, "step": 55980 }, { "epoch": 1.097843137254902, "grad_norm": 1.6793376207351685, "learning_rate": 4.605038092333363e-07, "loss": 0.1098, "step": 55990 }, { "epoch": 1.0980392156862746, "grad_norm": 1.9378938674926758, "learning_rate": 4.603422129699901e-07, "loss": 0.1529, "step": 56000 }, { "epoch": 1.098235294117647, "grad_norm": 1.2222367525100708, "learning_rate": 4.6018062087515207e-07, "loss": 0.0966, "step": 56010 }, { "epoch": 1.0984313725490196, "grad_norm": 1.883712887763977, "learning_rate": 4.6001903296580766e-07, "loss": 0.1439, "step": 56020 }, { "epoch": 1.0986274509803922, "grad_norm": 1.8261581659317017, "learning_rate": 4.5985744925894147e-07, "loss": 0.1152, "step": 56030 }, { "epoch": 1.0988235294117648, "grad_norm": 1.6406866312026978, "learning_rate": 4.5969586977153803e-07, "loss": 0.1275, "step": 56040 }, { "epoch": 1.0990196078431373, "grad_norm": 2.7366743087768555, "learning_rate": 4.595342945205813e-07, "loss": 0.1221, "step": 56050 }, { "epoch": 1.0992156862745097, "grad_norm": 1.5370292663574219, "learning_rate": 4.5937272352305433e-07, "loss": 0.1192, "step": 56060 }, { "epoch": 1.0994117647058823, "grad_norm": 1.7760977745056152, "learning_rate": 4.5921115679594093e-07, "loss": 0.1151, "step": 56070 }, { "epoch": 1.099607843137255, "grad_norm": 3.5872509479522705, "learning_rate": 4.590495943562232e-07, "loss": 0.1287, "step": 56080 }, { "epoch": 1.0998039215686275, "grad_norm": 3.711972236633301, "learning_rate": 4.588880362208834e-07, "loss": 0.1061, "step": 56090 }, { "epoch": 1.1, "grad_norm": 1.9497466087341309, "learning_rate": 4.5872648240690333e-07, "loss": 0.1031, "step": 56100 }, { "epoch": 1.1001960784313725, "grad_norm": 1.5869957208633423, "learning_rate": 4.5856493293126396e-07, "loss": 0.1173, "step": 56110 }, { "epoch": 1.100392156862745, "grad_norm": 3.126817464828491, "learning_rate": 4.584033878109463e-07, "loss": 0.1584, "step": 56120 }, { "epoch": 1.1005882352941176, "grad_norm": 1.4424481391906738, "learning_rate": 4.5824184706293067e-07, "loss": 0.1028, "step": 56130 }, { "epoch": 1.1007843137254902, "grad_norm": 4.741549015045166, "learning_rate": 4.5808031070419677e-07, "loss": 0.1195, "step": 56140 }, { "epoch": 1.1009803921568628, "grad_norm": 3.156744956970215, "learning_rate": 4.5791877875172414e-07, "loss": 0.1043, "step": 56150 }, { "epoch": 1.1011764705882352, "grad_norm": 1.2425661087036133, "learning_rate": 4.577572512224918e-07, "loss": 0.1326, "step": 56160 }, { "epoch": 1.1013725490196078, "grad_norm": 2.0296690464019775, "learning_rate": 4.5759572813347803e-07, "loss": 0.1485, "step": 56170 }, { "epoch": 1.1015686274509804, "grad_norm": 2.6312294006347656, "learning_rate": 4.57434209501661e-07, "loss": 0.1031, "step": 56180 }, { "epoch": 1.101764705882353, "grad_norm": 1.944903016090393, "learning_rate": 4.572726953440181e-07, "loss": 0.1043, "step": 56190 }, { "epoch": 1.1019607843137256, "grad_norm": 2.8385798931121826, "learning_rate": 4.5711118567752647e-07, "loss": 0.1329, "step": 56200 }, { "epoch": 1.102156862745098, "grad_norm": 2.3654420375823975, "learning_rate": 4.569496805191629e-07, "loss": 0.1418, "step": 56210 }, { "epoch": 1.1023529411764705, "grad_norm": 1.1247236728668213, "learning_rate": 4.567881798859032e-07, "loss": 0.1022, "step": 56220 }, { "epoch": 1.1025490196078431, "grad_norm": 3.0505950450897217, "learning_rate": 4.566266837947231e-07, "loss": 0.1309, "step": 56230 }, { "epoch": 1.1027450980392157, "grad_norm": 4.475377082824707, "learning_rate": 4.564651922625978e-07, "loss": 0.1354, "step": 56240 }, { "epoch": 1.1029411764705883, "grad_norm": 2.3912696838378906, "learning_rate": 4.5630370530650207e-07, "loss": 0.1103, "step": 56250 }, { "epoch": 1.1031372549019607, "grad_norm": 2.57004451751709, "learning_rate": 4.561422229434103e-07, "loss": 0.1357, "step": 56260 }, { "epoch": 1.1033333333333333, "grad_norm": 1.9369229078292847, "learning_rate": 4.559807451902958e-07, "loss": 0.1099, "step": 56270 }, { "epoch": 1.1035294117647059, "grad_norm": 1.4531179666519165, "learning_rate": 4.5581927206413206e-07, "loss": 0.0989, "step": 56280 }, { "epoch": 1.1037254901960785, "grad_norm": 2.2018227577209473, "learning_rate": 4.5565780358189185e-07, "loss": 0.1426, "step": 56290 }, { "epoch": 1.103921568627451, "grad_norm": 1.6325082778930664, "learning_rate": 4.5549633976054737e-07, "loss": 0.1203, "step": 56300 }, { "epoch": 1.1041176470588234, "grad_norm": 1.4403291940689087, "learning_rate": 4.5533488061707034e-07, "loss": 0.1185, "step": 56310 }, { "epoch": 1.104313725490196, "grad_norm": 2.592808485031128, "learning_rate": 4.5517342616843224e-07, "loss": 0.1066, "step": 56320 }, { "epoch": 1.1045098039215686, "grad_norm": 1.9907023906707764, "learning_rate": 4.5501197643160364e-07, "loss": 0.0997, "step": 56330 }, { "epoch": 1.1047058823529412, "grad_norm": 5.22540283203125, "learning_rate": 4.548505314235551e-07, "loss": 0.1166, "step": 56340 }, { "epoch": 1.1049019607843138, "grad_norm": 1.688703179359436, "learning_rate": 4.546890911612562e-07, "loss": 0.1033, "step": 56350 }, { "epoch": 1.1050980392156862, "grad_norm": 1.6741222143173218, "learning_rate": 4.5452765566167625e-07, "loss": 0.1142, "step": 56360 }, { "epoch": 1.1052941176470588, "grad_norm": 1.5159486532211304, "learning_rate": 4.5436622494178425e-07, "loss": 0.1173, "step": 56370 }, { "epoch": 1.1054901960784314, "grad_norm": 1.3378591537475586, "learning_rate": 4.5420479901854824e-07, "loss": 0.126, "step": 56380 }, { "epoch": 1.105686274509804, "grad_norm": 1.30527925491333, "learning_rate": 4.540433779089362e-07, "loss": 0.1356, "step": 56390 }, { "epoch": 1.1058823529411765, "grad_norm": 3.289459705352783, "learning_rate": 4.538819616299155e-07, "loss": 0.1109, "step": 56400 }, { "epoch": 1.106078431372549, "grad_norm": 3.8757846355438232, "learning_rate": 4.537205501984526e-07, "loss": 0.1079, "step": 56410 }, { "epoch": 1.1062745098039215, "grad_norm": 1.0133321285247803, "learning_rate": 4.5355914363151394e-07, "loss": 0.083, "step": 56420 }, { "epoch": 1.106470588235294, "grad_norm": 3.2989518642425537, "learning_rate": 4.533977419460654e-07, "loss": 0.1385, "step": 56430 }, { "epoch": 1.1066666666666667, "grad_norm": 5.105360984802246, "learning_rate": 4.532363451590722e-07, "loss": 0.1238, "step": 56440 }, { "epoch": 1.1068627450980393, "grad_norm": 1.147357702255249, "learning_rate": 4.53074953287499e-07, "loss": 0.1165, "step": 56450 }, { "epoch": 1.1070588235294117, "grad_norm": 1.618349552154541, "learning_rate": 4.5291356634831e-07, "loss": 0.1264, "step": 56460 }, { "epoch": 1.1072549019607842, "grad_norm": 3.281524181365967, "learning_rate": 4.5275218435846896e-07, "loss": 0.1077, "step": 56470 }, { "epoch": 1.1074509803921568, "grad_norm": 2.158872127532959, "learning_rate": 4.525908073349391e-07, "loss": 0.1055, "step": 56480 }, { "epoch": 1.1076470588235294, "grad_norm": 2.6913089752197266, "learning_rate": 4.524294352946829e-07, "loss": 0.1149, "step": 56490 }, { "epoch": 1.107843137254902, "grad_norm": 3.2736923694610596, "learning_rate": 4.5226806825466267e-07, "loss": 0.0932, "step": 56500 }, { "epoch": 1.1080392156862744, "grad_norm": 2.624269485473633, "learning_rate": 4.5210670623184e-07, "loss": 0.1126, "step": 56510 }, { "epoch": 1.108235294117647, "grad_norm": 4.272850513458252, "learning_rate": 4.5194534924317584e-07, "loss": 0.1376, "step": 56520 }, { "epoch": 1.1084313725490196, "grad_norm": 2.3619532585144043, "learning_rate": 4.5178399730563096e-07, "loss": 0.1194, "step": 56530 }, { "epoch": 1.1086274509803922, "grad_norm": 1.1028803586959839, "learning_rate": 4.516226504361651e-07, "loss": 0.1142, "step": 56540 }, { "epoch": 1.1088235294117648, "grad_norm": 1.8745627403259277, "learning_rate": 4.514613086517379e-07, "loss": 0.1047, "step": 56550 }, { "epoch": 1.1090196078431374, "grad_norm": 1.5833879709243774, "learning_rate": 4.512999719693084e-07, "loss": 0.0963, "step": 56560 }, { "epoch": 1.1092156862745097, "grad_norm": 1.3989081382751465, "learning_rate": 4.511386404058348e-07, "loss": 0.093, "step": 56570 }, { "epoch": 1.1094117647058823, "grad_norm": 2.3061420917510986, "learning_rate": 4.5097731397827517e-07, "loss": 0.1366, "step": 56580 }, { "epoch": 1.109607843137255, "grad_norm": 2.1047730445861816, "learning_rate": 4.508159927035869e-07, "loss": 0.1168, "step": 56590 }, { "epoch": 1.1098039215686275, "grad_norm": 2.889286756515503, "learning_rate": 4.5065467659872624e-07, "loss": 0.1164, "step": 56600 }, { "epoch": 1.11, "grad_norm": 2.672802209854126, "learning_rate": 4.5049336568065025e-07, "loss": 0.1271, "step": 56610 }, { "epoch": 1.1101960784313725, "grad_norm": 1.8987559080123901, "learning_rate": 4.503320599663142e-07, "loss": 0.1058, "step": 56620 }, { "epoch": 1.110392156862745, "grad_norm": 2.448155403137207, "learning_rate": 4.501707594726733e-07, "loss": 0.1098, "step": 56630 }, { "epoch": 1.1105882352941177, "grad_norm": 1.6362266540527344, "learning_rate": 4.500094642166823e-07, "loss": 0.1399, "step": 56640 }, { "epoch": 1.1107843137254902, "grad_norm": 3.0066730976104736, "learning_rate": 4.49848174215295e-07, "loss": 0.145, "step": 56650 }, { "epoch": 1.1109803921568628, "grad_norm": 1.9396260976791382, "learning_rate": 4.49686889485465e-07, "loss": 0.1174, "step": 56660 }, { "epoch": 1.1111764705882352, "grad_norm": 3.5712332725524902, "learning_rate": 4.495256100441455e-07, "loss": 0.103, "step": 56670 }, { "epoch": 1.1113725490196078, "grad_norm": 2.7304883003234863, "learning_rate": 4.493643359082885e-07, "loss": 0.1322, "step": 56680 }, { "epoch": 1.1115686274509804, "grad_norm": 3.115086317062378, "learning_rate": 4.492030670948463e-07, "loss": 0.1223, "step": 56690 }, { "epoch": 1.111764705882353, "grad_norm": 3.9860036373138428, "learning_rate": 4.490418036207697e-07, "loss": 0.1483, "step": 56700 }, { "epoch": 1.1119607843137256, "grad_norm": 1.8655683994293213, "learning_rate": 4.488805455030097e-07, "loss": 0.1149, "step": 56710 }, { "epoch": 1.112156862745098, "grad_norm": 1.1742582321166992, "learning_rate": 4.4871929275851656e-07, "loss": 0.1209, "step": 56720 }, { "epoch": 1.1123529411764705, "grad_norm": 1.307852864265442, "learning_rate": 4.485580454042396e-07, "loss": 0.1268, "step": 56730 }, { "epoch": 1.1125490196078431, "grad_norm": 1.9266037940979004, "learning_rate": 4.483968034571279e-07, "loss": 0.128, "step": 56740 }, { "epoch": 1.1127450980392157, "grad_norm": 2.1914288997650146, "learning_rate": 4.482355669341302e-07, "loss": 0.1279, "step": 56750 }, { "epoch": 1.1129411764705883, "grad_norm": 15.143533706665039, "learning_rate": 4.48074335852194e-07, "loss": 0.1368, "step": 56760 }, { "epoch": 1.1131372549019607, "grad_norm": 2.427741527557373, "learning_rate": 4.479131102282668e-07, "loss": 0.1024, "step": 56770 }, { "epoch": 1.1133333333333333, "grad_norm": 2.8225796222686768, "learning_rate": 4.477518900792953e-07, "loss": 0.11, "step": 56780 }, { "epoch": 1.1135294117647059, "grad_norm": 1.7920829057693481, "learning_rate": 4.475906754222258e-07, "loss": 0.118, "step": 56790 }, { "epoch": 1.1137254901960785, "grad_norm": 2.7173197269439697, "learning_rate": 4.4742946627400393e-07, "loss": 0.1376, "step": 56800 }, { "epoch": 1.113921568627451, "grad_norm": 3.064517021179199, "learning_rate": 4.472682626515745e-07, "loss": 0.1107, "step": 56810 }, { "epoch": 1.1141176470588234, "grad_norm": 2.616971969604492, "learning_rate": 4.471070645718819e-07, "loss": 0.0992, "step": 56820 }, { "epoch": 1.114313725490196, "grad_norm": 1.3538693189620972, "learning_rate": 4.4694587205187026e-07, "loss": 0.1398, "step": 56830 }, { "epoch": 1.1145098039215686, "grad_norm": 1.5661367177963257, "learning_rate": 4.4678468510848256e-07, "loss": 0.1121, "step": 56840 }, { "epoch": 1.1147058823529412, "grad_norm": 2.472689628601074, "learning_rate": 4.466235037586616e-07, "loss": 0.1145, "step": 56850 }, { "epoch": 1.1149019607843138, "grad_norm": 2.3690011501312256, "learning_rate": 4.464623280193496e-07, "loss": 0.1406, "step": 56860 }, { "epoch": 1.1150980392156862, "grad_norm": 1.0383256673812866, "learning_rate": 4.4630115790748773e-07, "loss": 0.1158, "step": 56870 }, { "epoch": 1.1152941176470588, "grad_norm": 1.6961288452148438, "learning_rate": 4.461399934400173e-07, "loss": 0.1091, "step": 56880 }, { "epoch": 1.1154901960784314, "grad_norm": 3.000361680984497, "learning_rate": 4.4597883463387825e-07, "loss": 0.116, "step": 56890 }, { "epoch": 1.115686274509804, "grad_norm": 2.5581486225128174, "learning_rate": 4.458176815060105e-07, "loss": 0.1189, "step": 56900 }, { "epoch": 1.1158823529411765, "grad_norm": 5.915740013122559, "learning_rate": 4.456565340733532e-07, "loss": 0.1151, "step": 56910 }, { "epoch": 1.116078431372549, "grad_norm": 2.1701319217681885, "learning_rate": 4.454953923528448e-07, "loss": 0.1364, "step": 56920 }, { "epoch": 1.1162745098039215, "grad_norm": 2.038266897201538, "learning_rate": 4.4533425636142315e-07, "loss": 0.0987, "step": 56930 }, { "epoch": 1.116470588235294, "grad_norm": 3.4988791942596436, "learning_rate": 4.451731261160258e-07, "loss": 0.1273, "step": 56940 }, { "epoch": 1.1166666666666667, "grad_norm": 1.1843886375427246, "learning_rate": 4.45012001633589e-07, "loss": 0.1182, "step": 56950 }, { "epoch": 1.1168627450980393, "grad_norm": 2.6390693187713623, "learning_rate": 4.4485088293104955e-07, "loss": 0.1258, "step": 56960 }, { "epoch": 1.1170588235294119, "grad_norm": 3.129866600036621, "learning_rate": 4.4468977002534244e-07, "loss": 0.1249, "step": 56970 }, { "epoch": 1.1172549019607843, "grad_norm": 1.5790716409683228, "learning_rate": 4.4452866293340277e-07, "loss": 0.1457, "step": 56980 }, { "epoch": 1.1174509803921568, "grad_norm": 2.448697090148926, "learning_rate": 4.4436756167216484e-07, "loss": 0.1247, "step": 56990 }, { "epoch": 1.1176470588235294, "grad_norm": 5.576593399047852, "learning_rate": 4.442064662585622e-07, "loss": 0.1385, "step": 57000 }, { "epoch": 1.117843137254902, "grad_norm": 1.9820785522460938, "learning_rate": 4.440453767095279e-07, "loss": 0.119, "step": 57010 }, { "epoch": 1.1180392156862746, "grad_norm": 3.158496379852295, "learning_rate": 4.438842930419946e-07, "loss": 0.1616, "step": 57020 }, { "epoch": 1.118235294117647, "grad_norm": 1.7211345434188843, "learning_rate": 4.437232152728938e-07, "loss": 0.1149, "step": 57030 }, { "epoch": 1.1184313725490196, "grad_norm": 2.914713144302368, "learning_rate": 4.4356214341915693e-07, "loss": 0.1209, "step": 57040 }, { "epoch": 1.1186274509803922, "grad_norm": 3.0572643280029297, "learning_rate": 4.434010774977146e-07, "loss": 0.1355, "step": 57050 }, { "epoch": 1.1188235294117648, "grad_norm": 1.813072681427002, "learning_rate": 4.4324001752549654e-07, "loss": 0.1577, "step": 57060 }, { "epoch": 1.1190196078431374, "grad_norm": 1.3149412870407104, "learning_rate": 4.4307896351943233e-07, "loss": 0.1337, "step": 57070 }, { "epoch": 1.1192156862745097, "grad_norm": 3.2570106983184814, "learning_rate": 4.429179154964505e-07, "loss": 0.1204, "step": 57080 }, { "epoch": 1.1194117647058823, "grad_norm": 3.6008946895599365, "learning_rate": 4.4275687347347915e-07, "loss": 0.1027, "step": 57090 }, { "epoch": 1.119607843137255, "grad_norm": 1.964137315750122, "learning_rate": 4.4259583746744587e-07, "loss": 0.1175, "step": 57100 }, { "epoch": 1.1198039215686275, "grad_norm": 2.105631113052368, "learning_rate": 4.424348074952773e-07, "loss": 0.138, "step": 57110 }, { "epoch": 1.12, "grad_norm": 2.2135136127471924, "learning_rate": 4.4227378357389967e-07, "loss": 0.1146, "step": 57120 }, { "epoch": 1.1201960784313725, "grad_norm": 2.9496169090270996, "learning_rate": 4.4211276572023837e-07, "loss": 0.1288, "step": 57130 }, { "epoch": 1.120392156862745, "grad_norm": 2.3328728675842285, "learning_rate": 4.4195175395121865e-07, "loss": 0.1179, "step": 57140 }, { "epoch": 1.1205882352941177, "grad_norm": 2.4963691234588623, "learning_rate": 4.417907482837646e-07, "loss": 0.1198, "step": 57150 }, { "epoch": 1.1207843137254903, "grad_norm": 4.018641471862793, "learning_rate": 4.416297487347998e-07, "loss": 0.1279, "step": 57160 }, { "epoch": 1.1209803921568628, "grad_norm": 1.5974138975143433, "learning_rate": 4.414687553212473e-07, "loss": 0.0996, "step": 57170 }, { "epoch": 1.1211764705882352, "grad_norm": 2.2803618907928467, "learning_rate": 4.4130776806002947e-07, "loss": 0.1376, "step": 57180 }, { "epoch": 1.1213725490196078, "grad_norm": 1.2488211393356323, "learning_rate": 4.411467869680678e-07, "loss": 0.1059, "step": 57190 }, { "epoch": 1.1215686274509804, "grad_norm": 4.288068771362305, "learning_rate": 4.409858120622835e-07, "loss": 0.1236, "step": 57200 }, { "epoch": 1.121764705882353, "grad_norm": 2.202702045440674, "learning_rate": 4.4082484335959703e-07, "loss": 0.1279, "step": 57210 }, { "epoch": 1.1219607843137256, "grad_norm": 1.4811500310897827, "learning_rate": 4.4066388087692783e-07, "loss": 0.1127, "step": 57220 }, { "epoch": 1.122156862745098, "grad_norm": 3.134615898132324, "learning_rate": 4.4050292463119537e-07, "loss": 0.1129, "step": 57230 }, { "epoch": 1.1223529411764706, "grad_norm": 3.0152199268341064, "learning_rate": 4.403419746393177e-07, "loss": 0.1332, "step": 57240 }, { "epoch": 1.1225490196078431, "grad_norm": 2.302363872528076, "learning_rate": 4.401810309182128e-07, "loss": 0.1472, "step": 57250 }, { "epoch": 1.1227450980392157, "grad_norm": 1.8501238822937012, "learning_rate": 4.4002009348479794e-07, "loss": 0.1211, "step": 57260 }, { "epoch": 1.1229411764705883, "grad_norm": 2.2417490482330322, "learning_rate": 4.398591623559892e-07, "loss": 0.1393, "step": 57270 }, { "epoch": 1.1231372549019607, "grad_norm": 1.0623137950897217, "learning_rate": 4.396982375487026e-07, "loss": 0.1097, "step": 57280 }, { "epoch": 1.1233333333333333, "grad_norm": 3.211698055267334, "learning_rate": 4.3953731907985323e-07, "loss": 0.1485, "step": 57290 }, { "epoch": 1.1235294117647059, "grad_norm": 3.2852113246917725, "learning_rate": 4.3937640696635553e-07, "loss": 0.1426, "step": 57300 }, { "epoch": 1.1237254901960785, "grad_norm": 2.099839210510254, "learning_rate": 4.3921550122512317e-07, "loss": 0.1153, "step": 57310 }, { "epoch": 1.123921568627451, "grad_norm": 1.404814600944519, "learning_rate": 4.390546018730695e-07, "loss": 0.1175, "step": 57320 }, { "epoch": 1.1241176470588234, "grad_norm": 2.211327314376831, "learning_rate": 4.388937089271068e-07, "loss": 0.1112, "step": 57330 }, { "epoch": 1.124313725490196, "grad_norm": 4.023961067199707, "learning_rate": 4.387328224041471e-07, "loss": 0.1049, "step": 57340 }, { "epoch": 1.1245098039215686, "grad_norm": 1.889855980873108, "learning_rate": 4.385719423211012e-07, "loss": 0.1265, "step": 57350 }, { "epoch": 1.1247058823529412, "grad_norm": 1.9716863632202148, "learning_rate": 4.3841106869487955e-07, "loss": 0.1272, "step": 57360 }, { "epoch": 1.1249019607843138, "grad_norm": 4.3349080085754395, "learning_rate": 4.3825020154239213e-07, "loss": 0.1332, "step": 57370 }, { "epoch": 1.1250980392156862, "grad_norm": 1.7391412258148193, "learning_rate": 4.380893408805478e-07, "loss": 0.1442, "step": 57380 }, { "epoch": 1.1252941176470588, "grad_norm": 2.139174699783325, "learning_rate": 4.3792848672625486e-07, "loss": 0.0996, "step": 57390 }, { "epoch": 1.1254901960784314, "grad_norm": 3.4281928539276123, "learning_rate": 4.3776763909642135e-07, "loss": 0.1238, "step": 57400 }, { "epoch": 1.125686274509804, "grad_norm": 1.2345081567764282, "learning_rate": 4.376067980079539e-07, "loss": 0.1073, "step": 57410 }, { "epoch": 1.1258823529411766, "grad_norm": 1.731722116470337, "learning_rate": 4.374459634777592e-07, "loss": 0.116, "step": 57420 }, { "epoch": 1.126078431372549, "grad_norm": 2.039785385131836, "learning_rate": 4.3728513552274247e-07, "loss": 0.1279, "step": 57430 }, { "epoch": 1.1262745098039215, "grad_norm": 1.6705540418624878, "learning_rate": 4.37124314159809e-07, "loss": 0.1318, "step": 57440 }, { "epoch": 1.1264705882352941, "grad_norm": 2.8535103797912598, "learning_rate": 4.3696349940586293e-07, "loss": 0.1282, "step": 57450 }, { "epoch": 1.1266666666666667, "grad_norm": 2.0242698192596436, "learning_rate": 4.368026912778077e-07, "loss": 0.0977, "step": 57460 }, { "epoch": 1.1268627450980393, "grad_norm": 1.5315910577774048, "learning_rate": 4.366418897925463e-07, "loss": 0.1054, "step": 57470 }, { "epoch": 1.1270588235294117, "grad_norm": 2.4203438758850098, "learning_rate": 4.364810949669809e-07, "loss": 0.1126, "step": 57480 }, { "epoch": 1.1272549019607843, "grad_norm": 1.3578569889068604, "learning_rate": 4.363203068180127e-07, "loss": 0.1386, "step": 57490 }, { "epoch": 1.1274509803921569, "grad_norm": 2.401380777359009, "learning_rate": 4.361595253625429e-07, "loss": 0.1421, "step": 57500 }, { "epoch": 1.1276470588235294, "grad_norm": 1.4961096048355103, "learning_rate": 4.359987506174713e-07, "loss": 0.1063, "step": 57510 }, { "epoch": 1.127843137254902, "grad_norm": 2.344400405883789, "learning_rate": 4.3583798259969726e-07, "loss": 0.1422, "step": 57520 }, { "epoch": 1.1280392156862744, "grad_norm": 1.949517011642456, "learning_rate": 4.356772213261196e-07, "loss": 0.1327, "step": 57530 }, { "epoch": 1.128235294117647, "grad_norm": 3.9289608001708984, "learning_rate": 4.3551646681363594e-07, "loss": 0.1195, "step": 57540 }, { "epoch": 1.1284313725490196, "grad_norm": 3.179199695587158, "learning_rate": 4.353557190791437e-07, "loss": 0.1321, "step": 57550 }, { "epoch": 1.1286274509803922, "grad_norm": 4.6054816246032715, "learning_rate": 4.351949781395395e-07, "loss": 0.1215, "step": 57560 }, { "epoch": 1.1288235294117648, "grad_norm": 2.016488790512085, "learning_rate": 4.3503424401171885e-07, "loss": 0.1147, "step": 57570 }, { "epoch": 1.1290196078431372, "grad_norm": 1.6272497177124023, "learning_rate": 4.348735167125771e-07, "loss": 0.1086, "step": 57580 }, { "epoch": 1.1292156862745097, "grad_norm": 2.2514264583587646, "learning_rate": 4.3471279625900837e-07, "loss": 0.1145, "step": 57590 }, { "epoch": 1.1294117647058823, "grad_norm": 3.8129677772521973, "learning_rate": 4.3455208266790646e-07, "loss": 0.1372, "step": 57600 }, { "epoch": 1.129607843137255, "grad_norm": 2.2880470752716064, "learning_rate": 4.3439137595616433e-07, "loss": 0.1128, "step": 57610 }, { "epoch": 1.1298039215686275, "grad_norm": 1.8551236391067505, "learning_rate": 4.3423067614067397e-07, "loss": 0.1012, "step": 57620 }, { "epoch": 1.13, "grad_norm": 2.3511505126953125, "learning_rate": 4.3406998323832713e-07, "loss": 0.1218, "step": 57630 }, { "epoch": 1.1301960784313725, "grad_norm": 3.888097047805786, "learning_rate": 4.3390929726601445e-07, "loss": 0.1119, "step": 57640 }, { "epoch": 1.130392156862745, "grad_norm": 2.324939250946045, "learning_rate": 4.3374861824062575e-07, "loss": 0.1174, "step": 57650 }, { "epoch": 1.1305882352941177, "grad_norm": 1.8748269081115723, "learning_rate": 4.335879461790506e-07, "loss": 0.1106, "step": 57660 }, { "epoch": 1.1307843137254903, "grad_norm": 2.98046612739563, "learning_rate": 4.334272810981773e-07, "loss": 0.1266, "step": 57670 }, { "epoch": 1.1309803921568626, "grad_norm": 1.6780846118927002, "learning_rate": 4.3326662301489385e-07, "loss": 0.1275, "step": 57680 }, { "epoch": 1.1311764705882352, "grad_norm": 1.2061961889266968, "learning_rate": 4.331059719460875e-07, "loss": 0.1095, "step": 57690 }, { "epoch": 1.1313725490196078, "grad_norm": 1.8448137044906616, "learning_rate": 4.329453279086443e-07, "loss": 0.1292, "step": 57700 }, { "epoch": 1.1315686274509804, "grad_norm": 3.635749101638794, "learning_rate": 4.3278469091945e-07, "loss": 0.1098, "step": 57710 }, { "epoch": 1.131764705882353, "grad_norm": 2.1816811561584473, "learning_rate": 4.326240609953895e-07, "loss": 0.1198, "step": 57720 }, { "epoch": 1.1319607843137254, "grad_norm": 3.265768051147461, "learning_rate": 4.324634381533467e-07, "loss": 0.1171, "step": 57730 }, { "epoch": 1.132156862745098, "grad_norm": 2.169647455215454, "learning_rate": 4.3230282241020534e-07, "loss": 0.1309, "step": 57740 }, { "epoch": 1.1323529411764706, "grad_norm": 1.4409621953964233, "learning_rate": 4.3214221378284785e-07, "loss": 0.1205, "step": 57750 }, { "epoch": 1.1325490196078432, "grad_norm": 3.517622947692871, "learning_rate": 4.3198161228815606e-07, "loss": 0.1571, "step": 57760 }, { "epoch": 1.1327450980392157, "grad_norm": 8.042048454284668, "learning_rate": 4.318210179430113e-07, "loss": 0.1247, "step": 57770 }, { "epoch": 1.1329411764705881, "grad_norm": 2.524143934249878, "learning_rate": 4.316604307642937e-07, "loss": 0.1172, "step": 57780 }, { "epoch": 1.1331372549019607, "grad_norm": 1.9622820615768433, "learning_rate": 4.314998507688831e-07, "loss": 0.1238, "step": 57790 }, { "epoch": 1.1333333333333333, "grad_norm": 1.387319803237915, "learning_rate": 4.313392779736584e-07, "loss": 0.1026, "step": 57800 }, { "epoch": 1.133529411764706, "grad_norm": 3.6856656074523926, "learning_rate": 4.3117871239549764e-07, "loss": 0.1422, "step": 57810 }, { "epoch": 1.1337254901960785, "grad_norm": 1.6654157638549805, "learning_rate": 4.3101815405127814e-07, "loss": 0.1215, "step": 57820 }, { "epoch": 1.1339215686274509, "grad_norm": 1.3374418020248413, "learning_rate": 4.3085760295787665e-07, "loss": 0.1027, "step": 57830 }, { "epoch": 1.1341176470588235, "grad_norm": 2.7676634788513184, "learning_rate": 4.306970591321688e-07, "loss": 0.138, "step": 57840 }, { "epoch": 1.134313725490196, "grad_norm": 5.442105770111084, "learning_rate": 4.305365225910297e-07, "loss": 0.1324, "step": 57850 }, { "epoch": 1.1345098039215686, "grad_norm": 2.9459211826324463, "learning_rate": 4.3037599335133385e-07, "loss": 0.1058, "step": 57860 }, { "epoch": 1.1347058823529412, "grad_norm": 1.867641806602478, "learning_rate": 4.302154714299547e-07, "loss": 0.1045, "step": 57870 }, { "epoch": 1.1349019607843136, "grad_norm": 3.32600736618042, "learning_rate": 4.3005495684376514e-07, "loss": 0.1222, "step": 57880 }, { "epoch": 1.1350980392156862, "grad_norm": 2.1638076305389404, "learning_rate": 4.2989444960963697e-07, "loss": 0.1248, "step": 57890 }, { "epoch": 1.1352941176470588, "grad_norm": 2.8135907649993896, "learning_rate": 4.2973394974444143e-07, "loss": 0.1131, "step": 57900 }, { "epoch": 1.1354901960784314, "grad_norm": 3.7386856079101562, "learning_rate": 4.2957345726504926e-07, "loss": 0.1319, "step": 57910 }, { "epoch": 1.135686274509804, "grad_norm": 3.048163414001465, "learning_rate": 4.294129721883298e-07, "loss": 0.1035, "step": 57920 }, { "epoch": 1.1358823529411766, "grad_norm": 2.9964115619659424, "learning_rate": 4.2925249453115213e-07, "loss": 0.1201, "step": 57930 }, { "epoch": 1.136078431372549, "grad_norm": 2.4473276138305664, "learning_rate": 4.2909202431038434e-07, "loss": 0.1419, "step": 57940 }, { "epoch": 1.1362745098039215, "grad_norm": 1.770851492881775, "learning_rate": 4.2893156154289366e-07, "loss": 0.1021, "step": 57950 }, { "epoch": 1.1364705882352941, "grad_norm": 1.6470597982406616, "learning_rate": 4.287711062455469e-07, "loss": 0.13, "step": 57960 }, { "epoch": 1.1366666666666667, "grad_norm": 2.2729947566986084, "learning_rate": 4.286106584352096e-07, "loss": 0.1175, "step": 57970 }, { "epoch": 1.1368627450980393, "grad_norm": 2.149923086166382, "learning_rate": 4.2845021812874684e-07, "loss": 0.131, "step": 57980 }, { "epoch": 1.1370588235294117, "grad_norm": 1.3493523597717285, "learning_rate": 4.282897853430229e-07, "loss": 0.126, "step": 57990 }, { "epoch": 1.1372549019607843, "grad_norm": 2.0368380546569824, "learning_rate": 4.28129360094901e-07, "loss": 0.1406, "step": 58000 }, { "epoch": 1.1374509803921569, "grad_norm": 1.811678171157837, "learning_rate": 4.2796894240124386e-07, "loss": 0.1436, "step": 58010 }, { "epoch": 1.1376470588235295, "grad_norm": 5.340487957000732, "learning_rate": 4.278085322789134e-07, "loss": 0.1204, "step": 58020 }, { "epoch": 1.137843137254902, "grad_norm": 2.250403642654419, "learning_rate": 4.2764812974477017e-07, "loss": 0.1057, "step": 58030 }, { "epoch": 1.1380392156862744, "grad_norm": 2.653878688812256, "learning_rate": 4.274877348156751e-07, "loss": 0.1222, "step": 58040 }, { "epoch": 1.138235294117647, "grad_norm": 2.1265177726745605, "learning_rate": 4.273273475084872e-07, "loss": 0.1226, "step": 58050 }, { "epoch": 1.1384313725490196, "grad_norm": 2.8553519248962402, "learning_rate": 4.2716696784006514e-07, "loss": 0.1166, "step": 58060 }, { "epoch": 1.1386274509803922, "grad_norm": 2.0843963623046875, "learning_rate": 4.2700659582726695e-07, "loss": 0.117, "step": 58070 }, { "epoch": 1.1388235294117648, "grad_norm": 2.4681310653686523, "learning_rate": 4.268462314869493e-07, "loss": 0.1349, "step": 58080 }, { "epoch": 1.1390196078431372, "grad_norm": 3.070858955383301, "learning_rate": 4.266858748359686e-07, "loss": 0.1241, "step": 58090 }, { "epoch": 1.1392156862745098, "grad_norm": 1.305370807647705, "learning_rate": 4.2652552589118037e-07, "loss": 0.1291, "step": 58100 }, { "epoch": 1.1394117647058823, "grad_norm": 3.2382972240448, "learning_rate": 4.2636518466943896e-07, "loss": 0.1264, "step": 58110 }, { "epoch": 1.139607843137255, "grad_norm": 3.195861577987671, "learning_rate": 4.2620485118759833e-07, "loss": 0.1227, "step": 58120 }, { "epoch": 1.1398039215686275, "grad_norm": 2.592710256576538, "learning_rate": 4.2604452546251125e-07, "loss": 0.1283, "step": 58130 }, { "epoch": 1.1400000000000001, "grad_norm": 1.6362384557724, "learning_rate": 4.2588420751103e-07, "loss": 0.1239, "step": 58140 }, { "epoch": 1.1401960784313725, "grad_norm": 2.39884614944458, "learning_rate": 4.257238973500061e-07, "loss": 0.1472, "step": 58150 }, { "epoch": 1.140392156862745, "grad_norm": 2.638500452041626, "learning_rate": 4.255635949962898e-07, "loss": 0.1257, "step": 58160 }, { "epoch": 1.1405882352941177, "grad_norm": 2.369736909866333, "learning_rate": 4.2540330046673076e-07, "loss": 0.1095, "step": 58170 }, { "epoch": 1.1407843137254903, "grad_norm": 2.3428263664245605, "learning_rate": 4.2524301377817815e-07, "loss": 0.1147, "step": 58180 }, { "epoch": 1.1409803921568629, "grad_norm": 4.427995204925537, "learning_rate": 4.2508273494747984e-07, "loss": 0.1241, "step": 58190 }, { "epoch": 1.1411764705882352, "grad_norm": 3.6954588890075684, "learning_rate": 4.249224639914828e-07, "loss": 0.116, "step": 58200 }, { "epoch": 1.1413725490196078, "grad_norm": 2.4992496967315674, "learning_rate": 4.2476220092703387e-07, "loss": 0.1289, "step": 58210 }, { "epoch": 1.1415686274509804, "grad_norm": 1.308821439743042, "learning_rate": 4.2460194577097843e-07, "loss": 0.1138, "step": 58220 }, { "epoch": 1.141764705882353, "grad_norm": 1.7893277406692505, "learning_rate": 4.2444169854016125e-07, "loss": 0.1458, "step": 58230 }, { "epoch": 1.1419607843137256, "grad_norm": 4.67758846282959, "learning_rate": 4.2428145925142615e-07, "loss": 0.1315, "step": 58240 }, { "epoch": 1.142156862745098, "grad_norm": 2.588898181915283, "learning_rate": 4.2412122792161623e-07, "loss": 0.1237, "step": 58250 }, { "epoch": 1.1423529411764706, "grad_norm": 3.1429550647735596, "learning_rate": 4.239610045675739e-07, "loss": 0.1356, "step": 58260 }, { "epoch": 1.1425490196078432, "grad_norm": 6.177664279937744, "learning_rate": 4.2380078920614026e-07, "loss": 0.1169, "step": 58270 }, { "epoch": 1.1427450980392158, "grad_norm": 3.368267297744751, "learning_rate": 4.236405818541561e-07, "loss": 0.1283, "step": 58280 }, { "epoch": 1.1429411764705883, "grad_norm": 1.8847028017044067, "learning_rate": 4.23480382528461e-07, "loss": 0.1578, "step": 58290 }, { "epoch": 1.1431372549019607, "grad_norm": 2.797987699508667, "learning_rate": 4.2332019124589373e-07, "loss": 0.1262, "step": 58300 }, { "epoch": 1.1433333333333333, "grad_norm": 1.461172103881836, "learning_rate": 4.231600080232926e-07, "loss": 0.1043, "step": 58310 }, { "epoch": 1.143529411764706, "grad_norm": 1.9165492057800293, "learning_rate": 4.229998328774945e-07, "loss": 0.0985, "step": 58320 }, { "epoch": 1.1437254901960785, "grad_norm": 3.6613330841064453, "learning_rate": 4.228396658253359e-07, "loss": 0.1311, "step": 58330 }, { "epoch": 1.143921568627451, "grad_norm": 1.8392459154129028, "learning_rate": 4.2267950688365233e-07, "loss": 0.1259, "step": 58340 }, { "epoch": 1.1441176470588235, "grad_norm": 3.2727136611938477, "learning_rate": 4.2251935606927824e-07, "loss": 0.1185, "step": 58350 }, { "epoch": 1.144313725490196, "grad_norm": 1.367798089981079, "learning_rate": 4.223592133990475e-07, "loss": 0.0912, "step": 58360 }, { "epoch": 1.1445098039215686, "grad_norm": 1.9088075160980225, "learning_rate": 4.2219907888979313e-07, "loss": 0.1217, "step": 58370 }, { "epoch": 1.1447058823529412, "grad_norm": 2.075138807296753, "learning_rate": 4.220389525583467e-07, "loss": 0.1341, "step": 58380 }, { "epoch": 1.1449019607843138, "grad_norm": 1.9149649143218994, "learning_rate": 4.218788344215401e-07, "loss": 0.113, "step": 58390 }, { "epoch": 1.1450980392156862, "grad_norm": 2.426027536392212, "learning_rate": 4.2171872449620325e-07, "loss": 0.0935, "step": 58400 }, { "epoch": 1.1452941176470588, "grad_norm": 3.8152718544006348, "learning_rate": 4.215586227991657e-07, "loss": 0.1231, "step": 58410 }, { "epoch": 1.1454901960784314, "grad_norm": 2.107529878616333, "learning_rate": 4.2139852934725615e-07, "loss": 0.1179, "step": 58420 }, { "epoch": 1.145686274509804, "grad_norm": 1.6567230224609375, "learning_rate": 4.212384441573022e-07, "loss": 0.1308, "step": 58430 }, { "epoch": 1.1458823529411766, "grad_norm": 4.155142307281494, "learning_rate": 4.210783672461308e-07, "loss": 0.1314, "step": 58440 }, { "epoch": 1.146078431372549, "grad_norm": 4.289221286773682, "learning_rate": 4.209182986305679e-07, "loss": 0.1368, "step": 58450 }, { "epoch": 1.1462745098039215, "grad_norm": 3.9872868061065674, "learning_rate": 4.207582383274386e-07, "loss": 0.1343, "step": 58460 }, { "epoch": 1.1464705882352941, "grad_norm": 1.2502537965774536, "learning_rate": 4.2059818635356736e-07, "loss": 0.1361, "step": 58470 }, { "epoch": 1.1466666666666667, "grad_norm": 1.7171475887298584, "learning_rate": 4.204381427257772e-07, "loss": 0.1088, "step": 58480 }, { "epoch": 1.1468627450980393, "grad_norm": 3.1180918216705322, "learning_rate": 4.2027810746089087e-07, "loss": 0.1214, "step": 58490 }, { "epoch": 1.1470588235294117, "grad_norm": 1.617401361465454, "learning_rate": 4.2011808057573004e-07, "loss": 0.0868, "step": 58500 }, { "epoch": 1.1472549019607843, "grad_norm": 2.7529489994049072, "learning_rate": 4.199580620871153e-07, "loss": 0.1407, "step": 58510 }, { "epoch": 1.1474509803921569, "grad_norm": 2.5163509845733643, "learning_rate": 4.197980520118665e-07, "loss": 0.131, "step": 58520 }, { "epoch": 1.1476470588235295, "grad_norm": 1.7994285821914673, "learning_rate": 4.1963805036680284e-07, "loss": 0.1237, "step": 58530 }, { "epoch": 1.147843137254902, "grad_norm": 3.0286343097686768, "learning_rate": 4.1947805716874213e-07, "loss": 0.1165, "step": 58540 }, { "epoch": 1.1480392156862744, "grad_norm": 2.3651113510131836, "learning_rate": 4.1931807243450166e-07, "loss": 0.1412, "step": 58550 }, { "epoch": 1.148235294117647, "grad_norm": 1.9831873178482056, "learning_rate": 4.191580961808977e-07, "loss": 0.1149, "step": 58560 }, { "epoch": 1.1484313725490196, "grad_norm": 2.2824082374572754, "learning_rate": 4.189981284247458e-07, "loss": 0.1189, "step": 58570 }, { "epoch": 1.1486274509803922, "grad_norm": 1.3988896608352661, "learning_rate": 4.188381691828606e-07, "loss": 0.1256, "step": 58580 }, { "epoch": 1.1488235294117648, "grad_norm": 3.17697811126709, "learning_rate": 4.1867821847205536e-07, "loss": 0.1165, "step": 58590 }, { "epoch": 1.1490196078431372, "grad_norm": 2.443286657333374, "learning_rate": 4.1851827630914306e-07, "loss": 0.1111, "step": 58600 }, { "epoch": 1.1492156862745098, "grad_norm": 1.8202590942382812, "learning_rate": 4.1835834271093563e-07, "loss": 0.1242, "step": 58610 }, { "epoch": 1.1494117647058824, "grad_norm": 1.6812026500701904, "learning_rate": 4.1819841769424366e-07, "loss": 0.075, "step": 58620 }, { "epoch": 1.149607843137255, "grad_norm": 2.9307498931884766, "learning_rate": 4.180385012758776e-07, "loss": 0.1412, "step": 58630 }, { "epoch": 1.1498039215686275, "grad_norm": 2.8227717876434326, "learning_rate": 4.1787859347264625e-07, "loss": 0.1346, "step": 58640 }, { "epoch": 1.15, "grad_norm": 2.4766225814819336, "learning_rate": 4.177186943013579e-07, "loss": 0.125, "step": 58650 }, { "epoch": 1.1501960784313725, "grad_norm": 2.67691969871521, "learning_rate": 4.175588037788201e-07, "loss": 0.1288, "step": 58660 }, { "epoch": 1.150392156862745, "grad_norm": 1.6099135875701904, "learning_rate": 4.173989219218389e-07, "loss": 0.113, "step": 58670 }, { "epoch": 1.1505882352941177, "grad_norm": 10.270928382873535, "learning_rate": 4.1723904874722005e-07, "loss": 0.1124, "step": 58680 }, { "epoch": 1.1507843137254903, "grad_norm": 1.5807297229766846, "learning_rate": 4.1707918427176806e-07, "loss": 0.1314, "step": 58690 }, { "epoch": 1.1509803921568627, "grad_norm": 2.4217607975006104, "learning_rate": 4.1691932851228664e-07, "loss": 0.1114, "step": 58700 }, { "epoch": 1.1511764705882352, "grad_norm": 2.1181747913360596, "learning_rate": 4.1675948148557837e-07, "loss": 0.123, "step": 58710 }, { "epoch": 1.1513725490196078, "grad_norm": 2.6922354698181152, "learning_rate": 4.165996432084454e-07, "loss": 0.1222, "step": 58720 }, { "epoch": 1.1515686274509804, "grad_norm": 1.768454670906067, "learning_rate": 4.1643981369768833e-07, "loss": 0.1145, "step": 58730 }, { "epoch": 1.151764705882353, "grad_norm": 3.1329238414764404, "learning_rate": 4.162799929701072e-07, "loss": 0.1314, "step": 58740 }, { "epoch": 1.1519607843137254, "grad_norm": 2.594435453414917, "learning_rate": 4.1612018104250127e-07, "loss": 0.1199, "step": 58750 }, { "epoch": 1.152156862745098, "grad_norm": 1.9060975313186646, "learning_rate": 4.1596037793166866e-07, "loss": 0.1252, "step": 58760 }, { "epoch": 1.1523529411764706, "grad_norm": 1.8022956848144531, "learning_rate": 4.158005836544066e-07, "loss": 0.1347, "step": 58770 }, { "epoch": 1.1525490196078432, "grad_norm": 1.9498085975646973, "learning_rate": 4.156407982275112e-07, "loss": 0.1, "step": 58780 }, { "epoch": 1.1527450980392158, "grad_norm": 1.7818598747253418, "learning_rate": 4.1548102166777786e-07, "loss": 0.1524, "step": 58790 }, { "epoch": 1.1529411764705881, "grad_norm": 1.0778934955596924, "learning_rate": 4.1532125399200124e-07, "loss": 0.1252, "step": 58800 }, { "epoch": 1.1531372549019607, "grad_norm": 2.491777181625366, "learning_rate": 4.1516149521697457e-07, "loss": 0.1462, "step": 58810 }, { "epoch": 1.1533333333333333, "grad_norm": 3.4163148403167725, "learning_rate": 4.150017453594906e-07, "loss": 0.1201, "step": 58820 }, { "epoch": 1.153529411764706, "grad_norm": 1.9770666360855103, "learning_rate": 4.1484200443634076e-07, "loss": 0.141, "step": 58830 }, { "epoch": 1.1537254901960785, "grad_norm": 8.045014381408691, "learning_rate": 4.1468227246431587e-07, "loss": 0.1131, "step": 58840 }, { "epoch": 1.1539215686274509, "grad_norm": 1.9330774545669556, "learning_rate": 4.145225494602057e-07, "loss": 0.1154, "step": 58850 }, { "epoch": 1.1541176470588235, "grad_norm": 1.4201686382293701, "learning_rate": 4.1436283544079894e-07, "loss": 0.1219, "step": 58860 }, { "epoch": 1.154313725490196, "grad_norm": 3.700789451599121, "learning_rate": 4.1420313042288344e-07, "loss": 0.1357, "step": 58870 }, { "epoch": 1.1545098039215687, "grad_norm": 4.0442609786987305, "learning_rate": 4.140434344232463e-07, "loss": 0.1229, "step": 58880 }, { "epoch": 1.1547058823529412, "grad_norm": 2.326788902282715, "learning_rate": 4.138837474586732e-07, "loss": 0.1384, "step": 58890 }, { "epoch": 1.1549019607843136, "grad_norm": 2.7427048683166504, "learning_rate": 4.1372406954594933e-07, "loss": 0.1359, "step": 58900 }, { "epoch": 1.1550980392156862, "grad_norm": 2.2910540103912354, "learning_rate": 4.135644007018589e-07, "loss": 0.0958, "step": 58910 }, { "epoch": 1.1552941176470588, "grad_norm": 4.114309310913086, "learning_rate": 4.134047409431844e-07, "loss": 0.1415, "step": 58920 }, { "epoch": 1.1554901960784314, "grad_norm": 1.9527952671051025, "learning_rate": 4.132450902867088e-07, "loss": 0.141, "step": 58930 }, { "epoch": 1.155686274509804, "grad_norm": 4.251617431640625, "learning_rate": 4.130854487492128e-07, "loss": 0.1751, "step": 58940 }, { "epoch": 1.1558823529411764, "grad_norm": 2.7397775650024414, "learning_rate": 4.1292581634747673e-07, "loss": 0.1124, "step": 58950 }, { "epoch": 1.156078431372549, "grad_norm": 2.8555288314819336, "learning_rate": 4.1276619309828003e-07, "loss": 0.1095, "step": 58960 }, { "epoch": 1.1562745098039215, "grad_norm": 1.9270182847976685, "learning_rate": 4.126065790184008e-07, "loss": 0.0987, "step": 58970 }, { "epoch": 1.1564705882352941, "grad_norm": 5.717408657073975, "learning_rate": 4.124469741246165e-07, "loss": 0.1306, "step": 58980 }, { "epoch": 1.1566666666666667, "grad_norm": 1.8862242698669434, "learning_rate": 4.122873784337034e-07, "loss": 0.1233, "step": 58990 }, { "epoch": 1.156862745098039, "grad_norm": 2.8591699600219727, "learning_rate": 4.1212779196243696e-07, "loss": 0.117, "step": 59000 }, { "epoch": 1.1570588235294117, "grad_norm": 3.1912379264831543, "learning_rate": 4.119682147275918e-07, "loss": 0.1338, "step": 59010 }, { "epoch": 1.1572549019607843, "grad_norm": 1.867116093635559, "learning_rate": 4.118086467459412e-07, "loss": 0.1319, "step": 59020 }, { "epoch": 1.1574509803921569, "grad_norm": 1.5679243803024292, "learning_rate": 4.116490880342577e-07, "loss": 0.1543, "step": 59030 }, { "epoch": 1.1576470588235295, "grad_norm": 3.203594446182251, "learning_rate": 4.1148953860931294e-07, "loss": 0.1407, "step": 59040 }, { "epoch": 1.1578431372549018, "grad_norm": 1.9908262491226196, "learning_rate": 4.1132999848787733e-07, "loss": 0.1046, "step": 59050 }, { "epoch": 1.1580392156862744, "grad_norm": 1.2357149124145508, "learning_rate": 4.1117046768672047e-07, "loss": 0.1129, "step": 59060 }, { "epoch": 1.158235294117647, "grad_norm": 2.3414392471313477, "learning_rate": 4.1101094622261113e-07, "loss": 0.1364, "step": 59070 }, { "epoch": 1.1584313725490196, "grad_norm": 2.111802101135254, "learning_rate": 4.108514341123167e-07, "loss": 0.1625, "step": 59080 }, { "epoch": 1.1586274509803922, "grad_norm": 2.1738336086273193, "learning_rate": 4.1069193137260386e-07, "loss": 0.1305, "step": 59090 }, { "epoch": 1.1588235294117646, "grad_norm": 1.8540394306182861, "learning_rate": 4.105324380202382e-07, "loss": 0.1178, "step": 59100 }, { "epoch": 1.1590196078431372, "grad_norm": 1.3940337896347046, "learning_rate": 4.103729540719847e-07, "loss": 0.1133, "step": 59110 }, { "epoch": 1.1592156862745098, "grad_norm": 1.0741530656814575, "learning_rate": 4.102134795446068e-07, "loss": 0.1231, "step": 59120 }, { "epoch": 1.1594117647058824, "grad_norm": 2.0449678897857666, "learning_rate": 4.100540144548672e-07, "loss": 0.1158, "step": 59130 }, { "epoch": 1.159607843137255, "grad_norm": 3.1363985538482666, "learning_rate": 4.098945588195275e-07, "loss": 0.1313, "step": 59140 }, { "epoch": 1.1598039215686273, "grad_norm": 1.6521828174591064, "learning_rate": 4.0973511265534867e-07, "loss": 0.0973, "step": 59150 }, { "epoch": 1.16, "grad_norm": 1.880245327949524, "learning_rate": 4.0957567597909e-07, "loss": 0.1254, "step": 59160 }, { "epoch": 1.1601960784313725, "grad_norm": 1.4397281408309937, "learning_rate": 4.0941624880751067e-07, "loss": 0.1135, "step": 59170 }, { "epoch": 1.160392156862745, "grad_norm": 3.2663562297821045, "learning_rate": 4.0925683115736797e-07, "loss": 0.1278, "step": 59180 }, { "epoch": 1.1605882352941177, "grad_norm": 4.115224838256836, "learning_rate": 4.0909742304541875e-07, "loss": 0.1378, "step": 59190 }, { "epoch": 1.1607843137254903, "grad_norm": 2.9812979698181152, "learning_rate": 4.0893802448841893e-07, "loss": 0.13, "step": 59200 }, { "epoch": 1.1609803921568627, "grad_norm": 2.283684730529785, "learning_rate": 4.087786355031228e-07, "loss": 0.125, "step": 59210 }, { "epoch": 1.1611764705882353, "grad_norm": 3.4101614952087402, "learning_rate": 4.086192561062843e-07, "loss": 0.1495, "step": 59220 }, { "epoch": 1.1613725490196078, "grad_norm": 2.5682666301727295, "learning_rate": 4.0845988631465617e-07, "loss": 0.1316, "step": 59230 }, { "epoch": 1.1615686274509804, "grad_norm": 2.9866387844085693, "learning_rate": 4.083005261449899e-07, "loss": 0.1342, "step": 59240 }, { "epoch": 1.161764705882353, "grad_norm": 2.169037103652954, "learning_rate": 4.081411756140362e-07, "loss": 0.1227, "step": 59250 }, { "epoch": 1.1619607843137254, "grad_norm": 2.0134851932525635, "learning_rate": 4.079818347385449e-07, "loss": 0.1407, "step": 59260 }, { "epoch": 1.162156862745098, "grad_norm": 1.9118402004241943, "learning_rate": 4.078225035352643e-07, "loss": 0.1103, "step": 59270 }, { "epoch": 1.1623529411764706, "grad_norm": 2.57450008392334, "learning_rate": 4.076631820209422e-07, "loss": 0.1265, "step": 59280 }, { "epoch": 1.1625490196078432, "grad_norm": 3.3651785850524902, "learning_rate": 4.0750387021232524e-07, "loss": 0.1279, "step": 59290 }, { "epoch": 1.1627450980392158, "grad_norm": 2.1813876628875732, "learning_rate": 4.0734456812615904e-07, "loss": 0.1385, "step": 59300 }, { "epoch": 1.1629411764705881, "grad_norm": 2.6954407691955566, "learning_rate": 4.071852757791881e-07, "loss": 0.1258, "step": 59310 }, { "epoch": 1.1631372549019607, "grad_norm": 1.5124086141586304, "learning_rate": 4.070259931881559e-07, "loss": 0.1476, "step": 59320 }, { "epoch": 1.1633333333333333, "grad_norm": 2.1428282260894775, "learning_rate": 4.0686672036980506e-07, "loss": 0.1122, "step": 59330 }, { "epoch": 1.163529411764706, "grad_norm": 3.959827184677124, "learning_rate": 4.067074573408769e-07, "loss": 0.1376, "step": 59340 }, { "epoch": 1.1637254901960785, "grad_norm": 1.9352320432662964, "learning_rate": 4.0654820411811186e-07, "loss": 0.1051, "step": 59350 }, { "epoch": 1.1639215686274509, "grad_norm": 2.8426289558410645, "learning_rate": 4.0638896071824965e-07, "loss": 0.119, "step": 59360 }, { "epoch": 1.1641176470588235, "grad_norm": 3.3966896533966064, "learning_rate": 4.0622972715802826e-07, "loss": 0.1112, "step": 59370 }, { "epoch": 1.164313725490196, "grad_norm": 2.840606212615967, "learning_rate": 4.0607050345418537e-07, "loss": 0.1234, "step": 59380 }, { "epoch": 1.1645098039215687, "grad_norm": 3.3164987564086914, "learning_rate": 4.059112896234572e-07, "loss": 0.1282, "step": 59390 }, { "epoch": 1.1647058823529413, "grad_norm": 1.3058021068572998, "learning_rate": 4.057520856825789e-07, "loss": 0.1403, "step": 59400 }, { "epoch": 1.1649019607843139, "grad_norm": 3.220930337905884, "learning_rate": 4.055928916482848e-07, "loss": 0.1336, "step": 59410 }, { "epoch": 1.1650980392156862, "grad_norm": 3.6014442443847656, "learning_rate": 4.0543370753730825e-07, "loss": 0.1464, "step": 59420 }, { "epoch": 1.1652941176470588, "grad_norm": 3.064793825149536, "learning_rate": 4.052745333663812e-07, "loss": 0.1098, "step": 59430 }, { "epoch": 1.1654901960784314, "grad_norm": 2.0056352615356445, "learning_rate": 4.0511536915223474e-07, "loss": 0.0986, "step": 59440 }, { "epoch": 1.165686274509804, "grad_norm": 1.4040818214416504, "learning_rate": 4.0495621491159917e-07, "loss": 0.1039, "step": 59450 }, { "epoch": 1.1658823529411766, "grad_norm": 3.429736375808716, "learning_rate": 4.0479707066120306e-07, "loss": 0.1096, "step": 59460 }, { "epoch": 1.166078431372549, "grad_norm": 1.8367340564727783, "learning_rate": 4.0463793641777495e-07, "loss": 0.1505, "step": 59470 }, { "epoch": 1.1662745098039216, "grad_norm": 3.8150365352630615, "learning_rate": 4.0447881219804134e-07, "loss": 0.1356, "step": 59480 }, { "epoch": 1.1664705882352941, "grad_norm": 1.5673866271972656, "learning_rate": 4.0431969801872835e-07, "loss": 0.1531, "step": 59490 }, { "epoch": 1.1666666666666667, "grad_norm": 1.5773961544036865, "learning_rate": 4.0416059389656073e-07, "loss": 0.1131, "step": 59500 }, { "epoch": 1.1668627450980393, "grad_norm": 2.525420665740967, "learning_rate": 4.04001499848262e-07, "loss": 0.1229, "step": 59510 }, { "epoch": 1.1670588235294117, "grad_norm": 3.3089613914489746, "learning_rate": 4.038424158905551e-07, "loss": 0.1407, "step": 59520 }, { "epoch": 1.1672549019607843, "grad_norm": 2.3779380321502686, "learning_rate": 4.036833420401615e-07, "loss": 0.1365, "step": 59530 }, { "epoch": 1.167450980392157, "grad_norm": 2.8707563877105713, "learning_rate": 4.035242783138019e-07, "loss": 0.1102, "step": 59540 }, { "epoch": 1.1676470588235295, "grad_norm": 2.362030029296875, "learning_rate": 4.033652247281958e-07, "loss": 0.1192, "step": 59550 }, { "epoch": 1.167843137254902, "grad_norm": 4.233871936798096, "learning_rate": 4.0320618130006146e-07, "loss": 0.1555, "step": 59560 }, { "epoch": 1.1680392156862744, "grad_norm": 3.3596715927124023, "learning_rate": 4.0304714804611637e-07, "loss": 0.1227, "step": 59570 }, { "epoch": 1.168235294117647, "grad_norm": 4.541404724121094, "learning_rate": 4.02888124983077e-07, "loss": 0.1041, "step": 59580 }, { "epoch": 1.1684313725490196, "grad_norm": 2.8756566047668457, "learning_rate": 4.0272911212765815e-07, "loss": 0.1242, "step": 59590 }, { "epoch": 1.1686274509803922, "grad_norm": 2.1615347862243652, "learning_rate": 4.0257010949657435e-07, "loss": 0.1215, "step": 59600 }, { "epoch": 1.1688235294117648, "grad_norm": 2.158951759338379, "learning_rate": 4.024111171065386e-07, "loss": 0.121, "step": 59610 }, { "epoch": 1.1690196078431372, "grad_norm": 3.17787766456604, "learning_rate": 4.0225213497426275e-07, "loss": 0.1141, "step": 59620 }, { "epoch": 1.1692156862745098, "grad_norm": 3.007934093475342, "learning_rate": 4.0209316311645767e-07, "loss": 0.1456, "step": 59630 }, { "epoch": 1.1694117647058824, "grad_norm": 2.3402183055877686, "learning_rate": 4.019342015498336e-07, "loss": 0.117, "step": 59640 }, { "epoch": 1.169607843137255, "grad_norm": 4.941279411315918, "learning_rate": 4.0177525029109904e-07, "loss": 0.1268, "step": 59650 }, { "epoch": 1.1698039215686276, "grad_norm": 3.4303996562957764, "learning_rate": 4.016163093569618e-07, "loss": 0.1322, "step": 59660 }, { "epoch": 1.17, "grad_norm": 3.5338544845581055, "learning_rate": 4.014573787641283e-07, "loss": 0.1441, "step": 59670 }, { "epoch": 1.1701960784313725, "grad_norm": 1.3974753618240356, "learning_rate": 4.0129845852930416e-07, "loss": 0.1138, "step": 59680 }, { "epoch": 1.1703921568627451, "grad_norm": 1.9523900747299194, "learning_rate": 4.0113954866919373e-07, "loss": 0.1404, "step": 59690 }, { "epoch": 1.1705882352941177, "grad_norm": 2.3204269409179688, "learning_rate": 4.0098064920050033e-07, "loss": 0.1348, "step": 59700 }, { "epoch": 1.1707843137254903, "grad_norm": 1.9924927949905396, "learning_rate": 4.008217601399264e-07, "loss": 0.1126, "step": 59710 }, { "epoch": 1.1709803921568627, "grad_norm": 4.100069046020508, "learning_rate": 4.0066288150417276e-07, "loss": 0.1305, "step": 59720 }, { "epoch": 1.1711764705882353, "grad_norm": 3.6202900409698486, "learning_rate": 4.0050401330993964e-07, "loss": 0.1129, "step": 59730 }, { "epoch": 1.1713725490196079, "grad_norm": 2.6426045894622803, "learning_rate": 4.0034515557392613e-07, "loss": 0.1159, "step": 59740 }, { "epoch": 1.1715686274509804, "grad_norm": 3.5876495838165283, "learning_rate": 4.001863083128298e-07, "loss": 0.1242, "step": 59750 }, { "epoch": 1.171764705882353, "grad_norm": 1.7723222970962524, "learning_rate": 4.000274715433475e-07, "loss": 0.1019, "step": 59760 }, { "epoch": 1.1719607843137254, "grad_norm": 4.263601303100586, "learning_rate": 3.99868645282175e-07, "loss": 0.1274, "step": 59770 }, { "epoch": 1.172156862745098, "grad_norm": 2.3134748935699463, "learning_rate": 3.9970982954600663e-07, "loss": 0.122, "step": 59780 }, { "epoch": 1.1723529411764706, "grad_norm": 2.2460219860076904, "learning_rate": 3.99551024351536e-07, "loss": 0.1231, "step": 59790 }, { "epoch": 1.1725490196078432, "grad_norm": 1.703170895576477, "learning_rate": 3.993922297154554e-07, "loss": 0.1061, "step": 59800 }, { "epoch": 1.1727450980392158, "grad_norm": 1.9416238069534302, "learning_rate": 3.9923344565445573e-07, "loss": 0.1407, "step": 59810 }, { "epoch": 1.1729411764705882, "grad_norm": 1.8935264348983765, "learning_rate": 3.990746721852277e-07, "loss": 0.1182, "step": 59820 }, { "epoch": 1.1731372549019607, "grad_norm": 1.7722547054290771, "learning_rate": 3.9891590932445985e-07, "loss": 0.132, "step": 59830 }, { "epoch": 1.1733333333333333, "grad_norm": 3.803239345550537, "learning_rate": 3.987571570888402e-07, "loss": 0.1218, "step": 59840 }, { "epoch": 1.173529411764706, "grad_norm": 2.498797655105591, "learning_rate": 3.9859841549505566e-07, "loss": 0.1182, "step": 59850 }, { "epoch": 1.1737254901960785, "grad_norm": 3.280061721801758, "learning_rate": 3.984396845597916e-07, "loss": 0.1233, "step": 59860 }, { "epoch": 1.173921568627451, "grad_norm": 2.2273242473602295, "learning_rate": 3.9828096429973277e-07, "loss": 0.1526, "step": 59870 }, { "epoch": 1.1741176470588235, "grad_norm": 2.785479784011841, "learning_rate": 3.9812225473156234e-07, "loss": 0.1385, "step": 59880 }, { "epoch": 1.174313725490196, "grad_norm": 3.3692710399627686, "learning_rate": 3.9796355587196274e-07, "loss": 0.1308, "step": 59890 }, { "epoch": 1.1745098039215687, "grad_norm": 2.546877384185791, "learning_rate": 3.9780486773761514e-07, "loss": 0.1487, "step": 59900 }, { "epoch": 1.1747058823529413, "grad_norm": 1.9816981554031372, "learning_rate": 3.976461903451994e-07, "loss": 0.1185, "step": 59910 }, { "epoch": 1.1749019607843136, "grad_norm": 2.101992130279541, "learning_rate": 3.9748752371139455e-07, "loss": 0.1641, "step": 59920 }, { "epoch": 1.1750980392156862, "grad_norm": 5.581902503967285, "learning_rate": 3.9732886785287847e-07, "loss": 0.1407, "step": 59930 }, { "epoch": 1.1752941176470588, "grad_norm": 1.5522030591964722, "learning_rate": 3.9717022278632743e-07, "loss": 0.1084, "step": 59940 }, { "epoch": 1.1754901960784314, "grad_norm": 1.8706660270690918, "learning_rate": 3.970115885284172e-07, "loss": 0.1329, "step": 59950 }, { "epoch": 1.175686274509804, "grad_norm": 2.1724343299865723, "learning_rate": 3.968529650958222e-07, "loss": 0.1106, "step": 59960 }, { "epoch": 1.1758823529411764, "grad_norm": 2.1692800521850586, "learning_rate": 3.9669435250521545e-07, "loss": 0.0976, "step": 59970 }, { "epoch": 1.176078431372549, "grad_norm": 3.140730857849121, "learning_rate": 3.9653575077326905e-07, "loss": 0.1277, "step": 59980 }, { "epoch": 1.1762745098039216, "grad_norm": 1.4888761043548584, "learning_rate": 3.9637715991665394e-07, "loss": 0.129, "step": 59990 }, { "epoch": 1.1764705882352942, "grad_norm": 2.354991912841797, "learning_rate": 3.962185799520401e-07, "loss": 0.1288, "step": 60000 }, { "epoch": 1.1766666666666667, "grad_norm": 1.8617191314697266, "learning_rate": 3.9606001089609616e-07, "loss": 0.117, "step": 60010 }, { "epoch": 1.1768627450980391, "grad_norm": 1.8140974044799805, "learning_rate": 3.959014527654894e-07, "loss": 0.113, "step": 60020 }, { "epoch": 1.1770588235294117, "grad_norm": 1.7770910263061523, "learning_rate": 3.957429055768865e-07, "loss": 0.1304, "step": 60030 }, { "epoch": 1.1772549019607843, "grad_norm": 2.180997371673584, "learning_rate": 3.955843693469523e-07, "loss": 0.1268, "step": 60040 }, { "epoch": 1.177450980392157, "grad_norm": 1.9081103801727295, "learning_rate": 3.9542584409235105e-07, "loss": 0.0993, "step": 60050 }, { "epoch": 1.1776470588235295, "grad_norm": 3.42874813079834, "learning_rate": 3.9526732982974576e-07, "loss": 0.1101, "step": 60060 }, { "epoch": 1.1778431372549019, "grad_norm": 2.3115234375, "learning_rate": 3.9510882657579793e-07, "loss": 0.1018, "step": 60070 }, { "epoch": 1.1780392156862745, "grad_norm": 7.936634063720703, "learning_rate": 3.949503343471682e-07, "loss": 0.1267, "step": 60080 }, { "epoch": 1.178235294117647, "grad_norm": 1.502160906791687, "learning_rate": 3.9479185316051635e-07, "loss": 0.1092, "step": 60090 }, { "epoch": 1.1784313725490196, "grad_norm": 1.1284857988357544, "learning_rate": 3.9463338303250015e-07, "loss": 0.0986, "step": 60100 }, { "epoch": 1.1786274509803922, "grad_norm": 1.5434117317199707, "learning_rate": 3.9447492397977696e-07, "loss": 0.0927, "step": 60110 }, { "epoch": 1.1788235294117646, "grad_norm": 2.065208911895752, "learning_rate": 3.943164760190028e-07, "loss": 0.1128, "step": 60120 }, { "epoch": 1.1790196078431372, "grad_norm": 2.5890824794769287, "learning_rate": 3.941580391668322e-07, "loss": 0.1073, "step": 60130 }, { "epoch": 1.1792156862745098, "grad_norm": 1.3607228994369507, "learning_rate": 3.93999613439919e-07, "loss": 0.1242, "step": 60140 }, { "epoch": 1.1794117647058824, "grad_norm": 2.2946252822875977, "learning_rate": 3.9384119885491564e-07, "loss": 0.1065, "step": 60150 }, { "epoch": 1.179607843137255, "grad_norm": 4.126425743103027, "learning_rate": 3.9368279542847314e-07, "loss": 0.1343, "step": 60160 }, { "epoch": 1.1798039215686273, "grad_norm": 3.280872344970703, "learning_rate": 3.935244031772417e-07, "loss": 0.1446, "step": 60170 }, { "epoch": 1.18, "grad_norm": 2.323307991027832, "learning_rate": 3.933660221178704e-07, "loss": 0.1294, "step": 60180 }, { "epoch": 1.1801960784313725, "grad_norm": 2.940072774887085, "learning_rate": 3.932076522670069e-07, "loss": 0.1433, "step": 60190 }, { "epoch": 1.1803921568627451, "grad_norm": 2.091552257537842, "learning_rate": 3.930492936412978e-07, "loss": 0.1095, "step": 60200 }, { "epoch": 1.1805882352941177, "grad_norm": 1.7848671674728394, "learning_rate": 3.928909462573884e-07, "loss": 0.1152, "step": 60210 }, { "epoch": 1.18078431372549, "grad_norm": 3.167189359664917, "learning_rate": 3.9273261013192304e-07, "loss": 0.1238, "step": 60220 }, { "epoch": 1.1809803921568627, "grad_norm": 1.8560925722122192, "learning_rate": 3.9257428528154455e-07, "loss": 0.1189, "step": 60230 }, { "epoch": 1.1811764705882353, "grad_norm": 1.9664406776428223, "learning_rate": 3.924159717228949e-07, "loss": 0.1097, "step": 60240 }, { "epoch": 1.1813725490196079, "grad_norm": 2.7852399349212646, "learning_rate": 3.922576694726148e-07, "loss": 0.1334, "step": 60250 }, { "epoch": 1.1815686274509805, "grad_norm": 1.0981653928756714, "learning_rate": 3.9209937854734347e-07, "loss": 0.1011, "step": 60260 }, { "epoch": 1.1817647058823528, "grad_norm": 2.7424814701080322, "learning_rate": 3.919410989637194e-07, "loss": 0.1064, "step": 60270 }, { "epoch": 1.1819607843137254, "grad_norm": 3.598031520843506, "learning_rate": 3.9178283073837966e-07, "loss": 0.1135, "step": 60280 }, { "epoch": 1.182156862745098, "grad_norm": 2.3220531940460205, "learning_rate": 3.9162457388796e-07, "loss": 0.1282, "step": 60290 }, { "epoch": 1.1823529411764706, "grad_norm": 3.0015616416931152, "learning_rate": 3.9146632842909516e-07, "loss": 0.112, "step": 60300 }, { "epoch": 1.1825490196078432, "grad_norm": 4.808025360107422, "learning_rate": 3.913080943784188e-07, "loss": 0.128, "step": 60310 }, { "epoch": 1.1827450980392156, "grad_norm": 2.45577335357666, "learning_rate": 3.911498717525629e-07, "loss": 0.1214, "step": 60320 }, { "epoch": 1.1829411764705882, "grad_norm": 0.8933539986610413, "learning_rate": 3.909916605681588e-07, "loss": 0.1142, "step": 60330 }, { "epoch": 1.1831372549019608, "grad_norm": 3.6043291091918945, "learning_rate": 3.9083346084183636e-07, "loss": 0.1001, "step": 60340 }, { "epoch": 1.1833333333333333, "grad_norm": 2.799798011779785, "learning_rate": 3.90675272590224e-07, "loss": 0.1389, "step": 60350 }, { "epoch": 1.183529411764706, "grad_norm": 2.2645578384399414, "learning_rate": 3.9051709582994965e-07, "loss": 0.1141, "step": 60360 }, { "epoch": 1.1837254901960783, "grad_norm": 2.395264148712158, "learning_rate": 3.9035893057763925e-07, "loss": 0.1048, "step": 60370 }, { "epoch": 1.183921568627451, "grad_norm": 4.221567153930664, "learning_rate": 3.902007768499181e-07, "loss": 0.1159, "step": 60380 }, { "epoch": 1.1841176470588235, "grad_norm": 13.22073745727539, "learning_rate": 3.900426346634098e-07, "loss": 0.1192, "step": 60390 }, { "epoch": 1.184313725490196, "grad_norm": 3.3440027236938477, "learning_rate": 3.898845040347371e-07, "loss": 0.1087, "step": 60400 }, { "epoch": 1.1845098039215687, "grad_norm": 3.038760185241699, "learning_rate": 3.897263849805216e-07, "loss": 0.135, "step": 60410 }, { "epoch": 1.184705882352941, "grad_norm": 2.2856032848358154, "learning_rate": 3.895682775173831e-07, "loss": 0.1018, "step": 60420 }, { "epoch": 1.1849019607843136, "grad_norm": 4.2612833976745605, "learning_rate": 3.8941018166194086e-07, "loss": 0.1463, "step": 60430 }, { "epoch": 1.1850980392156862, "grad_norm": 2.5442183017730713, "learning_rate": 3.892520974308127e-07, "loss": 0.1462, "step": 60440 }, { "epoch": 1.1852941176470588, "grad_norm": 2.4299819469451904, "learning_rate": 3.8909402484061496e-07, "loss": 0.1154, "step": 60450 }, { "epoch": 1.1854901960784314, "grad_norm": 2.1710662841796875, "learning_rate": 3.889359639079631e-07, "loss": 0.1089, "step": 60460 }, { "epoch": 1.185686274509804, "grad_norm": 2.3679463863372803, "learning_rate": 3.8877791464947127e-07, "loss": 0.1325, "step": 60470 }, { "epoch": 1.1858823529411764, "grad_norm": 1.1812154054641724, "learning_rate": 3.886198770817522e-07, "loss": 0.1174, "step": 60480 }, { "epoch": 1.186078431372549, "grad_norm": 1.2197060585021973, "learning_rate": 3.8846185122141755e-07, "loss": 0.1048, "step": 60490 }, { "epoch": 1.1862745098039216, "grad_norm": 1.6538773775100708, "learning_rate": 3.8830383708507784e-07, "loss": 0.1298, "step": 60500 }, { "epoch": 1.1864705882352942, "grad_norm": 1.1693739891052246, "learning_rate": 3.8814583468934217e-07, "loss": 0.124, "step": 60510 }, { "epoch": 1.1866666666666668, "grad_norm": 1.8708264827728271, "learning_rate": 3.879878440508184e-07, "loss": 0.1334, "step": 60520 }, { "epoch": 1.1868627450980391, "grad_norm": 4.392911434173584, "learning_rate": 3.878298651861133e-07, "loss": 0.1096, "step": 60530 }, { "epoch": 1.1870588235294117, "grad_norm": 2.1634042263031006, "learning_rate": 3.876718981118323e-07, "loss": 0.12, "step": 60540 }, { "epoch": 1.1872549019607843, "grad_norm": 4.3943328857421875, "learning_rate": 3.8751394284457995e-07, "loss": 0.1379, "step": 60550 }, { "epoch": 1.187450980392157, "grad_norm": 1.0943732261657715, "learning_rate": 3.8735599940095886e-07, "loss": 0.1297, "step": 60560 }, { "epoch": 1.1876470588235295, "grad_norm": 3.7615621089935303, "learning_rate": 3.8719806779757104e-07, "loss": 0.1235, "step": 60570 }, { "epoch": 1.1878431372549019, "grad_norm": 2.1347947120666504, "learning_rate": 3.870401480510167e-07, "loss": 0.1128, "step": 60580 }, { "epoch": 1.1880392156862745, "grad_norm": 1.707622766494751, "learning_rate": 3.8688224017789524e-07, "loss": 0.111, "step": 60590 }, { "epoch": 1.188235294117647, "grad_norm": 3.0331082344055176, "learning_rate": 3.867243441948048e-07, "loss": 0.1219, "step": 60600 }, { "epoch": 1.1884313725490196, "grad_norm": 3.0491700172424316, "learning_rate": 3.865664601183419e-07, "loss": 0.0985, "step": 60610 }, { "epoch": 1.1886274509803922, "grad_norm": 1.630362868309021, "learning_rate": 3.864085879651022e-07, "loss": 0.1346, "step": 60620 }, { "epoch": 1.1888235294117646, "grad_norm": 2.775050401687622, "learning_rate": 3.8625072775168e-07, "loss": 0.1128, "step": 60630 }, { "epoch": 1.1890196078431372, "grad_norm": 2.905442953109741, "learning_rate": 3.8609287949466814e-07, "loss": 0.1422, "step": 60640 }, { "epoch": 1.1892156862745098, "grad_norm": 1.3745776414871216, "learning_rate": 3.859350432106584e-07, "loss": 0.1322, "step": 60650 }, { "epoch": 1.1894117647058824, "grad_norm": 3.294236183166504, "learning_rate": 3.8577721891624144e-07, "loss": 0.0999, "step": 60660 }, { "epoch": 1.189607843137255, "grad_norm": 3.248297929763794, "learning_rate": 3.8561940662800616e-07, "loss": 0.1426, "step": 60670 }, { "epoch": 1.1898039215686274, "grad_norm": 3.4600417613983154, "learning_rate": 3.854616063625408e-07, "loss": 0.1322, "step": 60680 }, { "epoch": 1.19, "grad_norm": 3.969655990600586, "learning_rate": 3.853038181364321e-07, "loss": 0.1211, "step": 60690 }, { "epoch": 1.1901960784313725, "grad_norm": 1.9014726877212524, "learning_rate": 3.8514604196626513e-07, "loss": 0.1095, "step": 60700 }, { "epoch": 1.1903921568627451, "grad_norm": 4.435255527496338, "learning_rate": 3.8498827786862425e-07, "loss": 0.1279, "step": 60710 }, { "epoch": 1.1905882352941177, "grad_norm": 2.068410634994507, "learning_rate": 3.8483052586009243e-07, "loss": 0.1119, "step": 60720 }, { "epoch": 1.1907843137254903, "grad_norm": 2.9060890674591064, "learning_rate": 3.846727859572514e-07, "loss": 0.1181, "step": 60730 }, { "epoch": 1.1909803921568627, "grad_norm": 2.619347333908081, "learning_rate": 3.845150581766813e-07, "loss": 0.1224, "step": 60740 }, { "epoch": 1.1911764705882353, "grad_norm": 6.236408710479736, "learning_rate": 3.843573425349612e-07, "loss": 0.1252, "step": 60750 }, { "epoch": 1.1913725490196079, "grad_norm": 3.209684371948242, "learning_rate": 3.8419963904866916e-07, "loss": 0.1592, "step": 60760 }, { "epoch": 1.1915686274509805, "grad_norm": 3.4312403202056885, "learning_rate": 3.8404194773438134e-07, "loss": 0.1279, "step": 60770 }, { "epoch": 1.191764705882353, "grad_norm": 4.047525405883789, "learning_rate": 3.8388426860867316e-07, "loss": 0.1073, "step": 60780 }, { "epoch": 1.1919607843137254, "grad_norm": 2.606548547744751, "learning_rate": 3.8372660168811873e-07, "loss": 0.1001, "step": 60790 }, { "epoch": 1.192156862745098, "grad_norm": 3.050351619720459, "learning_rate": 3.835689469892904e-07, "loss": 0.1166, "step": 60800 }, { "epoch": 1.1923529411764706, "grad_norm": 3.686039686203003, "learning_rate": 3.834113045287599e-07, "loss": 0.1261, "step": 60810 }, { "epoch": 1.1925490196078432, "grad_norm": 2.8071866035461426, "learning_rate": 3.832536743230972e-07, "loss": 0.1229, "step": 60820 }, { "epoch": 1.1927450980392158, "grad_norm": 1.527093529701233, "learning_rate": 3.83096056388871e-07, "loss": 0.1286, "step": 60830 }, { "epoch": 1.1929411764705882, "grad_norm": 1.4409159421920776, "learning_rate": 3.8293845074264895e-07, "loss": 0.1022, "step": 60840 }, { "epoch": 1.1931372549019608, "grad_norm": 0.9405381679534912, "learning_rate": 3.827808574009974e-07, "loss": 0.0911, "step": 60850 }, { "epoch": 1.1933333333333334, "grad_norm": 2.668994903564453, "learning_rate": 3.826232763804811e-07, "loss": 0.1347, "step": 60860 }, { "epoch": 1.193529411764706, "grad_norm": 2.6085782051086426, "learning_rate": 3.8246570769766377e-07, "loss": 0.1144, "step": 60870 }, { "epoch": 1.1937254901960785, "grad_norm": 3.7709269523620605, "learning_rate": 3.8230815136910795e-07, "loss": 0.1427, "step": 60880 }, { "epoch": 1.193921568627451, "grad_norm": 1.5243793725967407, "learning_rate": 3.8215060741137415e-07, "loss": 0.0977, "step": 60890 }, { "epoch": 1.1941176470588235, "grad_norm": 3.389862298965454, "learning_rate": 3.819930758410228e-07, "loss": 0.1402, "step": 60900 }, { "epoch": 1.194313725490196, "grad_norm": 5.156234264373779, "learning_rate": 3.8183555667461204e-07, "loss": 0.1436, "step": 60910 }, { "epoch": 1.1945098039215687, "grad_norm": 3.962691068649292, "learning_rate": 3.816780499286991e-07, "loss": 0.1415, "step": 60920 }, { "epoch": 1.1947058823529413, "grad_norm": 3.2858176231384277, "learning_rate": 3.8152055561983965e-07, "loss": 0.1249, "step": 60930 }, { "epoch": 1.1949019607843137, "grad_norm": 1.974980115890503, "learning_rate": 3.813630737645883e-07, "loss": 0.1389, "step": 60940 }, { "epoch": 1.1950980392156862, "grad_norm": 1.6766096353530884, "learning_rate": 3.812056043794985e-07, "loss": 0.1212, "step": 60950 }, { "epoch": 1.1952941176470588, "grad_norm": 1.530430555343628, "learning_rate": 3.8104814748112185e-07, "loss": 0.1063, "step": 60960 }, { "epoch": 1.1954901960784314, "grad_norm": 1.9752280712127686, "learning_rate": 3.808907030860091e-07, "loss": 0.1428, "step": 60970 }, { "epoch": 1.195686274509804, "grad_norm": 1.631182312965393, "learning_rate": 3.8073327121070967e-07, "loss": 0.106, "step": 60980 }, { "epoch": 1.1958823529411764, "grad_norm": 1.9702413082122803, "learning_rate": 3.805758518717712e-07, "loss": 0.1325, "step": 60990 }, { "epoch": 1.196078431372549, "grad_norm": 1.187314748764038, "learning_rate": 3.8041844508574063e-07, "loss": 0.131, "step": 61000 }, { "epoch": 1.1962745098039216, "grad_norm": 2.779433012008667, "learning_rate": 3.8026105086916326e-07, "loss": 0.1323, "step": 61010 }, { "epoch": 1.1964705882352942, "grad_norm": 2.0645034313201904, "learning_rate": 3.80103669238583e-07, "loss": 0.1375, "step": 61020 }, { "epoch": 1.1966666666666668, "grad_norm": 3.1699934005737305, "learning_rate": 3.799463002105426e-07, "loss": 0.1152, "step": 61030 }, { "epoch": 1.1968627450980391, "grad_norm": 1.4613268375396729, "learning_rate": 3.7978894380158356e-07, "loss": 0.1171, "step": 61040 }, { "epoch": 1.1970588235294117, "grad_norm": 1.867820143699646, "learning_rate": 3.796316000282457e-07, "loss": 0.1217, "step": 61050 }, { "epoch": 1.1972549019607843, "grad_norm": 2.495070695877075, "learning_rate": 3.7947426890706764e-07, "loss": 0.106, "step": 61060 }, { "epoch": 1.197450980392157, "grad_norm": 2.7484288215637207, "learning_rate": 3.793169504545871e-07, "loss": 0.1049, "step": 61070 }, { "epoch": 1.1976470588235295, "grad_norm": 2.630584716796875, "learning_rate": 3.791596446873401e-07, "loss": 0.1128, "step": 61080 }, { "epoch": 1.1978431372549019, "grad_norm": 1.8305330276489258, "learning_rate": 3.7900235162186115e-07, "loss": 0.1173, "step": 61090 }, { "epoch": 1.1980392156862745, "grad_norm": 1.8870998620986938, "learning_rate": 3.7884507127468373e-07, "loss": 0.1166, "step": 61100 }, { "epoch": 1.198235294117647, "grad_norm": 2.5966129302978516, "learning_rate": 3.7868780366234004e-07, "loss": 0.1494, "step": 61110 }, { "epoch": 1.1984313725490197, "grad_norm": 1.3465678691864014, "learning_rate": 3.7853054880136047e-07, "loss": 0.1204, "step": 61120 }, { "epoch": 1.1986274509803923, "grad_norm": 0.9005642533302307, "learning_rate": 3.783733067082746e-07, "loss": 0.0769, "step": 61130 }, { "epoch": 1.1988235294117646, "grad_norm": 1.6347311735153198, "learning_rate": 3.7821607739961047e-07, "loss": 0.1061, "step": 61140 }, { "epoch": 1.1990196078431372, "grad_norm": 2.4074835777282715, "learning_rate": 3.7805886089189465e-07, "loss": 0.1313, "step": 61150 }, { "epoch": 1.1992156862745098, "grad_norm": 1.0773394107818604, "learning_rate": 3.779016572016525e-07, "loss": 0.0986, "step": 61160 }, { "epoch": 1.1994117647058824, "grad_norm": 2.0088977813720703, "learning_rate": 3.777444663454081e-07, "loss": 0.1335, "step": 61170 }, { "epoch": 1.199607843137255, "grad_norm": 1.9937776327133179, "learning_rate": 3.775872883396841e-07, "loss": 0.0991, "step": 61180 }, { "epoch": 1.1998039215686274, "grad_norm": 1.4456580877304077, "learning_rate": 3.7743012320100154e-07, "loss": 0.1126, "step": 61190 }, { "epoch": 1.2, "grad_norm": 2.427072763442993, "learning_rate": 3.7727297094588076e-07, "loss": 0.1228, "step": 61200 }, { "epoch": 1.2001960784313725, "grad_norm": 2.5643515586853027, "learning_rate": 3.7711583159084e-07, "loss": 0.1413, "step": 61210 }, { "epoch": 1.2003921568627451, "grad_norm": 1.73148775100708, "learning_rate": 3.769587051523966e-07, "loss": 0.1299, "step": 61220 }, { "epoch": 1.2005882352941177, "grad_norm": 2.352727174758911, "learning_rate": 3.7680159164706656e-07, "loss": 0.127, "step": 61230 }, { "epoch": 1.20078431372549, "grad_norm": 2.3505797386169434, "learning_rate": 3.7664449109136394e-07, "loss": 0.1463, "step": 61240 }, { "epoch": 1.2009803921568627, "grad_norm": 1.806278109550476, "learning_rate": 3.7648740350180255e-07, "loss": 0.1333, "step": 61250 }, { "epoch": 1.2011764705882353, "grad_norm": 1.444501280784607, "learning_rate": 3.763303288948937e-07, "loss": 0.1533, "step": 61260 }, { "epoch": 1.2013725490196079, "grad_norm": 2.7747366428375244, "learning_rate": 3.7617326728714814e-07, "loss": 0.1396, "step": 61270 }, { "epoch": 1.2015686274509805, "grad_norm": 2.4308955669403076, "learning_rate": 3.760162186950746e-07, "loss": 0.1077, "step": 61280 }, { "epoch": 1.2017647058823528, "grad_norm": 2.7178237438201904, "learning_rate": 3.7585918313518093e-07, "loss": 0.1186, "step": 61290 }, { "epoch": 1.2019607843137254, "grad_norm": 3.169229507446289, "learning_rate": 3.7570216062397355e-07, "loss": 0.1446, "step": 61300 }, { "epoch": 1.202156862745098, "grad_norm": 1.2202377319335938, "learning_rate": 3.7554515117795716e-07, "loss": 0.1299, "step": 61310 }, { "epoch": 1.2023529411764706, "grad_norm": 3.3603875637054443, "learning_rate": 3.753881548136355e-07, "loss": 0.1241, "step": 61320 }, { "epoch": 1.2025490196078432, "grad_norm": 1.9748222827911377, "learning_rate": 3.7523117154751084e-07, "loss": 0.1264, "step": 61330 }, { "epoch": 1.2027450980392156, "grad_norm": 2.7226266860961914, "learning_rate": 3.750742013960837e-07, "loss": 0.1209, "step": 61340 }, { "epoch": 1.2029411764705882, "grad_norm": 2.1994614601135254, "learning_rate": 3.749172443758538e-07, "loss": 0.1378, "step": 61350 }, { "epoch": 1.2031372549019608, "grad_norm": 2.27363657951355, "learning_rate": 3.747603005033192e-07, "loss": 0.1709, "step": 61360 }, { "epoch": 1.2033333333333334, "grad_norm": 2.7034075260162354, "learning_rate": 3.746033697949763e-07, "loss": 0.1178, "step": 61370 }, { "epoch": 1.203529411764706, "grad_norm": 2.393876075744629, "learning_rate": 3.744464522673207e-07, "loss": 0.1115, "step": 61380 }, { "epoch": 1.2037254901960783, "grad_norm": 1.2989838123321533, "learning_rate": 3.742895479368463e-07, "loss": 0.1393, "step": 61390 }, { "epoch": 1.203921568627451, "grad_norm": 3.7706668376922607, "learning_rate": 3.7413265682004534e-07, "loss": 0.1225, "step": 61400 }, { "epoch": 1.2041176470588235, "grad_norm": 1.8653398752212524, "learning_rate": 3.739757789334092e-07, "loss": 0.1304, "step": 61410 }, { "epoch": 1.204313725490196, "grad_norm": 3.0202090740203857, "learning_rate": 3.738189142934275e-07, "loss": 0.1306, "step": 61420 }, { "epoch": 1.2045098039215687, "grad_norm": 1.5986440181732178, "learning_rate": 3.7366206291658874e-07, "loss": 0.1039, "step": 61430 }, { "epoch": 1.204705882352941, "grad_norm": 3.732236862182617, "learning_rate": 3.735052248193798e-07, "loss": 0.1299, "step": 61440 }, { "epoch": 1.2049019607843137, "grad_norm": 2.734133243560791, "learning_rate": 3.733484000182862e-07, "loss": 0.1371, "step": 61450 }, { "epoch": 1.2050980392156863, "grad_norm": 2.2017955780029297, "learning_rate": 3.731915885297923e-07, "loss": 0.1335, "step": 61460 }, { "epoch": 1.2052941176470588, "grad_norm": 1.9385570287704468, "learning_rate": 3.7303479037038047e-07, "loss": 0.1375, "step": 61470 }, { "epoch": 1.2054901960784314, "grad_norm": 1.8152456283569336, "learning_rate": 3.7287800555653236e-07, "loss": 0.1242, "step": 61480 }, { "epoch": 1.2056862745098038, "grad_norm": 2.8002853393554688, "learning_rate": 3.7272123410472805e-07, "loss": 0.1266, "step": 61490 }, { "epoch": 1.2058823529411764, "grad_norm": 1.5047403573989868, "learning_rate": 3.725644760314458e-07, "loss": 0.0929, "step": 61500 }, { "epoch": 1.206078431372549, "grad_norm": 1.4549719095230103, "learning_rate": 3.7240773135316283e-07, "loss": 0.1187, "step": 61510 }, { "epoch": 1.2062745098039216, "grad_norm": 2.7746756076812744, "learning_rate": 3.722510000863551e-07, "loss": 0.1095, "step": 61520 }, { "epoch": 1.2064705882352942, "grad_norm": 3.3106093406677246, "learning_rate": 3.7209428224749664e-07, "loss": 0.1523, "step": 61530 }, { "epoch": 1.2066666666666666, "grad_norm": 3.225201368331909, "learning_rate": 3.7193757785306057e-07, "loss": 0.1199, "step": 61540 }, { "epoch": 1.2068627450980391, "grad_norm": 3.4992294311523438, "learning_rate": 3.7178088691951837e-07, "loss": 0.1214, "step": 61550 }, { "epoch": 1.2070588235294117, "grad_norm": 1.8169291019439697, "learning_rate": 3.7162420946334006e-07, "loss": 0.1177, "step": 61560 }, { "epoch": 1.2072549019607843, "grad_norm": 1.251281976699829, "learning_rate": 3.714675455009943e-07, "loss": 0.1184, "step": 61570 }, { "epoch": 1.207450980392157, "grad_norm": 2.1804051399230957, "learning_rate": 3.713108950489485e-07, "loss": 0.125, "step": 61580 }, { "epoch": 1.2076470588235293, "grad_norm": 1.0087846517562866, "learning_rate": 3.711542581236684e-07, "loss": 0.1246, "step": 61590 }, { "epoch": 1.2078431372549019, "grad_norm": 2.311981678009033, "learning_rate": 3.709976347416183e-07, "loss": 0.1189, "step": 61600 }, { "epoch": 1.2080392156862745, "grad_norm": 3.2084193229675293, "learning_rate": 3.708410249192613e-07, "loss": 0.119, "step": 61610 }, { "epoch": 1.208235294117647, "grad_norm": 1.239356279373169, "learning_rate": 3.706844286730592e-07, "loss": 0.1079, "step": 61620 }, { "epoch": 1.2084313725490197, "grad_norm": 1.8336689472198486, "learning_rate": 3.7052784601947185e-07, "loss": 0.1296, "step": 61630 }, { "epoch": 1.208627450980392, "grad_norm": 2.371990203857422, "learning_rate": 3.70371276974958e-07, "loss": 0.1246, "step": 61640 }, { "epoch": 1.2088235294117646, "grad_norm": 3.145918369293213, "learning_rate": 3.70214721555975e-07, "loss": 0.1144, "step": 61650 }, { "epoch": 1.2090196078431372, "grad_norm": 2.5425965785980225, "learning_rate": 3.700581797789786e-07, "loss": 0.1335, "step": 61660 }, { "epoch": 1.2092156862745098, "grad_norm": 1.9557334184646606, "learning_rate": 3.6990165166042333e-07, "loss": 0.1103, "step": 61670 }, { "epoch": 1.2094117647058824, "grad_norm": 4.200771331787109, "learning_rate": 3.6974513721676216e-07, "loss": 0.123, "step": 61680 }, { "epoch": 1.2096078431372548, "grad_norm": 2.8862311840057373, "learning_rate": 3.6958863646444647e-07, "loss": 0.1271, "step": 61690 }, { "epoch": 1.2098039215686274, "grad_norm": 3.6763198375701904, "learning_rate": 3.694321494199265e-07, "loss": 0.1237, "step": 61700 }, { "epoch": 1.21, "grad_norm": 2.2085959911346436, "learning_rate": 3.6927567609965106e-07, "loss": 0.0935, "step": 61710 }, { "epoch": 1.2101960784313726, "grad_norm": 2.944626569747925, "learning_rate": 3.6911921652006703e-07, "loss": 0.1469, "step": 61720 }, { "epoch": 1.2103921568627451, "grad_norm": 2.156829595565796, "learning_rate": 3.6896277069762044e-07, "loss": 0.13, "step": 61730 }, { "epoch": 1.2105882352941177, "grad_norm": 1.5298595428466797, "learning_rate": 3.6880633864875565e-07, "loss": 0.1273, "step": 61740 }, { "epoch": 1.2107843137254901, "grad_norm": 3.847585678100586, "learning_rate": 3.686499203899153e-07, "loss": 0.1288, "step": 61750 }, { "epoch": 1.2109803921568627, "grad_norm": 3.9754855632781982, "learning_rate": 3.6849351593754083e-07, "loss": 0.1035, "step": 61760 }, { "epoch": 1.2111764705882353, "grad_norm": 4.452577114105225, "learning_rate": 3.683371253080726e-07, "loss": 0.1099, "step": 61770 }, { "epoch": 1.211372549019608, "grad_norm": 2.482698678970337, "learning_rate": 3.681807485179486e-07, "loss": 0.1182, "step": 61780 }, { "epoch": 1.2115686274509805, "grad_norm": 1.8745335340499878, "learning_rate": 3.6802438558360634e-07, "loss": 0.1474, "step": 61790 }, { "epoch": 1.2117647058823529, "grad_norm": 1.1753153800964355, "learning_rate": 3.6786803652148123e-07, "loss": 0.1336, "step": 61800 }, { "epoch": 1.2119607843137254, "grad_norm": 3.543182849884033, "learning_rate": 3.677117013480076e-07, "loss": 0.1301, "step": 61810 }, { "epoch": 1.212156862745098, "grad_norm": 3.2706525325775146, "learning_rate": 3.6755538007961793e-07, "loss": 0.1114, "step": 61820 }, { "epoch": 1.2123529411764706, "grad_norm": 2.07944393157959, "learning_rate": 3.673990727327435e-07, "loss": 0.1411, "step": 61830 }, { "epoch": 1.2125490196078432, "grad_norm": 3.294851541519165, "learning_rate": 3.672427793238143e-07, "loss": 0.121, "step": 61840 }, { "epoch": 1.2127450980392156, "grad_norm": 1.1775050163269043, "learning_rate": 3.6708649986925845e-07, "loss": 0.1285, "step": 61850 }, { "epoch": 1.2129411764705882, "grad_norm": 3.175058603286743, "learning_rate": 3.669302343855027e-07, "loss": 0.1337, "step": 61860 }, { "epoch": 1.2131372549019608, "grad_norm": 2.6142349243164062, "learning_rate": 3.6677398288897266e-07, "loss": 0.1358, "step": 61870 }, { "epoch": 1.2133333333333334, "grad_norm": 3.272909164428711, "learning_rate": 3.6661774539609204e-07, "loss": 0.1376, "step": 61880 }, { "epoch": 1.213529411764706, "grad_norm": 2.5450778007507324, "learning_rate": 3.6646152192328336e-07, "loss": 0.0998, "step": 61890 }, { "epoch": 1.2137254901960783, "grad_norm": 1.4928735494613647, "learning_rate": 3.663053124869676e-07, "loss": 0.1544, "step": 61900 }, { "epoch": 1.213921568627451, "grad_norm": 2.1442179679870605, "learning_rate": 3.661491171035642e-07, "loss": 0.1436, "step": 61910 }, { "epoch": 1.2141176470588235, "grad_norm": 1.5750882625579834, "learning_rate": 3.659929357894911e-07, "loss": 0.1124, "step": 61920 }, { "epoch": 1.2143137254901961, "grad_norm": 0.8070198893547058, "learning_rate": 3.6583676856116497e-07, "loss": 0.1064, "step": 61930 }, { "epoch": 1.2145098039215687, "grad_norm": 1.9629875421524048, "learning_rate": 3.6568061543500074e-07, "loss": 0.1248, "step": 61940 }, { "epoch": 1.214705882352941, "grad_norm": 1.4228755235671997, "learning_rate": 3.65524476427412e-07, "loss": 0.1151, "step": 61950 }, { "epoch": 1.2149019607843137, "grad_norm": 1.283504605293274, "learning_rate": 3.653683515548107e-07, "loss": 0.1184, "step": 61960 }, { "epoch": 1.2150980392156863, "grad_norm": 1.9209688901901245, "learning_rate": 3.652122408336078e-07, "loss": 0.1099, "step": 61970 }, { "epoch": 1.2152941176470589, "grad_norm": 1.6614521741867065, "learning_rate": 3.6505614428021216e-07, "loss": 0.1185, "step": 61980 }, { "epoch": 1.2154901960784315, "grad_norm": 3.1566579341888428, "learning_rate": 3.649000619110313e-07, "loss": 0.1336, "step": 61990 }, { "epoch": 1.215686274509804, "grad_norm": 2.093726634979248, "learning_rate": 3.6474399374247165e-07, "loss": 0.1319, "step": 62000 }, { "epoch": 1.2158823529411764, "grad_norm": 3.4369096755981445, "learning_rate": 3.6458793979093755e-07, "loss": 0.1186, "step": 62010 }, { "epoch": 1.216078431372549, "grad_norm": 2.074662208557129, "learning_rate": 3.644319000728322e-07, "loss": 0.1123, "step": 62020 }, { "epoch": 1.2162745098039216, "grad_norm": 3.7817771434783936, "learning_rate": 3.642758746045574e-07, "loss": 0.1401, "step": 62030 }, { "epoch": 1.2164705882352942, "grad_norm": 2.9505927562713623, "learning_rate": 3.6411986340251317e-07, "loss": 0.1211, "step": 62040 }, { "epoch": 1.2166666666666668, "grad_norm": 2.089040517807007, "learning_rate": 3.639638664830981e-07, "loss": 0.1481, "step": 62050 }, { "epoch": 1.2168627450980392, "grad_norm": 1.396497130393982, "learning_rate": 3.638078838627095e-07, "loss": 0.1567, "step": 62060 }, { "epoch": 1.2170588235294117, "grad_norm": 2.3266139030456543, "learning_rate": 3.6365191555774284e-07, "loss": 0.1249, "step": 62070 }, { "epoch": 1.2172549019607843, "grad_norm": 2.5250298976898193, "learning_rate": 3.634959615845923e-07, "loss": 0.1122, "step": 62080 }, { "epoch": 1.217450980392157, "grad_norm": 1.5391076803207397, "learning_rate": 3.6334002195965063e-07, "loss": 0.0896, "step": 62090 }, { "epoch": 1.2176470588235295, "grad_norm": 1.0950390100479126, "learning_rate": 3.6318409669930885e-07, "loss": 0.1025, "step": 62100 }, { "epoch": 1.217843137254902, "grad_norm": 1.707134485244751, "learning_rate": 3.630281858199565e-07, "loss": 0.1355, "step": 62110 }, { "epoch": 1.2180392156862745, "grad_norm": 1.5626378059387207, "learning_rate": 3.628722893379819e-07, "loss": 0.0962, "step": 62120 }, { "epoch": 1.218235294117647, "grad_norm": 1.5432872772216797, "learning_rate": 3.6271640726977136e-07, "loss": 0.129, "step": 62130 }, { "epoch": 1.2184313725490197, "grad_norm": 2.379577159881592, "learning_rate": 3.6256053963171e-07, "loss": 0.1229, "step": 62140 }, { "epoch": 1.2186274509803923, "grad_norm": 3.072727680206299, "learning_rate": 3.624046864401816e-07, "loss": 0.1358, "step": 62150 }, { "epoch": 1.2188235294117646, "grad_norm": 2.0132479667663574, "learning_rate": 3.6224884771156814e-07, "loss": 0.12, "step": 62160 }, { "epoch": 1.2190196078431372, "grad_norm": 1.3626538515090942, "learning_rate": 3.6209302346225e-07, "loss": 0.1131, "step": 62170 }, { "epoch": 1.2192156862745098, "grad_norm": 2.9126975536346436, "learning_rate": 3.619372137086062e-07, "loss": 0.1268, "step": 62180 }, { "epoch": 1.2194117647058824, "grad_norm": 1.6669929027557373, "learning_rate": 3.617814184670144e-07, "loss": 0.1125, "step": 62190 }, { "epoch": 1.219607843137255, "grad_norm": 1.7605476379394531, "learning_rate": 3.616256377538502e-07, "loss": 0.1152, "step": 62200 }, { "epoch": 1.2198039215686274, "grad_norm": 2.923297882080078, "learning_rate": 3.6146987158548826e-07, "loss": 0.1324, "step": 62210 }, { "epoch": 1.22, "grad_norm": 2.6028995513916016, "learning_rate": 3.613141199783015e-07, "loss": 0.1239, "step": 62220 }, { "epoch": 1.2201960784313726, "grad_norm": 1.7061042785644531, "learning_rate": 3.6115838294866107e-07, "loss": 0.1493, "step": 62230 }, { "epoch": 1.2203921568627452, "grad_norm": 1.2697975635528564, "learning_rate": 3.61002660512937e-07, "loss": 0.1356, "step": 62240 }, { "epoch": 1.2205882352941178, "grad_norm": 3.5540335178375244, "learning_rate": 3.6084695268749753e-07, "loss": 0.1253, "step": 62250 }, { "epoch": 1.2207843137254901, "grad_norm": 1.9177312850952148, "learning_rate": 3.6069125948870937e-07, "loss": 0.127, "step": 62260 }, { "epoch": 1.2209803921568627, "grad_norm": 3.394228458404541, "learning_rate": 3.605355809329377e-07, "loss": 0.1216, "step": 62270 }, { "epoch": 1.2211764705882353, "grad_norm": 2.069456100463867, "learning_rate": 3.6037991703654637e-07, "loss": 0.1301, "step": 62280 }, { "epoch": 1.221372549019608, "grad_norm": 3.025212287902832, "learning_rate": 3.6022426781589734e-07, "loss": 0.1398, "step": 62290 }, { "epoch": 1.2215686274509805, "grad_norm": 1.704819917678833, "learning_rate": 3.6006863328735125e-07, "loss": 0.1036, "step": 62300 }, { "epoch": 1.2217647058823529, "grad_norm": 3.017970085144043, "learning_rate": 3.5991301346726734e-07, "loss": 0.1236, "step": 62310 }, { "epoch": 1.2219607843137255, "grad_norm": 2.700592041015625, "learning_rate": 3.5975740837200264e-07, "loss": 0.1122, "step": 62320 }, { "epoch": 1.222156862745098, "grad_norm": 3.295488119125366, "learning_rate": 3.5960181801791363e-07, "loss": 0.1485, "step": 62330 }, { "epoch": 1.2223529411764706, "grad_norm": 3.1969780921936035, "learning_rate": 3.5944624242135445e-07, "loss": 0.1482, "step": 62340 }, { "epoch": 1.2225490196078432, "grad_norm": 2.135328531265259, "learning_rate": 3.5929068159867814e-07, "loss": 0.1197, "step": 62350 }, { "epoch": 1.2227450980392156, "grad_norm": 2.8795554637908936, "learning_rate": 3.5913513556623576e-07, "loss": 0.1226, "step": 62360 }, { "epoch": 1.2229411764705882, "grad_norm": 3.480844020843506, "learning_rate": 3.589796043403772e-07, "loss": 0.1013, "step": 62370 }, { "epoch": 1.2231372549019608, "grad_norm": 4.490631580352783, "learning_rate": 3.588240879374508e-07, "loss": 0.0917, "step": 62380 }, { "epoch": 1.2233333333333334, "grad_norm": 2.9659197330474854, "learning_rate": 3.586685863738029e-07, "loss": 0.1202, "step": 62390 }, { "epoch": 1.223529411764706, "grad_norm": 3.7404041290283203, "learning_rate": 3.585130996657787e-07, "loss": 0.1213, "step": 62400 }, { "epoch": 1.2237254901960783, "grad_norm": 1.9540836811065674, "learning_rate": 3.583576278297219e-07, "loss": 0.1277, "step": 62410 }, { "epoch": 1.223921568627451, "grad_norm": 2.0736727714538574, "learning_rate": 3.5820217088197406e-07, "loss": 0.0942, "step": 62420 }, { "epoch": 1.2241176470588235, "grad_norm": 3.4770162105560303, "learning_rate": 3.5804672883887586e-07, "loss": 0.1216, "step": 62430 }, { "epoch": 1.2243137254901961, "grad_norm": 4.434560775756836, "learning_rate": 3.578913017167661e-07, "loss": 0.1424, "step": 62440 }, { "epoch": 1.2245098039215687, "grad_norm": 3.444671869277954, "learning_rate": 3.57735889531982e-07, "loss": 0.1149, "step": 62450 }, { "epoch": 1.224705882352941, "grad_norm": 2.866621255874634, "learning_rate": 3.5758049230085905e-07, "loss": 0.1705, "step": 62460 }, { "epoch": 1.2249019607843137, "grad_norm": 2.347701072692871, "learning_rate": 3.574251100397317e-07, "loss": 0.1351, "step": 62470 }, { "epoch": 1.2250980392156863, "grad_norm": 2.5603744983673096, "learning_rate": 3.572697427649322e-07, "loss": 0.1478, "step": 62480 }, { "epoch": 1.2252941176470589, "grad_norm": 1.7872065305709839, "learning_rate": 3.571143904927917e-07, "loss": 0.1276, "step": 62490 }, { "epoch": 1.2254901960784315, "grad_norm": 3.3424429893493652, "learning_rate": 3.5695905323963927e-07, "loss": 0.1418, "step": 62500 }, { "epoch": 1.2256862745098038, "grad_norm": 2.920271396636963, "learning_rate": 3.5680373102180326e-07, "loss": 0.1585, "step": 62510 }, { "epoch": 1.2258823529411764, "grad_norm": 2.2494750022888184, "learning_rate": 3.566484238556094e-07, "loss": 0.1095, "step": 62520 }, { "epoch": 1.226078431372549, "grad_norm": 5.0963592529296875, "learning_rate": 3.564931317573826e-07, "loss": 0.125, "step": 62530 }, { "epoch": 1.2262745098039216, "grad_norm": 2.637601613998413, "learning_rate": 3.563378547434459e-07, "loss": 0.1104, "step": 62540 }, { "epoch": 1.2264705882352942, "grad_norm": 1.899787187576294, "learning_rate": 3.5618259283012065e-07, "loss": 0.1456, "step": 62550 }, { "epoch": 1.2266666666666666, "grad_norm": 3.0227935314178467, "learning_rate": 3.560273460337268e-07, "loss": 0.1287, "step": 62560 }, { "epoch": 1.2268627450980392, "grad_norm": 1.238297939300537, "learning_rate": 3.5587211437058275e-07, "loss": 0.1101, "step": 62570 }, { "epoch": 1.2270588235294118, "grad_norm": 1.948943853378296, "learning_rate": 3.55716897857005e-07, "loss": 0.1087, "step": 62580 }, { "epoch": 1.2272549019607844, "grad_norm": 2.5733683109283447, "learning_rate": 3.555616965093087e-07, "loss": 0.1386, "step": 62590 }, { "epoch": 1.227450980392157, "grad_norm": 2.789691686630249, "learning_rate": 3.5540651034380757e-07, "loss": 0.1125, "step": 62600 }, { "epoch": 1.2276470588235293, "grad_norm": 1.253546953201294, "learning_rate": 3.5525133937681335e-07, "loss": 0.1036, "step": 62610 }, { "epoch": 1.227843137254902, "grad_norm": 1.648883581161499, "learning_rate": 3.550961836246363e-07, "loss": 0.134, "step": 62620 }, { "epoch": 1.2280392156862745, "grad_norm": 1.3680837154388428, "learning_rate": 3.549410431035855e-07, "loss": 0.1318, "step": 62630 }, { "epoch": 1.228235294117647, "grad_norm": 3.4574649333953857, "learning_rate": 3.547859178299677e-07, "loss": 0.1362, "step": 62640 }, { "epoch": 1.2284313725490197, "grad_norm": 1.3383252620697021, "learning_rate": 3.546308078200885e-07, "loss": 0.1226, "step": 62650 }, { "epoch": 1.228627450980392, "grad_norm": 1.9573332071304321, "learning_rate": 3.5447571309025204e-07, "loss": 0.0998, "step": 62660 }, { "epoch": 1.2288235294117646, "grad_norm": 2.4868249893188477, "learning_rate": 3.5432063365676026e-07, "loss": 0.138, "step": 62670 }, { "epoch": 1.2290196078431372, "grad_norm": 1.504896640777588, "learning_rate": 3.541655695359142e-07, "loss": 0.1077, "step": 62680 }, { "epoch": 1.2292156862745098, "grad_norm": 1.03737473487854, "learning_rate": 3.5401052074401284e-07, "loss": 0.1299, "step": 62690 }, { "epoch": 1.2294117647058824, "grad_norm": 2.0397331714630127, "learning_rate": 3.5385548729735373e-07, "loss": 0.1296, "step": 62700 }, { "epoch": 1.2296078431372548, "grad_norm": 1.8200743198394775, "learning_rate": 3.537004692122326e-07, "loss": 0.1165, "step": 62710 }, { "epoch": 1.2298039215686274, "grad_norm": 1.5689375400543213, "learning_rate": 3.5354546650494377e-07, "loss": 0.1232, "step": 62720 }, { "epoch": 1.23, "grad_norm": 4.617224216461182, "learning_rate": 3.5339047919177997e-07, "loss": 0.1271, "step": 62730 }, { "epoch": 1.2301960784313726, "grad_norm": 0.9333202838897705, "learning_rate": 3.5323550728903206e-07, "loss": 0.121, "step": 62740 }, { "epoch": 1.2303921568627452, "grad_norm": 3.689246416091919, "learning_rate": 3.530805508129895e-07, "loss": 0.1225, "step": 62750 }, { "epoch": 1.2305882352941175, "grad_norm": 2.7460103034973145, "learning_rate": 3.5292560977994024e-07, "loss": 0.152, "step": 62760 }, { "epoch": 1.2307843137254901, "grad_norm": 1.7379645109176636, "learning_rate": 3.527706842061702e-07, "loss": 0.1245, "step": 62770 }, { "epoch": 1.2309803921568627, "grad_norm": 0.8046085834503174, "learning_rate": 3.5261577410796394e-07, "loss": 0.1037, "step": 62780 }, { "epoch": 1.2311764705882353, "grad_norm": 2.226686716079712, "learning_rate": 3.5246087950160453e-07, "loss": 0.1325, "step": 62790 }, { "epoch": 1.231372549019608, "grad_norm": 4.919273376464844, "learning_rate": 3.523060004033731e-07, "loss": 0.1366, "step": 62800 }, { "epoch": 1.2315686274509803, "grad_norm": 1.5524264574050903, "learning_rate": 3.521511368295493e-07, "loss": 0.1329, "step": 62810 }, { "epoch": 1.2317647058823529, "grad_norm": 3.056757688522339, "learning_rate": 3.5199628879641136e-07, "loss": 0.1239, "step": 62820 }, { "epoch": 1.2319607843137255, "grad_norm": 2.981724500656128, "learning_rate": 3.518414563202353e-07, "loss": 0.1099, "step": 62830 }, { "epoch": 1.232156862745098, "grad_norm": 1.9795509576797485, "learning_rate": 3.5168663941729616e-07, "loss": 0.1252, "step": 62840 }, { "epoch": 1.2323529411764707, "grad_norm": 1.0995630025863647, "learning_rate": 3.5153183810386675e-07, "loss": 0.145, "step": 62850 }, { "epoch": 1.232549019607843, "grad_norm": 1.8678537607192993, "learning_rate": 3.5137705239621905e-07, "loss": 0.1475, "step": 62860 }, { "epoch": 1.2327450980392156, "grad_norm": 2.272477626800537, "learning_rate": 3.5122228231062245e-07, "loss": 0.1382, "step": 62870 }, { "epoch": 1.2329411764705882, "grad_norm": 4.809328079223633, "learning_rate": 3.510675278633453e-07, "loss": 0.1076, "step": 62880 }, { "epoch": 1.2331372549019608, "grad_norm": 2.03460693359375, "learning_rate": 3.5091278907065416e-07, "loss": 0.1326, "step": 62890 }, { "epoch": 1.2333333333333334, "grad_norm": 1.0988048315048218, "learning_rate": 3.507580659488138e-07, "loss": 0.111, "step": 62900 }, { "epoch": 1.2335294117647058, "grad_norm": 1.6846176385879517, "learning_rate": 3.5060335851408757e-07, "loss": 0.1187, "step": 62910 }, { "epoch": 1.2337254901960784, "grad_norm": 1.6351392269134521, "learning_rate": 3.504486667827372e-07, "loss": 0.1112, "step": 62920 }, { "epoch": 1.233921568627451, "grad_norm": 2.1100594997406006, "learning_rate": 3.5029399077102234e-07, "loss": 0.1234, "step": 62930 }, { "epoch": 1.2341176470588235, "grad_norm": 3.2904818058013916, "learning_rate": 3.501393304952015e-07, "loss": 0.1273, "step": 62940 }, { "epoch": 1.2343137254901961, "grad_norm": 4.238260746002197, "learning_rate": 3.499846859715314e-07, "loss": 0.1218, "step": 62950 }, { "epoch": 1.2345098039215685, "grad_norm": 1.9819778203964233, "learning_rate": 3.498300572162668e-07, "loss": 0.1193, "step": 62960 }, { "epoch": 1.234705882352941, "grad_norm": 3.654287815093994, "learning_rate": 3.496754442456611e-07, "loss": 0.1378, "step": 62970 }, { "epoch": 1.2349019607843137, "grad_norm": 2.207359790802002, "learning_rate": 3.495208470759662e-07, "loss": 0.1005, "step": 62980 }, { "epoch": 1.2350980392156863, "grad_norm": 1.1750481128692627, "learning_rate": 3.493662657234318e-07, "loss": 0.1121, "step": 62990 }, { "epoch": 1.2352941176470589, "grad_norm": 2.134307384490967, "learning_rate": 3.492117002043064e-07, "loss": 0.1163, "step": 63000 }, { "epoch": 1.2354901960784315, "grad_norm": 2.457047939300537, "learning_rate": 3.4905715053483673e-07, "loss": 0.1378, "step": 63010 }, { "epoch": 1.2356862745098038, "grad_norm": 1.9603041410446167, "learning_rate": 3.4890261673126776e-07, "loss": 0.1428, "step": 63020 }, { "epoch": 1.2358823529411764, "grad_norm": 2.0524771213531494, "learning_rate": 3.487480988098427e-07, "loss": 0.1156, "step": 63030 }, { "epoch": 1.236078431372549, "grad_norm": 3.5897538661956787, "learning_rate": 3.485935967868034e-07, "loss": 0.1355, "step": 63040 }, { "epoch": 1.2362745098039216, "grad_norm": 4.26077938079834, "learning_rate": 3.4843911067839e-07, "loss": 0.1256, "step": 63050 }, { "epoch": 1.2364705882352942, "grad_norm": 2.7314159870147705, "learning_rate": 3.482846405008406e-07, "loss": 0.1189, "step": 63060 }, { "epoch": 1.2366666666666666, "grad_norm": 2.7401692867279053, "learning_rate": 3.4813018627039194e-07, "loss": 0.1161, "step": 63070 }, { "epoch": 1.2368627450980392, "grad_norm": 2.658148765563965, "learning_rate": 3.479757480032791e-07, "loss": 0.1436, "step": 63080 }, { "epoch": 1.2370588235294118, "grad_norm": 1.8573330640792847, "learning_rate": 3.478213257157352e-07, "loss": 0.1251, "step": 63090 }, { "epoch": 1.2372549019607844, "grad_norm": 4.313286304473877, "learning_rate": 3.4766691942399196e-07, "loss": 0.121, "step": 63100 }, { "epoch": 1.237450980392157, "grad_norm": 2.108095169067383, "learning_rate": 3.4751252914427955e-07, "loss": 0.1211, "step": 63110 }, { "epoch": 1.2376470588235293, "grad_norm": 2.316350221633911, "learning_rate": 3.473581548928258e-07, "loss": 0.132, "step": 63120 }, { "epoch": 1.237843137254902, "grad_norm": 2.2482171058654785, "learning_rate": 3.4720379668585766e-07, "loss": 0.0976, "step": 63130 }, { "epoch": 1.2380392156862745, "grad_norm": 2.4428369998931885, "learning_rate": 3.4704945453959996e-07, "loss": 0.1119, "step": 63140 }, { "epoch": 1.238235294117647, "grad_norm": 2.018221139907837, "learning_rate": 3.468951284702757e-07, "loss": 0.1058, "step": 63150 }, { "epoch": 1.2384313725490197, "grad_norm": 0.8388502597808838, "learning_rate": 3.467408184941065e-07, "loss": 0.1356, "step": 63160 }, { "epoch": 1.238627450980392, "grad_norm": 2.4202046394348145, "learning_rate": 3.465865246273124e-07, "loss": 0.1202, "step": 63170 }, { "epoch": 1.2388235294117647, "grad_norm": 1.7992922067642212, "learning_rate": 3.464322468861113e-07, "loss": 0.12, "step": 63180 }, { "epoch": 1.2390196078431372, "grad_norm": 6.749047756195068, "learning_rate": 3.4627798528671964e-07, "loss": 0.127, "step": 63190 }, { "epoch": 1.2392156862745098, "grad_norm": 2.59001088142395, "learning_rate": 3.4612373984535237e-07, "loss": 0.1301, "step": 63200 }, { "epoch": 1.2394117647058824, "grad_norm": 1.1790019273757935, "learning_rate": 3.459695105782221e-07, "loss": 0.1248, "step": 63210 }, { "epoch": 1.2396078431372548, "grad_norm": 2.873767375946045, "learning_rate": 3.458152975015406e-07, "loss": 0.1234, "step": 63220 }, { "epoch": 1.2398039215686274, "grad_norm": 3.062335968017578, "learning_rate": 3.4566110063151733e-07, "loss": 0.1036, "step": 63230 }, { "epoch": 1.24, "grad_norm": 1.28623366355896, "learning_rate": 3.455069199843604e-07, "loss": 0.1128, "step": 63240 }, { "epoch": 1.2401960784313726, "grad_norm": 4.114004135131836, "learning_rate": 3.453527555762758e-07, "loss": 0.1355, "step": 63250 }, { "epoch": 1.2403921568627452, "grad_norm": 2.6290552616119385, "learning_rate": 3.4519860742346816e-07, "loss": 0.1486, "step": 63260 }, { "epoch": 1.2405882352941178, "grad_norm": 0.8570523858070374, "learning_rate": 3.450444755421403e-07, "loss": 0.1191, "step": 63270 }, { "epoch": 1.2407843137254901, "grad_norm": 4.728590965270996, "learning_rate": 3.4489035994849335e-07, "loss": 0.1328, "step": 63280 }, { "epoch": 1.2409803921568627, "grad_norm": 2.0189971923828125, "learning_rate": 3.447362606587266e-07, "loss": 0.1259, "step": 63290 }, { "epoch": 1.2411764705882353, "grad_norm": 3.5816900730133057, "learning_rate": 3.4458217768903794e-07, "loss": 0.1281, "step": 63300 }, { "epoch": 1.241372549019608, "grad_norm": 6.5985260009765625, "learning_rate": 3.444281110556231e-07, "loss": 0.1328, "step": 63310 }, { "epoch": 1.2415686274509805, "grad_norm": 2.1328749656677246, "learning_rate": 3.442740607746765e-07, "loss": 0.1155, "step": 63320 }, { "epoch": 1.2417647058823529, "grad_norm": 4.1898016929626465, "learning_rate": 3.4412002686239064e-07, "loss": 0.1164, "step": 63330 }, { "epoch": 1.2419607843137255, "grad_norm": 1.50349760055542, "learning_rate": 3.439660093349561e-07, "loss": 0.1153, "step": 63340 }, { "epoch": 1.242156862745098, "grad_norm": 1.1345394849777222, "learning_rate": 3.438120082085622e-07, "loss": 0.1234, "step": 63350 }, { "epoch": 1.2423529411764707, "grad_norm": 0.910362958908081, "learning_rate": 3.436580234993964e-07, "loss": 0.1064, "step": 63360 }, { "epoch": 1.2425490196078433, "grad_norm": 1.5770875215530396, "learning_rate": 3.435040552236441e-07, "loss": 0.1009, "step": 63370 }, { "epoch": 1.2427450980392156, "grad_norm": 2.324565887451172, "learning_rate": 3.4335010339748923e-07, "loss": 0.1243, "step": 63380 }, { "epoch": 1.2429411764705882, "grad_norm": 2.859635829925537, "learning_rate": 3.4319616803711394e-07, "loss": 0.116, "step": 63390 }, { "epoch": 1.2431372549019608, "grad_norm": 3.3094613552093506, "learning_rate": 3.4304224915869907e-07, "loss": 0.1093, "step": 63400 }, { "epoch": 1.2433333333333334, "grad_norm": 2.2824227809906006, "learning_rate": 3.428883467784228e-07, "loss": 0.1256, "step": 63410 }, { "epoch": 1.243529411764706, "grad_norm": 2.3705129623413086, "learning_rate": 3.427344609124625e-07, "loss": 0.1079, "step": 63420 }, { "epoch": 1.2437254901960784, "grad_norm": 2.270343065261841, "learning_rate": 3.425805915769933e-07, "loss": 0.1427, "step": 63430 }, { "epoch": 1.243921568627451, "grad_norm": 2.1368753910064697, "learning_rate": 3.424267387881886e-07, "loss": 0.1445, "step": 63440 }, { "epoch": 1.2441176470588236, "grad_norm": 2.4042983055114746, "learning_rate": 3.422729025622203e-07, "loss": 0.1196, "step": 63450 }, { "epoch": 1.2443137254901961, "grad_norm": 2.514547824859619, "learning_rate": 3.4211908291525843e-07, "loss": 0.1384, "step": 63460 }, { "epoch": 1.2445098039215687, "grad_norm": 1.5799102783203125, "learning_rate": 3.419652798634711e-07, "loss": 0.122, "step": 63470 }, { "epoch": 1.244705882352941, "grad_norm": 2.550236225128174, "learning_rate": 3.4181149342302507e-07, "loss": 0.1175, "step": 63480 }, { "epoch": 1.2449019607843137, "grad_norm": 1.4840339422225952, "learning_rate": 3.416577236100851e-07, "loss": 0.1222, "step": 63490 }, { "epoch": 1.2450980392156863, "grad_norm": 1.7415038347244263, "learning_rate": 3.4150397044081414e-07, "loss": 0.1098, "step": 63500 }, { "epoch": 1.2452941176470589, "grad_norm": 2.4808759689331055, "learning_rate": 3.4135023393137347e-07, "loss": 0.1309, "step": 63510 }, { "epoch": 1.2454901960784315, "grad_norm": 2.333073139190674, "learning_rate": 3.4119651409792285e-07, "loss": 0.1261, "step": 63520 }, { "epoch": 1.2456862745098038, "grad_norm": 2.11141300201416, "learning_rate": 3.410428109566198e-07, "loss": 0.1034, "step": 63530 }, { "epoch": 1.2458823529411764, "grad_norm": 1.908186674118042, "learning_rate": 3.408891245236205e-07, "loss": 0.1636, "step": 63540 }, { "epoch": 1.246078431372549, "grad_norm": 2.2221288681030273, "learning_rate": 3.407354548150793e-07, "loss": 0.1125, "step": 63550 }, { "epoch": 1.2462745098039216, "grad_norm": 3.8191676139831543, "learning_rate": 3.405818018471486e-07, "loss": 0.1493, "step": 63560 }, { "epoch": 1.2464705882352942, "grad_norm": 2.3194522857666016, "learning_rate": 3.4042816563597907e-07, "loss": 0.1198, "step": 63570 }, { "epoch": 1.2466666666666666, "grad_norm": 3.0288262367248535, "learning_rate": 3.4027454619772e-07, "loss": 0.1103, "step": 63580 }, { "epoch": 1.2468627450980392, "grad_norm": 1.4507880210876465, "learning_rate": 3.401209435485186e-07, "loss": 0.1074, "step": 63590 }, { "epoch": 1.2470588235294118, "grad_norm": 4.444198131561279, "learning_rate": 3.399673577045201e-07, "loss": 0.1148, "step": 63600 }, { "epoch": 1.2472549019607844, "grad_norm": 1.8610916137695312, "learning_rate": 3.398137886818684e-07, "loss": 0.1309, "step": 63610 }, { "epoch": 1.247450980392157, "grad_norm": 3.130845308303833, "learning_rate": 3.396602364967055e-07, "loss": 0.1511, "step": 63620 }, { "epoch": 1.2476470588235293, "grad_norm": 5.515585422515869, "learning_rate": 3.395067011651713e-07, "loss": 0.1354, "step": 63630 }, { "epoch": 1.247843137254902, "grad_norm": 2.128147840499878, "learning_rate": 3.393531827034044e-07, "loss": 0.0855, "step": 63640 }, { "epoch": 1.2480392156862745, "grad_norm": 1.551785945892334, "learning_rate": 3.391996811275415e-07, "loss": 0.1344, "step": 63650 }, { "epoch": 1.248235294117647, "grad_norm": 0.8708692193031311, "learning_rate": 3.390461964537173e-07, "loss": 0.1309, "step": 63660 }, { "epoch": 1.2484313725490197, "grad_norm": 2.8767380714416504, "learning_rate": 3.388927286980649e-07, "loss": 0.1082, "step": 63670 }, { "epoch": 1.248627450980392, "grad_norm": 2.6854915618896484, "learning_rate": 3.3873927787671573e-07, "loss": 0.1268, "step": 63680 }, { "epoch": 1.2488235294117647, "grad_norm": 3.594320297241211, "learning_rate": 3.385858440057991e-07, "loss": 0.111, "step": 63690 }, { "epoch": 1.2490196078431373, "grad_norm": 4.242233753204346, "learning_rate": 3.384324271014428e-07, "loss": 0.1447, "step": 63700 }, { "epoch": 1.2492156862745099, "grad_norm": 2.534431219100952, "learning_rate": 3.38279027179773e-07, "loss": 0.1269, "step": 63710 }, { "epoch": 1.2494117647058824, "grad_norm": 7.18101167678833, "learning_rate": 3.381256442569136e-07, "loss": 0.1608, "step": 63720 }, { "epoch": 1.2496078431372548, "grad_norm": 2.4050333499908447, "learning_rate": 3.379722783489871e-07, "loss": 0.1299, "step": 63730 }, { "epoch": 1.2498039215686274, "grad_norm": 1.7386589050292969, "learning_rate": 3.378189294721142e-07, "loss": 0.1023, "step": 63740 }, { "epoch": 1.25, "grad_norm": 2.6176810264587402, "learning_rate": 3.376655976424133e-07, "loss": 0.1203, "step": 63750 }, { "epoch": 1.2501960784313726, "grad_norm": 1.298193335533142, "learning_rate": 3.3751228287600187e-07, "loss": 0.1101, "step": 63760 }, { "epoch": 1.2503921568627452, "grad_norm": 2.0347416400909424, "learning_rate": 3.3735898518899494e-07, "loss": 0.1209, "step": 63770 }, { "epoch": 1.2505882352941176, "grad_norm": 2.526473045349121, "learning_rate": 3.3720570459750596e-07, "loss": 0.1291, "step": 63780 }, { "epoch": 1.2507843137254901, "grad_norm": 1.9353528022766113, "learning_rate": 3.370524411176465e-07, "loss": 0.132, "step": 63790 }, { "epoch": 1.2509803921568627, "grad_norm": 2.813096523284912, "learning_rate": 3.368991947655265e-07, "loss": 0.1151, "step": 63800 }, { "epoch": 1.2511764705882353, "grad_norm": 7.136311054229736, "learning_rate": 3.3674596555725395e-07, "loss": 0.1288, "step": 63810 }, { "epoch": 1.251372549019608, "grad_norm": 1.329174518585205, "learning_rate": 3.365927535089349e-07, "loss": 0.1344, "step": 63820 }, { "epoch": 1.2515686274509803, "grad_norm": 2.095106840133667, "learning_rate": 3.3643955863667395e-07, "loss": 0.1078, "step": 63830 }, { "epoch": 1.251764705882353, "grad_norm": 2.1806654930114746, "learning_rate": 3.362863809565738e-07, "loss": 0.1403, "step": 63840 }, { "epoch": 1.2519607843137255, "grad_norm": 0.9543010592460632, "learning_rate": 3.3613322048473495e-07, "loss": 0.1068, "step": 63850 }, { "epoch": 1.252156862745098, "grad_norm": 1.6206283569335938, "learning_rate": 3.359800772372566e-07, "loss": 0.1341, "step": 63860 }, { "epoch": 1.2523529411764707, "grad_norm": 2.1051113605499268, "learning_rate": 3.3582695123023606e-07, "loss": 0.1317, "step": 63870 }, { "epoch": 1.252549019607843, "grad_norm": 2.767571210861206, "learning_rate": 3.3567384247976846e-07, "loss": 0.1325, "step": 63880 }, { "epoch": 1.2527450980392156, "grad_norm": 1.4616225957870483, "learning_rate": 3.3552075100194754e-07, "loss": 0.1227, "step": 63890 }, { "epoch": 1.2529411764705882, "grad_norm": 10.870475769042969, "learning_rate": 3.353676768128649e-07, "loss": 0.1039, "step": 63900 }, { "epoch": 1.2531372549019608, "grad_norm": 3.784311532974243, "learning_rate": 3.352146199286106e-07, "loss": 0.1485, "step": 63910 }, { "epoch": 1.2533333333333334, "grad_norm": 3.5416414737701416, "learning_rate": 3.3506158036527263e-07, "loss": 0.1058, "step": 63920 }, { "epoch": 1.2535294117647058, "grad_norm": 2.227540969848633, "learning_rate": 3.3490855813893726e-07, "loss": 0.1411, "step": 63930 }, { "epoch": 1.2537254901960784, "grad_norm": 3.436495780944824, "learning_rate": 3.347555532656893e-07, "loss": 0.1139, "step": 63940 }, { "epoch": 1.253921568627451, "grad_norm": 2.3784406185150146, "learning_rate": 3.346025657616109e-07, "loss": 0.1194, "step": 63950 }, { "epoch": 1.2541176470588236, "grad_norm": 4.165017127990723, "learning_rate": 3.3444959564278323e-07, "loss": 0.1495, "step": 63960 }, { "epoch": 1.2543137254901962, "grad_norm": 1.7922272682189941, "learning_rate": 3.3429664292528526e-07, "loss": 0.1422, "step": 63970 }, { "epoch": 1.2545098039215685, "grad_norm": 3.402557134628296, "learning_rate": 3.3414370762519396e-07, "loss": 0.1371, "step": 63980 }, { "epoch": 1.2547058823529411, "grad_norm": 2.1559343338012695, "learning_rate": 3.3399078975858475e-07, "loss": 0.1408, "step": 63990 }, { "epoch": 1.2549019607843137, "grad_norm": 2.159982681274414, "learning_rate": 3.3383788934153125e-07, "loss": 0.1323, "step": 64000 }, { "epoch": 1.2550980392156863, "grad_norm": 2.6065680980682373, "learning_rate": 3.336850063901049e-07, "loss": 0.1245, "step": 64010 }, { "epoch": 1.255294117647059, "grad_norm": 2.615046977996826, "learning_rate": 3.335321409203756e-07, "loss": 0.1132, "step": 64020 }, { "epoch": 1.2554901960784313, "grad_norm": 1.2131637334823608, "learning_rate": 3.3337929294841163e-07, "loss": 0.1316, "step": 64030 }, { "epoch": 1.2556862745098039, "grad_norm": 2.4188339710235596, "learning_rate": 3.3322646249027866e-07, "loss": 0.1304, "step": 64040 }, { "epoch": 1.2558823529411764, "grad_norm": 2.7875568866729736, "learning_rate": 3.3307364956204125e-07, "loss": 0.1115, "step": 64050 }, { "epoch": 1.256078431372549, "grad_norm": 1.6063963174819946, "learning_rate": 3.32920854179762e-07, "loss": 0.134, "step": 64060 }, { "epoch": 1.2562745098039216, "grad_norm": 2.347970962524414, "learning_rate": 3.3276807635950126e-07, "loss": 0.138, "step": 64070 }, { "epoch": 1.256470588235294, "grad_norm": 1.6227918863296509, "learning_rate": 3.3261531611731795e-07, "loss": 0.1312, "step": 64080 }, { "epoch": 1.2566666666666666, "grad_norm": 1.5099282264709473, "learning_rate": 3.324625734692691e-07, "loss": 0.1224, "step": 64090 }, { "epoch": 1.2568627450980392, "grad_norm": 2.4087090492248535, "learning_rate": 3.323098484314095e-07, "loss": 0.1285, "step": 64100 }, { "epoch": 1.2570588235294118, "grad_norm": 3.348802089691162, "learning_rate": 3.3215714101979265e-07, "loss": 0.1289, "step": 64110 }, { "epoch": 1.2572549019607844, "grad_norm": 1.6276578903198242, "learning_rate": 3.320044512504698e-07, "loss": 0.1139, "step": 64120 }, { "epoch": 1.2574509803921567, "grad_norm": 2.992985248565674, "learning_rate": 3.3185177913949065e-07, "loss": 0.1455, "step": 64130 }, { "epoch": 1.2576470588235293, "grad_norm": 3.513991594314575, "learning_rate": 3.316991247029026e-07, "loss": 0.1421, "step": 64140 }, { "epoch": 1.257843137254902, "grad_norm": 2.119187355041504, "learning_rate": 3.3154648795675155e-07, "loss": 0.101, "step": 64150 }, { "epoch": 1.2580392156862745, "grad_norm": 0.8381035327911377, "learning_rate": 3.313938689170817e-07, "loss": 0.0986, "step": 64160 }, { "epoch": 1.2582352941176471, "grad_norm": 2.7300851345062256, "learning_rate": 3.312412675999347e-07, "loss": 0.1212, "step": 64170 }, { "epoch": 1.2584313725490195, "grad_norm": 1.9357597827911377, "learning_rate": 3.3108868402135103e-07, "loss": 0.14, "step": 64180 }, { "epoch": 1.258627450980392, "grad_norm": 2.9731931686401367, "learning_rate": 3.3093611819736924e-07, "loss": 0.1466, "step": 64190 }, { "epoch": 1.2588235294117647, "grad_norm": 1.6895833015441895, "learning_rate": 3.3078357014402533e-07, "loss": 0.1208, "step": 64200 }, { "epoch": 1.2590196078431373, "grad_norm": 2.369643449783325, "learning_rate": 3.306310398773543e-07, "loss": 0.1415, "step": 64210 }, { "epoch": 1.2592156862745099, "grad_norm": 2.971693277359009, "learning_rate": 3.304785274133889e-07, "loss": 0.1616, "step": 64220 }, { "epoch": 1.2594117647058822, "grad_norm": 3.6615049839019775, "learning_rate": 3.3032603276815986e-07, "loss": 0.1265, "step": 64230 }, { "epoch": 1.2596078431372548, "grad_norm": 3.9523396492004395, "learning_rate": 3.3017355595769614e-07, "loss": 0.1291, "step": 64240 }, { "epoch": 1.2598039215686274, "grad_norm": 2.3684792518615723, "learning_rate": 3.300210969980252e-07, "loss": 0.1477, "step": 64250 }, { "epoch": 1.26, "grad_norm": 1.9190387725830078, "learning_rate": 3.29868655905172e-07, "loss": 0.1216, "step": 64260 }, { "epoch": 1.2601960784313726, "grad_norm": 2.1134722232818604, "learning_rate": 3.2971623269515995e-07, "loss": 0.1517, "step": 64270 }, { "epoch": 1.260392156862745, "grad_norm": 2.041653871536255, "learning_rate": 3.2956382738401055e-07, "loss": 0.1336, "step": 64280 }, { "epoch": 1.2605882352941176, "grad_norm": 3.1162641048431396, "learning_rate": 3.2941143998774377e-07, "loss": 0.1035, "step": 64290 }, { "epoch": 1.2607843137254902, "grad_norm": 1.2901581525802612, "learning_rate": 3.292590705223769e-07, "loss": 0.1218, "step": 64300 }, { "epoch": 1.2609803921568628, "grad_norm": 2.0226030349731445, "learning_rate": 3.291067190039261e-07, "loss": 0.134, "step": 64310 }, { "epoch": 1.2611764705882353, "grad_norm": 1.7102550268173218, "learning_rate": 3.289543854484051e-07, "loss": 0.1182, "step": 64320 }, { "epoch": 1.2613725490196077, "grad_norm": 2.0491154193878174, "learning_rate": 3.2880206987182615e-07, "loss": 0.1365, "step": 64330 }, { "epoch": 1.2615686274509805, "grad_norm": 1.522523283958435, "learning_rate": 3.2864977229019935e-07, "loss": 0.1348, "step": 64340 }, { "epoch": 1.261764705882353, "grad_norm": 2.1259963512420654, "learning_rate": 3.284974927195331e-07, "loss": 0.1317, "step": 64350 }, { "epoch": 1.2619607843137255, "grad_norm": 1.5753257274627686, "learning_rate": 3.283452311758337e-07, "loss": 0.1223, "step": 64360 }, { "epoch": 1.262156862745098, "grad_norm": 3.681042432785034, "learning_rate": 3.2819298767510565e-07, "loss": 0.1449, "step": 64370 }, { "epoch": 1.2623529411764705, "grad_norm": 1.4049488306045532, "learning_rate": 3.2804076223335173e-07, "loss": 0.1066, "step": 64380 }, { "epoch": 1.2625490196078433, "grad_norm": 2.7785229682922363, "learning_rate": 3.278885548665725e-07, "loss": 0.1301, "step": 64390 }, { "epoch": 1.2627450980392156, "grad_norm": 4.354941368103027, "learning_rate": 3.2773636559076677e-07, "loss": 0.1201, "step": 64400 }, { "epoch": 1.2629411764705882, "grad_norm": 2.1658592224121094, "learning_rate": 3.275841944219316e-07, "loss": 0.1227, "step": 64410 }, { "epoch": 1.2631372549019608, "grad_norm": 2.66430401802063, "learning_rate": 3.2743204137606183e-07, "loss": 0.1605, "step": 64420 }, { "epoch": 1.2633333333333332, "grad_norm": 4.156144142150879, "learning_rate": 3.272799064691506e-07, "loss": 0.1198, "step": 64430 }, { "epoch": 1.263529411764706, "grad_norm": 0.5635537505149841, "learning_rate": 3.2712778971718927e-07, "loss": 0.1092, "step": 64440 }, { "epoch": 1.2637254901960784, "grad_norm": 1.7692019939422607, "learning_rate": 3.2697569113616697e-07, "loss": 0.1342, "step": 64450 }, { "epoch": 1.263921568627451, "grad_norm": 2.216235876083374, "learning_rate": 3.26823610742071e-07, "loss": 0.1191, "step": 64460 }, { "epoch": 1.2641176470588236, "grad_norm": 2.367858648300171, "learning_rate": 3.26671548550887e-07, "loss": 0.1189, "step": 64470 }, { "epoch": 1.2643137254901962, "grad_norm": 2.143179416656494, "learning_rate": 3.265195045785986e-07, "loss": 0.1157, "step": 64480 }, { "epoch": 1.2645098039215688, "grad_norm": 1.3841296434402466, "learning_rate": 3.2636747884118724e-07, "loss": 0.1177, "step": 64490 }, { "epoch": 1.2647058823529411, "grad_norm": 2.1065244674682617, "learning_rate": 3.262154713546328e-07, "loss": 0.1318, "step": 64500 }, { "epoch": 1.2649019607843137, "grad_norm": 1.845391869544983, "learning_rate": 3.26063482134913e-07, "loss": 0.1319, "step": 64510 }, { "epoch": 1.2650980392156863, "grad_norm": 1.6562708616256714, "learning_rate": 3.2591151119800376e-07, "loss": 0.1146, "step": 64520 }, { "epoch": 1.265294117647059, "grad_norm": 2.607816219329834, "learning_rate": 3.257595585598789e-07, "loss": 0.118, "step": 64530 }, { "epoch": 1.2654901960784315, "grad_norm": 2.9914002418518066, "learning_rate": 3.256076242365108e-07, "loss": 0.1218, "step": 64540 }, { "epoch": 1.2656862745098039, "grad_norm": 3.1999130249023438, "learning_rate": 3.2545570824386925e-07, "loss": 0.1293, "step": 64550 }, { "epoch": 1.2658823529411765, "grad_norm": 3.6329762935638428, "learning_rate": 3.2530381059792255e-07, "loss": 0.1333, "step": 64560 }, { "epoch": 1.266078431372549, "grad_norm": 2.054379463195801, "learning_rate": 3.2515193131463705e-07, "loss": 0.1514, "step": 64570 }, { "epoch": 1.2662745098039216, "grad_norm": 1.9330543279647827, "learning_rate": 3.250000704099769e-07, "loss": 0.1221, "step": 64580 }, { "epoch": 1.2664705882352942, "grad_norm": 2.3492166996002197, "learning_rate": 3.248482278999046e-07, "loss": 0.1377, "step": 64590 }, { "epoch": 1.2666666666666666, "grad_norm": 2.64753794670105, "learning_rate": 3.246964038003807e-07, "loss": 0.1042, "step": 64600 }, { "epoch": 1.2668627450980392, "grad_norm": 4.5319743156433105, "learning_rate": 3.2454459812736355e-07, "loss": 0.1249, "step": 64610 }, { "epoch": 1.2670588235294118, "grad_norm": 4.323459625244141, "learning_rate": 3.2439281089680993e-07, "loss": 0.1472, "step": 64620 }, { "epoch": 1.2672549019607844, "grad_norm": 1.7831532955169678, "learning_rate": 3.2424104212467443e-07, "loss": 0.1342, "step": 64630 }, { "epoch": 1.267450980392157, "grad_norm": 3.894855260848999, "learning_rate": 3.2408929182690946e-07, "loss": 0.1197, "step": 64640 }, { "epoch": 1.2676470588235293, "grad_norm": 2.296921730041504, "learning_rate": 3.2393756001946626e-07, "loss": 0.1464, "step": 64650 }, { "epoch": 1.267843137254902, "grad_norm": 1.4210642576217651, "learning_rate": 3.237858467182935e-07, "loss": 0.1221, "step": 64660 }, { "epoch": 1.2680392156862745, "grad_norm": 4.824569225311279, "learning_rate": 3.2363415193933806e-07, "loss": 0.1406, "step": 64670 }, { "epoch": 1.2682352941176471, "grad_norm": 1.6004520654678345, "learning_rate": 3.234824756985448e-07, "loss": 0.0967, "step": 64680 }, { "epoch": 1.2684313725490197, "grad_norm": 2.570357084274292, "learning_rate": 3.233308180118568e-07, "loss": 0.1275, "step": 64690 }, { "epoch": 1.268627450980392, "grad_norm": 2.9547030925750732, "learning_rate": 3.2317917889521514e-07, "loss": 0.1037, "step": 64700 }, { "epoch": 1.2688235294117647, "grad_norm": 2.531597852706909, "learning_rate": 3.230275583645587e-07, "loss": 0.1044, "step": 64710 }, { "epoch": 1.2690196078431373, "grad_norm": 2.4435198307037354, "learning_rate": 3.2287595643582475e-07, "loss": 0.1196, "step": 64720 }, { "epoch": 1.2692156862745099, "grad_norm": 1.8998775482177734, "learning_rate": 3.227243731249486e-07, "loss": 0.1128, "step": 64730 }, { "epoch": 1.2694117647058825, "grad_norm": 3.3820669651031494, "learning_rate": 3.2257280844786314e-07, "loss": 0.1117, "step": 64740 }, { "epoch": 1.2696078431372548, "grad_norm": 3.069061517715454, "learning_rate": 3.2242126242049995e-07, "loss": 0.1254, "step": 64750 }, { "epoch": 1.2698039215686274, "grad_norm": 1.9367578029632568, "learning_rate": 3.222697350587882e-07, "loss": 0.1221, "step": 64760 }, { "epoch": 1.27, "grad_norm": 3.679527759552002, "learning_rate": 3.2211822637865517e-07, "loss": 0.1296, "step": 64770 }, { "epoch": 1.2701960784313726, "grad_norm": 2.015181064605713, "learning_rate": 3.2196673639602627e-07, "loss": 0.1212, "step": 64780 }, { "epoch": 1.2703921568627452, "grad_norm": 1.900976300239563, "learning_rate": 3.218152651268251e-07, "loss": 0.1287, "step": 64790 }, { "epoch": 1.2705882352941176, "grad_norm": 3.0189874172210693, "learning_rate": 3.2166381258697276e-07, "loss": 0.1294, "step": 64800 }, { "epoch": 1.2707843137254902, "grad_norm": 4.346193790435791, "learning_rate": 3.21512378792389e-07, "loss": 0.1245, "step": 64810 }, { "epoch": 1.2709803921568628, "grad_norm": 1.2225062847137451, "learning_rate": 3.213609637589911e-07, "loss": 0.101, "step": 64820 }, { "epoch": 1.2711764705882354, "grad_norm": 2.8261988162994385, "learning_rate": 3.212095675026949e-07, "loss": 0.1195, "step": 64830 }, { "epoch": 1.271372549019608, "grad_norm": 3.219594717025757, "learning_rate": 3.210581900394137e-07, "loss": 0.1428, "step": 64840 }, { "epoch": 1.2715686274509803, "grad_norm": 4.1651811599731445, "learning_rate": 3.2090683138505924e-07, "loss": 0.1436, "step": 64850 }, { "epoch": 1.271764705882353, "grad_norm": 2.046488046646118, "learning_rate": 3.2075549155554103e-07, "loss": 0.1158, "step": 64860 }, { "epoch": 1.2719607843137255, "grad_norm": 2.996182441711426, "learning_rate": 3.206041705667668e-07, "loss": 0.1556, "step": 64870 }, { "epoch": 1.272156862745098, "grad_norm": 3.1067652702331543, "learning_rate": 3.20452868434642e-07, "loss": 0.1466, "step": 64880 }, { "epoch": 1.2723529411764707, "grad_norm": 3.4800772666931152, "learning_rate": 3.2030158517507055e-07, "loss": 0.1148, "step": 64890 }, { "epoch": 1.272549019607843, "grad_norm": 2.9566054344177246, "learning_rate": 3.2015032080395387e-07, "loss": 0.1356, "step": 64900 }, { "epoch": 1.2727450980392156, "grad_norm": 2.1403629779815674, "learning_rate": 3.1999907533719184e-07, "loss": 0.1307, "step": 64910 }, { "epoch": 1.2729411764705882, "grad_norm": 1.8315945863723755, "learning_rate": 3.1984784879068215e-07, "loss": 0.119, "step": 64920 }, { "epoch": 1.2731372549019608, "grad_norm": 2.51257061958313, "learning_rate": 3.196966411803205e-07, "loss": 0.1131, "step": 64930 }, { "epoch": 1.2733333333333334, "grad_norm": 1.816564917564392, "learning_rate": 3.195454525220005e-07, "loss": 0.1394, "step": 64940 }, { "epoch": 1.2735294117647058, "grad_norm": 3.8992385864257812, "learning_rate": 3.1939428283161417e-07, "loss": 0.1421, "step": 64950 }, { "epoch": 1.2737254901960784, "grad_norm": 1.7806711196899414, "learning_rate": 3.1924313212505094e-07, "loss": 0.1125, "step": 64960 }, { "epoch": 1.273921568627451, "grad_norm": 2.5127899646759033, "learning_rate": 3.1909200041819873e-07, "loss": 0.1244, "step": 64970 }, { "epoch": 1.2741176470588236, "grad_norm": 2.09289813041687, "learning_rate": 3.189408877269433e-07, "loss": 0.1476, "step": 64980 }, { "epoch": 1.2743137254901962, "grad_norm": 3.6365151405334473, "learning_rate": 3.1878979406716834e-07, "loss": 0.1318, "step": 64990 }, { "epoch": 1.2745098039215685, "grad_norm": 2.9835243225097656, "learning_rate": 3.1863871945475543e-07, "loss": 0.1107, "step": 65000 }, { "epoch": 1.2747058823529411, "grad_norm": 1.6191445589065552, "learning_rate": 3.1848766390558466e-07, "loss": 0.1238, "step": 65010 }, { "epoch": 1.2749019607843137, "grad_norm": 5.806128025054932, "learning_rate": 3.183366274355338e-07, "loss": 0.1251, "step": 65020 }, { "epoch": 1.2750980392156863, "grad_norm": 3.309255838394165, "learning_rate": 3.181856100604783e-07, "loss": 0.1185, "step": 65030 }, { "epoch": 1.275294117647059, "grad_norm": 2.720932722091675, "learning_rate": 3.1803461179629196e-07, "loss": 0.111, "step": 65040 }, { "epoch": 1.2754901960784313, "grad_norm": 2.34049654006958, "learning_rate": 3.1788363265884677e-07, "loss": 0.1056, "step": 65050 }, { "epoch": 1.2756862745098039, "grad_norm": 1.6578601598739624, "learning_rate": 3.1773267266401203e-07, "loss": 0.1282, "step": 65060 }, { "epoch": 1.2758823529411765, "grad_norm": 2.8535377979278564, "learning_rate": 3.175817318276557e-07, "loss": 0.1136, "step": 65070 }, { "epoch": 1.276078431372549, "grad_norm": 3.9023184776306152, "learning_rate": 3.1743081016564357e-07, "loss": 0.1144, "step": 65080 }, { "epoch": 1.2762745098039217, "grad_norm": 1.1242921352386475, "learning_rate": 3.1727990769383897e-07, "loss": 0.1089, "step": 65090 }, { "epoch": 1.276470588235294, "grad_norm": 1.6478755474090576, "learning_rate": 3.1712902442810383e-07, "loss": 0.123, "step": 65100 }, { "epoch": 1.2766666666666666, "grad_norm": 1.3423411846160889, "learning_rate": 3.169781603842978e-07, "loss": 0.1163, "step": 65110 }, { "epoch": 1.2768627450980392, "grad_norm": 2.962050676345825, "learning_rate": 3.1682731557827825e-07, "loss": 0.1098, "step": 65120 }, { "epoch": 1.2770588235294118, "grad_norm": 3.9813942909240723, "learning_rate": 3.16676490025901e-07, "loss": 0.1293, "step": 65130 }, { "epoch": 1.2772549019607844, "grad_norm": 3.7336390018463135, "learning_rate": 3.1652568374301963e-07, "loss": 0.1409, "step": 65140 }, { "epoch": 1.2774509803921568, "grad_norm": 2.333897352218628, "learning_rate": 3.163748967454855e-07, "loss": 0.1351, "step": 65150 }, { "epoch": 1.2776470588235294, "grad_norm": 4.598500728607178, "learning_rate": 3.1622412904914824e-07, "loss": 0.0983, "step": 65160 }, { "epoch": 1.277843137254902, "grad_norm": 2.2520525455474854, "learning_rate": 3.1607338066985544e-07, "loss": 0.1083, "step": 65170 }, { "epoch": 1.2780392156862745, "grad_norm": 2.693220376968384, "learning_rate": 3.159226516234523e-07, "loss": 0.1459, "step": 65180 }, { "epoch": 1.2782352941176471, "grad_norm": 1.847973346710205, "learning_rate": 3.157719419257825e-07, "loss": 0.1251, "step": 65190 }, { "epoch": 1.2784313725490195, "grad_norm": 4.1217498779296875, "learning_rate": 3.1562125159268727e-07, "loss": 0.1242, "step": 65200 }, { "epoch": 1.278627450980392, "grad_norm": 3.1612210273742676, "learning_rate": 3.154705806400062e-07, "loss": 0.1415, "step": 65210 }, { "epoch": 1.2788235294117647, "grad_norm": 2.637340784072876, "learning_rate": 3.153199290835763e-07, "loss": 0.1456, "step": 65220 }, { "epoch": 1.2790196078431373, "grad_norm": 19.924285888671875, "learning_rate": 3.151692969392331e-07, "loss": 0.1262, "step": 65230 }, { "epoch": 1.2792156862745099, "grad_norm": 2.9786341190338135, "learning_rate": 3.1501868422280976e-07, "loss": 0.1434, "step": 65240 }, { "epoch": 1.2794117647058822, "grad_norm": 1.6026616096496582, "learning_rate": 3.148680909501373e-07, "loss": 0.1242, "step": 65250 }, { "epoch": 1.2796078431372548, "grad_norm": 4.731010913848877, "learning_rate": 3.147175171370452e-07, "loss": 0.1109, "step": 65260 }, { "epoch": 1.2798039215686274, "grad_norm": 5.406619548797607, "learning_rate": 3.1456696279936043e-07, "loss": 0.1388, "step": 65270 }, { "epoch": 1.28, "grad_norm": 2.8185601234436035, "learning_rate": 3.144164279529079e-07, "loss": 0.1331, "step": 65280 }, { "epoch": 1.2801960784313726, "grad_norm": 2.9921042919158936, "learning_rate": 3.142659126135108e-07, "loss": 0.1157, "step": 65290 }, { "epoch": 1.280392156862745, "grad_norm": 2.8073925971984863, "learning_rate": 3.1411541679699014e-07, "loss": 0.0924, "step": 65300 }, { "epoch": 1.2805882352941176, "grad_norm": 1.2400574684143066, "learning_rate": 3.1396494051916466e-07, "loss": 0.1284, "step": 65310 }, { "epoch": 1.2807843137254902, "grad_norm": 2.4355812072753906, "learning_rate": 3.138144837958513e-07, "loss": 0.1329, "step": 65320 }, { "epoch": 1.2809803921568628, "grad_norm": 1.6786670684814453, "learning_rate": 3.1366404664286495e-07, "loss": 0.1043, "step": 65330 }, { "epoch": 1.2811764705882354, "grad_norm": 3.3366987705230713, "learning_rate": 3.135136290760182e-07, "loss": 0.1432, "step": 65340 }, { "epoch": 1.2813725490196077, "grad_norm": 2.2436599731445312, "learning_rate": 3.1336323111112184e-07, "loss": 0.131, "step": 65350 }, { "epoch": 1.2815686274509803, "grad_norm": 1.3203306198120117, "learning_rate": 3.1321285276398423e-07, "loss": 0.1393, "step": 65360 }, { "epoch": 1.281764705882353, "grad_norm": 2.4901347160339355, "learning_rate": 3.1306249405041246e-07, "loss": 0.1056, "step": 65370 }, { "epoch": 1.2819607843137255, "grad_norm": 2.094721555709839, "learning_rate": 3.129121549862106e-07, "loss": 0.1441, "step": 65380 }, { "epoch": 1.282156862745098, "grad_norm": 1.561299204826355, "learning_rate": 3.127618355871813e-07, "loss": 0.1121, "step": 65390 }, { "epoch": 1.2823529411764705, "grad_norm": 2.768383026123047, "learning_rate": 3.1261153586912494e-07, "loss": 0.1208, "step": 65400 }, { "epoch": 1.282549019607843, "grad_norm": 1.6607029438018799, "learning_rate": 3.1246125584783957e-07, "loss": 0.1465, "step": 65410 }, { "epoch": 1.2827450980392157, "grad_norm": 2.8780405521392822, "learning_rate": 3.123109955391217e-07, "loss": 0.1108, "step": 65420 }, { "epoch": 1.2829411764705883, "grad_norm": 2.9701898097991943, "learning_rate": 3.121607549587654e-07, "loss": 0.1262, "step": 65430 }, { "epoch": 1.2831372549019608, "grad_norm": 1.7901419401168823, "learning_rate": 3.120105341225626e-07, "loss": 0.126, "step": 65440 }, { "epoch": 1.2833333333333332, "grad_norm": 3.2203524112701416, "learning_rate": 3.1186033304630346e-07, "loss": 0.1397, "step": 65450 }, { "epoch": 1.2835294117647058, "grad_norm": 3.5293304920196533, "learning_rate": 3.11710151745776e-07, "loss": 0.1306, "step": 65460 }, { "epoch": 1.2837254901960784, "grad_norm": 1.2126959562301636, "learning_rate": 3.115599902367658e-07, "loss": 0.1234, "step": 65470 }, { "epoch": 1.283921568627451, "grad_norm": 2.8169219493865967, "learning_rate": 3.1140984853505683e-07, "loss": 0.15, "step": 65480 }, { "epoch": 1.2841176470588236, "grad_norm": 3.048896551132202, "learning_rate": 3.1125972665643085e-07, "loss": 0.1517, "step": 65490 }, { "epoch": 1.284313725490196, "grad_norm": 1.3856205940246582, "learning_rate": 3.1110962461666725e-07, "loss": 0.1176, "step": 65500 }, { "epoch": 1.2845098039215685, "grad_norm": 2.1463403701782227, "learning_rate": 3.1095954243154365e-07, "loss": 0.1139, "step": 65510 }, { "epoch": 1.2847058823529411, "grad_norm": 1.9735186100006104, "learning_rate": 3.1080948011683563e-07, "loss": 0.1333, "step": 65520 }, { "epoch": 1.2849019607843137, "grad_norm": 3.4032468795776367, "learning_rate": 3.106594376883162e-07, "loss": 0.1352, "step": 65530 }, { "epoch": 1.2850980392156863, "grad_norm": 3.5647544860839844, "learning_rate": 3.1050941516175686e-07, "loss": 0.093, "step": 65540 }, { "epoch": 1.2852941176470587, "grad_norm": 2.7513115406036377, "learning_rate": 3.103594125529268e-07, "loss": 0.1296, "step": 65550 }, { "epoch": 1.2854901960784313, "grad_norm": 2.137183666229248, "learning_rate": 3.102094298775931e-07, "loss": 0.1183, "step": 65560 }, { "epoch": 1.2856862745098039, "grad_norm": 1.1933571100234985, "learning_rate": 3.100594671515205e-07, "loss": 0.1238, "step": 65570 }, { "epoch": 1.2858823529411765, "grad_norm": 1.8441054821014404, "learning_rate": 3.0990952439047215e-07, "loss": 0.1287, "step": 65580 }, { "epoch": 1.286078431372549, "grad_norm": 2.2122020721435547, "learning_rate": 3.0975960161020875e-07, "loss": 0.1647, "step": 65590 }, { "epoch": 1.2862745098039214, "grad_norm": 2.475565195083618, "learning_rate": 3.096096988264888e-07, "loss": 0.1272, "step": 65600 }, { "epoch": 1.2864705882352943, "grad_norm": 1.9418741464614868, "learning_rate": 3.0945981605506904e-07, "loss": 0.1249, "step": 65610 }, { "epoch": 1.2866666666666666, "grad_norm": 3.2968180179595947, "learning_rate": 3.0930995331170404e-07, "loss": 0.1451, "step": 65620 }, { "epoch": 1.2868627450980392, "grad_norm": 4.674140453338623, "learning_rate": 3.0916011061214597e-07, "loss": 0.1035, "step": 65630 }, { "epoch": 1.2870588235294118, "grad_norm": 2.6399128437042236, "learning_rate": 3.0901028797214513e-07, "loss": 0.1526, "step": 65640 }, { "epoch": 1.2872549019607842, "grad_norm": 1.1313167810440063, "learning_rate": 3.0886048540744983e-07, "loss": 0.1346, "step": 65650 }, { "epoch": 1.287450980392157, "grad_norm": 2.302064895629883, "learning_rate": 3.08710702933806e-07, "loss": 0.1187, "step": 65660 }, { "epoch": 1.2876470588235294, "grad_norm": 2.3747692108154297, "learning_rate": 3.085609405669574e-07, "loss": 0.1398, "step": 65670 }, { "epoch": 1.287843137254902, "grad_norm": 2.0969040393829346, "learning_rate": 3.0841119832264623e-07, "loss": 0.1239, "step": 65680 }, { "epoch": 1.2880392156862746, "grad_norm": 2.505885601043701, "learning_rate": 3.082614762166118e-07, "loss": 0.1251, "step": 65690 }, { "epoch": 1.288235294117647, "grad_norm": 2.029587984085083, "learning_rate": 3.081117742645921e-07, "loss": 0.0899, "step": 65700 }, { "epoch": 1.2884313725490197, "grad_norm": 1.4790194034576416, "learning_rate": 3.07962092482322e-07, "loss": 0.1256, "step": 65710 }, { "epoch": 1.288627450980392, "grad_norm": 1.3949695825576782, "learning_rate": 3.078124308855355e-07, "loss": 0.1537, "step": 65720 }, { "epoch": 1.2888235294117647, "grad_norm": 2.037104368209839, "learning_rate": 3.076627894899635e-07, "loss": 0.123, "step": 65730 }, { "epoch": 1.2890196078431373, "grad_norm": 4.765023231506348, "learning_rate": 3.075131683113351e-07, "loss": 0.1104, "step": 65740 }, { "epoch": 1.2892156862745099, "grad_norm": 2.814674139022827, "learning_rate": 3.073635673653775e-07, "loss": 0.1203, "step": 65750 }, { "epoch": 1.2894117647058825, "grad_norm": 3.4641547203063965, "learning_rate": 3.072139866678152e-07, "loss": 0.1267, "step": 65760 }, { "epoch": 1.2896078431372549, "grad_norm": 2.9585351943969727, "learning_rate": 3.070644262343711e-07, "loss": 0.1183, "step": 65770 }, { "epoch": 1.2898039215686274, "grad_norm": 4.979754447937012, "learning_rate": 3.0691488608076594e-07, "loss": 0.1409, "step": 65780 }, { "epoch": 1.29, "grad_norm": 1.7923792600631714, "learning_rate": 3.067653662227179e-07, "loss": 0.1232, "step": 65790 }, { "epoch": 1.2901960784313726, "grad_norm": 2.4680395126342773, "learning_rate": 3.0661586667594353e-07, "loss": 0.1099, "step": 65800 }, { "epoch": 1.2903921568627452, "grad_norm": 1.802024006843567, "learning_rate": 3.0646638745615693e-07, "loss": 0.1509, "step": 65810 }, { "epoch": 1.2905882352941176, "grad_norm": 1.9146965742111206, "learning_rate": 3.063169285790701e-07, "loss": 0.1085, "step": 65820 }, { "epoch": 1.2907843137254902, "grad_norm": 2.8080689907073975, "learning_rate": 3.06167490060393e-07, "loss": 0.1405, "step": 65830 }, { "epoch": 1.2909803921568628, "grad_norm": 0.8539981842041016, "learning_rate": 3.060180719158335e-07, "loss": 0.1071, "step": 65840 }, { "epoch": 1.2911764705882354, "grad_norm": 2.965745687484741, "learning_rate": 3.0586867416109707e-07, "loss": 0.1093, "step": 65850 }, { "epoch": 1.291372549019608, "grad_norm": 1.7738044261932373, "learning_rate": 3.0571929681188733e-07, "loss": 0.1752, "step": 65860 }, { "epoch": 1.2915686274509803, "grad_norm": 2.4544591903686523, "learning_rate": 3.0556993988390566e-07, "loss": 0.1132, "step": 65870 }, { "epoch": 1.291764705882353, "grad_norm": 3.8983607292175293, "learning_rate": 3.0542060339285114e-07, "loss": 0.1279, "step": 65880 }, { "epoch": 1.2919607843137255, "grad_norm": 1.7085963487625122, "learning_rate": 3.0527128735442076e-07, "loss": 0.1317, "step": 65890 }, { "epoch": 1.2921568627450981, "grad_norm": 3.4944908618927, "learning_rate": 3.051219917843096e-07, "loss": 0.1155, "step": 65900 }, { "epoch": 1.2923529411764707, "grad_norm": 1.9715532064437866, "learning_rate": 3.0497271669821044e-07, "loss": 0.1074, "step": 65910 }, { "epoch": 1.292549019607843, "grad_norm": 2.6648061275482178, "learning_rate": 3.0482346211181376e-07, "loss": 0.1142, "step": 65920 }, { "epoch": 1.2927450980392157, "grad_norm": 2.0874040126800537, "learning_rate": 3.04674228040808e-07, "loss": 0.1275, "step": 65930 }, { "epoch": 1.2929411764705883, "grad_norm": 1.9656903743743896, "learning_rate": 3.0452501450087955e-07, "loss": 0.1101, "step": 65940 }, { "epoch": 1.2931372549019609, "grad_norm": 2.673274278640747, "learning_rate": 3.043758215077124e-07, "loss": 0.1549, "step": 65950 }, { "epoch": 1.2933333333333334, "grad_norm": 1.9403400421142578, "learning_rate": 3.042266490769886e-07, "loss": 0.1234, "step": 65960 }, { "epoch": 1.2935294117647058, "grad_norm": 2.387497901916504, "learning_rate": 3.04077497224388e-07, "loss": 0.1259, "step": 65970 }, { "epoch": 1.2937254901960784, "grad_norm": 2.0215904712677, "learning_rate": 3.039283659655881e-07, "loss": 0.1505, "step": 65980 }, { "epoch": 1.293921568627451, "grad_norm": 2.2553389072418213, "learning_rate": 3.037792553162645e-07, "loss": 0.1152, "step": 65990 }, { "epoch": 1.2941176470588236, "grad_norm": 3.852870464324951, "learning_rate": 3.036301652920906e-07, "loss": 0.1126, "step": 66000 }, { "epoch": 1.2943137254901962, "grad_norm": 1.4044456481933594, "learning_rate": 3.034810959087374e-07, "loss": 0.1399, "step": 66010 }, { "epoch": 1.2945098039215686, "grad_norm": 2.189096450805664, "learning_rate": 3.0333204718187377e-07, "loss": 0.0977, "step": 66020 }, { "epoch": 1.2947058823529412, "grad_norm": 2.5342791080474854, "learning_rate": 3.0318301912716684e-07, "loss": 0.1147, "step": 66030 }, { "epoch": 1.2949019607843137, "grad_norm": 2.2094874382019043, "learning_rate": 3.0303401176028097e-07, "loss": 0.1363, "step": 66040 }, { "epoch": 1.2950980392156863, "grad_norm": 1.6281942129135132, "learning_rate": 3.0288502509687874e-07, "loss": 0.1224, "step": 66050 }, { "epoch": 1.295294117647059, "grad_norm": 5.193877696990967, "learning_rate": 3.027360591526203e-07, "loss": 0.119, "step": 66060 }, { "epoch": 1.2954901960784313, "grad_norm": 9.054879188537598, "learning_rate": 3.025871139431639e-07, "loss": 0.1485, "step": 66070 }, { "epoch": 1.295686274509804, "grad_norm": 3.734135627746582, "learning_rate": 3.024381894841654e-07, "loss": 0.1237, "step": 66080 }, { "epoch": 1.2958823529411765, "grad_norm": 3.221576690673828, "learning_rate": 3.0228928579127853e-07, "loss": 0.1195, "step": 66090 }, { "epoch": 1.296078431372549, "grad_norm": 2.077666997909546, "learning_rate": 3.0214040288015507e-07, "loss": 0.1399, "step": 66100 }, { "epoch": 1.2962745098039217, "grad_norm": 2.38309645652771, "learning_rate": 3.019915407664441e-07, "loss": 0.0915, "step": 66110 }, { "epoch": 1.296470588235294, "grad_norm": 1.9453465938568115, "learning_rate": 3.0184269946579295e-07, "loss": 0.1182, "step": 66120 }, { "epoch": 1.2966666666666666, "grad_norm": 3.420548915863037, "learning_rate": 3.016938789938466e-07, "loss": 0.1057, "step": 66130 }, { "epoch": 1.2968627450980392, "grad_norm": 3.1795825958251953, "learning_rate": 3.015450793662478e-07, "loss": 0.1091, "step": 66140 }, { "epoch": 1.2970588235294118, "grad_norm": 1.8896856307983398, "learning_rate": 3.013963005986373e-07, "loss": 0.1313, "step": 66150 }, { "epoch": 1.2972549019607844, "grad_norm": 2.1288301944732666, "learning_rate": 3.0124754270665355e-07, "loss": 0.1103, "step": 66160 }, { "epoch": 1.2974509803921568, "grad_norm": 1.5767359733581543, "learning_rate": 3.010988057059325e-07, "loss": 0.1276, "step": 66170 }, { "epoch": 1.2976470588235294, "grad_norm": 2.6195530891418457, "learning_rate": 3.009500896121084e-07, "loss": 0.1319, "step": 66180 }, { "epoch": 1.297843137254902, "grad_norm": 3.2273426055908203, "learning_rate": 3.008013944408133e-07, "loss": 0.1482, "step": 66190 }, { "epoch": 1.2980392156862746, "grad_norm": 2.2708423137664795, "learning_rate": 3.006527202076764e-07, "loss": 0.109, "step": 66200 }, { "epoch": 1.2982352941176472, "grad_norm": 3.866255521774292, "learning_rate": 3.005040669283254e-07, "loss": 0.1159, "step": 66210 }, { "epoch": 1.2984313725490195, "grad_norm": 2.589477300643921, "learning_rate": 3.0035543461838565e-07, "loss": 0.1153, "step": 66220 }, { "epoch": 1.2986274509803921, "grad_norm": 2.002274990081787, "learning_rate": 3.0020682329347985e-07, "loss": 0.1146, "step": 66230 }, { "epoch": 1.2988235294117647, "grad_norm": 1.8811687231063843, "learning_rate": 3.0005823296922914e-07, "loss": 0.1247, "step": 66240 }, { "epoch": 1.2990196078431373, "grad_norm": 2.5688910484313965, "learning_rate": 2.9990966366125177e-07, "loss": 0.1361, "step": 66250 }, { "epoch": 1.29921568627451, "grad_norm": 2.177398204803467, "learning_rate": 2.997611153851647e-07, "loss": 0.1279, "step": 66260 }, { "epoch": 1.2994117647058823, "grad_norm": 1.0730743408203125, "learning_rate": 2.9961258815658165e-07, "loss": 0.1093, "step": 66270 }, { "epoch": 1.2996078431372549, "grad_norm": 1.3521814346313477, "learning_rate": 2.9946408199111483e-07, "loss": 0.1328, "step": 66280 }, { "epoch": 1.2998039215686275, "grad_norm": 2.511131763458252, "learning_rate": 2.9931559690437413e-07, "loss": 0.1114, "step": 66290 }, { "epoch": 1.3, "grad_norm": 1.6156588792800903, "learning_rate": 2.9916713291196683e-07, "loss": 0.1139, "step": 66300 }, { "epoch": 1.3001960784313726, "grad_norm": 2.503591537475586, "learning_rate": 2.990186900294984e-07, "loss": 0.1377, "step": 66310 }, { "epoch": 1.300392156862745, "grad_norm": 3.423478841781616, "learning_rate": 2.988702682725721e-07, "loss": 0.1319, "step": 66320 }, { "epoch": 1.3005882352941176, "grad_norm": 3.029318332672119, "learning_rate": 2.987218676567885e-07, "loss": 0.1089, "step": 66330 }, { "epoch": 1.3007843137254902, "grad_norm": 3.9978384971618652, "learning_rate": 2.985734881977466e-07, "loss": 0.0946, "step": 66340 }, { "epoch": 1.3009803921568628, "grad_norm": 2.2549920082092285, "learning_rate": 2.9842512991104277e-07, "loss": 0.1142, "step": 66350 }, { "epoch": 1.3011764705882354, "grad_norm": 2.0447425842285156, "learning_rate": 2.9827679281227114e-07, "loss": 0.1175, "step": 66360 }, { "epoch": 1.3013725490196077, "grad_norm": 1.6657249927520752, "learning_rate": 2.981284769170238e-07, "loss": 0.1112, "step": 66370 }, { "epoch": 1.3015686274509803, "grad_norm": 4.579926490783691, "learning_rate": 2.979801822408906e-07, "loss": 0.1362, "step": 66380 }, { "epoch": 1.301764705882353, "grad_norm": 1.7659850120544434, "learning_rate": 2.9783190879945897e-07, "loss": 0.1199, "step": 66390 }, { "epoch": 1.3019607843137255, "grad_norm": 2.8066344261169434, "learning_rate": 2.976836566083143e-07, "loss": 0.1177, "step": 66400 }, { "epoch": 1.3021568627450981, "grad_norm": 4.780801773071289, "learning_rate": 2.9753542568303946e-07, "loss": 0.1177, "step": 66410 }, { "epoch": 1.3023529411764705, "grad_norm": 2.727628469467163, "learning_rate": 2.9738721603921556e-07, "loss": 0.1522, "step": 66420 }, { "epoch": 1.302549019607843, "grad_norm": 1.244524359703064, "learning_rate": 2.9723902769242095e-07, "loss": 0.1398, "step": 66430 }, { "epoch": 1.3027450980392157, "grad_norm": 1.1453901529312134, "learning_rate": 2.970908606582323e-07, "loss": 0.1418, "step": 66440 }, { "epoch": 1.3029411764705883, "grad_norm": 2.261075973510742, "learning_rate": 2.9694271495222365e-07, "loss": 0.1271, "step": 66450 }, { "epoch": 1.3031372549019609, "grad_norm": 3.052316904067993, "learning_rate": 2.9679459058996673e-07, "loss": 0.1318, "step": 66460 }, { "epoch": 1.3033333333333332, "grad_norm": 1.6865837574005127, "learning_rate": 2.966464875870314e-07, "loss": 0.135, "step": 66470 }, { "epoch": 1.3035294117647058, "grad_norm": 2.0150582790374756, "learning_rate": 2.964984059589849e-07, "loss": 0.1254, "step": 66480 }, { "epoch": 1.3037254901960784, "grad_norm": 2.34409236907959, "learning_rate": 2.963503457213924e-07, "loss": 0.1457, "step": 66490 }, { "epoch": 1.303921568627451, "grad_norm": 3.4147772789001465, "learning_rate": 2.9620230688981687e-07, "loss": 0.1269, "step": 66500 }, { "epoch": 1.3041176470588236, "grad_norm": 1.4735594987869263, "learning_rate": 2.960542894798189e-07, "loss": 0.1091, "step": 66510 }, { "epoch": 1.304313725490196, "grad_norm": 2.273111343383789, "learning_rate": 2.9590629350695696e-07, "loss": 0.1294, "step": 66520 }, { "epoch": 1.3045098039215686, "grad_norm": 3.9868922233581543, "learning_rate": 2.9575831898678707e-07, "loss": 0.125, "step": 66530 }, { "epoch": 1.3047058823529412, "grad_norm": 2.830085515975952, "learning_rate": 2.956103659348633e-07, "loss": 0.1275, "step": 66540 }, { "epoch": 1.3049019607843138, "grad_norm": 2.6428205966949463, "learning_rate": 2.9546243436673717e-07, "loss": 0.1225, "step": 66550 }, { "epoch": 1.3050980392156863, "grad_norm": 2.3533191680908203, "learning_rate": 2.9531452429795804e-07, "loss": 0.1127, "step": 66560 }, { "epoch": 1.3052941176470587, "grad_norm": 2.223236322402954, "learning_rate": 2.9516663574407317e-07, "loss": 0.1171, "step": 66570 }, { "epoch": 1.3054901960784313, "grad_norm": 6.403404712677002, "learning_rate": 2.9501876872062724e-07, "loss": 0.1212, "step": 66580 }, { "epoch": 1.305686274509804, "grad_norm": 1.872247576713562, "learning_rate": 2.948709232431631e-07, "loss": 0.1095, "step": 66590 }, { "epoch": 1.3058823529411765, "grad_norm": 7.791156768798828, "learning_rate": 2.9472309932722074e-07, "loss": 0.1421, "step": 66600 }, { "epoch": 1.306078431372549, "grad_norm": 1.8881808519363403, "learning_rate": 2.945752969883383e-07, "loss": 0.1235, "step": 66610 }, { "epoch": 1.3062745098039215, "grad_norm": 2.825730800628662, "learning_rate": 2.9442751624205175e-07, "loss": 0.1001, "step": 66620 }, { "epoch": 1.306470588235294, "grad_norm": 1.8385916948318481, "learning_rate": 2.942797571038945e-07, "loss": 0.1232, "step": 66630 }, { "epoch": 1.3066666666666666, "grad_norm": 2.722876787185669, "learning_rate": 2.94132019589398e-07, "loss": 0.122, "step": 66640 }, { "epoch": 1.3068627450980392, "grad_norm": 1.8683834075927734, "learning_rate": 2.9398430371409095e-07, "loss": 0.1486, "step": 66650 }, { "epoch": 1.3070588235294118, "grad_norm": 4.433448314666748, "learning_rate": 2.938366094935002e-07, "loss": 0.1263, "step": 66660 }, { "epoch": 1.3072549019607842, "grad_norm": 3.868621349334717, "learning_rate": 2.936889369431502e-07, "loss": 0.1381, "step": 66670 }, { "epoch": 1.3074509803921568, "grad_norm": 1.623042345046997, "learning_rate": 2.9354128607856297e-07, "loss": 0.1095, "step": 66680 }, { "epoch": 1.3076470588235294, "grad_norm": 3.5171847343444824, "learning_rate": 2.933936569152585e-07, "loss": 0.1189, "step": 66690 }, { "epoch": 1.307843137254902, "grad_norm": 1.9762464761734009, "learning_rate": 2.9324604946875446e-07, "loss": 0.116, "step": 66700 }, { "epoch": 1.3080392156862746, "grad_norm": 2.6654164791107178, "learning_rate": 2.9309846375456594e-07, "loss": 0.1372, "step": 66710 }, { "epoch": 1.308235294117647, "grad_norm": 0.8824384212493896, "learning_rate": 2.929508997882061e-07, "loss": 0.1253, "step": 66720 }, { "epoch": 1.3084313725490195, "grad_norm": 2.725022554397583, "learning_rate": 2.928033575851858e-07, "loss": 0.1229, "step": 66730 }, { "epoch": 1.3086274509803921, "grad_norm": 7.44978666305542, "learning_rate": 2.9265583716101314e-07, "loss": 0.1535, "step": 66740 }, { "epoch": 1.3088235294117647, "grad_norm": 1.4518464803695679, "learning_rate": 2.925083385311947e-07, "loss": 0.1288, "step": 66750 }, { "epoch": 1.3090196078431373, "grad_norm": 2.8180882930755615, "learning_rate": 2.9236086171123404e-07, "loss": 0.092, "step": 66760 }, { "epoch": 1.3092156862745097, "grad_norm": 3.6313655376434326, "learning_rate": 2.922134067166328e-07, "loss": 0.1066, "step": 66770 }, { "epoch": 1.3094117647058823, "grad_norm": 2.3788340091705322, "learning_rate": 2.920659735628904e-07, "loss": 0.1331, "step": 66780 }, { "epoch": 1.3096078431372549, "grad_norm": 2.126757860183716, "learning_rate": 2.919185622655036e-07, "loss": 0.1485, "step": 66790 }, { "epoch": 1.3098039215686275, "grad_norm": 2.8906562328338623, "learning_rate": 2.917711728399672e-07, "loss": 0.1276, "step": 66800 }, { "epoch": 1.31, "grad_norm": 9.21432876586914, "learning_rate": 2.916238053017739e-07, "loss": 0.1551, "step": 66810 }, { "epoch": 1.3101960784313724, "grad_norm": 2.368805170059204, "learning_rate": 2.914764596664134e-07, "loss": 0.1156, "step": 66820 }, { "epoch": 1.310392156862745, "grad_norm": 2.8449313640594482, "learning_rate": 2.9132913594937347e-07, "loss": 0.1542, "step": 66830 }, { "epoch": 1.3105882352941176, "grad_norm": 2.246717691421509, "learning_rate": 2.9118183416613993e-07, "loss": 0.121, "step": 66840 }, { "epoch": 1.3107843137254902, "grad_norm": 5.197099208831787, "learning_rate": 2.910345543321958e-07, "loss": 0.1507, "step": 66850 }, { "epoch": 1.3109803921568628, "grad_norm": 1.9043453931808472, "learning_rate": 2.9088729646302174e-07, "loss": 0.1221, "step": 66860 }, { "epoch": 1.3111764705882352, "grad_norm": 2.830833911895752, "learning_rate": 2.9074006057409667e-07, "loss": 0.1315, "step": 66870 }, { "epoch": 1.311372549019608, "grad_norm": 2.2472593784332275, "learning_rate": 2.9059284668089666e-07, "loss": 0.1345, "step": 66880 }, { "epoch": 1.3115686274509804, "grad_norm": 2.1675331592559814, "learning_rate": 2.904456547988955e-07, "loss": 0.1152, "step": 66890 }, { "epoch": 1.311764705882353, "grad_norm": 2.4695398807525635, "learning_rate": 2.902984849435651e-07, "loss": 0.1246, "step": 66900 }, { "epoch": 1.3119607843137255, "grad_norm": 7.607154369354248, "learning_rate": 2.9015133713037475e-07, "loss": 0.1208, "step": 66910 }, { "epoch": 1.312156862745098, "grad_norm": 2.572021245956421, "learning_rate": 2.900042113747911e-07, "loss": 0.1057, "step": 66920 }, { "epoch": 1.3123529411764707, "grad_norm": 1.7273743152618408, "learning_rate": 2.8985710769227936e-07, "loss": 0.1384, "step": 66930 }, { "epoch": 1.312549019607843, "grad_norm": 2.0420660972595215, "learning_rate": 2.897100260983015e-07, "loss": 0.1088, "step": 66940 }, { "epoch": 1.3127450980392157, "grad_norm": 1.6035667657852173, "learning_rate": 2.8956296660831754e-07, "loss": 0.1018, "step": 66950 }, { "epoch": 1.3129411764705883, "grad_norm": 3.314302682876587, "learning_rate": 2.8941592923778535e-07, "loss": 0.1385, "step": 66960 }, { "epoch": 1.3131372549019606, "grad_norm": 1.7522486448287964, "learning_rate": 2.8926891400216044e-07, "loss": 0.135, "step": 66970 }, { "epoch": 1.3133333333333335, "grad_norm": 3.0983023643493652, "learning_rate": 2.891219209168957e-07, "loss": 0.1221, "step": 66980 }, { "epoch": 1.3135294117647058, "grad_norm": 2.59366774559021, "learning_rate": 2.8897494999744173e-07, "loss": 0.0947, "step": 66990 }, { "epoch": 1.3137254901960784, "grad_norm": 1.5243045091629028, "learning_rate": 2.888280012592472e-07, "loss": 0.1109, "step": 67000 }, { "epoch": 1.313921568627451, "grad_norm": 1.9616849422454834, "learning_rate": 2.886810747177582e-07, "loss": 0.1469, "step": 67010 }, { "epoch": 1.3141176470588236, "grad_norm": 4.6359357833862305, "learning_rate": 2.885341703884181e-07, "loss": 0.1419, "step": 67020 }, { "epoch": 1.3143137254901962, "grad_norm": 2.8564400672912598, "learning_rate": 2.883872882866687e-07, "loss": 0.1393, "step": 67030 }, { "epoch": 1.3145098039215686, "grad_norm": 2.161956787109375, "learning_rate": 2.8824042842794895e-07, "loss": 0.1439, "step": 67040 }, { "epoch": 1.3147058823529412, "grad_norm": 3.1774139404296875, "learning_rate": 2.880935908276954e-07, "loss": 0.1509, "step": 67050 }, { "epoch": 1.3149019607843138, "grad_norm": 3.9265780448913574, "learning_rate": 2.879467755013426e-07, "loss": 0.1143, "step": 67060 }, { "epoch": 1.3150980392156864, "grad_norm": 2.0504415035247803, "learning_rate": 2.8779998246432267e-07, "loss": 0.1636, "step": 67070 }, { "epoch": 1.315294117647059, "grad_norm": 2.3273000717163086, "learning_rate": 2.8765321173206507e-07, "loss": 0.1244, "step": 67080 }, { "epoch": 1.3154901960784313, "grad_norm": 3.1998796463012695, "learning_rate": 2.875064633199975e-07, "loss": 0.1364, "step": 67090 }, { "epoch": 1.315686274509804, "grad_norm": 3.3115642070770264, "learning_rate": 2.873597372435448e-07, "loss": 0.1457, "step": 67100 }, { "epoch": 1.3158823529411765, "grad_norm": 4.135552883148193, "learning_rate": 2.8721303351812935e-07, "loss": 0.1439, "step": 67110 }, { "epoch": 1.316078431372549, "grad_norm": 1.9410219192504883, "learning_rate": 2.87066352159172e-07, "loss": 0.1337, "step": 67120 }, { "epoch": 1.3162745098039217, "grad_norm": 2.4555318355560303, "learning_rate": 2.869196931820904e-07, "loss": 0.1507, "step": 67130 }, { "epoch": 1.316470588235294, "grad_norm": 3.112558603286743, "learning_rate": 2.8677305660230013e-07, "loss": 0.146, "step": 67140 }, { "epoch": 1.3166666666666667, "grad_norm": 1.6651315689086914, "learning_rate": 2.866264424352145e-07, "loss": 0.133, "step": 67150 }, { "epoch": 1.3168627450980392, "grad_norm": 3.034463882446289, "learning_rate": 2.864798506962447e-07, "loss": 0.1324, "step": 67160 }, { "epoch": 1.3170588235294118, "grad_norm": 5.151520729064941, "learning_rate": 2.8633328140079893e-07, "loss": 0.1196, "step": 67170 }, { "epoch": 1.3172549019607844, "grad_norm": 1.1436784267425537, "learning_rate": 2.861867345642834e-07, "loss": 0.126, "step": 67180 }, { "epoch": 1.3174509803921568, "grad_norm": 1.3420213460922241, "learning_rate": 2.860402102021022e-07, "loss": 0.1597, "step": 67190 }, { "epoch": 1.3176470588235294, "grad_norm": 2.183882713317871, "learning_rate": 2.858937083296565e-07, "loss": 0.1046, "step": 67200 }, { "epoch": 1.317843137254902, "grad_norm": 2.4254097938537598, "learning_rate": 2.857472289623455e-07, "loss": 0.1313, "step": 67210 }, { "epoch": 1.3180392156862746, "grad_norm": 3.622946262359619, "learning_rate": 2.8560077211556606e-07, "loss": 0.1267, "step": 67220 }, { "epoch": 1.3182352941176472, "grad_norm": 3.0357186794281006, "learning_rate": 2.854543378047124e-07, "loss": 0.1095, "step": 67230 }, { "epoch": 1.3184313725490195, "grad_norm": 2.352710008621216, "learning_rate": 2.8530792604517645e-07, "loss": 0.1144, "step": 67240 }, { "epoch": 1.3186274509803921, "grad_norm": 4.278421878814697, "learning_rate": 2.8516153685234814e-07, "loss": 0.1161, "step": 67250 }, { "epoch": 1.3188235294117647, "grad_norm": 1.6396925449371338, "learning_rate": 2.850151702416145e-07, "loss": 0.1202, "step": 67260 }, { "epoch": 1.3190196078431373, "grad_norm": 4.181994438171387, "learning_rate": 2.848688262283602e-07, "loss": 0.1304, "step": 67270 }, { "epoch": 1.31921568627451, "grad_norm": 3.246225595474243, "learning_rate": 2.847225048279683e-07, "loss": 0.1211, "step": 67280 }, { "epoch": 1.3194117647058823, "grad_norm": 1.371044397354126, "learning_rate": 2.8457620605581854e-07, "loss": 0.1144, "step": 67290 }, { "epoch": 1.3196078431372549, "grad_norm": 1.9213268756866455, "learning_rate": 2.8442992992728865e-07, "loss": 0.1104, "step": 67300 }, { "epoch": 1.3198039215686275, "grad_norm": 2.840080738067627, "learning_rate": 2.842836764577543e-07, "loss": 0.1383, "step": 67310 }, { "epoch": 1.32, "grad_norm": 1.2784991264343262, "learning_rate": 2.8413744566258803e-07, "loss": 0.1227, "step": 67320 }, { "epoch": 1.3201960784313727, "grad_norm": 2.431790590286255, "learning_rate": 2.8399123755716094e-07, "loss": 0.1191, "step": 67330 }, { "epoch": 1.320392156862745, "grad_norm": 3.5659565925598145, "learning_rate": 2.8384505215684084e-07, "loss": 0.1537, "step": 67340 }, { "epoch": 1.3205882352941176, "grad_norm": 2.060422658920288, "learning_rate": 2.836988894769939e-07, "loss": 0.1378, "step": 67350 }, { "epoch": 1.3207843137254902, "grad_norm": 4.201202869415283, "learning_rate": 2.835527495329835e-07, "loss": 0.1388, "step": 67360 }, { "epoch": 1.3209803921568628, "grad_norm": 3.2490766048431396, "learning_rate": 2.8340663234017037e-07, "loss": 0.1382, "step": 67370 }, { "epoch": 1.3211764705882354, "grad_norm": 2.928020477294922, "learning_rate": 2.832605379139137e-07, "loss": 0.131, "step": 67380 }, { "epoch": 1.3213725490196078, "grad_norm": 2.727191925048828, "learning_rate": 2.8311446626956935e-07, "loss": 0.1238, "step": 67390 }, { "epoch": 1.3215686274509804, "grad_norm": 1.1322013139724731, "learning_rate": 2.829684174224912e-07, "loss": 0.1126, "step": 67400 }, { "epoch": 1.321764705882353, "grad_norm": 2.052996873855591, "learning_rate": 2.828223913880311e-07, "loss": 0.1209, "step": 67410 }, { "epoch": 1.3219607843137255, "grad_norm": 1.2230697870254517, "learning_rate": 2.8267638818153786e-07, "loss": 0.141, "step": 67420 }, { "epoch": 1.3221568627450981, "grad_norm": 3.1427619457244873, "learning_rate": 2.82530407818358e-07, "loss": 0.115, "step": 67430 }, { "epoch": 1.3223529411764705, "grad_norm": 2.2790446281433105, "learning_rate": 2.823844503138363e-07, "loss": 0.1075, "step": 67440 }, { "epoch": 1.322549019607843, "grad_norm": 1.8045698404312134, "learning_rate": 2.8223851568331427e-07, "loss": 0.1206, "step": 67450 }, { "epoch": 1.3227450980392157, "grad_norm": 1.3936045169830322, "learning_rate": 2.820926039421313e-07, "loss": 0.0995, "step": 67460 }, { "epoch": 1.3229411764705883, "grad_norm": 2.6645750999450684, "learning_rate": 2.819467151056248e-07, "loss": 0.1377, "step": 67470 }, { "epoch": 1.3231372549019609, "grad_norm": 4.237121105194092, "learning_rate": 2.818008491891293e-07, "loss": 0.1422, "step": 67480 }, { "epoch": 1.3233333333333333, "grad_norm": 2.87370228767395, "learning_rate": 2.8165500620797675e-07, "loss": 0.1404, "step": 67490 }, { "epoch": 1.3235294117647058, "grad_norm": 3.795165777206421, "learning_rate": 2.815091861774973e-07, "loss": 0.1544, "step": 67500 }, { "epoch": 1.3237254901960784, "grad_norm": 5.244288444519043, "learning_rate": 2.8136338911301853e-07, "loss": 0.1366, "step": 67510 }, { "epoch": 1.323921568627451, "grad_norm": 2.4085323810577393, "learning_rate": 2.812176150298653e-07, "loss": 0.1212, "step": 67520 }, { "epoch": 1.3241176470588236, "grad_norm": 4.630978584289551, "learning_rate": 2.8107186394335997e-07, "loss": 0.1244, "step": 67530 }, { "epoch": 1.324313725490196, "grad_norm": 1.3033649921417236, "learning_rate": 2.8092613586882316e-07, "loss": 0.1146, "step": 67540 }, { "epoch": 1.3245098039215686, "grad_norm": 2.0154404640197754, "learning_rate": 2.8078043082157233e-07, "loss": 0.1252, "step": 67550 }, { "epoch": 1.3247058823529412, "grad_norm": 3.0737555027008057, "learning_rate": 2.806347488169228e-07, "loss": 0.1197, "step": 67560 }, { "epoch": 1.3249019607843138, "grad_norm": 1.356984257698059, "learning_rate": 2.8048908987018773e-07, "loss": 0.1303, "step": 67570 }, { "epoch": 1.3250980392156864, "grad_norm": 3.3547489643096924, "learning_rate": 2.8034345399667753e-07, "loss": 0.1185, "step": 67580 }, { "epoch": 1.3252941176470587, "grad_norm": 3.156801462173462, "learning_rate": 2.801978412117e-07, "loss": 0.1108, "step": 67590 }, { "epoch": 1.3254901960784313, "grad_norm": 1.689915657043457, "learning_rate": 2.800522515305612e-07, "loss": 0.1504, "step": 67600 }, { "epoch": 1.325686274509804, "grad_norm": 2.3596746921539307, "learning_rate": 2.7990668496856424e-07, "loss": 0.1085, "step": 67610 }, { "epoch": 1.3258823529411765, "grad_norm": 1.1942533254623413, "learning_rate": 2.797611415410096e-07, "loss": 0.1038, "step": 67620 }, { "epoch": 1.326078431372549, "grad_norm": 2.784607172012329, "learning_rate": 2.7961562126319607e-07, "loss": 0.139, "step": 67630 }, { "epoch": 1.3262745098039215, "grad_norm": 3.3445794582366943, "learning_rate": 2.7947012415041937e-07, "loss": 0.1218, "step": 67640 }, { "epoch": 1.326470588235294, "grad_norm": 2.087554693222046, "learning_rate": 2.793246502179728e-07, "loss": 0.1189, "step": 67650 }, { "epoch": 1.3266666666666667, "grad_norm": 2.583444118499756, "learning_rate": 2.7917919948114774e-07, "loss": 0.111, "step": 67660 }, { "epoch": 1.3268627450980393, "grad_norm": 3.182232141494751, "learning_rate": 2.7903377195523283e-07, "loss": 0.1329, "step": 67670 }, { "epoch": 1.3270588235294118, "grad_norm": 2.395320415496826, "learning_rate": 2.7888836765551375e-07, "loss": 0.1434, "step": 67680 }, { "epoch": 1.3272549019607842, "grad_norm": 2.975630760192871, "learning_rate": 2.787429865972746e-07, "loss": 0.1521, "step": 67690 }, { "epoch": 1.3274509803921568, "grad_norm": 2.675880193710327, "learning_rate": 2.7859762879579687e-07, "loss": 0.1418, "step": 67700 }, { "epoch": 1.3276470588235294, "grad_norm": 1.9757272005081177, "learning_rate": 2.784522942663591e-07, "loss": 0.1509, "step": 67710 }, { "epoch": 1.327843137254902, "grad_norm": 1.294633150100708, "learning_rate": 2.783069830242376e-07, "loss": 0.1086, "step": 67720 }, { "epoch": 1.3280392156862746, "grad_norm": 1.7843101024627686, "learning_rate": 2.781616950847068e-07, "loss": 0.131, "step": 67730 }, { "epoch": 1.328235294117647, "grad_norm": 3.6852405071258545, "learning_rate": 2.7801643046303767e-07, "loss": 0.1406, "step": 67740 }, { "epoch": 1.3284313725490196, "grad_norm": 2.3055267333984375, "learning_rate": 2.7787118917449944e-07, "loss": 0.1573, "step": 67750 }, { "epoch": 1.3286274509803921, "grad_norm": 3.605100154876709, "learning_rate": 2.7772597123435886e-07, "loss": 0.1126, "step": 67760 }, { "epoch": 1.3288235294117647, "grad_norm": 2.901970148086548, "learning_rate": 2.7758077665788005e-07, "loss": 0.1607, "step": 67770 }, { "epoch": 1.3290196078431373, "grad_norm": 1.7739229202270508, "learning_rate": 2.774356054603243e-07, "loss": 0.1148, "step": 67780 }, { "epoch": 1.3292156862745097, "grad_norm": 3.2680563926696777, "learning_rate": 2.772904576569514e-07, "loss": 0.1246, "step": 67790 }, { "epoch": 1.3294117647058823, "grad_norm": 2.8213095664978027, "learning_rate": 2.771453332630179e-07, "loss": 0.1246, "step": 67800 }, { "epoch": 1.3296078431372549, "grad_norm": 2.8718619346618652, "learning_rate": 2.7700023229377776e-07, "loss": 0.1233, "step": 67810 }, { "epoch": 1.3298039215686275, "grad_norm": 2.8864340782165527, "learning_rate": 2.7685515476448345e-07, "loss": 0.1236, "step": 67820 }, { "epoch": 1.33, "grad_norm": 3.0922129154205322, "learning_rate": 2.7671010069038404e-07, "loss": 0.1473, "step": 67830 }, { "epoch": 1.3301960784313724, "grad_norm": 2.650519371032715, "learning_rate": 2.765650700867262e-07, "loss": 0.1086, "step": 67840 }, { "epoch": 1.330392156862745, "grad_norm": 1.3945661783218384, "learning_rate": 2.7642006296875497e-07, "loss": 0.123, "step": 67850 }, { "epoch": 1.3305882352941176, "grad_norm": 2.6849470138549805, "learning_rate": 2.7627507935171177e-07, "loss": 0.1391, "step": 67860 }, { "epoch": 1.3307843137254902, "grad_norm": 3.197854995727539, "learning_rate": 2.761301192508365e-07, "loss": 0.1225, "step": 67870 }, { "epoch": 1.3309803921568628, "grad_norm": 5.966024875640869, "learning_rate": 2.7598518268136597e-07, "loss": 0.1318, "step": 67880 }, { "epoch": 1.3311764705882352, "grad_norm": 3.925966501235962, "learning_rate": 2.7584026965853496e-07, "loss": 0.1431, "step": 67890 }, { "epoch": 1.3313725490196078, "grad_norm": 1.9161927700042725, "learning_rate": 2.7569538019757545e-07, "loss": 0.115, "step": 67900 }, { "epoch": 1.3315686274509804, "grad_norm": 1.5504469871520996, "learning_rate": 2.755505143137168e-07, "loss": 0.1183, "step": 67910 }, { "epoch": 1.331764705882353, "grad_norm": 1.7363132238388062, "learning_rate": 2.754056720221867e-07, "loss": 0.1472, "step": 67920 }, { "epoch": 1.3319607843137256, "grad_norm": 2.266479969024658, "learning_rate": 2.7526085333820947e-07, "loss": 0.1222, "step": 67930 }, { "epoch": 1.332156862745098, "grad_norm": 3.3084182739257812, "learning_rate": 2.751160582770071e-07, "loss": 0.1177, "step": 67940 }, { "epoch": 1.3323529411764705, "grad_norm": 1.5256749391555786, "learning_rate": 2.7497128685379967e-07, "loss": 0.1141, "step": 67950 }, { "epoch": 1.332549019607843, "grad_norm": 2.047232151031494, "learning_rate": 2.7482653908380425e-07, "loss": 0.1158, "step": 67960 }, { "epoch": 1.3327450980392157, "grad_norm": 2.013540267944336, "learning_rate": 2.7468181498223543e-07, "loss": 0.1345, "step": 67970 }, { "epoch": 1.3329411764705883, "grad_norm": 1.6895416975021362, "learning_rate": 2.7453711456430567e-07, "loss": 0.1315, "step": 67980 }, { "epoch": 1.3331372549019607, "grad_norm": 2.391927480697632, "learning_rate": 2.7439243784522465e-07, "loss": 0.1584, "step": 67990 }, { "epoch": 1.3333333333333333, "grad_norm": 3.031461477279663, "learning_rate": 2.742477848401993e-07, "loss": 0.167, "step": 68000 }, { "epoch": 1.3335294117647059, "grad_norm": 1.9473506212234497, "learning_rate": 2.7410315556443497e-07, "loss": 0.1239, "step": 68010 }, { "epoch": 1.3337254901960784, "grad_norm": 1.046643853187561, "learning_rate": 2.739585500331336e-07, "loss": 0.1028, "step": 68020 }, { "epoch": 1.333921568627451, "grad_norm": 4.351035118103027, "learning_rate": 2.738139682614947e-07, "loss": 0.1047, "step": 68030 }, { "epoch": 1.3341176470588234, "grad_norm": 1.9918813705444336, "learning_rate": 2.736694102647159e-07, "loss": 0.1325, "step": 68040 }, { "epoch": 1.334313725490196, "grad_norm": 2.2621052265167236, "learning_rate": 2.7352487605799215e-07, "loss": 0.134, "step": 68050 }, { "epoch": 1.3345098039215686, "grad_norm": 2.5933170318603516, "learning_rate": 2.7338036565651536e-07, "loss": 0.1215, "step": 68060 }, { "epoch": 1.3347058823529412, "grad_norm": 4.522106170654297, "learning_rate": 2.732358790754753e-07, "loss": 0.1324, "step": 68070 }, { "epoch": 1.3349019607843138, "grad_norm": 2.3543283939361572, "learning_rate": 2.730914163300595e-07, "loss": 0.1213, "step": 68080 }, { "epoch": 1.3350980392156861, "grad_norm": 10.11464786529541, "learning_rate": 2.729469774354527e-07, "loss": 0.145, "step": 68090 }, { "epoch": 1.3352941176470587, "grad_norm": 2.813593864440918, "learning_rate": 2.7280256240683674e-07, "loss": 0.1521, "step": 68100 }, { "epoch": 1.3354901960784313, "grad_norm": 1.8308833837509155, "learning_rate": 2.726581712593919e-07, "loss": 0.1312, "step": 68110 }, { "epoch": 1.335686274509804, "grad_norm": 16.18933868408203, "learning_rate": 2.725138040082953e-07, "loss": 0.1285, "step": 68120 }, { "epoch": 1.3358823529411765, "grad_norm": 1.514593243598938, "learning_rate": 2.723694606687212e-07, "loss": 0.1623, "step": 68130 }, { "epoch": 1.336078431372549, "grad_norm": 4.2191243171691895, "learning_rate": 2.722251412558424e-07, "loss": 0.1402, "step": 68140 }, { "epoch": 1.3362745098039217, "grad_norm": 1.6773431301116943, "learning_rate": 2.720808457848284e-07, "loss": 0.1538, "step": 68150 }, { "epoch": 1.336470588235294, "grad_norm": 2.2501413822174072, "learning_rate": 2.7193657427084614e-07, "loss": 0.1202, "step": 68160 }, { "epoch": 1.3366666666666667, "grad_norm": 1.9873749017715454, "learning_rate": 2.717923267290608e-07, "loss": 0.1419, "step": 68170 }, { "epoch": 1.3368627450980393, "grad_norm": 2.432435989379883, "learning_rate": 2.716481031746341e-07, "loss": 0.1264, "step": 68180 }, { "epoch": 1.3370588235294116, "grad_norm": 2.3431174755096436, "learning_rate": 2.715039036227256e-07, "loss": 0.148, "step": 68190 }, { "epoch": 1.3372549019607844, "grad_norm": 2.4348385334014893, "learning_rate": 2.7135972808849285e-07, "loss": 0.1464, "step": 68200 }, { "epoch": 1.3374509803921568, "grad_norm": 3.373335361480713, "learning_rate": 2.7121557658709003e-07, "loss": 0.1394, "step": 68210 }, { "epoch": 1.3376470588235294, "grad_norm": 2.102126359939575, "learning_rate": 2.710714491336692e-07, "loss": 0.0935, "step": 68220 }, { "epoch": 1.337843137254902, "grad_norm": 1.4494131803512573, "learning_rate": 2.7092734574337995e-07, "loss": 0.1224, "step": 68230 }, { "epoch": 1.3380392156862744, "grad_norm": 3.787358045578003, "learning_rate": 2.707832664313696e-07, "loss": 0.1298, "step": 68240 }, { "epoch": 1.3382352941176472, "grad_norm": 2.1247787475585938, "learning_rate": 2.7063921121278224e-07, "loss": 0.1333, "step": 68250 }, { "epoch": 1.3384313725490196, "grad_norm": 2.3981547355651855, "learning_rate": 2.7049518010275974e-07, "loss": 0.125, "step": 68260 }, { "epoch": 1.3386274509803922, "grad_norm": 1.155108094215393, "learning_rate": 2.703511731164418e-07, "loss": 0.106, "step": 68270 }, { "epoch": 1.3388235294117647, "grad_norm": 0.5585775375366211, "learning_rate": 2.702071902689651e-07, "loss": 0.1171, "step": 68280 }, { "epoch": 1.3390196078431371, "grad_norm": 1.595794439315796, "learning_rate": 2.7006323157546385e-07, "loss": 0.1103, "step": 68290 }, { "epoch": 1.33921568627451, "grad_norm": 1.6427786350250244, "learning_rate": 2.699192970510701e-07, "loss": 0.129, "step": 68300 }, { "epoch": 1.3394117647058823, "grad_norm": 2.493964910507202, "learning_rate": 2.69775386710913e-07, "loss": 0.1291, "step": 68310 }, { "epoch": 1.339607843137255, "grad_norm": 3.5505101680755615, "learning_rate": 2.69631500570119e-07, "loss": 0.1254, "step": 68320 }, { "epoch": 1.3398039215686275, "grad_norm": 3.1844232082366943, "learning_rate": 2.694876386438126e-07, "loss": 0.1277, "step": 68330 }, { "epoch": 1.34, "grad_norm": 3.7292675971984863, "learning_rate": 2.6934380094711536e-07, "loss": 0.1121, "step": 68340 }, { "epoch": 1.3401960784313727, "grad_norm": 1.7608839273452759, "learning_rate": 2.69199987495146e-07, "loss": 0.1345, "step": 68350 }, { "epoch": 1.340392156862745, "grad_norm": 1.1715667247772217, "learning_rate": 2.690561983030215e-07, "loss": 0.13, "step": 68360 }, { "epoch": 1.3405882352941176, "grad_norm": 1.4098877906799316, "learning_rate": 2.6891243338585556e-07, "loss": 0.1129, "step": 68370 }, { "epoch": 1.3407843137254902, "grad_norm": 2.219813346862793, "learning_rate": 2.687686927587596e-07, "loss": 0.1027, "step": 68380 }, { "epoch": 1.3409803921568628, "grad_norm": 5.347517013549805, "learning_rate": 2.6862497643684234e-07, "loss": 0.1155, "step": 68390 }, { "epoch": 1.3411764705882354, "grad_norm": 1.6070830821990967, "learning_rate": 2.684812844352106e-07, "loss": 0.0996, "step": 68400 }, { "epoch": 1.3413725490196078, "grad_norm": 1.3541038036346436, "learning_rate": 2.6833761676896776e-07, "loss": 0.1123, "step": 68410 }, { "epoch": 1.3415686274509804, "grad_norm": 2.3086156845092773, "learning_rate": 2.681939734532149e-07, "loss": 0.1105, "step": 68420 }, { "epoch": 1.341764705882353, "grad_norm": 3.267338752746582, "learning_rate": 2.6805035450305094e-07, "loss": 0.131, "step": 68430 }, { "epoch": 1.3419607843137256, "grad_norm": 2.7957189083099365, "learning_rate": 2.6790675993357194e-07, "loss": 0.1354, "step": 68440 }, { "epoch": 1.3421568627450982, "grad_norm": 4.0759596824646, "learning_rate": 2.67763189759871e-07, "loss": 0.1344, "step": 68450 }, { "epoch": 1.3423529411764705, "grad_norm": 3.347731351852417, "learning_rate": 2.676196439970395e-07, "loss": 0.1441, "step": 68460 }, { "epoch": 1.3425490196078431, "grad_norm": 3.0790722370147705, "learning_rate": 2.674761226601658e-07, "loss": 0.1408, "step": 68470 }, { "epoch": 1.3427450980392157, "grad_norm": 1.7141015529632568, "learning_rate": 2.6733262576433537e-07, "loss": 0.1161, "step": 68480 }, { "epoch": 1.3429411764705883, "grad_norm": 2.3742916584014893, "learning_rate": 2.671891533246318e-07, "loss": 0.1184, "step": 68490 }, { "epoch": 1.343137254901961, "grad_norm": 4.90061092376709, "learning_rate": 2.670457053561356e-07, "loss": 0.132, "step": 68500 }, { "epoch": 1.3433333333333333, "grad_norm": 3.90012788772583, "learning_rate": 2.6690228187392476e-07, "loss": 0.1289, "step": 68510 }, { "epoch": 1.3435294117647059, "grad_norm": 2.7914717197418213, "learning_rate": 2.667588828930751e-07, "loss": 0.1392, "step": 68520 }, { "epoch": 1.3437254901960785, "grad_norm": 4.370547771453857, "learning_rate": 2.6661550842865944e-07, "loss": 0.1282, "step": 68530 }, { "epoch": 1.343921568627451, "grad_norm": 2.7600064277648926, "learning_rate": 2.664721584957479e-07, "loss": 0.1502, "step": 68540 }, { "epoch": 1.3441176470588236, "grad_norm": 3.692075252532959, "learning_rate": 2.663288331094088e-07, "loss": 0.1313, "step": 68550 }, { "epoch": 1.344313725490196, "grad_norm": 3.370267629623413, "learning_rate": 2.6618553228470696e-07, "loss": 0.1097, "step": 68560 }, { "epoch": 1.3445098039215686, "grad_norm": 4.476211071014404, "learning_rate": 2.660422560367049e-07, "loss": 0.1221, "step": 68570 }, { "epoch": 1.3447058823529412, "grad_norm": 2.940943479537964, "learning_rate": 2.6589900438046285e-07, "loss": 0.148, "step": 68580 }, { "epoch": 1.3449019607843138, "grad_norm": 3.3931775093078613, "learning_rate": 2.657557773310386e-07, "loss": 0.1475, "step": 68590 }, { "epoch": 1.3450980392156864, "grad_norm": 2.6339516639709473, "learning_rate": 2.6561257490348666e-07, "loss": 0.1101, "step": 68600 }, { "epoch": 1.3452941176470588, "grad_norm": 2.5873217582702637, "learning_rate": 2.6546939711285914e-07, "loss": 0.0974, "step": 68610 }, { "epoch": 1.3454901960784313, "grad_norm": 1.958488941192627, "learning_rate": 2.6532624397420625e-07, "loss": 0.1373, "step": 68620 }, { "epoch": 1.345686274509804, "grad_norm": 2.619717597961426, "learning_rate": 2.651831155025748e-07, "loss": 0.1344, "step": 68630 }, { "epoch": 1.3458823529411765, "grad_norm": 2.969738245010376, "learning_rate": 2.650400117130092e-07, "loss": 0.1196, "step": 68640 }, { "epoch": 1.3460784313725491, "grad_norm": 2.7526516914367676, "learning_rate": 2.648969326205516e-07, "loss": 0.1486, "step": 68650 }, { "epoch": 1.3462745098039215, "grad_norm": 3.251469612121582, "learning_rate": 2.6475387824024133e-07, "loss": 0.1239, "step": 68660 }, { "epoch": 1.346470588235294, "grad_norm": 1.6474940776824951, "learning_rate": 2.646108485871148e-07, "loss": 0.1388, "step": 68670 }, { "epoch": 1.3466666666666667, "grad_norm": 2.269892930984497, "learning_rate": 2.6446784367620657e-07, "loss": 0.1359, "step": 68680 }, { "epoch": 1.3468627450980393, "grad_norm": 2.9167232513427734, "learning_rate": 2.643248635225479e-07, "loss": 0.1414, "step": 68690 }, { "epoch": 1.3470588235294119, "grad_norm": 2.056692123413086, "learning_rate": 2.641819081411677e-07, "loss": 0.1434, "step": 68700 }, { "epoch": 1.3472549019607842, "grad_norm": 2.177448272705078, "learning_rate": 2.6403897754709246e-07, "loss": 0.1255, "step": 68710 }, { "epoch": 1.3474509803921568, "grad_norm": 2.104692220687866, "learning_rate": 2.638960717553459e-07, "loss": 0.1441, "step": 68720 }, { "epoch": 1.3476470588235294, "grad_norm": 2.532369375228882, "learning_rate": 2.6375319078094875e-07, "loss": 0.1312, "step": 68730 }, { "epoch": 1.347843137254902, "grad_norm": 2.5819647312164307, "learning_rate": 2.6361033463892003e-07, "loss": 0.1246, "step": 68740 }, { "epoch": 1.3480392156862746, "grad_norm": 3.213031530380249, "learning_rate": 2.6346750334427523e-07, "loss": 0.1472, "step": 68750 }, { "epoch": 1.348235294117647, "grad_norm": 1.7606016397476196, "learning_rate": 2.6332469691202795e-07, "loss": 0.1482, "step": 68760 }, { "epoch": 1.3484313725490196, "grad_norm": 3.674974203109741, "learning_rate": 2.631819153571885e-07, "loss": 0.1143, "step": 68770 }, { "epoch": 1.3486274509803922, "grad_norm": 2.9950716495513916, "learning_rate": 2.630391586947654e-07, "loss": 0.1388, "step": 68780 }, { "epoch": 1.3488235294117648, "grad_norm": 1.1262487173080444, "learning_rate": 2.628964269397638e-07, "loss": 0.1178, "step": 68790 }, { "epoch": 1.3490196078431373, "grad_norm": 3.410203695297241, "learning_rate": 2.6275372010718634e-07, "loss": 0.1289, "step": 68800 }, { "epoch": 1.3492156862745097, "grad_norm": 2.9276185035705566, "learning_rate": 2.626110382120336e-07, "loss": 0.1371, "step": 68810 }, { "epoch": 1.3494117647058823, "grad_norm": 2.067986249923706, "learning_rate": 2.6246838126930295e-07, "loss": 0.1381, "step": 68820 }, { "epoch": 1.349607843137255, "grad_norm": 2.961428165435791, "learning_rate": 2.6232574929398913e-07, "loss": 0.1264, "step": 68830 }, { "epoch": 1.3498039215686275, "grad_norm": 1.1932933330535889, "learning_rate": 2.6218314230108495e-07, "loss": 0.1317, "step": 68840 }, { "epoch": 1.35, "grad_norm": 2.762763738632202, "learning_rate": 2.6204056030557984e-07, "loss": 0.1272, "step": 68850 }, { "epoch": 1.3501960784313725, "grad_norm": 2.182406187057495, "learning_rate": 2.618980033224607e-07, "loss": 0.1473, "step": 68860 }, { "epoch": 1.350392156862745, "grad_norm": 3.6628472805023193, "learning_rate": 2.617554713667124e-07, "loss": 0.147, "step": 68870 }, { "epoch": 1.3505882352941176, "grad_norm": 1.9866188764572144, "learning_rate": 2.616129644533165e-07, "loss": 0.1439, "step": 68880 }, { "epoch": 1.3507843137254902, "grad_norm": 1.625122308731079, "learning_rate": 2.61470482597252e-07, "loss": 0.0946, "step": 68890 }, { "epoch": 1.3509803921568628, "grad_norm": 2.80393123626709, "learning_rate": 2.613280258134959e-07, "loss": 0.1186, "step": 68900 }, { "epoch": 1.3511764705882352, "grad_norm": 1.5199049711227417, "learning_rate": 2.6118559411702187e-07, "loss": 0.1264, "step": 68910 }, { "epoch": 1.3513725490196078, "grad_norm": 1.0940629243850708, "learning_rate": 2.6104318752280094e-07, "loss": 0.134, "step": 68920 }, { "epoch": 1.3515686274509804, "grad_norm": 1.9522647857666016, "learning_rate": 2.60900806045802e-07, "loss": 0.1249, "step": 68930 }, { "epoch": 1.351764705882353, "grad_norm": 1.8510907888412476, "learning_rate": 2.6075844970099134e-07, "loss": 0.11, "step": 68940 }, { "epoch": 1.3519607843137256, "grad_norm": 2.1117944717407227, "learning_rate": 2.606161185033319e-07, "loss": 0.1276, "step": 68950 }, { "epoch": 1.352156862745098, "grad_norm": 1.5816234350204468, "learning_rate": 2.604738124677844e-07, "loss": 0.1359, "step": 68960 }, { "epoch": 1.3523529411764705, "grad_norm": 1.4550716876983643, "learning_rate": 2.603315316093072e-07, "loss": 0.1336, "step": 68970 }, { "epoch": 1.3525490196078431, "grad_norm": 2.97025465965271, "learning_rate": 2.6018927594285553e-07, "loss": 0.1363, "step": 68980 }, { "epoch": 1.3527450980392157, "grad_norm": 1.5257400274276733, "learning_rate": 2.60047045483382e-07, "loss": 0.1286, "step": 68990 }, { "epoch": 1.3529411764705883, "grad_norm": 2.4379796981811523, "learning_rate": 2.5990484024583704e-07, "loss": 0.1095, "step": 69000 }, { "epoch": 1.3531372549019607, "grad_norm": 3.4288034439086914, "learning_rate": 2.597626602451681e-07, "loss": 0.1302, "step": 69010 }, { "epoch": 1.3533333333333333, "grad_norm": 4.969789028167725, "learning_rate": 2.5962050549631967e-07, "loss": 0.0989, "step": 69020 }, { "epoch": 1.3535294117647059, "grad_norm": 4.513338565826416, "learning_rate": 2.5947837601423426e-07, "loss": 0.1429, "step": 69030 }, { "epoch": 1.3537254901960785, "grad_norm": 2.9915313720703125, "learning_rate": 2.593362718138513e-07, "loss": 0.1236, "step": 69040 }, { "epoch": 1.353921568627451, "grad_norm": 2.612976551055908, "learning_rate": 2.5919419291010746e-07, "loss": 0.1147, "step": 69050 }, { "epoch": 1.3541176470588234, "grad_norm": 3.1380972862243652, "learning_rate": 2.590521393179372e-07, "loss": 0.1609, "step": 69060 }, { "epoch": 1.354313725490196, "grad_norm": 1.0771485567092896, "learning_rate": 2.58910111052272e-07, "loss": 0.142, "step": 69070 }, { "epoch": 1.3545098039215686, "grad_norm": 2.682370662689209, "learning_rate": 2.587681081280404e-07, "loss": 0.117, "step": 69080 }, { "epoch": 1.3547058823529412, "grad_norm": 2.6494345664978027, "learning_rate": 2.586261305601691e-07, "loss": 0.1398, "step": 69090 }, { "epoch": 1.3549019607843138, "grad_norm": 1.7999545335769653, "learning_rate": 2.584841783635815e-07, "loss": 0.1242, "step": 69100 }, { "epoch": 1.3550980392156862, "grad_norm": 3.940898895263672, "learning_rate": 2.583422515531981e-07, "loss": 0.1204, "step": 69110 }, { "epoch": 1.3552941176470588, "grad_norm": 3.505293607711792, "learning_rate": 2.582003501439374e-07, "loss": 0.1416, "step": 69120 }, { "epoch": 1.3554901960784314, "grad_norm": 3.450986385345459, "learning_rate": 2.5805847415071516e-07, "loss": 0.1486, "step": 69130 }, { "epoch": 1.355686274509804, "grad_norm": 4.3198771476745605, "learning_rate": 2.5791662358844395e-07, "loss": 0.1394, "step": 69140 }, { "epoch": 1.3558823529411765, "grad_norm": 2.944305658340454, "learning_rate": 2.577747984720339e-07, "loss": 0.1229, "step": 69150 }, { "epoch": 1.356078431372549, "grad_norm": 7.806429386138916, "learning_rate": 2.5763299881639283e-07, "loss": 0.1394, "step": 69160 }, { "epoch": 1.3562745098039215, "grad_norm": 1.094624638557434, "learning_rate": 2.574912246364254e-07, "loss": 0.13, "step": 69170 }, { "epoch": 1.356470588235294, "grad_norm": 2.2684872150421143, "learning_rate": 2.573494759470336e-07, "loss": 0.1428, "step": 69180 }, { "epoch": 1.3566666666666667, "grad_norm": 3.3258330821990967, "learning_rate": 2.5720775276311735e-07, "loss": 0.1238, "step": 69190 }, { "epoch": 1.3568627450980393, "grad_norm": 2.4836924076080322, "learning_rate": 2.570660550995731e-07, "loss": 0.1169, "step": 69200 }, { "epoch": 1.3570588235294117, "grad_norm": 2.4419779777526855, "learning_rate": 2.569243829712949e-07, "loss": 0.1455, "step": 69210 }, { "epoch": 1.3572549019607842, "grad_norm": 4.138843536376953, "learning_rate": 2.567827363931746e-07, "loss": 0.1381, "step": 69220 }, { "epoch": 1.3574509803921568, "grad_norm": 3.0257771015167236, "learning_rate": 2.566411153801007e-07, "loss": 0.1229, "step": 69230 }, { "epoch": 1.3576470588235294, "grad_norm": 2.5710301399230957, "learning_rate": 2.5649951994695907e-07, "loss": 0.1468, "step": 69240 }, { "epoch": 1.357843137254902, "grad_norm": 4.139385223388672, "learning_rate": 2.563579501086335e-07, "loss": 0.1215, "step": 69250 }, { "epoch": 1.3580392156862744, "grad_norm": 2.099061965942383, "learning_rate": 2.5621640588000443e-07, "loss": 0.1356, "step": 69260 }, { "epoch": 1.358235294117647, "grad_norm": 2.598263740539551, "learning_rate": 2.560748872759497e-07, "loss": 0.1358, "step": 69270 }, { "epoch": 1.3584313725490196, "grad_norm": 2.7940163612365723, "learning_rate": 2.5593339431134503e-07, "loss": 0.1018, "step": 69280 }, { "epoch": 1.3586274509803922, "grad_norm": 2.5463390350341797, "learning_rate": 2.5579192700106254e-07, "loss": 0.1281, "step": 69290 }, { "epoch": 1.3588235294117648, "grad_norm": 3.5416553020477295, "learning_rate": 2.556504853599727e-07, "loss": 0.1086, "step": 69300 }, { "epoch": 1.3590196078431371, "grad_norm": 2.2967522144317627, "learning_rate": 2.555090694029421e-07, "loss": 0.1244, "step": 69310 }, { "epoch": 1.3592156862745097, "grad_norm": 2.728795289993286, "learning_rate": 2.553676791448358e-07, "loss": 0.0985, "step": 69320 }, { "epoch": 1.3594117647058823, "grad_norm": 2.3172032833099365, "learning_rate": 2.5522631460051535e-07, "loss": 0.1187, "step": 69330 }, { "epoch": 1.359607843137255, "grad_norm": 1.9894630908966064, "learning_rate": 2.550849757848397e-07, "loss": 0.1545, "step": 69340 }, { "epoch": 1.3598039215686275, "grad_norm": 1.679476261138916, "learning_rate": 2.549436627126656e-07, "loss": 0.1215, "step": 69350 }, { "epoch": 1.3599999999999999, "grad_norm": 1.3358216285705566, "learning_rate": 2.548023753988466e-07, "loss": 0.1051, "step": 69360 }, { "epoch": 1.3601960784313725, "grad_norm": 3.531365156173706, "learning_rate": 2.546611138582334e-07, "loss": 0.1114, "step": 69370 }, { "epoch": 1.360392156862745, "grad_norm": 2.083967924118042, "learning_rate": 2.5451987810567484e-07, "loss": 0.1648, "step": 69380 }, { "epoch": 1.3605882352941177, "grad_norm": 2.2510828971862793, "learning_rate": 2.543786681560162e-07, "loss": 0.0953, "step": 69390 }, { "epoch": 1.3607843137254902, "grad_norm": 2.4256067276000977, "learning_rate": 2.542374840241e-07, "loss": 0.1287, "step": 69400 }, { "epoch": 1.3609803921568626, "grad_norm": 1.5707581043243408, "learning_rate": 2.54096325724767e-07, "loss": 0.1109, "step": 69410 }, { "epoch": 1.3611764705882354, "grad_norm": 1.585902214050293, "learning_rate": 2.539551932728543e-07, "loss": 0.1453, "step": 69420 }, { "epoch": 1.3613725490196078, "grad_norm": 1.3416290283203125, "learning_rate": 2.5381408668319647e-07, "loss": 0.1365, "step": 69430 }, { "epoch": 1.3615686274509804, "grad_norm": 2.7949371337890625, "learning_rate": 2.5367300597062587e-07, "loss": 0.1188, "step": 69440 }, { "epoch": 1.361764705882353, "grad_norm": 2.1079726219177246, "learning_rate": 2.535319511499716e-07, "loss": 0.1365, "step": 69450 }, { "epoch": 1.3619607843137254, "grad_norm": 1.6016407012939453, "learning_rate": 2.5339092223605996e-07, "loss": 0.1324, "step": 69460 }, { "epoch": 1.3621568627450982, "grad_norm": 1.1290037631988525, "learning_rate": 2.53249919243715e-07, "loss": 0.1182, "step": 69470 }, { "epoch": 1.3623529411764705, "grad_norm": 1.4457032680511475, "learning_rate": 2.5310894218775803e-07, "loss": 0.1267, "step": 69480 }, { "epoch": 1.3625490196078431, "grad_norm": 2.68906569480896, "learning_rate": 2.5296799108300726e-07, "loss": 0.1297, "step": 69490 }, { "epoch": 1.3627450980392157, "grad_norm": 1.9258151054382324, "learning_rate": 2.528270659442781e-07, "loss": 0.1476, "step": 69500 }, { "epoch": 1.362941176470588, "grad_norm": 4.481664180755615, "learning_rate": 2.526861667863839e-07, "loss": 0.1337, "step": 69510 }, { "epoch": 1.363137254901961, "grad_norm": 3.0276103019714355, "learning_rate": 2.525452936241346e-07, "loss": 0.1341, "step": 69520 }, { "epoch": 1.3633333333333333, "grad_norm": 1.8377227783203125, "learning_rate": 2.5240444647233743e-07, "loss": 0.1547, "step": 69530 }, { "epoch": 1.3635294117647059, "grad_norm": 2.921194314956665, "learning_rate": 2.522636253457976e-07, "loss": 0.1288, "step": 69540 }, { "epoch": 1.3637254901960785, "grad_norm": 1.2416492700576782, "learning_rate": 2.521228302593169e-07, "loss": 0.1214, "step": 69550 }, { "epoch": 1.3639215686274508, "grad_norm": 1.8688774108886719, "learning_rate": 2.519820612276943e-07, "loss": 0.1156, "step": 69560 }, { "epoch": 1.3641176470588237, "grad_norm": 2.2569022178649902, "learning_rate": 2.5184131826572676e-07, "loss": 0.119, "step": 69570 }, { "epoch": 1.364313725490196, "grad_norm": 1.5525929927825928, "learning_rate": 2.517006013882078e-07, "loss": 0.1068, "step": 69580 }, { "epoch": 1.3645098039215686, "grad_norm": 2.315380573272705, "learning_rate": 2.515599106099283e-07, "loss": 0.1415, "step": 69590 }, { "epoch": 1.3647058823529412, "grad_norm": 2.523371458053589, "learning_rate": 2.51419245945677e-07, "loss": 0.1338, "step": 69600 }, { "epoch": 1.3649019607843138, "grad_norm": 1.6981974840164185, "learning_rate": 2.512786074102391e-07, "loss": 0.1103, "step": 69610 }, { "epoch": 1.3650980392156864, "grad_norm": 2.5114784240722656, "learning_rate": 2.511379950183973e-07, "loss": 0.1119, "step": 69620 }, { "epoch": 1.3652941176470588, "grad_norm": 1.4566632509231567, "learning_rate": 2.5099740878493196e-07, "loss": 0.1204, "step": 69630 }, { "epoch": 1.3654901960784314, "grad_norm": 2.7473526000976562, "learning_rate": 2.5085684872462034e-07, "loss": 0.1499, "step": 69640 }, { "epoch": 1.365686274509804, "grad_norm": 5.8272013664245605, "learning_rate": 2.5071631485223657e-07, "loss": 0.1312, "step": 69650 }, { "epoch": 1.3658823529411765, "grad_norm": 1.65742826461792, "learning_rate": 2.505758071825528e-07, "loss": 0.1048, "step": 69660 }, { "epoch": 1.3660784313725491, "grad_norm": 3.015510082244873, "learning_rate": 2.5043532573033816e-07, "loss": 0.1244, "step": 69670 }, { "epoch": 1.3662745098039215, "grad_norm": 2.8957042694091797, "learning_rate": 2.5029487051035883e-07, "loss": 0.1228, "step": 69680 }, { "epoch": 1.366470588235294, "grad_norm": 2.369859218597412, "learning_rate": 2.501544415373781e-07, "loss": 0.1579, "step": 69690 }, { "epoch": 1.3666666666666667, "grad_norm": 2.415562391281128, "learning_rate": 2.500140388261571e-07, "loss": 0.1123, "step": 69700 }, { "epoch": 1.3668627450980393, "grad_norm": 2.5210506916046143, "learning_rate": 2.498736623914537e-07, "loss": 0.1216, "step": 69710 }, { "epoch": 1.3670588235294119, "grad_norm": 3.261364698410034, "learning_rate": 2.497333122480229e-07, "loss": 0.1483, "step": 69720 }, { "epoch": 1.3672549019607843, "grad_norm": 2.533245086669922, "learning_rate": 2.495929884106176e-07, "loss": 0.1542, "step": 69730 }, { "epoch": 1.3674509803921568, "grad_norm": 4.533980369567871, "learning_rate": 2.4945269089398723e-07, "loss": 0.1223, "step": 69740 }, { "epoch": 1.3676470588235294, "grad_norm": 2.8483052253723145, "learning_rate": 2.493124197128787e-07, "loss": 0.0989, "step": 69750 }, { "epoch": 1.367843137254902, "grad_norm": 2.0523200035095215, "learning_rate": 2.4917217488203653e-07, "loss": 0.1224, "step": 69760 }, { "epoch": 1.3680392156862746, "grad_norm": 1.9233890771865845, "learning_rate": 2.4903195641620176e-07, "loss": 0.094, "step": 69770 }, { "epoch": 1.368235294117647, "grad_norm": 2.753603935241699, "learning_rate": 2.488917643301131e-07, "loss": 0.1519, "step": 69780 }, { "epoch": 1.3684313725490196, "grad_norm": 2.5760984420776367, "learning_rate": 2.487515986385067e-07, "loss": 0.1433, "step": 69790 }, { "epoch": 1.3686274509803922, "grad_norm": 1.2147246599197388, "learning_rate": 2.4861145935611536e-07, "loss": 0.1277, "step": 69800 }, { "epoch": 1.3688235294117648, "grad_norm": 2.6055221557617188, "learning_rate": 2.4847134649766935e-07, "loss": 0.1152, "step": 69810 }, { "epoch": 1.3690196078431374, "grad_norm": 1.6876415014266968, "learning_rate": 2.483312600778965e-07, "loss": 0.1044, "step": 69820 }, { "epoch": 1.3692156862745097, "grad_norm": 2.4782581329345703, "learning_rate": 2.4819120011152127e-07, "loss": 0.1092, "step": 69830 }, { "epoch": 1.3694117647058823, "grad_norm": 3.59883975982666, "learning_rate": 2.4805116661326593e-07, "loss": 0.1254, "step": 69840 }, { "epoch": 1.369607843137255, "grad_norm": 2.144848346710205, "learning_rate": 2.479111595978494e-07, "loss": 0.1357, "step": 69850 }, { "epoch": 1.3698039215686275, "grad_norm": 2.1203715801239014, "learning_rate": 2.477711790799883e-07, "loss": 0.1019, "step": 69860 }, { "epoch": 1.37, "grad_norm": 2.1913015842437744, "learning_rate": 2.476312250743962e-07, "loss": 0.1275, "step": 69870 }, { "epoch": 1.3701960784313725, "grad_norm": 2.899608850479126, "learning_rate": 2.474912975957838e-07, "loss": 0.1261, "step": 69880 }, { "epoch": 1.370392156862745, "grad_norm": 3.082332134246826, "learning_rate": 2.473513966588594e-07, "loss": 0.1052, "step": 69890 }, { "epoch": 1.3705882352941177, "grad_norm": 4.059977054595947, "learning_rate": 2.4721152227832816e-07, "loss": 0.124, "step": 69900 }, { "epoch": 1.3707843137254903, "grad_norm": 2.107170820236206, "learning_rate": 2.4707167446889233e-07, "loss": 0.1039, "step": 69910 }, { "epoch": 1.3709803921568628, "grad_norm": 2.2051687240600586, "learning_rate": 2.46931853245252e-07, "loss": 0.1331, "step": 69920 }, { "epoch": 1.3711764705882352, "grad_norm": 2.494457244873047, "learning_rate": 2.4679205862210383e-07, "loss": 0.1329, "step": 69930 }, { "epoch": 1.3713725490196078, "grad_norm": 2.7158162593841553, "learning_rate": 2.466522906141417e-07, "loss": 0.1284, "step": 69940 }, { "epoch": 1.3715686274509804, "grad_norm": 3.7401235103607178, "learning_rate": 2.4651254923605737e-07, "loss": 0.1182, "step": 69950 }, { "epoch": 1.371764705882353, "grad_norm": 2.9363322257995605, "learning_rate": 2.4637283450253905e-07, "loss": 0.1436, "step": 69960 }, { "epoch": 1.3719607843137256, "grad_norm": 2.1006898880004883, "learning_rate": 2.462331464282723e-07, "loss": 0.1171, "step": 69970 }, { "epoch": 1.372156862745098, "grad_norm": 2.8534862995147705, "learning_rate": 2.460934850279403e-07, "loss": 0.1473, "step": 69980 }, { "epoch": 1.3723529411764706, "grad_norm": 3.91079044342041, "learning_rate": 2.4595385031622307e-07, "loss": 0.1679, "step": 69990 }, { "epoch": 1.3725490196078431, "grad_norm": 2.099531412124634, "learning_rate": 2.458142423077976e-07, "loss": 0.1398, "step": 70000 }, { "epoch": 1.3727450980392157, "grad_norm": 1.5693203210830688, "learning_rate": 2.456746610173387e-07, "loss": 0.1438, "step": 70010 }, { "epoch": 1.3729411764705883, "grad_norm": 2.8493807315826416, "learning_rate": 2.45535106459518e-07, "loss": 0.1281, "step": 70020 }, { "epoch": 1.3731372549019607, "grad_norm": 1.5731481313705444, "learning_rate": 2.4539557864900433e-07, "loss": 0.1425, "step": 70030 }, { "epoch": 1.3733333333333333, "grad_norm": 1.3346749544143677, "learning_rate": 2.452560776004635e-07, "loss": 0.1049, "step": 70040 }, { "epoch": 1.3735294117647059, "grad_norm": 1.4152575731277466, "learning_rate": 2.451166033285591e-07, "loss": 0.1181, "step": 70050 }, { "epoch": 1.3737254901960785, "grad_norm": 3.861933708190918, "learning_rate": 2.4497715584795136e-07, "loss": 0.1471, "step": 70060 }, { "epoch": 1.373921568627451, "grad_norm": 1.333160400390625, "learning_rate": 2.4483773517329776e-07, "loss": 0.1102, "step": 70070 }, { "epoch": 1.3741176470588234, "grad_norm": 1.832698941230774, "learning_rate": 2.446983413192533e-07, "loss": 0.1063, "step": 70080 }, { "epoch": 1.374313725490196, "grad_norm": 2.9307117462158203, "learning_rate": 2.4455897430046995e-07, "loss": 0.1286, "step": 70090 }, { "epoch": 1.3745098039215686, "grad_norm": 4.910470008850098, "learning_rate": 2.4441963413159653e-07, "loss": 0.1551, "step": 70100 }, { "epoch": 1.3747058823529412, "grad_norm": 2.2724320888519287, "learning_rate": 2.4428032082727983e-07, "loss": 0.1408, "step": 70110 }, { "epoch": 1.3749019607843138, "grad_norm": 2.125953435897827, "learning_rate": 2.4414103440216304e-07, "loss": 0.1103, "step": 70120 }, { "epoch": 1.3750980392156862, "grad_norm": 2.2304999828338623, "learning_rate": 2.4400177487088676e-07, "loss": 0.1481, "step": 70130 }, { "epoch": 1.3752941176470588, "grad_norm": 4.0316338539123535, "learning_rate": 2.4386254224808915e-07, "loss": 0.1431, "step": 70140 }, { "epoch": 1.3754901960784314, "grad_norm": 4.7366414070129395, "learning_rate": 2.4372333654840507e-07, "loss": 0.1353, "step": 70150 }, { "epoch": 1.375686274509804, "grad_norm": 3.814838409423828, "learning_rate": 2.435841577864664e-07, "loss": 0.115, "step": 70160 }, { "epoch": 1.3758823529411766, "grad_norm": 3.5241005420684814, "learning_rate": 2.43445005976903e-07, "loss": 0.1573, "step": 70170 }, { "epoch": 1.376078431372549, "grad_norm": 2.191573143005371, "learning_rate": 2.4330588113434103e-07, "loss": 0.1332, "step": 70180 }, { "epoch": 1.3762745098039215, "grad_norm": 3.2222306728363037, "learning_rate": 2.431667832734045e-07, "loss": 0.1155, "step": 70190 }, { "epoch": 1.3764705882352941, "grad_norm": 1.9151208400726318, "learning_rate": 2.430277124087138e-07, "loss": 0.1565, "step": 70200 }, { "epoch": 1.3766666666666667, "grad_norm": 2.2827136516571045, "learning_rate": 2.4288866855488754e-07, "loss": 0.1179, "step": 70210 }, { "epoch": 1.3768627450980393, "grad_norm": 2.383894205093384, "learning_rate": 2.427496517265405e-07, "loss": 0.1424, "step": 70220 }, { "epoch": 1.3770588235294117, "grad_norm": 3.002809762954712, "learning_rate": 2.426106619382849e-07, "loss": 0.1305, "step": 70230 }, { "epoch": 1.3772549019607843, "grad_norm": 2.147054672241211, "learning_rate": 2.424716992047306e-07, "loss": 0.1553, "step": 70240 }, { "epoch": 1.3774509803921569, "grad_norm": 1.616451382637024, "learning_rate": 2.423327635404841e-07, "loss": 0.1153, "step": 70250 }, { "epoch": 1.3776470588235294, "grad_norm": 1.3258620500564575, "learning_rate": 2.4219385496014893e-07, "loss": 0.1568, "step": 70260 }, { "epoch": 1.377843137254902, "grad_norm": 1.4683327674865723, "learning_rate": 2.420549734783265e-07, "loss": 0.1456, "step": 70270 }, { "epoch": 1.3780392156862744, "grad_norm": 4.081460952758789, "learning_rate": 2.419161191096147e-07, "loss": 0.0947, "step": 70280 }, { "epoch": 1.378235294117647, "grad_norm": 2.5224623680114746, "learning_rate": 2.417772918686085e-07, "loss": 0.1395, "step": 70290 }, { "epoch": 1.3784313725490196, "grad_norm": 1.7961852550506592, "learning_rate": 2.416384917699009e-07, "loss": 0.1342, "step": 70300 }, { "epoch": 1.3786274509803922, "grad_norm": 4.262200832366943, "learning_rate": 2.414997188280811e-07, "loss": 0.1174, "step": 70310 }, { "epoch": 1.3788235294117648, "grad_norm": 1.6346176862716675, "learning_rate": 2.4136097305773564e-07, "loss": 0.1177, "step": 70320 }, { "epoch": 1.3790196078431372, "grad_norm": 2.8583712577819824, "learning_rate": 2.4122225447344876e-07, "loss": 0.1338, "step": 70330 }, { "epoch": 1.3792156862745097, "grad_norm": 2.1072142124176025, "learning_rate": 2.410835630898012e-07, "loss": 0.1435, "step": 70340 }, { "epoch": 1.3794117647058823, "grad_norm": 4.6023783683776855, "learning_rate": 2.40944898921371e-07, "loss": 0.1261, "step": 70350 }, { "epoch": 1.379607843137255, "grad_norm": 1.4846529960632324, "learning_rate": 2.4080626198273354e-07, "loss": 0.1506, "step": 70360 }, { "epoch": 1.3798039215686275, "grad_norm": 2.7778477668762207, "learning_rate": 2.4066765228846136e-07, "loss": 0.1333, "step": 70370 }, { "epoch": 1.38, "grad_norm": 2.166180372238159, "learning_rate": 2.405290698531239e-07, "loss": 0.146, "step": 70380 }, { "epoch": 1.3801960784313725, "grad_norm": 1.684664011001587, "learning_rate": 2.403905146912877e-07, "loss": 0.1268, "step": 70390 }, { "epoch": 1.380392156862745, "grad_norm": 3.9760665893554688, "learning_rate": 2.402519868175168e-07, "loss": 0.1492, "step": 70400 }, { "epoch": 1.3805882352941177, "grad_norm": 3.818941354751587, "learning_rate": 2.40113486246372e-07, "loss": 0.121, "step": 70410 }, { "epoch": 1.3807843137254903, "grad_norm": 2.431816339492798, "learning_rate": 2.3997501299241124e-07, "loss": 0.1227, "step": 70420 }, { "epoch": 1.3809803921568626, "grad_norm": 1.719299077987671, "learning_rate": 2.3983656707019e-07, "loss": 0.1351, "step": 70430 }, { "epoch": 1.3811764705882352, "grad_norm": 2.2389743328094482, "learning_rate": 2.3969814849426047e-07, "loss": 0.1466, "step": 70440 }, { "epoch": 1.3813725490196078, "grad_norm": 2.8455615043640137, "learning_rate": 2.395597572791719e-07, "loss": 0.1264, "step": 70450 }, { "epoch": 1.3815686274509804, "grad_norm": 1.9212849140167236, "learning_rate": 2.3942139343947127e-07, "loss": 0.1245, "step": 70460 }, { "epoch": 1.381764705882353, "grad_norm": 3.3413467407226562, "learning_rate": 2.39283056989702e-07, "loss": 0.1331, "step": 70470 }, { "epoch": 1.3819607843137254, "grad_norm": 2.33007550239563, "learning_rate": 2.3914474794440486e-07, "loss": 0.1207, "step": 70480 }, { "epoch": 1.382156862745098, "grad_norm": 1.7083933353424072, "learning_rate": 2.3900646631811806e-07, "loss": 0.1321, "step": 70490 }, { "epoch": 1.3823529411764706, "grad_norm": 3.763195514678955, "learning_rate": 2.3886821212537643e-07, "loss": 0.1652, "step": 70500 }, { "epoch": 1.3825490196078432, "grad_norm": 2.345783233642578, "learning_rate": 2.3872998538071205e-07, "loss": 0.1067, "step": 70510 }, { "epoch": 1.3827450980392157, "grad_norm": 1.936809778213501, "learning_rate": 2.3859178609865455e-07, "loss": 0.119, "step": 70520 }, { "epoch": 1.3829411764705881, "grad_norm": 3.0618326663970947, "learning_rate": 2.384536142937301e-07, "loss": 0.1473, "step": 70530 }, { "epoch": 1.3831372549019607, "grad_norm": 4.682638645172119, "learning_rate": 2.3831546998046214e-07, "loss": 0.1342, "step": 70540 }, { "epoch": 1.3833333333333333, "grad_norm": 2.2219150066375732, "learning_rate": 2.3817735317337135e-07, "loss": 0.138, "step": 70550 }, { "epoch": 1.383529411764706, "grad_norm": 2.042125940322876, "learning_rate": 2.3803926388697564e-07, "loss": 0.1048, "step": 70560 }, { "epoch": 1.3837254901960785, "grad_norm": 1.693735122680664, "learning_rate": 2.3790120213578978e-07, "loss": 0.1394, "step": 70570 }, { "epoch": 1.3839215686274509, "grad_norm": 1.9643558263778687, "learning_rate": 2.3776316793432544e-07, "loss": 0.1084, "step": 70580 }, { "epoch": 1.3841176470588235, "grad_norm": 2.3197736740112305, "learning_rate": 2.3762516129709203e-07, "loss": 0.1382, "step": 70590 }, { "epoch": 1.384313725490196, "grad_norm": 3.2154133319854736, "learning_rate": 2.374871822385956e-07, "loss": 0.1362, "step": 70600 }, { "epoch": 1.3845098039215686, "grad_norm": 2.3293280601501465, "learning_rate": 2.3734923077333913e-07, "loss": 0.1137, "step": 70610 }, { "epoch": 1.3847058823529412, "grad_norm": 2.111626625061035, "learning_rate": 2.3721130691582342e-07, "loss": 0.1358, "step": 70620 }, { "epoch": 1.3849019607843136, "grad_norm": 4.120390892028809, "learning_rate": 2.370734106805456e-07, "loss": 0.1184, "step": 70630 }, { "epoch": 1.3850980392156862, "grad_norm": 1.9422860145568848, "learning_rate": 2.3693554208200018e-07, "loss": 0.1248, "step": 70640 }, { "epoch": 1.3852941176470588, "grad_norm": 3.9316885471343994, "learning_rate": 2.3679770113467906e-07, "loss": 0.1202, "step": 70650 }, { "epoch": 1.3854901960784314, "grad_norm": 5.522014141082764, "learning_rate": 2.366598878530709e-07, "loss": 0.1233, "step": 70660 }, { "epoch": 1.385686274509804, "grad_norm": 1.652029275894165, "learning_rate": 2.365221022516612e-07, "loss": 0.1478, "step": 70670 }, { "epoch": 1.3858823529411763, "grad_norm": 3.160660743713379, "learning_rate": 2.3638434434493337e-07, "loss": 0.1046, "step": 70680 }, { "epoch": 1.3860784313725492, "grad_norm": 1.5636640787124634, "learning_rate": 2.3624661414736712e-07, "loss": 0.1136, "step": 70690 }, { "epoch": 1.3862745098039215, "grad_norm": 2.18725323677063, "learning_rate": 2.3610891167343948e-07, "loss": 0.0958, "step": 70700 }, { "epoch": 1.3864705882352941, "grad_norm": 2.449559450149536, "learning_rate": 2.3597123693762495e-07, "loss": 0.1159, "step": 70710 }, { "epoch": 1.3866666666666667, "grad_norm": 3.9458677768707275, "learning_rate": 2.3583358995439438e-07, "loss": 0.1616, "step": 70720 }, { "epoch": 1.386862745098039, "grad_norm": 2.2565314769744873, "learning_rate": 2.3569597073821652e-07, "loss": 0.129, "step": 70730 }, { "epoch": 1.387058823529412, "grad_norm": 1.7338780164718628, "learning_rate": 2.3555837930355643e-07, "loss": 0.127, "step": 70740 }, { "epoch": 1.3872549019607843, "grad_norm": 1.8621855974197388, "learning_rate": 2.35420815664877e-07, "loss": 0.148, "step": 70750 }, { "epoch": 1.3874509803921569, "grad_norm": 2.518172264099121, "learning_rate": 2.3528327983663753e-07, "loss": 0.1474, "step": 70760 }, { "epoch": 1.3876470588235295, "grad_norm": 1.6365303993225098, "learning_rate": 2.3514577183329465e-07, "loss": 0.1149, "step": 70770 }, { "epoch": 1.3878431372549018, "grad_norm": 1.9661978483200073, "learning_rate": 2.350082916693024e-07, "loss": 0.1286, "step": 70780 }, { "epoch": 1.3880392156862746, "grad_norm": 2.279972791671753, "learning_rate": 2.3487083935911138e-07, "loss": 0.1298, "step": 70790 }, { "epoch": 1.388235294117647, "grad_norm": 2.8145594596862793, "learning_rate": 2.3473341491716935e-07, "loss": 0.0944, "step": 70800 }, { "epoch": 1.3884313725490196, "grad_norm": 2.3907699584960938, "learning_rate": 2.3459601835792158e-07, "loss": 0.0928, "step": 70810 }, { "epoch": 1.3886274509803922, "grad_norm": 2.890695333480835, "learning_rate": 2.3445864969580992e-07, "loss": 0.1097, "step": 70820 }, { "epoch": 1.3888235294117646, "grad_norm": 4.13673210144043, "learning_rate": 2.3432130894527323e-07, "loss": 0.1517, "step": 70830 }, { "epoch": 1.3890196078431374, "grad_norm": 1.5330902338027954, "learning_rate": 2.3418399612074819e-07, "loss": 0.108, "step": 70840 }, { "epoch": 1.3892156862745098, "grad_norm": 2.3072667121887207, "learning_rate": 2.3404671123666764e-07, "loss": 0.1278, "step": 70850 }, { "epoch": 1.3894117647058823, "grad_norm": 1.2525877952575684, "learning_rate": 2.3390945430746183e-07, "loss": 0.1025, "step": 70860 }, { "epoch": 1.389607843137255, "grad_norm": 1.7662973403930664, "learning_rate": 2.337722253475584e-07, "loss": 0.1248, "step": 70870 }, { "epoch": 1.3898039215686275, "grad_norm": 3.172543525695801, "learning_rate": 2.3363502437138162e-07, "loss": 0.1127, "step": 70880 }, { "epoch": 1.3900000000000001, "grad_norm": 3.8646299839019775, "learning_rate": 2.334978513933527e-07, "loss": 0.1118, "step": 70890 }, { "epoch": 1.3901960784313725, "grad_norm": 4.165194988250732, "learning_rate": 2.333607064278904e-07, "loss": 0.1323, "step": 70900 }, { "epoch": 1.390392156862745, "grad_norm": 2.2107207775115967, "learning_rate": 2.3322358948941052e-07, "loss": 0.1214, "step": 70910 }, { "epoch": 1.3905882352941177, "grad_norm": 3.4040586948394775, "learning_rate": 2.3308650059232544e-07, "loss": 0.1153, "step": 70920 }, { "epoch": 1.3907843137254903, "grad_norm": 2.200441837310791, "learning_rate": 2.3294943975104464e-07, "loss": 0.1191, "step": 70930 }, { "epoch": 1.3909803921568629, "grad_norm": 2.053452968597412, "learning_rate": 2.328124069799753e-07, "loss": 0.1387, "step": 70940 }, { "epoch": 1.3911764705882352, "grad_norm": 2.664736032485962, "learning_rate": 2.3267540229352102e-07, "loss": 0.1618, "step": 70950 }, { "epoch": 1.3913725490196078, "grad_norm": 2.8734514713287354, "learning_rate": 2.3253842570608234e-07, "loss": 0.1446, "step": 70960 }, { "epoch": 1.3915686274509804, "grad_norm": 2.9884285926818848, "learning_rate": 2.3240147723205762e-07, "loss": 0.1246, "step": 70970 }, { "epoch": 1.391764705882353, "grad_norm": 1.6302592754364014, "learning_rate": 2.3226455688584156e-07, "loss": 0.0998, "step": 70980 }, { "epoch": 1.3919607843137256, "grad_norm": 3.3067941665649414, "learning_rate": 2.3212766468182592e-07, "loss": 0.111, "step": 70990 }, { "epoch": 1.392156862745098, "grad_norm": 1.4165430068969727, "learning_rate": 2.3199080063440014e-07, "loss": 0.1283, "step": 71000 }, { "epoch": 1.3923529411764706, "grad_norm": 2.5399932861328125, "learning_rate": 2.3185396475795005e-07, "loss": 0.111, "step": 71010 }, { "epoch": 1.3925490196078432, "grad_norm": 4.74964714050293, "learning_rate": 2.317171570668585e-07, "loss": 0.1541, "step": 71020 }, { "epoch": 1.3927450980392158, "grad_norm": 3.1323275566101074, "learning_rate": 2.3158037757550608e-07, "loss": 0.1583, "step": 71030 }, { "epoch": 1.3929411764705883, "grad_norm": 3.434025526046753, "learning_rate": 2.3144362629826975e-07, "loss": 0.1508, "step": 71040 }, { "epoch": 1.3931372549019607, "grad_norm": 1.101574182510376, "learning_rate": 2.3130690324952345e-07, "loss": 0.1036, "step": 71050 }, { "epoch": 1.3933333333333333, "grad_norm": 1.7357598543167114, "learning_rate": 2.311702084436389e-07, "loss": 0.1361, "step": 71060 }, { "epoch": 1.393529411764706, "grad_norm": 2.8031673431396484, "learning_rate": 2.3103354189498402e-07, "loss": 0.1212, "step": 71070 }, { "epoch": 1.3937254901960785, "grad_norm": 2.3562371730804443, "learning_rate": 2.3089690361792403e-07, "loss": 0.1502, "step": 71080 }, { "epoch": 1.393921568627451, "grad_norm": 5.224374294281006, "learning_rate": 2.3076029362682141e-07, "loss": 0.1196, "step": 71090 }, { "epoch": 1.3941176470588235, "grad_norm": 1.5710721015930176, "learning_rate": 2.306237119360357e-07, "loss": 0.125, "step": 71100 }, { "epoch": 1.394313725490196, "grad_norm": 4.399165153503418, "learning_rate": 2.30487158559923e-07, "loss": 0.133, "step": 71110 }, { "epoch": 1.3945098039215686, "grad_norm": 2.0772745609283447, "learning_rate": 2.3035063351283667e-07, "loss": 0.1248, "step": 71120 }, { "epoch": 1.3947058823529412, "grad_norm": 2.181741237640381, "learning_rate": 2.302141368091274e-07, "loss": 0.1102, "step": 71130 }, { "epoch": 1.3949019607843138, "grad_norm": 1.3796086311340332, "learning_rate": 2.300776684631424e-07, "loss": 0.1277, "step": 71140 }, { "epoch": 1.3950980392156862, "grad_norm": 4.4115118980407715, "learning_rate": 2.299412284892261e-07, "loss": 0.1291, "step": 71150 }, { "epoch": 1.3952941176470588, "grad_norm": 3.037264585494995, "learning_rate": 2.2980481690172015e-07, "loss": 0.1311, "step": 71160 }, { "epoch": 1.3954901960784314, "grad_norm": 2.3566930294036865, "learning_rate": 2.2966843371496304e-07, "loss": 0.1232, "step": 71170 }, { "epoch": 1.395686274509804, "grad_norm": 1.6281102895736694, "learning_rate": 2.2953207894328992e-07, "loss": 0.1392, "step": 71180 }, { "epoch": 1.3958823529411766, "grad_norm": 3.599210262298584, "learning_rate": 2.2939575260103377e-07, "loss": 0.1121, "step": 71190 }, { "epoch": 1.396078431372549, "grad_norm": 1.8024452924728394, "learning_rate": 2.29259454702524e-07, "loss": 0.1109, "step": 71200 }, { "epoch": 1.3962745098039215, "grad_norm": 2.531104803085327, "learning_rate": 2.2912318526208686e-07, "loss": 0.1339, "step": 71210 }, { "epoch": 1.3964705882352941, "grad_norm": 2.688392162322998, "learning_rate": 2.2898694429404624e-07, "loss": 0.1233, "step": 71220 }, { "epoch": 1.3966666666666667, "grad_norm": 2.3573880195617676, "learning_rate": 2.2885073181272263e-07, "loss": 0.1511, "step": 71230 }, { "epoch": 1.3968627450980393, "grad_norm": 3.3283438682556152, "learning_rate": 2.2871454783243337e-07, "loss": 0.1458, "step": 71240 }, { "epoch": 1.3970588235294117, "grad_norm": 2.0109994411468506, "learning_rate": 2.2857839236749339e-07, "loss": 0.113, "step": 71250 }, { "epoch": 1.3972549019607843, "grad_norm": 2.795437812805176, "learning_rate": 2.2844226543221384e-07, "loss": 0.1326, "step": 71260 }, { "epoch": 1.3974509803921569, "grad_norm": 3.0797719955444336, "learning_rate": 2.2830616704090373e-07, "loss": 0.1291, "step": 71270 }, { "epoch": 1.3976470588235295, "grad_norm": 1.2765253782272339, "learning_rate": 2.2817009720786818e-07, "loss": 0.1475, "step": 71280 }, { "epoch": 1.397843137254902, "grad_norm": 2.3673524856567383, "learning_rate": 2.2803405594741022e-07, "loss": 0.105, "step": 71290 }, { "epoch": 1.3980392156862744, "grad_norm": 2.1303958892822266, "learning_rate": 2.2789804327382917e-07, "loss": 0.1372, "step": 71300 }, { "epoch": 1.398235294117647, "grad_norm": 1.8690216541290283, "learning_rate": 2.2776205920142144e-07, "loss": 0.1378, "step": 71310 }, { "epoch": 1.3984313725490196, "grad_norm": 3.1599855422973633, "learning_rate": 2.2762610374448093e-07, "loss": 0.127, "step": 71320 }, { "epoch": 1.3986274509803922, "grad_norm": 3.197730779647827, "learning_rate": 2.27490176917298e-07, "loss": 0.1256, "step": 71330 }, { "epoch": 1.3988235294117648, "grad_norm": 2.3402042388916016, "learning_rate": 2.2735427873416002e-07, "loss": 0.1315, "step": 71340 }, { "epoch": 1.3990196078431372, "grad_norm": 2.6513049602508545, "learning_rate": 2.2721840920935191e-07, "loss": 0.1251, "step": 71350 }, { "epoch": 1.3992156862745098, "grad_norm": 3.3482298851013184, "learning_rate": 2.2708256835715495e-07, "loss": 0.1573, "step": 71360 }, { "epoch": 1.3994117647058824, "grad_norm": 1.5145399570465088, "learning_rate": 2.2694675619184745e-07, "loss": 0.1428, "step": 71370 }, { "epoch": 1.399607843137255, "grad_norm": 3.1630866527557373, "learning_rate": 2.2681097272770528e-07, "loss": 0.1248, "step": 71380 }, { "epoch": 1.3998039215686275, "grad_norm": 2.272202491760254, "learning_rate": 2.266752179790008e-07, "loss": 0.1298, "step": 71390 }, { "epoch": 1.4, "grad_norm": 2.6968255043029785, "learning_rate": 2.265394919600031e-07, "loss": 0.1271, "step": 71400 }, { "epoch": 1.4001960784313725, "grad_norm": 3.7936201095581055, "learning_rate": 2.2640379468497917e-07, "loss": 0.1477, "step": 71410 }, { "epoch": 1.400392156862745, "grad_norm": 2.7571468353271484, "learning_rate": 2.2626812616819213e-07, "loss": 0.1577, "step": 71420 }, { "epoch": 1.4005882352941177, "grad_norm": 2.4170761108398438, "learning_rate": 2.2613248642390216e-07, "loss": 0.1239, "step": 71430 }, { "epoch": 1.4007843137254903, "grad_norm": 4.276594161987305, "learning_rate": 2.2599687546636687e-07, "loss": 0.1308, "step": 71440 }, { "epoch": 1.4009803921568627, "grad_norm": 3.749877452850342, "learning_rate": 2.2586129330984078e-07, "loss": 0.1154, "step": 71450 }, { "epoch": 1.4011764705882352, "grad_norm": 2.1780266761779785, "learning_rate": 2.2572573996857492e-07, "loss": 0.1154, "step": 71460 }, { "epoch": 1.4013725490196078, "grad_norm": 2.547633171081543, "learning_rate": 2.255902154568175e-07, "loss": 0.1254, "step": 71470 }, { "epoch": 1.4015686274509804, "grad_norm": 2.7959938049316406, "learning_rate": 2.2545471978881409e-07, "loss": 0.1135, "step": 71480 }, { "epoch": 1.401764705882353, "grad_norm": 2.5789244174957275, "learning_rate": 2.253192529788067e-07, "loss": 0.1364, "step": 71490 }, { "epoch": 1.4019607843137254, "grad_norm": 2.971668004989624, "learning_rate": 2.2518381504103433e-07, "loss": 0.1203, "step": 71500 }, { "epoch": 1.402156862745098, "grad_norm": 1.2501678466796875, "learning_rate": 2.2504840598973357e-07, "loss": 0.1139, "step": 71510 }, { "epoch": 1.4023529411764706, "grad_norm": 2.106171131134033, "learning_rate": 2.2491302583913729e-07, "loss": 0.137, "step": 71520 }, { "epoch": 1.4025490196078432, "grad_norm": 3.0951855182647705, "learning_rate": 2.2477767460347536e-07, "loss": 0.1618, "step": 71530 }, { "epoch": 1.4027450980392158, "grad_norm": 1.6062850952148438, "learning_rate": 2.2464235229697524e-07, "loss": 0.1412, "step": 71540 }, { "epoch": 1.4029411764705881, "grad_norm": 2.982443332672119, "learning_rate": 2.245070589338607e-07, "loss": 0.1495, "step": 71550 }, { "epoch": 1.4031372549019607, "grad_norm": 1.8722724914550781, "learning_rate": 2.2437179452835248e-07, "loss": 0.1369, "step": 71560 }, { "epoch": 1.4033333333333333, "grad_norm": 4.276647090911865, "learning_rate": 2.2423655909466894e-07, "loss": 0.1427, "step": 71570 }, { "epoch": 1.403529411764706, "grad_norm": 3.149244546890259, "learning_rate": 2.2410135264702473e-07, "loss": 0.1509, "step": 71580 }, { "epoch": 1.4037254901960785, "grad_norm": 3.2103464603424072, "learning_rate": 2.2396617519963146e-07, "loss": 0.104, "step": 71590 }, { "epoch": 1.4039215686274509, "grad_norm": 5.58598518371582, "learning_rate": 2.2383102676669823e-07, "loss": 0.1335, "step": 71600 }, { "epoch": 1.4041176470588235, "grad_norm": 2.7391583919525146, "learning_rate": 2.236959073624305e-07, "loss": 0.154, "step": 71610 }, { "epoch": 1.404313725490196, "grad_norm": 4.493603229522705, "learning_rate": 2.2356081700103118e-07, "loss": 0.1288, "step": 71620 }, { "epoch": 1.4045098039215687, "grad_norm": 2.025134325027466, "learning_rate": 2.2342575569669957e-07, "loss": 0.1339, "step": 71630 }, { "epoch": 1.4047058823529412, "grad_norm": 1.5310914516448975, "learning_rate": 2.2329072346363265e-07, "loss": 0.1154, "step": 71640 }, { "epoch": 1.4049019607843136, "grad_norm": 1.3984113931655884, "learning_rate": 2.231557203160237e-07, "loss": 0.1323, "step": 71650 }, { "epoch": 1.4050980392156862, "grad_norm": 1.7868176698684692, "learning_rate": 2.2302074626806295e-07, "loss": 0.1056, "step": 71660 }, { "epoch": 1.4052941176470588, "grad_norm": 2.5868468284606934, "learning_rate": 2.228858013339382e-07, "loss": 0.1154, "step": 71670 }, { "epoch": 1.4054901960784314, "grad_norm": 2.0474517345428467, "learning_rate": 2.2275088552783354e-07, "loss": 0.109, "step": 71680 }, { "epoch": 1.405686274509804, "grad_norm": 3.304607629776001, "learning_rate": 2.2261599886393012e-07, "loss": 0.1534, "step": 71690 }, { "epoch": 1.4058823529411764, "grad_norm": 2.801297426223755, "learning_rate": 2.224811413564065e-07, "loss": 0.1509, "step": 71700 }, { "epoch": 1.406078431372549, "grad_norm": 2.400465726852417, "learning_rate": 2.2234631301943763e-07, "loss": 0.1274, "step": 71710 }, { "epoch": 1.4062745098039215, "grad_norm": 4.049952030181885, "learning_rate": 2.2221151386719533e-07, "loss": 0.1451, "step": 71720 }, { "epoch": 1.4064705882352941, "grad_norm": 1.7976422309875488, "learning_rate": 2.2207674391384905e-07, "loss": 0.1047, "step": 71730 }, { "epoch": 1.4066666666666667, "grad_norm": 2.753192186355591, "learning_rate": 2.2194200317356454e-07, "loss": 0.1633, "step": 71740 }, { "epoch": 1.406862745098039, "grad_norm": 2.929530382156372, "learning_rate": 2.2180729166050443e-07, "loss": 0.1584, "step": 71750 }, { "epoch": 1.4070588235294117, "grad_norm": 2.564488649368286, "learning_rate": 2.2167260938882893e-07, "loss": 0.1307, "step": 71760 }, { "epoch": 1.4072549019607843, "grad_norm": 2.186443567276001, "learning_rate": 2.2153795637269456e-07, "loss": 0.1141, "step": 71770 }, { "epoch": 1.4074509803921569, "grad_norm": 3.434058666229248, "learning_rate": 2.21403332626255e-07, "loss": 0.1107, "step": 71780 }, { "epoch": 1.4076470588235295, "grad_norm": 2.742379903793335, "learning_rate": 2.2126873816366031e-07, "loss": 0.1675, "step": 71790 }, { "epoch": 1.4078431372549018, "grad_norm": 2.4803426265716553, "learning_rate": 2.2113417299905889e-07, "loss": 0.1225, "step": 71800 }, { "epoch": 1.4080392156862744, "grad_norm": 1.7785167694091797, "learning_rate": 2.209996371465947e-07, "loss": 0.1167, "step": 71810 }, { "epoch": 1.408235294117647, "grad_norm": 2.969883441925049, "learning_rate": 2.208651306204089e-07, "loss": 0.1417, "step": 71820 }, { "epoch": 1.4084313725490196, "grad_norm": 1.4842736721038818, "learning_rate": 2.2073065343464009e-07, "loss": 0.1142, "step": 71830 }, { "epoch": 1.4086274509803922, "grad_norm": 3.372262716293335, "learning_rate": 2.2059620560342323e-07, "loss": 0.1219, "step": 71840 }, { "epoch": 1.4088235294117646, "grad_norm": 2.5830559730529785, "learning_rate": 2.204617871408903e-07, "loss": 0.134, "step": 71850 }, { "epoch": 1.4090196078431372, "grad_norm": 2.7611758708953857, "learning_rate": 2.2032739806117057e-07, "loss": 0.147, "step": 71860 }, { "epoch": 1.4092156862745098, "grad_norm": 3.5976650714874268, "learning_rate": 2.2019303837838976e-07, "loss": 0.1398, "step": 71870 }, { "epoch": 1.4094117647058824, "grad_norm": 2.773303985595703, "learning_rate": 2.2005870810667054e-07, "loss": 0.1671, "step": 71880 }, { "epoch": 1.409607843137255, "grad_norm": 5.441235542297363, "learning_rate": 2.1992440726013295e-07, "loss": 0.1266, "step": 71890 }, { "epoch": 1.4098039215686273, "grad_norm": 3.6601691246032715, "learning_rate": 2.1979013585289346e-07, "loss": 0.1571, "step": 71900 }, { "epoch": 1.41, "grad_norm": 2.2895545959472656, "learning_rate": 2.1965589389906547e-07, "loss": 0.1131, "step": 71910 }, { "epoch": 1.4101960784313725, "grad_norm": 2.9551613330841064, "learning_rate": 2.1952168141275972e-07, "loss": 0.1169, "step": 71920 }, { "epoch": 1.410392156862745, "grad_norm": 1.2988402843475342, "learning_rate": 2.1938749840808336e-07, "loss": 0.1194, "step": 71930 }, { "epoch": 1.4105882352941177, "grad_norm": 1.7345197200775146, "learning_rate": 2.192533448991405e-07, "loss": 0.1105, "step": 71940 }, { "epoch": 1.41078431372549, "grad_norm": 2.980417251586914, "learning_rate": 2.191192209000326e-07, "loss": 0.1572, "step": 71950 }, { "epoch": 1.4109803921568629, "grad_norm": 2.1259796619415283, "learning_rate": 2.189851264248575e-07, "loss": 0.1609, "step": 71960 }, { "epoch": 1.4111764705882353, "grad_norm": 2.5360748767852783, "learning_rate": 2.1885106148771e-07, "loss": 0.1211, "step": 71970 }, { "epoch": 1.4113725490196078, "grad_norm": 2.40354323387146, "learning_rate": 2.187170261026821e-07, "loss": 0.1304, "step": 71980 }, { "epoch": 1.4115686274509804, "grad_norm": 1.4383628368377686, "learning_rate": 2.1858302028386272e-07, "loss": 0.1068, "step": 71990 }, { "epoch": 1.4117647058823528, "grad_norm": 3.6384170055389404, "learning_rate": 2.184490440453372e-07, "loss": 0.135, "step": 72000 }, { "epoch": 1.4119607843137256, "grad_norm": 3.7932679653167725, "learning_rate": 2.1831509740118804e-07, "loss": 0.1347, "step": 72010 }, { "epoch": 1.412156862745098, "grad_norm": 1.632944941520691, "learning_rate": 2.1818118036549483e-07, "loss": 0.1365, "step": 72020 }, { "epoch": 1.4123529411764706, "grad_norm": 2.426535129547119, "learning_rate": 2.1804729295233375e-07, "loss": 0.1498, "step": 72030 }, { "epoch": 1.4125490196078432, "grad_norm": 2.4056272506713867, "learning_rate": 2.179134351757778e-07, "loss": 0.159, "step": 72040 }, { "epoch": 1.4127450980392156, "grad_norm": 2.2639448642730713, "learning_rate": 2.1777960704989735e-07, "loss": 0.1348, "step": 72050 }, { "epoch": 1.4129411764705884, "grad_norm": 1.6979942321777344, "learning_rate": 2.176458085887592e-07, "loss": 0.1315, "step": 72060 }, { "epoch": 1.4131372549019607, "grad_norm": 2.8064732551574707, "learning_rate": 2.1751203980642696e-07, "loss": 0.1172, "step": 72070 }, { "epoch": 1.4133333333333333, "grad_norm": 1.2378225326538086, "learning_rate": 2.1737830071696173e-07, "loss": 0.1408, "step": 72080 }, { "epoch": 1.413529411764706, "grad_norm": 4.802685260772705, "learning_rate": 2.1724459133442085e-07, "loss": 0.1402, "step": 72090 }, { "epoch": 1.4137254901960783, "grad_norm": 2.1424849033355713, "learning_rate": 2.1711091167285866e-07, "loss": 0.1185, "step": 72100 }, { "epoch": 1.413921568627451, "grad_norm": 2.542191743850708, "learning_rate": 2.169772617463268e-07, "loss": 0.114, "step": 72110 }, { "epoch": 1.4141176470588235, "grad_norm": 3.462852716445923, "learning_rate": 2.1684364156887336e-07, "loss": 0.1265, "step": 72120 }, { "epoch": 1.414313725490196, "grad_norm": 2.6455724239349365, "learning_rate": 2.1671005115454332e-07, "loss": 0.13, "step": 72130 }, { "epoch": 1.4145098039215687, "grad_norm": 1.4519002437591553, "learning_rate": 2.1657649051737857e-07, "loss": 0.1561, "step": 72140 }, { "epoch": 1.4147058823529413, "grad_norm": 2.3515748977661133, "learning_rate": 2.1644295967141801e-07, "loss": 0.1387, "step": 72150 }, { "epoch": 1.4149019607843139, "grad_norm": 3.244818925857544, "learning_rate": 2.163094586306976e-07, "loss": 0.1264, "step": 72160 }, { "epoch": 1.4150980392156862, "grad_norm": 4.0979766845703125, "learning_rate": 2.1617598740924947e-07, "loss": 0.1429, "step": 72170 }, { "epoch": 1.4152941176470588, "grad_norm": 3.0484490394592285, "learning_rate": 2.1604254602110345e-07, "loss": 0.1091, "step": 72180 }, { "epoch": 1.4154901960784314, "grad_norm": 2.8520500659942627, "learning_rate": 2.1590913448028564e-07, "loss": 0.145, "step": 72190 }, { "epoch": 1.415686274509804, "grad_norm": 2.2111101150512695, "learning_rate": 2.1577575280081895e-07, "loss": 0.1323, "step": 72200 }, { "epoch": 1.4158823529411766, "grad_norm": 1.0907984972000122, "learning_rate": 2.1564240099672383e-07, "loss": 0.1247, "step": 72210 }, { "epoch": 1.416078431372549, "grad_norm": 2.7322473526000977, "learning_rate": 2.1550907908201687e-07, "loss": 0.1616, "step": 72220 }, { "epoch": 1.4162745098039216, "grad_norm": 1.2269114255905151, "learning_rate": 2.1537578707071174e-07, "loss": 0.1444, "step": 72230 }, { "epoch": 1.4164705882352941, "grad_norm": 1.8905340433120728, "learning_rate": 2.152425249768194e-07, "loss": 0.1223, "step": 72240 }, { "epoch": 1.4166666666666667, "grad_norm": 1.4420826435089111, "learning_rate": 2.1510929281434692e-07, "loss": 0.1323, "step": 72250 }, { "epoch": 1.4168627450980393, "grad_norm": 2.999685287475586, "learning_rate": 2.149760905972986e-07, "loss": 0.1168, "step": 72260 }, { "epoch": 1.4170588235294117, "grad_norm": 4.1738481521606445, "learning_rate": 2.148429183396759e-07, "loss": 0.1314, "step": 72270 }, { "epoch": 1.4172549019607843, "grad_norm": 2.3370654582977295, "learning_rate": 2.1470977605547664e-07, "loss": 0.1498, "step": 72280 }, { "epoch": 1.417450980392157, "grad_norm": 2.9171223640441895, "learning_rate": 2.1457666375869543e-07, "loss": 0.1147, "step": 72290 }, { "epoch": 1.4176470588235295, "grad_norm": 2.4362854957580566, "learning_rate": 2.1444358146332443e-07, "loss": 0.1466, "step": 72300 }, { "epoch": 1.417843137254902, "grad_norm": 3.125128984451294, "learning_rate": 2.143105291833519e-07, "loss": 0.1315, "step": 72310 }, { "epoch": 1.4180392156862744, "grad_norm": 1.537842869758606, "learning_rate": 2.1417750693276325e-07, "loss": 0.1023, "step": 72320 }, { "epoch": 1.418235294117647, "grad_norm": 1.802374005317688, "learning_rate": 2.140445147255404e-07, "loss": 0.1469, "step": 72330 }, { "epoch": 1.4184313725490196, "grad_norm": 3.0310511589050293, "learning_rate": 2.1391155257566313e-07, "loss": 0.118, "step": 72340 }, { "epoch": 1.4186274509803922, "grad_norm": 2.368821382522583, "learning_rate": 2.1377862049710698e-07, "loss": 0.1099, "step": 72350 }, { "epoch": 1.4188235294117648, "grad_norm": 1.319395661354065, "learning_rate": 2.136457185038445e-07, "loss": 0.1039, "step": 72360 }, { "epoch": 1.4190196078431372, "grad_norm": 2.7477691173553467, "learning_rate": 2.135128466098457e-07, "loss": 0.1116, "step": 72370 }, { "epoch": 1.4192156862745098, "grad_norm": 3.2994601726531982, "learning_rate": 2.1338000482907676e-07, "loss": 0.1317, "step": 72380 }, { "epoch": 1.4194117647058824, "grad_norm": 2.5507521629333496, "learning_rate": 2.1324719317550082e-07, "loss": 0.1217, "step": 72390 }, { "epoch": 1.419607843137255, "grad_norm": 2.498020648956299, "learning_rate": 2.1311441166307824e-07, "loss": 0.1738, "step": 72400 }, { "epoch": 1.4198039215686276, "grad_norm": 1.9268996715545654, "learning_rate": 2.1298166030576593e-07, "loss": 0.1456, "step": 72410 }, { "epoch": 1.42, "grad_norm": 3.5105528831481934, "learning_rate": 2.1284893911751734e-07, "loss": 0.1214, "step": 72420 }, { "epoch": 1.4201960784313725, "grad_norm": 2.529291868209839, "learning_rate": 2.1271624811228339e-07, "loss": 0.128, "step": 72430 }, { "epoch": 1.4203921568627451, "grad_norm": 3.826406478881836, "learning_rate": 2.1258358730401143e-07, "loss": 0.1423, "step": 72440 }, { "epoch": 1.4205882352941177, "grad_norm": 1.686355710029602, "learning_rate": 2.1245095670664547e-07, "loss": 0.12, "step": 72450 }, { "epoch": 1.4207843137254903, "grad_norm": 3.3731303215026855, "learning_rate": 2.1231835633412687e-07, "loss": 0.1268, "step": 72460 }, { "epoch": 1.4209803921568627, "grad_norm": 3.308316707611084, "learning_rate": 2.1218578620039345e-07, "loss": 0.1123, "step": 72470 }, { "epoch": 1.4211764705882353, "grad_norm": 3.054039716720581, "learning_rate": 2.1205324631937983e-07, "loss": 0.1387, "step": 72480 }, { "epoch": 1.4213725490196079, "grad_norm": 2.0762951374053955, "learning_rate": 2.1192073670501737e-07, "loss": 0.1154, "step": 72490 }, { "epoch": 1.4215686274509804, "grad_norm": 1.0577794313430786, "learning_rate": 2.1178825737123484e-07, "loss": 0.1127, "step": 72500 }, { "epoch": 1.421764705882353, "grad_norm": 3.574871063232422, "learning_rate": 2.1165580833195694e-07, "loss": 0.1302, "step": 72510 }, { "epoch": 1.4219607843137254, "grad_norm": 3.7162983417510986, "learning_rate": 2.1152338960110595e-07, "loss": 0.1463, "step": 72520 }, { "epoch": 1.422156862745098, "grad_norm": 1.1533277034759521, "learning_rate": 2.1139100119260073e-07, "loss": 0.1282, "step": 72530 }, { "epoch": 1.4223529411764706, "grad_norm": 1.8089925050735474, "learning_rate": 2.112586431203568e-07, "loss": 0.14, "step": 72540 }, { "epoch": 1.4225490196078432, "grad_norm": 3.51296067237854, "learning_rate": 2.1112631539828629e-07, "loss": 0.1531, "step": 72550 }, { "epoch": 1.4227450980392158, "grad_norm": 2.95794939994812, "learning_rate": 2.1099401804029892e-07, "loss": 0.1485, "step": 72560 }, { "epoch": 1.4229411764705882, "grad_norm": 1.9724252223968506, "learning_rate": 2.1086175106030046e-07, "loss": 0.1234, "step": 72570 }, { "epoch": 1.4231372549019607, "grad_norm": 2.2343556880950928, "learning_rate": 2.1072951447219362e-07, "loss": 0.1265, "step": 72580 }, { "epoch": 1.4233333333333333, "grad_norm": 2.8667280673980713, "learning_rate": 2.1059730828987836e-07, "loss": 0.1302, "step": 72590 }, { "epoch": 1.423529411764706, "grad_norm": 1.8110859394073486, "learning_rate": 2.1046513252725096e-07, "loss": 0.1491, "step": 72600 }, { "epoch": 1.4237254901960785, "grad_norm": 2.7590224742889404, "learning_rate": 2.1033298719820453e-07, "loss": 0.1455, "step": 72610 }, { "epoch": 1.423921568627451, "grad_norm": 3.3861844539642334, "learning_rate": 2.1020087231662947e-07, "loss": 0.1414, "step": 72620 }, { "epoch": 1.4241176470588235, "grad_norm": 2.4888086318969727, "learning_rate": 2.1006878789641247e-07, "loss": 0.1388, "step": 72630 }, { "epoch": 1.424313725490196, "grad_norm": 1.9386401176452637, "learning_rate": 2.0993673395143695e-07, "loss": 0.1287, "step": 72640 }, { "epoch": 1.4245098039215687, "grad_norm": 3.1867997646331787, "learning_rate": 2.0980471049558378e-07, "loss": 0.117, "step": 72650 }, { "epoch": 1.4247058823529413, "grad_norm": 7.194578647613525, "learning_rate": 2.0967271754272996e-07, "loss": 0.1251, "step": 72660 }, { "epoch": 1.4249019607843136, "grad_norm": 3.9907538890838623, "learning_rate": 2.0954075510674962e-07, "loss": 0.1319, "step": 72670 }, { "epoch": 1.4250980392156862, "grad_norm": 2.6852335929870605, "learning_rate": 2.0940882320151339e-07, "loss": 0.1482, "step": 72680 }, { "epoch": 1.4252941176470588, "grad_norm": 4.221119403839111, "learning_rate": 2.0927692184088903e-07, "loss": 0.1185, "step": 72690 }, { "epoch": 1.4254901960784314, "grad_norm": 2.1546788215637207, "learning_rate": 2.0914505103874107e-07, "loss": 0.1255, "step": 72700 }, { "epoch": 1.425686274509804, "grad_norm": 2.0501902103424072, "learning_rate": 2.0901321080893053e-07, "loss": 0.1268, "step": 72710 }, { "epoch": 1.4258823529411764, "grad_norm": 2.0783519744873047, "learning_rate": 2.088814011653156e-07, "loss": 0.131, "step": 72720 }, { "epoch": 1.426078431372549, "grad_norm": 1.5593198537826538, "learning_rate": 2.0874962212175095e-07, "loss": 0.152, "step": 72730 }, { "epoch": 1.4262745098039216, "grad_norm": 3.1471099853515625, "learning_rate": 2.0861787369208795e-07, "loss": 0.1126, "step": 72740 }, { "epoch": 1.4264705882352942, "grad_norm": 1.8246978521347046, "learning_rate": 2.0848615589017526e-07, "loss": 0.1234, "step": 72750 }, { "epoch": 1.4266666666666667, "grad_norm": 3.110276937484741, "learning_rate": 2.0835446872985784e-07, "loss": 0.1268, "step": 72760 }, { "epoch": 1.4268627450980391, "grad_norm": 1.803491473197937, "learning_rate": 2.0822281222497746e-07, "loss": 0.1148, "step": 72770 }, { "epoch": 1.4270588235294117, "grad_norm": 2.814429521560669, "learning_rate": 2.0809118638937307e-07, "loss": 0.1388, "step": 72780 }, { "epoch": 1.4272549019607843, "grad_norm": 2.028939962387085, "learning_rate": 2.0795959123687995e-07, "loss": 0.1159, "step": 72790 }, { "epoch": 1.427450980392157, "grad_norm": 4.727185249328613, "learning_rate": 2.0782802678133015e-07, "loss": 0.15, "step": 72800 }, { "epoch": 1.4276470588235295, "grad_norm": 3.5349488258361816, "learning_rate": 2.0769649303655307e-07, "loss": 0.1358, "step": 72810 }, { "epoch": 1.4278431372549019, "grad_norm": 3.3844048976898193, "learning_rate": 2.075649900163743e-07, "loss": 0.112, "step": 72820 }, { "epoch": 1.4280392156862745, "grad_norm": 3.397099018096924, "learning_rate": 2.0743351773461637e-07, "loss": 0.1309, "step": 72830 }, { "epoch": 1.428235294117647, "grad_norm": 2.367264747619629, "learning_rate": 2.0730207620509838e-07, "loss": 0.1557, "step": 72840 }, { "epoch": 1.4284313725490196, "grad_norm": 1.7856487035751343, "learning_rate": 2.071706654416368e-07, "loss": 0.1115, "step": 72850 }, { "epoch": 1.4286274509803922, "grad_norm": 2.2292346954345703, "learning_rate": 2.070392854580441e-07, "loss": 0.1287, "step": 72860 }, { "epoch": 1.4288235294117646, "grad_norm": 4.066153526306152, "learning_rate": 2.0690793626813007e-07, "loss": 0.1273, "step": 72870 }, { "epoch": 1.4290196078431372, "grad_norm": 1.6575746536254883, "learning_rate": 2.0677661788570128e-07, "loss": 0.109, "step": 72880 }, { "epoch": 1.4292156862745098, "grad_norm": 2.122418165206909, "learning_rate": 2.0664533032456067e-07, "loss": 0.135, "step": 72890 }, { "epoch": 1.4294117647058824, "grad_norm": 3.183856248855591, "learning_rate": 2.0651407359850791e-07, "loss": 0.132, "step": 72900 }, { "epoch": 1.429607843137255, "grad_norm": 3.580073833465576, "learning_rate": 2.0638284772134012e-07, "loss": 0.1329, "step": 72910 }, { "epoch": 1.4298039215686273, "grad_norm": 2.8902382850646973, "learning_rate": 2.0625165270685047e-07, "loss": 0.1135, "step": 72920 }, { "epoch": 1.43, "grad_norm": 2.5313022136688232, "learning_rate": 2.061204885688289e-07, "loss": 0.1253, "step": 72930 }, { "epoch": 1.4301960784313725, "grad_norm": 4.545020580291748, "learning_rate": 2.059893553210628e-07, "loss": 0.1244, "step": 72940 }, { "epoch": 1.4303921568627451, "grad_norm": 2.4390861988067627, "learning_rate": 2.0585825297733555e-07, "loss": 0.1293, "step": 72950 }, { "epoch": 1.4305882352941177, "grad_norm": 5.021145343780518, "learning_rate": 2.057271815514275e-07, "loss": 0.151, "step": 72960 }, { "epoch": 1.43078431372549, "grad_norm": 2.2906243801116943, "learning_rate": 2.0559614105711609e-07, "loss": 0.1256, "step": 72970 }, { "epoch": 1.4309803921568627, "grad_norm": 3.000657558441162, "learning_rate": 2.0546513150817513e-07, "loss": 0.1217, "step": 72980 }, { "epoch": 1.4311764705882353, "grad_norm": 3.0432467460632324, "learning_rate": 2.053341529183751e-07, "loss": 0.1183, "step": 72990 }, { "epoch": 1.4313725490196079, "grad_norm": 2.0179550647735596, "learning_rate": 2.0520320530148377e-07, "loss": 0.1351, "step": 73000 }, { "epoch": 1.4315686274509805, "grad_norm": 1.6733795404434204, "learning_rate": 2.0507228867126507e-07, "loss": 0.1229, "step": 73010 }, { "epoch": 1.4317647058823528, "grad_norm": 3.229912519454956, "learning_rate": 2.0494140304148007e-07, "loss": 0.1409, "step": 73020 }, { "epoch": 1.4319607843137254, "grad_norm": 4.07755184173584, "learning_rate": 2.0481054842588602e-07, "loss": 0.1172, "step": 73030 }, { "epoch": 1.432156862745098, "grad_norm": 1.9479248523712158, "learning_rate": 2.0467972483823766e-07, "loss": 0.1168, "step": 73040 }, { "epoch": 1.4323529411764706, "grad_norm": 2.7588934898376465, "learning_rate": 2.0454893229228614e-07, "loss": 0.127, "step": 73050 }, { "epoch": 1.4325490196078432, "grad_norm": 1.460852861404419, "learning_rate": 2.0441817080177913e-07, "loss": 0.1209, "step": 73060 }, { "epoch": 1.4327450980392156, "grad_norm": 1.7679892778396606, "learning_rate": 2.0428744038046146e-07, "loss": 0.146, "step": 73070 }, { "epoch": 1.4329411764705882, "grad_norm": 2.8496103286743164, "learning_rate": 2.0415674104207436e-07, "loss": 0.1363, "step": 73080 }, { "epoch": 1.4331372549019608, "grad_norm": 2.767158269882202, "learning_rate": 2.0402607280035571e-07, "loss": 0.127, "step": 73090 }, { "epoch": 1.4333333333333333, "grad_norm": 1.3350704908370972, "learning_rate": 2.038954356690406e-07, "loss": 0.1182, "step": 73100 }, { "epoch": 1.433529411764706, "grad_norm": 2.0911202430725098, "learning_rate": 2.0376482966186041e-07, "loss": 0.1252, "step": 73110 }, { "epoch": 1.4337254901960783, "grad_norm": 1.9956176280975342, "learning_rate": 2.0363425479254327e-07, "loss": 0.1475, "step": 73120 }, { "epoch": 1.433921568627451, "grad_norm": 2.8297042846679688, "learning_rate": 2.0350371107481445e-07, "loss": 0.1332, "step": 73130 }, { "epoch": 1.4341176470588235, "grad_norm": 3.5701870918273926, "learning_rate": 2.0337319852239553e-07, "loss": 0.1181, "step": 73140 }, { "epoch": 1.434313725490196, "grad_norm": 1.7956547737121582, "learning_rate": 2.032427171490047e-07, "loss": 0.1465, "step": 73150 }, { "epoch": 1.4345098039215687, "grad_norm": 2.7686095237731934, "learning_rate": 2.0311226696835754e-07, "loss": 0.1278, "step": 73160 }, { "epoch": 1.434705882352941, "grad_norm": 1.9710689783096313, "learning_rate": 2.0298184799416568e-07, "loss": 0.1459, "step": 73170 }, { "epoch": 1.4349019607843136, "grad_norm": 1.6626272201538086, "learning_rate": 2.0285146024013778e-07, "loss": 0.1367, "step": 73180 }, { "epoch": 1.4350980392156862, "grad_norm": 3.6782751083374023, "learning_rate": 2.0272110371997891e-07, "loss": 0.1682, "step": 73190 }, { "epoch": 1.4352941176470588, "grad_norm": 1.6070541143417358, "learning_rate": 2.0259077844739153e-07, "loss": 0.1048, "step": 73200 }, { "epoch": 1.4354901960784314, "grad_norm": 1.5108094215393066, "learning_rate": 2.0246048443607417e-07, "loss": 0.1098, "step": 73210 }, { "epoch": 1.4356862745098038, "grad_norm": 2.093479871749878, "learning_rate": 2.0233022169972192e-07, "loss": 0.123, "step": 73220 }, { "epoch": 1.4358823529411764, "grad_norm": 2.6865968704223633, "learning_rate": 2.021999902520277e-07, "loss": 0.1484, "step": 73230 }, { "epoch": 1.436078431372549, "grad_norm": 1.742870807647705, "learning_rate": 2.0206979010668e-07, "loss": 0.1467, "step": 73240 }, { "epoch": 1.4362745098039216, "grad_norm": 2.890423536300659, "learning_rate": 2.019396212773643e-07, "loss": 0.1231, "step": 73250 }, { "epoch": 1.4364705882352942, "grad_norm": 3.899416923522949, "learning_rate": 2.0180948377776318e-07, "loss": 0.1326, "step": 73260 }, { "epoch": 1.4366666666666665, "grad_norm": 5.392406940460205, "learning_rate": 2.016793776215555e-07, "loss": 0.1593, "step": 73270 }, { "epoch": 1.4368627450980394, "grad_norm": 2.5572669506073, "learning_rate": 2.0154930282241683e-07, "loss": 0.125, "step": 73280 }, { "epoch": 1.4370588235294117, "grad_norm": 2.201760768890381, "learning_rate": 2.0141925939401993e-07, "loss": 0.107, "step": 73290 }, { "epoch": 1.4372549019607843, "grad_norm": 3.354583263397217, "learning_rate": 2.012892473500337e-07, "loss": 0.1545, "step": 73300 }, { "epoch": 1.437450980392157, "grad_norm": 2.099836587905884, "learning_rate": 2.0115926670412387e-07, "loss": 0.1278, "step": 73310 }, { "epoch": 1.4376470588235293, "grad_norm": 3.023993968963623, "learning_rate": 2.010293174699532e-07, "loss": 0.1261, "step": 73320 }, { "epoch": 1.437843137254902, "grad_norm": 2.026628017425537, "learning_rate": 2.0089939966118087e-07, "loss": 0.118, "step": 73330 }, { "epoch": 1.4380392156862745, "grad_norm": 2.941028118133545, "learning_rate": 2.0076951329146247e-07, "loss": 0.1224, "step": 73340 }, { "epoch": 1.438235294117647, "grad_norm": 1.5107661485671997, "learning_rate": 2.0063965837445106e-07, "loss": 0.1189, "step": 73350 }, { "epoch": 1.4384313725490196, "grad_norm": 3.1688833236694336, "learning_rate": 2.005098349237958e-07, "loss": 0.1464, "step": 73360 }, { "epoch": 1.438627450980392, "grad_norm": 2.184624671936035, "learning_rate": 2.0038004295314258e-07, "loss": 0.1329, "step": 73370 }, { "epoch": 1.4388235294117648, "grad_norm": 2.0267183780670166, "learning_rate": 2.00250282476134e-07, "loss": 0.1381, "step": 73380 }, { "epoch": 1.4390196078431372, "grad_norm": 2.6559560298919678, "learning_rate": 2.0012055350640983e-07, "loss": 0.1687, "step": 73390 }, { "epoch": 1.4392156862745098, "grad_norm": 2.786552667617798, "learning_rate": 1.9999085605760564e-07, "loss": 0.1024, "step": 73400 }, { "epoch": 1.4394117647058824, "grad_norm": 1.9167094230651855, "learning_rate": 1.9986119014335444e-07, "loss": 0.145, "step": 73410 }, { "epoch": 1.439607843137255, "grad_norm": 2.196263313293457, "learning_rate": 1.9973155577728596e-07, "loss": 0.1291, "step": 73420 }, { "epoch": 1.4398039215686276, "grad_norm": 2.1371958255767822, "learning_rate": 1.9960195297302595e-07, "loss": 0.13, "step": 73430 }, { "epoch": 1.44, "grad_norm": 2.983785390853882, "learning_rate": 1.9947238174419712e-07, "loss": 0.1381, "step": 73440 }, { "epoch": 1.4401960784313725, "grad_norm": 1.8326914310455322, "learning_rate": 1.9934284210441939e-07, "loss": 0.1637, "step": 73450 }, { "epoch": 1.4403921568627451, "grad_norm": 3.338010549545288, "learning_rate": 1.9921333406730862e-07, "loss": 0.1247, "step": 73460 }, { "epoch": 1.4405882352941177, "grad_norm": 1.804635763168335, "learning_rate": 1.9908385764647756e-07, "loss": 0.1359, "step": 73470 }, { "epoch": 1.4407843137254903, "grad_norm": 1.8997437953948975, "learning_rate": 1.9895441285553605e-07, "loss": 0.1086, "step": 73480 }, { "epoch": 1.4409803921568627, "grad_norm": 2.7771546840667725, "learning_rate": 1.9882499970809014e-07, "loss": 0.1221, "step": 73490 }, { "epoch": 1.4411764705882353, "grad_norm": 4.268233776092529, "learning_rate": 1.986956182177425e-07, "loss": 0.1143, "step": 73500 }, { "epoch": 1.4413725490196079, "grad_norm": 1.9777783155441284, "learning_rate": 1.9856626839809304e-07, "loss": 0.1257, "step": 73510 }, { "epoch": 1.4415686274509805, "grad_norm": 2.1609883308410645, "learning_rate": 1.9843695026273776e-07, "loss": 0.1124, "step": 73520 }, { "epoch": 1.441764705882353, "grad_norm": 2.102041482925415, "learning_rate": 1.9830766382526965e-07, "loss": 0.1416, "step": 73530 }, { "epoch": 1.4419607843137254, "grad_norm": 0.9429334998130798, "learning_rate": 1.9817840909927797e-07, "loss": 0.0997, "step": 73540 }, { "epoch": 1.442156862745098, "grad_norm": 1.9833452701568604, "learning_rate": 1.980491860983493e-07, "loss": 0.1236, "step": 73550 }, { "epoch": 1.4423529411764706, "grad_norm": 3.2853288650512695, "learning_rate": 1.9791999483606648e-07, "loss": 0.1466, "step": 73560 }, { "epoch": 1.4425490196078432, "grad_norm": 3.6773056983947754, "learning_rate": 1.977908353260087e-07, "loss": 0.1439, "step": 73570 }, { "epoch": 1.4427450980392158, "grad_norm": 2.5235517024993896, "learning_rate": 1.9766170758175248e-07, "loss": 0.1146, "step": 73580 }, { "epoch": 1.4429411764705882, "grad_norm": 2.3812637329101562, "learning_rate": 1.9753261161687085e-07, "loss": 0.1267, "step": 73590 }, { "epoch": 1.4431372549019608, "grad_norm": 3.2054648399353027, "learning_rate": 1.9740354744493292e-07, "loss": 0.128, "step": 73600 }, { "epoch": 1.4433333333333334, "grad_norm": 2.411115884780884, "learning_rate": 1.9727451507950528e-07, "loss": 0.1098, "step": 73610 }, { "epoch": 1.443529411764706, "grad_norm": 2.2602200508117676, "learning_rate": 1.9714551453415056e-07, "loss": 0.1269, "step": 73620 }, { "epoch": 1.4437254901960785, "grad_norm": 3.532942056655884, "learning_rate": 1.9701654582242816e-07, "loss": 0.1416, "step": 73630 }, { "epoch": 1.443921568627451, "grad_norm": 3.5165910720825195, "learning_rate": 1.9688760895789457e-07, "loss": 0.1123, "step": 73640 }, { "epoch": 1.4441176470588235, "grad_norm": 2.6254308223724365, "learning_rate": 1.967587039541024e-07, "loss": 0.1362, "step": 73650 }, { "epoch": 1.444313725490196, "grad_norm": 3.1407229900360107, "learning_rate": 1.966298308246009e-07, "loss": 0.1116, "step": 73660 }, { "epoch": 1.4445098039215687, "grad_norm": 1.1839131116867065, "learning_rate": 1.9650098958293655e-07, "loss": 0.1378, "step": 73670 }, { "epoch": 1.4447058823529413, "grad_norm": 3.3794260025024414, "learning_rate": 1.9637218024265196e-07, "loss": 0.1612, "step": 73680 }, { "epoch": 1.4449019607843137, "grad_norm": 24.67474365234375, "learning_rate": 1.9624340281728636e-07, "loss": 0.1561, "step": 73690 }, { "epoch": 1.4450980392156862, "grad_norm": 2.8169915676116943, "learning_rate": 1.9611465732037608e-07, "loss": 0.1382, "step": 73700 }, { "epoch": 1.4452941176470588, "grad_norm": 2.1089229583740234, "learning_rate": 1.9598594376545375e-07, "loss": 0.1315, "step": 73710 }, { "epoch": 1.4454901960784314, "grad_norm": 3.4165256023406982, "learning_rate": 1.9585726216604864e-07, "loss": 0.1313, "step": 73720 }, { "epoch": 1.445686274509804, "grad_norm": 1.7261370420455933, "learning_rate": 1.9572861253568657e-07, "loss": 0.1371, "step": 73730 }, { "epoch": 1.4458823529411764, "grad_norm": 2.2590019702911377, "learning_rate": 1.9559999488789046e-07, "loss": 0.1166, "step": 73740 }, { "epoch": 1.446078431372549, "grad_norm": 2.779942512512207, "learning_rate": 1.9547140923617943e-07, "loss": 0.1515, "step": 73750 }, { "epoch": 1.4462745098039216, "grad_norm": 3.2245657444000244, "learning_rate": 1.9534285559406905e-07, "loss": 0.1409, "step": 73760 }, { "epoch": 1.4464705882352942, "grad_norm": 2.9565670490264893, "learning_rate": 1.9521433397507247e-07, "loss": 0.1445, "step": 73770 }, { "epoch": 1.4466666666666668, "grad_norm": 2.7793192863464355, "learning_rate": 1.9508584439269855e-07, "loss": 0.1173, "step": 73780 }, { "epoch": 1.4468627450980391, "grad_norm": 2.233349323272705, "learning_rate": 1.9495738686045293e-07, "loss": 0.1449, "step": 73790 }, { "epoch": 1.4470588235294117, "grad_norm": 1.0256422758102417, "learning_rate": 1.9482896139183836e-07, "loss": 0.1276, "step": 73800 }, { "epoch": 1.4472549019607843, "grad_norm": 1.8058687448501587, "learning_rate": 1.9470056800035367e-07, "loss": 0.163, "step": 73810 }, { "epoch": 1.447450980392157, "grad_norm": 2.6056272983551025, "learning_rate": 1.9457220669949448e-07, "loss": 0.1304, "step": 73820 }, { "epoch": 1.4476470588235295, "grad_norm": 3.542271852493286, "learning_rate": 1.9444387750275338e-07, "loss": 0.1407, "step": 73830 }, { "epoch": 1.4478431372549019, "grad_norm": 2.3449409008026123, "learning_rate": 1.9431558042361906e-07, "loss": 0.1219, "step": 73840 }, { "epoch": 1.4480392156862745, "grad_norm": 3.60882306098938, "learning_rate": 1.9418731547557705e-07, "loss": 0.1122, "step": 73850 }, { "epoch": 1.448235294117647, "grad_norm": 1.7274584770202637, "learning_rate": 1.940590826721098e-07, "loss": 0.1582, "step": 73860 }, { "epoch": 1.4484313725490197, "grad_norm": 2.4778029918670654, "learning_rate": 1.939308820266959e-07, "loss": 0.1343, "step": 73870 }, { "epoch": 1.4486274509803923, "grad_norm": 2.2562501430511475, "learning_rate": 1.9380271355281086e-07, "loss": 0.145, "step": 73880 }, { "epoch": 1.4488235294117646, "grad_norm": 3.327505350112915, "learning_rate": 1.9367457726392649e-07, "loss": 0.1297, "step": 73890 }, { "epoch": 1.4490196078431372, "grad_norm": 3.0081968307495117, "learning_rate": 1.9354647317351187e-07, "loss": 0.1227, "step": 73900 }, { "epoch": 1.4492156862745098, "grad_norm": 1.7991820573806763, "learning_rate": 1.93418401295032e-07, "loss": 0.1598, "step": 73910 }, { "epoch": 1.4494117647058824, "grad_norm": 1.3760805130004883, "learning_rate": 1.932903616419486e-07, "loss": 0.1276, "step": 73920 }, { "epoch": 1.449607843137255, "grad_norm": 2.2422008514404297, "learning_rate": 1.9316235422772058e-07, "loss": 0.1336, "step": 73930 }, { "epoch": 1.4498039215686274, "grad_norm": 2.0421154499053955, "learning_rate": 1.930343790658027e-07, "loss": 0.1217, "step": 73940 }, { "epoch": 1.45, "grad_norm": 5.014698505401611, "learning_rate": 1.9290643616964687e-07, "loss": 0.1804, "step": 73950 }, { "epoch": 1.4501960784313725, "grad_norm": 2.337230682373047, "learning_rate": 1.9277852555270152e-07, "loss": 0.1269, "step": 73960 }, { "epoch": 1.4503921568627451, "grad_norm": 4.605402946472168, "learning_rate": 1.9265064722841145e-07, "loss": 0.1606, "step": 73970 }, { "epoch": 1.4505882352941177, "grad_norm": 2.4112589359283447, "learning_rate": 1.9252280121021802e-07, "loss": 0.1068, "step": 73980 }, { "epoch": 1.45078431372549, "grad_norm": 3.804978132247925, "learning_rate": 1.9239498751155975e-07, "loss": 0.127, "step": 73990 }, { "epoch": 1.4509803921568627, "grad_norm": 2.5344033241271973, "learning_rate": 1.922672061458712e-07, "loss": 0.1262, "step": 74000 }, { "epoch": 1.4511764705882353, "grad_norm": 1.800079107284546, "learning_rate": 1.9213945712658358e-07, "loss": 0.1552, "step": 74010 }, { "epoch": 1.4513725490196079, "grad_norm": 0.9142264127731323, "learning_rate": 1.9201174046712514e-07, "loss": 0.1284, "step": 74020 }, { "epoch": 1.4515686274509805, "grad_norm": 2.405961751937866, "learning_rate": 1.918840561809203e-07, "loss": 0.1269, "step": 74030 }, { "epoch": 1.4517647058823528, "grad_norm": 3.932258129119873, "learning_rate": 1.9175640428138995e-07, "loss": 0.138, "step": 74040 }, { "epoch": 1.4519607843137254, "grad_norm": 1.8213069438934326, "learning_rate": 1.9162878478195232e-07, "loss": 0.1349, "step": 74050 }, { "epoch": 1.452156862745098, "grad_norm": 3.21279239654541, "learning_rate": 1.9150119769602141e-07, "loss": 0.1695, "step": 74060 }, { "epoch": 1.4523529411764706, "grad_norm": 1.6276588439941406, "learning_rate": 1.913736430370083e-07, "loss": 0.1098, "step": 74070 }, { "epoch": 1.4525490196078432, "grad_norm": 2.625190258026123, "learning_rate": 1.9124612081832025e-07, "loss": 0.1511, "step": 74080 }, { "epoch": 1.4527450980392156, "grad_norm": 3.027679920196533, "learning_rate": 1.9111863105336174e-07, "loss": 0.1468, "step": 74090 }, { "epoch": 1.4529411764705882, "grad_norm": 3.847750663757324, "learning_rate": 1.909911737555333e-07, "loss": 0.1157, "step": 74100 }, { "epoch": 1.4531372549019608, "grad_norm": 2.7040061950683594, "learning_rate": 1.9086374893823205e-07, "loss": 0.1015, "step": 74110 }, { "epoch": 1.4533333333333334, "grad_norm": 3.8659262657165527, "learning_rate": 1.90736356614852e-07, "loss": 0.1463, "step": 74120 }, { "epoch": 1.453529411764706, "grad_norm": 3.2263989448547363, "learning_rate": 1.9060899679878383e-07, "loss": 0.1562, "step": 74130 }, { "epoch": 1.4537254901960783, "grad_norm": 2.51619553565979, "learning_rate": 1.9048166950341422e-07, "loss": 0.1197, "step": 74140 }, { "epoch": 1.453921568627451, "grad_norm": 5.3260979652404785, "learning_rate": 1.9035437474212717e-07, "loss": 0.148, "step": 74150 }, { "epoch": 1.4541176470588235, "grad_norm": 2.511701822280884, "learning_rate": 1.9022711252830264e-07, "loss": 0.1432, "step": 74160 }, { "epoch": 1.454313725490196, "grad_norm": 1.3540104627609253, "learning_rate": 1.9009988287531724e-07, "loss": 0.1165, "step": 74170 }, { "epoch": 1.4545098039215687, "grad_norm": 2.2799293994903564, "learning_rate": 1.899726857965448e-07, "loss": 0.1167, "step": 74180 }, { "epoch": 1.454705882352941, "grad_norm": 2.3719522953033447, "learning_rate": 1.8984552130535498e-07, "loss": 0.1243, "step": 74190 }, { "epoch": 1.4549019607843137, "grad_norm": 2.1269588470458984, "learning_rate": 1.8971838941511408e-07, "loss": 0.1452, "step": 74200 }, { "epoch": 1.4550980392156863, "grad_norm": 3.184934616088867, "learning_rate": 1.8959129013918563e-07, "loss": 0.1251, "step": 74210 }, { "epoch": 1.4552941176470588, "grad_norm": 3.660442590713501, "learning_rate": 1.89464223490929e-07, "loss": 0.17, "step": 74220 }, { "epoch": 1.4554901960784314, "grad_norm": 3.606250047683716, "learning_rate": 1.8933718948370054e-07, "loss": 0.1143, "step": 74230 }, { "epoch": 1.4556862745098038, "grad_norm": 2.3107075691223145, "learning_rate": 1.8921018813085283e-07, "loss": 0.1203, "step": 74240 }, { "epoch": 1.4558823529411764, "grad_norm": 2.3192546367645264, "learning_rate": 1.8908321944573552e-07, "loss": 0.1315, "step": 74250 }, { "epoch": 1.456078431372549, "grad_norm": 2.148433208465576, "learning_rate": 1.889562834416944e-07, "loss": 0.1393, "step": 74260 }, { "epoch": 1.4562745098039216, "grad_norm": 1.2734261751174927, "learning_rate": 1.8882938013207184e-07, "loss": 0.1278, "step": 74270 }, { "epoch": 1.4564705882352942, "grad_norm": 3.3146936893463135, "learning_rate": 1.8870250953020717e-07, "loss": 0.1483, "step": 74280 }, { "epoch": 1.4566666666666666, "grad_norm": 1.6067525148391724, "learning_rate": 1.8857567164943573e-07, "loss": 0.1313, "step": 74290 }, { "epoch": 1.4568627450980391, "grad_norm": 4.235874176025391, "learning_rate": 1.8844886650308983e-07, "loss": 0.1173, "step": 74300 }, { "epoch": 1.4570588235294117, "grad_norm": 1.7067692279815674, "learning_rate": 1.8832209410449833e-07, "loss": 0.1309, "step": 74310 }, { "epoch": 1.4572549019607843, "grad_norm": 3.52600359916687, "learning_rate": 1.8819535446698648e-07, "loss": 0.1222, "step": 74320 }, { "epoch": 1.457450980392157, "grad_norm": 1.5951305627822876, "learning_rate": 1.8806864760387592e-07, "loss": 0.1507, "step": 74330 }, { "epoch": 1.4576470588235293, "grad_norm": 2.3367905616760254, "learning_rate": 1.8794197352848528e-07, "loss": 0.109, "step": 74340 }, { "epoch": 1.4578431372549019, "grad_norm": 2.491349697113037, "learning_rate": 1.8781533225412956e-07, "loss": 0.1502, "step": 74350 }, { "epoch": 1.4580392156862745, "grad_norm": 3.430820941925049, "learning_rate": 1.8768872379411992e-07, "loss": 0.12, "step": 74360 }, { "epoch": 1.458235294117647, "grad_norm": 2.533745050430298, "learning_rate": 1.8756214816176485e-07, "loss": 0.1541, "step": 74370 }, { "epoch": 1.4584313725490197, "grad_norm": 2.4187068939208984, "learning_rate": 1.874356053703688e-07, "loss": 0.1327, "step": 74380 }, { "epoch": 1.458627450980392, "grad_norm": 1.4382168054580688, "learning_rate": 1.8730909543323276e-07, "loss": 0.1096, "step": 74390 }, { "epoch": 1.4588235294117646, "grad_norm": 2.545210123062134, "learning_rate": 1.8718261836365473e-07, "loss": 0.1281, "step": 74400 }, { "epoch": 1.4590196078431372, "grad_norm": 24.5059814453125, "learning_rate": 1.870561741749288e-07, "loss": 0.144, "step": 74410 }, { "epoch": 1.4592156862745098, "grad_norm": 1.8354289531707764, "learning_rate": 1.8692976288034583e-07, "loss": 0.1328, "step": 74420 }, { "epoch": 1.4594117647058824, "grad_norm": 2.0708465576171875, "learning_rate": 1.86803384493193e-07, "loss": 0.1623, "step": 74430 }, { "epoch": 1.4596078431372548, "grad_norm": 1.4941675662994385, "learning_rate": 1.8667703902675442e-07, "loss": 0.1373, "step": 74440 }, { "epoch": 1.4598039215686274, "grad_norm": 1.5459843873977661, "learning_rate": 1.8655072649431043e-07, "loss": 0.1735, "step": 74450 }, { "epoch": 1.46, "grad_norm": 2.0748565196990967, "learning_rate": 1.8642444690913783e-07, "loss": 0.1129, "step": 74460 }, { "epoch": 1.4601960784313726, "grad_norm": 4.810415267944336, "learning_rate": 1.8629820028451022e-07, "loss": 0.1669, "step": 74470 }, { "epoch": 1.4603921568627451, "grad_norm": 1.9563502073287964, "learning_rate": 1.861719866336978e-07, "loss": 0.1317, "step": 74480 }, { "epoch": 1.4605882352941175, "grad_norm": 2.2345259189605713, "learning_rate": 1.8604580596996695e-07, "loss": 0.1377, "step": 74490 }, { "epoch": 1.4607843137254901, "grad_norm": 6.650888442993164, "learning_rate": 1.8591965830658092e-07, "loss": 0.1417, "step": 74500 }, { "epoch": 1.4609803921568627, "grad_norm": 2.6028549671173096, "learning_rate": 1.8579354365679923e-07, "loss": 0.1412, "step": 74510 }, { "epoch": 1.4611764705882353, "grad_norm": 3.9067556858062744, "learning_rate": 1.856674620338779e-07, "loss": 0.1629, "step": 74520 }, { "epoch": 1.461372549019608, "grad_norm": 2.2809090614318848, "learning_rate": 1.855414134510699e-07, "loss": 0.1427, "step": 74530 }, { "epoch": 1.4615686274509803, "grad_norm": 1.329213261604309, "learning_rate": 1.8541539792162437e-07, "loss": 0.1526, "step": 74540 }, { "epoch": 1.461764705882353, "grad_norm": 2.4355454444885254, "learning_rate": 1.852894154587868e-07, "loss": 0.1473, "step": 74550 }, { "epoch": 1.4619607843137254, "grad_norm": 3.553709030151367, "learning_rate": 1.851634660757998e-07, "loss": 0.1461, "step": 74560 }, { "epoch": 1.462156862745098, "grad_norm": 2.9792377948760986, "learning_rate": 1.8503754978590202e-07, "loss": 0.1224, "step": 74570 }, { "epoch": 1.4623529411764706, "grad_norm": 4.7297163009643555, "learning_rate": 1.8491166660232877e-07, "loss": 0.1553, "step": 74580 }, { "epoch": 1.462549019607843, "grad_norm": 2.670179605484009, "learning_rate": 1.8478581653831172e-07, "loss": 0.1085, "step": 74590 }, { "epoch": 1.4627450980392158, "grad_norm": 4.8703227043151855, "learning_rate": 1.846599996070795e-07, "loss": 0.1725, "step": 74600 }, { "epoch": 1.4629411764705882, "grad_norm": 2.1802449226379395, "learning_rate": 1.8453421582185684e-07, "loss": 0.1607, "step": 74610 }, { "epoch": 1.4631372549019608, "grad_norm": 2.5541138648986816, "learning_rate": 1.84408465195865e-07, "loss": 0.1338, "step": 74620 }, { "epoch": 1.4633333333333334, "grad_norm": 4.670895099639893, "learning_rate": 1.8428274774232212e-07, "loss": 0.1244, "step": 74630 }, { "epoch": 1.4635294117647057, "grad_norm": 1.6763640642166138, "learning_rate": 1.8415706347444255e-07, "loss": 0.1228, "step": 74640 }, { "epoch": 1.4637254901960786, "grad_norm": 3.4202308654785156, "learning_rate": 1.8403141240543674e-07, "loss": 0.2119, "step": 74650 }, { "epoch": 1.463921568627451, "grad_norm": 1.6012942790985107, "learning_rate": 1.8390579454851291e-07, "loss": 0.1546, "step": 74660 }, { "epoch": 1.4641176470588235, "grad_norm": 3.343010187149048, "learning_rate": 1.8378020991687458e-07, "loss": 0.1563, "step": 74670 }, { "epoch": 1.4643137254901961, "grad_norm": 4.653092384338379, "learning_rate": 1.8365465852372213e-07, "loss": 0.1413, "step": 74680 }, { "epoch": 1.4645098039215687, "grad_norm": 4.855581283569336, "learning_rate": 1.8352914038225282e-07, "loss": 0.1237, "step": 74690 }, { "epoch": 1.4647058823529413, "grad_norm": 3.069239854812622, "learning_rate": 1.8340365550565983e-07, "loss": 0.1425, "step": 74700 }, { "epoch": 1.4649019607843137, "grad_norm": 3.100480318069458, "learning_rate": 1.8327820390713311e-07, "loss": 0.1522, "step": 74710 }, { "epoch": 1.4650980392156863, "grad_norm": 2.920184850692749, "learning_rate": 1.8315278559985935e-07, "loss": 0.1417, "step": 74720 }, { "epoch": 1.4652941176470589, "grad_norm": 2.4603142738342285, "learning_rate": 1.8302740059702144e-07, "loss": 0.1141, "step": 74730 }, { "epoch": 1.4654901960784315, "grad_norm": 2.3625760078430176, "learning_rate": 1.8290204891179856e-07, "loss": 0.1158, "step": 74740 }, { "epoch": 1.465686274509804, "grad_norm": 1.9811047315597534, "learning_rate": 1.8277673055736713e-07, "loss": 0.1379, "step": 74750 }, { "epoch": 1.4658823529411764, "grad_norm": 3.508924722671509, "learning_rate": 1.8265144554689938e-07, "loss": 0.1454, "step": 74760 }, { "epoch": 1.466078431372549, "grad_norm": 2.8691213130950928, "learning_rate": 1.8252619389356421e-07, "loss": 0.1094, "step": 74770 }, { "epoch": 1.4662745098039216, "grad_norm": 1.8547929525375366, "learning_rate": 1.8240097561052704e-07, "loss": 0.1285, "step": 74780 }, { "epoch": 1.4664705882352942, "grad_norm": 1.2060304880142212, "learning_rate": 1.8227579071095005e-07, "loss": 0.1146, "step": 74790 }, { "epoch": 1.4666666666666668, "grad_norm": 2.0089244842529297, "learning_rate": 1.821506392079915e-07, "loss": 0.1343, "step": 74800 }, { "epoch": 1.4668627450980392, "grad_norm": 2.8013205528259277, "learning_rate": 1.8202552111480618e-07, "loss": 0.125, "step": 74810 }, { "epoch": 1.4670588235294117, "grad_norm": 3.9333720207214355, "learning_rate": 1.8190043644454584e-07, "loss": 0.1452, "step": 74820 }, { "epoch": 1.4672549019607843, "grad_norm": 4.380756378173828, "learning_rate": 1.81775385210358e-07, "loss": 0.1323, "step": 74830 }, { "epoch": 1.467450980392157, "grad_norm": 4.739504814147949, "learning_rate": 1.816503674253872e-07, "loss": 0.147, "step": 74840 }, { "epoch": 1.4676470588235295, "grad_norm": 3.1641581058502197, "learning_rate": 1.8152538310277455e-07, "loss": 0.114, "step": 74850 }, { "epoch": 1.467843137254902, "grad_norm": 1.438018560409546, "learning_rate": 1.814004322556572e-07, "loss": 0.1363, "step": 74860 }, { "epoch": 1.4680392156862745, "grad_norm": 3.7786548137664795, "learning_rate": 1.812755148971688e-07, "loss": 0.1458, "step": 74870 }, { "epoch": 1.468235294117647, "grad_norm": 2.3904640674591064, "learning_rate": 1.8115063104043994e-07, "loss": 0.1575, "step": 74880 }, { "epoch": 1.4684313725490197, "grad_norm": 1.710860013961792, "learning_rate": 1.8102578069859735e-07, "loss": 0.1276, "step": 74890 }, { "epoch": 1.4686274509803923, "grad_norm": 2.7049806118011475, "learning_rate": 1.8090096388476407e-07, "loss": 0.1251, "step": 74900 }, { "epoch": 1.4688235294117646, "grad_norm": 1.8849821090698242, "learning_rate": 1.8077618061206013e-07, "loss": 0.139, "step": 74910 }, { "epoch": 1.4690196078431372, "grad_norm": 1.9468613862991333, "learning_rate": 1.806514308936017e-07, "loss": 0.1433, "step": 74920 }, { "epoch": 1.4692156862745098, "grad_norm": 2.6455020904541016, "learning_rate": 1.8052671474250136e-07, "loss": 0.1334, "step": 74930 }, { "epoch": 1.4694117647058824, "grad_norm": 2.380258321762085, "learning_rate": 1.804020321718681e-07, "loss": 0.14, "step": 74940 }, { "epoch": 1.469607843137255, "grad_norm": 1.949549913406372, "learning_rate": 1.80277383194808e-07, "loss": 0.1171, "step": 74950 }, { "epoch": 1.4698039215686274, "grad_norm": 1.6801376342773438, "learning_rate": 1.801527678244229e-07, "loss": 0.1193, "step": 74960 }, { "epoch": 1.47, "grad_norm": 1.7720850706100464, "learning_rate": 1.800281860738112e-07, "loss": 0.1235, "step": 74970 }, { "epoch": 1.4701960784313726, "grad_norm": 3.049842357635498, "learning_rate": 1.7990363795606827e-07, "loss": 0.1211, "step": 74980 }, { "epoch": 1.4703921568627452, "grad_norm": 3.9697721004486084, "learning_rate": 1.7977912348428554e-07, "loss": 0.1517, "step": 74990 }, { "epoch": 1.4705882352941178, "grad_norm": 1.6891857385635376, "learning_rate": 1.7965464267155062e-07, "loss": 0.1421, "step": 75000 }, { "epoch": 1.4707843137254901, "grad_norm": 2.6315977573394775, "learning_rate": 1.7953019553094827e-07, "loss": 0.1467, "step": 75010 }, { "epoch": 1.4709803921568627, "grad_norm": 2.8312525749206543, "learning_rate": 1.7940578207555946e-07, "loss": 0.1602, "step": 75020 }, { "epoch": 1.4711764705882353, "grad_norm": 3.060295343399048, "learning_rate": 1.7928140231846123e-07, "loss": 0.1352, "step": 75030 }, { "epoch": 1.471372549019608, "grad_norm": 2.5026423931121826, "learning_rate": 1.791570562727277e-07, "loss": 0.1527, "step": 75040 }, { "epoch": 1.4715686274509805, "grad_norm": 2.674053192138672, "learning_rate": 1.7903274395142892e-07, "loss": 0.1337, "step": 75050 }, { "epoch": 1.4717647058823529, "grad_norm": 2.0541727542877197, "learning_rate": 1.7890846536763154e-07, "loss": 0.144, "step": 75060 }, { "epoch": 1.4719607843137255, "grad_norm": 1.5823417901992798, "learning_rate": 1.7878422053439896e-07, "loss": 0.1565, "step": 75070 }, { "epoch": 1.472156862745098, "grad_norm": 3.3855233192443848, "learning_rate": 1.7866000946479066e-07, "loss": 0.1498, "step": 75080 }, { "epoch": 1.4723529411764706, "grad_norm": 3.3421311378479004, "learning_rate": 1.7853583217186257e-07, "loss": 0.1479, "step": 75090 }, { "epoch": 1.4725490196078432, "grad_norm": 2.753316640853882, "learning_rate": 1.7841168866866756e-07, "loss": 0.123, "step": 75100 }, { "epoch": 1.4727450980392156, "grad_norm": 2.6792712211608887, "learning_rate": 1.7828757896825437e-07, "loss": 0.1807, "step": 75110 }, { "epoch": 1.4729411764705882, "grad_norm": 1.4439517259597778, "learning_rate": 1.7816350308366846e-07, "loss": 0.1343, "step": 75120 }, { "epoch": 1.4731372549019608, "grad_norm": 2.8273391723632812, "learning_rate": 1.7803946102795153e-07, "loss": 0.1485, "step": 75130 }, { "epoch": 1.4733333333333334, "grad_norm": 4.478643894195557, "learning_rate": 1.7791545281414217e-07, "loss": 0.1337, "step": 75140 }, { "epoch": 1.473529411764706, "grad_norm": 3.204296588897705, "learning_rate": 1.7779147845527504e-07, "loss": 0.1358, "step": 75150 }, { "epoch": 1.4737254901960783, "grad_norm": 3.6306521892547607, "learning_rate": 1.7766753796438106e-07, "loss": 0.1278, "step": 75160 }, { "epoch": 1.473921568627451, "grad_norm": 2.153322696685791, "learning_rate": 1.7754363135448824e-07, "loss": 0.1318, "step": 75170 }, { "epoch": 1.4741176470588235, "grad_norm": 1.8763412237167358, "learning_rate": 1.7741975863862052e-07, "loss": 0.1449, "step": 75180 }, { "epoch": 1.4743137254901961, "grad_norm": 4.760504245758057, "learning_rate": 1.7729591982979802e-07, "loss": 0.1357, "step": 75190 }, { "epoch": 1.4745098039215687, "grad_norm": 2.9911842346191406, "learning_rate": 1.771721149410384e-07, "loss": 0.1206, "step": 75200 }, { "epoch": 1.474705882352941, "grad_norm": 2.0125207901000977, "learning_rate": 1.7704834398535468e-07, "loss": 0.1292, "step": 75210 }, { "epoch": 1.4749019607843137, "grad_norm": 2.4696402549743652, "learning_rate": 1.7692460697575644e-07, "loss": 0.1428, "step": 75220 }, { "epoch": 1.4750980392156863, "grad_norm": 3.0476667881011963, "learning_rate": 1.7680090392525031e-07, "loss": 0.0982, "step": 75230 }, { "epoch": 1.4752941176470589, "grad_norm": 2.4536964893341064, "learning_rate": 1.766772348468389e-07, "loss": 0.1502, "step": 75240 }, { "epoch": 1.4754901960784315, "grad_norm": 2.8639888763427734, "learning_rate": 1.7655359975352092e-07, "loss": 0.1239, "step": 75250 }, { "epoch": 1.4756862745098038, "grad_norm": 2.1143646240234375, "learning_rate": 1.764299986582924e-07, "loss": 0.1539, "step": 75260 }, { "epoch": 1.4758823529411764, "grad_norm": 2.0032615661621094, "learning_rate": 1.7630643157414505e-07, "loss": 0.1701, "step": 75270 }, { "epoch": 1.476078431372549, "grad_norm": 1.8544723987579346, "learning_rate": 1.7618289851406732e-07, "loss": 0.1146, "step": 75280 }, { "epoch": 1.4762745098039216, "grad_norm": 2.284815788269043, "learning_rate": 1.7605939949104376e-07, "loss": 0.121, "step": 75290 }, { "epoch": 1.4764705882352942, "grad_norm": 2.2578649520874023, "learning_rate": 1.7593593451805598e-07, "loss": 0.1486, "step": 75300 }, { "epoch": 1.4766666666666666, "grad_norm": 1.9680193662643433, "learning_rate": 1.7581250360808152e-07, "loss": 0.1275, "step": 75310 }, { "epoch": 1.4768627450980392, "grad_norm": 2.3240795135498047, "learning_rate": 1.7568910677409417e-07, "loss": 0.1443, "step": 75320 }, { "epoch": 1.4770588235294118, "grad_norm": 4.746493816375732, "learning_rate": 1.7556574402906477e-07, "loss": 0.1395, "step": 75330 }, { "epoch": 1.4772549019607844, "grad_norm": 5.402227401733398, "learning_rate": 1.754424153859601e-07, "loss": 0.1494, "step": 75340 }, { "epoch": 1.477450980392157, "grad_norm": 2.042562246322632, "learning_rate": 1.753191208577433e-07, "loss": 0.1215, "step": 75350 }, { "epoch": 1.4776470588235293, "grad_norm": 2.4178242683410645, "learning_rate": 1.751958604573744e-07, "loss": 0.1602, "step": 75360 }, { "epoch": 1.477843137254902, "grad_norm": 6.3348612785339355, "learning_rate": 1.7507263419780926e-07, "loss": 0.1469, "step": 75370 }, { "epoch": 1.4780392156862745, "grad_norm": 2.2069764137268066, "learning_rate": 1.7494944209200057e-07, "loss": 0.1268, "step": 75380 }, { "epoch": 1.478235294117647, "grad_norm": 3.7041609287261963, "learning_rate": 1.7482628415289747e-07, "loss": 0.1586, "step": 75390 }, { "epoch": 1.4784313725490197, "grad_norm": 4.0052337646484375, "learning_rate": 1.7470316039344513e-07, "loss": 0.1134, "step": 75400 }, { "epoch": 1.478627450980392, "grad_norm": 4.188941955566406, "learning_rate": 1.745800708265852e-07, "loss": 0.1305, "step": 75410 }, { "epoch": 1.4788235294117646, "grad_norm": 1.6759568452835083, "learning_rate": 1.744570154652562e-07, "loss": 0.127, "step": 75420 }, { "epoch": 1.4790196078431372, "grad_norm": 3.0596537590026855, "learning_rate": 1.743339943223926e-07, "loss": 0.1351, "step": 75430 }, { "epoch": 1.4792156862745098, "grad_norm": 2.4324867725372314, "learning_rate": 1.7421100741092514e-07, "loss": 0.1144, "step": 75440 }, { "epoch": 1.4794117647058824, "grad_norm": 3.389589548110962, "learning_rate": 1.7408805474378164e-07, "loss": 0.1288, "step": 75450 }, { "epoch": 1.4796078431372548, "grad_norm": 1.6620147228240967, "learning_rate": 1.7396513633388564e-07, "loss": 0.1159, "step": 75460 }, { "epoch": 1.4798039215686274, "grad_norm": 4.5850830078125, "learning_rate": 1.7384225219415739e-07, "loss": 0.1418, "step": 75470 }, { "epoch": 1.48, "grad_norm": 4.487499713897705, "learning_rate": 1.737194023375133e-07, "loss": 0.1345, "step": 75480 }, { "epoch": 1.4801960784313726, "grad_norm": 1.354351282119751, "learning_rate": 1.7359658677686672e-07, "loss": 0.1219, "step": 75490 }, { "epoch": 1.4803921568627452, "grad_norm": 2.135150671005249, "learning_rate": 1.7347380552512685e-07, "loss": 0.1071, "step": 75500 }, { "epoch": 1.4805882352941175, "grad_norm": 2.3063275814056396, "learning_rate": 1.7335105859519928e-07, "loss": 0.1384, "step": 75510 }, { "epoch": 1.4807843137254901, "grad_norm": 1.1137495040893555, "learning_rate": 1.7322834599998653e-07, "loss": 0.1523, "step": 75520 }, { "epoch": 1.4809803921568627, "grad_norm": 1.9127286672592163, "learning_rate": 1.7310566775238705e-07, "loss": 0.1232, "step": 75530 }, { "epoch": 1.4811764705882353, "grad_norm": 0.9684649109840393, "learning_rate": 1.7298302386529557e-07, "loss": 0.1348, "step": 75540 }, { "epoch": 1.481372549019608, "grad_norm": 4.034304618835449, "learning_rate": 1.7286041435160358e-07, "loss": 0.1246, "step": 75550 }, { "epoch": 1.4815686274509803, "grad_norm": 2.2282650470733643, "learning_rate": 1.7273783922419905e-07, "loss": 0.1369, "step": 75560 }, { "epoch": 1.4817647058823529, "grad_norm": 2.409097671508789, "learning_rate": 1.7261529849596567e-07, "loss": 0.1357, "step": 75570 }, { "epoch": 1.4819607843137255, "grad_norm": 2.2588508129119873, "learning_rate": 1.7249279217978435e-07, "loss": 0.1278, "step": 75580 }, { "epoch": 1.482156862745098, "grad_norm": 4.949796199798584, "learning_rate": 1.7237032028853176e-07, "loss": 0.1256, "step": 75590 }, { "epoch": 1.4823529411764707, "grad_norm": 3.3845744132995605, "learning_rate": 1.72247882835081e-07, "loss": 0.1387, "step": 75600 }, { "epoch": 1.482549019607843, "grad_norm": 2.4285924434661865, "learning_rate": 1.7212547983230207e-07, "loss": 0.1193, "step": 75610 }, { "epoch": 1.4827450980392156, "grad_norm": 3.1936469078063965, "learning_rate": 1.720031112930608e-07, "loss": 0.114, "step": 75620 }, { "epoch": 1.4829411764705882, "grad_norm": 2.874737024307251, "learning_rate": 1.7188077723021955e-07, "loss": 0.1181, "step": 75630 }, { "epoch": 1.4831372549019608, "grad_norm": 3.376587390899658, "learning_rate": 1.7175847765663705e-07, "loss": 0.16, "step": 75640 }, { "epoch": 1.4833333333333334, "grad_norm": 4.421582221984863, "learning_rate": 1.7163621258516863e-07, "loss": 0.1266, "step": 75650 }, { "epoch": 1.4835294117647058, "grad_norm": 1.75730562210083, "learning_rate": 1.7151398202866574e-07, "loss": 0.1297, "step": 75660 }, { "epoch": 1.4837254901960784, "grad_norm": 1.786429524421692, "learning_rate": 1.7139178599997612e-07, "loss": 0.1434, "step": 75670 }, { "epoch": 1.483921568627451, "grad_norm": 3.572390556335449, "learning_rate": 1.7126962451194426e-07, "loss": 0.1499, "step": 75680 }, { "epoch": 1.4841176470588235, "grad_norm": 13.34300422668457, "learning_rate": 1.7114749757741076e-07, "loss": 0.136, "step": 75690 }, { "epoch": 1.4843137254901961, "grad_norm": 1.6877844333648682, "learning_rate": 1.7102540520921242e-07, "loss": 0.1243, "step": 75700 }, { "epoch": 1.4845098039215685, "grad_norm": 1.7312217950820923, "learning_rate": 1.7090334742018292e-07, "loss": 0.1335, "step": 75710 }, { "epoch": 1.484705882352941, "grad_norm": 3.8346338272094727, "learning_rate": 1.7078132422315177e-07, "loss": 0.1576, "step": 75720 }, { "epoch": 1.4849019607843137, "grad_norm": 3.8978731632232666, "learning_rate": 1.7065933563094488e-07, "loss": 0.1217, "step": 75730 }, { "epoch": 1.4850980392156863, "grad_norm": 4.923848628997803, "learning_rate": 1.7053738165638525e-07, "loss": 0.133, "step": 75740 }, { "epoch": 1.4852941176470589, "grad_norm": 1.7735018730163574, "learning_rate": 1.7041546231229143e-07, "loss": 0.1295, "step": 75750 }, { "epoch": 1.4854901960784312, "grad_norm": 2.8926315307617188, "learning_rate": 1.7029357761147844e-07, "loss": 0.1114, "step": 75760 }, { "epoch": 1.4856862745098038, "grad_norm": 3.9534289836883545, "learning_rate": 1.7017172756675812e-07, "loss": 0.1244, "step": 75770 }, { "epoch": 1.4858823529411764, "grad_norm": 1.4039864540100098, "learning_rate": 1.7004991219093824e-07, "loss": 0.1284, "step": 75780 }, { "epoch": 1.486078431372549, "grad_norm": 3.38309645652771, "learning_rate": 1.6992813149682284e-07, "loss": 0.165, "step": 75790 }, { "epoch": 1.4862745098039216, "grad_norm": 4.081636905670166, "learning_rate": 1.6980638549721287e-07, "loss": 0.1456, "step": 75800 }, { "epoch": 1.486470588235294, "grad_norm": 2.7960104942321777, "learning_rate": 1.6968467420490517e-07, "loss": 0.1277, "step": 75810 }, { "epoch": 1.4866666666666668, "grad_norm": 2.163067579269409, "learning_rate": 1.6956299763269304e-07, "loss": 0.1727, "step": 75820 }, { "epoch": 1.4868627450980392, "grad_norm": 5.015048027038574, "learning_rate": 1.69441355793366e-07, "loss": 0.119, "step": 75830 }, { "epoch": 1.4870588235294118, "grad_norm": 2.5567216873168945, "learning_rate": 1.6931974869971028e-07, "loss": 0.1278, "step": 75840 }, { "epoch": 1.4872549019607844, "grad_norm": 2.109712600708008, "learning_rate": 1.6919817636450822e-07, "loss": 0.1292, "step": 75850 }, { "epoch": 1.4874509803921567, "grad_norm": 11.58813762664795, "learning_rate": 1.6907663880053829e-07, "loss": 0.1541, "step": 75860 }, { "epoch": 1.4876470588235295, "grad_norm": 3.337066888809204, "learning_rate": 1.6895513602057587e-07, "loss": 0.1449, "step": 75870 }, { "epoch": 1.487843137254902, "grad_norm": 3.4876956939697266, "learning_rate": 1.6883366803739224e-07, "loss": 0.1527, "step": 75880 }, { "epoch": 1.4880392156862745, "grad_norm": 3.2099146842956543, "learning_rate": 1.687122348637549e-07, "loss": 0.1387, "step": 75890 }, { "epoch": 1.488235294117647, "grad_norm": 3.692063808441162, "learning_rate": 1.6859083651242816e-07, "loss": 0.1594, "step": 75900 }, { "epoch": 1.4884313725490195, "grad_norm": 4.740246772766113, "learning_rate": 1.684694729961726e-07, "loss": 0.1318, "step": 75910 }, { "epoch": 1.4886274509803923, "grad_norm": 3.328282356262207, "learning_rate": 1.683481443277446e-07, "loss": 0.1378, "step": 75920 }, { "epoch": 1.4888235294117647, "grad_norm": 2.69073486328125, "learning_rate": 1.682268505198976e-07, "loss": 0.1205, "step": 75930 }, { "epoch": 1.4890196078431372, "grad_norm": 2.8363213539123535, "learning_rate": 1.681055915853809e-07, "loss": 0.1455, "step": 75940 }, { "epoch": 1.4892156862745098, "grad_norm": 4.189370632171631, "learning_rate": 1.6798436753694013e-07, "loss": 0.1562, "step": 75950 }, { "epoch": 1.4894117647058824, "grad_norm": 1.8601635694503784, "learning_rate": 1.678631783873176e-07, "loss": 0.1832, "step": 75960 }, { "epoch": 1.489607843137255, "grad_norm": 4.202711582183838, "learning_rate": 1.677420241492517e-07, "loss": 0.1369, "step": 75970 }, { "epoch": 1.4898039215686274, "grad_norm": 3.772278308868408, "learning_rate": 1.6762090483547713e-07, "loss": 0.1207, "step": 75980 }, { "epoch": 1.49, "grad_norm": 2.6849193572998047, "learning_rate": 1.674998204587248e-07, "loss": 0.14, "step": 75990 }, { "epoch": 1.4901960784313726, "grad_norm": 4.505634784698486, "learning_rate": 1.6737877103172254e-07, "loss": 0.1069, "step": 76000 }, { "epoch": 1.4903921568627452, "grad_norm": 1.4936951398849487, "learning_rate": 1.6725775656719382e-07, "loss": 0.1463, "step": 76010 }, { "epoch": 1.4905882352941178, "grad_norm": 1.9121609926223755, "learning_rate": 1.6713677707785861e-07, "loss": 0.1215, "step": 76020 }, { "epoch": 1.4907843137254901, "grad_norm": 4.520946502685547, "learning_rate": 1.670158325764336e-07, "loss": 0.1652, "step": 76030 }, { "epoch": 1.4909803921568627, "grad_norm": 1.8298925161361694, "learning_rate": 1.6689492307563136e-07, "loss": 0.1518, "step": 76040 }, { "epoch": 1.4911764705882353, "grad_norm": 2.4144601821899414, "learning_rate": 1.667740485881608e-07, "loss": 0.1384, "step": 76050 }, { "epoch": 1.491372549019608, "grad_norm": 2.2154088020324707, "learning_rate": 1.6665320912672753e-07, "loss": 0.1161, "step": 76060 }, { "epoch": 1.4915686274509805, "grad_norm": 2.722443103790283, "learning_rate": 1.6653240470403312e-07, "loss": 0.1544, "step": 76070 }, { "epoch": 1.4917647058823529, "grad_norm": 2.381964921951294, "learning_rate": 1.6641163533277518e-07, "loss": 0.1412, "step": 76080 }, { "epoch": 1.4919607843137255, "grad_norm": 5.8045220375061035, "learning_rate": 1.662909010256487e-07, "loss": 0.1598, "step": 76090 }, { "epoch": 1.492156862745098, "grad_norm": 1.379010796546936, "learning_rate": 1.6617020179534398e-07, "loss": 0.1352, "step": 76100 }, { "epoch": 1.4923529411764707, "grad_norm": 4.775455951690674, "learning_rate": 1.6604953765454778e-07, "loss": 0.1568, "step": 76110 }, { "epoch": 1.4925490196078433, "grad_norm": 2.744558572769165, "learning_rate": 1.659289086159437e-07, "loss": 0.136, "step": 76120 }, { "epoch": 1.4927450980392156, "grad_norm": 2.7336063385009766, "learning_rate": 1.6580831469221107e-07, "loss": 0.1406, "step": 76130 }, { "epoch": 1.4929411764705882, "grad_norm": 2.4820380210876465, "learning_rate": 1.6568775589602568e-07, "loss": 0.1404, "step": 76140 }, { "epoch": 1.4931372549019608, "grad_norm": 2.7851765155792236, "learning_rate": 1.6556723224005997e-07, "loss": 0.164, "step": 76150 }, { "epoch": 1.4933333333333334, "grad_norm": 5.606532096862793, "learning_rate": 1.6544674373698225e-07, "loss": 0.1579, "step": 76160 }, { "epoch": 1.493529411764706, "grad_norm": 2.6605045795440674, "learning_rate": 1.6532629039945732e-07, "loss": 0.1517, "step": 76170 }, { "epoch": 1.4937254901960784, "grad_norm": 5.246696949005127, "learning_rate": 1.6520587224014614e-07, "loss": 0.1345, "step": 76180 }, { "epoch": 1.493921568627451, "grad_norm": 3.595540761947632, "learning_rate": 1.6508548927170636e-07, "loss": 0.1231, "step": 76190 }, { "epoch": 1.4941176470588236, "grad_norm": 1.9881423711776733, "learning_rate": 1.6496514150679152e-07, "loss": 0.141, "step": 76200 }, { "epoch": 1.4943137254901961, "grad_norm": 16.223608016967773, "learning_rate": 1.6484482895805152e-07, "loss": 0.1441, "step": 76210 }, { "epoch": 1.4945098039215687, "grad_norm": 2.702725887298584, "learning_rate": 1.6472455163813292e-07, "loss": 0.1449, "step": 76220 }, { "epoch": 1.494705882352941, "grad_norm": 6.380324840545654, "learning_rate": 1.6460430955967812e-07, "loss": 0.1329, "step": 76230 }, { "epoch": 1.4949019607843137, "grad_norm": 3.6976678371429443, "learning_rate": 1.6448410273532588e-07, "loss": 0.1333, "step": 76240 }, { "epoch": 1.4950980392156863, "grad_norm": 3.743689775466919, "learning_rate": 1.6436393117771168e-07, "loss": 0.1611, "step": 76250 }, { "epoch": 1.4952941176470589, "grad_norm": 2.780235528945923, "learning_rate": 1.6424379489946666e-07, "loss": 0.1565, "step": 76260 }, { "epoch": 1.4954901960784315, "grad_norm": 2.1263468265533447, "learning_rate": 1.641236939132188e-07, "loss": 0.1303, "step": 76270 }, { "epoch": 1.4956862745098038, "grad_norm": 3.164792776107788, "learning_rate": 1.6400362823159225e-07, "loss": 0.1314, "step": 76280 }, { "epoch": 1.4958823529411764, "grad_norm": 4.2647833824157715, "learning_rate": 1.6388359786720724e-07, "loss": 0.132, "step": 76290 }, { "epoch": 1.496078431372549, "grad_norm": 2.2111713886260986, "learning_rate": 1.6376360283268015e-07, "loss": 0.1205, "step": 76300 }, { "epoch": 1.4962745098039216, "grad_norm": 1.6657564640045166, "learning_rate": 1.6364364314062429e-07, "loss": 0.0996, "step": 76310 }, { "epoch": 1.4964705882352942, "grad_norm": 3.9859259128570557, "learning_rate": 1.6352371880364867e-07, "loss": 0.1177, "step": 76320 }, { "epoch": 1.4966666666666666, "grad_norm": 3.666992425918579, "learning_rate": 1.6340382983435886e-07, "loss": 0.141, "step": 76330 }, { "epoch": 1.4968627450980392, "grad_norm": 2.0414719581604004, "learning_rate": 1.6328397624535634e-07, "loss": 0.1258, "step": 76340 }, { "epoch": 1.4970588235294118, "grad_norm": 3.581366539001465, "learning_rate": 1.631641580492395e-07, "loss": 0.1361, "step": 76350 }, { "epoch": 1.4972549019607844, "grad_norm": 2.627896547317505, "learning_rate": 1.630443752586026e-07, "loss": 0.1104, "step": 76360 }, { "epoch": 1.497450980392157, "grad_norm": 2.69380784034729, "learning_rate": 1.6292462788603594e-07, "loss": 0.1027, "step": 76370 }, { "epoch": 1.4976470588235293, "grad_norm": 1.448268175125122, "learning_rate": 1.6280491594412687e-07, "loss": 0.1246, "step": 76380 }, { "epoch": 1.497843137254902, "grad_norm": 2.7326438426971436, "learning_rate": 1.626852394454583e-07, "loss": 0.1618, "step": 76390 }, { "epoch": 1.4980392156862745, "grad_norm": 1.8727549314498901, "learning_rate": 1.6256559840260958e-07, "loss": 0.1435, "step": 76400 }, { "epoch": 1.498235294117647, "grad_norm": 4.014744758605957, "learning_rate": 1.6244599282815663e-07, "loss": 0.1161, "step": 76410 }, { "epoch": 1.4984313725490197, "grad_norm": 2.513834238052368, "learning_rate": 1.623264227346713e-07, "loss": 0.1172, "step": 76420 }, { "epoch": 1.498627450980392, "grad_norm": 1.9204463958740234, "learning_rate": 1.622068881347217e-07, "loss": 0.132, "step": 76430 }, { "epoch": 1.4988235294117647, "grad_norm": 4.561342239379883, "learning_rate": 1.6208738904087254e-07, "loss": 0.1683, "step": 76440 }, { "epoch": 1.4990196078431373, "grad_norm": 3.3719232082366943, "learning_rate": 1.6196792546568472e-07, "loss": 0.1318, "step": 76450 }, { "epoch": 1.4992156862745099, "grad_norm": 8.08061408996582, "learning_rate": 1.6184849742171503e-07, "loss": 0.1295, "step": 76460 }, { "epoch": 1.4994117647058824, "grad_norm": 1.597764015197754, "learning_rate": 1.6172910492151699e-07, "loss": 0.144, "step": 76470 }, { "epoch": 1.4996078431372548, "grad_norm": 4.795874118804932, "learning_rate": 1.616097479776401e-07, "loss": 0.1359, "step": 76480 }, { "epoch": 1.4998039215686274, "grad_norm": 2.617490291595459, "learning_rate": 1.6149042660263007e-07, "loss": 0.1413, "step": 76490 }, { "epoch": 1.5, "grad_norm": 0.7514553070068359, "learning_rate": 1.6137114080902925e-07, "loss": 0.1568, "step": 76500 }, { "epoch": 1.5001960784313726, "grad_norm": 3.6833975315093994, "learning_rate": 1.6125189060937593e-07, "loss": 0.1407, "step": 76510 }, { "epoch": 1.5003921568627452, "grad_norm": 3.334094524383545, "learning_rate": 1.6113267601620467e-07, "loss": 0.1439, "step": 76520 }, { "epoch": 1.5005882352941176, "grad_norm": 4.263716220855713, "learning_rate": 1.610134970420462e-07, "loss": 0.1437, "step": 76530 }, { "epoch": 1.5007843137254901, "grad_norm": 2.5064146518707275, "learning_rate": 1.60894353699428e-07, "loss": 0.1325, "step": 76540 }, { "epoch": 1.5009803921568627, "grad_norm": 4.450088024139404, "learning_rate": 1.6077524600087327e-07, "loss": 0.1593, "step": 76550 }, { "epoch": 1.5011764705882353, "grad_norm": 2.3275890350341797, "learning_rate": 1.6065617395890157e-07, "loss": 0.1031, "step": 76560 }, { "epoch": 1.501372549019608, "grad_norm": 3.965893507003784, "learning_rate": 1.6053713758602898e-07, "loss": 0.1349, "step": 76570 }, { "epoch": 1.5015686274509803, "grad_norm": 2.990725517272949, "learning_rate": 1.6041813689476764e-07, "loss": 0.1102, "step": 76580 }, { "epoch": 1.501764705882353, "grad_norm": 3.8641035556793213, "learning_rate": 1.602991718976257e-07, "loss": 0.1191, "step": 76590 }, { "epoch": 1.5019607843137255, "grad_norm": 1.6902812719345093, "learning_rate": 1.6018024260710817e-07, "loss": 0.151, "step": 76600 }, { "epoch": 1.502156862745098, "grad_norm": 2.2397594451904297, "learning_rate": 1.6006134903571572e-07, "loss": 0.1333, "step": 76610 }, { "epoch": 1.5023529411764707, "grad_norm": 1.6531291007995605, "learning_rate": 1.5994249119594516e-07, "loss": 0.1407, "step": 76620 }, { "epoch": 1.502549019607843, "grad_norm": 3.4720003604888916, "learning_rate": 1.598236691002906e-07, "loss": 0.1456, "step": 76630 }, { "epoch": 1.5027450980392156, "grad_norm": 3.747227191925049, "learning_rate": 1.5970488276124133e-07, "loss": 0.1435, "step": 76640 }, { "epoch": 1.5029411764705882, "grad_norm": 2.3198492527008057, "learning_rate": 1.5958613219128298e-07, "loss": 0.1228, "step": 76650 }, { "epoch": 1.5031372549019608, "grad_norm": 1.9448282718658447, "learning_rate": 1.5946741740289804e-07, "loss": 0.1117, "step": 76660 }, { "epoch": 1.5033333333333334, "grad_norm": 4.177764892578125, "learning_rate": 1.593487384085647e-07, "loss": 0.1328, "step": 76670 }, { "epoch": 1.5035294117647058, "grad_norm": 4.630558490753174, "learning_rate": 1.592300952207575e-07, "loss": 0.162, "step": 76680 }, { "epoch": 1.5037254901960786, "grad_norm": 1.7471181154251099, "learning_rate": 1.591114878519471e-07, "loss": 0.1543, "step": 76690 }, { "epoch": 1.503921568627451, "grad_norm": 2.1130778789520264, "learning_rate": 1.5899291631460094e-07, "loss": 0.1032, "step": 76700 }, { "epoch": 1.5041176470588236, "grad_norm": 1.0411797761917114, "learning_rate": 1.588743806211821e-07, "loss": 0.1234, "step": 76710 }, { "epoch": 1.5043137254901962, "grad_norm": 2.537968873977661, "learning_rate": 1.5875588078414992e-07, "loss": 0.1427, "step": 76720 }, { "epoch": 1.5045098039215685, "grad_norm": 3.0130813121795654, "learning_rate": 1.586374168159605e-07, "loss": 0.1462, "step": 76730 }, { "epoch": 1.5047058823529413, "grad_norm": 3.0032711029052734, "learning_rate": 1.5851898872906566e-07, "loss": 0.1075, "step": 76740 }, { "epoch": 1.5049019607843137, "grad_norm": 6.917596340179443, "learning_rate": 1.5840059653591346e-07, "loss": 0.1703, "step": 76750 }, { "epoch": 1.5050980392156863, "grad_norm": 2.009040117263794, "learning_rate": 1.5828224024894866e-07, "loss": 0.141, "step": 76760 }, { "epoch": 1.505294117647059, "grad_norm": 2.7240724563598633, "learning_rate": 1.5816391988061173e-07, "loss": 0.1318, "step": 76770 }, { "epoch": 1.5054901960784313, "grad_norm": 1.826346755027771, "learning_rate": 1.5804563544333938e-07, "loss": 0.1251, "step": 76780 }, { "epoch": 1.505686274509804, "grad_norm": 2.087123394012451, "learning_rate": 1.5792738694956515e-07, "loss": 0.1383, "step": 76790 }, { "epoch": 1.5058823529411764, "grad_norm": 3.7098307609558105, "learning_rate": 1.5780917441171792e-07, "loss": 0.1443, "step": 76800 }, { "epoch": 1.506078431372549, "grad_norm": 3.0217483043670654, "learning_rate": 1.5769099784222346e-07, "loss": 0.1262, "step": 76810 }, { "epoch": 1.5062745098039216, "grad_norm": 2.1060800552368164, "learning_rate": 1.5757285725350371e-07, "loss": 0.1524, "step": 76820 }, { "epoch": 1.506470588235294, "grad_norm": 2.076265811920166, "learning_rate": 1.5745475265797652e-07, "loss": 0.1449, "step": 76830 }, { "epoch": 1.5066666666666668, "grad_norm": 1.8097902536392212, "learning_rate": 1.573366840680559e-07, "loss": 0.122, "step": 76840 }, { "epoch": 1.5068627450980392, "grad_norm": 4.659647464752197, "learning_rate": 1.572186514961526e-07, "loss": 0.1638, "step": 76850 }, { "epoch": 1.5070588235294118, "grad_norm": 3.757636785507202, "learning_rate": 1.5710065495467311e-07, "loss": 0.1691, "step": 76860 }, { "epoch": 1.5072549019607844, "grad_norm": 2.654047727584839, "learning_rate": 1.5698269445602025e-07, "loss": 0.1187, "step": 76870 }, { "epoch": 1.5074509803921567, "grad_norm": 3.610267400741577, "learning_rate": 1.5686477001259295e-07, "loss": 0.1444, "step": 76880 }, { "epoch": 1.5076470588235296, "grad_norm": 2.6361639499664307, "learning_rate": 1.5674688163678678e-07, "loss": 0.1179, "step": 76890 }, { "epoch": 1.507843137254902, "grad_norm": 2.3373968601226807, "learning_rate": 1.566290293409931e-07, "loss": 0.1174, "step": 76900 }, { "epoch": 1.5080392156862745, "grad_norm": 3.225525379180908, "learning_rate": 1.565112131375994e-07, "loss": 0.1279, "step": 76910 }, { "epoch": 1.5082352941176471, "grad_norm": 5.134041786193848, "learning_rate": 1.5639343303898988e-07, "loss": 0.1423, "step": 76920 }, { "epoch": 1.5084313725490195, "grad_norm": 1.7988613843917847, "learning_rate": 1.562756890575445e-07, "loss": 0.1477, "step": 76930 }, { "epoch": 1.5086274509803923, "grad_norm": 2.7496535778045654, "learning_rate": 1.5615798120563945e-07, "loss": 0.1452, "step": 76940 }, { "epoch": 1.5088235294117647, "grad_norm": 1.5442137718200684, "learning_rate": 1.5604030949564744e-07, "loss": 0.1445, "step": 76950 }, { "epoch": 1.5090196078431373, "grad_norm": 3.196664571762085, "learning_rate": 1.5592267393993713e-07, "loss": 0.1492, "step": 76960 }, { "epoch": 1.5092156862745099, "grad_norm": 2.599721908569336, "learning_rate": 1.5580507455087316e-07, "loss": 0.1186, "step": 76970 }, { "epoch": 1.5094117647058822, "grad_norm": 1.7898547649383545, "learning_rate": 1.5568751134081693e-07, "loss": 0.1444, "step": 76980 }, { "epoch": 1.509607843137255, "grad_norm": 2.7087626457214355, "learning_rate": 1.5556998432212576e-07, "loss": 0.1176, "step": 76990 }, { "epoch": 1.5098039215686274, "grad_norm": 3.704984664916992, "learning_rate": 1.5545249350715295e-07, "loss": 0.1608, "step": 77000 }, { "epoch": 1.51, "grad_norm": 4.029630661010742, "learning_rate": 1.553350389082484e-07, "loss": 0.1437, "step": 77010 }, { "epoch": 1.5101960784313726, "grad_norm": 4.347019195556641, "learning_rate": 1.5521762053775788e-07, "loss": 0.1337, "step": 77020 }, { "epoch": 1.510392156862745, "grad_norm": 3.019008159637451, "learning_rate": 1.551002384080235e-07, "loss": 0.1249, "step": 77030 }, { "epoch": 1.5105882352941178, "grad_norm": 2.985170364379883, "learning_rate": 1.5498289253138335e-07, "loss": 0.1563, "step": 77040 }, { "epoch": 1.5107843137254902, "grad_norm": 1.4687999486923218, "learning_rate": 1.5486558292017215e-07, "loss": 0.1298, "step": 77050 }, { "epoch": 1.5109803921568628, "grad_norm": 2.178864002227783, "learning_rate": 1.5474830958672047e-07, "loss": 0.1369, "step": 77060 }, { "epoch": 1.5111764705882353, "grad_norm": 3.168283224105835, "learning_rate": 1.5463107254335493e-07, "loss": 0.1185, "step": 77070 }, { "epoch": 1.5113725490196077, "grad_norm": 2.223497152328491, "learning_rate": 1.5451387180239888e-07, "loss": 0.166, "step": 77080 }, { "epoch": 1.5115686274509805, "grad_norm": 3.202212333679199, "learning_rate": 1.5439670737617133e-07, "loss": 0.1382, "step": 77090 }, { "epoch": 1.511764705882353, "grad_norm": 4.2547197341918945, "learning_rate": 1.542795792769876e-07, "loss": 0.1317, "step": 77100 }, { "epoch": 1.5119607843137255, "grad_norm": 2.2544846534729004, "learning_rate": 1.5416248751715943e-07, "loss": 0.1243, "step": 77110 }, { "epoch": 1.512156862745098, "grad_norm": 1.9854037761688232, "learning_rate": 1.5404543210899456e-07, "loss": 0.1417, "step": 77120 }, { "epoch": 1.5123529411764705, "grad_norm": 1.987125039100647, "learning_rate": 1.5392841306479666e-07, "loss": 0.151, "step": 77130 }, { "epoch": 1.5125490196078433, "grad_norm": 3.4229047298431396, "learning_rate": 1.5381143039686612e-07, "loss": 0.1269, "step": 77140 }, { "epoch": 1.5127450980392156, "grad_norm": 3.357185125350952, "learning_rate": 1.536944841174992e-07, "loss": 0.1403, "step": 77150 }, { "epoch": 1.5129411764705882, "grad_norm": 2.2036852836608887, "learning_rate": 1.5357757423898787e-07, "loss": 0.1374, "step": 77160 }, { "epoch": 1.5131372549019608, "grad_norm": 4.934015274047852, "learning_rate": 1.534607007736215e-07, "loss": 0.1293, "step": 77170 }, { "epoch": 1.5133333333333332, "grad_norm": 2.1556191444396973, "learning_rate": 1.533438637336845e-07, "loss": 0.1311, "step": 77180 }, { "epoch": 1.513529411764706, "grad_norm": 3.719811201095581, "learning_rate": 1.5322706313145778e-07, "loss": 0.1453, "step": 77190 }, { "epoch": 1.5137254901960784, "grad_norm": 2.0339629650115967, "learning_rate": 1.5311029897921879e-07, "loss": 0.1368, "step": 77200 }, { "epoch": 1.513921568627451, "grad_norm": 1.494247317314148, "learning_rate": 1.5299357128924055e-07, "loss": 0.154, "step": 77210 }, { "epoch": 1.5141176470588236, "grad_norm": 1.8634825944900513, "learning_rate": 1.5287688007379268e-07, "loss": 0.1513, "step": 77220 }, { "epoch": 1.514313725490196, "grad_norm": 2.591132879257202, "learning_rate": 1.5276022534514062e-07, "loss": 0.1557, "step": 77230 }, { "epoch": 1.5145098039215688, "grad_norm": 2.7993977069854736, "learning_rate": 1.5264360711554642e-07, "loss": 0.1472, "step": 77240 }, { "epoch": 1.5147058823529411, "grad_norm": 7.842361927032471, "learning_rate": 1.5252702539726802e-07, "loss": 0.1222, "step": 77250 }, { "epoch": 1.5149019607843137, "grad_norm": 2.158365249633789, "learning_rate": 1.5241048020255932e-07, "loss": 0.1907, "step": 77260 }, { "epoch": 1.5150980392156863, "grad_norm": 4.724609375, "learning_rate": 1.522939715436709e-07, "loss": 0.1677, "step": 77270 }, { "epoch": 1.5152941176470587, "grad_norm": 1.9862793684005737, "learning_rate": 1.521774994328492e-07, "loss": 0.0937, "step": 77280 }, { "epoch": 1.5154901960784315, "grad_norm": 2.3954977989196777, "learning_rate": 1.5206106388233653e-07, "loss": 0.1588, "step": 77290 }, { "epoch": 1.5156862745098039, "grad_norm": 2.186244487762451, "learning_rate": 1.5194466490437202e-07, "loss": 0.1254, "step": 77300 }, { "epoch": 1.5158823529411765, "grad_norm": 1.5631318092346191, "learning_rate": 1.5182830251119043e-07, "loss": 0.1305, "step": 77310 }, { "epoch": 1.516078431372549, "grad_norm": 2.312382459640503, "learning_rate": 1.5171197671502272e-07, "loss": 0.1431, "step": 77320 }, { "epoch": 1.5162745098039214, "grad_norm": 2.2165029048919678, "learning_rate": 1.5159568752809633e-07, "loss": 0.1399, "step": 77330 }, { "epoch": 1.5164705882352942, "grad_norm": 2.883729934692383, "learning_rate": 1.5147943496263464e-07, "loss": 0.145, "step": 77340 }, { "epoch": 1.5166666666666666, "grad_norm": 3.9207940101623535, "learning_rate": 1.5136321903085696e-07, "loss": 0.1625, "step": 77350 }, { "epoch": 1.5168627450980392, "grad_norm": 1.7325365543365479, "learning_rate": 1.512470397449794e-07, "loss": 0.1113, "step": 77360 }, { "epoch": 1.5170588235294118, "grad_norm": 1.7753785848617554, "learning_rate": 1.5113089711721345e-07, "loss": 0.1106, "step": 77370 }, { "epoch": 1.5172549019607842, "grad_norm": 1.9940595626831055, "learning_rate": 1.5101479115976718e-07, "loss": 0.1265, "step": 77380 }, { "epoch": 1.517450980392157, "grad_norm": 2.525458335876465, "learning_rate": 1.5089872188484453e-07, "loss": 0.1152, "step": 77390 }, { "epoch": 1.5176470588235293, "grad_norm": 3.716296911239624, "learning_rate": 1.507826893046461e-07, "loss": 0.1426, "step": 77400 }, { "epoch": 1.517843137254902, "grad_norm": 3.946384906768799, "learning_rate": 1.5066669343136824e-07, "loss": 0.1124, "step": 77410 }, { "epoch": 1.5180392156862745, "grad_norm": 2.501760959625244, "learning_rate": 1.5055073427720315e-07, "loss": 0.1282, "step": 77420 }, { "epoch": 1.518235294117647, "grad_norm": 2.905280113220215, "learning_rate": 1.5043481185434005e-07, "loss": 0.1221, "step": 77430 }, { "epoch": 1.5184313725490197, "grad_norm": 2.4078357219696045, "learning_rate": 1.503189261749635e-07, "loss": 0.1164, "step": 77440 }, { "epoch": 1.518627450980392, "grad_norm": 2.0787999629974365, "learning_rate": 1.502030772512543e-07, "loss": 0.1422, "step": 77450 }, { "epoch": 1.5188235294117647, "grad_norm": 2.5617637634277344, "learning_rate": 1.5008726509538988e-07, "loss": 0.1502, "step": 77460 }, { "epoch": 1.5190196078431373, "grad_norm": 1.9065279960632324, "learning_rate": 1.4997148971954344e-07, "loss": 0.1213, "step": 77470 }, { "epoch": 1.5192156862745096, "grad_norm": 1.6011641025543213, "learning_rate": 1.4985575113588405e-07, "loss": 0.142, "step": 77480 }, { "epoch": 1.5194117647058825, "grad_norm": 3.2447032928466797, "learning_rate": 1.4974004935657759e-07, "loss": 0.1456, "step": 77490 }, { "epoch": 1.5196078431372548, "grad_norm": 2.173506021499634, "learning_rate": 1.496243843937856e-07, "loss": 0.1365, "step": 77500 }, { "epoch": 1.5198039215686274, "grad_norm": 1.0311870574951172, "learning_rate": 1.4950875625966547e-07, "loss": 0.1501, "step": 77510 }, { "epoch": 1.52, "grad_norm": 1.7883278131484985, "learning_rate": 1.4939316496637177e-07, "loss": 0.122, "step": 77520 }, { "epoch": 1.5201960784313724, "grad_norm": 2.0340185165405273, "learning_rate": 1.492776105260542e-07, "loss": 0.1188, "step": 77530 }, { "epoch": 1.5203921568627452, "grad_norm": 3.5291061401367188, "learning_rate": 1.4916209295085874e-07, "loss": 0.1296, "step": 77540 }, { "epoch": 1.5205882352941176, "grad_norm": 3.2614264488220215, "learning_rate": 1.4904661225292798e-07, "loss": 0.1481, "step": 77550 }, { "epoch": 1.5207843137254902, "grad_norm": 2.800096035003662, "learning_rate": 1.4893116844440023e-07, "loss": 0.1387, "step": 77560 }, { "epoch": 1.5209803921568628, "grad_norm": 2.4347052574157715, "learning_rate": 1.4881576153740993e-07, "loss": 0.1605, "step": 77570 }, { "epoch": 1.5211764705882351, "grad_norm": 3.3690645694732666, "learning_rate": 1.4870039154408763e-07, "loss": 0.1388, "step": 77580 }, { "epoch": 1.521372549019608, "grad_norm": 2.540653705596924, "learning_rate": 1.4858505847656038e-07, "loss": 0.1804, "step": 77590 }, { "epoch": 1.5215686274509803, "grad_norm": 2.9876184463500977, "learning_rate": 1.4846976234695096e-07, "loss": 0.1274, "step": 77600 }, { "epoch": 1.521764705882353, "grad_norm": 2.3390676975250244, "learning_rate": 1.483545031673782e-07, "loss": 0.1601, "step": 77610 }, { "epoch": 1.5219607843137255, "grad_norm": 1.7503972053527832, "learning_rate": 1.4823928094995747e-07, "loss": 0.1445, "step": 77620 }, { "epoch": 1.522156862745098, "grad_norm": 1.204720139503479, "learning_rate": 1.481240957067999e-07, "loss": 0.1531, "step": 77630 }, { "epoch": 1.5223529411764707, "grad_norm": 4.629509925842285, "learning_rate": 1.480089474500127e-07, "loss": 0.1244, "step": 77640 }, { "epoch": 1.522549019607843, "grad_norm": 2.482980251312256, "learning_rate": 1.478938361916996e-07, "loss": 0.1746, "step": 77650 }, { "epoch": 1.5227450980392156, "grad_norm": 2.2507269382476807, "learning_rate": 1.4777876194396006e-07, "loss": 0.1464, "step": 77660 }, { "epoch": 1.5229411764705882, "grad_norm": 3.1021389961242676, "learning_rate": 1.4766372471888954e-07, "loss": 0.1374, "step": 77670 }, { "epoch": 1.5231372549019608, "grad_norm": 3.793592691421509, "learning_rate": 1.475487245285802e-07, "loss": 0.1498, "step": 77680 }, { "epoch": 1.5233333333333334, "grad_norm": 2.406883716583252, "learning_rate": 1.474337613851196e-07, "loss": 0.1483, "step": 77690 }, { "epoch": 1.5235294117647058, "grad_norm": 2.0795836448669434, "learning_rate": 1.4731883530059186e-07, "loss": 0.1299, "step": 77700 }, { "epoch": 1.5237254901960784, "grad_norm": 1.7708884477615356, "learning_rate": 1.4720394628707732e-07, "loss": 0.1468, "step": 77710 }, { "epoch": 1.523921568627451, "grad_norm": 2.7037205696105957, "learning_rate": 1.4708909435665206e-07, "loss": 0.1042, "step": 77720 }, { "epoch": 1.5241176470588236, "grad_norm": 4.131012439727783, "learning_rate": 1.4697427952138825e-07, "loss": 0.1331, "step": 77730 }, { "epoch": 1.5243137254901962, "grad_norm": 2.3955938816070557, "learning_rate": 1.4685950179335427e-07, "loss": 0.1938, "step": 77740 }, { "epoch": 1.5245098039215685, "grad_norm": 1.8688963651657104, "learning_rate": 1.4674476118461492e-07, "loss": 0.125, "step": 77750 }, { "epoch": 1.5247058823529411, "grad_norm": 2.257736921310425, "learning_rate": 1.4663005770723064e-07, "loss": 0.1323, "step": 77760 }, { "epoch": 1.5249019607843137, "grad_norm": 5.749514579772949, "learning_rate": 1.4651539137325796e-07, "loss": 0.1298, "step": 77770 }, { "epoch": 1.5250980392156863, "grad_norm": 1.6397608518600464, "learning_rate": 1.4640076219475005e-07, "loss": 0.1166, "step": 77780 }, { "epoch": 1.525294117647059, "grad_norm": 2.5202696323394775, "learning_rate": 1.4628617018375562e-07, "loss": 0.1304, "step": 77790 }, { "epoch": 1.5254901960784313, "grad_norm": 2.482609272003174, "learning_rate": 1.4617161535231954e-07, "loss": 0.1367, "step": 77800 }, { "epoch": 1.5256862745098039, "grad_norm": 4.095216274261475, "learning_rate": 1.4605709771248314e-07, "loss": 0.1282, "step": 77810 }, { "epoch": 1.5258823529411765, "grad_norm": 2.480715274810791, "learning_rate": 1.459426172762835e-07, "loss": 0.1293, "step": 77820 }, { "epoch": 1.526078431372549, "grad_norm": 4.9740190505981445, "learning_rate": 1.4582817405575372e-07, "loss": 0.1342, "step": 77830 }, { "epoch": 1.5262745098039217, "grad_norm": 2.322859287261963, "learning_rate": 1.4571376806292341e-07, "loss": 0.1359, "step": 77840 }, { "epoch": 1.526470588235294, "grad_norm": 2.489748954772949, "learning_rate": 1.4559939930981791e-07, "loss": 0.133, "step": 77850 }, { "epoch": 1.5266666666666666, "grad_norm": 4.901412487030029, "learning_rate": 1.454850678084586e-07, "loss": 0.1478, "step": 77860 }, { "epoch": 1.5268627450980392, "grad_norm": 3.000319242477417, "learning_rate": 1.4537077357086318e-07, "loss": 0.1301, "step": 77870 }, { "epoch": 1.5270588235294118, "grad_norm": 2.819558620452881, "learning_rate": 1.4525651660904552e-07, "loss": 0.1292, "step": 77880 }, { "epoch": 1.5272549019607844, "grad_norm": 2.972402572631836, "learning_rate": 1.4514229693501516e-07, "loss": 0.1168, "step": 77890 }, { "epoch": 1.5274509803921568, "grad_norm": 4.296419143676758, "learning_rate": 1.4502811456077817e-07, "loss": 0.1281, "step": 77900 }, { "epoch": 1.5276470588235294, "grad_norm": 2.100459575653076, "learning_rate": 1.449139694983364e-07, "loss": 0.1171, "step": 77910 }, { "epoch": 1.527843137254902, "grad_norm": 2.736449956893921, "learning_rate": 1.4479986175968783e-07, "loss": 0.1404, "step": 77920 }, { "epoch": 1.5280392156862745, "grad_norm": 11.130356788635254, "learning_rate": 1.4468579135682635e-07, "loss": 0.1364, "step": 77930 }, { "epoch": 1.5282352941176471, "grad_norm": 1.398632526397705, "learning_rate": 1.4457175830174252e-07, "loss": 0.1464, "step": 77940 }, { "epoch": 1.5284313725490195, "grad_norm": 5.197631359100342, "learning_rate": 1.444577626064224e-07, "loss": 0.1488, "step": 77950 }, { "epoch": 1.5286274509803923, "grad_norm": 1.9198044538497925, "learning_rate": 1.4434380428284809e-07, "loss": 0.1266, "step": 77960 }, { "epoch": 1.5288235294117647, "grad_norm": 1.7998343706130981, "learning_rate": 1.4422988334299836e-07, "loss": 0.131, "step": 77970 }, { "epoch": 1.5290196078431373, "grad_norm": 2.5795094966888428, "learning_rate": 1.4411599979884743e-07, "loss": 0.1404, "step": 77980 }, { "epoch": 1.5292156862745099, "grad_norm": 5.467197418212891, "learning_rate": 1.4400215366236573e-07, "loss": 0.1108, "step": 77990 }, { "epoch": 1.5294117647058822, "grad_norm": 3.364102363586426, "learning_rate": 1.4388834494552017e-07, "loss": 0.1252, "step": 78000 }, { "epoch": 1.529607843137255, "grad_norm": 2.8805534839630127, "learning_rate": 1.4377457366027324e-07, "loss": 0.1412, "step": 78010 }, { "epoch": 1.5298039215686274, "grad_norm": 1.607322335243225, "learning_rate": 1.4366083981858347e-07, "loss": 0.1286, "step": 78020 }, { "epoch": 1.53, "grad_norm": 1.0492029190063477, "learning_rate": 1.4354714343240598e-07, "loss": 0.1277, "step": 78030 }, { "epoch": 1.5301960784313726, "grad_norm": 2.5320076942443848, "learning_rate": 1.4343348451369152e-07, "loss": 0.1406, "step": 78040 }, { "epoch": 1.530392156862745, "grad_norm": 3.8913393020629883, "learning_rate": 1.433198630743866e-07, "loss": 0.1323, "step": 78050 }, { "epoch": 1.5305882352941178, "grad_norm": 2.1798110008239746, "learning_rate": 1.432062791264349e-07, "loss": 0.1464, "step": 78060 }, { "epoch": 1.5307843137254902, "grad_norm": 5.878293514251709, "learning_rate": 1.4309273268177508e-07, "loss": 0.1284, "step": 78070 }, { "epoch": 1.5309803921568628, "grad_norm": 3.8093791007995605, "learning_rate": 1.4297922375234222e-07, "loss": 0.1235, "step": 78080 }, { "epoch": 1.5311764705882354, "grad_norm": 1.7363550662994385, "learning_rate": 1.4286575235006742e-07, "loss": 0.1349, "step": 78090 }, { "epoch": 1.5313725490196077, "grad_norm": 3.3070437908172607, "learning_rate": 1.4275231848687808e-07, "loss": 0.1512, "step": 78100 }, { "epoch": 1.5315686274509805, "grad_norm": 4.400396823883057, "learning_rate": 1.4263892217469736e-07, "loss": 0.1224, "step": 78110 }, { "epoch": 1.531764705882353, "grad_norm": 2.7671141624450684, "learning_rate": 1.4252556342544437e-07, "loss": 0.1276, "step": 78120 }, { "epoch": 1.5319607843137255, "grad_norm": 4.041668891906738, "learning_rate": 1.4241224225103483e-07, "loss": 0.1546, "step": 78130 }, { "epoch": 1.532156862745098, "grad_norm": 1.2447408437728882, "learning_rate": 1.4229895866337998e-07, "loss": 0.1366, "step": 78140 }, { "epoch": 1.5323529411764705, "grad_norm": 4.1678385734558105, "learning_rate": 1.421857126743871e-07, "loss": 0.141, "step": 78150 }, { "epoch": 1.5325490196078433, "grad_norm": 4.059925079345703, "learning_rate": 1.4207250429596007e-07, "loss": 0.1563, "step": 78160 }, { "epoch": 1.5327450980392157, "grad_norm": 1.6643964052200317, "learning_rate": 1.4195933353999822e-07, "loss": 0.1281, "step": 78170 }, { "epoch": 1.5329411764705883, "grad_norm": 2.9132871627807617, "learning_rate": 1.4184620041839706e-07, "loss": 0.1397, "step": 78180 }, { "epoch": 1.5331372549019608, "grad_norm": 3.56577730178833, "learning_rate": 1.4173310494304852e-07, "loss": 0.1183, "step": 78190 }, { "epoch": 1.5333333333333332, "grad_norm": 4.040253162384033, "learning_rate": 1.416200471258401e-07, "loss": 0.1217, "step": 78200 }, { "epoch": 1.533529411764706, "grad_norm": 3.4774258136749268, "learning_rate": 1.4150702697865542e-07, "loss": 0.125, "step": 78210 }, { "epoch": 1.5337254901960784, "grad_norm": 3.3224191665649414, "learning_rate": 1.413940445133745e-07, "loss": 0.1638, "step": 78220 }, { "epoch": 1.533921568627451, "grad_norm": 4.992483139038086, "learning_rate": 1.4128109974187295e-07, "loss": 0.1442, "step": 78230 }, { "epoch": 1.5341176470588236, "grad_norm": 1.512109637260437, "learning_rate": 1.4116819267602265e-07, "loss": 0.1183, "step": 78240 }, { "epoch": 1.534313725490196, "grad_norm": 3.8109841346740723, "learning_rate": 1.4105532332769176e-07, "loss": 0.1475, "step": 78250 }, { "epoch": 1.5345098039215688, "grad_norm": 2.3369507789611816, "learning_rate": 1.4094249170874394e-07, "loss": 0.1085, "step": 78260 }, { "epoch": 1.5347058823529411, "grad_norm": 2.228425979614258, "learning_rate": 1.408296978310392e-07, "loss": 0.1505, "step": 78270 }, { "epoch": 1.5349019607843137, "grad_norm": 2.4495134353637695, "learning_rate": 1.4071694170643334e-07, "loss": 0.1192, "step": 78280 }, { "epoch": 1.5350980392156863, "grad_norm": 1.990591287612915, "learning_rate": 1.4060422334677875e-07, "loss": 0.1054, "step": 78290 }, { "epoch": 1.5352941176470587, "grad_norm": 1.6252150535583496, "learning_rate": 1.4049154276392323e-07, "loss": 0.177, "step": 78300 }, { "epoch": 1.5354901960784315, "grad_norm": 2.345982551574707, "learning_rate": 1.4037889996971076e-07, "loss": 0.1473, "step": 78310 }, { "epoch": 1.5356862745098039, "grad_norm": 4.906802177429199, "learning_rate": 1.4026629497598175e-07, "loss": 0.1204, "step": 78320 }, { "epoch": 1.5358823529411765, "grad_norm": 3.678150177001953, "learning_rate": 1.401537277945722e-07, "loss": 0.1765, "step": 78330 }, { "epoch": 1.536078431372549, "grad_norm": 2.519106149673462, "learning_rate": 1.400411984373141e-07, "loss": 0.1393, "step": 78340 }, { "epoch": 1.5362745098039214, "grad_norm": 2.833735704421997, "learning_rate": 1.3992870691603593e-07, "loss": 0.1283, "step": 78350 }, { "epoch": 1.5364705882352943, "grad_norm": 2.400789976119995, "learning_rate": 1.3981625324256168e-07, "loss": 0.113, "step": 78360 }, { "epoch": 1.5366666666666666, "grad_norm": 2.510859727859497, "learning_rate": 1.3970383742871155e-07, "loss": 0.135, "step": 78370 }, { "epoch": 1.5368627450980392, "grad_norm": 3.2595934867858887, "learning_rate": 1.3959145948630202e-07, "loss": 0.1103, "step": 78380 }, { "epoch": 1.5370588235294118, "grad_norm": 3.0567080974578857, "learning_rate": 1.3947911942714525e-07, "loss": 0.1435, "step": 78390 }, { "epoch": 1.5372549019607842, "grad_norm": 5.225748538970947, "learning_rate": 1.3936681726304934e-07, "loss": 0.1465, "step": 78400 }, { "epoch": 1.537450980392157, "grad_norm": 4.691697120666504, "learning_rate": 1.3925455300581867e-07, "loss": 0.1633, "step": 78410 }, { "epoch": 1.5376470588235294, "grad_norm": 3.9652175903320312, "learning_rate": 1.3914232666725384e-07, "loss": 0.1179, "step": 78420 }, { "epoch": 1.537843137254902, "grad_norm": 2.0052177906036377, "learning_rate": 1.39030138259151e-07, "loss": 0.1222, "step": 78430 }, { "epoch": 1.5380392156862746, "grad_norm": 2.0145044326782227, "learning_rate": 1.3891798779330227e-07, "loss": 0.1335, "step": 78440 }, { "epoch": 1.538235294117647, "grad_norm": 2.3279340267181396, "learning_rate": 1.388058752814964e-07, "loss": 0.1197, "step": 78450 }, { "epoch": 1.5384313725490197, "grad_norm": 2.049452304840088, "learning_rate": 1.386938007355175e-07, "loss": 0.1645, "step": 78460 }, { "epoch": 1.538627450980392, "grad_norm": 4.130347728729248, "learning_rate": 1.3858176416714595e-07, "loss": 0.1557, "step": 78470 }, { "epoch": 1.5388235294117647, "grad_norm": 2.551325559616089, "learning_rate": 1.3846976558815825e-07, "loss": 0.1366, "step": 78480 }, { "epoch": 1.5390196078431373, "grad_norm": 1.4414169788360596, "learning_rate": 1.383578050103268e-07, "loss": 0.1574, "step": 78490 }, { "epoch": 1.5392156862745097, "grad_norm": 2.1234657764434814, "learning_rate": 1.382458824454198e-07, "loss": 0.1259, "step": 78500 }, { "epoch": 1.5394117647058825, "grad_norm": 2.3759706020355225, "learning_rate": 1.3813399790520185e-07, "loss": 0.1271, "step": 78510 }, { "epoch": 1.5396078431372549, "grad_norm": 2.1882593631744385, "learning_rate": 1.3802215140143335e-07, "loss": 0.1177, "step": 78520 }, { "epoch": 1.5398039215686274, "grad_norm": 3.150327444076538, "learning_rate": 1.3791034294587051e-07, "loss": 0.1182, "step": 78530 }, { "epoch": 1.54, "grad_norm": 2.403712511062622, "learning_rate": 1.37798572550266e-07, "loss": 0.1315, "step": 78540 }, { "epoch": 1.5401960784313724, "grad_norm": 3.8526337146759033, "learning_rate": 1.376868402263681e-07, "loss": 0.1486, "step": 78550 }, { "epoch": 1.5403921568627452, "grad_norm": 1.5373719930648804, "learning_rate": 1.3757514598592108e-07, "loss": 0.1332, "step": 78560 }, { "epoch": 1.5405882352941176, "grad_norm": 2.0322704315185547, "learning_rate": 1.3746348984066563e-07, "loss": 0.1081, "step": 78570 }, { "epoch": 1.5407843137254902, "grad_norm": 2.7997632026672363, "learning_rate": 1.37351871802338e-07, "loss": 0.1391, "step": 78580 }, { "epoch": 1.5409803921568628, "grad_norm": 5.837028980255127, "learning_rate": 1.3724029188267022e-07, "loss": 0.1387, "step": 78590 }, { "epoch": 1.5411764705882351, "grad_norm": 4.123704433441162, "learning_rate": 1.3712875009339142e-07, "loss": 0.1314, "step": 78600 }, { "epoch": 1.541372549019608, "grad_norm": 1.5960155725479126, "learning_rate": 1.3701724644622554e-07, "loss": 0.1079, "step": 78610 }, { "epoch": 1.5415686274509803, "grad_norm": 2.1171979904174805, "learning_rate": 1.3690578095289303e-07, "loss": 0.1496, "step": 78620 }, { "epoch": 1.541764705882353, "grad_norm": 1.5141867399215698, "learning_rate": 1.3679435362510998e-07, "loss": 0.1327, "step": 78630 }, { "epoch": 1.5419607843137255, "grad_norm": 2.168471097946167, "learning_rate": 1.366829644745892e-07, "loss": 0.1356, "step": 78640 }, { "epoch": 1.542156862745098, "grad_norm": 2.0153584480285645, "learning_rate": 1.3657161351303882e-07, "loss": 0.1499, "step": 78650 }, { "epoch": 1.5423529411764707, "grad_norm": 2.251620054244995, "learning_rate": 1.3646030075216292e-07, "loss": 0.1362, "step": 78660 }, { "epoch": 1.542549019607843, "grad_norm": 2.686110734939575, "learning_rate": 1.3634902620366223e-07, "loss": 0.1614, "step": 78670 }, { "epoch": 1.5427450980392157, "grad_norm": 2.1897692680358887, "learning_rate": 1.3623778987923278e-07, "loss": 0.1212, "step": 78680 }, { "epoch": 1.5429411764705883, "grad_norm": 2.028498888015747, "learning_rate": 1.3612659179056672e-07, "loss": 0.1528, "step": 78690 }, { "epoch": 1.5431372549019606, "grad_norm": 1.791712999343872, "learning_rate": 1.3601543194935267e-07, "loss": 0.1334, "step": 78700 }, { "epoch": 1.5433333333333334, "grad_norm": 2.456624746322632, "learning_rate": 1.359043103672747e-07, "loss": 0.1112, "step": 78710 }, { "epoch": 1.5435294117647058, "grad_norm": 2.3755056858062744, "learning_rate": 1.357932270560127e-07, "loss": 0.1101, "step": 78720 }, { "epoch": 1.5437254901960784, "grad_norm": 3.7177090644836426, "learning_rate": 1.3568218202724334e-07, "loss": 0.1328, "step": 78730 }, { "epoch": 1.543921568627451, "grad_norm": 4.176877975463867, "learning_rate": 1.3557117529263862e-07, "loss": 0.1393, "step": 78740 }, { "epoch": 1.5441176470588234, "grad_norm": 2.1788551807403564, "learning_rate": 1.3546020686386643e-07, "loss": 0.1342, "step": 78750 }, { "epoch": 1.5443137254901962, "grad_norm": 3.6962368488311768, "learning_rate": 1.353492767525911e-07, "loss": 0.1689, "step": 78760 }, { "epoch": 1.5445098039215686, "grad_norm": 2.4940271377563477, "learning_rate": 1.3523838497047286e-07, "loss": 0.1213, "step": 78770 }, { "epoch": 1.5447058823529412, "grad_norm": 2.1557302474975586, "learning_rate": 1.3512753152916756e-07, "loss": 0.1639, "step": 78780 }, { "epoch": 1.5449019607843137, "grad_norm": 2.7918899059295654, "learning_rate": 1.3501671644032714e-07, "loss": 0.1508, "step": 78790 }, { "epoch": 1.5450980392156861, "grad_norm": 4.212100505828857, "learning_rate": 1.3490593971559982e-07, "loss": 0.1377, "step": 78800 }, { "epoch": 1.545294117647059, "grad_norm": 1.7669055461883545, "learning_rate": 1.3479520136662953e-07, "loss": 0.1236, "step": 78810 }, { "epoch": 1.5454901960784313, "grad_norm": 2.296621084213257, "learning_rate": 1.3468450140505587e-07, "loss": 0.1261, "step": 78820 }, { "epoch": 1.545686274509804, "grad_norm": 2.7142324447631836, "learning_rate": 1.3457383984251524e-07, "loss": 0.1199, "step": 78830 }, { "epoch": 1.5458823529411765, "grad_norm": 4.890938758850098, "learning_rate": 1.344632166906392e-07, "loss": 0.1645, "step": 78840 }, { "epoch": 1.5460784313725489, "grad_norm": 1.9409188032150269, "learning_rate": 1.3435263196105546e-07, "loss": 0.1573, "step": 78850 }, { "epoch": 1.5462745098039217, "grad_norm": 1.9002586603164673, "learning_rate": 1.342420856653881e-07, "loss": 0.1268, "step": 78860 }, { "epoch": 1.546470588235294, "grad_norm": 2.689175605773926, "learning_rate": 1.3413157781525664e-07, "loss": 0.145, "step": 78870 }, { "epoch": 1.5466666666666666, "grad_norm": 4.460127353668213, "learning_rate": 1.3402110842227676e-07, "loss": 0.16, "step": 78880 }, { "epoch": 1.5468627450980392, "grad_norm": 2.46279239654541, "learning_rate": 1.3391067749806028e-07, "loss": 0.1566, "step": 78890 }, { "epoch": 1.5470588235294118, "grad_norm": 2.2187182903289795, "learning_rate": 1.3380028505421482e-07, "loss": 0.1456, "step": 78900 }, { "epoch": 1.5472549019607844, "grad_norm": 1.189678430557251, "learning_rate": 1.3368993110234367e-07, "loss": 0.1522, "step": 78910 }, { "epoch": 1.5474509803921568, "grad_norm": 2.7932074069976807, "learning_rate": 1.335796156540467e-07, "loss": 0.1392, "step": 78920 }, { "epoch": 1.5476470588235294, "grad_norm": 2.515692710876465, "learning_rate": 1.334693387209192e-07, "loss": 0.1547, "step": 78930 }, { "epoch": 1.547843137254902, "grad_norm": 2.9800148010253906, "learning_rate": 1.3335910031455232e-07, "loss": 0.1218, "step": 78940 }, { "epoch": 1.5480392156862746, "grad_norm": 3.0855984687805176, "learning_rate": 1.3324890044653402e-07, "loss": 0.1799, "step": 78950 }, { "epoch": 1.5482352941176472, "grad_norm": 1.8410835266113281, "learning_rate": 1.3313873912844738e-07, "loss": 0.1526, "step": 78960 }, { "epoch": 1.5484313725490195, "grad_norm": 2.12221097946167, "learning_rate": 1.3302861637187163e-07, "loss": 0.1258, "step": 78970 }, { "epoch": 1.5486274509803921, "grad_norm": 1.6668808460235596, "learning_rate": 1.3291853218838185e-07, "loss": 0.1059, "step": 78980 }, { "epoch": 1.5488235294117647, "grad_norm": 1.3991525173187256, "learning_rate": 1.3280848658954946e-07, "loss": 0.1478, "step": 78990 }, { "epoch": 1.5490196078431373, "grad_norm": 1.6969367265701294, "learning_rate": 1.3269847958694148e-07, "loss": 0.131, "step": 79000 }, { "epoch": 1.54921568627451, "grad_norm": 3.8317551612854004, "learning_rate": 1.3258851119212083e-07, "loss": 0.1389, "step": 79010 }, { "epoch": 1.5494117647058823, "grad_norm": 2.387209177017212, "learning_rate": 1.3247858141664674e-07, "loss": 0.1393, "step": 79020 }, { "epoch": 1.5496078431372549, "grad_norm": 3.6420037746429443, "learning_rate": 1.3236869027207402e-07, "loss": 0.1608, "step": 79030 }, { "epoch": 1.5498039215686275, "grad_norm": 1.4944807291030884, "learning_rate": 1.3225883776995338e-07, "loss": 0.1379, "step": 79040 }, { "epoch": 1.55, "grad_norm": 3.5336921215057373, "learning_rate": 1.3214902392183202e-07, "loss": 0.1223, "step": 79050 }, { "epoch": 1.5501960784313726, "grad_norm": 3.156830310821533, "learning_rate": 1.3203924873925243e-07, "loss": 0.1486, "step": 79060 }, { "epoch": 1.550392156862745, "grad_norm": 2.3339614868164062, "learning_rate": 1.3192951223375321e-07, "loss": 0.1203, "step": 79070 }, { "epoch": 1.5505882352941176, "grad_norm": 1.944219946861267, "learning_rate": 1.3181981441686924e-07, "loss": 0.1654, "step": 79080 }, { "epoch": 1.5507843137254902, "grad_norm": 2.6498336791992188, "learning_rate": 1.3171015530013092e-07, "loss": 0.1106, "step": 79090 }, { "epoch": 1.5509803921568628, "grad_norm": 4.5890092849731445, "learning_rate": 1.3160053489506463e-07, "loss": 0.1509, "step": 79100 }, { "epoch": 1.5511764705882354, "grad_norm": 3.4022507667541504, "learning_rate": 1.314909532131931e-07, "loss": 0.1481, "step": 79110 }, { "epoch": 1.5513725490196077, "grad_norm": 2.572608232498169, "learning_rate": 1.3138141026603426e-07, "loss": 0.1528, "step": 79120 }, { "epoch": 1.5515686274509803, "grad_norm": 3.0532705783843994, "learning_rate": 1.3127190606510286e-07, "loss": 0.1645, "step": 79130 }, { "epoch": 1.551764705882353, "grad_norm": 2.536008834838867, "learning_rate": 1.3116244062190866e-07, "loss": 0.1486, "step": 79140 }, { "epoch": 1.5519607843137255, "grad_norm": 2.570452928543091, "learning_rate": 1.310530139479582e-07, "loss": 0.141, "step": 79150 }, { "epoch": 1.5521568627450981, "grad_norm": 2.4503886699676514, "learning_rate": 1.309436260547533e-07, "loss": 0.1468, "step": 79160 }, { "epoch": 1.5523529411764705, "grad_norm": 4.0282883644104, "learning_rate": 1.3083427695379172e-07, "loss": 0.1336, "step": 79170 }, { "epoch": 1.552549019607843, "grad_norm": 1.6623274087905884, "learning_rate": 1.3072496665656786e-07, "loss": 0.1156, "step": 79180 }, { "epoch": 1.5527450980392157, "grad_norm": 2.229245901107788, "learning_rate": 1.3061569517457123e-07, "loss": 0.1256, "step": 79190 }, { "epoch": 1.5529411764705883, "grad_norm": 3.3004298210144043, "learning_rate": 1.305064625192875e-07, "loss": 0.1459, "step": 79200 }, { "epoch": 1.5531372549019609, "grad_norm": 3.920109272003174, "learning_rate": 1.303972687021986e-07, "loss": 0.1633, "step": 79210 }, { "epoch": 1.5533333333333332, "grad_norm": 2.414316415786743, "learning_rate": 1.3028811373478194e-07, "loss": 0.136, "step": 79220 }, { "epoch": 1.553529411764706, "grad_norm": 3.1110758781433105, "learning_rate": 1.3017899762851092e-07, "loss": 0.1524, "step": 79230 }, { "epoch": 1.5537254901960784, "grad_norm": 3.62896728515625, "learning_rate": 1.3006992039485516e-07, "loss": 0.1164, "step": 79240 }, { "epoch": 1.553921568627451, "grad_norm": 3.1761016845703125, "learning_rate": 1.2996088204527995e-07, "loss": 0.1292, "step": 79250 }, { "epoch": 1.5541176470588236, "grad_norm": 1.5606075525283813, "learning_rate": 1.2985188259124624e-07, "loss": 0.1396, "step": 79260 }, { "epoch": 1.554313725490196, "grad_norm": 2.3264238834381104, "learning_rate": 1.297429220442116e-07, "loss": 0.1521, "step": 79270 }, { "epoch": 1.5545098039215688, "grad_norm": 4.080442428588867, "learning_rate": 1.2963400041562888e-07, "loss": 0.1218, "step": 79280 }, { "epoch": 1.5547058823529412, "grad_norm": 5.7939229011535645, "learning_rate": 1.2952511771694692e-07, "loss": 0.1621, "step": 79290 }, { "epoch": 1.5549019607843138, "grad_norm": 2.1579391956329346, "learning_rate": 1.2941627395961068e-07, "loss": 0.1234, "step": 79300 }, { "epoch": 1.5550980392156863, "grad_norm": 2.1267690658569336, "learning_rate": 1.293074691550612e-07, "loss": 0.1232, "step": 79310 }, { "epoch": 1.5552941176470587, "grad_norm": 1.5500560998916626, "learning_rate": 1.2919870331473492e-07, "loss": 0.1713, "step": 79320 }, { "epoch": 1.5554901960784315, "grad_norm": 3.964118003845215, "learning_rate": 1.2908997645006436e-07, "loss": 0.1493, "step": 79330 }, { "epoch": 1.555686274509804, "grad_norm": 1.6260799169540405, "learning_rate": 1.2898128857247825e-07, "loss": 0.1232, "step": 79340 }, { "epoch": 1.5558823529411765, "grad_norm": 2.60280179977417, "learning_rate": 1.2887263969340084e-07, "loss": 0.1091, "step": 79350 }, { "epoch": 1.556078431372549, "grad_norm": 1.2277146577835083, "learning_rate": 1.287640298242524e-07, "loss": 0.1407, "step": 79360 }, { "epoch": 1.5562745098039215, "grad_norm": 7.901689529418945, "learning_rate": 1.2865545897644926e-07, "loss": 0.1165, "step": 79370 }, { "epoch": 1.5564705882352943, "grad_norm": 2.98565673828125, "learning_rate": 1.2854692716140342e-07, "loss": 0.1456, "step": 79380 }, { "epoch": 1.5566666666666666, "grad_norm": 3.006843090057373, "learning_rate": 1.2843843439052275e-07, "loss": 0.162, "step": 79390 }, { "epoch": 1.5568627450980392, "grad_norm": 2.282667636871338, "learning_rate": 1.2832998067521146e-07, "loss": 0.1264, "step": 79400 }, { "epoch": 1.5570588235294118, "grad_norm": 2.6037771701812744, "learning_rate": 1.282215660268691e-07, "loss": 0.1468, "step": 79410 }, { "epoch": 1.5572549019607842, "grad_norm": 2.5387794971466064, "learning_rate": 1.2811319045689135e-07, "loss": 0.145, "step": 79420 }, { "epoch": 1.557450980392157, "grad_norm": 2.3209071159362793, "learning_rate": 1.2800485397666987e-07, "loss": 0.129, "step": 79430 }, { "epoch": 1.5576470588235294, "grad_norm": 11.148244857788086, "learning_rate": 1.2789655659759212e-07, "loss": 0.1474, "step": 79440 }, { "epoch": 1.557843137254902, "grad_norm": 3.0567901134490967, "learning_rate": 1.277882983310413e-07, "loss": 0.1484, "step": 79450 }, { "epoch": 1.5580392156862746, "grad_norm": 5.187744617462158, "learning_rate": 1.276800791883969e-07, "loss": 0.1458, "step": 79460 }, { "epoch": 1.558235294117647, "grad_norm": 1.5878801345825195, "learning_rate": 1.275718991810339e-07, "loss": 0.1227, "step": 79470 }, { "epoch": 1.5584313725490198, "grad_norm": 4.183772563934326, "learning_rate": 1.2746375832032314e-07, "loss": 0.1062, "step": 79480 }, { "epoch": 1.5586274509803921, "grad_norm": 2.3052384853363037, "learning_rate": 1.2735565661763174e-07, "loss": 0.1258, "step": 79490 }, { "epoch": 1.5588235294117647, "grad_norm": 2.7037241458892822, "learning_rate": 1.2724759408432262e-07, "loss": 0.1511, "step": 79500 }, { "epoch": 1.5590196078431373, "grad_norm": 3.2798545360565186, "learning_rate": 1.2713957073175424e-07, "loss": 0.1617, "step": 79510 }, { "epoch": 1.5592156862745097, "grad_norm": 2.2282402515411377, "learning_rate": 1.2703158657128106e-07, "loss": 0.1333, "step": 79520 }, { "epoch": 1.5594117647058825, "grad_norm": 2.1102044582366943, "learning_rate": 1.2692364161425374e-07, "loss": 0.1208, "step": 79530 }, { "epoch": 1.5596078431372549, "grad_norm": 2.4709880352020264, "learning_rate": 1.268157358720185e-07, "loss": 0.1652, "step": 79540 }, { "epoch": 1.5598039215686275, "grad_norm": 5.194530963897705, "learning_rate": 1.2670786935591732e-07, "loss": 0.171, "step": 79550 }, { "epoch": 1.56, "grad_norm": 5.543421268463135, "learning_rate": 1.2660004207728864e-07, "loss": 0.1403, "step": 79560 }, { "epoch": 1.5601960784313724, "grad_norm": 2.3447465896606445, "learning_rate": 1.2649225404746615e-07, "loss": 0.1423, "step": 79570 }, { "epoch": 1.5603921568627452, "grad_norm": 4.453362464904785, "learning_rate": 1.2638450527777962e-07, "loss": 0.1408, "step": 79580 }, { "epoch": 1.5605882352941176, "grad_norm": 3.236701726913452, "learning_rate": 1.2627679577955498e-07, "loss": 0.1212, "step": 79590 }, { "epoch": 1.5607843137254902, "grad_norm": 2.3059022426605225, "learning_rate": 1.2616912556411357e-07, "loss": 0.1349, "step": 79600 }, { "epoch": 1.5609803921568628, "grad_norm": 3.245685338973999, "learning_rate": 1.2606149464277278e-07, "loss": 0.1278, "step": 79610 }, { "epoch": 1.5611764705882352, "grad_norm": 3.877159595489502, "learning_rate": 1.2595390302684616e-07, "loss": 0.1369, "step": 79620 }, { "epoch": 1.561372549019608, "grad_norm": 2.275117874145508, "learning_rate": 1.2584635072764278e-07, "loss": 0.1347, "step": 79630 }, { "epoch": 1.5615686274509804, "grad_norm": 2.609736680984497, "learning_rate": 1.2573883775646742e-07, "loss": 0.1333, "step": 79640 }, { "epoch": 1.561764705882353, "grad_norm": 1.7641218900680542, "learning_rate": 1.2563136412462133e-07, "loss": 0.1463, "step": 79650 }, { "epoch": 1.5619607843137255, "grad_norm": 3.4814319610595703, "learning_rate": 1.2552392984340104e-07, "loss": 0.1107, "step": 79660 }, { "epoch": 1.562156862745098, "grad_norm": 3.694629669189453, "learning_rate": 1.2541653492409936e-07, "loss": 0.1409, "step": 79670 }, { "epoch": 1.5623529411764707, "grad_norm": 2.1275508403778076, "learning_rate": 1.2530917937800461e-07, "loss": 0.2028, "step": 79680 }, { "epoch": 1.562549019607843, "grad_norm": 3.5859994888305664, "learning_rate": 1.252018632164013e-07, "loss": 0.1416, "step": 79690 }, { "epoch": 1.5627450980392157, "grad_norm": 3.9001760482788086, "learning_rate": 1.2509458645056965e-07, "loss": 0.1387, "step": 79700 }, { "epoch": 1.5629411764705883, "grad_norm": 2.862001895904541, "learning_rate": 1.2498734909178544e-07, "loss": 0.1248, "step": 79710 }, { "epoch": 1.5631372549019606, "grad_norm": 5.533148765563965, "learning_rate": 1.24880151151321e-07, "loss": 0.184, "step": 79720 }, { "epoch": 1.5633333333333335, "grad_norm": 2.0680932998657227, "learning_rate": 1.247729926404439e-07, "loss": 0.1281, "step": 79730 }, { "epoch": 1.5635294117647058, "grad_norm": 3.316175699234009, "learning_rate": 1.246658735704177e-07, "loss": 0.1746, "step": 79740 }, { "epoch": 1.5637254901960784, "grad_norm": 2.905388832092285, "learning_rate": 1.2455879395250212e-07, "loss": 0.1335, "step": 79750 }, { "epoch": 1.563921568627451, "grad_norm": 3.435713768005371, "learning_rate": 1.2445175379795243e-07, "loss": 0.1674, "step": 79760 }, { "epoch": 1.5641176470588234, "grad_norm": 2.388033151626587, "learning_rate": 1.2434475311801966e-07, "loss": 0.1703, "step": 79770 }, { "epoch": 1.5643137254901962, "grad_norm": 5.434759616851807, "learning_rate": 1.2423779192395112e-07, "loss": 0.1249, "step": 79780 }, { "epoch": 1.5645098039215686, "grad_norm": 1.936819314956665, "learning_rate": 1.241308702269896e-07, "loss": 0.1357, "step": 79790 }, { "epoch": 1.5647058823529412, "grad_norm": 2.09151554107666, "learning_rate": 1.2402398803837372e-07, "loss": 0.1418, "step": 79800 }, { "epoch": 1.5649019607843138, "grad_norm": 2.847257614135742, "learning_rate": 1.2391714536933835e-07, "loss": 0.1509, "step": 79810 }, { "epoch": 1.5650980392156861, "grad_norm": 3.695518732070923, "learning_rate": 1.2381034223111375e-07, "loss": 0.1451, "step": 79820 }, { "epoch": 1.565294117647059, "grad_norm": 1.9851179122924805, "learning_rate": 1.2370357863492614e-07, "loss": 0.1224, "step": 79830 }, { "epoch": 1.5654901960784313, "grad_norm": 2.02195143699646, "learning_rate": 1.2359685459199775e-07, "loss": 0.1228, "step": 79840 }, { "epoch": 1.565686274509804, "grad_norm": 1.6522630453109741, "learning_rate": 1.2349017011354674e-07, "loss": 0.119, "step": 79850 }, { "epoch": 1.5658823529411765, "grad_norm": 4.16756534576416, "learning_rate": 1.233835252107867e-07, "loss": 0.1268, "step": 79860 }, { "epoch": 1.5660784313725489, "grad_norm": 2.57039213180542, "learning_rate": 1.2327691989492723e-07, "loss": 0.1959, "step": 79870 }, { "epoch": 1.5662745098039217, "grad_norm": 4.682147979736328, "learning_rate": 1.2317035417717404e-07, "loss": 0.1188, "step": 79880 }, { "epoch": 1.566470588235294, "grad_norm": 3.110137462615967, "learning_rate": 1.230638280687284e-07, "loss": 0.1253, "step": 79890 }, { "epoch": 1.5666666666666667, "grad_norm": 2.8864798545837402, "learning_rate": 1.2295734158078725e-07, "loss": 0.1271, "step": 79900 }, { "epoch": 1.5668627450980392, "grad_norm": 5.248527526855469, "learning_rate": 1.2285089472454392e-07, "loss": 0.1248, "step": 79910 }, { "epoch": 1.5670588235294116, "grad_norm": 1.9495189189910889, "learning_rate": 1.227444875111871e-07, "loss": 0.1586, "step": 79920 }, { "epoch": 1.5672549019607844, "grad_norm": 4.111243724822998, "learning_rate": 1.2263811995190137e-07, "loss": 0.1281, "step": 79930 }, { "epoch": 1.5674509803921568, "grad_norm": 3.7891762256622314, "learning_rate": 1.2253179205786745e-07, "loss": 0.155, "step": 79940 }, { "epoch": 1.5676470588235294, "grad_norm": 4.340931415557861, "learning_rate": 1.224255038402615e-07, "loss": 0.1409, "step": 79950 }, { "epoch": 1.567843137254902, "grad_norm": 3.2212395668029785, "learning_rate": 1.2231925531025572e-07, "loss": 0.1518, "step": 79960 }, { "epoch": 1.5680392156862744, "grad_norm": 3.640178918838501, "learning_rate": 1.222130464790182e-07, "loss": 0.1595, "step": 79970 }, { "epoch": 1.5682352941176472, "grad_norm": 1.3655366897583008, "learning_rate": 1.2210687735771268e-07, "loss": 0.1071, "step": 79980 }, { "epoch": 1.5684313725490195, "grad_norm": 1.8032116889953613, "learning_rate": 1.2200074795749866e-07, "loss": 0.1229, "step": 79990 }, { "epoch": 1.5686274509803921, "grad_norm": 1.6142385005950928, "learning_rate": 1.2189465828953193e-07, "loss": 0.1472, "step": 80000 }, { "epoch": 1.5688235294117647, "grad_norm": 1.384979486465454, "learning_rate": 1.2178860836496356e-07, "loss": 0.1527, "step": 80010 }, { "epoch": 1.569019607843137, "grad_norm": 3.3440163135528564, "learning_rate": 1.2168259819494065e-07, "loss": 0.1512, "step": 80020 }, { "epoch": 1.56921568627451, "grad_norm": 3.7889633178710938, "learning_rate": 1.2157662779060618e-07, "loss": 0.1589, "step": 80030 }, { "epoch": 1.5694117647058823, "grad_norm": 2.5802533626556396, "learning_rate": 1.214706971630991e-07, "loss": 0.1359, "step": 80040 }, { "epoch": 1.5696078431372549, "grad_norm": 1.9671980142593384, "learning_rate": 1.2136480632355384e-07, "loss": 0.1601, "step": 80050 }, { "epoch": 1.5698039215686275, "grad_norm": 3.487351655960083, "learning_rate": 1.212589552831006e-07, "loss": 0.1377, "step": 80060 }, { "epoch": 1.5699999999999998, "grad_norm": 2.5576610565185547, "learning_rate": 1.211531440528659e-07, "loss": 0.1357, "step": 80070 }, { "epoch": 1.5701960784313727, "grad_norm": 5.95650053024292, "learning_rate": 1.2104737264397169e-07, "loss": 0.1562, "step": 80080 }, { "epoch": 1.570392156862745, "grad_norm": 1.6161049604415894, "learning_rate": 1.2094164106753557e-07, "loss": 0.1558, "step": 80090 }, { "epoch": 1.5705882352941176, "grad_norm": 2.7989611625671387, "learning_rate": 1.2083594933467157e-07, "loss": 0.1295, "step": 80100 }, { "epoch": 1.5707843137254902, "grad_norm": 4.531503200531006, "learning_rate": 1.2073029745648894e-07, "loss": 0.1301, "step": 80110 }, { "epoch": 1.5709803921568626, "grad_norm": 2.125936508178711, "learning_rate": 1.2062468544409282e-07, "loss": 0.1163, "step": 80120 }, { "epoch": 1.5711764705882354, "grad_norm": 3.219247341156006, "learning_rate": 1.2051911330858467e-07, "loss": 0.1506, "step": 80130 }, { "epoch": 1.5713725490196078, "grad_norm": 2.443721055984497, "learning_rate": 1.2041358106106115e-07, "loss": 0.127, "step": 80140 }, { "epoch": 1.5715686274509804, "grad_norm": 1.0962753295898438, "learning_rate": 1.2030808871261488e-07, "loss": 0.1534, "step": 80150 }, { "epoch": 1.571764705882353, "grad_norm": 0.8608381748199463, "learning_rate": 1.2020263627433458e-07, "loss": 0.1329, "step": 80160 }, { "epoch": 1.5719607843137255, "grad_norm": 5.437244415283203, "learning_rate": 1.2009722375730453e-07, "loss": 0.1121, "step": 80170 }, { "epoch": 1.5721568627450981, "grad_norm": 1.657752513885498, "learning_rate": 1.1999185117260458e-07, "loss": 0.132, "step": 80180 }, { "epoch": 1.5723529411764705, "grad_norm": 2.3997089862823486, "learning_rate": 1.1988651853131088e-07, "loss": 0.1277, "step": 80190 }, { "epoch": 1.572549019607843, "grad_norm": 1.63324773311615, "learning_rate": 1.197812258444953e-07, "loss": 0.1382, "step": 80200 }, { "epoch": 1.5727450980392157, "grad_norm": 3.2089412212371826, "learning_rate": 1.1967597312322514e-07, "loss": 0.1631, "step": 80210 }, { "epoch": 1.5729411764705883, "grad_norm": 3.0540771484375, "learning_rate": 1.1957076037856368e-07, "loss": 0.1291, "step": 80220 }, { "epoch": 1.5731372549019609, "grad_norm": 5.337319850921631, "learning_rate": 1.1946558762157022e-07, "loss": 0.1448, "step": 80230 }, { "epoch": 1.5733333333333333, "grad_norm": 2.8898732662200928, "learning_rate": 1.193604548632996e-07, "loss": 0.1089, "step": 80240 }, { "epoch": 1.5735294117647058, "grad_norm": 3.0900299549102783, "learning_rate": 1.1925536211480235e-07, "loss": 0.1182, "step": 80250 }, { "epoch": 1.5737254901960784, "grad_norm": 3.84377121925354, "learning_rate": 1.1915030938712533e-07, "loss": 0.1157, "step": 80260 }, { "epoch": 1.573921568627451, "grad_norm": 4.277869701385498, "learning_rate": 1.1904529669131063e-07, "loss": 0.1314, "step": 80270 }, { "epoch": 1.5741176470588236, "grad_norm": 2.132689952850342, "learning_rate": 1.1894032403839627e-07, "loss": 0.1388, "step": 80280 }, { "epoch": 1.574313725490196, "grad_norm": 3.1703219413757324, "learning_rate": 1.1883539143941635e-07, "loss": 0.1397, "step": 80290 }, { "epoch": 1.5745098039215686, "grad_norm": 1.7917170524597168, "learning_rate": 1.1873049890540038e-07, "loss": 0.1221, "step": 80300 }, { "epoch": 1.5747058823529412, "grad_norm": 3.0993432998657227, "learning_rate": 1.1862564644737372e-07, "loss": 0.1322, "step": 80310 }, { "epoch": 1.5749019607843138, "grad_norm": 3.879319906234741, "learning_rate": 1.1852083407635793e-07, "loss": 0.1531, "step": 80320 }, { "epoch": 1.5750980392156864, "grad_norm": 2.5782785415649414, "learning_rate": 1.1841606180336988e-07, "loss": 0.141, "step": 80330 }, { "epoch": 1.5752941176470587, "grad_norm": 4.242121696472168, "learning_rate": 1.183113296394222e-07, "loss": 0.1596, "step": 80340 }, { "epoch": 1.5754901960784313, "grad_norm": 3.4411263465881348, "learning_rate": 1.1820663759552385e-07, "loss": 0.125, "step": 80350 }, { "epoch": 1.575686274509804, "grad_norm": 2.1732664108276367, "learning_rate": 1.1810198568267903e-07, "loss": 0.1603, "step": 80360 }, { "epoch": 1.5758823529411765, "grad_norm": 2.029867649078369, "learning_rate": 1.179973739118878e-07, "loss": 0.1371, "step": 80370 }, { "epoch": 1.576078431372549, "grad_norm": 2.6079518795013428, "learning_rate": 1.1789280229414623e-07, "loss": 0.1517, "step": 80380 }, { "epoch": 1.5762745098039215, "grad_norm": 2.5077552795410156, "learning_rate": 1.1778827084044618e-07, "loss": 0.1332, "step": 80390 }, { "epoch": 1.576470588235294, "grad_norm": 2.415924072265625, "learning_rate": 1.1768377956177506e-07, "loss": 0.132, "step": 80400 }, { "epoch": 1.5766666666666667, "grad_norm": 5.852514266967773, "learning_rate": 1.17579328469116e-07, "loss": 0.1656, "step": 80410 }, { "epoch": 1.5768627450980393, "grad_norm": 4.2276482582092285, "learning_rate": 1.174749175734483e-07, "loss": 0.1634, "step": 80420 }, { "epoch": 1.5770588235294118, "grad_norm": 3.063913345336914, "learning_rate": 1.1737054688574666e-07, "loss": 0.1369, "step": 80430 }, { "epoch": 1.5772549019607842, "grad_norm": 3.8468923568725586, "learning_rate": 1.1726621641698159e-07, "loss": 0.1487, "step": 80440 }, { "epoch": 1.5774509803921568, "grad_norm": 1.5149158239364624, "learning_rate": 1.1716192617811976e-07, "loss": 0.1062, "step": 80450 }, { "epoch": 1.5776470588235294, "grad_norm": 2.39127254486084, "learning_rate": 1.1705767618012314e-07, "loss": 0.1259, "step": 80460 }, { "epoch": 1.577843137254902, "grad_norm": 2.941812515258789, "learning_rate": 1.1695346643394948e-07, "loss": 0.1688, "step": 80470 }, { "epoch": 1.5780392156862746, "grad_norm": 1.6723427772521973, "learning_rate": 1.1684929695055285e-07, "loss": 0.0944, "step": 80480 }, { "epoch": 1.578235294117647, "grad_norm": 1.3049827814102173, "learning_rate": 1.167451677408825e-07, "loss": 0.1049, "step": 80490 }, { "epoch": 1.5784313725490198, "grad_norm": 3.786797523498535, "learning_rate": 1.1664107881588353e-07, "loss": 0.1281, "step": 80500 }, { "epoch": 1.5786274509803921, "grad_norm": 4.045886039733887, "learning_rate": 1.1653703018649725e-07, "loss": 0.1081, "step": 80510 }, { "epoch": 1.5788235294117647, "grad_norm": 2.76483416557312, "learning_rate": 1.1643302186366022e-07, "loss": 0.1376, "step": 80520 }, { "epoch": 1.5790196078431373, "grad_norm": 1.9081753492355347, "learning_rate": 1.1632905385830483e-07, "loss": 0.1399, "step": 80530 }, { "epoch": 1.5792156862745097, "grad_norm": 5.4599409103393555, "learning_rate": 1.1622512618135966e-07, "loss": 0.1673, "step": 80540 }, { "epoch": 1.5794117647058825, "grad_norm": 3.5933070182800293, "learning_rate": 1.1612123884374847e-07, "loss": 0.1362, "step": 80550 }, { "epoch": 1.5796078431372549, "grad_norm": 1.8453232049942017, "learning_rate": 1.1601739185639137e-07, "loss": 0.1584, "step": 80560 }, { "epoch": 1.5798039215686275, "grad_norm": 3.014385461807251, "learning_rate": 1.1591358523020361e-07, "loss": 0.1553, "step": 80570 }, { "epoch": 1.58, "grad_norm": 1.5282284021377563, "learning_rate": 1.1580981897609676e-07, "loss": 0.121, "step": 80580 }, { "epoch": 1.5801960784313724, "grad_norm": 4.908100605010986, "learning_rate": 1.1570609310497786e-07, "loss": 0.1525, "step": 80590 }, { "epoch": 1.5803921568627453, "grad_norm": 4.922686576843262, "learning_rate": 1.1560240762774947e-07, "loss": 0.1335, "step": 80600 }, { "epoch": 1.5805882352941176, "grad_norm": 3.814509868621826, "learning_rate": 1.1549876255531054e-07, "loss": 0.1165, "step": 80610 }, { "epoch": 1.5807843137254902, "grad_norm": 3.274844169616699, "learning_rate": 1.1539515789855526e-07, "loss": 0.1609, "step": 80620 }, { "epoch": 1.5809803921568628, "grad_norm": 3.064671516418457, "learning_rate": 1.1529159366837355e-07, "loss": 0.1435, "step": 80630 }, { "epoch": 1.5811764705882352, "grad_norm": 2.756584882736206, "learning_rate": 1.1518806987565149e-07, "loss": 0.128, "step": 80640 }, { "epoch": 1.581372549019608, "grad_norm": 3.214625835418701, "learning_rate": 1.1508458653127063e-07, "loss": 0.1353, "step": 80650 }, { "epoch": 1.5815686274509804, "grad_norm": 2.3172430992126465, "learning_rate": 1.1498114364610811e-07, "loss": 0.1537, "step": 80660 }, { "epoch": 1.581764705882353, "grad_norm": 2.536471366882324, "learning_rate": 1.1487774123103733e-07, "loss": 0.1197, "step": 80670 }, { "epoch": 1.5819607843137256, "grad_norm": 3.693798542022705, "learning_rate": 1.1477437929692695e-07, "loss": 0.1289, "step": 80680 }, { "epoch": 1.582156862745098, "grad_norm": 3.2342052459716797, "learning_rate": 1.1467105785464142e-07, "loss": 0.142, "step": 80690 }, { "epoch": 1.5823529411764707, "grad_norm": 2.735646963119507, "learning_rate": 1.1456777691504139e-07, "loss": 0.1311, "step": 80700 }, { "epoch": 1.582549019607843, "grad_norm": 3.527329444885254, "learning_rate": 1.1446453648898263e-07, "loss": 0.1528, "step": 80710 }, { "epoch": 1.5827450980392157, "grad_norm": 1.9826452732086182, "learning_rate": 1.14361336587317e-07, "loss": 0.1375, "step": 80720 }, { "epoch": 1.5829411764705883, "grad_norm": 2.8013219833374023, "learning_rate": 1.1425817722089204e-07, "loss": 0.1115, "step": 80730 }, { "epoch": 1.5831372549019607, "grad_norm": 1.5762181282043457, "learning_rate": 1.1415505840055123e-07, "loss": 0.1416, "step": 80740 }, { "epoch": 1.5833333333333335, "grad_norm": 2.044611692428589, "learning_rate": 1.1405198013713346e-07, "loss": 0.1374, "step": 80750 }, { "epoch": 1.5835294117647059, "grad_norm": 2.7009410858154297, "learning_rate": 1.1394894244147329e-07, "loss": 0.1474, "step": 80760 }, { "epoch": 1.5837254901960784, "grad_norm": 5.333556175231934, "learning_rate": 1.1384594532440161e-07, "loss": 0.1319, "step": 80770 }, { "epoch": 1.583921568627451, "grad_norm": 2.5537147521972656, "learning_rate": 1.1374298879674443e-07, "loss": 0.1447, "step": 80780 }, { "epoch": 1.5841176470588234, "grad_norm": 2.153672456741333, "learning_rate": 1.1364007286932359e-07, "loss": 0.1309, "step": 80790 }, { "epoch": 1.5843137254901962, "grad_norm": 3.08455491065979, "learning_rate": 1.1353719755295704e-07, "loss": 0.1302, "step": 80800 }, { "epoch": 1.5845098039215686, "grad_norm": 2.047212839126587, "learning_rate": 1.1343436285845808e-07, "loss": 0.1179, "step": 80810 }, { "epoch": 1.5847058823529412, "grad_norm": 5.117534160614014, "learning_rate": 1.1333156879663575e-07, "loss": 0.1372, "step": 80820 }, { "epoch": 1.5849019607843138, "grad_norm": 2.8169972896575928, "learning_rate": 1.132288153782952e-07, "loss": 0.1275, "step": 80830 }, { "epoch": 1.5850980392156861, "grad_norm": 2.415891170501709, "learning_rate": 1.1312610261423694e-07, "loss": 0.1342, "step": 80840 }, { "epoch": 1.585294117647059, "grad_norm": 2.4349911212921143, "learning_rate": 1.1302343051525715e-07, "loss": 0.1375, "step": 80850 }, { "epoch": 1.5854901960784313, "grad_norm": 2.0876986980438232, "learning_rate": 1.1292079909214814e-07, "loss": 0.1409, "step": 80860 }, { "epoch": 1.585686274509804, "grad_norm": 3.847200393676758, "learning_rate": 1.1281820835569761e-07, "loss": 0.1613, "step": 80870 }, { "epoch": 1.5858823529411765, "grad_norm": 1.9814465045928955, "learning_rate": 1.127156583166889e-07, "loss": 0.1242, "step": 80880 }, { "epoch": 1.586078431372549, "grad_norm": 3.382814407348633, "learning_rate": 1.1261314898590152e-07, "loss": 0.1466, "step": 80890 }, { "epoch": 1.5862745098039217, "grad_norm": 1.6999329328536987, "learning_rate": 1.1251068037411032e-07, "loss": 0.1379, "step": 80900 }, { "epoch": 1.586470588235294, "grad_norm": 3.438810110092163, "learning_rate": 1.1240825249208585e-07, "loss": 0.1512, "step": 80910 }, { "epoch": 1.5866666666666667, "grad_norm": 2.796187162399292, "learning_rate": 1.1230586535059461e-07, "loss": 0.1682, "step": 80920 }, { "epoch": 1.5868627450980393, "grad_norm": 2.250906229019165, "learning_rate": 1.1220351896039887e-07, "loss": 0.1465, "step": 80930 }, { "epoch": 1.5870588235294116, "grad_norm": 1.7748490571975708, "learning_rate": 1.1210121333225631e-07, "loss": 0.137, "step": 80940 }, { "epoch": 1.5872549019607844, "grad_norm": 3.8918957710266113, "learning_rate": 1.1199894847692037e-07, "loss": 0.1504, "step": 80950 }, { "epoch": 1.5874509803921568, "grad_norm": 8.031425476074219, "learning_rate": 1.1189672440514053e-07, "loss": 0.1224, "step": 80960 }, { "epoch": 1.5876470588235294, "grad_norm": 3.404632568359375, "learning_rate": 1.1179454112766167e-07, "loss": 0.1535, "step": 80970 }, { "epoch": 1.587843137254902, "grad_norm": 4.457058906555176, "learning_rate": 1.1169239865522434e-07, "loss": 0.1265, "step": 80980 }, { "epoch": 1.5880392156862744, "grad_norm": 2.424490213394165, "learning_rate": 1.115902969985652e-07, "loss": 0.149, "step": 80990 }, { "epoch": 1.5882352941176472, "grad_norm": 2.993483543395996, "learning_rate": 1.1148823616841624e-07, "loss": 0.1621, "step": 81000 }, { "epoch": 1.5884313725490196, "grad_norm": 3.221224784851074, "learning_rate": 1.1138621617550509e-07, "loss": 0.157, "step": 81010 }, { "epoch": 1.5886274509803922, "grad_norm": 3.6362295150756836, "learning_rate": 1.1128423703055557e-07, "loss": 0.1394, "step": 81020 }, { "epoch": 1.5888235294117647, "grad_norm": 2.0525283813476562, "learning_rate": 1.1118229874428681e-07, "loss": 0.1453, "step": 81030 }, { "epoch": 1.5890196078431371, "grad_norm": 1.567894697189331, "learning_rate": 1.1108040132741353e-07, "loss": 0.1559, "step": 81040 }, { "epoch": 1.58921568627451, "grad_norm": 3.9815287590026855, "learning_rate": 1.1097854479064672e-07, "loss": 0.1538, "step": 81050 }, { "epoch": 1.5894117647058823, "grad_norm": 1.9024536609649658, "learning_rate": 1.1087672914469254e-07, "loss": 0.1615, "step": 81060 }, { "epoch": 1.589607843137255, "grad_norm": 2.1288654804229736, "learning_rate": 1.1077495440025292e-07, "loss": 0.1399, "step": 81070 }, { "epoch": 1.5898039215686275, "grad_norm": 2.5915069580078125, "learning_rate": 1.1067322056802586e-07, "loss": 0.1336, "step": 81080 }, { "epoch": 1.5899999999999999, "grad_norm": 1.7182351350784302, "learning_rate": 1.1057152765870447e-07, "loss": 0.1423, "step": 81090 }, { "epoch": 1.5901960784313727, "grad_norm": 3.8320729732513428, "learning_rate": 1.1046987568297833e-07, "loss": 0.1517, "step": 81100 }, { "epoch": 1.590392156862745, "grad_norm": 3.0352330207824707, "learning_rate": 1.1036826465153187e-07, "loss": 0.1678, "step": 81110 }, { "epoch": 1.5905882352941176, "grad_norm": 2.222116708755493, "learning_rate": 1.1026669457504595e-07, "loss": 0.1343, "step": 81120 }, { "epoch": 1.5907843137254902, "grad_norm": 2.552889823913574, "learning_rate": 1.1016516546419663e-07, "loss": 0.1422, "step": 81130 }, { "epoch": 1.5909803921568626, "grad_norm": 3.545452117919922, "learning_rate": 1.1006367732965571e-07, "loss": 0.1795, "step": 81140 }, { "epoch": 1.5911764705882354, "grad_norm": 2.5743281841278076, "learning_rate": 1.0996223018209111e-07, "loss": 0.1463, "step": 81150 }, { "epoch": 1.5913725490196078, "grad_norm": 1.8954182863235474, "learning_rate": 1.0986082403216596e-07, "loss": 0.1132, "step": 81160 }, { "epoch": 1.5915686274509804, "grad_norm": 2.2940855026245117, "learning_rate": 1.0975945889053917e-07, "loss": 0.1756, "step": 81170 }, { "epoch": 1.591764705882353, "grad_norm": 4.008917331695557, "learning_rate": 1.0965813476786567e-07, "loss": 0.1421, "step": 81180 }, { "epoch": 1.5919607843137253, "grad_norm": 3.4327967166900635, "learning_rate": 1.0955685167479573e-07, "loss": 0.1328, "step": 81190 }, { "epoch": 1.5921568627450982, "grad_norm": 2.099069356918335, "learning_rate": 1.0945560962197526e-07, "loss": 0.1443, "step": 81200 }, { "epoch": 1.5923529411764705, "grad_norm": 2.827967882156372, "learning_rate": 1.093544086200463e-07, "loss": 0.1377, "step": 81210 }, { "epoch": 1.5925490196078431, "grad_norm": 2.039731979370117, "learning_rate": 1.0925324867964614e-07, "loss": 0.1524, "step": 81220 }, { "epoch": 1.5927450980392157, "grad_norm": 1.600409746170044, "learning_rate": 1.091521298114077e-07, "loss": 0.152, "step": 81230 }, { "epoch": 1.592941176470588, "grad_norm": 3.0369510650634766, "learning_rate": 1.0905105202596021e-07, "loss": 0.1445, "step": 81240 }, { "epoch": 1.593137254901961, "grad_norm": 2.1108500957489014, "learning_rate": 1.089500153339279e-07, "loss": 0.1789, "step": 81250 }, { "epoch": 1.5933333333333333, "grad_norm": 3.1458942890167236, "learning_rate": 1.0884901974593075e-07, "loss": 0.1371, "step": 81260 }, { "epoch": 1.5935294117647059, "grad_norm": 2.6292850971221924, "learning_rate": 1.087480652725849e-07, "loss": 0.1657, "step": 81270 }, { "epoch": 1.5937254901960785, "grad_norm": 4.762749671936035, "learning_rate": 1.0864715192450192e-07, "loss": 0.1699, "step": 81280 }, { "epoch": 1.5939215686274508, "grad_norm": 2.1262848377227783, "learning_rate": 1.0854627971228886e-07, "loss": 0.1434, "step": 81290 }, { "epoch": 1.5941176470588236, "grad_norm": 3.877840280532837, "learning_rate": 1.0844544864654853e-07, "loss": 0.1653, "step": 81300 }, { "epoch": 1.594313725490196, "grad_norm": 2.7882766723632812, "learning_rate": 1.0834465873787967e-07, "loss": 0.1637, "step": 81310 }, { "epoch": 1.5945098039215686, "grad_norm": 2.352811098098755, "learning_rate": 1.0824390999687644e-07, "loss": 0.1647, "step": 81320 }, { "epoch": 1.5947058823529412, "grad_norm": 4.774430274963379, "learning_rate": 1.081432024341285e-07, "loss": 0.1403, "step": 81330 }, { "epoch": 1.5949019607843136, "grad_norm": 2.8109939098358154, "learning_rate": 1.0804253606022184e-07, "loss": 0.174, "step": 81340 }, { "epoch": 1.5950980392156864, "grad_norm": 3.870290756225586, "learning_rate": 1.0794191088573745e-07, "loss": 0.1323, "step": 81350 }, { "epoch": 1.5952941176470588, "grad_norm": 2.6769800186157227, "learning_rate": 1.078413269212522e-07, "loss": 0.1473, "step": 81360 }, { "epoch": 1.5954901960784313, "grad_norm": 3.1130661964416504, "learning_rate": 1.0774078417733884e-07, "loss": 0.147, "step": 81370 }, { "epoch": 1.595686274509804, "grad_norm": 1.96239173412323, "learning_rate": 1.0764028266456554e-07, "loss": 0.15, "step": 81380 }, { "epoch": 1.5958823529411763, "grad_norm": 5.596366882324219, "learning_rate": 1.0753982239349608e-07, "loss": 0.1528, "step": 81390 }, { "epoch": 1.5960784313725491, "grad_norm": 3.400111436843872, "learning_rate": 1.0743940337469026e-07, "loss": 0.1286, "step": 81400 }, { "epoch": 1.5962745098039215, "grad_norm": 3.6347267627716064, "learning_rate": 1.0733902561870323e-07, "loss": 0.1452, "step": 81410 }, { "epoch": 1.596470588235294, "grad_norm": 2.140779733657837, "learning_rate": 1.0723868913608575e-07, "loss": 0.1663, "step": 81420 }, { "epoch": 1.5966666666666667, "grad_norm": 1.3942822217941284, "learning_rate": 1.0713839393738467e-07, "loss": 0.1469, "step": 81430 }, { "epoch": 1.596862745098039, "grad_norm": 3.9382147789001465, "learning_rate": 1.0703814003314204e-07, "loss": 0.198, "step": 81440 }, { "epoch": 1.5970588235294119, "grad_norm": 2.9857490062713623, "learning_rate": 1.0693792743389562e-07, "loss": 0.1523, "step": 81450 }, { "epoch": 1.5972549019607842, "grad_norm": 5.030546188354492, "learning_rate": 1.0683775615017915e-07, "loss": 0.1368, "step": 81460 }, { "epoch": 1.5974509803921568, "grad_norm": 3.1289093494415283, "learning_rate": 1.0673762619252186e-07, "loss": 0.1182, "step": 81470 }, { "epoch": 1.5976470588235294, "grad_norm": 3.86401104927063, "learning_rate": 1.0663753757144856e-07, "loss": 0.1599, "step": 81480 }, { "epoch": 1.597843137254902, "grad_norm": 3.899214267730713, "learning_rate": 1.065374902974796e-07, "loss": 0.1535, "step": 81490 }, { "epoch": 1.5980392156862746, "grad_norm": 2.6564083099365234, "learning_rate": 1.0643748438113143e-07, "loss": 0.1764, "step": 81500 }, { "epoch": 1.598235294117647, "grad_norm": 3.0297579765319824, "learning_rate": 1.0633751983291572e-07, "loss": 0.1202, "step": 81510 }, { "epoch": 1.5984313725490196, "grad_norm": 2.597321033477783, "learning_rate": 1.0623759666333976e-07, "loss": 0.1398, "step": 81520 }, { "epoch": 1.5986274509803922, "grad_norm": 5.476846218109131, "learning_rate": 1.0613771488290702e-07, "loss": 0.1481, "step": 81530 }, { "epoch": 1.5988235294117648, "grad_norm": 2.0298399925231934, "learning_rate": 1.060378745021161e-07, "loss": 0.1628, "step": 81540 }, { "epoch": 1.5990196078431373, "grad_norm": 3.4045369625091553, "learning_rate": 1.0593807553146128e-07, "loss": 0.1414, "step": 81550 }, { "epoch": 1.5992156862745097, "grad_norm": 2.1965041160583496, "learning_rate": 1.0583831798143295e-07, "loss": 0.1289, "step": 81560 }, { "epoch": 1.5994117647058823, "grad_norm": 3.2219364643096924, "learning_rate": 1.0573860186251659e-07, "loss": 0.14, "step": 81570 }, { "epoch": 1.599607843137255, "grad_norm": 1.477235198020935, "learning_rate": 1.0563892718519345e-07, "loss": 0.172, "step": 81580 }, { "epoch": 1.5998039215686275, "grad_norm": 2.8042612075805664, "learning_rate": 1.0553929395994088e-07, "loss": 0.136, "step": 81590 }, { "epoch": 1.6, "grad_norm": 1.8836666345596313, "learning_rate": 1.0543970219723136e-07, "loss": 0.1245, "step": 81600 }, { "epoch": 1.6001960784313725, "grad_norm": 4.079176425933838, "learning_rate": 1.0534015190753298e-07, "loss": 0.171, "step": 81610 }, { "epoch": 1.600392156862745, "grad_norm": 4.457946300506592, "learning_rate": 1.0524064310130981e-07, "loss": 0.1272, "step": 81620 }, { "epoch": 1.6005882352941176, "grad_norm": 6.158318996429443, "learning_rate": 1.0514117578902165e-07, "loss": 0.14, "step": 81630 }, { "epoch": 1.6007843137254902, "grad_norm": 2.6831204891204834, "learning_rate": 1.0504174998112348e-07, "loss": 0.1535, "step": 81640 }, { "epoch": 1.6009803921568628, "grad_norm": 3.5173513889312744, "learning_rate": 1.0494236568806597e-07, "loss": 0.1396, "step": 81650 }, { "epoch": 1.6011764705882352, "grad_norm": 3.4372103214263916, "learning_rate": 1.0484302292029601e-07, "loss": 0.1326, "step": 81660 }, { "epoch": 1.6013725490196078, "grad_norm": 2.432494878768921, "learning_rate": 1.0474372168825547e-07, "loss": 0.1339, "step": 81670 }, { "epoch": 1.6015686274509804, "grad_norm": 2.7150163650512695, "learning_rate": 1.0464446200238198e-07, "loss": 0.1563, "step": 81680 }, { "epoch": 1.601764705882353, "grad_norm": 1.4288533926010132, "learning_rate": 1.045452438731092e-07, "loss": 0.1357, "step": 81690 }, { "epoch": 1.6019607843137256, "grad_norm": 3.8998684883117676, "learning_rate": 1.0444606731086603e-07, "loss": 0.1309, "step": 81700 }, { "epoch": 1.602156862745098, "grad_norm": 1.9306493997573853, "learning_rate": 1.0434693232607695e-07, "loss": 0.1102, "step": 81710 }, { "epoch": 1.6023529411764705, "grad_norm": 2.1276843547821045, "learning_rate": 1.0424783892916256e-07, "loss": 0.1291, "step": 81720 }, { "epoch": 1.6025490196078431, "grad_norm": 3.611034393310547, "learning_rate": 1.0414878713053853e-07, "loss": 0.1433, "step": 81730 }, { "epoch": 1.6027450980392157, "grad_norm": 2.274879217147827, "learning_rate": 1.0404977694061629e-07, "loss": 0.1408, "step": 81740 }, { "epoch": 1.6029411764705883, "grad_norm": 3.09014630317688, "learning_rate": 1.0395080836980336e-07, "loss": 0.1405, "step": 81750 }, { "epoch": 1.6031372549019607, "grad_norm": 3.8335347175598145, "learning_rate": 1.0385188142850227e-07, "loss": 0.1285, "step": 81760 }, { "epoch": 1.6033333333333335, "grad_norm": 2.845978021621704, "learning_rate": 1.037529961271113e-07, "loss": 0.1246, "step": 81770 }, { "epoch": 1.6035294117647059, "grad_norm": 5.128767490386963, "learning_rate": 1.0365415247602483e-07, "loss": 0.1586, "step": 81780 }, { "epoch": 1.6037254901960785, "grad_norm": 2.4691925048828125, "learning_rate": 1.035553504856323e-07, "loss": 0.123, "step": 81790 }, { "epoch": 1.603921568627451, "grad_norm": 4.189206600189209, "learning_rate": 1.0345659016631886e-07, "loss": 0.1362, "step": 81800 }, { "epoch": 1.6041176470588234, "grad_norm": 2.7019121646881104, "learning_rate": 1.0335787152846559e-07, "loss": 0.1208, "step": 81810 }, { "epoch": 1.6043137254901962, "grad_norm": 4.529266357421875, "learning_rate": 1.0325919458244908e-07, "loss": 0.1522, "step": 81820 }, { "epoch": 1.6045098039215686, "grad_norm": 2.11472487449646, "learning_rate": 1.0316055933864137e-07, "loss": 0.1504, "step": 81830 }, { "epoch": 1.6047058823529412, "grad_norm": 2.0200631618499756, "learning_rate": 1.0306196580741e-07, "loss": 0.1335, "step": 81840 }, { "epoch": 1.6049019607843138, "grad_norm": 1.980359673500061, "learning_rate": 1.0296341399911868e-07, "loss": 0.1275, "step": 81850 }, { "epoch": 1.6050980392156862, "grad_norm": 2.795957326889038, "learning_rate": 1.0286490392412616e-07, "loss": 0.1549, "step": 81860 }, { "epoch": 1.605294117647059, "grad_norm": 2.8227603435516357, "learning_rate": 1.0276643559278697e-07, "loss": 0.155, "step": 81870 }, { "epoch": 1.6054901960784314, "grad_norm": 2.6824588775634766, "learning_rate": 1.0266800901545153e-07, "loss": 0.1214, "step": 81880 }, { "epoch": 1.605686274509804, "grad_norm": 3.3682122230529785, "learning_rate": 1.0256962420246557e-07, "loss": 0.1145, "step": 81890 }, { "epoch": 1.6058823529411765, "grad_norm": 3.5982086658477783, "learning_rate": 1.0247128116417037e-07, "loss": 0.1286, "step": 81900 }, { "epoch": 1.606078431372549, "grad_norm": 3.8526463508605957, "learning_rate": 1.0237297991090315e-07, "loss": 0.1457, "step": 81910 }, { "epoch": 1.6062745098039217, "grad_norm": 3.091006278991699, "learning_rate": 1.0227472045299651e-07, "loss": 0.1187, "step": 81920 }, { "epoch": 1.606470588235294, "grad_norm": 1.691012978553772, "learning_rate": 1.0217650280077855e-07, "loss": 0.1326, "step": 81930 }, { "epoch": 1.6066666666666667, "grad_norm": 4.133618354797363, "learning_rate": 1.0207832696457335e-07, "loss": 0.1268, "step": 81940 }, { "epoch": 1.6068627450980393, "grad_norm": 2.5266592502593994, "learning_rate": 1.0198019295470023e-07, "loss": 0.1287, "step": 81950 }, { "epoch": 1.6070588235294117, "grad_norm": 3.650395631790161, "learning_rate": 1.0188210078147414e-07, "loss": 0.1376, "step": 81960 }, { "epoch": 1.6072549019607845, "grad_norm": 3.529294490814209, "learning_rate": 1.0178405045520599e-07, "loss": 0.1453, "step": 81970 }, { "epoch": 1.6074509803921568, "grad_norm": 3.1288084983825684, "learning_rate": 1.0168604198620179e-07, "loss": 0.1251, "step": 81980 }, { "epoch": 1.6076470588235294, "grad_norm": 3.1376049518585205, "learning_rate": 1.0158807538476366e-07, "loss": 0.1392, "step": 81990 }, { "epoch": 1.607843137254902, "grad_norm": 3.0909523963928223, "learning_rate": 1.0149015066118883e-07, "loss": 0.1404, "step": 82000 }, { "epoch": 1.6080392156862744, "grad_norm": 3.5062360763549805, "learning_rate": 1.0139226782577054e-07, "loss": 0.1545, "step": 82010 }, { "epoch": 1.6082352941176472, "grad_norm": 2.49005126953125, "learning_rate": 1.0129442688879742e-07, "loss": 0.1341, "step": 82020 }, { "epoch": 1.6084313725490196, "grad_norm": 2.6208102703094482, "learning_rate": 1.0119662786055355e-07, "loss": 0.1562, "step": 82030 }, { "epoch": 1.6086274509803922, "grad_norm": 5.145352363586426, "learning_rate": 1.0109887075131901e-07, "loss": 0.1351, "step": 82040 }, { "epoch": 1.6088235294117648, "grad_norm": 2.8056223392486572, "learning_rate": 1.0100115557136918e-07, "loss": 0.1339, "step": 82050 }, { "epoch": 1.6090196078431371, "grad_norm": 2.66703462600708, "learning_rate": 1.009034823309749e-07, "loss": 0.1099, "step": 82060 }, { "epoch": 1.60921568627451, "grad_norm": 1.3929418325424194, "learning_rate": 1.0080585104040313e-07, "loss": 0.1368, "step": 82070 }, { "epoch": 1.6094117647058823, "grad_norm": 1.2914105653762817, "learning_rate": 1.0070826170991586e-07, "loss": 0.1327, "step": 82080 }, { "epoch": 1.609607843137255, "grad_norm": 2.285640239715576, "learning_rate": 1.006107143497708e-07, "loss": 0.1428, "step": 82090 }, { "epoch": 1.6098039215686275, "grad_norm": 2.392228126525879, "learning_rate": 1.0051320897022164e-07, "loss": 0.1367, "step": 82100 }, { "epoch": 1.6099999999999999, "grad_norm": 1.560536503791809, "learning_rate": 1.0041574558151721e-07, "loss": 0.1481, "step": 82110 }, { "epoch": 1.6101960784313727, "grad_norm": 2.8604705333709717, "learning_rate": 1.0031832419390201e-07, "loss": 0.1462, "step": 82120 }, { "epoch": 1.610392156862745, "grad_norm": 3.162263870239258, "learning_rate": 1.0022094481761634e-07, "loss": 0.1333, "step": 82130 }, { "epoch": 1.6105882352941177, "grad_norm": 2.0133254528045654, "learning_rate": 1.0012360746289589e-07, "loss": 0.1171, "step": 82140 }, { "epoch": 1.6107843137254902, "grad_norm": 4.001467227935791, "learning_rate": 1.0002631213997182e-07, "loss": 0.1276, "step": 82150 }, { "epoch": 1.6109803921568626, "grad_norm": 2.9682726860046387, "learning_rate": 9.992905885907121e-08, "loss": 0.1523, "step": 82160 }, { "epoch": 1.6111764705882354, "grad_norm": 2.783081293106079, "learning_rate": 9.983184763041657e-08, "loss": 0.1535, "step": 82170 }, { "epoch": 1.6113725490196078, "grad_norm": 3.4858617782592773, "learning_rate": 9.973467846422596e-08, "loss": 0.1253, "step": 82180 }, { "epoch": 1.6115686274509804, "grad_norm": 2.2477049827575684, "learning_rate": 9.963755137071278e-08, "loss": 0.139, "step": 82190 }, { "epoch": 1.611764705882353, "grad_norm": 1.9032362699508667, "learning_rate": 9.954046636008662e-08, "loss": 0.1555, "step": 82200 }, { "epoch": 1.6119607843137254, "grad_norm": 2.875558614730835, "learning_rate": 9.944342344255208e-08, "loss": 0.1576, "step": 82210 }, { "epoch": 1.6121568627450982, "grad_norm": 1.8000099658966064, "learning_rate": 9.934642262830939e-08, "loss": 0.1486, "step": 82220 }, { "epoch": 1.6123529411764705, "grad_norm": 1.5237786769866943, "learning_rate": 9.924946392755479e-08, "loss": 0.128, "step": 82230 }, { "epoch": 1.6125490196078431, "grad_norm": 2.7033634185791016, "learning_rate": 9.915254735047961e-08, "loss": 0.1325, "step": 82240 }, { "epoch": 1.6127450980392157, "grad_norm": 2.9176669120788574, "learning_rate": 9.905567290727096e-08, "loss": 0.138, "step": 82250 }, { "epoch": 1.612941176470588, "grad_norm": 2.4741628170013428, "learning_rate": 9.89588406081116e-08, "loss": 0.1353, "step": 82260 }, { "epoch": 1.613137254901961, "grad_norm": 1.6014059782028198, "learning_rate": 9.886205046317969e-08, "loss": 0.1765, "step": 82270 }, { "epoch": 1.6133333333333333, "grad_norm": 2.302802562713623, "learning_rate": 9.876530248264898e-08, "loss": 0.1389, "step": 82280 }, { "epoch": 1.6135294117647059, "grad_norm": 2.797407865524292, "learning_rate": 9.866859667668898e-08, "loss": 0.1451, "step": 82290 }, { "epoch": 1.6137254901960785, "grad_norm": 1.9785107374191284, "learning_rate": 9.857193305546458e-08, "loss": 0.1447, "step": 82300 }, { "epoch": 1.6139215686274508, "grad_norm": 1.427921175956726, "learning_rate": 9.847531162913608e-08, "loss": 0.1338, "step": 82310 }, { "epoch": 1.6141176470588237, "grad_norm": 3.397850275039673, "learning_rate": 9.837873240785993e-08, "loss": 0.1468, "step": 82320 }, { "epoch": 1.614313725490196, "grad_norm": 1.7880396842956543, "learning_rate": 9.828219540178751e-08, "loss": 0.1706, "step": 82330 }, { "epoch": 1.6145098039215686, "grad_norm": 4.224120616912842, "learning_rate": 9.81857006210659e-08, "loss": 0.1533, "step": 82340 }, { "epoch": 1.6147058823529412, "grad_norm": 1.4583827257156372, "learning_rate": 9.808924807583807e-08, "loss": 0.1502, "step": 82350 }, { "epoch": 1.6149019607843136, "grad_norm": 4.157112121582031, "learning_rate": 9.799283777624246e-08, "loss": 0.1257, "step": 82360 }, { "epoch": 1.6150980392156864, "grad_norm": 3.720313787460327, "learning_rate": 9.789646973241273e-08, "loss": 0.1564, "step": 82370 }, { "epoch": 1.6152941176470588, "grad_norm": 4.78391170501709, "learning_rate": 9.780014395447817e-08, "loss": 0.1584, "step": 82380 }, { "epoch": 1.6154901960784314, "grad_norm": 2.7544751167297363, "learning_rate": 9.770386045256418e-08, "loss": 0.1257, "step": 82390 }, { "epoch": 1.615686274509804, "grad_norm": 3.028548240661621, "learning_rate": 9.760761923679106e-08, "loss": 0.135, "step": 82400 }, { "epoch": 1.6158823529411763, "grad_norm": 2.2908146381378174, "learning_rate": 9.751142031727478e-08, "loss": 0.1419, "step": 82410 }, { "epoch": 1.6160784313725491, "grad_norm": 4.049477577209473, "learning_rate": 9.741526370412734e-08, "loss": 0.1417, "step": 82420 }, { "epoch": 1.6162745098039215, "grad_norm": 4.192707061767578, "learning_rate": 9.731914940745579e-08, "loss": 0.118, "step": 82430 }, { "epoch": 1.616470588235294, "grad_norm": 3.2558000087738037, "learning_rate": 9.72230774373627e-08, "loss": 0.1646, "step": 82440 }, { "epoch": 1.6166666666666667, "grad_norm": 3.760279893875122, "learning_rate": 9.71270478039467e-08, "loss": 0.136, "step": 82450 }, { "epoch": 1.616862745098039, "grad_norm": 2.3922555446624756, "learning_rate": 9.703106051730153e-08, "loss": 0.1237, "step": 82460 }, { "epoch": 1.6170588235294119, "grad_norm": 2.415221691131592, "learning_rate": 9.693511558751643e-08, "loss": 0.1286, "step": 82470 }, { "epoch": 1.6172549019607843, "grad_norm": 3.2045998573303223, "learning_rate": 9.683921302467668e-08, "loss": 0.1393, "step": 82480 }, { "epoch": 1.6174509803921568, "grad_norm": 2.701259136199951, "learning_rate": 9.674335283886264e-08, "loss": 0.1401, "step": 82490 }, { "epoch": 1.6176470588235294, "grad_norm": 2.133366584777832, "learning_rate": 9.664753504015022e-08, "loss": 0.1423, "step": 82500 }, { "epoch": 1.6178431372549018, "grad_norm": 2.729686737060547, "learning_rate": 9.655175963861123e-08, "loss": 0.165, "step": 82510 }, { "epoch": 1.6180392156862746, "grad_norm": 1.5022571086883545, "learning_rate": 9.645602664431263e-08, "loss": 0.1501, "step": 82520 }, { "epoch": 1.618235294117647, "grad_norm": 3.314688205718994, "learning_rate": 9.636033606731736e-08, "loss": 0.1167, "step": 82530 }, { "epoch": 1.6184313725490196, "grad_norm": 5.011789798736572, "learning_rate": 9.626468791768333e-08, "loss": 0.1886, "step": 82540 }, { "epoch": 1.6186274509803922, "grad_norm": 3.016845464706421, "learning_rate": 9.616908220546461e-08, "loss": 0.1608, "step": 82550 }, { "epoch": 1.6188235294117646, "grad_norm": 2.4550724029541016, "learning_rate": 9.607351894071036e-08, "loss": 0.1288, "step": 82560 }, { "epoch": 1.6190196078431374, "grad_norm": 3.2144343852996826, "learning_rate": 9.597799813346524e-08, "loss": 0.1589, "step": 82570 }, { "epoch": 1.6192156862745097, "grad_norm": 4.140003681182861, "learning_rate": 9.588251979377e-08, "loss": 0.14, "step": 82580 }, { "epoch": 1.6194117647058823, "grad_norm": 2.5377542972564697, "learning_rate": 9.578708393166035e-08, "loss": 0.1544, "step": 82590 }, { "epoch": 1.619607843137255, "grad_norm": 1.4429064989089966, "learning_rate": 9.569169055716769e-08, "loss": 0.1501, "step": 82600 }, { "epoch": 1.6198039215686273, "grad_norm": 3.1604537963867188, "learning_rate": 9.559633968031916e-08, "loss": 0.1133, "step": 82610 }, { "epoch": 1.62, "grad_norm": 1.9297999143600464, "learning_rate": 9.550103131113718e-08, "loss": 0.1592, "step": 82620 }, { "epoch": 1.6201960784313725, "grad_norm": 2.6294310092926025, "learning_rate": 9.540576545963968e-08, "loss": 0.162, "step": 82630 }, { "epoch": 1.620392156862745, "grad_norm": 1.3562066555023193, "learning_rate": 9.531054213584056e-08, "loss": 0.1565, "step": 82640 }, { "epoch": 1.6205882352941177, "grad_norm": 6.242929935455322, "learning_rate": 9.521536134974873e-08, "loss": 0.1351, "step": 82650 }, { "epoch": 1.62078431372549, "grad_norm": 1.6231976747512817, "learning_rate": 9.51202231113687e-08, "loss": 0.1533, "step": 82660 }, { "epoch": 1.6209803921568628, "grad_norm": 2.7215323448181152, "learning_rate": 9.502512743070096e-08, "loss": 0.1559, "step": 82670 }, { "epoch": 1.6211764705882352, "grad_norm": 2.1315808296203613, "learning_rate": 9.493007431774103e-08, "loss": 0.1531, "step": 82680 }, { "epoch": 1.6213725490196078, "grad_norm": 2.2689666748046875, "learning_rate": 9.483506378247996e-08, "loss": 0.136, "step": 82690 }, { "epoch": 1.6215686274509804, "grad_norm": 1.9923396110534668, "learning_rate": 9.474009583490467e-08, "loss": 0.151, "step": 82700 }, { "epoch": 1.6217647058823528, "grad_norm": 2.3521716594696045, "learning_rate": 9.46451704849976e-08, "loss": 0.1234, "step": 82710 }, { "epoch": 1.6219607843137256, "grad_norm": 3.955672025680542, "learning_rate": 9.455028774273633e-08, "loss": 0.1742, "step": 82720 }, { "epoch": 1.622156862745098, "grad_norm": 2.42781925201416, "learning_rate": 9.445544761809409e-08, "loss": 0.1421, "step": 82730 }, { "epoch": 1.6223529411764706, "grad_norm": 2.321718692779541, "learning_rate": 9.436065012104e-08, "loss": 0.1525, "step": 82740 }, { "epoch": 1.6225490196078431, "grad_norm": 2.589130163192749, "learning_rate": 9.426589526153823e-08, "loss": 0.1565, "step": 82750 }, { "epoch": 1.6227450980392157, "grad_norm": 2.1944994926452637, "learning_rate": 9.417118304954852e-08, "loss": 0.157, "step": 82760 }, { "epoch": 1.6229411764705883, "grad_norm": 4.409675121307373, "learning_rate": 9.407651349502649e-08, "loss": 0.1369, "step": 82770 }, { "epoch": 1.6231372549019607, "grad_norm": 3.351317882537842, "learning_rate": 9.3981886607923e-08, "loss": 0.1499, "step": 82780 }, { "epoch": 1.6233333333333333, "grad_norm": 2.007736921310425, "learning_rate": 9.388730239818432e-08, "loss": 0.1345, "step": 82790 }, { "epoch": 1.6235294117647059, "grad_norm": 5.097679615020752, "learning_rate": 9.379276087575255e-08, "loss": 0.1576, "step": 82800 }, { "epoch": 1.6237254901960785, "grad_norm": 4.476324081420898, "learning_rate": 9.369826205056513e-08, "loss": 0.159, "step": 82810 }, { "epoch": 1.623921568627451, "grad_norm": 3.1304092407226562, "learning_rate": 9.36038059325548e-08, "loss": 0.1316, "step": 82820 }, { "epoch": 1.6241176470588234, "grad_norm": 3.5368103981018066, "learning_rate": 9.350939253165035e-08, "loss": 0.1189, "step": 82830 }, { "epoch": 1.624313725490196, "grad_norm": 3.064030408859253, "learning_rate": 9.341502185777555e-08, "loss": 0.1509, "step": 82840 }, { "epoch": 1.6245098039215686, "grad_norm": 2.489629030227661, "learning_rate": 9.332069392084985e-08, "loss": 0.1507, "step": 82850 }, { "epoch": 1.6247058823529412, "grad_norm": 2.0302202701568604, "learning_rate": 9.322640873078846e-08, "loss": 0.1211, "step": 82860 }, { "epoch": 1.6249019607843138, "grad_norm": 3.4993629455566406, "learning_rate": 9.313216629750176e-08, "loss": 0.18, "step": 82870 }, { "epoch": 1.6250980392156862, "grad_norm": 1.2534568309783936, "learning_rate": 9.303796663089557e-08, "loss": 0.1322, "step": 82880 }, { "epoch": 1.6252941176470588, "grad_norm": 2.205672264099121, "learning_rate": 9.294380974087168e-08, "loss": 0.1197, "step": 82890 }, { "epoch": 1.6254901960784314, "grad_norm": 3.0427935123443604, "learning_rate": 9.284969563732709e-08, "loss": 0.1449, "step": 82900 }, { "epoch": 1.625686274509804, "grad_norm": 3.0570671558380127, "learning_rate": 9.275562433015421e-08, "loss": 0.1267, "step": 82910 }, { "epoch": 1.6258823529411766, "grad_norm": 3.6334872245788574, "learning_rate": 9.266159582924099e-08, "loss": 0.1309, "step": 82920 }, { "epoch": 1.626078431372549, "grad_norm": 1.6033633947372437, "learning_rate": 9.256761014447117e-08, "loss": 0.1783, "step": 82930 }, { "epoch": 1.6262745098039215, "grad_norm": 2.9864108562469482, "learning_rate": 9.24736672857237e-08, "loss": 0.149, "step": 82940 }, { "epoch": 1.6264705882352941, "grad_norm": 2.468981981277466, "learning_rate": 9.237976726287289e-08, "loss": 0.1638, "step": 82950 }, { "epoch": 1.6266666666666667, "grad_norm": 0.9607957005500793, "learning_rate": 9.228591008578901e-08, "loss": 0.1468, "step": 82960 }, { "epoch": 1.6268627450980393, "grad_norm": 4.605683326721191, "learning_rate": 9.219209576433757e-08, "loss": 0.1337, "step": 82970 }, { "epoch": 1.6270588235294117, "grad_norm": 3.2735769748687744, "learning_rate": 9.209832430837928e-08, "loss": 0.1405, "step": 82980 }, { "epoch": 1.6272549019607843, "grad_norm": 2.0058939456939697, "learning_rate": 9.200459572777103e-08, "loss": 0.1203, "step": 82990 }, { "epoch": 1.6274509803921569, "grad_norm": 2.637845277786255, "learning_rate": 9.191091003236463e-08, "loss": 0.1459, "step": 83000 }, { "epoch": 1.6276470588235294, "grad_norm": 3.831618547439575, "learning_rate": 9.181726723200744e-08, "loss": 0.1308, "step": 83010 }, { "epoch": 1.627843137254902, "grad_norm": 2.264291763305664, "learning_rate": 9.172366733654268e-08, "loss": 0.1936, "step": 83020 }, { "epoch": 1.6280392156862744, "grad_norm": 3.2978100776672363, "learning_rate": 9.163011035580881e-08, "loss": 0.1515, "step": 83030 }, { "epoch": 1.6282352941176472, "grad_norm": 1.2674640417099, "learning_rate": 9.15365962996395e-08, "loss": 0.1402, "step": 83040 }, { "epoch": 1.6284313725490196, "grad_norm": 3.7067644596099854, "learning_rate": 9.144312517786451e-08, "loss": 0.1361, "step": 83050 }, { "epoch": 1.6286274509803922, "grad_norm": 4.983558177947998, "learning_rate": 9.134969700030854e-08, "loss": 0.1476, "step": 83060 }, { "epoch": 1.6288235294117648, "grad_norm": 2.5266551971435547, "learning_rate": 9.125631177679227e-08, "loss": 0.1457, "step": 83070 }, { "epoch": 1.6290196078431372, "grad_norm": 3.469972848892212, "learning_rate": 9.116296951713131e-08, "loss": 0.1298, "step": 83080 }, { "epoch": 1.62921568627451, "grad_norm": 2.3757436275482178, "learning_rate": 9.106967023113737e-08, "loss": 0.1534, "step": 83090 }, { "epoch": 1.6294117647058823, "grad_norm": 2.9316229820251465, "learning_rate": 9.09764139286171e-08, "loss": 0.1287, "step": 83100 }, { "epoch": 1.629607843137255, "grad_norm": 1.464553952217102, "learning_rate": 9.088320061937282e-08, "loss": 0.1588, "step": 83110 }, { "epoch": 1.6298039215686275, "grad_norm": 1.1713261604309082, "learning_rate": 9.079003031320259e-08, "loss": 0.1575, "step": 83120 }, { "epoch": 1.63, "grad_norm": 3.484830379486084, "learning_rate": 9.069690301989957e-08, "loss": 0.1571, "step": 83130 }, { "epoch": 1.6301960784313727, "grad_norm": 2.0322141647338867, "learning_rate": 9.060381874925244e-08, "loss": 0.1326, "step": 83140 }, { "epoch": 1.630392156862745, "grad_norm": 3.1409032344818115, "learning_rate": 9.051077751104569e-08, "loss": 0.1307, "step": 83150 }, { "epoch": 1.6305882352941177, "grad_norm": 3.213338613510132, "learning_rate": 9.041777931505907e-08, "loss": 0.1919, "step": 83160 }, { "epoch": 1.6307843137254903, "grad_norm": 4.342961311340332, "learning_rate": 9.032482417106752e-08, "loss": 0.1287, "step": 83170 }, { "epoch": 1.6309803921568626, "grad_norm": 1.9436830282211304, "learning_rate": 9.023191208884207e-08, "loss": 0.1619, "step": 83180 }, { "epoch": 1.6311764705882354, "grad_norm": 3.9318783283233643, "learning_rate": 9.013904307814874e-08, "loss": 0.1392, "step": 83190 }, { "epoch": 1.6313725490196078, "grad_norm": 3.508897542953491, "learning_rate": 9.00462171487491e-08, "loss": 0.1372, "step": 83200 }, { "epoch": 1.6315686274509804, "grad_norm": 2.997346878051758, "learning_rate": 8.995343431040048e-08, "loss": 0.1325, "step": 83210 }, { "epoch": 1.631764705882353, "grad_norm": 5.147337913513184, "learning_rate": 8.986069457285528e-08, "loss": 0.1821, "step": 83220 }, { "epoch": 1.6319607843137254, "grad_norm": 8.436978340148926, "learning_rate": 8.976799794586154e-08, "loss": 0.1562, "step": 83230 }, { "epoch": 1.6321568627450982, "grad_norm": 5.043769359588623, "learning_rate": 8.967534443916286e-08, "loss": 0.1512, "step": 83240 }, { "epoch": 1.6323529411764706, "grad_norm": 1.6177935600280762, "learning_rate": 8.958273406249838e-08, "loss": 0.1569, "step": 83250 }, { "epoch": 1.6325490196078432, "grad_norm": 2.2489848136901855, "learning_rate": 8.94901668256024e-08, "loss": 0.1501, "step": 83260 }, { "epoch": 1.6327450980392157, "grad_norm": 3.891209602355957, "learning_rate": 8.939764273820477e-08, "loss": 0.1448, "step": 83270 }, { "epoch": 1.6329411764705881, "grad_norm": 1.185903549194336, "learning_rate": 8.930516181003106e-08, "loss": 0.1218, "step": 83280 }, { "epoch": 1.633137254901961, "grad_norm": 3.4268462657928467, "learning_rate": 8.921272405080204e-08, "loss": 0.1434, "step": 83290 }, { "epoch": 1.6333333333333333, "grad_norm": 1.2549998760223389, "learning_rate": 8.91203294702339e-08, "loss": 0.1191, "step": 83300 }, { "epoch": 1.633529411764706, "grad_norm": 1.8907809257507324, "learning_rate": 8.902797807803869e-08, "loss": 0.1333, "step": 83310 }, { "epoch": 1.6337254901960785, "grad_norm": 1.8025859594345093, "learning_rate": 8.893566988392343e-08, "loss": 0.1445, "step": 83320 }, { "epoch": 1.6339215686274509, "grad_norm": 2.634967565536499, "learning_rate": 8.884340489759079e-08, "loss": 0.1435, "step": 83330 }, { "epoch": 1.6341176470588237, "grad_norm": 3.724515676498413, "learning_rate": 8.875118312873914e-08, "loss": 0.1458, "step": 83340 }, { "epoch": 1.634313725490196, "grad_norm": 2.526797294616699, "learning_rate": 8.8659004587062e-08, "loss": 0.1604, "step": 83350 }, { "epoch": 1.6345098039215686, "grad_norm": 4.517382621765137, "learning_rate": 8.856686928224822e-08, "loss": 0.1428, "step": 83360 }, { "epoch": 1.6347058823529412, "grad_norm": 2.780508041381836, "learning_rate": 8.847477722398267e-08, "loss": 0.149, "step": 83370 }, { "epoch": 1.6349019607843136, "grad_norm": 2.1330363750457764, "learning_rate": 8.838272842194516e-08, "loss": 0.1533, "step": 83380 }, { "epoch": 1.6350980392156864, "grad_norm": 2.173766613006592, "learning_rate": 8.829072288581097e-08, "loss": 0.1395, "step": 83390 }, { "epoch": 1.6352941176470588, "grad_norm": 2.714606523513794, "learning_rate": 8.819876062525133e-08, "loss": 0.1462, "step": 83400 }, { "epoch": 1.6354901960784314, "grad_norm": 1.549850344657898, "learning_rate": 8.810684164993221e-08, "loss": 0.1974, "step": 83410 }, { "epoch": 1.635686274509804, "grad_norm": 2.4309279918670654, "learning_rate": 8.801496596951574e-08, "loss": 0.1221, "step": 83420 }, { "epoch": 1.6358823529411763, "grad_norm": 3.829111337661743, "learning_rate": 8.792313359365882e-08, "loss": 0.1259, "step": 83430 }, { "epoch": 1.6360784313725492, "grad_norm": 2.799696207046509, "learning_rate": 8.783134453201441e-08, "loss": 0.1471, "step": 83440 }, { "epoch": 1.6362745098039215, "grad_norm": 4.084920406341553, "learning_rate": 8.773959879423054e-08, "loss": 0.1473, "step": 83450 }, { "epoch": 1.6364705882352941, "grad_norm": 1.7610037326812744, "learning_rate": 8.764789638995069e-08, "loss": 0.1228, "step": 83460 }, { "epoch": 1.6366666666666667, "grad_norm": 3.362905740737915, "learning_rate": 8.755623732881407e-08, "loss": 0.1113, "step": 83470 }, { "epoch": 1.636862745098039, "grad_norm": 5.075390815734863, "learning_rate": 8.746462162045498e-08, "loss": 0.1711, "step": 83480 }, { "epoch": 1.637058823529412, "grad_norm": 3.8316731452941895, "learning_rate": 8.737304927450334e-08, "loss": 0.1817, "step": 83490 }, { "epoch": 1.6372549019607843, "grad_norm": 1.8360552787780762, "learning_rate": 8.728152030058462e-08, "loss": 0.1683, "step": 83500 }, { "epoch": 1.6374509803921569, "grad_norm": 1.7557146549224854, "learning_rate": 8.719003470831954e-08, "loss": 0.159, "step": 83510 }, { "epoch": 1.6376470588235295, "grad_norm": 2.8061845302581787, "learning_rate": 8.709859250732421e-08, "loss": 0.1595, "step": 83520 }, { "epoch": 1.6378431372549018, "grad_norm": 2.337405204772949, "learning_rate": 8.700719370721055e-08, "loss": 0.1304, "step": 83530 }, { "epoch": 1.6380392156862746, "grad_norm": 1.852823257446289, "learning_rate": 8.691583831758553e-08, "loss": 0.1839, "step": 83540 }, { "epoch": 1.638235294117647, "grad_norm": 4.392087936401367, "learning_rate": 8.682452634805154e-08, "loss": 0.2082, "step": 83550 }, { "epoch": 1.6384313725490196, "grad_norm": 2.051410675048828, "learning_rate": 8.67332578082069e-08, "loss": 0.1183, "step": 83560 }, { "epoch": 1.6386274509803922, "grad_norm": 3.1256160736083984, "learning_rate": 8.664203270764481e-08, "loss": 0.157, "step": 83570 }, { "epoch": 1.6388235294117646, "grad_norm": 4.191614151000977, "learning_rate": 8.655085105595405e-08, "loss": 0.138, "step": 83580 }, { "epoch": 1.6390196078431374, "grad_norm": 2.4111344814300537, "learning_rate": 8.645971286271903e-08, "loss": 0.1628, "step": 83590 }, { "epoch": 1.6392156862745098, "grad_norm": 4.0711822509765625, "learning_rate": 8.636861813751956e-08, "loss": 0.1566, "step": 83600 }, { "epoch": 1.6394117647058823, "grad_norm": 2.412484884262085, "learning_rate": 8.627756688993065e-08, "loss": 0.1223, "step": 83610 }, { "epoch": 1.639607843137255, "grad_norm": 10.781309127807617, "learning_rate": 8.618655912952283e-08, "loss": 0.1504, "step": 83620 }, { "epoch": 1.6398039215686273, "grad_norm": 2.7006144523620605, "learning_rate": 8.609559486586227e-08, "loss": 0.1223, "step": 83630 }, { "epoch": 1.6400000000000001, "grad_norm": 3.0489237308502197, "learning_rate": 8.600467410851026e-08, "loss": 0.1713, "step": 83640 }, { "epoch": 1.6401960784313725, "grad_norm": 1.1319565773010254, "learning_rate": 8.591379686702365e-08, "loss": 0.1458, "step": 83650 }, { "epoch": 1.640392156862745, "grad_norm": 3.765763759613037, "learning_rate": 8.582296315095489e-08, "loss": 0.179, "step": 83660 }, { "epoch": 1.6405882352941177, "grad_norm": 3.917611598968506, "learning_rate": 8.573217296985158e-08, "loss": 0.1715, "step": 83670 }, { "epoch": 1.64078431372549, "grad_norm": 4.497402191162109, "learning_rate": 8.564142633325671e-08, "loss": 0.1531, "step": 83680 }, { "epoch": 1.6409803921568629, "grad_norm": 1.5784128904342651, "learning_rate": 8.555072325070916e-08, "loss": 0.128, "step": 83690 }, { "epoch": 1.6411764705882352, "grad_norm": 1.7821215391159058, "learning_rate": 8.54600637317427e-08, "loss": 0.1296, "step": 83700 }, { "epoch": 1.6413725490196078, "grad_norm": 3.650923490524292, "learning_rate": 8.536944778588667e-08, "loss": 0.1134, "step": 83710 }, { "epoch": 1.6415686274509804, "grad_norm": 3.4837112426757812, "learning_rate": 8.52788754226661e-08, "loss": 0.1517, "step": 83720 }, { "epoch": 1.6417647058823528, "grad_norm": 3.288219690322876, "learning_rate": 8.518834665160113e-08, "loss": 0.1325, "step": 83730 }, { "epoch": 1.6419607843137256, "grad_norm": 3.105612277984619, "learning_rate": 8.509786148220732e-08, "loss": 0.1642, "step": 83740 }, { "epoch": 1.642156862745098, "grad_norm": 0.769385576248169, "learning_rate": 8.500741992399591e-08, "loss": 0.1457, "step": 83750 }, { "epoch": 1.6423529411764706, "grad_norm": 2.4422028064727783, "learning_rate": 8.491702198647333e-08, "loss": 0.1179, "step": 83760 }, { "epoch": 1.6425490196078432, "grad_norm": 2.37018084526062, "learning_rate": 8.482666767914132e-08, "loss": 0.1374, "step": 83770 }, { "epoch": 1.6427450980392155, "grad_norm": 3.4627397060394287, "learning_rate": 8.473635701149734e-08, "loss": 0.1391, "step": 83780 }, { "epoch": 1.6429411764705883, "grad_norm": 3.8847100734710693, "learning_rate": 8.46460899930343e-08, "loss": 0.1317, "step": 83790 }, { "epoch": 1.6431372549019607, "grad_norm": 3.9726173877716064, "learning_rate": 8.455586663324016e-08, "loss": 0.1384, "step": 83800 }, { "epoch": 1.6433333333333333, "grad_norm": 2.7296454906463623, "learning_rate": 8.446568694159834e-08, "loss": 0.1593, "step": 83810 }, { "epoch": 1.643529411764706, "grad_norm": 3.460960626602173, "learning_rate": 8.437555092758808e-08, "loss": 0.1554, "step": 83820 }, { "epoch": 1.6437254901960783, "grad_norm": 3.1000964641571045, "learning_rate": 8.428545860068364e-08, "loss": 0.1324, "step": 83830 }, { "epoch": 1.643921568627451, "grad_norm": 3.6919267177581787, "learning_rate": 8.419540997035473e-08, "loss": 0.1652, "step": 83840 }, { "epoch": 1.6441176470588235, "grad_norm": 2.653611898422241, "learning_rate": 8.410540504606667e-08, "loss": 0.1458, "step": 83850 }, { "epoch": 1.644313725490196, "grad_norm": 3.805859088897705, "learning_rate": 8.401544383727999e-08, "loss": 0.1627, "step": 83860 }, { "epoch": 1.6445098039215686, "grad_norm": 3.7992966175079346, "learning_rate": 8.392552635345051e-08, "loss": 0.1728, "step": 83870 }, { "epoch": 1.644705882352941, "grad_norm": 3.6081807613372803, "learning_rate": 8.383565260403003e-08, "loss": 0.1645, "step": 83880 }, { "epoch": 1.6449019607843138, "grad_norm": 6.571556091308594, "learning_rate": 8.374582259846507e-08, "loss": 0.1325, "step": 83890 }, { "epoch": 1.6450980392156862, "grad_norm": 2.661325693130493, "learning_rate": 8.365603634619784e-08, "loss": 0.1411, "step": 83900 }, { "epoch": 1.6452941176470588, "grad_norm": 2.5039000511169434, "learning_rate": 8.356629385666608e-08, "loss": 0.1436, "step": 83910 }, { "epoch": 1.6454901960784314, "grad_norm": 3.846479892730713, "learning_rate": 8.347659513930277e-08, "loss": 0.1446, "step": 83920 }, { "epoch": 1.6456862745098038, "grad_norm": 2.875473737716675, "learning_rate": 8.338694020353615e-08, "loss": 0.143, "step": 83930 }, { "epoch": 1.6458823529411766, "grad_norm": 3.913328170776367, "learning_rate": 8.329732905879028e-08, "loss": 0.1694, "step": 83940 }, { "epoch": 1.646078431372549, "grad_norm": 1.6830860376358032, "learning_rate": 8.320776171448418e-08, "loss": 0.1356, "step": 83950 }, { "epoch": 1.6462745098039215, "grad_norm": 2.8380188941955566, "learning_rate": 8.311823818003261e-08, "loss": 0.163, "step": 83960 }, { "epoch": 1.6464705882352941, "grad_norm": 2.106247901916504, "learning_rate": 8.30287584648453e-08, "loss": 0.1343, "step": 83970 }, { "epoch": 1.6466666666666665, "grad_norm": 2.538419723510742, "learning_rate": 8.293932257832797e-08, "loss": 0.1475, "step": 83980 }, { "epoch": 1.6468627450980393, "grad_norm": 5.967103004455566, "learning_rate": 8.284993052988126e-08, "loss": 0.1351, "step": 83990 }, { "epoch": 1.6470588235294117, "grad_norm": 2.743112087249756, "learning_rate": 8.27605823289012e-08, "loss": 0.1487, "step": 84000 }, { "epoch": 1.6472549019607843, "grad_norm": 2.102877378463745, "learning_rate": 8.267127798477963e-08, "loss": 0.1373, "step": 84010 }, { "epoch": 1.6474509803921569, "grad_norm": 1.8756741285324097, "learning_rate": 8.258201750690331e-08, "loss": 0.1674, "step": 84020 }, { "epoch": 1.6476470588235295, "grad_norm": 2.6736867427825928, "learning_rate": 8.249280090465455e-08, "loss": 0.158, "step": 84030 }, { "epoch": 1.647843137254902, "grad_norm": 2.148125648498535, "learning_rate": 8.240362818741126e-08, "loss": 0.1561, "step": 84040 }, { "epoch": 1.6480392156862744, "grad_norm": 2.05061411857605, "learning_rate": 8.231449936454654e-08, "loss": 0.13, "step": 84050 }, { "epoch": 1.648235294117647, "grad_norm": 2.4657018184661865, "learning_rate": 8.222541444542864e-08, "loss": 0.1196, "step": 84060 }, { "epoch": 1.6484313725490196, "grad_norm": 3.6528327465057373, "learning_rate": 8.213637343942176e-08, "loss": 0.132, "step": 84070 }, { "epoch": 1.6486274509803922, "grad_norm": 3.291351079940796, "learning_rate": 8.20473763558851e-08, "loss": 0.1469, "step": 84080 }, { "epoch": 1.6488235294117648, "grad_norm": 1.4967148303985596, "learning_rate": 8.195842320417307e-08, "loss": 0.163, "step": 84090 }, { "epoch": 1.6490196078431372, "grad_norm": 3.427520513534546, "learning_rate": 8.186951399363612e-08, "loss": 0.1497, "step": 84100 }, { "epoch": 1.6492156862745098, "grad_norm": 2.2883718013763428, "learning_rate": 8.17806487336194e-08, "loss": 0.1547, "step": 84110 }, { "epoch": 1.6494117647058824, "grad_norm": 1.8351792097091675, "learning_rate": 8.16918274334637e-08, "loss": 0.0971, "step": 84120 }, { "epoch": 1.649607843137255, "grad_norm": 1.8924827575683594, "learning_rate": 8.160305010250528e-08, "loss": 0.1463, "step": 84130 }, { "epoch": 1.6498039215686275, "grad_norm": 1.537138819694519, "learning_rate": 8.15143167500758e-08, "loss": 0.1502, "step": 84140 }, { "epoch": 1.65, "grad_norm": 3.1892387866973877, "learning_rate": 8.142562738550213e-08, "loss": 0.1233, "step": 84150 }, { "epoch": 1.6501960784313725, "grad_norm": 3.92061185836792, "learning_rate": 8.13369820181064e-08, "loss": 0.1589, "step": 84160 }, { "epoch": 1.650392156862745, "grad_norm": 2.737957715988159, "learning_rate": 8.124838065720657e-08, "loss": 0.1494, "step": 84170 }, { "epoch": 1.6505882352941177, "grad_norm": 1.5317621231079102, "learning_rate": 8.115982331211563e-08, "loss": 0.1206, "step": 84180 }, { "epoch": 1.6507843137254903, "grad_norm": 4.29496955871582, "learning_rate": 8.10713099921418e-08, "loss": 0.1832, "step": 84190 }, { "epoch": 1.6509803921568627, "grad_norm": 2.8943214416503906, "learning_rate": 8.098284070658923e-08, "loss": 0.1364, "step": 84200 }, { "epoch": 1.6511764705882352, "grad_norm": 4.410364151000977, "learning_rate": 8.08944154647569e-08, "loss": 0.1458, "step": 84210 }, { "epoch": 1.6513725490196078, "grad_norm": 2.088629961013794, "learning_rate": 8.080603427593929e-08, "loss": 0.139, "step": 84220 }, { "epoch": 1.6515686274509804, "grad_norm": 2.7399919033050537, "learning_rate": 8.071769714942656e-08, "loss": 0.1556, "step": 84230 }, { "epoch": 1.651764705882353, "grad_norm": 3.9622113704681396, "learning_rate": 8.062940409450386e-08, "loss": 0.1836, "step": 84240 }, { "epoch": 1.6519607843137254, "grad_norm": 2.421043872833252, "learning_rate": 8.054115512045174e-08, "loss": 0.132, "step": 84250 }, { "epoch": 1.652156862745098, "grad_norm": 2.9454236030578613, "learning_rate": 8.045295023654641e-08, "loss": 0.1272, "step": 84260 }, { "epoch": 1.6523529411764706, "grad_norm": 3.5367329120635986, "learning_rate": 8.036478945205922e-08, "loss": 0.113, "step": 84270 }, { "epoch": 1.6525490196078432, "grad_norm": 5.057125568389893, "learning_rate": 8.027667277625677e-08, "loss": 0.1887, "step": 84280 }, { "epoch": 1.6527450980392158, "grad_norm": 3.7985312938690186, "learning_rate": 8.01886002184014e-08, "loss": 0.1315, "step": 84290 }, { "epoch": 1.6529411764705881, "grad_norm": 2.5162336826324463, "learning_rate": 8.010057178775054e-08, "loss": 0.1523, "step": 84300 }, { "epoch": 1.653137254901961, "grad_norm": 3.084120035171509, "learning_rate": 8.001258749355682e-08, "loss": 0.1589, "step": 84310 }, { "epoch": 1.6533333333333333, "grad_norm": 3.735044240951538, "learning_rate": 7.992464734506865e-08, "loss": 0.1579, "step": 84320 }, { "epoch": 1.653529411764706, "grad_norm": 2.039736747741699, "learning_rate": 7.983675135152962e-08, "loss": 0.1493, "step": 84330 }, { "epoch": 1.6537254901960785, "grad_norm": 2.8429667949676514, "learning_rate": 7.974889952217862e-08, "loss": 0.1373, "step": 84340 }, { "epoch": 1.6539215686274509, "grad_norm": 3.5107345581054688, "learning_rate": 7.96610918662497e-08, "loss": 0.1636, "step": 84350 }, { "epoch": 1.6541176470588237, "grad_norm": 1.6930837631225586, "learning_rate": 7.957332839297287e-08, "loss": 0.1458, "step": 84360 }, { "epoch": 1.654313725490196, "grad_norm": 2.0675158500671387, "learning_rate": 7.948560911157287e-08, "loss": 0.1262, "step": 84370 }, { "epoch": 1.6545098039215687, "grad_norm": 4.800974369049072, "learning_rate": 7.939793403127004e-08, "loss": 0.1448, "step": 84380 }, { "epoch": 1.6547058823529412, "grad_norm": 1.8982322216033936, "learning_rate": 7.931030316128024e-08, "loss": 0.1473, "step": 84390 }, { "epoch": 1.6549019607843136, "grad_norm": 1.8980823755264282, "learning_rate": 7.922271651081446e-08, "loss": 0.1366, "step": 84400 }, { "epoch": 1.6550980392156864, "grad_norm": 2.080777645111084, "learning_rate": 7.9135174089079e-08, "loss": 0.1308, "step": 84410 }, { "epoch": 1.6552941176470588, "grad_norm": 2.184858560562134, "learning_rate": 7.904767590527572e-08, "loss": 0.1409, "step": 84420 }, { "epoch": 1.6554901960784314, "grad_norm": 4.459595680236816, "learning_rate": 7.89602219686018e-08, "loss": 0.1502, "step": 84430 }, { "epoch": 1.655686274509804, "grad_norm": 3.988987922668457, "learning_rate": 7.887281228824943e-08, "loss": 0.1345, "step": 84440 }, { "epoch": 1.6558823529411764, "grad_norm": 2.7782208919525146, "learning_rate": 7.878544687340671e-08, "loss": 0.1465, "step": 84450 }, { "epoch": 1.6560784313725492, "grad_norm": 2.980189085006714, "learning_rate": 7.869812573325669e-08, "loss": 0.1331, "step": 84460 }, { "epoch": 1.6562745098039215, "grad_norm": 4.805995941162109, "learning_rate": 7.861084887697766e-08, "loss": 0.134, "step": 84470 }, { "epoch": 1.6564705882352941, "grad_norm": 1.2071763277053833, "learning_rate": 7.852361631374383e-08, "loss": 0.1226, "step": 84480 }, { "epoch": 1.6566666666666667, "grad_norm": 2.371898889541626, "learning_rate": 7.843642805272405e-08, "loss": 0.1497, "step": 84490 }, { "epoch": 1.656862745098039, "grad_norm": 2.656053066253662, "learning_rate": 7.834928410308311e-08, "loss": 0.178, "step": 84500 }, { "epoch": 1.657058823529412, "grad_norm": 2.147796630859375, "learning_rate": 7.82621844739807e-08, "loss": 0.1422, "step": 84510 }, { "epoch": 1.6572549019607843, "grad_norm": 4.692461013793945, "learning_rate": 7.81751291745722e-08, "loss": 0.1285, "step": 84520 }, { "epoch": 1.6574509803921569, "grad_norm": 2.955658197402954, "learning_rate": 7.808811821400807e-08, "loss": 0.1409, "step": 84530 }, { "epoch": 1.6576470588235295, "grad_norm": 2.7267279624938965, "learning_rate": 7.80011516014341e-08, "loss": 0.1322, "step": 84540 }, { "epoch": 1.6578431372549018, "grad_norm": 4.695857048034668, "learning_rate": 7.791422934599173e-08, "loss": 0.1804, "step": 84550 }, { "epoch": 1.6580392156862747, "grad_norm": 2.054504871368408, "learning_rate": 7.782735145681746e-08, "loss": 0.1537, "step": 84560 }, { "epoch": 1.658235294117647, "grad_norm": 2.6846234798431396, "learning_rate": 7.774051794304304e-08, "loss": 0.1547, "step": 84570 }, { "epoch": 1.6584313725490196, "grad_norm": 2.9893593788146973, "learning_rate": 7.76537288137959e-08, "loss": 0.1488, "step": 84580 }, { "epoch": 1.6586274509803922, "grad_norm": 1.8971396684646606, "learning_rate": 7.756698407819861e-08, "loss": 0.1498, "step": 84590 }, { "epoch": 1.6588235294117646, "grad_norm": 1.7779343128204346, "learning_rate": 7.748028374536891e-08, "loss": 0.1553, "step": 84600 }, { "epoch": 1.6590196078431374, "grad_norm": 2.54830265045166, "learning_rate": 7.739362782442022e-08, "loss": 0.1584, "step": 84610 }, { "epoch": 1.6592156862745098, "grad_norm": 3.427030086517334, "learning_rate": 7.730701632446113e-08, "loss": 0.1642, "step": 84620 }, { "epoch": 1.6594117647058824, "grad_norm": 4.467417240142822, "learning_rate": 7.722044925459531e-08, "loss": 0.1781, "step": 84630 }, { "epoch": 1.659607843137255, "grad_norm": 1.1506890058517456, "learning_rate": 7.713392662392226e-08, "loss": 0.1333, "step": 84640 }, { "epoch": 1.6598039215686273, "grad_norm": 2.2186279296875, "learning_rate": 7.704744844153644e-08, "loss": 0.1175, "step": 84650 }, { "epoch": 1.6600000000000001, "grad_norm": 1.6769664287567139, "learning_rate": 7.696101471652761e-08, "loss": 0.1486, "step": 84660 }, { "epoch": 1.6601960784313725, "grad_norm": 2.6580748558044434, "learning_rate": 7.687462545798113e-08, "loss": 0.1381, "step": 84670 }, { "epoch": 1.660392156862745, "grad_norm": 3.0513317584991455, "learning_rate": 7.678828067497767e-08, "loss": 0.1332, "step": 84680 }, { "epoch": 1.6605882352941177, "grad_norm": 2.4234678745269775, "learning_rate": 7.670198037659298e-08, "loss": 0.1534, "step": 84690 }, { "epoch": 1.66078431372549, "grad_norm": 3.7833755016326904, "learning_rate": 7.661572457189807e-08, "loss": 0.1518, "step": 84700 }, { "epoch": 1.6609803921568629, "grad_norm": 3.245516061782837, "learning_rate": 7.652951326995977e-08, "loss": 0.1429, "step": 84710 }, { "epoch": 1.6611764705882353, "grad_norm": 4.199728012084961, "learning_rate": 7.644334647983975e-08, "loss": 0.1637, "step": 84720 }, { "epoch": 1.6613725490196078, "grad_norm": 6.455078601837158, "learning_rate": 7.635722421059515e-08, "loss": 0.164, "step": 84730 }, { "epoch": 1.6615686274509804, "grad_norm": 3.097874402999878, "learning_rate": 7.627114647127853e-08, "loss": 0.1178, "step": 84740 }, { "epoch": 1.6617647058823528, "grad_norm": 3.2407450675964355, "learning_rate": 7.618511327093775e-08, "loss": 0.1463, "step": 84750 }, { "epoch": 1.6619607843137256, "grad_norm": 3.71929931640625, "learning_rate": 7.609912461861568e-08, "loss": 0.2087, "step": 84760 }, { "epoch": 1.662156862745098, "grad_norm": 3.2790334224700928, "learning_rate": 7.601318052335098e-08, "loss": 0.1187, "step": 84770 }, { "epoch": 1.6623529411764706, "grad_norm": 3.761132001876831, "learning_rate": 7.592728099417745e-08, "loss": 0.1591, "step": 84780 }, { "epoch": 1.6625490196078432, "grad_norm": 4.275231838226318, "learning_rate": 7.584142604012384e-08, "loss": 0.1709, "step": 84790 }, { "epoch": 1.6627450980392156, "grad_norm": 3.178645133972168, "learning_rate": 7.575561567021487e-08, "loss": 0.1587, "step": 84800 }, { "epoch": 1.6629411764705884, "grad_norm": 2.423557758331299, "learning_rate": 7.566984989347019e-08, "loss": 0.1191, "step": 84810 }, { "epoch": 1.6631372549019607, "grad_norm": 2.562812566757202, "learning_rate": 7.558412871890452e-08, "loss": 0.1821, "step": 84820 }, { "epoch": 1.6633333333333333, "grad_norm": 2.0338985919952393, "learning_rate": 7.549845215552858e-08, "loss": 0.14, "step": 84830 }, { "epoch": 1.663529411764706, "grad_norm": 3.6154327392578125, "learning_rate": 7.541282021234769e-08, "loss": 0.1488, "step": 84840 }, { "epoch": 1.6637254901960783, "grad_norm": 3.2630085945129395, "learning_rate": 7.532723289836302e-08, "loss": 0.1391, "step": 84850 }, { "epoch": 1.663921568627451, "grad_norm": 1.7219468355178833, "learning_rate": 7.524169022257054e-08, "loss": 0.1586, "step": 84860 }, { "epoch": 1.6641176470588235, "grad_norm": 3.3451766967773438, "learning_rate": 7.515619219396219e-08, "loss": 0.1807, "step": 84870 }, { "epoch": 1.664313725490196, "grad_norm": 2.269044876098633, "learning_rate": 7.507073882152459e-08, "loss": 0.1411, "step": 84880 }, { "epoch": 1.6645098039215687, "grad_norm": 2.8452391624450684, "learning_rate": 7.498533011423985e-08, "loss": 0.1389, "step": 84890 }, { "epoch": 1.664705882352941, "grad_norm": 4.142491817474365, "learning_rate": 7.489996608108568e-08, "loss": 0.1456, "step": 84900 }, { "epoch": 1.6649019607843139, "grad_norm": 2.3151516914367676, "learning_rate": 7.481464673103477e-08, "loss": 0.1515, "step": 84910 }, { "epoch": 1.6650980392156862, "grad_norm": 1.5481057167053223, "learning_rate": 7.472937207305502e-08, "loss": 0.1257, "step": 84920 }, { "epoch": 1.6652941176470588, "grad_norm": 4.128742218017578, "learning_rate": 7.464414211611009e-08, "loss": 0.1519, "step": 84930 }, { "epoch": 1.6654901960784314, "grad_norm": 3.30147123336792, "learning_rate": 7.455895686915853e-08, "loss": 0.1462, "step": 84940 }, { "epoch": 1.6656862745098038, "grad_norm": 3.8010833263397217, "learning_rate": 7.447381634115424e-08, "loss": 0.1231, "step": 84950 }, { "epoch": 1.6658823529411766, "grad_norm": 1.6232742071151733, "learning_rate": 7.438872054104672e-08, "loss": 0.1315, "step": 84960 }, { "epoch": 1.666078431372549, "grad_norm": 2.378615379333496, "learning_rate": 7.430366947778049e-08, "loss": 0.1409, "step": 84970 }, { "epoch": 1.6662745098039216, "grad_norm": 2.5554001331329346, "learning_rate": 7.421866316029529e-08, "loss": 0.1469, "step": 84980 }, { "epoch": 1.6664705882352941, "grad_norm": 2.3465840816497803, "learning_rate": 7.413370159752647e-08, "loss": 0.1484, "step": 84990 }, { "epoch": 1.6666666666666665, "grad_norm": 3.1945652961730957, "learning_rate": 7.404878479840448e-08, "loss": 0.1124, "step": 85000 }, { "epoch": 1.6668627450980393, "grad_norm": 2.790224552154541, "learning_rate": 7.396391277185499e-08, "loss": 0.1379, "step": 85010 }, { "epoch": 1.6670588235294117, "grad_norm": 2.1737124919891357, "learning_rate": 7.387908552679905e-08, "loss": 0.1478, "step": 85020 }, { "epoch": 1.6672549019607843, "grad_norm": 4.739912033081055, "learning_rate": 7.379430307215329e-08, "loss": 0.1677, "step": 85030 }, { "epoch": 1.667450980392157, "grad_norm": 4.494415283203125, "learning_rate": 7.370956541682916e-08, "loss": 0.1618, "step": 85040 }, { "epoch": 1.6676470588235293, "grad_norm": 2.730346918106079, "learning_rate": 7.362487256973348e-08, "loss": 0.1616, "step": 85050 }, { "epoch": 1.667843137254902, "grad_norm": 2.0052335262298584, "learning_rate": 7.354022453976877e-08, "loss": 0.1596, "step": 85060 }, { "epoch": 1.6680392156862744, "grad_norm": 4.202548503875732, "learning_rate": 7.345562133583238e-08, "loss": 0.1612, "step": 85070 }, { "epoch": 1.668235294117647, "grad_norm": 4.0304789543151855, "learning_rate": 7.337106296681705e-08, "loss": 0.1605, "step": 85080 }, { "epoch": 1.6684313725490196, "grad_norm": 1.32792329788208, "learning_rate": 7.328654944161106e-08, "loss": 0.128, "step": 85090 }, { "epoch": 1.668627450980392, "grad_norm": 4.887015342712402, "learning_rate": 7.320208076909773e-08, "loss": 0.1702, "step": 85100 }, { "epoch": 1.6688235294117648, "grad_norm": 3.7021868228912354, "learning_rate": 7.311765695815559e-08, "loss": 0.1402, "step": 85110 }, { "epoch": 1.6690196078431372, "grad_norm": 1.5985187292099, "learning_rate": 7.303327801765879e-08, "loss": 0.1395, "step": 85120 }, { "epoch": 1.6692156862745098, "grad_norm": 3.4331789016723633, "learning_rate": 7.294894395647649e-08, "loss": 0.1552, "step": 85130 }, { "epoch": 1.6694117647058824, "grad_norm": 1.3232232332229614, "learning_rate": 7.286465478347309e-08, "loss": 0.1429, "step": 85140 }, { "epoch": 1.6696078431372547, "grad_norm": 3.676354169845581, "learning_rate": 7.27804105075086e-08, "loss": 0.1541, "step": 85150 }, { "epoch": 1.6698039215686276, "grad_norm": 6.2402238845825195, "learning_rate": 7.269621113743807e-08, "loss": 0.1278, "step": 85160 }, { "epoch": 1.67, "grad_norm": 3.945234537124634, "learning_rate": 7.261205668211162e-08, "loss": 0.1459, "step": 85170 }, { "epoch": 1.6701960784313725, "grad_norm": 1.7745367288589478, "learning_rate": 7.25279471503752e-08, "loss": 0.1495, "step": 85180 }, { "epoch": 1.6703921568627451, "grad_norm": 1.647571325302124, "learning_rate": 7.24438825510696e-08, "loss": 0.1345, "step": 85190 }, { "epoch": 1.6705882352941175, "grad_norm": 1.9953951835632324, "learning_rate": 7.235986289303087e-08, "loss": 0.169, "step": 85200 }, { "epoch": 1.6707843137254903, "grad_norm": 2.1291956901550293, "learning_rate": 7.227588818509067e-08, "loss": 0.1562, "step": 85210 }, { "epoch": 1.6709803921568627, "grad_norm": 2.2071523666381836, "learning_rate": 7.219195843607578e-08, "loss": 0.1219, "step": 85220 }, { "epoch": 1.6711764705882353, "grad_norm": 1.3020099401474, "learning_rate": 7.210807365480815e-08, "loss": 0.1512, "step": 85230 }, { "epoch": 1.6713725490196079, "grad_norm": 2.142954111099243, "learning_rate": 7.202423385010497e-08, "loss": 0.1397, "step": 85240 }, { "epoch": 1.6715686274509802, "grad_norm": 5.0357136726379395, "learning_rate": 7.1940439030779e-08, "loss": 0.1609, "step": 85250 }, { "epoch": 1.671764705882353, "grad_norm": 2.6341567039489746, "learning_rate": 7.185668920563798e-08, "loss": 0.148, "step": 85260 }, { "epoch": 1.6719607843137254, "grad_norm": 3.3128371238708496, "learning_rate": 7.177298438348494e-08, "loss": 0.1454, "step": 85270 }, { "epoch": 1.672156862745098, "grad_norm": 10.653000831604004, "learning_rate": 7.168932457311844e-08, "loss": 0.1523, "step": 85280 }, { "epoch": 1.6723529411764706, "grad_norm": 3.7393958568573, "learning_rate": 7.160570978333202e-08, "loss": 0.1217, "step": 85290 }, { "epoch": 1.6725490196078432, "grad_norm": 2.1613621711730957, "learning_rate": 7.15221400229145e-08, "loss": 0.1241, "step": 85300 }, { "epoch": 1.6727450980392158, "grad_norm": 2.8701653480529785, "learning_rate": 7.143861530065032e-08, "loss": 0.1167, "step": 85310 }, { "epoch": 1.6729411764705882, "grad_norm": 1.647568702697754, "learning_rate": 7.135513562531869e-08, "loss": 0.1353, "step": 85320 }, { "epoch": 1.6731372549019607, "grad_norm": 2.7118613719940186, "learning_rate": 7.127170100569436e-08, "loss": 0.1108, "step": 85330 }, { "epoch": 1.6733333333333333, "grad_norm": 2.5277445316314697, "learning_rate": 7.118831145054748e-08, "loss": 0.1373, "step": 85340 }, { "epoch": 1.673529411764706, "grad_norm": 4.3367109298706055, "learning_rate": 7.110496696864316e-08, "loss": 0.1523, "step": 85350 }, { "epoch": 1.6737254901960785, "grad_norm": 2.7555460929870605, "learning_rate": 7.102166756874179e-08, "loss": 0.1459, "step": 85360 }, { "epoch": 1.673921568627451, "grad_norm": 3.5117688179016113, "learning_rate": 7.093841325959932e-08, "loss": 0.1654, "step": 85370 }, { "epoch": 1.6741176470588235, "grad_norm": 1.6712006330490112, "learning_rate": 7.085520404996664e-08, "loss": 0.1314, "step": 85380 }, { "epoch": 1.674313725490196, "grad_norm": 11.596540451049805, "learning_rate": 7.077203994859021e-08, "loss": 0.1531, "step": 85390 }, { "epoch": 1.6745098039215687, "grad_norm": 3.049121141433716, "learning_rate": 7.068892096421136e-08, "loss": 0.1124, "step": 85400 }, { "epoch": 1.6747058823529413, "grad_norm": 2.9749677181243896, "learning_rate": 7.060584710556711e-08, "loss": 0.1409, "step": 85410 }, { "epoch": 1.6749019607843136, "grad_norm": 2.4975366592407227, "learning_rate": 7.05228183813894e-08, "loss": 0.1402, "step": 85420 }, { "epoch": 1.6750980392156862, "grad_norm": 3.9862992763519287, "learning_rate": 7.043983480040545e-08, "loss": 0.1523, "step": 85430 }, { "epoch": 1.6752941176470588, "grad_norm": 4.290897846221924, "learning_rate": 7.035689637133802e-08, "loss": 0.1568, "step": 85440 }, { "epoch": 1.6754901960784314, "grad_norm": 3.0340442657470703, "learning_rate": 7.027400310290482e-08, "loss": 0.1358, "step": 85450 }, { "epoch": 1.675686274509804, "grad_norm": 4.97999906539917, "learning_rate": 7.019115500381889e-08, "loss": 0.1656, "step": 85460 }, { "epoch": 1.6758823529411764, "grad_norm": 4.982423782348633, "learning_rate": 7.010835208278864e-08, "loss": 0.1916, "step": 85470 }, { "epoch": 1.676078431372549, "grad_norm": 1.76437246799469, "learning_rate": 7.002559434851768e-08, "loss": 0.1305, "step": 85480 }, { "epoch": 1.6762745098039216, "grad_norm": 3.0510787963867188, "learning_rate": 6.994288180970464e-08, "loss": 0.1319, "step": 85490 }, { "epoch": 1.6764705882352942, "grad_norm": 3.8892786502838135, "learning_rate": 6.986021447504387e-08, "loss": 0.1495, "step": 85500 }, { "epoch": 1.6766666666666667, "grad_norm": 2.7944507598876953, "learning_rate": 6.97775923532245e-08, "loss": 0.1589, "step": 85510 }, { "epoch": 1.6768627450980391, "grad_norm": 4.144799709320068, "learning_rate": 6.96950154529311e-08, "loss": 0.1639, "step": 85520 }, { "epoch": 1.6770588235294117, "grad_norm": 1.714669108390808, "learning_rate": 6.961248378284367e-08, "loss": 0.1633, "step": 85530 }, { "epoch": 1.6772549019607843, "grad_norm": 12.522821426391602, "learning_rate": 6.952999735163717e-08, "loss": 0.1578, "step": 85540 }, { "epoch": 1.677450980392157, "grad_norm": 3.5324041843414307, "learning_rate": 6.944755616798175e-08, "loss": 0.1443, "step": 85550 }, { "epoch": 1.6776470588235295, "grad_norm": 4.116796493530273, "learning_rate": 6.936516024054318e-08, "loss": 0.2219, "step": 85560 }, { "epoch": 1.6778431372549019, "grad_norm": 2.048912286758423, "learning_rate": 6.928280957798227e-08, "loss": 0.1583, "step": 85570 }, { "epoch": 1.6780392156862745, "grad_norm": 3.2676050662994385, "learning_rate": 6.920050418895501e-08, "loss": 0.1733, "step": 85580 }, { "epoch": 1.678235294117647, "grad_norm": 5.309493541717529, "learning_rate": 6.911824408211253e-08, "loss": 0.1444, "step": 85590 }, { "epoch": 1.6784313725490196, "grad_norm": 3.2296180725097656, "learning_rate": 6.903602926610158e-08, "loss": 0.1295, "step": 85600 }, { "epoch": 1.6786274509803922, "grad_norm": 3.073791980743408, "learning_rate": 6.895385974956386e-08, "loss": 0.149, "step": 85610 }, { "epoch": 1.6788235294117646, "grad_norm": 2.335235834121704, "learning_rate": 6.887173554113624e-08, "loss": 0.1962, "step": 85620 }, { "epoch": 1.6790196078431374, "grad_norm": 3.932365655899048, "learning_rate": 6.878965664945108e-08, "loss": 0.1587, "step": 85630 }, { "epoch": 1.6792156862745098, "grad_norm": 2.4316117763519287, "learning_rate": 6.870762308313588e-08, "loss": 0.1494, "step": 85640 }, { "epoch": 1.6794117647058824, "grad_norm": 3.025506019592285, "learning_rate": 6.862563485081318e-08, "loss": 0.1692, "step": 85650 }, { "epoch": 1.679607843137255, "grad_norm": 3.7482192516326904, "learning_rate": 6.854369196110115e-08, "loss": 0.1631, "step": 85660 }, { "epoch": 1.6798039215686273, "grad_norm": 2.2527015209198, "learning_rate": 6.84617944226129e-08, "loss": 0.153, "step": 85670 }, { "epoch": 1.6800000000000002, "grad_norm": 2.5173819065093994, "learning_rate": 6.837994224395665e-08, "loss": 0.1433, "step": 85680 }, { "epoch": 1.6801960784313725, "grad_norm": 3.6212985515594482, "learning_rate": 6.82981354337363e-08, "loss": 0.1571, "step": 85690 }, { "epoch": 1.6803921568627451, "grad_norm": 4.163003444671631, "learning_rate": 6.821637400055064e-08, "loss": 0.1396, "step": 85700 }, { "epoch": 1.6805882352941177, "grad_norm": 1.9504295587539673, "learning_rate": 6.813465795299361e-08, "loss": 0.142, "step": 85710 }, { "epoch": 1.68078431372549, "grad_norm": 4.343625068664551, "learning_rate": 6.805298729965486e-08, "loss": 0.1323, "step": 85720 }, { "epoch": 1.680980392156863, "grad_norm": 3.2256476879119873, "learning_rate": 6.797136204911874e-08, "loss": 0.1376, "step": 85730 }, { "epoch": 1.6811764705882353, "grad_norm": 4.468620300292969, "learning_rate": 6.7889782209965e-08, "loss": 0.1354, "step": 85740 }, { "epoch": 1.6813725490196079, "grad_norm": 11.072484970092773, "learning_rate": 6.780824779076871e-08, "loss": 0.184, "step": 85750 }, { "epoch": 1.6815686274509805, "grad_norm": 3.6795942783355713, "learning_rate": 6.772675880010031e-08, "loss": 0.1358, "step": 85760 }, { "epoch": 1.6817647058823528, "grad_norm": 3.745248556137085, "learning_rate": 6.764531524652506e-08, "loss": 0.1713, "step": 85770 }, { "epoch": 1.6819607843137256, "grad_norm": 1.8605940341949463, "learning_rate": 6.756391713860365e-08, "loss": 0.1603, "step": 85780 }, { "epoch": 1.682156862745098, "grad_norm": 4.469517707824707, "learning_rate": 6.748256448489209e-08, "loss": 0.139, "step": 85790 }, { "epoch": 1.6823529411764706, "grad_norm": 3.666275978088379, "learning_rate": 6.740125729394148e-08, "loss": 0.1384, "step": 85800 }, { "epoch": 1.6825490196078432, "grad_norm": 7.030455589294434, "learning_rate": 6.731999557429807e-08, "loss": 0.1974, "step": 85810 }, { "epoch": 1.6827450980392156, "grad_norm": 3.441981315612793, "learning_rate": 6.723877933450366e-08, "loss": 0.1482, "step": 85820 }, { "epoch": 1.6829411764705884, "grad_norm": 5.459099292755127, "learning_rate": 6.715760858309493e-08, "loss": 0.1728, "step": 85830 }, { "epoch": 1.6831372549019608, "grad_norm": 3.3663558959960938, "learning_rate": 6.707648332860383e-08, "loss": 0.1424, "step": 85840 }, { "epoch": 1.6833333333333333, "grad_norm": 4.0517377853393555, "learning_rate": 6.699540357955768e-08, "loss": 0.1446, "step": 85850 }, { "epoch": 1.683529411764706, "grad_norm": 1.5113979578018188, "learning_rate": 6.691436934447897e-08, "loss": 0.166, "step": 85860 }, { "epoch": 1.6837254901960783, "grad_norm": 1.4699238538742065, "learning_rate": 6.68333806318852e-08, "loss": 0.1472, "step": 85870 }, { "epoch": 1.6839215686274511, "grad_norm": 3.2687504291534424, "learning_rate": 6.675243745028946e-08, "loss": 0.1394, "step": 85880 }, { "epoch": 1.6841176470588235, "grad_norm": 4.309233665466309, "learning_rate": 6.667153980819978e-08, "loss": 0.1301, "step": 85890 }, { "epoch": 1.684313725490196, "grad_norm": 5.719812393188477, "learning_rate": 6.659068771411935e-08, "loss": 0.1587, "step": 85900 }, { "epoch": 1.6845098039215687, "grad_norm": 5.304452896118164, "learning_rate": 6.650988117654688e-08, "loss": 0.173, "step": 85910 }, { "epoch": 1.684705882352941, "grad_norm": 2.6753952503204346, "learning_rate": 6.64291202039759e-08, "loss": 0.1574, "step": 85920 }, { "epoch": 1.6849019607843139, "grad_norm": 1.949116587638855, "learning_rate": 6.634840480489556e-08, "loss": 0.1715, "step": 85930 }, { "epoch": 1.6850980392156862, "grad_norm": 2.2455050945281982, "learning_rate": 6.626773498778987e-08, "loss": 0.1743, "step": 85940 }, { "epoch": 1.6852941176470588, "grad_norm": 1.8049718141555786, "learning_rate": 6.618711076113831e-08, "loss": 0.176, "step": 85950 }, { "epoch": 1.6854901960784314, "grad_norm": 4.575697898864746, "learning_rate": 6.610653213341539e-08, "loss": 0.1566, "step": 85960 }, { "epoch": 1.6856862745098038, "grad_norm": 1.8573737144470215, "learning_rate": 6.602599911309082e-08, "loss": 0.1487, "step": 85970 }, { "epoch": 1.6858823529411766, "grad_norm": 3.3659842014312744, "learning_rate": 6.594551170862972e-08, "loss": 0.1657, "step": 85980 }, { "epoch": 1.686078431372549, "grad_norm": 4.43519401550293, "learning_rate": 6.586506992849222e-08, "loss": 0.1512, "step": 85990 }, { "epoch": 1.6862745098039216, "grad_norm": 1.5318859815597534, "learning_rate": 6.578467378113362e-08, "loss": 0.1205, "step": 86000 }, { "epoch": 1.6864705882352942, "grad_norm": 4.618016719818115, "learning_rate": 6.570432327500469e-08, "loss": 0.1389, "step": 86010 }, { "epoch": 1.6866666666666665, "grad_norm": 1.5206937789916992, "learning_rate": 6.562401841855114e-08, "loss": 0.1368, "step": 86020 }, { "epoch": 1.6868627450980394, "grad_norm": 1.9883068799972534, "learning_rate": 6.554375922021388e-08, "loss": 0.1605, "step": 86030 }, { "epoch": 1.6870588235294117, "grad_norm": 1.4669994115829468, "learning_rate": 6.546354568842932e-08, "loss": 0.1518, "step": 86040 }, { "epoch": 1.6872549019607843, "grad_norm": 1.8371739387512207, "learning_rate": 6.538337783162873e-08, "loss": 0.1467, "step": 86050 }, { "epoch": 1.687450980392157, "grad_norm": 6.613672256469727, "learning_rate": 6.530325565823863e-08, "loss": 0.1278, "step": 86060 }, { "epoch": 1.6876470588235293, "grad_norm": 4.294214248657227, "learning_rate": 6.522317917668102e-08, "loss": 0.1556, "step": 86070 }, { "epoch": 1.687843137254902, "grad_norm": 4.289769172668457, "learning_rate": 6.514314839537282e-08, "loss": 0.1458, "step": 86080 }, { "epoch": 1.6880392156862745, "grad_norm": 3.654547691345215, "learning_rate": 6.50631633227261e-08, "loss": 0.1848, "step": 86090 }, { "epoch": 1.688235294117647, "grad_norm": 3.366018056869507, "learning_rate": 6.498322396714834e-08, "loss": 0.1746, "step": 86100 }, { "epoch": 1.6884313725490196, "grad_norm": 4.141428470611572, "learning_rate": 6.490333033704226e-08, "loss": 0.1366, "step": 86110 }, { "epoch": 1.688627450980392, "grad_norm": 3.116755247116089, "learning_rate": 6.482348244080549e-08, "loss": 0.1287, "step": 86120 }, { "epoch": 1.6888235294117648, "grad_norm": 3.28308367729187, "learning_rate": 6.474368028683092e-08, "loss": 0.1655, "step": 86130 }, { "epoch": 1.6890196078431372, "grad_norm": 3.758060932159424, "learning_rate": 6.466392388350694e-08, "loss": 0.1858, "step": 86140 }, { "epoch": 1.6892156862745098, "grad_norm": 2.070408344268799, "learning_rate": 6.458421323921675e-08, "loss": 0.1415, "step": 86150 }, { "epoch": 1.6894117647058824, "grad_norm": 4.741847991943359, "learning_rate": 6.450454836233881e-08, "loss": 0.1421, "step": 86160 }, { "epoch": 1.6896078431372548, "grad_norm": 4.1563920974731445, "learning_rate": 6.442492926124709e-08, "loss": 0.1658, "step": 86170 }, { "epoch": 1.6898039215686276, "grad_norm": 2.3083176612854004, "learning_rate": 6.43453559443104e-08, "loss": 0.1759, "step": 86180 }, { "epoch": 1.69, "grad_norm": 2.501211643218994, "learning_rate": 6.42658284198927e-08, "loss": 0.1557, "step": 86190 }, { "epoch": 1.6901960784313725, "grad_norm": 4.059484481811523, "learning_rate": 6.418634669635353e-08, "loss": 0.1733, "step": 86200 }, { "epoch": 1.6903921568627451, "grad_norm": 1.8528660535812378, "learning_rate": 6.410691078204721e-08, "loss": 0.1456, "step": 86210 }, { "epoch": 1.6905882352941175, "grad_norm": 3.2806761264801025, "learning_rate": 6.402752068532341e-08, "loss": 0.147, "step": 86220 }, { "epoch": 1.6907843137254903, "grad_norm": 2.0624167919158936, "learning_rate": 6.39481764145271e-08, "loss": 0.1359, "step": 86230 }, { "epoch": 1.6909803921568627, "grad_norm": 2.4936106204986572, "learning_rate": 6.386887797799822e-08, "loss": 0.1349, "step": 86240 }, { "epoch": 1.6911764705882353, "grad_norm": 5.147985458374023, "learning_rate": 6.378962538407196e-08, "loss": 0.149, "step": 86250 }, { "epoch": 1.6913725490196079, "grad_norm": 1.9978753328323364, "learning_rate": 6.371041864107879e-08, "loss": 0.1512, "step": 86260 }, { "epoch": 1.6915686274509802, "grad_norm": 2.8112895488739014, "learning_rate": 6.363125775734418e-08, "loss": 0.1424, "step": 86270 }, { "epoch": 1.691764705882353, "grad_norm": 2.7643322944641113, "learning_rate": 6.355214274118908e-08, "loss": 0.183, "step": 86280 }, { "epoch": 1.6919607843137254, "grad_norm": 2.626595973968506, "learning_rate": 6.347307360092918e-08, "loss": 0.1396, "step": 86290 }, { "epoch": 1.692156862745098, "grad_norm": 3.6138038635253906, "learning_rate": 6.339405034487583e-08, "loss": 0.1477, "step": 86300 }, { "epoch": 1.6923529411764706, "grad_norm": 2.8556222915649414, "learning_rate": 6.33150729813352e-08, "loss": 0.1437, "step": 86310 }, { "epoch": 1.692549019607843, "grad_norm": 3.750307559967041, "learning_rate": 6.323614151860867e-08, "loss": 0.1618, "step": 86320 }, { "epoch": 1.6927450980392158, "grad_norm": 5.702624320983887, "learning_rate": 6.315725596499304e-08, "loss": 0.1685, "step": 86330 }, { "epoch": 1.6929411764705882, "grad_norm": 1.935115933418274, "learning_rate": 6.307841632878013e-08, "loss": 0.1581, "step": 86340 }, { "epoch": 1.6931372549019608, "grad_norm": 2.3703725337982178, "learning_rate": 6.299962261825676e-08, "loss": 0.1585, "step": 86350 }, { "epoch": 1.6933333333333334, "grad_norm": 6.199554920196533, "learning_rate": 6.292087484170527e-08, "loss": 0.1582, "step": 86360 }, { "epoch": 1.6935294117647057, "grad_norm": 3.9627602100372314, "learning_rate": 6.284217300740292e-08, "loss": 0.1416, "step": 86370 }, { "epoch": 1.6937254901960785, "grad_norm": 3.795642137527466, "learning_rate": 6.27635171236221e-08, "loss": 0.1543, "step": 86380 }, { "epoch": 1.693921568627451, "grad_norm": 4.7159104347229, "learning_rate": 6.268490719863074e-08, "loss": 0.1682, "step": 86390 }, { "epoch": 1.6941176470588235, "grad_norm": 1.7262064218521118, "learning_rate": 6.260634324069153e-08, "loss": 0.1658, "step": 86400 }, { "epoch": 1.694313725490196, "grad_norm": 1.5974992513656616, "learning_rate": 6.252782525806238e-08, "loss": 0.1622, "step": 86410 }, { "epoch": 1.6945098039215685, "grad_norm": 1.8875383138656616, "learning_rate": 6.24493532589967e-08, "loss": 0.1773, "step": 86420 }, { "epoch": 1.6947058823529413, "grad_norm": 3.203242540359497, "learning_rate": 6.23709272517427e-08, "loss": 0.1804, "step": 86430 }, { "epoch": 1.6949019607843137, "grad_norm": 2.5047004222869873, "learning_rate": 6.229254724454386e-08, "loss": 0.1585, "step": 86440 }, { "epoch": 1.6950980392156862, "grad_norm": 5.7715959548950195, "learning_rate": 6.221421324563891e-08, "loss": 0.1654, "step": 86450 }, { "epoch": 1.6952941176470588, "grad_norm": 1.6892155408859253, "learning_rate": 6.213592526326178e-08, "loss": 0.1341, "step": 86460 }, { "epoch": 1.6954901960784312, "grad_norm": 2.382819414138794, "learning_rate": 6.205768330564137e-08, "loss": 0.1354, "step": 86470 }, { "epoch": 1.695686274509804, "grad_norm": 2.4587063789367676, "learning_rate": 6.197948738100178e-08, "loss": 0.1469, "step": 86480 }, { "epoch": 1.6958823529411764, "grad_norm": 2.2240712642669678, "learning_rate": 6.190133749756254e-08, "loss": 0.15, "step": 86490 }, { "epoch": 1.696078431372549, "grad_norm": 1.5010521411895752, "learning_rate": 6.182323366353797e-08, "loss": 0.1488, "step": 86500 }, { "epoch": 1.6962745098039216, "grad_norm": 4.467418193817139, "learning_rate": 6.174517588713773e-08, "loss": 0.1374, "step": 86510 }, { "epoch": 1.696470588235294, "grad_norm": 5.285540580749512, "learning_rate": 6.166716417656675e-08, "loss": 0.1738, "step": 86520 }, { "epoch": 1.6966666666666668, "grad_norm": 2.576927900314331, "learning_rate": 6.158919854002493e-08, "loss": 0.1596, "step": 86530 }, { "epoch": 1.6968627450980391, "grad_norm": 3.5936567783355713, "learning_rate": 6.151127898570724e-08, "loss": 0.1209, "step": 86540 }, { "epoch": 1.6970588235294117, "grad_norm": 3.437441825866699, "learning_rate": 6.143340552180421e-08, "loss": 0.1697, "step": 86550 }, { "epoch": 1.6972549019607843, "grad_norm": 3.8322763442993164, "learning_rate": 6.13555781565011e-08, "loss": 0.1793, "step": 86560 }, { "epoch": 1.697450980392157, "grad_norm": 2.2004804611206055, "learning_rate": 6.127779689797851e-08, "loss": 0.1459, "step": 86570 }, { "epoch": 1.6976470588235295, "grad_norm": 2.248389959335327, "learning_rate": 6.120006175441233e-08, "loss": 0.1612, "step": 86580 }, { "epoch": 1.6978431372549019, "grad_norm": 2.9927542209625244, "learning_rate": 6.11223727339733e-08, "loss": 0.157, "step": 86590 }, { "epoch": 1.6980392156862745, "grad_norm": 2.61301851272583, "learning_rate": 6.104472984482744e-08, "loss": 0.1488, "step": 86600 }, { "epoch": 1.698235294117647, "grad_norm": 3.4693679809570312, "learning_rate": 6.096713309513618e-08, "loss": 0.1583, "step": 86610 }, { "epoch": 1.6984313725490197, "grad_norm": 2.3878426551818848, "learning_rate": 6.088958249305565e-08, "loss": 0.1312, "step": 86620 }, { "epoch": 1.6986274509803923, "grad_norm": 2.9984655380249023, "learning_rate": 6.081207804673733e-08, "loss": 0.1728, "step": 86630 }, { "epoch": 1.6988235294117646, "grad_norm": 4.114070892333984, "learning_rate": 6.073461976432792e-08, "loss": 0.1565, "step": 86640 }, { "epoch": 1.6990196078431372, "grad_norm": 2.3991878032684326, "learning_rate": 6.06572076539693e-08, "loss": 0.1498, "step": 86650 }, { "epoch": 1.6992156862745098, "grad_norm": 2.4810309410095215, "learning_rate": 6.057984172379842e-08, "loss": 0.1765, "step": 86660 }, { "epoch": 1.6994117647058824, "grad_norm": 2.4725894927978516, "learning_rate": 6.050252198194717e-08, "loss": 0.1417, "step": 86670 }, { "epoch": 1.699607843137255, "grad_norm": 3.5095462799072266, "learning_rate": 6.042524843654295e-08, "loss": 0.1461, "step": 86680 }, { "epoch": 1.6998039215686274, "grad_norm": 2.7913053035736084, "learning_rate": 6.034802109570813e-08, "loss": 0.1563, "step": 86690 }, { "epoch": 1.7, "grad_norm": 3.127902030944824, "learning_rate": 6.027083996756011e-08, "loss": 0.1641, "step": 86700 }, { "epoch": 1.7001960784313725, "grad_norm": 5.864881992340088, "learning_rate": 6.019370506021165e-08, "loss": 0.1478, "step": 86710 }, { "epoch": 1.7003921568627451, "grad_norm": 3.343505859375, "learning_rate": 6.011661638177057e-08, "loss": 0.1526, "step": 86720 }, { "epoch": 1.7005882352941177, "grad_norm": 10.503191947937012, "learning_rate": 6.00395739403397e-08, "loss": 0.1524, "step": 86730 }, { "epoch": 1.70078431372549, "grad_norm": 4.5355048179626465, "learning_rate": 5.996257774401726e-08, "loss": 0.1569, "step": 86740 }, { "epoch": 1.7009803921568627, "grad_norm": 2.435079336166382, "learning_rate": 5.98856278008964e-08, "loss": 0.1682, "step": 86750 }, { "epoch": 1.7011764705882353, "grad_norm": 1.9073008298873901, "learning_rate": 5.980872411906535e-08, "loss": 0.1412, "step": 86760 }, { "epoch": 1.7013725490196079, "grad_norm": 4.041545867919922, "learning_rate": 5.97318667066079e-08, "loss": 0.1453, "step": 86770 }, { "epoch": 1.7015686274509805, "grad_norm": 4.9481000900268555, "learning_rate": 5.965505557160249e-08, "loss": 0.1641, "step": 86780 }, { "epoch": 1.7017647058823528, "grad_norm": 8.4861478805542, "learning_rate": 5.957829072212284e-08, "loss": 0.1519, "step": 86790 }, { "epoch": 1.7019607843137254, "grad_norm": 3.5525434017181396, "learning_rate": 5.950157216623808e-08, "loss": 0.1699, "step": 86800 }, { "epoch": 1.702156862745098, "grad_norm": 2.4999730587005615, "learning_rate": 5.942489991201194e-08, "loss": 0.1638, "step": 86810 }, { "epoch": 1.7023529411764706, "grad_norm": 3.0025131702423096, "learning_rate": 5.9348273967503916e-08, "loss": 0.162, "step": 86820 }, { "epoch": 1.7025490196078432, "grad_norm": 1.5101652145385742, "learning_rate": 5.927169434076806e-08, "loss": 0.1697, "step": 86830 }, { "epoch": 1.7027450980392156, "grad_norm": 2.5858840942382812, "learning_rate": 5.919516103985395e-08, "loss": 0.1447, "step": 86840 }, { "epoch": 1.7029411764705882, "grad_norm": 2.303257465362549, "learning_rate": 5.9118674072806184e-08, "loss": 0.1664, "step": 86850 }, { "epoch": 1.7031372549019608, "grad_norm": 1.482443928718567, "learning_rate": 5.9042233447664237e-08, "loss": 0.1609, "step": 86860 }, { "epoch": 1.7033333333333334, "grad_norm": 3.6886658668518066, "learning_rate": 5.896583917246317e-08, "loss": 0.1433, "step": 86870 }, { "epoch": 1.703529411764706, "grad_norm": 3.202353000640869, "learning_rate": 5.888949125523285e-08, "loss": 0.1579, "step": 86880 }, { "epoch": 1.7037254901960783, "grad_norm": 7.114976406097412, "learning_rate": 5.881318970399829e-08, "loss": 0.1431, "step": 86890 }, { "epoch": 1.7039215686274511, "grad_norm": 2.2792768478393555, "learning_rate": 5.87369345267798e-08, "loss": 0.1621, "step": 86900 }, { "epoch": 1.7041176470588235, "grad_norm": 2.929497003555298, "learning_rate": 5.86607257315927e-08, "loss": 0.149, "step": 86910 }, { "epoch": 1.704313725490196, "grad_norm": 1.657928466796875, "learning_rate": 5.8584563326447346e-08, "loss": 0.1808, "step": 86920 }, { "epoch": 1.7045098039215687, "grad_norm": 3.9306247234344482, "learning_rate": 5.850844731934945e-08, "loss": 0.1831, "step": 86930 }, { "epoch": 1.704705882352941, "grad_norm": 3.637009620666504, "learning_rate": 5.8432377718299677e-08, "loss": 0.1366, "step": 86940 }, { "epoch": 1.7049019607843139, "grad_norm": 3.8683278560638428, "learning_rate": 5.835635453129373e-08, "loss": 0.1828, "step": 86950 }, { "epoch": 1.7050980392156863, "grad_norm": 2.939152956008911, "learning_rate": 5.828037776632277e-08, "loss": 0.1411, "step": 86960 }, { "epoch": 1.7052941176470588, "grad_norm": 1.7030110359191895, "learning_rate": 5.820444743137271e-08, "loss": 0.1666, "step": 86970 }, { "epoch": 1.7054901960784314, "grad_norm": 3.4245736598968506, "learning_rate": 5.812856353442469e-08, "loss": 0.1359, "step": 86980 }, { "epoch": 1.7056862745098038, "grad_norm": 2.375988483428955, "learning_rate": 5.805272608345513e-08, "loss": 0.1845, "step": 86990 }, { "epoch": 1.7058823529411766, "grad_norm": 3.421820878982544, "learning_rate": 5.797693508643553e-08, "loss": 0.1499, "step": 87000 }, { "epoch": 1.706078431372549, "grad_norm": 2.73046612739563, "learning_rate": 5.790119055133236e-08, "loss": 0.1509, "step": 87010 }, { "epoch": 1.7062745098039216, "grad_norm": 3.1248586177825928, "learning_rate": 5.7825492486107075e-08, "loss": 0.1331, "step": 87020 }, { "epoch": 1.7064705882352942, "grad_norm": 3.54040265083313, "learning_rate": 5.774984089871676e-08, "loss": 0.1635, "step": 87030 }, { "epoch": 1.7066666666666666, "grad_norm": 4.310431003570557, "learning_rate": 5.767423579711317e-08, "loss": 0.164, "step": 87040 }, { "epoch": 1.7068627450980394, "grad_norm": 3.4311363697052, "learning_rate": 5.759867718924322e-08, "loss": 0.1626, "step": 87050 }, { "epoch": 1.7070588235294117, "grad_norm": 4.306209564208984, "learning_rate": 5.7523165083049166e-08, "loss": 0.1539, "step": 87060 }, { "epoch": 1.7072549019607843, "grad_norm": 2.393920660018921, "learning_rate": 5.744769948646821e-08, "loss": 0.1624, "step": 87070 }, { "epoch": 1.707450980392157, "grad_norm": 2.025158166885376, "learning_rate": 5.7372280407432525e-08, "loss": 0.1525, "step": 87080 }, { "epoch": 1.7076470588235293, "grad_norm": 4.102408409118652, "learning_rate": 5.7296907853869814e-08, "loss": 0.1457, "step": 87090 }, { "epoch": 1.707843137254902, "grad_norm": 3.920598030090332, "learning_rate": 5.722158183370251e-08, "loss": 0.1521, "step": 87100 }, { "epoch": 1.7080392156862745, "grad_norm": 6.495209217071533, "learning_rate": 5.714630235484813e-08, "loss": 0.185, "step": 87110 }, { "epoch": 1.708235294117647, "grad_norm": 1.6513246297836304, "learning_rate": 5.707106942521972e-08, "loss": 0.1625, "step": 87120 }, { "epoch": 1.7084313725490197, "grad_norm": 4.227635383605957, "learning_rate": 5.699588305272507e-08, "loss": 0.1435, "step": 87130 }, { "epoch": 1.708627450980392, "grad_norm": 2.5918405055999756, "learning_rate": 5.6920743245267024e-08, "loss": 0.1436, "step": 87140 }, { "epoch": 1.7088235294117649, "grad_norm": 2.7767434120178223, "learning_rate": 5.684565001074387e-08, "loss": 0.1396, "step": 87150 }, { "epoch": 1.7090196078431372, "grad_norm": 2.333869218826294, "learning_rate": 5.6770603357048695e-08, "loss": 0.1763, "step": 87160 }, { "epoch": 1.7092156862745098, "grad_norm": 3.545480489730835, "learning_rate": 5.6695603292069794e-08, "loss": 0.1611, "step": 87170 }, { "epoch": 1.7094117647058824, "grad_norm": 4.582231521606445, "learning_rate": 5.662064982369058e-08, "loss": 0.159, "step": 87180 }, { "epoch": 1.7096078431372548, "grad_norm": 1.5335817337036133, "learning_rate": 5.65457429597897e-08, "loss": 0.1543, "step": 87190 }, { "epoch": 1.7098039215686276, "grad_norm": 3.8980629444122314, "learning_rate": 5.647088270824063e-08, "loss": 0.1458, "step": 87200 }, { "epoch": 1.71, "grad_norm": 2.285064935684204, "learning_rate": 5.6396069076911964e-08, "loss": 0.151, "step": 87210 }, { "epoch": 1.7101960784313726, "grad_norm": 3.0326290130615234, "learning_rate": 5.632130207366781e-08, "loss": 0.1593, "step": 87220 }, { "epoch": 1.7103921568627451, "grad_norm": 1.8516477346420288, "learning_rate": 5.624658170636687e-08, "loss": 0.1721, "step": 87230 }, { "epoch": 1.7105882352941175, "grad_norm": 2.8234293460845947, "learning_rate": 5.6171907982863146e-08, "loss": 0.1607, "step": 87240 }, { "epoch": 1.7107843137254903, "grad_norm": 8.332621574401855, "learning_rate": 5.60972809110058e-08, "loss": 0.1383, "step": 87250 }, { "epoch": 1.7109803921568627, "grad_norm": 1.9347103834152222, "learning_rate": 5.602270049863905e-08, "loss": 0.1455, "step": 87260 }, { "epoch": 1.7111764705882353, "grad_norm": 2.371662139892578, "learning_rate": 5.594816675360209e-08, "loss": 0.1341, "step": 87270 }, { "epoch": 1.711372549019608, "grad_norm": 2.4535932540893555, "learning_rate": 5.587367968372941e-08, "loss": 0.1682, "step": 87280 }, { "epoch": 1.7115686274509803, "grad_norm": 2.575770139694214, "learning_rate": 5.579923929685049e-08, "loss": 0.1762, "step": 87290 }, { "epoch": 1.711764705882353, "grad_norm": 3.292515754699707, "learning_rate": 5.5724845600789774e-08, "loss": 0.1309, "step": 87300 }, { "epoch": 1.7119607843137254, "grad_norm": 8.05392837524414, "learning_rate": 5.5650498603367124e-08, "loss": 0.1792, "step": 87310 }, { "epoch": 1.712156862745098, "grad_norm": 1.6379774808883667, "learning_rate": 5.5576198312397237e-08, "loss": 0.1681, "step": 87320 }, { "epoch": 1.7123529411764706, "grad_norm": 2.7672669887542725, "learning_rate": 5.550194473568981e-08, "loss": 0.1527, "step": 87330 }, { "epoch": 1.712549019607843, "grad_norm": 3.3887321949005127, "learning_rate": 5.5427737881049986e-08, "loss": 0.1406, "step": 87340 }, { "epoch": 1.7127450980392158, "grad_norm": 4.180313587188721, "learning_rate": 5.5353577756277637e-08, "loss": 0.1364, "step": 87350 }, { "epoch": 1.7129411764705882, "grad_norm": 4.256516456604004, "learning_rate": 5.5279464369168025e-08, "loss": 0.1581, "step": 87360 }, { "epoch": 1.7131372549019608, "grad_norm": 2.571380376815796, "learning_rate": 5.520539772751115e-08, "loss": 0.164, "step": 87370 }, { "epoch": 1.7133333333333334, "grad_norm": 3.196058988571167, "learning_rate": 5.5131377839092606e-08, "loss": 0.1588, "step": 87380 }, { "epoch": 1.7135294117647057, "grad_norm": 3.619016408920288, "learning_rate": 5.5057404711692565e-08, "loss": 0.1519, "step": 87390 }, { "epoch": 1.7137254901960786, "grad_norm": 4.184840202331543, "learning_rate": 5.498347835308642e-08, "loss": 0.1724, "step": 87400 }, { "epoch": 1.713921568627451, "grad_norm": 6.230337619781494, "learning_rate": 5.490959877104495e-08, "loss": 0.1745, "step": 87410 }, { "epoch": 1.7141176470588235, "grad_norm": 2.104922294616699, "learning_rate": 5.4835765973333616e-08, "loss": 0.1699, "step": 87420 }, { "epoch": 1.7143137254901961, "grad_norm": 1.5053819417953491, "learning_rate": 5.476197996771309e-08, "loss": 0.1576, "step": 87430 }, { "epoch": 1.7145098039215685, "grad_norm": 2.175311326980591, "learning_rate": 5.468824076193934e-08, "loss": 0.1225, "step": 87440 }, { "epoch": 1.7147058823529413, "grad_norm": 2.6770708560943604, "learning_rate": 5.461454836376317e-08, "loss": 0.1314, "step": 87450 }, { "epoch": 1.7149019607843137, "grad_norm": 7.693430423736572, "learning_rate": 5.454090278093038e-08, "loss": 0.1491, "step": 87460 }, { "epoch": 1.7150980392156863, "grad_norm": 2.293501615524292, "learning_rate": 5.4467304021182214e-08, "loss": 0.1451, "step": 87470 }, { "epoch": 1.7152941176470589, "grad_norm": 4.551978588104248, "learning_rate": 5.4393752092254664e-08, "loss": 0.1729, "step": 87480 }, { "epoch": 1.7154901960784312, "grad_norm": 9.04611873626709, "learning_rate": 5.432024700187887e-08, "loss": 0.1808, "step": 87490 }, { "epoch": 1.715686274509804, "grad_norm": 5.138297080993652, "learning_rate": 5.4246788757781325e-08, "loss": 0.1386, "step": 87500 }, { "epoch": 1.7158823529411764, "grad_norm": 2.4710047245025635, "learning_rate": 5.417337736768318e-08, "loss": 0.1383, "step": 87510 }, { "epoch": 1.716078431372549, "grad_norm": 2.6334733963012695, "learning_rate": 5.4100012839300766e-08, "loss": 0.1332, "step": 87520 }, { "epoch": 1.7162745098039216, "grad_norm": 2.6291840076446533, "learning_rate": 5.402669518034575e-08, "loss": 0.1455, "step": 87530 }, { "epoch": 1.716470588235294, "grad_norm": 3.4288482666015625, "learning_rate": 5.3953424398524736e-08, "loss": 0.1576, "step": 87540 }, { "epoch": 1.7166666666666668, "grad_norm": 2.5887820720672607, "learning_rate": 5.388020050153924e-08, "loss": 0.1352, "step": 87550 }, { "epoch": 1.7168627450980392, "grad_norm": 1.9171994924545288, "learning_rate": 5.3807023497085935e-08, "loss": 0.1531, "step": 87560 }, { "epoch": 1.7170588235294117, "grad_norm": 5.49196195602417, "learning_rate": 5.373389339285678e-08, "loss": 0.1667, "step": 87570 }, { "epoch": 1.7172549019607843, "grad_norm": 2.4029133319854736, "learning_rate": 5.366081019653845e-08, "loss": 0.1365, "step": 87580 }, { "epoch": 1.7174509803921567, "grad_norm": 2.0698442459106445, "learning_rate": 5.358777391581287e-08, "loss": 0.1745, "step": 87590 }, { "epoch": 1.7176470588235295, "grad_norm": 5.885772705078125, "learning_rate": 5.3514784558357174e-08, "loss": 0.1705, "step": 87600 }, { "epoch": 1.717843137254902, "grad_norm": 2.7556138038635254, "learning_rate": 5.344184213184327e-08, "loss": 0.157, "step": 87610 }, { "epoch": 1.7180392156862745, "grad_norm": 2.72415828704834, "learning_rate": 5.336894664393832e-08, "loss": 0.1413, "step": 87620 }, { "epoch": 1.718235294117647, "grad_norm": 1.5568289756774902, "learning_rate": 5.329609810230457e-08, "loss": 0.1622, "step": 87630 }, { "epoch": 1.7184313725490195, "grad_norm": 0.8687751889228821, "learning_rate": 5.322329651459928e-08, "loss": 0.1273, "step": 87640 }, { "epoch": 1.7186274509803923, "grad_norm": 2.2961156368255615, "learning_rate": 5.3150541888474563e-08, "loss": 0.1617, "step": 87650 }, { "epoch": 1.7188235294117646, "grad_norm": 3.16353178024292, "learning_rate": 5.307783423157808e-08, "loss": 0.1483, "step": 87660 }, { "epoch": 1.7190196078431372, "grad_norm": 5.998159408569336, "learning_rate": 5.300517355155215e-08, "loss": 0.108, "step": 87670 }, { "epoch": 1.7192156862745098, "grad_norm": 14.260655403137207, "learning_rate": 5.293255985603423e-08, "loss": 0.1507, "step": 87680 }, { "epoch": 1.7194117647058822, "grad_norm": 5.497095108032227, "learning_rate": 5.285999315265699e-08, "loss": 0.1554, "step": 87690 }, { "epoch": 1.719607843137255, "grad_norm": 2.4596362113952637, "learning_rate": 5.2787473449047936e-08, "loss": 0.1239, "step": 87700 }, { "epoch": 1.7198039215686274, "grad_norm": 2.3221116065979004, "learning_rate": 5.271500075282997e-08, "loss": 0.1684, "step": 87710 }, { "epoch": 1.72, "grad_norm": 2.9253318309783936, "learning_rate": 5.2642575071620564e-08, "loss": 0.1727, "step": 87720 }, { "epoch": 1.7201960784313726, "grad_norm": 3.4399797916412354, "learning_rate": 5.257019641303273e-08, "loss": 0.148, "step": 87730 }, { "epoch": 1.720392156862745, "grad_norm": 1.8955638408660889, "learning_rate": 5.249786478467433e-08, "loss": 0.1625, "step": 87740 }, { "epoch": 1.7205882352941178, "grad_norm": 3.783743143081665, "learning_rate": 5.242558019414811e-08, "loss": 0.1979, "step": 87750 }, { "epoch": 1.7207843137254901, "grad_norm": 3.880866527557373, "learning_rate": 5.2353342649052226e-08, "loss": 0.1345, "step": 87760 }, { "epoch": 1.7209803921568627, "grad_norm": 2.9080920219421387, "learning_rate": 5.228115215697965e-08, "loss": 0.1284, "step": 87770 }, { "epoch": 1.7211764705882353, "grad_norm": 3.6873533725738525, "learning_rate": 5.2209008725518365e-08, "loss": 0.1724, "step": 87780 }, { "epoch": 1.7213725490196077, "grad_norm": 2.3055524826049805, "learning_rate": 5.21369123622517e-08, "loss": 0.154, "step": 87790 }, { "epoch": 1.7215686274509805, "grad_norm": 1.6736721992492676, "learning_rate": 5.2064863074757714e-08, "loss": 0.1653, "step": 87800 }, { "epoch": 1.7217647058823529, "grad_norm": 4.337638854980469, "learning_rate": 5.199286087060961e-08, "loss": 0.1475, "step": 87810 }, { "epoch": 1.7219607843137255, "grad_norm": 2.798610210418701, "learning_rate": 5.19209057573759e-08, "loss": 0.1139, "step": 87820 }, { "epoch": 1.722156862745098, "grad_norm": 2.363173246383667, "learning_rate": 5.184899774261969e-08, "loss": 0.1268, "step": 87830 }, { "epoch": 1.7223529411764706, "grad_norm": 3.54876971244812, "learning_rate": 5.177713683389945e-08, "loss": 0.1521, "step": 87840 }, { "epoch": 1.7225490196078432, "grad_norm": 6.633046627044678, "learning_rate": 5.170532303876868e-08, "loss": 0.164, "step": 87850 }, { "epoch": 1.7227450980392156, "grad_norm": 3.2167155742645264, "learning_rate": 5.1633556364775846e-08, "loss": 0.1279, "step": 87860 }, { "epoch": 1.7229411764705882, "grad_norm": 1.8337936401367188, "learning_rate": 5.156183681946436e-08, "loss": 0.1411, "step": 87870 }, { "epoch": 1.7231372549019608, "grad_norm": 1.2564119100570679, "learning_rate": 5.149016441037296e-08, "loss": 0.1307, "step": 87880 }, { "epoch": 1.7233333333333334, "grad_norm": 3.8320260047912598, "learning_rate": 5.141853914503524e-08, "loss": 0.1488, "step": 87890 }, { "epoch": 1.723529411764706, "grad_norm": 4.5245513916015625, "learning_rate": 5.1346961030979906e-08, "loss": 0.1706, "step": 87900 }, { "epoch": 1.7237254901960783, "grad_norm": 3.7010250091552734, "learning_rate": 5.127543007573054e-08, "loss": 0.1713, "step": 87910 }, { "epoch": 1.723921568627451, "grad_norm": 2.494797945022583, "learning_rate": 5.120394628680608e-08, "loss": 0.1637, "step": 87920 }, { "epoch": 1.7241176470588235, "grad_norm": 5.824625015258789, "learning_rate": 5.1132509671720226e-08, "loss": 0.1258, "step": 87930 }, { "epoch": 1.7243137254901961, "grad_norm": 2.2423269748687744, "learning_rate": 5.1061120237981703e-08, "loss": 0.179, "step": 87940 }, { "epoch": 1.7245098039215687, "grad_norm": 1.2539669275283813, "learning_rate": 5.0989777993094674e-08, "loss": 0.1489, "step": 87950 }, { "epoch": 1.724705882352941, "grad_norm": 3.4341042041778564, "learning_rate": 5.091848294455792e-08, "loss": 0.1676, "step": 87960 }, { "epoch": 1.7249019607843137, "grad_norm": 2.987725019454956, "learning_rate": 5.0847235099865215e-08, "loss": 0.1635, "step": 87970 }, { "epoch": 1.7250980392156863, "grad_norm": 2.912872552871704, "learning_rate": 5.077603446650591e-08, "loss": 0.1481, "step": 87980 }, { "epoch": 1.7252941176470589, "grad_norm": 3.811583995819092, "learning_rate": 5.070488105196385e-08, "loss": 0.13, "step": 87990 }, { "epoch": 1.7254901960784315, "grad_norm": 3.979534864425659, "learning_rate": 5.0633774863718056e-08, "loss": 0.1975, "step": 88000 }, { "epoch": 1.7256862745098038, "grad_norm": 2.1373450756073, "learning_rate": 5.0562715909242825e-08, "loss": 0.1331, "step": 88010 }, { "epoch": 1.7258823529411764, "grad_norm": 3.576913595199585, "learning_rate": 5.0491704196007234e-08, "loss": 0.1397, "step": 88020 }, { "epoch": 1.726078431372549, "grad_norm": 2.589442253112793, "learning_rate": 5.042073973147526e-08, "loss": 0.1436, "step": 88030 }, { "epoch": 1.7262745098039216, "grad_norm": 4.658156394958496, "learning_rate": 5.034982252310643e-08, "loss": 0.1701, "step": 88040 }, { "epoch": 1.7264705882352942, "grad_norm": 2.1909453868865967, "learning_rate": 5.027895257835485e-08, "loss": 0.1946, "step": 88050 }, { "epoch": 1.7266666666666666, "grad_norm": 2.573780059814453, "learning_rate": 5.0208129904669715e-08, "loss": 0.1214, "step": 88060 }, { "epoch": 1.7268627450980392, "grad_norm": 3.1040756702423096, "learning_rate": 5.0137354509495355e-08, "loss": 0.1548, "step": 88070 }, { "epoch": 1.7270588235294118, "grad_norm": 2.689335346221924, "learning_rate": 5.0066626400271316e-08, "loss": 0.1795, "step": 88080 }, { "epoch": 1.7272549019607844, "grad_norm": 2.513221502304077, "learning_rate": 4.9995945584431874e-08, "loss": 0.15, "step": 88090 }, { "epoch": 1.727450980392157, "grad_norm": 4.799485206604004, "learning_rate": 4.9925312069406256e-08, "loss": 0.1433, "step": 88100 }, { "epoch": 1.7276470588235293, "grad_norm": 3.023890256881714, "learning_rate": 4.9854725862619075e-08, "loss": 0.1409, "step": 88110 }, { "epoch": 1.727843137254902, "grad_norm": 5.564763069152832, "learning_rate": 4.9784186971489785e-08, "loss": 0.133, "step": 88120 }, { "epoch": 1.7280392156862745, "grad_norm": 2.127617120742798, "learning_rate": 4.971369540343268e-08, "loss": 0.1321, "step": 88130 }, { "epoch": 1.728235294117647, "grad_norm": 2.3750505447387695, "learning_rate": 4.964325116585749e-08, "loss": 0.1404, "step": 88140 }, { "epoch": 1.7284313725490197, "grad_norm": 4.479709148406982, "learning_rate": 4.95728542661687e-08, "loss": 0.1388, "step": 88150 }, { "epoch": 1.728627450980392, "grad_norm": 4.324331283569336, "learning_rate": 4.95025047117657e-08, "loss": 0.1768, "step": 88160 }, { "epoch": 1.7288235294117649, "grad_norm": 3.699504852294922, "learning_rate": 4.943220251004332e-08, "loss": 0.142, "step": 88170 }, { "epoch": 1.7290196078431372, "grad_norm": 4.435176849365234, "learning_rate": 4.936194766839103e-08, "loss": 0.1737, "step": 88180 }, { "epoch": 1.7292156862745098, "grad_norm": 6.184327125549316, "learning_rate": 4.929174019419341e-08, "loss": 0.1647, "step": 88190 }, { "epoch": 1.7294117647058824, "grad_norm": 4.107227325439453, "learning_rate": 4.922158009483018e-08, "loss": 0.1438, "step": 88200 }, { "epoch": 1.7296078431372548, "grad_norm": 2.967963457107544, "learning_rate": 4.915146737767606e-08, "loss": 0.1608, "step": 88210 }, { "epoch": 1.7298039215686276, "grad_norm": 2.827653408050537, "learning_rate": 4.908140205010053e-08, "loss": 0.1546, "step": 88220 }, { "epoch": 1.73, "grad_norm": 2.6012802124023438, "learning_rate": 4.901138411946848e-08, "loss": 0.1379, "step": 88230 }, { "epoch": 1.7301960784313726, "grad_norm": 3.108269214630127, "learning_rate": 4.894141359313958e-08, "loss": 0.1743, "step": 88240 }, { "epoch": 1.7303921568627452, "grad_norm": 3.654479742050171, "learning_rate": 4.8871490478468615e-08, "loss": 0.1907, "step": 88250 }, { "epoch": 1.7305882352941175, "grad_norm": 3.297222852706909, "learning_rate": 4.8801614782805204e-08, "loss": 0.1634, "step": 88260 }, { "epoch": 1.7307843137254904, "grad_norm": 3.04890513420105, "learning_rate": 4.87317865134943e-08, "loss": 0.1523, "step": 88270 }, { "epoch": 1.7309803921568627, "grad_norm": 1.410651445388794, "learning_rate": 4.86620056778756e-08, "loss": 0.1485, "step": 88280 }, { "epoch": 1.7311764705882353, "grad_norm": 1.4587992429733276, "learning_rate": 4.859227228328383e-08, "loss": 0.1495, "step": 88290 }, { "epoch": 1.731372549019608, "grad_norm": 2.599762439727783, "learning_rate": 4.852258633704898e-08, "loss": 0.153, "step": 88300 }, { "epoch": 1.7315686274509803, "grad_norm": 3.0039772987365723, "learning_rate": 4.8452947846495786e-08, "loss": 0.1398, "step": 88310 }, { "epoch": 1.731764705882353, "grad_norm": 2.5515124797821045, "learning_rate": 4.8383356818943957e-08, "loss": 0.1329, "step": 88320 }, { "epoch": 1.7319607843137255, "grad_norm": 2.5047130584716797, "learning_rate": 4.831381326170858e-08, "loss": 0.1488, "step": 88330 }, { "epoch": 1.732156862745098, "grad_norm": 3.7566075325012207, "learning_rate": 4.824431718209943e-08, "loss": 0.1784, "step": 88340 }, { "epoch": 1.7323529411764707, "grad_norm": 4.231027126312256, "learning_rate": 4.817486858742126e-08, "loss": 0.1261, "step": 88350 }, { "epoch": 1.732549019607843, "grad_norm": 3.294365167617798, "learning_rate": 4.8105467484974074e-08, "loss": 0.1495, "step": 88360 }, { "epoch": 1.7327450980392158, "grad_norm": 3.5758752822875977, "learning_rate": 4.803611388205275e-08, "loss": 0.1671, "step": 88370 }, { "epoch": 1.7329411764705882, "grad_norm": 4.559340953826904, "learning_rate": 4.7966807785947085e-08, "loss": 0.1514, "step": 88380 }, { "epoch": 1.7331372549019608, "grad_norm": 2.72615122795105, "learning_rate": 4.789754920394212e-08, "loss": 0.1609, "step": 88390 }, { "epoch": 1.7333333333333334, "grad_norm": 3.590738296508789, "learning_rate": 4.7828338143317715e-08, "loss": 0.1563, "step": 88400 }, { "epoch": 1.7335294117647058, "grad_norm": 2.3426804542541504, "learning_rate": 4.775917461134865e-08, "loss": 0.1634, "step": 88410 }, { "epoch": 1.7337254901960786, "grad_norm": 3.2052764892578125, "learning_rate": 4.7690058615304953e-08, "loss": 0.1698, "step": 88420 }, { "epoch": 1.733921568627451, "grad_norm": 4.06210470199585, "learning_rate": 4.762099016245164e-08, "loss": 0.149, "step": 88430 }, { "epoch": 1.7341176470588235, "grad_norm": 4.002612113952637, "learning_rate": 4.7551969260048506e-08, "loss": 0.1761, "step": 88440 }, { "epoch": 1.7343137254901961, "grad_norm": 2.763173818588257, "learning_rate": 4.7482995915350433e-08, "loss": 0.1611, "step": 88450 }, { "epoch": 1.7345098039215685, "grad_norm": 4.906870365142822, "learning_rate": 4.74140701356075e-08, "loss": 0.1606, "step": 88460 }, { "epoch": 1.7347058823529413, "grad_norm": 4.8577141761779785, "learning_rate": 4.734519192806452e-08, "loss": 0.1408, "step": 88470 }, { "epoch": 1.7349019607843137, "grad_norm": 3.554136037826538, "learning_rate": 4.727636129996138e-08, "loss": 0.17, "step": 88480 }, { "epoch": 1.7350980392156863, "grad_norm": 2.0951836109161377, "learning_rate": 4.720757825853311e-08, "loss": 0.1641, "step": 88490 }, { "epoch": 1.7352941176470589, "grad_norm": 4.709864139556885, "learning_rate": 4.713884281100961e-08, "loss": 0.1678, "step": 88500 }, { "epoch": 1.7354901960784312, "grad_norm": 3.1579928398132324, "learning_rate": 4.707015496461569e-08, "loss": 0.1786, "step": 88510 }, { "epoch": 1.735686274509804, "grad_norm": 1.6972800493240356, "learning_rate": 4.700151472657143e-08, "loss": 0.1416, "step": 88520 }, { "epoch": 1.7358823529411764, "grad_norm": 1.897004246711731, "learning_rate": 4.693292210409166e-08, "loss": 0.143, "step": 88530 }, { "epoch": 1.736078431372549, "grad_norm": 4.263427257537842, "learning_rate": 4.6864377104386176e-08, "loss": 0.1498, "step": 88540 }, { "epoch": 1.7362745098039216, "grad_norm": 3.465970993041992, "learning_rate": 4.679587973466009e-08, "loss": 0.136, "step": 88550 }, { "epoch": 1.736470588235294, "grad_norm": 3.0995428562164307, "learning_rate": 4.672743000211321e-08, "loss": 0.1422, "step": 88560 }, { "epoch": 1.7366666666666668, "grad_norm": 4.040424346923828, "learning_rate": 4.665902791394027e-08, "loss": 0.1708, "step": 88570 }, { "epoch": 1.7368627450980392, "grad_norm": 2.0610086917877197, "learning_rate": 4.659067347733142e-08, "loss": 0.1518, "step": 88580 }, { "epoch": 1.7370588235294118, "grad_norm": 2.698354959487915, "learning_rate": 4.6522366699471406e-08, "loss": 0.2118, "step": 88590 }, { "epoch": 1.7372549019607844, "grad_norm": 3.4318530559539795, "learning_rate": 4.645410758753993e-08, "loss": 0.1637, "step": 88600 }, { "epoch": 1.7374509803921567, "grad_norm": 2.6116621494293213, "learning_rate": 4.638589614871202e-08, "loss": 0.1312, "step": 88610 }, { "epoch": 1.7376470588235295, "grad_norm": 3.1113815307617188, "learning_rate": 4.631773239015757e-08, "loss": 0.1574, "step": 88620 }, { "epoch": 1.737843137254902, "grad_norm": 1.536422610282898, "learning_rate": 4.624961631904128e-08, "loss": 0.1532, "step": 88630 }, { "epoch": 1.7380392156862745, "grad_norm": 5.125415325164795, "learning_rate": 4.6181547942522936e-08, "loss": 0.1444, "step": 88640 }, { "epoch": 1.738235294117647, "grad_norm": 3.594794273376465, "learning_rate": 4.611352726775752e-08, "loss": 0.1584, "step": 88650 }, { "epoch": 1.7384313725490195, "grad_norm": 3.3892266750335693, "learning_rate": 4.6045554301894716e-08, "loss": 0.1373, "step": 88660 }, { "epoch": 1.7386274509803923, "grad_norm": 3.2960617542266846, "learning_rate": 4.597762905207914e-08, "loss": 0.144, "step": 88670 }, { "epoch": 1.7388235294117647, "grad_norm": 3.512291669845581, "learning_rate": 4.590975152545085e-08, "loss": 0.144, "step": 88680 }, { "epoch": 1.7390196078431372, "grad_norm": 2.934527635574341, "learning_rate": 4.5841921729144416e-08, "loss": 0.1812, "step": 88690 }, { "epoch": 1.7392156862745098, "grad_norm": 12.23814868927002, "learning_rate": 4.577413967028948e-08, "loss": 0.1615, "step": 88700 }, { "epoch": 1.7394117647058822, "grad_norm": 3.139305830001831, "learning_rate": 4.570640535601089e-08, "loss": 0.1707, "step": 88710 }, { "epoch": 1.739607843137255, "grad_norm": 3.137928009033203, "learning_rate": 4.563871879342835e-08, "loss": 0.1555, "step": 88720 }, { "epoch": 1.7398039215686274, "grad_norm": 4.451283931732178, "learning_rate": 4.557107998965637e-08, "loss": 0.1606, "step": 88730 }, { "epoch": 1.74, "grad_norm": 3.495777130126953, "learning_rate": 4.550348895180478e-08, "loss": 0.1779, "step": 88740 }, { "epoch": 1.7401960784313726, "grad_norm": 2.5854744911193848, "learning_rate": 4.5435945686978114e-08, "loss": 0.143, "step": 88750 }, { "epoch": 1.740392156862745, "grad_norm": 2.4094338417053223, "learning_rate": 4.5368450202275866e-08, "loss": 0.152, "step": 88760 }, { "epoch": 1.7405882352941178, "grad_norm": 2.445521831512451, "learning_rate": 4.530100250479291e-08, "loss": 0.152, "step": 88770 }, { "epoch": 1.7407843137254901, "grad_norm": 9.82052993774414, "learning_rate": 4.523360260161846e-08, "loss": 0.1671, "step": 88780 }, { "epoch": 1.7409803921568627, "grad_norm": 2.3731191158294678, "learning_rate": 4.5166250499837354e-08, "loss": 0.1729, "step": 88790 }, { "epoch": 1.7411764705882353, "grad_norm": 3.4974493980407715, "learning_rate": 4.509894620652893e-08, "loss": 0.1755, "step": 88800 }, { "epoch": 1.7413725490196077, "grad_norm": 3.1818296909332275, "learning_rate": 4.503168972876775e-08, "loss": 0.1591, "step": 88810 }, { "epoch": 1.7415686274509805, "grad_norm": 3.042832851409912, "learning_rate": 4.4964481073623326e-08, "loss": 0.147, "step": 88820 }, { "epoch": 1.7417647058823529, "grad_norm": 5.042718887329102, "learning_rate": 4.48973202481599e-08, "loss": 0.1497, "step": 88830 }, { "epoch": 1.7419607843137255, "grad_norm": 2.6624391078948975, "learning_rate": 4.4830207259437035e-08, "loss": 0.1678, "step": 88840 }, { "epoch": 1.742156862745098, "grad_norm": 3.2024929523468018, "learning_rate": 4.47631421145091e-08, "loss": 0.1441, "step": 88850 }, { "epoch": 1.7423529411764704, "grad_norm": 2.688992500305176, "learning_rate": 4.469612482042534e-08, "loss": 0.1682, "step": 88860 }, { "epoch": 1.7425490196078433, "grad_norm": 2.9331564903259277, "learning_rate": 4.4629155384230234e-08, "loss": 0.186, "step": 88870 }, { "epoch": 1.7427450980392156, "grad_norm": 5.217560291290283, "learning_rate": 4.4562233812962976e-08, "loss": 0.1791, "step": 88880 }, { "epoch": 1.7429411764705882, "grad_norm": 3.7787511348724365, "learning_rate": 4.4495360113657784e-08, "loss": 0.1532, "step": 88890 }, { "epoch": 1.7431372549019608, "grad_norm": 1.2491652965545654, "learning_rate": 4.442853429334398e-08, "loss": 0.1357, "step": 88900 }, { "epoch": 1.7433333333333332, "grad_norm": 2.9711380004882812, "learning_rate": 4.436175635904571e-08, "loss": 0.1542, "step": 88910 }, { "epoch": 1.743529411764706, "grad_norm": 4.088237762451172, "learning_rate": 4.4295026317782094e-08, "loss": 0.1295, "step": 88920 }, { "epoch": 1.7437254901960784, "grad_norm": 1.9621955156326294, "learning_rate": 4.422834417656735e-08, "loss": 0.1878, "step": 88930 }, { "epoch": 1.743921568627451, "grad_norm": 1.6785863637924194, "learning_rate": 4.416170994241053e-08, "loss": 0.1272, "step": 88940 }, { "epoch": 1.7441176470588236, "grad_norm": 5.556687355041504, "learning_rate": 4.40951236223156e-08, "loss": 0.1235, "step": 88950 }, { "epoch": 1.744313725490196, "grad_norm": 3.159242630004883, "learning_rate": 4.402858522328168e-08, "loss": 0.1908, "step": 88960 }, { "epoch": 1.7445098039215687, "grad_norm": 3.7792441844940186, "learning_rate": 4.396209475230278e-08, "loss": 0.1774, "step": 88970 }, { "epoch": 1.744705882352941, "grad_norm": 4.360814571380615, "learning_rate": 4.3895652216367876e-08, "loss": 0.1683, "step": 88980 }, { "epoch": 1.7449019607843137, "grad_norm": 3.2598819732666016, "learning_rate": 4.3829257622460645e-08, "loss": 0.1542, "step": 88990 }, { "epoch": 1.7450980392156863, "grad_norm": 1.9609800577163696, "learning_rate": 4.3762910977560186e-08, "loss": 0.185, "step": 89000 }, { "epoch": 1.7452941176470587, "grad_norm": 6.618499279022217, "learning_rate": 4.369661228864025e-08, "loss": 0.1507, "step": 89010 }, { "epoch": 1.7454901960784315, "grad_norm": 3.8271660804748535, "learning_rate": 4.363036156266958e-08, "loss": 0.1392, "step": 89020 }, { "epoch": 1.7456862745098038, "grad_norm": 2.338658094406128, "learning_rate": 4.356415880661196e-08, "loss": 0.1496, "step": 89030 }, { "epoch": 1.7458823529411764, "grad_norm": 1.644309163093567, "learning_rate": 4.349800402742615e-08, "loss": 0.1402, "step": 89040 }, { "epoch": 1.746078431372549, "grad_norm": 4.7960968017578125, "learning_rate": 4.34318972320657e-08, "loss": 0.1736, "step": 89050 }, { "epoch": 1.7462745098039214, "grad_norm": 3.6141977310180664, "learning_rate": 4.336583842747932e-08, "loss": 0.186, "step": 89060 }, { "epoch": 1.7464705882352942, "grad_norm": 1.6324443817138672, "learning_rate": 4.329982762061052e-08, "loss": 0.1311, "step": 89070 }, { "epoch": 1.7466666666666666, "grad_norm": 5.2328267097473145, "learning_rate": 4.3233864818397746e-08, "loss": 0.1649, "step": 89080 }, { "epoch": 1.7468627450980392, "grad_norm": 4.157143592834473, "learning_rate": 4.316795002777468e-08, "loss": 0.1708, "step": 89090 }, { "epoch": 1.7470588235294118, "grad_norm": 6.66314697265625, "learning_rate": 4.3102083255669663e-08, "loss": 0.1597, "step": 89100 }, { "epoch": 1.7472549019607844, "grad_norm": 4.868808746337891, "learning_rate": 4.3036264509005994e-08, "loss": 0.1844, "step": 89110 }, { "epoch": 1.747450980392157, "grad_norm": 2.2543160915374756, "learning_rate": 4.2970493794702135e-08, "loss": 0.162, "step": 89120 }, { "epoch": 1.7476470588235293, "grad_norm": 2.6000170707702637, "learning_rate": 4.290477111967128e-08, "loss": 0.1137, "step": 89130 }, { "epoch": 1.747843137254902, "grad_norm": 2.260303020477295, "learning_rate": 4.283909649082179e-08, "loss": 0.151, "step": 89140 }, { "epoch": 1.7480392156862745, "grad_norm": 3.9368062019348145, "learning_rate": 4.2773469915056703e-08, "loss": 0.1801, "step": 89150 }, { "epoch": 1.748235294117647, "grad_norm": 2.9523448944091797, "learning_rate": 4.2707891399274274e-08, "loss": 0.151, "step": 89160 }, { "epoch": 1.7484313725490197, "grad_norm": 2.049017906188965, "learning_rate": 4.2642360950367606e-08, "loss": 0.1793, "step": 89170 }, { "epoch": 1.748627450980392, "grad_norm": 2.995009422302246, "learning_rate": 4.257687857522457e-08, "loss": 0.1438, "step": 89180 }, { "epoch": 1.7488235294117647, "grad_norm": 3.5070366859436035, "learning_rate": 4.2511444280728336e-08, "loss": 0.168, "step": 89190 }, { "epoch": 1.7490196078431373, "grad_norm": 1.3448339700698853, "learning_rate": 4.2446058073756784e-08, "loss": 0.163, "step": 89200 }, { "epoch": 1.7492156862745099, "grad_norm": 2.418210506439209, "learning_rate": 4.238071996118264e-08, "loss": 0.1535, "step": 89210 }, { "epoch": 1.7494117647058824, "grad_norm": 1.6053050756454468, "learning_rate": 4.231542994987397e-08, "loss": 0.1542, "step": 89220 }, { "epoch": 1.7496078431372548, "grad_norm": 1.109533429145813, "learning_rate": 4.225018804669334e-08, "loss": 0.1579, "step": 89230 }, { "epoch": 1.7498039215686274, "grad_norm": 4.999990940093994, "learning_rate": 4.2184994258498476e-08, "loss": 0.1564, "step": 89240 }, { "epoch": 1.75, "grad_norm": 3.120966672897339, "learning_rate": 4.2119848592142126e-08, "loss": 0.1623, "step": 89250 }, { "epoch": 1.7501960784313726, "grad_norm": 2.501770496368408, "learning_rate": 4.2054751054471873e-08, "loss": 0.1604, "step": 89260 }, { "epoch": 1.7503921568627452, "grad_norm": 5.1664652824401855, "learning_rate": 4.1989701652330064e-08, "loss": 0.1575, "step": 89270 }, { "epoch": 1.7505882352941176, "grad_norm": 2.0505852699279785, "learning_rate": 4.192470039255441e-08, "loss": 0.1597, "step": 89280 }, { "epoch": 1.7507843137254901, "grad_norm": 3.186802387237549, "learning_rate": 4.185974728197722e-08, "loss": 0.161, "step": 89290 }, { "epoch": 1.7509803921568627, "grad_norm": 1.9112207889556885, "learning_rate": 4.179484232742575e-08, "loss": 0.1455, "step": 89300 }, { "epoch": 1.7511764705882353, "grad_norm": 6.239572525024414, "learning_rate": 4.172998553572238e-08, "loss": 0.1723, "step": 89310 }, { "epoch": 1.751372549019608, "grad_norm": 1.273717999458313, "learning_rate": 4.1665176913684374e-08, "loss": 0.1532, "step": 89320 }, { "epoch": 1.7515686274509803, "grad_norm": 3.066347360610962, "learning_rate": 4.16004164681239e-08, "loss": 0.1338, "step": 89330 }, { "epoch": 1.751764705882353, "grad_norm": 2.554032564163208, "learning_rate": 4.15357042058479e-08, "loss": 0.145, "step": 89340 }, { "epoch": 1.7519607843137255, "grad_norm": 3.673780679702759, "learning_rate": 4.147104013365865e-08, "loss": 0.1572, "step": 89350 }, { "epoch": 1.752156862745098, "grad_norm": 1.9966932535171509, "learning_rate": 4.1406424258352947e-08, "loss": 0.1575, "step": 89360 }, { "epoch": 1.7523529411764707, "grad_norm": 2.1317903995513916, "learning_rate": 4.1341856586722624e-08, "loss": 0.1718, "step": 89370 }, { "epoch": 1.752549019607843, "grad_norm": 4.400059223175049, "learning_rate": 4.12773371255547e-08, "loss": 0.1456, "step": 89380 }, { "epoch": 1.7527450980392156, "grad_norm": 3.5288681983947754, "learning_rate": 4.121286588163092e-08, "loss": 0.1652, "step": 89390 }, { "epoch": 1.7529411764705882, "grad_norm": 4.036856174468994, "learning_rate": 4.1148442861727805e-08, "loss": 0.1792, "step": 89400 }, { "epoch": 1.7531372549019608, "grad_norm": 2.984321117401123, "learning_rate": 4.1084068072617214e-08, "loss": 0.1749, "step": 89410 }, { "epoch": 1.7533333333333334, "grad_norm": 2.214045286178589, "learning_rate": 4.101974152106558e-08, "loss": 0.1514, "step": 89420 }, { "epoch": 1.7535294117647058, "grad_norm": 2.4145922660827637, "learning_rate": 4.0955463213834365e-08, "loss": 0.1499, "step": 89430 }, { "epoch": 1.7537254901960786, "grad_norm": 3.293874502182007, "learning_rate": 4.0891233157680117e-08, "loss": 0.1452, "step": 89440 }, { "epoch": 1.753921568627451, "grad_norm": 1.8079872131347656, "learning_rate": 4.082705135935405e-08, "loss": 0.1422, "step": 89450 }, { "epoch": 1.7541176470588236, "grad_norm": 1.9279390573501587, "learning_rate": 4.076291782560248e-08, "loss": 0.1346, "step": 89460 }, { "epoch": 1.7543137254901962, "grad_norm": 1.797584056854248, "learning_rate": 4.0698832563166695e-08, "loss": 0.1565, "step": 89470 }, { "epoch": 1.7545098039215685, "grad_norm": 2.1454427242279053, "learning_rate": 4.0634795578782745e-08, "loss": 0.1949, "step": 89480 }, { "epoch": 1.7547058823529413, "grad_norm": 2.211793899536133, "learning_rate": 4.057080687918163e-08, "loss": 0.1473, "step": 89490 }, { "epoch": 1.7549019607843137, "grad_norm": 4.521178722381592, "learning_rate": 4.050686647108936e-08, "loss": 0.1448, "step": 89500 }, { "epoch": 1.7550980392156863, "grad_norm": 4.8486199378967285, "learning_rate": 4.044297436122701e-08, "loss": 0.1587, "step": 89510 }, { "epoch": 1.755294117647059, "grad_norm": 5.273141384124756, "learning_rate": 4.0379130556310205e-08, "loss": 0.1788, "step": 89520 }, { "epoch": 1.7554901960784313, "grad_norm": 5.280218601226807, "learning_rate": 4.031533506304968e-08, "loss": 0.1433, "step": 89530 }, { "epoch": 1.755686274509804, "grad_norm": 1.927119493484497, "learning_rate": 4.025158788815131e-08, "loss": 0.1559, "step": 89540 }, { "epoch": 1.7558823529411764, "grad_norm": 2.9933037757873535, "learning_rate": 4.018788903831549e-08, "loss": 0.1695, "step": 89550 }, { "epoch": 1.756078431372549, "grad_norm": 7.203047752380371, "learning_rate": 4.0124238520237775e-08, "loss": 0.1929, "step": 89560 }, { "epoch": 1.7562745098039216, "grad_norm": 3.2043397426605225, "learning_rate": 4.006063634060863e-08, "loss": 0.1671, "step": 89570 }, { "epoch": 1.756470588235294, "grad_norm": 3.917602062225342, "learning_rate": 3.9997082506113446e-08, "loss": 0.1723, "step": 89580 }, { "epoch": 1.7566666666666668, "grad_norm": 6.005856990814209, "learning_rate": 3.9933577023432365e-08, "loss": 0.1607, "step": 89590 }, { "epoch": 1.7568627450980392, "grad_norm": 2.044699192047119, "learning_rate": 3.987011989924072e-08, "loss": 0.1646, "step": 89600 }, { "epoch": 1.7570588235294118, "grad_norm": 3.6356875896453857, "learning_rate": 3.98067111402085e-08, "loss": 0.1757, "step": 89610 }, { "epoch": 1.7572549019607844, "grad_norm": 2.9455082416534424, "learning_rate": 3.974335075300073e-08, "loss": 0.182, "step": 89620 }, { "epoch": 1.7574509803921567, "grad_norm": 3.4787261486053467, "learning_rate": 3.96800387442775e-08, "loss": 0.1803, "step": 89630 }, { "epoch": 1.7576470588235296, "grad_norm": 1.9724187850952148, "learning_rate": 3.96167751206935e-08, "loss": 0.1826, "step": 89640 }, { "epoch": 1.757843137254902, "grad_norm": 5.137234210968018, "learning_rate": 3.955355988889841e-08, "loss": 0.134, "step": 89650 }, { "epoch": 1.7580392156862745, "grad_norm": 5.1223273277282715, "learning_rate": 3.94903930555372e-08, "loss": 0.173, "step": 89660 }, { "epoch": 1.7582352941176471, "grad_norm": 2.9715194702148438, "learning_rate": 3.9427274627249154e-08, "loss": 0.1507, "step": 89670 }, { "epoch": 1.7584313725490195, "grad_norm": 4.446134567260742, "learning_rate": 3.9364204610669025e-08, "loss": 0.149, "step": 89680 }, { "epoch": 1.7586274509803923, "grad_norm": 3.5301499366760254, "learning_rate": 3.930118301242602e-08, "loss": 0.1358, "step": 89690 }, { "epoch": 1.7588235294117647, "grad_norm": 4.050116062164307, "learning_rate": 3.923820983914466e-08, "loss": 0.1675, "step": 89700 }, { "epoch": 1.7590196078431373, "grad_norm": 1.6335031986236572, "learning_rate": 3.917528509744411e-08, "loss": 0.142, "step": 89710 }, { "epoch": 1.7592156862745099, "grad_norm": 2.608938694000244, "learning_rate": 3.9112408793938345e-08, "loss": 0.1888, "step": 89720 }, { "epoch": 1.7594117647058822, "grad_norm": 3.09283447265625, "learning_rate": 3.904958093523669e-08, "loss": 0.18, "step": 89730 }, { "epoch": 1.759607843137255, "grad_norm": 3.6766245365142822, "learning_rate": 3.898680152794293e-08, "loss": 0.1546, "step": 89740 }, { "epoch": 1.7598039215686274, "grad_norm": 10.011228561401367, "learning_rate": 3.892407057865593e-08, "loss": 0.1609, "step": 89750 }, { "epoch": 1.76, "grad_norm": 2.7248454093933105, "learning_rate": 3.886138809396955e-08, "loss": 0.1462, "step": 89760 }, { "epoch": 1.7601960784313726, "grad_norm": 5.4654154777526855, "learning_rate": 3.8798754080472454e-08, "loss": 0.1546, "step": 89770 }, { "epoch": 1.760392156862745, "grad_norm": 3.6403415203094482, "learning_rate": 3.873616854474815e-08, "loss": 0.1518, "step": 89780 }, { "epoch": 1.7605882352941178, "grad_norm": 1.9286788702011108, "learning_rate": 3.8673631493375216e-08, "loss": 0.1473, "step": 89790 }, { "epoch": 1.7607843137254902, "grad_norm": 1.628743290901184, "learning_rate": 3.8611142932927e-08, "loss": 0.145, "step": 89800 }, { "epoch": 1.7609803921568628, "grad_norm": 2.3837873935699463, "learning_rate": 3.854870286997175e-08, "loss": 0.1711, "step": 89810 }, { "epoch": 1.7611764705882353, "grad_norm": 3.2112205028533936, "learning_rate": 3.848631131107277e-08, "loss": 0.1607, "step": 89820 }, { "epoch": 1.7613725490196077, "grad_norm": 3.0977468490600586, "learning_rate": 3.842396826278815e-08, "loss": 0.1441, "step": 89830 }, { "epoch": 1.7615686274509805, "grad_norm": 4.415966987609863, "learning_rate": 3.836167373167071e-08, "loss": 0.2051, "step": 89840 }, { "epoch": 1.761764705882353, "grad_norm": 4.366090774536133, "learning_rate": 3.829942772426853e-08, "loss": 0.1537, "step": 89850 }, { "epoch": 1.7619607843137255, "grad_norm": 2.0449001789093018, "learning_rate": 3.8237230247124444e-08, "loss": 0.1607, "step": 89860 }, { "epoch": 1.762156862745098, "grad_norm": 1.7748358249664307, "learning_rate": 3.81750813067761e-08, "loss": 0.1628, "step": 89870 }, { "epoch": 1.7623529411764705, "grad_norm": 3.6981616020202637, "learning_rate": 3.811298090975601e-08, "loss": 0.1674, "step": 89880 }, { "epoch": 1.7625490196078433, "grad_norm": 1.5931743383407593, "learning_rate": 3.805092906259177e-08, "loss": 0.1492, "step": 89890 }, { "epoch": 1.7627450980392156, "grad_norm": 3.5254263877868652, "learning_rate": 3.798892577180574e-08, "loss": 0.1406, "step": 89900 }, { "epoch": 1.7629411764705882, "grad_norm": 5.4333720207214355, "learning_rate": 3.7926971043915125e-08, "loss": 0.2008, "step": 89910 }, { "epoch": 1.7631372549019608, "grad_norm": 1.1800049543380737, "learning_rate": 3.786506488543229e-08, "loss": 0.169, "step": 89920 }, { "epoch": 1.7633333333333332, "grad_norm": 4.828118324279785, "learning_rate": 3.780320730286418e-08, "loss": 0.1896, "step": 89930 }, { "epoch": 1.763529411764706, "grad_norm": 1.9117470979690552, "learning_rate": 3.774139830271272e-08, "loss": 0.1615, "step": 89940 }, { "epoch": 1.7637254901960784, "grad_norm": 3.1659955978393555, "learning_rate": 3.767963789147499e-08, "loss": 0.1303, "step": 89950 }, { "epoch": 1.763921568627451, "grad_norm": 1.7899202108383179, "learning_rate": 3.761792607564257e-08, "loss": 0.1446, "step": 89960 }, { "epoch": 1.7641176470588236, "grad_norm": 7.99686861038208, "learning_rate": 3.75562628617021e-08, "loss": 0.1603, "step": 89970 }, { "epoch": 1.764313725490196, "grad_norm": 2.1898910999298096, "learning_rate": 3.749464825613524e-08, "loss": 0.1533, "step": 89980 }, { "epoch": 1.7645098039215688, "grad_norm": 3.7389423847198486, "learning_rate": 3.7433082265418305e-08, "loss": 0.1568, "step": 89990 }, { "epoch": 1.7647058823529411, "grad_norm": 3.022203207015991, "learning_rate": 3.7371564896022665e-08, "loss": 0.1588, "step": 90000 }, { "epoch": 1.7649019607843137, "grad_norm": 3.156451940536499, "learning_rate": 3.731009615441461e-08, "loss": 0.1461, "step": 90010 }, { "epoch": 1.7650980392156863, "grad_norm": 2.600426435470581, "learning_rate": 3.7248676047055116e-08, "loss": 0.1509, "step": 90020 }, { "epoch": 1.7652941176470587, "grad_norm": 4.740198135375977, "learning_rate": 3.71873045804002e-08, "loss": 0.1354, "step": 90030 }, { "epoch": 1.7654901960784315, "grad_norm": 2.807250499725342, "learning_rate": 3.712598176090076e-08, "loss": 0.1634, "step": 90040 }, { "epoch": 1.7656862745098039, "grad_norm": 4.735527992248535, "learning_rate": 3.706470759500263e-08, "loss": 0.1659, "step": 90050 }, { "epoch": 1.7658823529411765, "grad_norm": 3.4901273250579834, "learning_rate": 3.7003482089146385e-08, "loss": 0.1673, "step": 90060 }, { "epoch": 1.766078431372549, "grad_norm": 2.103353977203369, "learning_rate": 3.694230524976749e-08, "loss": 0.1473, "step": 90070 }, { "epoch": 1.7662745098039214, "grad_norm": 4.6770710945129395, "learning_rate": 3.688117708329652e-08, "loss": 0.1476, "step": 90080 }, { "epoch": 1.7664705882352942, "grad_norm": 2.7267918586730957, "learning_rate": 3.682009759615867e-08, "loss": 0.1653, "step": 90090 }, { "epoch": 1.7666666666666666, "grad_norm": 2.0846235752105713, "learning_rate": 3.6759066794774085e-08, "loss": 0.1602, "step": 90100 }, { "epoch": 1.7668627450980392, "grad_norm": 2.8061444759368896, "learning_rate": 3.6698084685558016e-08, "loss": 0.1458, "step": 90110 }, { "epoch": 1.7670588235294118, "grad_norm": 4.591136455535889, "learning_rate": 3.663715127492023e-08, "loss": 0.159, "step": 90120 }, { "epoch": 1.7672549019607842, "grad_norm": 2.68310546875, "learning_rate": 3.65762665692656e-08, "loss": 0.1806, "step": 90130 }, { "epoch": 1.767450980392157, "grad_norm": 2.6010072231292725, "learning_rate": 3.65154305749939e-08, "loss": 0.1286, "step": 90140 }, { "epoch": 1.7676470588235293, "grad_norm": 5.942532062530518, "learning_rate": 3.645464329849968e-08, "loss": 0.1765, "step": 90150 }, { "epoch": 1.767843137254902, "grad_norm": 3.803825855255127, "learning_rate": 3.6393904746172324e-08, "loss": 0.1602, "step": 90160 }, { "epoch": 1.7680392156862745, "grad_norm": 3.6850438117980957, "learning_rate": 3.633321492439639e-08, "loss": 0.1605, "step": 90170 }, { "epoch": 1.768235294117647, "grad_norm": 2.628077268600464, "learning_rate": 3.6272573839550956e-08, "loss": 0.1452, "step": 90180 }, { "epoch": 1.7684313725490197, "grad_norm": 2.359551191329956, "learning_rate": 3.621198149801008e-08, "loss": 0.1622, "step": 90190 }, { "epoch": 1.768627450980392, "grad_norm": 4.892025947570801, "learning_rate": 3.615143790614289e-08, "loss": 0.217, "step": 90200 }, { "epoch": 1.7688235294117647, "grad_norm": 1.9985703229904175, "learning_rate": 3.609094307031313e-08, "loss": 0.1764, "step": 90210 }, { "epoch": 1.7690196078431373, "grad_norm": 7.473964691162109, "learning_rate": 3.60304969968796e-08, "loss": 0.1656, "step": 90220 }, { "epoch": 1.7692156862745096, "grad_norm": 4.372125625610352, "learning_rate": 3.597009969219583e-08, "loss": 0.1738, "step": 90230 }, { "epoch": 1.7694117647058825, "grad_norm": 4.792016506195068, "learning_rate": 3.590975116261041e-08, "loss": 0.1681, "step": 90240 }, { "epoch": 1.7696078431372548, "grad_norm": 1.6725348234176636, "learning_rate": 3.58494514144666e-08, "loss": 0.1258, "step": 90250 }, { "epoch": 1.7698039215686274, "grad_norm": 4.085906505584717, "learning_rate": 3.578920045410261e-08, "loss": 0.1475, "step": 90260 }, { "epoch": 1.77, "grad_norm": 3.6890432834625244, "learning_rate": 3.572899828785164e-08, "loss": 0.1717, "step": 90270 }, { "epoch": 1.7701960784313724, "grad_norm": 12.038270950317383, "learning_rate": 3.566884492204164e-08, "loss": 0.1903, "step": 90280 }, { "epoch": 1.7703921568627452, "grad_norm": 2.219543218612671, "learning_rate": 3.560874036299533e-08, "loss": 0.1666, "step": 90290 }, { "epoch": 1.7705882352941176, "grad_norm": 3.683680534362793, "learning_rate": 3.5548684617030535e-08, "loss": 0.1221, "step": 90300 }, { "epoch": 1.7707843137254902, "grad_norm": 3.3675429821014404, "learning_rate": 3.548867769045982e-08, "loss": 0.1793, "step": 90310 }, { "epoch": 1.7709803921568628, "grad_norm": 4.274959087371826, "learning_rate": 3.5428719589590537e-08, "loss": 0.1652, "step": 90320 }, { "epoch": 1.7711764705882351, "grad_norm": 2.537874221801758, "learning_rate": 3.536881032072514e-08, "loss": 0.1586, "step": 90330 }, { "epoch": 1.771372549019608, "grad_norm": 3.0446765422821045, "learning_rate": 3.53089498901607e-08, "loss": 0.1684, "step": 90340 }, { "epoch": 1.7715686274509803, "grad_norm": 5.453998565673828, "learning_rate": 3.5249138304189296e-08, "loss": 0.1483, "step": 90350 }, { "epoch": 1.771764705882353, "grad_norm": 1.9484513998031616, "learning_rate": 3.51893755690979e-08, "loss": 0.1468, "step": 90360 }, { "epoch": 1.7719607843137255, "grad_norm": 6.579421043395996, "learning_rate": 3.512966169116821e-08, "loss": 0.1827, "step": 90370 }, { "epoch": 1.772156862745098, "grad_norm": 3.54195237159729, "learning_rate": 3.5069996676676875e-08, "loss": 0.1476, "step": 90380 }, { "epoch": 1.7723529411764707, "grad_norm": 3.5584568977355957, "learning_rate": 3.501038053189542e-08, "loss": 0.1538, "step": 90390 }, { "epoch": 1.772549019607843, "grad_norm": 2.8322770595550537, "learning_rate": 3.4950813263090296e-08, "loss": 0.1566, "step": 90400 }, { "epoch": 1.7727450980392156, "grad_norm": 2.717524290084839, "learning_rate": 3.489129487652265e-08, "loss": 0.1547, "step": 90410 }, { "epoch": 1.7729411764705882, "grad_norm": 3.1074740886688232, "learning_rate": 3.4831825378448535e-08, "loss": 0.1767, "step": 90420 }, { "epoch": 1.7731372549019608, "grad_norm": 3.2371068000793457, "learning_rate": 3.4772404775119054e-08, "loss": 0.1658, "step": 90430 }, { "epoch": 1.7733333333333334, "grad_norm": 5.042556285858154, "learning_rate": 3.471303307277995e-08, "loss": 0.2054, "step": 90440 }, { "epoch": 1.7735294117647058, "grad_norm": 5.057046413421631, "learning_rate": 3.4653710277671774e-08, "loss": 0.1639, "step": 90450 }, { "epoch": 1.7737254901960784, "grad_norm": 4.3223137855529785, "learning_rate": 3.4594436396030213e-08, "loss": 0.158, "step": 90460 }, { "epoch": 1.773921568627451, "grad_norm": 2.6329963207244873, "learning_rate": 3.453521143408567e-08, "loss": 0.1642, "step": 90470 }, { "epoch": 1.7741176470588236, "grad_norm": 4.434403419494629, "learning_rate": 3.4476035398063264e-08, "loss": 0.1522, "step": 90480 }, { "epoch": 1.7743137254901962, "grad_norm": 2.562260866165161, "learning_rate": 3.441690829418326e-08, "loss": 0.1185, "step": 90490 }, { "epoch": 1.7745098039215685, "grad_norm": 4.002668857574463, "learning_rate": 3.435783012866056e-08, "loss": 0.1519, "step": 90500 }, { "epoch": 1.7747058823529411, "grad_norm": 2.267773389816284, "learning_rate": 3.429880090770493e-08, "loss": 0.1941, "step": 90510 }, { "epoch": 1.7749019607843137, "grad_norm": 3.7974791526794434, "learning_rate": 3.4239820637521124e-08, "loss": 0.1785, "step": 90520 }, { "epoch": 1.7750980392156863, "grad_norm": 3.91811203956604, "learning_rate": 3.418088932430868e-08, "loss": 0.1374, "step": 90530 }, { "epoch": 1.775294117647059, "grad_norm": 3.8314688205718994, "learning_rate": 3.4122006974261877e-08, "loss": 0.1565, "step": 90540 }, { "epoch": 1.7754901960784313, "grad_norm": 1.6824779510498047, "learning_rate": 3.406317359357014e-08, "loss": 0.1436, "step": 90550 }, { "epoch": 1.7756862745098039, "grad_norm": 4.979233264923096, "learning_rate": 3.4004389188417304e-08, "loss": 0.2289, "step": 90560 }, { "epoch": 1.7758823529411765, "grad_norm": 2.903376817703247, "learning_rate": 3.3945653764982596e-08, "loss": 0.1567, "step": 90570 }, { "epoch": 1.776078431372549, "grad_norm": 3.0524396896362305, "learning_rate": 3.388696732943958e-08, "loss": 0.1519, "step": 90580 }, { "epoch": 1.7762745098039217, "grad_norm": 5.051602363586426, "learning_rate": 3.3828329887957085e-08, "loss": 0.165, "step": 90590 }, { "epoch": 1.776470588235294, "grad_norm": 2.4675450325012207, "learning_rate": 3.376974144669853e-08, "loss": 0.1361, "step": 90600 }, { "epoch": 1.7766666666666666, "grad_norm": 2.3303582668304443, "learning_rate": 3.3711202011822146e-08, "loss": 0.1546, "step": 90610 }, { "epoch": 1.7768627450980392, "grad_norm": 3.01674485206604, "learning_rate": 3.365271158948135e-08, "loss": 0.1438, "step": 90620 }, { "epoch": 1.7770588235294118, "grad_norm": 3.9697399139404297, "learning_rate": 3.3594270185824116e-08, "loss": 0.1356, "step": 90630 }, { "epoch": 1.7772549019607844, "grad_norm": 4.383273124694824, "learning_rate": 3.353587780699313e-08, "loss": 0.1629, "step": 90640 }, { "epoch": 1.7774509803921568, "grad_norm": 1.723038911819458, "learning_rate": 3.3477534459126434e-08, "loss": 0.1796, "step": 90650 }, { "epoch": 1.7776470588235294, "grad_norm": 2.79323673248291, "learning_rate": 3.341924014835645e-08, "loss": 0.1725, "step": 90660 }, { "epoch": 1.777843137254902, "grad_norm": 2.7886438369750977, "learning_rate": 3.3360994880810565e-08, "loss": 0.1641, "step": 90670 }, { "epoch": 1.7780392156862745, "grad_norm": 2.6024746894836426, "learning_rate": 3.330279866261121e-08, "loss": 0.1729, "step": 90680 }, { "epoch": 1.7782352941176471, "grad_norm": 4.022749900817871, "learning_rate": 3.324465149987538e-08, "loss": 0.1401, "step": 90690 }, { "epoch": 1.7784313725490195, "grad_norm": 2.596130847930908, "learning_rate": 3.318655339871501e-08, "loss": 0.1244, "step": 90700 }, { "epoch": 1.7786274509803923, "grad_norm": 2.1087563037872314, "learning_rate": 3.3128504365237066e-08, "loss": 0.1378, "step": 90710 }, { "epoch": 1.7788235294117647, "grad_norm": 3.1036789417266846, "learning_rate": 3.3070504405543096e-08, "loss": 0.1627, "step": 90720 }, { "epoch": 1.7790196078431373, "grad_norm": 2.9237723350524902, "learning_rate": 3.301255352572946e-08, "loss": 0.1512, "step": 90730 }, { "epoch": 1.7792156862745099, "grad_norm": 1.9524983167648315, "learning_rate": 3.295465173188766e-08, "loss": 0.1592, "step": 90740 }, { "epoch": 1.7794117647058822, "grad_norm": 5.605836391448975, "learning_rate": 3.289679903010395e-08, "loss": 0.1989, "step": 90750 }, { "epoch": 1.779607843137255, "grad_norm": 6.989656925201416, "learning_rate": 3.283899542645918e-08, "loss": 0.1794, "step": 90760 }, { "epoch": 1.7798039215686274, "grad_norm": 3.0595507621765137, "learning_rate": 3.2781240927029165e-08, "loss": 0.1863, "step": 90770 }, { "epoch": 1.78, "grad_norm": 3.9380996227264404, "learning_rate": 3.272353553788476e-08, "loss": 0.1806, "step": 90780 }, { "epoch": 1.7801960784313726, "grad_norm": 3.791478157043457, "learning_rate": 3.26658792650914e-08, "loss": 0.1369, "step": 90790 }, { "epoch": 1.780392156862745, "grad_norm": 5.398431777954102, "learning_rate": 3.260827211470935e-08, "loss": 0.1871, "step": 90800 }, { "epoch": 1.7805882352941178, "grad_norm": 2.5539450645446777, "learning_rate": 3.255071409279403e-08, "loss": 0.1542, "step": 90810 }, { "epoch": 1.7807843137254902, "grad_norm": 4.607982158660889, "learning_rate": 3.249320520539533e-08, "loss": 0.1923, "step": 90820 }, { "epoch": 1.7809803921568628, "grad_norm": 8.833721160888672, "learning_rate": 3.2435745458558096e-08, "loss": 0.1784, "step": 90830 }, { "epoch": 1.7811764705882354, "grad_norm": 5.1393351554870605, "learning_rate": 3.23783348583222e-08, "loss": 0.1492, "step": 90840 }, { "epoch": 1.7813725490196077, "grad_norm": 3.729701519012451, "learning_rate": 3.2320973410722044e-08, "loss": 0.1494, "step": 90850 }, { "epoch": 1.7815686274509805, "grad_norm": 1.784274935722351, "learning_rate": 3.2263661121786964e-08, "loss": 0.1436, "step": 90860 }, { "epoch": 1.781764705882353, "grad_norm": 5.565689563751221, "learning_rate": 3.2206397997541327e-08, "loss": 0.1428, "step": 90870 }, { "epoch": 1.7819607843137255, "grad_norm": 2.36617112159729, "learning_rate": 3.2149184044004074e-08, "loss": 0.1431, "step": 90880 }, { "epoch": 1.782156862745098, "grad_norm": 4.024158477783203, "learning_rate": 3.2092019267189016e-08, "loss": 0.1586, "step": 90890 }, { "epoch": 1.7823529411764705, "grad_norm": 4.205783843994141, "learning_rate": 3.2034903673105016e-08, "loss": 0.1496, "step": 90900 }, { "epoch": 1.7825490196078433, "grad_norm": 3.2636756896972656, "learning_rate": 3.1977837267755546e-08, "loss": 0.1544, "step": 90910 }, { "epoch": 1.7827450980392157, "grad_norm": 8.004817962646484, "learning_rate": 3.192082005713881e-08, "loss": 0.1758, "step": 90920 }, { "epoch": 1.7829411764705883, "grad_norm": 2.2403414249420166, "learning_rate": 3.186385204724817e-08, "loss": 0.1533, "step": 90930 }, { "epoch": 1.7831372549019608, "grad_norm": 7.196314334869385, "learning_rate": 3.1806933244071634e-08, "loss": 0.1476, "step": 90940 }, { "epoch": 1.7833333333333332, "grad_norm": 4.708137035369873, "learning_rate": 3.175006365359206e-08, "loss": 0.1893, "step": 90950 }, { "epoch": 1.783529411764706, "grad_norm": 4.026877403259277, "learning_rate": 3.1693243281787016e-08, "loss": 0.19, "step": 90960 }, { "epoch": 1.7837254901960784, "grad_norm": 3.406883478164673, "learning_rate": 3.163647213462917e-08, "loss": 0.1801, "step": 90970 }, { "epoch": 1.783921568627451, "grad_norm": 3.26440691947937, "learning_rate": 3.1579750218085674e-08, "loss": 0.2021, "step": 90980 }, { "epoch": 1.7841176470588236, "grad_norm": 2.909762382507324, "learning_rate": 3.1523077538118726e-08, "loss": 0.1394, "step": 90990 }, { "epoch": 1.784313725490196, "grad_norm": 3.3888721466064453, "learning_rate": 3.1466454100685434e-08, "loss": 0.1471, "step": 91000 }, { "epoch": 1.7845098039215688, "grad_norm": 4.679535865783691, "learning_rate": 3.140987991173749e-08, "loss": 0.1527, "step": 91010 }, { "epoch": 1.7847058823529411, "grad_norm": 3.9777228832244873, "learning_rate": 3.135335497722147e-08, "loss": 0.1661, "step": 91020 }, { "epoch": 1.7849019607843137, "grad_norm": 5.906140327453613, "learning_rate": 3.1296879303078894e-08, "loss": 0.1294, "step": 91030 }, { "epoch": 1.7850980392156863, "grad_norm": 2.9600110054016113, "learning_rate": 3.1240452895246075e-08, "loss": 0.1617, "step": 91040 }, { "epoch": 1.7852941176470587, "grad_norm": 3.1579015254974365, "learning_rate": 3.1184075759653985e-08, "loss": 0.181, "step": 91050 }, { "epoch": 1.7854901960784315, "grad_norm": 4.0643157958984375, "learning_rate": 3.1127747902228606e-08, "loss": 0.1732, "step": 91060 }, { "epoch": 1.7856862745098039, "grad_norm": 3.070841073989868, "learning_rate": 3.107146932889071e-08, "loss": 0.1828, "step": 91070 }, { "epoch": 1.7858823529411765, "grad_norm": 3.1671109199523926, "learning_rate": 3.101524004555572e-08, "loss": 0.156, "step": 91080 }, { "epoch": 1.786078431372549, "grad_norm": 2.0807735919952393, "learning_rate": 3.09590600581342e-08, "loss": 0.157, "step": 91090 }, { "epoch": 1.7862745098039214, "grad_norm": 2.90020751953125, "learning_rate": 3.090292937253108e-08, "loss": 0.1502, "step": 91100 }, { "epoch": 1.7864705882352943, "grad_norm": 3.758770704269409, "learning_rate": 3.0846847994646655e-08, "loss": 0.1724, "step": 91110 }, { "epoch": 1.7866666666666666, "grad_norm": 4.748383522033691, "learning_rate": 3.079081593037552e-08, "loss": 0.1635, "step": 91120 }, { "epoch": 1.7868627450980392, "grad_norm": 2.416781425476074, "learning_rate": 3.073483318560749e-08, "loss": 0.1329, "step": 91130 }, { "epoch": 1.7870588235294118, "grad_norm": 2.000581741333008, "learning_rate": 3.0678899766226895e-08, "loss": 0.1212, "step": 91140 }, { "epoch": 1.7872549019607842, "grad_norm": 1.9800293445587158, "learning_rate": 3.062301567811298e-08, "loss": 0.177, "step": 91150 }, { "epoch": 1.787450980392157, "grad_norm": 3.0576746463775635, "learning_rate": 3.056718092714e-08, "loss": 0.1901, "step": 91160 }, { "epoch": 1.7876470588235294, "grad_norm": 2.922712802886963, "learning_rate": 3.051139551917675e-08, "loss": 0.1387, "step": 91170 }, { "epoch": 1.787843137254902, "grad_norm": 8.087736129760742, "learning_rate": 3.045565946008688e-08, "loss": 0.1846, "step": 91180 }, { "epoch": 1.7880392156862746, "grad_norm": 3.2319483757019043, "learning_rate": 3.039997275572903e-08, "loss": 0.1474, "step": 91190 }, { "epoch": 1.788235294117647, "grad_norm": 2.3461086750030518, "learning_rate": 3.034433541195658e-08, "loss": 0.1632, "step": 91200 }, { "epoch": 1.7884313725490197, "grad_norm": 2.7658021450042725, "learning_rate": 3.028874743461746e-08, "loss": 0.146, "step": 91210 }, { "epoch": 1.788627450980392, "grad_norm": 5.065966606140137, "learning_rate": 3.023320882955482e-08, "loss": 0.1586, "step": 91220 }, { "epoch": 1.7888235294117647, "grad_norm": 1.8612316846847534, "learning_rate": 3.0177719602606457e-08, "loss": 0.157, "step": 91230 }, { "epoch": 1.7890196078431373, "grad_norm": 4.099499225616455, "learning_rate": 3.0122279759604744e-08, "loss": 0.1419, "step": 91240 }, { "epoch": 1.7892156862745097, "grad_norm": 5.284189701080322, "learning_rate": 3.00668893063773e-08, "loss": 0.1355, "step": 91250 }, { "epoch": 1.7894117647058825, "grad_norm": 3.0422184467315674, "learning_rate": 3.0011548248746246e-08, "loss": 0.1231, "step": 91260 }, { "epoch": 1.7896078431372549, "grad_norm": 4.719525337219238, "learning_rate": 2.995625659252848e-08, "loss": 0.1671, "step": 91270 }, { "epoch": 1.7898039215686274, "grad_norm": 1.7230113744735718, "learning_rate": 2.9901014343535926e-08, "loss": 0.129, "step": 91280 }, { "epoch": 1.79, "grad_norm": 1.721483826637268, "learning_rate": 2.984582150757525e-08, "loss": 0.1623, "step": 91290 }, { "epoch": 1.7901960784313724, "grad_norm": 4.062942028045654, "learning_rate": 2.9790678090447884e-08, "loss": 0.1561, "step": 91300 }, { "epoch": 1.7903921568627452, "grad_norm": 6.327994346618652, "learning_rate": 2.973558409794985e-08, "loss": 0.1348, "step": 91310 }, { "epoch": 1.7905882352941176, "grad_norm": 2.417030096054077, "learning_rate": 2.968053953587246e-08, "loss": 0.18, "step": 91320 }, { "epoch": 1.7907843137254902, "grad_norm": 2.856907606124878, "learning_rate": 2.9625544410001425e-08, "loss": 0.1342, "step": 91330 }, { "epoch": 1.7909803921568628, "grad_norm": 4.516937732696533, "learning_rate": 2.9570598726117278e-08, "loss": 0.1602, "step": 91340 }, { "epoch": 1.7911764705882351, "grad_norm": 4.5722856521606445, "learning_rate": 2.9515702489995688e-08, "loss": 0.1573, "step": 91350 }, { "epoch": 1.791372549019608, "grad_norm": 2.6209917068481445, "learning_rate": 2.9460855707406863e-08, "loss": 0.1602, "step": 91360 }, { "epoch": 1.7915686274509803, "grad_norm": 3.999408483505249, "learning_rate": 2.9406058384115638e-08, "loss": 0.1857, "step": 91370 }, { "epoch": 1.791764705882353, "grad_norm": 2.3800675868988037, "learning_rate": 2.9351310525882188e-08, "loss": 0.1434, "step": 91380 }, { "epoch": 1.7919607843137255, "grad_norm": 3.6264336109161377, "learning_rate": 2.929661213846096e-08, "loss": 0.163, "step": 91390 }, { "epoch": 1.792156862745098, "grad_norm": 7.550563335418701, "learning_rate": 2.9241963227601362e-08, "loss": 0.1604, "step": 91400 }, { "epoch": 1.7923529411764707, "grad_norm": 4.979345798492432, "learning_rate": 2.9187363799047904e-08, "loss": 0.1665, "step": 91410 }, { "epoch": 1.792549019607843, "grad_norm": 3.7906696796417236, "learning_rate": 2.913281385853944e-08, "loss": 0.1783, "step": 91420 }, { "epoch": 1.7927450980392157, "grad_norm": 3.1415693759918213, "learning_rate": 2.9078313411809775e-08, "loss": 0.1763, "step": 91430 }, { "epoch": 1.7929411764705883, "grad_norm": 1.3484605550765991, "learning_rate": 2.9023862464587713e-08, "loss": 0.1441, "step": 91440 }, { "epoch": 1.7931372549019606, "grad_norm": 4.59124755859375, "learning_rate": 2.8969461022596676e-08, "loss": 0.1826, "step": 91450 }, { "epoch": 1.7933333333333334, "grad_norm": 2.8047573566436768, "learning_rate": 2.891510909155476e-08, "loss": 0.1691, "step": 91460 }, { "epoch": 1.7935294117647058, "grad_norm": 3.1097211837768555, "learning_rate": 2.8860806677175164e-08, "loss": 0.1334, "step": 91470 }, { "epoch": 1.7937254901960784, "grad_norm": 2.5227346420288086, "learning_rate": 2.880655378516572e-08, "loss": 0.1684, "step": 91480 }, { "epoch": 1.793921568627451, "grad_norm": 2.330803871154785, "learning_rate": 2.8752350421228976e-08, "loss": 0.1661, "step": 91490 }, { "epoch": 1.7941176470588234, "grad_norm": 5.601288318634033, "learning_rate": 2.8698196591062317e-08, "loss": 0.1531, "step": 91500 }, { "epoch": 1.7943137254901962, "grad_norm": 2.245849609375, "learning_rate": 2.8644092300358136e-08, "loss": 0.1396, "step": 91510 }, { "epoch": 1.7945098039215686, "grad_norm": 1.990762710571289, "learning_rate": 2.859003755480327e-08, "loss": 0.1821, "step": 91520 }, { "epoch": 1.7947058823529412, "grad_norm": 2.6170170307159424, "learning_rate": 2.8536032360079564e-08, "loss": 0.1627, "step": 91530 }, { "epoch": 1.7949019607843137, "grad_norm": 1.8337239027023315, "learning_rate": 2.8482076721863647e-08, "loss": 0.1558, "step": 91540 }, { "epoch": 1.7950980392156861, "grad_norm": 4.895341396331787, "learning_rate": 2.842817064582692e-08, "loss": 0.1779, "step": 91550 }, { "epoch": 1.795294117647059, "grad_norm": 2.0304253101348877, "learning_rate": 2.837431413763547e-08, "loss": 0.1577, "step": 91560 }, { "epoch": 1.7954901960784313, "grad_norm": 2.1932504177093506, "learning_rate": 2.8320507202950372e-08, "loss": 0.1838, "step": 91570 }, { "epoch": 1.795686274509804, "grad_norm": 2.9300029277801514, "learning_rate": 2.8266749847427275e-08, "loss": 0.1498, "step": 91580 }, { "epoch": 1.7958823529411765, "grad_norm": 6.369248867034912, "learning_rate": 2.821304207671671e-08, "loss": 0.1458, "step": 91590 }, { "epoch": 1.7960784313725489, "grad_norm": 6.248647212982178, "learning_rate": 2.8159383896464163e-08, "loss": 0.1406, "step": 91600 }, { "epoch": 1.7962745098039217, "grad_norm": 5.0916924476623535, "learning_rate": 2.8105775312309622e-08, "loss": 0.178, "step": 91610 }, { "epoch": 1.796470588235294, "grad_norm": 2.2361199855804443, "learning_rate": 2.8052216329887968e-08, "loss": 0.1328, "step": 91620 }, { "epoch": 1.7966666666666666, "grad_norm": 3.463287115097046, "learning_rate": 2.7998706954828978e-08, "loss": 0.1463, "step": 91630 }, { "epoch": 1.7968627450980392, "grad_norm": 4.0109333992004395, "learning_rate": 2.7945247192757037e-08, "loss": 0.1812, "step": 91640 }, { "epoch": 1.7970588235294118, "grad_norm": 3.4945878982543945, "learning_rate": 2.7891837049291544e-08, "loss": 0.1586, "step": 91650 }, { "epoch": 1.7972549019607844, "grad_norm": 8.165363311767578, "learning_rate": 2.7838476530046395e-08, "loss": 0.1564, "step": 91660 }, { "epoch": 1.7974509803921568, "grad_norm": 1.466667652130127, "learning_rate": 2.778516564063055e-08, "loss": 0.1372, "step": 91670 }, { "epoch": 1.7976470588235294, "grad_norm": 2.1540260314941406, "learning_rate": 2.773190438664752e-08, "loss": 0.1481, "step": 91680 }, { "epoch": 1.797843137254902, "grad_norm": 2.777458906173706, "learning_rate": 2.7678692773695723e-08, "loss": 0.1351, "step": 91690 }, { "epoch": 1.7980392156862746, "grad_norm": 3.223940849304199, "learning_rate": 2.7625530807368404e-08, "loss": 0.1671, "step": 91700 }, { "epoch": 1.7982352941176472, "grad_norm": 2.2273600101470947, "learning_rate": 2.7572418493253425e-08, "loss": 0.1461, "step": 91710 }, { "epoch": 1.7984313725490195, "grad_norm": 7.729284286499023, "learning_rate": 2.7519355836933546e-08, "loss": 0.1253, "step": 91720 }, { "epoch": 1.7986274509803921, "grad_norm": 3.173243522644043, "learning_rate": 2.7466342843986357e-08, "loss": 0.1636, "step": 91730 }, { "epoch": 1.7988235294117647, "grad_norm": 3.506786823272705, "learning_rate": 2.7413379519984127e-08, "loss": 0.1593, "step": 91740 }, { "epoch": 1.7990196078431373, "grad_norm": 1.9809508323669434, "learning_rate": 2.7360465870493898e-08, "loss": 0.1595, "step": 91750 }, { "epoch": 1.79921568627451, "grad_norm": 2.501143217086792, "learning_rate": 2.730760190107756e-08, "loss": 0.1535, "step": 91760 }, { "epoch": 1.7994117647058823, "grad_norm": 3.4060535430908203, "learning_rate": 2.7254787617291774e-08, "loss": 0.1556, "step": 91770 }, { "epoch": 1.7996078431372549, "grad_norm": 2.5175554752349854, "learning_rate": 2.7202023024687826e-08, "loss": 0.1844, "step": 91780 }, { "epoch": 1.7998039215686275, "grad_norm": 5.314558506011963, "learning_rate": 2.7149308128812055e-08, "loss": 0.1775, "step": 91790 }, { "epoch": 1.8, "grad_norm": 4.061481952667236, "learning_rate": 2.7096642935205417e-08, "loss": 0.1842, "step": 91800 }, { "epoch": 1.8001960784313726, "grad_norm": 2.6031064987182617, "learning_rate": 2.704402744940354e-08, "loss": 0.1575, "step": 91810 }, { "epoch": 1.800392156862745, "grad_norm": 3.422520875930786, "learning_rate": 2.6991461676937e-08, "loss": 0.1579, "step": 91820 }, { "epoch": 1.8005882352941176, "grad_norm": 4.099550724029541, "learning_rate": 2.6938945623331155e-08, "loss": 0.1579, "step": 91830 }, { "epoch": 1.8007843137254902, "grad_norm": 3.5688304901123047, "learning_rate": 2.6886479294106036e-08, "loss": 0.1476, "step": 91840 }, { "epoch": 1.8009803921568628, "grad_norm": 3.9006381034851074, "learning_rate": 2.6834062694776393e-08, "loss": 0.1785, "step": 91850 }, { "epoch": 1.8011764705882354, "grad_norm": 2.1616201400756836, "learning_rate": 2.6781695830851935e-08, "loss": 0.1986, "step": 91860 }, { "epoch": 1.8013725490196077, "grad_norm": 3.443020820617676, "learning_rate": 2.672937870783709e-08, "loss": 0.1807, "step": 91870 }, { "epoch": 1.8015686274509803, "grad_norm": 6.341641902923584, "learning_rate": 2.667711133123085e-08, "loss": 0.1481, "step": 91880 }, { "epoch": 1.801764705882353, "grad_norm": 3.7640554904937744, "learning_rate": 2.6624893706527262e-08, "loss": 0.2102, "step": 91890 }, { "epoch": 1.8019607843137255, "grad_norm": 3.7292051315307617, "learning_rate": 2.657272583921505e-08, "loss": 0.2137, "step": 91900 }, { "epoch": 1.8021568627450981, "grad_norm": 2.1021666526794434, "learning_rate": 2.6520607734777544e-08, "loss": 0.1464, "step": 91910 }, { "epoch": 1.8023529411764705, "grad_norm": 2.242141008377075, "learning_rate": 2.6468539398693146e-08, "loss": 0.1565, "step": 91920 }, { "epoch": 1.802549019607843, "grad_norm": 3.0395405292510986, "learning_rate": 2.6416520836434752e-08, "loss": 0.1719, "step": 91930 }, { "epoch": 1.8027450980392157, "grad_norm": 4.237236499786377, "learning_rate": 2.6364552053470157e-08, "loss": 0.1815, "step": 91940 }, { "epoch": 1.8029411764705883, "grad_norm": 4.050787448883057, "learning_rate": 2.6312633055261936e-08, "loss": 0.1739, "step": 91950 }, { "epoch": 1.8031372549019609, "grad_norm": 4.984015464782715, "learning_rate": 2.6260763847267387e-08, "loss": 0.1516, "step": 91960 }, { "epoch": 1.8033333333333332, "grad_norm": 3.8019604682922363, "learning_rate": 2.6208944434938483e-08, "loss": 0.1521, "step": 91970 }, { "epoch": 1.803529411764706, "grad_norm": 2.4274215698242188, "learning_rate": 2.6157174823722207e-08, "loss": 0.1681, "step": 91980 }, { "epoch": 1.8037254901960784, "grad_norm": 6.137797832489014, "learning_rate": 2.6105455019060085e-08, "loss": 0.1817, "step": 91990 }, { "epoch": 1.803921568627451, "grad_norm": 2.996764659881592, "learning_rate": 2.6053785026388554e-08, "loss": 0.1963, "step": 92000 }, { "epoch": 1.8041176470588236, "grad_norm": 2.623760461807251, "learning_rate": 2.600216485113865e-08, "loss": 0.1628, "step": 92010 }, { "epoch": 1.804313725490196, "grad_norm": 2.7234930992126465, "learning_rate": 2.595059449873638e-08, "loss": 0.1787, "step": 92020 }, { "epoch": 1.8045098039215688, "grad_norm": 10.046682357788086, "learning_rate": 2.5899073974602392e-08, "loss": 0.1677, "step": 92030 }, { "epoch": 1.8047058823529412, "grad_norm": 3.1116745471954346, "learning_rate": 2.5847603284151976e-08, "loss": 0.169, "step": 92040 }, { "epoch": 1.8049019607843138, "grad_norm": 3.7486298084259033, "learning_rate": 2.5796182432795465e-08, "loss": 0.1714, "step": 92050 }, { "epoch": 1.8050980392156863, "grad_norm": 2.8579611778259277, "learning_rate": 2.5744811425937763e-08, "loss": 0.1715, "step": 92060 }, { "epoch": 1.8052941176470587, "grad_norm": 4.555144309997559, "learning_rate": 2.5693490268978546e-08, "loss": 0.1534, "step": 92070 }, { "epoch": 1.8054901960784315, "grad_norm": 4.445648670196533, "learning_rate": 2.5642218967312336e-08, "loss": 0.1467, "step": 92080 }, { "epoch": 1.805686274509804, "grad_norm": 3.385164499282837, "learning_rate": 2.5590997526328317e-08, "loss": 0.1925, "step": 92090 }, { "epoch": 1.8058823529411765, "grad_norm": 2.768993616104126, "learning_rate": 2.5539825951410464e-08, "loss": 0.1708, "step": 92100 }, { "epoch": 1.806078431372549, "grad_norm": 2.641998052597046, "learning_rate": 2.5488704247937587e-08, "loss": 0.1601, "step": 92110 }, { "epoch": 1.8062745098039215, "grad_norm": 2.4745242595672607, "learning_rate": 2.5437632421283163e-08, "loss": 0.1684, "step": 92120 }, { "epoch": 1.8064705882352943, "grad_norm": 1.8203332424163818, "learning_rate": 2.538661047681534e-08, "loss": 0.1747, "step": 92130 }, { "epoch": 1.8066666666666666, "grad_norm": 2.850705146789551, "learning_rate": 2.5335638419897333e-08, "loss": 0.1668, "step": 92140 }, { "epoch": 1.8068627450980392, "grad_norm": 3.078371286392212, "learning_rate": 2.5284716255886797e-08, "loss": 0.1577, "step": 92150 }, { "epoch": 1.8070588235294118, "grad_norm": 2.1159708499908447, "learning_rate": 2.5233843990136228e-08, "loss": 0.1533, "step": 92160 }, { "epoch": 1.8072549019607842, "grad_norm": 2.023362398147583, "learning_rate": 2.518302162799296e-08, "loss": 0.1462, "step": 92170 }, { "epoch": 1.807450980392157, "grad_norm": 2.542008399963379, "learning_rate": 2.5132249174799115e-08, "loss": 0.1423, "step": 92180 }, { "epoch": 1.8076470588235294, "grad_norm": 3.170121669769287, "learning_rate": 2.5081526635891414e-08, "loss": 0.1669, "step": 92190 }, { "epoch": 1.807843137254902, "grad_norm": 4.843275547027588, "learning_rate": 2.5030854016601267e-08, "loss": 0.1805, "step": 92200 }, { "epoch": 1.8080392156862746, "grad_norm": 2.4509425163269043, "learning_rate": 2.498023132225524e-08, "loss": 0.1512, "step": 92210 }, { "epoch": 1.808235294117647, "grad_norm": 2.495920181274414, "learning_rate": 2.4929658558174194e-08, "loss": 0.139, "step": 92220 }, { "epoch": 1.8084313725490198, "grad_norm": 3.1415467262268066, "learning_rate": 2.487913572967393e-08, "loss": 0.1577, "step": 92230 }, { "epoch": 1.8086274509803921, "grad_norm": 4.403148174285889, "learning_rate": 2.482866284206514e-08, "loss": 0.1343, "step": 92240 }, { "epoch": 1.8088235294117647, "grad_norm": 3.0959744453430176, "learning_rate": 2.477823990065303e-08, "loss": 0.1682, "step": 92250 }, { "epoch": 1.8090196078431373, "grad_norm": 4.005143642425537, "learning_rate": 2.4727866910737582e-08, "loss": 0.1875, "step": 92260 }, { "epoch": 1.8092156862745097, "grad_norm": 3.1250808238983154, "learning_rate": 2.4677543877613782e-08, "loss": 0.1475, "step": 92270 }, { "epoch": 1.8094117647058825, "grad_norm": 6.1797661781311035, "learning_rate": 2.4627270806571123e-08, "loss": 0.166, "step": 92280 }, { "epoch": 1.8096078431372549, "grad_norm": 6.363940715789795, "learning_rate": 2.457704770289376e-08, "loss": 0.1708, "step": 92290 }, { "epoch": 1.8098039215686275, "grad_norm": 2.8254446983337402, "learning_rate": 2.4526874571860922e-08, "loss": 0.156, "step": 92300 }, { "epoch": 1.81, "grad_norm": 2.36224365234375, "learning_rate": 2.4476751418746333e-08, "loss": 0.1391, "step": 92310 }, { "epoch": 1.8101960784313724, "grad_norm": 2.98189640045166, "learning_rate": 2.44266782488185e-08, "loss": 0.1679, "step": 92320 }, { "epoch": 1.8103921568627452, "grad_norm": 1.927029013633728, "learning_rate": 2.4376655067340767e-08, "loss": 0.1512, "step": 92330 }, { "epoch": 1.8105882352941176, "grad_norm": 8.174689292907715, "learning_rate": 2.432668187957121e-08, "loss": 0.1747, "step": 92340 }, { "epoch": 1.8107843137254902, "grad_norm": 6.197778701782227, "learning_rate": 2.427675869076251e-08, "loss": 0.1591, "step": 92350 }, { "epoch": 1.8109803921568628, "grad_norm": 2.2223541736602783, "learning_rate": 2.42268855061622e-08, "loss": 0.2115, "step": 92360 }, { "epoch": 1.8111764705882352, "grad_norm": 5.035487174987793, "learning_rate": 2.4177062331012633e-08, "loss": 0.1735, "step": 92370 }, { "epoch": 1.811372549019608, "grad_norm": 3.291673421859741, "learning_rate": 2.412728917055079e-08, "loss": 0.171, "step": 92380 }, { "epoch": 1.8115686274509804, "grad_norm": 2.5459978580474854, "learning_rate": 2.407756603000838e-08, "loss": 0.1886, "step": 92390 }, { "epoch": 1.811764705882353, "grad_norm": 2.436527729034424, "learning_rate": 2.4027892914611992e-08, "loss": 0.1741, "step": 92400 }, { "epoch": 1.8119607843137255, "grad_norm": 4.680018901824951, "learning_rate": 2.3978269829582786e-08, "loss": 0.1548, "step": 92410 }, { "epoch": 1.812156862745098, "grad_norm": 5.6535725593566895, "learning_rate": 2.3928696780136705e-08, "loss": 0.146, "step": 92420 }, { "epoch": 1.8123529411764707, "grad_norm": 1.6202852725982666, "learning_rate": 2.3879173771484574e-08, "loss": 0.1634, "step": 92430 }, { "epoch": 1.812549019607843, "grad_norm": 2.4540555477142334, "learning_rate": 2.3829700808831787e-08, "loss": 0.1562, "step": 92440 }, { "epoch": 1.8127450980392157, "grad_norm": 3.9740548133850098, "learning_rate": 2.378027789737852e-08, "loss": 0.1401, "step": 92450 }, { "epoch": 1.8129411764705883, "grad_norm": 2.8627350330352783, "learning_rate": 2.3730905042319836e-08, "loss": 0.1813, "step": 92460 }, { "epoch": 1.8131372549019606, "grad_norm": 2.1283085346221924, "learning_rate": 2.3681582248845365e-08, "loss": 0.1763, "step": 92470 }, { "epoch": 1.8133333333333335, "grad_norm": 3.223050355911255, "learning_rate": 2.3632309522139347e-08, "loss": 0.163, "step": 92480 }, { "epoch": 1.8135294117647058, "grad_norm": 2.103266716003418, "learning_rate": 2.3583086867381198e-08, "loss": 0.1257, "step": 92490 }, { "epoch": 1.8137254901960784, "grad_norm": 3.751681089401245, "learning_rate": 2.3533914289744716e-08, "loss": 0.1278, "step": 92500 }, { "epoch": 1.813921568627451, "grad_norm": 2.747363328933716, "learning_rate": 2.3484791794398384e-08, "loss": 0.1551, "step": 92510 }, { "epoch": 1.8141176470588234, "grad_norm": 1.3960938453674316, "learning_rate": 2.343571938650579e-08, "loss": 0.1386, "step": 92520 }, { "epoch": 1.8143137254901962, "grad_norm": 6.305624008178711, "learning_rate": 2.3386697071224816e-08, "loss": 0.2249, "step": 92530 }, { "epoch": 1.8145098039215686, "grad_norm": 4.404764175415039, "learning_rate": 2.3337724853708495e-08, "loss": 0.1696, "step": 92540 }, { "epoch": 1.8147058823529412, "grad_norm": 1.7062835693359375, "learning_rate": 2.3288802739104218e-08, "loss": 0.1631, "step": 92550 }, { "epoch": 1.8149019607843138, "grad_norm": 4.098809719085693, "learning_rate": 2.323993073255448e-08, "loss": 0.1443, "step": 92560 }, { "epoch": 1.8150980392156861, "grad_norm": 8.082550048828125, "learning_rate": 2.3191108839196172e-08, "loss": 0.1867, "step": 92570 }, { "epoch": 1.815294117647059, "grad_norm": 3.3389742374420166, "learning_rate": 2.3142337064161023e-08, "loss": 0.1611, "step": 92580 }, { "epoch": 1.8154901960784313, "grad_norm": 3.3402223587036133, "learning_rate": 2.309361541257565e-08, "loss": 0.1473, "step": 92590 }, { "epoch": 1.815686274509804, "grad_norm": 5.9545578956604, "learning_rate": 2.304494388956124e-08, "loss": 0.186, "step": 92600 }, { "epoch": 1.8158823529411765, "grad_norm": 1.9836760759353638, "learning_rate": 2.2996322500233644e-08, "loss": 0.1564, "step": 92610 }, { "epoch": 1.8160784313725489, "grad_norm": 2.9638264179229736, "learning_rate": 2.2947751249703772e-08, "loss": 0.1477, "step": 92620 }, { "epoch": 1.8162745098039217, "grad_norm": 3.7895030975341797, "learning_rate": 2.2899230143076874e-08, "loss": 0.1605, "step": 92630 }, { "epoch": 1.816470588235294, "grad_norm": 3.2335145473480225, "learning_rate": 2.285075918545315e-08, "loss": 0.13, "step": 92640 }, { "epoch": 1.8166666666666667, "grad_norm": 3.226975202560425, "learning_rate": 2.2802338381927467e-08, "loss": 0.1653, "step": 92650 }, { "epoch": 1.8168627450980392, "grad_norm": 2.586106538772583, "learning_rate": 2.2753967737589474e-08, "loss": 0.1857, "step": 92660 }, { "epoch": 1.8170588235294116, "grad_norm": 3.086845636367798, "learning_rate": 2.2705647257523442e-08, "loss": 0.1271, "step": 92670 }, { "epoch": 1.8172549019607844, "grad_norm": 4.565742492675781, "learning_rate": 2.2657376946808472e-08, "loss": 0.1741, "step": 92680 }, { "epoch": 1.8174509803921568, "grad_norm": 5.871712684631348, "learning_rate": 2.2609156810518337e-08, "loss": 0.2107, "step": 92690 }, { "epoch": 1.8176470588235294, "grad_norm": 3.7791285514831543, "learning_rate": 2.256098685372154e-08, "loss": 0.1956, "step": 92700 }, { "epoch": 1.817843137254902, "grad_norm": 10.885112762451172, "learning_rate": 2.2512867081481256e-08, "loss": 0.2118, "step": 92710 }, { "epoch": 1.8180392156862744, "grad_norm": 3.4824187755584717, "learning_rate": 2.2464797498855658e-08, "loss": 0.1718, "step": 92720 }, { "epoch": 1.8182352941176472, "grad_norm": 3.9986379146575928, "learning_rate": 2.241677811089726e-08, "loss": 0.1611, "step": 92730 }, { "epoch": 1.8184313725490195, "grad_norm": 1.9289559125900269, "learning_rate": 2.2368808922653526e-08, "loss": 0.1589, "step": 92740 }, { "epoch": 1.8186274509803921, "grad_norm": 4.2516865730285645, "learning_rate": 2.2320889939166587e-08, "loss": 0.1319, "step": 92750 }, { "epoch": 1.8188235294117647, "grad_norm": 4.241572380065918, "learning_rate": 2.2273021165473303e-08, "loss": 0.1784, "step": 92760 }, { "epoch": 1.819019607843137, "grad_norm": 3.033608913421631, "learning_rate": 2.222520260660521e-08, "loss": 0.153, "step": 92770 }, { "epoch": 1.81921568627451, "grad_norm": 2.042746067047119, "learning_rate": 2.217743426758867e-08, "loss": 0.1623, "step": 92780 }, { "epoch": 1.8194117647058823, "grad_norm": 2.177009344100952, "learning_rate": 2.2129716153444723e-08, "loss": 0.1461, "step": 92790 }, { "epoch": 1.8196078431372549, "grad_norm": 4.320364952087402, "learning_rate": 2.2082048269189024e-08, "loss": 0.1595, "step": 92800 }, { "epoch": 1.8198039215686275, "grad_norm": 2.391594886779785, "learning_rate": 2.2034430619832177e-08, "loss": 0.1287, "step": 92810 }, { "epoch": 1.8199999999999998, "grad_norm": 5.694700717926025, "learning_rate": 2.1986863210379237e-08, "loss": 0.2031, "step": 92820 }, { "epoch": 1.8201960784313727, "grad_norm": 2.2221949100494385, "learning_rate": 2.1939346045830144e-08, "loss": 0.15, "step": 92830 }, { "epoch": 1.820392156862745, "grad_norm": 3.13676381111145, "learning_rate": 2.189187913117957e-08, "loss": 0.1534, "step": 92840 }, { "epoch": 1.8205882352941176, "grad_norm": 2.274855375289917, "learning_rate": 2.18444624714168e-08, "loss": 0.1757, "step": 92850 }, { "epoch": 1.8207843137254902, "grad_norm": 3.4062156677246094, "learning_rate": 2.1797096071525855e-08, "loss": 0.1371, "step": 92860 }, { "epoch": 1.8209803921568626, "grad_norm": 5.2059407234191895, "learning_rate": 2.174977993648569e-08, "loss": 0.1556, "step": 92870 }, { "epoch": 1.8211764705882354, "grad_norm": 3.151927947998047, "learning_rate": 2.1702514071269607e-08, "loss": 0.1162, "step": 92880 }, { "epoch": 1.8213725490196078, "grad_norm": 3.013315439224243, "learning_rate": 2.165529848084585e-08, "loss": 0.1761, "step": 92890 }, { "epoch": 1.8215686274509804, "grad_norm": 4.165760040283203, "learning_rate": 2.1608133170177346e-08, "loss": 0.1626, "step": 92900 }, { "epoch": 1.821764705882353, "grad_norm": 2.5993242263793945, "learning_rate": 2.1561018144221897e-08, "loss": 0.2079, "step": 92910 }, { "epoch": 1.8219607843137255, "grad_norm": 1.828224539756775, "learning_rate": 2.151395340793166e-08, "loss": 0.1503, "step": 92920 }, { "epoch": 1.8221568627450981, "grad_norm": 1.967411756515503, "learning_rate": 2.1466938966253734e-08, "loss": 0.16, "step": 92930 }, { "epoch": 1.8223529411764705, "grad_norm": 4.428292751312256, "learning_rate": 2.1419974824129938e-08, "loss": 0.1329, "step": 92940 }, { "epoch": 1.822549019607843, "grad_norm": 3.8943188190460205, "learning_rate": 2.1373060986496828e-08, "loss": 0.1508, "step": 92950 }, { "epoch": 1.8227450980392157, "grad_norm": 4.3071513175964355, "learning_rate": 2.1326197458285454e-08, "loss": 0.1631, "step": 92960 }, { "epoch": 1.8229411764705883, "grad_norm": 1.4045288562774658, "learning_rate": 2.1279384244421883e-08, "loss": 0.1463, "step": 92970 }, { "epoch": 1.8231372549019609, "grad_norm": 4.121142864227295, "learning_rate": 2.1232621349826672e-08, "loss": 0.1855, "step": 92980 }, { "epoch": 1.8233333333333333, "grad_norm": 3.4228241443634033, "learning_rate": 2.1185908779415117e-08, "loss": 0.1663, "step": 92990 }, { "epoch": 1.8235294117647058, "grad_norm": 2.348532199859619, "learning_rate": 2.1139246538097344e-08, "loss": 0.1653, "step": 93000 }, { "epoch": 1.8237254901960784, "grad_norm": 2.2005605697631836, "learning_rate": 2.109263463077815e-08, "loss": 0.1475, "step": 93010 }, { "epoch": 1.823921568627451, "grad_norm": 3.3137855529785156, "learning_rate": 2.104607306235684e-08, "loss": 0.1728, "step": 93020 }, { "epoch": 1.8241176470588236, "grad_norm": 3.833616256713867, "learning_rate": 2.0999561837727776e-08, "loss": 0.145, "step": 93030 }, { "epoch": 1.824313725490196, "grad_norm": 4.694371223449707, "learning_rate": 2.0953100961779824e-08, "loss": 0.1544, "step": 93040 }, { "epoch": 1.8245098039215686, "grad_norm": 3.6640334129333496, "learning_rate": 2.0906690439396356e-08, "loss": 0.153, "step": 93050 }, { "epoch": 1.8247058823529412, "grad_norm": 2.8828680515289307, "learning_rate": 2.0860330275455974e-08, "loss": 0.1601, "step": 93060 }, { "epoch": 1.8249019607843138, "grad_norm": 3.1693687438964844, "learning_rate": 2.0814020474831496e-08, "loss": 0.1735, "step": 93070 }, { "epoch": 1.8250980392156864, "grad_norm": 3.2539217472076416, "learning_rate": 2.0767761042390698e-08, "loss": 0.1902, "step": 93080 }, { "epoch": 1.8252941176470587, "grad_norm": 4.6056389808654785, "learning_rate": 2.0721551982995967e-08, "loss": 0.1593, "step": 93090 }, { "epoch": 1.8254901960784313, "grad_norm": 5.355638027191162, "learning_rate": 2.067539330150453e-08, "loss": 0.1456, "step": 93100 }, { "epoch": 1.825686274509804, "grad_norm": 2.196620464324951, "learning_rate": 2.0629285002768114e-08, "loss": 0.1358, "step": 93110 }, { "epoch": 1.8258823529411765, "grad_norm": 3.589651346206665, "learning_rate": 2.0583227091633238e-08, "loss": 0.1764, "step": 93120 }, { "epoch": 1.826078431372549, "grad_norm": 3.440211296081543, "learning_rate": 2.053721957294119e-08, "loss": 0.1652, "step": 93130 }, { "epoch": 1.8262745098039215, "grad_norm": 2.859363555908203, "learning_rate": 2.049126245152799e-08, "loss": 0.1695, "step": 93140 }, { "epoch": 1.826470588235294, "grad_norm": 2.9985642433166504, "learning_rate": 2.0445355732224056e-08, "loss": 0.1616, "step": 93150 }, { "epoch": 1.8266666666666667, "grad_norm": 5.023952007293701, "learning_rate": 2.0399499419854972e-08, "loss": 0.1682, "step": 93160 }, { "epoch": 1.8268627450980393, "grad_norm": 5.251572132110596, "learning_rate": 2.0353693519240712e-08, "loss": 0.1657, "step": 93170 }, { "epoch": 1.8270588235294118, "grad_norm": 3.174997329711914, "learning_rate": 2.0307938035195925e-08, "loss": 0.167, "step": 93180 }, { "epoch": 1.8272549019607842, "grad_norm": 2.446481227874756, "learning_rate": 2.0262232972530157e-08, "loss": 0.1446, "step": 93190 }, { "epoch": 1.8274509803921568, "grad_norm": 4.4559407234191895, "learning_rate": 2.0216578336047563e-08, "loss": 0.1804, "step": 93200 }, { "epoch": 1.8276470588235294, "grad_norm": 2.4473931789398193, "learning_rate": 2.0170974130546912e-08, "loss": 0.1767, "step": 93210 }, { "epoch": 1.827843137254902, "grad_norm": 9.654277801513672, "learning_rate": 2.0125420360821876e-08, "loss": 0.1632, "step": 93220 }, { "epoch": 1.8280392156862746, "grad_norm": 3.6164283752441406, "learning_rate": 2.007991703166062e-08, "loss": 0.1573, "step": 93230 }, { "epoch": 1.828235294117647, "grad_norm": 4.152737617492676, "learning_rate": 2.0034464147846042e-08, "loss": 0.1372, "step": 93240 }, { "epoch": 1.8284313725490198, "grad_norm": 6.009202003479004, "learning_rate": 1.998906171415582e-08, "loss": 0.1753, "step": 93250 }, { "epoch": 1.8286274509803921, "grad_norm": 2.543933391571045, "learning_rate": 1.9943709735362357e-08, "loss": 0.1526, "step": 93260 }, { "epoch": 1.8288235294117647, "grad_norm": 5.224937915802002, "learning_rate": 1.9898408216232677e-08, "loss": 0.1727, "step": 93270 }, { "epoch": 1.8290196078431373, "grad_norm": 2.2439920902252197, "learning_rate": 1.9853157161528468e-08, "loss": 0.1543, "step": 93280 }, { "epoch": 1.8292156862745097, "grad_norm": 2.3855156898498535, "learning_rate": 1.9807956576006148e-08, "loss": 0.1271, "step": 93290 }, { "epoch": 1.8294117647058825, "grad_norm": 3.418869972229004, "learning_rate": 1.9762806464416916e-08, "loss": 0.1736, "step": 93300 }, { "epoch": 1.8296078431372549, "grad_norm": 4.6224517822265625, "learning_rate": 1.9717706831506477e-08, "loss": 0.1407, "step": 93310 }, { "epoch": 1.8298039215686275, "grad_norm": 3.026339054107666, "learning_rate": 1.9672657682015425e-08, "loss": 0.146, "step": 93320 }, { "epoch": 1.83, "grad_norm": 5.899168968200684, "learning_rate": 1.962765902067898e-08, "loss": 0.1966, "step": 93330 }, { "epoch": 1.8301960784313724, "grad_norm": 3.211107015609741, "learning_rate": 1.95827108522269e-08, "loss": 0.1551, "step": 93340 }, { "epoch": 1.8303921568627453, "grad_norm": 4.208362579345703, "learning_rate": 1.9537813181383922e-08, "loss": 0.171, "step": 93350 }, { "epoch": 1.8305882352941176, "grad_norm": 2.900691032409668, "learning_rate": 1.9492966012869372e-08, "loss": 0.1568, "step": 93360 }, { "epoch": 1.8307843137254902, "grad_norm": 4.205105304718018, "learning_rate": 1.9448169351396982e-08, "loss": 0.1719, "step": 93370 }, { "epoch": 1.8309803921568628, "grad_norm": 2.9664053916931152, "learning_rate": 1.9403423201675707e-08, "loss": 0.1492, "step": 93380 }, { "epoch": 1.8311764705882352, "grad_norm": 2.8182766437530518, "learning_rate": 1.935872756840873e-08, "loss": 0.1647, "step": 93390 }, { "epoch": 1.831372549019608, "grad_norm": 1.6129220724105835, "learning_rate": 1.9314082456294068e-08, "loss": 0.1456, "step": 93400 }, { "epoch": 1.8315686274509804, "grad_norm": 2.2973809242248535, "learning_rate": 1.926948787002458e-08, "loss": 0.1348, "step": 93410 }, { "epoch": 1.831764705882353, "grad_norm": 4.173738956451416, "learning_rate": 1.922494381428763e-08, "loss": 0.1777, "step": 93420 }, { "epoch": 1.8319607843137256, "grad_norm": 2.3755438327789307, "learning_rate": 1.918045029376536e-08, "loss": 0.1548, "step": 93430 }, { "epoch": 1.832156862745098, "grad_norm": 2.663383960723877, "learning_rate": 1.913600731313447e-08, "loss": 0.1167, "step": 93440 }, { "epoch": 1.8323529411764707, "grad_norm": 2.3785512447357178, "learning_rate": 1.909161487706662e-08, "loss": 0.1723, "step": 93450 }, { "epoch": 1.832549019607843, "grad_norm": 2.9129207134246826, "learning_rate": 1.9047272990227846e-08, "loss": 0.1396, "step": 93460 }, { "epoch": 1.8327450980392157, "grad_norm": 8.288172721862793, "learning_rate": 1.9002981657279037e-08, "loss": 0.1407, "step": 93470 }, { "epoch": 1.8329411764705883, "grad_norm": 3.443512201309204, "learning_rate": 1.8958740882875803e-08, "loss": 0.1563, "step": 93480 }, { "epoch": 1.8331372549019607, "grad_norm": 6.514165878295898, "learning_rate": 1.891455067166836e-08, "loss": 0.1696, "step": 93490 }, { "epoch": 1.8333333333333335, "grad_norm": 3.4093198776245117, "learning_rate": 1.887041102830156e-08, "loss": 0.17, "step": 93500 }, { "epoch": 1.8335294117647059, "grad_norm": 7.223086833953857, "learning_rate": 1.882632195741507e-08, "loss": 0.1726, "step": 93510 }, { "epoch": 1.8337254901960784, "grad_norm": 4.967798233032227, "learning_rate": 1.8782283463643243e-08, "loss": 0.1918, "step": 93520 }, { "epoch": 1.833921568627451, "grad_norm": 3.670401096343994, "learning_rate": 1.8738295551614825e-08, "loss": 0.2065, "step": 93530 }, { "epoch": 1.8341176470588234, "grad_norm": 3.2633049488067627, "learning_rate": 1.8694358225953778e-08, "loss": 0.1613, "step": 93540 }, { "epoch": 1.8343137254901962, "grad_norm": 3.2338008880615234, "learning_rate": 1.8650471491278186e-08, "loss": 0.1713, "step": 93550 }, { "epoch": 1.8345098039215686, "grad_norm": 2.614150047302246, "learning_rate": 1.8606635352201195e-08, "loss": 0.1654, "step": 93560 }, { "epoch": 1.8347058823529412, "grad_norm": 4.869675636291504, "learning_rate": 1.8562849813330506e-08, "loss": 0.1599, "step": 93570 }, { "epoch": 1.8349019607843138, "grad_norm": 3.0150251388549805, "learning_rate": 1.8519114879268495e-08, "loss": 0.188, "step": 93580 }, { "epoch": 1.8350980392156861, "grad_norm": 2.6274569034576416, "learning_rate": 1.8475430554612094e-08, "loss": 0.1536, "step": 93590 }, { "epoch": 1.835294117647059, "grad_norm": 3.153616428375244, "learning_rate": 1.8431796843953242e-08, "loss": 0.1657, "step": 93600 }, { "epoch": 1.8354901960784313, "grad_norm": 2.5738584995269775, "learning_rate": 1.8388213751878322e-08, "loss": 0.1628, "step": 93610 }, { "epoch": 1.835686274509804, "grad_norm": 3.831946611404419, "learning_rate": 1.83446812829684e-08, "loss": 0.1423, "step": 93620 }, { "epoch": 1.8358823529411765, "grad_norm": 2.9769351482391357, "learning_rate": 1.8301199441799197e-08, "loss": 0.1672, "step": 93630 }, { "epoch": 1.836078431372549, "grad_norm": 3.998730421066284, "learning_rate": 1.8257768232941283e-08, "loss": 0.1433, "step": 93640 }, { "epoch": 1.8362745098039217, "grad_norm": 3.0666608810424805, "learning_rate": 1.821438766095984e-08, "loss": 0.1709, "step": 93650 }, { "epoch": 1.836470588235294, "grad_norm": 3.5686583518981934, "learning_rate": 1.8171057730414496e-08, "loss": 0.1595, "step": 93660 }, { "epoch": 1.8366666666666667, "grad_norm": 4.858696460723877, "learning_rate": 1.8127778445859887e-08, "loss": 0.16, "step": 93670 }, { "epoch": 1.8368627450980393, "grad_norm": 2.913323163986206, "learning_rate": 1.8084549811845152e-08, "loss": 0.1773, "step": 93680 }, { "epoch": 1.8370588235294116, "grad_norm": 2.5534589290618896, "learning_rate": 1.8041371832914098e-08, "loss": 0.1398, "step": 93690 }, { "epoch": 1.8372549019607844, "grad_norm": 5.043476104736328, "learning_rate": 1.7998244513605378e-08, "loss": 0.1439, "step": 93700 }, { "epoch": 1.8374509803921568, "grad_norm": 3.934819459915161, "learning_rate": 1.795516785845208e-08, "loss": 0.1598, "step": 93710 }, { "epoch": 1.8376470588235294, "grad_norm": 1.552584171295166, "learning_rate": 1.7912141871982035e-08, "loss": 0.1546, "step": 93720 }, { "epoch": 1.837843137254902, "grad_norm": 4.06858491897583, "learning_rate": 1.7869166558717897e-08, "loss": 0.1592, "step": 93730 }, { "epoch": 1.8380392156862744, "grad_norm": 2.8887431621551514, "learning_rate": 1.7826241923176833e-08, "loss": 0.1756, "step": 93740 }, { "epoch": 1.8382352941176472, "grad_norm": 3.907710313796997, "learning_rate": 1.7783367969870677e-08, "loss": 0.1468, "step": 93750 }, { "epoch": 1.8384313725490196, "grad_norm": 2.304746150970459, "learning_rate": 1.7740544703306104e-08, "loss": 0.1911, "step": 93760 }, { "epoch": 1.8386274509803922, "grad_norm": 3.643524646759033, "learning_rate": 1.7697772127984347e-08, "loss": 0.1569, "step": 93770 }, { "epoch": 1.8388235294117647, "grad_norm": 2.343214750289917, "learning_rate": 1.7655050248401194e-08, "loss": 0.1428, "step": 93780 }, { "epoch": 1.8390196078431371, "grad_norm": 3.1381750106811523, "learning_rate": 1.7612379069047334e-08, "loss": 0.1896, "step": 93790 }, { "epoch": 1.83921568627451, "grad_norm": 3.284815788269043, "learning_rate": 1.7569758594408015e-08, "loss": 0.1636, "step": 93800 }, { "epoch": 1.8394117647058823, "grad_norm": 3.5201005935668945, "learning_rate": 1.7527188828963146e-08, "loss": 0.1667, "step": 93810 }, { "epoch": 1.839607843137255, "grad_norm": 2.9132471084594727, "learning_rate": 1.7484669777187267e-08, "loss": 0.1786, "step": 93820 }, { "epoch": 1.8398039215686275, "grad_norm": 1.6082954406738281, "learning_rate": 1.7442201443549686e-08, "loss": 0.1844, "step": 93830 }, { "epoch": 1.8399999999999999, "grad_norm": 4.0245890617370605, "learning_rate": 1.7399783832514393e-08, "loss": 0.1523, "step": 93840 }, { "epoch": 1.8401960784313727, "grad_norm": 2.6036365032196045, "learning_rate": 1.7357416948539815e-08, "loss": 0.2097, "step": 93850 }, { "epoch": 1.840392156862745, "grad_norm": 6.315070629119873, "learning_rate": 1.731510079607945e-08, "loss": 0.171, "step": 93860 }, { "epoch": 1.8405882352941176, "grad_norm": 3.7944962978363037, "learning_rate": 1.7272835379581075e-08, "loss": 0.1624, "step": 93870 }, { "epoch": 1.8407843137254902, "grad_norm": 7.742101192474365, "learning_rate": 1.7230620703487242e-08, "loss": 0.1437, "step": 93880 }, { "epoch": 1.8409803921568626, "grad_norm": 3.840319871902466, "learning_rate": 1.7188456772235403e-08, "loss": 0.1676, "step": 93890 }, { "epoch": 1.8411764705882354, "grad_norm": 0.9951925873756409, "learning_rate": 1.714634359025735e-08, "loss": 0.1448, "step": 93900 }, { "epoch": 1.8413725490196078, "grad_norm": 5.2677998542785645, "learning_rate": 1.7104281161979704e-08, "loss": 0.1706, "step": 93910 }, { "epoch": 1.8415686274509804, "grad_norm": 3.7487573623657227, "learning_rate": 1.7062269491823765e-08, "loss": 0.1542, "step": 93920 }, { "epoch": 1.841764705882353, "grad_norm": 65.20537567138672, "learning_rate": 1.70203085842055e-08, "loss": 0.1465, "step": 93930 }, { "epoch": 1.8419607843137253, "grad_norm": 5.985004901885986, "learning_rate": 1.6978398443535326e-08, "loss": 0.1988, "step": 93940 }, { "epoch": 1.8421568627450982, "grad_norm": 4.7297821044921875, "learning_rate": 1.6936539074218714e-08, "loss": 0.1922, "step": 93950 }, { "epoch": 1.8423529411764705, "grad_norm": 2.174078941345215, "learning_rate": 1.6894730480655483e-08, "loss": 0.1608, "step": 93960 }, { "epoch": 1.8425490196078431, "grad_norm": 4.725101470947266, "learning_rate": 1.6852972667240228e-08, "loss": 0.171, "step": 93970 }, { "epoch": 1.8427450980392157, "grad_norm": 2.704026699066162, "learning_rate": 1.681126563836216e-08, "loss": 0.1506, "step": 93980 }, { "epoch": 1.842941176470588, "grad_norm": 5.914866924285889, "learning_rate": 1.676960939840527e-08, "loss": 0.1429, "step": 93990 }, { "epoch": 1.843137254901961, "grad_norm": 4.006988048553467, "learning_rate": 1.6728003951748117e-08, "loss": 0.1889, "step": 94000 }, { "epoch": 1.8433333333333333, "grad_norm": 4.445137023925781, "learning_rate": 1.6686449302763806e-08, "loss": 0.1764, "step": 94010 }, { "epoch": 1.8435294117647059, "grad_norm": 1.8630023002624512, "learning_rate": 1.6644945455820348e-08, "loss": 0.1743, "step": 94020 }, { "epoch": 1.8437254901960785, "grad_norm": 3.657540798187256, "learning_rate": 1.6603492415280307e-08, "loss": 0.1861, "step": 94030 }, { "epoch": 1.8439215686274508, "grad_norm": 1.5305033922195435, "learning_rate": 1.656209018550081e-08, "loss": 0.1944, "step": 94040 }, { "epoch": 1.8441176470588236, "grad_norm": 1.8260923624038696, "learning_rate": 1.6520738770833763e-08, "loss": 0.1315, "step": 94050 }, { "epoch": 1.844313725490196, "grad_norm": 2.429561138153076, "learning_rate": 1.647943817562575e-08, "loss": 0.1296, "step": 94060 }, { "epoch": 1.8445098039215686, "grad_norm": 1.3997759819030762, "learning_rate": 1.6438188404217845e-08, "loss": 0.1641, "step": 94070 }, { "epoch": 1.8447058823529412, "grad_norm": 4.248830318450928, "learning_rate": 1.6396989460945977e-08, "loss": 0.1454, "step": 94080 }, { "epoch": 1.8449019607843136, "grad_norm": 3.146409749984741, "learning_rate": 1.6355841350140676e-08, "loss": 0.1499, "step": 94090 }, { "epoch": 1.8450980392156864, "grad_norm": 3.5534942150115967, "learning_rate": 1.631474407612693e-08, "loss": 0.1749, "step": 94100 }, { "epoch": 1.8452941176470588, "grad_norm": 4.522618770599365, "learning_rate": 1.6273697643224783e-08, "loss": 0.1876, "step": 94110 }, { "epoch": 1.8454901960784313, "grad_norm": 2.254260778427124, "learning_rate": 1.6232702055748616e-08, "loss": 0.1695, "step": 94120 }, { "epoch": 1.845686274509804, "grad_norm": 2.4869556427001953, "learning_rate": 1.6191757318007425e-08, "loss": 0.1417, "step": 94130 }, { "epoch": 1.8458823529411763, "grad_norm": 2.616736650466919, "learning_rate": 1.615086343430516e-08, "loss": 0.1563, "step": 94140 }, { "epoch": 1.8460784313725491, "grad_norm": 7.149663925170898, "learning_rate": 1.6110020408940218e-08, "loss": 0.1763, "step": 94150 }, { "epoch": 1.8462745098039215, "grad_norm": 7.309041976928711, "learning_rate": 1.6069228246205713e-08, "loss": 0.1447, "step": 94160 }, { "epoch": 1.846470588235294, "grad_norm": 3.167119026184082, "learning_rate": 1.6028486950389276e-08, "loss": 0.1312, "step": 94170 }, { "epoch": 1.8466666666666667, "grad_norm": 3.941758871078491, "learning_rate": 1.598779652577348e-08, "loss": 0.1717, "step": 94180 }, { "epoch": 1.846862745098039, "grad_norm": 3.9888107776641846, "learning_rate": 1.5947156976635236e-08, "loss": 0.1717, "step": 94190 }, { "epoch": 1.8470588235294119, "grad_norm": 2.078476667404175, "learning_rate": 1.590656830724624e-08, "loss": 0.1716, "step": 94200 }, { "epoch": 1.8472549019607842, "grad_norm": 1.9366987943649292, "learning_rate": 1.5866030521872964e-08, "loss": 0.1992, "step": 94210 }, { "epoch": 1.8474509803921568, "grad_norm": 4.9565935134887695, "learning_rate": 1.5825543624776338e-08, "loss": 0.1551, "step": 94220 }, { "epoch": 1.8476470588235294, "grad_norm": 3.8525943756103516, "learning_rate": 1.5785107620211956e-08, "loss": 0.154, "step": 94230 }, { "epoch": 1.847843137254902, "grad_norm": 4.6634416580200195, "learning_rate": 1.5744722512430254e-08, "loss": 0.1616, "step": 94240 }, { "epoch": 1.8480392156862746, "grad_norm": 1.8345588445663452, "learning_rate": 1.5704388305676166e-08, "loss": 0.156, "step": 94250 }, { "epoch": 1.848235294117647, "grad_norm": 3.8785042762756348, "learning_rate": 1.5664105004189188e-08, "loss": 0.1699, "step": 94260 }, { "epoch": 1.8484313725490196, "grad_norm": 3.8435003757476807, "learning_rate": 1.562387261220377e-08, "loss": 0.1973, "step": 94270 }, { "epoch": 1.8486274509803922, "grad_norm": 3.6961779594421387, "learning_rate": 1.558369113394864e-08, "loss": 0.1896, "step": 94280 }, { "epoch": 1.8488235294117648, "grad_norm": 4.3440423011779785, "learning_rate": 1.554356057364736e-08, "loss": 0.1762, "step": 94290 }, { "epoch": 1.8490196078431373, "grad_norm": 2.966097354888916, "learning_rate": 1.5503480935518288e-08, "loss": 0.1822, "step": 94300 }, { "epoch": 1.8492156862745097, "grad_norm": 2.8534023761749268, "learning_rate": 1.546345222377421e-08, "loss": 0.1897, "step": 94310 }, { "epoch": 1.8494117647058823, "grad_norm": 2.714473009109497, "learning_rate": 1.5423474442622487e-08, "loss": 0.1514, "step": 94320 }, { "epoch": 1.849607843137255, "grad_norm": 3.635420560836792, "learning_rate": 1.5383547596265368e-08, "loss": 0.1455, "step": 94330 }, { "epoch": 1.8498039215686275, "grad_norm": 4.046818256378174, "learning_rate": 1.534367168889966e-08, "loss": 0.174, "step": 94340 }, { "epoch": 1.85, "grad_norm": 3.31522274017334, "learning_rate": 1.530384672471685e-08, "loss": 0.1493, "step": 94350 }, { "epoch": 1.8501960784313725, "grad_norm": 2.079287052154541, "learning_rate": 1.526407270790281e-08, "loss": 0.1544, "step": 94360 }, { "epoch": 1.850392156862745, "grad_norm": 2.583326816558838, "learning_rate": 1.5224349642638524e-08, "loss": 0.1541, "step": 94370 }, { "epoch": 1.8505882352941176, "grad_norm": 2.5440874099731445, "learning_rate": 1.518467753309921e-08, "loss": 0.174, "step": 94380 }, { "epoch": 1.8507843137254902, "grad_norm": 3.0016255378723145, "learning_rate": 1.5145056383454868e-08, "loss": 0.1623, "step": 94390 }, { "epoch": 1.8509803921568628, "grad_norm": 3.8074254989624023, "learning_rate": 1.5105486197870276e-08, "loss": 0.1753, "step": 94400 }, { "epoch": 1.8511764705882352, "grad_norm": 1.7442172765731812, "learning_rate": 1.5065966980504607e-08, "loss": 0.1562, "step": 94410 }, { "epoch": 1.8513725490196078, "grad_norm": 2.743460178375244, "learning_rate": 1.5026498735511873e-08, "loss": 0.1431, "step": 94420 }, { "epoch": 1.8515686274509804, "grad_norm": 1.9323476552963257, "learning_rate": 1.498708146704064e-08, "loss": 0.1762, "step": 94430 }, { "epoch": 1.851764705882353, "grad_norm": 6.1632609367370605, "learning_rate": 1.4947715179234155e-08, "loss": 0.1942, "step": 94440 }, { "epoch": 1.8519607843137256, "grad_norm": 3.611480712890625, "learning_rate": 1.4908399876230214e-08, "loss": 0.1827, "step": 94450 }, { "epoch": 1.852156862745098, "grad_norm": 5.459424018859863, "learning_rate": 1.4869135562161406e-08, "loss": 0.1789, "step": 94460 }, { "epoch": 1.8523529411764705, "grad_norm": 3.5282764434814453, "learning_rate": 1.4829922241154869e-08, "loss": 0.1262, "step": 94470 }, { "epoch": 1.8525490196078431, "grad_norm": 3.0608110427856445, "learning_rate": 1.4790759917332307e-08, "loss": 0.1562, "step": 94480 }, { "epoch": 1.8527450980392157, "grad_norm": 4.17458963394165, "learning_rate": 1.4751648594810318e-08, "loss": 0.1891, "step": 94490 }, { "epoch": 1.8529411764705883, "grad_norm": 3.0577707290649414, "learning_rate": 1.4712588277699778e-08, "loss": 0.1834, "step": 94500 }, { "epoch": 1.8531372549019607, "grad_norm": 3.7238497734069824, "learning_rate": 1.467357897010657e-08, "loss": 0.1654, "step": 94510 }, { "epoch": 1.8533333333333335, "grad_norm": 4.215824604034424, "learning_rate": 1.463462067613086e-08, "loss": 0.1393, "step": 94520 }, { "epoch": 1.8535294117647059, "grad_norm": 2.510939359664917, "learning_rate": 1.4595713399867815e-08, "loss": 0.1443, "step": 94530 }, { "epoch": 1.8537254901960785, "grad_norm": 2.756190538406372, "learning_rate": 1.4556857145406942e-08, "loss": 0.165, "step": 94540 }, { "epoch": 1.853921568627451, "grad_norm": 2.166990041732788, "learning_rate": 1.4518051916832475e-08, "loss": 0.1809, "step": 94550 }, { "epoch": 1.8541176470588234, "grad_norm": 2.57747220993042, "learning_rate": 1.4479297718223427e-08, "loss": 0.1775, "step": 94560 }, { "epoch": 1.8543137254901962, "grad_norm": 2.475611925125122, "learning_rate": 1.4440594553653262e-08, "loss": 0.1481, "step": 94570 }, { "epoch": 1.8545098039215686, "grad_norm": 2.465827226638794, "learning_rate": 1.4401942427190117e-08, "loss": 0.1478, "step": 94580 }, { "epoch": 1.8547058823529412, "grad_norm": 6.110325813293457, "learning_rate": 1.4363341342896795e-08, "loss": 0.1555, "step": 94590 }, { "epoch": 1.8549019607843138, "grad_norm": 2.7788760662078857, "learning_rate": 1.432479130483083e-08, "loss": 0.172, "step": 94600 }, { "epoch": 1.8550980392156862, "grad_norm": 4.374547958374023, "learning_rate": 1.4286292317044146e-08, "loss": 0.1818, "step": 94610 }, { "epoch": 1.855294117647059, "grad_norm": 2.454836368560791, "learning_rate": 1.4247844383583563e-08, "loss": 0.1811, "step": 94620 }, { "epoch": 1.8554901960784314, "grad_norm": 2.3931243419647217, "learning_rate": 1.4209447508490346e-08, "loss": 0.1599, "step": 94630 }, { "epoch": 1.855686274509804, "grad_norm": 5.0294694900512695, "learning_rate": 1.4171101695800491e-08, "loss": 0.1657, "step": 94640 }, { "epoch": 1.8558823529411765, "grad_norm": 3.6917877197265625, "learning_rate": 1.4132806949544662e-08, "loss": 0.1669, "step": 94650 }, { "epoch": 1.856078431372549, "grad_norm": 4.291083335876465, "learning_rate": 1.4094563273748028e-08, "loss": 0.1641, "step": 94660 }, { "epoch": 1.8562745098039217, "grad_norm": 4.357166767120361, "learning_rate": 1.4056370672430428e-08, "loss": 0.1711, "step": 94670 }, { "epoch": 1.856470588235294, "grad_norm": 3.363449811935425, "learning_rate": 1.4018229149606376e-08, "loss": 0.1596, "step": 94680 }, { "epoch": 1.8566666666666667, "grad_norm": 5.628708839416504, "learning_rate": 1.3980138709285105e-08, "loss": 0.1613, "step": 94690 }, { "epoch": 1.8568627450980393, "grad_norm": 5.962939739227295, "learning_rate": 1.3942099355470305e-08, "loss": 0.1681, "step": 94700 }, { "epoch": 1.8570588235294117, "grad_norm": 3.160982847213745, "learning_rate": 1.3904111092160276e-08, "loss": 0.1936, "step": 94710 }, { "epoch": 1.8572549019607845, "grad_norm": 2.944708824157715, "learning_rate": 1.3866173923348212e-08, "loss": 0.1923, "step": 94720 }, { "epoch": 1.8574509803921568, "grad_norm": 2.64400315284729, "learning_rate": 1.3828287853021648e-08, "loss": 0.1322, "step": 94730 }, { "epoch": 1.8576470588235294, "grad_norm": 3.291283369064331, "learning_rate": 1.3790452885162895e-08, "loss": 0.1508, "step": 94740 }, { "epoch": 1.857843137254902, "grad_norm": 2.0802390575408936, "learning_rate": 1.3752669023748886e-08, "loss": 0.181, "step": 94750 }, { "epoch": 1.8580392156862744, "grad_norm": 2.243785858154297, "learning_rate": 1.3714936272751165e-08, "loss": 0.1885, "step": 94760 }, { "epoch": 1.8582352941176472, "grad_norm": 3.6829590797424316, "learning_rate": 1.3677254636135783e-08, "loss": 0.1482, "step": 94770 }, { "epoch": 1.8584313725490196, "grad_norm": 3.1253511905670166, "learning_rate": 1.3639624117863625e-08, "loss": 0.1876, "step": 94780 }, { "epoch": 1.8586274509803922, "grad_norm": 1.6762951612472534, "learning_rate": 1.3602044721890138e-08, "loss": 0.1622, "step": 94790 }, { "epoch": 1.8588235294117648, "grad_norm": 2.3924715518951416, "learning_rate": 1.3564516452165276e-08, "loss": 0.1925, "step": 94800 }, { "epoch": 1.8590196078431371, "grad_norm": 3.2735841274261475, "learning_rate": 1.3527039312633826e-08, "loss": 0.1665, "step": 94810 }, { "epoch": 1.85921568627451, "grad_norm": 3.796565055847168, "learning_rate": 1.3489613307234971e-08, "loss": 0.1722, "step": 94820 }, { "epoch": 1.8594117647058823, "grad_norm": 4.422581672668457, "learning_rate": 1.3452238439902619e-08, "loss": 0.1664, "step": 94830 }, { "epoch": 1.859607843137255, "grad_norm": 7.691451549530029, "learning_rate": 1.3414914714565462e-08, "loss": 0.1858, "step": 94840 }, { "epoch": 1.8598039215686275, "grad_norm": 4.174158096313477, "learning_rate": 1.3377642135146528e-08, "loss": 0.1782, "step": 94850 }, { "epoch": 1.8599999999999999, "grad_norm": 2.0063772201538086, "learning_rate": 1.3340420705563737e-08, "loss": 0.1617, "step": 94860 }, { "epoch": 1.8601960784313727, "grad_norm": 5.160975933074951, "learning_rate": 1.3303250429729351e-08, "loss": 0.1561, "step": 94870 }, { "epoch": 1.860392156862745, "grad_norm": 3.4423625469207764, "learning_rate": 1.3266131311550522e-08, "loss": 0.1449, "step": 94880 }, { "epoch": 1.8605882352941177, "grad_norm": 2.8039770126342773, "learning_rate": 1.322906335492896e-08, "loss": 0.1782, "step": 94890 }, { "epoch": 1.8607843137254902, "grad_norm": 4.206528663635254, "learning_rate": 1.3192046563760772e-08, "loss": 0.1669, "step": 94900 }, { "epoch": 1.8609803921568626, "grad_norm": 3.8079113960266113, "learning_rate": 1.3155080941937013e-08, "loss": 0.2108, "step": 94910 }, { "epoch": 1.8611764705882354, "grad_norm": 2.9253971576690674, "learning_rate": 1.3118166493343241e-08, "loss": 0.1596, "step": 94920 }, { "epoch": 1.8613725490196078, "grad_norm": 11.609301567077637, "learning_rate": 1.3081303221859408e-08, "loss": 0.183, "step": 94930 }, { "epoch": 1.8615686274509804, "grad_norm": 1.2139862775802612, "learning_rate": 1.3044491131360525e-08, "loss": 0.1672, "step": 94940 }, { "epoch": 1.861764705882353, "grad_norm": 2.2780022621154785, "learning_rate": 1.3007730225715829e-08, "loss": 0.1574, "step": 94950 }, { "epoch": 1.8619607843137254, "grad_norm": 3.4108128547668457, "learning_rate": 1.2971020508789344e-08, "loss": 0.2102, "step": 94960 }, { "epoch": 1.8621568627450982, "grad_norm": 3.6700103282928467, "learning_rate": 1.2934361984439813e-08, "loss": 0.1605, "step": 94970 }, { "epoch": 1.8623529411764705, "grad_norm": 2.687793016433716, "learning_rate": 1.2897754656520376e-08, "loss": 0.1526, "step": 94980 }, { "epoch": 1.8625490196078431, "grad_norm": 2.8003270626068115, "learning_rate": 1.2861198528878848e-08, "loss": 0.1481, "step": 94990 }, { "epoch": 1.8627450980392157, "grad_norm": 3.950212240219116, "learning_rate": 1.2824693605357873e-08, "loss": 0.1569, "step": 95000 }, { "epoch": 1.862941176470588, "grad_norm": 1.8666270971298218, "learning_rate": 1.2788239889794494e-08, "loss": 0.1801, "step": 95010 }, { "epoch": 1.863137254901961, "grad_norm": 3.0910327434539795, "learning_rate": 1.275183738602037e-08, "loss": 0.1429, "step": 95020 }, { "epoch": 1.8633333333333333, "grad_norm": 4.943366050720215, "learning_rate": 1.2715486097861883e-08, "loss": 0.1871, "step": 95030 }, { "epoch": 1.8635294117647059, "grad_norm": 2.0795435905456543, "learning_rate": 1.2679186029140031e-08, "loss": 0.1617, "step": 95040 }, { "epoch": 1.8637254901960785, "grad_norm": 4.4141716957092285, "learning_rate": 1.2642937183670377e-08, "loss": 0.1564, "step": 95050 }, { "epoch": 1.8639215686274508, "grad_norm": 5.027094841003418, "learning_rate": 1.2606739565263036e-08, "loss": 0.1612, "step": 95060 }, { "epoch": 1.8641176470588237, "grad_norm": 2.4957399368286133, "learning_rate": 1.2570593177722855e-08, "loss": 0.2124, "step": 95070 }, { "epoch": 1.864313725490196, "grad_norm": 4.96834135055542, "learning_rate": 1.2534498024849293e-08, "loss": 0.1702, "step": 95080 }, { "epoch": 1.8645098039215686, "grad_norm": 4.535289764404297, "learning_rate": 1.2498454110436319e-08, "loss": 0.155, "step": 95090 }, { "epoch": 1.8647058823529412, "grad_norm": 3.428528070449829, "learning_rate": 1.246246143827262e-08, "loss": 0.1623, "step": 95100 }, { "epoch": 1.8649019607843136, "grad_norm": 1.5540807247161865, "learning_rate": 1.242652001214145e-08, "loss": 0.1538, "step": 95110 }, { "epoch": 1.8650980392156864, "grad_norm": 4.127627372741699, "learning_rate": 1.239062983582062e-08, "loss": 0.1485, "step": 95120 }, { "epoch": 1.8652941176470588, "grad_norm": 4.767012119293213, "learning_rate": 1.2354790913082725e-08, "loss": 0.1574, "step": 95130 }, { "epoch": 1.8654901960784314, "grad_norm": 11.9371919631958, "learning_rate": 1.2319003247694804e-08, "loss": 0.1692, "step": 95140 }, { "epoch": 1.865686274509804, "grad_norm": 5.098299026489258, "learning_rate": 1.2283266843418516e-08, "loss": 0.1702, "step": 95150 }, { "epoch": 1.8658823529411763, "grad_norm": 3.3960816860198975, "learning_rate": 1.2247581704010301e-08, "loss": 0.1836, "step": 95160 }, { "epoch": 1.8660784313725491, "grad_norm": 1.9781080484390259, "learning_rate": 1.2211947833221048e-08, "loss": 0.187, "step": 95170 }, { "epoch": 1.8662745098039215, "grad_norm": 4.979678630828857, "learning_rate": 1.2176365234796259e-08, "loss": 0.1619, "step": 95180 }, { "epoch": 1.866470588235294, "grad_norm": 2.7016539573669434, "learning_rate": 1.2140833912476168e-08, "loss": 0.1752, "step": 95190 }, { "epoch": 1.8666666666666667, "grad_norm": 10.765118598937988, "learning_rate": 1.2105353869995504e-08, "loss": 0.1608, "step": 95200 }, { "epoch": 1.866862745098039, "grad_norm": 3.005810499191284, "learning_rate": 1.2069925111083568e-08, "loss": 0.1367, "step": 95210 }, { "epoch": 1.8670588235294119, "grad_norm": 2.706528663635254, "learning_rate": 1.2034547639464433e-08, "loss": 0.1526, "step": 95220 }, { "epoch": 1.8672549019607843, "grad_norm": 2.362985372543335, "learning_rate": 1.1999221458856734e-08, "loss": 0.1413, "step": 95230 }, { "epoch": 1.8674509803921568, "grad_norm": 2.92110538482666, "learning_rate": 1.1963946572973672e-08, "loss": 0.1565, "step": 95240 }, { "epoch": 1.8676470588235294, "grad_norm": 7.428976058959961, "learning_rate": 1.1928722985522888e-08, "loss": 0.1729, "step": 95250 }, { "epoch": 1.8678431372549018, "grad_norm": 3.300654172897339, "learning_rate": 1.1893550700206978e-08, "loss": 0.2165, "step": 95260 }, { "epoch": 1.8680392156862746, "grad_norm": 4.051749229431152, "learning_rate": 1.1858429720722984e-08, "loss": 0.1627, "step": 95270 }, { "epoch": 1.868235294117647, "grad_norm": 2.038412570953369, "learning_rate": 1.1823360050762344e-08, "loss": 0.1634, "step": 95280 }, { "epoch": 1.8684313725490196, "grad_norm": 4.248924255371094, "learning_rate": 1.178834169401155e-08, "loss": 0.1864, "step": 95290 }, { "epoch": 1.8686274509803922, "grad_norm": 3.2019309997558594, "learning_rate": 1.1753374654151327e-08, "loss": 0.1787, "step": 95300 }, { "epoch": 1.8688235294117646, "grad_norm": 1.8557815551757812, "learning_rate": 1.1718458934857067e-08, "loss": 0.1474, "step": 95310 }, { "epoch": 1.8690196078431374, "grad_norm": 3.0665249824523926, "learning_rate": 1.168359453979889e-08, "loss": 0.1673, "step": 95320 }, { "epoch": 1.8692156862745097, "grad_norm": 4.17194938659668, "learning_rate": 1.1648781472641533e-08, "loss": 0.1507, "step": 95330 }, { "epoch": 1.8694117647058823, "grad_norm": 2.967473030090332, "learning_rate": 1.1614019737044179e-08, "loss": 0.1731, "step": 95340 }, { "epoch": 1.869607843137255, "grad_norm": 4.237368106842041, "learning_rate": 1.1579309336660736e-08, "loss": 0.1452, "step": 95350 }, { "epoch": 1.8698039215686273, "grad_norm": 3.7859926223754883, "learning_rate": 1.1544650275139678e-08, "loss": 0.1774, "step": 95360 }, { "epoch": 1.87, "grad_norm": 2.0808863639831543, "learning_rate": 1.1510042556124033e-08, "loss": 0.1683, "step": 95370 }, { "epoch": 1.8701960784313725, "grad_norm": 1.8998138904571533, "learning_rate": 1.147548618325156e-08, "loss": 0.17, "step": 95380 }, { "epoch": 1.870392156862745, "grad_norm": 2.6103482246398926, "learning_rate": 1.1440981160154517e-08, "loss": 0.1608, "step": 95390 }, { "epoch": 1.8705882352941177, "grad_norm": 5.777167320251465, "learning_rate": 1.1406527490459839e-08, "loss": 0.1807, "step": 95400 }, { "epoch": 1.87078431372549, "grad_norm": 5.507325172424316, "learning_rate": 1.1372125177788905e-08, "loss": 0.1634, "step": 95410 }, { "epoch": 1.8709803921568628, "grad_norm": 3.1138973236083984, "learning_rate": 1.133777422575799e-08, "loss": 0.1488, "step": 95420 }, { "epoch": 1.8711764705882352, "grad_norm": 3.9041860103607178, "learning_rate": 1.130347463797765e-08, "loss": 0.1785, "step": 95430 }, { "epoch": 1.8713725490196078, "grad_norm": 2.5902416706085205, "learning_rate": 1.1269226418053168e-08, "loss": 0.178, "step": 95440 }, { "epoch": 1.8715686274509804, "grad_norm": 2.4944145679473877, "learning_rate": 1.1235029569584553e-08, "loss": 0.1618, "step": 95450 }, { "epoch": 1.8717647058823528, "grad_norm": 2.4671666622161865, "learning_rate": 1.1200884096166263e-08, "loss": 0.1443, "step": 95460 }, { "epoch": 1.8719607843137256, "grad_norm": 3.2039949893951416, "learning_rate": 1.1166790001387317e-08, "loss": 0.1681, "step": 95470 }, { "epoch": 1.872156862745098, "grad_norm": 3.010450601577759, "learning_rate": 1.1132747288831512e-08, "loss": 0.1885, "step": 95480 }, { "epoch": 1.8723529411764706, "grad_norm": 4.002710342407227, "learning_rate": 1.10987559620771e-08, "loss": 0.1541, "step": 95490 }, { "epoch": 1.8725490196078431, "grad_norm": 3.6678273677825928, "learning_rate": 1.1064816024696944e-08, "loss": 0.1637, "step": 95500 }, { "epoch": 1.8727450980392157, "grad_norm": 2.596648931503296, "learning_rate": 1.1030927480258578e-08, "loss": 0.1753, "step": 95510 }, { "epoch": 1.8729411764705883, "grad_norm": 3.1792562007904053, "learning_rate": 1.0997090332324155e-08, "loss": 0.1737, "step": 95520 }, { "epoch": 1.8731372549019607, "grad_norm": 4.330287456512451, "learning_rate": 1.0963304584450217e-08, "loss": 0.189, "step": 95530 }, { "epoch": 1.8733333333333333, "grad_norm": 2.311363935470581, "learning_rate": 1.0929570240188202e-08, "loss": 0.1449, "step": 95540 }, { "epoch": 1.8735294117647059, "grad_norm": 3.1446821689605713, "learning_rate": 1.0895887303083884e-08, "loss": 0.1702, "step": 95550 }, { "epoch": 1.8737254901960785, "grad_norm": 2.1856069564819336, "learning_rate": 1.086225577667782e-08, "loss": 0.1796, "step": 95560 }, { "epoch": 1.873921568627451, "grad_norm": 5.384634971618652, "learning_rate": 1.082867566450496e-08, "loss": 0.1734, "step": 95570 }, { "epoch": 1.8741176470588234, "grad_norm": 3.2999203205108643, "learning_rate": 1.0795146970095148e-08, "loss": 0.1405, "step": 95580 }, { "epoch": 1.874313725490196, "grad_norm": 3.052093505859375, "learning_rate": 1.0761669696972563e-08, "loss": 0.1999, "step": 95590 }, { "epoch": 1.8745098039215686, "grad_norm": 2.2526590824127197, "learning_rate": 1.072824384865606e-08, "loss": 0.1766, "step": 95600 }, { "epoch": 1.8747058823529412, "grad_norm": 2.209197521209717, "learning_rate": 1.0694869428659104e-08, "loss": 0.1689, "step": 95610 }, { "epoch": 1.8749019607843138, "grad_norm": 2.6817378997802734, "learning_rate": 1.0661546440489777e-08, "loss": 0.1703, "step": 95620 }, { "epoch": 1.8750980392156862, "grad_norm": 4.155731678009033, "learning_rate": 1.0628274887650667e-08, "loss": 0.1751, "step": 95630 }, { "epoch": 1.8752941176470588, "grad_norm": 3.97983717918396, "learning_rate": 1.059505477363909e-08, "loss": 0.1475, "step": 95640 }, { "epoch": 1.8754901960784314, "grad_norm": 4.25549840927124, "learning_rate": 1.0561886101946805e-08, "loss": 0.1778, "step": 95650 }, { "epoch": 1.875686274509804, "grad_norm": 2.7060439586639404, "learning_rate": 1.0528768876060245e-08, "loss": 0.1692, "step": 95660 }, { "epoch": 1.8758823529411766, "grad_norm": 4.474353790283203, "learning_rate": 1.0495703099460518e-08, "loss": 0.1809, "step": 95670 }, { "epoch": 1.876078431372549, "grad_norm": 4.851702690124512, "learning_rate": 1.0462688775623119e-08, "loss": 0.1801, "step": 95680 }, { "epoch": 1.8762745098039215, "grad_norm": 3.467587471008301, "learning_rate": 1.0429725908018216e-08, "loss": 0.1786, "step": 95690 }, { "epoch": 1.8764705882352941, "grad_norm": 3.4767866134643555, "learning_rate": 1.0396814500110762e-08, "loss": 0.1573, "step": 95700 }, { "epoch": 1.8766666666666667, "grad_norm": 4.723526954650879, "learning_rate": 1.0363954555360043e-08, "loss": 0.2038, "step": 95710 }, { "epoch": 1.8768627450980393, "grad_norm": 2.52561092376709, "learning_rate": 1.033114607722002e-08, "loss": 0.1432, "step": 95720 }, { "epoch": 1.8770588235294117, "grad_norm": 2.455678701400757, "learning_rate": 1.0298389069139268e-08, "loss": 0.1823, "step": 95730 }, { "epoch": 1.8772549019607843, "grad_norm": 2.7887349128723145, "learning_rate": 1.0265683534561031e-08, "loss": 0.1706, "step": 95740 }, { "epoch": 1.8774509803921569, "grad_norm": 2.702260732650757, "learning_rate": 1.0233029476922838e-08, "loss": 0.1615, "step": 95750 }, { "epoch": 1.8776470588235294, "grad_norm": 5.094602584838867, "learning_rate": 1.0200426899657222e-08, "loss": 0.168, "step": 95760 }, { "epoch": 1.877843137254902, "grad_norm": 3.9772698879241943, "learning_rate": 1.0167875806191107e-08, "loss": 0.1552, "step": 95770 }, { "epoch": 1.8780392156862744, "grad_norm": 2.697794198989868, "learning_rate": 1.0135376199945866e-08, "loss": 0.1665, "step": 95780 }, { "epoch": 1.8782352941176472, "grad_norm": 2.776536464691162, "learning_rate": 1.0102928084337659e-08, "loss": 0.1621, "step": 95790 }, { "epoch": 1.8784313725490196, "grad_norm": 4.764627933502197, "learning_rate": 1.00705314627772e-08, "loss": 0.1745, "step": 95800 }, { "epoch": 1.8786274509803922, "grad_norm": 4.80532169342041, "learning_rate": 1.0038186338669762e-08, "loss": 0.1438, "step": 95810 }, { "epoch": 1.8788235294117648, "grad_norm": 4.791679859161377, "learning_rate": 1.0005892715415132e-08, "loss": 0.1633, "step": 95820 }, { "epoch": 1.8790196078431372, "grad_norm": 1.8927056789398193, "learning_rate": 9.973650596407867e-09, "loss": 0.1419, "step": 95830 }, { "epoch": 1.87921568627451, "grad_norm": 3.0216786861419678, "learning_rate": 9.941459985036927e-09, "loss": 0.1452, "step": 95840 }, { "epoch": 1.8794117647058823, "grad_norm": 2.6957638263702393, "learning_rate": 9.909320884685879e-09, "loss": 0.1543, "step": 95850 }, { "epoch": 1.879607843137255, "grad_norm": 5.374926567077637, "learning_rate": 9.877233298733078e-09, "loss": 0.147, "step": 95860 }, { "epoch": 1.8798039215686275, "grad_norm": 2.2076961994171143, "learning_rate": 9.845197230551161e-09, "loss": 0.1523, "step": 95870 }, { "epoch": 1.88, "grad_norm": 3.1601452827453613, "learning_rate": 9.813212683507598e-09, "loss": 0.1575, "step": 95880 }, { "epoch": 1.8801960784313727, "grad_norm": 4.195461273193359, "learning_rate": 9.781279660964314e-09, "loss": 0.1639, "step": 95890 }, { "epoch": 1.880392156862745, "grad_norm": 3.79569411277771, "learning_rate": 9.749398166277901e-09, "loss": 0.2017, "step": 95900 }, { "epoch": 1.8805882352941177, "grad_norm": 2.0467236042022705, "learning_rate": 9.717568202799397e-09, "loss": 0.1558, "step": 95910 }, { "epoch": 1.8807843137254903, "grad_norm": 5.5135297775268555, "learning_rate": 9.68578977387452e-09, "loss": 0.1651, "step": 95920 }, { "epoch": 1.8809803921568626, "grad_norm": 3.1249473094940186, "learning_rate": 9.654062882843594e-09, "loss": 0.1762, "step": 95930 }, { "epoch": 1.8811764705882354, "grad_norm": 2.9273111820220947, "learning_rate": 9.622387533041565e-09, "loss": 0.1565, "step": 95940 }, { "epoch": 1.8813725490196078, "grad_norm": 3.5781922340393066, "learning_rate": 9.590763727797823e-09, "loss": 0.1614, "step": 95950 }, { "epoch": 1.8815686274509804, "grad_norm": 8.498175621032715, "learning_rate": 9.55919147043638e-09, "loss": 0.1675, "step": 95960 }, { "epoch": 1.881764705882353, "grad_norm": 2.87109375, "learning_rate": 9.527670764275964e-09, "loss": 0.1682, "step": 95970 }, { "epoch": 1.8819607843137254, "grad_norm": 1.7195956707000732, "learning_rate": 9.496201612629595e-09, "loss": 0.1374, "step": 95980 }, { "epoch": 1.8821568627450982, "grad_norm": 1.7528210878372192, "learning_rate": 9.464784018805238e-09, "loss": 0.1953, "step": 95990 }, { "epoch": 1.8823529411764706, "grad_norm": 2.2058703899383545, "learning_rate": 9.433417986105197e-09, "loss": 0.1645, "step": 96000 }, { "epoch": 1.8825490196078432, "grad_norm": 3.189485788345337, "learning_rate": 9.402103517826388e-09, "loss": 0.1721, "step": 96010 }, { "epoch": 1.8827450980392157, "grad_norm": 3.8839492797851562, "learning_rate": 9.370840617260345e-09, "loss": 0.1509, "step": 96020 }, { "epoch": 1.8829411764705881, "grad_norm": 3.9173014163970947, "learning_rate": 9.339629287693217e-09, "loss": 0.1614, "step": 96030 }, { "epoch": 1.883137254901961, "grad_norm": 3.6700878143310547, "learning_rate": 9.308469532405605e-09, "loss": 0.1863, "step": 96040 }, { "epoch": 1.8833333333333333, "grad_norm": 4.528023719787598, "learning_rate": 9.277361354672885e-09, "loss": 0.1428, "step": 96050 }, { "epoch": 1.883529411764706, "grad_norm": 3.142563819885254, "learning_rate": 9.246304757764833e-09, "loss": 0.1995, "step": 96060 }, { "epoch": 1.8837254901960785, "grad_norm": 6.572561740875244, "learning_rate": 9.215299744945838e-09, "loss": 0.1547, "step": 96070 }, { "epoch": 1.8839215686274509, "grad_norm": 2.4878008365631104, "learning_rate": 9.184346319474956e-09, "loss": 0.1734, "step": 96080 }, { "epoch": 1.8841176470588237, "grad_norm": 3.341557502746582, "learning_rate": 9.153444484605754e-09, "loss": 0.1823, "step": 96090 }, { "epoch": 1.884313725490196, "grad_norm": 3.2290940284729004, "learning_rate": 9.122594243586356e-09, "loss": 0.1883, "step": 96100 }, { "epoch": 1.8845098039215686, "grad_norm": 2.3283698558807373, "learning_rate": 9.091795599659447e-09, "loss": 0.1635, "step": 96110 }, { "epoch": 1.8847058823529412, "grad_norm": 2.239290237426758, "learning_rate": 9.06104855606249e-09, "loss": 0.1705, "step": 96120 }, { "epoch": 1.8849019607843136, "grad_norm": 2.2274746894836426, "learning_rate": 9.030353116027234e-09, "loss": 0.1606, "step": 96130 }, { "epoch": 1.8850980392156864, "grad_norm": 4.92426872253418, "learning_rate": 8.999709282780154e-09, "loss": 0.1755, "step": 96140 }, { "epoch": 1.8852941176470588, "grad_norm": 3.620441436767578, "learning_rate": 8.969117059542342e-09, "loss": 0.1788, "step": 96150 }, { "epoch": 1.8854901960784314, "grad_norm": 4.358010768890381, "learning_rate": 8.938576449529388e-09, "loss": 0.1641, "step": 96160 }, { "epoch": 1.885686274509804, "grad_norm": 2.7375001907348633, "learning_rate": 8.908087455951396e-09, "loss": 0.158, "step": 96170 }, { "epoch": 1.8858823529411763, "grad_norm": 5.430164813995361, "learning_rate": 8.877650082013243e-09, "loss": 0.1974, "step": 96180 }, { "epoch": 1.8860784313725492, "grad_norm": 2.838944673538208, "learning_rate": 8.847264330914206e-09, "loss": 0.1697, "step": 96190 }, { "epoch": 1.8862745098039215, "grad_norm": 5.9820990562438965, "learning_rate": 8.816930205848117e-09, "loss": 0.1447, "step": 96200 }, { "epoch": 1.8864705882352941, "grad_norm": 3.4121642112731934, "learning_rate": 8.786647710003591e-09, "loss": 0.1602, "step": 96210 }, { "epoch": 1.8866666666666667, "grad_norm": 1.5386650562286377, "learning_rate": 8.75641684656364e-09, "loss": 0.1734, "step": 96220 }, { "epoch": 1.886862745098039, "grad_norm": 5.218724250793457, "learning_rate": 8.726237618705834e-09, "loss": 0.1761, "step": 96230 }, { "epoch": 1.887058823529412, "grad_norm": 1.968654990196228, "learning_rate": 8.696110029602411e-09, "loss": 0.1791, "step": 96240 }, { "epoch": 1.8872549019607843, "grad_norm": 5.521032333374023, "learning_rate": 8.666034082420115e-09, "loss": 0.1582, "step": 96250 }, { "epoch": 1.8874509803921569, "grad_norm": 1.4827532768249512, "learning_rate": 8.636009780320364e-09, "loss": 0.1527, "step": 96260 }, { "epoch": 1.8876470588235295, "grad_norm": 3.7956702709198, "learning_rate": 8.60603712645891e-09, "loss": 0.1792, "step": 96270 }, { "epoch": 1.8878431372549018, "grad_norm": 4.415364742279053, "learning_rate": 8.576116123986399e-09, "loss": 0.2008, "step": 96280 }, { "epoch": 1.8880392156862746, "grad_norm": 2.297551393508911, "learning_rate": 8.546246776047872e-09, "loss": 0.1485, "step": 96290 }, { "epoch": 1.888235294117647, "grad_norm": 4.528631210327148, "learning_rate": 8.516429085782818e-09, "loss": 0.213, "step": 96300 }, { "epoch": 1.8884313725490196, "grad_norm": 2.3690290451049805, "learning_rate": 8.486663056325615e-09, "loss": 0.135, "step": 96310 }, { "epoch": 1.8886274509803922, "grad_norm": 1.606754183769226, "learning_rate": 8.456948690804932e-09, "loss": 0.1514, "step": 96320 }, { "epoch": 1.8888235294117646, "grad_norm": 3.5151782035827637, "learning_rate": 8.427285992344103e-09, "loss": 0.1892, "step": 96330 }, { "epoch": 1.8890196078431374, "grad_norm": 4.707324028015137, "learning_rate": 8.397674964061074e-09, "loss": 0.1574, "step": 96340 }, { "epoch": 1.8892156862745098, "grad_norm": 3.8755500316619873, "learning_rate": 8.368115609068248e-09, "loss": 0.1724, "step": 96350 }, { "epoch": 1.8894117647058823, "grad_norm": 3.051909923553467, "learning_rate": 8.338607930472697e-09, "loss": 0.1741, "step": 96360 }, { "epoch": 1.889607843137255, "grad_norm": 4.9777116775512695, "learning_rate": 8.309151931376101e-09, "loss": 0.2159, "step": 96370 }, { "epoch": 1.8898039215686273, "grad_norm": 4.6760406494140625, "learning_rate": 8.2797476148746e-09, "loss": 0.145, "step": 96380 }, { "epoch": 1.8900000000000001, "grad_norm": 2.0587525367736816, "learning_rate": 8.250394984058828e-09, "loss": 0.1614, "step": 96390 }, { "epoch": 1.8901960784313725, "grad_norm": 4.286766529083252, "learning_rate": 8.221094042014265e-09, "loss": 0.1672, "step": 96400 }, { "epoch": 1.890392156862745, "grad_norm": 2.1829490661621094, "learning_rate": 8.191844791820724e-09, "loss": 0.1676, "step": 96410 }, { "epoch": 1.8905882352941177, "grad_norm": 1.3427704572677612, "learning_rate": 8.162647236552577e-09, "loss": 0.139, "step": 96420 }, { "epoch": 1.89078431372549, "grad_norm": 2.6500492095947266, "learning_rate": 8.133501379278929e-09, "loss": 0.167, "step": 96430 }, { "epoch": 1.8909803921568629, "grad_norm": 3.3085179328918457, "learning_rate": 8.10440722306338e-09, "loss": 0.1249, "step": 96440 }, { "epoch": 1.8911764705882352, "grad_norm": 2.8384194374084473, "learning_rate": 8.075364770963988e-09, "loss": 0.1662, "step": 96450 }, { "epoch": 1.8913725490196078, "grad_norm": 3.8116116523742676, "learning_rate": 8.046374026033476e-09, "loss": 0.1344, "step": 96460 }, { "epoch": 1.8915686274509804, "grad_norm": 3.1871683597564697, "learning_rate": 8.017434991319182e-09, "loss": 0.1626, "step": 96470 }, { "epoch": 1.8917647058823528, "grad_norm": 4.597091197967529, "learning_rate": 7.9885476698629e-09, "loss": 0.16, "step": 96480 }, { "epoch": 1.8919607843137256, "grad_norm": 3.8825552463531494, "learning_rate": 7.95971206470103e-09, "loss": 0.1663, "step": 96490 }, { "epoch": 1.892156862745098, "grad_norm": 3.2079875469207764, "learning_rate": 7.930928178864593e-09, "loss": 0.1468, "step": 96500 }, { "epoch": 1.8923529411764706, "grad_norm": 4.870492458343506, "learning_rate": 7.90219601537906e-09, "loss": 0.1871, "step": 96510 }, { "epoch": 1.8925490196078432, "grad_norm": 1.636338710784912, "learning_rate": 7.873515577264512e-09, "loss": 0.1245, "step": 96520 }, { "epoch": 1.8927450980392155, "grad_norm": 3.5748488903045654, "learning_rate": 7.84488686753565e-09, "loss": 0.1759, "step": 96530 }, { "epoch": 1.8929411764705883, "grad_norm": 4.021389961242676, "learning_rate": 7.816309889201678e-09, "loss": 0.1479, "step": 96540 }, { "epoch": 1.8931372549019607, "grad_norm": 3.015880584716797, "learning_rate": 7.787784645266416e-09, "loss": 0.1326, "step": 96550 }, { "epoch": 1.8933333333333333, "grad_norm": 4.514860153198242, "learning_rate": 7.759311138728132e-09, "loss": 0.19, "step": 96560 }, { "epoch": 1.893529411764706, "grad_norm": 4.545989990234375, "learning_rate": 7.730889372579819e-09, "loss": 0.1901, "step": 96570 }, { "epoch": 1.8937254901960783, "grad_norm": 7.574948787689209, "learning_rate": 7.702519349808868e-09, "loss": 0.1624, "step": 96580 }, { "epoch": 1.893921568627451, "grad_norm": 7.4201579093933105, "learning_rate": 7.674201073397335e-09, "loss": 0.1665, "step": 96590 }, { "epoch": 1.8941176470588235, "grad_norm": 5.447049140930176, "learning_rate": 7.645934546321841e-09, "loss": 0.1731, "step": 96600 }, { "epoch": 1.894313725490196, "grad_norm": 3.2013180255889893, "learning_rate": 7.617719771553511e-09, "loss": 0.1677, "step": 96610 }, { "epoch": 1.8945098039215686, "grad_norm": 3.268371105194092, "learning_rate": 7.589556752058024e-09, "loss": 0.1996, "step": 96620 }, { "epoch": 1.894705882352941, "grad_norm": 3.067328929901123, "learning_rate": 7.561445490795737e-09, "loss": 0.1455, "step": 96630 }, { "epoch": 1.8949019607843138, "grad_norm": 2.3241381645202637, "learning_rate": 7.533385990721342e-09, "loss": 0.1803, "step": 96640 }, { "epoch": 1.8950980392156862, "grad_norm": 3.885265827178955, "learning_rate": 7.50537825478431e-09, "loss": 0.1335, "step": 96650 }, { "epoch": 1.8952941176470588, "grad_norm": 2.990527868270874, "learning_rate": 7.47742228592868e-09, "loss": 0.2199, "step": 96660 }, { "epoch": 1.8954901960784314, "grad_norm": 2.977550983428955, "learning_rate": 7.449518087092821e-09, "loss": 0.1669, "step": 96670 }, { "epoch": 1.8956862745098038, "grad_norm": 3.2300825119018555, "learning_rate": 7.421665661209886e-09, "loss": 0.16, "step": 96680 }, { "epoch": 1.8958823529411766, "grad_norm": 1.7541759014129639, "learning_rate": 7.393865011207423e-09, "loss": 0.1416, "step": 96690 }, { "epoch": 1.896078431372549, "grad_norm": 2.8778111934661865, "learning_rate": 7.36611614000765e-09, "loss": 0.146, "step": 96700 }, { "epoch": 1.8962745098039215, "grad_norm": 3.40997314453125, "learning_rate": 7.338419050527289e-09, "loss": 0.1618, "step": 96710 }, { "epoch": 1.8964705882352941, "grad_norm": 2.9164130687713623, "learning_rate": 7.310773745677734e-09, "loss": 0.1664, "step": 96720 }, { "epoch": 1.8966666666666665, "grad_norm": 3.1957836151123047, "learning_rate": 7.283180228364716e-09, "loss": 0.1818, "step": 96730 }, { "epoch": 1.8968627450980393, "grad_norm": 2.731229066848755, "learning_rate": 7.255638501488637e-09, "loss": 0.163, "step": 96740 }, { "epoch": 1.8970588235294117, "grad_norm": 3.2894012928009033, "learning_rate": 7.22814856794457e-09, "loss": 0.1377, "step": 96750 }, { "epoch": 1.8972549019607843, "grad_norm": 10.349311828613281, "learning_rate": 7.200710430621926e-09, "loss": 0.1789, "step": 96760 }, { "epoch": 1.8974509803921569, "grad_norm": 3.4425759315490723, "learning_rate": 7.173324092404842e-09, "loss": 0.1627, "step": 96770 }, { "epoch": 1.8976470588235295, "grad_norm": 3.293480634689331, "learning_rate": 7.145989556171961e-09, "loss": 0.1801, "step": 96780 }, { "epoch": 1.897843137254902, "grad_norm": 3.5625901222229004, "learning_rate": 7.118706824796483e-09, "loss": 0.1691, "step": 96790 }, { "epoch": 1.8980392156862744, "grad_norm": 3.3956408500671387, "learning_rate": 7.09147590114606e-09, "loss": 0.2073, "step": 96800 }, { "epoch": 1.898235294117647, "grad_norm": 1.7273836135864258, "learning_rate": 7.0642967880830686e-09, "loss": 0.1596, "step": 96810 }, { "epoch": 1.8984313725490196, "grad_norm": 5.472630500793457, "learning_rate": 7.037169488464279e-09, "loss": 0.1636, "step": 96820 }, { "epoch": 1.8986274509803922, "grad_norm": 4.290827751159668, "learning_rate": 7.010094005141187e-09, "loss": 0.1617, "step": 96830 }, { "epoch": 1.8988235294117648, "grad_norm": 3.325751781463623, "learning_rate": 6.983070340959685e-09, "loss": 0.1366, "step": 96840 }, { "epoch": 1.8990196078431372, "grad_norm": 2.7087883949279785, "learning_rate": 6.9560984987603875e-09, "loss": 0.1676, "step": 96850 }, { "epoch": 1.8992156862745098, "grad_norm": 3.810818672180176, "learning_rate": 6.9291784813782505e-09, "loss": 0.1375, "step": 96860 }, { "epoch": 1.8994117647058824, "grad_norm": 4.560551166534424, "learning_rate": 6.902310291642899e-09, "loss": 0.1709, "step": 96870 }, { "epoch": 1.899607843137255, "grad_norm": 4.722829818725586, "learning_rate": 6.8754939323785175e-09, "loss": 0.1583, "step": 96880 }, { "epoch": 1.8998039215686275, "grad_norm": 3.39774227142334, "learning_rate": 6.848729406403908e-09, "loss": 0.1824, "step": 96890 }, { "epoch": 1.9, "grad_norm": 4.792853355407715, "learning_rate": 6.822016716532153e-09, "loss": 0.1681, "step": 96900 }, { "epoch": 1.9001960784313725, "grad_norm": 1.6344029903411865, "learning_rate": 6.795355865571284e-09, "loss": 0.1632, "step": 96910 }, { "epoch": 1.900392156862745, "grad_norm": 1.622955560684204, "learning_rate": 6.76874685632356e-09, "loss": 0.1756, "step": 96920 }, { "epoch": 1.9005882352941177, "grad_norm": 3.4208738803863525, "learning_rate": 6.742189691585909e-09, "loss": 0.1709, "step": 96930 }, { "epoch": 1.9007843137254903, "grad_norm": 2.9772982597351074, "learning_rate": 6.7156843741498216e-09, "loss": 0.167, "step": 96940 }, { "epoch": 1.9009803921568627, "grad_norm": 1.7661901712417603, "learning_rate": 6.6892309068013465e-09, "loss": 0.1405, "step": 96950 }, { "epoch": 1.9011764705882352, "grad_norm": 5.449833393096924, "learning_rate": 6.6628292923210375e-09, "loss": 0.1712, "step": 96960 }, { "epoch": 1.9013725490196078, "grad_norm": 2.620677947998047, "learning_rate": 6.636479533484007e-09, "loss": 0.1651, "step": 96970 }, { "epoch": 1.9015686274509804, "grad_norm": 2.7455883026123047, "learning_rate": 6.6101816330599856e-09, "loss": 0.1774, "step": 96980 }, { "epoch": 1.901764705882353, "grad_norm": 2.1404852867126465, "learning_rate": 6.58393559381315e-09, "loss": 0.1842, "step": 96990 }, { "epoch": 1.9019607843137254, "grad_norm": 4.345264434814453, "learning_rate": 6.557741418502238e-09, "loss": 0.186, "step": 97000 }, { "epoch": 1.902156862745098, "grad_norm": 3.2579355239868164, "learning_rate": 6.531599109880659e-09, "loss": 0.1413, "step": 97010 }, { "epoch": 1.9023529411764706, "grad_norm": 1.990606427192688, "learning_rate": 6.505508670696269e-09, "loss": 0.1465, "step": 97020 }, { "epoch": 1.9025490196078432, "grad_norm": 2.8857581615448, "learning_rate": 6.479470103691376e-09, "loss": 0.185, "step": 97030 }, { "epoch": 1.9027450980392158, "grad_norm": 1.9568125009536743, "learning_rate": 6.453483411603122e-09, "loss": 0.1846, "step": 97040 }, { "epoch": 1.9029411764705881, "grad_norm": 1.6720383167266846, "learning_rate": 6.4275485971628794e-09, "loss": 0.1755, "step": 97050 }, { "epoch": 1.903137254901961, "grad_norm": 3.855743885040283, "learning_rate": 6.401665663096745e-09, "loss": 0.1323, "step": 97060 }, { "epoch": 1.9033333333333333, "grad_norm": 3.4690675735473633, "learning_rate": 6.375834612125319e-09, "loss": 0.1554, "step": 97070 }, { "epoch": 1.903529411764706, "grad_norm": 2.3810689449310303, "learning_rate": 6.3500554469638204e-09, "loss": 0.1895, "step": 97080 }, { "epoch": 1.9037254901960785, "grad_norm": 2.168771505355835, "learning_rate": 6.324328170321858e-09, "loss": 0.1541, "step": 97090 }, { "epoch": 1.9039215686274509, "grad_norm": 3.4526333808898926, "learning_rate": 6.298652784903713e-09, "loss": 0.1567, "step": 97100 }, { "epoch": 1.9041176470588237, "grad_norm": 2.9333419799804688, "learning_rate": 6.273029293408172e-09, "loss": 0.1996, "step": 97110 }, { "epoch": 1.904313725490196, "grad_norm": 4.550887584686279, "learning_rate": 6.24745769852858e-09, "loss": 0.2249, "step": 97120 }, { "epoch": 1.9045098039215687, "grad_norm": 2.822293519973755, "learning_rate": 6.221938002952787e-09, "loss": 0.1698, "step": 97130 }, { "epoch": 1.9047058823529412, "grad_norm": 3.3494725227355957, "learning_rate": 6.196470209363314e-09, "loss": 0.1574, "step": 97140 }, { "epoch": 1.9049019607843136, "grad_norm": 39.47426986694336, "learning_rate": 6.171054320437019e-09, "loss": 0.1605, "step": 97150 }, { "epoch": 1.9050980392156864, "grad_norm": 5.581935882568359, "learning_rate": 6.145690338845433e-09, "loss": 0.1669, "step": 97160 }, { "epoch": 1.9052941176470588, "grad_norm": 10.061897277832031, "learning_rate": 6.1203782672547e-09, "loss": 0.1744, "step": 97170 }, { "epoch": 1.9054901960784314, "grad_norm": 4.190247058868408, "learning_rate": 6.095118108325304e-09, "loss": 0.1501, "step": 97180 }, { "epoch": 1.905686274509804, "grad_norm": 3.3461599349975586, "learning_rate": 6.069909864712453e-09, "loss": 0.2183, "step": 97190 }, { "epoch": 1.9058823529411764, "grad_norm": 2.2876646518707275, "learning_rate": 6.044753539065861e-09, "loss": 0.1793, "step": 97200 }, { "epoch": 1.9060784313725492, "grad_norm": 3.0969691276550293, "learning_rate": 6.019649134029692e-09, "loss": 0.2094, "step": 97210 }, { "epoch": 1.9062745098039215, "grad_norm": 4.144609451293945, "learning_rate": 5.994596652242778e-09, "loss": 0.1634, "step": 97220 }, { "epoch": 1.9064705882352941, "grad_norm": 3.498976469039917, "learning_rate": 5.969596096338403e-09, "loss": 0.1597, "step": 97230 }, { "epoch": 1.9066666666666667, "grad_norm": 2.789281129837036, "learning_rate": 5.944647468944464e-09, "loss": 0.2014, "step": 97240 }, { "epoch": 1.906862745098039, "grad_norm": 1.9846278429031372, "learning_rate": 5.919750772683308e-09, "loss": 0.1533, "step": 97250 }, { "epoch": 1.907058823529412, "grad_norm": 3.0577280521392822, "learning_rate": 5.894906010171896e-09, "loss": 0.1766, "step": 97260 }, { "epoch": 1.9072549019607843, "grad_norm": 3.649109363555908, "learning_rate": 5.870113184021752e-09, "loss": 0.1671, "step": 97270 }, { "epoch": 1.9074509803921569, "grad_norm": 5.60776948928833, "learning_rate": 5.845372296838791e-09, "loss": 0.1574, "step": 97280 }, { "epoch": 1.9076470588235295, "grad_norm": 3.7116549015045166, "learning_rate": 5.8206833512236545e-09, "loss": 0.1761, "step": 97290 }, { "epoch": 1.9078431372549018, "grad_norm": 1.4471269845962524, "learning_rate": 5.796046349771488e-09, "loss": 0.1397, "step": 97300 }, { "epoch": 1.9080392156862747, "grad_norm": 4.74782133102417, "learning_rate": 5.771461295071833e-09, "loss": 0.15, "step": 97310 }, { "epoch": 1.908235294117647, "grad_norm": 3.4602582454681396, "learning_rate": 5.746928189708955e-09, "loss": 0.1516, "step": 97320 }, { "epoch": 1.9084313725490196, "grad_norm": 8.479333877563477, "learning_rate": 5.722447036261568e-09, "loss": 0.1699, "step": 97330 }, { "epoch": 1.9086274509803922, "grad_norm": 5.047855377197266, "learning_rate": 5.698017837302893e-09, "loss": 0.1647, "step": 97340 }, { "epoch": 1.9088235294117646, "grad_norm": 2.3692262172698975, "learning_rate": 5.673640595400708e-09, "loss": 0.1488, "step": 97350 }, { "epoch": 1.9090196078431374, "grad_norm": 8.533371925354004, "learning_rate": 5.649315313117409e-09, "loss": 0.1634, "step": 97360 }, { "epoch": 1.9092156862745098, "grad_norm": 2.427799701690674, "learning_rate": 5.625041993009894e-09, "loss": 0.1563, "step": 97370 }, { "epoch": 1.9094117647058824, "grad_norm": 3.714107036590576, "learning_rate": 5.6008206376295105e-09, "loss": 0.1877, "step": 97380 }, { "epoch": 1.909607843137255, "grad_norm": 5.348755359649658, "learning_rate": 5.57665124952228e-09, "loss": 0.1556, "step": 97390 }, { "epoch": 1.9098039215686273, "grad_norm": 6.772092342376709, "learning_rate": 5.5525338312286676e-09, "loss": 0.1661, "step": 97400 }, { "epoch": 1.9100000000000001, "grad_norm": 1.867067813873291, "learning_rate": 5.528468385283647e-09, "loss": 0.1651, "step": 97410 }, { "epoch": 1.9101960784313725, "grad_norm": 3.930208444595337, "learning_rate": 5.504454914216861e-09, "loss": 0.1946, "step": 97420 }, { "epoch": 1.910392156862745, "grad_norm": 6.399804592132568, "learning_rate": 5.4804934205524565e-09, "loss": 0.1828, "step": 97430 }, { "epoch": 1.9105882352941177, "grad_norm": 2.675964593887329, "learning_rate": 5.45658390680892e-09, "loss": 0.1683, "step": 97440 }, { "epoch": 1.91078431372549, "grad_norm": 2.644547462463379, "learning_rate": 5.432726375499519e-09, "loss": 0.158, "step": 97450 }, { "epoch": 1.9109803921568629, "grad_norm": 4.23090124130249, "learning_rate": 5.408920829132024e-09, "loss": 0.1838, "step": 97460 }, { "epoch": 1.9111764705882353, "grad_norm": 2.6519968509674072, "learning_rate": 5.385167270208546e-09, "loss": 0.193, "step": 97470 }, { "epoch": 1.9113725490196078, "grad_norm": 3.3616652488708496, "learning_rate": 5.3614657012259755e-09, "loss": 0.1526, "step": 97480 }, { "epoch": 1.9115686274509804, "grad_norm": 4.368190288543701, "learning_rate": 5.337816124675598e-09, "loss": 0.1894, "step": 97490 }, { "epoch": 1.9117647058823528, "grad_norm": 5.9945197105407715, "learning_rate": 5.314218543043259e-09, "loss": 0.1547, "step": 97500 }, { "epoch": 1.9119607843137256, "grad_norm": 11.842249870300293, "learning_rate": 5.290672958809361e-09, "loss": 0.1732, "step": 97510 }, { "epoch": 1.912156862745098, "grad_norm": 5.258184909820557, "learning_rate": 5.2671793744488154e-09, "loss": 0.1737, "step": 97520 }, { "epoch": 1.9123529411764706, "grad_norm": 6.847755432128906, "learning_rate": 5.243737792431091e-09, "loss": 0.1283, "step": 97530 }, { "epoch": 1.9125490196078432, "grad_norm": 4.1299920082092285, "learning_rate": 5.220348215220105e-09, "loss": 0.1705, "step": 97540 }, { "epoch": 1.9127450980392156, "grad_norm": 2.0535573959350586, "learning_rate": 5.197010645274502e-09, "loss": 0.1775, "step": 97550 }, { "epoch": 1.9129411764705884, "grad_norm": 3.0791590213775635, "learning_rate": 5.1737250850473205e-09, "loss": 0.1478, "step": 97560 }, { "epoch": 1.9131372549019607, "grad_norm": 2.0073699951171875, "learning_rate": 5.150491536986046e-09, "loss": 0.1605, "step": 97570 }, { "epoch": 1.9133333333333333, "grad_norm": 1.920637845993042, "learning_rate": 5.127310003532948e-09, "loss": 0.1481, "step": 97580 }, { "epoch": 1.913529411764706, "grad_norm": 4.9446916580200195, "learning_rate": 5.104180487124577e-09, "loss": 0.1781, "step": 97590 }, { "epoch": 1.9137254901960783, "grad_norm": 4.141665458679199, "learning_rate": 5.081102990192099e-09, "loss": 0.1799, "step": 97600 }, { "epoch": 1.913921568627451, "grad_norm": 4.423698902130127, "learning_rate": 5.058077515161352e-09, "loss": 0.1642, "step": 97610 }, { "epoch": 1.9141176470588235, "grad_norm": 3.4906258583068848, "learning_rate": 5.035104064452567e-09, "loss": 0.1624, "step": 97620 }, { "epoch": 1.914313725490196, "grad_norm": 3.5681092739105225, "learning_rate": 5.0121826404804225e-09, "loss": 0.1791, "step": 97630 }, { "epoch": 1.9145098039215687, "grad_norm": 4.184510707855225, "learning_rate": 4.989313245654325e-09, "loss": 0.1831, "step": 97640 }, { "epoch": 1.914705882352941, "grad_norm": 5.335333347320557, "learning_rate": 4.966495882378075e-09, "loss": 0.1802, "step": 97650 }, { "epoch": 1.9149019607843139, "grad_norm": 5.291140556335449, "learning_rate": 4.943730553050085e-09, "loss": 0.1651, "step": 97660 }, { "epoch": 1.9150980392156862, "grad_norm": 3.6772446632385254, "learning_rate": 4.921017260063276e-09, "loss": 0.1906, "step": 97670 }, { "epoch": 1.9152941176470588, "grad_norm": 1.765798807144165, "learning_rate": 4.898356005805016e-09, "loss": 0.1682, "step": 97680 }, { "epoch": 1.9154901960784314, "grad_norm": 2.408818244934082, "learning_rate": 4.875746792657398e-09, "loss": 0.1483, "step": 97690 }, { "epoch": 1.9156862745098038, "grad_norm": 1.9210132360458374, "learning_rate": 4.853189622996745e-09, "loss": 0.1651, "step": 97700 }, { "epoch": 1.9158823529411766, "grad_norm": 2.0807504653930664, "learning_rate": 4.8306844991942706e-09, "loss": 0.1709, "step": 97710 }, { "epoch": 1.916078431372549, "grad_norm": 4.696357727050781, "learning_rate": 4.8082314236153605e-09, "loss": 0.1421, "step": 97720 }, { "epoch": 1.9162745098039216, "grad_norm": 4.681709289550781, "learning_rate": 4.785830398620238e-09, "loss": 0.178, "step": 97730 }, { "epoch": 1.9164705882352941, "grad_norm": 2.669921636581421, "learning_rate": 4.763481426563465e-09, "loss": 0.182, "step": 97740 }, { "epoch": 1.9166666666666665, "grad_norm": 2.5012776851654053, "learning_rate": 4.741184509794216e-09, "loss": 0.1613, "step": 97750 }, { "epoch": 1.9168627450980393, "grad_norm": 1.5868545770645142, "learning_rate": 4.718939650656062e-09, "loss": 0.1696, "step": 97760 }, { "epoch": 1.9170588235294117, "grad_norm": 6.44674015045166, "learning_rate": 4.6967468514873545e-09, "loss": 0.149, "step": 97770 }, { "epoch": 1.9172549019607843, "grad_norm": 2.375676155090332, "learning_rate": 4.674606114620783e-09, "loss": 0.1703, "step": 97780 }, { "epoch": 1.917450980392157, "grad_norm": 4.672613143920898, "learning_rate": 4.652517442383486e-09, "loss": 0.1668, "step": 97790 }, { "epoch": 1.9176470588235293, "grad_norm": 1.653153657913208, "learning_rate": 4.6304808370973285e-09, "loss": 0.1639, "step": 97800 }, { "epoch": 1.917843137254902, "grad_norm": 5.015472888946533, "learning_rate": 4.608496301078679e-09, "loss": 0.1834, "step": 97810 }, { "epoch": 1.9180392156862744, "grad_norm": 2.9541003704071045, "learning_rate": 4.586563836638302e-09, "loss": 0.1531, "step": 97820 }, { "epoch": 1.918235294117647, "grad_norm": 4.3947672843933105, "learning_rate": 4.564683446081574e-09, "loss": 0.1601, "step": 97830 }, { "epoch": 1.9184313725490196, "grad_norm": 2.7924752235412598, "learning_rate": 4.542855131708379e-09, "loss": 0.157, "step": 97840 }, { "epoch": 1.918627450980392, "grad_norm": 2.1752235889434814, "learning_rate": 4.521078895813213e-09, "loss": 0.1731, "step": 97850 }, { "epoch": 1.9188235294117648, "grad_norm": 3.773127794265747, "learning_rate": 4.499354740684858e-09, "loss": 0.1623, "step": 97860 }, { "epoch": 1.9190196078431372, "grad_norm": 2.3152568340301514, "learning_rate": 4.47768266860693e-09, "loss": 0.1779, "step": 97870 }, { "epoch": 1.9192156862745098, "grad_norm": 2.9606826305389404, "learning_rate": 4.456062681857387e-09, "loss": 0.1677, "step": 97880 }, { "epoch": 1.9194117647058824, "grad_norm": 2.7219550609588623, "learning_rate": 4.434494782708687e-09, "loss": 0.1761, "step": 97890 }, { "epoch": 1.9196078431372547, "grad_norm": 7.619016170501709, "learning_rate": 4.412978973427905e-09, "loss": 0.1641, "step": 97900 }, { "epoch": 1.9198039215686276, "grad_norm": 4.889416217803955, "learning_rate": 4.391515256276679e-09, "loss": 0.2051, "step": 97910 }, { "epoch": 1.92, "grad_norm": 2.25229811668396, "learning_rate": 4.370103633511035e-09, "loss": 0.1407, "step": 97920 }, { "epoch": 1.9201960784313725, "grad_norm": 2.821127414703369, "learning_rate": 4.348744107381619e-09, "loss": 0.175, "step": 97930 }, { "epoch": 1.9203921568627451, "grad_norm": 2.0972084999084473, "learning_rate": 4.327436680133578e-09, "loss": 0.1513, "step": 97940 }, { "epoch": 1.9205882352941175, "grad_norm": 2.0973949432373047, "learning_rate": 4.30618135400651e-09, "loss": 0.1719, "step": 97950 }, { "epoch": 1.9207843137254903, "grad_norm": 4.277551651000977, "learning_rate": 4.284978131234629e-09, "loss": 0.2037, "step": 97960 }, { "epoch": 1.9209803921568627, "grad_norm": 3.181502342224121, "learning_rate": 4.263827014046761e-09, "loss": 0.1878, "step": 97970 }, { "epoch": 1.9211764705882353, "grad_norm": 5.247817516326904, "learning_rate": 4.242728004665963e-09, "loss": 0.1693, "step": 97980 }, { "epoch": 1.9213725490196079, "grad_norm": 4.8475542068481445, "learning_rate": 4.221681105310126e-09, "loss": 0.1626, "step": 97990 }, { "epoch": 1.9215686274509802, "grad_norm": 3.1119110584259033, "learning_rate": 4.200686318191537e-09, "loss": 0.151, "step": 98000 } ], "logging_steps": 10, "max_steps": 102000, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 4.3478335300975985e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }