{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9018036072144289, "eval_steps": 100, "global_step": 1350, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "learning_rate": 1.111111111111111e-08, "loss": 0.0176, "step": 2, "train/R_acc": 0.003125, "train/R_penalty": -0.671875, "train/R_reason": 0.1771926448030963, "train/R_vocab": 0.27042410714285714 }, { "learning_rate": 3.3333333333333334e-08, "loss": 0.0255, "step": 4, "train/R_acc": 0.00625, "train/R_penalty": -0.578125, "train/R_reason": 0.20292794300576886, "train/R_vocab": 0.13051835317460317 }, { "learning_rate": 5.555555555555555e-08, "loss": -0.0402, "step": 6, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.21681085047252255, "train/R_vocab": 0.3141617063492063 }, { "learning_rate": 7.777777777777778e-08, "loss": -0.0512, "step": 8, "train/R_acc": 0.034375, "train/R_penalty": -0.53125, "train/R_reason": 0.22407962891811065, "train/R_vocab": 0.2578125 }, { "learning_rate": 1e-07, "loss": 0.0653, "step": 10, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.24027254265023745, "train/R_vocab": 0.290625 }, { "learning_rate": 1.2222222222222222e-07, "loss": 0.0747, "step": 12, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.2086565261070774, "train/R_vocab": 0.2722098214285714 }, { "learning_rate": 1.4444444444444442e-07, "loss": -0.0073, "step": 14, "train/R_acc": 0.034375, "train/R_penalty": -0.78125, "train/R_reason": 0.18557344633053446, "train/R_vocab": 0.23597470238095236 }, { "learning_rate": 1.6666666666666665e-07, "loss": 0.0011, "step": 16, "train/R_acc": 0.00625, "train/R_penalty": -0.5625, "train/R_reason": 0.18460204443700487, "train/R_vocab": 0.28095238095238095 }, { "learning_rate": 1.8888888888888888e-07, "loss": 0.0511, "step": 18, "train/R_acc": 0.003125, "train/R_penalty": -0.6875, "train/R_reason": 0.1691650605005834, "train/R_vocab": 0.22633928571428572 }, { "learning_rate": 2.111111111111111e-07, "loss": -0.0247, "step": 20, "train/R_acc": 0.03125, "train/R_penalty": -0.484375, "train/R_reason": 0.23429908163761645, "train/R_vocab": 0.3410714285714286 }, { "learning_rate": 2.3333333333333333e-07, "loss": 0.0088, "step": 22, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.15088100811244767, "train/R_vocab": 0.2056919642857143 }, { "learning_rate": 2.5555555555555553e-07, "loss": 0.0366, "step": 24, "train/R_acc": 0.0, "train/R_penalty": -0.6875, "train/R_reason": 0.2605457318222165, "train/R_vocab": 0.29765625 }, { "learning_rate": 2.7777777777777776e-07, "loss": 0.0558, "step": 26, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.5, "train/R_reason": 0.24175666877540153, "train/R_vocab": 0.26484375000000004 }, { "learning_rate": 3e-07, "loss": 0.0588, "step": 28, "train/R_acc": 0.0, "train/R_penalty": -0.540625, "train/R_reason": 0.2054192775694245, "train/R_vocab": 0.2109747023809524 }, { "learning_rate": 3.222222222222222e-07, "loss": -0.1103, "step": 30, "train/R_acc": 0.003125, "train/R_penalty": -0.59375, "train/R_reason": 0.18972739597143928, "train/R_vocab": 0.2044642857142857 }, { "learning_rate": 3.4444444444444444e-07, "loss": 0.0162, "step": 32, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.16386884044279187, "train/R_vocab": 0.26328125 }, { "learning_rate": 3.666666666666666e-07, "loss": 0.0241, "step": 34, "train/R_acc": 0.065625, "train/R_penalty": -0.6875, "train/R_reason": 0.1991015810053379, "train/R_vocab": 0.3072916666666667 }, { "learning_rate": 3.888888888888889e-07, "loss": 0.0114, "step": 36, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.1542880167619275, "train/R_vocab": 0.15633680555555557 }, { "learning_rate": 4.1111111111111107e-07, "loss": 0.0418, "step": 38, "train/R_acc": 0.03125, "train/R_penalty": -0.59375, "train/R_reason": 0.21628990497316553, "train/R_vocab": 0.32326388888888885 }, { "learning_rate": 4.3333333333333335e-07, "loss": -0.0085, "step": 40, "train/R_acc": 0.065625, "train/R_penalty": -0.540625, "train/R_reason": 0.16669781447815488, "train/R_vocab": 0.21458333333333335 }, { "learning_rate": 4.555555555555555e-07, "loss": -0.0614, "step": 42, "train/R_acc": 0.003125, "train/R_penalty": -0.671875, "train/R_reason": 0.23833932777613548, "train/R_vocab": 0.24285714285714288 }, { "learning_rate": 4.777777777777778e-07, "loss": 0.004, "step": 44, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.2297590496740362, "train/R_vocab": 0.1607142857142857 }, { "learning_rate": 5e-07, "loss": 0.0491, "step": 46, "train/R_acc": 0.03125, "train/R_penalty": -0.6875, "train/R_reason": 0.2037377285506986, "train/R_vocab": 0.2598958333333333 }, { "learning_rate": 4.999976528920434e-07, "loss": -0.083, "step": 48, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.22468232799435314, "train/R_vocab": 0.2745907738095238 }, { "learning_rate": 4.999906116122447e-07, "loss": -0.0036, "step": 50, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.22593521940016348, "train/R_vocab": 0.2805803571428572 }, { "learning_rate": 4.999788762928172e-07, "loss": 0.0107, "step": 52, "train/R_acc": 0.003125, "train/R_penalty": -0.821875, "train/R_reason": 0.20522814513215548, "train/R_vocab": 0.14702380952380953 }, { "learning_rate": 4.999624471541134e-07, "loss": 0.029, "step": 54, "train/R_acc": 0.0, "train/R_penalty": -0.78125, "train/R_reason": 0.21737758529192291, "train/R_vocab": 0.19330357142857144 }, { "learning_rate": 4.99941324504621e-07, "loss": -0.0246, "step": 56, "train/R_acc": 0.00625, "train/R_penalty": -0.546875, "train/R_reason": 0.2251848744560886, "train/R_vocab": 0.28158482142857144 }, { "learning_rate": 4.999155087409571e-07, "loss": -0.0517, "step": 58, "train/R_acc": 0.03125, "train/R_penalty": -0.571875, "train/R_reason": 0.18310813181259872, "train/R_vocab": 0.24776785714285715 }, { "learning_rate": 4.998850003478606e-07, "loss": 0.0598, "step": 60, "train/R_acc": 0.034375, "train/R_penalty": -0.703125, "train/R_reason": 0.2082140947052848, "train/R_vocab": 0.26149553571428574 }, { "learning_rate": 4.998497998981838e-07, "loss": 0.0027, "step": 62, "train/R_acc": 0.0, "train/R_penalty": -0.828125, "train/R_reason": 0.2611822631590372, "train/R_vocab": 0.39639136904761907 }, { "learning_rate": 4.998099080528803e-07, "loss": 0.0385, "step": 64, "train/R_acc": 0.0, "train/R_penalty": -0.53125, "train/R_reason": 0.2213369943749453, "train/R_vocab": 0.3606026785714286 }, { "learning_rate": 4.997653255609941e-07, "loss": -0.0523, "step": 66, "train/R_acc": 0.065625, "train/R_penalty": -0.578125, "train/R_reason": 0.20860609604648056, "train/R_vocab": 0.2251488095238095 }, { "learning_rate": 4.997160532596447e-07, "loss": 0.0352, "step": 68, "train/R_acc": 0.003125, "train/R_penalty": -0.790625, "train/R_reason": 0.19482757334666323, "train/R_vocab": 0.23776041666666667 }, { "learning_rate": 4.996620920740111e-07, "loss": 0.0103, "step": 70, "train/R_acc": 0.03125, "train/R_penalty": -0.671875, "train/R_reason": 0.25304049429757086, "train/R_vocab": 0.299516369047619 }, { "learning_rate": 4.996034430173153e-07, "loss": 0.0247, "step": 72, "train/R_acc": 0.003125, "train/R_penalty": -0.58125, "train/R_reason": 0.27724208008758466, "train/R_vocab": 0.3654637896825397 }, { "learning_rate": 4.995401071908025e-07, "loss": -0.0689, "step": 74, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.18864259893071442, "train/R_vocab": 0.265625 }, { "learning_rate": 4.99472085783721e-07, "loss": 0.0669, "step": 76, "train/R_acc": 0.003125, "train/R_penalty": -0.775, "train/R_reason": 0.24872595893355823, "train/R_vocab": 0.3293526785714286 }, { "learning_rate": 4.993993800732995e-07, "loss": -0.0369, "step": 78, "train/R_acc": 0.034375, "train/R_penalty": -0.546875, "train/R_reason": 0.26486544822313024, "train/R_vocab": 0.3398065476190476 }, { "learning_rate": 4.993219914247231e-07, "loss": 0.0685, "step": 80, "train/R_acc": 0.03125, "train/R_penalty": -0.75, "train/R_reason": 0.1711866679882719, "train/R_vocab": 0.2575892857142857 }, { "learning_rate": 4.992399212911082e-07, "loss": 0.0382, "step": 82, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.640625, "train/R_reason": 0.22934020748837214, "train/R_vocab": 0.31875 }, { "learning_rate": 4.99153171213474e-07, "loss": 0.083, "step": 84, "train/R_acc": 0.003125, "train/R_penalty": -0.703125, "train/R_reason": 0.16303826493835916, "train/R_vocab": 0.18002232142857144 }, { "learning_rate": 4.990617428207153e-07, "loss": 0.0433, "step": 86, "train/R_acc": 0.03125, "train/R_penalty": -0.84375, "train/R_reason": 0.21139857471142573, "train/R_vocab": 0.28266369047619044 }, { "learning_rate": 4.989656378295704e-07, "loss": 0.0554, "step": 88, "train/R_acc": 0.00625, "train/R_penalty": -0.5, "train/R_reason": 0.2142568618009861, "train/R_vocab": 0.3390625 }, { "learning_rate": 4.988648580445895e-07, "loss": 0.0648, "step": 90, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.17382571925607848, "train/R_vocab": 0.1894345238095238 }, { "learning_rate": 4.987594053581012e-07, "loss": -0.0241, "step": 92, "train/R_acc": 0.003125, "train/R_penalty": -0.703125, "train/R_reason": 0.21781325533595836, "train/R_vocab": 0.33205605158730156 }, { "learning_rate": 4.98649281750176e-07, "loss": 0.1022, "step": 94, "train/R_acc": 0.003125, "train/R_penalty": -0.703125, "train/R_reason": 0.1999237909283059, "train/R_vocab": 0.32064732142857144 }, { "learning_rate": 4.985344892885899e-07, "loss": 0.0821, "step": 96, "train/R_acc": 0.00625, "train/R_penalty": -0.578125, "train/R_reason": 0.23594255343132425, "train/R_vocab": 0.30199652777777775 }, { "learning_rate": 4.984150301287853e-07, "loss": 0.0096, "step": 98, "train/R_acc": 0.0, "train/R_penalty": -0.80625, "train/R_reason": 0.1900727268525148, "train/R_vocab": 0.20625000000000002 }, { "learning_rate": 4.982909065138305e-07, "loss": 0.0247, "step": 100, "train/R_acc": 0.003125, "train/R_penalty": -0.6875, "train/R_reason": 0.21411060850160504, "train/R_vocab": 0.22946428571428573 }, { "eval/R_acc": 0.002777777777777778, "eval/R_penalty": -0.6847222222222222, "eval/R_reason": 0.2572056199703138, "eval/R_vocab": 0.3255787037037037, "step": 100 }, { "step": 100 }, { "learning_rate": 4.981621207743778e-07, "loss": -0.0004, "step": 102, "train/R_acc": 0.0, "train/R_penalty": -0.765625, "train/R_reason": 0.1735908404636029, "train/R_vocab": 0.15747767857142858 }, { "learning_rate": 4.980286753286194e-07, "loss": 0.0155, "step": 104, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.2016416402158515, "train/R_vocab": 0.3698660714285714 }, { "learning_rate": 4.978905726822423e-07, "loss": -0.0257, "step": 106, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.2019264344648219, "train/R_vocab": 0.2499627976190476 }, { "learning_rate": 4.97747815428381e-07, "loss": 0.0757, "step": 108, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.2609304111036691, "train/R_vocab": 0.22472718253968255 }, { "learning_rate": 4.976004062475691e-07, "loss": -0.0376, "step": 110, "train/R_acc": 0.03125, "train/R_penalty": -0.55625, "train/R_reason": 0.24456802402586483, "train/R_vocab": 0.30167410714285714 }, { "learning_rate": 4.974483479076885e-07, "loss": -0.0155, "step": 112, "train/R_acc": 0.0, "train/R_penalty": -0.734375, "train/R_reason": 0.15732482982112658, "train/R_vocab": 0.20078125 }, { "learning_rate": 4.972916432639181e-07, "loss": 0.0107, "step": 114, "train/R_acc": 0.0625, "train/R_penalty": -0.6125, "train/R_reason": 0.19719755508097864, "train/R_vocab": 0.18292410714285715 }, { "learning_rate": 4.971302952586796e-07, "loss": -0.0536, "step": 116, "train/R_acc": 0.003125, "train/R_penalty": -0.6875, "train/R_reason": 0.1615191195936923, "train/R_vocab": 0.20714285714285713 }, { "learning_rate": 4.969643069215823e-07, "loss": 0.03, "step": 118, "train/R_acc": 0.003125, "train/R_penalty": -0.703125, "train/R_reason": 0.2057090874466078, "train/R_vocab": 0.26006944444444446 }, { "learning_rate": 4.967936813693668e-07, "loss": 0.0669, "step": 120, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.17123245123013492, "train/R_vocab": 0.22881944444444446 }, { "learning_rate": 4.966184218058457e-07, "loss": -0.01, "step": 122, "train/R_acc": 0.065625, "train/R_penalty": -0.640625, "train/R_reason": 0.25008522769707353, "train/R_vocab": 0.2768725198412698 }, { "learning_rate": 4.96438531521844e-07, "loss": 0.0227, "step": 124, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.228776542397296, "train/R_vocab": 0.19776785714285716 }, { "learning_rate": 4.962540138951371e-07, "loss": 0.0342, "step": 126, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.20827985353158485, "train/R_vocab": 0.3024440836940837 }, { "learning_rate": 4.960648723903872e-07, "loss": 0.0418, "step": 128, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.19681757054800564, "train/R_vocab": 0.22455357142857144 }, { "learning_rate": 4.958711105590785e-07, "loss": -0.0196, "step": 130, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.5625, "train/R_reason": 0.22729280568853277, "train/R_vocab": 0.19475446428571427 }, { "learning_rate": 4.956727320394506e-07, "loss": 0.0963, "step": 132, "train/R_acc": 0.003125, "train/R_penalty": -0.84375, "train/R_reason": 0.20146228335738753, "train/R_vocab": 0.24609995039682542 }, { "learning_rate": 4.954697405564299e-07, "loss": 0.0259, "step": 134, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.625, "train/R_reason": 0.228208707691525, "train/R_vocab": 0.23055217352092353 }, { "learning_rate": 4.952621399215597e-07, "loss": 0.0781, "step": 136, "train/R_acc": 0.003125, "train/R_penalty": -0.8125, "train/R_reason": 0.1811115376120021, "train/R_vocab": 0.2921875 }, { "learning_rate": 4.95049934032929e-07, "loss": 0.0659, "step": 138, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.2199008430371848, "train/R_vocab": 0.22447916666666667 }, { "learning_rate": 4.948331268750988e-07, "loss": -0.0087, "step": 140, "train/R_acc": 0.03125, "train/R_penalty": -0.796875, "train/R_reason": 0.16414720324183496, "train/R_vocab": 0.20345982142857144 }, { "learning_rate": 4.946117225190273e-07, "loss": 0.0109, "step": 142, "train/R_acc": 0.03125, "train/R_penalty": -0.634375, "train/R_reason": 0.20413731109417932, "train/R_vocab": 0.27366071428571426 }, { "learning_rate": 4.943857251219942e-07, "loss": -0.0291, "step": 144, "train/R_acc": 0.0, "train/R_penalty": -0.703125, "train/R_reason": 0.2114743293926566, "train/R_vocab": 0.24921875 }, { "learning_rate": 4.941551389275217e-07, "loss": -0.0249, "step": 146, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.19512092302735518, "train/R_vocab": 0.17879464285714286 }, { "learning_rate": 4.939199682652952e-07, "loss": -0.03, "step": 148, "train/R_acc": 0.003125, "train/R_penalty": -0.6125, "train/R_reason": 0.19263372463344114, "train/R_vocab": 0.2714285714285714 }, { "learning_rate": 4.936802175510823e-07, "loss": -0.1025, "step": 150, "train/R_acc": 0.00625, "train/R_penalty": -0.665625, "train/R_reason": 0.21358490777949846, "train/R_vocab": 0.3302455357142857 }, { "learning_rate": 4.934358912866497e-07, "loss": 0.0293, "step": 152, "train/R_acc": 0.00625, "train/R_penalty": -0.484375, "train/R_reason": 0.16662975968037924, "train/R_vocab": 0.1699280753968254 }, { "learning_rate": 4.931869940596779e-07, "loss": 0.0662, "step": 154, "train/R_acc": 0.00625, "train/R_penalty": -0.59375, "train/R_reason": 0.20091746530920346, "train/R_vocab": 0.27994791666666663 }, { "learning_rate": 4.929335305436764e-07, "loss": 0.0355, "step": 156, "train/R_acc": 0.003125, "train/R_penalty": -0.6875, "train/R_reason": 0.22218261967350866, "train/R_vocab": 0.3209821428571429 }, { "learning_rate": 4.926755054978951e-07, "loss": 0.0457, "step": 158, "train/R_acc": 0.0, "train/R_penalty": -0.703125, "train/R_reason": 0.22181716803364154, "train/R_vocab": 0.29573863636363634 }, { "learning_rate": 4.92412923767235e-07, "loss": -0.0201, "step": 160, "train/R_acc": 0.00625, "train/R_penalty": -0.625, "train/R_reason": 0.24103408318415312, "train/R_vocab": 0.23645833333333333 }, { "learning_rate": 4.921457902821578e-07, "loss": -0.0013, "step": 162, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.21205287864979985, "train/R_vocab": 0.24444444444444446 }, { "learning_rate": 4.918741100585921e-07, "loss": -0.0549, "step": 164, "train/R_acc": 0.0, "train/R_penalty": -0.75, "train/R_reason": 0.21139952333180925, "train/R_vocab": 0.36577380952380956 }, { "learning_rate": 4.915978881978406e-07, "loss": 0.0116, "step": 166, "train/R_acc": 0.003125, "train/R_penalty": -0.671875, "train/R_reason": 0.21990626286225867, "train/R_vocab": 0.3359623015873016 }, { "learning_rate": 4.913171298864836e-07, "loss": -0.0194, "step": 168, "train/R_acc": 0.0, "train/R_penalty": -0.6875, "train/R_reason": 0.19859495503742927, "train/R_vocab": 0.3138020833333333 }, { "learning_rate": 4.910318403962813e-07, "loss": 0.0306, "step": 170, "train/R_acc": 0.0, "train/R_penalty": -0.75, "train/R_reason": 0.15012936918136355, "train/R_vocab": 0.2549107142857143 }, { "learning_rate": 4.907420250840761e-07, "loss": 0.0817, "step": 172, "train/R_acc": 0.03125, "train/R_penalty": -0.68125, "train/R_reason": 0.20995580823960475, "train/R_vocab": 0.20602678571428573 }, { "learning_rate": 4.9044768939169e-07, "loss": -0.104, "step": 174, "train/R_acc": 0.03125, "train/R_penalty": -0.625, "train/R_reason": 0.18253049494341367, "train/R_vocab": 0.20379464285714285 }, { "learning_rate": 4.901488388458247e-07, "loss": 0.0418, "step": 176, "train/R_acc": 0.0625, "train/R_penalty": -0.5625, "train/R_reason": 0.2305279249641788, "train/R_vocab": 0.38442460317460314 }, { "learning_rate": 4.898454790579558e-07, "loss": -0.0033, "step": 178, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.18641040374410697, "train/R_vocab": 0.25188492063492063 }, { "learning_rate": 4.895376157242288e-07, "loss": 0.0036, "step": 180, "train/R_acc": 0.003125, "train/R_penalty": -0.634375, "train/R_reason": 0.19709747248451465, "train/R_vocab": 0.28185763888888893 }, { "learning_rate": 4.892252546253514e-07, "loss": -0.0037, "step": 182, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.2449271246503006, "train/R_vocab": 0.30446428571428574 }, { "learning_rate": 4.889084016264858e-07, "loss": 0.0279, "step": 184, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.24078946289847405, "train/R_vocab": 0.3054935515873016 }, { "learning_rate": 4.88587062677137e-07, "loss": 0.0824, "step": 186, "train/R_acc": 0.00625, "train/R_penalty": -0.625, "train/R_reason": 0.16765366722116085, "train/R_vocab": 0.22786458333333331 }, { "learning_rate": 4.882612438110429e-07, "loss": 0.0193, "step": 188, "train/R_acc": 0.003125, "train/R_penalty": -0.71875, "train/R_reason": 0.21543693992639024, "train/R_vocab": 0.3402901785714285 }, { "learning_rate": 4.8793095114606e-07, "loss": -0.0257, "step": 190, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.154932479362758, "train/R_vocab": 0.2091393849206349 }, { "learning_rate": 4.875961908840485e-07, "loss": -0.0201, "step": 192, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.20005776839850115, "train/R_vocab": 0.26316964285714284 }, { "learning_rate": 4.872569693107563e-07, "loss": 0.0163, "step": 194, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.2028767825912614, "train/R_vocab": 0.22935267857142855 }, { "learning_rate": 4.869132927957006e-07, "loss": -0.0312, "step": 196, "train/R_acc": 0.0625, "train/R_penalty": -0.6875, "train/R_reason": 0.20598954317804136, "train/R_vocab": 0.2894097222222222 }, { "learning_rate": 4.865651677920483e-07, "loss": -0.041, "step": 198, "train/R_acc": 0.003125, "train/R_penalty": -0.5625, "train/R_reason": 0.19447689115700523, "train/R_vocab": 0.21484375 }, { "learning_rate": 4.862126008364954e-07, "loss": -0.0021, "step": 200, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.24708335127072298, "train/R_vocab": 0.3235491071428571 }, { "eval/R_acc": 0.001388888888888889, "eval/R_penalty": -0.6875, "eval/R_reason": 0.26545862362041284, "eval/R_vocab": 0.3180500440917107, "step": 200 }, { "step": 200 }, { "learning_rate": 4.858555985491434e-07, "loss": 0.0539, "step": 202, "train/R_acc": 0.003125, "train/R_penalty": -0.71875, "train/R_reason": 0.12246968403633852, "train/R_vocab": 0.12857142857142856 }, { "learning_rate": 4.854941676333756e-07, "loss": -0.0105, "step": 204, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.59375, "train/R_reason": 0.20647794674779765, "train/R_vocab": 0.2294642857142857 }, { "learning_rate": 4.85128314875731e-07, "loss": 0.0418, "step": 206, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.22341563734148095, "train/R_vocab": 0.24494047619047618 }, { "learning_rate": 4.84758047145777e-07, "loss": 0.0338, "step": 208, "train/R_acc": 0.0, "train/R_penalty": -0.55625, "train/R_reason": 0.21091575917752947, "train/R_vocab": 0.25234375 }, { "learning_rate": 4.843833713959802e-07, "loss": 0.0127, "step": 210, "train/R_acc": 0.003125, "train/R_penalty": -0.78125, "train/R_reason": 0.15006747203000467, "train/R_vocab": 0.23049355158730156 }, { "learning_rate": 4.840042946615761e-07, "loss": 0.0429, "step": 212, "train/R_acc": 0.003125, "train/R_penalty": -0.59375, "train/R_reason": 0.23417842297662064, "train/R_vocab": 0.28020833333333334 }, { "learning_rate": 4.836208240604368e-07, "loss": -0.0, "step": 214, "train/R_acc": 0.0, "train/R_penalty": -0.65, "train/R_reason": 0.20300438739011717, "train/R_vocab": 0.30078125 }, { "learning_rate": 4.832329667929376e-07, "loss": 0.0223, "step": 216, "train/R_acc": 0.0, "train/R_penalty": -0.696875, "train/R_reason": 0.21329021924994432, "train/R_vocab": 0.2752604166666667 }, { "learning_rate": 4.828407301418217e-07, "loss": -0.0077, "step": 218, "train/R_acc": 0.0, "train/R_penalty": -0.75, "train/R_reason": 0.22684321233993932, "train/R_vocab": 0.2927331349206349 }, { "learning_rate": 4.824441214720628e-07, "loss": 0.0196, "step": 220, "train/R_acc": 0.003125, "train/R_penalty": -0.68125, "train/R_reason": 0.22135178051709037, "train/R_vocab": 0.3354166666666667 }, { "learning_rate": 4.820431482307281e-07, "loss": 0.0335, "step": 222, "train/R_acc": 0.034375, "train/R_penalty": -0.609375, "train/R_reason": 0.23161254238918147, "train/R_vocab": 0.3784598214285714 }, { "learning_rate": 4.816378179468374e-07, "loss": 0.0416, "step": 224, "train/R_acc": 0.003125, "train/R_penalty": -0.728125, "train/R_reason": 0.20212408310335858, "train/R_vocab": 0.2705357142857143 }, { "learning_rate": 4.812281382312223e-07, "loss": 0.0156, "step": 226, "train/R_acc": 0.03125, "train/R_penalty": -0.640625, "train/R_reason": 0.19926301994985757, "train/R_vocab": 0.35390625000000003 }, { "learning_rate": 4.808141167763826e-07, "loss": 0.0111, "step": 228, "train/R_acc": 0.003125, "train/R_penalty": -0.484375, "train/R_reason": 0.24773363470223553, "train/R_vocab": 0.3498015873015873 }, { "learning_rate": 4.803957613563431e-07, "loss": 0.1022, "step": 230, "train/R_acc": 0.00625, "train/R_penalty": -0.609375, "train/R_reason": 0.19832487010930644, "train/R_vocab": 0.2538690476190476 }, { "learning_rate": 4.799730798265063e-07, "loss": 0.1914, "step": 232, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.22378486607186504, "train/R_vocab": 0.28828125000000004 }, { "learning_rate": 4.795460801235058e-07, "loss": -0.036, "step": 234, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.17550109595621496, "train/R_vocab": 0.24095982142857145 }, { "learning_rate": 4.791147702650565e-07, "loss": 0.1009, "step": 236, "train/R_acc": 0.034375, "train/R_penalty": -0.61875, "train/R_reason": 0.2616533979112148, "train/R_vocab": 0.3763888888888889 }, { "learning_rate": 4.786791583498051e-07, "loss": -0.0144, "step": 238, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.20754413577165548, "train/R_vocab": 0.22721974206349208 }, { "learning_rate": 4.78239252557177e-07, "loss": 0.0564, "step": 240, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.2701327994702766, "train/R_vocab": 0.3259424603174603 }, { "learning_rate": 4.777950611472233e-07, "loss": -0.0099, "step": 242, "train/R_acc": 0.0, "train/R_penalty": -0.703125, "train/R_reason": 0.21945774918879585, "train/R_vocab": 0.2623883928571429 }, { "learning_rate": 4.773465924604656e-07, "loss": 0.0141, "step": 244, "train/R_acc": 0.065625, "train/R_penalty": -0.671875, "train/R_reason": 0.2201195420166109, "train/R_vocab": 0.26729910714285715 }, { "learning_rate": 4.768938549177392e-07, "loss": 0.052, "step": 246, "train/R_acc": 0.0, "train/R_penalty": -0.6875, "train/R_reason": 0.2231016124642528, "train/R_vocab": 0.26313244047619044 }, { "learning_rate": 4.764368570200353e-07, "loss": 0.0239, "step": 248, "train/R_acc": 0.003125, "train/R_penalty": -0.5625, "train/R_reason": 0.12548248807792622, "train/R_vocab": 0.15256696428571428 }, { "learning_rate": 4.759756073483411e-07, "loss": 0.0264, "step": 250, "train/R_acc": 0.0, "train/R_penalty": -0.765625, "train/R_reason": 0.17495784675781162, "train/R_vocab": 0.13359375 }, { "learning_rate": 4.7551011456347876e-07, "loss": 0.0271, "step": 252, "train/R_acc": 0.065625, "train/R_penalty": -0.759375, "train/R_reason": 0.2455444828427244, "train/R_vocab": 0.3765625 }, { "learning_rate": 4.750403874059428e-07, "loss": 0.035, "step": 254, "train/R_acc": 0.003125, "train/R_penalty": -0.61875, "train/R_reason": 0.2624393029582254, "train/R_vocab": 0.3123883928571428 }, { "learning_rate": 4.745664346957361e-07, "loss": -0.0167, "step": 256, "train/R_acc": 0.0, "train/R_penalty": -0.734375, "train/R_reason": 0.2061035486282405, "train/R_vocab": 0.2132688492063492 }, { "learning_rate": 4.740882653322039e-07, "loss": 0.0071, "step": 258, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.25927699094783563, "train/R_vocab": 0.3599330357142857 }, { "learning_rate": 4.7360588829386736e-07, "loss": -0.0231, "step": 260, "train/R_acc": 0.0, "train/R_penalty": -0.7125, "train/R_reason": 0.2040610708874982, "train/R_vocab": 0.23035714285714287 }, { "learning_rate": 4.7311931263825434e-07, "loss": 0.0593, "step": 262, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.14067559915832142, "train/R_vocab": 0.171875 }, { "learning_rate": 4.726285475017294e-07, "loss": 0.0213, "step": 264, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.2728806346759708, "train/R_vocab": 0.34206349206349207 }, { "learning_rate": 4.721336020993228e-07, "loss": -0.0176, "step": 266, "train/R_acc": 0.003125, "train/R_penalty": -0.6875, "train/R_reason": 0.22780919072556438, "train/R_vocab": 0.26149553571428574 }, { "learning_rate": 4.716344857245567e-07, "loss": 0.0807, "step": 268, "train/R_acc": 0.0, "train/R_penalty": -0.61875, "train/R_reason": 0.23184996778054867, "train/R_vocab": 0.27254464285714286 }, { "learning_rate": 4.7113120774927136e-07, "loss": 0.0114, "step": 270, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.24027473582911105, "train/R_vocab": 0.41076388888888893 }, { "learning_rate": 4.706237776234486e-07, "loss": 0.0187, "step": 272, "train/R_acc": 0.03125, "train/R_penalty": -0.609375, "train/R_reason": 0.1528155999223981, "train/R_vocab": 0.1519097222222222 }, { "learning_rate": 4.7011220487503476e-07, "loss": 0.0688, "step": 274, "train/R_acc": 0.034375, "train/R_penalty": -0.609375, "train/R_reason": 0.15041989042686232, "train/R_vocab": 0.1591889880952381 }, { "learning_rate": 4.695964991097616e-07, "loss": 0.0498, "step": 276, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.20384797152303086, "train/R_vocab": 0.35360863095238093 }, { "learning_rate": 4.6907667001096585e-07, "loss": -0.0992, "step": 278, "train/R_acc": 0.00625, "train/R_penalty": -0.578125, "train/R_reason": 0.23776322399396135, "train/R_vocab": 0.4219246031746032 }, { "learning_rate": 4.685527273394078e-07, "loss": -0.0568, "step": 280, "train/R_acc": 0.003125, "train/R_penalty": -0.5625, "train/R_reason": 0.2305205389308027, "train/R_vocab": 0.21640624999999997 }, { "learning_rate": 4.680246809330874e-07, "loss": -0.0174, "step": 282, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.27570159502906066, "train/R_vocab": 0.23663194444444444 }, { "learning_rate": 4.6749254070706013e-07, "loss": -0.053, "step": 284, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.17788848371405203, "train/R_vocab": 0.159375 }, { "learning_rate": 4.669563166532503e-07, "loss": 0.0595, "step": 286, "train/R_acc": 0.03125, "train/R_penalty": -0.6875, "train/R_reason": 0.22202849960365972, "train/R_vocab": 0.2922619047619048 }, { "learning_rate": 4.6641601884026407e-07, "loss": -0.0247, "step": 288, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.20278024644809003, "train/R_vocab": 0.2072544642857143 }, { "learning_rate": 4.6587165741319967e-07, "loss": 0.0055, "step": 290, "train/R_acc": 0.0, "train/R_penalty": -0.71875, "train/R_reason": 0.20358950838718215, "train/R_vocab": 0.30781250000000004 }, { "learning_rate": 4.6532324259345743e-07, "loss": 0.013, "step": 292, "train/R_acc": 0.0, "train/R_penalty": -0.68125, "train/R_reason": 0.19116956517383865, "train/R_vocab": 0.19854910714285712 }, { "learning_rate": 4.647707846785477e-07, "loss": -0.0187, "step": 294, "train/R_acc": 0.00625, "train/R_penalty": -0.71875, "train/R_reason": 0.18667014063523668, "train/R_vocab": 0.27327178030303034 }, { "learning_rate": 4.642142940418973e-07, "loss": 0.0376, "step": 296, "train/R_acc": 0.034375, "train/R_penalty": -0.546875, "train/R_reason": 0.193981736932492, "train/R_vocab": 0.19051339285714286 }, { "learning_rate": 4.6365378113265505e-07, "loss": -0.0587, "step": 298, "train/R_acc": 0.003125, "train/R_penalty": -0.76875, "train/R_reason": 0.21676354110088458, "train/R_vocab": 0.2875 }, { "learning_rate": 4.630892564754956e-07, "loss": 0.0012, "step": 300, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.24459838346088703, "train/R_vocab": 0.33515625 }, { "eval/R_acc": 0.001388888888888889, "eval/R_penalty": -0.6499999999999999, "eval/R_reason": 0.2340619843391708, "eval/R_vocab": 0.2772872574955908, "step": 300 }, { "step": 300 }, { "learning_rate": 4.6252073067042127e-07, "loss": -0.0006, "step": 302, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.20949847516693765, "train/R_vocab": 0.24609375 }, { "learning_rate": 4.6194821439256373e-07, "loss": 0.0001, "step": 304, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.20496328530187652, "train/R_vocab": 0.20011160714285714 }, { "learning_rate": 4.6137171839198297e-07, "loss": 0.0792, "step": 306, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.1994046532974053, "train/R_vocab": 0.2868551587301587 }, { "learning_rate": 4.6079125349346576e-07, "loss": 0.0068, "step": 308, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.18547339942836427, "train/R_vocab": 0.2426711309523809 }, { "learning_rate": 4.602068305963224e-07, "loss": 0.0044, "step": 310, "train/R_acc": 0.0, "train/R_penalty": -0.71875, "train/R_reason": 0.2112060437994439, "train/R_vocab": 0.2083829365079365 }, { "learning_rate": 4.59618460674182e-07, "loss": 0.0279, "step": 312, "train/R_acc": 0.034375, "train/R_penalty": -0.53125, "train/R_reason": 0.272254341125526, "train/R_vocab": 0.45736607142857144 }, { "learning_rate": 4.5902615477478636e-07, "loss": -0.0181, "step": 314, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.1449225433066495, "train/R_vocab": 0.16674107142857142 }, { "learning_rate": 4.5842992401978256e-07, "loss": 0.0225, "step": 316, "train/R_acc": 0.0, "train/R_penalty": -0.828125, "train/R_reason": 0.2286200296039895, "train/R_vocab": 0.2551339285714286 }, { "learning_rate": 4.5782977960451414e-07, "loss": 0.0192, "step": 318, "train/R_acc": 0.003125, "train/R_penalty": -0.68125, "train/R_reason": 0.21916647791965457, "train/R_vocab": 0.2505580357142857 }, { "learning_rate": 4.57225732797811e-07, "loss": 0.0485, "step": 320, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.22755603236630983, "train/R_vocab": 0.22947668650793654 }, { "learning_rate": 4.566177949417777e-07, "loss": 0.0163, "step": 322, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.22636069312789864, "train/R_vocab": 0.21104910714285716 }, { "learning_rate": 4.560059774515804e-07, "loss": 0.0143, "step": 324, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.15769657140460813, "train/R_vocab": 0.22265625000000003 }, { "learning_rate": 4.5539029181523284e-07, "loss": -0.0796, "step": 326, "train/R_acc": 0.03125, "train/R_penalty": -0.75, "train/R_reason": 0.1961929211100274, "train/R_vocab": 0.27142857142857146 }, { "learning_rate": 4.5477074959338015e-07, "loss": 0.0627, "step": 328, "train/R_acc": 0.003125, "train/R_penalty": -0.75, "train/R_reason": 0.22035420658260274, "train/R_vocab": 0.33661954365079366 }, { "learning_rate": 4.5414736241908214e-07, "loss": -0.015, "step": 330, "train/R_acc": 0.0, "train/R_penalty": -0.703125, "train/R_reason": 0.21972588070640325, "train/R_vocab": 0.30012400793650795 }, { "learning_rate": 4.535201419975948e-07, "loss": 0.0515, "step": 332, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.46875, "train/R_reason": 0.17121075988051718, "train/R_vocab": 0.234375 }, { "learning_rate": 4.5288910010615053e-07, "loss": 0.0479, "step": 334, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.5625, "train/R_reason": 0.21868874143295755, "train/R_vocab": 0.2630580357142857 }, { "learning_rate": 4.5225424859373684e-07, "loss": 0.0442, "step": 336, "train/R_acc": 0.00625, "train/R_penalty": -0.578125, "train/R_reason": 0.22348722064953433, "train/R_vocab": 0.3254464285714286 }, { "learning_rate": 4.51615599380874e-07, "loss": 0.0299, "step": 338, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.18059697532933602, "train/R_vocab": 0.21510416666666665 }, { "learning_rate": 4.5097316445939124e-07, "loss": 0.0359, "step": 340, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.1941301451022681, "train/R_vocab": 0.32220982142857146 }, { "learning_rate": 4.503269558922015e-07, "loss": -0.0107, "step": 342, "train/R_acc": 0.034375, "train/R_penalty": -0.625, "train/R_reason": 0.21181567459611045, "train/R_vocab": 0.2482514880952381 }, { "learning_rate": 4.496769858130748e-07, "loss": 0.0179, "step": 344, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.23230602042254933, "train/R_vocab": 0.3271701388888889 }, { "learning_rate": 4.490232664264109e-07, "loss": 0.042, "step": 346, "train/R_acc": 0.03125, "train/R_penalty": -0.696875, "train/R_reason": 0.2775747137079635, "train/R_vocab": 0.3890625 }, { "learning_rate": 4.4836581000700944e-07, "loss": 0.0274, "step": 348, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.22799018822395078, "train/R_vocab": 0.27528521825396823 }, { "learning_rate": 4.477046288998401e-07, "loss": 0.0259, "step": 350, "train/R_acc": 0.13125, "train/R_penalty": -0.546875, "train/R_reason": 0.20698908553226353, "train/R_vocab": 0.29140625 }, { "learning_rate": 4.470397355198102e-07, "loss": 0.0362, "step": 352, "train/R_acc": 0.0125, "train/R_penalty": -0.525, "train/R_reason": 0.210767134421379, "train/R_vocab": 0.3446180555555556 }, { "learning_rate": 4.463711423515323e-07, "loss": 0.0077, "step": 354, "train/R_acc": 0.0, "train/R_penalty": -0.796875, "train/R_reason": 0.16874434359341295, "train/R_vocab": 0.20223214285714286 }, { "learning_rate": 4.456988619490889e-07, "loss": 0.0364, "step": 356, "train/R_acc": 0.00625, "train/R_penalty": -0.665625, "train/R_reason": 0.2266184174999622, "train/R_vocab": 0.32075892857142857 }, { "learning_rate": 4.450229069357977e-07, "loss": 0.0624, "step": 358, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.19035143640898555, "train/R_vocab": 0.1884548611111111 }, { "learning_rate": 4.4434329000397363e-07, "loss": -0.0017, "step": 360, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.18822223579328265, "train/R_vocab": 0.2318452380952381 }, { "learning_rate": 4.4366002391469126e-07, "loss": 0.0383, "step": 362, "train/R_acc": 0.034375, "train/R_penalty": -0.71875, "train/R_reason": 0.18399620476664255, "train/R_vocab": 0.17989831349206348 }, { "learning_rate": 4.4297312149754477e-07, "loss": 0.0709, "step": 364, "train/R_acc": 0.03125, "train/R_penalty": -0.65625, "train/R_reason": 0.2374170544763758, "train/R_vocab": 0.2957589285714286 }, { "learning_rate": 4.422825956504072e-07, "loss": 0.012, "step": 366, "train/R_acc": 0.034375, "train/R_penalty": -0.775, "train/R_reason": 0.18171410574759275, "train/R_vocab": 0.27645089285714286 }, { "learning_rate": 4.415884593391882e-07, "loss": -0.0291, "step": 368, "train/R_acc": 0.0, "train/R_penalty": -0.728125, "train/R_reason": 0.20255323919085794, "train/R_vocab": 0.3200396825396825 }, { "learning_rate": 4.4089072559759065e-07, "loss": 0.0228, "step": 370, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.25184199643437555, "train/R_vocab": 0.31328125 }, { "learning_rate": 4.40189407526866e-07, "loss": -0.022, "step": 372, "train/R_acc": 0.03125, "train/R_penalty": -0.65625, "train/R_reason": 0.21868958770018052, "train/R_vocab": 0.21980406746031747 }, { "learning_rate": 4.3948451829556775e-07, "loss": 0.0437, "step": 374, "train/R_acc": 0.0, "train/R_penalty": -0.603125, "train/R_reason": 0.23191700315429237, "train/R_vocab": 0.25416666666666665 }, { "learning_rate": 4.3877607113930516e-07, "loss": -0.0032, "step": 376, "train/R_acc": 0.0, "train/R_penalty": -0.5, "train/R_reason": 0.1725605699858892, "train/R_vocab": 0.23932291666666666 }, { "learning_rate": 4.380640793604938e-07, "loss": -0.0522, "step": 378, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.1570635238253007, "train/R_vocab": 0.209375 }, { "learning_rate": 4.373485563281062e-07, "loss": 0.0522, "step": 380, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.25503011482100246, "train/R_vocab": 0.2538690476190476 }, { "learning_rate": 4.3662951547742075e-07, "loss": 0.0447, "step": 382, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.1739243830311305, "train/R_vocab": 0.23214285714285715 }, { "learning_rate": 4.3590697030976965e-07, "loss": -0.006, "step": 384, "train/R_acc": 0.03125, "train/R_penalty": -0.65625, "train/R_reason": 0.15723816285274234, "train/R_vocab": 0.2829996392496392 }, { "learning_rate": 4.3518093439228474e-07, "loss": 0.0162, "step": 386, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.16239281076258544, "train/R_vocab": 0.1482142857142857 }, { "learning_rate": 4.3445142135764367e-07, "loss": -0.0038, "step": 388, "train/R_acc": 0.034375, "train/R_penalty": -0.6875, "train/R_reason": 0.19918031683009202, "train/R_vocab": 0.27239583333333334 }, { "learning_rate": 4.33718444903813e-07, "loss": 0.0341, "step": 390, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.20130303529176516, "train/R_vocab": 0.29593253968253963 }, { "learning_rate": 4.329820187937919e-07, "loss": 0.0876, "step": 392, "train/R_acc": 0.034375, "train/R_penalty": -0.5, "train/R_reason": 0.23979456377125202, "train/R_vocab": 0.24832589285714285 }, { "learning_rate": 4.3224215685535287e-07, "loss": 0.0139, "step": 394, "train/R_acc": 0.0, "train/R_penalty": -0.8125, "train/R_reason": 0.1709315188515677, "train/R_vocab": 0.18359375 }, { "learning_rate": 4.314988729807827e-07, "loss": -0.021, "step": 396, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.17646043188459717, "train/R_vocab": 0.17589285714285713 }, { "learning_rate": 4.3075218112662135e-07, "loss": 0.0513, "step": 398, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.1973938417696664, "train/R_vocab": 0.19837549603174603 }, { "learning_rate": 4.3000209531339996e-07, "loss": 0.0043, "step": 400, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.24184598292042236, "train/R_vocab": 0.33288690476190474 }, { "eval/R_acc": 0.0, "eval/R_penalty": -0.7916666666666666, "eval/R_reason": 0.27821714712822276, "eval/R_vocab": 0.32563381834215166, "step": 400 }, { "step": 400 }, { "learning_rate": 4.2943367562278896e-07, "loss": -0.0155, "step": 402, "train/R_acc": 0.065625, "train/R_penalty": -0.515625, "train/R_reason": 0.20335033205254746, "train/R_vocab": 0.24464285714285713 }, { "learning_rate": 4.286787117443108e-07, "loss": -0.036, "step": 404, "train/R_acc": 0.04062500000000001, "train/R_penalty": -0.5625, "train/R_reason": 0.20780436238210176, "train/R_vocab": 0.2703993055555556 }, { "learning_rate": 4.2792040207614e-07, "loss": -0.0044, "step": 406, "train/R_acc": 0.03125, "train/R_penalty": -0.734375, "train/R_reason": 0.2366017128148703, "train/R_vocab": 0.3632440476190476 }, { "learning_rate": 4.27158760817756e-07, "loss": 0.0245, "step": 408, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.19598878934576208, "train/R_vocab": 0.28031994047619047 }, { "learning_rate": 4.263938022310226e-07, "loss": -0.036, "step": 410, "train/R_acc": 0.034375, "train/R_penalty": -0.671875, "train/R_reason": 0.23507477335166782, "train/R_vocab": 0.27189980158730165 }, { "learning_rate": 4.2562554063992127e-07, "loss": 0.024, "step": 412, "train/R_acc": 0.0625, "train/R_penalty": -0.7125, "train/R_reason": 0.19770138255909644, "train/R_vocab": 0.24754464285714284 }, { "learning_rate": 4.248539904302829e-07, "loss": 0.0091, "step": 414, "train/R_acc": 0.00625, "train/R_penalty": -0.59375, "train/R_reason": 0.2098416268444374, "train/R_vocab": 0.2533234126984127 }, { "learning_rate": 4.240791660495182e-07, "loss": 0.0125, "step": 416, "train/R_acc": 0.003125, "train/R_penalty": -0.603125, "train/R_reason": 0.164986564236729, "train/R_vocab": 0.1863219246031746 }, { "learning_rate": 4.2330108200634723e-07, "loss": 0.0353, "step": 418, "train/R_acc": 0.03125, "train/R_penalty": -0.703125, "train/R_reason": 0.16427460147948922, "train/R_vocab": 0.1800595238095238 }, { "learning_rate": 4.22519752870528e-07, "loss": -0.0002, "step": 420, "train/R_acc": 0.003125, "train/R_penalty": -0.703125, "train/R_reason": 0.21440123784001752, "train/R_vocab": 0.26026785714285716 }, { "learning_rate": 4.2173519327258325e-07, "loss": 0.0336, "step": 422, "train/R_acc": 0.003125, "train/R_penalty": -0.59375, "train/R_reason": 0.2248050023022144, "train/R_vocab": 0.22937184343434344 }, { "learning_rate": 4.2094741790352673e-07, "loss": 0.0256, "step": 424, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.22803044267397005, "train/R_vocab": 0.28350694444444446 }, { "learning_rate": 4.2015644151458827e-07, "loss": -0.0319, "step": 426, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.20821167099100335, "train/R_vocab": 0.21744791666666666 }, { "learning_rate": 4.19362278916937e-07, "loss": 0.0344, "step": 428, "train/R_acc": 0.03125, "train/R_penalty": -0.671875, "train/R_reason": 0.1891472011902095, "train/R_vocab": 0.20641233766233769 }, { "learning_rate": 4.185649449814045e-07, "loss": 0.0712, "step": 430, "train/R_acc": 0.00625, "train/R_penalty": -0.6875, "train/R_reason": 0.2279409653415437, "train/R_vocab": 0.2996651785714286 }, { "learning_rate": 4.177644546382063e-07, "loss": 0.0645, "step": 432, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.625, "train/R_reason": 0.21120852841137872, "train/R_vocab": 0.19676339285714284 }, { "learning_rate": 4.1696082287666217e-07, "loss": 0.0481, "step": 434, "train/R_acc": 0.003125, "train/R_penalty": -0.59375, "train/R_reason": 0.20179254286468842, "train/R_vocab": 0.3659598214285714 }, { "learning_rate": 4.1615406474491535e-07, "loss": -0.0453, "step": 436, "train/R_acc": 0.034375, "train/R_penalty": -0.78125, "train/R_reason": 0.21685467943636602, "train/R_vocab": 0.2571428571428571 }, { "learning_rate": 4.1534419534965105e-07, "loss": 0.0985, "step": 438, "train/R_acc": 0.03125, "train/R_penalty": -0.696875, "train/R_reason": 0.20741733073153354, "train/R_vocab": 0.27142857142857146 }, { "learning_rate": 4.145312298558133e-07, "loss": -0.0621, "step": 440, "train/R_acc": 0.003125, "train/R_penalty": -0.5625, "train/R_reason": 0.17062418658581344, "train/R_vocab": 0.2 }, { "learning_rate": 4.137151834863213e-07, "loss": -0.0448, "step": 442, "train/R_acc": 0.03125, "train/R_penalty": -0.5, "train/R_reason": 0.23174115282287933, "train/R_vocab": 0.36713789682539677 }, { "learning_rate": 4.128960715217839e-07, "loss": 0.0187, "step": 444, "train/R_acc": 0.003125, "train/R_penalty": -0.59375, "train/R_reason": 0.15167707696946453, "train/R_vocab": 0.20052083333333334 }, { "learning_rate": 4.1207390930021394e-07, "loss": -0.0156, "step": 446, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.2000569888049995, "train/R_vocab": 0.24027777777777778 }, { "learning_rate": 4.11248712216741e-07, "loss": 0.0897, "step": 448, "train/R_acc": 0.0, "train/R_penalty": -0.68125, "train/R_reason": 0.21749146764720112, "train/R_vocab": 0.2636160714285714 }, { "learning_rate": 4.104204957233225e-07, "loss": 0.0032, "step": 450, "train/R_acc": 0.00625, "train/R_penalty": -0.609375, "train/R_reason": 0.19073873404612746, "train/R_vocab": 0.30033482142857143 }, { "learning_rate": 4.095892753284553e-07, "loss": -0.0985, "step": 452, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.25767563611959865, "train/R_vocab": 0.41929563492063493 }, { "learning_rate": 4.087550665968846e-07, "loss": 0.0021, "step": 454, "train/R_acc": 0.0, "train/R_penalty": -0.53125, "train/R_reason": 0.1972933583803858, "train/R_vocab": 0.23854166666666668 }, { "learning_rate": 4.079178851493127e-07, "loss": 0.0433, "step": 456, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.2107089160600319, "train/R_vocab": 0.25 }, { "learning_rate": 4.070777466621067e-07, "loss": -0.0103, "step": 458, "train/R_acc": 0.03125, "train/R_penalty": -0.515625, "train/R_reason": 0.2429741695318347, "train/R_vocab": 0.38958333333333334 }, { "learning_rate": 4.062346668670046e-07, "loss": -0.0092, "step": 460, "train/R_acc": 0.003125, "train/R_penalty": -0.734375, "train/R_reason": 0.2328364572562428, "train/R_vocab": 0.2795758928571429 }, { "learning_rate": 4.0538866155082094e-07, "loss": -0.0198, "step": 462, "train/R_acc": 0.040625, "train/R_penalty": -0.4625, "train/R_reason": 0.2483761704003692, "train/R_vocab": 0.2885168650793651 }, { "learning_rate": 4.045397465551513e-07, "loss": -0.0249, "step": 464, "train/R_acc": 0.003125, "train/R_penalty": -0.5875, "train/R_reason": 0.17307827497369785, "train/R_vocab": 0.2544642857142857 }, { "learning_rate": 4.036879377760752e-07, "loss": 0.0683, "step": 466, "train/R_acc": 0.03125, "train/R_penalty": -0.671875, "train/R_reason": 0.22330642900293768, "train/R_vocab": 0.37135416666666665 }, { "learning_rate": 4.02833251163859e-07, "loss": -0.0192, "step": 468, "train/R_acc": 0.003125, "train/R_penalty": -0.65, "train/R_reason": 0.21129365536420686, "train/R_vocab": 0.2589285714285714 }, { "learning_rate": 4.0197570272265704e-07, "loss": 0.0497, "step": 470, "train/R_acc": 0.003125, "train/R_penalty": -0.515625, "train/R_reason": 0.2052435879372262, "train/R_vocab": 0.2556547619047619 }, { "learning_rate": 4.011153085102116e-07, "loss": 0.0309, "step": 472, "train/R_acc": 0.003125, "train/R_penalty": -0.671875, "train/R_reason": 0.21023509045617045, "train/R_vocab": 0.2963169642857143 }, { "learning_rate": 4.0025208463755274e-07, "loss": 0.0424, "step": 474, "train/R_acc": 0.0, "train/R_penalty": -0.784375, "train/R_reason": 0.1959827825898999, "train/R_vocab": 0.35357142857142854 }, { "learning_rate": 3.9938604726869636e-07, "loss": -0.0506, "step": 476, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.19807364686139228, "train/R_vocab": 0.17491319444444442 }, { "learning_rate": 3.9851721262034157e-07, "loss": 0.0724, "step": 478, "train/R_acc": 0.003125, "train/R_penalty": -0.5, "train/R_reason": 0.20997218794964279, "train/R_vocab": 0.27433035714285714 }, { "learning_rate": 3.9764559696156697e-07, "loss": 0.0338, "step": 480, "train/R_acc": 0.003125, "train/R_penalty": -0.665625, "train/R_reason": 0.21410868507940445, "train/R_vocab": 0.25691964285714286 }, { "learning_rate": 3.9677121661352607e-07, "loss": -0.0087, "step": 482, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.21506411602738199, "train/R_vocab": 0.3130952380952381 }, { "learning_rate": 3.958940879491418e-07, "loss": -0.0375, "step": 484, "train/R_acc": 0.0, "train/R_penalty": -0.71875, "train/R_reason": 0.18895673088287218, "train/R_vocab": 0.21395089285714286 }, { "learning_rate": 3.9501422739279953e-07, "loss": -0.0485, "step": 486, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.20816202695918376, "train/R_vocab": 0.22013888888888888 }, { "learning_rate": 3.9413165142004e-07, "loss": -0.0479, "step": 488, "train/R_acc": 0.03125, "train/R_penalty": -0.5625, "train/R_reason": 0.1929871719837118, "train/R_vocab": 0.3703125 }, { "learning_rate": 3.932463765572505e-07, "loss": -0.0312, "step": 490, "train/R_acc": 0.0, "train/R_penalty": -0.75, "train/R_reason": 0.2513287746728735, "train/R_vocab": 0.27569444444444446 }, { "learning_rate": 3.923584193813555e-07, "loss": 0.03, "step": 492, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.23851168147176444, "train/R_vocab": 0.313219246031746 }, { "learning_rate": 3.914677965195062e-07, "loss": -0.0009, "step": 494, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.15729111795287057, "train/R_vocab": 0.1830357142857143 }, { "learning_rate": 3.9057452464876946e-07, "loss": 0.049, "step": 496, "train/R_acc": 0.003125, "train/R_penalty": -0.703125, "train/R_reason": 0.15483905838498802, "train/R_vocab": 0.18125000000000002 }, { "learning_rate": 3.89678620495815e-07, "loss": 0.0521, "step": 498, "train/R_acc": 0.0, "train/R_penalty": -0.75, "train/R_reason": 0.2600686346082261, "train/R_vocab": 0.253125 }, { "learning_rate": 3.887801008366025e-07, "loss": -0.005, "step": 500, "train/R_acc": 0.003125, "train/R_penalty": -0.790625, "train/R_reason": 0.22427641116364, "train/R_vocab": 0.3156125992063492 }, { "eval/R_acc": 0.0025, "eval/R_penalty": -0.7075, "eval/R_reason": 0.2610032668269429, "eval/R_vocab": 0.21906295093795097, "step": 500 }, { "step": 500 }, { "learning_rate": 3.8787898249606767e-07, "loss": 0.0252, "step": 502, "train/R_acc": 0.0625, "train/R_penalty": -0.609375, "train/R_reason": 0.2684211960918422, "train/R_vocab": 0.3091517857142857 }, { "learning_rate": 3.8697528234780674e-07, "loss": 0.019, "step": 504, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.23017420694962337, "train/R_vocab": 0.29288194444444443 }, { "learning_rate": 3.86069017313761e-07, "loss": 0.013, "step": 506, "train/R_acc": 0.03125, "train/R_penalty": -0.578125, "train/R_reason": 0.22614899753752782, "train/R_vocab": 0.3089905753968254 }, { "learning_rate": 3.851602043638994e-07, "loss": -0.0248, "step": 508, "train/R_acc": 0.034375, "train/R_penalty": -0.515625, "train/R_reason": 0.2017065165275432, "train/R_vocab": 0.26004464285714285 }, { "learning_rate": 3.8424886051590115e-07, "loss": 0.0296, "step": 510, "train/R_acc": 0.0, "train/R_penalty": -0.7375, "train/R_reason": 0.17507956245908884, "train/R_vocab": 0.17238343253968252 }, { "learning_rate": 3.83335002834837e-07, "loss": -0.0622, "step": 512, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.22332916829962618, "train/R_vocab": 0.34873511904761906 }, { "learning_rate": 3.8241864843284964e-07, "loss": -0.0075, "step": 514, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.132802326682096, "train/R_vocab": 0.16104910714285714 }, { "learning_rate": 3.814998144688333e-07, "loss": 0.0227, "step": 516, "train/R_acc": 0.034375, "train/R_penalty": -0.59375, "train/R_reason": 0.1754689583879656, "train/R_vocab": 0.19609374999999998 }, { "learning_rate": 3.805785181481123e-07, "loss": -0.0118, "step": 518, "train/R_acc": 0.003125, "train/R_penalty": -0.540625, "train/R_reason": 0.23541827525810197, "train/R_vocab": 0.25967261904761907 }, { "learning_rate": 3.796547767221193e-07, "loss": -0.0924, "step": 520, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.2376371992311745, "train/R_vocab": 0.3128968253968254 }, { "learning_rate": 3.787286074880718e-07, "loss": -0.0122, "step": 522, "train/R_acc": 0.0, "train/R_penalty": -0.55625, "train/R_reason": 0.22978465999581174, "train/R_vocab": 0.2617931547619048 }, { "learning_rate": 3.778000277886483e-07, "loss": -0.0054, "step": 524, "train/R_acc": 0.0, "train/R_penalty": -0.53125, "train/R_reason": 0.22061871520509158, "train/R_vocab": 0.30894209956709956 }, { "learning_rate": 3.768690550116639e-07, "loss": 0.0092, "step": 526, "train/R_acc": 0.03125, "train/R_penalty": -0.546875, "train/R_reason": 0.22469643175166784, "train/R_vocab": 0.34073660714285714 }, { "learning_rate": 3.7593570658974436e-07, "loss": 0.0264, "step": 528, "train/R_acc": 0.03125, "train/R_penalty": -0.640625, "train/R_reason": 0.26427900699025786, "train/R_vocab": 0.32444196428571426 }, { "learning_rate": 3.75e-07, "loss": 0.0264, "step": 530, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.16190914297300985, "train/R_vocab": 0.18397817460317462 }, { "learning_rate": 3.740619527636979e-07, "loss": -0.0014, "step": 532, "train/R_acc": 0.003125, "train/R_penalty": -0.540625, "train/R_reason": 0.2013910995332763, "train/R_vocab": 0.16294642857142855 }, { "learning_rate": 3.731215824459344e-07, "loss": 0.0653, "step": 534, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.640625, "train/R_reason": 0.1980956136240478, "train/R_vocab": 0.26245039682539684 }, { "learning_rate": 3.7217890665530587e-07, "loss": -0.0052, "step": 536, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.193118823374168, "train/R_vocab": 0.18813131313131315 }, { "learning_rate": 3.712339430435792e-07, "loss": 0.044, "step": 538, "train/R_acc": 0.00625, "train/R_penalty": -0.65625, "train/R_reason": 0.21065931054504308, "train/R_vocab": 0.2915054563492063 }, { "learning_rate": 3.7028670930536077e-07, "loss": 0.0443, "step": 540, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.21326431514536787, "train/R_vocab": 0.23958333333333334 }, { "learning_rate": 3.693372231777658e-07, "loss": 0.0062, "step": 542, "train/R_acc": 0.003125, "train/R_penalty": -0.86875, "train/R_reason": 0.2361538113041483, "train/R_vocab": 0.26145833333333335 }, { "learning_rate": 3.6838550244008573e-07, "loss": 0.0011, "step": 544, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.2560180166234486, "train/R_vocab": 0.3419642857142857 }, { "learning_rate": 3.6743156491345564e-07, "loss": -0.1304, "step": 546, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.22700866641314227, "train/R_vocab": 0.3409474206349206 }, { "learning_rate": 3.6647542846052003e-07, "loss": 0.0374, "step": 548, "train/R_acc": 0.003125, "train/R_penalty": -0.515625, "train/R_reason": 0.18355650415358743, "train/R_vocab": 0.23628472222222222 }, { "learning_rate": 3.65517110985099e-07, "loss": -0.0193, "step": 550, "train/R_acc": 0.0625, "train/R_penalty": -0.546875, "train/R_reason": 0.1497471083396625, "train/R_vocab": 0.1660342261904762 }, { "learning_rate": 3.645566304318526e-07, "loss": -0.0168, "step": 552, "train/R_acc": 0.00625, "train/R_penalty": -0.71875, "train/R_reason": 0.17760395766642206, "train/R_vocab": 0.20212053571428573 }, { "learning_rate": 3.6359400478594473e-07, "loss": 0.0519, "step": 554, "train/R_acc": 0.03125, "train/R_penalty": -0.68125, "train/R_reason": 0.2032957807067698, "train/R_vocab": 0.3547991071428571 }, { "learning_rate": 3.6262925207270666e-07, "loss": -0.0077, "step": 556, "train/R_acc": 0.034375, "train/R_penalty": -0.53125, "train/R_reason": 0.24608095001365107, "train/R_vocab": 0.2532366071428571 }, { "learning_rate": 3.616623903572994e-07, "loss": -0.0273, "step": 558, "train/R_acc": 0.03125, "train/R_penalty": -0.46875, "train/R_reason": 0.25668514206671234, "train/R_vocab": 0.32946428571428565 }, { "learning_rate": 3.6069343774437516e-07, "loss": 0.0356, "step": 560, "train/R_acc": 0.00625, "train/R_penalty": -0.853125, "train/R_reason": 0.22578641617093248, "train/R_vocab": 0.3117063492063492 }, { "learning_rate": 3.597224123777389e-07, "loss": 0.0074, "step": 562, "train/R_acc": 0.00625, "train/R_penalty": -0.61875, "train/R_reason": 0.1608838946497146, "train/R_vocab": 0.22940228174603175 }, { "learning_rate": 3.58749332440008e-07, "loss": 0.0218, "step": 564, "train/R_acc": 0.0, "train/R_penalty": -0.5, "train/R_reason": 0.2042184580919646, "train/R_vocab": 0.23277529761904764 }, { "learning_rate": 3.5777421615227207e-07, "loss": 0.0446, "step": 566, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.21492820594376136, "train/R_vocab": 0.2902901785714286 }, { "learning_rate": 3.567970817737518e-07, "loss": 0.0771, "step": 568, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.2497326839973281, "train/R_vocab": 0.30859375 }, { "learning_rate": 3.5581794760145696e-07, "loss": 0.0354, "step": 570, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.20181240026890324, "train/R_vocab": 0.22098214285714285 }, { "learning_rate": 3.548368319698437e-07, "loss": 0.0731, "step": 572, "train/R_acc": 0.00625, "train/R_penalty": -0.696875, "train/R_reason": 0.16856543760413237, "train/R_vocab": 0.1800595238095238 }, { "learning_rate": 3.5385375325047163e-07, "loss": 0.0291, "step": 574, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.2099860707148752, "train/R_vocab": 0.2753348214285714 }, { "learning_rate": 3.528687298516591e-07, "loss": -0.0002, "step": 576, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.17694118112711849, "train/R_vocab": 0.12705853174603174 }, { "learning_rate": 3.5188178021813925e-07, "loss": 0.028, "step": 578, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.609375, "train/R_reason": 0.24354058691683123, "train/R_vocab": 0.39140624999999996 }, { "learning_rate": 3.5089292283071417e-07, "loss": 0.0398, "step": 580, "train/R_acc": 0.00625, "train/R_penalty": -0.53125, "train/R_reason": 0.22343115261429752, "train/R_vocab": 0.32678571428571423 }, { "learning_rate": 3.499021762059089e-07, "loss": 0.0565, "step": 582, "train/R_acc": 0.03125, "train/R_penalty": -0.7125, "train/R_reason": 0.19496747632032374, "train/R_vocab": 0.26284722222222223 }, { "learning_rate": 3.489095588956249e-07, "loss": 0.0092, "step": 584, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.2074570772291119, "train/R_vocab": 0.30972222222222223 }, { "learning_rate": 3.479150894867926e-07, "loss": -0.0015, "step": 586, "train/R_acc": 0.0625, "train/R_penalty": -0.703125, "train/R_reason": 0.20443537938998835, "train/R_vocab": 0.2761904761904762 }, { "learning_rate": 3.46918786601023e-07, "loss": 0.1059, "step": 588, "train/R_acc": 0.0, "train/R_penalty": -0.68125, "train/R_reason": 0.20239388628972388, "train/R_vocab": 0.22630208333333335 }, { "learning_rate": 3.459206688942596e-07, "loss": 0.0608, "step": 590, "train/R_acc": 0.0, "train/R_penalty": -0.75, "train/R_reason": 0.24682579728602413, "train/R_vocab": 0.4008928571428571 }, { "learning_rate": 3.4492075505642847e-07, "loss": 0.055, "step": 592, "train/R_acc": 0.0, "train/R_penalty": -0.5, "train/R_reason": 0.19593385729165314, "train/R_vocab": 0.31875 }, { "learning_rate": 3.439190638110888e-07, "loss": 0.041, "step": 594, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.1711366333418505, "train/R_vocab": 0.26268601190476193 }, { "learning_rate": 3.4291561391508185e-07, "loss": 0.0807, "step": 596, "train/R_acc": 0.003125, "train/R_penalty": -0.59375, "train/R_reason": 0.240933143584435, "train/R_vocab": 0.4119791666666667 }, { "learning_rate": 3.4191042415818e-07, "loss": -0.0247, "step": 598, "train/R_acc": 0.003125, "train/R_penalty": -0.74375, "train/R_reason": 0.2539401489710603, "train/R_vocab": 0.2673363095238095 }, { "learning_rate": 3.4090351336273474e-07, "loss": -0.0238, "step": 600, "train/R_acc": 0.0, "train/R_penalty": -0.703125, "train/R_reason": 0.19356852065501937, "train/R_vocab": 0.24319196428571427 }, { "eval/R_acc": 0.005, "eval/R_penalty": -0.625, "eval/R_reason": 0.24855003165355338, "eval/R_vocab": 0.22628517316017316, "step": 600 }, { "step": 600 }, { "learning_rate": 3.398949003833246e-07, "loss": 0.0122, "step": 602, "train/R_acc": 0.00625, "train/R_penalty": -0.578125, "train/R_reason": 0.22168416427879684, "train/R_vocab": 0.24791666666666667 }, { "learning_rate": 3.388846041064012e-07, "loss": 0.0797, "step": 604, "train/R_acc": 0.09375, "train/R_penalty": -0.546875, "train/R_reason": 0.1777229928874741, "train/R_vocab": 0.29464285714285715 }, { "learning_rate": 3.378726434499368e-07, "loss": 0.033, "step": 606, "train/R_acc": 0.0, "train/R_penalty": -0.7125, "train/R_reason": 0.1956222769670857, "train/R_vocab": 0.3197916666666667 }, { "learning_rate": 3.368590373630692e-07, "loss": -0.0275, "step": 608, "train/R_acc": 0.06875, "train/R_penalty": -0.46875, "train/R_reason": 0.17503223220226105, "train/R_vocab": 0.2591517857142857 }, { "learning_rate": 3.3584380482574717e-07, "loss": -0.058, "step": 610, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.1902514010231657, "train/R_vocab": 0.23125 }, { "learning_rate": 3.348269648483749e-07, "loss": -0.054, "step": 612, "train/R_acc": 0.0, "train/R_penalty": -0.53125, "train/R_reason": 0.17467640906171372, "train/R_vocab": 0.190625 }, { "learning_rate": 3.3380853647145656e-07, "loss": -0.0371, "step": 614, "train/R_acc": 0.00625, "train/R_penalty": -0.65625, "train/R_reason": 0.17121010785732182, "train/R_vocab": 0.2569444444444444 }, { "learning_rate": 3.327885387652391e-07, "loss": -0.0013, "step": 616, "train/R_acc": 0.0, "train/R_penalty": -0.6875, "train/R_reason": 0.19415455674882087, "train/R_vocab": 0.23277529761904764 }, { "learning_rate": 3.317669908293554e-07, "loss": 0.0253, "step": 618, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.24193135939319937, "train/R_vocab": 0.25 }, { "learning_rate": 3.307439117924668e-07, "loss": 0.0588, "step": 620, "train/R_acc": 0.0, "train/R_penalty": -0.6875, "train/R_reason": 0.20416377806011923, "train/R_vocab": 0.2056423611111111 }, { "learning_rate": 3.297193208119047e-07, "loss": -0.0064, "step": 622, "train/R_acc": 0.0, "train/R_penalty": -0.484375, "train/R_reason": 0.211192554787049, "train/R_vocab": 0.30390625000000004 }, { "learning_rate": 3.2869323707331176e-07, "loss": 0.0335, "step": 624, "train/R_acc": 0.003125, "train/R_penalty": -0.59375, "train/R_reason": 0.23403307094336184, "train/R_vocab": 0.35301339285714284 }, { "learning_rate": 3.2766567979028324e-07, "loss": -0.0308, "step": 626, "train/R_acc": 0.00625, "train/R_penalty": -0.578125, "train/R_reason": 0.25211879339329, "train/R_vocab": 0.278125 }, { "learning_rate": 3.2663666820400625e-07, "loss": 0.0239, "step": 628, "train/R_acc": 0.03125, "train/R_penalty": -0.609375, "train/R_reason": 0.23123528000410434, "train/R_vocab": 0.328844246031746 }, { "learning_rate": 3.2560622158290025e-07, "loss": 0.0127, "step": 630, "train/R_acc": 0.003125, "train/R_penalty": -0.5, "train/R_reason": 0.22103995585722944, "train/R_vocab": 0.26785714285714285 }, { "learning_rate": 3.2457435922225603e-07, "loss": 0.0445, "step": 632, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.185424618889319, "train/R_vocab": 0.28928571428571426 }, { "learning_rate": 3.235411004438741e-07, "loss": 0.1006, "step": 634, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.23390812732295468, "train/R_vocab": 0.29355158730158726 }, { "learning_rate": 3.2250646459570343e-07, "loss": 0.0611, "step": 636, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.17366692788791044, "train/R_vocab": 0.21026785714285715 }, { "learning_rate": 3.214704710514786e-07, "loss": 0.0463, "step": 638, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.23856693280209076, "train/R_vocab": 0.3723214285714286 }, { "learning_rate": 3.204331392103574e-07, "loss": -0.0048, "step": 640, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.2248622406893548, "train/R_vocab": 0.38828125 }, { "learning_rate": 3.193944884965576e-07, "loss": 0.0172, "step": 642, "train/R_acc": 0.065625, "train/R_penalty": -0.671875, "train/R_reason": 0.2000315806348117, "train/R_vocab": 0.32477678571428575 }, { "learning_rate": 3.183545383589927e-07, "loss": -0.0164, "step": 644, "train/R_acc": 0.003125, "train/R_penalty": -0.55625, "train/R_reason": 0.1901960813142597, "train/R_vocab": 0.2540178571428572 }, { "learning_rate": 3.173133082709086e-07, "loss": 0.0815, "step": 646, "train/R_acc": 0.003125, "train/R_penalty": -0.68125, "train/R_reason": 0.20267428182016378, "train/R_vocab": 0.2115079365079365 }, { "learning_rate": 3.1627081772951815e-07, "loss": 0.0279, "step": 648, "train/R_acc": 0.03125, "train/R_penalty": -0.484375, "train/R_reason": 0.18044257928711105, "train/R_vocab": 0.21806795634920634 }, { "learning_rate": 3.152270862556367e-07, "loss": 0.0707, "step": 650, "train/R_acc": 0.003125, "train/R_penalty": -0.515625, "train/R_reason": 0.26331484572063035, "train/R_vocab": 0.24206349206349206 }, { "learning_rate": 3.1418213339331576e-07, "loss": -0.0396, "step": 652, "train/R_acc": 0.003125, "train/R_penalty": -0.53125, "train/R_reason": 0.23778099952829604, "train/R_vocab": 0.284375 }, { "learning_rate": 3.1313597870947816e-07, "loss": 0.0883, "step": 654, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.22377350102735885, "train/R_vocab": 0.3117931547619047 }, { "learning_rate": 3.1208864179355074e-07, "loss": 0.0009, "step": 656, "train/R_acc": 0.03125, "train/R_penalty": -0.571875, "train/R_reason": 0.20656730207307256, "train/R_vocab": 0.2512276785714286 }, { "learning_rate": 3.1104014225709784e-07, "loss": 0.0828, "step": 658, "train/R_acc": 0.003125, "train/R_penalty": -0.65, "train/R_reason": 0.2410817269424846, "train/R_vocab": 0.265625 }, { "learning_rate": 3.099904997334541e-07, "loss": -0.0317, "step": 660, "train/R_acc": 0.03125, "train/R_penalty": -0.5875, "train/R_reason": 0.25759275591542546, "train/R_vocab": 0.24017857142857144 }, { "learning_rate": 3.0893973387735683e-07, "loss": -0.0235, "step": 662, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.16711049206280193, "train/R_vocab": 0.21428571428571427 }, { "learning_rate": 3.078878643645778e-07, "loss": 0.0242, "step": 664, "train/R_acc": 0.00625, "train/R_penalty": -0.696875, "train/R_reason": 0.22549504519978125, "train/R_vocab": 0.33912450396825394 }, { "learning_rate": 3.068349108915553e-07, "loss": 0.0147, "step": 666, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.2824812700031851, "train/R_vocab": 0.24285714285714285 }, { "learning_rate": 3.0578089317502436e-07, "loss": -0.0134, "step": 668, "train/R_acc": 0.034375, "train/R_penalty": -0.525, "train/R_reason": 0.1605276972684076, "train/R_vocab": 0.22779017857142858 }, { "learning_rate": 3.0472583095164873e-07, "loss": -0.0496, "step": 670, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.19456119250844006, "train/R_vocab": 0.23214285714285715 }, { "learning_rate": 3.036697439776504e-07, "loss": 0.0554, "step": 672, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.2515289920392415, "train/R_vocab": 0.2724330357142857 }, { "learning_rate": 3.0261265202844016e-07, "loss": 0.02, "step": 674, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.18580953264755184, "train/R_vocab": 0.25388144841269844 }, { "learning_rate": 3.01554574898247e-07, "loss": 0.0155, "step": 676, "train/R_acc": 0.0, "train/R_penalty": -0.55625, "train/R_reason": 0.22948436053810994, "train/R_vocab": 0.22013888888888888 }, { "learning_rate": 3.004955323997478e-07, "loss": -0.0179, "step": 678, "train/R_acc": 0.00625, "train/R_penalty": -0.59375, "train/R_reason": 0.21153443586764287, "train/R_vocab": 0.23524305555555558 }, { "learning_rate": 2.994355443636958e-07, "loss": -0.0362, "step": 680, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.1840845890109734, "train/R_vocab": 0.31015625 }, { "learning_rate": 2.983746306385499e-07, "loss": 0.0297, "step": 682, "train/R_acc": 0.0625, "train/R_penalty": -0.453125, "train/R_reason": 0.2084560166468069, "train/R_vocab": 0.31228918650793647 }, { "learning_rate": 2.9731281109010253e-07, "loss": 0.0534, "step": 684, "train/R_acc": 0.0, "train/R_penalty": -0.71875, "train/R_reason": 0.16422497838360026, "train/R_vocab": 0.23582589285714284 }, { "learning_rate": 2.9625010560110787e-07, "loss": 0.0183, "step": 686, "train/R_acc": 0.003125, "train/R_penalty": -0.634375, "train/R_reason": 0.24390149995516267, "train/R_vocab": 0.29910714285714285 }, { "learning_rate": 2.951865340709095e-07, "loss": 0.0683, "step": 688, "train/R_acc": 0.034375, "train/R_penalty": -0.64375, "train/R_reason": 0.23521821731229603, "train/R_vocab": 0.3359747023809524 }, { "learning_rate": 2.9412211641506773e-07, "loss": -0.001, "step": 690, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.18437836963522442, "train/R_vocab": 0.209375 }, { "learning_rate": 2.930568725649867e-07, "loss": 0.0194, "step": 692, "train/R_acc": 0.03125, "train/R_penalty": -0.65, "train/R_reason": 0.21411232248047601, "train/R_vocab": 0.2759548611111111 }, { "learning_rate": 2.919908224675412e-07, "loss": 0.0579, "step": 694, "train/R_acc": 0.003125, "train/R_penalty": -0.5625, "train/R_reason": 0.1540599514957384, "train/R_vocab": 0.18344494047619048 }, { "learning_rate": 2.9092398608470326e-07, "loss": -0.0106, "step": 696, "train/R_acc": 0.03125, "train/R_penalty": -0.59375, "train/R_reason": 0.2501350490350204, "train/R_vocab": 0.3291294642857143 }, { "learning_rate": 2.898563833931681e-07, "loss": 0.0156, "step": 698, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.20865047120336916, "train/R_vocab": 0.2130332341269841 }, { "learning_rate": 2.8878803438398015e-07, "loss": 0.0192, "step": 700, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.19110667770899964, "train/R_vocab": 0.24427083333333333 }, { "eval/R_acc": 0.0025, "eval/R_penalty": -0.5875, "eval/R_reason": 0.25186260885534084, "eval/R_vocab": 0.2177263708513708, "step": 700 }, { "step": 700 }, { "learning_rate": 2.8771895906215895e-07, "loss": 0.05, "step": 702, "train/R_acc": 0.003125, "train/R_penalty": -0.53125, "train/R_reason": 0.2448799886759996, "train/R_vocab": 0.3176339285714286 }, { "learning_rate": 2.8664917744632423e-07, "loss": 0.0057, "step": 704, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.22652983918213118, "train/R_vocab": 0.41149553571428577 }, { "learning_rate": 2.8557870956832133e-07, "loss": 0.0356, "step": 706, "train/R_acc": 0.0, "train/R_penalty": -0.71875, "train/R_reason": 0.22872856053626722, "train/R_vocab": 0.2622767857142857 }, { "learning_rate": 2.8450757547284576e-07, "loss": 0.0236, "step": 708, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.19584017346681087, "train/R_vocab": 0.20625000000000002 }, { "learning_rate": 2.8343579521706823e-07, "loss": -0.0244, "step": 710, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.17909964580588045, "train/R_vocab": 0.23783482142857143 }, { "learning_rate": 2.8236338887025886e-07, "loss": 0.0155, "step": 712, "train/R_acc": 0.034375, "train/R_penalty": -0.484375, "train/R_reason": 0.26302313003150035, "train/R_vocab": 0.3813988095238095 }, { "learning_rate": 2.812903765134115e-07, "loss": -0.0178, "step": 714, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.2631388468076138, "train/R_vocab": 0.4032242063492063 }, { "learning_rate": 2.8021677823886754e-07, "loss": -0.0467, "step": 716, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.19911790925777817, "train/R_vocab": 0.23608630952380952 }, { "learning_rate": 2.7914261414993976e-07, "loss": 0.0054, "step": 718, "train/R_acc": 0.003125, "train/R_penalty": -0.603125, "train/R_reason": 0.2387146239629917, "train/R_vocab": 0.2571428571428571 }, { "learning_rate": 2.780679043605361e-07, "loss": -0.0082, "step": 720, "train/R_acc": 0.03125, "train/R_penalty": -0.571875, "train/R_reason": 0.2269094768328706, "train/R_vocab": 0.31436011904761907 }, { "learning_rate": 2.7699266899478274e-07, "loss": 0.0178, "step": 722, "train/R_acc": 0.034375, "train/R_penalty": -0.703125, "train/R_reason": 0.24130363361070928, "train/R_vocab": 0.22456597222222222 }, { "learning_rate": 2.759169281866472e-07, "loss": 0.0661, "step": 724, "train/R_acc": 0.03125, "train/R_penalty": -0.64375, "train/R_reason": 0.21308869490175653, "train/R_vocab": 0.2772321428571428 }, { "learning_rate": 2.748407020795617e-07, "loss": 0.1056, "step": 726, "train/R_acc": 0.096875, "train/R_penalty": -0.421875, "train/R_reason": 0.23621705469287618, "train/R_vocab": 0.2981770833333333 }, { "learning_rate": 2.737640108260456e-07, "loss": 0.0812, "step": 728, "train/R_acc": 0.03125, "train/R_penalty": -0.6875, "train/R_reason": 0.2310603125451809, "train/R_vocab": 0.30590277777777775 }, { "learning_rate": 2.726868745873286e-07, "loss": -0.0637, "step": 730, "train/R_acc": 0.034375, "train/R_penalty": -0.696875, "train/R_reason": 0.19990588785781663, "train/R_vocab": 0.265625 }, { "learning_rate": 2.716093135329722e-07, "loss": 0.0348, "step": 732, "train/R_acc": 0.0375, "train/R_penalty": -0.421875, "train/R_reason": 0.2095006410516756, "train/R_vocab": 0.32751736111111107 }, { "learning_rate": 2.7053134784049316e-07, "loss": 0.0875, "step": 734, "train/R_acc": 0.003125, "train/R_penalty": -0.71875, "train/R_reason": 0.17118221219882784, "train/R_vocab": 0.17343750000000002 }, { "learning_rate": 2.694529976949849e-07, "loss": -0.0448, "step": 736, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.609375, "train/R_reason": 0.20588894905097077, "train/R_vocab": 0.34609375 }, { "learning_rate": 2.6837428328874007e-07, "loss": -0.0244, "step": 738, "train/R_acc": 0.06875, "train/R_penalty": -0.6125, "train/R_reason": 0.26183079845209467, "train/R_vocab": 0.2656994047619048 }, { "learning_rate": 2.6729522482087165e-07, "loss": 0.0854, "step": 740, "train/R_acc": 0.003125, "train/R_penalty": -0.59375, "train/R_reason": 0.20793241180273003, "train/R_vocab": 0.21996527777777775 }, { "learning_rate": 2.662158424969357e-07, "loss": 0.0149, "step": 742, "train/R_acc": 0.003125, "train/R_penalty": -0.46875, "train/R_reason": 0.18410838343666375, "train/R_vocab": 0.1974330357142857 }, { "learning_rate": 2.6513615652855246e-07, "loss": -0.0144, "step": 744, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.24293661434589603, "train/R_vocab": 0.3267857142857143 }, { "learning_rate": 2.6405618713302783e-07, "loss": -0.0678, "step": 746, "train/R_acc": 0.00625, "train/R_penalty": -0.61875, "train/R_reason": 0.135718109682479, "train/R_vocab": 0.22979910714285715 }, { "learning_rate": 2.629759545329749e-07, "loss": 0.1246, "step": 748, "train/R_acc": 0.0, "train/R_penalty": -0.634375, "train/R_reason": 0.20181279701198124, "train/R_vocab": 0.3018849206349207 }, { "learning_rate": 2.618954789559356e-07, "loss": 0.0933, "step": 750, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.20252563848526106, "train/R_vocab": 0.2291294642857143 }, { "learning_rate": 2.6081478063400124e-07, "loss": -0.0234, "step": 752, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.19106041589978606, "train/R_vocab": 0.20889136904761904 }, { "learning_rate": 2.597338798034344e-07, "loss": -0.0631, "step": 754, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.1867345308525643, "train/R_vocab": 0.22708333333333333 }, { "learning_rate": 2.5865279670428956e-07, "loss": 0.0483, "step": 756, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.17135153722070606, "train/R_vocab": 0.18147321428571428 }, { "learning_rate": 2.5757155158003414e-07, "loss": 0.0295, "step": 758, "train/R_acc": 0.0, "train/R_penalty": -0.703125, "train/R_reason": 0.1663155804467654, "train/R_vocab": 0.22377232142857145 }, { "learning_rate": 2.5649016467716954e-07, "loss": 0.0283, "step": 760, "train/R_acc": 0.04062500000000001, "train/R_penalty": -0.375, "train/R_reason": 0.19205683929152886, "train/R_vocab": 0.2728422619047619 }, { "learning_rate": 2.5540865624485216e-07, "loss": -0.0491, "step": 762, "train/R_acc": 0.0, "train/R_penalty": -0.703125, "train/R_reason": 0.2550380457423676, "train/R_vocab": 0.2734375 }, { "learning_rate": 2.5432704653451374e-07, "loss": 0.0441, "step": 764, "train/R_acc": 0.03125, "train/R_penalty": -0.53125, "train/R_reason": 0.20927672505440928, "train/R_vocab": 0.30055803571428574 }, { "learning_rate": 2.532453557994827e-07, "loss": 0.048, "step": 766, "train/R_acc": 0.003125, "train/R_penalty": -0.515625, "train/R_reason": 0.20293042206204454, "train/R_vocab": 0.2447668650793651 }, { "learning_rate": 2.521636042946046e-07, "loss": 0.0256, "step": 768, "train/R_acc": 0.04062500000000001, "train/R_penalty": -0.55625, "train/R_reason": 0.2228340738544955, "train/R_vocab": 0.21646825396825398 }, { "learning_rate": 2.5108181227586307e-07, "loss": -0.0186, "step": 770, "train/R_acc": 0.03125, "train/R_penalty": -0.703125, "train/R_reason": 0.15339300119866717, "train/R_vocab": 0.24015376984126985 }, { "learning_rate": 2.5e-07, "loss": 0.0082, "step": 772, "train/R_acc": 0.034375, "train/R_penalty": -0.59375, "train/R_reason": 0.25634859172881913, "train/R_vocab": 0.2591517857142857 }, { "learning_rate": 2.48918187724137e-07, "loss": -0.0372, "step": 774, "train/R_acc": 0.0, "train/R_penalty": -0.484375, "train/R_reason": 0.15909137824419298, "train/R_vocab": 0.2725953733766234 }, { "learning_rate": 2.4783639570539535e-07, "loss": -0.1172, "step": 776, "train/R_acc": 0.00625, "train/R_penalty": -0.46875, "train/R_reason": 0.21610677086076593, "train/R_vocab": 0.21164434523809522 }, { "learning_rate": 2.467546442005173e-07, "loss": 0.0088, "step": 778, "train/R_acc": 0.06875, "train/R_penalty": -0.390625, "train/R_reason": 0.19689824765738337, "train/R_vocab": 0.1890625 }, { "learning_rate": 2.4567295346548634e-07, "loss": 0.1308, "step": 780, "train/R_acc": 0.03125, "train/R_penalty": -0.625, "train/R_reason": 0.16858555636857672, "train/R_vocab": 0.29520089285714285 }, { "learning_rate": 2.4459134375514787e-07, "loss": -0.0356, "step": 782, "train/R_acc": 0.003125, "train/R_penalty": -0.5875, "train/R_reason": 0.20190652121448122, "train/R_vocab": 0.2455357142857143 }, { "learning_rate": 2.4350983532283043e-07, "loss": -0.0182, "step": 784, "train/R_acc": 0.003125, "train/R_penalty": -0.703125, "train/R_reason": 0.1856742513120948, "train/R_vocab": 0.275 }, { "learning_rate": 2.4242844841996583e-07, "loss": 0.0167, "step": 786, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.20132693663438478, "train/R_vocab": 0.18971974206349207 }, { "learning_rate": 2.4134720329571047e-07, "loss": 0.0408, "step": 788, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.571875, "train/R_reason": 0.22452081729968582, "train/R_vocab": 0.21640625 }, { "learning_rate": 2.4026612019656556e-07, "loss": 0.0107, "step": 790, "train/R_acc": 0.034375, "train/R_penalty": -0.609375, "train/R_reason": 0.18352974837528543, "train/R_vocab": 0.25814732142857144 }, { "learning_rate": 2.3918521936599874e-07, "loss": 0.036, "step": 792, "train/R_acc": 0.03125, "train/R_penalty": -0.609375, "train/R_reason": 0.27166955319028896, "train/R_vocab": 0.35963541666666665 }, { "learning_rate": 2.381045210440644e-07, "loss": -0.0346, "step": 794, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.22416891439419612, "train/R_vocab": 0.32734375000000004 }, { "learning_rate": 2.3702404546702509e-07, "loss": 0.0911, "step": 796, "train/R_acc": 0.034375, "train/R_penalty": -0.546875, "train/R_reason": 0.24303349052516854, "train/R_vocab": 0.27745535714285713 }, { "learning_rate": 2.3594381286697215e-07, "loss": 0.0121, "step": 798, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.23618598427535661, "train/R_vocab": 0.28122519841269844 }, { "learning_rate": 2.3486384347144752e-07, "loss": 0.0064, "step": 800, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.25934359162168097, "train/R_vocab": 0.4203125 }, { "eval/R_acc": 0.0025, "eval/R_penalty": -0.5875, "eval/R_reason": 0.29083729720667373, "eval/R_vocab": 0.22747113997114, "step": 800 }, { "step": 800 }, { "learning_rate": 2.337841575030642e-07, "loss": -0.0305, "step": 802, "train/R_acc": 0.03125, "train/R_penalty": -0.59375, "train/R_reason": 0.20417958820529003, "train/R_vocab": 0.2702008928571429 }, { "learning_rate": 2.3270477517912835e-07, "loss": 0.0361, "step": 804, "train/R_acc": 0.00625, "train/R_penalty": -0.621875, "train/R_reason": 0.2872431102010518, "train/R_vocab": 0.37351190476190477 }, { "learning_rate": 2.3162571671126e-07, "loss": 0.0184, "step": 806, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.2281768901636334, "train/R_vocab": 0.30616319444444445 }, { "learning_rate": 2.3054700230501502e-07, "loss": 0.0128, "step": 808, "train/R_acc": 0.03125, "train/R_penalty": -0.671875, "train/R_reason": 0.20789617508651104, "train/R_vocab": 0.3257440476190476 }, { "learning_rate": 2.2946865215950685e-07, "loss": 0.0158, "step": 810, "train/R_acc": 0.003125, "train/R_penalty": -0.671875, "train/R_reason": 0.22696589858153038, "train/R_vocab": 0.21540178571428575 }, { "learning_rate": 2.2839068646702786e-07, "loss": 1.5239, "step": 812, "train/R_acc": 0.0375, "train/R_penalty": -0.734375, "train/R_reason": 0.1726926027420783, "train/R_vocab": 0.25167410714285715 }, { "learning_rate": 2.2731312541267143e-07, "loss": 0.0348, "step": 814, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.53125, "train/R_reason": 0.18792369164667982, "train/R_vocab": 0.2892485119047619 }, { "learning_rate": 2.2623598917395436e-07, "loss": 0.0032, "step": 816, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.16929829805580582, "train/R_vocab": 0.16987847222222222 }, { "learning_rate": 2.251592979204383e-07, "loss": 0.0127, "step": 818, "train/R_acc": 0.03125, "train/R_penalty": -0.390625, "train/R_reason": 0.28610113589945263, "train/R_vocab": 0.35 }, { "learning_rate": 2.2408307181335285e-07, "loss": 0.0528, "step": 820, "train/R_acc": 0.0, "train/R_penalty": -0.509375, "train/R_reason": 0.18837133598647154, "train/R_vocab": 0.21272321428571428 }, { "learning_rate": 2.2300733100521732e-07, "loss": 0.0454, "step": 822, "train/R_acc": 0.003125, "train/R_penalty": -0.671875, "train/R_reason": 0.1814823048344199, "train/R_vocab": 0.13814484126984128 }, { "learning_rate": 2.2193209563946382e-07, "loss": 0.0355, "step": 824, "train/R_acc": 0.0, "train/R_penalty": -0.6875, "train/R_reason": 0.18663805046129311, "train/R_vocab": 0.1985863095238095 }, { "learning_rate": 2.2085738585006021e-07, "loss": 0.0026, "step": 826, "train/R_acc": 0.0, "train/R_penalty": -0.49375, "train/R_reason": 0.18273024857160083, "train/R_vocab": 0.21136363636363636 }, { "learning_rate": 2.1978322176113251e-07, "loss": -0.0318, "step": 828, "train/R_acc": 0.034375, "train/R_penalty": -0.578125, "train/R_reason": 0.19883211536778234, "train/R_vocab": 0.27878224206349206 }, { "learning_rate": 2.1870962348658852e-07, "loss": -0.0335, "step": 830, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.546875, "train/R_reason": 0.1985312923562238, "train/R_vocab": 0.2785714285714286 }, { "learning_rate": 2.1763661112974115e-07, "loss": -0.0012, "step": 832, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.2787590605964265, "train/R_vocab": 0.35 }, { "learning_rate": 2.165642047829318e-07, "loss": -0.0375, "step": 834, "train/R_acc": 0.03125, "train/R_penalty": -0.578125, "train/R_reason": 0.15530706954823303, "train/R_vocab": 0.22005208333333334 }, { "learning_rate": 2.1549242452715427e-07, "loss": -0.0387, "step": 836, "train/R_acc": 0.003125, "train/R_penalty": -0.5, "train/R_reason": 0.24527110960119072, "train/R_vocab": 0.38571428571428573 }, { "learning_rate": 2.1442129043167873e-07, "loss": 0.0321, "step": 838, "train/R_acc": 0.003125, "train/R_penalty": -0.703125, "train/R_reason": 0.21774443607948502, "train/R_vocab": 0.26517857142857143 }, { "learning_rate": 2.1335082255367572e-07, "loss": 0.0288, "step": 840, "train/R_acc": 0.03125, "train/R_penalty": -0.546875, "train/R_reason": 0.22299875389864712, "train/R_vocab": 0.27976190476190477 }, { "learning_rate": 2.1228104093784108e-07, "loss": 0.0203, "step": 842, "train/R_acc": 0.003125, "train/R_penalty": -0.515625, "train/R_reason": 0.25459030399590693, "train/R_vocab": 0.36163194444444446 }, { "learning_rate": 2.112119656160199e-07, "loss": 0.0337, "step": 844, "train/R_acc": 0.003125, "train/R_penalty": -0.515625, "train/R_reason": 0.2547007960169786, "train/R_vocab": 0.34011656746031743 }, { "learning_rate": 2.1014361660683194e-07, "loss": -0.0236, "step": 846, "train/R_acc": 0.03125, "train/R_penalty": -0.609375, "train/R_reason": 0.23028272641397346, "train/R_vocab": 0.25 }, { "learning_rate": 2.0907601391529677e-07, "loss": -0.0204, "step": 848, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.18682793317030633, "train/R_vocab": 0.21324066558441557 }, { "learning_rate": 2.0800917753245875e-07, "loss": 0.1057, "step": 850, "train/R_acc": 0.00625, "train/R_penalty": -0.609375, "train/R_reason": 0.2020052493242343, "train/R_vocab": 0.27734375 }, { "learning_rate": 2.0694312743501334e-07, "loss": -0.0585, "step": 852, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.23457174043904147, "train/R_vocab": 0.3036830357142857 }, { "learning_rate": 2.0587788358493235e-07, "loss": -0.0318, "step": 854, "train/R_acc": 0.003125, "train/R_penalty": -0.34375, "train/R_reason": 0.2538105699194751, "train/R_vocab": 0.40078125 }, { "learning_rate": 2.048134659290905e-07, "loss": -0.0723, "step": 856, "train/R_acc": 0.0, "train/R_penalty": -0.6875, "train/R_reason": 0.17298681086107648, "train/R_vocab": 0.15267857142857144 }, { "learning_rate": 2.037498943988921e-07, "loss": -0.0058, "step": 858, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.28539022868157543, "train/R_vocab": 0.3465593434343434 }, { "learning_rate": 2.0268718890989752e-07, "loss": -0.0416, "step": 860, "train/R_acc": 0.03125, "train/R_penalty": -0.509375, "train/R_reason": 0.20216319087856421, "train/R_vocab": 0.3190848214285714 }, { "learning_rate": 2.0162536936145008e-07, "loss": -0.0399, "step": 862, "train/R_acc": 0.003125, "train/R_penalty": -0.696875, "train/R_reason": 0.2420341418631836, "train/R_vocab": 0.27738095238095234 }, { "learning_rate": 2.0056445563630423e-07, "loss": 0.0372, "step": 864, "train/R_acc": 0.03125, "train/R_penalty": -0.55625, "train/R_reason": 0.21099684301910737, "train/R_vocab": 0.26659226190476193 }, { "learning_rate": 1.9950446760025219e-07, "loss": 0.0326, "step": 866, "train/R_acc": 0.0625, "train/R_penalty": -0.61875, "train/R_reason": 0.19938542697531797, "train/R_vocab": 0.25535714285714284 }, { "learning_rate": 1.98445425101753e-07, "loss": 0.0211, "step": 868, "train/R_acc": 0.003125, "train/R_penalty": -0.671875, "train/R_reason": 0.22302710668155856, "train/R_vocab": 0.2036830357142857 }, { "learning_rate": 1.9738734797155987e-07, "loss": 0.1391, "step": 870, "train/R_acc": 0.0125, "train/R_penalty": -0.703125, "train/R_reason": 0.20507300880803225, "train/R_vocab": 0.29453125 }, { "learning_rate": 1.963302560223496e-07, "loss": 0.0314, "step": 872, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.1946319904099672, "train/R_vocab": 0.32533482142857145 }, { "learning_rate": 1.9527416904835132e-07, "loss": 0.0562, "step": 874, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.2123123501156333, "train/R_vocab": 0.24363839285714287 }, { "learning_rate": 1.9421910682497572e-07, "loss": 0.0335, "step": 876, "train/R_acc": 0.0625, "train/R_penalty": -0.578125, "train/R_reason": 0.24626829087757965, "train/R_vocab": 0.28229166666666666 }, { "learning_rate": 1.9316508910844476e-07, "loss": 0.0255, "step": 878, "train/R_acc": 0.034375, "train/R_penalty": -0.59375, "train/R_reason": 0.18789237365926498, "train/R_vocab": 0.278125 }, { "learning_rate": 1.9211213563542218e-07, "loss": -0.0622, "step": 880, "train/R_acc": 0.0, "train/R_penalty": -0.5, "train/R_reason": 0.19592139018937305, "train/R_vocab": 0.26670386904761906 }, { "learning_rate": 1.9106026612264315e-07, "loss": 0.0249, "step": 882, "train/R_acc": 0.065625, "train/R_penalty": -0.5, "train/R_reason": 0.17202561289857762, "train/R_vocab": 0.24394841269841272 }, { "learning_rate": 1.900095002665459e-07, "loss": 0.0627, "step": 884, "train/R_acc": 0.003125, "train/R_penalty": -0.5625, "train/R_reason": 0.20212251182315438, "train/R_vocab": 0.234375 }, { "learning_rate": 1.889598577429022e-07, "loss": 0.0251, "step": 886, "train/R_acc": 0.00625, "train/R_penalty": -0.6875, "train/R_reason": 0.21931575978869752, "train/R_vocab": 0.20535714285714285 }, { "learning_rate": 1.8791135820644922e-07, "loss": 0.0228, "step": 888, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.23473900860844063, "train/R_vocab": 0.2440104166666667 }, { "learning_rate": 1.8686402129052181e-07, "loss": 0.0602, "step": 890, "train/R_acc": 0.034375, "train/R_penalty": -0.5625, "train/R_reason": 0.2188358452056008, "train/R_vocab": 0.24036458333333333 }, { "learning_rate": 1.858178666066843e-07, "loss": 0.0042, "step": 892, "train/R_acc": 0.034375, "train/R_penalty": -0.59375, "train/R_reason": 0.1951297685148197, "train/R_vocab": 0.23270089285714285 }, { "learning_rate": 1.8477291374436338e-07, "loss": -0.0364, "step": 894, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.11877722404091096, "train/R_vocab": 0.12317708333333333 }, { "learning_rate": 1.8372918227048183e-07, "loss": 0.3093, "step": 896, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.1855972158801493, "train/R_vocab": 0.25017361111111114 }, { "learning_rate": 1.8268669172909136e-07, "loss": 0.0017, "step": 898, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.18804079477258429, "train/R_vocab": 0.16622023809523812 }, { "learning_rate": 1.8164546164100728e-07, "loss": -0.0186, "step": 900, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.24246348014867208, "train/R_vocab": 0.29110863095238093 }, { "eval/R_acc": 0.030000000000000006, "eval/R_penalty": -0.6, "eval/R_reason": 0.29157417644976924, "eval/R_vocab": 0.37875, "step": 900 }, { "step": 900 }, { "learning_rate": 1.806055115034425e-07, "loss": -0.0579, "step": 902, "train/R_acc": 0.003125, "train/R_penalty": -0.515625, "train/R_reason": 0.23201039495721226, "train/R_vocab": 0.2525669642857143 }, { "learning_rate": 1.7956686078964255e-07, "loss": 0.0383, "step": 904, "train/R_acc": 0.03125, "train/R_penalty": -0.61875, "train/R_reason": 0.19469584258694483, "train/R_vocab": 0.21830357142857146 }, { "learning_rate": 1.7852952894852148e-07, "loss": 0.0293, "step": 906, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.625, "train/R_reason": 0.25663321405687106, "train/R_vocab": 0.32265625000000003 }, { "learning_rate": 1.7749353540429663e-07, "loss": -0.0106, "step": 908, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.19230140159133588, "train/R_vocab": 0.16744791666666667 }, { "learning_rate": 1.7645889955612592e-07, "loss": -0.0295, "step": 910, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.12090342349306576, "train/R_vocab": 0.1583457341269841 }, { "learning_rate": 1.7542564077774403e-07, "loss": 0.0045, "step": 912, "train/R_acc": 0.034375, "train/R_penalty": -0.675, "train/R_reason": 0.2114914680859179, "train/R_vocab": 0.3030505952380952 }, { "learning_rate": 1.7439377841709973e-07, "loss": -0.0017, "step": 914, "train/R_acc": 0.0625, "train/R_penalty": -0.609375, "train/R_reason": 0.217770400372361, "train/R_vocab": 0.25592757936507937 }, { "learning_rate": 1.7336333179599378e-07, "loss": 0.0849, "step": 916, "train/R_acc": 0.00625, "train/R_penalty": -0.609375, "train/R_reason": 0.2602397752431368, "train/R_vocab": 0.4150173611111111 }, { "learning_rate": 1.7233432020971684e-07, "loss": 0.0793, "step": 918, "train/R_acc": 0.0, "train/R_penalty": -0.55625, "train/R_reason": 0.2766625564419999, "train/R_vocab": 0.23649553571428572 }, { "learning_rate": 1.7130676292668816e-07, "loss": 0.0446, "step": 920, "train/R_acc": 0.003125, "train/R_penalty": -0.6875, "train/R_reason": 0.22695619844003265, "train/R_vocab": 0.20081845238095236 }, { "learning_rate": 1.7028067918809535e-07, "loss": -0.036, "step": 922, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.17932276860695562, "train/R_vocab": 0.2564732142857143 }, { "learning_rate": 1.6925608820753325e-07, "loss": -0.0333, "step": 924, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.1710392126096163, "train/R_vocab": 0.11149553571428572 }, { "learning_rate": 1.6823300917064458e-07, "loss": -0.0427, "step": 926, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.1853385177642702, "train/R_vocab": 0.20937499999999998 }, { "learning_rate": 1.6721146123476093e-07, "loss": 0.122, "step": 928, "train/R_acc": 0.03125, "train/R_penalty": -0.634375, "train/R_reason": 0.21111988917807223, "train/R_vocab": 0.28467261904761904 }, { "learning_rate": 1.6619146352854334e-07, "loss": 0.0335, "step": 930, "train/R_acc": 0.00625, "train/R_penalty": -0.525, "train/R_reason": 0.2012206313966573, "train/R_vocab": 0.2486359126984127 }, { "learning_rate": 1.6517303515162505e-07, "loss": -0.0597, "step": 932, "train/R_acc": 0.034375, "train/R_penalty": -0.5625, "train/R_reason": 0.16395669599613677, "train/R_vocab": 0.2521329365079365 }, { "learning_rate": 1.6415619517425294e-07, "loss": 0.0165, "step": 934, "train/R_acc": 0.003125, "train/R_penalty": -0.65, "train/R_reason": 0.14773459331491504, "train/R_vocab": 0.2604910714285714 }, { "learning_rate": 1.6314096263693078e-07, "loss": -0.0171, "step": 936, "train/R_acc": 0.0, "train/R_penalty": -0.61875, "train/R_reason": 0.24617103354461017, "train/R_vocab": 0.25078125 }, { "learning_rate": 1.6212735655006322e-07, "loss": -0.0329, "step": 938, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.2476563881397631, "train/R_vocab": 0.22760416666666666 }, { "learning_rate": 1.6111539589359885e-07, "loss": -0.0157, "step": 940, "train/R_acc": 0.034375, "train/R_penalty": -0.55625, "train/R_reason": 0.18716442508640668, "train/R_vocab": 0.21462053571428572 }, { "learning_rate": 1.6010509961667546e-07, "loss": 0.0015, "step": 942, "train/R_acc": 0.00625, "train/R_penalty": -0.5625, "train/R_reason": 0.2163363343470736, "train/R_vocab": 0.3014508928571429 }, { "learning_rate": 1.590964866372652e-07, "loss": -0.0064, "step": 944, "train/R_acc": 0.0, "train/R_penalty": -0.5875, "train/R_reason": 0.22361495696140293, "train/R_vocab": 0.278125 }, { "learning_rate": 1.5808957584181994e-07, "loss": -0.0227, "step": 946, "train/R_acc": 0.03125, "train/R_penalty": -0.596875, "train/R_reason": 0.2240925844127057, "train/R_vocab": 0.30379464285714286 }, { "learning_rate": 1.5708438608491815e-07, "loss": 0.0113, "step": 948, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.19887587350757036, "train/R_vocab": 0.2051846590909091 }, { "learning_rate": 1.560809361889112e-07, "loss": 0.0187, "step": 950, "train/R_acc": 0.0, "train/R_penalty": -0.671875, "train/R_reason": 0.1577485102536454, "train/R_vocab": 0.20122767857142856 }, { "learning_rate": 1.5507924494357154e-07, "loss": -0.0467, "step": 952, "train/R_acc": 0.034375, "train/R_penalty": -0.5625, "train/R_reason": 0.14574324556072643, "train/R_vocab": 0.20066964285714284 }, { "learning_rate": 1.5407933110574044e-07, "loss": 0.0123, "step": 954, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.21002736065355815, "train/R_vocab": 0.27578125 }, { "learning_rate": 1.5308121339897705e-07, "loss": 0.0131, "step": 956, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.24345409494067652, "train/R_vocab": 0.33035714285714285 }, { "learning_rate": 1.5208491051320744e-07, "loss": -0.1197, "step": 958, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.578125, "train/R_reason": 0.21702138753711128, "train/R_vocab": 0.28426339285714286 }, { "learning_rate": 1.510904411043751e-07, "loss": -0.0818, "step": 960, "train/R_acc": 0.0, "train/R_penalty": -0.71875, "train/R_reason": 0.23023067131726802, "train/R_vocab": 0.3091517857142857 }, { "learning_rate": 1.5009782379409108e-07, "loss": -0.1087, "step": 962, "train/R_acc": 0.003125, "train/R_penalty": -0.765625, "train/R_reason": 0.17687649620358462, "train/R_vocab": 0.21863839285714284 }, { "learning_rate": 1.4910707716928586e-07, "loss": 0.0956, "step": 964, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.46875, "train/R_reason": 0.22474928414381412, "train/R_vocab": 0.26093750000000004 }, { "learning_rate": 1.4811821978186078e-07, "loss": 0.0536, "step": 966, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.22651502510275934, "train/R_vocab": 0.20833333333333331 }, { "learning_rate": 1.4713127014834088e-07, "loss": 0.0205, "step": 968, "train/R_acc": 0.0, "train/R_penalty": -0.5625, "train/R_reason": 0.2021908958276862, "train/R_vocab": 0.1978794642857143 }, { "learning_rate": 1.461462467495284e-07, "loss": 0.0427, "step": 970, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.15024320937786767, "train/R_vocab": 0.20468750000000002 }, { "learning_rate": 1.4516316803015627e-07, "loss": 0.0126, "step": 972, "train/R_acc": 0.00625, "train/R_penalty": -0.59375, "train/R_reason": 0.18895812905862736, "train/R_vocab": 0.14308035714285713 }, { "learning_rate": 1.441820523985431e-07, "loss": -0.0349, "step": 974, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.2382146965819557, "train/R_vocab": 0.2640997023809524 }, { "learning_rate": 1.432029182262482e-07, "loss": 0.0445, "step": 976, "train/R_acc": 0.09375, "train/R_penalty": -0.59375, "train/R_reason": 0.2085522733677449, "train/R_vocab": 0.26365327380952386 }, { "learning_rate": 1.42225783847728e-07, "loss": 0.0309, "step": 978, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.27484186892341045, "train/R_vocab": 0.3541170634920635 }, { "learning_rate": 1.41250667559992e-07, "loss": 0.0122, "step": 980, "train/R_acc": 0.034375, "train/R_penalty": -0.53125, "train/R_reason": 0.22480109886784352, "train/R_vocab": 0.30301339285714285 }, { "learning_rate": 1.4027758762226107e-07, "loss": -0.0861, "step": 982, "train/R_acc": 0.003125, "train/R_penalty": -0.665625, "train/R_reason": 0.21184840132463886, "train/R_vocab": 0.34201388888888884 }, { "learning_rate": 1.3930656225562474e-07, "loss": 0.0047, "step": 984, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.2253757708880576, "train/R_vocab": 0.30904017857142857 }, { "learning_rate": 1.3833760964270058e-07, "loss": 0.0296, "step": 986, "train/R_acc": 0.03125, "train/R_penalty": -0.546875, "train/R_reason": 0.2222406698557359, "train/R_vocab": 0.305406746031746 }, { "learning_rate": 1.3737074792729332e-07, "loss": 0.0262, "step": 988, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.2037288188497659, "train/R_vocab": 0.17431795634920633 }, { "learning_rate": 1.3640599521405522e-07, "loss": 0.0671, "step": 990, "train/R_acc": 0.040625, "train/R_penalty": -0.53125, "train/R_reason": 0.19595219511870288, "train/R_vocab": 0.2941964285714286 }, { "learning_rate": 1.354433695681474e-07, "loss": -0.0366, "step": 992, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.1895803639379184, "train/R_vocab": 0.2979910714285714 }, { "learning_rate": 1.3448288901490092e-07, "loss": 0.0929, "step": 994, "train/R_acc": 0.034375, "train/R_penalty": -0.665625, "train/R_reason": 0.21089602634201932, "train/R_vocab": 0.19340277777777776 }, { "learning_rate": 1.3352457153947997e-07, "loss": 0.0515, "step": 996, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.546875, "train/R_reason": 0.21176803496084237, "train/R_vocab": 0.3035094246031746 }, { "learning_rate": 1.3256843508654444e-07, "loss": 0.0103, "step": 998, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.15247972660677223, "train/R_vocab": 0.21183035714285717 }, { "learning_rate": 1.3161449755991425e-07, "loss": -0.082, "step": 1000, "train/R_acc": 0.13125, "train/R_penalty": -0.515625, "train/R_reason": 0.2110448967120055, "train/R_vocab": 0.25863095238095235 }, { "eval/R_acc": 0.0025, "eval/R_penalty": -0.55, "eval/R_reason": 0.2655294246752871, "eval/R_vocab": 0.34914772727272725, "step": 1000 }, { "step": 1000 }, { "learning_rate": 1.3066277682223425e-07, "loss": -0.0096, "step": 1002, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.181312457358332, "train/R_vocab": 0.21517857142857144 }, { "learning_rate": 1.2971329069463932e-07, "loss": -0.0918, "step": 1004, "train/R_acc": 0.003125, "train/R_penalty": -0.75, "train/R_reason": 0.12020329779967928, "train/R_vocab": 0.12425595238095238 }, { "learning_rate": 1.2876605695642084e-07, "loss": -0.0157, "step": 1006, "train/R_acc": 0.03125, "train/R_penalty": -0.625, "train/R_reason": 0.2288382776444276, "train/R_vocab": 0.3042410714285715 }, { "learning_rate": 1.2782109334469409e-07, "loss": 0.0782, "step": 1008, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.23151180047594122, "train/R_vocab": 0.30937499999999996 }, { "learning_rate": 1.2687841755406554e-07, "loss": -0.0145, "step": 1010, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.23526236082480567, "train/R_vocab": 0.30009920634920634 }, { "learning_rate": 1.2593804723630209e-07, "loss": 0.0254, "step": 1012, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.2194206796665749, "train/R_vocab": 0.3340401785714286 }, { "learning_rate": 1.2500000000000005e-07, "loss": 0.0277, "step": 1014, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.18922638449707688, "train/R_vocab": 0.3610491071428572 }, { "learning_rate": 1.2406429341025556e-07, "loss": 0.0668, "step": 1016, "train/R_acc": 0.00625, "train/R_penalty": -0.55625, "train/R_reason": 0.2346744057550262, "train/R_vocab": 0.2239955357142857 }, { "learning_rate": 1.231309449883361e-07, "loss": 0.0338, "step": 1018, "train/R_acc": 0.003125, "train/R_penalty": -0.53125, "train/R_reason": 0.19074387902354337, "train/R_vocab": 0.21744791666666666 }, { "learning_rate": 1.2219997221135177e-07, "loss": 0.0082, "step": 1020, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.24590138406147055, "train/R_vocab": 0.2600446428571429 }, { "learning_rate": 1.2127139251192823e-07, "loss": 0.0268, "step": 1022, "train/R_acc": 0.0, "train/R_penalty": -0.53125, "train/R_reason": 0.24386600100476474, "train/R_vocab": 0.3195560515873016 }, { "learning_rate": 1.203452232778807e-07, "loss": 0.0183, "step": 1024, "train/R_acc": 0.00625, "train/R_penalty": -0.671875, "train/R_reason": 0.171127249096387, "train/R_vocab": 0.29609375000000004 }, { "learning_rate": 1.1942148185188763e-07, "loss": 0.0272, "step": 1026, "train/R_acc": 0.00625, "train/R_penalty": -0.5625, "train/R_reason": 0.17080492182756957, "train/R_vocab": 0.21568627450980393 }, { "learning_rate": 1.1850018553116676e-07, "loss": -0.0052, "step": 1028, "train/R_acc": 0.003125, "train/R_penalty": -0.5, "train/R_reason": 0.2604828445455006, "train/R_vocab": 0.23158482142857142 }, { "learning_rate": 1.1758135156715041e-07, "loss": -0.0154, "step": 1030, "train/R_acc": 0.065625, "train/R_penalty": -0.578125, "train/R_reason": 0.1946716410122738, "train/R_vocab": 0.22801339285714284 }, { "learning_rate": 1.1666499716516302e-07, "loss": 0.0101, "step": 1032, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.18422370796707505, "train/R_vocab": 0.21763392857142855 }, { "learning_rate": 1.157511394840989e-07, "loss": -0.0049, "step": 1034, "train/R_acc": 0.00625, "train/R_penalty": -0.7125, "train/R_reason": 0.17663980590936243, "train/R_vocab": 0.24196428571428574 }, { "learning_rate": 1.1483979563610069e-07, "loss": 0.0098, "step": 1036, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.26512284555646376, "train/R_vocab": 0.375 }, { "learning_rate": 1.1393098268623902e-07, "loss": 0.1176, "step": 1038, "train/R_acc": 0.0625, "train/R_penalty": -0.634375, "train/R_reason": 0.26504285268931027, "train/R_vocab": 0.30766369047619047 }, { "learning_rate": 1.1302471765219327e-07, "loss": 0.0428, "step": 1040, "train/R_acc": 0.03125, "train/R_penalty": -0.59375, "train/R_reason": 0.2008627221719819, "train/R_vocab": 0.253125 }, { "learning_rate": 1.1212101750393235e-07, "loss": 0.0175, "step": 1042, "train/R_acc": 0.0625, "train/R_penalty": -0.409375, "train/R_reason": 0.19803424974179828, "train/R_vocab": 0.17265625 }, { "learning_rate": 1.1121989916339756e-07, "loss": 0.0292, "step": 1044, "train/R_acc": 0.0, "train/R_penalty": -0.61875, "train/R_reason": 0.21600135050493885, "train/R_vocab": 0.3185267857142857 }, { "learning_rate": 1.1032137950418514e-07, "loss": 0.0669, "step": 1046, "train/R_acc": 0.00625, "train/R_penalty": -0.59375, "train/R_reason": 0.23236107648223334, "train/R_vocab": 0.22142857142857142 }, { "learning_rate": 1.0942547535123057e-07, "loss": -0.0038, "step": 1048, "train/R_acc": 0.00625, "train/R_penalty": -0.509375, "train/R_reason": 0.24041357438290228, "train/R_vocab": 0.2955357142857143 }, { "learning_rate": 1.085322034804938e-07, "loss": 0.1258, "step": 1050, "train/R_acc": 0.003125, "train/R_penalty": -0.665625, "train/R_reason": 0.20610390132448592, "train/R_vocab": 0.24229910714285713 }, { "learning_rate": 1.0764158061864448e-07, "loss": -0.0495, "step": 1052, "train/R_acc": 0.003125, "train/R_penalty": -0.65, "train/R_reason": 0.2606824569537362, "train/R_vocab": 0.39392361111111107 }, { "learning_rate": 1.0675362344274952e-07, "loss": -0.0224, "step": 1054, "train/R_acc": 0.040625, "train/R_penalty": -0.53125, "train/R_reason": 0.2362594455371677, "train/R_vocab": 0.19696631493506495 }, { "learning_rate": 1.0586834857995996e-07, "loss": 0.0405, "step": 1056, "train/R_acc": 0.0, "train/R_penalty": -0.59375, "train/R_reason": 0.22383388089744485, "train/R_vocab": 0.2839285714285714 }, { "learning_rate": 1.0498577260720048e-07, "loss": 0.006, "step": 1058, "train/R_acc": 0.003125, "train/R_penalty": -0.5625, "train/R_reason": 0.2154017687323606, "train/R_vocab": 0.18723958333333335 }, { "learning_rate": 1.0410591205085817e-07, "loss": 0.0547, "step": 1060, "train/R_acc": 0.03125, "train/R_penalty": -0.64375, "train/R_reason": 0.22723913109425195, "train/R_vocab": 0.32499999999999996 }, { "learning_rate": 1.0322878338647392e-07, "loss": 0.1005, "step": 1062, "train/R_acc": 0.0, "train/R_penalty": -0.65, "train/R_reason": 0.21853491707209402, "train/R_vocab": 0.19251217532467532 }, { "learning_rate": 1.0235440303843302e-07, "loss": 0.0902, "step": 1064, "train/R_acc": 0.03125, "train/R_penalty": -0.546875, "train/R_reason": 0.21816371211780805, "train/R_vocab": 0.22801339285714284 }, { "learning_rate": 1.0148278737965844e-07, "loss": -0.0553, "step": 1066, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.19269070794229914, "train/R_vocab": 0.18387896825396824 }, { "learning_rate": 1.0061395273130368e-07, "loss": 0.039, "step": 1068, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.21827443734801666, "train/R_vocab": 0.2911830357142857 }, { "learning_rate": 9.974791536244726e-08, "loss": 0.0343, "step": 1070, "train/R_acc": 0.0, "train/R_penalty": -0.61875, "train/R_reason": 0.2733091389175897, "train/R_vocab": 0.3764136904761905 }, { "learning_rate": 9.888469148978843e-08, "loss": -0.0473, "step": 1072, "train/R_acc": 0.003125, "train/R_penalty": -0.484375, "train/R_reason": 0.24157353834104733, "train/R_vocab": 0.3257496843434343 }, { "learning_rate": 9.802429727734294e-08, "loss": -0.0036, "step": 1074, "train/R_acc": 0.040625, "train/R_penalty": -0.421875, "train/R_reason": 0.2823229532013337, "train/R_vocab": 0.43273809523809526 }, { "learning_rate": 9.716674883614092e-08, "loss": 0.0124, "step": 1076, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.453125, "train/R_reason": 0.2879044257474095, "train/R_vocab": 0.33738839285714284 }, { "learning_rate": 9.631206222392479e-08, "loss": -0.0157, "step": 1078, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.26644330439869424, "train/R_vocab": 0.27585565476190477 }, { "learning_rate": 9.546025344484868e-08, "loss": 0.0013, "step": 1080, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.17772787504592696, "train/R_vocab": 0.22031250000000002 }, { "learning_rate": 9.461133844917901e-08, "loss": -0.0364, "step": 1082, "train/R_acc": 0.0625, "train/R_penalty": -0.696875, "train/R_reason": 0.14647117758570943, "train/R_vocab": 0.19616815476190477 }, { "learning_rate": 9.376533313299542e-08, "loss": -0.0051, "step": 1084, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.18199736371919578, "train/R_vocab": 0.2505580357142857 }, { "learning_rate": 9.292225333789327e-08, "loss": 0.0575, "step": 1086, "train/R_acc": 0.03125, "train/R_penalty": -0.603125, "train/R_reason": 0.1892425633255841, "train/R_vocab": 0.31540178571428573 }, { "learning_rate": 9.208211485068729e-08, "loss": 0.0395, "step": 1088, "train/R_acc": 0.003125, "train/R_penalty": -0.578125, "train/R_reason": 0.15220888566600485, "train/R_vocab": 0.21587301587301586 }, { "learning_rate": 9.124493340311537e-08, "loss": -0.016, "step": 1090, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.18263365867722547, "train/R_vocab": 0.25424107142857144 }, { "learning_rate": 9.041072467154471e-08, "loss": -0.0291, "step": 1092, "train/R_acc": 0.0, "train/R_penalty": -0.375, "train/R_reason": 0.25466359857700704, "train/R_vocab": 0.34862351190476193 }, { "learning_rate": 8.957950427667751e-08, "loss": 0.0296, "step": 1094, "train/R_acc": 0.03125, "train/R_penalty": -0.53125, "train/R_reason": 0.23575074339334445, "train/R_vocab": 0.2983010912698413 }, { "learning_rate": 8.875128778325902e-08, "loss": -0.0389, "step": 1096, "train/R_acc": 0.0, "train/R_penalty": -0.53125, "train/R_reason": 0.1681449082491554, "train/R_vocab": 0.23660714285714285 }, { "learning_rate": 8.792609069978602e-08, "loss": 0.0201, "step": 1098, "train/R_acc": 0.0125, "train/R_penalty": -0.765625, "train/R_reason": 0.24261987678361474, "train/R_vocab": 0.23020833333333335 }, { "learning_rate": 8.710392847821615e-08, "loss": -0.0508, "step": 1100, "train/R_acc": 0.03125, "train/R_penalty": -0.53125, "train/R_reason": 0.19600584280495498, "train/R_vocab": 0.24397321428571428 }, { "eval/R_acc": 0.0025, "eval/R_penalty": -0.4875, "eval/R_reason": 0.24036428951274935, "eval/R_vocab": 0.29469967532467534, "step": 1100 }, { "step": 1100 }, { "learning_rate": 8.628481651367875e-08, "loss": 0.0336, "step": 1102, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.22866834230086636, "train/R_vocab": 0.19769345238095237 }, { "learning_rate": 8.546877014418671e-08, "loss": -0.0387, "step": 1104, "train/R_acc": 0.03125, "train/R_penalty": -0.609375, "train/R_reason": 0.21012454085545956, "train/R_vocab": 0.23225446428571428 }, { "learning_rate": 8.465580465034894e-08, "loss": 0.0231, "step": 1106, "train/R_acc": 0.03125, "train/R_penalty": -0.609375, "train/R_reason": 0.20315946724213801, "train/R_vocab": 0.2909598214285714 }, { "learning_rate": 8.384593525508466e-08, "loss": 0.0343, "step": 1108, "train/R_acc": 0.0, "train/R_penalty": -0.509375, "train/R_reason": 0.20224091353679885, "train/R_vocab": 0.19077380952380954 }, { "learning_rate": 8.303917712333788e-08, "loss": 0.0208, "step": 1110, "train/R_acc": 0.065625, "train/R_penalty": -0.5625, "train/R_reason": 0.23770102329722845, "train/R_vocab": 0.21218998015873014 }, { "learning_rate": 8.223554536179367e-08, "loss": 0.0016, "step": 1112, "train/R_acc": 0.0, "train/R_penalty": -0.5, "train/R_reason": 0.24531693423546427, "train/R_vocab": 0.3191964285714286 }, { "learning_rate": 8.143505501859551e-08, "loss": 0.0352, "step": 1114, "train/R_acc": 0.0, "train/R_penalty": -0.49375, "train/R_reason": 0.2098236201971303, "train/R_vocab": 0.2838169642857143 }, { "learning_rate": 8.06377210830631e-08, "loss": 0.0216, "step": 1116, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.20338203537474456, "train/R_vocab": 0.3302455357142857 }, { "learning_rate": 7.984355848541175e-08, "loss": -0.0481, "step": 1118, "train/R_acc": 0.003125, "train/R_penalty": -0.53125, "train/R_reason": 0.17974317204155837, "train/R_vocab": 0.32105654761904767 }, { "learning_rate": 7.905258209647325e-08, "loss": 0.0544, "step": 1120, "train/R_acc": 0.03125, "train/R_penalty": -0.603125, "train/R_reason": 0.24957606202713686, "train/R_vocab": 0.30379464285714286 }, { "learning_rate": 7.826480672741676e-08, "loss": -0.0196, "step": 1122, "train/R_acc": 0.0625, "train/R_penalty": -0.5625, "train/R_reason": 0.21989961574561373, "train/R_vocab": 0.24270833333333333 }, { "learning_rate": 7.748024712947204e-08, "loss": 0.0036, "step": 1124, "train/R_acc": 0.003125, "train/R_penalty": -0.53125, "train/R_reason": 0.20629454131030475, "train/R_vocab": 0.2872767857142857 }, { "learning_rate": 7.669891799365282e-08, "loss": 0.0386, "step": 1126, "train/R_acc": 0.00625, "train/R_penalty": -0.671875, "train/R_reason": 0.22259933007352922, "train/R_vocab": 0.36199156746031746 }, { "learning_rate": 7.592083395048182e-08, "loss": 0.0643, "step": 1128, "train/R_acc": 0.03125, "train/R_penalty": -0.49375, "train/R_reason": 0.18965226315756645, "train/R_vocab": 0.2122767857142857 }, { "learning_rate": 7.514600956971712e-08, "loss": 0.0678, "step": 1130, "train/R_acc": 0.0, "train/R_penalty": -0.5875, "train/R_reason": 0.23261509390210205, "train/R_vocab": 0.23307291666666669 }, { "learning_rate": 7.437445936007867e-08, "loss": 0.01, "step": 1132, "train/R_acc": 0.03125, "train/R_penalty": -0.59375, "train/R_reason": 0.2093342551771361, "train/R_vocab": 0.2585565476190476 }, { "learning_rate": 7.360619776897742e-08, "loss": -0.0392, "step": 1134, "train/R_acc": 0.0, "train/R_penalty": -0.525, "train/R_reason": 0.22079774921664902, "train/R_vocab": 0.27708333333333335 }, { "learning_rate": 7.284123918224397e-08, "loss": 0.0161, "step": 1136, "train/R_acc": 0.03125, "train/R_penalty": -0.59375, "train/R_reason": 0.22520667797905555, "train/R_vocab": 0.3407118055555556 }, { "learning_rate": 7.207959792385998e-08, "loss": 0.0479, "step": 1138, "train/R_acc": 0.00625, "train/R_penalty": -0.515625, "train/R_reason": 0.19075869235061965, "train/R_vocab": 0.2729166666666667 }, { "learning_rate": 7.132128825568914e-08, "loss": -0.0027, "step": 1140, "train/R_acc": 0.03125, "train/R_penalty": -0.634375, "train/R_reason": 0.23292150900288705, "train/R_vocab": 0.21964285714285714 }, { "learning_rate": 7.056632437721099e-08, "loss": -0.0379, "step": 1142, "train/R_acc": 0.0625, "train/R_penalty": -0.578125, "train/R_reason": 0.17068707794196553, "train/R_vocab": 0.23448660714285716 }, { "learning_rate": 6.981472042525416e-08, "loss": -0.0097, "step": 1144, "train/R_acc": 0.003125, "train/R_penalty": -0.484375, "train/R_reason": 0.1982660319129968, "train/R_vocab": 0.18515624999999997 }, { "learning_rate": 6.906649047373245e-08, "loss": -0.0136, "step": 1146, "train/R_acc": 0.0, "train/R_penalty": -0.571875, "train/R_reason": 0.18697098939403184, "train/R_vocab": 0.22890624999999998 }, { "learning_rate": 6.832164853338066e-08, "loss": 0.0075, "step": 1148, "train/R_acc": 0.065625, "train/R_penalty": -0.515625, "train/R_reason": 0.26961296459836603, "train/R_vocab": 0.3282118055555555 }, { "learning_rate": 6.758020855149249e-08, "loss": -0.0081, "step": 1150, "train/R_acc": 0.03125, "train/R_penalty": -0.55625, "train/R_reason": 0.24886312239764444, "train/R_vocab": 0.3654017857142857 }, { "learning_rate": 6.684218441165962e-08, "loss": 0.0666, "step": 1152, "train/R_acc": 0.0, "train/R_penalty": -0.61875, "train/R_reason": 0.2088855316383599, "train/R_vocab": 0.22730654761904762 }, { "learning_rate": 6.610758993351123e-08, "loss": 0.0395, "step": 1154, "train/R_acc": 0.034375, "train/R_penalty": -0.53125, "train/R_reason": 0.22720806482016326, "train/R_vocab": 0.24931795634920634 }, { "learning_rate": 6.537643887245573e-08, "loss": 0.0465, "step": 1156, "train/R_acc": 0.03125, "train/R_penalty": -0.515625, "train/R_reason": 0.1968795889130503, "train/R_vocab": 0.2591517857142857 }, { "learning_rate": 6.464874491942279e-08, "loss": 0.0203, "step": 1158, "train/R_acc": 0.0, "train/R_penalty": -0.665625, "train/R_reason": 0.19558365060317057, "train/R_vocab": 0.1140625 }, { "learning_rate": 6.392452170060706e-08, "loss": 0.0173, "step": 1160, "train/R_acc": 0.03125, "train/R_penalty": -0.540625, "train/R_reason": 0.25112434395380084, "train/R_vocab": 0.23022073412698413 }, { "learning_rate": 6.320378277721342e-08, "loss": -0.0386, "step": 1162, "train/R_acc": 0.003125, "train/R_penalty": -0.478125, "train/R_reason": 0.24359373155022102, "train/R_vocab": 0.3531498015873016 }, { "learning_rate": 6.248654164520237e-08, "loss": -0.0419, "step": 1164, "train/R_acc": 0.003125, "train/R_penalty": -0.71875, "train/R_reason": 0.20550711716499057, "train/R_vocab": 0.3539434523809524 }, { "learning_rate": 6.177281173503778e-08, "loss": 0.0405, "step": 1166, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.20957454414924856, "train/R_vocab": 0.2280133928571429 }, { "learning_rate": 6.106260641143546e-08, "loss": -0.0833, "step": 1168, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.20663920365929042, "train/R_vocab": 0.21852678571428574 }, { "learning_rate": 6.035593897311239e-08, "loss": -0.0156, "step": 1170, "train/R_acc": 0.003125, "train/R_penalty": -0.61875, "train/R_reason": 0.2371425092299592, "train/R_vocab": 0.3323784722222222 }, { "learning_rate": 5.965282265253838e-08, "loss": 0.0651, "step": 1172, "train/R_acc": 0.034375, "train/R_penalty": -0.390625, "train/R_reason": 0.22832625758068797, "train/R_vocab": 0.30301339285714285 }, { "learning_rate": 5.895327061568775e-08, "loss": -0.0379, "step": 1174, "train/R_acc": 0.03125, "train/R_penalty": -0.578125, "train/R_reason": 0.19140193598424352, "train/R_vocab": 0.2755952380952381 }, { "learning_rate": 5.8257295961792936e-08, "loss": -0.1099, "step": 1176, "train/R_acc": 0.0, "train/R_penalty": -0.640625, "train/R_reason": 0.2013441433056279, "train/R_vocab": 0.2722222222222222 }, { "learning_rate": 5.756491172309952e-08, "loss": -0.0236, "step": 1178, "train/R_acc": 0.003125, "train/R_penalty": -0.53125, "train/R_reason": 0.2409798833837463, "train/R_vocab": 0.2924107142857143 }, { "learning_rate": 5.68761308646217e-08, "loss": -0.0326, "step": 1180, "train/R_acc": 0.003125, "train/R_penalty": -0.665625, "train/R_reason": 0.16268249986159153, "train/R_vocab": 0.17838541666666669 }, { "learning_rate": 5.6190966283899774e-08, "loss": 0.0286, "step": 1182, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.2404998117876481, "train/R_vocab": 0.34227430555555555 }, { "learning_rate": 5.5509430810758817e-08, "loss": 0.004, "step": 1184, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.22824560961553947, "train/R_vocab": 0.3375 }, { "learning_rate": 5.483153720706798e-08, "loss": -0.0074, "step": 1186, "train/R_acc": 0.1, "train/R_penalty": -0.53125, "train/R_reason": 0.21326554603657105, "train/R_vocab": 0.20881696428571428 }, { "learning_rate": 5.415729816650202e-08, "loss": 0.0246, "step": 1188, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.2444787127154557, "train/R_vocab": 0.35230654761904756 }, { "learning_rate": 5.3486726314303175e-08, "loss": 0.0987, "step": 1190, "train/R_acc": 0.003125, "train/R_penalty": -0.634375, "train/R_reason": 0.22288328491646994, "train/R_vocab": 0.2479910714285714 }, { "learning_rate": 5.2819834207044974e-08, "loss": -0.0208, "step": 1192, "train/R_acc": 0.128125, "train/R_penalty": -0.4375, "train/R_reason": 0.27609449087124366, "train/R_vocab": 0.3087425595238095 }, { "learning_rate": 5.215663433239728e-08, "loss": -0.006, "step": 1194, "train/R_acc": 0.034375, "train/R_penalty": -0.609375, "train/R_reason": 0.22109386819415358, "train/R_vocab": 0.24497767857142858 }, { "learning_rate": 5.149713910889206e-08, "loss": -0.0461, "step": 1196, "train/R_acc": 0.034375, "train/R_penalty": -0.421875, "train/R_reason": 0.2398768408665738, "train/R_vocab": 0.3560267857142857 }, { "learning_rate": 5.0841360885690996e-08, "loss": 0.0184, "step": 1198, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.5625, "train/R_reason": 0.21147620601357403, "train/R_vocab": 0.20399305555555555 }, { "learning_rate": 5.01893119423546e-08, "loss": 0.0505, "step": 1200, "train/R_acc": 0.065625, "train/R_penalty": -0.671875, "train/R_reason": 0.2487143037586896, "train/R_vocab": 0.3544642857142857 }, { "eval/R_acc": 0.0, "eval/R_penalty": -0.5875, "eval/R_reason": 0.20266877330422212, "eval/R_vocab": 0.2709090909090909, "step": 1200 }, { "step": 1200 }, { "learning_rate": 4.954100448861165e-08, "loss": -0.009, "step": 1202, "train/R_acc": 0.034375, "train/R_penalty": -0.446875, "train/R_reason": 0.2773491666866028, "train/R_vocab": 0.29509943181818177 }, { "learning_rate": 4.889645066413112e-08, "loss": 0.0306, "step": 1204, "train/R_acc": 0.003125, "train/R_penalty": -0.734375, "train/R_reason": 0.2607235600350966, "train/R_vocab": 0.1830729166666667 }, { "learning_rate": 4.82556625382945e-08, "loss": -0.0188, "step": 1206, "train/R_acc": 0.03125, "train/R_penalty": -0.665625, "train/R_reason": 0.20154795093112635, "train/R_vocab": 0.19722222222222224 }, { "learning_rate": 4.76186521099699e-08, "loss": -0.0385, "step": 1208, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.2240002350916589, "train/R_vocab": 0.3001116071428571 }, { "learning_rate": 4.698543130728755e-08, "loss": 0.0571, "step": 1210, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.15973930176485518, "train/R_vocab": 0.21350446428571426 }, { "learning_rate": 4.635601198741607e-08, "loss": 0.0042, "step": 1212, "train/R_acc": 0.06875, "train/R_penalty": -0.5, "train/R_reason": 0.19491433422937576, "train/R_vocab": 0.24547483766233766 }, { "learning_rate": 4.573040593634092e-08, "loss": 0.0204, "step": 1214, "train/R_acc": 0.0, "train/R_penalty": -0.7125, "train/R_reason": 0.2036836381635984, "train/R_vocab": 0.2583333333333333 }, { "learning_rate": 4.510862486864317e-08, "loss": -0.0139, "step": 1216, "train/R_acc": 0.03125, "train/R_penalty": -0.5, "train/R_reason": 0.1935253778055071, "train/R_vocab": 0.2539434523809524 }, { "learning_rate": 4.449068042728063e-08, "loss": -0.0221, "step": 1218, "train/R_acc": 0.03125, "train/R_penalty": -0.671875, "train/R_reason": 0.17478719789421077, "train/R_vocab": 0.2385044642857143 }, { "learning_rate": 4.3876584183369454e-08, "loss": 0.0401, "step": 1220, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.12396243023486506, "train/R_vocab": 0.16953125 }, { "learning_rate": 4.326634763596784e-08, "loss": 0.0362, "step": 1222, "train/R_acc": 0.003125, "train/R_penalty": -0.634375, "train/R_reason": 0.21970153796136732, "train/R_vocab": 0.23229166666666667 }, { "learning_rate": 4.265998221186023e-08, "loss": 0.0579, "step": 1224, "train/R_acc": 0.003125, "train/R_penalty": -0.696875, "train/R_reason": 0.25118929349410624, "train/R_vocab": 0.3527901785714286 }, { "learning_rate": 4.205749926534391e-08, "loss": 0.0325, "step": 1226, "train/R_acc": 0.0, "train/R_penalty": -0.53125, "train/R_reason": 0.20960885628620934, "train/R_vocab": 0.24062499999999998 }, { "learning_rate": 4.145891007801588e-08, "loss": -0.012, "step": 1228, "train/R_acc": 0.037500000000000006, "train/R_penalty": -0.375, "train/R_reason": 0.2440708995563252, "train/R_vocab": 0.31674107142857144 }, { "learning_rate": 4.086422585856189e-08, "loss": 0.0635, "step": 1230, "train/R_acc": 0.03125, "train/R_penalty": -0.5, "train/R_reason": 0.16689617467212345, "train/R_vocab": 0.24068700396825393 }, { "learning_rate": 4.0273457742546566e-08, "loss": -0.0426, "step": 1232, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.192806526715886, "train/R_vocab": 0.2780505952380953 }, { "learning_rate": 3.968661679220467e-08, "loss": 0.0384, "step": 1234, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.17224863139672328, "train/R_vocab": 0.23426339285714287 }, { "learning_rate": 3.9103713996234355e-08, "loss": 0.0036, "step": 1236, "train/R_acc": 0.03125, "train/R_penalty": -0.734375, "train/R_reason": 0.17680126111067432, "train/R_vocab": 0.2612723214285714 }, { "learning_rate": 3.852476026959089e-08, "loss": 0.0172, "step": 1238, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.640625, "train/R_reason": 0.23625775975212537, "train/R_vocab": 0.2917410714285715 }, { "learning_rate": 3.794976645328266e-08, "loss": 0.0608, "step": 1240, "train/R_acc": 0.00625, "train/R_penalty": -0.49375, "train/R_reason": 0.20085612003970227, "train/R_vocab": 0.29375 }, { "learning_rate": 3.737874331416807e-08, "loss": -0.0099, "step": 1242, "train/R_acc": 0.06875, "train/R_penalty": -0.53125, "train/R_reason": 0.2067606031031175, "train/R_vocab": 0.1792782738095238 }, { "learning_rate": 3.681170154475391e-08, "loss": -0.0105, "step": 1244, "train/R_acc": 0.003125, "train/R_penalty": -0.609375, "train/R_reason": 0.24829657601476124, "train/R_vocab": 0.3546875 }, { "learning_rate": 3.624865176299499e-08, "loss": -0.0273, "step": 1246, "train/R_acc": 0.003125, "train/R_penalty": -0.484375, "train/R_reason": 0.2728674749829552, "train/R_vocab": 0.42109375000000004 }, { "learning_rate": 3.568960451209574e-08, "loss": 0.0602, "step": 1248, "train/R_acc": 0.0375, "train/R_penalty": -0.546875, "train/R_reason": 0.2555940677098582, "train/R_vocab": 0.41767113095238095 }, { "learning_rate": 3.513457026031216e-08, "loss": -0.0627, "step": 1250, "train/R_acc": 0.0, "train/R_penalty": -0.53125, "train/R_reason": 0.20621535738161093, "train/R_vocab": 0.16438492063492063 }, { "learning_rate": 3.458355940075652e-08, "loss": 0.0626, "step": 1252, "train/R_acc": 0.03125, "train/R_penalty": -0.53125, "train/R_reason": 0.25723943167289565, "train/R_vocab": 0.290625 }, { "learning_rate": 3.403658225120212e-08, "loss": -0.0653, "step": 1254, "train/R_acc": 0.03125, "train/R_penalty": -0.53125, "train/R_reason": 0.2280530069453393, "train/R_vocab": 0.26473214285714286 }, { "learning_rate": 3.349364905389032e-08, "loss": -0.013, "step": 1256, "train/R_acc": 0.034375, "train/R_penalty": -0.4375, "train/R_reason": 0.20137595355438587, "train/R_vocab": 0.3265625 }, { "learning_rate": 3.295476997533905e-08, "loss": 0.0298, "step": 1258, "train/R_acc": 0.034375, "train/R_penalty": -0.59375, "train/R_reason": 0.20510172934689136, "train/R_vocab": 0.3380580357142857 }, { "learning_rate": 3.2419955106151893e-08, "loss": -0.0064, "step": 1260, "train/R_acc": 0.00625, "train/R_penalty": -0.53125, "train/R_reason": 0.24581030179992447, "train/R_vocab": 0.27968750000000003 }, { "learning_rate": 3.188921446082946e-08, "loss": 0.0136, "step": 1262, "train/R_acc": 0.0, "train/R_penalty": -0.5, "train/R_reason": 0.19498686377833271, "train/R_vocab": 0.21171875 }, { "learning_rate": 3.1362557977582e-08, "loss": 0.0327, "step": 1264, "train/R_acc": 0.003125, "train/R_penalty": -0.65625, "train/R_reason": 0.2546377151605037, "train/R_vocab": 0.38132440476190477 }, { "learning_rate": 3.0839995518142844e-08, "loss": 0.0813, "step": 1266, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.578125, "train/R_reason": 0.23407475867218072, "train/R_vocab": 0.25262896825396824 }, { "learning_rate": 3.032153686758432e-08, "loss": 0.0111, "step": 1268, "train/R_acc": 0.0, "train/R_penalty": -0.75, "train/R_reason": 0.13893382364177903, "train/R_vocab": 0.1873139880952381 }, { "learning_rate": 2.980719173413396e-08, "loss": -0.0655, "step": 1270, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.241975792190938, "train/R_vocab": 0.359375 }, { "learning_rate": 2.9296969748993093e-08, "loss": 0.0107, "step": 1272, "train/R_acc": 0.03125, "train/R_penalty": -0.53125, "train/R_reason": 0.24308125423194502, "train/R_vocab": 0.18370535714285716 }, { "learning_rate": 2.8790880466156464e-08, "loss": 0.0368, "step": 1274, "train/R_acc": 0.0, "train/R_penalty": -0.571875, "train/R_reason": 0.21833878646152022, "train/R_vocab": 0.2832217261904762 }, { "learning_rate": 2.8288933362233148e-08, "loss": 0.0016, "step": 1276, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.205687843726902, "train/R_vocab": 0.23705357142857142 }, { "learning_rate": 2.7791137836269158e-08, "loss": -0.0029, "step": 1278, "train/R_acc": 0.003125, "train/R_penalty": -0.640625, "train/R_reason": 0.2181107503992247, "train/R_vocab": 0.20520833333333333 }, { "learning_rate": 2.7297503209571665e-08, "loss": 0.1301, "step": 1280, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.16072625927287004, "train/R_vocab": 0.22611607142857143 }, { "learning_rate": 2.680803872553408e-08, "loss": 0.0204, "step": 1282, "train/R_acc": 0.00625, "train/R_penalty": -0.525, "train/R_reason": 0.18120856948997982, "train/R_vocab": 0.2267981150793651 }, { "learning_rate": 2.632275354946342e-08, "loss": 0.0174, "step": 1284, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.55625, "train/R_reason": 0.19887443410752542, "train/R_vocab": 0.20744047619047618 }, { "learning_rate": 2.5841656768408216e-08, "loss": 0.0125, "step": 1286, "train/R_acc": 0.003125, "train/R_penalty": -0.53125, "train/R_reason": 0.23872779990823945, "train/R_vocab": 0.28046875 }, { "learning_rate": 2.5364757390988616e-08, "loss": 0.0652, "step": 1288, "train/R_acc": 0.003125, "train/R_penalty": -0.546875, "train/R_reason": 0.19387491540524032, "train/R_vocab": 0.23203125 }, { "learning_rate": 2.489206434722785e-08, "loss": -0.012, "step": 1290, "train/R_acc": 0.0, "train/R_penalty": -0.515625, "train/R_reason": 0.18941429133549656, "train/R_vocab": 0.23415178571428572 }, { "learning_rate": 2.4423586488384524e-08, "loss": 0.0179, "step": 1292, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.248278845576544, "train/R_vocab": 0.22329545454545455 }, { "learning_rate": 2.395933258678745e-08, "loss": 0.0178, "step": 1294, "train/R_acc": 0.0, "train/R_penalty": -0.609375, "train/R_reason": 0.20548558553050647, "train/R_vocab": 0.3214905753968254 }, { "learning_rate": 2.349931133567096e-08, "loss": -0.0004, "step": 1296, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.2386328952490884, "train/R_vocab": 0.23169642857142858 }, { "learning_rate": 2.3043531349012353e-08, "loss": -0.0659, "step": 1298, "train/R_acc": 0.034375, "train/R_penalty": -0.671875, "train/R_reason": 0.15007159038519916, "train/R_vocab": 0.19921875 }, { "learning_rate": 2.259200116137039e-08, "loss": 0.0095, "step": 1300, "train/R_acc": 0.0, "train/R_penalty": -0.65625, "train/R_reason": 0.20368911805265966, "train/R_vocab": 0.2474330357142857 }, { "eval/R_acc": 0.0025, "eval/R_penalty": -0.5625, "eval/R_reason": 0.22623212762462228, "eval/R_vocab": 0.3164285714285714, "step": 1300 }, { "step": 1300 }, { "learning_rate": 2.214472922772584e-08, "loss": 0.0957, "step": 1302, "train/R_acc": 0.003125, "train/R_penalty": -0.625, "train/R_reason": 0.22116478834904366, "train/R_vocab": 0.29776785714285714 }, { "learning_rate": 2.1701723923322673e-08, "loss": -0.0749, "step": 1304, "train/R_acc": 0.065625, "train/R_penalty": -0.53125, "train/R_reason": 0.21324021225069947, "train/R_vocab": 0.2928977272727273 }, { "learning_rate": 2.1262993543511715e-08, "loss": 9.305, "step": 1306, "train/R_acc": 0.0375, "train/R_penalty": -0.484375, "train/R_reason": 0.23529231846467052, "train/R_vocab": 0.3332589285714286 }, { "learning_rate": 2.082854630359493e-08, "loss": 0.0331, "step": 1308, "train/R_acc": 0.0, "train/R_penalty": -0.421875, "train/R_reason": 0.23986047807887131, "train/R_vocab": 0.36618303571428573 }, { "learning_rate": 2.03983903386718e-08, "loss": 0.0118, "step": 1310, "train/R_acc": 0.0, "train/R_penalty": -0.578125, "train/R_reason": 0.2656667530369191, "train/R_vocab": 0.22192460317460316 }, { "learning_rate": 1.997253370348703e-08, "loss": 0.0556, "step": 1312, "train/R_acc": 0.0, "train/R_penalty": -0.546875, "train/R_reason": 0.2476736629649698, "train/R_vocab": 0.31614583333333335 }, { "learning_rate": 1.9550984372279476e-08, "loss": -0.037, "step": 1314, "train/R_acc": 0.00625, "train/R_penalty": -0.53125, "train/R_reason": 0.19878604497907437, "train/R_vocab": 0.27142857142857146 }, { "learning_rate": 1.9133750238633117e-08, "loss": 0.0175, "step": 1316, "train/R_acc": 0.00625, "train/R_penalty": -0.634375, "train/R_reason": 0.2233584354922284, "train/R_vocab": 0.23705357142857142 }, { "learning_rate": 1.872083911532907e-08, "loss": 0.0006, "step": 1318, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.53125, "train/R_reason": 0.2794139552851759, "train/R_vocab": 0.37043650793650795 }, { "learning_rate": 1.8312258734199253e-08, "loss": -0.0277, "step": 1320, "train/R_acc": 0.03125, "train/R_penalty": -0.46875, "train/R_reason": 0.19046029498161032, "train/R_vocab": 0.2759548611111111 }, { "learning_rate": 1.7908016745981856e-08, "loss": 0.0615, "step": 1322, "train/R_acc": 0.00625, "train/R_penalty": -0.515625, "train/R_reason": 0.1939617321953623, "train/R_vocab": 0.2484623015873016 }, { "learning_rate": 1.7508120720177795e-08, "loss": -0.0239, "step": 1324, "train/R_acc": 0.00625, "train/R_penalty": -0.53125, "train/R_reason": 0.21866868750973806, "train/R_vocab": 0.2526785714285714 }, { "learning_rate": 1.7112578144909047e-08, "loss": -0.0744, "step": 1326, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.23171087837305318, "train/R_vocab": 0.28158482142857144 }, { "learning_rate": 1.6721396426778667e-08, "loss": 0.0113, "step": 1328, "train/R_acc": 0.00625, "train/R_penalty": -0.49375, "train/R_reason": 0.2527234820261308, "train/R_vocab": 0.290687003968254 }, { "learning_rate": 1.6334582890731697e-08, "loss": 0.0423, "step": 1330, "train/R_acc": 0.0, "train/R_penalty": -0.625, "train/R_reason": 0.18629708667365014, "train/R_vocab": 0.1819940476190476 }, { "learning_rate": 1.595214477991841e-08, "loss": 0.027, "step": 1332, "train/R_acc": 0.065625, "train/R_penalty": -0.5625, "train/R_reason": 0.23647010980361977, "train/R_vocab": 0.2343625992063492 }, { "learning_rate": 1.557408925555831e-08, "loss": 0.0619, "step": 1334, "train/R_acc": 0.03125, "train/R_penalty": -0.640625, "train/R_reason": 0.2178143312883945, "train/R_vocab": 0.28158482142857144 }, { "learning_rate": 1.5200423396806272e-08, "loss": -0.006, "step": 1336, "train/R_acc": 0.003125, "train/R_penalty": -0.40625, "train/R_reason": 0.26096178007275994, "train/R_vocab": 0.3358134920634921 }, { "learning_rate": 1.4831154200620033e-08, "loss": -0.0105, "step": 1338, "train/R_acc": 0.034375, "train/R_penalty": -0.578125, "train/R_reason": 0.18847235645846208, "train/R_vocab": 0.23511904761904762 }, { "learning_rate": 1.4466288581628906e-08, "loss": 0.0143, "step": 1340, "train/R_acc": 0.009375000000000001, "train/R_penalty": -0.578125, "train/R_reason": 0.23179656802725537, "train/R_vocab": 0.2564732142857143 }, { "learning_rate": 1.4105833372004523e-08, "loss": 0.0246, "step": 1342, "train/R_acc": 0.03125, "train/R_penalty": -0.546875, "train/R_reason": 0.1816783870148443, "train/R_vocab": 0.21495535714285713 }, { "learning_rate": 1.3749795321332885e-08, "loss": -0.0062, "step": 1344, "train/R_acc": 0.03125, "train/R_penalty": -0.546875, "train/R_reason": 0.22414561936584018, "train/R_vocab": 0.2232886904761905 }, { "learning_rate": 1.3398181096487882e-08, "loss": 0.0182, "step": 1346, "train/R_acc": 0.065625, "train/R_penalty": -0.5625, "train/R_reason": 0.2680587101569149, "train/R_vocab": 0.3170138888888889 }, { "learning_rate": 1.3050997281506549e-08, "loss": 0.0831, "step": 1348, "train/R_acc": 0.003125, "train/R_penalty": -0.68125, "train/R_reason": 0.2494493257892117, "train/R_vocab": 0.18958333333333333 }, { "learning_rate": 1.2708250377465645e-08, "loss": -0.004, "step": 1350, "train/R_acc": 0.00625, "train/R_penalty": -0.665625, "train/R_reason": 0.2093549139749035, "train/R_vocab": 0.1791497564935065 } ], "logging_steps": 2, "max_steps": 1497, "num_input_tokens_seen": 13748095, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }