| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9018036072144289, | |
| "eval_steps": 100, | |
| "global_step": 1350, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "learning_rate": 1.111111111111111e-08, | |
| "loss": 0.0176, | |
| "step": 2, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.1771926448030963, | |
| "train/R_vocab": 0.27042410714285714 | |
| }, | |
| { | |
| "learning_rate": 3.3333333333333334e-08, | |
| "loss": 0.0255, | |
| "step": 4, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.20292794300576886, | |
| "train/R_vocab": 0.13051835317460317 | |
| }, | |
| { | |
| "learning_rate": 5.555555555555555e-08, | |
| "loss": -0.0402, | |
| "step": 6, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.21681085047252255, | |
| "train/R_vocab": 0.3141617063492063 | |
| }, | |
| { | |
| "learning_rate": 7.777777777777778e-08, | |
| "loss": -0.0512, | |
| "step": 8, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.22407962891811065, | |
| "train/R_vocab": 0.2578125 | |
| }, | |
| { | |
| "learning_rate": 1e-07, | |
| "loss": 0.0653, | |
| "step": 10, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.24027254265023745, | |
| "train/R_vocab": 0.290625 | |
| }, | |
| { | |
| "learning_rate": 1.2222222222222222e-07, | |
| "loss": 0.0747, | |
| "step": 12, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2086565261070774, | |
| "train/R_vocab": 0.2722098214285714 | |
| }, | |
| { | |
| "learning_rate": 1.4444444444444442e-07, | |
| "loss": -0.0073, | |
| "step": 14, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.78125, | |
| "train/R_reason": 0.18557344633053446, | |
| "train/R_vocab": 0.23597470238095236 | |
| }, | |
| { | |
| "learning_rate": 1.6666666666666665e-07, | |
| "loss": 0.0011, | |
| "step": 16, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.18460204443700487, | |
| "train/R_vocab": 0.28095238095238095 | |
| }, | |
| { | |
| "learning_rate": 1.8888888888888888e-07, | |
| "loss": 0.0511, | |
| "step": 18, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.1691650605005834, | |
| "train/R_vocab": 0.22633928571428572 | |
| }, | |
| { | |
| "learning_rate": 2.111111111111111e-07, | |
| "loss": -0.0247, | |
| "step": 20, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.23429908163761645, | |
| "train/R_vocab": 0.3410714285714286 | |
| }, | |
| { | |
| "learning_rate": 2.3333333333333333e-07, | |
| "loss": 0.0088, | |
| "step": 22, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.15088100811244767, | |
| "train/R_vocab": 0.2056919642857143 | |
| }, | |
| { | |
| "learning_rate": 2.5555555555555553e-07, | |
| "loss": 0.0366, | |
| "step": 24, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.2605457318222165, | |
| "train/R_vocab": 0.29765625 | |
| }, | |
| { | |
| "learning_rate": 2.7777777777777776e-07, | |
| "loss": 0.0558, | |
| "step": 26, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.24175666877540153, | |
| "train/R_vocab": 0.26484375000000004 | |
| }, | |
| { | |
| "learning_rate": 3e-07, | |
| "loss": 0.0588, | |
| "step": 28, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.540625, | |
| "train/R_reason": 0.2054192775694245, | |
| "train/R_vocab": 0.2109747023809524 | |
| }, | |
| { | |
| "learning_rate": 3.222222222222222e-07, | |
| "loss": -0.1103, | |
| "step": 30, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.18972739597143928, | |
| "train/R_vocab": 0.2044642857142857 | |
| }, | |
| { | |
| "learning_rate": 3.4444444444444444e-07, | |
| "loss": 0.0162, | |
| "step": 32, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.16386884044279187, | |
| "train/R_vocab": 0.26328125 | |
| }, | |
| { | |
| "learning_rate": 3.666666666666666e-07, | |
| "loss": 0.0241, | |
| "step": 34, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.1991015810053379, | |
| "train/R_vocab": 0.3072916666666667 | |
| }, | |
| { | |
| "learning_rate": 3.888888888888889e-07, | |
| "loss": 0.0114, | |
| "step": 36, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.1542880167619275, | |
| "train/R_vocab": 0.15633680555555557 | |
| }, | |
| { | |
| "learning_rate": 4.1111111111111107e-07, | |
| "loss": 0.0418, | |
| "step": 38, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.21628990497316553, | |
| "train/R_vocab": 0.32326388888888885 | |
| }, | |
| { | |
| "learning_rate": 4.3333333333333335e-07, | |
| "loss": -0.0085, | |
| "step": 40, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.540625, | |
| "train/R_reason": 0.16669781447815488, | |
| "train/R_vocab": 0.21458333333333335 | |
| }, | |
| { | |
| "learning_rate": 4.555555555555555e-07, | |
| "loss": -0.0614, | |
| "step": 42, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.23833932777613548, | |
| "train/R_vocab": 0.24285714285714288 | |
| }, | |
| { | |
| "learning_rate": 4.777777777777778e-07, | |
| "loss": 0.004, | |
| "step": 44, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.2297590496740362, | |
| "train/R_vocab": 0.1607142857142857 | |
| }, | |
| { | |
| "learning_rate": 5e-07, | |
| "loss": 0.0491, | |
| "step": 46, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.2037377285506986, | |
| "train/R_vocab": 0.2598958333333333 | |
| }, | |
| { | |
| "learning_rate": 4.999976528920434e-07, | |
| "loss": -0.083, | |
| "step": 48, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.22468232799435314, | |
| "train/R_vocab": 0.2745907738095238 | |
| }, | |
| { | |
| "learning_rate": 4.999906116122447e-07, | |
| "loss": -0.0036, | |
| "step": 50, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.22593521940016348, | |
| "train/R_vocab": 0.2805803571428572 | |
| }, | |
| { | |
| "learning_rate": 4.999788762928172e-07, | |
| "loss": 0.0107, | |
| "step": 52, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.821875, | |
| "train/R_reason": 0.20522814513215548, | |
| "train/R_vocab": 0.14702380952380953 | |
| }, | |
| { | |
| "learning_rate": 4.999624471541134e-07, | |
| "loss": 0.029, | |
| "step": 54, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.78125, | |
| "train/R_reason": 0.21737758529192291, | |
| "train/R_vocab": 0.19330357142857144 | |
| }, | |
| { | |
| "learning_rate": 4.99941324504621e-07, | |
| "loss": -0.0246, | |
| "step": 56, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2251848744560886, | |
| "train/R_vocab": 0.28158482142857144 | |
| }, | |
| { | |
| "learning_rate": 4.999155087409571e-07, | |
| "loss": -0.0517, | |
| "step": 58, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.571875, | |
| "train/R_reason": 0.18310813181259872, | |
| "train/R_vocab": 0.24776785714285715 | |
| }, | |
| { | |
| "learning_rate": 4.998850003478606e-07, | |
| "loss": 0.0598, | |
| "step": 60, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.2082140947052848, | |
| "train/R_vocab": 0.26149553571428574 | |
| }, | |
| { | |
| "learning_rate": 4.998497998981838e-07, | |
| "loss": 0.0027, | |
| "step": 62, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.828125, | |
| "train/R_reason": 0.2611822631590372, | |
| "train/R_vocab": 0.39639136904761907 | |
| }, | |
| { | |
| "learning_rate": 4.998099080528803e-07, | |
| "loss": 0.0385, | |
| "step": 64, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.2213369943749453, | |
| "train/R_vocab": 0.3606026785714286 | |
| }, | |
| { | |
| "learning_rate": 4.997653255609941e-07, | |
| "loss": -0.0523, | |
| "step": 66, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.20860609604648056, | |
| "train/R_vocab": 0.2251488095238095 | |
| }, | |
| { | |
| "learning_rate": 4.997160532596447e-07, | |
| "loss": 0.0352, | |
| "step": 68, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.790625, | |
| "train/R_reason": 0.19482757334666323, | |
| "train/R_vocab": 0.23776041666666667 | |
| }, | |
| { | |
| "learning_rate": 4.996620920740111e-07, | |
| "loss": 0.0103, | |
| "step": 70, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.25304049429757086, | |
| "train/R_vocab": 0.299516369047619 | |
| }, | |
| { | |
| "learning_rate": 4.996034430173153e-07, | |
| "loss": 0.0247, | |
| "step": 72, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.58125, | |
| "train/R_reason": 0.27724208008758466, | |
| "train/R_vocab": 0.3654637896825397 | |
| }, | |
| { | |
| "learning_rate": 4.995401071908025e-07, | |
| "loss": -0.0689, | |
| "step": 74, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.18864259893071442, | |
| "train/R_vocab": 0.265625 | |
| }, | |
| { | |
| "learning_rate": 4.99472085783721e-07, | |
| "loss": 0.0669, | |
| "step": 76, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.775, | |
| "train/R_reason": 0.24872595893355823, | |
| "train/R_vocab": 0.3293526785714286 | |
| }, | |
| { | |
| "learning_rate": 4.993993800732995e-07, | |
| "loss": -0.0369, | |
| "step": 78, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.26486544822313024, | |
| "train/R_vocab": 0.3398065476190476 | |
| }, | |
| { | |
| "learning_rate": 4.993219914247231e-07, | |
| "loss": 0.0685, | |
| "step": 80, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.1711866679882719, | |
| "train/R_vocab": 0.2575892857142857 | |
| }, | |
| { | |
| "learning_rate": 4.992399212911082e-07, | |
| "loss": 0.0382, | |
| "step": 82, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.22934020748837214, | |
| "train/R_vocab": 0.31875 | |
| }, | |
| { | |
| "learning_rate": 4.99153171213474e-07, | |
| "loss": 0.083, | |
| "step": 84, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.16303826493835916, | |
| "train/R_vocab": 0.18002232142857144 | |
| }, | |
| { | |
| "learning_rate": 4.990617428207153e-07, | |
| "loss": 0.0433, | |
| "step": 86, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.84375, | |
| "train/R_reason": 0.21139857471142573, | |
| "train/R_vocab": 0.28266369047619044 | |
| }, | |
| { | |
| "learning_rate": 4.989656378295704e-07, | |
| "loss": 0.0554, | |
| "step": 88, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.2142568618009861, | |
| "train/R_vocab": 0.3390625 | |
| }, | |
| { | |
| "learning_rate": 4.988648580445895e-07, | |
| "loss": 0.0648, | |
| "step": 90, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.17382571925607848, | |
| "train/R_vocab": 0.1894345238095238 | |
| }, | |
| { | |
| "learning_rate": 4.987594053581012e-07, | |
| "loss": -0.0241, | |
| "step": 92, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.21781325533595836, | |
| "train/R_vocab": 0.33205605158730156 | |
| }, | |
| { | |
| "learning_rate": 4.98649281750176e-07, | |
| "loss": 0.1022, | |
| "step": 94, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.1999237909283059, | |
| "train/R_vocab": 0.32064732142857144 | |
| }, | |
| { | |
| "learning_rate": 4.985344892885899e-07, | |
| "loss": 0.0821, | |
| "step": 96, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.23594255343132425, | |
| "train/R_vocab": 0.30199652777777775 | |
| }, | |
| { | |
| "learning_rate": 4.984150301287853e-07, | |
| "loss": 0.0096, | |
| "step": 98, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.80625, | |
| "train/R_reason": 0.1900727268525148, | |
| "train/R_vocab": 0.20625000000000002 | |
| }, | |
| { | |
| "learning_rate": 4.982909065138305e-07, | |
| "loss": 0.0247, | |
| "step": 100, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.21411060850160504, | |
| "train/R_vocab": 0.22946428571428573 | |
| }, | |
| { | |
| "eval/R_acc": 0.002777777777777778, | |
| "eval/R_penalty": -0.6847222222222222, | |
| "eval/R_reason": 0.2572056199703138, | |
| "eval/R_vocab": 0.3255787037037037, | |
| "step": 100 | |
| }, | |
| { | |
| "step": 100 | |
| }, | |
| { | |
| "learning_rate": 4.981621207743778e-07, | |
| "loss": -0.0004, | |
| "step": 102, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.765625, | |
| "train/R_reason": 0.1735908404636029, | |
| "train/R_vocab": 0.15747767857142858 | |
| }, | |
| { | |
| "learning_rate": 4.980286753286194e-07, | |
| "loss": 0.0155, | |
| "step": 104, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.2016416402158515, | |
| "train/R_vocab": 0.3698660714285714 | |
| }, | |
| { | |
| "learning_rate": 4.978905726822423e-07, | |
| "loss": -0.0257, | |
| "step": 106, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.2019264344648219, | |
| "train/R_vocab": 0.2499627976190476 | |
| }, | |
| { | |
| "learning_rate": 4.97747815428381e-07, | |
| "loss": 0.0757, | |
| "step": 108, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.2609304111036691, | |
| "train/R_vocab": 0.22472718253968255 | |
| }, | |
| { | |
| "learning_rate": 4.976004062475691e-07, | |
| "loss": -0.0376, | |
| "step": 110, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.24456802402586483, | |
| "train/R_vocab": 0.30167410714285714 | |
| }, | |
| { | |
| "learning_rate": 4.974483479076885e-07, | |
| "loss": -0.0155, | |
| "step": 112, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.734375, | |
| "train/R_reason": 0.15732482982112658, | |
| "train/R_vocab": 0.20078125 | |
| }, | |
| { | |
| "learning_rate": 4.972916432639181e-07, | |
| "loss": 0.0107, | |
| "step": 114, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.6125, | |
| "train/R_reason": 0.19719755508097864, | |
| "train/R_vocab": 0.18292410714285715 | |
| }, | |
| { | |
| "learning_rate": 4.971302952586796e-07, | |
| "loss": -0.0536, | |
| "step": 116, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.1615191195936923, | |
| "train/R_vocab": 0.20714285714285713 | |
| }, | |
| { | |
| "learning_rate": 4.969643069215823e-07, | |
| "loss": 0.03, | |
| "step": 118, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.2057090874466078, | |
| "train/R_vocab": 0.26006944444444446 | |
| }, | |
| { | |
| "learning_rate": 4.967936813693668e-07, | |
| "loss": 0.0669, | |
| "step": 120, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.17123245123013492, | |
| "train/R_vocab": 0.22881944444444446 | |
| }, | |
| { | |
| "learning_rate": 4.966184218058457e-07, | |
| "loss": -0.01, | |
| "step": 122, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.25008522769707353, | |
| "train/R_vocab": 0.2768725198412698 | |
| }, | |
| { | |
| "learning_rate": 4.96438531521844e-07, | |
| "loss": 0.0227, | |
| "step": 124, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.228776542397296, | |
| "train/R_vocab": 0.19776785714285716 | |
| }, | |
| { | |
| "learning_rate": 4.962540138951371e-07, | |
| "loss": 0.0342, | |
| "step": 126, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.20827985353158485, | |
| "train/R_vocab": 0.3024440836940837 | |
| }, | |
| { | |
| "learning_rate": 4.960648723903872e-07, | |
| "loss": 0.0418, | |
| "step": 128, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.19681757054800564, | |
| "train/R_vocab": 0.22455357142857144 | |
| }, | |
| { | |
| "learning_rate": 4.958711105590785e-07, | |
| "loss": -0.0196, | |
| "step": 130, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.22729280568853277, | |
| "train/R_vocab": 0.19475446428571427 | |
| }, | |
| { | |
| "learning_rate": 4.956727320394506e-07, | |
| "loss": 0.0963, | |
| "step": 132, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.84375, | |
| "train/R_reason": 0.20146228335738753, | |
| "train/R_vocab": 0.24609995039682542 | |
| }, | |
| { | |
| "learning_rate": 4.954697405564299e-07, | |
| "loss": 0.0259, | |
| "step": 134, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.228208707691525, | |
| "train/R_vocab": 0.23055217352092353 | |
| }, | |
| { | |
| "learning_rate": 4.952621399215597e-07, | |
| "loss": 0.0781, | |
| "step": 136, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.8125, | |
| "train/R_reason": 0.1811115376120021, | |
| "train/R_vocab": 0.2921875 | |
| }, | |
| { | |
| "learning_rate": 4.95049934032929e-07, | |
| "loss": 0.0659, | |
| "step": 138, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.2199008430371848, | |
| "train/R_vocab": 0.22447916666666667 | |
| }, | |
| { | |
| "learning_rate": 4.948331268750988e-07, | |
| "loss": -0.0087, | |
| "step": 140, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.796875, | |
| "train/R_reason": 0.16414720324183496, | |
| "train/R_vocab": 0.20345982142857144 | |
| }, | |
| { | |
| "learning_rate": 4.946117225190273e-07, | |
| "loss": 0.0109, | |
| "step": 142, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.20413731109417932, | |
| "train/R_vocab": 0.27366071428571426 | |
| }, | |
| { | |
| "learning_rate": 4.943857251219942e-07, | |
| "loss": -0.0291, | |
| "step": 144, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.2114743293926566, | |
| "train/R_vocab": 0.24921875 | |
| }, | |
| { | |
| "learning_rate": 4.941551389275217e-07, | |
| "loss": -0.0249, | |
| "step": 146, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.19512092302735518, | |
| "train/R_vocab": 0.17879464285714286 | |
| }, | |
| { | |
| "learning_rate": 4.939199682652952e-07, | |
| "loss": -0.03, | |
| "step": 148, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.6125, | |
| "train/R_reason": 0.19263372463344114, | |
| "train/R_vocab": 0.2714285714285714 | |
| }, | |
| { | |
| "learning_rate": 4.936802175510823e-07, | |
| "loss": -0.1025, | |
| "step": 150, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.21358490777949846, | |
| "train/R_vocab": 0.3302455357142857 | |
| }, | |
| { | |
| "learning_rate": 4.934358912866497e-07, | |
| "loss": 0.0293, | |
| "step": 152, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.16662975968037924, | |
| "train/R_vocab": 0.1699280753968254 | |
| }, | |
| { | |
| "learning_rate": 4.931869940596779e-07, | |
| "loss": 0.0662, | |
| "step": 154, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.20091746530920346, | |
| "train/R_vocab": 0.27994791666666663 | |
| }, | |
| { | |
| "learning_rate": 4.929335305436764e-07, | |
| "loss": 0.0355, | |
| "step": 156, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.22218261967350866, | |
| "train/R_vocab": 0.3209821428571429 | |
| }, | |
| { | |
| "learning_rate": 4.926755054978951e-07, | |
| "loss": 0.0457, | |
| "step": 158, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.22181716803364154, | |
| "train/R_vocab": 0.29573863636363634 | |
| }, | |
| { | |
| "learning_rate": 4.92412923767235e-07, | |
| "loss": -0.0201, | |
| "step": 160, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.24103408318415312, | |
| "train/R_vocab": 0.23645833333333333 | |
| }, | |
| { | |
| "learning_rate": 4.921457902821578e-07, | |
| "loss": -0.0013, | |
| "step": 162, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.21205287864979985, | |
| "train/R_vocab": 0.24444444444444446 | |
| }, | |
| { | |
| "learning_rate": 4.918741100585921e-07, | |
| "loss": -0.0549, | |
| "step": 164, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.21139952333180925, | |
| "train/R_vocab": 0.36577380952380956 | |
| }, | |
| { | |
| "learning_rate": 4.915978881978406e-07, | |
| "loss": 0.0116, | |
| "step": 166, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.21990626286225867, | |
| "train/R_vocab": 0.3359623015873016 | |
| }, | |
| { | |
| "learning_rate": 4.913171298864836e-07, | |
| "loss": -0.0194, | |
| "step": 168, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.19859495503742927, | |
| "train/R_vocab": 0.3138020833333333 | |
| }, | |
| { | |
| "learning_rate": 4.910318403962813e-07, | |
| "loss": 0.0306, | |
| "step": 170, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.15012936918136355, | |
| "train/R_vocab": 0.2549107142857143 | |
| }, | |
| { | |
| "learning_rate": 4.907420250840761e-07, | |
| "loss": 0.0817, | |
| "step": 172, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.20995580823960475, | |
| "train/R_vocab": 0.20602678571428573 | |
| }, | |
| { | |
| "learning_rate": 4.9044768939169e-07, | |
| "loss": -0.104, | |
| "step": 174, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.18253049494341367, | |
| "train/R_vocab": 0.20379464285714285 | |
| }, | |
| { | |
| "learning_rate": 4.901488388458247e-07, | |
| "loss": 0.0418, | |
| "step": 176, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.2305279249641788, | |
| "train/R_vocab": 0.38442460317460314 | |
| }, | |
| { | |
| "learning_rate": 4.898454790579558e-07, | |
| "loss": -0.0033, | |
| "step": 178, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.18641040374410697, | |
| "train/R_vocab": 0.25188492063492063 | |
| }, | |
| { | |
| "learning_rate": 4.895376157242288e-07, | |
| "loss": 0.0036, | |
| "step": 180, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.19709747248451465, | |
| "train/R_vocab": 0.28185763888888893 | |
| }, | |
| { | |
| "learning_rate": 4.892252546253514e-07, | |
| "loss": -0.0037, | |
| "step": 182, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2449271246503006, | |
| "train/R_vocab": 0.30446428571428574 | |
| }, | |
| { | |
| "learning_rate": 4.889084016264858e-07, | |
| "loss": 0.0279, | |
| "step": 184, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.24078946289847405, | |
| "train/R_vocab": 0.3054935515873016 | |
| }, | |
| { | |
| "learning_rate": 4.88587062677137e-07, | |
| "loss": 0.0824, | |
| "step": 186, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.16765366722116085, | |
| "train/R_vocab": 0.22786458333333331 | |
| }, | |
| { | |
| "learning_rate": 4.882612438110429e-07, | |
| "loss": 0.0193, | |
| "step": 188, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.21543693992639024, | |
| "train/R_vocab": 0.3402901785714285 | |
| }, | |
| { | |
| "learning_rate": 4.8793095114606e-07, | |
| "loss": -0.0257, | |
| "step": 190, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.154932479362758, | |
| "train/R_vocab": 0.2091393849206349 | |
| }, | |
| { | |
| "learning_rate": 4.875961908840485e-07, | |
| "loss": -0.0201, | |
| "step": 192, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.20005776839850115, | |
| "train/R_vocab": 0.26316964285714284 | |
| }, | |
| { | |
| "learning_rate": 4.872569693107563e-07, | |
| "loss": 0.0163, | |
| "step": 194, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.2028767825912614, | |
| "train/R_vocab": 0.22935267857142855 | |
| }, | |
| { | |
| "learning_rate": 4.869132927957006e-07, | |
| "loss": -0.0312, | |
| "step": 196, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.20598954317804136, | |
| "train/R_vocab": 0.2894097222222222 | |
| }, | |
| { | |
| "learning_rate": 4.865651677920483e-07, | |
| "loss": -0.041, | |
| "step": 198, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.19447689115700523, | |
| "train/R_vocab": 0.21484375 | |
| }, | |
| { | |
| "learning_rate": 4.862126008364954e-07, | |
| "loss": -0.0021, | |
| "step": 200, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.24708335127072298, | |
| "train/R_vocab": 0.3235491071428571 | |
| }, | |
| { | |
| "eval/R_acc": 0.001388888888888889, | |
| "eval/R_penalty": -0.6875, | |
| "eval/R_reason": 0.26545862362041284, | |
| "eval/R_vocab": 0.3180500440917107, | |
| "step": 200 | |
| }, | |
| { | |
| "step": 200 | |
| }, | |
| { | |
| "learning_rate": 4.858555985491434e-07, | |
| "loss": 0.0539, | |
| "step": 202, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.12246968403633852, | |
| "train/R_vocab": 0.12857142857142856 | |
| }, | |
| { | |
| "learning_rate": 4.854941676333756e-07, | |
| "loss": -0.0105, | |
| "step": 204, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.20647794674779765, | |
| "train/R_vocab": 0.2294642857142857 | |
| }, | |
| { | |
| "learning_rate": 4.85128314875731e-07, | |
| "loss": 0.0418, | |
| "step": 206, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.22341563734148095, | |
| "train/R_vocab": 0.24494047619047618 | |
| }, | |
| { | |
| "learning_rate": 4.84758047145777e-07, | |
| "loss": 0.0338, | |
| "step": 208, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.21091575917752947, | |
| "train/R_vocab": 0.25234375 | |
| }, | |
| { | |
| "learning_rate": 4.843833713959802e-07, | |
| "loss": 0.0127, | |
| "step": 210, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.78125, | |
| "train/R_reason": 0.15006747203000467, | |
| "train/R_vocab": 0.23049355158730156 | |
| }, | |
| { | |
| "learning_rate": 4.840042946615761e-07, | |
| "loss": 0.0429, | |
| "step": 212, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.23417842297662064, | |
| "train/R_vocab": 0.28020833333333334 | |
| }, | |
| { | |
| "learning_rate": 4.836208240604368e-07, | |
| "loss": -0.0, | |
| "step": 214, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65, | |
| "train/R_reason": 0.20300438739011717, | |
| "train/R_vocab": 0.30078125 | |
| }, | |
| { | |
| "learning_rate": 4.832329667929376e-07, | |
| "loss": 0.0223, | |
| "step": 216, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.21329021924994432, | |
| "train/R_vocab": 0.2752604166666667 | |
| }, | |
| { | |
| "learning_rate": 4.828407301418217e-07, | |
| "loss": -0.0077, | |
| "step": 218, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.22684321233993932, | |
| "train/R_vocab": 0.2927331349206349 | |
| }, | |
| { | |
| "learning_rate": 4.824441214720628e-07, | |
| "loss": 0.0196, | |
| "step": 220, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.22135178051709037, | |
| "train/R_vocab": 0.3354166666666667 | |
| }, | |
| { | |
| "learning_rate": 4.820431482307281e-07, | |
| "loss": 0.0335, | |
| "step": 222, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.23161254238918147, | |
| "train/R_vocab": 0.3784598214285714 | |
| }, | |
| { | |
| "learning_rate": 4.816378179468374e-07, | |
| "loss": 0.0416, | |
| "step": 224, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.728125, | |
| "train/R_reason": 0.20212408310335858, | |
| "train/R_vocab": 0.2705357142857143 | |
| }, | |
| { | |
| "learning_rate": 4.812281382312223e-07, | |
| "loss": 0.0156, | |
| "step": 226, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.19926301994985757, | |
| "train/R_vocab": 0.35390625000000003 | |
| }, | |
| { | |
| "learning_rate": 4.808141167763826e-07, | |
| "loss": 0.0111, | |
| "step": 228, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.24773363470223553, | |
| "train/R_vocab": 0.3498015873015873 | |
| }, | |
| { | |
| "learning_rate": 4.803957613563431e-07, | |
| "loss": 0.1022, | |
| "step": 230, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.19832487010930644, | |
| "train/R_vocab": 0.2538690476190476 | |
| }, | |
| { | |
| "learning_rate": 4.799730798265063e-07, | |
| "loss": 0.1914, | |
| "step": 232, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.22378486607186504, | |
| "train/R_vocab": 0.28828125000000004 | |
| }, | |
| { | |
| "learning_rate": 4.795460801235058e-07, | |
| "loss": -0.036, | |
| "step": 234, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.17550109595621496, | |
| "train/R_vocab": 0.24095982142857145 | |
| }, | |
| { | |
| "learning_rate": 4.791147702650565e-07, | |
| "loss": 0.1009, | |
| "step": 236, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.2616533979112148, | |
| "train/R_vocab": 0.3763888888888889 | |
| }, | |
| { | |
| "learning_rate": 4.786791583498051e-07, | |
| "loss": -0.0144, | |
| "step": 238, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.20754413577165548, | |
| "train/R_vocab": 0.22721974206349208 | |
| }, | |
| { | |
| "learning_rate": 4.78239252557177e-07, | |
| "loss": 0.0564, | |
| "step": 240, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2701327994702766, | |
| "train/R_vocab": 0.3259424603174603 | |
| }, | |
| { | |
| "learning_rate": 4.777950611472233e-07, | |
| "loss": -0.0099, | |
| "step": 242, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.21945774918879585, | |
| "train/R_vocab": 0.2623883928571429 | |
| }, | |
| { | |
| "learning_rate": 4.773465924604656e-07, | |
| "loss": 0.0141, | |
| "step": 244, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.2201195420166109, | |
| "train/R_vocab": 0.26729910714285715 | |
| }, | |
| { | |
| "learning_rate": 4.768938549177392e-07, | |
| "loss": 0.052, | |
| "step": 246, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.2231016124642528, | |
| "train/R_vocab": 0.26313244047619044 | |
| }, | |
| { | |
| "learning_rate": 4.764368570200353e-07, | |
| "loss": 0.0239, | |
| "step": 248, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.12548248807792622, | |
| "train/R_vocab": 0.15256696428571428 | |
| }, | |
| { | |
| "learning_rate": 4.759756073483411e-07, | |
| "loss": 0.0264, | |
| "step": 250, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.765625, | |
| "train/R_reason": 0.17495784675781162, | |
| "train/R_vocab": 0.13359375 | |
| }, | |
| { | |
| "learning_rate": 4.7551011456347876e-07, | |
| "loss": 0.0271, | |
| "step": 252, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.759375, | |
| "train/R_reason": 0.2455444828427244, | |
| "train/R_vocab": 0.3765625 | |
| }, | |
| { | |
| "learning_rate": 4.750403874059428e-07, | |
| "loss": 0.035, | |
| "step": 254, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.2624393029582254, | |
| "train/R_vocab": 0.3123883928571428 | |
| }, | |
| { | |
| "learning_rate": 4.745664346957361e-07, | |
| "loss": -0.0167, | |
| "step": 256, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.734375, | |
| "train/R_reason": 0.2061035486282405, | |
| "train/R_vocab": 0.2132688492063492 | |
| }, | |
| { | |
| "learning_rate": 4.740882653322039e-07, | |
| "loss": 0.0071, | |
| "step": 258, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.25927699094783563, | |
| "train/R_vocab": 0.3599330357142857 | |
| }, | |
| { | |
| "learning_rate": 4.7360588829386736e-07, | |
| "loss": -0.0231, | |
| "step": 260, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.7125, | |
| "train/R_reason": 0.2040610708874982, | |
| "train/R_vocab": 0.23035714285714287 | |
| }, | |
| { | |
| "learning_rate": 4.7311931263825434e-07, | |
| "loss": 0.0593, | |
| "step": 262, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.14067559915832142, | |
| "train/R_vocab": 0.171875 | |
| }, | |
| { | |
| "learning_rate": 4.726285475017294e-07, | |
| "loss": 0.0213, | |
| "step": 264, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.2728806346759708, | |
| "train/R_vocab": 0.34206349206349207 | |
| }, | |
| { | |
| "learning_rate": 4.721336020993228e-07, | |
| "loss": -0.0176, | |
| "step": 266, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.22780919072556438, | |
| "train/R_vocab": 0.26149553571428574 | |
| }, | |
| { | |
| "learning_rate": 4.716344857245567e-07, | |
| "loss": 0.0807, | |
| "step": 268, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.23184996778054867, | |
| "train/R_vocab": 0.27254464285714286 | |
| }, | |
| { | |
| "learning_rate": 4.7113120774927136e-07, | |
| "loss": 0.0114, | |
| "step": 270, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.24027473582911105, | |
| "train/R_vocab": 0.41076388888888893 | |
| }, | |
| { | |
| "learning_rate": 4.706237776234486e-07, | |
| "loss": 0.0187, | |
| "step": 272, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.1528155999223981, | |
| "train/R_vocab": 0.1519097222222222 | |
| }, | |
| { | |
| "learning_rate": 4.7011220487503476e-07, | |
| "loss": 0.0688, | |
| "step": 274, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.15041989042686232, | |
| "train/R_vocab": 0.1591889880952381 | |
| }, | |
| { | |
| "learning_rate": 4.695964991097616e-07, | |
| "loss": 0.0498, | |
| "step": 276, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.20384797152303086, | |
| "train/R_vocab": 0.35360863095238093 | |
| }, | |
| { | |
| "learning_rate": 4.6907667001096585e-07, | |
| "loss": -0.0992, | |
| "step": 278, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.23776322399396135, | |
| "train/R_vocab": 0.4219246031746032 | |
| }, | |
| { | |
| "learning_rate": 4.685527273394078e-07, | |
| "loss": -0.0568, | |
| "step": 280, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.2305205389308027, | |
| "train/R_vocab": 0.21640624999999997 | |
| }, | |
| { | |
| "learning_rate": 4.680246809330874e-07, | |
| "loss": -0.0174, | |
| "step": 282, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.27570159502906066, | |
| "train/R_vocab": 0.23663194444444444 | |
| }, | |
| { | |
| "learning_rate": 4.6749254070706013e-07, | |
| "loss": -0.053, | |
| "step": 284, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.17788848371405203, | |
| "train/R_vocab": 0.159375 | |
| }, | |
| { | |
| "learning_rate": 4.669563166532503e-07, | |
| "loss": 0.0595, | |
| "step": 286, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.22202849960365972, | |
| "train/R_vocab": 0.2922619047619048 | |
| }, | |
| { | |
| "learning_rate": 4.6641601884026407e-07, | |
| "loss": -0.0247, | |
| "step": 288, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.20278024644809003, | |
| "train/R_vocab": 0.2072544642857143 | |
| }, | |
| { | |
| "learning_rate": 4.6587165741319967e-07, | |
| "loss": 0.0055, | |
| "step": 290, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.20358950838718215, | |
| "train/R_vocab": 0.30781250000000004 | |
| }, | |
| { | |
| "learning_rate": 4.6532324259345743e-07, | |
| "loss": 0.013, | |
| "step": 292, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.19116956517383865, | |
| "train/R_vocab": 0.19854910714285712 | |
| }, | |
| { | |
| "learning_rate": 4.647707846785477e-07, | |
| "loss": -0.0187, | |
| "step": 294, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.18667014063523668, | |
| "train/R_vocab": 0.27327178030303034 | |
| }, | |
| { | |
| "learning_rate": 4.642142940418973e-07, | |
| "loss": 0.0376, | |
| "step": 296, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.193981736932492, | |
| "train/R_vocab": 0.19051339285714286 | |
| }, | |
| { | |
| "learning_rate": 4.6365378113265505e-07, | |
| "loss": -0.0587, | |
| "step": 298, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.76875, | |
| "train/R_reason": 0.21676354110088458, | |
| "train/R_vocab": 0.2875 | |
| }, | |
| { | |
| "learning_rate": 4.630892564754956e-07, | |
| "loss": 0.0012, | |
| "step": 300, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.24459838346088703, | |
| "train/R_vocab": 0.33515625 | |
| }, | |
| { | |
| "eval/R_acc": 0.001388888888888889, | |
| "eval/R_penalty": -0.6499999999999999, | |
| "eval/R_reason": 0.2340619843391708, | |
| "eval/R_vocab": 0.2772872574955908, | |
| "step": 300 | |
| }, | |
| { | |
| "step": 300 | |
| }, | |
| { | |
| "learning_rate": 4.6252073067042127e-07, | |
| "loss": -0.0006, | |
| "step": 302, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.20949847516693765, | |
| "train/R_vocab": 0.24609375 | |
| }, | |
| { | |
| "learning_rate": 4.6194821439256373e-07, | |
| "loss": 0.0001, | |
| "step": 304, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.20496328530187652, | |
| "train/R_vocab": 0.20011160714285714 | |
| }, | |
| { | |
| "learning_rate": 4.6137171839198297e-07, | |
| "loss": 0.0792, | |
| "step": 306, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.1994046532974053, | |
| "train/R_vocab": 0.2868551587301587 | |
| }, | |
| { | |
| "learning_rate": 4.6079125349346576e-07, | |
| "loss": 0.0068, | |
| "step": 308, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.18547339942836427, | |
| "train/R_vocab": 0.2426711309523809 | |
| }, | |
| { | |
| "learning_rate": 4.602068305963224e-07, | |
| "loss": 0.0044, | |
| "step": 310, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.2112060437994439, | |
| "train/R_vocab": 0.2083829365079365 | |
| }, | |
| { | |
| "learning_rate": 4.59618460674182e-07, | |
| "loss": 0.0279, | |
| "step": 312, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.272254341125526, | |
| "train/R_vocab": 0.45736607142857144 | |
| }, | |
| { | |
| "learning_rate": 4.5902615477478636e-07, | |
| "loss": -0.0181, | |
| "step": 314, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.1449225433066495, | |
| "train/R_vocab": 0.16674107142857142 | |
| }, | |
| { | |
| "learning_rate": 4.5842992401978256e-07, | |
| "loss": 0.0225, | |
| "step": 316, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.828125, | |
| "train/R_reason": 0.2286200296039895, | |
| "train/R_vocab": 0.2551339285714286 | |
| }, | |
| { | |
| "learning_rate": 4.5782977960451414e-07, | |
| "loss": 0.0192, | |
| "step": 318, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.21916647791965457, | |
| "train/R_vocab": 0.2505580357142857 | |
| }, | |
| { | |
| "learning_rate": 4.57225732797811e-07, | |
| "loss": 0.0485, | |
| "step": 320, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.22755603236630983, | |
| "train/R_vocab": 0.22947668650793654 | |
| }, | |
| { | |
| "learning_rate": 4.566177949417777e-07, | |
| "loss": 0.0163, | |
| "step": 322, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.22636069312789864, | |
| "train/R_vocab": 0.21104910714285716 | |
| }, | |
| { | |
| "learning_rate": 4.560059774515804e-07, | |
| "loss": 0.0143, | |
| "step": 324, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.15769657140460813, | |
| "train/R_vocab": 0.22265625000000003 | |
| }, | |
| { | |
| "learning_rate": 4.5539029181523284e-07, | |
| "loss": -0.0796, | |
| "step": 326, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.1961929211100274, | |
| "train/R_vocab": 0.27142857142857146 | |
| }, | |
| { | |
| "learning_rate": 4.5477074959338015e-07, | |
| "loss": 0.0627, | |
| "step": 328, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.22035420658260274, | |
| "train/R_vocab": 0.33661954365079366 | |
| }, | |
| { | |
| "learning_rate": 4.5414736241908214e-07, | |
| "loss": -0.015, | |
| "step": 330, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.21972588070640325, | |
| "train/R_vocab": 0.30012400793650795 | |
| }, | |
| { | |
| "learning_rate": 4.535201419975948e-07, | |
| "loss": 0.0515, | |
| "step": 332, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.46875, | |
| "train/R_reason": 0.17121075988051718, | |
| "train/R_vocab": 0.234375 | |
| }, | |
| { | |
| "learning_rate": 4.5288910010615053e-07, | |
| "loss": 0.0479, | |
| "step": 334, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.21868874143295755, | |
| "train/R_vocab": 0.2630580357142857 | |
| }, | |
| { | |
| "learning_rate": 4.5225424859373684e-07, | |
| "loss": 0.0442, | |
| "step": 336, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.22348722064953433, | |
| "train/R_vocab": 0.3254464285714286 | |
| }, | |
| { | |
| "learning_rate": 4.51615599380874e-07, | |
| "loss": 0.0299, | |
| "step": 338, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.18059697532933602, | |
| "train/R_vocab": 0.21510416666666665 | |
| }, | |
| { | |
| "learning_rate": 4.5097316445939124e-07, | |
| "loss": 0.0359, | |
| "step": 340, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.1941301451022681, | |
| "train/R_vocab": 0.32220982142857146 | |
| }, | |
| { | |
| "learning_rate": 4.503269558922015e-07, | |
| "loss": -0.0107, | |
| "step": 342, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.21181567459611045, | |
| "train/R_vocab": 0.2482514880952381 | |
| }, | |
| { | |
| "learning_rate": 4.496769858130748e-07, | |
| "loss": 0.0179, | |
| "step": 344, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.23230602042254933, | |
| "train/R_vocab": 0.3271701388888889 | |
| }, | |
| { | |
| "learning_rate": 4.490232664264109e-07, | |
| "loss": 0.042, | |
| "step": 346, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.2775747137079635, | |
| "train/R_vocab": 0.3890625 | |
| }, | |
| { | |
| "learning_rate": 4.4836581000700944e-07, | |
| "loss": 0.0274, | |
| "step": 348, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.22799018822395078, | |
| "train/R_vocab": 0.27528521825396823 | |
| }, | |
| { | |
| "learning_rate": 4.477046288998401e-07, | |
| "loss": 0.0259, | |
| "step": 350, | |
| "train/R_acc": 0.13125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.20698908553226353, | |
| "train/R_vocab": 0.29140625 | |
| }, | |
| { | |
| "learning_rate": 4.470397355198102e-07, | |
| "loss": 0.0362, | |
| "step": 352, | |
| "train/R_acc": 0.0125, | |
| "train/R_penalty": -0.525, | |
| "train/R_reason": 0.210767134421379, | |
| "train/R_vocab": 0.3446180555555556 | |
| }, | |
| { | |
| "learning_rate": 4.463711423515323e-07, | |
| "loss": 0.0077, | |
| "step": 354, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.796875, | |
| "train/R_reason": 0.16874434359341295, | |
| "train/R_vocab": 0.20223214285714286 | |
| }, | |
| { | |
| "learning_rate": 4.456988619490889e-07, | |
| "loss": 0.0364, | |
| "step": 356, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.2266184174999622, | |
| "train/R_vocab": 0.32075892857142857 | |
| }, | |
| { | |
| "learning_rate": 4.450229069357977e-07, | |
| "loss": 0.0624, | |
| "step": 358, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.19035143640898555, | |
| "train/R_vocab": 0.1884548611111111 | |
| }, | |
| { | |
| "learning_rate": 4.4434329000397363e-07, | |
| "loss": -0.0017, | |
| "step": 360, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.18822223579328265, | |
| "train/R_vocab": 0.2318452380952381 | |
| }, | |
| { | |
| "learning_rate": 4.4366002391469126e-07, | |
| "loss": 0.0383, | |
| "step": 362, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.18399620476664255, | |
| "train/R_vocab": 0.17989831349206348 | |
| }, | |
| { | |
| "learning_rate": 4.4297312149754477e-07, | |
| "loss": 0.0709, | |
| "step": 364, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.2374170544763758, | |
| "train/R_vocab": 0.2957589285714286 | |
| }, | |
| { | |
| "learning_rate": 4.422825956504072e-07, | |
| "loss": 0.012, | |
| "step": 366, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.775, | |
| "train/R_reason": 0.18171410574759275, | |
| "train/R_vocab": 0.27645089285714286 | |
| }, | |
| { | |
| "learning_rate": 4.415884593391882e-07, | |
| "loss": -0.0291, | |
| "step": 368, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.728125, | |
| "train/R_reason": 0.20255323919085794, | |
| "train/R_vocab": 0.3200396825396825 | |
| }, | |
| { | |
| "learning_rate": 4.4089072559759065e-07, | |
| "loss": 0.0228, | |
| "step": 370, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.25184199643437555, | |
| "train/R_vocab": 0.31328125 | |
| }, | |
| { | |
| "learning_rate": 4.40189407526866e-07, | |
| "loss": -0.022, | |
| "step": 372, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.21868958770018052, | |
| "train/R_vocab": 0.21980406746031747 | |
| }, | |
| { | |
| "learning_rate": 4.3948451829556775e-07, | |
| "loss": 0.0437, | |
| "step": 374, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.603125, | |
| "train/R_reason": 0.23191700315429237, | |
| "train/R_vocab": 0.25416666666666665 | |
| }, | |
| { | |
| "learning_rate": 4.3877607113930516e-07, | |
| "loss": -0.0032, | |
| "step": 376, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.1725605699858892, | |
| "train/R_vocab": 0.23932291666666666 | |
| }, | |
| { | |
| "learning_rate": 4.380640793604938e-07, | |
| "loss": -0.0522, | |
| "step": 378, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.1570635238253007, | |
| "train/R_vocab": 0.209375 | |
| }, | |
| { | |
| "learning_rate": 4.373485563281062e-07, | |
| "loss": 0.0522, | |
| "step": 380, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.25503011482100246, | |
| "train/R_vocab": 0.2538690476190476 | |
| }, | |
| { | |
| "learning_rate": 4.3662951547742075e-07, | |
| "loss": 0.0447, | |
| "step": 382, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.1739243830311305, | |
| "train/R_vocab": 0.23214285714285715 | |
| }, | |
| { | |
| "learning_rate": 4.3590697030976965e-07, | |
| "loss": -0.006, | |
| "step": 384, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.15723816285274234, | |
| "train/R_vocab": 0.2829996392496392 | |
| }, | |
| { | |
| "learning_rate": 4.3518093439228474e-07, | |
| "loss": 0.0162, | |
| "step": 386, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.16239281076258544, | |
| "train/R_vocab": 0.1482142857142857 | |
| }, | |
| { | |
| "learning_rate": 4.3445142135764367e-07, | |
| "loss": -0.0038, | |
| "step": 388, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.19918031683009202, | |
| "train/R_vocab": 0.27239583333333334 | |
| }, | |
| { | |
| "learning_rate": 4.33718444903813e-07, | |
| "loss": 0.0341, | |
| "step": 390, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.20130303529176516, | |
| "train/R_vocab": 0.29593253968253963 | |
| }, | |
| { | |
| "learning_rate": 4.329820187937919e-07, | |
| "loss": 0.0876, | |
| "step": 392, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.23979456377125202, | |
| "train/R_vocab": 0.24832589285714285 | |
| }, | |
| { | |
| "learning_rate": 4.3224215685535287e-07, | |
| "loss": 0.0139, | |
| "step": 394, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.8125, | |
| "train/R_reason": 0.1709315188515677, | |
| "train/R_vocab": 0.18359375 | |
| }, | |
| { | |
| "learning_rate": 4.314988729807827e-07, | |
| "loss": -0.021, | |
| "step": 396, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.17646043188459717, | |
| "train/R_vocab": 0.17589285714285713 | |
| }, | |
| { | |
| "learning_rate": 4.3075218112662135e-07, | |
| "loss": 0.0513, | |
| "step": 398, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.1973938417696664, | |
| "train/R_vocab": 0.19837549603174603 | |
| }, | |
| { | |
| "learning_rate": 4.3000209531339996e-07, | |
| "loss": 0.0043, | |
| "step": 400, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.24184598292042236, | |
| "train/R_vocab": 0.33288690476190474 | |
| }, | |
| { | |
| "eval/R_acc": 0.0, | |
| "eval/R_penalty": -0.7916666666666666, | |
| "eval/R_reason": 0.27821714712822276, | |
| "eval/R_vocab": 0.32563381834215166, | |
| "step": 400 | |
| }, | |
| { | |
| "step": 400 | |
| }, | |
| { | |
| "learning_rate": 4.2943367562278896e-07, | |
| "loss": -0.0155, | |
| "step": 402, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.20335033205254746, | |
| "train/R_vocab": 0.24464285714285713 | |
| }, | |
| { | |
| "learning_rate": 4.286787117443108e-07, | |
| "loss": -0.036, | |
| "step": 404, | |
| "train/R_acc": 0.04062500000000001, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.20780436238210176, | |
| "train/R_vocab": 0.2703993055555556 | |
| }, | |
| { | |
| "learning_rate": 4.2792040207614e-07, | |
| "loss": -0.0044, | |
| "step": 406, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.734375, | |
| "train/R_reason": 0.2366017128148703, | |
| "train/R_vocab": 0.3632440476190476 | |
| }, | |
| { | |
| "learning_rate": 4.27158760817756e-07, | |
| "loss": 0.0245, | |
| "step": 408, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.19598878934576208, | |
| "train/R_vocab": 0.28031994047619047 | |
| }, | |
| { | |
| "learning_rate": 4.263938022310226e-07, | |
| "loss": -0.036, | |
| "step": 410, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.23507477335166782, | |
| "train/R_vocab": 0.27189980158730165 | |
| }, | |
| { | |
| "learning_rate": 4.2562554063992127e-07, | |
| "loss": 0.024, | |
| "step": 412, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.7125, | |
| "train/R_reason": 0.19770138255909644, | |
| "train/R_vocab": 0.24754464285714284 | |
| }, | |
| { | |
| "learning_rate": 4.248539904302829e-07, | |
| "loss": 0.0091, | |
| "step": 414, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2098416268444374, | |
| "train/R_vocab": 0.2533234126984127 | |
| }, | |
| { | |
| "learning_rate": 4.240791660495182e-07, | |
| "loss": 0.0125, | |
| "step": 416, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.603125, | |
| "train/R_reason": 0.164986564236729, | |
| "train/R_vocab": 0.1863219246031746 | |
| }, | |
| { | |
| "learning_rate": 4.2330108200634723e-07, | |
| "loss": 0.0353, | |
| "step": 418, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.16427460147948922, | |
| "train/R_vocab": 0.1800595238095238 | |
| }, | |
| { | |
| "learning_rate": 4.22519752870528e-07, | |
| "loss": -0.0002, | |
| "step": 420, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.21440123784001752, | |
| "train/R_vocab": 0.26026785714285716 | |
| }, | |
| { | |
| "learning_rate": 4.2173519327258325e-07, | |
| "loss": 0.0336, | |
| "step": 422, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2248050023022144, | |
| "train/R_vocab": 0.22937184343434344 | |
| }, | |
| { | |
| "learning_rate": 4.2094741790352673e-07, | |
| "loss": 0.0256, | |
| "step": 424, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.22803044267397005, | |
| "train/R_vocab": 0.28350694444444446 | |
| }, | |
| { | |
| "learning_rate": 4.2015644151458827e-07, | |
| "loss": -0.0319, | |
| "step": 426, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.20821167099100335, | |
| "train/R_vocab": 0.21744791666666666 | |
| }, | |
| { | |
| "learning_rate": 4.19362278916937e-07, | |
| "loss": 0.0344, | |
| "step": 428, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.1891472011902095, | |
| "train/R_vocab": 0.20641233766233769 | |
| }, | |
| { | |
| "learning_rate": 4.185649449814045e-07, | |
| "loss": 0.0712, | |
| "step": 430, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.2279409653415437, | |
| "train/R_vocab": 0.2996651785714286 | |
| }, | |
| { | |
| "learning_rate": 4.177644546382063e-07, | |
| "loss": 0.0645, | |
| "step": 432, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.21120852841137872, | |
| "train/R_vocab": 0.19676339285714284 | |
| }, | |
| { | |
| "learning_rate": 4.1696082287666217e-07, | |
| "loss": 0.0481, | |
| "step": 434, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.20179254286468842, | |
| "train/R_vocab": 0.3659598214285714 | |
| }, | |
| { | |
| "learning_rate": 4.1615406474491535e-07, | |
| "loss": -0.0453, | |
| "step": 436, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.78125, | |
| "train/R_reason": 0.21685467943636602, | |
| "train/R_vocab": 0.2571428571428571 | |
| }, | |
| { | |
| "learning_rate": 4.1534419534965105e-07, | |
| "loss": 0.0985, | |
| "step": 438, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.20741733073153354, | |
| "train/R_vocab": 0.27142857142857146 | |
| }, | |
| { | |
| "learning_rate": 4.145312298558133e-07, | |
| "loss": -0.0621, | |
| "step": 440, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.17062418658581344, | |
| "train/R_vocab": 0.2 | |
| }, | |
| { | |
| "learning_rate": 4.137151834863213e-07, | |
| "loss": -0.0448, | |
| "step": 442, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.23174115282287933, | |
| "train/R_vocab": 0.36713789682539677 | |
| }, | |
| { | |
| "learning_rate": 4.128960715217839e-07, | |
| "loss": 0.0187, | |
| "step": 444, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.15167707696946453, | |
| "train/R_vocab": 0.20052083333333334 | |
| }, | |
| { | |
| "learning_rate": 4.1207390930021394e-07, | |
| "loss": -0.0156, | |
| "step": 446, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2000569888049995, | |
| "train/R_vocab": 0.24027777777777778 | |
| }, | |
| { | |
| "learning_rate": 4.11248712216741e-07, | |
| "loss": 0.0897, | |
| "step": 448, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.21749146764720112, | |
| "train/R_vocab": 0.2636160714285714 | |
| }, | |
| { | |
| "learning_rate": 4.104204957233225e-07, | |
| "loss": 0.0032, | |
| "step": 450, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.19073873404612746, | |
| "train/R_vocab": 0.30033482142857143 | |
| }, | |
| { | |
| "learning_rate": 4.095892753284553e-07, | |
| "loss": -0.0985, | |
| "step": 452, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.25767563611959865, | |
| "train/R_vocab": 0.41929563492063493 | |
| }, | |
| { | |
| "learning_rate": 4.087550665968846e-07, | |
| "loss": 0.0021, | |
| "step": 454, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.1972933583803858, | |
| "train/R_vocab": 0.23854166666666668 | |
| }, | |
| { | |
| "learning_rate": 4.079178851493127e-07, | |
| "loss": 0.0433, | |
| "step": 456, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.2107089160600319, | |
| "train/R_vocab": 0.25 | |
| }, | |
| { | |
| "learning_rate": 4.070777466621067e-07, | |
| "loss": -0.0103, | |
| "step": 458, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.2429741695318347, | |
| "train/R_vocab": 0.38958333333333334 | |
| }, | |
| { | |
| "learning_rate": 4.062346668670046e-07, | |
| "loss": -0.0092, | |
| "step": 460, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.734375, | |
| "train/R_reason": 0.2328364572562428, | |
| "train/R_vocab": 0.2795758928571429 | |
| }, | |
| { | |
| "learning_rate": 4.0538866155082094e-07, | |
| "loss": -0.0198, | |
| "step": 462, | |
| "train/R_acc": 0.040625, | |
| "train/R_penalty": -0.4625, | |
| "train/R_reason": 0.2483761704003692, | |
| "train/R_vocab": 0.2885168650793651 | |
| }, | |
| { | |
| "learning_rate": 4.045397465551513e-07, | |
| "loss": -0.0249, | |
| "step": 464, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5875, | |
| "train/R_reason": 0.17307827497369785, | |
| "train/R_vocab": 0.2544642857142857 | |
| }, | |
| { | |
| "learning_rate": 4.036879377760752e-07, | |
| "loss": 0.0683, | |
| "step": 466, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.22330642900293768, | |
| "train/R_vocab": 0.37135416666666665 | |
| }, | |
| { | |
| "learning_rate": 4.02833251163859e-07, | |
| "loss": -0.0192, | |
| "step": 468, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65, | |
| "train/R_reason": 0.21129365536420686, | |
| "train/R_vocab": 0.2589285714285714 | |
| }, | |
| { | |
| "learning_rate": 4.0197570272265704e-07, | |
| "loss": 0.0497, | |
| "step": 470, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.2052435879372262, | |
| "train/R_vocab": 0.2556547619047619 | |
| }, | |
| { | |
| "learning_rate": 4.011153085102116e-07, | |
| "loss": 0.0309, | |
| "step": 472, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.21023509045617045, | |
| "train/R_vocab": 0.2963169642857143 | |
| }, | |
| { | |
| "learning_rate": 4.0025208463755274e-07, | |
| "loss": 0.0424, | |
| "step": 474, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.784375, | |
| "train/R_reason": 0.1959827825898999, | |
| "train/R_vocab": 0.35357142857142854 | |
| }, | |
| { | |
| "learning_rate": 3.9938604726869636e-07, | |
| "loss": -0.0506, | |
| "step": 476, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.19807364686139228, | |
| "train/R_vocab": 0.17491319444444442 | |
| }, | |
| { | |
| "learning_rate": 3.9851721262034157e-07, | |
| "loss": 0.0724, | |
| "step": 478, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.20997218794964279, | |
| "train/R_vocab": 0.27433035714285714 | |
| }, | |
| { | |
| "learning_rate": 3.9764559696156697e-07, | |
| "loss": 0.0338, | |
| "step": 480, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.21410868507940445, | |
| "train/R_vocab": 0.25691964285714286 | |
| }, | |
| { | |
| "learning_rate": 3.9677121661352607e-07, | |
| "loss": -0.0087, | |
| "step": 482, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.21506411602738199, | |
| "train/R_vocab": 0.3130952380952381 | |
| }, | |
| { | |
| "learning_rate": 3.958940879491418e-07, | |
| "loss": -0.0375, | |
| "step": 484, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.18895673088287218, | |
| "train/R_vocab": 0.21395089285714286 | |
| }, | |
| { | |
| "learning_rate": 3.9501422739279953e-07, | |
| "loss": -0.0485, | |
| "step": 486, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.20816202695918376, | |
| "train/R_vocab": 0.22013888888888888 | |
| }, | |
| { | |
| "learning_rate": 3.9413165142004e-07, | |
| "loss": -0.0479, | |
| "step": 488, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.1929871719837118, | |
| "train/R_vocab": 0.3703125 | |
| }, | |
| { | |
| "learning_rate": 3.932463765572505e-07, | |
| "loss": -0.0312, | |
| "step": 490, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.2513287746728735, | |
| "train/R_vocab": 0.27569444444444446 | |
| }, | |
| { | |
| "learning_rate": 3.923584193813555e-07, | |
| "loss": 0.03, | |
| "step": 492, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.23851168147176444, | |
| "train/R_vocab": 0.313219246031746 | |
| }, | |
| { | |
| "learning_rate": 3.914677965195062e-07, | |
| "loss": -0.0009, | |
| "step": 494, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.15729111795287057, | |
| "train/R_vocab": 0.1830357142857143 | |
| }, | |
| { | |
| "learning_rate": 3.9057452464876946e-07, | |
| "loss": 0.049, | |
| "step": 496, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.15483905838498802, | |
| "train/R_vocab": 0.18125000000000002 | |
| }, | |
| { | |
| "learning_rate": 3.89678620495815e-07, | |
| "loss": 0.0521, | |
| "step": 498, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.2600686346082261, | |
| "train/R_vocab": 0.253125 | |
| }, | |
| { | |
| "learning_rate": 3.887801008366025e-07, | |
| "loss": -0.005, | |
| "step": 500, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.790625, | |
| "train/R_reason": 0.22427641116364, | |
| "train/R_vocab": 0.3156125992063492 | |
| }, | |
| { | |
| "eval/R_acc": 0.0025, | |
| "eval/R_penalty": -0.7075, | |
| "eval/R_reason": 0.2610032668269429, | |
| "eval/R_vocab": 0.21906295093795097, | |
| "step": 500 | |
| }, | |
| { | |
| "step": 500 | |
| }, | |
| { | |
| "learning_rate": 3.8787898249606767e-07, | |
| "loss": 0.0252, | |
| "step": 502, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.2684211960918422, | |
| "train/R_vocab": 0.3091517857142857 | |
| }, | |
| { | |
| "learning_rate": 3.8697528234780674e-07, | |
| "loss": 0.019, | |
| "step": 504, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.23017420694962337, | |
| "train/R_vocab": 0.29288194444444443 | |
| }, | |
| { | |
| "learning_rate": 3.86069017313761e-07, | |
| "loss": 0.013, | |
| "step": 506, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.22614899753752782, | |
| "train/R_vocab": 0.3089905753968254 | |
| }, | |
| { | |
| "learning_rate": 3.851602043638994e-07, | |
| "loss": -0.0248, | |
| "step": 508, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.2017065165275432, | |
| "train/R_vocab": 0.26004464285714285 | |
| }, | |
| { | |
| "learning_rate": 3.8424886051590115e-07, | |
| "loss": 0.0296, | |
| "step": 510, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.7375, | |
| "train/R_reason": 0.17507956245908884, | |
| "train/R_vocab": 0.17238343253968252 | |
| }, | |
| { | |
| "learning_rate": 3.83335002834837e-07, | |
| "loss": -0.0622, | |
| "step": 512, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.22332916829962618, | |
| "train/R_vocab": 0.34873511904761906 | |
| }, | |
| { | |
| "learning_rate": 3.8241864843284964e-07, | |
| "loss": -0.0075, | |
| "step": 514, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.132802326682096, | |
| "train/R_vocab": 0.16104910714285714 | |
| }, | |
| { | |
| "learning_rate": 3.814998144688333e-07, | |
| "loss": 0.0227, | |
| "step": 516, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.1754689583879656, | |
| "train/R_vocab": 0.19609374999999998 | |
| }, | |
| { | |
| "learning_rate": 3.805785181481123e-07, | |
| "loss": -0.0118, | |
| "step": 518, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.540625, | |
| "train/R_reason": 0.23541827525810197, | |
| "train/R_vocab": 0.25967261904761907 | |
| }, | |
| { | |
| "learning_rate": 3.796547767221193e-07, | |
| "loss": -0.0924, | |
| "step": 520, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2376371992311745, | |
| "train/R_vocab": 0.3128968253968254 | |
| }, | |
| { | |
| "learning_rate": 3.787286074880718e-07, | |
| "loss": -0.0122, | |
| "step": 522, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.22978465999581174, | |
| "train/R_vocab": 0.2617931547619048 | |
| }, | |
| { | |
| "learning_rate": 3.778000277886483e-07, | |
| "loss": -0.0054, | |
| "step": 524, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.22061871520509158, | |
| "train/R_vocab": 0.30894209956709956 | |
| }, | |
| { | |
| "learning_rate": 3.768690550116639e-07, | |
| "loss": 0.0092, | |
| "step": 526, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.22469643175166784, | |
| "train/R_vocab": 0.34073660714285714 | |
| }, | |
| { | |
| "learning_rate": 3.7593570658974436e-07, | |
| "loss": 0.0264, | |
| "step": 528, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.26427900699025786, | |
| "train/R_vocab": 0.32444196428571426 | |
| }, | |
| { | |
| "learning_rate": 3.75e-07, | |
| "loss": 0.0264, | |
| "step": 530, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.16190914297300985, | |
| "train/R_vocab": 0.18397817460317462 | |
| }, | |
| { | |
| "learning_rate": 3.740619527636979e-07, | |
| "loss": -0.0014, | |
| "step": 532, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.540625, | |
| "train/R_reason": 0.2013910995332763, | |
| "train/R_vocab": 0.16294642857142855 | |
| }, | |
| { | |
| "learning_rate": 3.731215824459344e-07, | |
| "loss": 0.0653, | |
| "step": 534, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.1980956136240478, | |
| "train/R_vocab": 0.26245039682539684 | |
| }, | |
| { | |
| "learning_rate": 3.7217890665530587e-07, | |
| "loss": -0.0052, | |
| "step": 536, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.193118823374168, | |
| "train/R_vocab": 0.18813131313131315 | |
| }, | |
| { | |
| "learning_rate": 3.712339430435792e-07, | |
| "loss": 0.044, | |
| "step": 538, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.21065931054504308, | |
| "train/R_vocab": 0.2915054563492063 | |
| }, | |
| { | |
| "learning_rate": 3.7028670930536077e-07, | |
| "loss": 0.0443, | |
| "step": 540, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.21326431514536787, | |
| "train/R_vocab": 0.23958333333333334 | |
| }, | |
| { | |
| "learning_rate": 3.693372231777658e-07, | |
| "loss": 0.0062, | |
| "step": 542, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.86875, | |
| "train/R_reason": 0.2361538113041483, | |
| "train/R_vocab": 0.26145833333333335 | |
| }, | |
| { | |
| "learning_rate": 3.6838550244008573e-07, | |
| "loss": 0.0011, | |
| "step": 544, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.2560180166234486, | |
| "train/R_vocab": 0.3419642857142857 | |
| }, | |
| { | |
| "learning_rate": 3.6743156491345564e-07, | |
| "loss": -0.1304, | |
| "step": 546, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.22700866641314227, | |
| "train/R_vocab": 0.3409474206349206 | |
| }, | |
| { | |
| "learning_rate": 3.6647542846052003e-07, | |
| "loss": 0.0374, | |
| "step": 548, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.18355650415358743, | |
| "train/R_vocab": 0.23628472222222222 | |
| }, | |
| { | |
| "learning_rate": 3.65517110985099e-07, | |
| "loss": -0.0193, | |
| "step": 550, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.1497471083396625, | |
| "train/R_vocab": 0.1660342261904762 | |
| }, | |
| { | |
| "learning_rate": 3.645566304318526e-07, | |
| "loss": -0.0168, | |
| "step": 552, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.17760395766642206, | |
| "train/R_vocab": 0.20212053571428573 | |
| }, | |
| { | |
| "learning_rate": 3.6359400478594473e-07, | |
| "loss": 0.0519, | |
| "step": 554, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.2032957807067698, | |
| "train/R_vocab": 0.3547991071428571 | |
| }, | |
| { | |
| "learning_rate": 3.6262925207270666e-07, | |
| "loss": -0.0077, | |
| "step": 556, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.24608095001365107, | |
| "train/R_vocab": 0.2532366071428571 | |
| }, | |
| { | |
| "learning_rate": 3.616623903572994e-07, | |
| "loss": -0.0273, | |
| "step": 558, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.46875, | |
| "train/R_reason": 0.25668514206671234, | |
| "train/R_vocab": 0.32946428571428565 | |
| }, | |
| { | |
| "learning_rate": 3.6069343774437516e-07, | |
| "loss": 0.0356, | |
| "step": 560, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.853125, | |
| "train/R_reason": 0.22578641617093248, | |
| "train/R_vocab": 0.3117063492063492 | |
| }, | |
| { | |
| "learning_rate": 3.597224123777389e-07, | |
| "loss": 0.0074, | |
| "step": 562, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.1608838946497146, | |
| "train/R_vocab": 0.22940228174603175 | |
| }, | |
| { | |
| "learning_rate": 3.58749332440008e-07, | |
| "loss": 0.0218, | |
| "step": 564, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.2042184580919646, | |
| "train/R_vocab": 0.23277529761904764 | |
| }, | |
| { | |
| "learning_rate": 3.5777421615227207e-07, | |
| "loss": 0.0446, | |
| "step": 566, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.21492820594376136, | |
| "train/R_vocab": 0.2902901785714286 | |
| }, | |
| { | |
| "learning_rate": 3.567970817737518e-07, | |
| "loss": 0.0771, | |
| "step": 568, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.2497326839973281, | |
| "train/R_vocab": 0.30859375 | |
| }, | |
| { | |
| "learning_rate": 3.5581794760145696e-07, | |
| "loss": 0.0354, | |
| "step": 570, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.20181240026890324, | |
| "train/R_vocab": 0.22098214285714285 | |
| }, | |
| { | |
| "learning_rate": 3.548368319698437e-07, | |
| "loss": 0.0731, | |
| "step": 572, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.16856543760413237, | |
| "train/R_vocab": 0.1800595238095238 | |
| }, | |
| { | |
| "learning_rate": 3.5385375325047163e-07, | |
| "loss": 0.0291, | |
| "step": 574, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2099860707148752, | |
| "train/R_vocab": 0.2753348214285714 | |
| }, | |
| { | |
| "learning_rate": 3.528687298516591e-07, | |
| "loss": -0.0002, | |
| "step": 576, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.17694118112711849, | |
| "train/R_vocab": 0.12705853174603174 | |
| }, | |
| { | |
| "learning_rate": 3.5188178021813925e-07, | |
| "loss": 0.028, | |
| "step": 578, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.24354058691683123, | |
| "train/R_vocab": 0.39140624999999996 | |
| }, | |
| { | |
| "learning_rate": 3.5089292283071417e-07, | |
| "loss": 0.0398, | |
| "step": 580, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.22343115261429752, | |
| "train/R_vocab": 0.32678571428571423 | |
| }, | |
| { | |
| "learning_rate": 3.499021762059089e-07, | |
| "loss": 0.0565, | |
| "step": 582, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.7125, | |
| "train/R_reason": 0.19496747632032374, | |
| "train/R_vocab": 0.26284722222222223 | |
| }, | |
| { | |
| "learning_rate": 3.489095588956249e-07, | |
| "loss": 0.0092, | |
| "step": 584, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.2074570772291119, | |
| "train/R_vocab": 0.30972222222222223 | |
| }, | |
| { | |
| "learning_rate": 3.479150894867926e-07, | |
| "loss": -0.0015, | |
| "step": 586, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.20443537938998835, | |
| "train/R_vocab": 0.2761904761904762 | |
| }, | |
| { | |
| "learning_rate": 3.46918786601023e-07, | |
| "loss": 0.1059, | |
| "step": 588, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.20239388628972388, | |
| "train/R_vocab": 0.22630208333333335 | |
| }, | |
| { | |
| "learning_rate": 3.459206688942596e-07, | |
| "loss": 0.0608, | |
| "step": 590, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.24682579728602413, | |
| "train/R_vocab": 0.4008928571428571 | |
| }, | |
| { | |
| "learning_rate": 3.4492075505642847e-07, | |
| "loss": 0.055, | |
| "step": 592, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.19593385729165314, | |
| "train/R_vocab": 0.31875 | |
| }, | |
| { | |
| "learning_rate": 3.439190638110888e-07, | |
| "loss": 0.041, | |
| "step": 594, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.1711366333418505, | |
| "train/R_vocab": 0.26268601190476193 | |
| }, | |
| { | |
| "learning_rate": 3.4291561391508185e-07, | |
| "loss": 0.0807, | |
| "step": 596, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.240933143584435, | |
| "train/R_vocab": 0.4119791666666667 | |
| }, | |
| { | |
| "learning_rate": 3.4191042415818e-07, | |
| "loss": -0.0247, | |
| "step": 598, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.74375, | |
| "train/R_reason": 0.2539401489710603, | |
| "train/R_vocab": 0.2673363095238095 | |
| }, | |
| { | |
| "learning_rate": 3.4090351336273474e-07, | |
| "loss": -0.0238, | |
| "step": 600, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.19356852065501937, | |
| "train/R_vocab": 0.24319196428571427 | |
| }, | |
| { | |
| "eval/R_acc": 0.005, | |
| "eval/R_penalty": -0.625, | |
| "eval/R_reason": 0.24855003165355338, | |
| "eval/R_vocab": 0.22628517316017316, | |
| "step": 600 | |
| }, | |
| { | |
| "step": 600 | |
| }, | |
| { | |
| "learning_rate": 3.398949003833246e-07, | |
| "loss": 0.0122, | |
| "step": 602, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.22168416427879684, | |
| "train/R_vocab": 0.24791666666666667 | |
| }, | |
| { | |
| "learning_rate": 3.388846041064012e-07, | |
| "loss": 0.0797, | |
| "step": 604, | |
| "train/R_acc": 0.09375, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.1777229928874741, | |
| "train/R_vocab": 0.29464285714285715 | |
| }, | |
| { | |
| "learning_rate": 3.378726434499368e-07, | |
| "loss": 0.033, | |
| "step": 606, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.7125, | |
| "train/R_reason": 0.1956222769670857, | |
| "train/R_vocab": 0.3197916666666667 | |
| }, | |
| { | |
| "learning_rate": 3.368590373630692e-07, | |
| "loss": -0.0275, | |
| "step": 608, | |
| "train/R_acc": 0.06875, | |
| "train/R_penalty": -0.46875, | |
| "train/R_reason": 0.17503223220226105, | |
| "train/R_vocab": 0.2591517857142857 | |
| }, | |
| { | |
| "learning_rate": 3.3584380482574717e-07, | |
| "loss": -0.058, | |
| "step": 610, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.1902514010231657, | |
| "train/R_vocab": 0.23125 | |
| }, | |
| { | |
| "learning_rate": 3.348269648483749e-07, | |
| "loss": -0.054, | |
| "step": 612, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.17467640906171372, | |
| "train/R_vocab": 0.190625 | |
| }, | |
| { | |
| "learning_rate": 3.3380853647145656e-07, | |
| "loss": -0.0371, | |
| "step": 614, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.17121010785732182, | |
| "train/R_vocab": 0.2569444444444444 | |
| }, | |
| { | |
| "learning_rate": 3.327885387652391e-07, | |
| "loss": -0.0013, | |
| "step": 616, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.19415455674882087, | |
| "train/R_vocab": 0.23277529761904764 | |
| }, | |
| { | |
| "learning_rate": 3.317669908293554e-07, | |
| "loss": 0.0253, | |
| "step": 618, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.24193135939319937, | |
| "train/R_vocab": 0.25 | |
| }, | |
| { | |
| "learning_rate": 3.307439117924668e-07, | |
| "loss": 0.0588, | |
| "step": 620, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.20416377806011923, | |
| "train/R_vocab": 0.2056423611111111 | |
| }, | |
| { | |
| "learning_rate": 3.297193208119047e-07, | |
| "loss": -0.0064, | |
| "step": 622, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.211192554787049, | |
| "train/R_vocab": 0.30390625000000004 | |
| }, | |
| { | |
| "learning_rate": 3.2869323707331176e-07, | |
| "loss": 0.0335, | |
| "step": 624, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.23403307094336184, | |
| "train/R_vocab": 0.35301339285714284 | |
| }, | |
| { | |
| "learning_rate": 3.2766567979028324e-07, | |
| "loss": -0.0308, | |
| "step": 626, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.25211879339329, | |
| "train/R_vocab": 0.278125 | |
| }, | |
| { | |
| "learning_rate": 3.2663666820400625e-07, | |
| "loss": 0.0239, | |
| "step": 628, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.23123528000410434, | |
| "train/R_vocab": 0.328844246031746 | |
| }, | |
| { | |
| "learning_rate": 3.2560622158290025e-07, | |
| "loss": 0.0127, | |
| "step": 630, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.22103995585722944, | |
| "train/R_vocab": 0.26785714285714285 | |
| }, | |
| { | |
| "learning_rate": 3.2457435922225603e-07, | |
| "loss": 0.0445, | |
| "step": 632, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.185424618889319, | |
| "train/R_vocab": 0.28928571428571426 | |
| }, | |
| { | |
| "learning_rate": 3.235411004438741e-07, | |
| "loss": 0.1006, | |
| "step": 634, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.23390812732295468, | |
| "train/R_vocab": 0.29355158730158726 | |
| }, | |
| { | |
| "learning_rate": 3.2250646459570343e-07, | |
| "loss": 0.0611, | |
| "step": 636, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.17366692788791044, | |
| "train/R_vocab": 0.21026785714285715 | |
| }, | |
| { | |
| "learning_rate": 3.214704710514786e-07, | |
| "loss": 0.0463, | |
| "step": 638, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.23856693280209076, | |
| "train/R_vocab": 0.3723214285714286 | |
| }, | |
| { | |
| "learning_rate": 3.204331392103574e-07, | |
| "loss": -0.0048, | |
| "step": 640, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.2248622406893548, | |
| "train/R_vocab": 0.38828125 | |
| }, | |
| { | |
| "learning_rate": 3.193944884965576e-07, | |
| "loss": 0.0172, | |
| "step": 642, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.2000315806348117, | |
| "train/R_vocab": 0.32477678571428575 | |
| }, | |
| { | |
| "learning_rate": 3.183545383589927e-07, | |
| "loss": -0.0164, | |
| "step": 644, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.1901960813142597, | |
| "train/R_vocab": 0.2540178571428572 | |
| }, | |
| { | |
| "learning_rate": 3.173133082709086e-07, | |
| "loss": 0.0815, | |
| "step": 646, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.20267428182016378, | |
| "train/R_vocab": 0.2115079365079365 | |
| }, | |
| { | |
| "learning_rate": 3.1627081772951815e-07, | |
| "loss": 0.0279, | |
| "step": 648, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.18044257928711105, | |
| "train/R_vocab": 0.21806795634920634 | |
| }, | |
| { | |
| "learning_rate": 3.152270862556367e-07, | |
| "loss": 0.0707, | |
| "step": 650, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.26331484572063035, | |
| "train/R_vocab": 0.24206349206349206 | |
| }, | |
| { | |
| "learning_rate": 3.1418213339331576e-07, | |
| "loss": -0.0396, | |
| "step": 652, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.23778099952829604, | |
| "train/R_vocab": 0.284375 | |
| }, | |
| { | |
| "learning_rate": 3.1313597870947816e-07, | |
| "loss": 0.0883, | |
| "step": 654, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.22377350102735885, | |
| "train/R_vocab": 0.3117931547619047 | |
| }, | |
| { | |
| "learning_rate": 3.1208864179355074e-07, | |
| "loss": 0.0009, | |
| "step": 656, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.571875, | |
| "train/R_reason": 0.20656730207307256, | |
| "train/R_vocab": 0.2512276785714286 | |
| }, | |
| { | |
| "learning_rate": 3.1104014225709784e-07, | |
| "loss": 0.0828, | |
| "step": 658, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65, | |
| "train/R_reason": 0.2410817269424846, | |
| "train/R_vocab": 0.265625 | |
| }, | |
| { | |
| "learning_rate": 3.099904997334541e-07, | |
| "loss": -0.0317, | |
| "step": 660, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.5875, | |
| "train/R_reason": 0.25759275591542546, | |
| "train/R_vocab": 0.24017857142857144 | |
| }, | |
| { | |
| "learning_rate": 3.0893973387735683e-07, | |
| "loss": -0.0235, | |
| "step": 662, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.16711049206280193, | |
| "train/R_vocab": 0.21428571428571427 | |
| }, | |
| { | |
| "learning_rate": 3.078878643645778e-07, | |
| "loss": 0.0242, | |
| "step": 664, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.22549504519978125, | |
| "train/R_vocab": 0.33912450396825394 | |
| }, | |
| { | |
| "learning_rate": 3.068349108915553e-07, | |
| "loss": 0.0147, | |
| "step": 666, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2824812700031851, | |
| "train/R_vocab": 0.24285714285714285 | |
| }, | |
| { | |
| "learning_rate": 3.0578089317502436e-07, | |
| "loss": -0.0134, | |
| "step": 668, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.525, | |
| "train/R_reason": 0.1605276972684076, | |
| "train/R_vocab": 0.22779017857142858 | |
| }, | |
| { | |
| "learning_rate": 3.0472583095164873e-07, | |
| "loss": -0.0496, | |
| "step": 670, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.19456119250844006, | |
| "train/R_vocab": 0.23214285714285715 | |
| }, | |
| { | |
| "learning_rate": 3.036697439776504e-07, | |
| "loss": 0.0554, | |
| "step": 672, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.2515289920392415, | |
| "train/R_vocab": 0.2724330357142857 | |
| }, | |
| { | |
| "learning_rate": 3.0261265202844016e-07, | |
| "loss": 0.02, | |
| "step": 674, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.18580953264755184, | |
| "train/R_vocab": 0.25388144841269844 | |
| }, | |
| { | |
| "learning_rate": 3.01554574898247e-07, | |
| "loss": 0.0155, | |
| "step": 676, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.22948436053810994, | |
| "train/R_vocab": 0.22013888888888888 | |
| }, | |
| { | |
| "learning_rate": 3.004955323997478e-07, | |
| "loss": -0.0179, | |
| "step": 678, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.21153443586764287, | |
| "train/R_vocab": 0.23524305555555558 | |
| }, | |
| { | |
| "learning_rate": 2.994355443636958e-07, | |
| "loss": -0.0362, | |
| "step": 680, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.1840845890109734, | |
| "train/R_vocab": 0.31015625 | |
| }, | |
| { | |
| "learning_rate": 2.983746306385499e-07, | |
| "loss": 0.0297, | |
| "step": 682, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.453125, | |
| "train/R_reason": 0.2084560166468069, | |
| "train/R_vocab": 0.31228918650793647 | |
| }, | |
| { | |
| "learning_rate": 2.9731281109010253e-07, | |
| "loss": 0.0534, | |
| "step": 684, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.16422497838360026, | |
| "train/R_vocab": 0.23582589285714284 | |
| }, | |
| { | |
| "learning_rate": 2.9625010560110787e-07, | |
| "loss": 0.0183, | |
| "step": 686, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.24390149995516267, | |
| "train/R_vocab": 0.29910714285714285 | |
| }, | |
| { | |
| "learning_rate": 2.951865340709095e-07, | |
| "loss": 0.0683, | |
| "step": 688, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.64375, | |
| "train/R_reason": 0.23521821731229603, | |
| "train/R_vocab": 0.3359747023809524 | |
| }, | |
| { | |
| "learning_rate": 2.9412211641506773e-07, | |
| "loss": -0.001, | |
| "step": 690, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.18437836963522442, | |
| "train/R_vocab": 0.209375 | |
| }, | |
| { | |
| "learning_rate": 2.930568725649867e-07, | |
| "loss": 0.0194, | |
| "step": 692, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.65, | |
| "train/R_reason": 0.21411232248047601, | |
| "train/R_vocab": 0.2759548611111111 | |
| }, | |
| { | |
| "learning_rate": 2.919908224675412e-07, | |
| "loss": 0.0579, | |
| "step": 694, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.1540599514957384, | |
| "train/R_vocab": 0.18344494047619048 | |
| }, | |
| { | |
| "learning_rate": 2.9092398608470326e-07, | |
| "loss": -0.0106, | |
| "step": 696, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2501350490350204, | |
| "train/R_vocab": 0.3291294642857143 | |
| }, | |
| { | |
| "learning_rate": 2.898563833931681e-07, | |
| "loss": 0.0156, | |
| "step": 698, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.20865047120336916, | |
| "train/R_vocab": 0.2130332341269841 | |
| }, | |
| { | |
| "learning_rate": 2.8878803438398015e-07, | |
| "loss": 0.0192, | |
| "step": 700, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.19110667770899964, | |
| "train/R_vocab": 0.24427083333333333 | |
| }, | |
| { | |
| "eval/R_acc": 0.0025, | |
| "eval/R_penalty": -0.5875, | |
| "eval/R_reason": 0.25186260885534084, | |
| "eval/R_vocab": 0.2177263708513708, | |
| "step": 700 | |
| }, | |
| { | |
| "step": 700 | |
| }, | |
| { | |
| "learning_rate": 2.8771895906215895e-07, | |
| "loss": 0.05, | |
| "step": 702, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.2448799886759996, | |
| "train/R_vocab": 0.3176339285714286 | |
| }, | |
| { | |
| "learning_rate": 2.8664917744632423e-07, | |
| "loss": 0.0057, | |
| "step": 704, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.22652983918213118, | |
| "train/R_vocab": 0.41149553571428577 | |
| }, | |
| { | |
| "learning_rate": 2.8557870956832133e-07, | |
| "loss": 0.0356, | |
| "step": 706, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.22872856053626722, | |
| "train/R_vocab": 0.2622767857142857 | |
| }, | |
| { | |
| "learning_rate": 2.8450757547284576e-07, | |
| "loss": 0.0236, | |
| "step": 708, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.19584017346681087, | |
| "train/R_vocab": 0.20625000000000002 | |
| }, | |
| { | |
| "learning_rate": 2.8343579521706823e-07, | |
| "loss": -0.0244, | |
| "step": 710, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.17909964580588045, | |
| "train/R_vocab": 0.23783482142857143 | |
| }, | |
| { | |
| "learning_rate": 2.8236338887025886e-07, | |
| "loss": 0.0155, | |
| "step": 712, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.26302313003150035, | |
| "train/R_vocab": 0.3813988095238095 | |
| }, | |
| { | |
| "learning_rate": 2.812903765134115e-07, | |
| "loss": -0.0178, | |
| "step": 714, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2631388468076138, | |
| "train/R_vocab": 0.4032242063492063 | |
| }, | |
| { | |
| "learning_rate": 2.8021677823886754e-07, | |
| "loss": -0.0467, | |
| "step": 716, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.19911790925777817, | |
| "train/R_vocab": 0.23608630952380952 | |
| }, | |
| { | |
| "learning_rate": 2.7914261414993976e-07, | |
| "loss": 0.0054, | |
| "step": 718, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.603125, | |
| "train/R_reason": 0.2387146239629917, | |
| "train/R_vocab": 0.2571428571428571 | |
| }, | |
| { | |
| "learning_rate": 2.780679043605361e-07, | |
| "loss": -0.0082, | |
| "step": 720, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.571875, | |
| "train/R_reason": 0.2269094768328706, | |
| "train/R_vocab": 0.31436011904761907 | |
| }, | |
| { | |
| "learning_rate": 2.7699266899478274e-07, | |
| "loss": 0.0178, | |
| "step": 722, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.24130363361070928, | |
| "train/R_vocab": 0.22456597222222222 | |
| }, | |
| { | |
| "learning_rate": 2.759169281866472e-07, | |
| "loss": 0.0661, | |
| "step": 724, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.64375, | |
| "train/R_reason": 0.21308869490175653, | |
| "train/R_vocab": 0.2772321428571428 | |
| }, | |
| { | |
| "learning_rate": 2.748407020795617e-07, | |
| "loss": 0.1056, | |
| "step": 726, | |
| "train/R_acc": 0.096875, | |
| "train/R_penalty": -0.421875, | |
| "train/R_reason": 0.23621705469287618, | |
| "train/R_vocab": 0.2981770833333333 | |
| }, | |
| { | |
| "learning_rate": 2.737640108260456e-07, | |
| "loss": 0.0812, | |
| "step": 728, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.2310603125451809, | |
| "train/R_vocab": 0.30590277777777775 | |
| }, | |
| { | |
| "learning_rate": 2.726868745873286e-07, | |
| "loss": -0.0637, | |
| "step": 730, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.19990588785781663, | |
| "train/R_vocab": 0.265625 | |
| }, | |
| { | |
| "learning_rate": 2.716093135329722e-07, | |
| "loss": 0.0348, | |
| "step": 732, | |
| "train/R_acc": 0.0375, | |
| "train/R_penalty": -0.421875, | |
| "train/R_reason": 0.2095006410516756, | |
| "train/R_vocab": 0.32751736111111107 | |
| }, | |
| { | |
| "learning_rate": 2.7053134784049316e-07, | |
| "loss": 0.0875, | |
| "step": 734, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.17118221219882784, | |
| "train/R_vocab": 0.17343750000000002 | |
| }, | |
| { | |
| "learning_rate": 2.694529976949849e-07, | |
| "loss": -0.0448, | |
| "step": 736, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.20588894905097077, | |
| "train/R_vocab": 0.34609375 | |
| }, | |
| { | |
| "learning_rate": 2.6837428328874007e-07, | |
| "loss": -0.0244, | |
| "step": 738, | |
| "train/R_acc": 0.06875, | |
| "train/R_penalty": -0.6125, | |
| "train/R_reason": 0.26183079845209467, | |
| "train/R_vocab": 0.2656994047619048 | |
| }, | |
| { | |
| "learning_rate": 2.6729522482087165e-07, | |
| "loss": 0.0854, | |
| "step": 740, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.20793241180273003, | |
| "train/R_vocab": 0.21996527777777775 | |
| }, | |
| { | |
| "learning_rate": 2.662158424969357e-07, | |
| "loss": 0.0149, | |
| "step": 742, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.46875, | |
| "train/R_reason": 0.18410838343666375, | |
| "train/R_vocab": 0.1974330357142857 | |
| }, | |
| { | |
| "learning_rate": 2.6513615652855246e-07, | |
| "loss": -0.0144, | |
| "step": 744, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.24293661434589603, | |
| "train/R_vocab": 0.3267857142857143 | |
| }, | |
| { | |
| "learning_rate": 2.6405618713302783e-07, | |
| "loss": -0.0678, | |
| "step": 746, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.135718109682479, | |
| "train/R_vocab": 0.22979910714285715 | |
| }, | |
| { | |
| "learning_rate": 2.629759545329749e-07, | |
| "loss": 0.1246, | |
| "step": 748, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.20181279701198124, | |
| "train/R_vocab": 0.3018849206349207 | |
| }, | |
| { | |
| "learning_rate": 2.618954789559356e-07, | |
| "loss": 0.0933, | |
| "step": 750, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.20252563848526106, | |
| "train/R_vocab": 0.2291294642857143 | |
| }, | |
| { | |
| "learning_rate": 2.6081478063400124e-07, | |
| "loss": -0.0234, | |
| "step": 752, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.19106041589978606, | |
| "train/R_vocab": 0.20889136904761904 | |
| }, | |
| { | |
| "learning_rate": 2.597338798034344e-07, | |
| "loss": -0.0631, | |
| "step": 754, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.1867345308525643, | |
| "train/R_vocab": 0.22708333333333333 | |
| }, | |
| { | |
| "learning_rate": 2.5865279670428956e-07, | |
| "loss": 0.0483, | |
| "step": 756, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.17135153722070606, | |
| "train/R_vocab": 0.18147321428571428 | |
| }, | |
| { | |
| "learning_rate": 2.5757155158003414e-07, | |
| "loss": 0.0295, | |
| "step": 758, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.1663155804467654, | |
| "train/R_vocab": 0.22377232142857145 | |
| }, | |
| { | |
| "learning_rate": 2.5649016467716954e-07, | |
| "loss": 0.0283, | |
| "step": 760, | |
| "train/R_acc": 0.04062500000000001, | |
| "train/R_penalty": -0.375, | |
| "train/R_reason": 0.19205683929152886, | |
| "train/R_vocab": 0.2728422619047619 | |
| }, | |
| { | |
| "learning_rate": 2.5540865624485216e-07, | |
| "loss": -0.0491, | |
| "step": 762, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.2550380457423676, | |
| "train/R_vocab": 0.2734375 | |
| }, | |
| { | |
| "learning_rate": 2.5432704653451374e-07, | |
| "loss": 0.0441, | |
| "step": 764, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.20927672505440928, | |
| "train/R_vocab": 0.30055803571428574 | |
| }, | |
| { | |
| "learning_rate": 2.532453557994827e-07, | |
| "loss": 0.048, | |
| "step": 766, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.20293042206204454, | |
| "train/R_vocab": 0.2447668650793651 | |
| }, | |
| { | |
| "learning_rate": 2.521636042946046e-07, | |
| "loss": 0.0256, | |
| "step": 768, | |
| "train/R_acc": 0.04062500000000001, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.2228340738544955, | |
| "train/R_vocab": 0.21646825396825398 | |
| }, | |
| { | |
| "learning_rate": 2.5108181227586307e-07, | |
| "loss": -0.0186, | |
| "step": 770, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.15339300119866717, | |
| "train/R_vocab": 0.24015376984126985 | |
| }, | |
| { | |
| "learning_rate": 2.5e-07, | |
| "loss": 0.0082, | |
| "step": 772, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.25634859172881913, | |
| "train/R_vocab": 0.2591517857142857 | |
| }, | |
| { | |
| "learning_rate": 2.48918187724137e-07, | |
| "loss": -0.0372, | |
| "step": 774, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.15909137824419298, | |
| "train/R_vocab": 0.2725953733766234 | |
| }, | |
| { | |
| "learning_rate": 2.4783639570539535e-07, | |
| "loss": -0.1172, | |
| "step": 776, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.46875, | |
| "train/R_reason": 0.21610677086076593, | |
| "train/R_vocab": 0.21164434523809522 | |
| }, | |
| { | |
| "learning_rate": 2.467546442005173e-07, | |
| "loss": 0.0088, | |
| "step": 778, | |
| "train/R_acc": 0.06875, | |
| "train/R_penalty": -0.390625, | |
| "train/R_reason": 0.19689824765738337, | |
| "train/R_vocab": 0.1890625 | |
| }, | |
| { | |
| "learning_rate": 2.4567295346548634e-07, | |
| "loss": 0.1308, | |
| "step": 780, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.16858555636857672, | |
| "train/R_vocab": 0.29520089285714285 | |
| }, | |
| { | |
| "learning_rate": 2.4459134375514787e-07, | |
| "loss": -0.0356, | |
| "step": 782, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5875, | |
| "train/R_reason": 0.20190652121448122, | |
| "train/R_vocab": 0.2455357142857143 | |
| }, | |
| { | |
| "learning_rate": 2.4350983532283043e-07, | |
| "loss": -0.0182, | |
| "step": 784, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.1856742513120948, | |
| "train/R_vocab": 0.275 | |
| }, | |
| { | |
| "learning_rate": 2.4242844841996583e-07, | |
| "loss": 0.0167, | |
| "step": 786, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.20132693663438478, | |
| "train/R_vocab": 0.18971974206349207 | |
| }, | |
| { | |
| "learning_rate": 2.4134720329571047e-07, | |
| "loss": 0.0408, | |
| "step": 788, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.571875, | |
| "train/R_reason": 0.22452081729968582, | |
| "train/R_vocab": 0.21640625 | |
| }, | |
| { | |
| "learning_rate": 2.4026612019656556e-07, | |
| "loss": 0.0107, | |
| "step": 790, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.18352974837528543, | |
| "train/R_vocab": 0.25814732142857144 | |
| }, | |
| { | |
| "learning_rate": 2.3918521936599874e-07, | |
| "loss": 0.036, | |
| "step": 792, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.27166955319028896, | |
| "train/R_vocab": 0.35963541666666665 | |
| }, | |
| { | |
| "learning_rate": 2.381045210440644e-07, | |
| "loss": -0.0346, | |
| "step": 794, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.22416891439419612, | |
| "train/R_vocab": 0.32734375000000004 | |
| }, | |
| { | |
| "learning_rate": 2.3702404546702509e-07, | |
| "loss": 0.0911, | |
| "step": 796, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.24303349052516854, | |
| "train/R_vocab": 0.27745535714285713 | |
| }, | |
| { | |
| "learning_rate": 2.3594381286697215e-07, | |
| "loss": 0.0121, | |
| "step": 798, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.23618598427535661, | |
| "train/R_vocab": 0.28122519841269844 | |
| }, | |
| { | |
| "learning_rate": 2.3486384347144752e-07, | |
| "loss": 0.0064, | |
| "step": 800, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.25934359162168097, | |
| "train/R_vocab": 0.4203125 | |
| }, | |
| { | |
| "eval/R_acc": 0.0025, | |
| "eval/R_penalty": -0.5875, | |
| "eval/R_reason": 0.29083729720667373, | |
| "eval/R_vocab": 0.22747113997114, | |
| "step": 800 | |
| }, | |
| { | |
| "step": 800 | |
| }, | |
| { | |
| "learning_rate": 2.337841575030642e-07, | |
| "loss": -0.0305, | |
| "step": 802, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.20417958820529003, | |
| "train/R_vocab": 0.2702008928571429 | |
| }, | |
| { | |
| "learning_rate": 2.3270477517912835e-07, | |
| "loss": 0.0361, | |
| "step": 804, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.621875, | |
| "train/R_reason": 0.2872431102010518, | |
| "train/R_vocab": 0.37351190476190477 | |
| }, | |
| { | |
| "learning_rate": 2.3162571671126e-07, | |
| "loss": 0.0184, | |
| "step": 806, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2281768901636334, | |
| "train/R_vocab": 0.30616319444444445 | |
| }, | |
| { | |
| "learning_rate": 2.3054700230501502e-07, | |
| "loss": 0.0128, | |
| "step": 808, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.20789617508651104, | |
| "train/R_vocab": 0.3257440476190476 | |
| }, | |
| { | |
| "learning_rate": 2.2946865215950685e-07, | |
| "loss": 0.0158, | |
| "step": 810, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.22696589858153038, | |
| "train/R_vocab": 0.21540178571428575 | |
| }, | |
| { | |
| "learning_rate": 2.2839068646702786e-07, | |
| "loss": 1.5239, | |
| "step": 812, | |
| "train/R_acc": 0.0375, | |
| "train/R_penalty": -0.734375, | |
| "train/R_reason": 0.1726926027420783, | |
| "train/R_vocab": 0.25167410714285715 | |
| }, | |
| { | |
| "learning_rate": 2.2731312541267143e-07, | |
| "loss": 0.0348, | |
| "step": 814, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.18792369164667982, | |
| "train/R_vocab": 0.2892485119047619 | |
| }, | |
| { | |
| "learning_rate": 2.2623598917395436e-07, | |
| "loss": 0.0032, | |
| "step": 816, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.16929829805580582, | |
| "train/R_vocab": 0.16987847222222222 | |
| }, | |
| { | |
| "learning_rate": 2.251592979204383e-07, | |
| "loss": 0.0127, | |
| "step": 818, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.390625, | |
| "train/R_reason": 0.28610113589945263, | |
| "train/R_vocab": 0.35 | |
| }, | |
| { | |
| "learning_rate": 2.2408307181335285e-07, | |
| "loss": 0.0528, | |
| "step": 820, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.509375, | |
| "train/R_reason": 0.18837133598647154, | |
| "train/R_vocab": 0.21272321428571428 | |
| }, | |
| { | |
| "learning_rate": 2.2300733100521732e-07, | |
| "loss": 0.0454, | |
| "step": 822, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.1814823048344199, | |
| "train/R_vocab": 0.13814484126984128 | |
| }, | |
| { | |
| "learning_rate": 2.2193209563946382e-07, | |
| "loss": 0.0355, | |
| "step": 824, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.18663805046129311, | |
| "train/R_vocab": 0.1985863095238095 | |
| }, | |
| { | |
| "learning_rate": 2.2085738585006021e-07, | |
| "loss": 0.0026, | |
| "step": 826, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.49375, | |
| "train/R_reason": 0.18273024857160083, | |
| "train/R_vocab": 0.21136363636363636 | |
| }, | |
| { | |
| "learning_rate": 2.1978322176113251e-07, | |
| "loss": -0.0318, | |
| "step": 828, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.19883211536778234, | |
| "train/R_vocab": 0.27878224206349206 | |
| }, | |
| { | |
| "learning_rate": 2.1870962348658852e-07, | |
| "loss": -0.0335, | |
| "step": 830, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.1985312923562238, | |
| "train/R_vocab": 0.2785714285714286 | |
| }, | |
| { | |
| "learning_rate": 2.1763661112974115e-07, | |
| "loss": -0.0012, | |
| "step": 832, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2787590605964265, | |
| "train/R_vocab": 0.35 | |
| }, | |
| { | |
| "learning_rate": 2.165642047829318e-07, | |
| "loss": -0.0375, | |
| "step": 834, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.15530706954823303, | |
| "train/R_vocab": 0.22005208333333334 | |
| }, | |
| { | |
| "learning_rate": 2.1549242452715427e-07, | |
| "loss": -0.0387, | |
| "step": 836, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.24527110960119072, | |
| "train/R_vocab": 0.38571428571428573 | |
| }, | |
| { | |
| "learning_rate": 2.1442129043167873e-07, | |
| "loss": 0.0321, | |
| "step": 838, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.21774443607948502, | |
| "train/R_vocab": 0.26517857142857143 | |
| }, | |
| { | |
| "learning_rate": 2.1335082255367572e-07, | |
| "loss": 0.0288, | |
| "step": 840, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.22299875389864712, | |
| "train/R_vocab": 0.27976190476190477 | |
| }, | |
| { | |
| "learning_rate": 2.1228104093784108e-07, | |
| "loss": 0.0203, | |
| "step": 842, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.25459030399590693, | |
| "train/R_vocab": 0.36163194444444446 | |
| }, | |
| { | |
| "learning_rate": 2.112119656160199e-07, | |
| "loss": 0.0337, | |
| "step": 844, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.2547007960169786, | |
| "train/R_vocab": 0.34011656746031743 | |
| }, | |
| { | |
| "learning_rate": 2.1014361660683194e-07, | |
| "loss": -0.0236, | |
| "step": 846, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.23028272641397346, | |
| "train/R_vocab": 0.25 | |
| }, | |
| { | |
| "learning_rate": 2.0907601391529677e-07, | |
| "loss": -0.0204, | |
| "step": 848, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.18682793317030633, | |
| "train/R_vocab": 0.21324066558441557 | |
| }, | |
| { | |
| "learning_rate": 2.0800917753245875e-07, | |
| "loss": 0.1057, | |
| "step": 850, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.2020052493242343, | |
| "train/R_vocab": 0.27734375 | |
| }, | |
| { | |
| "learning_rate": 2.0694312743501334e-07, | |
| "loss": -0.0585, | |
| "step": 852, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.23457174043904147, | |
| "train/R_vocab": 0.3036830357142857 | |
| }, | |
| { | |
| "learning_rate": 2.0587788358493235e-07, | |
| "loss": -0.0318, | |
| "step": 854, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.34375, | |
| "train/R_reason": 0.2538105699194751, | |
| "train/R_vocab": 0.40078125 | |
| }, | |
| { | |
| "learning_rate": 2.048134659290905e-07, | |
| "loss": -0.0723, | |
| "step": 856, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.17298681086107648, | |
| "train/R_vocab": 0.15267857142857144 | |
| }, | |
| { | |
| "learning_rate": 2.037498943988921e-07, | |
| "loss": -0.0058, | |
| "step": 858, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.28539022868157543, | |
| "train/R_vocab": 0.3465593434343434 | |
| }, | |
| { | |
| "learning_rate": 2.0268718890989752e-07, | |
| "loss": -0.0416, | |
| "step": 860, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.509375, | |
| "train/R_reason": 0.20216319087856421, | |
| "train/R_vocab": 0.3190848214285714 | |
| }, | |
| { | |
| "learning_rate": 2.0162536936145008e-07, | |
| "loss": -0.0399, | |
| "step": 862, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.2420341418631836, | |
| "train/R_vocab": 0.27738095238095234 | |
| }, | |
| { | |
| "learning_rate": 2.0056445563630423e-07, | |
| "loss": 0.0372, | |
| "step": 864, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.21099684301910737, | |
| "train/R_vocab": 0.26659226190476193 | |
| }, | |
| { | |
| "learning_rate": 1.9950446760025219e-07, | |
| "loss": 0.0326, | |
| "step": 866, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.19938542697531797, | |
| "train/R_vocab": 0.25535714285714284 | |
| }, | |
| { | |
| "learning_rate": 1.98445425101753e-07, | |
| "loss": 0.0211, | |
| "step": 868, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.22302710668155856, | |
| "train/R_vocab": 0.2036830357142857 | |
| }, | |
| { | |
| "learning_rate": 1.9738734797155987e-07, | |
| "loss": 0.1391, | |
| "step": 870, | |
| "train/R_acc": 0.0125, | |
| "train/R_penalty": -0.703125, | |
| "train/R_reason": 0.20507300880803225, | |
| "train/R_vocab": 0.29453125 | |
| }, | |
| { | |
| "learning_rate": 1.963302560223496e-07, | |
| "loss": 0.0314, | |
| "step": 872, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.1946319904099672, | |
| "train/R_vocab": 0.32533482142857145 | |
| }, | |
| { | |
| "learning_rate": 1.9527416904835132e-07, | |
| "loss": 0.0562, | |
| "step": 874, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.2123123501156333, | |
| "train/R_vocab": 0.24363839285714287 | |
| }, | |
| { | |
| "learning_rate": 1.9421910682497572e-07, | |
| "loss": 0.0335, | |
| "step": 876, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.24626829087757965, | |
| "train/R_vocab": 0.28229166666666666 | |
| }, | |
| { | |
| "learning_rate": 1.9316508910844476e-07, | |
| "loss": 0.0255, | |
| "step": 878, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.18789237365926498, | |
| "train/R_vocab": 0.278125 | |
| }, | |
| { | |
| "learning_rate": 1.9211213563542218e-07, | |
| "loss": -0.0622, | |
| "step": 880, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.19592139018937305, | |
| "train/R_vocab": 0.26670386904761906 | |
| }, | |
| { | |
| "learning_rate": 1.9106026612264315e-07, | |
| "loss": 0.0249, | |
| "step": 882, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.17202561289857762, | |
| "train/R_vocab": 0.24394841269841272 | |
| }, | |
| { | |
| "learning_rate": 1.900095002665459e-07, | |
| "loss": 0.0627, | |
| "step": 884, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.20212251182315438, | |
| "train/R_vocab": 0.234375 | |
| }, | |
| { | |
| "learning_rate": 1.889598577429022e-07, | |
| "loss": 0.0251, | |
| "step": 886, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.21931575978869752, | |
| "train/R_vocab": 0.20535714285714285 | |
| }, | |
| { | |
| "learning_rate": 1.8791135820644922e-07, | |
| "loss": 0.0228, | |
| "step": 888, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.23473900860844063, | |
| "train/R_vocab": 0.2440104166666667 | |
| }, | |
| { | |
| "learning_rate": 1.8686402129052181e-07, | |
| "loss": 0.0602, | |
| "step": 890, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.2188358452056008, | |
| "train/R_vocab": 0.24036458333333333 | |
| }, | |
| { | |
| "learning_rate": 1.858178666066843e-07, | |
| "loss": 0.0042, | |
| "step": 892, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.1951297685148197, | |
| "train/R_vocab": 0.23270089285714285 | |
| }, | |
| { | |
| "learning_rate": 1.8477291374436338e-07, | |
| "loss": -0.0364, | |
| "step": 894, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.11877722404091096, | |
| "train/R_vocab": 0.12317708333333333 | |
| }, | |
| { | |
| "learning_rate": 1.8372918227048183e-07, | |
| "loss": 0.3093, | |
| "step": 896, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.1855972158801493, | |
| "train/R_vocab": 0.25017361111111114 | |
| }, | |
| { | |
| "learning_rate": 1.8268669172909136e-07, | |
| "loss": 0.0017, | |
| "step": 898, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.18804079477258429, | |
| "train/R_vocab": 0.16622023809523812 | |
| }, | |
| { | |
| "learning_rate": 1.8164546164100728e-07, | |
| "loss": -0.0186, | |
| "step": 900, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.24246348014867208, | |
| "train/R_vocab": 0.29110863095238093 | |
| }, | |
| { | |
| "eval/R_acc": 0.030000000000000006, | |
| "eval/R_penalty": -0.6, | |
| "eval/R_reason": 0.29157417644976924, | |
| "eval/R_vocab": 0.37875, | |
| "step": 900 | |
| }, | |
| { | |
| "step": 900 | |
| }, | |
| { | |
| "learning_rate": 1.806055115034425e-07, | |
| "loss": -0.0579, | |
| "step": 902, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.23201039495721226, | |
| "train/R_vocab": 0.2525669642857143 | |
| }, | |
| { | |
| "learning_rate": 1.7956686078964255e-07, | |
| "loss": 0.0383, | |
| "step": 904, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.19469584258694483, | |
| "train/R_vocab": 0.21830357142857146 | |
| }, | |
| { | |
| "learning_rate": 1.7852952894852148e-07, | |
| "loss": 0.0293, | |
| "step": 906, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.25663321405687106, | |
| "train/R_vocab": 0.32265625000000003 | |
| }, | |
| { | |
| "learning_rate": 1.7749353540429663e-07, | |
| "loss": -0.0106, | |
| "step": 908, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.19230140159133588, | |
| "train/R_vocab": 0.16744791666666667 | |
| }, | |
| { | |
| "learning_rate": 1.7645889955612592e-07, | |
| "loss": -0.0295, | |
| "step": 910, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.12090342349306576, | |
| "train/R_vocab": 0.1583457341269841 | |
| }, | |
| { | |
| "learning_rate": 1.7542564077774403e-07, | |
| "loss": 0.0045, | |
| "step": 912, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.675, | |
| "train/R_reason": 0.2114914680859179, | |
| "train/R_vocab": 0.3030505952380952 | |
| }, | |
| { | |
| "learning_rate": 1.7439377841709973e-07, | |
| "loss": -0.0017, | |
| "step": 914, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.217770400372361, | |
| "train/R_vocab": 0.25592757936507937 | |
| }, | |
| { | |
| "learning_rate": 1.7336333179599378e-07, | |
| "loss": 0.0849, | |
| "step": 916, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.2602397752431368, | |
| "train/R_vocab": 0.4150173611111111 | |
| }, | |
| { | |
| "learning_rate": 1.7233432020971684e-07, | |
| "loss": 0.0793, | |
| "step": 918, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.2766625564419999, | |
| "train/R_vocab": 0.23649553571428572 | |
| }, | |
| { | |
| "learning_rate": 1.7130676292668816e-07, | |
| "loss": 0.0446, | |
| "step": 920, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.6875, | |
| "train/R_reason": 0.22695619844003265, | |
| "train/R_vocab": 0.20081845238095236 | |
| }, | |
| { | |
| "learning_rate": 1.7028067918809535e-07, | |
| "loss": -0.036, | |
| "step": 922, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.17932276860695562, | |
| "train/R_vocab": 0.2564732142857143 | |
| }, | |
| { | |
| "learning_rate": 1.6925608820753325e-07, | |
| "loss": -0.0333, | |
| "step": 924, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.1710392126096163, | |
| "train/R_vocab": 0.11149553571428572 | |
| }, | |
| { | |
| "learning_rate": 1.6823300917064458e-07, | |
| "loss": -0.0427, | |
| "step": 926, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.1853385177642702, | |
| "train/R_vocab": 0.20937499999999998 | |
| }, | |
| { | |
| "learning_rate": 1.6721146123476093e-07, | |
| "loss": 0.122, | |
| "step": 928, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.21111988917807223, | |
| "train/R_vocab": 0.28467261904761904 | |
| }, | |
| { | |
| "learning_rate": 1.6619146352854334e-07, | |
| "loss": 0.0335, | |
| "step": 930, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.525, | |
| "train/R_reason": 0.2012206313966573, | |
| "train/R_vocab": 0.2486359126984127 | |
| }, | |
| { | |
| "learning_rate": 1.6517303515162505e-07, | |
| "loss": -0.0597, | |
| "step": 932, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.16395669599613677, | |
| "train/R_vocab": 0.2521329365079365 | |
| }, | |
| { | |
| "learning_rate": 1.6415619517425294e-07, | |
| "loss": 0.0165, | |
| "step": 934, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65, | |
| "train/R_reason": 0.14773459331491504, | |
| "train/R_vocab": 0.2604910714285714 | |
| }, | |
| { | |
| "learning_rate": 1.6314096263693078e-07, | |
| "loss": -0.0171, | |
| "step": 936, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.24617103354461017, | |
| "train/R_vocab": 0.25078125 | |
| }, | |
| { | |
| "learning_rate": 1.6212735655006322e-07, | |
| "loss": -0.0329, | |
| "step": 938, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.2476563881397631, | |
| "train/R_vocab": 0.22760416666666666 | |
| }, | |
| { | |
| "learning_rate": 1.6111539589359885e-07, | |
| "loss": -0.0157, | |
| "step": 940, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.18716442508640668, | |
| "train/R_vocab": 0.21462053571428572 | |
| }, | |
| { | |
| "learning_rate": 1.6010509961667546e-07, | |
| "loss": 0.0015, | |
| "step": 942, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.2163363343470736, | |
| "train/R_vocab": 0.3014508928571429 | |
| }, | |
| { | |
| "learning_rate": 1.590964866372652e-07, | |
| "loss": -0.0064, | |
| "step": 944, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5875, | |
| "train/R_reason": 0.22361495696140293, | |
| "train/R_vocab": 0.278125 | |
| }, | |
| { | |
| "learning_rate": 1.5808957584181994e-07, | |
| "loss": -0.0227, | |
| "step": 946, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.596875, | |
| "train/R_reason": 0.2240925844127057, | |
| "train/R_vocab": 0.30379464285714286 | |
| }, | |
| { | |
| "learning_rate": 1.5708438608491815e-07, | |
| "loss": 0.0113, | |
| "step": 948, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.19887587350757036, | |
| "train/R_vocab": 0.2051846590909091 | |
| }, | |
| { | |
| "learning_rate": 1.560809361889112e-07, | |
| "loss": 0.0187, | |
| "step": 950, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.1577485102536454, | |
| "train/R_vocab": 0.20122767857142856 | |
| }, | |
| { | |
| "learning_rate": 1.5507924494357154e-07, | |
| "loss": -0.0467, | |
| "step": 952, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.14574324556072643, | |
| "train/R_vocab": 0.20066964285714284 | |
| }, | |
| { | |
| "learning_rate": 1.5407933110574044e-07, | |
| "loss": 0.0123, | |
| "step": 954, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.21002736065355815, | |
| "train/R_vocab": 0.27578125 | |
| }, | |
| { | |
| "learning_rate": 1.5308121339897705e-07, | |
| "loss": 0.0131, | |
| "step": 956, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.24345409494067652, | |
| "train/R_vocab": 0.33035714285714285 | |
| }, | |
| { | |
| "learning_rate": 1.5208491051320744e-07, | |
| "loss": -0.1197, | |
| "step": 958, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.21702138753711128, | |
| "train/R_vocab": 0.28426339285714286 | |
| }, | |
| { | |
| "learning_rate": 1.510904411043751e-07, | |
| "loss": -0.0818, | |
| "step": 960, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.23023067131726802, | |
| "train/R_vocab": 0.3091517857142857 | |
| }, | |
| { | |
| "learning_rate": 1.5009782379409108e-07, | |
| "loss": -0.1087, | |
| "step": 962, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.765625, | |
| "train/R_reason": 0.17687649620358462, | |
| "train/R_vocab": 0.21863839285714284 | |
| }, | |
| { | |
| "learning_rate": 1.4910707716928586e-07, | |
| "loss": 0.0956, | |
| "step": 964, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.46875, | |
| "train/R_reason": 0.22474928414381412, | |
| "train/R_vocab": 0.26093750000000004 | |
| }, | |
| { | |
| "learning_rate": 1.4811821978186078e-07, | |
| "loss": 0.0536, | |
| "step": 966, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.22651502510275934, | |
| "train/R_vocab": 0.20833333333333331 | |
| }, | |
| { | |
| "learning_rate": 1.4713127014834088e-07, | |
| "loss": 0.0205, | |
| "step": 968, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.2021908958276862, | |
| "train/R_vocab": 0.1978794642857143 | |
| }, | |
| { | |
| "learning_rate": 1.461462467495284e-07, | |
| "loss": 0.0427, | |
| "step": 970, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.15024320937786767, | |
| "train/R_vocab": 0.20468750000000002 | |
| }, | |
| { | |
| "learning_rate": 1.4516316803015627e-07, | |
| "loss": 0.0126, | |
| "step": 972, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.18895812905862736, | |
| "train/R_vocab": 0.14308035714285713 | |
| }, | |
| { | |
| "learning_rate": 1.441820523985431e-07, | |
| "loss": -0.0349, | |
| "step": 974, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.2382146965819557, | |
| "train/R_vocab": 0.2640997023809524 | |
| }, | |
| { | |
| "learning_rate": 1.432029182262482e-07, | |
| "loss": 0.0445, | |
| "step": 976, | |
| "train/R_acc": 0.09375, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2085522733677449, | |
| "train/R_vocab": 0.26365327380952386 | |
| }, | |
| { | |
| "learning_rate": 1.42225783847728e-07, | |
| "loss": 0.0309, | |
| "step": 978, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.27484186892341045, | |
| "train/R_vocab": 0.3541170634920635 | |
| }, | |
| { | |
| "learning_rate": 1.41250667559992e-07, | |
| "loss": 0.0122, | |
| "step": 980, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.22480109886784352, | |
| "train/R_vocab": 0.30301339285714285 | |
| }, | |
| { | |
| "learning_rate": 1.4027758762226107e-07, | |
| "loss": -0.0861, | |
| "step": 982, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.21184840132463886, | |
| "train/R_vocab": 0.34201388888888884 | |
| }, | |
| { | |
| "learning_rate": 1.3930656225562474e-07, | |
| "loss": 0.0047, | |
| "step": 984, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.2253757708880576, | |
| "train/R_vocab": 0.30904017857142857 | |
| }, | |
| { | |
| "learning_rate": 1.3833760964270058e-07, | |
| "loss": 0.0296, | |
| "step": 986, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2222406698557359, | |
| "train/R_vocab": 0.305406746031746 | |
| }, | |
| { | |
| "learning_rate": 1.3737074792729332e-07, | |
| "loss": 0.0262, | |
| "step": 988, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2037288188497659, | |
| "train/R_vocab": 0.17431795634920633 | |
| }, | |
| { | |
| "learning_rate": 1.3640599521405522e-07, | |
| "loss": 0.0671, | |
| "step": 990, | |
| "train/R_acc": 0.040625, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.19595219511870288, | |
| "train/R_vocab": 0.2941964285714286 | |
| }, | |
| { | |
| "learning_rate": 1.354433695681474e-07, | |
| "loss": -0.0366, | |
| "step": 992, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.1895803639379184, | |
| "train/R_vocab": 0.2979910714285714 | |
| }, | |
| { | |
| "learning_rate": 1.3448288901490092e-07, | |
| "loss": 0.0929, | |
| "step": 994, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.21089602634201932, | |
| "train/R_vocab": 0.19340277777777776 | |
| }, | |
| { | |
| "learning_rate": 1.3352457153947997e-07, | |
| "loss": 0.0515, | |
| "step": 996, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.21176803496084237, | |
| "train/R_vocab": 0.3035094246031746 | |
| }, | |
| { | |
| "learning_rate": 1.3256843508654444e-07, | |
| "loss": 0.0103, | |
| "step": 998, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.15247972660677223, | |
| "train/R_vocab": 0.21183035714285717 | |
| }, | |
| { | |
| "learning_rate": 1.3161449755991425e-07, | |
| "loss": -0.082, | |
| "step": 1000, | |
| "train/R_acc": 0.13125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.2110448967120055, | |
| "train/R_vocab": 0.25863095238095235 | |
| }, | |
| { | |
| "eval/R_acc": 0.0025, | |
| "eval/R_penalty": -0.55, | |
| "eval/R_reason": 0.2655294246752871, | |
| "eval/R_vocab": 0.34914772727272725, | |
| "step": 1000 | |
| }, | |
| { | |
| "step": 1000 | |
| }, | |
| { | |
| "learning_rate": 1.3066277682223425e-07, | |
| "loss": -0.0096, | |
| "step": 1002, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.181312457358332, | |
| "train/R_vocab": 0.21517857142857144 | |
| }, | |
| { | |
| "learning_rate": 1.2971329069463932e-07, | |
| "loss": -0.0918, | |
| "step": 1004, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.12020329779967928, | |
| "train/R_vocab": 0.12425595238095238 | |
| }, | |
| { | |
| "learning_rate": 1.2876605695642084e-07, | |
| "loss": -0.0157, | |
| "step": 1006, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.2288382776444276, | |
| "train/R_vocab": 0.3042410714285715 | |
| }, | |
| { | |
| "learning_rate": 1.2782109334469409e-07, | |
| "loss": 0.0782, | |
| "step": 1008, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.23151180047594122, | |
| "train/R_vocab": 0.30937499999999996 | |
| }, | |
| { | |
| "learning_rate": 1.2687841755406554e-07, | |
| "loss": -0.0145, | |
| "step": 1010, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.23526236082480567, | |
| "train/R_vocab": 0.30009920634920634 | |
| }, | |
| { | |
| "learning_rate": 1.2593804723630209e-07, | |
| "loss": 0.0254, | |
| "step": 1012, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.2194206796665749, | |
| "train/R_vocab": 0.3340401785714286 | |
| }, | |
| { | |
| "learning_rate": 1.2500000000000005e-07, | |
| "loss": 0.0277, | |
| "step": 1014, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.18922638449707688, | |
| "train/R_vocab": 0.3610491071428572 | |
| }, | |
| { | |
| "learning_rate": 1.2406429341025556e-07, | |
| "loss": 0.0668, | |
| "step": 1016, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.2346744057550262, | |
| "train/R_vocab": 0.2239955357142857 | |
| }, | |
| { | |
| "learning_rate": 1.231309449883361e-07, | |
| "loss": 0.0338, | |
| "step": 1018, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.19074387902354337, | |
| "train/R_vocab": 0.21744791666666666 | |
| }, | |
| { | |
| "learning_rate": 1.2219997221135177e-07, | |
| "loss": 0.0082, | |
| "step": 1020, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.24590138406147055, | |
| "train/R_vocab": 0.2600446428571429 | |
| }, | |
| { | |
| "learning_rate": 1.2127139251192823e-07, | |
| "loss": 0.0268, | |
| "step": 1022, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.24386600100476474, | |
| "train/R_vocab": 0.3195560515873016 | |
| }, | |
| { | |
| "learning_rate": 1.203452232778807e-07, | |
| "loss": 0.0183, | |
| "step": 1024, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.171127249096387, | |
| "train/R_vocab": 0.29609375000000004 | |
| }, | |
| { | |
| "learning_rate": 1.1942148185188763e-07, | |
| "loss": 0.0272, | |
| "step": 1026, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.17080492182756957, | |
| "train/R_vocab": 0.21568627450980393 | |
| }, | |
| { | |
| "learning_rate": 1.1850018553116676e-07, | |
| "loss": -0.0052, | |
| "step": 1028, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.2604828445455006, | |
| "train/R_vocab": 0.23158482142857142 | |
| }, | |
| { | |
| "learning_rate": 1.1758135156715041e-07, | |
| "loss": -0.0154, | |
| "step": 1030, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.1946716410122738, | |
| "train/R_vocab": 0.22801339285714284 | |
| }, | |
| { | |
| "learning_rate": 1.1666499716516302e-07, | |
| "loss": 0.0101, | |
| "step": 1032, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.18422370796707505, | |
| "train/R_vocab": 0.21763392857142855 | |
| }, | |
| { | |
| "learning_rate": 1.157511394840989e-07, | |
| "loss": -0.0049, | |
| "step": 1034, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.7125, | |
| "train/R_reason": 0.17663980590936243, | |
| "train/R_vocab": 0.24196428571428574 | |
| }, | |
| { | |
| "learning_rate": 1.1483979563610069e-07, | |
| "loss": 0.0098, | |
| "step": 1036, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.26512284555646376, | |
| "train/R_vocab": 0.375 | |
| }, | |
| { | |
| "learning_rate": 1.1393098268623902e-07, | |
| "loss": 0.1176, | |
| "step": 1038, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.26504285268931027, | |
| "train/R_vocab": 0.30766369047619047 | |
| }, | |
| { | |
| "learning_rate": 1.1302471765219327e-07, | |
| "loss": 0.0428, | |
| "step": 1040, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2008627221719819, | |
| "train/R_vocab": 0.253125 | |
| }, | |
| { | |
| "learning_rate": 1.1212101750393235e-07, | |
| "loss": 0.0175, | |
| "step": 1042, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.409375, | |
| "train/R_reason": 0.19803424974179828, | |
| "train/R_vocab": 0.17265625 | |
| }, | |
| { | |
| "learning_rate": 1.1121989916339756e-07, | |
| "loss": 0.0292, | |
| "step": 1044, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.21600135050493885, | |
| "train/R_vocab": 0.3185267857142857 | |
| }, | |
| { | |
| "learning_rate": 1.1032137950418514e-07, | |
| "loss": 0.0669, | |
| "step": 1046, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.23236107648223334, | |
| "train/R_vocab": 0.22142857142857142 | |
| }, | |
| { | |
| "learning_rate": 1.0942547535123057e-07, | |
| "loss": -0.0038, | |
| "step": 1048, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.509375, | |
| "train/R_reason": 0.24041357438290228, | |
| "train/R_vocab": 0.2955357142857143 | |
| }, | |
| { | |
| "learning_rate": 1.085322034804938e-07, | |
| "loss": 0.1258, | |
| "step": 1050, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.20610390132448592, | |
| "train/R_vocab": 0.24229910714285713 | |
| }, | |
| { | |
| "learning_rate": 1.0764158061864448e-07, | |
| "loss": -0.0495, | |
| "step": 1052, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65, | |
| "train/R_reason": 0.2606824569537362, | |
| "train/R_vocab": 0.39392361111111107 | |
| }, | |
| { | |
| "learning_rate": 1.0675362344274952e-07, | |
| "loss": -0.0224, | |
| "step": 1054, | |
| "train/R_acc": 0.040625, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.2362594455371677, | |
| "train/R_vocab": 0.19696631493506495 | |
| }, | |
| { | |
| "learning_rate": 1.0586834857995996e-07, | |
| "loss": 0.0405, | |
| "step": 1056, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.22383388089744485, | |
| "train/R_vocab": 0.2839285714285714 | |
| }, | |
| { | |
| "learning_rate": 1.0498577260720048e-07, | |
| "loss": 0.006, | |
| "step": 1058, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.2154017687323606, | |
| "train/R_vocab": 0.18723958333333335 | |
| }, | |
| { | |
| "learning_rate": 1.0410591205085817e-07, | |
| "loss": 0.0547, | |
| "step": 1060, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.64375, | |
| "train/R_reason": 0.22723913109425195, | |
| "train/R_vocab": 0.32499999999999996 | |
| }, | |
| { | |
| "learning_rate": 1.0322878338647392e-07, | |
| "loss": 0.1005, | |
| "step": 1062, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65, | |
| "train/R_reason": 0.21853491707209402, | |
| "train/R_vocab": 0.19251217532467532 | |
| }, | |
| { | |
| "learning_rate": 1.0235440303843302e-07, | |
| "loss": 0.0902, | |
| "step": 1064, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.21816371211780805, | |
| "train/R_vocab": 0.22801339285714284 | |
| }, | |
| { | |
| "learning_rate": 1.0148278737965844e-07, | |
| "loss": -0.0553, | |
| "step": 1066, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.19269070794229914, | |
| "train/R_vocab": 0.18387896825396824 | |
| }, | |
| { | |
| "learning_rate": 1.0061395273130368e-07, | |
| "loss": 0.039, | |
| "step": 1068, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.21827443734801666, | |
| "train/R_vocab": 0.2911830357142857 | |
| }, | |
| { | |
| "learning_rate": 9.974791536244726e-08, | |
| "loss": 0.0343, | |
| "step": 1070, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.2733091389175897, | |
| "train/R_vocab": 0.3764136904761905 | |
| }, | |
| { | |
| "learning_rate": 9.888469148978843e-08, | |
| "loss": -0.0473, | |
| "step": 1072, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.24157353834104733, | |
| "train/R_vocab": 0.3257496843434343 | |
| }, | |
| { | |
| "learning_rate": 9.802429727734294e-08, | |
| "loss": -0.0036, | |
| "step": 1074, | |
| "train/R_acc": 0.040625, | |
| "train/R_penalty": -0.421875, | |
| "train/R_reason": 0.2823229532013337, | |
| "train/R_vocab": 0.43273809523809526 | |
| }, | |
| { | |
| "learning_rate": 9.716674883614092e-08, | |
| "loss": 0.0124, | |
| "step": 1076, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.453125, | |
| "train/R_reason": 0.2879044257474095, | |
| "train/R_vocab": 0.33738839285714284 | |
| }, | |
| { | |
| "learning_rate": 9.631206222392479e-08, | |
| "loss": -0.0157, | |
| "step": 1078, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.26644330439869424, | |
| "train/R_vocab": 0.27585565476190477 | |
| }, | |
| { | |
| "learning_rate": 9.546025344484868e-08, | |
| "loss": 0.0013, | |
| "step": 1080, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.17772787504592696, | |
| "train/R_vocab": 0.22031250000000002 | |
| }, | |
| { | |
| "learning_rate": 9.461133844917901e-08, | |
| "loss": -0.0364, | |
| "step": 1082, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.14647117758570943, | |
| "train/R_vocab": 0.19616815476190477 | |
| }, | |
| { | |
| "learning_rate": 9.376533313299542e-08, | |
| "loss": -0.0051, | |
| "step": 1084, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.18199736371919578, | |
| "train/R_vocab": 0.2505580357142857 | |
| }, | |
| { | |
| "learning_rate": 9.292225333789327e-08, | |
| "loss": 0.0575, | |
| "step": 1086, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.603125, | |
| "train/R_reason": 0.1892425633255841, | |
| "train/R_vocab": 0.31540178571428573 | |
| }, | |
| { | |
| "learning_rate": 9.208211485068729e-08, | |
| "loss": 0.0395, | |
| "step": 1088, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.15220888566600485, | |
| "train/R_vocab": 0.21587301587301586 | |
| }, | |
| { | |
| "learning_rate": 9.124493340311537e-08, | |
| "loss": -0.016, | |
| "step": 1090, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.18263365867722547, | |
| "train/R_vocab": 0.25424107142857144 | |
| }, | |
| { | |
| "learning_rate": 9.041072467154471e-08, | |
| "loss": -0.0291, | |
| "step": 1092, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.375, | |
| "train/R_reason": 0.25466359857700704, | |
| "train/R_vocab": 0.34862351190476193 | |
| }, | |
| { | |
| "learning_rate": 8.957950427667751e-08, | |
| "loss": 0.0296, | |
| "step": 1094, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.23575074339334445, | |
| "train/R_vocab": 0.2983010912698413 | |
| }, | |
| { | |
| "learning_rate": 8.875128778325902e-08, | |
| "loss": -0.0389, | |
| "step": 1096, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.1681449082491554, | |
| "train/R_vocab": 0.23660714285714285 | |
| }, | |
| { | |
| "learning_rate": 8.792609069978602e-08, | |
| "loss": 0.0201, | |
| "step": 1098, | |
| "train/R_acc": 0.0125, | |
| "train/R_penalty": -0.765625, | |
| "train/R_reason": 0.24261987678361474, | |
| "train/R_vocab": 0.23020833333333335 | |
| }, | |
| { | |
| "learning_rate": 8.710392847821615e-08, | |
| "loss": -0.0508, | |
| "step": 1100, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.19600584280495498, | |
| "train/R_vocab": 0.24397321428571428 | |
| }, | |
| { | |
| "eval/R_acc": 0.0025, | |
| "eval/R_penalty": -0.4875, | |
| "eval/R_reason": 0.24036428951274935, | |
| "eval/R_vocab": 0.29469967532467534, | |
| "step": 1100 | |
| }, | |
| { | |
| "step": 1100 | |
| }, | |
| { | |
| "learning_rate": 8.628481651367875e-08, | |
| "loss": 0.0336, | |
| "step": 1102, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.22866834230086636, | |
| "train/R_vocab": 0.19769345238095237 | |
| }, | |
| { | |
| "learning_rate": 8.546877014418671e-08, | |
| "loss": -0.0387, | |
| "step": 1104, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.21012454085545956, | |
| "train/R_vocab": 0.23225446428571428 | |
| }, | |
| { | |
| "learning_rate": 8.465580465034894e-08, | |
| "loss": 0.0231, | |
| "step": 1106, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.20315946724213801, | |
| "train/R_vocab": 0.2909598214285714 | |
| }, | |
| { | |
| "learning_rate": 8.384593525508466e-08, | |
| "loss": 0.0343, | |
| "step": 1108, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.509375, | |
| "train/R_reason": 0.20224091353679885, | |
| "train/R_vocab": 0.19077380952380954 | |
| }, | |
| { | |
| "learning_rate": 8.303917712333788e-08, | |
| "loss": 0.0208, | |
| "step": 1110, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.23770102329722845, | |
| "train/R_vocab": 0.21218998015873014 | |
| }, | |
| { | |
| "learning_rate": 8.223554536179367e-08, | |
| "loss": 0.0016, | |
| "step": 1112, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.24531693423546427, | |
| "train/R_vocab": 0.3191964285714286 | |
| }, | |
| { | |
| "learning_rate": 8.143505501859551e-08, | |
| "loss": 0.0352, | |
| "step": 1114, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.49375, | |
| "train/R_reason": 0.2098236201971303, | |
| "train/R_vocab": 0.2838169642857143 | |
| }, | |
| { | |
| "learning_rate": 8.06377210830631e-08, | |
| "loss": 0.0216, | |
| "step": 1116, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.20338203537474456, | |
| "train/R_vocab": 0.3302455357142857 | |
| }, | |
| { | |
| "learning_rate": 7.984355848541175e-08, | |
| "loss": -0.0481, | |
| "step": 1118, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.17974317204155837, | |
| "train/R_vocab": 0.32105654761904767 | |
| }, | |
| { | |
| "learning_rate": 7.905258209647325e-08, | |
| "loss": 0.0544, | |
| "step": 1120, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.603125, | |
| "train/R_reason": 0.24957606202713686, | |
| "train/R_vocab": 0.30379464285714286 | |
| }, | |
| { | |
| "learning_rate": 7.826480672741676e-08, | |
| "loss": -0.0196, | |
| "step": 1122, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.21989961574561373, | |
| "train/R_vocab": 0.24270833333333333 | |
| }, | |
| { | |
| "learning_rate": 7.748024712947204e-08, | |
| "loss": 0.0036, | |
| "step": 1124, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.20629454131030475, | |
| "train/R_vocab": 0.2872767857142857 | |
| }, | |
| { | |
| "learning_rate": 7.669891799365282e-08, | |
| "loss": 0.0386, | |
| "step": 1126, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.22259933007352922, | |
| "train/R_vocab": 0.36199156746031746 | |
| }, | |
| { | |
| "learning_rate": 7.592083395048182e-08, | |
| "loss": 0.0643, | |
| "step": 1128, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.49375, | |
| "train/R_reason": 0.18965226315756645, | |
| "train/R_vocab": 0.2122767857142857 | |
| }, | |
| { | |
| "learning_rate": 7.514600956971712e-08, | |
| "loss": 0.0678, | |
| "step": 1130, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5875, | |
| "train/R_reason": 0.23261509390210205, | |
| "train/R_vocab": 0.23307291666666669 | |
| }, | |
| { | |
| "learning_rate": 7.437445936007867e-08, | |
| "loss": 0.01, | |
| "step": 1132, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.2093342551771361, | |
| "train/R_vocab": 0.2585565476190476 | |
| }, | |
| { | |
| "learning_rate": 7.360619776897742e-08, | |
| "loss": -0.0392, | |
| "step": 1134, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.525, | |
| "train/R_reason": 0.22079774921664902, | |
| "train/R_vocab": 0.27708333333333335 | |
| }, | |
| { | |
| "learning_rate": 7.284123918224397e-08, | |
| "loss": 0.0161, | |
| "step": 1136, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.22520667797905555, | |
| "train/R_vocab": 0.3407118055555556 | |
| }, | |
| { | |
| "learning_rate": 7.207959792385998e-08, | |
| "loss": 0.0479, | |
| "step": 1138, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.19075869235061965, | |
| "train/R_vocab": 0.2729166666666667 | |
| }, | |
| { | |
| "learning_rate": 7.132128825568914e-08, | |
| "loss": -0.0027, | |
| "step": 1140, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.23292150900288705, | |
| "train/R_vocab": 0.21964285714285714 | |
| }, | |
| { | |
| "learning_rate": 7.056632437721099e-08, | |
| "loss": -0.0379, | |
| "step": 1142, | |
| "train/R_acc": 0.0625, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.17068707794196553, | |
| "train/R_vocab": 0.23448660714285716 | |
| }, | |
| { | |
| "learning_rate": 6.981472042525416e-08, | |
| "loss": -0.0097, | |
| "step": 1144, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.1982660319129968, | |
| "train/R_vocab": 0.18515624999999997 | |
| }, | |
| { | |
| "learning_rate": 6.906649047373245e-08, | |
| "loss": -0.0136, | |
| "step": 1146, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.571875, | |
| "train/R_reason": 0.18697098939403184, | |
| "train/R_vocab": 0.22890624999999998 | |
| }, | |
| { | |
| "learning_rate": 6.832164853338066e-08, | |
| "loss": 0.0075, | |
| "step": 1148, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.26961296459836603, | |
| "train/R_vocab": 0.3282118055555555 | |
| }, | |
| { | |
| "learning_rate": 6.758020855149249e-08, | |
| "loss": -0.0081, | |
| "step": 1150, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.24886312239764444, | |
| "train/R_vocab": 0.3654017857142857 | |
| }, | |
| { | |
| "learning_rate": 6.684218441165962e-08, | |
| "loss": 0.0666, | |
| "step": 1152, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.2088855316383599, | |
| "train/R_vocab": 0.22730654761904762 | |
| }, | |
| { | |
| "learning_rate": 6.610758993351123e-08, | |
| "loss": 0.0395, | |
| "step": 1154, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.22720806482016326, | |
| "train/R_vocab": 0.24931795634920634 | |
| }, | |
| { | |
| "learning_rate": 6.537643887245573e-08, | |
| "loss": 0.0465, | |
| "step": 1156, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.1968795889130503, | |
| "train/R_vocab": 0.2591517857142857 | |
| }, | |
| { | |
| "learning_rate": 6.464874491942279e-08, | |
| "loss": 0.0203, | |
| "step": 1158, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.19558365060317057, | |
| "train/R_vocab": 0.1140625 | |
| }, | |
| { | |
| "learning_rate": 6.392452170060706e-08, | |
| "loss": 0.0173, | |
| "step": 1160, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.540625, | |
| "train/R_reason": 0.25112434395380084, | |
| "train/R_vocab": 0.23022073412698413 | |
| }, | |
| { | |
| "learning_rate": 6.320378277721342e-08, | |
| "loss": -0.0386, | |
| "step": 1162, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.478125, | |
| "train/R_reason": 0.24359373155022102, | |
| "train/R_vocab": 0.3531498015873016 | |
| }, | |
| { | |
| "learning_rate": 6.248654164520237e-08, | |
| "loss": -0.0419, | |
| "step": 1164, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.71875, | |
| "train/R_reason": 0.20550711716499057, | |
| "train/R_vocab": 0.3539434523809524 | |
| }, | |
| { | |
| "learning_rate": 6.177281173503778e-08, | |
| "loss": 0.0405, | |
| "step": 1166, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.20957454414924856, | |
| "train/R_vocab": 0.2280133928571429 | |
| }, | |
| { | |
| "learning_rate": 6.106260641143546e-08, | |
| "loss": -0.0833, | |
| "step": 1168, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.20663920365929042, | |
| "train/R_vocab": 0.21852678571428574 | |
| }, | |
| { | |
| "learning_rate": 6.035593897311239e-08, | |
| "loss": -0.0156, | |
| "step": 1170, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.61875, | |
| "train/R_reason": 0.2371425092299592, | |
| "train/R_vocab": 0.3323784722222222 | |
| }, | |
| { | |
| "learning_rate": 5.965282265253838e-08, | |
| "loss": 0.0651, | |
| "step": 1172, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.390625, | |
| "train/R_reason": 0.22832625758068797, | |
| "train/R_vocab": 0.30301339285714285 | |
| }, | |
| { | |
| "learning_rate": 5.895327061568775e-08, | |
| "loss": -0.0379, | |
| "step": 1174, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.19140193598424352, | |
| "train/R_vocab": 0.2755952380952381 | |
| }, | |
| { | |
| "learning_rate": 5.8257295961792936e-08, | |
| "loss": -0.1099, | |
| "step": 1176, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.2013441433056279, | |
| "train/R_vocab": 0.2722222222222222 | |
| }, | |
| { | |
| "learning_rate": 5.756491172309952e-08, | |
| "loss": -0.0236, | |
| "step": 1178, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.2409798833837463, | |
| "train/R_vocab": 0.2924107142857143 | |
| }, | |
| { | |
| "learning_rate": 5.68761308646217e-08, | |
| "loss": -0.0326, | |
| "step": 1180, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.16268249986159153, | |
| "train/R_vocab": 0.17838541666666669 | |
| }, | |
| { | |
| "learning_rate": 5.6190966283899774e-08, | |
| "loss": 0.0286, | |
| "step": 1182, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.2404998117876481, | |
| "train/R_vocab": 0.34227430555555555 | |
| }, | |
| { | |
| "learning_rate": 5.5509430810758817e-08, | |
| "loss": 0.004, | |
| "step": 1184, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.22824560961553947, | |
| "train/R_vocab": 0.3375 | |
| }, | |
| { | |
| "learning_rate": 5.483153720706798e-08, | |
| "loss": -0.0074, | |
| "step": 1186, | |
| "train/R_acc": 0.1, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.21326554603657105, | |
| "train/R_vocab": 0.20881696428571428 | |
| }, | |
| { | |
| "learning_rate": 5.415729816650202e-08, | |
| "loss": 0.0246, | |
| "step": 1188, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.2444787127154557, | |
| "train/R_vocab": 0.35230654761904756 | |
| }, | |
| { | |
| "learning_rate": 5.3486726314303175e-08, | |
| "loss": 0.0987, | |
| "step": 1190, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.22288328491646994, | |
| "train/R_vocab": 0.2479910714285714 | |
| }, | |
| { | |
| "learning_rate": 5.2819834207044974e-08, | |
| "loss": -0.0208, | |
| "step": 1192, | |
| "train/R_acc": 0.128125, | |
| "train/R_penalty": -0.4375, | |
| "train/R_reason": 0.27609449087124366, | |
| "train/R_vocab": 0.3087425595238095 | |
| }, | |
| { | |
| "learning_rate": 5.215663433239728e-08, | |
| "loss": -0.006, | |
| "step": 1194, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.22109386819415358, | |
| "train/R_vocab": 0.24497767857142858 | |
| }, | |
| { | |
| "learning_rate": 5.149713910889206e-08, | |
| "loss": -0.0461, | |
| "step": 1196, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.421875, | |
| "train/R_reason": 0.2398768408665738, | |
| "train/R_vocab": 0.3560267857142857 | |
| }, | |
| { | |
| "learning_rate": 5.0841360885690996e-08, | |
| "loss": 0.0184, | |
| "step": 1198, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.21147620601357403, | |
| "train/R_vocab": 0.20399305555555555 | |
| }, | |
| { | |
| "learning_rate": 5.01893119423546e-08, | |
| "loss": 0.0505, | |
| "step": 1200, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.2487143037586896, | |
| "train/R_vocab": 0.3544642857142857 | |
| }, | |
| { | |
| "eval/R_acc": 0.0, | |
| "eval/R_penalty": -0.5875, | |
| "eval/R_reason": 0.20266877330422212, | |
| "eval/R_vocab": 0.2709090909090909, | |
| "step": 1200 | |
| }, | |
| { | |
| "step": 1200 | |
| }, | |
| { | |
| "learning_rate": 4.954100448861165e-08, | |
| "loss": -0.009, | |
| "step": 1202, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.446875, | |
| "train/R_reason": 0.2773491666866028, | |
| "train/R_vocab": 0.29509943181818177 | |
| }, | |
| { | |
| "learning_rate": 4.889645066413112e-08, | |
| "loss": 0.0306, | |
| "step": 1204, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.734375, | |
| "train/R_reason": 0.2607235600350966, | |
| "train/R_vocab": 0.1830729166666667 | |
| }, | |
| { | |
| "learning_rate": 4.82556625382945e-08, | |
| "loss": -0.0188, | |
| "step": 1206, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.20154795093112635, | |
| "train/R_vocab": 0.19722222222222224 | |
| }, | |
| { | |
| "learning_rate": 4.76186521099699e-08, | |
| "loss": -0.0385, | |
| "step": 1208, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.2240002350916589, | |
| "train/R_vocab": 0.3001116071428571 | |
| }, | |
| { | |
| "learning_rate": 4.698543130728755e-08, | |
| "loss": 0.0571, | |
| "step": 1210, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.15973930176485518, | |
| "train/R_vocab": 0.21350446428571426 | |
| }, | |
| { | |
| "learning_rate": 4.635601198741607e-08, | |
| "loss": 0.0042, | |
| "step": 1212, | |
| "train/R_acc": 0.06875, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.19491433422937576, | |
| "train/R_vocab": 0.24547483766233766 | |
| }, | |
| { | |
| "learning_rate": 4.573040593634092e-08, | |
| "loss": 0.0204, | |
| "step": 1214, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.7125, | |
| "train/R_reason": 0.2036836381635984, | |
| "train/R_vocab": 0.2583333333333333 | |
| }, | |
| { | |
| "learning_rate": 4.510862486864317e-08, | |
| "loss": -0.0139, | |
| "step": 1216, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.1935253778055071, | |
| "train/R_vocab": 0.2539434523809524 | |
| }, | |
| { | |
| "learning_rate": 4.449068042728063e-08, | |
| "loss": -0.0221, | |
| "step": 1218, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.17478719789421077, | |
| "train/R_vocab": 0.2385044642857143 | |
| }, | |
| { | |
| "learning_rate": 4.3876584183369454e-08, | |
| "loss": 0.0401, | |
| "step": 1220, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.12396243023486506, | |
| "train/R_vocab": 0.16953125 | |
| }, | |
| { | |
| "learning_rate": 4.326634763596784e-08, | |
| "loss": 0.0362, | |
| "step": 1222, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.21970153796136732, | |
| "train/R_vocab": 0.23229166666666667 | |
| }, | |
| { | |
| "learning_rate": 4.265998221186023e-08, | |
| "loss": 0.0579, | |
| "step": 1224, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.696875, | |
| "train/R_reason": 0.25118929349410624, | |
| "train/R_vocab": 0.3527901785714286 | |
| }, | |
| { | |
| "learning_rate": 4.205749926534391e-08, | |
| "loss": 0.0325, | |
| "step": 1226, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.20960885628620934, | |
| "train/R_vocab": 0.24062499999999998 | |
| }, | |
| { | |
| "learning_rate": 4.145891007801588e-08, | |
| "loss": -0.012, | |
| "step": 1228, | |
| "train/R_acc": 0.037500000000000006, | |
| "train/R_penalty": -0.375, | |
| "train/R_reason": 0.2440708995563252, | |
| "train/R_vocab": 0.31674107142857144 | |
| }, | |
| { | |
| "learning_rate": 4.086422585856189e-08, | |
| "loss": 0.0635, | |
| "step": 1230, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.16689617467212345, | |
| "train/R_vocab": 0.24068700396825393 | |
| }, | |
| { | |
| "learning_rate": 4.0273457742546566e-08, | |
| "loss": -0.0426, | |
| "step": 1232, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.192806526715886, | |
| "train/R_vocab": 0.2780505952380953 | |
| }, | |
| { | |
| "learning_rate": 3.968661679220467e-08, | |
| "loss": 0.0384, | |
| "step": 1234, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.17224863139672328, | |
| "train/R_vocab": 0.23426339285714287 | |
| }, | |
| { | |
| "learning_rate": 3.9103713996234355e-08, | |
| "loss": 0.0036, | |
| "step": 1236, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.734375, | |
| "train/R_reason": 0.17680126111067432, | |
| "train/R_vocab": 0.2612723214285714 | |
| }, | |
| { | |
| "learning_rate": 3.852476026959089e-08, | |
| "loss": 0.0172, | |
| "step": 1238, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.23625775975212537, | |
| "train/R_vocab": 0.2917410714285715 | |
| }, | |
| { | |
| "learning_rate": 3.794976645328266e-08, | |
| "loss": 0.0608, | |
| "step": 1240, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.49375, | |
| "train/R_reason": 0.20085612003970227, | |
| "train/R_vocab": 0.29375 | |
| }, | |
| { | |
| "learning_rate": 3.737874331416807e-08, | |
| "loss": -0.0099, | |
| "step": 1242, | |
| "train/R_acc": 0.06875, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.2067606031031175, | |
| "train/R_vocab": 0.1792782738095238 | |
| }, | |
| { | |
| "learning_rate": 3.681170154475391e-08, | |
| "loss": -0.0105, | |
| "step": 1244, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.24829657601476124, | |
| "train/R_vocab": 0.3546875 | |
| }, | |
| { | |
| "learning_rate": 3.624865176299499e-08, | |
| "loss": -0.0273, | |
| "step": 1246, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.2728674749829552, | |
| "train/R_vocab": 0.42109375000000004 | |
| }, | |
| { | |
| "learning_rate": 3.568960451209574e-08, | |
| "loss": 0.0602, | |
| "step": 1248, | |
| "train/R_acc": 0.0375, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2555940677098582, | |
| "train/R_vocab": 0.41767113095238095 | |
| }, | |
| { | |
| "learning_rate": 3.513457026031216e-08, | |
| "loss": -0.0627, | |
| "step": 1250, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.20621535738161093, | |
| "train/R_vocab": 0.16438492063492063 | |
| }, | |
| { | |
| "learning_rate": 3.458355940075652e-08, | |
| "loss": 0.0626, | |
| "step": 1252, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.25723943167289565, | |
| "train/R_vocab": 0.290625 | |
| }, | |
| { | |
| "learning_rate": 3.403658225120212e-08, | |
| "loss": -0.0653, | |
| "step": 1254, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.2280530069453393, | |
| "train/R_vocab": 0.26473214285714286 | |
| }, | |
| { | |
| "learning_rate": 3.349364905389032e-08, | |
| "loss": -0.013, | |
| "step": 1256, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.4375, | |
| "train/R_reason": 0.20137595355438587, | |
| "train/R_vocab": 0.3265625 | |
| }, | |
| { | |
| "learning_rate": 3.295476997533905e-08, | |
| "loss": 0.0298, | |
| "step": 1258, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.59375, | |
| "train/R_reason": 0.20510172934689136, | |
| "train/R_vocab": 0.3380580357142857 | |
| }, | |
| { | |
| "learning_rate": 3.2419955106151893e-08, | |
| "loss": -0.0064, | |
| "step": 1260, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.24581030179992447, | |
| "train/R_vocab": 0.27968750000000003 | |
| }, | |
| { | |
| "learning_rate": 3.188921446082946e-08, | |
| "loss": 0.0136, | |
| "step": 1262, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.5, | |
| "train/R_reason": 0.19498686377833271, | |
| "train/R_vocab": 0.21171875 | |
| }, | |
| { | |
| "learning_rate": 3.1362557977582e-08, | |
| "loss": 0.0327, | |
| "step": 1264, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.2546377151605037, | |
| "train/R_vocab": 0.38132440476190477 | |
| }, | |
| { | |
| "learning_rate": 3.0839995518142844e-08, | |
| "loss": 0.0813, | |
| "step": 1266, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.23407475867218072, | |
| "train/R_vocab": 0.25262896825396824 | |
| }, | |
| { | |
| "learning_rate": 3.032153686758432e-08, | |
| "loss": 0.0111, | |
| "step": 1268, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.75, | |
| "train/R_reason": 0.13893382364177903, | |
| "train/R_vocab": 0.1873139880952381 | |
| }, | |
| { | |
| "learning_rate": 2.980719173413396e-08, | |
| "loss": -0.0655, | |
| "step": 1270, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.241975792190938, | |
| "train/R_vocab": 0.359375 | |
| }, | |
| { | |
| "learning_rate": 2.9296969748993093e-08, | |
| "loss": 0.0107, | |
| "step": 1272, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.24308125423194502, | |
| "train/R_vocab": 0.18370535714285716 | |
| }, | |
| { | |
| "learning_rate": 2.8790880466156464e-08, | |
| "loss": 0.0368, | |
| "step": 1274, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.571875, | |
| "train/R_reason": 0.21833878646152022, | |
| "train/R_vocab": 0.2832217261904762 | |
| }, | |
| { | |
| "learning_rate": 2.8288933362233148e-08, | |
| "loss": 0.0016, | |
| "step": 1276, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.205687843726902, | |
| "train/R_vocab": 0.23705357142857142 | |
| }, | |
| { | |
| "learning_rate": 2.7791137836269158e-08, | |
| "loss": -0.0029, | |
| "step": 1278, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.2181107503992247, | |
| "train/R_vocab": 0.20520833333333333 | |
| }, | |
| { | |
| "learning_rate": 2.7297503209571665e-08, | |
| "loss": 0.1301, | |
| "step": 1280, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.16072625927287004, | |
| "train/R_vocab": 0.22611607142857143 | |
| }, | |
| { | |
| "learning_rate": 2.680803872553408e-08, | |
| "loss": 0.0204, | |
| "step": 1282, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.525, | |
| "train/R_reason": 0.18120856948997982, | |
| "train/R_vocab": 0.2267981150793651 | |
| }, | |
| { | |
| "learning_rate": 2.632275354946342e-08, | |
| "loss": 0.0174, | |
| "step": 1284, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.55625, | |
| "train/R_reason": 0.19887443410752542, | |
| "train/R_vocab": 0.20744047619047618 | |
| }, | |
| { | |
| "learning_rate": 2.5841656768408216e-08, | |
| "loss": 0.0125, | |
| "step": 1286, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.23872779990823945, | |
| "train/R_vocab": 0.28046875 | |
| }, | |
| { | |
| "learning_rate": 2.5364757390988616e-08, | |
| "loss": 0.0652, | |
| "step": 1288, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.19387491540524032, | |
| "train/R_vocab": 0.23203125 | |
| }, | |
| { | |
| "learning_rate": 2.489206434722785e-08, | |
| "loss": -0.012, | |
| "step": 1290, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.18941429133549656, | |
| "train/R_vocab": 0.23415178571428572 | |
| }, | |
| { | |
| "learning_rate": 2.4423586488384524e-08, | |
| "loss": 0.0179, | |
| "step": 1292, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.248278845576544, | |
| "train/R_vocab": 0.22329545454545455 | |
| }, | |
| { | |
| "learning_rate": 2.395933258678745e-08, | |
| "loss": 0.0178, | |
| "step": 1294, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.609375, | |
| "train/R_reason": 0.20548558553050647, | |
| "train/R_vocab": 0.3214905753968254 | |
| }, | |
| { | |
| "learning_rate": 2.349931133567096e-08, | |
| "loss": -0.0004, | |
| "step": 1296, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.2386328952490884, | |
| "train/R_vocab": 0.23169642857142858 | |
| }, | |
| { | |
| "learning_rate": 2.3043531349012353e-08, | |
| "loss": -0.0659, | |
| "step": 1298, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.671875, | |
| "train/R_reason": 0.15007159038519916, | |
| "train/R_vocab": 0.19921875 | |
| }, | |
| { | |
| "learning_rate": 2.259200116137039e-08, | |
| "loss": 0.0095, | |
| "step": 1300, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.65625, | |
| "train/R_reason": 0.20368911805265966, | |
| "train/R_vocab": 0.2474330357142857 | |
| }, | |
| { | |
| "eval/R_acc": 0.0025, | |
| "eval/R_penalty": -0.5625, | |
| "eval/R_reason": 0.22623212762462228, | |
| "eval/R_vocab": 0.3164285714285714, | |
| "step": 1300 | |
| }, | |
| { | |
| "step": 1300 | |
| }, | |
| { | |
| "learning_rate": 2.214472922772584e-08, | |
| "loss": 0.0957, | |
| "step": 1302, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.22116478834904366, | |
| "train/R_vocab": 0.29776785714285714 | |
| }, | |
| { | |
| "learning_rate": 2.1701723923322673e-08, | |
| "loss": -0.0749, | |
| "step": 1304, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.21324021225069947, | |
| "train/R_vocab": 0.2928977272727273 | |
| }, | |
| { | |
| "learning_rate": 2.1262993543511715e-08, | |
| "loss": 9.305, | |
| "step": 1306, | |
| "train/R_acc": 0.0375, | |
| "train/R_penalty": -0.484375, | |
| "train/R_reason": 0.23529231846467052, | |
| "train/R_vocab": 0.3332589285714286 | |
| }, | |
| { | |
| "learning_rate": 2.082854630359493e-08, | |
| "loss": 0.0331, | |
| "step": 1308, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.421875, | |
| "train/R_reason": 0.23986047807887131, | |
| "train/R_vocab": 0.36618303571428573 | |
| }, | |
| { | |
| "learning_rate": 2.03983903386718e-08, | |
| "loss": 0.0118, | |
| "step": 1310, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.2656667530369191, | |
| "train/R_vocab": 0.22192460317460316 | |
| }, | |
| { | |
| "learning_rate": 1.997253370348703e-08, | |
| "loss": 0.0556, | |
| "step": 1312, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.2476736629649698, | |
| "train/R_vocab": 0.31614583333333335 | |
| }, | |
| { | |
| "learning_rate": 1.9550984372279476e-08, | |
| "loss": -0.037, | |
| "step": 1314, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.19878604497907437, | |
| "train/R_vocab": 0.27142857142857146 | |
| }, | |
| { | |
| "learning_rate": 1.9133750238633117e-08, | |
| "loss": 0.0175, | |
| "step": 1316, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.634375, | |
| "train/R_reason": 0.2233584354922284, | |
| "train/R_vocab": 0.23705357142857142 | |
| }, | |
| { | |
| "learning_rate": 1.872083911532907e-08, | |
| "loss": 0.0006, | |
| "step": 1318, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.2794139552851759, | |
| "train/R_vocab": 0.37043650793650795 | |
| }, | |
| { | |
| "learning_rate": 1.8312258734199253e-08, | |
| "loss": -0.0277, | |
| "step": 1320, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.46875, | |
| "train/R_reason": 0.19046029498161032, | |
| "train/R_vocab": 0.2759548611111111 | |
| }, | |
| { | |
| "learning_rate": 1.7908016745981856e-08, | |
| "loss": 0.0615, | |
| "step": 1322, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.515625, | |
| "train/R_reason": 0.1939617321953623, | |
| "train/R_vocab": 0.2484623015873016 | |
| }, | |
| { | |
| "learning_rate": 1.7508120720177795e-08, | |
| "loss": -0.0239, | |
| "step": 1324, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.53125, | |
| "train/R_reason": 0.21866868750973806, | |
| "train/R_vocab": 0.2526785714285714 | |
| }, | |
| { | |
| "learning_rate": 1.7112578144909047e-08, | |
| "loss": -0.0744, | |
| "step": 1326, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.23171087837305318, | |
| "train/R_vocab": 0.28158482142857144 | |
| }, | |
| { | |
| "learning_rate": 1.6721396426778667e-08, | |
| "loss": 0.0113, | |
| "step": 1328, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.49375, | |
| "train/R_reason": 0.2527234820261308, | |
| "train/R_vocab": 0.290687003968254 | |
| }, | |
| { | |
| "learning_rate": 1.6334582890731697e-08, | |
| "loss": 0.0423, | |
| "step": 1330, | |
| "train/R_acc": 0.0, | |
| "train/R_penalty": -0.625, | |
| "train/R_reason": 0.18629708667365014, | |
| "train/R_vocab": 0.1819940476190476 | |
| }, | |
| { | |
| "learning_rate": 1.595214477991841e-08, | |
| "loss": 0.027, | |
| "step": 1332, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.23647010980361977, | |
| "train/R_vocab": 0.2343625992063492 | |
| }, | |
| { | |
| "learning_rate": 1.557408925555831e-08, | |
| "loss": 0.0619, | |
| "step": 1334, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.640625, | |
| "train/R_reason": 0.2178143312883945, | |
| "train/R_vocab": 0.28158482142857144 | |
| }, | |
| { | |
| "learning_rate": 1.5200423396806272e-08, | |
| "loss": -0.006, | |
| "step": 1336, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.40625, | |
| "train/R_reason": 0.26096178007275994, | |
| "train/R_vocab": 0.3358134920634921 | |
| }, | |
| { | |
| "learning_rate": 1.4831154200620033e-08, | |
| "loss": -0.0105, | |
| "step": 1338, | |
| "train/R_acc": 0.034375, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.18847235645846208, | |
| "train/R_vocab": 0.23511904761904762 | |
| }, | |
| { | |
| "learning_rate": 1.4466288581628906e-08, | |
| "loss": 0.0143, | |
| "step": 1340, | |
| "train/R_acc": 0.009375000000000001, | |
| "train/R_penalty": -0.578125, | |
| "train/R_reason": 0.23179656802725537, | |
| "train/R_vocab": 0.2564732142857143 | |
| }, | |
| { | |
| "learning_rate": 1.4105833372004523e-08, | |
| "loss": 0.0246, | |
| "step": 1342, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.1816783870148443, | |
| "train/R_vocab": 0.21495535714285713 | |
| }, | |
| { | |
| "learning_rate": 1.3749795321332885e-08, | |
| "loss": -0.0062, | |
| "step": 1344, | |
| "train/R_acc": 0.03125, | |
| "train/R_penalty": -0.546875, | |
| "train/R_reason": 0.22414561936584018, | |
| "train/R_vocab": 0.2232886904761905 | |
| }, | |
| { | |
| "learning_rate": 1.3398181096487882e-08, | |
| "loss": 0.0182, | |
| "step": 1346, | |
| "train/R_acc": 0.065625, | |
| "train/R_penalty": -0.5625, | |
| "train/R_reason": 0.2680587101569149, | |
| "train/R_vocab": 0.3170138888888889 | |
| }, | |
| { | |
| "learning_rate": 1.3050997281506549e-08, | |
| "loss": 0.0831, | |
| "step": 1348, | |
| "train/R_acc": 0.003125, | |
| "train/R_penalty": -0.68125, | |
| "train/R_reason": 0.2494493257892117, | |
| "train/R_vocab": 0.18958333333333333 | |
| }, | |
| { | |
| "learning_rate": 1.2708250377465645e-08, | |
| "loss": -0.004, | |
| "step": 1350, | |
| "train/R_acc": 0.00625, | |
| "train/R_penalty": -0.665625, | |
| "train/R_reason": 0.2093549139749035, | |
| "train/R_vocab": 0.1791497564935065 | |
| } | |
| ], | |
| "logging_steps": 2, | |
| "max_steps": 1497, | |
| "num_input_tokens_seen": 13748095, | |
| "num_train_epochs": 1, | |
| "save_steps": 50, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |