rlhf / checkpoint_1350 /trainer_state.json
whaleL's picture
Upload folder using huggingface_hub
763824c verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9018036072144289,
"eval_steps": 100,
"global_step": 1350,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"learning_rate": 1.111111111111111e-08,
"loss": 0.0176,
"step": 2,
"train/R_acc": 0.003125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.1771926448030963,
"train/R_vocab": 0.27042410714285714
},
{
"learning_rate": 3.3333333333333334e-08,
"loss": 0.0255,
"step": 4,
"train/R_acc": 0.00625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.20292794300576886,
"train/R_vocab": 0.13051835317460317
},
{
"learning_rate": 5.555555555555555e-08,
"loss": -0.0402,
"step": 6,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.21681085047252255,
"train/R_vocab": 0.3141617063492063
},
{
"learning_rate": 7.777777777777778e-08,
"loss": -0.0512,
"step": 8,
"train/R_acc": 0.034375,
"train/R_penalty": -0.53125,
"train/R_reason": 0.22407962891811065,
"train/R_vocab": 0.2578125
},
{
"learning_rate": 1e-07,
"loss": 0.0653,
"step": 10,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.24027254265023745,
"train/R_vocab": 0.290625
},
{
"learning_rate": 1.2222222222222222e-07,
"loss": 0.0747,
"step": 12,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2086565261070774,
"train/R_vocab": 0.2722098214285714
},
{
"learning_rate": 1.4444444444444442e-07,
"loss": -0.0073,
"step": 14,
"train/R_acc": 0.034375,
"train/R_penalty": -0.78125,
"train/R_reason": 0.18557344633053446,
"train/R_vocab": 0.23597470238095236
},
{
"learning_rate": 1.6666666666666665e-07,
"loss": 0.0011,
"step": 16,
"train/R_acc": 0.00625,
"train/R_penalty": -0.5625,
"train/R_reason": 0.18460204443700487,
"train/R_vocab": 0.28095238095238095
},
{
"learning_rate": 1.8888888888888888e-07,
"loss": 0.0511,
"step": 18,
"train/R_acc": 0.003125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.1691650605005834,
"train/R_vocab": 0.22633928571428572
},
{
"learning_rate": 2.111111111111111e-07,
"loss": -0.0247,
"step": 20,
"train/R_acc": 0.03125,
"train/R_penalty": -0.484375,
"train/R_reason": 0.23429908163761645,
"train/R_vocab": 0.3410714285714286
},
{
"learning_rate": 2.3333333333333333e-07,
"loss": 0.0088,
"step": 22,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.15088100811244767,
"train/R_vocab": 0.2056919642857143
},
{
"learning_rate": 2.5555555555555553e-07,
"loss": 0.0366,
"step": 24,
"train/R_acc": 0.0,
"train/R_penalty": -0.6875,
"train/R_reason": 0.2605457318222165,
"train/R_vocab": 0.29765625
},
{
"learning_rate": 2.7777777777777776e-07,
"loss": 0.0558,
"step": 26,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.5,
"train/R_reason": 0.24175666877540153,
"train/R_vocab": 0.26484375000000004
},
{
"learning_rate": 3e-07,
"loss": 0.0588,
"step": 28,
"train/R_acc": 0.0,
"train/R_penalty": -0.540625,
"train/R_reason": 0.2054192775694245,
"train/R_vocab": 0.2109747023809524
},
{
"learning_rate": 3.222222222222222e-07,
"loss": -0.1103,
"step": 30,
"train/R_acc": 0.003125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.18972739597143928,
"train/R_vocab": 0.2044642857142857
},
{
"learning_rate": 3.4444444444444444e-07,
"loss": 0.0162,
"step": 32,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.16386884044279187,
"train/R_vocab": 0.26328125
},
{
"learning_rate": 3.666666666666666e-07,
"loss": 0.0241,
"step": 34,
"train/R_acc": 0.065625,
"train/R_penalty": -0.6875,
"train/R_reason": 0.1991015810053379,
"train/R_vocab": 0.3072916666666667
},
{
"learning_rate": 3.888888888888889e-07,
"loss": 0.0114,
"step": 36,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.1542880167619275,
"train/R_vocab": 0.15633680555555557
},
{
"learning_rate": 4.1111111111111107e-07,
"loss": 0.0418,
"step": 38,
"train/R_acc": 0.03125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.21628990497316553,
"train/R_vocab": 0.32326388888888885
},
{
"learning_rate": 4.3333333333333335e-07,
"loss": -0.0085,
"step": 40,
"train/R_acc": 0.065625,
"train/R_penalty": -0.540625,
"train/R_reason": 0.16669781447815488,
"train/R_vocab": 0.21458333333333335
},
{
"learning_rate": 4.555555555555555e-07,
"loss": -0.0614,
"step": 42,
"train/R_acc": 0.003125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.23833932777613548,
"train/R_vocab": 0.24285714285714288
},
{
"learning_rate": 4.777777777777778e-07,
"loss": 0.004,
"step": 44,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.2297590496740362,
"train/R_vocab": 0.1607142857142857
},
{
"learning_rate": 5e-07,
"loss": 0.0491,
"step": 46,
"train/R_acc": 0.03125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.2037377285506986,
"train/R_vocab": 0.2598958333333333
},
{
"learning_rate": 4.999976528920434e-07,
"loss": -0.083,
"step": 48,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.22468232799435314,
"train/R_vocab": 0.2745907738095238
},
{
"learning_rate": 4.999906116122447e-07,
"loss": -0.0036,
"step": 50,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.22593521940016348,
"train/R_vocab": 0.2805803571428572
},
{
"learning_rate": 4.999788762928172e-07,
"loss": 0.0107,
"step": 52,
"train/R_acc": 0.003125,
"train/R_penalty": -0.821875,
"train/R_reason": 0.20522814513215548,
"train/R_vocab": 0.14702380952380953
},
{
"learning_rate": 4.999624471541134e-07,
"loss": 0.029,
"step": 54,
"train/R_acc": 0.0,
"train/R_penalty": -0.78125,
"train/R_reason": 0.21737758529192291,
"train/R_vocab": 0.19330357142857144
},
{
"learning_rate": 4.99941324504621e-07,
"loss": -0.0246,
"step": 56,
"train/R_acc": 0.00625,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2251848744560886,
"train/R_vocab": 0.28158482142857144
},
{
"learning_rate": 4.999155087409571e-07,
"loss": -0.0517,
"step": 58,
"train/R_acc": 0.03125,
"train/R_penalty": -0.571875,
"train/R_reason": 0.18310813181259872,
"train/R_vocab": 0.24776785714285715
},
{
"learning_rate": 4.998850003478606e-07,
"loss": 0.0598,
"step": 60,
"train/R_acc": 0.034375,
"train/R_penalty": -0.703125,
"train/R_reason": 0.2082140947052848,
"train/R_vocab": 0.26149553571428574
},
{
"learning_rate": 4.998497998981838e-07,
"loss": 0.0027,
"step": 62,
"train/R_acc": 0.0,
"train/R_penalty": -0.828125,
"train/R_reason": 0.2611822631590372,
"train/R_vocab": 0.39639136904761907
},
{
"learning_rate": 4.998099080528803e-07,
"loss": 0.0385,
"step": 64,
"train/R_acc": 0.0,
"train/R_penalty": -0.53125,
"train/R_reason": 0.2213369943749453,
"train/R_vocab": 0.3606026785714286
},
{
"learning_rate": 4.997653255609941e-07,
"loss": -0.0523,
"step": 66,
"train/R_acc": 0.065625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.20860609604648056,
"train/R_vocab": 0.2251488095238095
},
{
"learning_rate": 4.997160532596447e-07,
"loss": 0.0352,
"step": 68,
"train/R_acc": 0.003125,
"train/R_penalty": -0.790625,
"train/R_reason": 0.19482757334666323,
"train/R_vocab": 0.23776041666666667
},
{
"learning_rate": 4.996620920740111e-07,
"loss": 0.0103,
"step": 70,
"train/R_acc": 0.03125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.25304049429757086,
"train/R_vocab": 0.299516369047619
},
{
"learning_rate": 4.996034430173153e-07,
"loss": 0.0247,
"step": 72,
"train/R_acc": 0.003125,
"train/R_penalty": -0.58125,
"train/R_reason": 0.27724208008758466,
"train/R_vocab": 0.3654637896825397
},
{
"learning_rate": 4.995401071908025e-07,
"loss": -0.0689,
"step": 74,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.18864259893071442,
"train/R_vocab": 0.265625
},
{
"learning_rate": 4.99472085783721e-07,
"loss": 0.0669,
"step": 76,
"train/R_acc": 0.003125,
"train/R_penalty": -0.775,
"train/R_reason": 0.24872595893355823,
"train/R_vocab": 0.3293526785714286
},
{
"learning_rate": 4.993993800732995e-07,
"loss": -0.0369,
"step": 78,
"train/R_acc": 0.034375,
"train/R_penalty": -0.546875,
"train/R_reason": 0.26486544822313024,
"train/R_vocab": 0.3398065476190476
},
{
"learning_rate": 4.993219914247231e-07,
"loss": 0.0685,
"step": 80,
"train/R_acc": 0.03125,
"train/R_penalty": -0.75,
"train/R_reason": 0.1711866679882719,
"train/R_vocab": 0.2575892857142857
},
{
"learning_rate": 4.992399212911082e-07,
"loss": 0.0382,
"step": 82,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.640625,
"train/R_reason": 0.22934020748837214,
"train/R_vocab": 0.31875
},
{
"learning_rate": 4.99153171213474e-07,
"loss": 0.083,
"step": 84,
"train/R_acc": 0.003125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.16303826493835916,
"train/R_vocab": 0.18002232142857144
},
{
"learning_rate": 4.990617428207153e-07,
"loss": 0.0433,
"step": 86,
"train/R_acc": 0.03125,
"train/R_penalty": -0.84375,
"train/R_reason": 0.21139857471142573,
"train/R_vocab": 0.28266369047619044
},
{
"learning_rate": 4.989656378295704e-07,
"loss": 0.0554,
"step": 88,
"train/R_acc": 0.00625,
"train/R_penalty": -0.5,
"train/R_reason": 0.2142568618009861,
"train/R_vocab": 0.3390625
},
{
"learning_rate": 4.988648580445895e-07,
"loss": 0.0648,
"step": 90,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.17382571925607848,
"train/R_vocab": 0.1894345238095238
},
{
"learning_rate": 4.987594053581012e-07,
"loss": -0.0241,
"step": 92,
"train/R_acc": 0.003125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.21781325533595836,
"train/R_vocab": 0.33205605158730156
},
{
"learning_rate": 4.98649281750176e-07,
"loss": 0.1022,
"step": 94,
"train/R_acc": 0.003125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.1999237909283059,
"train/R_vocab": 0.32064732142857144
},
{
"learning_rate": 4.985344892885899e-07,
"loss": 0.0821,
"step": 96,
"train/R_acc": 0.00625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.23594255343132425,
"train/R_vocab": 0.30199652777777775
},
{
"learning_rate": 4.984150301287853e-07,
"loss": 0.0096,
"step": 98,
"train/R_acc": 0.0,
"train/R_penalty": -0.80625,
"train/R_reason": 0.1900727268525148,
"train/R_vocab": 0.20625000000000002
},
{
"learning_rate": 4.982909065138305e-07,
"loss": 0.0247,
"step": 100,
"train/R_acc": 0.003125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.21411060850160504,
"train/R_vocab": 0.22946428571428573
},
{
"eval/R_acc": 0.002777777777777778,
"eval/R_penalty": -0.6847222222222222,
"eval/R_reason": 0.2572056199703138,
"eval/R_vocab": 0.3255787037037037,
"step": 100
},
{
"step": 100
},
{
"learning_rate": 4.981621207743778e-07,
"loss": -0.0004,
"step": 102,
"train/R_acc": 0.0,
"train/R_penalty": -0.765625,
"train/R_reason": 0.1735908404636029,
"train/R_vocab": 0.15747767857142858
},
{
"learning_rate": 4.980286753286194e-07,
"loss": 0.0155,
"step": 104,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.2016416402158515,
"train/R_vocab": 0.3698660714285714
},
{
"learning_rate": 4.978905726822423e-07,
"loss": -0.0257,
"step": 106,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.2019264344648219,
"train/R_vocab": 0.2499627976190476
},
{
"learning_rate": 4.97747815428381e-07,
"loss": 0.0757,
"step": 108,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.2609304111036691,
"train/R_vocab": 0.22472718253968255
},
{
"learning_rate": 4.976004062475691e-07,
"loss": -0.0376,
"step": 110,
"train/R_acc": 0.03125,
"train/R_penalty": -0.55625,
"train/R_reason": 0.24456802402586483,
"train/R_vocab": 0.30167410714285714
},
{
"learning_rate": 4.974483479076885e-07,
"loss": -0.0155,
"step": 112,
"train/R_acc": 0.0,
"train/R_penalty": -0.734375,
"train/R_reason": 0.15732482982112658,
"train/R_vocab": 0.20078125
},
{
"learning_rate": 4.972916432639181e-07,
"loss": 0.0107,
"step": 114,
"train/R_acc": 0.0625,
"train/R_penalty": -0.6125,
"train/R_reason": 0.19719755508097864,
"train/R_vocab": 0.18292410714285715
},
{
"learning_rate": 4.971302952586796e-07,
"loss": -0.0536,
"step": 116,
"train/R_acc": 0.003125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.1615191195936923,
"train/R_vocab": 0.20714285714285713
},
{
"learning_rate": 4.969643069215823e-07,
"loss": 0.03,
"step": 118,
"train/R_acc": 0.003125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.2057090874466078,
"train/R_vocab": 0.26006944444444446
},
{
"learning_rate": 4.967936813693668e-07,
"loss": 0.0669,
"step": 120,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.17123245123013492,
"train/R_vocab": 0.22881944444444446
},
{
"learning_rate": 4.966184218058457e-07,
"loss": -0.01,
"step": 122,
"train/R_acc": 0.065625,
"train/R_penalty": -0.640625,
"train/R_reason": 0.25008522769707353,
"train/R_vocab": 0.2768725198412698
},
{
"learning_rate": 4.96438531521844e-07,
"loss": 0.0227,
"step": 124,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.228776542397296,
"train/R_vocab": 0.19776785714285716
},
{
"learning_rate": 4.962540138951371e-07,
"loss": 0.0342,
"step": 126,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.20827985353158485,
"train/R_vocab": 0.3024440836940837
},
{
"learning_rate": 4.960648723903872e-07,
"loss": 0.0418,
"step": 128,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.19681757054800564,
"train/R_vocab": 0.22455357142857144
},
{
"learning_rate": 4.958711105590785e-07,
"loss": -0.0196,
"step": 130,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.5625,
"train/R_reason": 0.22729280568853277,
"train/R_vocab": 0.19475446428571427
},
{
"learning_rate": 4.956727320394506e-07,
"loss": 0.0963,
"step": 132,
"train/R_acc": 0.003125,
"train/R_penalty": -0.84375,
"train/R_reason": 0.20146228335738753,
"train/R_vocab": 0.24609995039682542
},
{
"learning_rate": 4.954697405564299e-07,
"loss": 0.0259,
"step": 134,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.625,
"train/R_reason": 0.228208707691525,
"train/R_vocab": 0.23055217352092353
},
{
"learning_rate": 4.952621399215597e-07,
"loss": 0.0781,
"step": 136,
"train/R_acc": 0.003125,
"train/R_penalty": -0.8125,
"train/R_reason": 0.1811115376120021,
"train/R_vocab": 0.2921875
},
{
"learning_rate": 4.95049934032929e-07,
"loss": 0.0659,
"step": 138,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.2199008430371848,
"train/R_vocab": 0.22447916666666667
},
{
"learning_rate": 4.948331268750988e-07,
"loss": -0.0087,
"step": 140,
"train/R_acc": 0.03125,
"train/R_penalty": -0.796875,
"train/R_reason": 0.16414720324183496,
"train/R_vocab": 0.20345982142857144
},
{
"learning_rate": 4.946117225190273e-07,
"loss": 0.0109,
"step": 142,
"train/R_acc": 0.03125,
"train/R_penalty": -0.634375,
"train/R_reason": 0.20413731109417932,
"train/R_vocab": 0.27366071428571426
},
{
"learning_rate": 4.943857251219942e-07,
"loss": -0.0291,
"step": 144,
"train/R_acc": 0.0,
"train/R_penalty": -0.703125,
"train/R_reason": 0.2114743293926566,
"train/R_vocab": 0.24921875
},
{
"learning_rate": 4.941551389275217e-07,
"loss": -0.0249,
"step": 146,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.19512092302735518,
"train/R_vocab": 0.17879464285714286
},
{
"learning_rate": 4.939199682652952e-07,
"loss": -0.03,
"step": 148,
"train/R_acc": 0.003125,
"train/R_penalty": -0.6125,
"train/R_reason": 0.19263372463344114,
"train/R_vocab": 0.2714285714285714
},
{
"learning_rate": 4.936802175510823e-07,
"loss": -0.1025,
"step": 150,
"train/R_acc": 0.00625,
"train/R_penalty": -0.665625,
"train/R_reason": 0.21358490777949846,
"train/R_vocab": 0.3302455357142857
},
{
"learning_rate": 4.934358912866497e-07,
"loss": 0.0293,
"step": 152,
"train/R_acc": 0.00625,
"train/R_penalty": -0.484375,
"train/R_reason": 0.16662975968037924,
"train/R_vocab": 0.1699280753968254
},
{
"learning_rate": 4.931869940596779e-07,
"loss": 0.0662,
"step": 154,
"train/R_acc": 0.00625,
"train/R_penalty": -0.59375,
"train/R_reason": 0.20091746530920346,
"train/R_vocab": 0.27994791666666663
},
{
"learning_rate": 4.929335305436764e-07,
"loss": 0.0355,
"step": 156,
"train/R_acc": 0.003125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.22218261967350866,
"train/R_vocab": 0.3209821428571429
},
{
"learning_rate": 4.926755054978951e-07,
"loss": 0.0457,
"step": 158,
"train/R_acc": 0.0,
"train/R_penalty": -0.703125,
"train/R_reason": 0.22181716803364154,
"train/R_vocab": 0.29573863636363634
},
{
"learning_rate": 4.92412923767235e-07,
"loss": -0.0201,
"step": 160,
"train/R_acc": 0.00625,
"train/R_penalty": -0.625,
"train/R_reason": 0.24103408318415312,
"train/R_vocab": 0.23645833333333333
},
{
"learning_rate": 4.921457902821578e-07,
"loss": -0.0013,
"step": 162,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.21205287864979985,
"train/R_vocab": 0.24444444444444446
},
{
"learning_rate": 4.918741100585921e-07,
"loss": -0.0549,
"step": 164,
"train/R_acc": 0.0,
"train/R_penalty": -0.75,
"train/R_reason": 0.21139952333180925,
"train/R_vocab": 0.36577380952380956
},
{
"learning_rate": 4.915978881978406e-07,
"loss": 0.0116,
"step": 166,
"train/R_acc": 0.003125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.21990626286225867,
"train/R_vocab": 0.3359623015873016
},
{
"learning_rate": 4.913171298864836e-07,
"loss": -0.0194,
"step": 168,
"train/R_acc": 0.0,
"train/R_penalty": -0.6875,
"train/R_reason": 0.19859495503742927,
"train/R_vocab": 0.3138020833333333
},
{
"learning_rate": 4.910318403962813e-07,
"loss": 0.0306,
"step": 170,
"train/R_acc": 0.0,
"train/R_penalty": -0.75,
"train/R_reason": 0.15012936918136355,
"train/R_vocab": 0.2549107142857143
},
{
"learning_rate": 4.907420250840761e-07,
"loss": 0.0817,
"step": 172,
"train/R_acc": 0.03125,
"train/R_penalty": -0.68125,
"train/R_reason": 0.20995580823960475,
"train/R_vocab": 0.20602678571428573
},
{
"learning_rate": 4.9044768939169e-07,
"loss": -0.104,
"step": 174,
"train/R_acc": 0.03125,
"train/R_penalty": -0.625,
"train/R_reason": 0.18253049494341367,
"train/R_vocab": 0.20379464285714285
},
{
"learning_rate": 4.901488388458247e-07,
"loss": 0.0418,
"step": 176,
"train/R_acc": 0.0625,
"train/R_penalty": -0.5625,
"train/R_reason": 0.2305279249641788,
"train/R_vocab": 0.38442460317460314
},
{
"learning_rate": 4.898454790579558e-07,
"loss": -0.0033,
"step": 178,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.18641040374410697,
"train/R_vocab": 0.25188492063492063
},
{
"learning_rate": 4.895376157242288e-07,
"loss": 0.0036,
"step": 180,
"train/R_acc": 0.003125,
"train/R_penalty": -0.634375,
"train/R_reason": 0.19709747248451465,
"train/R_vocab": 0.28185763888888893
},
{
"learning_rate": 4.892252546253514e-07,
"loss": -0.0037,
"step": 182,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2449271246503006,
"train/R_vocab": 0.30446428571428574
},
{
"learning_rate": 4.889084016264858e-07,
"loss": 0.0279,
"step": 184,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.24078946289847405,
"train/R_vocab": 0.3054935515873016
},
{
"learning_rate": 4.88587062677137e-07,
"loss": 0.0824,
"step": 186,
"train/R_acc": 0.00625,
"train/R_penalty": -0.625,
"train/R_reason": 0.16765366722116085,
"train/R_vocab": 0.22786458333333331
},
{
"learning_rate": 4.882612438110429e-07,
"loss": 0.0193,
"step": 188,
"train/R_acc": 0.003125,
"train/R_penalty": -0.71875,
"train/R_reason": 0.21543693992639024,
"train/R_vocab": 0.3402901785714285
},
{
"learning_rate": 4.8793095114606e-07,
"loss": -0.0257,
"step": 190,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.154932479362758,
"train/R_vocab": 0.2091393849206349
},
{
"learning_rate": 4.875961908840485e-07,
"loss": -0.0201,
"step": 192,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.20005776839850115,
"train/R_vocab": 0.26316964285714284
},
{
"learning_rate": 4.872569693107563e-07,
"loss": 0.0163,
"step": 194,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.2028767825912614,
"train/R_vocab": 0.22935267857142855
},
{
"learning_rate": 4.869132927957006e-07,
"loss": -0.0312,
"step": 196,
"train/R_acc": 0.0625,
"train/R_penalty": -0.6875,
"train/R_reason": 0.20598954317804136,
"train/R_vocab": 0.2894097222222222
},
{
"learning_rate": 4.865651677920483e-07,
"loss": -0.041,
"step": 198,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5625,
"train/R_reason": 0.19447689115700523,
"train/R_vocab": 0.21484375
},
{
"learning_rate": 4.862126008364954e-07,
"loss": -0.0021,
"step": 200,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.24708335127072298,
"train/R_vocab": 0.3235491071428571
},
{
"eval/R_acc": 0.001388888888888889,
"eval/R_penalty": -0.6875,
"eval/R_reason": 0.26545862362041284,
"eval/R_vocab": 0.3180500440917107,
"step": 200
},
{
"step": 200
},
{
"learning_rate": 4.858555985491434e-07,
"loss": 0.0539,
"step": 202,
"train/R_acc": 0.003125,
"train/R_penalty": -0.71875,
"train/R_reason": 0.12246968403633852,
"train/R_vocab": 0.12857142857142856
},
{
"learning_rate": 4.854941676333756e-07,
"loss": -0.0105,
"step": 204,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.59375,
"train/R_reason": 0.20647794674779765,
"train/R_vocab": 0.2294642857142857
},
{
"learning_rate": 4.85128314875731e-07,
"loss": 0.0418,
"step": 206,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.22341563734148095,
"train/R_vocab": 0.24494047619047618
},
{
"learning_rate": 4.84758047145777e-07,
"loss": 0.0338,
"step": 208,
"train/R_acc": 0.0,
"train/R_penalty": -0.55625,
"train/R_reason": 0.21091575917752947,
"train/R_vocab": 0.25234375
},
{
"learning_rate": 4.843833713959802e-07,
"loss": 0.0127,
"step": 210,
"train/R_acc": 0.003125,
"train/R_penalty": -0.78125,
"train/R_reason": 0.15006747203000467,
"train/R_vocab": 0.23049355158730156
},
{
"learning_rate": 4.840042946615761e-07,
"loss": 0.0429,
"step": 212,
"train/R_acc": 0.003125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.23417842297662064,
"train/R_vocab": 0.28020833333333334
},
{
"learning_rate": 4.836208240604368e-07,
"loss": -0.0,
"step": 214,
"train/R_acc": 0.0,
"train/R_penalty": -0.65,
"train/R_reason": 0.20300438739011717,
"train/R_vocab": 0.30078125
},
{
"learning_rate": 4.832329667929376e-07,
"loss": 0.0223,
"step": 216,
"train/R_acc": 0.0,
"train/R_penalty": -0.696875,
"train/R_reason": 0.21329021924994432,
"train/R_vocab": 0.2752604166666667
},
{
"learning_rate": 4.828407301418217e-07,
"loss": -0.0077,
"step": 218,
"train/R_acc": 0.0,
"train/R_penalty": -0.75,
"train/R_reason": 0.22684321233993932,
"train/R_vocab": 0.2927331349206349
},
{
"learning_rate": 4.824441214720628e-07,
"loss": 0.0196,
"step": 220,
"train/R_acc": 0.003125,
"train/R_penalty": -0.68125,
"train/R_reason": 0.22135178051709037,
"train/R_vocab": 0.3354166666666667
},
{
"learning_rate": 4.820431482307281e-07,
"loss": 0.0335,
"step": 222,
"train/R_acc": 0.034375,
"train/R_penalty": -0.609375,
"train/R_reason": 0.23161254238918147,
"train/R_vocab": 0.3784598214285714
},
{
"learning_rate": 4.816378179468374e-07,
"loss": 0.0416,
"step": 224,
"train/R_acc": 0.003125,
"train/R_penalty": -0.728125,
"train/R_reason": 0.20212408310335858,
"train/R_vocab": 0.2705357142857143
},
{
"learning_rate": 4.812281382312223e-07,
"loss": 0.0156,
"step": 226,
"train/R_acc": 0.03125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.19926301994985757,
"train/R_vocab": 0.35390625000000003
},
{
"learning_rate": 4.808141167763826e-07,
"loss": 0.0111,
"step": 228,
"train/R_acc": 0.003125,
"train/R_penalty": -0.484375,
"train/R_reason": 0.24773363470223553,
"train/R_vocab": 0.3498015873015873
},
{
"learning_rate": 4.803957613563431e-07,
"loss": 0.1022,
"step": 230,
"train/R_acc": 0.00625,
"train/R_penalty": -0.609375,
"train/R_reason": 0.19832487010930644,
"train/R_vocab": 0.2538690476190476
},
{
"learning_rate": 4.799730798265063e-07,
"loss": 0.1914,
"step": 232,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.22378486607186504,
"train/R_vocab": 0.28828125000000004
},
{
"learning_rate": 4.795460801235058e-07,
"loss": -0.036,
"step": 234,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.17550109595621496,
"train/R_vocab": 0.24095982142857145
},
{
"learning_rate": 4.791147702650565e-07,
"loss": 0.1009,
"step": 236,
"train/R_acc": 0.034375,
"train/R_penalty": -0.61875,
"train/R_reason": 0.2616533979112148,
"train/R_vocab": 0.3763888888888889
},
{
"learning_rate": 4.786791583498051e-07,
"loss": -0.0144,
"step": 238,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.20754413577165548,
"train/R_vocab": 0.22721974206349208
},
{
"learning_rate": 4.78239252557177e-07,
"loss": 0.0564,
"step": 240,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2701327994702766,
"train/R_vocab": 0.3259424603174603
},
{
"learning_rate": 4.777950611472233e-07,
"loss": -0.0099,
"step": 242,
"train/R_acc": 0.0,
"train/R_penalty": -0.703125,
"train/R_reason": 0.21945774918879585,
"train/R_vocab": 0.2623883928571429
},
{
"learning_rate": 4.773465924604656e-07,
"loss": 0.0141,
"step": 244,
"train/R_acc": 0.065625,
"train/R_penalty": -0.671875,
"train/R_reason": 0.2201195420166109,
"train/R_vocab": 0.26729910714285715
},
{
"learning_rate": 4.768938549177392e-07,
"loss": 0.052,
"step": 246,
"train/R_acc": 0.0,
"train/R_penalty": -0.6875,
"train/R_reason": 0.2231016124642528,
"train/R_vocab": 0.26313244047619044
},
{
"learning_rate": 4.764368570200353e-07,
"loss": 0.0239,
"step": 248,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5625,
"train/R_reason": 0.12548248807792622,
"train/R_vocab": 0.15256696428571428
},
{
"learning_rate": 4.759756073483411e-07,
"loss": 0.0264,
"step": 250,
"train/R_acc": 0.0,
"train/R_penalty": -0.765625,
"train/R_reason": 0.17495784675781162,
"train/R_vocab": 0.13359375
},
{
"learning_rate": 4.7551011456347876e-07,
"loss": 0.0271,
"step": 252,
"train/R_acc": 0.065625,
"train/R_penalty": -0.759375,
"train/R_reason": 0.2455444828427244,
"train/R_vocab": 0.3765625
},
{
"learning_rate": 4.750403874059428e-07,
"loss": 0.035,
"step": 254,
"train/R_acc": 0.003125,
"train/R_penalty": -0.61875,
"train/R_reason": 0.2624393029582254,
"train/R_vocab": 0.3123883928571428
},
{
"learning_rate": 4.745664346957361e-07,
"loss": -0.0167,
"step": 256,
"train/R_acc": 0.0,
"train/R_penalty": -0.734375,
"train/R_reason": 0.2061035486282405,
"train/R_vocab": 0.2132688492063492
},
{
"learning_rate": 4.740882653322039e-07,
"loss": 0.0071,
"step": 258,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.25927699094783563,
"train/R_vocab": 0.3599330357142857
},
{
"learning_rate": 4.7360588829386736e-07,
"loss": -0.0231,
"step": 260,
"train/R_acc": 0.0,
"train/R_penalty": -0.7125,
"train/R_reason": 0.2040610708874982,
"train/R_vocab": 0.23035714285714287
},
{
"learning_rate": 4.7311931263825434e-07,
"loss": 0.0593,
"step": 262,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.14067559915832142,
"train/R_vocab": 0.171875
},
{
"learning_rate": 4.726285475017294e-07,
"loss": 0.0213,
"step": 264,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.2728806346759708,
"train/R_vocab": 0.34206349206349207
},
{
"learning_rate": 4.721336020993228e-07,
"loss": -0.0176,
"step": 266,
"train/R_acc": 0.003125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.22780919072556438,
"train/R_vocab": 0.26149553571428574
},
{
"learning_rate": 4.716344857245567e-07,
"loss": 0.0807,
"step": 268,
"train/R_acc": 0.0,
"train/R_penalty": -0.61875,
"train/R_reason": 0.23184996778054867,
"train/R_vocab": 0.27254464285714286
},
{
"learning_rate": 4.7113120774927136e-07,
"loss": 0.0114,
"step": 270,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.24027473582911105,
"train/R_vocab": 0.41076388888888893
},
{
"learning_rate": 4.706237776234486e-07,
"loss": 0.0187,
"step": 272,
"train/R_acc": 0.03125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.1528155999223981,
"train/R_vocab": 0.1519097222222222
},
{
"learning_rate": 4.7011220487503476e-07,
"loss": 0.0688,
"step": 274,
"train/R_acc": 0.034375,
"train/R_penalty": -0.609375,
"train/R_reason": 0.15041989042686232,
"train/R_vocab": 0.1591889880952381
},
{
"learning_rate": 4.695964991097616e-07,
"loss": 0.0498,
"step": 276,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.20384797152303086,
"train/R_vocab": 0.35360863095238093
},
{
"learning_rate": 4.6907667001096585e-07,
"loss": -0.0992,
"step": 278,
"train/R_acc": 0.00625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.23776322399396135,
"train/R_vocab": 0.4219246031746032
},
{
"learning_rate": 4.685527273394078e-07,
"loss": -0.0568,
"step": 280,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5625,
"train/R_reason": 0.2305205389308027,
"train/R_vocab": 0.21640624999999997
},
{
"learning_rate": 4.680246809330874e-07,
"loss": -0.0174,
"step": 282,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.27570159502906066,
"train/R_vocab": 0.23663194444444444
},
{
"learning_rate": 4.6749254070706013e-07,
"loss": -0.053,
"step": 284,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.17788848371405203,
"train/R_vocab": 0.159375
},
{
"learning_rate": 4.669563166532503e-07,
"loss": 0.0595,
"step": 286,
"train/R_acc": 0.03125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.22202849960365972,
"train/R_vocab": 0.2922619047619048
},
{
"learning_rate": 4.6641601884026407e-07,
"loss": -0.0247,
"step": 288,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.20278024644809003,
"train/R_vocab": 0.2072544642857143
},
{
"learning_rate": 4.6587165741319967e-07,
"loss": 0.0055,
"step": 290,
"train/R_acc": 0.0,
"train/R_penalty": -0.71875,
"train/R_reason": 0.20358950838718215,
"train/R_vocab": 0.30781250000000004
},
{
"learning_rate": 4.6532324259345743e-07,
"loss": 0.013,
"step": 292,
"train/R_acc": 0.0,
"train/R_penalty": -0.68125,
"train/R_reason": 0.19116956517383865,
"train/R_vocab": 0.19854910714285712
},
{
"learning_rate": 4.647707846785477e-07,
"loss": -0.0187,
"step": 294,
"train/R_acc": 0.00625,
"train/R_penalty": -0.71875,
"train/R_reason": 0.18667014063523668,
"train/R_vocab": 0.27327178030303034
},
{
"learning_rate": 4.642142940418973e-07,
"loss": 0.0376,
"step": 296,
"train/R_acc": 0.034375,
"train/R_penalty": -0.546875,
"train/R_reason": 0.193981736932492,
"train/R_vocab": 0.19051339285714286
},
{
"learning_rate": 4.6365378113265505e-07,
"loss": -0.0587,
"step": 298,
"train/R_acc": 0.003125,
"train/R_penalty": -0.76875,
"train/R_reason": 0.21676354110088458,
"train/R_vocab": 0.2875
},
{
"learning_rate": 4.630892564754956e-07,
"loss": 0.0012,
"step": 300,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.24459838346088703,
"train/R_vocab": 0.33515625
},
{
"eval/R_acc": 0.001388888888888889,
"eval/R_penalty": -0.6499999999999999,
"eval/R_reason": 0.2340619843391708,
"eval/R_vocab": 0.2772872574955908,
"step": 300
},
{
"step": 300
},
{
"learning_rate": 4.6252073067042127e-07,
"loss": -0.0006,
"step": 302,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.20949847516693765,
"train/R_vocab": 0.24609375
},
{
"learning_rate": 4.6194821439256373e-07,
"loss": 0.0001,
"step": 304,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.20496328530187652,
"train/R_vocab": 0.20011160714285714
},
{
"learning_rate": 4.6137171839198297e-07,
"loss": 0.0792,
"step": 306,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.1994046532974053,
"train/R_vocab": 0.2868551587301587
},
{
"learning_rate": 4.6079125349346576e-07,
"loss": 0.0068,
"step": 308,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.18547339942836427,
"train/R_vocab": 0.2426711309523809
},
{
"learning_rate": 4.602068305963224e-07,
"loss": 0.0044,
"step": 310,
"train/R_acc": 0.0,
"train/R_penalty": -0.71875,
"train/R_reason": 0.2112060437994439,
"train/R_vocab": 0.2083829365079365
},
{
"learning_rate": 4.59618460674182e-07,
"loss": 0.0279,
"step": 312,
"train/R_acc": 0.034375,
"train/R_penalty": -0.53125,
"train/R_reason": 0.272254341125526,
"train/R_vocab": 0.45736607142857144
},
{
"learning_rate": 4.5902615477478636e-07,
"loss": -0.0181,
"step": 314,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.1449225433066495,
"train/R_vocab": 0.16674107142857142
},
{
"learning_rate": 4.5842992401978256e-07,
"loss": 0.0225,
"step": 316,
"train/R_acc": 0.0,
"train/R_penalty": -0.828125,
"train/R_reason": 0.2286200296039895,
"train/R_vocab": 0.2551339285714286
},
{
"learning_rate": 4.5782977960451414e-07,
"loss": 0.0192,
"step": 318,
"train/R_acc": 0.003125,
"train/R_penalty": -0.68125,
"train/R_reason": 0.21916647791965457,
"train/R_vocab": 0.2505580357142857
},
{
"learning_rate": 4.57225732797811e-07,
"loss": 0.0485,
"step": 320,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.22755603236630983,
"train/R_vocab": 0.22947668650793654
},
{
"learning_rate": 4.566177949417777e-07,
"loss": 0.0163,
"step": 322,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.22636069312789864,
"train/R_vocab": 0.21104910714285716
},
{
"learning_rate": 4.560059774515804e-07,
"loss": 0.0143,
"step": 324,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.15769657140460813,
"train/R_vocab": 0.22265625000000003
},
{
"learning_rate": 4.5539029181523284e-07,
"loss": -0.0796,
"step": 326,
"train/R_acc": 0.03125,
"train/R_penalty": -0.75,
"train/R_reason": 0.1961929211100274,
"train/R_vocab": 0.27142857142857146
},
{
"learning_rate": 4.5477074959338015e-07,
"loss": 0.0627,
"step": 328,
"train/R_acc": 0.003125,
"train/R_penalty": -0.75,
"train/R_reason": 0.22035420658260274,
"train/R_vocab": 0.33661954365079366
},
{
"learning_rate": 4.5414736241908214e-07,
"loss": -0.015,
"step": 330,
"train/R_acc": 0.0,
"train/R_penalty": -0.703125,
"train/R_reason": 0.21972588070640325,
"train/R_vocab": 0.30012400793650795
},
{
"learning_rate": 4.535201419975948e-07,
"loss": 0.0515,
"step": 332,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.46875,
"train/R_reason": 0.17121075988051718,
"train/R_vocab": 0.234375
},
{
"learning_rate": 4.5288910010615053e-07,
"loss": 0.0479,
"step": 334,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.5625,
"train/R_reason": 0.21868874143295755,
"train/R_vocab": 0.2630580357142857
},
{
"learning_rate": 4.5225424859373684e-07,
"loss": 0.0442,
"step": 336,
"train/R_acc": 0.00625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.22348722064953433,
"train/R_vocab": 0.3254464285714286
},
{
"learning_rate": 4.51615599380874e-07,
"loss": 0.0299,
"step": 338,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.18059697532933602,
"train/R_vocab": 0.21510416666666665
},
{
"learning_rate": 4.5097316445939124e-07,
"loss": 0.0359,
"step": 340,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.1941301451022681,
"train/R_vocab": 0.32220982142857146
},
{
"learning_rate": 4.503269558922015e-07,
"loss": -0.0107,
"step": 342,
"train/R_acc": 0.034375,
"train/R_penalty": -0.625,
"train/R_reason": 0.21181567459611045,
"train/R_vocab": 0.2482514880952381
},
{
"learning_rate": 4.496769858130748e-07,
"loss": 0.0179,
"step": 344,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.23230602042254933,
"train/R_vocab": 0.3271701388888889
},
{
"learning_rate": 4.490232664264109e-07,
"loss": 0.042,
"step": 346,
"train/R_acc": 0.03125,
"train/R_penalty": -0.696875,
"train/R_reason": 0.2775747137079635,
"train/R_vocab": 0.3890625
},
{
"learning_rate": 4.4836581000700944e-07,
"loss": 0.0274,
"step": 348,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.22799018822395078,
"train/R_vocab": 0.27528521825396823
},
{
"learning_rate": 4.477046288998401e-07,
"loss": 0.0259,
"step": 350,
"train/R_acc": 0.13125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.20698908553226353,
"train/R_vocab": 0.29140625
},
{
"learning_rate": 4.470397355198102e-07,
"loss": 0.0362,
"step": 352,
"train/R_acc": 0.0125,
"train/R_penalty": -0.525,
"train/R_reason": 0.210767134421379,
"train/R_vocab": 0.3446180555555556
},
{
"learning_rate": 4.463711423515323e-07,
"loss": 0.0077,
"step": 354,
"train/R_acc": 0.0,
"train/R_penalty": -0.796875,
"train/R_reason": 0.16874434359341295,
"train/R_vocab": 0.20223214285714286
},
{
"learning_rate": 4.456988619490889e-07,
"loss": 0.0364,
"step": 356,
"train/R_acc": 0.00625,
"train/R_penalty": -0.665625,
"train/R_reason": 0.2266184174999622,
"train/R_vocab": 0.32075892857142857
},
{
"learning_rate": 4.450229069357977e-07,
"loss": 0.0624,
"step": 358,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.19035143640898555,
"train/R_vocab": 0.1884548611111111
},
{
"learning_rate": 4.4434329000397363e-07,
"loss": -0.0017,
"step": 360,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.18822223579328265,
"train/R_vocab": 0.2318452380952381
},
{
"learning_rate": 4.4366002391469126e-07,
"loss": 0.0383,
"step": 362,
"train/R_acc": 0.034375,
"train/R_penalty": -0.71875,
"train/R_reason": 0.18399620476664255,
"train/R_vocab": 0.17989831349206348
},
{
"learning_rate": 4.4297312149754477e-07,
"loss": 0.0709,
"step": 364,
"train/R_acc": 0.03125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.2374170544763758,
"train/R_vocab": 0.2957589285714286
},
{
"learning_rate": 4.422825956504072e-07,
"loss": 0.012,
"step": 366,
"train/R_acc": 0.034375,
"train/R_penalty": -0.775,
"train/R_reason": 0.18171410574759275,
"train/R_vocab": 0.27645089285714286
},
{
"learning_rate": 4.415884593391882e-07,
"loss": -0.0291,
"step": 368,
"train/R_acc": 0.0,
"train/R_penalty": -0.728125,
"train/R_reason": 0.20255323919085794,
"train/R_vocab": 0.3200396825396825
},
{
"learning_rate": 4.4089072559759065e-07,
"loss": 0.0228,
"step": 370,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.25184199643437555,
"train/R_vocab": 0.31328125
},
{
"learning_rate": 4.40189407526866e-07,
"loss": -0.022,
"step": 372,
"train/R_acc": 0.03125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.21868958770018052,
"train/R_vocab": 0.21980406746031747
},
{
"learning_rate": 4.3948451829556775e-07,
"loss": 0.0437,
"step": 374,
"train/R_acc": 0.0,
"train/R_penalty": -0.603125,
"train/R_reason": 0.23191700315429237,
"train/R_vocab": 0.25416666666666665
},
{
"learning_rate": 4.3877607113930516e-07,
"loss": -0.0032,
"step": 376,
"train/R_acc": 0.0,
"train/R_penalty": -0.5,
"train/R_reason": 0.1725605699858892,
"train/R_vocab": 0.23932291666666666
},
{
"learning_rate": 4.380640793604938e-07,
"loss": -0.0522,
"step": 378,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.1570635238253007,
"train/R_vocab": 0.209375
},
{
"learning_rate": 4.373485563281062e-07,
"loss": 0.0522,
"step": 380,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.25503011482100246,
"train/R_vocab": 0.2538690476190476
},
{
"learning_rate": 4.3662951547742075e-07,
"loss": 0.0447,
"step": 382,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.1739243830311305,
"train/R_vocab": 0.23214285714285715
},
{
"learning_rate": 4.3590697030976965e-07,
"loss": -0.006,
"step": 384,
"train/R_acc": 0.03125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.15723816285274234,
"train/R_vocab": 0.2829996392496392
},
{
"learning_rate": 4.3518093439228474e-07,
"loss": 0.0162,
"step": 386,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.16239281076258544,
"train/R_vocab": 0.1482142857142857
},
{
"learning_rate": 4.3445142135764367e-07,
"loss": -0.0038,
"step": 388,
"train/R_acc": 0.034375,
"train/R_penalty": -0.6875,
"train/R_reason": 0.19918031683009202,
"train/R_vocab": 0.27239583333333334
},
{
"learning_rate": 4.33718444903813e-07,
"loss": 0.0341,
"step": 390,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.20130303529176516,
"train/R_vocab": 0.29593253968253963
},
{
"learning_rate": 4.329820187937919e-07,
"loss": 0.0876,
"step": 392,
"train/R_acc": 0.034375,
"train/R_penalty": -0.5,
"train/R_reason": 0.23979456377125202,
"train/R_vocab": 0.24832589285714285
},
{
"learning_rate": 4.3224215685535287e-07,
"loss": 0.0139,
"step": 394,
"train/R_acc": 0.0,
"train/R_penalty": -0.8125,
"train/R_reason": 0.1709315188515677,
"train/R_vocab": 0.18359375
},
{
"learning_rate": 4.314988729807827e-07,
"loss": -0.021,
"step": 396,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.17646043188459717,
"train/R_vocab": 0.17589285714285713
},
{
"learning_rate": 4.3075218112662135e-07,
"loss": 0.0513,
"step": 398,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.1973938417696664,
"train/R_vocab": 0.19837549603174603
},
{
"learning_rate": 4.3000209531339996e-07,
"loss": 0.0043,
"step": 400,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.24184598292042236,
"train/R_vocab": 0.33288690476190474
},
{
"eval/R_acc": 0.0,
"eval/R_penalty": -0.7916666666666666,
"eval/R_reason": 0.27821714712822276,
"eval/R_vocab": 0.32563381834215166,
"step": 400
},
{
"step": 400
},
{
"learning_rate": 4.2943367562278896e-07,
"loss": -0.0155,
"step": 402,
"train/R_acc": 0.065625,
"train/R_penalty": -0.515625,
"train/R_reason": 0.20335033205254746,
"train/R_vocab": 0.24464285714285713
},
{
"learning_rate": 4.286787117443108e-07,
"loss": -0.036,
"step": 404,
"train/R_acc": 0.04062500000000001,
"train/R_penalty": -0.5625,
"train/R_reason": 0.20780436238210176,
"train/R_vocab": 0.2703993055555556
},
{
"learning_rate": 4.2792040207614e-07,
"loss": -0.0044,
"step": 406,
"train/R_acc": 0.03125,
"train/R_penalty": -0.734375,
"train/R_reason": 0.2366017128148703,
"train/R_vocab": 0.3632440476190476
},
{
"learning_rate": 4.27158760817756e-07,
"loss": 0.0245,
"step": 408,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.19598878934576208,
"train/R_vocab": 0.28031994047619047
},
{
"learning_rate": 4.263938022310226e-07,
"loss": -0.036,
"step": 410,
"train/R_acc": 0.034375,
"train/R_penalty": -0.671875,
"train/R_reason": 0.23507477335166782,
"train/R_vocab": 0.27189980158730165
},
{
"learning_rate": 4.2562554063992127e-07,
"loss": 0.024,
"step": 412,
"train/R_acc": 0.0625,
"train/R_penalty": -0.7125,
"train/R_reason": 0.19770138255909644,
"train/R_vocab": 0.24754464285714284
},
{
"learning_rate": 4.248539904302829e-07,
"loss": 0.0091,
"step": 414,
"train/R_acc": 0.00625,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2098416268444374,
"train/R_vocab": 0.2533234126984127
},
{
"learning_rate": 4.240791660495182e-07,
"loss": 0.0125,
"step": 416,
"train/R_acc": 0.003125,
"train/R_penalty": -0.603125,
"train/R_reason": 0.164986564236729,
"train/R_vocab": 0.1863219246031746
},
{
"learning_rate": 4.2330108200634723e-07,
"loss": 0.0353,
"step": 418,
"train/R_acc": 0.03125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.16427460147948922,
"train/R_vocab": 0.1800595238095238
},
{
"learning_rate": 4.22519752870528e-07,
"loss": -0.0002,
"step": 420,
"train/R_acc": 0.003125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.21440123784001752,
"train/R_vocab": 0.26026785714285716
},
{
"learning_rate": 4.2173519327258325e-07,
"loss": 0.0336,
"step": 422,
"train/R_acc": 0.003125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2248050023022144,
"train/R_vocab": 0.22937184343434344
},
{
"learning_rate": 4.2094741790352673e-07,
"loss": 0.0256,
"step": 424,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.22803044267397005,
"train/R_vocab": 0.28350694444444446
},
{
"learning_rate": 4.2015644151458827e-07,
"loss": -0.0319,
"step": 426,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.20821167099100335,
"train/R_vocab": 0.21744791666666666
},
{
"learning_rate": 4.19362278916937e-07,
"loss": 0.0344,
"step": 428,
"train/R_acc": 0.03125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.1891472011902095,
"train/R_vocab": 0.20641233766233769
},
{
"learning_rate": 4.185649449814045e-07,
"loss": 0.0712,
"step": 430,
"train/R_acc": 0.00625,
"train/R_penalty": -0.6875,
"train/R_reason": 0.2279409653415437,
"train/R_vocab": 0.2996651785714286
},
{
"learning_rate": 4.177644546382063e-07,
"loss": 0.0645,
"step": 432,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.625,
"train/R_reason": 0.21120852841137872,
"train/R_vocab": 0.19676339285714284
},
{
"learning_rate": 4.1696082287666217e-07,
"loss": 0.0481,
"step": 434,
"train/R_acc": 0.003125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.20179254286468842,
"train/R_vocab": 0.3659598214285714
},
{
"learning_rate": 4.1615406474491535e-07,
"loss": -0.0453,
"step": 436,
"train/R_acc": 0.034375,
"train/R_penalty": -0.78125,
"train/R_reason": 0.21685467943636602,
"train/R_vocab": 0.2571428571428571
},
{
"learning_rate": 4.1534419534965105e-07,
"loss": 0.0985,
"step": 438,
"train/R_acc": 0.03125,
"train/R_penalty": -0.696875,
"train/R_reason": 0.20741733073153354,
"train/R_vocab": 0.27142857142857146
},
{
"learning_rate": 4.145312298558133e-07,
"loss": -0.0621,
"step": 440,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5625,
"train/R_reason": 0.17062418658581344,
"train/R_vocab": 0.2
},
{
"learning_rate": 4.137151834863213e-07,
"loss": -0.0448,
"step": 442,
"train/R_acc": 0.03125,
"train/R_penalty": -0.5,
"train/R_reason": 0.23174115282287933,
"train/R_vocab": 0.36713789682539677
},
{
"learning_rate": 4.128960715217839e-07,
"loss": 0.0187,
"step": 444,
"train/R_acc": 0.003125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.15167707696946453,
"train/R_vocab": 0.20052083333333334
},
{
"learning_rate": 4.1207390930021394e-07,
"loss": -0.0156,
"step": 446,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2000569888049995,
"train/R_vocab": 0.24027777777777778
},
{
"learning_rate": 4.11248712216741e-07,
"loss": 0.0897,
"step": 448,
"train/R_acc": 0.0,
"train/R_penalty": -0.68125,
"train/R_reason": 0.21749146764720112,
"train/R_vocab": 0.2636160714285714
},
{
"learning_rate": 4.104204957233225e-07,
"loss": 0.0032,
"step": 450,
"train/R_acc": 0.00625,
"train/R_penalty": -0.609375,
"train/R_reason": 0.19073873404612746,
"train/R_vocab": 0.30033482142857143
},
{
"learning_rate": 4.095892753284553e-07,
"loss": -0.0985,
"step": 452,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.25767563611959865,
"train/R_vocab": 0.41929563492063493
},
{
"learning_rate": 4.087550665968846e-07,
"loss": 0.0021,
"step": 454,
"train/R_acc": 0.0,
"train/R_penalty": -0.53125,
"train/R_reason": 0.1972933583803858,
"train/R_vocab": 0.23854166666666668
},
{
"learning_rate": 4.079178851493127e-07,
"loss": 0.0433,
"step": 456,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.2107089160600319,
"train/R_vocab": 0.25
},
{
"learning_rate": 4.070777466621067e-07,
"loss": -0.0103,
"step": 458,
"train/R_acc": 0.03125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.2429741695318347,
"train/R_vocab": 0.38958333333333334
},
{
"learning_rate": 4.062346668670046e-07,
"loss": -0.0092,
"step": 460,
"train/R_acc": 0.003125,
"train/R_penalty": -0.734375,
"train/R_reason": 0.2328364572562428,
"train/R_vocab": 0.2795758928571429
},
{
"learning_rate": 4.0538866155082094e-07,
"loss": -0.0198,
"step": 462,
"train/R_acc": 0.040625,
"train/R_penalty": -0.4625,
"train/R_reason": 0.2483761704003692,
"train/R_vocab": 0.2885168650793651
},
{
"learning_rate": 4.045397465551513e-07,
"loss": -0.0249,
"step": 464,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5875,
"train/R_reason": 0.17307827497369785,
"train/R_vocab": 0.2544642857142857
},
{
"learning_rate": 4.036879377760752e-07,
"loss": 0.0683,
"step": 466,
"train/R_acc": 0.03125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.22330642900293768,
"train/R_vocab": 0.37135416666666665
},
{
"learning_rate": 4.02833251163859e-07,
"loss": -0.0192,
"step": 468,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65,
"train/R_reason": 0.21129365536420686,
"train/R_vocab": 0.2589285714285714
},
{
"learning_rate": 4.0197570272265704e-07,
"loss": 0.0497,
"step": 470,
"train/R_acc": 0.003125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.2052435879372262,
"train/R_vocab": 0.2556547619047619
},
{
"learning_rate": 4.011153085102116e-07,
"loss": 0.0309,
"step": 472,
"train/R_acc": 0.003125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.21023509045617045,
"train/R_vocab": 0.2963169642857143
},
{
"learning_rate": 4.0025208463755274e-07,
"loss": 0.0424,
"step": 474,
"train/R_acc": 0.0,
"train/R_penalty": -0.784375,
"train/R_reason": 0.1959827825898999,
"train/R_vocab": 0.35357142857142854
},
{
"learning_rate": 3.9938604726869636e-07,
"loss": -0.0506,
"step": 476,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.19807364686139228,
"train/R_vocab": 0.17491319444444442
},
{
"learning_rate": 3.9851721262034157e-07,
"loss": 0.0724,
"step": 478,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5,
"train/R_reason": 0.20997218794964279,
"train/R_vocab": 0.27433035714285714
},
{
"learning_rate": 3.9764559696156697e-07,
"loss": 0.0338,
"step": 480,
"train/R_acc": 0.003125,
"train/R_penalty": -0.665625,
"train/R_reason": 0.21410868507940445,
"train/R_vocab": 0.25691964285714286
},
{
"learning_rate": 3.9677121661352607e-07,
"loss": -0.0087,
"step": 482,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.21506411602738199,
"train/R_vocab": 0.3130952380952381
},
{
"learning_rate": 3.958940879491418e-07,
"loss": -0.0375,
"step": 484,
"train/R_acc": 0.0,
"train/R_penalty": -0.71875,
"train/R_reason": 0.18895673088287218,
"train/R_vocab": 0.21395089285714286
},
{
"learning_rate": 3.9501422739279953e-07,
"loss": -0.0485,
"step": 486,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.20816202695918376,
"train/R_vocab": 0.22013888888888888
},
{
"learning_rate": 3.9413165142004e-07,
"loss": -0.0479,
"step": 488,
"train/R_acc": 0.03125,
"train/R_penalty": -0.5625,
"train/R_reason": 0.1929871719837118,
"train/R_vocab": 0.3703125
},
{
"learning_rate": 3.932463765572505e-07,
"loss": -0.0312,
"step": 490,
"train/R_acc": 0.0,
"train/R_penalty": -0.75,
"train/R_reason": 0.2513287746728735,
"train/R_vocab": 0.27569444444444446
},
{
"learning_rate": 3.923584193813555e-07,
"loss": 0.03,
"step": 492,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.23851168147176444,
"train/R_vocab": 0.313219246031746
},
{
"learning_rate": 3.914677965195062e-07,
"loss": -0.0009,
"step": 494,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.15729111795287057,
"train/R_vocab": 0.1830357142857143
},
{
"learning_rate": 3.9057452464876946e-07,
"loss": 0.049,
"step": 496,
"train/R_acc": 0.003125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.15483905838498802,
"train/R_vocab": 0.18125000000000002
},
{
"learning_rate": 3.89678620495815e-07,
"loss": 0.0521,
"step": 498,
"train/R_acc": 0.0,
"train/R_penalty": -0.75,
"train/R_reason": 0.2600686346082261,
"train/R_vocab": 0.253125
},
{
"learning_rate": 3.887801008366025e-07,
"loss": -0.005,
"step": 500,
"train/R_acc": 0.003125,
"train/R_penalty": -0.790625,
"train/R_reason": 0.22427641116364,
"train/R_vocab": 0.3156125992063492
},
{
"eval/R_acc": 0.0025,
"eval/R_penalty": -0.7075,
"eval/R_reason": 0.2610032668269429,
"eval/R_vocab": 0.21906295093795097,
"step": 500
},
{
"step": 500
},
{
"learning_rate": 3.8787898249606767e-07,
"loss": 0.0252,
"step": 502,
"train/R_acc": 0.0625,
"train/R_penalty": -0.609375,
"train/R_reason": 0.2684211960918422,
"train/R_vocab": 0.3091517857142857
},
{
"learning_rate": 3.8697528234780674e-07,
"loss": 0.019,
"step": 504,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.23017420694962337,
"train/R_vocab": 0.29288194444444443
},
{
"learning_rate": 3.86069017313761e-07,
"loss": 0.013,
"step": 506,
"train/R_acc": 0.03125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.22614899753752782,
"train/R_vocab": 0.3089905753968254
},
{
"learning_rate": 3.851602043638994e-07,
"loss": -0.0248,
"step": 508,
"train/R_acc": 0.034375,
"train/R_penalty": -0.515625,
"train/R_reason": 0.2017065165275432,
"train/R_vocab": 0.26004464285714285
},
{
"learning_rate": 3.8424886051590115e-07,
"loss": 0.0296,
"step": 510,
"train/R_acc": 0.0,
"train/R_penalty": -0.7375,
"train/R_reason": 0.17507956245908884,
"train/R_vocab": 0.17238343253968252
},
{
"learning_rate": 3.83335002834837e-07,
"loss": -0.0622,
"step": 512,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.22332916829962618,
"train/R_vocab": 0.34873511904761906
},
{
"learning_rate": 3.8241864843284964e-07,
"loss": -0.0075,
"step": 514,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.132802326682096,
"train/R_vocab": 0.16104910714285714
},
{
"learning_rate": 3.814998144688333e-07,
"loss": 0.0227,
"step": 516,
"train/R_acc": 0.034375,
"train/R_penalty": -0.59375,
"train/R_reason": 0.1754689583879656,
"train/R_vocab": 0.19609374999999998
},
{
"learning_rate": 3.805785181481123e-07,
"loss": -0.0118,
"step": 518,
"train/R_acc": 0.003125,
"train/R_penalty": -0.540625,
"train/R_reason": 0.23541827525810197,
"train/R_vocab": 0.25967261904761907
},
{
"learning_rate": 3.796547767221193e-07,
"loss": -0.0924,
"step": 520,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2376371992311745,
"train/R_vocab": 0.3128968253968254
},
{
"learning_rate": 3.787286074880718e-07,
"loss": -0.0122,
"step": 522,
"train/R_acc": 0.0,
"train/R_penalty": -0.55625,
"train/R_reason": 0.22978465999581174,
"train/R_vocab": 0.2617931547619048
},
{
"learning_rate": 3.778000277886483e-07,
"loss": -0.0054,
"step": 524,
"train/R_acc": 0.0,
"train/R_penalty": -0.53125,
"train/R_reason": 0.22061871520509158,
"train/R_vocab": 0.30894209956709956
},
{
"learning_rate": 3.768690550116639e-07,
"loss": 0.0092,
"step": 526,
"train/R_acc": 0.03125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.22469643175166784,
"train/R_vocab": 0.34073660714285714
},
{
"learning_rate": 3.7593570658974436e-07,
"loss": 0.0264,
"step": 528,
"train/R_acc": 0.03125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.26427900699025786,
"train/R_vocab": 0.32444196428571426
},
{
"learning_rate": 3.75e-07,
"loss": 0.0264,
"step": 530,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.16190914297300985,
"train/R_vocab": 0.18397817460317462
},
{
"learning_rate": 3.740619527636979e-07,
"loss": -0.0014,
"step": 532,
"train/R_acc": 0.003125,
"train/R_penalty": -0.540625,
"train/R_reason": 0.2013910995332763,
"train/R_vocab": 0.16294642857142855
},
{
"learning_rate": 3.731215824459344e-07,
"loss": 0.0653,
"step": 534,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.640625,
"train/R_reason": 0.1980956136240478,
"train/R_vocab": 0.26245039682539684
},
{
"learning_rate": 3.7217890665530587e-07,
"loss": -0.0052,
"step": 536,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.193118823374168,
"train/R_vocab": 0.18813131313131315
},
{
"learning_rate": 3.712339430435792e-07,
"loss": 0.044,
"step": 538,
"train/R_acc": 0.00625,
"train/R_penalty": -0.65625,
"train/R_reason": 0.21065931054504308,
"train/R_vocab": 0.2915054563492063
},
{
"learning_rate": 3.7028670930536077e-07,
"loss": 0.0443,
"step": 540,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.21326431514536787,
"train/R_vocab": 0.23958333333333334
},
{
"learning_rate": 3.693372231777658e-07,
"loss": 0.0062,
"step": 542,
"train/R_acc": 0.003125,
"train/R_penalty": -0.86875,
"train/R_reason": 0.2361538113041483,
"train/R_vocab": 0.26145833333333335
},
{
"learning_rate": 3.6838550244008573e-07,
"loss": 0.0011,
"step": 544,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.2560180166234486,
"train/R_vocab": 0.3419642857142857
},
{
"learning_rate": 3.6743156491345564e-07,
"loss": -0.1304,
"step": 546,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.22700866641314227,
"train/R_vocab": 0.3409474206349206
},
{
"learning_rate": 3.6647542846052003e-07,
"loss": 0.0374,
"step": 548,
"train/R_acc": 0.003125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.18355650415358743,
"train/R_vocab": 0.23628472222222222
},
{
"learning_rate": 3.65517110985099e-07,
"loss": -0.0193,
"step": 550,
"train/R_acc": 0.0625,
"train/R_penalty": -0.546875,
"train/R_reason": 0.1497471083396625,
"train/R_vocab": 0.1660342261904762
},
{
"learning_rate": 3.645566304318526e-07,
"loss": -0.0168,
"step": 552,
"train/R_acc": 0.00625,
"train/R_penalty": -0.71875,
"train/R_reason": 0.17760395766642206,
"train/R_vocab": 0.20212053571428573
},
{
"learning_rate": 3.6359400478594473e-07,
"loss": 0.0519,
"step": 554,
"train/R_acc": 0.03125,
"train/R_penalty": -0.68125,
"train/R_reason": 0.2032957807067698,
"train/R_vocab": 0.3547991071428571
},
{
"learning_rate": 3.6262925207270666e-07,
"loss": -0.0077,
"step": 556,
"train/R_acc": 0.034375,
"train/R_penalty": -0.53125,
"train/R_reason": 0.24608095001365107,
"train/R_vocab": 0.2532366071428571
},
{
"learning_rate": 3.616623903572994e-07,
"loss": -0.0273,
"step": 558,
"train/R_acc": 0.03125,
"train/R_penalty": -0.46875,
"train/R_reason": 0.25668514206671234,
"train/R_vocab": 0.32946428571428565
},
{
"learning_rate": 3.6069343774437516e-07,
"loss": 0.0356,
"step": 560,
"train/R_acc": 0.00625,
"train/R_penalty": -0.853125,
"train/R_reason": 0.22578641617093248,
"train/R_vocab": 0.3117063492063492
},
{
"learning_rate": 3.597224123777389e-07,
"loss": 0.0074,
"step": 562,
"train/R_acc": 0.00625,
"train/R_penalty": -0.61875,
"train/R_reason": 0.1608838946497146,
"train/R_vocab": 0.22940228174603175
},
{
"learning_rate": 3.58749332440008e-07,
"loss": 0.0218,
"step": 564,
"train/R_acc": 0.0,
"train/R_penalty": -0.5,
"train/R_reason": 0.2042184580919646,
"train/R_vocab": 0.23277529761904764
},
{
"learning_rate": 3.5777421615227207e-07,
"loss": 0.0446,
"step": 566,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.21492820594376136,
"train/R_vocab": 0.2902901785714286
},
{
"learning_rate": 3.567970817737518e-07,
"loss": 0.0771,
"step": 568,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.2497326839973281,
"train/R_vocab": 0.30859375
},
{
"learning_rate": 3.5581794760145696e-07,
"loss": 0.0354,
"step": 570,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.20181240026890324,
"train/R_vocab": 0.22098214285714285
},
{
"learning_rate": 3.548368319698437e-07,
"loss": 0.0731,
"step": 572,
"train/R_acc": 0.00625,
"train/R_penalty": -0.696875,
"train/R_reason": 0.16856543760413237,
"train/R_vocab": 0.1800595238095238
},
{
"learning_rate": 3.5385375325047163e-07,
"loss": 0.0291,
"step": 574,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2099860707148752,
"train/R_vocab": 0.2753348214285714
},
{
"learning_rate": 3.528687298516591e-07,
"loss": -0.0002,
"step": 576,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.17694118112711849,
"train/R_vocab": 0.12705853174603174
},
{
"learning_rate": 3.5188178021813925e-07,
"loss": 0.028,
"step": 578,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.609375,
"train/R_reason": 0.24354058691683123,
"train/R_vocab": 0.39140624999999996
},
{
"learning_rate": 3.5089292283071417e-07,
"loss": 0.0398,
"step": 580,
"train/R_acc": 0.00625,
"train/R_penalty": -0.53125,
"train/R_reason": 0.22343115261429752,
"train/R_vocab": 0.32678571428571423
},
{
"learning_rate": 3.499021762059089e-07,
"loss": 0.0565,
"step": 582,
"train/R_acc": 0.03125,
"train/R_penalty": -0.7125,
"train/R_reason": 0.19496747632032374,
"train/R_vocab": 0.26284722222222223
},
{
"learning_rate": 3.489095588956249e-07,
"loss": 0.0092,
"step": 584,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.2074570772291119,
"train/R_vocab": 0.30972222222222223
},
{
"learning_rate": 3.479150894867926e-07,
"loss": -0.0015,
"step": 586,
"train/R_acc": 0.0625,
"train/R_penalty": -0.703125,
"train/R_reason": 0.20443537938998835,
"train/R_vocab": 0.2761904761904762
},
{
"learning_rate": 3.46918786601023e-07,
"loss": 0.1059,
"step": 588,
"train/R_acc": 0.0,
"train/R_penalty": -0.68125,
"train/R_reason": 0.20239388628972388,
"train/R_vocab": 0.22630208333333335
},
{
"learning_rate": 3.459206688942596e-07,
"loss": 0.0608,
"step": 590,
"train/R_acc": 0.0,
"train/R_penalty": -0.75,
"train/R_reason": 0.24682579728602413,
"train/R_vocab": 0.4008928571428571
},
{
"learning_rate": 3.4492075505642847e-07,
"loss": 0.055,
"step": 592,
"train/R_acc": 0.0,
"train/R_penalty": -0.5,
"train/R_reason": 0.19593385729165314,
"train/R_vocab": 0.31875
},
{
"learning_rate": 3.439190638110888e-07,
"loss": 0.041,
"step": 594,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.1711366333418505,
"train/R_vocab": 0.26268601190476193
},
{
"learning_rate": 3.4291561391508185e-07,
"loss": 0.0807,
"step": 596,
"train/R_acc": 0.003125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.240933143584435,
"train/R_vocab": 0.4119791666666667
},
{
"learning_rate": 3.4191042415818e-07,
"loss": -0.0247,
"step": 598,
"train/R_acc": 0.003125,
"train/R_penalty": -0.74375,
"train/R_reason": 0.2539401489710603,
"train/R_vocab": 0.2673363095238095
},
{
"learning_rate": 3.4090351336273474e-07,
"loss": -0.0238,
"step": 600,
"train/R_acc": 0.0,
"train/R_penalty": -0.703125,
"train/R_reason": 0.19356852065501937,
"train/R_vocab": 0.24319196428571427
},
{
"eval/R_acc": 0.005,
"eval/R_penalty": -0.625,
"eval/R_reason": 0.24855003165355338,
"eval/R_vocab": 0.22628517316017316,
"step": 600
},
{
"step": 600
},
{
"learning_rate": 3.398949003833246e-07,
"loss": 0.0122,
"step": 602,
"train/R_acc": 0.00625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.22168416427879684,
"train/R_vocab": 0.24791666666666667
},
{
"learning_rate": 3.388846041064012e-07,
"loss": 0.0797,
"step": 604,
"train/R_acc": 0.09375,
"train/R_penalty": -0.546875,
"train/R_reason": 0.1777229928874741,
"train/R_vocab": 0.29464285714285715
},
{
"learning_rate": 3.378726434499368e-07,
"loss": 0.033,
"step": 606,
"train/R_acc": 0.0,
"train/R_penalty": -0.7125,
"train/R_reason": 0.1956222769670857,
"train/R_vocab": 0.3197916666666667
},
{
"learning_rate": 3.368590373630692e-07,
"loss": -0.0275,
"step": 608,
"train/R_acc": 0.06875,
"train/R_penalty": -0.46875,
"train/R_reason": 0.17503223220226105,
"train/R_vocab": 0.2591517857142857
},
{
"learning_rate": 3.3584380482574717e-07,
"loss": -0.058,
"step": 610,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.1902514010231657,
"train/R_vocab": 0.23125
},
{
"learning_rate": 3.348269648483749e-07,
"loss": -0.054,
"step": 612,
"train/R_acc": 0.0,
"train/R_penalty": -0.53125,
"train/R_reason": 0.17467640906171372,
"train/R_vocab": 0.190625
},
{
"learning_rate": 3.3380853647145656e-07,
"loss": -0.0371,
"step": 614,
"train/R_acc": 0.00625,
"train/R_penalty": -0.65625,
"train/R_reason": 0.17121010785732182,
"train/R_vocab": 0.2569444444444444
},
{
"learning_rate": 3.327885387652391e-07,
"loss": -0.0013,
"step": 616,
"train/R_acc": 0.0,
"train/R_penalty": -0.6875,
"train/R_reason": 0.19415455674882087,
"train/R_vocab": 0.23277529761904764
},
{
"learning_rate": 3.317669908293554e-07,
"loss": 0.0253,
"step": 618,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.24193135939319937,
"train/R_vocab": 0.25
},
{
"learning_rate": 3.307439117924668e-07,
"loss": 0.0588,
"step": 620,
"train/R_acc": 0.0,
"train/R_penalty": -0.6875,
"train/R_reason": 0.20416377806011923,
"train/R_vocab": 0.2056423611111111
},
{
"learning_rate": 3.297193208119047e-07,
"loss": -0.0064,
"step": 622,
"train/R_acc": 0.0,
"train/R_penalty": -0.484375,
"train/R_reason": 0.211192554787049,
"train/R_vocab": 0.30390625000000004
},
{
"learning_rate": 3.2869323707331176e-07,
"loss": 0.0335,
"step": 624,
"train/R_acc": 0.003125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.23403307094336184,
"train/R_vocab": 0.35301339285714284
},
{
"learning_rate": 3.2766567979028324e-07,
"loss": -0.0308,
"step": 626,
"train/R_acc": 0.00625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.25211879339329,
"train/R_vocab": 0.278125
},
{
"learning_rate": 3.2663666820400625e-07,
"loss": 0.0239,
"step": 628,
"train/R_acc": 0.03125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.23123528000410434,
"train/R_vocab": 0.328844246031746
},
{
"learning_rate": 3.2560622158290025e-07,
"loss": 0.0127,
"step": 630,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5,
"train/R_reason": 0.22103995585722944,
"train/R_vocab": 0.26785714285714285
},
{
"learning_rate": 3.2457435922225603e-07,
"loss": 0.0445,
"step": 632,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.185424618889319,
"train/R_vocab": 0.28928571428571426
},
{
"learning_rate": 3.235411004438741e-07,
"loss": 0.1006,
"step": 634,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.23390812732295468,
"train/R_vocab": 0.29355158730158726
},
{
"learning_rate": 3.2250646459570343e-07,
"loss": 0.0611,
"step": 636,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.17366692788791044,
"train/R_vocab": 0.21026785714285715
},
{
"learning_rate": 3.214704710514786e-07,
"loss": 0.0463,
"step": 638,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.23856693280209076,
"train/R_vocab": 0.3723214285714286
},
{
"learning_rate": 3.204331392103574e-07,
"loss": -0.0048,
"step": 640,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.2248622406893548,
"train/R_vocab": 0.38828125
},
{
"learning_rate": 3.193944884965576e-07,
"loss": 0.0172,
"step": 642,
"train/R_acc": 0.065625,
"train/R_penalty": -0.671875,
"train/R_reason": 0.2000315806348117,
"train/R_vocab": 0.32477678571428575
},
{
"learning_rate": 3.183545383589927e-07,
"loss": -0.0164,
"step": 644,
"train/R_acc": 0.003125,
"train/R_penalty": -0.55625,
"train/R_reason": 0.1901960813142597,
"train/R_vocab": 0.2540178571428572
},
{
"learning_rate": 3.173133082709086e-07,
"loss": 0.0815,
"step": 646,
"train/R_acc": 0.003125,
"train/R_penalty": -0.68125,
"train/R_reason": 0.20267428182016378,
"train/R_vocab": 0.2115079365079365
},
{
"learning_rate": 3.1627081772951815e-07,
"loss": 0.0279,
"step": 648,
"train/R_acc": 0.03125,
"train/R_penalty": -0.484375,
"train/R_reason": 0.18044257928711105,
"train/R_vocab": 0.21806795634920634
},
{
"learning_rate": 3.152270862556367e-07,
"loss": 0.0707,
"step": 650,
"train/R_acc": 0.003125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.26331484572063035,
"train/R_vocab": 0.24206349206349206
},
{
"learning_rate": 3.1418213339331576e-07,
"loss": -0.0396,
"step": 652,
"train/R_acc": 0.003125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.23778099952829604,
"train/R_vocab": 0.284375
},
{
"learning_rate": 3.1313597870947816e-07,
"loss": 0.0883,
"step": 654,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.22377350102735885,
"train/R_vocab": 0.3117931547619047
},
{
"learning_rate": 3.1208864179355074e-07,
"loss": 0.0009,
"step": 656,
"train/R_acc": 0.03125,
"train/R_penalty": -0.571875,
"train/R_reason": 0.20656730207307256,
"train/R_vocab": 0.2512276785714286
},
{
"learning_rate": 3.1104014225709784e-07,
"loss": 0.0828,
"step": 658,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65,
"train/R_reason": 0.2410817269424846,
"train/R_vocab": 0.265625
},
{
"learning_rate": 3.099904997334541e-07,
"loss": -0.0317,
"step": 660,
"train/R_acc": 0.03125,
"train/R_penalty": -0.5875,
"train/R_reason": 0.25759275591542546,
"train/R_vocab": 0.24017857142857144
},
{
"learning_rate": 3.0893973387735683e-07,
"loss": -0.0235,
"step": 662,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.16711049206280193,
"train/R_vocab": 0.21428571428571427
},
{
"learning_rate": 3.078878643645778e-07,
"loss": 0.0242,
"step": 664,
"train/R_acc": 0.00625,
"train/R_penalty": -0.696875,
"train/R_reason": 0.22549504519978125,
"train/R_vocab": 0.33912450396825394
},
{
"learning_rate": 3.068349108915553e-07,
"loss": 0.0147,
"step": 666,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2824812700031851,
"train/R_vocab": 0.24285714285714285
},
{
"learning_rate": 3.0578089317502436e-07,
"loss": -0.0134,
"step": 668,
"train/R_acc": 0.034375,
"train/R_penalty": -0.525,
"train/R_reason": 0.1605276972684076,
"train/R_vocab": 0.22779017857142858
},
{
"learning_rate": 3.0472583095164873e-07,
"loss": -0.0496,
"step": 670,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.19456119250844006,
"train/R_vocab": 0.23214285714285715
},
{
"learning_rate": 3.036697439776504e-07,
"loss": 0.0554,
"step": 672,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.2515289920392415,
"train/R_vocab": 0.2724330357142857
},
{
"learning_rate": 3.0261265202844016e-07,
"loss": 0.02,
"step": 674,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.18580953264755184,
"train/R_vocab": 0.25388144841269844
},
{
"learning_rate": 3.01554574898247e-07,
"loss": 0.0155,
"step": 676,
"train/R_acc": 0.0,
"train/R_penalty": -0.55625,
"train/R_reason": 0.22948436053810994,
"train/R_vocab": 0.22013888888888888
},
{
"learning_rate": 3.004955323997478e-07,
"loss": -0.0179,
"step": 678,
"train/R_acc": 0.00625,
"train/R_penalty": -0.59375,
"train/R_reason": 0.21153443586764287,
"train/R_vocab": 0.23524305555555558
},
{
"learning_rate": 2.994355443636958e-07,
"loss": -0.0362,
"step": 680,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.1840845890109734,
"train/R_vocab": 0.31015625
},
{
"learning_rate": 2.983746306385499e-07,
"loss": 0.0297,
"step": 682,
"train/R_acc": 0.0625,
"train/R_penalty": -0.453125,
"train/R_reason": 0.2084560166468069,
"train/R_vocab": 0.31228918650793647
},
{
"learning_rate": 2.9731281109010253e-07,
"loss": 0.0534,
"step": 684,
"train/R_acc": 0.0,
"train/R_penalty": -0.71875,
"train/R_reason": 0.16422497838360026,
"train/R_vocab": 0.23582589285714284
},
{
"learning_rate": 2.9625010560110787e-07,
"loss": 0.0183,
"step": 686,
"train/R_acc": 0.003125,
"train/R_penalty": -0.634375,
"train/R_reason": 0.24390149995516267,
"train/R_vocab": 0.29910714285714285
},
{
"learning_rate": 2.951865340709095e-07,
"loss": 0.0683,
"step": 688,
"train/R_acc": 0.034375,
"train/R_penalty": -0.64375,
"train/R_reason": 0.23521821731229603,
"train/R_vocab": 0.3359747023809524
},
{
"learning_rate": 2.9412211641506773e-07,
"loss": -0.001,
"step": 690,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.18437836963522442,
"train/R_vocab": 0.209375
},
{
"learning_rate": 2.930568725649867e-07,
"loss": 0.0194,
"step": 692,
"train/R_acc": 0.03125,
"train/R_penalty": -0.65,
"train/R_reason": 0.21411232248047601,
"train/R_vocab": 0.2759548611111111
},
{
"learning_rate": 2.919908224675412e-07,
"loss": 0.0579,
"step": 694,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5625,
"train/R_reason": 0.1540599514957384,
"train/R_vocab": 0.18344494047619048
},
{
"learning_rate": 2.9092398608470326e-07,
"loss": -0.0106,
"step": 696,
"train/R_acc": 0.03125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2501350490350204,
"train/R_vocab": 0.3291294642857143
},
{
"learning_rate": 2.898563833931681e-07,
"loss": 0.0156,
"step": 698,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.20865047120336916,
"train/R_vocab": 0.2130332341269841
},
{
"learning_rate": 2.8878803438398015e-07,
"loss": 0.0192,
"step": 700,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.19110667770899964,
"train/R_vocab": 0.24427083333333333
},
{
"eval/R_acc": 0.0025,
"eval/R_penalty": -0.5875,
"eval/R_reason": 0.25186260885534084,
"eval/R_vocab": 0.2177263708513708,
"step": 700
},
{
"step": 700
},
{
"learning_rate": 2.8771895906215895e-07,
"loss": 0.05,
"step": 702,
"train/R_acc": 0.003125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.2448799886759996,
"train/R_vocab": 0.3176339285714286
},
{
"learning_rate": 2.8664917744632423e-07,
"loss": 0.0057,
"step": 704,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.22652983918213118,
"train/R_vocab": 0.41149553571428577
},
{
"learning_rate": 2.8557870956832133e-07,
"loss": 0.0356,
"step": 706,
"train/R_acc": 0.0,
"train/R_penalty": -0.71875,
"train/R_reason": 0.22872856053626722,
"train/R_vocab": 0.2622767857142857
},
{
"learning_rate": 2.8450757547284576e-07,
"loss": 0.0236,
"step": 708,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.19584017346681087,
"train/R_vocab": 0.20625000000000002
},
{
"learning_rate": 2.8343579521706823e-07,
"loss": -0.0244,
"step": 710,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.17909964580588045,
"train/R_vocab": 0.23783482142857143
},
{
"learning_rate": 2.8236338887025886e-07,
"loss": 0.0155,
"step": 712,
"train/R_acc": 0.034375,
"train/R_penalty": -0.484375,
"train/R_reason": 0.26302313003150035,
"train/R_vocab": 0.3813988095238095
},
{
"learning_rate": 2.812903765134115e-07,
"loss": -0.0178,
"step": 714,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2631388468076138,
"train/R_vocab": 0.4032242063492063
},
{
"learning_rate": 2.8021677823886754e-07,
"loss": -0.0467,
"step": 716,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.19911790925777817,
"train/R_vocab": 0.23608630952380952
},
{
"learning_rate": 2.7914261414993976e-07,
"loss": 0.0054,
"step": 718,
"train/R_acc": 0.003125,
"train/R_penalty": -0.603125,
"train/R_reason": 0.2387146239629917,
"train/R_vocab": 0.2571428571428571
},
{
"learning_rate": 2.780679043605361e-07,
"loss": -0.0082,
"step": 720,
"train/R_acc": 0.03125,
"train/R_penalty": -0.571875,
"train/R_reason": 0.2269094768328706,
"train/R_vocab": 0.31436011904761907
},
{
"learning_rate": 2.7699266899478274e-07,
"loss": 0.0178,
"step": 722,
"train/R_acc": 0.034375,
"train/R_penalty": -0.703125,
"train/R_reason": 0.24130363361070928,
"train/R_vocab": 0.22456597222222222
},
{
"learning_rate": 2.759169281866472e-07,
"loss": 0.0661,
"step": 724,
"train/R_acc": 0.03125,
"train/R_penalty": -0.64375,
"train/R_reason": 0.21308869490175653,
"train/R_vocab": 0.2772321428571428
},
{
"learning_rate": 2.748407020795617e-07,
"loss": 0.1056,
"step": 726,
"train/R_acc": 0.096875,
"train/R_penalty": -0.421875,
"train/R_reason": 0.23621705469287618,
"train/R_vocab": 0.2981770833333333
},
{
"learning_rate": 2.737640108260456e-07,
"loss": 0.0812,
"step": 728,
"train/R_acc": 0.03125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.2310603125451809,
"train/R_vocab": 0.30590277777777775
},
{
"learning_rate": 2.726868745873286e-07,
"loss": -0.0637,
"step": 730,
"train/R_acc": 0.034375,
"train/R_penalty": -0.696875,
"train/R_reason": 0.19990588785781663,
"train/R_vocab": 0.265625
},
{
"learning_rate": 2.716093135329722e-07,
"loss": 0.0348,
"step": 732,
"train/R_acc": 0.0375,
"train/R_penalty": -0.421875,
"train/R_reason": 0.2095006410516756,
"train/R_vocab": 0.32751736111111107
},
{
"learning_rate": 2.7053134784049316e-07,
"loss": 0.0875,
"step": 734,
"train/R_acc": 0.003125,
"train/R_penalty": -0.71875,
"train/R_reason": 0.17118221219882784,
"train/R_vocab": 0.17343750000000002
},
{
"learning_rate": 2.694529976949849e-07,
"loss": -0.0448,
"step": 736,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.609375,
"train/R_reason": 0.20588894905097077,
"train/R_vocab": 0.34609375
},
{
"learning_rate": 2.6837428328874007e-07,
"loss": -0.0244,
"step": 738,
"train/R_acc": 0.06875,
"train/R_penalty": -0.6125,
"train/R_reason": 0.26183079845209467,
"train/R_vocab": 0.2656994047619048
},
{
"learning_rate": 2.6729522482087165e-07,
"loss": 0.0854,
"step": 740,
"train/R_acc": 0.003125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.20793241180273003,
"train/R_vocab": 0.21996527777777775
},
{
"learning_rate": 2.662158424969357e-07,
"loss": 0.0149,
"step": 742,
"train/R_acc": 0.003125,
"train/R_penalty": -0.46875,
"train/R_reason": 0.18410838343666375,
"train/R_vocab": 0.1974330357142857
},
{
"learning_rate": 2.6513615652855246e-07,
"loss": -0.0144,
"step": 744,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.24293661434589603,
"train/R_vocab": 0.3267857142857143
},
{
"learning_rate": 2.6405618713302783e-07,
"loss": -0.0678,
"step": 746,
"train/R_acc": 0.00625,
"train/R_penalty": -0.61875,
"train/R_reason": 0.135718109682479,
"train/R_vocab": 0.22979910714285715
},
{
"learning_rate": 2.629759545329749e-07,
"loss": 0.1246,
"step": 748,
"train/R_acc": 0.0,
"train/R_penalty": -0.634375,
"train/R_reason": 0.20181279701198124,
"train/R_vocab": 0.3018849206349207
},
{
"learning_rate": 2.618954789559356e-07,
"loss": 0.0933,
"step": 750,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.20252563848526106,
"train/R_vocab": 0.2291294642857143
},
{
"learning_rate": 2.6081478063400124e-07,
"loss": -0.0234,
"step": 752,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.19106041589978606,
"train/R_vocab": 0.20889136904761904
},
{
"learning_rate": 2.597338798034344e-07,
"loss": -0.0631,
"step": 754,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.1867345308525643,
"train/R_vocab": 0.22708333333333333
},
{
"learning_rate": 2.5865279670428956e-07,
"loss": 0.0483,
"step": 756,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.17135153722070606,
"train/R_vocab": 0.18147321428571428
},
{
"learning_rate": 2.5757155158003414e-07,
"loss": 0.0295,
"step": 758,
"train/R_acc": 0.0,
"train/R_penalty": -0.703125,
"train/R_reason": 0.1663155804467654,
"train/R_vocab": 0.22377232142857145
},
{
"learning_rate": 2.5649016467716954e-07,
"loss": 0.0283,
"step": 760,
"train/R_acc": 0.04062500000000001,
"train/R_penalty": -0.375,
"train/R_reason": 0.19205683929152886,
"train/R_vocab": 0.2728422619047619
},
{
"learning_rate": 2.5540865624485216e-07,
"loss": -0.0491,
"step": 762,
"train/R_acc": 0.0,
"train/R_penalty": -0.703125,
"train/R_reason": 0.2550380457423676,
"train/R_vocab": 0.2734375
},
{
"learning_rate": 2.5432704653451374e-07,
"loss": 0.0441,
"step": 764,
"train/R_acc": 0.03125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.20927672505440928,
"train/R_vocab": 0.30055803571428574
},
{
"learning_rate": 2.532453557994827e-07,
"loss": 0.048,
"step": 766,
"train/R_acc": 0.003125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.20293042206204454,
"train/R_vocab": 0.2447668650793651
},
{
"learning_rate": 2.521636042946046e-07,
"loss": 0.0256,
"step": 768,
"train/R_acc": 0.04062500000000001,
"train/R_penalty": -0.55625,
"train/R_reason": 0.2228340738544955,
"train/R_vocab": 0.21646825396825398
},
{
"learning_rate": 2.5108181227586307e-07,
"loss": -0.0186,
"step": 770,
"train/R_acc": 0.03125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.15339300119866717,
"train/R_vocab": 0.24015376984126985
},
{
"learning_rate": 2.5e-07,
"loss": 0.0082,
"step": 772,
"train/R_acc": 0.034375,
"train/R_penalty": -0.59375,
"train/R_reason": 0.25634859172881913,
"train/R_vocab": 0.2591517857142857
},
{
"learning_rate": 2.48918187724137e-07,
"loss": -0.0372,
"step": 774,
"train/R_acc": 0.0,
"train/R_penalty": -0.484375,
"train/R_reason": 0.15909137824419298,
"train/R_vocab": 0.2725953733766234
},
{
"learning_rate": 2.4783639570539535e-07,
"loss": -0.1172,
"step": 776,
"train/R_acc": 0.00625,
"train/R_penalty": -0.46875,
"train/R_reason": 0.21610677086076593,
"train/R_vocab": 0.21164434523809522
},
{
"learning_rate": 2.467546442005173e-07,
"loss": 0.0088,
"step": 778,
"train/R_acc": 0.06875,
"train/R_penalty": -0.390625,
"train/R_reason": 0.19689824765738337,
"train/R_vocab": 0.1890625
},
{
"learning_rate": 2.4567295346548634e-07,
"loss": 0.1308,
"step": 780,
"train/R_acc": 0.03125,
"train/R_penalty": -0.625,
"train/R_reason": 0.16858555636857672,
"train/R_vocab": 0.29520089285714285
},
{
"learning_rate": 2.4459134375514787e-07,
"loss": -0.0356,
"step": 782,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5875,
"train/R_reason": 0.20190652121448122,
"train/R_vocab": 0.2455357142857143
},
{
"learning_rate": 2.4350983532283043e-07,
"loss": -0.0182,
"step": 784,
"train/R_acc": 0.003125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.1856742513120948,
"train/R_vocab": 0.275
},
{
"learning_rate": 2.4242844841996583e-07,
"loss": 0.0167,
"step": 786,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.20132693663438478,
"train/R_vocab": 0.18971974206349207
},
{
"learning_rate": 2.4134720329571047e-07,
"loss": 0.0408,
"step": 788,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.571875,
"train/R_reason": 0.22452081729968582,
"train/R_vocab": 0.21640625
},
{
"learning_rate": 2.4026612019656556e-07,
"loss": 0.0107,
"step": 790,
"train/R_acc": 0.034375,
"train/R_penalty": -0.609375,
"train/R_reason": 0.18352974837528543,
"train/R_vocab": 0.25814732142857144
},
{
"learning_rate": 2.3918521936599874e-07,
"loss": 0.036,
"step": 792,
"train/R_acc": 0.03125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.27166955319028896,
"train/R_vocab": 0.35963541666666665
},
{
"learning_rate": 2.381045210440644e-07,
"loss": -0.0346,
"step": 794,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.22416891439419612,
"train/R_vocab": 0.32734375000000004
},
{
"learning_rate": 2.3702404546702509e-07,
"loss": 0.0911,
"step": 796,
"train/R_acc": 0.034375,
"train/R_penalty": -0.546875,
"train/R_reason": 0.24303349052516854,
"train/R_vocab": 0.27745535714285713
},
{
"learning_rate": 2.3594381286697215e-07,
"loss": 0.0121,
"step": 798,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.23618598427535661,
"train/R_vocab": 0.28122519841269844
},
{
"learning_rate": 2.3486384347144752e-07,
"loss": 0.0064,
"step": 800,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.25934359162168097,
"train/R_vocab": 0.4203125
},
{
"eval/R_acc": 0.0025,
"eval/R_penalty": -0.5875,
"eval/R_reason": 0.29083729720667373,
"eval/R_vocab": 0.22747113997114,
"step": 800
},
{
"step": 800
},
{
"learning_rate": 2.337841575030642e-07,
"loss": -0.0305,
"step": 802,
"train/R_acc": 0.03125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.20417958820529003,
"train/R_vocab": 0.2702008928571429
},
{
"learning_rate": 2.3270477517912835e-07,
"loss": 0.0361,
"step": 804,
"train/R_acc": 0.00625,
"train/R_penalty": -0.621875,
"train/R_reason": 0.2872431102010518,
"train/R_vocab": 0.37351190476190477
},
{
"learning_rate": 2.3162571671126e-07,
"loss": 0.0184,
"step": 806,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2281768901636334,
"train/R_vocab": 0.30616319444444445
},
{
"learning_rate": 2.3054700230501502e-07,
"loss": 0.0128,
"step": 808,
"train/R_acc": 0.03125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.20789617508651104,
"train/R_vocab": 0.3257440476190476
},
{
"learning_rate": 2.2946865215950685e-07,
"loss": 0.0158,
"step": 810,
"train/R_acc": 0.003125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.22696589858153038,
"train/R_vocab": 0.21540178571428575
},
{
"learning_rate": 2.2839068646702786e-07,
"loss": 1.5239,
"step": 812,
"train/R_acc": 0.0375,
"train/R_penalty": -0.734375,
"train/R_reason": 0.1726926027420783,
"train/R_vocab": 0.25167410714285715
},
{
"learning_rate": 2.2731312541267143e-07,
"loss": 0.0348,
"step": 814,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.53125,
"train/R_reason": 0.18792369164667982,
"train/R_vocab": 0.2892485119047619
},
{
"learning_rate": 2.2623598917395436e-07,
"loss": 0.0032,
"step": 816,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.16929829805580582,
"train/R_vocab": 0.16987847222222222
},
{
"learning_rate": 2.251592979204383e-07,
"loss": 0.0127,
"step": 818,
"train/R_acc": 0.03125,
"train/R_penalty": -0.390625,
"train/R_reason": 0.28610113589945263,
"train/R_vocab": 0.35
},
{
"learning_rate": 2.2408307181335285e-07,
"loss": 0.0528,
"step": 820,
"train/R_acc": 0.0,
"train/R_penalty": -0.509375,
"train/R_reason": 0.18837133598647154,
"train/R_vocab": 0.21272321428571428
},
{
"learning_rate": 2.2300733100521732e-07,
"loss": 0.0454,
"step": 822,
"train/R_acc": 0.003125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.1814823048344199,
"train/R_vocab": 0.13814484126984128
},
{
"learning_rate": 2.2193209563946382e-07,
"loss": 0.0355,
"step": 824,
"train/R_acc": 0.0,
"train/R_penalty": -0.6875,
"train/R_reason": 0.18663805046129311,
"train/R_vocab": 0.1985863095238095
},
{
"learning_rate": 2.2085738585006021e-07,
"loss": 0.0026,
"step": 826,
"train/R_acc": 0.0,
"train/R_penalty": -0.49375,
"train/R_reason": 0.18273024857160083,
"train/R_vocab": 0.21136363636363636
},
{
"learning_rate": 2.1978322176113251e-07,
"loss": -0.0318,
"step": 828,
"train/R_acc": 0.034375,
"train/R_penalty": -0.578125,
"train/R_reason": 0.19883211536778234,
"train/R_vocab": 0.27878224206349206
},
{
"learning_rate": 2.1870962348658852e-07,
"loss": -0.0335,
"step": 830,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.546875,
"train/R_reason": 0.1985312923562238,
"train/R_vocab": 0.2785714285714286
},
{
"learning_rate": 2.1763661112974115e-07,
"loss": -0.0012,
"step": 832,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2787590605964265,
"train/R_vocab": 0.35
},
{
"learning_rate": 2.165642047829318e-07,
"loss": -0.0375,
"step": 834,
"train/R_acc": 0.03125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.15530706954823303,
"train/R_vocab": 0.22005208333333334
},
{
"learning_rate": 2.1549242452715427e-07,
"loss": -0.0387,
"step": 836,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5,
"train/R_reason": 0.24527110960119072,
"train/R_vocab": 0.38571428571428573
},
{
"learning_rate": 2.1442129043167873e-07,
"loss": 0.0321,
"step": 838,
"train/R_acc": 0.003125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.21774443607948502,
"train/R_vocab": 0.26517857142857143
},
{
"learning_rate": 2.1335082255367572e-07,
"loss": 0.0288,
"step": 840,
"train/R_acc": 0.03125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.22299875389864712,
"train/R_vocab": 0.27976190476190477
},
{
"learning_rate": 2.1228104093784108e-07,
"loss": 0.0203,
"step": 842,
"train/R_acc": 0.003125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.25459030399590693,
"train/R_vocab": 0.36163194444444446
},
{
"learning_rate": 2.112119656160199e-07,
"loss": 0.0337,
"step": 844,
"train/R_acc": 0.003125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.2547007960169786,
"train/R_vocab": 0.34011656746031743
},
{
"learning_rate": 2.1014361660683194e-07,
"loss": -0.0236,
"step": 846,
"train/R_acc": 0.03125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.23028272641397346,
"train/R_vocab": 0.25
},
{
"learning_rate": 2.0907601391529677e-07,
"loss": -0.0204,
"step": 848,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.18682793317030633,
"train/R_vocab": 0.21324066558441557
},
{
"learning_rate": 2.0800917753245875e-07,
"loss": 0.1057,
"step": 850,
"train/R_acc": 0.00625,
"train/R_penalty": -0.609375,
"train/R_reason": 0.2020052493242343,
"train/R_vocab": 0.27734375
},
{
"learning_rate": 2.0694312743501334e-07,
"loss": -0.0585,
"step": 852,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.23457174043904147,
"train/R_vocab": 0.3036830357142857
},
{
"learning_rate": 2.0587788358493235e-07,
"loss": -0.0318,
"step": 854,
"train/R_acc": 0.003125,
"train/R_penalty": -0.34375,
"train/R_reason": 0.2538105699194751,
"train/R_vocab": 0.40078125
},
{
"learning_rate": 2.048134659290905e-07,
"loss": -0.0723,
"step": 856,
"train/R_acc": 0.0,
"train/R_penalty": -0.6875,
"train/R_reason": 0.17298681086107648,
"train/R_vocab": 0.15267857142857144
},
{
"learning_rate": 2.037498943988921e-07,
"loss": -0.0058,
"step": 858,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.28539022868157543,
"train/R_vocab": 0.3465593434343434
},
{
"learning_rate": 2.0268718890989752e-07,
"loss": -0.0416,
"step": 860,
"train/R_acc": 0.03125,
"train/R_penalty": -0.509375,
"train/R_reason": 0.20216319087856421,
"train/R_vocab": 0.3190848214285714
},
{
"learning_rate": 2.0162536936145008e-07,
"loss": -0.0399,
"step": 862,
"train/R_acc": 0.003125,
"train/R_penalty": -0.696875,
"train/R_reason": 0.2420341418631836,
"train/R_vocab": 0.27738095238095234
},
{
"learning_rate": 2.0056445563630423e-07,
"loss": 0.0372,
"step": 864,
"train/R_acc": 0.03125,
"train/R_penalty": -0.55625,
"train/R_reason": 0.21099684301910737,
"train/R_vocab": 0.26659226190476193
},
{
"learning_rate": 1.9950446760025219e-07,
"loss": 0.0326,
"step": 866,
"train/R_acc": 0.0625,
"train/R_penalty": -0.61875,
"train/R_reason": 0.19938542697531797,
"train/R_vocab": 0.25535714285714284
},
{
"learning_rate": 1.98445425101753e-07,
"loss": 0.0211,
"step": 868,
"train/R_acc": 0.003125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.22302710668155856,
"train/R_vocab": 0.2036830357142857
},
{
"learning_rate": 1.9738734797155987e-07,
"loss": 0.1391,
"step": 870,
"train/R_acc": 0.0125,
"train/R_penalty": -0.703125,
"train/R_reason": 0.20507300880803225,
"train/R_vocab": 0.29453125
},
{
"learning_rate": 1.963302560223496e-07,
"loss": 0.0314,
"step": 872,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.1946319904099672,
"train/R_vocab": 0.32533482142857145
},
{
"learning_rate": 1.9527416904835132e-07,
"loss": 0.0562,
"step": 874,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.2123123501156333,
"train/R_vocab": 0.24363839285714287
},
{
"learning_rate": 1.9421910682497572e-07,
"loss": 0.0335,
"step": 876,
"train/R_acc": 0.0625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.24626829087757965,
"train/R_vocab": 0.28229166666666666
},
{
"learning_rate": 1.9316508910844476e-07,
"loss": 0.0255,
"step": 878,
"train/R_acc": 0.034375,
"train/R_penalty": -0.59375,
"train/R_reason": 0.18789237365926498,
"train/R_vocab": 0.278125
},
{
"learning_rate": 1.9211213563542218e-07,
"loss": -0.0622,
"step": 880,
"train/R_acc": 0.0,
"train/R_penalty": -0.5,
"train/R_reason": 0.19592139018937305,
"train/R_vocab": 0.26670386904761906
},
{
"learning_rate": 1.9106026612264315e-07,
"loss": 0.0249,
"step": 882,
"train/R_acc": 0.065625,
"train/R_penalty": -0.5,
"train/R_reason": 0.17202561289857762,
"train/R_vocab": 0.24394841269841272
},
{
"learning_rate": 1.900095002665459e-07,
"loss": 0.0627,
"step": 884,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5625,
"train/R_reason": 0.20212251182315438,
"train/R_vocab": 0.234375
},
{
"learning_rate": 1.889598577429022e-07,
"loss": 0.0251,
"step": 886,
"train/R_acc": 0.00625,
"train/R_penalty": -0.6875,
"train/R_reason": 0.21931575978869752,
"train/R_vocab": 0.20535714285714285
},
{
"learning_rate": 1.8791135820644922e-07,
"loss": 0.0228,
"step": 888,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.23473900860844063,
"train/R_vocab": 0.2440104166666667
},
{
"learning_rate": 1.8686402129052181e-07,
"loss": 0.0602,
"step": 890,
"train/R_acc": 0.034375,
"train/R_penalty": -0.5625,
"train/R_reason": 0.2188358452056008,
"train/R_vocab": 0.24036458333333333
},
{
"learning_rate": 1.858178666066843e-07,
"loss": 0.0042,
"step": 892,
"train/R_acc": 0.034375,
"train/R_penalty": -0.59375,
"train/R_reason": 0.1951297685148197,
"train/R_vocab": 0.23270089285714285
},
{
"learning_rate": 1.8477291374436338e-07,
"loss": -0.0364,
"step": 894,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.11877722404091096,
"train/R_vocab": 0.12317708333333333
},
{
"learning_rate": 1.8372918227048183e-07,
"loss": 0.3093,
"step": 896,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.1855972158801493,
"train/R_vocab": 0.25017361111111114
},
{
"learning_rate": 1.8268669172909136e-07,
"loss": 0.0017,
"step": 898,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.18804079477258429,
"train/R_vocab": 0.16622023809523812
},
{
"learning_rate": 1.8164546164100728e-07,
"loss": -0.0186,
"step": 900,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.24246348014867208,
"train/R_vocab": 0.29110863095238093
},
{
"eval/R_acc": 0.030000000000000006,
"eval/R_penalty": -0.6,
"eval/R_reason": 0.29157417644976924,
"eval/R_vocab": 0.37875,
"step": 900
},
{
"step": 900
},
{
"learning_rate": 1.806055115034425e-07,
"loss": -0.0579,
"step": 902,
"train/R_acc": 0.003125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.23201039495721226,
"train/R_vocab": 0.2525669642857143
},
{
"learning_rate": 1.7956686078964255e-07,
"loss": 0.0383,
"step": 904,
"train/R_acc": 0.03125,
"train/R_penalty": -0.61875,
"train/R_reason": 0.19469584258694483,
"train/R_vocab": 0.21830357142857146
},
{
"learning_rate": 1.7852952894852148e-07,
"loss": 0.0293,
"step": 906,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.625,
"train/R_reason": 0.25663321405687106,
"train/R_vocab": 0.32265625000000003
},
{
"learning_rate": 1.7749353540429663e-07,
"loss": -0.0106,
"step": 908,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.19230140159133588,
"train/R_vocab": 0.16744791666666667
},
{
"learning_rate": 1.7645889955612592e-07,
"loss": -0.0295,
"step": 910,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.12090342349306576,
"train/R_vocab": 0.1583457341269841
},
{
"learning_rate": 1.7542564077774403e-07,
"loss": 0.0045,
"step": 912,
"train/R_acc": 0.034375,
"train/R_penalty": -0.675,
"train/R_reason": 0.2114914680859179,
"train/R_vocab": 0.3030505952380952
},
{
"learning_rate": 1.7439377841709973e-07,
"loss": -0.0017,
"step": 914,
"train/R_acc": 0.0625,
"train/R_penalty": -0.609375,
"train/R_reason": 0.217770400372361,
"train/R_vocab": 0.25592757936507937
},
{
"learning_rate": 1.7336333179599378e-07,
"loss": 0.0849,
"step": 916,
"train/R_acc": 0.00625,
"train/R_penalty": -0.609375,
"train/R_reason": 0.2602397752431368,
"train/R_vocab": 0.4150173611111111
},
{
"learning_rate": 1.7233432020971684e-07,
"loss": 0.0793,
"step": 918,
"train/R_acc": 0.0,
"train/R_penalty": -0.55625,
"train/R_reason": 0.2766625564419999,
"train/R_vocab": 0.23649553571428572
},
{
"learning_rate": 1.7130676292668816e-07,
"loss": 0.0446,
"step": 920,
"train/R_acc": 0.003125,
"train/R_penalty": -0.6875,
"train/R_reason": 0.22695619844003265,
"train/R_vocab": 0.20081845238095236
},
{
"learning_rate": 1.7028067918809535e-07,
"loss": -0.036,
"step": 922,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.17932276860695562,
"train/R_vocab": 0.2564732142857143
},
{
"learning_rate": 1.6925608820753325e-07,
"loss": -0.0333,
"step": 924,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.1710392126096163,
"train/R_vocab": 0.11149553571428572
},
{
"learning_rate": 1.6823300917064458e-07,
"loss": -0.0427,
"step": 926,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.1853385177642702,
"train/R_vocab": 0.20937499999999998
},
{
"learning_rate": 1.6721146123476093e-07,
"loss": 0.122,
"step": 928,
"train/R_acc": 0.03125,
"train/R_penalty": -0.634375,
"train/R_reason": 0.21111988917807223,
"train/R_vocab": 0.28467261904761904
},
{
"learning_rate": 1.6619146352854334e-07,
"loss": 0.0335,
"step": 930,
"train/R_acc": 0.00625,
"train/R_penalty": -0.525,
"train/R_reason": 0.2012206313966573,
"train/R_vocab": 0.2486359126984127
},
{
"learning_rate": 1.6517303515162505e-07,
"loss": -0.0597,
"step": 932,
"train/R_acc": 0.034375,
"train/R_penalty": -0.5625,
"train/R_reason": 0.16395669599613677,
"train/R_vocab": 0.2521329365079365
},
{
"learning_rate": 1.6415619517425294e-07,
"loss": 0.0165,
"step": 934,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65,
"train/R_reason": 0.14773459331491504,
"train/R_vocab": 0.2604910714285714
},
{
"learning_rate": 1.6314096263693078e-07,
"loss": -0.0171,
"step": 936,
"train/R_acc": 0.0,
"train/R_penalty": -0.61875,
"train/R_reason": 0.24617103354461017,
"train/R_vocab": 0.25078125
},
{
"learning_rate": 1.6212735655006322e-07,
"loss": -0.0329,
"step": 938,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.2476563881397631,
"train/R_vocab": 0.22760416666666666
},
{
"learning_rate": 1.6111539589359885e-07,
"loss": -0.0157,
"step": 940,
"train/R_acc": 0.034375,
"train/R_penalty": -0.55625,
"train/R_reason": 0.18716442508640668,
"train/R_vocab": 0.21462053571428572
},
{
"learning_rate": 1.6010509961667546e-07,
"loss": 0.0015,
"step": 942,
"train/R_acc": 0.00625,
"train/R_penalty": -0.5625,
"train/R_reason": 0.2163363343470736,
"train/R_vocab": 0.3014508928571429
},
{
"learning_rate": 1.590964866372652e-07,
"loss": -0.0064,
"step": 944,
"train/R_acc": 0.0,
"train/R_penalty": -0.5875,
"train/R_reason": 0.22361495696140293,
"train/R_vocab": 0.278125
},
{
"learning_rate": 1.5808957584181994e-07,
"loss": -0.0227,
"step": 946,
"train/R_acc": 0.03125,
"train/R_penalty": -0.596875,
"train/R_reason": 0.2240925844127057,
"train/R_vocab": 0.30379464285714286
},
{
"learning_rate": 1.5708438608491815e-07,
"loss": 0.0113,
"step": 948,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.19887587350757036,
"train/R_vocab": 0.2051846590909091
},
{
"learning_rate": 1.560809361889112e-07,
"loss": 0.0187,
"step": 950,
"train/R_acc": 0.0,
"train/R_penalty": -0.671875,
"train/R_reason": 0.1577485102536454,
"train/R_vocab": 0.20122767857142856
},
{
"learning_rate": 1.5507924494357154e-07,
"loss": -0.0467,
"step": 952,
"train/R_acc": 0.034375,
"train/R_penalty": -0.5625,
"train/R_reason": 0.14574324556072643,
"train/R_vocab": 0.20066964285714284
},
{
"learning_rate": 1.5407933110574044e-07,
"loss": 0.0123,
"step": 954,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.21002736065355815,
"train/R_vocab": 0.27578125
},
{
"learning_rate": 1.5308121339897705e-07,
"loss": 0.0131,
"step": 956,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.24345409494067652,
"train/R_vocab": 0.33035714285714285
},
{
"learning_rate": 1.5208491051320744e-07,
"loss": -0.1197,
"step": 958,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.578125,
"train/R_reason": 0.21702138753711128,
"train/R_vocab": 0.28426339285714286
},
{
"learning_rate": 1.510904411043751e-07,
"loss": -0.0818,
"step": 960,
"train/R_acc": 0.0,
"train/R_penalty": -0.71875,
"train/R_reason": 0.23023067131726802,
"train/R_vocab": 0.3091517857142857
},
{
"learning_rate": 1.5009782379409108e-07,
"loss": -0.1087,
"step": 962,
"train/R_acc": 0.003125,
"train/R_penalty": -0.765625,
"train/R_reason": 0.17687649620358462,
"train/R_vocab": 0.21863839285714284
},
{
"learning_rate": 1.4910707716928586e-07,
"loss": 0.0956,
"step": 964,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.46875,
"train/R_reason": 0.22474928414381412,
"train/R_vocab": 0.26093750000000004
},
{
"learning_rate": 1.4811821978186078e-07,
"loss": 0.0536,
"step": 966,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.22651502510275934,
"train/R_vocab": 0.20833333333333331
},
{
"learning_rate": 1.4713127014834088e-07,
"loss": 0.0205,
"step": 968,
"train/R_acc": 0.0,
"train/R_penalty": -0.5625,
"train/R_reason": 0.2021908958276862,
"train/R_vocab": 0.1978794642857143
},
{
"learning_rate": 1.461462467495284e-07,
"loss": 0.0427,
"step": 970,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.15024320937786767,
"train/R_vocab": 0.20468750000000002
},
{
"learning_rate": 1.4516316803015627e-07,
"loss": 0.0126,
"step": 972,
"train/R_acc": 0.00625,
"train/R_penalty": -0.59375,
"train/R_reason": 0.18895812905862736,
"train/R_vocab": 0.14308035714285713
},
{
"learning_rate": 1.441820523985431e-07,
"loss": -0.0349,
"step": 974,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.2382146965819557,
"train/R_vocab": 0.2640997023809524
},
{
"learning_rate": 1.432029182262482e-07,
"loss": 0.0445,
"step": 976,
"train/R_acc": 0.09375,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2085522733677449,
"train/R_vocab": 0.26365327380952386
},
{
"learning_rate": 1.42225783847728e-07,
"loss": 0.0309,
"step": 978,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.27484186892341045,
"train/R_vocab": 0.3541170634920635
},
{
"learning_rate": 1.41250667559992e-07,
"loss": 0.0122,
"step": 980,
"train/R_acc": 0.034375,
"train/R_penalty": -0.53125,
"train/R_reason": 0.22480109886784352,
"train/R_vocab": 0.30301339285714285
},
{
"learning_rate": 1.4027758762226107e-07,
"loss": -0.0861,
"step": 982,
"train/R_acc": 0.003125,
"train/R_penalty": -0.665625,
"train/R_reason": 0.21184840132463886,
"train/R_vocab": 0.34201388888888884
},
{
"learning_rate": 1.3930656225562474e-07,
"loss": 0.0047,
"step": 984,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.2253757708880576,
"train/R_vocab": 0.30904017857142857
},
{
"learning_rate": 1.3833760964270058e-07,
"loss": 0.0296,
"step": 986,
"train/R_acc": 0.03125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2222406698557359,
"train/R_vocab": 0.305406746031746
},
{
"learning_rate": 1.3737074792729332e-07,
"loss": 0.0262,
"step": 988,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2037288188497659,
"train/R_vocab": 0.17431795634920633
},
{
"learning_rate": 1.3640599521405522e-07,
"loss": 0.0671,
"step": 990,
"train/R_acc": 0.040625,
"train/R_penalty": -0.53125,
"train/R_reason": 0.19595219511870288,
"train/R_vocab": 0.2941964285714286
},
{
"learning_rate": 1.354433695681474e-07,
"loss": -0.0366,
"step": 992,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.1895803639379184,
"train/R_vocab": 0.2979910714285714
},
{
"learning_rate": 1.3448288901490092e-07,
"loss": 0.0929,
"step": 994,
"train/R_acc": 0.034375,
"train/R_penalty": -0.665625,
"train/R_reason": 0.21089602634201932,
"train/R_vocab": 0.19340277777777776
},
{
"learning_rate": 1.3352457153947997e-07,
"loss": 0.0515,
"step": 996,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.546875,
"train/R_reason": 0.21176803496084237,
"train/R_vocab": 0.3035094246031746
},
{
"learning_rate": 1.3256843508654444e-07,
"loss": 0.0103,
"step": 998,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.15247972660677223,
"train/R_vocab": 0.21183035714285717
},
{
"learning_rate": 1.3161449755991425e-07,
"loss": -0.082,
"step": 1000,
"train/R_acc": 0.13125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.2110448967120055,
"train/R_vocab": 0.25863095238095235
},
{
"eval/R_acc": 0.0025,
"eval/R_penalty": -0.55,
"eval/R_reason": 0.2655294246752871,
"eval/R_vocab": 0.34914772727272725,
"step": 1000
},
{
"step": 1000
},
{
"learning_rate": 1.3066277682223425e-07,
"loss": -0.0096,
"step": 1002,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.181312457358332,
"train/R_vocab": 0.21517857142857144
},
{
"learning_rate": 1.2971329069463932e-07,
"loss": -0.0918,
"step": 1004,
"train/R_acc": 0.003125,
"train/R_penalty": -0.75,
"train/R_reason": 0.12020329779967928,
"train/R_vocab": 0.12425595238095238
},
{
"learning_rate": 1.2876605695642084e-07,
"loss": -0.0157,
"step": 1006,
"train/R_acc": 0.03125,
"train/R_penalty": -0.625,
"train/R_reason": 0.2288382776444276,
"train/R_vocab": 0.3042410714285715
},
{
"learning_rate": 1.2782109334469409e-07,
"loss": 0.0782,
"step": 1008,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.23151180047594122,
"train/R_vocab": 0.30937499999999996
},
{
"learning_rate": 1.2687841755406554e-07,
"loss": -0.0145,
"step": 1010,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.23526236082480567,
"train/R_vocab": 0.30009920634920634
},
{
"learning_rate": 1.2593804723630209e-07,
"loss": 0.0254,
"step": 1012,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.2194206796665749,
"train/R_vocab": 0.3340401785714286
},
{
"learning_rate": 1.2500000000000005e-07,
"loss": 0.0277,
"step": 1014,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.18922638449707688,
"train/R_vocab": 0.3610491071428572
},
{
"learning_rate": 1.2406429341025556e-07,
"loss": 0.0668,
"step": 1016,
"train/R_acc": 0.00625,
"train/R_penalty": -0.55625,
"train/R_reason": 0.2346744057550262,
"train/R_vocab": 0.2239955357142857
},
{
"learning_rate": 1.231309449883361e-07,
"loss": 0.0338,
"step": 1018,
"train/R_acc": 0.003125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.19074387902354337,
"train/R_vocab": 0.21744791666666666
},
{
"learning_rate": 1.2219997221135177e-07,
"loss": 0.0082,
"step": 1020,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.24590138406147055,
"train/R_vocab": 0.2600446428571429
},
{
"learning_rate": 1.2127139251192823e-07,
"loss": 0.0268,
"step": 1022,
"train/R_acc": 0.0,
"train/R_penalty": -0.53125,
"train/R_reason": 0.24386600100476474,
"train/R_vocab": 0.3195560515873016
},
{
"learning_rate": 1.203452232778807e-07,
"loss": 0.0183,
"step": 1024,
"train/R_acc": 0.00625,
"train/R_penalty": -0.671875,
"train/R_reason": 0.171127249096387,
"train/R_vocab": 0.29609375000000004
},
{
"learning_rate": 1.1942148185188763e-07,
"loss": 0.0272,
"step": 1026,
"train/R_acc": 0.00625,
"train/R_penalty": -0.5625,
"train/R_reason": 0.17080492182756957,
"train/R_vocab": 0.21568627450980393
},
{
"learning_rate": 1.1850018553116676e-07,
"loss": -0.0052,
"step": 1028,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5,
"train/R_reason": 0.2604828445455006,
"train/R_vocab": 0.23158482142857142
},
{
"learning_rate": 1.1758135156715041e-07,
"loss": -0.0154,
"step": 1030,
"train/R_acc": 0.065625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.1946716410122738,
"train/R_vocab": 0.22801339285714284
},
{
"learning_rate": 1.1666499716516302e-07,
"loss": 0.0101,
"step": 1032,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.18422370796707505,
"train/R_vocab": 0.21763392857142855
},
{
"learning_rate": 1.157511394840989e-07,
"loss": -0.0049,
"step": 1034,
"train/R_acc": 0.00625,
"train/R_penalty": -0.7125,
"train/R_reason": 0.17663980590936243,
"train/R_vocab": 0.24196428571428574
},
{
"learning_rate": 1.1483979563610069e-07,
"loss": 0.0098,
"step": 1036,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.26512284555646376,
"train/R_vocab": 0.375
},
{
"learning_rate": 1.1393098268623902e-07,
"loss": 0.1176,
"step": 1038,
"train/R_acc": 0.0625,
"train/R_penalty": -0.634375,
"train/R_reason": 0.26504285268931027,
"train/R_vocab": 0.30766369047619047
},
{
"learning_rate": 1.1302471765219327e-07,
"loss": 0.0428,
"step": 1040,
"train/R_acc": 0.03125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2008627221719819,
"train/R_vocab": 0.253125
},
{
"learning_rate": 1.1212101750393235e-07,
"loss": 0.0175,
"step": 1042,
"train/R_acc": 0.0625,
"train/R_penalty": -0.409375,
"train/R_reason": 0.19803424974179828,
"train/R_vocab": 0.17265625
},
{
"learning_rate": 1.1121989916339756e-07,
"loss": 0.0292,
"step": 1044,
"train/R_acc": 0.0,
"train/R_penalty": -0.61875,
"train/R_reason": 0.21600135050493885,
"train/R_vocab": 0.3185267857142857
},
{
"learning_rate": 1.1032137950418514e-07,
"loss": 0.0669,
"step": 1046,
"train/R_acc": 0.00625,
"train/R_penalty": -0.59375,
"train/R_reason": 0.23236107648223334,
"train/R_vocab": 0.22142857142857142
},
{
"learning_rate": 1.0942547535123057e-07,
"loss": -0.0038,
"step": 1048,
"train/R_acc": 0.00625,
"train/R_penalty": -0.509375,
"train/R_reason": 0.24041357438290228,
"train/R_vocab": 0.2955357142857143
},
{
"learning_rate": 1.085322034804938e-07,
"loss": 0.1258,
"step": 1050,
"train/R_acc": 0.003125,
"train/R_penalty": -0.665625,
"train/R_reason": 0.20610390132448592,
"train/R_vocab": 0.24229910714285713
},
{
"learning_rate": 1.0764158061864448e-07,
"loss": -0.0495,
"step": 1052,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65,
"train/R_reason": 0.2606824569537362,
"train/R_vocab": 0.39392361111111107
},
{
"learning_rate": 1.0675362344274952e-07,
"loss": -0.0224,
"step": 1054,
"train/R_acc": 0.040625,
"train/R_penalty": -0.53125,
"train/R_reason": 0.2362594455371677,
"train/R_vocab": 0.19696631493506495
},
{
"learning_rate": 1.0586834857995996e-07,
"loss": 0.0405,
"step": 1056,
"train/R_acc": 0.0,
"train/R_penalty": -0.59375,
"train/R_reason": 0.22383388089744485,
"train/R_vocab": 0.2839285714285714
},
{
"learning_rate": 1.0498577260720048e-07,
"loss": 0.006,
"step": 1058,
"train/R_acc": 0.003125,
"train/R_penalty": -0.5625,
"train/R_reason": 0.2154017687323606,
"train/R_vocab": 0.18723958333333335
},
{
"learning_rate": 1.0410591205085817e-07,
"loss": 0.0547,
"step": 1060,
"train/R_acc": 0.03125,
"train/R_penalty": -0.64375,
"train/R_reason": 0.22723913109425195,
"train/R_vocab": 0.32499999999999996
},
{
"learning_rate": 1.0322878338647392e-07,
"loss": 0.1005,
"step": 1062,
"train/R_acc": 0.0,
"train/R_penalty": -0.65,
"train/R_reason": 0.21853491707209402,
"train/R_vocab": 0.19251217532467532
},
{
"learning_rate": 1.0235440303843302e-07,
"loss": 0.0902,
"step": 1064,
"train/R_acc": 0.03125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.21816371211780805,
"train/R_vocab": 0.22801339285714284
},
{
"learning_rate": 1.0148278737965844e-07,
"loss": -0.0553,
"step": 1066,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.19269070794229914,
"train/R_vocab": 0.18387896825396824
},
{
"learning_rate": 1.0061395273130368e-07,
"loss": 0.039,
"step": 1068,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.21827443734801666,
"train/R_vocab": 0.2911830357142857
},
{
"learning_rate": 9.974791536244726e-08,
"loss": 0.0343,
"step": 1070,
"train/R_acc": 0.0,
"train/R_penalty": -0.61875,
"train/R_reason": 0.2733091389175897,
"train/R_vocab": 0.3764136904761905
},
{
"learning_rate": 9.888469148978843e-08,
"loss": -0.0473,
"step": 1072,
"train/R_acc": 0.003125,
"train/R_penalty": -0.484375,
"train/R_reason": 0.24157353834104733,
"train/R_vocab": 0.3257496843434343
},
{
"learning_rate": 9.802429727734294e-08,
"loss": -0.0036,
"step": 1074,
"train/R_acc": 0.040625,
"train/R_penalty": -0.421875,
"train/R_reason": 0.2823229532013337,
"train/R_vocab": 0.43273809523809526
},
{
"learning_rate": 9.716674883614092e-08,
"loss": 0.0124,
"step": 1076,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.453125,
"train/R_reason": 0.2879044257474095,
"train/R_vocab": 0.33738839285714284
},
{
"learning_rate": 9.631206222392479e-08,
"loss": -0.0157,
"step": 1078,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.26644330439869424,
"train/R_vocab": 0.27585565476190477
},
{
"learning_rate": 9.546025344484868e-08,
"loss": 0.0013,
"step": 1080,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.17772787504592696,
"train/R_vocab": 0.22031250000000002
},
{
"learning_rate": 9.461133844917901e-08,
"loss": -0.0364,
"step": 1082,
"train/R_acc": 0.0625,
"train/R_penalty": -0.696875,
"train/R_reason": 0.14647117758570943,
"train/R_vocab": 0.19616815476190477
},
{
"learning_rate": 9.376533313299542e-08,
"loss": -0.0051,
"step": 1084,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.18199736371919578,
"train/R_vocab": 0.2505580357142857
},
{
"learning_rate": 9.292225333789327e-08,
"loss": 0.0575,
"step": 1086,
"train/R_acc": 0.03125,
"train/R_penalty": -0.603125,
"train/R_reason": 0.1892425633255841,
"train/R_vocab": 0.31540178571428573
},
{
"learning_rate": 9.208211485068729e-08,
"loss": 0.0395,
"step": 1088,
"train/R_acc": 0.003125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.15220888566600485,
"train/R_vocab": 0.21587301587301586
},
{
"learning_rate": 9.124493340311537e-08,
"loss": -0.016,
"step": 1090,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.18263365867722547,
"train/R_vocab": 0.25424107142857144
},
{
"learning_rate": 9.041072467154471e-08,
"loss": -0.0291,
"step": 1092,
"train/R_acc": 0.0,
"train/R_penalty": -0.375,
"train/R_reason": 0.25466359857700704,
"train/R_vocab": 0.34862351190476193
},
{
"learning_rate": 8.957950427667751e-08,
"loss": 0.0296,
"step": 1094,
"train/R_acc": 0.03125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.23575074339334445,
"train/R_vocab": 0.2983010912698413
},
{
"learning_rate": 8.875128778325902e-08,
"loss": -0.0389,
"step": 1096,
"train/R_acc": 0.0,
"train/R_penalty": -0.53125,
"train/R_reason": 0.1681449082491554,
"train/R_vocab": 0.23660714285714285
},
{
"learning_rate": 8.792609069978602e-08,
"loss": 0.0201,
"step": 1098,
"train/R_acc": 0.0125,
"train/R_penalty": -0.765625,
"train/R_reason": 0.24261987678361474,
"train/R_vocab": 0.23020833333333335
},
{
"learning_rate": 8.710392847821615e-08,
"loss": -0.0508,
"step": 1100,
"train/R_acc": 0.03125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.19600584280495498,
"train/R_vocab": 0.24397321428571428
},
{
"eval/R_acc": 0.0025,
"eval/R_penalty": -0.4875,
"eval/R_reason": 0.24036428951274935,
"eval/R_vocab": 0.29469967532467534,
"step": 1100
},
{
"step": 1100
},
{
"learning_rate": 8.628481651367875e-08,
"loss": 0.0336,
"step": 1102,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.22866834230086636,
"train/R_vocab": 0.19769345238095237
},
{
"learning_rate": 8.546877014418671e-08,
"loss": -0.0387,
"step": 1104,
"train/R_acc": 0.03125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.21012454085545956,
"train/R_vocab": 0.23225446428571428
},
{
"learning_rate": 8.465580465034894e-08,
"loss": 0.0231,
"step": 1106,
"train/R_acc": 0.03125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.20315946724213801,
"train/R_vocab": 0.2909598214285714
},
{
"learning_rate": 8.384593525508466e-08,
"loss": 0.0343,
"step": 1108,
"train/R_acc": 0.0,
"train/R_penalty": -0.509375,
"train/R_reason": 0.20224091353679885,
"train/R_vocab": 0.19077380952380954
},
{
"learning_rate": 8.303917712333788e-08,
"loss": 0.0208,
"step": 1110,
"train/R_acc": 0.065625,
"train/R_penalty": -0.5625,
"train/R_reason": 0.23770102329722845,
"train/R_vocab": 0.21218998015873014
},
{
"learning_rate": 8.223554536179367e-08,
"loss": 0.0016,
"step": 1112,
"train/R_acc": 0.0,
"train/R_penalty": -0.5,
"train/R_reason": 0.24531693423546427,
"train/R_vocab": 0.3191964285714286
},
{
"learning_rate": 8.143505501859551e-08,
"loss": 0.0352,
"step": 1114,
"train/R_acc": 0.0,
"train/R_penalty": -0.49375,
"train/R_reason": 0.2098236201971303,
"train/R_vocab": 0.2838169642857143
},
{
"learning_rate": 8.06377210830631e-08,
"loss": 0.0216,
"step": 1116,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.20338203537474456,
"train/R_vocab": 0.3302455357142857
},
{
"learning_rate": 7.984355848541175e-08,
"loss": -0.0481,
"step": 1118,
"train/R_acc": 0.003125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.17974317204155837,
"train/R_vocab": 0.32105654761904767
},
{
"learning_rate": 7.905258209647325e-08,
"loss": 0.0544,
"step": 1120,
"train/R_acc": 0.03125,
"train/R_penalty": -0.603125,
"train/R_reason": 0.24957606202713686,
"train/R_vocab": 0.30379464285714286
},
{
"learning_rate": 7.826480672741676e-08,
"loss": -0.0196,
"step": 1122,
"train/R_acc": 0.0625,
"train/R_penalty": -0.5625,
"train/R_reason": 0.21989961574561373,
"train/R_vocab": 0.24270833333333333
},
{
"learning_rate": 7.748024712947204e-08,
"loss": 0.0036,
"step": 1124,
"train/R_acc": 0.003125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.20629454131030475,
"train/R_vocab": 0.2872767857142857
},
{
"learning_rate": 7.669891799365282e-08,
"loss": 0.0386,
"step": 1126,
"train/R_acc": 0.00625,
"train/R_penalty": -0.671875,
"train/R_reason": 0.22259933007352922,
"train/R_vocab": 0.36199156746031746
},
{
"learning_rate": 7.592083395048182e-08,
"loss": 0.0643,
"step": 1128,
"train/R_acc": 0.03125,
"train/R_penalty": -0.49375,
"train/R_reason": 0.18965226315756645,
"train/R_vocab": 0.2122767857142857
},
{
"learning_rate": 7.514600956971712e-08,
"loss": 0.0678,
"step": 1130,
"train/R_acc": 0.0,
"train/R_penalty": -0.5875,
"train/R_reason": 0.23261509390210205,
"train/R_vocab": 0.23307291666666669
},
{
"learning_rate": 7.437445936007867e-08,
"loss": 0.01,
"step": 1132,
"train/R_acc": 0.03125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.2093342551771361,
"train/R_vocab": 0.2585565476190476
},
{
"learning_rate": 7.360619776897742e-08,
"loss": -0.0392,
"step": 1134,
"train/R_acc": 0.0,
"train/R_penalty": -0.525,
"train/R_reason": 0.22079774921664902,
"train/R_vocab": 0.27708333333333335
},
{
"learning_rate": 7.284123918224397e-08,
"loss": 0.0161,
"step": 1136,
"train/R_acc": 0.03125,
"train/R_penalty": -0.59375,
"train/R_reason": 0.22520667797905555,
"train/R_vocab": 0.3407118055555556
},
{
"learning_rate": 7.207959792385998e-08,
"loss": 0.0479,
"step": 1138,
"train/R_acc": 0.00625,
"train/R_penalty": -0.515625,
"train/R_reason": 0.19075869235061965,
"train/R_vocab": 0.2729166666666667
},
{
"learning_rate": 7.132128825568914e-08,
"loss": -0.0027,
"step": 1140,
"train/R_acc": 0.03125,
"train/R_penalty": -0.634375,
"train/R_reason": 0.23292150900288705,
"train/R_vocab": 0.21964285714285714
},
{
"learning_rate": 7.056632437721099e-08,
"loss": -0.0379,
"step": 1142,
"train/R_acc": 0.0625,
"train/R_penalty": -0.578125,
"train/R_reason": 0.17068707794196553,
"train/R_vocab": 0.23448660714285716
},
{
"learning_rate": 6.981472042525416e-08,
"loss": -0.0097,
"step": 1144,
"train/R_acc": 0.003125,
"train/R_penalty": -0.484375,
"train/R_reason": 0.1982660319129968,
"train/R_vocab": 0.18515624999999997
},
{
"learning_rate": 6.906649047373245e-08,
"loss": -0.0136,
"step": 1146,
"train/R_acc": 0.0,
"train/R_penalty": -0.571875,
"train/R_reason": 0.18697098939403184,
"train/R_vocab": 0.22890624999999998
},
{
"learning_rate": 6.832164853338066e-08,
"loss": 0.0075,
"step": 1148,
"train/R_acc": 0.065625,
"train/R_penalty": -0.515625,
"train/R_reason": 0.26961296459836603,
"train/R_vocab": 0.3282118055555555
},
{
"learning_rate": 6.758020855149249e-08,
"loss": -0.0081,
"step": 1150,
"train/R_acc": 0.03125,
"train/R_penalty": -0.55625,
"train/R_reason": 0.24886312239764444,
"train/R_vocab": 0.3654017857142857
},
{
"learning_rate": 6.684218441165962e-08,
"loss": 0.0666,
"step": 1152,
"train/R_acc": 0.0,
"train/R_penalty": -0.61875,
"train/R_reason": 0.2088855316383599,
"train/R_vocab": 0.22730654761904762
},
{
"learning_rate": 6.610758993351123e-08,
"loss": 0.0395,
"step": 1154,
"train/R_acc": 0.034375,
"train/R_penalty": -0.53125,
"train/R_reason": 0.22720806482016326,
"train/R_vocab": 0.24931795634920634
},
{
"learning_rate": 6.537643887245573e-08,
"loss": 0.0465,
"step": 1156,
"train/R_acc": 0.03125,
"train/R_penalty": -0.515625,
"train/R_reason": 0.1968795889130503,
"train/R_vocab": 0.2591517857142857
},
{
"learning_rate": 6.464874491942279e-08,
"loss": 0.0203,
"step": 1158,
"train/R_acc": 0.0,
"train/R_penalty": -0.665625,
"train/R_reason": 0.19558365060317057,
"train/R_vocab": 0.1140625
},
{
"learning_rate": 6.392452170060706e-08,
"loss": 0.0173,
"step": 1160,
"train/R_acc": 0.03125,
"train/R_penalty": -0.540625,
"train/R_reason": 0.25112434395380084,
"train/R_vocab": 0.23022073412698413
},
{
"learning_rate": 6.320378277721342e-08,
"loss": -0.0386,
"step": 1162,
"train/R_acc": 0.003125,
"train/R_penalty": -0.478125,
"train/R_reason": 0.24359373155022102,
"train/R_vocab": 0.3531498015873016
},
{
"learning_rate": 6.248654164520237e-08,
"loss": -0.0419,
"step": 1164,
"train/R_acc": 0.003125,
"train/R_penalty": -0.71875,
"train/R_reason": 0.20550711716499057,
"train/R_vocab": 0.3539434523809524
},
{
"learning_rate": 6.177281173503778e-08,
"loss": 0.0405,
"step": 1166,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.20957454414924856,
"train/R_vocab": 0.2280133928571429
},
{
"learning_rate": 6.106260641143546e-08,
"loss": -0.0833,
"step": 1168,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.20663920365929042,
"train/R_vocab": 0.21852678571428574
},
{
"learning_rate": 6.035593897311239e-08,
"loss": -0.0156,
"step": 1170,
"train/R_acc": 0.003125,
"train/R_penalty": -0.61875,
"train/R_reason": 0.2371425092299592,
"train/R_vocab": 0.3323784722222222
},
{
"learning_rate": 5.965282265253838e-08,
"loss": 0.0651,
"step": 1172,
"train/R_acc": 0.034375,
"train/R_penalty": -0.390625,
"train/R_reason": 0.22832625758068797,
"train/R_vocab": 0.30301339285714285
},
{
"learning_rate": 5.895327061568775e-08,
"loss": -0.0379,
"step": 1174,
"train/R_acc": 0.03125,
"train/R_penalty": -0.578125,
"train/R_reason": 0.19140193598424352,
"train/R_vocab": 0.2755952380952381
},
{
"learning_rate": 5.8257295961792936e-08,
"loss": -0.1099,
"step": 1176,
"train/R_acc": 0.0,
"train/R_penalty": -0.640625,
"train/R_reason": 0.2013441433056279,
"train/R_vocab": 0.2722222222222222
},
{
"learning_rate": 5.756491172309952e-08,
"loss": -0.0236,
"step": 1178,
"train/R_acc": 0.003125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.2409798833837463,
"train/R_vocab": 0.2924107142857143
},
{
"learning_rate": 5.68761308646217e-08,
"loss": -0.0326,
"step": 1180,
"train/R_acc": 0.003125,
"train/R_penalty": -0.665625,
"train/R_reason": 0.16268249986159153,
"train/R_vocab": 0.17838541666666669
},
{
"learning_rate": 5.6190966283899774e-08,
"loss": 0.0286,
"step": 1182,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.2404998117876481,
"train/R_vocab": 0.34227430555555555
},
{
"learning_rate": 5.5509430810758817e-08,
"loss": 0.004,
"step": 1184,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.22824560961553947,
"train/R_vocab": 0.3375
},
{
"learning_rate": 5.483153720706798e-08,
"loss": -0.0074,
"step": 1186,
"train/R_acc": 0.1,
"train/R_penalty": -0.53125,
"train/R_reason": 0.21326554603657105,
"train/R_vocab": 0.20881696428571428
},
{
"learning_rate": 5.415729816650202e-08,
"loss": 0.0246,
"step": 1188,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.2444787127154557,
"train/R_vocab": 0.35230654761904756
},
{
"learning_rate": 5.3486726314303175e-08,
"loss": 0.0987,
"step": 1190,
"train/R_acc": 0.003125,
"train/R_penalty": -0.634375,
"train/R_reason": 0.22288328491646994,
"train/R_vocab": 0.2479910714285714
},
{
"learning_rate": 5.2819834207044974e-08,
"loss": -0.0208,
"step": 1192,
"train/R_acc": 0.128125,
"train/R_penalty": -0.4375,
"train/R_reason": 0.27609449087124366,
"train/R_vocab": 0.3087425595238095
},
{
"learning_rate": 5.215663433239728e-08,
"loss": -0.006,
"step": 1194,
"train/R_acc": 0.034375,
"train/R_penalty": -0.609375,
"train/R_reason": 0.22109386819415358,
"train/R_vocab": 0.24497767857142858
},
{
"learning_rate": 5.149713910889206e-08,
"loss": -0.0461,
"step": 1196,
"train/R_acc": 0.034375,
"train/R_penalty": -0.421875,
"train/R_reason": 0.2398768408665738,
"train/R_vocab": 0.3560267857142857
},
{
"learning_rate": 5.0841360885690996e-08,
"loss": 0.0184,
"step": 1198,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.5625,
"train/R_reason": 0.21147620601357403,
"train/R_vocab": 0.20399305555555555
},
{
"learning_rate": 5.01893119423546e-08,
"loss": 0.0505,
"step": 1200,
"train/R_acc": 0.065625,
"train/R_penalty": -0.671875,
"train/R_reason": 0.2487143037586896,
"train/R_vocab": 0.3544642857142857
},
{
"eval/R_acc": 0.0,
"eval/R_penalty": -0.5875,
"eval/R_reason": 0.20266877330422212,
"eval/R_vocab": 0.2709090909090909,
"step": 1200
},
{
"step": 1200
},
{
"learning_rate": 4.954100448861165e-08,
"loss": -0.009,
"step": 1202,
"train/R_acc": 0.034375,
"train/R_penalty": -0.446875,
"train/R_reason": 0.2773491666866028,
"train/R_vocab": 0.29509943181818177
},
{
"learning_rate": 4.889645066413112e-08,
"loss": 0.0306,
"step": 1204,
"train/R_acc": 0.003125,
"train/R_penalty": -0.734375,
"train/R_reason": 0.2607235600350966,
"train/R_vocab": 0.1830729166666667
},
{
"learning_rate": 4.82556625382945e-08,
"loss": -0.0188,
"step": 1206,
"train/R_acc": 0.03125,
"train/R_penalty": -0.665625,
"train/R_reason": 0.20154795093112635,
"train/R_vocab": 0.19722222222222224
},
{
"learning_rate": 4.76186521099699e-08,
"loss": -0.0385,
"step": 1208,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.2240002350916589,
"train/R_vocab": 0.3001116071428571
},
{
"learning_rate": 4.698543130728755e-08,
"loss": 0.0571,
"step": 1210,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.15973930176485518,
"train/R_vocab": 0.21350446428571426
},
{
"learning_rate": 4.635601198741607e-08,
"loss": 0.0042,
"step": 1212,
"train/R_acc": 0.06875,
"train/R_penalty": -0.5,
"train/R_reason": 0.19491433422937576,
"train/R_vocab": 0.24547483766233766
},
{
"learning_rate": 4.573040593634092e-08,
"loss": 0.0204,
"step": 1214,
"train/R_acc": 0.0,
"train/R_penalty": -0.7125,
"train/R_reason": 0.2036836381635984,
"train/R_vocab": 0.2583333333333333
},
{
"learning_rate": 4.510862486864317e-08,
"loss": -0.0139,
"step": 1216,
"train/R_acc": 0.03125,
"train/R_penalty": -0.5,
"train/R_reason": 0.1935253778055071,
"train/R_vocab": 0.2539434523809524
},
{
"learning_rate": 4.449068042728063e-08,
"loss": -0.0221,
"step": 1218,
"train/R_acc": 0.03125,
"train/R_penalty": -0.671875,
"train/R_reason": 0.17478719789421077,
"train/R_vocab": 0.2385044642857143
},
{
"learning_rate": 4.3876584183369454e-08,
"loss": 0.0401,
"step": 1220,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.12396243023486506,
"train/R_vocab": 0.16953125
},
{
"learning_rate": 4.326634763596784e-08,
"loss": 0.0362,
"step": 1222,
"train/R_acc": 0.003125,
"train/R_penalty": -0.634375,
"train/R_reason": 0.21970153796136732,
"train/R_vocab": 0.23229166666666667
},
{
"learning_rate": 4.265998221186023e-08,
"loss": 0.0579,
"step": 1224,
"train/R_acc": 0.003125,
"train/R_penalty": -0.696875,
"train/R_reason": 0.25118929349410624,
"train/R_vocab": 0.3527901785714286
},
{
"learning_rate": 4.205749926534391e-08,
"loss": 0.0325,
"step": 1226,
"train/R_acc": 0.0,
"train/R_penalty": -0.53125,
"train/R_reason": 0.20960885628620934,
"train/R_vocab": 0.24062499999999998
},
{
"learning_rate": 4.145891007801588e-08,
"loss": -0.012,
"step": 1228,
"train/R_acc": 0.037500000000000006,
"train/R_penalty": -0.375,
"train/R_reason": 0.2440708995563252,
"train/R_vocab": 0.31674107142857144
},
{
"learning_rate": 4.086422585856189e-08,
"loss": 0.0635,
"step": 1230,
"train/R_acc": 0.03125,
"train/R_penalty": -0.5,
"train/R_reason": 0.16689617467212345,
"train/R_vocab": 0.24068700396825393
},
{
"learning_rate": 4.0273457742546566e-08,
"loss": -0.0426,
"step": 1232,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.192806526715886,
"train/R_vocab": 0.2780505952380953
},
{
"learning_rate": 3.968661679220467e-08,
"loss": 0.0384,
"step": 1234,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.17224863139672328,
"train/R_vocab": 0.23426339285714287
},
{
"learning_rate": 3.9103713996234355e-08,
"loss": 0.0036,
"step": 1236,
"train/R_acc": 0.03125,
"train/R_penalty": -0.734375,
"train/R_reason": 0.17680126111067432,
"train/R_vocab": 0.2612723214285714
},
{
"learning_rate": 3.852476026959089e-08,
"loss": 0.0172,
"step": 1238,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.640625,
"train/R_reason": 0.23625775975212537,
"train/R_vocab": 0.2917410714285715
},
{
"learning_rate": 3.794976645328266e-08,
"loss": 0.0608,
"step": 1240,
"train/R_acc": 0.00625,
"train/R_penalty": -0.49375,
"train/R_reason": 0.20085612003970227,
"train/R_vocab": 0.29375
},
{
"learning_rate": 3.737874331416807e-08,
"loss": -0.0099,
"step": 1242,
"train/R_acc": 0.06875,
"train/R_penalty": -0.53125,
"train/R_reason": 0.2067606031031175,
"train/R_vocab": 0.1792782738095238
},
{
"learning_rate": 3.681170154475391e-08,
"loss": -0.0105,
"step": 1244,
"train/R_acc": 0.003125,
"train/R_penalty": -0.609375,
"train/R_reason": 0.24829657601476124,
"train/R_vocab": 0.3546875
},
{
"learning_rate": 3.624865176299499e-08,
"loss": -0.0273,
"step": 1246,
"train/R_acc": 0.003125,
"train/R_penalty": -0.484375,
"train/R_reason": 0.2728674749829552,
"train/R_vocab": 0.42109375000000004
},
{
"learning_rate": 3.568960451209574e-08,
"loss": 0.0602,
"step": 1248,
"train/R_acc": 0.0375,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2555940677098582,
"train/R_vocab": 0.41767113095238095
},
{
"learning_rate": 3.513457026031216e-08,
"loss": -0.0627,
"step": 1250,
"train/R_acc": 0.0,
"train/R_penalty": -0.53125,
"train/R_reason": 0.20621535738161093,
"train/R_vocab": 0.16438492063492063
},
{
"learning_rate": 3.458355940075652e-08,
"loss": 0.0626,
"step": 1252,
"train/R_acc": 0.03125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.25723943167289565,
"train/R_vocab": 0.290625
},
{
"learning_rate": 3.403658225120212e-08,
"loss": -0.0653,
"step": 1254,
"train/R_acc": 0.03125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.2280530069453393,
"train/R_vocab": 0.26473214285714286
},
{
"learning_rate": 3.349364905389032e-08,
"loss": -0.013,
"step": 1256,
"train/R_acc": 0.034375,
"train/R_penalty": -0.4375,
"train/R_reason": 0.20137595355438587,
"train/R_vocab": 0.3265625
},
{
"learning_rate": 3.295476997533905e-08,
"loss": 0.0298,
"step": 1258,
"train/R_acc": 0.034375,
"train/R_penalty": -0.59375,
"train/R_reason": 0.20510172934689136,
"train/R_vocab": 0.3380580357142857
},
{
"learning_rate": 3.2419955106151893e-08,
"loss": -0.0064,
"step": 1260,
"train/R_acc": 0.00625,
"train/R_penalty": -0.53125,
"train/R_reason": 0.24581030179992447,
"train/R_vocab": 0.27968750000000003
},
{
"learning_rate": 3.188921446082946e-08,
"loss": 0.0136,
"step": 1262,
"train/R_acc": 0.0,
"train/R_penalty": -0.5,
"train/R_reason": 0.19498686377833271,
"train/R_vocab": 0.21171875
},
{
"learning_rate": 3.1362557977582e-08,
"loss": 0.0327,
"step": 1264,
"train/R_acc": 0.003125,
"train/R_penalty": -0.65625,
"train/R_reason": 0.2546377151605037,
"train/R_vocab": 0.38132440476190477
},
{
"learning_rate": 3.0839995518142844e-08,
"loss": 0.0813,
"step": 1266,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.578125,
"train/R_reason": 0.23407475867218072,
"train/R_vocab": 0.25262896825396824
},
{
"learning_rate": 3.032153686758432e-08,
"loss": 0.0111,
"step": 1268,
"train/R_acc": 0.0,
"train/R_penalty": -0.75,
"train/R_reason": 0.13893382364177903,
"train/R_vocab": 0.1873139880952381
},
{
"learning_rate": 2.980719173413396e-08,
"loss": -0.0655,
"step": 1270,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.241975792190938,
"train/R_vocab": 0.359375
},
{
"learning_rate": 2.9296969748993093e-08,
"loss": 0.0107,
"step": 1272,
"train/R_acc": 0.03125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.24308125423194502,
"train/R_vocab": 0.18370535714285716
},
{
"learning_rate": 2.8790880466156464e-08,
"loss": 0.0368,
"step": 1274,
"train/R_acc": 0.0,
"train/R_penalty": -0.571875,
"train/R_reason": 0.21833878646152022,
"train/R_vocab": 0.2832217261904762
},
{
"learning_rate": 2.8288933362233148e-08,
"loss": 0.0016,
"step": 1276,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.205687843726902,
"train/R_vocab": 0.23705357142857142
},
{
"learning_rate": 2.7791137836269158e-08,
"loss": -0.0029,
"step": 1278,
"train/R_acc": 0.003125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.2181107503992247,
"train/R_vocab": 0.20520833333333333
},
{
"learning_rate": 2.7297503209571665e-08,
"loss": 0.1301,
"step": 1280,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.16072625927287004,
"train/R_vocab": 0.22611607142857143
},
{
"learning_rate": 2.680803872553408e-08,
"loss": 0.0204,
"step": 1282,
"train/R_acc": 0.00625,
"train/R_penalty": -0.525,
"train/R_reason": 0.18120856948997982,
"train/R_vocab": 0.2267981150793651
},
{
"learning_rate": 2.632275354946342e-08,
"loss": 0.0174,
"step": 1284,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.55625,
"train/R_reason": 0.19887443410752542,
"train/R_vocab": 0.20744047619047618
},
{
"learning_rate": 2.5841656768408216e-08,
"loss": 0.0125,
"step": 1286,
"train/R_acc": 0.003125,
"train/R_penalty": -0.53125,
"train/R_reason": 0.23872779990823945,
"train/R_vocab": 0.28046875
},
{
"learning_rate": 2.5364757390988616e-08,
"loss": 0.0652,
"step": 1288,
"train/R_acc": 0.003125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.19387491540524032,
"train/R_vocab": 0.23203125
},
{
"learning_rate": 2.489206434722785e-08,
"loss": -0.012,
"step": 1290,
"train/R_acc": 0.0,
"train/R_penalty": -0.515625,
"train/R_reason": 0.18941429133549656,
"train/R_vocab": 0.23415178571428572
},
{
"learning_rate": 2.4423586488384524e-08,
"loss": 0.0179,
"step": 1292,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.248278845576544,
"train/R_vocab": 0.22329545454545455
},
{
"learning_rate": 2.395933258678745e-08,
"loss": 0.0178,
"step": 1294,
"train/R_acc": 0.0,
"train/R_penalty": -0.609375,
"train/R_reason": 0.20548558553050647,
"train/R_vocab": 0.3214905753968254
},
{
"learning_rate": 2.349931133567096e-08,
"loss": -0.0004,
"step": 1296,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.2386328952490884,
"train/R_vocab": 0.23169642857142858
},
{
"learning_rate": 2.3043531349012353e-08,
"loss": -0.0659,
"step": 1298,
"train/R_acc": 0.034375,
"train/R_penalty": -0.671875,
"train/R_reason": 0.15007159038519916,
"train/R_vocab": 0.19921875
},
{
"learning_rate": 2.259200116137039e-08,
"loss": 0.0095,
"step": 1300,
"train/R_acc": 0.0,
"train/R_penalty": -0.65625,
"train/R_reason": 0.20368911805265966,
"train/R_vocab": 0.2474330357142857
},
{
"eval/R_acc": 0.0025,
"eval/R_penalty": -0.5625,
"eval/R_reason": 0.22623212762462228,
"eval/R_vocab": 0.3164285714285714,
"step": 1300
},
{
"step": 1300
},
{
"learning_rate": 2.214472922772584e-08,
"loss": 0.0957,
"step": 1302,
"train/R_acc": 0.003125,
"train/R_penalty": -0.625,
"train/R_reason": 0.22116478834904366,
"train/R_vocab": 0.29776785714285714
},
{
"learning_rate": 2.1701723923322673e-08,
"loss": -0.0749,
"step": 1304,
"train/R_acc": 0.065625,
"train/R_penalty": -0.53125,
"train/R_reason": 0.21324021225069947,
"train/R_vocab": 0.2928977272727273
},
{
"learning_rate": 2.1262993543511715e-08,
"loss": 9.305,
"step": 1306,
"train/R_acc": 0.0375,
"train/R_penalty": -0.484375,
"train/R_reason": 0.23529231846467052,
"train/R_vocab": 0.3332589285714286
},
{
"learning_rate": 2.082854630359493e-08,
"loss": 0.0331,
"step": 1308,
"train/R_acc": 0.0,
"train/R_penalty": -0.421875,
"train/R_reason": 0.23986047807887131,
"train/R_vocab": 0.36618303571428573
},
{
"learning_rate": 2.03983903386718e-08,
"loss": 0.0118,
"step": 1310,
"train/R_acc": 0.0,
"train/R_penalty": -0.578125,
"train/R_reason": 0.2656667530369191,
"train/R_vocab": 0.22192460317460316
},
{
"learning_rate": 1.997253370348703e-08,
"loss": 0.0556,
"step": 1312,
"train/R_acc": 0.0,
"train/R_penalty": -0.546875,
"train/R_reason": 0.2476736629649698,
"train/R_vocab": 0.31614583333333335
},
{
"learning_rate": 1.9550984372279476e-08,
"loss": -0.037,
"step": 1314,
"train/R_acc": 0.00625,
"train/R_penalty": -0.53125,
"train/R_reason": 0.19878604497907437,
"train/R_vocab": 0.27142857142857146
},
{
"learning_rate": 1.9133750238633117e-08,
"loss": 0.0175,
"step": 1316,
"train/R_acc": 0.00625,
"train/R_penalty": -0.634375,
"train/R_reason": 0.2233584354922284,
"train/R_vocab": 0.23705357142857142
},
{
"learning_rate": 1.872083911532907e-08,
"loss": 0.0006,
"step": 1318,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.53125,
"train/R_reason": 0.2794139552851759,
"train/R_vocab": 0.37043650793650795
},
{
"learning_rate": 1.8312258734199253e-08,
"loss": -0.0277,
"step": 1320,
"train/R_acc": 0.03125,
"train/R_penalty": -0.46875,
"train/R_reason": 0.19046029498161032,
"train/R_vocab": 0.2759548611111111
},
{
"learning_rate": 1.7908016745981856e-08,
"loss": 0.0615,
"step": 1322,
"train/R_acc": 0.00625,
"train/R_penalty": -0.515625,
"train/R_reason": 0.1939617321953623,
"train/R_vocab": 0.2484623015873016
},
{
"learning_rate": 1.7508120720177795e-08,
"loss": -0.0239,
"step": 1324,
"train/R_acc": 0.00625,
"train/R_penalty": -0.53125,
"train/R_reason": 0.21866868750973806,
"train/R_vocab": 0.2526785714285714
},
{
"learning_rate": 1.7112578144909047e-08,
"loss": -0.0744,
"step": 1326,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.23171087837305318,
"train/R_vocab": 0.28158482142857144
},
{
"learning_rate": 1.6721396426778667e-08,
"loss": 0.0113,
"step": 1328,
"train/R_acc": 0.00625,
"train/R_penalty": -0.49375,
"train/R_reason": 0.2527234820261308,
"train/R_vocab": 0.290687003968254
},
{
"learning_rate": 1.6334582890731697e-08,
"loss": 0.0423,
"step": 1330,
"train/R_acc": 0.0,
"train/R_penalty": -0.625,
"train/R_reason": 0.18629708667365014,
"train/R_vocab": 0.1819940476190476
},
{
"learning_rate": 1.595214477991841e-08,
"loss": 0.027,
"step": 1332,
"train/R_acc": 0.065625,
"train/R_penalty": -0.5625,
"train/R_reason": 0.23647010980361977,
"train/R_vocab": 0.2343625992063492
},
{
"learning_rate": 1.557408925555831e-08,
"loss": 0.0619,
"step": 1334,
"train/R_acc": 0.03125,
"train/R_penalty": -0.640625,
"train/R_reason": 0.2178143312883945,
"train/R_vocab": 0.28158482142857144
},
{
"learning_rate": 1.5200423396806272e-08,
"loss": -0.006,
"step": 1336,
"train/R_acc": 0.003125,
"train/R_penalty": -0.40625,
"train/R_reason": 0.26096178007275994,
"train/R_vocab": 0.3358134920634921
},
{
"learning_rate": 1.4831154200620033e-08,
"loss": -0.0105,
"step": 1338,
"train/R_acc": 0.034375,
"train/R_penalty": -0.578125,
"train/R_reason": 0.18847235645846208,
"train/R_vocab": 0.23511904761904762
},
{
"learning_rate": 1.4466288581628906e-08,
"loss": 0.0143,
"step": 1340,
"train/R_acc": 0.009375000000000001,
"train/R_penalty": -0.578125,
"train/R_reason": 0.23179656802725537,
"train/R_vocab": 0.2564732142857143
},
{
"learning_rate": 1.4105833372004523e-08,
"loss": 0.0246,
"step": 1342,
"train/R_acc": 0.03125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.1816783870148443,
"train/R_vocab": 0.21495535714285713
},
{
"learning_rate": 1.3749795321332885e-08,
"loss": -0.0062,
"step": 1344,
"train/R_acc": 0.03125,
"train/R_penalty": -0.546875,
"train/R_reason": 0.22414561936584018,
"train/R_vocab": 0.2232886904761905
},
{
"learning_rate": 1.3398181096487882e-08,
"loss": 0.0182,
"step": 1346,
"train/R_acc": 0.065625,
"train/R_penalty": -0.5625,
"train/R_reason": 0.2680587101569149,
"train/R_vocab": 0.3170138888888889
},
{
"learning_rate": 1.3050997281506549e-08,
"loss": 0.0831,
"step": 1348,
"train/R_acc": 0.003125,
"train/R_penalty": -0.68125,
"train/R_reason": 0.2494493257892117,
"train/R_vocab": 0.18958333333333333
},
{
"learning_rate": 1.2708250377465645e-08,
"loss": -0.004,
"step": 1350,
"train/R_acc": 0.00625,
"train/R_penalty": -0.665625,
"train/R_reason": 0.2093549139749035,
"train/R_vocab": 0.1791497564935065
}
],
"logging_steps": 2,
"max_steps": 1497,
"num_input_tokens_seen": 13748095,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}