| [ | |
| { | |
| "current_steps": 0, | |
| "loss": 3.6926, | |
| "learning_rate": 0.0, | |
| "epoch": 0.01 | |
| }, | |
| { | |
| "current_steps": 1, | |
| "loss": 3.0107, | |
| "learning_rate": 0.0, | |
| "epoch": 0.03 | |
| }, | |
| { | |
| "current_steps": 2, | |
| "loss": 3.0688, | |
| "learning_rate": 0.0, | |
| "epoch": 0.04 | |
| }, | |
| { | |
| "current_steps": 3, | |
| "loss": 4.0176, | |
| "learning_rate": 0.0, | |
| "epoch": 0.06 | |
| }, | |
| { | |
| "current_steps": 4, | |
| "loss": 4.0609, | |
| "learning_rate": 0.0, | |
| "epoch": 0.07 | |
| }, | |
| { | |
| "current_steps": 5, | |
| "loss": 3.754, | |
| "learning_rate": 0.0, | |
| "epoch": 0.09 | |
| }, | |
| { | |
| "current_steps": 6, | |
| "loss": 3.1981, | |
| "learning_rate": 0.0, | |
| "epoch": 0.1 | |
| }, | |
| { | |
| "current_steps": 7, | |
| "loss": 2.7138, | |
| "learning_rate": 0.0, | |
| "epoch": 0.11 | |
| }, | |
| { | |
| "current_steps": 8, | |
| "loss": 3.8803, | |
| "learning_rate": 0.0, | |
| "epoch": 0.13 | |
| }, | |
| { | |
| "current_steps": 9, | |
| "loss": 3.5793, | |
| "learning_rate": 2.9999999999999997e-06, | |
| "epoch": 0.14 | |
| }, | |
| { | |
| "current_steps": 10, | |
| "loss": 2.8589, | |
| "learning_rate": 5.999999999999999e-06, | |
| "epoch": 0.16 | |
| }, | |
| { | |
| "current_steps": 11, | |
| "loss": 2.6536, | |
| "learning_rate": 8.999999999999999e-06, | |
| "epoch": 0.17 | |
| }, | |
| { | |
| "current_steps": 12, | |
| "loss": 4.1845, | |
| "learning_rate": 1.1999999999999999e-05, | |
| "epoch": 0.19 | |
| }, | |
| { | |
| "current_steps": 13, | |
| "loss": 3.7447, | |
| "learning_rate": 1.4999999999999999e-05, | |
| "epoch": 0.2 | |
| }, | |
| { | |
| "current_steps": 14, | |
| "loss": 2.9558, | |
| "learning_rate": 1.7999999999999997e-05, | |
| "epoch": 0.21 | |
| }, | |
| { | |
| "current_steps": 15, | |
| "loss": 3.7552, | |
| "learning_rate": 2.1e-05, | |
| "epoch": 0.23 | |
| }, | |
| { | |
| "current_steps": 16, | |
| "loss": 3.7336, | |
| "learning_rate": 2.3999999999999997e-05, | |
| "epoch": 0.24 | |
| }, | |
| { | |
| "current_steps": 17, | |
| "loss": 2.484, | |
| "learning_rate": 2.6999999999999996e-05, | |
| "epoch": 0.26 | |
| }, | |
| { | |
| "current_steps": 18, | |
| "loss": 3.7203, | |
| "learning_rate": 2.9999999999999997e-05, | |
| "epoch": 0.27 | |
| }, | |
| { | |
| "current_steps": 19, | |
| "loss": 3.6129, | |
| "learning_rate": 2.9999999999999997e-05, | |
| "epoch": 0.29 | |
| }, | |
| { | |
| "current_steps": 20, | |
| "loss": 2.2497, | |
| "learning_rate": 3.2999999999999996e-05, | |
| "epoch": 0.3 | |
| }, | |
| { | |
| "current_steps": 21, | |
| "loss": 3.2732, | |
| "learning_rate": 3.5999999999999994e-05, | |
| "epoch": 0.31 | |
| }, | |
| { | |
| "current_steps": 22, | |
| "loss": 2.4979, | |
| "learning_rate": 3.9e-05, | |
| "epoch": 0.33 | |
| }, | |
| { | |
| "current_steps": 23, | |
| "loss": 3.043, | |
| "learning_rate": 4.2e-05, | |
| "epoch": 0.34 | |
| }, | |
| { | |
| "current_steps": 24, | |
| "loss": 3.2435, | |
| "learning_rate": 4.4999999999999996e-05, | |
| "epoch": 0.36 | |
| }, | |
| { | |
| "current_steps": 25, | |
| "loss": 3.3493, | |
| "learning_rate": 4.7999999999999994e-05, | |
| "epoch": 0.37 | |
| }, | |
| { | |
| "current_steps": 26, | |
| "loss": 2.6339, | |
| "learning_rate": 5.1e-05, | |
| "epoch": 0.39 | |
| }, | |
| { | |
| "current_steps": 27, | |
| "loss": 2.9486, | |
| "learning_rate": 5.399999999999999e-05, | |
| "epoch": 0.4 | |
| }, | |
| { | |
| "current_steps": 28, | |
| "loss": 2.3865, | |
| "learning_rate": 5.6999999999999996e-05, | |
| "epoch": 0.41 | |
| }, | |
| { | |
| "current_steps": 29, | |
| "loss": 2.3687, | |
| "learning_rate": 5.9999999999999995e-05, | |
| "epoch": 0.43 | |
| }, | |
| { | |
| "current_steps": 30, | |
| "loss": 2.3417, | |
| "learning_rate": 6.299999999999999e-05, | |
| "epoch": 0.44 | |
| }, | |
| { | |
| "current_steps": 31, | |
| "loss": 2.5443, | |
| "learning_rate": 6.599999999999999e-05, | |
| "epoch": 0.46 | |
| }, | |
| { | |
| "current_steps": 32, | |
| "loss": 2.4717, | |
| "learning_rate": 6.9e-05, | |
| "epoch": 0.47 | |
| }, | |
| { | |
| "current_steps": 33, | |
| "loss": 2.2291, | |
| "learning_rate": 7.199999999999999e-05, | |
| "epoch": 0.49 | |
| }, | |
| { | |
| "current_steps": 34, | |
| "loss": 2.4815, | |
| "learning_rate": 7.5e-05, | |
| "epoch": 0.5 | |
| }, | |
| { | |
| "current_steps": 35, | |
| "loss": 2.4468, | |
| "learning_rate": 7.8e-05, | |
| "epoch": 0.51 | |
| }, | |
| { | |
| "current_steps": 36, | |
| "loss": 2.4246, | |
| "learning_rate": 8.1e-05, | |
| "epoch": 0.53 | |
| }, | |
| { | |
| "current_steps": 37, | |
| "loss": 2.3154, | |
| "learning_rate": 8.4e-05, | |
| "epoch": 0.54 | |
| }, | |
| { | |
| "current_steps": 38, | |
| "loss": 1.9939, | |
| "learning_rate": 8.699999999999999e-05, | |
| "epoch": 0.56 | |
| }, | |
| { | |
| "current_steps": 39, | |
| "loss": 1.8099, | |
| "learning_rate": 8.999999999999999e-05, | |
| "epoch": 0.57 | |
| }, | |
| { | |
| "current_steps": 40, | |
| "loss": 1.865, | |
| "learning_rate": 9.3e-05, | |
| "epoch": 0.59 | |
| }, | |
| { | |
| "current_steps": 41, | |
| "loss": 1.4322, | |
| "learning_rate": 9.599999999999999e-05, | |
| "epoch": 0.6 | |
| }, | |
| { | |
| "current_steps": 42, | |
| "loss": 2.3252, | |
| "learning_rate": 9.9e-05, | |
| "epoch": 0.61 | |
| }, | |
| { | |
| "current_steps": 43, | |
| "loss": 1.7521, | |
| "learning_rate": 0.000102, | |
| "epoch": 0.63 | |
| }, | |
| { | |
| "current_steps": 44, | |
| "loss": 2.145, | |
| "learning_rate": 0.00010499999999999999, | |
| "epoch": 0.64 | |
| }, | |
| { | |
| "current_steps": 45, | |
| "loss": 1.6641, | |
| "learning_rate": 0.00010799999999999998, | |
| "epoch": 0.66 | |
| }, | |
| { | |
| "current_steps": 46, | |
| "loss": 1.8938, | |
| "learning_rate": 0.00011099999999999999, | |
| "epoch": 0.67 | |
| }, | |
| { | |
| "current_steps": 47, | |
| "loss": 1.5194, | |
| "learning_rate": 0.00011399999999999999, | |
| "epoch": 0.69 | |
| }, | |
| { | |
| "current_steps": 48, | |
| "loss": 1.4478, | |
| "learning_rate": 0.000117, | |
| "epoch": 0.7 | |
| }, | |
| { | |
| "current_steps": 49, | |
| "loss": 1.9414, | |
| "learning_rate": 0.00011999999999999999, | |
| "epoch": 0.71 | |
| }, | |
| { | |
| "current_steps": 50, | |
| "loss": 1.5601, | |
| "learning_rate": 0.00012299999999999998, | |
| "epoch": 0.73 | |
| }, | |
| { | |
| "current_steps": 51, | |
| "loss": 1.9015, | |
| "learning_rate": 0.00012599999999999997, | |
| "epoch": 0.74 | |
| }, | |
| { | |
| "current_steps": 52, | |
| "loss": 1.7384, | |
| "learning_rate": 0.000129, | |
| "epoch": 0.76 | |
| }, | |
| { | |
| "current_steps": 53, | |
| "loss": 1.7282, | |
| "learning_rate": 0.00013199999999999998, | |
| "epoch": 0.77 | |
| }, | |
| { | |
| "current_steps": 54, | |
| "loss": 1.6877, | |
| "learning_rate": 0.000135, | |
| "epoch": 0.79 | |
| }, | |
| { | |
| "current_steps": 55, | |
| "loss": 1.0647, | |
| "learning_rate": 0.000138, | |
| "epoch": 0.8 | |
| }, | |
| { | |
| "current_steps": 56, | |
| "loss": 1.755, | |
| "learning_rate": 0.00014099999999999998, | |
| "epoch": 0.81 | |
| }, | |
| { | |
| "current_steps": 57, | |
| "loss": 1.516, | |
| "learning_rate": 0.00014399999999999998, | |
| "epoch": 0.83 | |
| }, | |
| { | |
| "current_steps": 58, | |
| "loss": 1.3498, | |
| "learning_rate": 0.000147, | |
| "epoch": 0.84 | |
| }, | |
| { | |
| "current_steps": 59, | |
| "loss": 1.6789, | |
| "learning_rate": 0.00015, | |
| "epoch": 0.86 | |
| }, | |
| { | |
| "current_steps": 60, | |
| "loss": 1.3651, | |
| "learning_rate": 0.00015299999999999998, | |
| "epoch": 0.87 | |
| }, | |
| { | |
| "current_steps": 61, | |
| "loss": 1.2967, | |
| "learning_rate": 0.000156, | |
| "epoch": 0.89 | |
| }, | |
| { | |
| "current_steps": 62, | |
| "loss": 1.3998, | |
| "learning_rate": 0.000159, | |
| "epoch": 0.9 | |
| }, | |
| { | |
| "current_steps": 63, | |
| "loss": 1.1989, | |
| "learning_rate": 0.000162, | |
| "epoch": 0.91 | |
| }, | |
| { | |
| "current_steps": 64, | |
| "loss": 1.029, | |
| "learning_rate": 0.000165, | |
| "epoch": 0.93 | |
| }, | |
| { | |
| "current_steps": 65, | |
| "loss": 1.4441, | |
| "learning_rate": 0.000168, | |
| "epoch": 0.94 | |
| }, | |
| { | |
| "current_steps": 66, | |
| "loss": 1.1804, | |
| "learning_rate": 0.00017099999999999998, | |
| "epoch": 0.96 | |
| }, | |
| { | |
| "current_steps": 67, | |
| "loss": 1.2868, | |
| "learning_rate": 0.00017399999999999997, | |
| "epoch": 0.97 | |
| }, | |
| { | |
| "current_steps": 68, | |
| "loss": 1.1724, | |
| "learning_rate": 0.00017699999999999997, | |
| "epoch": 0.99 | |
| }, | |
| { | |
| "current_steps": 69, | |
| "loss": 1.4346, | |
| "learning_rate": 0.00017999999999999998, | |
| "epoch": 1.0 | |
| }, | |
| { | |
| "current_steps": 70, | |
| "loss": 1.0183, | |
| "learning_rate": 0.00018299999999999998, | |
| "epoch": 1.01 | |
| }, | |
| { | |
| "current_steps": 71, | |
| "loss": 1.1259, | |
| "learning_rate": 0.000186, | |
| "epoch": 1.03 | |
| }, | |
| { | |
| "current_steps": 72, | |
| "loss": 1.1713, | |
| "learning_rate": 0.00018899999999999999, | |
| "epoch": 1.04 | |
| }, | |
| { | |
| "current_steps": 73, | |
| "loss": 1.0773, | |
| "learning_rate": 0.00019199999999999998, | |
| "epoch": 1.06 | |
| }, | |
| { | |
| "current_steps": 74, | |
| "loss": 1.2956, | |
| "learning_rate": 0.000195, | |
| "epoch": 1.07 | |
| }, | |
| { | |
| "current_steps": 75, | |
| "loss": 0.9359, | |
| "learning_rate": 0.000198, | |
| "epoch": 1.09 | |
| }, | |
| { | |
| "current_steps": 76, | |
| "loss": 0.9838, | |
| "learning_rate": 0.000201, | |
| "epoch": 1.1 | |
| }, | |
| { | |
| "current_steps": 77, | |
| "loss": 0.9488, | |
| "learning_rate": 0.000204, | |
| "epoch": 1.11 | |
| }, | |
| { | |
| "current_steps": 78, | |
| "loss": 1.0748, | |
| "learning_rate": 0.00020699999999999996, | |
| "epoch": 1.13 | |
| }, | |
| { | |
| "current_steps": 79, | |
| "loss": 1.0914, | |
| "learning_rate": 0.00020999999999999998, | |
| "epoch": 1.14 | |
| }, | |
| { | |
| "current_steps": 80, | |
| "loss": 0.5364, | |
| "learning_rate": 0.00021299999999999997, | |
| "epoch": 1.16 | |
| }, | |
| { | |
| "current_steps": 81, | |
| "loss": 0.4909, | |
| "learning_rate": 0.00021599999999999996, | |
| "epoch": 1.17 | |
| }, | |
| { | |
| "current_steps": 82, | |
| "loss": 1.0176, | |
| "learning_rate": 0.00021899999999999998, | |
| "epoch": 1.19 | |
| }, | |
| { | |
| "current_steps": 83, | |
| "loss": 1.0543, | |
| "learning_rate": 0.00022199999999999998, | |
| "epoch": 1.2 | |
| }, | |
| { | |
| "current_steps": 84, | |
| "loss": 0.9092, | |
| "learning_rate": 0.000225, | |
| "epoch": 1.21 | |
| }, | |
| { | |
| "current_steps": 85, | |
| "loss": 0.5678, | |
| "learning_rate": 0.00022799999999999999, | |
| "epoch": 1.23 | |
| }, | |
| { | |
| "current_steps": 86, | |
| "loss": 0.8385, | |
| "learning_rate": 0.00023099999999999998, | |
| "epoch": 1.24 | |
| }, | |
| { | |
| "current_steps": 87, | |
| "loss": 1.6241, | |
| "learning_rate": 0.000234, | |
| "epoch": 1.26 | |
| }, | |
| { | |
| "current_steps": 88, | |
| "loss": 0.8294, | |
| "learning_rate": 0.000237, | |
| "epoch": 1.27 | |
| }, | |
| { | |
| "current_steps": 89, | |
| "loss": 1.2953, | |
| "learning_rate": 0.00023999999999999998, | |
| "epoch": 1.29 | |
| }, | |
| { | |
| "current_steps": 90, | |
| "loss": 0.5297, | |
| "learning_rate": 0.000243, | |
| "epoch": 1.3 | |
| }, | |
| { | |
| "current_steps": 91, | |
| "loss": 1.1592, | |
| "learning_rate": 0.00024599999999999996, | |
| "epoch": 1.31 | |
| }, | |
| { | |
| "current_steps": 92, | |
| "loss": 1.1055, | |
| "learning_rate": 0.000249, | |
| "epoch": 1.33 | |
| }, | |
| { | |
| "current_steps": 93, | |
| "loss": 0.9804, | |
| "learning_rate": 0.00025199999999999995, | |
| "epoch": 1.34 | |
| }, | |
| { | |
| "current_steps": 94, | |
| "loss": 0.8683, | |
| "learning_rate": 0.00025499999999999996, | |
| "epoch": 1.36 | |
| }, | |
| { | |
| "current_steps": 95, | |
| "loss": 0.9698, | |
| "learning_rate": 0.000258, | |
| "epoch": 1.37 | |
| }, | |
| { | |
| "current_steps": 96, | |
| "loss": 1.037, | |
| "learning_rate": 0.000261, | |
| "epoch": 1.39 | |
| }, | |
| { | |
| "current_steps": 97, | |
| "loss": 0.5479, | |
| "learning_rate": 0.00026399999999999997, | |
| "epoch": 1.4 | |
| }, | |
| { | |
| "current_steps": 98, | |
| "loss": 0.598, | |
| "learning_rate": 0.000267, | |
| "epoch": 1.41 | |
| }, | |
| { | |
| "current_steps": 99, | |
| "loss": 0.3627, | |
| "learning_rate": 0.00027, | |
| "epoch": 1.43 | |
| }, | |
| { | |
| "current_steps": 100, | |
| "loss": 0.8042, | |
| "learning_rate": 0.00027299999999999997, | |
| "epoch": 1.44 | |
| }, | |
| { | |
| "current_steps": 101, | |
| "loss": 1.0378, | |
| "learning_rate": 0.000276, | |
| "epoch": 1.46 | |
| }, | |
| { | |
| "current_steps": 102, | |
| "loss": 1.0192, | |
| "learning_rate": 0.000279, | |
| "epoch": 1.47 | |
| }, | |
| { | |
| "current_steps": 103, | |
| "loss": 1.4654, | |
| "learning_rate": 0.00028199999999999997, | |
| "epoch": 1.49 | |
| }, | |
| { | |
| "current_steps": 104, | |
| "loss": 1.0614, | |
| "learning_rate": 0.000285, | |
| "epoch": 1.5 | |
| }, | |
| { | |
| "current_steps": 105, | |
| "loss": 0.779, | |
| "learning_rate": 0.00028799999999999995, | |
| "epoch": 1.51 | |
| }, | |
| { | |
| "current_steps": 106, | |
| "loss": 1.0255, | |
| "learning_rate": 0.00029099999999999997, | |
| "epoch": 1.53 | |
| }, | |
| { | |
| "current_steps": 107, | |
| "loss": 1.0924, | |
| "learning_rate": 0.000294, | |
| "epoch": 1.54 | |
| }, | |
| { | |
| "current_steps": 108, | |
| "loss": 1.1016, | |
| "learning_rate": 0.00029699999999999996, | |
| "epoch": 1.56 | |
| }, | |
| { | |
| "current_steps": 109, | |
| "loss": 0.8748, | |
| "learning_rate": 0.0003, | |
| "epoch": 1.57 | |
| }, | |
| { | |
| "current_steps": 110, | |
| "loss": 0.8048, | |
| "learning_rate": 0.00029727272727272724, | |
| "epoch": 1.59 | |
| }, | |
| { | |
| "current_steps": 111, | |
| "loss": 1.0229, | |
| "learning_rate": 0.0002945454545454545, | |
| "epoch": 1.6 | |
| }, | |
| { | |
| "current_steps": 112, | |
| "loss": 0.9677, | |
| "learning_rate": 0.0002918181818181818, | |
| "epoch": 1.61 | |
| }, | |
| { | |
| "current_steps": 113, | |
| "loss": 0.5605, | |
| "learning_rate": 0.00028909090909090904, | |
| "epoch": 1.63 | |
| }, | |
| { | |
| "current_steps": 114, | |
| "loss": 0.9392, | |
| "learning_rate": 0.00028636363636363636, | |
| "epoch": 1.64 | |
| }, | |
| { | |
| "current_steps": 115, | |
| "loss": 1.0068, | |
| "learning_rate": 0.0002836363636363636, | |
| "epoch": 1.66 | |
| }, | |
| { | |
| "current_steps": 116, | |
| "loss": 0.898, | |
| "learning_rate": 0.0002809090909090909, | |
| "epoch": 1.67 | |
| }, | |
| { | |
| "current_steps": 117, | |
| "loss": 1.1297, | |
| "learning_rate": 0.00027818181818181815, | |
| "epoch": 1.69 | |
| }, | |
| { | |
| "current_steps": 118, | |
| "loss": 0.8696, | |
| "learning_rate": 0.0002754545454545454, | |
| "epoch": 1.7 | |
| }, | |
| { | |
| "current_steps": 119, | |
| "loss": 0.975, | |
| "learning_rate": 0.0002727272727272727, | |
| "epoch": 1.71 | |
| }, | |
| { | |
| "current_steps": 120, | |
| "loss": 1.107, | |
| "learning_rate": 0.00027, | |
| "epoch": 1.73 | |
| }, | |
| { | |
| "current_steps": 121, | |
| "loss": 1.1696, | |
| "learning_rate": 0.0002672727272727272, | |
| "epoch": 1.74 | |
| }, | |
| { | |
| "current_steps": 122, | |
| "loss": 0.7181, | |
| "learning_rate": 0.00026454545454545453, | |
| "epoch": 1.76 | |
| }, | |
| { | |
| "current_steps": 123, | |
| "loss": 0.4008, | |
| "learning_rate": 0.0002618181818181818, | |
| "epoch": 1.77 | |
| }, | |
| { | |
| "current_steps": 124, | |
| "loss": 0.6227, | |
| "learning_rate": 0.00025909090909090907, | |
| "epoch": 1.79 | |
| }, | |
| { | |
| "current_steps": 125, | |
| "loss": 1.4091, | |
| "learning_rate": 0.00025636363636363633, | |
| "epoch": 1.8 | |
| }, | |
| { | |
| "current_steps": 126, | |
| "loss": 0.7953, | |
| "learning_rate": 0.0002536363636363636, | |
| "epoch": 1.81 | |
| }, | |
| { | |
| "current_steps": 127, | |
| "loss": 1.1635, | |
| "learning_rate": 0.00025090909090909086, | |
| "epoch": 1.83 | |
| }, | |
| { | |
| "current_steps": 128, | |
| "loss": 0.6109, | |
| "learning_rate": 0.0002481818181818182, | |
| "epoch": 1.84 | |
| }, | |
| { | |
| "current_steps": 129, | |
| "loss": 0.7244, | |
| "learning_rate": 0.00024545454545454545, | |
| "epoch": 1.86 | |
| }, | |
| { | |
| "current_steps": 130, | |
| "loss": 1.0868, | |
| "learning_rate": 0.0002427272727272727, | |
| "epoch": 1.87 | |
| }, | |
| { | |
| "current_steps": 131, | |
| "loss": 1.1968, | |
| "learning_rate": 0.00023999999999999998, | |
| "epoch": 1.89 | |
| }, | |
| { | |
| "current_steps": 132, | |
| "loss": 1.109, | |
| "learning_rate": 0.00023727272727272724, | |
| "epoch": 1.9 | |
| }, | |
| { | |
| "current_steps": 133, | |
| "loss": 0.8744, | |
| "learning_rate": 0.00023454545454545454, | |
| "epoch": 1.91 | |
| }, | |
| { | |
| "current_steps": 134, | |
| "loss": 0.6971, | |
| "learning_rate": 0.0002318181818181818, | |
| "epoch": 1.93 | |
| }, | |
| { | |
| "current_steps": 135, | |
| "loss": 0.827, | |
| "learning_rate": 0.00022909090909090907, | |
| "epoch": 1.94 | |
| }, | |
| { | |
| "current_steps": 136, | |
| "loss": 0.8474, | |
| "learning_rate": 0.00022636363636363633, | |
| "epoch": 1.96 | |
| }, | |
| { | |
| "current_steps": 137, | |
| "loss": 0.8051, | |
| "learning_rate": 0.00022363636363636363, | |
| "epoch": 1.97 | |
| }, | |
| { | |
| "current_steps": 138, | |
| "loss": 0.8532, | |
| "learning_rate": 0.0002209090909090909, | |
| "epoch": 1.99 | |
| }, | |
| { | |
| "current_steps": 139, | |
| "loss": 0.8277, | |
| "learning_rate": 0.00021818181818181816, | |
| "epoch": 2.0 | |
| }, | |
| { | |
| "current_steps": 140, | |
| "loss": 0.3926, | |
| "learning_rate": 0.00021545454545454542, | |
| "epoch": 2.01 | |
| }, | |
| { | |
| "current_steps": 141, | |
| "loss": 0.5717, | |
| "learning_rate": 0.00021272727272727272, | |
| "epoch": 2.03 | |
| }, | |
| { | |
| "current_steps": 142, | |
| "loss": 0.6956, | |
| "learning_rate": 0.00020999999999999998, | |
| "epoch": 2.04 | |
| }, | |
| { | |
| "current_steps": 143, | |
| "loss": 0.6353, | |
| "learning_rate": 0.00020727272727272725, | |
| "epoch": 2.06 | |
| }, | |
| { | |
| "current_steps": 144, | |
| "loss": 0.4248, | |
| "learning_rate": 0.0002045454545454545, | |
| "epoch": 2.07 | |
| }, | |
| { | |
| "current_steps": 145, | |
| "loss": 0.6299, | |
| "learning_rate": 0.0002018181818181818, | |
| "epoch": 2.09 | |
| }, | |
| { | |
| "current_steps": 146, | |
| "loss": 0.3415, | |
| "learning_rate": 0.0001990909090909091, | |
| "epoch": 2.1 | |
| }, | |
| { | |
| "current_steps": 147, | |
| "loss": 0.4788, | |
| "learning_rate": 0.00019636363636363634, | |
| "epoch": 2.11 | |
| }, | |
| { | |
| "current_steps": 148, | |
| "loss": 0.6898, | |
| "learning_rate": 0.00019363636363636363, | |
| "epoch": 2.13 | |
| }, | |
| { | |
| "current_steps": 149, | |
| "loss": 0.4212, | |
| "learning_rate": 0.0001909090909090909, | |
| "epoch": 2.14 | |
| }, | |
| { | |
| "current_steps": 150, | |
| "loss": 0.6855, | |
| "learning_rate": 0.0001881818181818182, | |
| "epoch": 2.16 | |
| }, | |
| { | |
| "current_steps": 151, | |
| "loss": 0.4411, | |
| "learning_rate": 0.00018545454545454543, | |
| "epoch": 2.17 | |
| }, | |
| { | |
| "current_steps": 152, | |
| "loss": 0.4706, | |
| "learning_rate": 0.00018272727272727272, | |
| "epoch": 2.19 | |
| }, | |
| { | |
| "current_steps": 153, | |
| "loss": 0.6222, | |
| "learning_rate": 0.00017999999999999998, | |
| "epoch": 2.2 | |
| }, | |
| { | |
| "current_steps": 154, | |
| "loss": 0.6584, | |
| "learning_rate": 0.00017727272727272728, | |
| "epoch": 2.21 | |
| }, | |
| { | |
| "current_steps": 155, | |
| "loss": 0.6037, | |
| "learning_rate": 0.00017454545454545452, | |
| "epoch": 2.23 | |
| }, | |
| { | |
| "current_steps": 156, | |
| "loss": 0.6478, | |
| "learning_rate": 0.0001718181818181818, | |
| "epoch": 2.24 | |
| }, | |
| { | |
| "current_steps": 157, | |
| "loss": 0.4591, | |
| "learning_rate": 0.00016909090909090907, | |
| "epoch": 2.26 | |
| }, | |
| { | |
| "current_steps": 158, | |
| "loss": 0.4085, | |
| "learning_rate": 0.00016636363636363637, | |
| "epoch": 2.27 | |
| }, | |
| { | |
| "current_steps": 159, | |
| "loss": 0.6747, | |
| "learning_rate": 0.0001636363636363636, | |
| "epoch": 2.29 | |
| }, | |
| { | |
| "current_steps": 160, | |
| "loss": 0.4008, | |
| "learning_rate": 0.0001609090909090909, | |
| "epoch": 2.3 | |
| }, | |
| { | |
| "current_steps": 161, | |
| "loss": 0.671, | |
| "learning_rate": 0.00015818181818181816, | |
| "epoch": 2.31 | |
| }, | |
| { | |
| "current_steps": 162, | |
| "loss": 0.593, | |
| "learning_rate": 0.00015545454545454546, | |
| "epoch": 2.33 | |
| }, | |
| { | |
| "current_steps": 163, | |
| "loss": 0.4881, | |
| "learning_rate": 0.0001527272727272727, | |
| "epoch": 2.34 | |
| }, | |
| { | |
| "current_steps": 164, | |
| "loss": 0.6749, | |
| "learning_rate": 0.00015, | |
| "epoch": 2.36 | |
| }, | |
| { | |
| "current_steps": 165, | |
| "loss": 0.7904, | |
| "learning_rate": 0.00014727272727272725, | |
| "epoch": 2.37 | |
| }, | |
| { | |
| "current_steps": 166, | |
| "loss": 0.7036, | |
| "learning_rate": 0.00014454545454545452, | |
| "epoch": 2.39 | |
| }, | |
| { | |
| "current_steps": 167, | |
| "loss": 0.5043, | |
| "learning_rate": 0.0001418181818181818, | |
| "epoch": 2.4 | |
| }, | |
| { | |
| "current_steps": 168, | |
| "loss": 0.7245, | |
| "learning_rate": 0.00013909090909090908, | |
| "epoch": 2.41 | |
| }, | |
| { | |
| "current_steps": 169, | |
| "loss": 0.7101, | |
| "learning_rate": 0.00013636363636363634, | |
| "epoch": 2.43 | |
| }, | |
| { | |
| "current_steps": 170, | |
| "loss": 0.3898, | |
| "learning_rate": 0.0001336363636363636, | |
| "epoch": 2.44 | |
| }, | |
| { | |
| "current_steps": 171, | |
| "loss": 0.4639, | |
| "learning_rate": 0.0001309090909090909, | |
| "epoch": 2.46 | |
| }, | |
| { | |
| "current_steps": 172, | |
| "loss": 0.686, | |
| "learning_rate": 0.00012818181818181817, | |
| "epoch": 2.47 | |
| }, | |
| { | |
| "current_steps": 173, | |
| "loss": 0.4952, | |
| "learning_rate": 0.00012545454545454543, | |
| "epoch": 2.49 | |
| }, | |
| { | |
| "current_steps": 174, | |
| "loss": 0.2727, | |
| "learning_rate": 0.00012272727272727272, | |
| "epoch": 2.5 | |
| }, | |
| { | |
| "current_steps": 175, | |
| "loss": 0.3428, | |
| "learning_rate": 0.00011999999999999999, | |
| "epoch": 2.51 | |
| }, | |
| { | |
| "current_steps": 176, | |
| "loss": 0.253, | |
| "learning_rate": 0.00011727272727272727, | |
| "epoch": 2.53 | |
| }, | |
| { | |
| "current_steps": 177, | |
| "loss": 0.5778, | |
| "learning_rate": 0.00011454545454545453, | |
| "epoch": 2.54 | |
| }, | |
| { | |
| "current_steps": 178, | |
| "loss": 0.639, | |
| "learning_rate": 0.00011181818181818181, | |
| "epoch": 2.56 | |
| }, | |
| { | |
| "current_steps": 179, | |
| "loss": 0.7327, | |
| "learning_rate": 0.00010909090909090908, | |
| "epoch": 2.57 | |
| }, | |
| { | |
| "current_steps": 180, | |
| "loss": 0.4956, | |
| "learning_rate": 0.00010636363636363636, | |
| "epoch": 2.59 | |
| }, | |
| { | |
| "current_steps": 181, | |
| "loss": 0.364, | |
| "learning_rate": 0.00010363636363636362, | |
| "epoch": 2.6 | |
| }, | |
| { | |
| "current_steps": 182, | |
| "loss": 0.4497, | |
| "learning_rate": 0.0001009090909090909, | |
| "epoch": 2.61 | |
| }, | |
| { | |
| "current_steps": 183, | |
| "loss": 0.7653, | |
| "learning_rate": 9.818181818181817e-05, | |
| "epoch": 2.63 | |
| }, | |
| { | |
| "current_steps": 184, | |
| "loss": 0.6586, | |
| "learning_rate": 9.545454545454545e-05, | |
| "epoch": 2.64 | |
| }, | |
| { | |
| "current_steps": 185, | |
| "loss": 0.4404, | |
| "learning_rate": 9.272727272727271e-05, | |
| "epoch": 2.66 | |
| }, | |
| { | |
| "current_steps": 186, | |
| "loss": 0.7484, | |
| "learning_rate": 8.999999999999999e-05, | |
| "epoch": 2.67 | |
| }, | |
| { | |
| "current_steps": 187, | |
| "loss": 0.6176, | |
| "learning_rate": 8.727272727272726e-05, | |
| "epoch": 2.69 | |
| }, | |
| { | |
| "current_steps": 188, | |
| "loss": 0.7404, | |
| "learning_rate": 8.454545454545454e-05, | |
| "epoch": 2.7 | |
| }, | |
| { | |
| "current_steps": 189, | |
| "loss": 0.8124, | |
| "learning_rate": 8.18181818181818e-05, | |
| "epoch": 2.71 | |
| }, | |
| { | |
| "current_steps": 190, | |
| "loss": 0.6525, | |
| "learning_rate": 7.909090909090908e-05, | |
| "epoch": 2.73 | |
| }, | |
| { | |
| "current_steps": 191, | |
| "loss": 0.5986, | |
| "learning_rate": 7.636363636363635e-05, | |
| "epoch": 2.74 | |
| }, | |
| { | |
| "current_steps": 192, | |
| "loss": 0.294, | |
| "learning_rate": 7.363636363636363e-05, | |
| "epoch": 2.76 | |
| }, | |
| { | |
| "current_steps": 193, | |
| "loss": 0.2973, | |
| "learning_rate": 7.09090909090909e-05, | |
| "epoch": 2.77 | |
| }, | |
| { | |
| "current_steps": 194, | |
| "loss": 0.3589, | |
| "learning_rate": 6.818181818181817e-05, | |
| "epoch": 2.79 | |
| }, | |
| { | |
| "current_steps": 195, | |
| "loss": 0.78, | |
| "learning_rate": 6.545454545454545e-05, | |
| "epoch": 2.8 | |
| }, | |
| { | |
| "current_steps": 196, | |
| "loss": 0.4056, | |
| "learning_rate": 6.272727272727272e-05, | |
| "epoch": 2.81 | |
| }, | |
| { | |
| "current_steps": 197, | |
| "loss": 0.5843, | |
| "learning_rate": 5.9999999999999995e-05, | |
| "epoch": 2.83 | |
| }, | |
| { | |
| "current_steps": 198, | |
| "loss": 0.5745, | |
| "learning_rate": 5.727272727272727e-05, | |
| "epoch": 2.84 | |
| }, | |
| { | |
| "current_steps": 199, | |
| "loss": 0.2763, | |
| "learning_rate": 5.454545454545454e-05, | |
| "epoch": 2.86 | |
| }, | |
| { | |
| "current_steps": 200, | |
| "loss": 0.5494, | |
| "learning_rate": 5.181818181818181e-05, | |
| "epoch": 2.87 | |
| }, | |
| { | |
| "current_steps": 201, | |
| "loss": 0.4962, | |
| "learning_rate": 4.9090909090909084e-05, | |
| "epoch": 2.89 | |
| }, | |
| { | |
| "current_steps": 202, | |
| "loss": 0.5864, | |
| "learning_rate": 4.6363636363636356e-05, | |
| "epoch": 2.9 | |
| }, | |
| { | |
| "current_steps": 203, | |
| "loss": 0.3165, | |
| "learning_rate": 4.363636363636363e-05, | |
| "epoch": 2.91 | |
| }, | |
| { | |
| "current_steps": 204, | |
| "loss": 0.5969, | |
| "learning_rate": 4.09090909090909e-05, | |
| "epoch": 2.93 | |
| }, | |
| { | |
| "current_steps": 205, | |
| "loss": 0.3003, | |
| "learning_rate": 3.8181818181818174e-05, | |
| "epoch": 2.94 | |
| }, | |
| { | |
| "current_steps": 206, | |
| "loss": 0.6398, | |
| "learning_rate": 3.545454545454545e-05, | |
| "epoch": 2.96 | |
| }, | |
| { | |
| "current_steps": 207, | |
| "loss": 0.604, | |
| "learning_rate": 3.2727272727272725e-05, | |
| "epoch": 2.97 | |
| }, | |
| { | |
| "current_steps": 208, | |
| "loss": 0.6389, | |
| "learning_rate": 2.9999999999999997e-05, | |
| "epoch": 2.99 | |
| }, | |
| { | |
| "current_steps": 209, | |
| "loss": 0.5865, | |
| "learning_rate": 2.727272727272727e-05, | |
| "epoch": 3.0 | |
| }, | |
| { | |
| "current_steps": 209, | |
| "loss": 0.5865, | |
| "learning_rate": 2.727272727272727e-05, | |
| "epoch": 3.0 | |
| } | |
| ] |