| { | |
| "best_metric": 0.5186672210693359, | |
| "best_model_checkpoint": "/workspace/dolphin-2.6-mistral-7b/checkpoint-730", | |
| "epoch": 2.3532627263409633, | |
| "eval_steps": 73, | |
| "global_step": 876, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 5.000000000000001e-07, | |
| "loss": 1.1255, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 1.1614, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.5e-06, | |
| "loss": 1.1455, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.0631, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.9595, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3e-06, | |
| "loss": 0.9296, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.5e-06, | |
| "loss": 0.8787, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.8616, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.5e-06, | |
| "loss": 0.852, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8329, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.999994132223223e-06, | |
| "loss": 0.8109, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.999976528920434e-06, | |
| "loss": 0.7917, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.999947190174268e-06, | |
| "loss": 0.8021, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.999906116122448e-06, | |
| "loss": 0.7972, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.999853306957783e-06, | |
| "loss": 0.7847, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.9997887629281724e-06, | |
| "loss": 0.7669, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.9997124843366004e-06, | |
| "loss": 0.7892, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.999624471541135e-06, | |
| "loss": 0.7752, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.999524724954928e-06, | |
| "loss": 0.7702, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.99941324504621e-06, | |
| "loss": 0.7345, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.999290032338296e-06, | |
| "loss": 0.7482, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9991550874095715e-06, | |
| "loss": 0.745, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9990084108934975e-06, | |
| "loss": 0.7291, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.998850003478607e-06, | |
| "loss": 0.7275, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.998679865908499e-06, | |
| "loss": 0.7432, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.998497998981838e-06, | |
| "loss": 0.718, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.998304403552346e-06, | |
| "loss": 0.7137, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.998099080528803e-06, | |
| "loss": 0.7057, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.997882030875043e-06, | |
| "loss": 0.7137, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.9976532556099425e-06, | |
| "loss": 0.6866, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.9974127558074236e-06, | |
| "loss": 0.6794, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.997160532596447e-06, | |
| "loss": 0.6774, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.996896587161004e-06, | |
| "loss": 0.6922, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.996620920740111e-06, | |
| "loss": 0.6962, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.99633353462781e-06, | |
| "loss": 0.687, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.996034430173154e-06, | |
| "loss": 0.6773, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.995723608780204e-06, | |
| "loss": 0.6822, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.995401071908026e-06, | |
| "loss": 0.6876, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.9950668210706795e-06, | |
| "loss": 0.6959, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.994720857837211e-06, | |
| "loss": 0.6893, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.99436318383165e-06, | |
| "loss": 0.6734, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.9939938007329955e-06, | |
| "loss": 0.6591, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.993612710275216e-06, | |
| "loss": 0.6988, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.993219914247233e-06, | |
| "loss": 0.66, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.992815414492917e-06, | |
| "loss": 0.6642, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.992399212911082e-06, | |
| "loss": 0.6652, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.991971311455468e-06, | |
| "loss": 0.6777, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.99153171213474e-06, | |
| "loss": 0.6647, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.991080417012476e-06, | |
| "loss": 0.6737, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.990617428207153e-06, | |
| "loss": 0.6442, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.990142747892145e-06, | |
| "loss": 0.6534, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.989656378295704e-06, | |
| "loss": 0.6519, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.989158321700959e-06, | |
| "loss": 0.669, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.988648580445896e-06, | |
| "loss": 0.6409, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.988127156923355e-06, | |
| "loss": 0.6372, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.987594053581013e-06, | |
| "loss": 0.6687, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.987049272921374e-06, | |
| "loss": 0.6508, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.98649281750176e-06, | |
| "loss": 0.649, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.985924689934297e-06, | |
| "loss": 0.6501, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.985344892885899e-06, | |
| "loss": 0.6478, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.984753429078264e-06, | |
| "loss": 0.6582, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.984150301287854e-06, | |
| "loss": 0.6382, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.9835355123458815e-06, | |
| "loss": 0.6563, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.9829090651383064e-06, | |
| "loss": 0.6414, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.9822709626058065e-06, | |
| "loss": 0.6518, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.981621207743779e-06, | |
| "loss": 0.6379, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.980959803602315e-06, | |
| "loss": 0.6356, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.980286753286196e-06, | |
| "loss": 0.6546, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.979602059954863e-06, | |
| "loss": 0.6511, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.978905726822424e-06, | |
| "loss": 0.6521, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.978197757157618e-06, | |
| "loss": 0.6404, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.9774781542838115e-06, | |
| "loss": 0.6349, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.97674692157898e-06, | |
| "loss": 0.6397, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "eval_loss": 0.6273066401481628, | |
| "eval_runtime": 70.5832, | |
| "eval_samples_per_second": 71.745, | |
| "eval_steps_per_second": 5.979, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.976004062475691e-06, | |
| "loss": 0.616, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.975249580461092e-06, | |
| "loss": 0.6359, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.9744834790768855e-06, | |
| "loss": 0.6313, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.973705761919324e-06, | |
| "loss": 0.6318, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.972916432639182e-06, | |
| "loss": 0.621, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.972115494941746e-06, | |
| "loss": 0.6438, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.971302952586796e-06, | |
| "loss": 0.6325, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.970478809388585e-06, | |
| "loss": 0.6426, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.969643069215824e-06, | |
| "loss": 0.6391, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.968795735991661e-06, | |
| "loss": 0.6215, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.967936813693668e-06, | |
| "loss": 0.6167, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.967066306353816e-06, | |
| "loss": 0.6232, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.966184218058457e-06, | |
| "loss": 0.6141, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.965290552948312e-06, | |
| "loss": 0.6269, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.964385315218441e-06, | |
| "loss": 0.6144, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.9634685091182306e-06, | |
| "loss": 0.633, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.962540138951371e-06, | |
| "loss": 0.6173, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.9616002090758385e-06, | |
| "loss": 0.6209, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.960648723903872e-06, | |
| "loss": 0.6298, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.9596856879019524e-06, | |
| "loss": 0.6239, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.958711105590786e-06, | |
| "loss": 0.6347, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.957724981545276e-06, | |
| "loss": 0.6137, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.956727320394507e-06, | |
| "loss": 0.6205, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.9557181268217225e-06, | |
| "loss": 0.588, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.954697405564299e-06, | |
| "loss": 0.6211, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.953665161413731e-06, | |
| "loss": 0.6264, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.952621399215598e-06, | |
| "loss": 0.6031, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.951566123869553e-06, | |
| "loss": 0.6186, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.950499340329291e-06, | |
| "loss": 0.6177, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.9494210536025306e-06, | |
| "loss": 0.6364, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.948331268750988e-06, | |
| "loss": 0.6214, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.947229990890356e-06, | |
| "loss": 0.6159, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.946117225190274e-06, | |
| "loss": 0.5957, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.944992976874313e-06, | |
| "loss": 0.6088, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.9438572512199425e-06, | |
| "loss": 0.6163, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.942710053558511e-06, | |
| "loss": 0.5964, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.941551389275217e-06, | |
| "loss": 0.6132, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.940381263809089e-06, | |
| "loss": 0.605, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.9391996826529535e-06, | |
| "loss": 0.6063, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.938006651353414e-06, | |
| "loss": 0.5971, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.936802175510824e-06, | |
| "loss": 0.607, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.935586260779261e-06, | |
| "loss": 0.6128, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.934358912866497e-06, | |
| "loss": 0.5985, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.933120137533975e-06, | |
| "loss": 0.6132, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.931869940596779e-06, | |
| "loss": 0.611, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.9306083279236124e-06, | |
| "loss": 0.5972, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.929335305436764e-06, | |
| "loss": 0.6025, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.928050879112083e-06, | |
| "loss": 0.618, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.926755054978951e-06, | |
| "loss": 0.6072, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.925447839120254e-06, | |
| "loss": 0.6045, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.924129237672351e-06, | |
| "loss": 0.6133, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.922799256825052e-06, | |
| "loss": 0.6027, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.921457902821578e-06, | |
| "loss": 0.5949, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.920105181958545e-06, | |
| "loss": 0.5807, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.918741100585921e-06, | |
| "loss": 0.6019, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.917365665107009e-06, | |
| "loss": 0.5998, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.915978881978407e-06, | |
| "loss": 0.6053, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.91458075770998e-06, | |
| "loss": 0.6015, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.913171298864836e-06, | |
| "loss": 0.5926, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.911750512059285e-06, | |
| "loss": 0.5923, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.910318403962814e-06, | |
| "loss": 0.5809, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.908874981298058e-06, | |
| "loss": 0.6, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.907420250840761e-06, | |
| "loss": 0.5861, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.90595421941975e-06, | |
| "loss": 0.5969, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.904476893916901e-06, | |
| "loss": 0.5932, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.902988281267108e-06, | |
| "loss": 0.6009, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.901488388458247e-06, | |
| "loss": 0.5853, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.899977222531148e-06, | |
| "loss": 0.605, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.898454790579558e-06, | |
| "loss": 0.5774, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.89692109975011e-06, | |
| "loss": 0.5982, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.895376157242288e-06, | |
| "loss": 0.5874, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.893819970308394e-06, | |
| "loss": 0.5886, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.8922525462535154e-06, | |
| "loss": 0.5784, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "eval_loss": 0.5756770372390747, | |
| "eval_runtime": 70.7014, | |
| "eval_samples_per_second": 71.625, | |
| "eval_steps_per_second": 5.969, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.890673892435487e-06, | |
| "loss": 0.589, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.889084016264858e-06, | |
| "loss": 0.5799, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.887482925204861e-06, | |
| "loss": 0.5827, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.885870626771371e-06, | |
| "loss": 0.5763, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.884247128532874e-06, | |
| "loss": 0.584, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.88261243811043e-06, | |
| "loss": 0.5953, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.880966563177638e-06, | |
| "loss": 0.5745, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.879309511460601e-06, | |
| "loss": 0.5961, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.8776412907378845e-06, | |
| "loss": 0.5848, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.875961908840486e-06, | |
| "loss": 0.5758, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.874271373651798e-06, | |
| "loss": 0.5779, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.872569693107564e-06, | |
| "loss": 0.5827, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.8708568751958516e-06, | |
| "loss": 0.5683, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.869132927957007e-06, | |
| "loss": 0.587, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.867397859483619e-06, | |
| "loss": 0.5712, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.865651677920484e-06, | |
| "loss": 0.5786, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.863894391464566e-06, | |
| "loss": 0.5795, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.862126008364954e-06, | |
| "loss": 0.5824, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.860346536922834e-06, | |
| "loss": 0.5698, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.858555985491434e-06, | |
| "loss": 0.5971, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.856754362476003e-06, | |
| "loss": 0.5527, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.854941676333756e-06, | |
| "loss": 0.573, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.853117935573844e-06, | |
| "loss": 0.5791, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.85128314875731e-06, | |
| "loss": 0.5793, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.849437324497049e-06, | |
| "loss": 0.5636, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.8475804714577704e-06, | |
| "loss": 0.5799, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.8457125983559506e-06, | |
| "loss": 0.5865, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.843833713959803e-06, | |
| "loss": 0.5781, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.841943827089223e-06, | |
| "loss": 0.5907, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.840042946615761e-06, | |
| "loss": 0.5689, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.8381310814625705e-06, | |
| "loss": 0.5756, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.836208240604369e-06, | |
| "loss": 0.5755, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.834274433067398e-06, | |
| "loss": 0.5661, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.832329667929378e-06, | |
| "loss": 0.584, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.830373954319467e-06, | |
| "loss": 0.5673, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.828407301418218e-06, | |
| "loss": 0.5537, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.826429718457534e-06, | |
| "loss": 0.5636, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.824441214720629e-06, | |
| "loss": 0.5688, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.822441799541979e-06, | |
| "loss": 0.573, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.820431482307282e-06, | |
| "loss": 0.5746, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.818410272453411e-06, | |
| "loss": 0.5639, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.816378179468375e-06, | |
| "loss": 0.5686, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.814335212891266e-06, | |
| "loss": 0.5671, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.812281382312222e-06, | |
| "loss": 0.5697, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.81021669737238e-06, | |
| "loss": 0.5853, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.808141167763827e-06, | |
| "loss": 0.5627, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.806054803229559e-06, | |
| "loss": 0.5818, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.8039576135634314e-06, | |
| "loss": 0.5935, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.801849608610119e-06, | |
| "loss": 0.5656, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.799730798265064e-06, | |
| "loss": 0.5716, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.79760119247443e-06, | |
| "loss": 0.5605, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.795460801235058e-06, | |
| "loss": 0.5776, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.793309634594418e-06, | |
| "loss": 0.555, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.7911477026505656e-06, | |
| "loss": 0.5692, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.788975015552085e-06, | |
| "loss": 0.5529, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.786791583498051e-06, | |
| "loss": 0.5721, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.784597416737978e-06, | |
| "loss": 0.5488, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.782392525571771e-06, | |
| "loss": 0.5598, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.780176920349675e-06, | |
| "loss": 0.5603, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.777950611472234e-06, | |
| "loss": 0.5513, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.775713609390234e-06, | |
| "loss": 0.5668, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.773465924604657e-06, | |
| "loss": 0.5585, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.771207567666635e-06, | |
| "loss": 0.5636, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.7689385491773934e-06, | |
| "loss": 0.5654, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.766658879788208e-06, | |
| "loss": 0.5672, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.7643685702003536e-06, | |
| "loss": 0.5657, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.762067631165049e-06, | |
| "loss": 0.5686, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.759756073483412e-06, | |
| "loss": 0.547, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.7574339080064046e-06, | |
| "loss": 0.5575, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.755101145634788e-06, | |
| "loss": 0.5702, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.752757797319064e-06, | |
| "loss": 0.5498, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.750403874059428e-06, | |
| "loss": 0.5484, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.748039386905719e-06, | |
| "loss": 0.576, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "eval_loss": 0.5469194054603577, | |
| "eval_runtime": 70.7492, | |
| "eval_samples_per_second": 71.577, | |
| "eval_steps_per_second": 5.965, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.745664346957362e-06, | |
| "loss": 0.5703, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.743278765363319e-06, | |
| "loss": 0.5642, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.74088265332204e-06, | |
| "loss": 0.5533, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.738476022081405e-06, | |
| "loss": 0.5696, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.736058882938674e-06, | |
| "loss": 0.5581, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.733631247240435e-06, | |
| "loss": 0.5368, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.731193126382544e-06, | |
| "loss": 0.5589, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.728744531810082e-06, | |
| "loss": 0.5446, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.726285475017295e-06, | |
| "loss": 0.5661, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.723815967547539e-06, | |
| "loss": 0.5493, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.721336020993228e-06, | |
| "loss": 0.5648, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.718845646995782e-06, | |
| "loss": 0.5503, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.716344857245568e-06, | |
| "loss": 0.5604, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.713833663481845e-06, | |
| "loss": 0.5536, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.711312077492714e-06, | |
| "loss": 0.5531, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.708780111115058e-06, | |
| "loss": 0.5584, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.706237776234487e-06, | |
| "loss": 0.5694, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.7036850847852835e-06, | |
| "loss": 0.562, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.7011220487503485e-06, | |
| "loss": 0.5542, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.698548680161138e-06, | |
| "loss": 0.5655, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.6959649910976165e-06, | |
| "loss": 0.5529, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.693370993688191e-06, | |
| "loss": 0.5675, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.690766700109659e-06, | |
| "loss": 0.556, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.688152122587153e-06, | |
| "loss": 0.5566, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.685527273394078e-06, | |
| "loss": 0.5462, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.682892164852057e-06, | |
| "loss": 0.5596, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.680246809330874e-06, | |
| "loss": 0.5605, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.677591219248413e-06, | |
| "loss": 0.5508, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.674925407070602e-06, | |
| "loss": 0.5716, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.672249385311353e-06, | |
| "loss": 0.5368, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.669563166532504e-06, | |
| "loss": 0.5474, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.666866763343762e-06, | |
| "loss": 0.557, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.664160188402641e-06, | |
| "loss": 0.5523, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.661443454414403e-06, | |
| "loss": 0.5547, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.6587165741319975e-06, | |
| "loss": 0.5347, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.655979560356006e-06, | |
| "loss": 0.5491, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.6532324259345755e-06, | |
| "loss": 0.5724, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.650475183763365e-06, | |
| "loss": 0.5611, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.647707846785478e-06, | |
| "loss": 0.5471, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.644930427991408e-06, | |
| "loss": 0.5585, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.642142940418973e-06, | |
| "loss": 0.5425, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.639345397153259e-06, | |
| "loss": 0.5544, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.636537811326551e-06, | |
| "loss": 0.5396, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.633720196118281e-06, | |
| "loss": 0.5348, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.630892564754957e-06, | |
| "loss": 0.5505, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.6280549305101065e-06, | |
| "loss": 0.5358, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.625207306704213e-06, | |
| "loss": 0.556, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.6223497067046555e-06, | |
| "loss": 0.5546, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.619482143925637e-06, | |
| "loss": 0.535, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.616604631828137e-06, | |
| "loss": 0.5645, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.61371718391983e-06, | |
| "loss": 0.5615, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.610819813755038e-06, | |
| "loss": 0.546, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.607912534934658e-06, | |
| "loss": 0.5517, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.604995361106101e-06, | |
| "loss": 0.5346, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.602068305963225e-06, | |
| "loss": 0.5422, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.599131383246277e-06, | |
| "loss": 0.5463, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.596184606741821e-06, | |
| "loss": 0.5452, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.593227990282679e-06, | |
| "loss": 0.5416, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.590261547747864e-06, | |
| "loss": 0.5486, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.587285293062514e-06, | |
| "loss": 0.5444, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.584299240197826e-06, | |
| "loss": 0.5399, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.581303403170995e-06, | |
| "loss": 0.5495, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.578297796045142e-06, | |
| "loss": 0.5382, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.5752824329292534e-06, | |
| "loss": 0.5349, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.572257327978111e-06, | |
| "loss": 0.5627, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.569222495392227e-06, | |
| "loss": 0.5395, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.566177949417778e-06, | |
| "loss": 0.5602, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.5631237043465356e-06, | |
| "loss": 0.5432, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.560059774515805e-06, | |
| "loss": 0.5407, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.556986174308349e-06, | |
| "loss": 0.5434, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.553902918152329e-06, | |
| "loss": 0.5449, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.550810020521231e-06, | |
| "loss": 0.5487, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.547707495933802e-06, | |
| "loss": 0.5619, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "eval_loss": 0.536051869392395, | |
| "eval_runtime": 70.7206, | |
| "eval_samples_per_second": 71.606, | |
| "eval_steps_per_second": 5.967, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.5445953589539784e-06, | |
| "loss": 0.5463, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.541473624190822e-06, | |
| "loss": 0.5608, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.5383423062984455e-06, | |
| "loss": 0.5606, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.535201419975948e-06, | |
| "loss": 0.5392, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.5320509799673476e-06, | |
| "loss": 0.5356, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.528891001061506e-06, | |
| "loss": 0.5771, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.5257214980920636e-06, | |
| "loss": 0.5488, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.522542485937369e-06, | |
| "loss": 0.5658, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.51935397952041e-06, | |
| "loss": 0.5461, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.516155993808741e-06, | |
| "loss": 0.5524, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.512948543814415e-06, | |
| "loss": 0.5534, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.5097316445939135e-06, | |
| "loss": 0.5446, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.5065053112480725e-06, | |
| "loss": 0.5407, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.503269558922015e-06, | |
| "loss": 0.544, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.500024402805081e-06, | |
| "loss": 0.5447, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.496769858130749e-06, | |
| "loss": 0.5609, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.493505940176574e-06, | |
| "loss": 0.535, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.49023266426411e-06, | |
| "loss": 0.5494, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.486950045758838e-06, | |
| "loss": 0.5495, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.483658100070095e-06, | |
| "loss": 0.5522, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.480356842651006e-06, | |
| "loss": 0.556, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.477046288998401e-06, | |
| "loss": 0.5581, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.473726454652755e-06, | |
| "loss": 0.545, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.470397355198103e-06, | |
| "loss": 0.5489, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.467059006261976e-06, | |
| "loss": 0.5504, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.463711423515324e-06, | |
| "loss": 0.5629, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.46035462267244e-06, | |
| "loss": 0.5424, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.45698861949089e-06, | |
| "loss": 0.5585, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.45361342977144e-06, | |
| "loss": 0.5507, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.450229069357978e-06, | |
| "loss": 0.5627, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.446835554137438e-06, | |
| "loss": 0.5503, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.443432900039737e-06, | |
| "loss": 0.538, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.440021123037683e-06, | |
| "loss": 0.5408, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.436600239146913e-06, | |
| "loss": 0.5494, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.433170264425814e-06, | |
| "loss": 0.5485, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.429731214975448e-06, | |
| "loss": 0.5497, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.426283106939474e-06, | |
| "loss": 0.5406, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.422825956504073e-06, | |
| "loss": 0.5421, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.4193597798978756e-06, | |
| "loss": 0.5504, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.415884593391882e-06, | |
| "loss": 0.5368, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.412400413299389e-06, | |
| "loss": 0.5456, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.408907255975907e-06, | |
| "loss": 0.5644, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.4054051378190915e-06, | |
| "loss": 0.5358, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.40189407526866e-06, | |
| "loss": 0.5566, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.3983740848063175e-06, | |
| "loss": 0.5332, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.394845182955678e-06, | |
| "loss": 0.5389, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.39130738628219e-06, | |
| "loss": 0.5406, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.387760711393052e-06, | |
| "loss": 0.537, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.3842051749371435e-06, | |
| "loss": 0.5301, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.380640793604938e-06, | |
| "loss": 0.547, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.377067584128432e-06, | |
| "loss": 0.555, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.3734855632810624e-06, | |
| "loss": 0.5568, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.369894747877627e-06, | |
| "loss": 0.5529, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.3662951547742085e-06, | |
| "loss": 0.5668, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.3626868008680946e-06, | |
| "loss": 0.5351, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.359069703097697e-06, | |
| "loss": 0.5553, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.355443878442474e-06, | |
| "loss": 0.5262, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.3518093439228484e-06, | |
| "loss": 0.5428, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.348166116600131e-06, | |
| "loss": 0.5474, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.3445142135764375e-06, | |
| "loss": 0.5279, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.340853651994608e-06, | |
| "loss": 0.5614, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.337184449038131e-06, | |
| "loss": 0.5533, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.333506621931056e-06, | |
| "loss": 0.539, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.32982018793792e-06, | |
| "loss": 0.5574, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.326125164363658e-06, | |
| "loss": 0.5527, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.322421568553529e-06, | |
| "loss": 0.5425, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.318709417893033e-06, | |
| "loss": 0.551, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.3149887298078275e-06, | |
| "loss": 0.5664, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.311259521763645e-06, | |
| "loss": 0.5376, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.307521811266214e-06, | |
| "loss": 0.5583, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.303775615861175e-06, | |
| "loss": 0.5444, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.300020953134e-06, | |
| "loss": 0.5359, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.296257840709906e-06, | |
| "loss": 0.5338, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.5293887853622437, | |
| "eval_runtime": 71.2151, | |
| "eval_samples_per_second": 71.108, | |
| "eval_steps_per_second": 5.926, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.292486296253776e-06, | |
| "loss": 0.541, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.288706337470076e-06, | |
| "loss": 0.5501, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.2849179821027684e-06, | |
| "loss": 0.5359, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.281121247935231e-06, | |
| "loss": 0.5362, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.277316152790177e-06, | |
| "loss": 0.5382, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.273502714529564e-06, | |
| "loss": 0.5298, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.269680951054517e-06, | |
| "loss": 0.5776, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.265850880305238e-06, | |
| "loss": 0.5497, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.262012520260928e-06, | |
| "loss": 0.5263, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.2581658889397e-06, | |
| "loss": 0.528, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.254311004398492e-06, | |
| "loss": 0.5341, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.250447884732986e-06, | |
| "loss": 0.5362, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.2465765480775215e-06, | |
| "loss": 0.5306, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.242697012605008e-06, | |
| "loss": 0.5191, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.238809296526847e-06, | |
| "loss": 0.5287, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.234913418092837e-06, | |
| "loss": 0.5421, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.231009395591093e-06, | |
| "loss": 0.5201, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.227097247347962e-06, | |
| "loss": 0.5283, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.223176991727935e-06, | |
| "loss": 0.5288, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.219248647133559e-06, | |
| "loss": 0.5204, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.215312232005353e-06, | |
| "loss": 0.5182, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.211367764821722e-06, | |
| "loss": 0.5431, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.207415264098868e-06, | |
| "loss": 0.5424, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.203454748390705e-06, | |
| "loss": 0.5213, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.19948623628877e-06, | |
| "loss": 0.5278, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.195509746422138e-06, | |
| "loss": 0.5443, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.191525297457332e-06, | |
| "loss": 0.5378, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.187532908098239e-06, | |
| "loss": 0.5287, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.183532597086018e-06, | |
| "loss": 0.5301, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.179524383199016e-06, | |
| "loss": 0.5317, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.175508285252674e-06, | |
| "loss": 0.534, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.171484322099446e-06, | |
| "loss": 0.5177, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.167452512628707e-06, | |
| "loss": 0.5344, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.163412875766661e-06, | |
| "loss": 0.5329, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.159365430476262e-06, | |
| "loss": 0.5297, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.15531019575711e-06, | |
| "loss": 0.5154, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.151247190645378e-06, | |
| "loss": 0.5344, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.1471764342137095e-06, | |
| "loss": 0.5193, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.143097945571137e-06, | |
| "loss": 0.5023, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.139011743862991e-06, | |
| "loss": 0.5235, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.1349178482708044e-06, | |
| "loss": 0.5086, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.130816278012231e-06, | |
| "loss": 0.5083, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.126707052340949e-06, | |
| "loss": 0.5234, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.122590190546576e-06, | |
| "loss": 0.5257, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.11846571195457e-06, | |
| "loss": 0.5202, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.114333635926147e-06, | |
| "loss": 0.5215, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.1101939818581885e-06, | |
| "loss": 0.5431, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.106046769183146e-06, | |
| "loss": 0.5299, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.1018920173689555e-06, | |
| "loss": 0.5259, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.0977297459189405e-06, | |
| "loss": 0.5023, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.093559974371725e-06, | |
| "loss": 0.5258, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.08938272230114e-06, | |
| "loss": 0.5287, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.085198009316132e-06, | |
| "loss": 0.5117, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.0810058550606706e-06, | |
| "loss": 0.5191, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.076806279213656e-06, | |
| "loss": 0.5234, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.072599301488826e-06, | |
| "loss": 0.5298, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.068384941634665e-06, | |
| "loss": 0.5183, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.064163219434315e-06, | |
| "loss": 0.5244, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.059934154705472e-06, | |
| "loss": 0.5057, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.055697767300302e-06, | |
| "loss": 0.5116, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.051454077105347e-06, | |
| "loss": 0.5208, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.047203104041427e-06, | |
| "loss": 0.5038, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.042944868063552e-06, | |
| "loss": 0.4986, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.038679389160823e-06, | |
| "loss": 0.5153, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.034406687356344e-06, | |
| "loss": 0.5256, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 4.030126782707123e-06, | |
| "loss": 0.5164, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 4.02583969530398e-06, | |
| "loss": 0.5315, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 4.021545445271453e-06, | |
| "loss": 0.5049, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.0172440527677e-06, | |
| "loss": 0.5222, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.012935537984414e-06, | |
| "loss": 0.5012, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.0086199211467135e-06, | |
| "loss": 0.5325, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.004297222513062e-06, | |
| "loss": 0.5121, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.9999674623751625e-06, | |
| "loss": 0.5227, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "eval_loss": 0.526366651058197, | |
| "eval_runtime": 70.7579, | |
| "eval_samples_per_second": 71.568, | |
| "eval_steps_per_second": 5.964, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.995630661057869e-06, | |
| "loss": 0.5069, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.991286838919086e-06, | |
| "loss": 0.5114, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.986936016349677e-06, | |
| "loss": 0.5242, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.982578213773368e-06, | |
| "loss": 0.5121, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.978213451646646e-06, | |
| "loss": 0.5265, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.9738417504586734e-06, | |
| "loss": 0.5144, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.969463130731183e-06, | |
| "loss": 0.5083, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.965077613018385e-06, | |
| "loss": 0.5113, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.960685217906871e-06, | |
| "loss": 0.4961, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.956285966015516e-06, | |
| "loss": 0.5018, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.951879877995381e-06, | |
| "loss": 0.5033, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.947466974529622e-06, | |
| "loss": 0.5151, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.943047276333383e-06, | |
| "loss": 0.5079, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.9386208041537055e-06, | |
| "loss": 0.4961, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.93418757876943e-06, | |
| "loss": 0.5098, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.9297476209911e-06, | |
| "loss": 0.5238, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.925300951660859e-06, | |
| "loss": 0.5197, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.920847591652357e-06, | |
| "loss": 0.5014, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.916387561870653e-06, | |
| "loss": 0.4977, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.911920883252114e-06, | |
| "loss": 0.5065, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.9074475767643175e-06, | |
| "loss": 0.4941, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.9029676634059565e-06, | |
| "loss": 0.5123, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.898481164206734e-06, | |
| "loss": 0.4956, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.893988100227273e-06, | |
| "loss": 0.5138, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.8894884925590095e-06, | |
| "loss": 0.5094, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.884982362324098e-06, | |
| "loss": 0.4993, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.880469730675311e-06, | |
| "loss": 0.5128, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.875950618795943e-06, | |
| "loss": 0.5091, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.871425047899704e-06, | |
| "loss": 0.5187, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.866893039230626e-06, | |
| "loss": 0.4973, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.862354614062961e-06, | |
| "loss": 0.5104, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.857809793701082e-06, | |
| "loss": 0.494, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.853258599479383e-06, | |
| "loss": 0.4944, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.848701052762176e-06, | |
| "loss": 0.5071, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.844137174943594e-06, | |
| "loss": 0.5059, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.839566987447492e-06, | |
| "loss": 0.4971, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.834990511727341e-06, | |
| "loss": 0.5109, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.8304077692661305e-06, | |
| "loss": 0.5275, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.82581878157627e-06, | |
| "loss": 0.5171, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.821223570199483e-06, | |
| "loss": 0.5085, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.816622156706709e-06, | |
| "loss": 0.4948, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.812014562698002e-06, | |
| "loss": 0.4959, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.8074008098024282e-06, | |
| "loss": 0.5087, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.802780919677966e-06, | |
| "loss": 0.4993, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.7981549140114015e-06, | |
| "loss": 0.5091, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.7935228145182314e-06, | |
| "loss": 0.4971, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.788884642942555e-06, | |
| "loss": 0.4976, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.784240421056976e-06, | |
| "loss": 0.4976, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.7795901706625025e-06, | |
| "loss": 0.496, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.7749339135884373e-06, | |
| "loss": 0.5072, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.7702716716922825e-06, | |
| "loss": 0.5074, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.765603466859635e-06, | |
| "loss": 0.5106, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.760929321004082e-06, | |
| "loss": 0.5052, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.756249256067097e-06, | |
| "loss": 0.4925, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.751563294017944e-06, | |
| "loss": 0.4922, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.7468714568535643e-06, | |
| "loss": 0.5246, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.7421737665984807e-06, | |
| "loss": 0.5016, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.737470245304692e-06, | |
| "loss": 0.5108, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.7327609150515685e-06, | |
| "loss": 0.5088, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.728045797945749e-06, | |
| "loss": 0.502, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.723324916121038e-06, | |
| "loss": 0.5099, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.7185982917382986e-06, | |
| "loss": 0.4917, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.7138659469853535e-06, | |
| "loss": 0.5076, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.7091279040768752e-06, | |
| "loss": 0.4993, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.7043841852542884e-06, | |
| "loss": 0.5042, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.6996348127856584e-06, | |
| "loss": 0.508, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.6948798089655913e-06, | |
| "loss": 0.5046, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.6901191961151285e-06, | |
| "loss": 0.5037, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.6853529965816394e-06, | |
| "loss": 0.488, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.6805812327387212e-06, | |
| "loss": 0.5089, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.67580392698609e-06, | |
| "loss": 0.4839, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.671021101749476e-06, | |
| "loss": 0.5099, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.6662327794805203e-06, | |
| "loss": 0.507, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "eval_loss": 0.5247712135314941, | |
| "eval_runtime": 70.9408, | |
| "eval_samples_per_second": 71.383, | |
| "eval_steps_per_second": 5.949, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.6614389826566678e-06, | |
| "loss": 0.4947, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.656639733781062e-06, | |
| "loss": 0.5022, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.6518350553824406e-06, | |
| "loss": 0.5206, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.6470249700150273e-06, | |
| "loss": 0.5084, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.642209500258428e-06, | |
| "loss": 0.4943, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.637388668717523e-06, | |
| "loss": 0.5118, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.632562498022365e-06, | |
| "loss": 0.5063, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.6277310108280662e-06, | |
| "loss": 0.4828, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.6228942298146985e-06, | |
| "loss": 0.498, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.6180521776871815e-06, | |
| "loss": 0.5064, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.613204877175181e-06, | |
| "loss": 0.5094, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.608352351032999e-06, | |
| "loss": 0.4858, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.6034946220394667e-06, | |
| "loss": 0.5, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.598631712997841e-06, | |
| "loss": 0.5078, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.593763646735693e-06, | |
| "loss": 0.5025, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.5888904461048037e-06, | |
| "loss": 0.5095, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.5840121339810558e-06, | |
| "loss": 0.5085, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.579128733264326e-06, | |
| "loss": 0.5069, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.5742402668783797e-06, | |
| "loss": 0.4872, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.5693467577707597e-06, | |
| "loss": 0.5029, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.564448228912682e-06, | |
| "loss": 0.4977, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.559544703298925e-06, | |
| "loss": 0.5014, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.5546362039477244e-06, | |
| "loss": 0.4964, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.549722753900662e-06, | |
| "loss": 0.5073, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.54480437622256e-06, | |
| "loss": 0.4972, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.539881094001375e-06, | |
| "loss": 0.508, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.534952930348081e-06, | |
| "loss": 0.5, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.530019908396571e-06, | |
| "loss": 0.5208, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.5250820513035403e-06, | |
| "loss": 0.4958, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.520139382248386e-06, | |
| "loss": 0.4894, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.515191924433089e-06, | |
| "loss": 0.4965, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.5102397010821133e-06, | |
| "loss": 0.5107, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.5052827354422913e-06, | |
| "loss": 0.4939, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.500321050782717e-06, | |
| "loss": 0.4948, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.4953546703946366e-06, | |
| "loss": 0.5096, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.4903836175913402e-06, | |
| "loss": 0.5124, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.4854079157080502e-06, | |
| "loss": 0.5176, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.480427588101812e-06, | |
| "loss": 0.5004, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.4754426581513866e-06, | |
| "loss": 0.5058, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.470453149257139e-06, | |
| "loss": 0.4973, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.4654590848409296e-06, | |
| "loss": 0.5085, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.460460488346002e-06, | |
| "loss": 0.5075, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.4554573832368743e-06, | |
| "loss": 0.5046, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.45044979299923e-06, | |
| "loss": 0.4852, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.445437741139807e-06, | |
| "loss": 0.4921, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.440421251186286e-06, | |
| "loss": 0.4943, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.435400346687182e-06, | |
| "loss": 0.5066, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 3.4303750512117324e-06, | |
| "loss": 0.5039, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 3.425345388349787e-06, | |
| "loss": 0.4971, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 3.4203113817116955e-06, | |
| "loss": 0.5009, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.4152730549282007e-06, | |
| "loss": 0.4939, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.4102304316503256e-06, | |
| "loss": 0.504, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.4051835355492578e-06, | |
| "loss": 0.5192, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.4001323903162476e-06, | |
| "loss": 0.5001, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 3.395077019662487e-06, | |
| "loss": 0.5132, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 3.3900174473190066e-06, | |
| "loss": 0.5215, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 3.3849536970365586e-06, | |
| "loss": 0.5056, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 3.3798857925855083e-06, | |
| "loss": 0.5085, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 3.3748137577557216e-06, | |
| "loss": 0.4918, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 3.369737616356452e-06, | |
| "loss": 0.5194, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 3.364657392216233e-06, | |
| "loss": 0.4927, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 3.35957310918276e-06, | |
| "loss": 0.5003, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 3.354484791122784e-06, | |
| "loss": 0.51, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 3.3493924619219964e-06, | |
| "loss": 0.4934, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 3.344296145484918e-06, | |
| "loss": 0.4863, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 3.339195865734788e-06, | |
| "loss": 0.4977, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 3.3340916466134458e-06, | |
| "loss": 0.5064, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 3.3289835120812297e-06, | |
| "loss": 0.4869, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 3.3238714861168513e-06, | |
| "loss": 0.4952, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 3.318755592717294e-06, | |
| "loss": 0.4991, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 3.313635855897693e-06, | |
| "loss": 0.4963, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 3.3085122996912284e-06, | |
| "loss": 0.5106, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 3.3033849481490067e-06, | |
| "loss": 0.4995, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "eval_loss": 0.5225731134414673, | |
| "eval_runtime": 70.7593, | |
| "eval_samples_per_second": 71.567, | |
| "eval_steps_per_second": 5.964, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 3.29825382533995e-06, | |
| "loss": 0.5121, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 3.293118955350687e-06, | |
| "loss": 0.4968, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 3.287980362285433e-06, | |
| "loss": 0.4936, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 3.2828380702658823e-06, | |
| "loss": 0.4992, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 3.2776921034310917e-06, | |
| "loss": 0.5006, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 3.272542485937369e-06, | |
| "loss": 0.4895, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 3.2673892419581595e-06, | |
| "loss": 0.4973, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 3.262232395683931e-06, | |
| "loss": 0.4921, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 3.257071971322062e-06, | |
| "loss": 0.5214, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 3.251907993096728e-06, | |
| "loss": 0.5045, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 3.2467404852487846e-06, | |
| "loss": 0.5033, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 3.241569472035658e-06, | |
| "loss": 0.494, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 3.23639497773123e-06, | |
| "loss": 0.5056, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 3.2312170266257225e-06, | |
| "loss": 0.4868, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 3.2260356430255845e-06, | |
| "loss": 0.5029, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 3.2208508512533777e-06, | |
| "loss": 0.4825, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 3.2156626756476624e-06, | |
| "loss": 0.4996, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 3.210471140562883e-06, | |
| "loss": 0.4989, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 3.2052762703692574e-06, | |
| "loss": 0.4891, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 3.200078089452654e-06, | |
| "loss": 0.507, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 3.1948766222144863e-06, | |
| "loss": 0.4998, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 3.189671893071592e-06, | |
| "loss": 0.4925, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 3.184463926456124e-06, | |
| "loss": 0.4935, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 3.179252746815429e-06, | |
| "loss": 0.4964, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 3.174038378611939e-06, | |
| "loss": 0.504, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 3.168820846323053e-06, | |
| "loss": 0.5163, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 3.1636001744410218e-06, | |
| "loss": 0.4929, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 3.1583763874728364e-06, | |
| "loss": 0.5051, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 3.153149509940108e-06, | |
| "loss": 0.4999, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 3.1479195663789596e-06, | |
| "loss": 0.5059, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 3.142686581339902e-06, | |
| "loss": 0.4932, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 3.137450579387727e-06, | |
| "loss": 0.5076, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 3.1322115851013867e-06, | |
| "loss": 0.4981, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 3.12696962307388e-06, | |
| "loss": 0.4902, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 3.121724717912138e-06, | |
| "loss": 0.5056, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 3.1164768942369058e-06, | |
| "loss": 0.4962, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 3.1112261766826303e-06, | |
| "loss": 0.5111, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 3.105972589897342e-06, | |
| "loss": 0.492, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 3.1007161585425417e-06, | |
| "loss": 0.4836, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 3.0954569072930813e-06, | |
| "loss": 0.4998, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 3.0901948608370503e-06, | |
| "loss": 0.5057, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 3.0849300438756607e-06, | |
| "loss": 0.4881, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 3.0796624811231286e-06, | |
| "loss": 0.4896, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 3.07439219730656e-06, | |
| "loss": 0.493, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 3.0691192171658335e-06, | |
| "loss": 0.5068, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 3.063843565453486e-06, | |
| "loss": 0.5064, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 3.0585652669345944e-06, | |
| "loss": 0.4997, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 3.053284346386662e-06, | |
| "loss": 0.486, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 3.0480008285994976e-06, | |
| "loss": 0.4983, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 3.0427147383751043e-06, | |
| "loss": 0.5115, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 3.0374261005275606e-06, | |
| "loss": 0.4859, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 3.0321349398829026e-06, | |
| "loss": 0.4691, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 3.02684128127901e-06, | |
| "loss": 0.4983, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 3.021545149565489e-06, | |
| "loss": 0.4794, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 3.0162465696035536e-06, | |
| "loss": 0.5014, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 3.0109455662659126e-06, | |
| "loss": 0.5093, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 3.0056421644366486e-06, | |
| "loss": 0.4815, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 3.0003363890111042e-06, | |
| "loss": 0.4993, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.9950282648957646e-06, | |
| "loss": 0.5145, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.989717817008139e-06, | |
| "loss": 0.4909, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.984405070276646e-06, | |
| "loss": 0.4992, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.979090049640495e-06, | |
| "loss": 0.4954, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.9737727800495706e-06, | |
| "loss": 0.4714, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.9684532864643123e-06, | |
| "loss": 0.4903, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.963131593855601e-06, | |
| "loss": 0.4932, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.9578077272046407e-06, | |
| "loss": 0.4887, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.95248171150284e-06, | |
| "loss": 0.5008, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.947153571751696e-06, | |
| "loss": 0.4912, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.9418233329626765e-06, | |
| "loss": 0.4917, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.936491020157103e-06, | |
| "loss": 0.4844, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.931156658366032e-06, | |
| "loss": 0.4927, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.925820272630138e-06, | |
| "loss": 0.4881, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.9204818879995995e-06, | |
| "loss": 0.4988, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "eval_loss": 0.521266520023346, | |
| "eval_runtime": 70.745, | |
| "eval_samples_per_second": 71.581, | |
| "eval_steps_per_second": 5.965, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.9151415295339734e-06, | |
| "loss": 0.4906, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.909799222302087e-06, | |
| "loss": 0.5025, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.9044549913819125e-06, | |
| "loss": 0.4962, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.8991088618604522e-06, | |
| "loss": 0.4934, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.8937608588336234e-06, | |
| "loss": 0.4848, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.8884110074061357e-06, | |
| "loss": 0.4976, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.883059332691377e-06, | |
| "loss": 0.491, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.877705859811292e-06, | |
| "loss": 0.5074, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.8723506138962694e-06, | |
| "loss": 0.4974, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.866993620085018e-06, | |
| "loss": 0.5048, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.8616349035244535e-06, | |
| "loss": 0.5067, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.856274489369577e-06, | |
| "loss": 0.4891, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.850912402783361e-06, | |
| "loss": 0.4988, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.8455486689366262e-06, | |
| "loss": 0.4999, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.8401833130079277e-06, | |
| "loss": 0.5119, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.834816360183434e-06, | |
| "loss": 0.504, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.8294478356568096e-06, | |
| "loss": 0.4941, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.8240777646290973e-06, | |
| "loss": 0.4963, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.8187061723086e-06, | |
| "loss": 0.5062, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.813333083910761e-06, | |
| "loss": 0.506, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.8079585246580476e-06, | |
| "loss": 0.4828, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.80258251977983e-06, | |
| "loss": 0.4919, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.7972050945122666e-06, | |
| "loss": 0.5074, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.791826274098181e-06, | |
| "loss": 0.4908, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.7864460837869483e-06, | |
| "loss": 0.4934, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.7810645488343737e-06, | |
| "loss": 0.499, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.775681694502572e-06, | |
| "loss": 0.4893, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.7702975460598545e-06, | |
| "loss": 0.5136, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.7649121287806064e-06, | |
| "loss": 0.5038, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.759525467945168e-06, | |
| "loss": 0.4965, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.754137588839718e-06, | |
| "loss": 0.4993, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.7487485167561546e-06, | |
| "loss": 0.4945, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.7433582769919752e-06, | |
| "loss": 0.5069, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.7379668948501575e-06, | |
| "loss": 0.4958, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.7325743956390456e-06, | |
| "loss": 0.5077, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.7271808046722224e-06, | |
| "loss": 0.4892, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.7217861472684004e-06, | |
| "loss": 0.5081, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.716390448751294e-06, | |
| "loss": 0.497, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.7109937344495076e-06, | |
| "loss": 0.5157, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.7055960296964134e-06, | |
| "loss": 0.4954, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.700197359830032e-06, | |
| "loss": 0.4886, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.6947977501929153e-06, | |
| "loss": 0.4987, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.6893972261320265e-06, | |
| "loss": 0.4989, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.6839958129986215e-06, | |
| "loss": 0.4992, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.678593536148129e-06, | |
| "loss": 0.4941, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.673190420940034e-06, | |
| "loss": 0.4832, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.667786492737755e-06, | |
| "loss": 0.5095, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.6623817769085268e-06, | |
| "loss": 0.4958, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.6569762988232838e-06, | |
| "loss": 0.481, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.6515700838565355e-06, | |
| "loss": 0.5198, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.6461631573862528e-06, | |
| "loss": 0.4827, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.6407555447937455e-06, | |
| "loss": 0.5017, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.6353472714635443e-06, | |
| "loss": 0.5014, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.629938362783282e-06, | |
| "loss": 0.4826, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.6245288441435734e-06, | |
| "loss": 0.4939, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.619118740937898e-06, | |
| "loss": 0.489, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.613708078562476e-06, | |
| "loss": 0.4769, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.6082968824161558e-06, | |
| "loss": 0.4963, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.60288517790029e-06, | |
| "loss": 0.499, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.5974729904186187e-06, | |
| "loss": 0.505, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.5920603453771475e-06, | |
| "loss": 0.5134, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.586647268184031e-06, | |
| "loss": 0.5112, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.5812337842494517e-06, | |
| "loss": 0.496, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.575819918985502e-06, | |
| "loss": 0.5056, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.5704056978060643e-06, | |
| "loss": 0.4804, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.5649911461266923e-06, | |
| "loss": 0.4924, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.55957628936449e-06, | |
| "loss": 0.4994, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.554161152937994e-06, | |
| "loss": 0.4814, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.548745762267053e-06, | |
| "loss": 0.498, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.543330142772711e-06, | |
| "loss": 0.5082, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.537914319877084e-06, | |
| "loss": 0.5103, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.532498319003245e-06, | |
| "loss": 0.4912, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.5270821655750997e-06, | |
| "loss": 0.4983, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "eval_loss": 0.5186672210693359, | |
| "eval_runtime": 70.712, | |
| "eval_samples_per_second": 71.614, | |
| "eval_steps_per_second": 5.968, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.5216658850172737e-06, | |
| "loss": 0.5086, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.5162495027549862e-06, | |
| "loss": 0.5001, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.510833044213934e-06, | |
| "loss": 0.516, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.505416534820174e-06, | |
| "loss": 0.4995, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.4981, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.4945834651798266e-06, | |
| "loss": 0.5008, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.4891669557860664e-06, | |
| "loss": 0.4864, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.483750497245015e-06, | |
| "loss": 0.5013, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.478334114982727e-06, | |
| "loss": 0.49, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.4729178344249007e-06, | |
| "loss": 0.4878, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.467501680996756e-06, | |
| "loss": 0.4954, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.4620856801229166e-06, | |
| "loss": 0.4922, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.4566698572272902e-06, | |
| "loss": 0.4923, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.451254237732948e-06, | |
| "loss": 0.4844, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 2.4458388470620066e-06, | |
| "loss": 0.5174, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 2.4404237106355104e-06, | |
| "loss": 0.5085, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.4350088538733086e-06, | |
| "loss": 0.4873, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.429594302193936e-06, | |
| "loss": 0.4799, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.4241800810144987e-06, | |
| "loss": 0.493, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.418766215750549e-06, | |
| "loss": 0.4899, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.41335273181597e-06, | |
| "loss": 0.4776, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.4079396546228537e-06, | |
| "loss": 0.4784, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 2.402527009581382e-06, | |
| "loss": 0.4842, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 2.3971148220997103e-06, | |
| "loss": 0.4924, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 2.3917031175838447e-06, | |
| "loss": 0.4846, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 2.3862919214375244e-06, | |
| "loss": 0.4832, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 2.380881259062104e-06, | |
| "loss": 0.478, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 2.3754711558564274e-06, | |
| "loss": 0.4786, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 2.370061637216719e-06, | |
| "loss": 0.4766, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 2.3646527285364565e-06, | |
| "loss": 0.4826, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 2.3592444552062557e-06, | |
| "loss": 0.4917, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 2.353836842613748e-06, | |
| "loss": 0.494, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 2.3484299161434653e-06, | |
| "loss": 0.4787, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 2.3430237011767166e-06, | |
| "loss": 0.4948, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 2.3376182230914728e-06, | |
| "loss": 0.4891, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 2.3322135072622463e-06, | |
| "loss": 0.4881, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 2.326809579059967e-06, | |
| "loss": 0.479, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 2.3214064638518714e-06, | |
| "loss": 0.4843, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 2.3160041870013794e-06, | |
| "loss": 0.5038, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 2.3106027738679743e-06, | |
| "loss": 0.4713, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 2.305202249807086e-06, | |
| "loss": 0.4877, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 2.2998026401699693e-06, | |
| "loss": 0.4998, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 2.2944039703035874e-06, | |
| "loss": 0.4752, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 2.289006265550493e-06, | |
| "loss": 0.4736, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 2.2836095512487063e-06, | |
| "loss": 0.4884, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 2.278213852731601e-06, | |
| "loss": 0.4754, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 2.272819195327778e-06, | |
| "loss": 0.4737, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 2.2674256043609553e-06, | |
| "loss": 0.4739, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 2.2620331051498425e-06, | |
| "loss": 0.4585, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 2.256641723008026e-06, | |
| "loss": 0.468, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 2.2512514832438466e-06, | |
| "loss": 0.4789, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 2.245862411160283e-06, | |
| "loss": 0.4844, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 2.2404745320548326e-06, | |
| "loss": 0.4731, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 2.235087871219394e-06, | |
| "loss": 0.474, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 2.2297024539401463e-06, | |
| "loss": 0.4877, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 2.224318305497429e-06, | |
| "loss": 0.4919, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 2.2189354511656267e-06, | |
| "loss": 0.4884, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 2.2135539162130516e-06, | |
| "loss": 0.4605, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 2.208173725901819e-06, | |
| "loss": 0.4737, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 2.2027949054877342e-06, | |
| "loss": 0.4881, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 2.1974174802201708e-06, | |
| "loss": 0.4724, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 2.192041475341953e-06, | |
| "loss": 0.4779, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 2.186666916089239e-06, | |
| "loss": 0.471, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 2.1812938276914016e-06, | |
| "loss": 0.4874, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 2.175922235370904e-06, | |
| "loss": 0.4786, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 2.1705521643431916e-06, | |
| "loss": 0.485, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 2.165183639816567e-06, | |
| "loss": 0.4651, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 2.1598166869920723e-06, | |
| "loss": 0.4594, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 2.154451331063374e-06, | |
| "loss": 0.4779, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 2.1490875972166394e-06, | |
| "loss": 0.4697, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 2.1437255106304232e-06, | |
| "loss": 0.4598, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 2.1383650964755473e-06, | |
| "loss": 0.4685, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 2.1330063799149825e-06, | |
| "loss": 0.4808, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "eval_loss": 0.520827054977417, | |
| "eval_runtime": 70.6736, | |
| "eval_samples_per_second": 71.653, | |
| "eval_steps_per_second": 5.971, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 2.127649386103732e-06, | |
| "loss": 0.4769, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 2.1222941401887087e-06, | |
| "loss": 0.4773, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 2.116940667308624e-06, | |
| "loss": 0.4658, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 2.1115889925938642e-06, | |
| "loss": 0.4772, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 2.106239141166377e-06, | |
| "loss": 0.4744, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 2.100891138139548e-06, | |
| "loss": 0.4728, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 2.0955450086180883e-06, | |
| "loss": 0.4777, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 2.0902007776979133e-06, | |
| "loss": 0.4752, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 2.0848584704660266e-06, | |
| "loss": 0.4649, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 2.079518112000402e-06, | |
| "loss": 0.4732, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 2.0741797273698627e-06, | |
| "loss": 0.4756, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 2.0688433416339694e-06, | |
| "loss": 0.4684, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 2.0635089798428976e-06, | |
| "loss": 0.4799, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 2.058176667037324e-06, | |
| "loss": 0.483, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 2.0528464282483047e-06, | |
| "loss": 0.4611, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 2.0475182884971606e-06, | |
| "loss": 0.4668, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 2.0421922727953597e-06, | |
| "loss": 0.4638, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 2.0368684061443995e-06, | |
| "loss": 0.4475, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 2.031546713535688e-06, | |
| "loss": 0.4765, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 2.0262272199504306e-06, | |
| "loss": 0.4625, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 2.0209099503595052e-06, | |
| "loss": 0.4708, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 2.0155949297233542e-06, | |
| "loss": 0.4522, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 2.010282182991861e-06, | |
| "loss": 0.4681, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 2.0049717351042366e-06, | |
| "loss": 0.4742, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.9996636109888966e-06, | |
| "loss": 0.4777, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.9943578355633523e-06, | |
| "loss": 0.4621, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.9890544337340882e-06, | |
| "loss": 0.4574, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.983753430396447e-06, | |
| "loss": 0.4681, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.9784548504345123e-06, | |
| "loss": 0.4549, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.973158718720991e-06, | |
| "loss": 0.4592, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.9678650601170986e-06, | |
| "loss": 0.4699, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.96257389947244e-06, | |
| "loss": 0.4619, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.9572852616248957e-06, | |
| "loss": 0.4725, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.951999171400503e-06, | |
| "loss": 0.4565, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.946715653613339e-06, | |
| "loss": 0.4708, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.941434733065406e-06, | |
| "loss": 0.4544, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.936156434546515e-06, | |
| "loss": 0.4817, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.930880782834168e-06, | |
| "loss": 0.4692, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.9256078026934417e-06, | |
| "loss": 0.4655, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.9203375188768726e-06, | |
| "loss": 0.4584, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.9150699561243397e-06, | |
| "loss": 0.4458, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.90980513916295e-06, | |
| "loss": 0.468, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.90454309270692e-06, | |
| "loss": 0.4652, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.8992838414574593e-06, | |
| "loss": 0.4526, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.8940274101026584e-06, | |
| "loss": 0.4693, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.8887738233173708e-06, | |
| "loss": 0.4787, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.8835231057630955e-06, | |
| "loss": 0.4896, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.8782752820878636e-06, | |
| "loss": 0.4637, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.873030376926121e-06, | |
| "loss": 0.4592, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.867788414898614e-06, | |
| "loss": 0.4505, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.8625494206122732e-06, | |
| "loss": 0.4714, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.8573134186600978e-06, | |
| "loss": 0.4629, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.8520804336210413e-06, | |
| "loss": 0.4598, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.8468504900598922e-06, | |
| "loss": 0.4632, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.8416236125271647e-06, | |
| "loss": 0.454, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.836399825558979e-06, | |
| "loss": 0.4589, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.8311791536769485e-06, | |
| "loss": 0.4514, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.8259616213880622e-06, | |
| "loss": 0.4682, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.8207472531845716e-06, | |
| "loss": 0.4674, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.8155360735438766e-06, | |
| "loss": 0.4664, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.810328106928408e-06, | |
| "loss": 0.4622, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.805123377785515e-06, | |
| "loss": 0.4618, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.799921910547347e-06, | |
| "loss": 0.4491, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.7947237296307434e-06, | |
| "loss": 0.4786, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.7895288594371168e-06, | |
| "loss": 0.4658, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.7843373243523385e-06, | |
| "loss": 0.4683, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.7791491487466234e-06, | |
| "loss": 0.4687, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.7739643569744163e-06, | |
| "loss": 0.4692, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.7687829733742779e-06, | |
| "loss": 0.47, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.76360502226877e-06, | |
| "loss": 0.4536, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.7584305279643433e-06, | |
| "loss": 0.4499, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.7532595147512167e-06, | |
| "loss": 0.4632, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.748092006903273e-06, | |
| "loss": 0.4647, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "eval_loss": 0.52239590883255, | |
| "eval_runtime": 71.0114, | |
| "eval_samples_per_second": 71.313, | |
| "eval_steps_per_second": 5.943, | |
| "step": 876 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 1460, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 4, | |
| "save_steps": 73, | |
| "total_flos": 2200684871024640.0, | |
| "train_batch_size": 3, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |