| { | |
| "best_metric": 0.3944380581378937, | |
| "best_model_checkpoint": "./FT_models/[LDH]0226_origin_data/checkpoint-3000", | |
| "epoch": 2.8483985765124555, | |
| "eval_steps": 500, | |
| "global_step": 3000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.009489916963226572, | |
| "grad_norm": 0.9249091148376465, | |
| "learning_rate": 0.0001999995055317446, | |
| "loss": 2.3789, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.018979833926453145, | |
| "grad_norm": 0.7498548030853271, | |
| "learning_rate": 0.0001999955498150411, | |
| "loss": 1.3794, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.028469750889679714, | |
| "grad_norm": 0.7051133513450623, | |
| "learning_rate": 0.00019998763853811184, | |
| "loss": 1.2474, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.03795966785290629, | |
| "grad_norm": 0.7239620685577393, | |
| "learning_rate": 0.00019997577201390606, | |
| "loss": 1.2512, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.04744958481613286, | |
| "grad_norm": 0.5843775272369385, | |
| "learning_rate": 0.0001999599507118322, | |
| "loss": 1.1068, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05693950177935943, | |
| "grad_norm": 0.59147709608078, | |
| "learning_rate": 0.00019994017525773913, | |
| "loss": 1.1164, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.066429418742586, | |
| "grad_norm": 0.6587395668029785, | |
| "learning_rate": 0.0001999164464338918, | |
| "loss": 1.063, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.07591933570581258, | |
| "grad_norm": 0.5964196920394897, | |
| "learning_rate": 0.0001998887651789398, | |
| "loss": 1.0046, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.08540925266903915, | |
| "grad_norm": 0.6036216020584106, | |
| "learning_rate": 0.0001998571325878806, | |
| "loss": 1.0529, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.09489916963226572, | |
| "grad_norm": 0.7564125061035156, | |
| "learning_rate": 0.00019982154991201608, | |
| "loss": 1.0009, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.10438908659549229, | |
| "grad_norm": 0.7184653282165527, | |
| "learning_rate": 0.00019978201855890308, | |
| "loss": 0.9723, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.11387900355871886, | |
| "grad_norm": 0.6793704628944397, | |
| "learning_rate": 0.00019973854009229763, | |
| "loss": 0.9341, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.12336892052194544, | |
| "grad_norm": 0.6832011342048645, | |
| "learning_rate": 0.00019969111623209323, | |
| "loss": 0.9873, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.132858837485172, | |
| "grad_norm": 0.85871422290802, | |
| "learning_rate": 0.00019963974885425266, | |
| "loss": 0.9299, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.1423487544483986, | |
| "grad_norm": 0.7056658267974854, | |
| "learning_rate": 0.00019958443999073397, | |
| "loss": 0.9533, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.15183867141162516, | |
| "grad_norm": 0.7040665149688721, | |
| "learning_rate": 0.00019952519182940993, | |
| "loss": 0.8815, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.16132858837485173, | |
| "grad_norm": 0.6971840262413025, | |
| "learning_rate": 0.0001994620067139815, | |
| "loss": 0.9461, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.1708185053380783, | |
| "grad_norm": 0.7504563331604004, | |
| "learning_rate": 0.00019939488714388524, | |
| "loss": 0.8483, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.18030842230130487, | |
| "grad_norm": 0.684840977191925, | |
| "learning_rate": 0.00019932383577419432, | |
| "loss": 0.9122, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.18979833926453143, | |
| "grad_norm": 0.6745721101760864, | |
| "learning_rate": 0.0001992488554155135, | |
| "loss": 0.8271, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.199288256227758, | |
| "grad_norm": 0.7717174887657166, | |
| "learning_rate": 0.0001991699490338681, | |
| "loss": 0.8124, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.20877817319098457, | |
| "grad_norm": 0.7041401267051697, | |
| "learning_rate": 0.00019908711975058637, | |
| "loss": 0.8219, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.21826809015421114, | |
| "grad_norm": 0.6780515909194946, | |
| "learning_rate": 0.00019900037084217637, | |
| "loss": 0.7733, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.2277580071174377, | |
| "grad_norm": 0.7086532115936279, | |
| "learning_rate": 0.00019890970574019617, | |
| "loss": 0.7575, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.2372479240806643, | |
| "grad_norm": 0.7180762887001038, | |
| "learning_rate": 0.00019881512803111796, | |
| "loss": 0.8863, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.24673784104389088, | |
| "grad_norm": 0.8088333010673523, | |
| "learning_rate": 0.00019871664145618657, | |
| "loss": 0.7497, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.25622775800711745, | |
| "grad_norm": 0.6337082386016846, | |
| "learning_rate": 0.00019861424991127115, | |
| "loss": 0.8269, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.265717674970344, | |
| "grad_norm": 0.8026911020278931, | |
| "learning_rate": 0.00019850795744671116, | |
| "loss": 0.8229, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.2752075919335706, | |
| "grad_norm": 0.8139302730560303, | |
| "learning_rate": 0.00019839776826715614, | |
| "loss": 0.78, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.2846975088967972, | |
| "grad_norm": 0.8401020765304565, | |
| "learning_rate": 0.00019828368673139947, | |
| "loss": 0.6933, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.2941874258600237, | |
| "grad_norm": 0.7387434840202332, | |
| "learning_rate": 0.00019816571735220583, | |
| "loss": 0.8241, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.3036773428232503, | |
| "grad_norm": 0.7867814302444458, | |
| "learning_rate": 0.0001980438647961327, | |
| "loss": 0.695, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.31316725978647686, | |
| "grad_norm": 0.6026145219802856, | |
| "learning_rate": 0.00019791813388334581, | |
| "loss": 0.8118, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.32265717674970346, | |
| "grad_norm": 0.7333958148956299, | |
| "learning_rate": 0.00019778852958742853, | |
| "loss": 0.7278, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.33214709371293, | |
| "grad_norm": 0.7646610140800476, | |
| "learning_rate": 0.00019765505703518496, | |
| "loss": 0.6802, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3416370106761566, | |
| "grad_norm": 0.8131152987480164, | |
| "learning_rate": 0.00019751772150643722, | |
| "loss": 0.766, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.35112692763938314, | |
| "grad_norm": 0.7349848747253418, | |
| "learning_rate": 0.0001973765284338167, | |
| "loss": 0.6832, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.36061684460260973, | |
| "grad_norm": 0.6821774244308472, | |
| "learning_rate": 0.00019723148340254892, | |
| "loss": 0.7976, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.3701067615658363, | |
| "grad_norm": 0.6879362463951111, | |
| "learning_rate": 0.0001970825921502328, | |
| "loss": 0.7195, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.37959667852906287, | |
| "grad_norm": 0.699598491191864, | |
| "learning_rate": 0.00019692986056661356, | |
| "loss": 0.7, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.38908659549228947, | |
| "grad_norm": 0.6818081736564636, | |
| "learning_rate": 0.0001967732946933499, | |
| "loss": 0.6991, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.398576512455516, | |
| "grad_norm": 0.6942121386528015, | |
| "learning_rate": 0.00019661290072377482, | |
| "loss": 0.7161, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.4080664294187426, | |
| "grad_norm": 0.6413360834121704, | |
| "learning_rate": 0.0001964486850026507, | |
| "loss": 0.7084, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.41755634638196915, | |
| "grad_norm": 0.6501904129981995, | |
| "learning_rate": 0.00019628065402591845, | |
| "loss": 0.7431, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.42704626334519574, | |
| "grad_norm": 0.6739936470985413, | |
| "learning_rate": 0.0001961088144404403, | |
| "loss": 0.7081, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.4365361803084223, | |
| "grad_norm": 0.7370271682739258, | |
| "learning_rate": 0.00019593317304373705, | |
| "loss": 0.7203, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.4460260972716489, | |
| "grad_norm": 0.5882676243782043, | |
| "learning_rate": 0.00019575373678371909, | |
| "loss": 0.679, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.4555160142348754, | |
| "grad_norm": 0.6047580242156982, | |
| "learning_rate": 0.0001955705127584117, | |
| "loss": 0.6712, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.465005931198102, | |
| "grad_norm": 0.6153593063354492, | |
| "learning_rate": 0.00019538350821567404, | |
| "loss": 0.6788, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.4744958481613286, | |
| "grad_norm": 0.6934164762496948, | |
| "learning_rate": 0.00019519273055291266, | |
| "loss": 0.6903, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.4744958481613286, | |
| "eval_loss": 0.6663702130317688, | |
| "eval_runtime": 20.5638, | |
| "eval_samples_per_second": 15.415, | |
| "eval_steps_per_second": 7.732, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.48398576512455516, | |
| "grad_norm": 0.6835016012191772, | |
| "learning_rate": 0.00019499818731678873, | |
| "loss": 0.6843, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.49347568208778175, | |
| "grad_norm": 0.666349470615387, | |
| "learning_rate": 0.00019479988620291956, | |
| "loss": 0.6961, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.5029655990510083, | |
| "grad_norm": 0.7474802732467651, | |
| "learning_rate": 0.00019459783505557424, | |
| "loss": 0.6614, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.5124555160142349, | |
| "grad_norm": 0.6815286874771118, | |
| "learning_rate": 0.0001943920418673633, | |
| "loss": 0.6693, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.5219454329774614, | |
| "grad_norm": 0.5686200261116028, | |
| "learning_rate": 0.0001941825147789225, | |
| "loss": 0.6946, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.531435349940688, | |
| "grad_norm": 0.643159806728363, | |
| "learning_rate": 0.00019396926207859084, | |
| "loss": 0.6486, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.5409252669039146, | |
| "grad_norm": 0.7110168933868408, | |
| "learning_rate": 0.00019375229220208276, | |
| "loss": 0.6459, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.5504151838671412, | |
| "grad_norm": 0.6624748706817627, | |
| "learning_rate": 0.0001935316137321543, | |
| "loss": 0.6692, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5599051008303677, | |
| "grad_norm": 0.761031985282898, | |
| "learning_rate": 0.00019330723539826375, | |
| "loss": 0.6046, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.5693950177935944, | |
| "grad_norm": 0.6096632480621338, | |
| "learning_rate": 0.0001930791660762262, | |
| "loss": 0.6775, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.5788849347568209, | |
| "grad_norm": 0.6898393630981445, | |
| "learning_rate": 0.0001928474147878626, | |
| "loss": 0.6293, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5883748517200474, | |
| "grad_norm": 0.6664237380027771, | |
| "learning_rate": 0.0001926119907006426, | |
| "loss": 0.6274, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.597864768683274, | |
| "grad_norm": 0.7034701704978943, | |
| "learning_rate": 0.00019237290312732226, | |
| "loss": 0.6334, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.6073546856465006, | |
| "grad_norm": 0.5845565795898438, | |
| "learning_rate": 0.0001921301615255754, | |
| "loss": 0.6791, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.6168446026097272, | |
| "grad_norm": 0.6939854025840759, | |
| "learning_rate": 0.00019188377549761963, | |
| "loss": 0.6554, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.6263345195729537, | |
| "grad_norm": 0.6801786422729492, | |
| "learning_rate": 0.00019163375478983632, | |
| "loss": 0.6543, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.6358244365361803, | |
| "grad_norm": 0.583102285861969, | |
| "learning_rate": 0.00019138010929238534, | |
| "loss": 0.6268, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.6453143534994069, | |
| "grad_norm": 0.6107549667358398, | |
| "learning_rate": 0.0001911228490388136, | |
| "loss": 0.6274, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.6548042704626335, | |
| "grad_norm": 0.6578108072280884, | |
| "learning_rate": 0.00019086198420565823, | |
| "loss": 0.6171, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.66429418742586, | |
| "grad_norm": 0.6483505368232727, | |
| "learning_rate": 0.000190597525112044, | |
| "loss": 0.5673, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.6737841043890866, | |
| "grad_norm": 0.698631763458252, | |
| "learning_rate": 0.00019032948221927524, | |
| "loss": 0.5675, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.6832740213523132, | |
| "grad_norm": 0.6782714128494263, | |
| "learning_rate": 0.00019005786613042185, | |
| "loss": 0.5904, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.6927639383155397, | |
| "grad_norm": 0.673217236995697, | |
| "learning_rate": 0.00018978268758989991, | |
| "loss": 0.622, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.7022538552787663, | |
| "grad_norm": 0.6975003480911255, | |
| "learning_rate": 0.00018950395748304678, | |
| "loss": 0.6558, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.7117437722419929, | |
| "grad_norm": 0.6466664671897888, | |
| "learning_rate": 0.0001892216868356904, | |
| "loss": 0.673, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.7212336892052195, | |
| "grad_norm": 0.6829777359962463, | |
| "learning_rate": 0.00018893588681371303, | |
| "loss": 0.5808, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.730723606168446, | |
| "grad_norm": 0.6941152811050415, | |
| "learning_rate": 0.00018864656872260985, | |
| "loss": 0.6232, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.7402135231316725, | |
| "grad_norm": 0.7243316769599915, | |
| "learning_rate": 0.00018835374400704154, | |
| "loss": 0.5933, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.7497034400948992, | |
| "grad_norm": 0.6251444816589355, | |
| "learning_rate": 0.00018805742425038145, | |
| "loss": 0.6116, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.7591933570581257, | |
| "grad_norm": 0.646567165851593, | |
| "learning_rate": 0.00018775762117425777, | |
| "loss": 0.601, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.7686832740213523, | |
| "grad_norm": 0.6085071563720703, | |
| "learning_rate": 0.00018745434663808942, | |
| "loss": 0.6603, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.7781731909845789, | |
| "grad_norm": 0.6589152812957764, | |
| "learning_rate": 0.00018714761263861728, | |
| "loss": 0.5672, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.7876631079478055, | |
| "grad_norm": 0.7008639574050903, | |
| "learning_rate": 0.00018683743130942928, | |
| "loss": 0.6363, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.797153024911032, | |
| "grad_norm": 0.6070975065231323, | |
| "learning_rate": 0.00018652381492048083, | |
| "loss": 0.6075, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.8066429418742586, | |
| "grad_norm": 0.7204874753952026, | |
| "learning_rate": 0.00018620677587760916, | |
| "loss": 0.5797, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.8161328588374852, | |
| "grad_norm": 0.6779175400733948, | |
| "learning_rate": 0.00018588632672204264, | |
| "loss": 0.5729, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.8256227758007118, | |
| "grad_norm": 0.5760260224342346, | |
| "learning_rate": 0.00018556248012990468, | |
| "loss": 0.5568, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.8351126927639383, | |
| "grad_norm": 0.6586163640022278, | |
| "learning_rate": 0.0001852352489117124, | |
| "loss": 0.5734, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.8446026097271648, | |
| "grad_norm": 0.6545649766921997, | |
| "learning_rate": 0.0001849046460118698, | |
| "loss": 0.6065, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.8540925266903915, | |
| "grad_norm": 0.5810338258743286, | |
| "learning_rate": 0.00018457068450815562, | |
| "loss": 0.5435, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.863582443653618, | |
| "grad_norm": 0.5936874151229858, | |
| "learning_rate": 0.00018423337761120618, | |
| "loss": 0.5878, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.8730723606168446, | |
| "grad_norm": 0.6365971565246582, | |
| "learning_rate": 0.00018389273866399275, | |
| "loss": 0.5696, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.8825622775800712, | |
| "grad_norm": 0.5335951447486877, | |
| "learning_rate": 0.00018354878114129367, | |
| "loss": 0.5848, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.8920521945432978, | |
| "grad_norm": 0.6080604791641235, | |
| "learning_rate": 0.00018320151864916135, | |
| "loss": 0.5739, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.9015421115065243, | |
| "grad_norm": 0.5476921200752258, | |
| "learning_rate": 0.00018285096492438424, | |
| "loss": 0.5474, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.9110320284697508, | |
| "grad_norm": 0.6528366804122925, | |
| "learning_rate": 0.00018249713383394303, | |
| "loss": 0.542, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.9205219454329775, | |
| "grad_norm": 0.5690432786941528, | |
| "learning_rate": 0.00018214003937446253, | |
| "loss": 0.5771, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.930011862396204, | |
| "grad_norm": 0.7033817172050476, | |
| "learning_rate": 0.0001817796956716578, | |
| "loss": 0.537, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.9395017793594306, | |
| "grad_norm": 0.6648123264312744, | |
| "learning_rate": 0.00018141611697977529, | |
| "loss": 0.5579, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.9489916963226572, | |
| "grad_norm": 0.5434956550598145, | |
| "learning_rate": 0.0001810493176810292, | |
| "loss": 0.58, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.9489916963226572, | |
| "eval_loss": 0.5363849401473999, | |
| "eval_runtime": 20.4233, | |
| "eval_samples_per_second": 15.521, | |
| "eval_steps_per_second": 7.785, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.9584816132858838, | |
| "grad_norm": 0.6258216500282288, | |
| "learning_rate": 0.00018067931228503246, | |
| "loss": 0.6065, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.9679715302491103, | |
| "grad_norm": 0.49021461606025696, | |
| "learning_rate": 0.00018030611542822257, | |
| "loss": 0.5184, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.9774614472123369, | |
| "grad_norm": 0.6789380311965942, | |
| "learning_rate": 0.00017992974187328305, | |
| "loss": 0.558, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.9869513641755635, | |
| "grad_norm": 0.6128289103507996, | |
| "learning_rate": 0.000179550206508559, | |
| "loss": 0.5901, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.99644128113879, | |
| "grad_norm": 0.5243271589279175, | |
| "learning_rate": 0.00017916752434746856, | |
| "loss": 0.5455, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.0066429418742586, | |
| "grad_norm": 0.6114349365234375, | |
| "learning_rate": 0.00017878171052790868, | |
| "loss": 0.5706, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.0161328588374852, | |
| "grad_norm": 0.5785910487174988, | |
| "learning_rate": 0.00017839278031165658, | |
| "loss": 0.4914, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.0256227758007117, | |
| "grad_norm": 0.6278082132339478, | |
| "learning_rate": 0.00017800074908376584, | |
| "loss": 0.4585, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.0351126927639382, | |
| "grad_norm": 0.5266045331954956, | |
| "learning_rate": 0.0001776056323519579, | |
| "loss": 0.4257, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.0446026097271648, | |
| "grad_norm": 0.5933576226234436, | |
| "learning_rate": 0.00017720744574600863, | |
| "loss": 0.4579, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.0540925266903916, | |
| "grad_norm": 0.568048357963562, | |
| "learning_rate": 0.00017680620501712996, | |
| "loss": 0.4736, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.063582443653618, | |
| "grad_norm": 0.586031436920166, | |
| "learning_rate": 0.00017640192603734692, | |
| "loss": 0.4208, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.0730723606168446, | |
| "grad_norm": 0.632996141910553, | |
| "learning_rate": 0.00017599462479886974, | |
| "loss": 0.4518, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.0825622775800712, | |
| "grad_norm": 0.6920955777168274, | |
| "learning_rate": 0.00017558431741346122, | |
| "loss": 0.4593, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.0920521945432977, | |
| "grad_norm": 0.5045620799064636, | |
| "learning_rate": 0.00017517102011179933, | |
| "loss": 0.472, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.1015421115065243, | |
| "grad_norm": 0.5956369638442993, | |
| "learning_rate": 0.00017475474924283536, | |
| "loss": 0.446, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.1110320284697508, | |
| "grad_norm": 0.7154461741447449, | |
| "learning_rate": 0.000174335521273147, | |
| "loss": 0.4513, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.1205219454329776, | |
| "grad_norm": 0.5512189269065857, | |
| "learning_rate": 0.00017391335278628712, | |
| "loss": 0.4584, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.130011862396204, | |
| "grad_norm": 0.9251359105110168, | |
| "learning_rate": 0.0001734882604821276, | |
| "loss": 0.4276, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.1395017793594306, | |
| "grad_norm": 0.640958845615387, | |
| "learning_rate": 0.00017306026117619889, | |
| "loss": 0.4481, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.1489916963226572, | |
| "grad_norm": 0.5504414439201355, | |
| "learning_rate": 0.00017262937179902472, | |
| "loss": 0.4697, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.1584816132858837, | |
| "grad_norm": 0.575148344039917, | |
| "learning_rate": 0.00017219560939545246, | |
| "loss": 0.4614, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.1679715302491103, | |
| "grad_norm": 0.5111013650894165, | |
| "learning_rate": 0.0001717589911239788, | |
| "loss": 0.4314, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.1774614472123368, | |
| "grad_norm": 0.5844498872756958, | |
| "learning_rate": 0.00017131953425607104, | |
| "loss": 0.4628, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.1869513641755636, | |
| "grad_norm": 0.5282189846038818, | |
| "learning_rate": 0.00017087725617548385, | |
| "loss": 0.443, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.1964412811387901, | |
| "grad_norm": 0.5334432721138, | |
| "learning_rate": 0.00017043217437757164, | |
| "loss": 0.4927, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.2059311981020167, | |
| "grad_norm": 0.8024120926856995, | |
| "learning_rate": 0.00016998430646859654, | |
| "loss": 0.4808, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.2154211150652432, | |
| "grad_norm": 0.5527599453926086, | |
| "learning_rate": 0.00016953367016503182, | |
| "loss": 0.433, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.2249110320284697, | |
| "grad_norm": 0.6859824657440186, | |
| "learning_rate": 0.00016908028329286112, | |
| "loss": 0.4774, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.2344009489916963, | |
| "grad_norm": 0.592665433883667, | |
| "learning_rate": 0.0001686241637868734, | |
| "loss": 0.4595, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.2438908659549228, | |
| "grad_norm": 0.643675684928894, | |
| "learning_rate": 0.00016816532968995328, | |
| "loss": 0.4377, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.2533807829181494, | |
| "grad_norm": 0.5149793028831482, | |
| "learning_rate": 0.00016770379915236766, | |
| "loss": 0.4127, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.262870699881376, | |
| "grad_norm": 0.6323621273040771, | |
| "learning_rate": 0.00016723959043104728, | |
| "loss": 0.4821, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.2723606168446027, | |
| "grad_norm": 0.6195471882820129, | |
| "learning_rate": 0.00016677272188886483, | |
| "loss": 0.479, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.2818505338078292, | |
| "grad_norm": 0.6360680460929871, | |
| "learning_rate": 0.00016630321199390867, | |
| "loss": 0.4322, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.2913404507710557, | |
| "grad_norm": 0.614509642124176, | |
| "learning_rate": 0.00016583107931875192, | |
| "loss": 0.4992, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.3008303677342823, | |
| "grad_norm": 0.6960734724998474, | |
| "learning_rate": 0.00016535634253971794, | |
| "loss": 0.4765, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.3103202846975088, | |
| "grad_norm": 0.5039885640144348, | |
| "learning_rate": 0.00016487902043614173, | |
| "loss": 0.4437, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.3198102016607356, | |
| "grad_norm": 0.629076361656189, | |
| "learning_rate": 0.00016439913188962685, | |
| "loss": 0.4626, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.3293001186239621, | |
| "grad_norm": 0.538341760635376, | |
| "learning_rate": 0.0001639166958832985, | |
| "loss": 0.4166, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.3387900355871887, | |
| "grad_norm": 0.6629465222358704, | |
| "learning_rate": 0.00016343173150105278, | |
| "loss": 0.4736, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.3482799525504152, | |
| "grad_norm": 0.6603752374649048, | |
| "learning_rate": 0.0001629442579268016, | |
| "loss": 0.4522, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.3577698695136418, | |
| "grad_norm": 0.7585862874984741, | |
| "learning_rate": 0.0001624542944437139, | |
| "loss": 0.439, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.3672597864768683, | |
| "grad_norm": 0.5442929863929749, | |
| "learning_rate": 0.00016196186043345288, | |
| "loss": 0.4181, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.3767497034400948, | |
| "grad_norm": 0.5769705772399902, | |
| "learning_rate": 0.00016146697537540924, | |
| "loss": 0.4431, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.3862396204033214, | |
| "grad_norm": 0.6381351351737976, | |
| "learning_rate": 0.0001609696588459307, | |
| "loss": 0.4476, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.395729537366548, | |
| "grad_norm": 0.573300838470459, | |
| "learning_rate": 0.00016046993051754756, | |
| "loss": 0.4839, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.4052194543297747, | |
| "grad_norm": 0.607252836227417, | |
| "learning_rate": 0.0001599678101581945, | |
| "loss": 0.4528, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.4147093712930012, | |
| "grad_norm": 0.6059923768043518, | |
| "learning_rate": 0.00015946331763042867, | |
| "loss": 0.4493, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.4241992882562278, | |
| "grad_norm": 0.6414983868598938, | |
| "learning_rate": 0.00015895647289064396, | |
| "loss": 0.4455, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.4241992882562278, | |
| "eval_loss": 0.4805718660354614, | |
| "eval_runtime": 20.4306, | |
| "eval_samples_per_second": 15.516, | |
| "eval_steps_per_second": 7.782, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.4336892052194543, | |
| "grad_norm": 0.6877856850624084, | |
| "learning_rate": 0.0001584472959882815, | |
| "loss": 0.4553, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.4431791221826809, | |
| "grad_norm": 0.6739844679832458, | |
| "learning_rate": 0.0001579358070650367, | |
| "loss": 0.4682, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.4526690391459074, | |
| "grad_norm": 0.6030142903327942, | |
| "learning_rate": 0.00015742202635406235, | |
| "loss": 0.4521, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.4621589561091342, | |
| "grad_norm": 0.6368103623390198, | |
| "learning_rate": 0.0001569059741791684, | |
| "loss": 0.4458, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.4716488730723607, | |
| "grad_norm": 0.649683952331543, | |
| "learning_rate": 0.0001563876709540178, | |
| "loss": 0.4568, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.4811387900355872, | |
| "grad_norm": 0.5779871940612793, | |
| "learning_rate": 0.00015586713718131922, | |
| "loss": 0.4808, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.4906287069988138, | |
| "grad_norm": 0.5961576700210571, | |
| "learning_rate": 0.0001553443934520159, | |
| "loss": 0.4595, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.5001186239620403, | |
| "grad_norm": 0.5500461459159851, | |
| "learning_rate": 0.00015481946044447099, | |
| "loss": 0.4584, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.5096085409252669, | |
| "grad_norm": 0.7123764157295227, | |
| "learning_rate": 0.00015429235892364994, | |
| "loss": 0.4236, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.5190984578884934, | |
| "grad_norm": 0.6884598731994629, | |
| "learning_rate": 0.00015376310974029873, | |
| "loss": 0.4661, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.52858837485172, | |
| "grad_norm": 0.5527105331420898, | |
| "learning_rate": 0.0001532317338301192, | |
| "loss": 0.4307, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.5380782918149465, | |
| "grad_norm": 0.5720965266227722, | |
| "learning_rate": 0.00015269825221294098, | |
| "loss": 0.415, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.547568208778173, | |
| "grad_norm": 0.716526985168457, | |
| "learning_rate": 0.0001521626859918898, | |
| "loss": 0.48, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.5570581257413998, | |
| "grad_norm": 0.5414557456970215, | |
| "learning_rate": 0.00015162505635255287, | |
| "loss": 0.4943, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.5665480427046263, | |
| "grad_norm": 0.6315144896507263, | |
| "learning_rate": 0.0001510853845621409, | |
| "loss": 0.4427, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.5760379596678529, | |
| "grad_norm": 0.5746013522148132, | |
| "learning_rate": 0.00015054369196864644, | |
| "loss": 0.4148, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.5855278766310796, | |
| "grad_norm": 0.6795936226844788, | |
| "learning_rate": 0.00015000000000000001, | |
| "loss": 0.4553, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.5950177935943062, | |
| "grad_norm": 0.6202004551887512, | |
| "learning_rate": 0.0001494543301632219, | |
| "loss": 0.4593, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.6045077105575327, | |
| "grad_norm": 0.49426236748695374, | |
| "learning_rate": 0.0001489067040435717, | |
| "loss": 0.4162, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.6139976275207593, | |
| "grad_norm": 0.6348617076873779, | |
| "learning_rate": 0.00014835714330369446, | |
| "loss": 0.4421, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.6234875444839858, | |
| "grad_norm": 0.5943213701248169, | |
| "learning_rate": 0.0001478056696827636, | |
| "loss": 0.491, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.6329774614472123, | |
| "grad_norm": 0.6268736124038696, | |
| "learning_rate": 0.00014725230499562119, | |
| "loss": 0.4144, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.6424673784104389, | |
| "grad_norm": 0.6147856712341309, | |
| "learning_rate": 0.00014669707113191483, | |
| "loss": 0.4089, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.6519572953736654, | |
| "grad_norm": 0.5978316068649292, | |
| "learning_rate": 0.00014613999005523174, | |
| "loss": 0.467, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.661447212336892, | |
| "grad_norm": 0.6331775784492493, | |
| "learning_rate": 0.00014558108380223012, | |
| "loss": 0.4489, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.6709371293001185, | |
| "grad_norm": 0.6398030519485474, | |
| "learning_rate": 0.00014502037448176734, | |
| "loss": 0.4383, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.680427046263345, | |
| "grad_norm": 0.6606591939926147, | |
| "learning_rate": 0.00014445788427402528, | |
| "loss": 0.4222, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.6899169632265718, | |
| "grad_norm": 0.644334077835083, | |
| "learning_rate": 0.00014389363542963306, | |
| "loss": 0.4403, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.6994068801897984, | |
| "grad_norm": 0.6094640493392944, | |
| "learning_rate": 0.00014332765026878687, | |
| "loss": 0.4433, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.708896797153025, | |
| "grad_norm": 0.6047070026397705, | |
| "learning_rate": 0.00014275995118036693, | |
| "loss": 0.4254, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.7183867141162514, | |
| "grad_norm": 0.612147331237793, | |
| "learning_rate": 0.00014219056062105193, | |
| "loss": 0.4544, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.7278766310794782, | |
| "grad_norm": 0.563838541507721, | |
| "learning_rate": 0.00014161950111443077, | |
| "loss": 0.4405, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.7373665480427047, | |
| "grad_norm": 0.6529393196105957, | |
| "learning_rate": 0.0001410467952501114, | |
| "loss": 0.4312, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.7468564650059313, | |
| "grad_norm": 0.663934588432312, | |
| "learning_rate": 0.00014047246568282736, | |
| "loss": 0.4421, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.7563463819691578, | |
| "grad_norm": 0.6207594871520996, | |
| "learning_rate": 0.00013989653513154165, | |
| "loss": 0.4365, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.7658362989323844, | |
| "grad_norm": 0.5731400847434998, | |
| "learning_rate": 0.0001393190263785479, | |
| "loss": 0.4202, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.775326215895611, | |
| "grad_norm": 0.5179402828216553, | |
| "learning_rate": 0.00013873996226856933, | |
| "loss": 0.4719, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.7848161328588374, | |
| "grad_norm": 0.40121665596961975, | |
| "learning_rate": 0.00013815936570785487, | |
| "loss": 0.4148, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.794306049822064, | |
| "grad_norm": 0.8414328694343567, | |
| "learning_rate": 0.00013757725966327322, | |
| "loss": 0.4648, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.8037959667852905, | |
| "grad_norm": 0.3349032998085022, | |
| "learning_rate": 0.00013699366716140435, | |
| "loss": 0.4168, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.813285883748517, | |
| "grad_norm": 0.6793477535247803, | |
| "learning_rate": 0.0001364086112876284, | |
| "loss": 0.4483, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.8227758007117436, | |
| "grad_norm": 0.5308493971824646, | |
| "learning_rate": 0.00013582211518521273, | |
| "loss": 0.4022, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.8322657176749704, | |
| "grad_norm": 0.5604270696640015, | |
| "learning_rate": 0.00013523420205439646, | |
| "loss": 0.4075, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.841755634638197, | |
| "grad_norm": 0.5878490805625916, | |
| "learning_rate": 0.00013464489515147238, | |
| "loss": 0.4337, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.8512455516014235, | |
| "grad_norm": 0.6031716465950012, | |
| "learning_rate": 0.00013405421778786737, | |
| "loss": 0.4655, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.8607354685646502, | |
| "grad_norm": 0.6215988993644714, | |
| "learning_rate": 0.00013346219332922016, | |
| "loss": 0.4715, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.8702253855278768, | |
| "grad_norm": 0.565021276473999, | |
| "learning_rate": 0.0001328688451944569, | |
| "loss": 0.4118, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.8797153024911033, | |
| "grad_norm": 0.5580743551254272, | |
| "learning_rate": 0.00013227419685486492, | |
| "loss": 0.4451, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.8892052194543298, | |
| "grad_norm": 0.7063333988189697, | |
| "learning_rate": 0.0001316782718331643, | |
| "loss": 0.4114, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.8986951364175564, | |
| "grad_norm": 0.534269392490387, | |
| "learning_rate": 0.00013108109370257712, | |
| "loss": 0.4029, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.8986951364175564, | |
| "eval_loss": 0.4455092251300812, | |
| "eval_runtime": 20.4087, | |
| "eval_samples_per_second": 15.533, | |
| "eval_steps_per_second": 7.791, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.908185053380783, | |
| "grad_norm": 0.5529626607894897, | |
| "learning_rate": 0.00013048268608589533, | |
| "loss": 0.4108, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.9176749703440095, | |
| "grad_norm": 0.673758327960968, | |
| "learning_rate": 0.00012988307265454597, | |
| "loss": 0.4219, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.927164887307236, | |
| "grad_norm": 0.6164252758026123, | |
| "learning_rate": 0.00012928227712765504, | |
| "loss": 0.4611, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.9366548042704625, | |
| "grad_norm": 0.6192930340766907, | |
| "learning_rate": 0.00012868032327110904, | |
| "loss": 0.4601, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.946144721233689, | |
| "grad_norm": 0.5321446061134338, | |
| "learning_rate": 0.00012807723489661495, | |
| "loss": 0.4112, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.9556346381969156, | |
| "grad_norm": 0.5794662237167358, | |
| "learning_rate": 0.0001274730358607583, | |
| "loss": 0.4915, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.9651245551601424, | |
| "grad_norm": 0.6252449750900269, | |
| "learning_rate": 0.00012686775006405946, | |
| "loss": 0.4307, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.974614472123369, | |
| "grad_norm": 0.6203203201293945, | |
| "learning_rate": 0.0001262614014500282, | |
| "loss": 0.4375, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.9841043890865955, | |
| "grad_norm": 0.617151141166687, | |
| "learning_rate": 0.00012565401400421651, | |
| "loss": 0.4517, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.993594306049822, | |
| "grad_norm": 0.681266188621521, | |
| "learning_rate": 0.00012504561175326985, | |
| "loss": 0.4245, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 2.0037959667852907, | |
| "grad_norm": 0.5491665005683899, | |
| "learning_rate": 0.0001244362187639767, | |
| "loss": 0.4221, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 2.0132858837485172, | |
| "grad_norm": 0.7722651958465576, | |
| "learning_rate": 0.0001238258591423165, | |
| "loss": 0.3264, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 2.022775800711744, | |
| "grad_norm": 0.6458817720413208, | |
| "learning_rate": 0.00012321455703250616, | |
| "loss": 0.3128, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 2.0322657176749703, | |
| "grad_norm": 0.5134835243225098, | |
| "learning_rate": 0.0001226023366160449, | |
| "loss": 0.3112, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 2.041755634638197, | |
| "grad_norm": 0.5023068785667419, | |
| "learning_rate": 0.00012198922211075778, | |
| "loss": 0.2929, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 2.0512455516014234, | |
| "grad_norm": 0.5612622499465942, | |
| "learning_rate": 0.00012137523776983757, | |
| "loss": 0.2943, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 2.06073546856465, | |
| "grad_norm": 0.5554709434509277, | |
| "learning_rate": 0.00012076040788088554, | |
| "loss": 0.3099, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 2.0702253855278765, | |
| "grad_norm": 0.5206819772720337, | |
| "learning_rate": 0.00012014475676495052, | |
| "loss": 0.2933, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 2.079715302491103, | |
| "grad_norm": 0.5239592790603638, | |
| "learning_rate": 0.000119528308775567, | |
| "loss": 0.2991, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 2.0892052194543296, | |
| "grad_norm": 0.5854765772819519, | |
| "learning_rate": 0.00011891108829779165, | |
| "loss": 0.3288, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 2.0986951364175566, | |
| "grad_norm": 0.7295845746994019, | |
| "learning_rate": 0.00011829311974723867, | |
| "loss": 0.3275, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 2.108185053380783, | |
| "grad_norm": 0.6106508374214172, | |
| "learning_rate": 0.00011767442756911417, | |
| "loss": 0.3111, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 2.1176749703440096, | |
| "grad_norm": 0.5574607849121094, | |
| "learning_rate": 0.00011705503623724898, | |
| "loss": 0.3194, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 2.127164887307236, | |
| "grad_norm": 0.5289790630340576, | |
| "learning_rate": 0.00011643497025313061, | |
| "loss": 0.3088, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 2.1366548042704627, | |
| "grad_norm": 0.5656944513320923, | |
| "learning_rate": 0.0001158142541449341, | |
| "loss": 0.3099, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 2.1461447212336893, | |
| "grad_norm": 0.594753086566925, | |
| "learning_rate": 0.0001151929124665516, | |
| "loss": 0.309, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 2.155634638196916, | |
| "grad_norm": 0.6289036870002747, | |
| "learning_rate": 0.00011457096979662114, | |
| "loss": 0.2948, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 2.1651245551601424, | |
| "grad_norm": 0.5498007535934448, | |
| "learning_rate": 0.00011394845073755455, | |
| "loss": 0.3082, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 2.174614472123369, | |
| "grad_norm": 0.5636598467826843, | |
| "learning_rate": 0.00011332537991456398, | |
| "loss": 0.2858, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 2.1841043890865954, | |
| "grad_norm": 0.6761645674705505, | |
| "learning_rate": 0.00011270178197468789, | |
| "loss": 0.332, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 2.193594306049822, | |
| "grad_norm": 0.5729044675827026, | |
| "learning_rate": 0.00011207768158581613, | |
| "loss": 0.326, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 2.2030842230130485, | |
| "grad_norm": 0.4295574724674225, | |
| "learning_rate": 0.00011145310343571411, | |
| "loss": 0.304, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 2.212574139976275, | |
| "grad_norm": 0.5968920588493347, | |
| "learning_rate": 0.0001108280722310462, | |
| "loss": 0.3205, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 2.2220640569395016, | |
| "grad_norm": 0.6211205720901489, | |
| "learning_rate": 0.00011020261269639842, | |
| "loss": 0.2823, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 2.231553973902728, | |
| "grad_norm": 0.6108773946762085, | |
| "learning_rate": 0.00010957674957330042, | |
| "loss": 0.324, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 2.241043890865955, | |
| "grad_norm": 0.8250358700752258, | |
| "learning_rate": 0.00010895050761924668, | |
| "loss": 0.3359, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 2.2505338078291817, | |
| "grad_norm": 0.5284693837165833, | |
| "learning_rate": 0.00010832391160671729, | |
| "loss": 0.3122, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 2.260023724792408, | |
| "grad_norm": 0.5513204336166382, | |
| "learning_rate": 0.00010769698632219794, | |
| "loss": 0.3264, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 2.2695136417556347, | |
| "grad_norm": 0.6842665672302246, | |
| "learning_rate": 0.00010706975656519946, | |
| "loss": 0.3047, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 2.2790035587188613, | |
| "grad_norm": 0.545037031173706, | |
| "learning_rate": 0.00010644224714727681, | |
| "loss": 0.3115, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.288493475682088, | |
| "grad_norm": 0.5622695088386536, | |
| "learning_rate": 0.00010581448289104758, | |
| "loss": 0.3158, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 2.2979833926453144, | |
| "grad_norm": 0.6832267642021179, | |
| "learning_rate": 0.00010518648862921012, | |
| "loss": 0.3298, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 2.307473309608541, | |
| "grad_norm": 0.5637592673301697, | |
| "learning_rate": 0.00010455828920356115, | |
| "loss": 0.3103, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 2.3169632265717675, | |
| "grad_norm": 0.5744633078575134, | |
| "learning_rate": 0.00010392990946401313, | |
| "loss": 0.3271, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 2.326453143534994, | |
| "grad_norm": 0.7118310928344727, | |
| "learning_rate": 0.00010330137426761135, | |
| "loss": 0.3276, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.3359430604982205, | |
| "grad_norm": 0.5940092206001282, | |
| "learning_rate": 0.00010267270847755048, | |
| "loss": 0.3152, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 2.345432977461447, | |
| "grad_norm": 0.6454032063484192, | |
| "learning_rate": 0.00010204393696219117, | |
| "loss": 0.2916, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 2.3549228944246736, | |
| "grad_norm": 0.5936978459358215, | |
| "learning_rate": 0.00010141508459407623, | |
| "loss": 0.3146, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 2.3644128113879006, | |
| "grad_norm": 0.5801250338554382, | |
| "learning_rate": 0.00010078617624894684, | |
| "loss": 0.3129, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 2.373902728351127, | |
| "grad_norm": 0.6835585832595825, | |
| "learning_rate": 0.00010015723680475846, | |
| "loss": 0.3028, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.373902728351127, | |
| "eval_loss": 0.4224609136581421, | |
| "eval_runtime": 20.4149, | |
| "eval_samples_per_second": 15.528, | |
| "eval_steps_per_second": 7.788, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.3833926453143537, | |
| "grad_norm": 0.5739105939865112, | |
| "learning_rate": 9.95282911406968e-05, | |
| "loss": 0.2934, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 2.3928825622775802, | |
| "grad_norm": 0.4810318350791931, | |
| "learning_rate": 9.889936413619356e-05, | |
| "loss": 0.3147, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 2.4023724792408068, | |
| "grad_norm": 0.5853947997093201, | |
| "learning_rate": 9.827048066994225e-05, | |
| "loss": 0.3008, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 2.4118623962040333, | |
| "grad_norm": 0.5773667693138123, | |
| "learning_rate": 9.764166561891432e-05, | |
| "loss": 0.2926, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 2.42135231316726, | |
| "grad_norm": 0.7114972472190857, | |
| "learning_rate": 9.70129438573747e-05, | |
| "loss": 0.3169, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.4308422301304864, | |
| "grad_norm": 0.6915512681007385, | |
| "learning_rate": 9.63843402558981e-05, | |
| "loss": 0.3068, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 2.440332147093713, | |
| "grad_norm": 0.594857394695282, | |
| "learning_rate": 9.57558796803852e-05, | |
| "loss": 0.3196, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 2.4498220640569395, | |
| "grad_norm": 0.6631921529769897, | |
| "learning_rate": 9.512758699107879e-05, | |
| "loss": 0.2999, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 2.459311981020166, | |
| "grad_norm": 0.6086694002151489, | |
| "learning_rate": 9.449948704158071e-05, | |
| "loss": 0.3128, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 2.4688018979833926, | |
| "grad_norm": 0.6541831493377686, | |
| "learning_rate": 9.38716046778684e-05, | |
| "loss": 0.3279, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.478291814946619, | |
| "grad_norm": 0.5784711241722107, | |
| "learning_rate": 9.324396473731217e-05, | |
| "loss": 0.3026, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 2.4877817319098456, | |
| "grad_norm": 0.6732935905456543, | |
| "learning_rate": 9.261659204769284e-05, | |
| "loss": 0.3212, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 2.497271648873072, | |
| "grad_norm": 0.6389648914337158, | |
| "learning_rate": 9.198951142621929e-05, | |
| "loss": 0.3184, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 2.5067615658362987, | |
| "grad_norm": 0.5656270980834961, | |
| "learning_rate": 9.136274767854716e-05, | |
| "loss": 0.31, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 2.5162514827995253, | |
| "grad_norm": 0.61983323097229, | |
| "learning_rate": 9.07363255977973e-05, | |
| "loss": 0.3148, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.525741399762752, | |
| "grad_norm": 0.46594876050949097, | |
| "learning_rate": 9.011026996357503e-05, | |
| "loss": 0.303, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 2.535231316725979, | |
| "grad_norm": 0.6862909197807312, | |
| "learning_rate": 8.948460554099018e-05, | |
| "loss": 0.3072, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 2.5447212336892053, | |
| "grad_norm": 0.6350931525230408, | |
| "learning_rate": 8.885935707967716e-05, | |
| "loss": 0.2909, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 2.554211150652432, | |
| "grad_norm": 0.6156574487686157, | |
| "learning_rate": 8.823454931281616e-05, | |
| "loss": 0.3375, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 2.5637010676156584, | |
| "grad_norm": 0.5949609875679016, | |
| "learning_rate": 8.76102069561545e-05, | |
| "loss": 0.3335, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 2.573190984578885, | |
| "grad_norm": 0.5457854866981506, | |
| "learning_rate": 8.698635470702923e-05, | |
| "loss": 0.3119, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 2.5826809015421115, | |
| "grad_norm": 0.5788692235946655, | |
| "learning_rate": 8.636301724339004e-05, | |
| "loss": 0.2898, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 2.592170818505338, | |
| "grad_norm": 0.5722967982292175, | |
| "learning_rate": 8.574021922282292e-05, | |
| "loss": 0.3079, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 2.6016607354685646, | |
| "grad_norm": 0.6501619815826416, | |
| "learning_rate": 8.511798528157512e-05, | |
| "loss": 0.2971, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 2.611150652431791, | |
| "grad_norm": 0.6138727068901062, | |
| "learning_rate": 8.449634003358022e-05, | |
| "loss": 0.3286, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 2.6206405693950177, | |
| "grad_norm": 0.5789212584495544, | |
| "learning_rate": 8.387530806948476e-05, | |
| "loss": 0.3101, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 2.630130486358244, | |
| "grad_norm": 0.6013932228088379, | |
| "learning_rate": 8.325491395567541e-05, | |
| "loss": 0.2997, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 2.639620403321471, | |
| "grad_norm": 0.5596510767936707, | |
| "learning_rate": 8.263518223330697e-05, | |
| "loss": 0.2928, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 2.6491103202846977, | |
| "grad_norm": 0.7271096706390381, | |
| "learning_rate": 8.201613741733203e-05, | |
| "loss": 0.3144, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 2.6586002372479243, | |
| "grad_norm": 0.715353786945343, | |
| "learning_rate": 8.13978039955308e-05, | |
| "loss": 0.3341, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 2.668090154211151, | |
| "grad_norm": 0.6036480665206909, | |
| "learning_rate": 8.078020642754274e-05, | |
| "loss": 0.3176, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 2.6775800711743774, | |
| "grad_norm": 0.5531415939331055, | |
| "learning_rate": 8.016336914389874e-05, | |
| "loss": 0.3043, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 2.687069988137604, | |
| "grad_norm": 0.5626965165138245, | |
| "learning_rate": 7.954731654505491e-05, | |
| "loss": 0.316, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 2.6965599051008304, | |
| "grad_norm": 0.6845198273658752, | |
| "learning_rate": 7.89320730004274e-05, | |
| "loss": 0.3167, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 2.706049822064057, | |
| "grad_norm": 0.5867395997047424, | |
| "learning_rate": 7.831766284742807e-05, | |
| "loss": 0.3189, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 2.7155397390272835, | |
| "grad_norm": 0.5502896308898926, | |
| "learning_rate": 7.77041103905023e-05, | |
| "loss": 0.3085, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 2.72502965599051, | |
| "grad_norm": 0.6936707496643066, | |
| "learning_rate": 7.709143990016702e-05, | |
| "loss": 0.2824, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 2.7345195729537366, | |
| "grad_norm": 0.6040688157081604, | |
| "learning_rate": 7.6479675612051e-05, | |
| "loss": 0.3005, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 2.744009489916963, | |
| "grad_norm": 0.6335172057151794, | |
| "learning_rate": 7.586884172593609e-05, | |
| "loss": 0.3048, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 2.7534994068801897, | |
| "grad_norm": 0.5545411705970764, | |
| "learning_rate": 7.525896240479976e-05, | |
| "loss": 0.3137, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 2.7629893238434162, | |
| "grad_norm": 0.56629878282547, | |
| "learning_rate": 7.465006177385953e-05, | |
| "loss": 0.3164, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 2.7724792408066428, | |
| "grad_norm": 0.6280866861343384, | |
| "learning_rate": 7.404216391961847e-05, | |
| "loss": 0.3266, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 2.7819691577698693, | |
| "grad_norm": 0.6720747947692871, | |
| "learning_rate": 7.343529288891239e-05, | |
| "loss": 0.2954, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 2.791459074733096, | |
| "grad_norm": 0.7422773838043213, | |
| "learning_rate": 7.282947268795877e-05, | |
| "loss": 0.3037, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 2.8009489916963224, | |
| "grad_norm": 0.6017013192176819, | |
| "learning_rate": 7.222472728140695e-05, | |
| "loss": 0.3007, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.8104389086595494, | |
| "grad_norm": 0.5282939672470093, | |
| "learning_rate": 7.162108059139032e-05, | |
| "loss": 0.2987, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 2.819928825622776, | |
| "grad_norm": 0.6391469240188599, | |
| "learning_rate": 7.101855649657991e-05, | |
| "loss": 0.306, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 2.8294187425860025, | |
| "grad_norm": 0.7014105319976807, | |
| "learning_rate": 7.041717883123977e-05, | |
| "loss": 0.2957, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 2.838908659549229, | |
| "grad_norm": 0.7407575845718384, | |
| "learning_rate": 6.981697138428434e-05, | |
| "loss": 0.2931, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 2.8483985765124555, | |
| "grad_norm": 0.6662490367889404, | |
| "learning_rate": 6.921795789833723e-05, | |
| "loss": 0.2909, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.8483985765124555, | |
| "eval_loss": 0.3944380581378937, | |
| "eval_runtime": 20.3916, | |
| "eval_samples_per_second": 15.546, | |
| "eval_steps_per_second": 7.797, | |
| "step": 3000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 5000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.9630137686215885e+17, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |