{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9981298423724285, "eval_steps": 200, "global_step": 467, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0021373230029388193, "grad_norm": 0.0, "learning_rate": 0.0, "logits": -2.7276527881622314, "logps": -123.19757843017578, "loss": 1.0, "step": 1 }, { "epoch": 0.010686615014694095, "grad_norm": 0.0, "learning_rate": 0.0, "logits": -2.8715224266052246, "logps": -234.59034729003906, "loss": 1.0, "step": 5 }, { "epoch": 0.02137323002938819, "grad_norm": 48.2801432920563, "learning_rate": 5.3191489361702123e-08, "logits": -2.8460676670074463, "logps": -248.16787719726562, "loss": 0.9998, "step": 10 }, { "epoch": 0.03205984504408229, "grad_norm": 50.29241141091707, "learning_rate": 9.574468085106382e-08, "logits": -2.7791006565093994, "logps": -229.27816772460938, "loss": 0.9874, "step": 15 }, { "epoch": 0.04274646005877638, "grad_norm": 46.57864781408822, "learning_rate": 1.3829787234042553e-07, "logits": -2.7698864936828613, "logps": -204.43753051757812, "loss": 0.9444, "step": 20 }, { "epoch": 0.053433075073470476, "grad_norm": 36.984896584442474, "learning_rate": 1.9148936170212765e-07, "logits": -2.9412951469421387, "logps": -291.7972717285156, "loss": 0.795, "step": 25 }, { "epoch": 0.06411969008816458, "grad_norm": 25.64585807242414, "learning_rate": 2.4468085106382976e-07, "logits": -2.913182258605957, "logps": -280.78009033203125, "loss": 0.6761, "step": 30 }, { "epoch": 0.07480630510285867, "grad_norm": 22.189719283428744, "learning_rate": 2.978723404255319e-07, "logits": -2.8854992389678955, "logps": -233.5765380859375, "loss": 0.6236, "step": 35 }, { "epoch": 0.08549292011755276, "grad_norm": 22.990475128782094, "learning_rate": 3.5106382978723405e-07, "logits": -2.8781936168670654, "logps": -233.80117797851562, "loss": 0.5835, "step": 40 }, { "epoch": 0.09617953513224686, "grad_norm": 21.917107832657997, "learning_rate": 4.0425531914893614e-07, "logits": -2.7908034324645996, "logps": -257.6076965332031, "loss": 0.5822, "step": 45 }, { "epoch": 0.10686615014694095, "grad_norm": 22.029695238746953, "learning_rate": 4.574468085106383e-07, "logits": -2.614025592803955, "logps": -270.81671142578125, "loss": 0.5573, "step": 50 }, { "epoch": 0.11755276516163506, "grad_norm": 21.89479772573872, "learning_rate": 4.999930062653174e-07, "logits": -2.7972469329833984, "logps": -286.5457458496094, "loss": 0.5318, "step": 55 }, { "epoch": 0.12823938017632916, "grad_norm": 21.44902343753467, "learning_rate": 4.997482666353286e-07, "logits": -2.690673828125, "logps": -253.75869750976562, "loss": 0.5219, "step": 60 }, { "epoch": 0.13892599519102325, "grad_norm": 21.630693894892058, "learning_rate": 4.991542314714122e-07, "logits": -2.73010516166687, "logps": -293.25762939453125, "loss": 0.5026, "step": 65 }, { "epoch": 0.14961261020571734, "grad_norm": 25.306412397060555, "learning_rate": 4.982117315854593e-07, "logits": -2.5588488578796387, "logps": -263.66851806640625, "loss": 0.5258, "step": 70 }, { "epoch": 0.16029922522041143, "grad_norm": 22.279892008137498, "learning_rate": 4.969220851487844e-07, "logits": -2.401634693145752, "logps": -262.5799255371094, "loss": 0.4803, "step": 75 }, { "epoch": 0.17098584023510552, "grad_norm": 23.85473503890347, "learning_rate": 4.952870958485431e-07, "logits": -2.7048754692077637, "logps": -305.823486328125, "loss": 0.4825, "step": 80 }, { "epoch": 0.18167245524979964, "grad_norm": 26.91529025786325, "learning_rate": 4.933090503651128e-07, "logits": -2.3958709239959717, "logps": -296.0675964355469, "loss": 0.4779, "step": 85 }, { "epoch": 0.19235907026449373, "grad_norm": 23.370663292540115, "learning_rate": 4.909907151739633e-07, "logits": -2.552422046661377, "logps": -235.0645751953125, "loss": 0.479, "step": 90 }, { "epoch": 0.20304568527918782, "grad_norm": 24.61367431074327, "learning_rate": 4.883353326764906e-07, "logits": -2.3946549892425537, "logps": -295.76959228515625, "loss": 0.4708, "step": 95 }, { "epoch": 0.2137323002938819, "grad_norm": 25.53290336307927, "learning_rate": 4.853466166652258e-07, "logits": -2.442629098892212, "logps": -241.5957794189453, "loss": 0.4609, "step": 100 }, { "epoch": 0.224418915308576, "grad_norm": 22.113276785856915, "learning_rate": 4.820287471297597e-07, "logits": -2.474640369415283, "logps": -290.6409912109375, "loss": 0.4479, "step": 105 }, { "epoch": 0.2351055303232701, "grad_norm": 21.079178522203513, "learning_rate": 4.783863644106502e-07, "logits": -2.515345335006714, "logps": -306.5741882324219, "loss": 0.4529, "step": 110 }, { "epoch": 0.2457921453379642, "grad_norm": 21.06551183306965, "learning_rate": 4.752422169756047e-07, "logits": -2.4756886959075928, "logps": -281.9891662597656, "loss": 0.4696, "step": 115 }, { "epoch": 0.2564787603526583, "grad_norm": 22.63241646221903, "learning_rate": 4.710288483761524e-07, "logits": -2.4553544521331787, "logps": -262.58184814453125, "loss": 0.4712, "step": 120 }, { "epoch": 0.2671653753673524, "grad_norm": 24.782221235525792, "learning_rate": 4.6650635094610966e-07, "logits": -2.3455586433410645, "logps": -286.12481689453125, "loss": 0.439, "step": 125 }, { "epoch": 0.2778519903820465, "grad_norm": 22.240996330864814, "learning_rate": 4.6168104980707103e-07, "logits": -2.454857349395752, "logps": -278.26177978515625, "loss": 0.4606, "step": 130 }, { "epoch": 0.2885386053967406, "grad_norm": 23.029344848112892, "learning_rate": 4.565596935789987e-07, "logits": -2.5119175910949707, "logps": -271.8935546875, "loss": 0.4583, "step": 135 }, { "epoch": 0.2992252204114347, "grad_norm": 23.114231775090644, "learning_rate": 4.511494449416671e-07, "logits": -2.4763400554656982, "logps": -305.2304992675781, "loss": 0.4444, "step": 140 }, { "epoch": 0.30991183542612877, "grad_norm": 22.307052406807973, "learning_rate": 4.4545787061700746e-07, "logits": -2.279383659362793, "logps": -314.3868103027344, "loss": 0.4433, "step": 145 }, { "epoch": 0.32059845044082286, "grad_norm": 22.621940716026984, "learning_rate": 4.394929307863632e-07, "logits": -2.4016177654266357, "logps": -280.4060363769531, "loss": 0.4603, "step": 150 }, { "epoch": 0.33128506545551695, "grad_norm": 27.05574310923442, "learning_rate": 4.332629679574565e-07, "logits": -2.4656858444213867, "logps": -310.09173583984375, "loss": 0.4505, "step": 155 }, { "epoch": 0.34197168047021104, "grad_norm": 23.53583613618808, "learning_rate": 4.2677669529663686e-07, "logits": -2.337423801422119, "logps": -317.040771484375, "loss": 0.4395, "step": 160 }, { "epoch": 0.3526582954849052, "grad_norm": 23.875484303552483, "learning_rate": 4.200431844427298e-07, "logits": -2.5105397701263428, "logps": -305.39630126953125, "loss": 0.4456, "step": 165 }, { "epoch": 0.36334491049959927, "grad_norm": 22.450514799974165, "learning_rate": 4.130718528195303e-07, "logits": -2.3933465480804443, "logps": -275.6124267578125, "loss": 0.443, "step": 170 }, { "epoch": 0.37403152551429336, "grad_norm": 23.74049455633747, "learning_rate": 4.058724504646834e-07, "logits": -2.3835606575012207, "logps": -277.36956787109375, "loss": 0.4339, "step": 175 }, { "epoch": 0.38471814052898745, "grad_norm": 23.64677950778481, "learning_rate": 3.9845504639337535e-07, "logits": -2.4487411975860596, "logps": -290.956298828125, "loss": 0.4573, "step": 180 }, { "epoch": 0.39540475554368154, "grad_norm": 21.598937457276417, "learning_rate": 3.908300145159055e-07, "logits": -2.413297176361084, "logps": -235.90109252929688, "loss": 0.4279, "step": 185 }, { "epoch": 0.40609137055837563, "grad_norm": 23.712081510750036, "learning_rate": 3.8300801912883414e-07, "logits": -2.4640491008758545, "logps": -287.1435546875, "loss": 0.4354, "step": 190 }, { "epoch": 0.4167779855730697, "grad_norm": 44.372947453047885, "learning_rate": 3.75e-07, "logits": -2.3341147899627686, "logps": -343.4080505371094, "loss": 0.4329, "step": 195 }, { "epoch": 0.4274646005877638, "grad_norm": 23.012230002196524, "learning_rate": 3.668171570682655e-07, "logits": -2.340860366821289, "logps": -262.5423889160156, "loss": 0.431, "step": 200 }, { "epoch": 0.4274646005877638, "eval_logits": -2.3263635635375977, "eval_logps": -301.7860412597656, "eval_loss": 0.4314996898174286, "eval_runtime": 507.0323, "eval_samples_per_second": 3.881, "eval_steps_per_second": 0.243, "step": 200 }, { "epoch": 0.4381512156024579, "grad_norm": 22.412220925103565, "learning_rate": 3.584709347793895e-07, "logits": -2.2411904335021973, "logps": -277.8357849121094, "loss": 0.4425, "step": 205 }, { "epoch": 0.448837830617152, "grad_norm": 23.021158257266347, "learning_rate": 3.499730060799352e-07, "logits": -2.1322011947631836, "logps": -289.6743469238281, "loss": 0.4292, "step": 210 }, { "epoch": 0.45952444563184613, "grad_norm": 23.724850433619796, "learning_rate": 3.413352560915988e-07, "logits": -2.0033411979675293, "logps": -272.5525207519531, "loss": 0.4121, "step": 215 }, { "epoch": 0.4702110606465402, "grad_norm": 21.651299269242514, "learning_rate": 3.325697654887918e-07, "logits": -2.051339626312256, "logps": -235.84262084960938, "loss": 0.4222, "step": 220 }, { "epoch": 0.4808976756612343, "grad_norm": 25.440752125826293, "learning_rate": 3.2368879360272606e-07, "logits": -2.34849214553833, "logps": -312.7589111328125, "loss": 0.424, "step": 225 }, { "epoch": 0.4915842906759284, "grad_norm": 21.05394354899474, "learning_rate": 3.147047612756302e-07, "logits": -2.335400104522705, "logps": -325.1053161621094, "loss": 0.4261, "step": 230 }, { "epoch": 0.5022709056906225, "grad_norm": 22.293568633261653, "learning_rate": 3.056302334890786e-07, "logits": -2.2490334510803223, "logps": -266.56219482421875, "loss": 0.4322, "step": 235 }, { "epoch": 0.5129575207053166, "grad_norm": 22.256364087698653, "learning_rate": 2.964779017907287e-07, "logits": -2.16744065284729, "logps": -309.4420471191406, "loss": 0.4222, "step": 240 }, { "epoch": 0.5236441357200107, "grad_norm": 25.954148860680444, "learning_rate": 2.872605665440436e-07, "logits": -2.451477527618408, "logps": -282.71319580078125, "loss": 0.4096, "step": 245 }, { "epoch": 0.5343307507347048, "grad_norm": 21.39918862577877, "learning_rate": 2.7799111902582693e-07, "logits": -2.6052417755126953, "logps": -273.9092102050781, "loss": 0.4155, "step": 250 }, { "epoch": 0.5450173657493989, "grad_norm": 24.967131244901736, "learning_rate": 2.6868252339660607e-07, "logits": -2.225090503692627, "logps": -280.70782470703125, "loss": 0.416, "step": 255 }, { "epoch": 0.555703980764093, "grad_norm": 24.077110482464793, "learning_rate": 2.593477985690815e-07, "logits": -2.2254879474639893, "logps": -286.8822937011719, "loss": 0.4099, "step": 260 }, { "epoch": 0.566390595778787, "grad_norm": 22.910756064363806, "learning_rate": 2.5e-07, "logits": -2.3557419776916504, "logps": -280.0397033691406, "loss": 0.4125, "step": 265 }, { "epoch": 0.5770772107934812, "grad_norm": 27.974836628803857, "learning_rate": 2.406522014309186e-07, "logits": -2.041544198989868, "logps": -273.45318603515625, "loss": 0.4291, "step": 270 }, { "epoch": 0.5877638258081752, "grad_norm": 22.765553156216026, "learning_rate": 2.3131747660339394e-07, "logits": -2.3228344917297363, "logps": -282.9244079589844, "loss": 0.4141, "step": 275 }, { "epoch": 0.5984504408228694, "grad_norm": 22.498986802146685, "learning_rate": 2.2200888097417302e-07, "logits": -1.89029061794281, "logps": -315.2174072265625, "loss": 0.42, "step": 280 }, { "epoch": 0.6091370558375635, "grad_norm": 23.58891369147268, "learning_rate": 2.1273943345595635e-07, "logits": -1.903066873550415, "logps": -261.4866638183594, "loss": 0.4062, "step": 285 }, { "epoch": 0.6198236708522575, "grad_norm": 21.26751900271618, "learning_rate": 2.0352209820927135e-07, "logits": -2.048973560333252, "logps": -312.31048583984375, "loss": 0.4142, "step": 290 }, { "epoch": 0.6305102858669517, "grad_norm": 22.071346201309105, "learning_rate": 1.9436976651092142e-07, "logits": -1.7636706829071045, "logps": -257.8403625488281, "loss": 0.4124, "step": 295 }, { "epoch": 0.6411969008816457, "grad_norm": 21.12974848393809, "learning_rate": 1.8529523872436977e-07, "logits": -2.014808177947998, "logps": -279.1208190917969, "loss": 0.4112, "step": 300 }, { "epoch": 0.6518835158963399, "grad_norm": 23.594905022803008, "learning_rate": 1.763112063972739e-07, "logits": -2.050807237625122, "logps": -296.3996887207031, "loss": 0.4112, "step": 305 }, { "epoch": 0.6625701309110339, "grad_norm": 21.591511525006933, "learning_rate": 1.674302345112083e-07, "logits": -1.986161470413208, "logps": -275.21893310546875, "loss": 0.419, "step": 310 }, { "epoch": 0.673256745925728, "grad_norm": 20.6885856376684, "learning_rate": 1.5866474390840124e-07, "logits": -2.2522599697113037, "logps": -319.27911376953125, "loss": 0.4106, "step": 315 }, { "epoch": 0.6839433609404221, "grad_norm": 22.909433086566732, "learning_rate": 1.500269939200648e-07, "logits": -2.207468032836914, "logps": -266.585205078125, "loss": 0.3909, "step": 320 }, { "epoch": 0.6946299759551162, "grad_norm": 22.282735602270293, "learning_rate": 1.4152906522061047e-07, "logits": -2.35794997215271, "logps": -316.83807373046875, "loss": 0.4048, "step": 325 }, { "epoch": 0.7053165909698104, "grad_norm": 23.16656248515272, "learning_rate": 1.3318284293173449e-07, "logits": -2.242034912109375, "logps": -307.3307189941406, "loss": 0.4042, "step": 330 }, { "epoch": 0.7160032059845044, "grad_norm": 21.510083315851972, "learning_rate": 1.2500000000000005e-07, "logits": -1.9815715551376343, "logps": -338.7240905761719, "loss": 0.3934, "step": 335 }, { "epoch": 0.7266898209991985, "grad_norm": 21.492689286630217, "learning_rate": 1.1699198087116588e-07, "logits": -2.1296348571777344, "logps": -271.06878662109375, "loss": 0.4011, "step": 340 }, { "epoch": 0.7373764360138926, "grad_norm": 23.01568750995505, "learning_rate": 1.0916998548409447e-07, "logits": -2.244135618209839, "logps": -310.14501953125, "loss": 0.412, "step": 345 }, { "epoch": 0.7480630510285867, "grad_norm": 22.53687051047034, "learning_rate": 1.0154495360662463e-07, "logits": -2.1778364181518555, "logps": -269.3367614746094, "loss": 0.3901, "step": 350 }, { "epoch": 0.7587496660432808, "grad_norm": 23.184829487659847, "learning_rate": 9.412754953531663e-08, "logits": -2.286181926727295, "logps": -271.3564453125, "loss": 0.4172, "step": 355 }, { "epoch": 0.7694362810579749, "grad_norm": 21.040610617000496, "learning_rate": 8.692814718046978e-08, "logits": -2.055994987487793, "logps": -304.63885498046875, "loss": 0.4083, "step": 360 }, { "epoch": 0.7801228960726689, "grad_norm": 23.09698562414689, "learning_rate": 7.99568155572701e-08, "logits": -2.213494062423706, "logps": -277.4697570800781, "loss": 0.4155, "step": 365 }, { "epoch": 0.7908095110873631, "grad_norm": 21.236358981933506, "learning_rate": 7.322330470336313e-08, "logits": -2.0819058418273926, "logps": -281.2724914550781, "loss": 0.4037, "step": 370 }, { "epoch": 0.8014961261020572, "grad_norm": 26.06067361683024, "learning_rate": 6.673703204254347e-08, "logits": -2.030791997909546, "logps": -284.4011535644531, "loss": 0.4121, "step": 375 }, { "epoch": 0.8121827411167513, "grad_norm": 20.224636468750884, "learning_rate": 6.050706921363672e-08, "logits": -1.9256960153579712, "logps": -294.9930725097656, "loss": 0.4027, "step": 380 }, { "epoch": 0.8228693561314454, "grad_norm": 24.14425420434055, "learning_rate": 5.454212938299255e-08, "logits": -1.7693021297454834, "logps": -286.45452880859375, "loss": 0.4136, "step": 385 }, { "epoch": 0.8335559711461394, "grad_norm": 22.762479440937984, "learning_rate": 4.885055505833291e-08, "logits": -1.9304530620574951, "logps": -291.02880859375, "loss": 0.3988, "step": 390 }, { "epoch": 0.8442425861608336, "grad_norm": 21.312231537009946, "learning_rate": 4.3440306421001324e-08, "logits": -2.025493860244751, "logps": -268.2240905761719, "loss": 0.4083, "step": 395 }, { "epoch": 0.8549292011755276, "grad_norm": 23.03482969018928, "learning_rate": 3.831895019292897e-08, "logits": -2.1968250274658203, "logps": -294.18292236328125, "loss": 0.4118, "step": 400 }, { "epoch": 0.8549292011755276, "eval_logits": -2.2095947265625, "eval_logps": -310.0013122558594, "eval_loss": 0.40863949060440063, "eval_runtime": 537.6528, "eval_samples_per_second": 3.66, "eval_steps_per_second": 0.229, "step": 400 }, { "epoch": 0.8656158161902218, "grad_norm": 21.5208583380133, "learning_rate": 3.349364905389032e-08, "logits": -2.188328981399536, "logps": -306.3011169433594, "loss": 0.4113, "step": 405 }, { "epoch": 0.8763024312049158, "grad_norm": 20.54362762134308, "learning_rate": 2.8971151623847584e-08, "logits": -1.99441659450531, "logps": -286.16876220703125, "loss": 0.4171, "step": 410 }, { "epoch": 0.88698904621961, "grad_norm": 22.86656655422917, "learning_rate": 2.475778302439524e-08, "logits": -2.1187539100646973, "logps": -266.64959716796875, "loss": 0.4053, "step": 415 }, { "epoch": 0.897675661234304, "grad_norm": 22.18589761074244, "learning_rate": 2.085943603250595e-08, "logits": -2.3276543617248535, "logps": -284.54974365234375, "loss": 0.4153, "step": 420 }, { "epoch": 0.9083622762489981, "grad_norm": 20.63324288656286, "learning_rate": 1.7281562838948966e-08, "logits": -2.1507515907287598, "logps": -281.44158935546875, "loss": 0.4074, "step": 425 }, { "epoch": 0.9190488912636923, "grad_norm": 20.358726620730323, "learning_rate": 1.4029167422908105e-08, "logits": -2.1837050914764404, "logps": -310.3664245605469, "loss": 0.3928, "step": 430 }, { "epoch": 0.9297355062783863, "grad_norm": 21.126634047747054, "learning_rate": 1.1106798553464802e-08, "logits": -2.260655164718628, "logps": -281.1075439453125, "loss": 0.413, "step": 435 }, { "epoch": 0.9404221212930804, "grad_norm": 23.274529002886094, "learning_rate": 8.518543427732949e-09, "logits": -2.16852068901062, "logps": -262.402099609375, "loss": 0.3922, "step": 440 }, { "epoch": 0.9511087363077745, "grad_norm": 26.435895904221407, "learning_rate": 6.268021954544095e-09, "logits": -2.1322624683380127, "logps": -324.4895324707031, "loss": 0.4172, "step": 445 }, { "epoch": 0.9617953513224686, "grad_norm": 24.896799022635523, "learning_rate": 4.358381691677931e-09, "logits": -2.1141066551208496, "logps": -289.3265380859375, "loss": 0.4068, "step": 450 }, { "epoch": 0.9724819663371627, "grad_norm": 21.21043979653372, "learning_rate": 2.7922934437178692e-09, "logits": -1.9867866039276123, "logps": -283.8013000488281, "loss": 0.391, "step": 455 }, { "epoch": 0.9831685813518568, "grad_norm": 21.01754870372027, "learning_rate": 1.5719475266893489e-09, "logits": -2.012530565261841, "logps": -245.16806030273438, "loss": 0.4041, "step": 460 }, { "epoch": 0.9938551963665508, "grad_norm": 22.43911437197679, "learning_rate": 6.990507047049676e-10, "logits": -2.1750540733337402, "logps": -284.3227233886719, "loss": 0.4094, "step": 465 }, { "epoch": 0.9981298423724285, "step": 467, "total_flos": 0.0, "train_loss": 0.4667486662017201, "train_runtime": 41187.6602, "train_samples_per_second": 1.454, "train_steps_per_second": 0.011 } ], "logging_steps": 5, "max_steps": 467, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }