functionary-small-v3.1 / trainer_state.json
khaimai's picture
Upload folder using huggingface_hub
a8a2ec0 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9993174061433447,
"eval_steps": 183,
"global_step": 549,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0018202502844141069,
"grad_norm": 7.828993836368125,
"learning_rate": 4.705882352941176e-07,
"loss": 0.7927,
"step": 1
},
{
"epoch": 0.0036405005688282138,
"grad_norm": 6.714284407371177,
"learning_rate": 9.411764705882352e-07,
"loss": 0.7939,
"step": 2
},
{
"epoch": 0.005460750853242321,
"grad_norm": 6.364056661574126,
"learning_rate": 1.411764705882353e-06,
"loss": 0.8186,
"step": 3
},
{
"epoch": 0.0072810011376564275,
"grad_norm": 6.001686432641966,
"learning_rate": 1.8823529411764705e-06,
"loss": 0.7232,
"step": 4
},
{
"epoch": 0.009101251422070534,
"grad_norm": 5.534730246504558,
"learning_rate": 2.352941176470588e-06,
"loss": 0.7891,
"step": 5
},
{
"epoch": 0.010921501706484642,
"grad_norm": 3.4121020699424713,
"learning_rate": 2.823529411764706e-06,
"loss": 0.7351,
"step": 6
},
{
"epoch": 0.01274175199089875,
"grad_norm": 3.9324785806724476,
"learning_rate": 3.294117647058823e-06,
"loss": 0.6899,
"step": 7
},
{
"epoch": 0.014562002275312855,
"grad_norm": 2.5947504275059496,
"learning_rate": 3.764705882352941e-06,
"loss": 0.6103,
"step": 8
},
{
"epoch": 0.016382252559726963,
"grad_norm": 2.5403086872594955,
"learning_rate": 4.235294117647058e-06,
"loss": 0.6377,
"step": 9
},
{
"epoch": 0.01820250284414107,
"grad_norm": 2.339440216808695,
"learning_rate": 4.705882352941176e-06,
"loss": 0.6723,
"step": 10
},
{
"epoch": 0.020022753128555178,
"grad_norm": 2.1983237574211922,
"learning_rate": 5.176470588235294e-06,
"loss": 0.684,
"step": 11
},
{
"epoch": 0.021843003412969283,
"grad_norm": 1.6424427461297102,
"learning_rate": 5.647058823529412e-06,
"loss": 0.6027,
"step": 12
},
{
"epoch": 0.02366325369738339,
"grad_norm": 1.6005922788724476,
"learning_rate": 6.1176470588235285e-06,
"loss": 0.7195,
"step": 13
},
{
"epoch": 0.0254835039817975,
"grad_norm": 1.8438931230594375,
"learning_rate": 6.588235294117646e-06,
"loss": 0.6329,
"step": 14
},
{
"epoch": 0.027303754266211604,
"grad_norm": 1.8086010126416687,
"learning_rate": 7.058823529411764e-06,
"loss": 0.6812,
"step": 15
},
{
"epoch": 0.02912400455062571,
"grad_norm": 1.6390713129791323,
"learning_rate": 7.529411764705882e-06,
"loss": 0.6848,
"step": 16
},
{
"epoch": 0.03094425483503982,
"grad_norm": 1.5725949727834614,
"learning_rate": 8e-06,
"loss": 0.6415,
"step": 17
},
{
"epoch": 0.032764505119453925,
"grad_norm": 1.5078400807519021,
"learning_rate": 7.999930256262932e-06,
"loss": 0.6284,
"step": 18
},
{
"epoch": 0.03458475540386803,
"grad_norm": 1.5292651701417765,
"learning_rate": 7.999721027483818e-06,
"loss": 0.6503,
"step": 19
},
{
"epoch": 0.03640500568828214,
"grad_norm": 1.4983054037314452,
"learning_rate": 7.999372320958861e-06,
"loss": 0.6167,
"step": 20
},
{
"epoch": 0.03822525597269624,
"grad_norm": 1.4412298959675836,
"learning_rate": 7.998884148848109e-06,
"loss": 0.6245,
"step": 21
},
{
"epoch": 0.040045506257110355,
"grad_norm": 1.2365167423542114,
"learning_rate": 7.998256528175033e-06,
"loss": 0.5953,
"step": 22
},
{
"epoch": 0.04186575654152446,
"grad_norm": 1.5357829692377438,
"learning_rate": 7.997489480825941e-06,
"loss": 0.6367,
"step": 23
},
{
"epoch": 0.04368600682593857,
"grad_norm": 1.3171822289121358,
"learning_rate": 7.996583033549204e-06,
"loss": 0.5577,
"step": 24
},
{
"epoch": 0.04550625711035267,
"grad_norm": 1.3245922461587984,
"learning_rate": 7.995537217954335e-06,
"loss": 0.5706,
"step": 25
},
{
"epoch": 0.04732650739476678,
"grad_norm": 1.4166210756319584,
"learning_rate": 7.994352070510876e-06,
"loss": 0.6612,
"step": 26
},
{
"epoch": 0.049146757679180884,
"grad_norm": 1.4367785667776511,
"learning_rate": 7.993027632547137e-06,
"loss": 0.5766,
"step": 27
},
{
"epoch": 0.050967007963595,
"grad_norm": 1.2803978303957413,
"learning_rate": 7.991563950248739e-06,
"loss": 0.6023,
"step": 28
},
{
"epoch": 0.0527872582480091,
"grad_norm": 1.3201964359281728,
"learning_rate": 7.989961074657023e-06,
"loss": 0.6026,
"step": 29
},
{
"epoch": 0.05460750853242321,
"grad_norm": 1.5466402476684098,
"learning_rate": 7.988219061667252e-06,
"loss": 0.5979,
"step": 30
},
{
"epoch": 0.056427758816837315,
"grad_norm": 1.372807482234666,
"learning_rate": 7.986337972026678e-06,
"loss": 0.5928,
"step": 31
},
{
"epoch": 0.05824800910125142,
"grad_norm": 1.2476758007512014,
"learning_rate": 7.98431787133241e-06,
"loss": 0.5506,
"step": 32
},
{
"epoch": 0.060068259385665526,
"grad_norm": 1.3332407362456573,
"learning_rate": 7.982158830029133e-06,
"loss": 0.5252,
"step": 33
},
{
"epoch": 0.06188850967007964,
"grad_norm": 1.2956829613552345,
"learning_rate": 7.979860923406654e-06,
"loss": 0.6162,
"step": 34
},
{
"epoch": 0.06370875995449374,
"grad_norm": 1.3358171325973744,
"learning_rate": 7.977424231597266e-06,
"loss": 0.6323,
"step": 35
},
{
"epoch": 0.06552901023890785,
"grad_norm": 1.2668022187917536,
"learning_rate": 7.97484883957297e-06,
"loss": 0.5481,
"step": 36
},
{
"epoch": 0.06734926052332196,
"grad_norm": 1.3032578120954865,
"learning_rate": 7.972134837142497e-06,
"loss": 0.6982,
"step": 37
},
{
"epoch": 0.06916951080773606,
"grad_norm": 1.438077454657845,
"learning_rate": 7.969282318948179e-06,
"loss": 0.6386,
"step": 38
},
{
"epoch": 0.07098976109215017,
"grad_norm": 1.2348261155030926,
"learning_rate": 7.966291384462662e-06,
"loss": 0.5691,
"step": 39
},
{
"epoch": 0.07281001137656427,
"grad_norm": 1.7781144758356542,
"learning_rate": 7.963162137985416e-06,
"loss": 0.6133,
"step": 40
},
{
"epoch": 0.07463026166097839,
"grad_norm": 1.3915614778891998,
"learning_rate": 7.959894688639114e-06,
"loss": 0.6097,
"step": 41
},
{
"epoch": 0.07645051194539249,
"grad_norm": 1.3714026109891253,
"learning_rate": 7.956489150365818e-06,
"loss": 0.7127,
"step": 42
},
{
"epoch": 0.0782707622298066,
"grad_norm": 1.2639378240340353,
"learning_rate": 7.952945641923014e-06,
"loss": 0.5649,
"step": 43
},
{
"epoch": 0.08009101251422071,
"grad_norm": 1.3658987432003644,
"learning_rate": 7.949264286879461e-06,
"loss": 0.5975,
"step": 44
},
{
"epoch": 0.08191126279863481,
"grad_norm": 1.3869528512684368,
"learning_rate": 7.94544521361089e-06,
"loss": 0.5851,
"step": 45
},
{
"epoch": 0.08373151308304892,
"grad_norm": 1.3457362580768018,
"learning_rate": 7.941488555295519e-06,
"loss": 0.6241,
"step": 46
},
{
"epoch": 0.08555176336746302,
"grad_norm": 1.3016443215121218,
"learning_rate": 7.937394449909417e-06,
"loss": 0.5603,
"step": 47
},
{
"epoch": 0.08737201365187713,
"grad_norm": 1.4891468132527028,
"learning_rate": 7.933163040221691e-06,
"loss": 0.6103,
"step": 48
},
{
"epoch": 0.08919226393629125,
"grad_norm": 1.3296810848522813,
"learning_rate": 7.928794473789502e-06,
"loss": 0.5823,
"step": 49
},
{
"epoch": 0.09101251422070535,
"grad_norm": 1.5647412617424314,
"learning_rate": 7.924288902952924e-06,
"loss": 0.6222,
"step": 50
},
{
"epoch": 0.09283276450511946,
"grad_norm": 1.224983623587384,
"learning_rate": 7.91964648482963e-06,
"loss": 0.5779,
"step": 51
},
{
"epoch": 0.09465301478953356,
"grad_norm": 1.3384527596977494,
"learning_rate": 7.914867381309417e-06,
"loss": 0.5721,
"step": 52
},
{
"epoch": 0.09647326507394767,
"grad_norm": 1.3725597146198896,
"learning_rate": 7.909951759048553e-06,
"loss": 0.6531,
"step": 53
},
{
"epoch": 0.09829351535836177,
"grad_norm": 1.4825184286499173,
"learning_rate": 7.904899789463974e-06,
"loss": 0.5767,
"step": 54
},
{
"epoch": 0.10011376564277588,
"grad_norm": 1.2710681313032695,
"learning_rate": 7.899711648727295e-06,
"loss": 0.5447,
"step": 55
},
{
"epoch": 0.10193401592719,
"grad_norm": 1.468311538305991,
"learning_rate": 7.894387517758679e-06,
"loss": 0.6303,
"step": 56
},
{
"epoch": 0.1037542662116041,
"grad_norm": 1.1661278644804365,
"learning_rate": 7.888927582220521e-06,
"loss": 0.606,
"step": 57
},
{
"epoch": 0.1055745164960182,
"grad_norm": 1.3217868795788283,
"learning_rate": 7.883332032510978e-06,
"loss": 0.5329,
"step": 58
},
{
"epoch": 0.1073947667804323,
"grad_norm": 1.33515774996865,
"learning_rate": 7.877601063757322e-06,
"loss": 0.5335,
"step": 59
},
{
"epoch": 0.10921501706484642,
"grad_norm": 1.4916296734659389,
"learning_rate": 7.871734875809141e-06,
"loss": 0.5705,
"step": 60
},
{
"epoch": 0.11103526734926053,
"grad_norm": 1.3630503455015932,
"learning_rate": 7.86573367323137e-06,
"loss": 0.6279,
"step": 61
},
{
"epoch": 0.11285551763367463,
"grad_norm": 1.2409472798363426,
"learning_rate": 7.859597665297158e-06,
"loss": 0.5096,
"step": 62
},
{
"epoch": 0.11467576791808874,
"grad_norm": 1.241131166674189,
"learning_rate": 7.853327065980567e-06,
"loss": 0.5792,
"step": 63
},
{
"epoch": 0.11649601820250284,
"grad_norm": 1.2488483644853932,
"learning_rate": 7.84692209394911e-06,
"loss": 0.5191,
"step": 64
},
{
"epoch": 0.11831626848691695,
"grad_norm": 1.4362376863954367,
"learning_rate": 7.84038297255613e-06,
"loss": 0.5749,
"step": 65
},
{
"epoch": 0.12013651877133105,
"grad_norm": 1.301660597552013,
"learning_rate": 7.83370992983301e-06,
"loss": 0.5598,
"step": 66
},
{
"epoch": 0.12195676905574517,
"grad_norm": 1.2871188302606258,
"learning_rate": 7.826903198481218e-06,
"loss": 0.6357,
"step": 67
},
{
"epoch": 0.12377701934015928,
"grad_norm": 1.2412889792540172,
"learning_rate": 7.819963015864195e-06,
"loss": 0.6025,
"step": 68
},
{
"epoch": 0.12559726962457338,
"grad_norm": 1.2417123308251397,
"learning_rate": 7.812889623999077e-06,
"loss": 0.5973,
"step": 69
},
{
"epoch": 0.12741751990898748,
"grad_norm": 1.334500041209377,
"learning_rate": 7.805683269548253e-06,
"loss": 0.5339,
"step": 70
},
{
"epoch": 0.1292377701934016,
"grad_norm": 1.2800707133087328,
"learning_rate": 7.798344203810772e-06,
"loss": 0.5506,
"step": 71
},
{
"epoch": 0.1310580204778157,
"grad_norm": 1.2634004680746123,
"learning_rate": 7.790872682713567e-06,
"loss": 0.554,
"step": 72
},
{
"epoch": 0.1328782707622298,
"grad_norm": 1.3715962042311087,
"learning_rate": 7.783268966802538e-06,
"loss": 0.5949,
"step": 73
},
{
"epoch": 0.13469852104664393,
"grad_norm": 1.4002060242886838,
"learning_rate": 7.77553332123347e-06,
"loss": 0.6422,
"step": 74
},
{
"epoch": 0.13651877133105803,
"grad_norm": 1.2759192431077615,
"learning_rate": 7.767666015762775e-06,
"loss": 0.607,
"step": 75
},
{
"epoch": 0.13833902161547212,
"grad_norm": 1.6921865669723448,
"learning_rate": 7.7596673247381e-06,
"loss": 0.6002,
"step": 76
},
{
"epoch": 0.14015927189988622,
"grad_norm": 1.3571851968954738,
"learning_rate": 7.751537527088742e-06,
"loss": 0.5215,
"step": 77
},
{
"epoch": 0.14197952218430035,
"grad_norm": 1.4785928356534102,
"learning_rate": 7.743276906315936e-06,
"loss": 0.6101,
"step": 78
},
{
"epoch": 0.14379977246871445,
"grad_norm": 1.465222696303414,
"learning_rate": 7.734885750482967e-06,
"loss": 0.6187,
"step": 79
},
{
"epoch": 0.14562002275312855,
"grad_norm": 1.264573511241066,
"learning_rate": 7.726364352205116e-06,
"loss": 0.5673,
"step": 80
},
{
"epoch": 0.14744027303754267,
"grad_norm": 1.238555624330946,
"learning_rate": 7.717713008639463e-06,
"loss": 0.6066,
"step": 81
},
{
"epoch": 0.14926052332195677,
"grad_norm": 1.20038139697854,
"learning_rate": 7.708932021474524e-06,
"loss": 0.5678,
"step": 82
},
{
"epoch": 0.15108077360637087,
"grad_norm": 1.3190323958334018,
"learning_rate": 7.70002169691973e-06,
"loss": 0.5544,
"step": 83
},
{
"epoch": 0.15290102389078497,
"grad_norm": 1.3333392166861238,
"learning_rate": 7.690982345694746e-06,
"loss": 0.5212,
"step": 84
},
{
"epoch": 0.1547212741751991,
"grad_norm": 1.5189079377057624,
"learning_rate": 7.68181428301864e-06,
"loss": 0.5411,
"step": 85
},
{
"epoch": 0.1565415244596132,
"grad_norm": 1.2341512401643826,
"learning_rate": 7.67251782859889e-06,
"loss": 0.5,
"step": 86
},
{
"epoch": 0.1583617747440273,
"grad_norm": 1.3710782832898465,
"learning_rate": 7.663093306620228e-06,
"loss": 0.567,
"step": 87
},
{
"epoch": 0.16018202502844142,
"grad_norm": 1.3224943188093254,
"learning_rate": 7.653541045733351e-06,
"loss": 0.6514,
"step": 88
},
{
"epoch": 0.16200227531285552,
"grad_norm": 1.3825136099871158,
"learning_rate": 7.643861379043442e-06,
"loss": 0.49,
"step": 89
},
{
"epoch": 0.16382252559726962,
"grad_norm": 1.6656667334345212,
"learning_rate": 7.634054644098566e-06,
"loss": 0.649,
"step": 90
},
{
"epoch": 0.16564277588168372,
"grad_norm": 1.3683202937271444,
"learning_rate": 7.624121182877892e-06,
"loss": 0.497,
"step": 91
},
{
"epoch": 0.16746302616609784,
"grad_norm": 1.5512864267072353,
"learning_rate": 7.614061341779777e-06,
"loss": 0.6176,
"step": 92
},
{
"epoch": 0.16928327645051194,
"grad_norm": 1.5790193819370095,
"learning_rate": 7.6038754716096755e-06,
"loss": 0.5634,
"step": 93
},
{
"epoch": 0.17110352673492604,
"grad_norm": 1.4344008312589909,
"learning_rate": 7.593563927567915e-06,
"loss": 0.5932,
"step": 94
},
{
"epoch": 0.17292377701934017,
"grad_norm": 1.2510278162330568,
"learning_rate": 7.583127069237302e-06,
"loss": 0.5604,
"step": 95
},
{
"epoch": 0.17474402730375427,
"grad_norm": 1.1926891094591303,
"learning_rate": 7.5725652605705876e-06,
"loss": 0.5746,
"step": 96
},
{
"epoch": 0.17656427758816837,
"grad_norm": 1.3458065119541616,
"learning_rate": 7.561878869877778e-06,
"loss": 0.5,
"step": 97
},
{
"epoch": 0.1783845278725825,
"grad_norm": 1.311426359460282,
"learning_rate": 7.551068269813282e-06,
"loss": 0.503,
"step": 98
},
{
"epoch": 0.1802047781569966,
"grad_norm": 1.2792152183438508,
"learning_rate": 7.540133837362924e-06,
"loss": 0.5279,
"step": 99
},
{
"epoch": 0.1820250284414107,
"grad_norm": 1.2349765362905594,
"learning_rate": 7.5290759538307944e-06,
"loss": 0.5159,
"step": 100
},
{
"epoch": 0.1838452787258248,
"grad_norm": 1.3587018583909733,
"learning_rate": 7.517895004825955e-06,
"loss": 0.573,
"step": 101
},
{
"epoch": 0.18566552901023892,
"grad_norm": 1.3554993825796526,
"learning_rate": 7.506591380248991e-06,
"loss": 0.5801,
"step": 102
},
{
"epoch": 0.18748577929465302,
"grad_norm": 1.2364653944665345,
"learning_rate": 7.495165474278411e-06,
"loss": 0.5618,
"step": 103
},
{
"epoch": 0.18930602957906711,
"grad_norm": 1.226975873209754,
"learning_rate": 7.483617685356906e-06,
"loss": 0.6663,
"step": 104
},
{
"epoch": 0.19112627986348124,
"grad_norm": 1.4312274290636884,
"learning_rate": 7.471948416177452e-06,
"loss": 0.5473,
"step": 105
},
{
"epoch": 0.19294653014789534,
"grad_norm": 1.467104665014613,
"learning_rate": 7.460158073669271e-06,
"loss": 0.5418,
"step": 106
},
{
"epoch": 0.19476678043230944,
"grad_norm": 1.1804815586636788,
"learning_rate": 7.448247068983638e-06,
"loss": 0.5378,
"step": 107
},
{
"epoch": 0.19658703071672354,
"grad_norm": 1.3602412614708939,
"learning_rate": 7.43621581747954e-06,
"loss": 0.5026,
"step": 108
},
{
"epoch": 0.19840728100113766,
"grad_norm": 1.2691524680339796,
"learning_rate": 7.4240647387092e-06,
"loss": 0.591,
"step": 109
},
{
"epoch": 0.20022753128555176,
"grad_norm": 1.2783869708675566,
"learning_rate": 7.411794256403439e-06,
"loss": 0.5085,
"step": 110
},
{
"epoch": 0.20204778156996586,
"grad_norm": 1.2804361153159327,
"learning_rate": 7.399404798456901e-06,
"loss": 0.6244,
"step": 111
},
{
"epoch": 0.20386803185438,
"grad_norm": 1.2444898246213776,
"learning_rate": 7.3868967969131364e-06,
"loss": 0.5313,
"step": 112
},
{
"epoch": 0.2056882821387941,
"grad_norm": 1.3015010555018793,
"learning_rate": 7.374270687949531e-06,
"loss": 0.5512,
"step": 113
},
{
"epoch": 0.2075085324232082,
"grad_norm": 1.2989535634921763,
"learning_rate": 7.3615269118620945e-06,
"loss": 0.5612,
"step": 114
},
{
"epoch": 0.20932878270762229,
"grad_norm": 1.2893299441070913,
"learning_rate": 7.348665913050114e-06,
"loss": 0.4779,
"step": 115
},
{
"epoch": 0.2111490329920364,
"grad_norm": 1.4374586177315487,
"learning_rate": 7.3356881400006485e-06,
"loss": 0.6057,
"step": 116
},
{
"epoch": 0.2129692832764505,
"grad_norm": 1.2587242021503462,
"learning_rate": 7.3225940452728915e-06,
"loss": 0.5679,
"step": 117
},
{
"epoch": 0.2147895335608646,
"grad_norm": 1.4250046519243573,
"learning_rate": 7.309384085482396e-06,
"loss": 0.5,
"step": 118
},
{
"epoch": 0.21660978384527874,
"grad_norm": 1.222873795275555,
"learning_rate": 7.29605872128514e-06,
"loss": 0.6714,
"step": 119
},
{
"epoch": 0.21843003412969283,
"grad_norm": 1.4243471522403268,
"learning_rate": 7.282618417361476e-06,
"loss": 0.6238,
"step": 120
},
{
"epoch": 0.22025028441410693,
"grad_norm": 1.3065783409043068,
"learning_rate": 7.269063642399912e-06,
"loss": 0.5464,
"step": 121
},
{
"epoch": 0.22207053469852106,
"grad_norm": 1.4517652323683794,
"learning_rate": 7.25539486908078e-06,
"loss": 0.4985,
"step": 122
},
{
"epoch": 0.22389078498293516,
"grad_norm": 1.1483510986901082,
"learning_rate": 7.241612574059745e-06,
"loss": 0.4978,
"step": 123
},
{
"epoch": 0.22571103526734926,
"grad_norm": 1.4389217565677268,
"learning_rate": 7.227717237951189e-06,
"loss": 0.6112,
"step": 124
},
{
"epoch": 0.22753128555176336,
"grad_norm": 1.3389550682905482,
"learning_rate": 7.213709345311444e-06,
"loss": 0.6476,
"step": 125
},
{
"epoch": 0.22935153583617748,
"grad_norm": 1.4169245520418259,
"learning_rate": 7.1995893846219035e-06,
"loss": 0.5354,
"step": 126
},
{
"epoch": 0.23117178612059158,
"grad_norm": 1.2613504469980097,
"learning_rate": 7.185357848271977e-06,
"loss": 0.5467,
"step": 127
},
{
"epoch": 0.23299203640500568,
"grad_norm": 1.1666125829814091,
"learning_rate": 7.17101523254193e-06,
"loss": 0.4698,
"step": 128
},
{
"epoch": 0.2348122866894198,
"grad_norm": 1.352110005355786,
"learning_rate": 7.156562037585575e-06,
"loss": 0.6109,
"step": 129
},
{
"epoch": 0.2366325369738339,
"grad_norm": 1.2180780408157383,
"learning_rate": 7.1419987674128225e-06,
"loss": 0.5332,
"step": 130
},
{
"epoch": 0.238452787258248,
"grad_norm": 1.3933377570677665,
"learning_rate": 7.127325929872119e-06,
"loss": 0.6671,
"step": 131
},
{
"epoch": 0.2402730375426621,
"grad_norm": 1.5246008222442193,
"learning_rate": 7.1125440366327245e-06,
"loss": 0.6212,
"step": 132
},
{
"epoch": 0.24209328782707623,
"grad_norm": 1.2745899656103845,
"learning_rate": 7.0976536031668775e-06,
"loss": 0.6395,
"step": 133
},
{
"epoch": 0.24391353811149033,
"grad_norm": 1.226190343569202,
"learning_rate": 7.082655148731815e-06,
"loss": 0.5761,
"step": 134
},
{
"epoch": 0.24573378839590443,
"grad_norm": 1.1905155038972701,
"learning_rate": 7.067549196351669e-06,
"loss": 0.5418,
"step": 135
},
{
"epoch": 0.24755403868031856,
"grad_norm": 1.2116012422540454,
"learning_rate": 7.052336272799226e-06,
"loss": 0.5273,
"step": 136
},
{
"epoch": 0.24937428896473265,
"grad_norm": 1.282002116739219,
"learning_rate": 7.037016908577555e-06,
"loss": 0.4506,
"step": 137
},
{
"epoch": 0.25119453924914675,
"grad_norm": 1.2499521919340497,
"learning_rate": 7.02159163790151e-06,
"loss": 0.5606,
"step": 138
},
{
"epoch": 0.2530147895335609,
"grad_norm": 1.7373606199315674,
"learning_rate": 7.006060998679105e-06,
"loss": 0.559,
"step": 139
},
{
"epoch": 0.25483503981797495,
"grad_norm": 1.4650449441633262,
"learning_rate": 6.990425532492747e-06,
"loss": 0.5135,
"step": 140
},
{
"epoch": 0.2566552901023891,
"grad_norm": 1.154885986165648,
"learning_rate": 6.974685784580359e-06,
"loss": 0.5039,
"step": 141
},
{
"epoch": 0.2584755403868032,
"grad_norm": 1.2096650723076037,
"learning_rate": 6.958842303816359e-06,
"loss": 0.5079,
"step": 142
},
{
"epoch": 0.2602957906712173,
"grad_norm": 1.1620810226211598,
"learning_rate": 6.942895642692527e-06,
"loss": 0.5245,
"step": 143
},
{
"epoch": 0.2621160409556314,
"grad_norm": 1.3200231462656904,
"learning_rate": 6.926846357298732e-06,
"loss": 0.5935,
"step": 144
},
{
"epoch": 0.26393629124004553,
"grad_norm": 1.2355460824632627,
"learning_rate": 6.910695007303544e-06,
"loss": 0.5543,
"step": 145
},
{
"epoch": 0.2657565415244596,
"grad_norm": 1.2033075815432748,
"learning_rate": 6.894442155934719e-06,
"loss": 0.4831,
"step": 146
},
{
"epoch": 0.2675767918088737,
"grad_norm": 1.1907917135137838,
"learning_rate": 6.878088369959553e-06,
"loss": 0.5221,
"step": 147
},
{
"epoch": 0.26939704209328785,
"grad_norm": 1.3255509879738674,
"learning_rate": 6.861634219665117e-06,
"loss": 0.6086,
"step": 148
},
{
"epoch": 0.2712172923777019,
"grad_norm": 1.2271219658785495,
"learning_rate": 6.845080278838381e-06,
"loss": 0.5825,
"step": 149
},
{
"epoch": 0.27303754266211605,
"grad_norm": 1.1690589610793065,
"learning_rate": 6.82842712474619e-06,
"loss": 0.5807,
"step": 150
},
{
"epoch": 0.2748577929465301,
"grad_norm": 1.2888207762623227,
"learning_rate": 6.811675338115146e-06,
"loss": 0.6188,
"step": 151
},
{
"epoch": 0.27667804323094425,
"grad_norm": 1.2514369430789496,
"learning_rate": 6.7948255031113505e-06,
"loss": 0.5913,
"step": 152
},
{
"epoch": 0.2784982935153584,
"grad_norm": 1.3076941421175066,
"learning_rate": 6.777878207320034e-06,
"loss": 0.5054,
"step": 153
},
{
"epoch": 0.28031854379977245,
"grad_norm": 1.2943985087075844,
"learning_rate": 6.760834041725068e-06,
"loss": 0.4915,
"step": 154
},
{
"epoch": 0.2821387940841866,
"grad_norm": 1.2887581327542428,
"learning_rate": 6.743693600688352e-06,
"loss": 0.5538,
"step": 155
},
{
"epoch": 0.2839590443686007,
"grad_norm": 1.157113796843012,
"learning_rate": 6.726457481929095e-06,
"loss": 0.537,
"step": 156
},
{
"epoch": 0.28577929465301477,
"grad_norm": 1.1766314672266696,
"learning_rate": 6.7091262865029645e-06,
"loss": 0.5896,
"step": 157
},
{
"epoch": 0.2875995449374289,
"grad_norm": 1.2292202798354899,
"learning_rate": 6.691700618781126e-06,
"loss": 0.6347,
"step": 158
},
{
"epoch": 0.289419795221843,
"grad_norm": 1.1513406588801496,
"learning_rate": 6.674181086429177e-06,
"loss": 0.4663,
"step": 159
},
{
"epoch": 0.2912400455062571,
"grad_norm": 1.2946840584006447,
"learning_rate": 6.656568300385944e-06,
"loss": 0.6247,
"step": 160
},
{
"epoch": 0.2930602957906712,
"grad_norm": 1.2952632442735728,
"learning_rate": 6.6388628748421895e-06,
"loss": 0.4728,
"step": 161
},
{
"epoch": 0.29488054607508535,
"grad_norm": 1.2684543190366842,
"learning_rate": 6.62106542721918e-06,
"loss": 0.5,
"step": 162
},
{
"epoch": 0.2967007963594994,
"grad_norm": 1.2434590334770437,
"learning_rate": 6.603176578147174e-06,
"loss": 0.552,
"step": 163
},
{
"epoch": 0.29852104664391355,
"grad_norm": 1.2853162665121605,
"learning_rate": 6.585196951443763e-06,
"loss": 0.5311,
"step": 164
},
{
"epoch": 0.3003412969283277,
"grad_norm": 1.3910181828529422,
"learning_rate": 6.5671271740921266e-06,
"loss": 0.5595,
"step": 165
},
{
"epoch": 0.30216154721274174,
"grad_norm": 1.318698152941268,
"learning_rate": 6.548967876219163e-06,
"loss": 0.5323,
"step": 166
},
{
"epoch": 0.30398179749715587,
"grad_norm": 1.259567167046916,
"learning_rate": 6.530719691073521e-06,
"loss": 0.5773,
"step": 167
},
{
"epoch": 0.30580204778156994,
"grad_norm": 1.3201679730014977,
"learning_rate": 6.5123832550035165e-06,
"loss": 0.5143,
"step": 168
},
{
"epoch": 0.30762229806598407,
"grad_norm": 1.3232034824966301,
"learning_rate": 6.493959207434934e-06,
"loss": 0.553,
"step": 169
},
{
"epoch": 0.3094425483503982,
"grad_norm": 1.3960220649200046,
"learning_rate": 6.47544819084874e-06,
"loss": 0.561,
"step": 170
},
{
"epoch": 0.31126279863481227,
"grad_norm": 1.3348060079340793,
"learning_rate": 6.4568508507586715e-06,
"loss": 0.5047,
"step": 171
},
{
"epoch": 0.3130830489192264,
"grad_norm": 1.2868651237482562,
"learning_rate": 6.438167835688725e-06,
"loss": 0.5094,
"step": 172
},
{
"epoch": 0.3149032992036405,
"grad_norm": 1.2603952904899627,
"learning_rate": 6.41939979715055e-06,
"loss": 0.5323,
"step": 173
},
{
"epoch": 0.3167235494880546,
"grad_norm": 1.2921556438401538,
"learning_rate": 6.400547389620716e-06,
"loss": 0.5554,
"step": 174
},
{
"epoch": 0.3185437997724687,
"grad_norm": 1.2590564886848532,
"learning_rate": 6.3816112705178984e-06,
"loss": 0.5288,
"step": 175
},
{
"epoch": 0.32036405005688284,
"grad_norm": 1.3084436554782835,
"learning_rate": 6.362592100179958e-06,
"loss": 0.6402,
"step": 176
},
{
"epoch": 0.3221843003412969,
"grad_norm": 1.4261334165831296,
"learning_rate": 6.343490541840899e-06,
"loss": 0.489,
"step": 177
},
{
"epoch": 0.32400455062571104,
"grad_norm": 1.4986503257367303,
"learning_rate": 6.3243072616077535e-06,
"loss": 0.5957,
"step": 178
},
{
"epoch": 0.32582480091012517,
"grad_norm": 1.2030242787629297,
"learning_rate": 6.3050429284373465e-06,
"loss": 0.4974,
"step": 179
},
{
"epoch": 0.32764505119453924,
"grad_norm": 1.3717989144113625,
"learning_rate": 6.285698214112974e-06,
"loss": 0.593,
"step": 180
},
{
"epoch": 0.32946530147895337,
"grad_norm": 1.2021644407962897,
"learning_rate": 6.2662737932209695e-06,
"loss": 0.616,
"step": 181
},
{
"epoch": 0.33128555176336744,
"grad_norm": 1.281410126238882,
"learning_rate": 6.246770343127185e-06,
"loss": 0.5598,
"step": 182
},
{
"epoch": 0.33310580204778156,
"grad_norm": 1.39438795848328,
"learning_rate": 6.227188543953368e-06,
"loss": 0.5932,
"step": 183
},
{
"epoch": 0.33310580204778156,
"eval_accuracy": 0.8092656088844726,
"eval_accuracy_first_token": 0.7541679610645128,
"eval_accuracy_first_token_<": 0.9296587926509187,
"eval_accuracy_first_token_<_total": 1905,
"eval_accuracy_first_token_<|python_tag|>": 0.8752515090543259,
"eval_accuracy_first_token_<|python_tag|>_total": 994,
"eval_accuracy_first_token_Certainly": 0.7024793388429752,
"eval_accuracy_first_token_Certainly_total": 363,
"eval_accuracy_first_token_The": 0.9059161873459326,
"eval_accuracy_first_token_The_total": 2434,
"eval_accuracy_first_token_To": 0.8237179487179487,
"eval_accuracy_first_token_To_total": 936,
"eval_loss": 0.5801064372062683,
"eval_perplexity": 1.1141803737974993,
"eval_runtime": 508.2626,
"eval_samples_per_second": 1.371,
"eval_steps_per_second": 0.687,
"eval_total_number_first_token": 9657,
"step": 183
},
{
"epoch": 0.3349260523321957,
"grad_norm": 1.3081524231573554,
"learning_rate": 6.207529078553444e-06,
"loss": 0.5457,
"step": 184
},
{
"epoch": 0.33674630261660976,
"grad_norm": 1.3716078335539046,
"learning_rate": 6.1877926324897085e-06,
"loss": 0.5473,
"step": 185
},
{
"epoch": 0.3385665529010239,
"grad_norm": 1.4145939624062198,
"learning_rate": 6.16797989400891e-06,
"loss": 0.5786,
"step": 186
},
{
"epoch": 0.340386803185438,
"grad_norm": 1.2294721126569037,
"learning_rate": 6.148091554018264e-06,
"loss": 0.5902,
"step": 187
},
{
"epoch": 0.3422070534698521,
"grad_norm": 1.2925378088030424,
"learning_rate": 6.128128306061346e-06,
"loss": 0.5142,
"step": 188
},
{
"epoch": 0.3440273037542662,
"grad_norm": 1.2279588518524418,
"learning_rate": 6.108090846293915e-06,
"loss": 0.5135,
"step": 189
},
{
"epoch": 0.34584755403868034,
"grad_norm": 1.268146835786646,
"learning_rate": 6.087979873459634e-06,
"loss": 0.5447,
"step": 190
},
{
"epoch": 0.3476678043230944,
"grad_norm": 1.318995573559777,
"learning_rate": 6.0677960888657015e-06,
"loss": 0.6744,
"step": 191
},
{
"epoch": 0.34948805460750854,
"grad_norm": 1.231758642404661,
"learning_rate": 6.047540196358404e-06,
"loss": 0.5809,
"step": 192
},
{
"epoch": 0.35130830489192266,
"grad_norm": 1.2372891673165372,
"learning_rate": 6.02721290229856e-06,
"loss": 0.5807,
"step": 193
},
{
"epoch": 0.35312855517633673,
"grad_norm": 1.481210652387573,
"learning_rate": 6.006814915536894e-06,
"loss": 0.5936,
"step": 194
},
{
"epoch": 0.35494880546075086,
"grad_norm": 1.2305803524181071,
"learning_rate": 5.9863469473893225e-06,
"loss": 0.5438,
"step": 195
},
{
"epoch": 0.356769055745165,
"grad_norm": 1.2206928256434937,
"learning_rate": 5.965809711612137e-06,
"loss": 0.5005,
"step": 196
},
{
"epoch": 0.35858930602957906,
"grad_norm": 1.086573502847394,
"learning_rate": 5.945203924377125e-06,
"loss": 0.4889,
"step": 197
},
{
"epoch": 0.3604095563139932,
"grad_norm": 1.3513470624112347,
"learning_rate": 5.92453030424659e-06,
"loss": 0.5599,
"step": 198
},
{
"epoch": 0.36222980659840726,
"grad_norm": 1.1113956980921844,
"learning_rate": 5.903789572148294e-06,
"loss": 0.5182,
"step": 199
},
{
"epoch": 0.3640500568828214,
"grad_norm": 1.4891052629080104,
"learning_rate": 5.88298245135032e-06,
"loss": 0.5716,
"step": 200
},
{
"epoch": 0.3658703071672355,
"grad_norm": 1.3005774833983796,
"learning_rate": 5.862109667435853e-06,
"loss": 0.5665,
"step": 201
},
{
"epoch": 0.3676905574516496,
"grad_norm": 1.2151067893045482,
"learning_rate": 5.8411719482778645e-06,
"loss": 0.4965,
"step": 202
},
{
"epoch": 0.3695108077360637,
"grad_norm": 1.5031392413729012,
"learning_rate": 5.820170024013746e-06,
"loss": 0.5398,
"step": 203
},
{
"epoch": 0.37133105802047783,
"grad_norm": 1.1627104663425107,
"learning_rate": 5.79910462701984e-06,
"loss": 0.4461,
"step": 204
},
{
"epoch": 0.3731513083048919,
"grad_norm": 1.3656640622390992,
"learning_rate": 5.777976491885903e-06,
"loss": 0.6048,
"step": 205
},
{
"epoch": 0.37497155858930603,
"grad_norm": 1.2327820864728312,
"learning_rate": 5.756786355389481e-06,
"loss": 0.5052,
"step": 206
},
{
"epoch": 0.37679180887372016,
"grad_norm": 1.3098660955959893,
"learning_rate": 5.735534956470232e-06,
"loss": 0.5507,
"step": 207
},
{
"epoch": 0.37861205915813423,
"grad_norm": 1.2828934352712993,
"learning_rate": 5.714223036204144e-06,
"loss": 0.5973,
"step": 208
},
{
"epoch": 0.38043230944254836,
"grad_norm": 1.1860097743128348,
"learning_rate": 5.6928513377777e-06,
"loss": 0.4965,
"step": 209
},
{
"epoch": 0.3822525597269625,
"grad_norm": 1.2517621268060033,
"learning_rate": 5.671420606461956e-06,
"loss": 0.5487,
"step": 210
},
{
"epoch": 0.38407281001137655,
"grad_norm": 1.271873202647325,
"learning_rate": 5.649931589586557e-06,
"loss": 0.5979,
"step": 211
},
{
"epoch": 0.3858930602957907,
"grad_norm": 1.2859574150365818,
"learning_rate": 5.628385036513676e-06,
"loss": 0.4776,
"step": 212
},
{
"epoch": 0.38771331058020475,
"grad_norm": 1.354537404525919,
"learning_rate": 5.606781698611878e-06,
"loss": 0.4877,
"step": 213
},
{
"epoch": 0.3895335608646189,
"grad_norm": 1.3284236870109494,
"learning_rate": 5.585122329229923e-06,
"loss": 0.5859,
"step": 214
},
{
"epoch": 0.391353811149033,
"grad_norm": 1.2340130409038237,
"learning_rate": 5.56340768367049e-06,
"loss": 0.5305,
"step": 215
},
{
"epoch": 0.3931740614334471,
"grad_norm": 1.453487991090255,
"learning_rate": 5.541638519163849e-06,
"loss": 0.55,
"step": 216
},
{
"epoch": 0.3949943117178612,
"grad_norm": 1.298062739902415,
"learning_rate": 5.51981559484144e-06,
"loss": 0.5169,
"step": 217
},
{
"epoch": 0.39681456200227533,
"grad_norm": 1.311968473611326,
"learning_rate": 5.49793967170941e-06,
"loss": 0.5751,
"step": 218
},
{
"epoch": 0.3986348122866894,
"grad_norm": 1.2412899795871963,
"learning_rate": 5.476011512622076e-06,
"loss": 0.6166,
"step": 219
},
{
"epoch": 0.4004550625711035,
"grad_norm": 1.242116646999028,
"learning_rate": 5.454031882255319e-06,
"loss": 0.5578,
"step": 220
},
{
"epoch": 0.40227531285551765,
"grad_norm": 1.4168825327120473,
"learning_rate": 5.43200154707992e-06,
"loss": 0.5662,
"step": 221
},
{
"epoch": 0.4040955631399317,
"grad_norm": 1.426587220049501,
"learning_rate": 5.4099212753348294e-06,
"loss": 0.5169,
"step": 222
},
{
"epoch": 0.40591581342434585,
"grad_norm": 1.3014887756398712,
"learning_rate": 5.3877918370003806e-06,
"loss": 0.5117,
"step": 223
},
{
"epoch": 0.40773606370876,
"grad_norm": 1.3406748606110184,
"learning_rate": 5.365614003771439e-06,
"loss": 0.5549,
"step": 224
},
{
"epoch": 0.40955631399317405,
"grad_norm": 1.2661574936204552,
"learning_rate": 5.343388549030491e-06,
"loss": 0.5163,
"step": 225
},
{
"epoch": 0.4113765642775882,
"grad_norm": 1.085801335463159,
"learning_rate": 5.321116247820669e-06,
"loss": 0.5244,
"step": 226
},
{
"epoch": 0.4131968145620023,
"grad_norm": 1.1238570077454868,
"learning_rate": 5.298797876818734e-06,
"loss": 0.4877,
"step": 227
},
{
"epoch": 0.4150170648464164,
"grad_norm": 1.2501588033198834,
"learning_rate": 5.276434214307986e-06,
"loss": 0.5175,
"step": 228
},
{
"epoch": 0.4168373151308305,
"grad_norm": 1.1104132675236253,
"learning_rate": 5.2540260401511255e-06,
"loss": 0.4912,
"step": 229
},
{
"epoch": 0.41865756541524457,
"grad_norm": 1.3469476360006967,
"learning_rate": 5.231574135763052e-06,
"loss": 0.5119,
"step": 230
},
{
"epoch": 0.4204778156996587,
"grad_norm": 1.3917881004299013,
"learning_rate": 5.209079284083626e-06,
"loss": 0.5893,
"step": 231
},
{
"epoch": 0.4222980659840728,
"grad_norm": 1.3049155919134754,
"learning_rate": 5.186542269550359e-06,
"loss": 0.5863,
"step": 232
},
{
"epoch": 0.4241183162684869,
"grad_norm": 1.4688437185729748,
"learning_rate": 5.163963878071058e-06,
"loss": 0.6134,
"step": 233
},
{
"epoch": 0.425938566552901,
"grad_norm": 1.334594097803803,
"learning_rate": 5.141344896996421e-06,
"loss": 0.4778,
"step": 234
},
{
"epoch": 0.42775881683731515,
"grad_norm": 1.349733087487026,
"learning_rate": 5.1186861150925844e-06,
"loss": 0.5989,
"step": 235
},
{
"epoch": 0.4295790671217292,
"grad_norm": 1.1945891754612503,
"learning_rate": 5.09598832251361e-06,
"loss": 0.4466,
"step": 236
},
{
"epoch": 0.43139931740614335,
"grad_norm": 1.388728814777883,
"learning_rate": 5.073252310773939e-06,
"loss": 0.6193,
"step": 237
},
{
"epoch": 0.43321956769055747,
"grad_norm": 1.4284168159961905,
"learning_rate": 5.050478872720782e-06,
"loss": 0.5535,
"step": 238
},
{
"epoch": 0.43503981797497154,
"grad_norm": 1.3786811940208537,
"learning_rate": 5.027668802506476e-06,
"loss": 0.4974,
"step": 239
},
{
"epoch": 0.43686006825938567,
"grad_norm": 1.2927644778322318,
"learning_rate": 5.004822895560794e-06,
"loss": 0.5029,
"step": 240
},
{
"epoch": 0.4386803185437998,
"grad_norm": 1.252387489439096,
"learning_rate": 4.981941948563196e-06,
"loss": 0.5278,
"step": 241
},
{
"epoch": 0.44050056882821387,
"grad_norm": 1.339410297479304,
"learning_rate": 4.959026759415061e-06,
"loss": 0.4939,
"step": 242
},
{
"epoch": 0.442320819112628,
"grad_norm": 1.364314280772671,
"learning_rate": 4.936078127211849e-06,
"loss": 0.5951,
"step": 243
},
{
"epoch": 0.4441410693970421,
"grad_norm": 1.3075257340817037,
"learning_rate": 4.913096852215248e-06,
"loss": 0.6049,
"step": 244
},
{
"epoch": 0.4459613196814562,
"grad_norm": 1.4449459168578944,
"learning_rate": 4.890083735825257e-06,
"loss": 0.6495,
"step": 245
},
{
"epoch": 0.4477815699658703,
"grad_norm": 1.250520396487667,
"learning_rate": 4.867039580552247e-06,
"loss": 0.5094,
"step": 246
},
{
"epoch": 0.4496018202502844,
"grad_norm": 1.3729253299629682,
"learning_rate": 4.843965189988969e-06,
"loss": 0.601,
"step": 247
},
{
"epoch": 0.4514220705346985,
"grad_norm": 1.3588267674574899,
"learning_rate": 4.820861368782537e-06,
"loss": 0.6282,
"step": 248
},
{
"epoch": 0.45324232081911264,
"grad_norm": 1.2605268123722788,
"learning_rate": 4.79772892260637e-06,
"loss": 0.5305,
"step": 249
},
{
"epoch": 0.4550625711035267,
"grad_norm": 1.2681522997617236,
"learning_rate": 4.774568658132086e-06,
"loss": 0.5748,
"step": 250
},
{
"epoch": 0.45688282138794084,
"grad_norm": 1.3533665486552986,
"learning_rate": 4.751381383001386e-06,
"loss": 0.4689,
"step": 251
},
{
"epoch": 0.45870307167235497,
"grad_norm": 1.267434752321521,
"learning_rate": 4.728167905797877e-06,
"loss": 0.534,
"step": 252
},
{
"epoch": 0.46052332195676904,
"grad_norm": 1.2225430366963792,
"learning_rate": 4.7049290360188875e-06,
"loss": 0.5003,
"step": 253
},
{
"epoch": 0.46234357224118316,
"grad_norm": 1.3388181688742944,
"learning_rate": 4.681665584047227e-06,
"loss": 0.5219,
"step": 254
},
{
"epoch": 0.4641638225255973,
"grad_norm": 1.3028544028541067,
"learning_rate": 4.658378361122936e-06,
"loss": 0.5452,
"step": 255
},
{
"epoch": 0.46598407281001136,
"grad_norm": 1.157965088527389,
"learning_rate": 4.6350681793149884e-06,
"loss": 0.5229,
"step": 256
},
{
"epoch": 0.4678043230944255,
"grad_norm": 1.3045701677810966,
"learning_rate": 4.611735851492984e-06,
"loss": 0.5728,
"step": 257
},
{
"epoch": 0.4696245733788396,
"grad_norm": 1.2850295730378811,
"learning_rate": 4.588382191298787e-06,
"loss": 0.5537,
"step": 258
},
{
"epoch": 0.4714448236632537,
"grad_norm": 1.1541773263319788,
"learning_rate": 4.5650080131181675e-06,
"loss": 0.538,
"step": 259
},
{
"epoch": 0.4732650739476678,
"grad_norm": 1.2910982881953428,
"learning_rate": 4.541614132052393e-06,
"loss": 0.5612,
"step": 260
},
{
"epoch": 0.4750853242320819,
"grad_norm": 1.3654547247389965,
"learning_rate": 4.51820136388981e-06,
"loss": 0.4475,
"step": 261
},
{
"epoch": 0.476905574516496,
"grad_norm": 1.2698111645568326,
"learning_rate": 4.494770525077391e-06,
"loss": 0.5621,
"step": 262
},
{
"epoch": 0.47872582480091014,
"grad_norm": 1.3935553561279155,
"learning_rate": 4.4713224326922655e-06,
"loss": 0.599,
"step": 263
},
{
"epoch": 0.4805460750853242,
"grad_norm": 1.3156202456183232,
"learning_rate": 4.447857904413231e-06,
"loss": 0.532,
"step": 264
},
{
"epoch": 0.48236632536973834,
"grad_norm": 1.1337315757674442,
"learning_rate": 4.424377758492232e-06,
"loss": 0.5353,
"step": 265
},
{
"epoch": 0.48418657565415246,
"grad_norm": 1.4101220722381445,
"learning_rate": 4.40088281372583e-06,
"loss": 0.5006,
"step": 266
},
{
"epoch": 0.48600682593856653,
"grad_norm": 1.2300783759396936,
"learning_rate": 4.377373889426649e-06,
"loss": 0.5438,
"step": 267
},
{
"epoch": 0.48782707622298066,
"grad_norm": 1.17585511931677,
"learning_rate": 4.353851805394808e-06,
"loss": 0.5369,
"step": 268
},
{
"epoch": 0.4896473265073948,
"grad_norm": 1.2115284188121456,
"learning_rate": 4.33031738188933e-06,
"loss": 0.4524,
"step": 269
},
{
"epoch": 0.49146757679180886,
"grad_norm": 1.4317839486717954,
"learning_rate": 4.306771439599534e-06,
"loss": 0.6436,
"step": 270
},
{
"epoch": 0.493287827076223,
"grad_norm": 1.4621543028546877,
"learning_rate": 4.283214799616428e-06,
"loss": 0.4368,
"step": 271
},
{
"epoch": 0.4951080773606371,
"grad_norm": 1.2936498719418335,
"learning_rate": 4.259648283404062e-06,
"loss": 0.5541,
"step": 272
},
{
"epoch": 0.4969283276450512,
"grad_norm": 1.120893056901481,
"learning_rate": 4.236072712770891e-06,
"loss": 0.5822,
"step": 273
},
{
"epoch": 0.4987485779294653,
"grad_norm": 1.429920580926913,
"learning_rate": 4.2124889098411175e-06,
"loss": 0.5302,
"step": 274
},
{
"epoch": 0.5005688282138794,
"grad_norm": 1.312546385183068,
"learning_rate": 4.1888976970260135e-06,
"loss": 0.4835,
"step": 275
},
{
"epoch": 0.5023890784982935,
"grad_norm": 1.3107231912852029,
"learning_rate": 4.165299896995252e-06,
"loss": 0.5421,
"step": 276
},
{
"epoch": 0.5042093287827076,
"grad_norm": 1.232720241327702,
"learning_rate": 4.141696332648216e-06,
"loss": 0.5012,
"step": 277
},
{
"epoch": 0.5060295790671218,
"grad_norm": 1.2132932647482422,
"learning_rate": 4.118087827085294e-06,
"loss": 0.5463,
"step": 278
},
{
"epoch": 0.5078498293515359,
"grad_norm": 1.270716742837696,
"learning_rate": 4.094475203579191e-06,
"loss": 0.5383,
"step": 279
},
{
"epoch": 0.5096700796359499,
"grad_norm": 1.2438093689244545,
"learning_rate": 4.070859285546209e-06,
"loss": 0.5556,
"step": 280
},
{
"epoch": 0.511490329920364,
"grad_norm": 1.3516997697288733,
"learning_rate": 4.047240896517539e-06,
"loss": 0.6018,
"step": 281
},
{
"epoch": 0.5133105802047782,
"grad_norm": 1.2439174788361766,
"learning_rate": 4.023620860110533e-06,
"loss": 0.4133,
"step": 282
},
{
"epoch": 0.5151308304891923,
"grad_norm": 1.3778918541026397,
"learning_rate": 4e-06,
"loss": 0.6007,
"step": 283
},
{
"epoch": 0.5169510807736064,
"grad_norm": 1.2268835899477202,
"learning_rate": 3.976379139889467e-06,
"loss": 0.5331,
"step": 284
},
{
"epoch": 0.5187713310580204,
"grad_norm": 1.3429317349537357,
"learning_rate": 3.9527591034824616e-06,
"loss": 0.5311,
"step": 285
},
{
"epoch": 0.5205915813424346,
"grad_norm": 1.2134030361634403,
"learning_rate": 3.929140714453791e-06,
"loss": 0.481,
"step": 286
},
{
"epoch": 0.5224118316268487,
"grad_norm": 1.2462618006711519,
"learning_rate": 3.9055247964208075e-06,
"loss": 0.5273,
"step": 287
},
{
"epoch": 0.5242320819112628,
"grad_norm": 1.5559356945744065,
"learning_rate": 3.8819121729147055e-06,
"loss": 0.6021,
"step": 288
},
{
"epoch": 0.5260523321956769,
"grad_norm": 1.3832902722589653,
"learning_rate": 3.8583036673517845e-06,
"loss": 0.4454,
"step": 289
},
{
"epoch": 0.5278725824800911,
"grad_norm": 1.3978629990846738,
"learning_rate": 3.834700103004747e-06,
"loss": 0.5124,
"step": 290
},
{
"epoch": 0.5296928327645051,
"grad_norm": 1.2666994035299775,
"learning_rate": 3.8111023029739866e-06,
"loss": 0.4667,
"step": 291
},
{
"epoch": 0.5315130830489192,
"grad_norm": 1.3305625183535323,
"learning_rate": 3.787511090158884e-06,
"loss": 0.5368,
"step": 292
},
{
"epoch": 0.5333333333333333,
"grad_norm": 1.2554353067602742,
"learning_rate": 3.763927287229109e-06,
"loss": 0.5499,
"step": 293
},
{
"epoch": 0.5351535836177475,
"grad_norm": 1.307288214215709,
"learning_rate": 3.740351716595939e-06,
"loss": 0.5055,
"step": 294
},
{
"epoch": 0.5369738339021616,
"grad_norm": 1.3279514539943822,
"learning_rate": 3.7167852003835723e-06,
"loss": 0.511,
"step": 295
},
{
"epoch": 0.5387940841865757,
"grad_norm": 1.223828271060312,
"learning_rate": 3.6932285604004656e-06,
"loss": 0.4595,
"step": 296
},
{
"epoch": 0.5406143344709897,
"grad_norm": 1.332815030257366,
"learning_rate": 3.669682618110671e-06,
"loss": 0.6227,
"step": 297
},
{
"epoch": 0.5424345847554038,
"grad_norm": 1.2271598713957907,
"learning_rate": 3.646148194605191e-06,
"loss": 0.5925,
"step": 298
},
{
"epoch": 0.544254835039818,
"grad_norm": 1.2856112828358344,
"learning_rate": 3.622626110573351e-06,
"loss": 0.4888,
"step": 299
},
{
"epoch": 0.5460750853242321,
"grad_norm": 1.2933732045646906,
"learning_rate": 3.5991171862741713e-06,
"loss": 0.5072,
"step": 300
},
{
"epoch": 0.5478953356086462,
"grad_norm": 1.928545669129003,
"learning_rate": 3.575622241507768e-06,
"loss": 0.525,
"step": 301
},
{
"epoch": 0.5497155858930602,
"grad_norm": 1.1113802518291283,
"learning_rate": 3.5521420955867683e-06,
"loss": 0.5977,
"step": 302
},
{
"epoch": 0.5515358361774744,
"grad_norm": 1.3617289948905469,
"learning_rate": 3.5286775673077332e-06,
"loss": 0.5839,
"step": 303
},
{
"epoch": 0.5533560864618885,
"grad_norm": 1.391130826033813,
"learning_rate": 3.505229474922609e-06,
"loss": 0.5181,
"step": 304
},
{
"epoch": 0.5551763367463026,
"grad_norm": 1.2316652646361441,
"learning_rate": 3.481798636110191e-06,
"loss": 0.4945,
"step": 305
},
{
"epoch": 0.5569965870307167,
"grad_norm": 1.2873086430494702,
"learning_rate": 3.458385867947607e-06,
"loss": 0.4924,
"step": 306
},
{
"epoch": 0.5588168373151309,
"grad_norm": 1.430369532857129,
"learning_rate": 3.434991986881833e-06,
"loss": 0.4821,
"step": 307
},
{
"epoch": 0.5606370875995449,
"grad_norm": 1.1646260708946579,
"learning_rate": 3.4116178087012136e-06,
"loss": 0.5052,
"step": 308
},
{
"epoch": 0.562457337883959,
"grad_norm": 1.2327288356772756,
"learning_rate": 3.388264148507016e-06,
"loss": 0.5057,
"step": 309
},
{
"epoch": 0.5642775881683731,
"grad_norm": 1.3536008520463172,
"learning_rate": 3.3649318206850116e-06,
"loss": 0.5178,
"step": 310
},
{
"epoch": 0.5660978384527873,
"grad_norm": 1.1896041830424324,
"learning_rate": 3.3416216388770635e-06,
"loss": 0.5417,
"step": 311
},
{
"epoch": 0.5679180887372014,
"grad_norm": 1.2975919761075365,
"learning_rate": 3.3183344159527736e-06,
"loss": 0.5234,
"step": 312
},
{
"epoch": 0.5697383390216155,
"grad_norm": 1.1688323545338841,
"learning_rate": 3.2950709639811134e-06,
"loss": 0.4888,
"step": 313
},
{
"epoch": 0.5715585893060295,
"grad_norm": 1.3939258642019638,
"learning_rate": 3.271832094202123e-06,
"loss": 0.5183,
"step": 314
},
{
"epoch": 0.5733788395904437,
"grad_norm": 1.1897624823605304,
"learning_rate": 3.2486186169986153e-06,
"loss": 0.5454,
"step": 315
},
{
"epoch": 0.5751990898748578,
"grad_norm": 1.1944714828344472,
"learning_rate": 3.2254313418679154e-06,
"loss": 0.4807,
"step": 316
},
{
"epoch": 0.5770193401592719,
"grad_norm": 1.2256094296723554,
"learning_rate": 3.2022710773936304e-06,
"loss": 0.5223,
"step": 317
},
{
"epoch": 0.578839590443686,
"grad_norm": 1.2296819521179183,
"learning_rate": 3.1791386312174633e-06,
"loss": 0.4951,
"step": 318
},
{
"epoch": 0.5806598407281001,
"grad_norm": 1.4525163093513003,
"learning_rate": 3.1560348100110315e-06,
"loss": 0.4874,
"step": 319
},
{
"epoch": 0.5824800910125142,
"grad_norm": 1.4766150215295402,
"learning_rate": 3.1329604194477535e-06,
"loss": 0.5186,
"step": 320
},
{
"epoch": 0.5843003412969283,
"grad_norm": 1.5620410502792905,
"learning_rate": 3.1099162641747427e-06,
"loss": 0.5542,
"step": 321
},
{
"epoch": 0.5861205915813424,
"grad_norm": 1.1792003162185065,
"learning_rate": 3.0869031477847507e-06,
"loss": 0.4751,
"step": 322
},
{
"epoch": 0.5879408418657566,
"grad_norm": 1.4371388899245734,
"learning_rate": 3.0639218727881508e-06,
"loss": 0.5066,
"step": 323
},
{
"epoch": 0.5897610921501707,
"grad_norm": 1.3321828422036859,
"learning_rate": 3.04097324058494e-06,
"loss": 0.418,
"step": 324
},
{
"epoch": 0.5915813424345847,
"grad_norm": 1.2642329608748821,
"learning_rate": 3.0180580514368034e-06,
"loss": 0.6167,
"step": 325
},
{
"epoch": 0.5934015927189988,
"grad_norm": 1.1538255682096556,
"learning_rate": 2.9951771044392066e-06,
"loss": 0.5297,
"step": 326
},
{
"epoch": 0.595221843003413,
"grad_norm": 1.2453988044078719,
"learning_rate": 2.972331197493523e-06,
"loss": 0.4552,
"step": 327
},
{
"epoch": 0.5970420932878271,
"grad_norm": 1.3576907607149231,
"learning_rate": 2.949521127279218e-06,
"loss": 0.5003,
"step": 328
},
{
"epoch": 0.5988623435722412,
"grad_norm": 1.3497348777364608,
"learning_rate": 2.926747689226062e-06,
"loss": 0.5561,
"step": 329
},
{
"epoch": 0.6006825938566553,
"grad_norm": 1.259164100959422,
"learning_rate": 2.9040116774863896e-06,
"loss": 0.4856,
"step": 330
},
{
"epoch": 0.6025028441410694,
"grad_norm": 1.2529485220686642,
"learning_rate": 2.881313884907416e-06,
"loss": 0.5575,
"step": 331
},
{
"epoch": 0.6043230944254835,
"grad_norm": 1.0669208953569564,
"learning_rate": 2.8586551030035797e-06,
"loss": 0.4644,
"step": 332
},
{
"epoch": 0.6061433447098976,
"grad_norm": 1.3366563458096783,
"learning_rate": 2.836036121928942e-06,
"loss": 0.453,
"step": 333
},
{
"epoch": 0.6079635949943117,
"grad_norm": 1.1476061284968695,
"learning_rate": 2.813457730449641e-06,
"loss": 0.4207,
"step": 334
},
{
"epoch": 0.6097838452787259,
"grad_norm": 1.3411031407126155,
"learning_rate": 2.790920715916372e-06,
"loss": 0.5404,
"step": 335
},
{
"epoch": 0.6116040955631399,
"grad_norm": 1.1482077310699785,
"learning_rate": 2.7684258642369484e-06,
"loss": 0.5205,
"step": 336
},
{
"epoch": 0.613424345847554,
"grad_norm": 1.42649538957105,
"learning_rate": 2.7459739598488762e-06,
"loss": 0.5013,
"step": 337
},
{
"epoch": 0.6152445961319681,
"grad_norm": 1.2243282105614175,
"learning_rate": 2.723565785692013e-06,
"loss": 0.5464,
"step": 338
},
{
"epoch": 0.6170648464163823,
"grad_norm": 1.2484628027395077,
"learning_rate": 2.701202123181266e-06,
"loss": 0.5519,
"step": 339
},
{
"epoch": 0.6188850967007964,
"grad_norm": 1.3240902412697022,
"learning_rate": 2.6788837521793328e-06,
"loss": 0.5205,
"step": 340
},
{
"epoch": 0.6207053469852105,
"grad_norm": 1.2873575493742448,
"learning_rate": 2.6566114509695096e-06,
"loss": 0.4761,
"step": 341
},
{
"epoch": 0.6225255972696245,
"grad_norm": 1.2093203753299095,
"learning_rate": 2.634385996228561e-06,
"loss": 0.4753,
"step": 342
},
{
"epoch": 0.6243458475540387,
"grad_norm": 1.300179853101682,
"learning_rate": 2.6122081629996195e-06,
"loss": 0.4934,
"step": 343
},
{
"epoch": 0.6261660978384528,
"grad_norm": 1.2890047749069995,
"learning_rate": 2.5900787246651715e-06,
"loss": 0.4873,
"step": 344
},
{
"epoch": 0.6279863481228669,
"grad_norm": 1.4341436462807016,
"learning_rate": 2.567998452920081e-06,
"loss": 0.5213,
"step": 345
},
{
"epoch": 0.629806598407281,
"grad_norm": 1.299103431190263,
"learning_rate": 2.5459681177446797e-06,
"loss": 0.4783,
"step": 346
},
{
"epoch": 0.631626848691695,
"grad_norm": 1.331377030608932,
"learning_rate": 2.523988487377924e-06,
"loss": 0.5045,
"step": 347
},
{
"epoch": 0.6334470989761092,
"grad_norm": 1.3571947037370755,
"learning_rate": 2.50206032829059e-06,
"loss": 0.5005,
"step": 348
},
{
"epoch": 0.6352673492605233,
"grad_norm": 1.5380704302051296,
"learning_rate": 2.4801844051585604e-06,
"loss": 0.5238,
"step": 349
},
{
"epoch": 0.6370875995449374,
"grad_norm": 1.3042046637167102,
"learning_rate": 2.4583614808361508e-06,
"loss": 0.4785,
"step": 350
},
{
"epoch": 0.6389078498293516,
"grad_norm": 1.1803375558623432,
"learning_rate": 2.4365923163295083e-06,
"loss": 0.5518,
"step": 351
},
{
"epoch": 0.6407281001137657,
"grad_norm": 1.1305498748692666,
"learning_rate": 2.4148776707700775e-06,
"loss": 0.4627,
"step": 352
},
{
"epoch": 0.6425483503981797,
"grad_norm": 1.3696351562191598,
"learning_rate": 2.393218301388123e-06,
"loss": 0.4714,
"step": 353
},
{
"epoch": 0.6443686006825938,
"grad_norm": 1.1646766669675297,
"learning_rate": 2.3716149634863244e-06,
"loss": 0.461,
"step": 354
},
{
"epoch": 0.646188850967008,
"grad_norm": 1.432326181786707,
"learning_rate": 2.3500684104134433e-06,
"loss": 0.4775,
"step": 355
},
{
"epoch": 0.6480091012514221,
"grad_norm": 1.2811226649238618,
"learning_rate": 2.328579393538046e-06,
"loss": 0.4473,
"step": 356
},
{
"epoch": 0.6498293515358362,
"grad_norm": 1.2847418645420832,
"learning_rate": 2.3071486622223e-06,
"loss": 0.473,
"step": 357
},
{
"epoch": 0.6516496018202503,
"grad_norm": 1.1991535525500763,
"learning_rate": 2.2857769637958554e-06,
"loss": 0.4548,
"step": 358
},
{
"epoch": 0.6534698521046644,
"grad_norm": 1.3510869929117142,
"learning_rate": 2.2644650435297675e-06,
"loss": 0.474,
"step": 359
},
{
"epoch": 0.6552901023890785,
"grad_norm": 1.2247454838152558,
"learning_rate": 2.243213644610519e-06,
"loss": 0.4063,
"step": 360
},
{
"epoch": 0.6571103526734926,
"grad_norm": 1.224682187747472,
"learning_rate": 2.2220235081140985e-06,
"loss": 0.5137,
"step": 361
},
{
"epoch": 0.6589306029579067,
"grad_norm": 1.5257557938450914,
"learning_rate": 2.2008953729801583e-06,
"loss": 0.4591,
"step": 362
},
{
"epoch": 0.6607508532423209,
"grad_norm": 1.221121328273825,
"learning_rate": 2.1798299759862545e-06,
"loss": 0.5614,
"step": 363
},
{
"epoch": 0.6625711035267349,
"grad_norm": 1.2199344648755224,
"learning_rate": 2.158828051722137e-06,
"loss": 0.5104,
"step": 364
},
{
"epoch": 0.664391353811149,
"grad_norm": 1.3052741269665118,
"learning_rate": 2.137890332564147e-06,
"loss": 0.4732,
"step": 365
},
{
"epoch": 0.6662116040955631,
"grad_norm": 1.1647023893060888,
"learning_rate": 2.117017548649678e-06,
"loss": 0.5229,
"step": 366
},
{
"epoch": 0.6662116040955631,
"eval_accuracy": 0.8129295561130228,
"eval_accuracy_first_token": 0.7684581132856995,
"eval_accuracy_first_token_<": 0.9595800524934384,
"eval_accuracy_first_token_<_total": 1905,
"eval_accuracy_first_token_<|python_tag|>": 0.9094567404426559,
"eval_accuracy_first_token_<|python_tag|>_total": 994,
"eval_accuracy_first_token_Certainly": 0.7741046831955923,
"eval_accuracy_first_token_Certainly_total": 363,
"eval_accuracy_first_token_The": 0.8948233360723089,
"eval_accuracy_first_token_The_total": 2434,
"eval_accuracy_first_token_To": 0.8044871794871795,
"eval_accuracy_first_token_To_total": 936,
"eval_loss": 0.5655013918876648,
"eval_perplexity": 1.1114110979501997,
"eval_runtime": 507.2948,
"eval_samples_per_second": 1.374,
"eval_steps_per_second": 0.688,
"eval_total_number_first_token": 9657,
"step": 366
},
{
"epoch": 0.6680318543799773,
"grad_norm": 1.2605323109478153,
"learning_rate": 2.0962104278517058e-06,
"loss": 0.4634,
"step": 367
},
{
"epoch": 0.6698521046643914,
"grad_norm": 1.153711484102447,
"learning_rate": 2.0754696957534105e-06,
"loss": 0.4578,
"step": 368
},
{
"epoch": 0.6716723549488055,
"grad_norm": 1.4112272127644152,
"learning_rate": 2.0547960756228746e-06,
"loss": 0.5903,
"step": 369
},
{
"epoch": 0.6734926052332195,
"grad_norm": 1.3058143917601592,
"learning_rate": 2.0341902883878626e-06,
"loss": 0.4261,
"step": 370
},
{
"epoch": 0.6753128555176336,
"grad_norm": 1.241032329122879,
"learning_rate": 2.013653052610678e-06,
"loss": 0.4901,
"step": 371
},
{
"epoch": 0.6771331058020478,
"grad_norm": 1.134116834066691,
"learning_rate": 1.993185084463106e-06,
"loss": 0.5478,
"step": 372
},
{
"epoch": 0.6789533560864619,
"grad_norm": 1.2621524843864569,
"learning_rate": 1.97278709770144e-06,
"loss": 0.4521,
"step": 373
},
{
"epoch": 0.680773606370876,
"grad_norm": 1.2445963074217277,
"learning_rate": 1.952459803641597e-06,
"loss": 0.5048,
"step": 374
},
{
"epoch": 0.6825938566552902,
"grad_norm": 1.3367185945909759,
"learning_rate": 1.9322039111342977e-06,
"loss": 0.4859,
"step": 375
},
{
"epoch": 0.6844141069397042,
"grad_norm": 1.502198228897516,
"learning_rate": 1.912020126540366e-06,
"loss": 0.5483,
"step": 376
},
{
"epoch": 0.6862343572241183,
"grad_norm": 1.5682296957615942,
"learning_rate": 1.8919091537060847e-06,
"loss": 0.5403,
"step": 377
},
{
"epoch": 0.6880546075085324,
"grad_norm": 1.2186683041461865,
"learning_rate": 1.8718716939386541e-06,
"loss": 0.4953,
"step": 378
},
{
"epoch": 0.6898748577929465,
"grad_norm": 1.2104649746142353,
"learning_rate": 1.8519084459817362e-06,
"loss": 0.4599,
"step": 379
},
{
"epoch": 0.6916951080773607,
"grad_norm": 1.1390426306451955,
"learning_rate": 1.83202010599109e-06,
"loss": 0.4164,
"step": 380
},
{
"epoch": 0.6935153583617747,
"grad_norm": 1.2956325376708957,
"learning_rate": 1.8122073675102932e-06,
"loss": 0.5417,
"step": 381
},
{
"epoch": 0.6953356086461888,
"grad_norm": 1.1586136644085798,
"learning_rate": 1.792470921446557e-06,
"loss": 0.4365,
"step": 382
},
{
"epoch": 0.697155858930603,
"grad_norm": 1.1975210529143365,
"learning_rate": 1.7728114560466324e-06,
"loss": 0.4956,
"step": 383
},
{
"epoch": 0.6989761092150171,
"grad_norm": 1.5675227569116297,
"learning_rate": 1.753229656872815e-06,
"loss": 0.4646,
"step": 384
},
{
"epoch": 0.7007963594994312,
"grad_norm": 1.1981622083221466,
"learning_rate": 1.7337262067790319e-06,
"loss": 0.5042,
"step": 385
},
{
"epoch": 0.7026166097838453,
"grad_norm": 1.291822326824022,
"learning_rate": 1.7143017858870259e-06,
"loss": 0.5786,
"step": 386
},
{
"epoch": 0.7044368600682593,
"grad_norm": 1.3381873610330526,
"learning_rate": 1.6949570715626532e-06,
"loss": 0.3987,
"step": 387
},
{
"epoch": 0.7062571103526735,
"grad_norm": 1.5233756050791378,
"learning_rate": 1.675692738392247e-06,
"loss": 0.5373,
"step": 388
},
{
"epoch": 0.7080773606370876,
"grad_norm": 1.405379762218711,
"learning_rate": 1.6565094581591015e-06,
"loss": 0.5151,
"step": 389
},
{
"epoch": 0.7098976109215017,
"grad_norm": 1.3827588130238773,
"learning_rate": 1.6374078998200424e-06,
"loss": 0.4868,
"step": 390
},
{
"epoch": 0.7117178612059158,
"grad_norm": 1.3281467896725871,
"learning_rate": 1.6183887294820995e-06,
"loss": 0.4892,
"step": 391
},
{
"epoch": 0.71353811149033,
"grad_norm": 1.5562464103926885,
"learning_rate": 1.5994526103792852e-06,
"loss": 0.5977,
"step": 392
},
{
"epoch": 0.715358361774744,
"grad_norm": 1.2645130650718202,
"learning_rate": 1.5806002028494509e-06,
"loss": 0.4245,
"step": 393
},
{
"epoch": 0.7171786120591581,
"grad_norm": 1.3281593922925885,
"learning_rate": 1.5618321643112738e-06,
"loss": 0.5813,
"step": 394
},
{
"epoch": 0.7189988623435722,
"grad_norm": 1.1215366227811656,
"learning_rate": 1.5431491492413286e-06,
"loss": 0.4276,
"step": 395
},
{
"epoch": 0.7208191126279864,
"grad_norm": 1.3212838118308114,
"learning_rate": 1.52455180915126e-06,
"loss": 0.5774,
"step": 396
},
{
"epoch": 0.7226393629124005,
"grad_norm": 1.2852914600481689,
"learning_rate": 1.506040792565066e-06,
"loss": 0.5057,
"step": 397
},
{
"epoch": 0.7244596131968145,
"grad_norm": 1.280275275618163,
"learning_rate": 1.487616744996484e-06,
"loss": 0.444,
"step": 398
},
{
"epoch": 0.7262798634812286,
"grad_norm": 1.1583238977099228,
"learning_rate": 1.4692803089264772e-06,
"loss": 0.5377,
"step": 399
},
{
"epoch": 0.7281001137656428,
"grad_norm": 1.435157708312753,
"learning_rate": 1.4510321237808377e-06,
"loss": 0.5444,
"step": 400
},
{
"epoch": 0.7299203640500569,
"grad_norm": 1.3208185752900872,
"learning_rate": 1.4328728259078746e-06,
"loss": 0.5566,
"step": 401
},
{
"epoch": 0.731740614334471,
"grad_norm": 1.2130339190915678,
"learning_rate": 1.414803048556236e-06,
"loss": 0.4988,
"step": 402
},
{
"epoch": 0.7335608646188851,
"grad_norm": 1.1363530661008532,
"learning_rate": 1.396823421852825e-06,
"loss": 0.6129,
"step": 403
},
{
"epoch": 0.7353811149032992,
"grad_norm": 1.3222588910481998,
"learning_rate": 1.3789345727808207e-06,
"loss": 0.546,
"step": 404
},
{
"epoch": 0.7372013651877133,
"grad_norm": 1.3949194783709729,
"learning_rate": 1.3611371251578114e-06,
"loss": 0.5583,
"step": 405
},
{
"epoch": 0.7390216154721274,
"grad_norm": 1.2917335175784925,
"learning_rate": 1.3434316996140553e-06,
"loss": 0.5151,
"step": 406
},
{
"epoch": 0.7408418657565415,
"grad_norm": 1.2895735708732046,
"learning_rate": 1.3258189135708229e-06,
"loss": 0.5098,
"step": 407
},
{
"epoch": 0.7426621160409557,
"grad_norm": 1.2978294874532978,
"learning_rate": 1.3082993812188735e-06,
"loss": 0.5414,
"step": 408
},
{
"epoch": 0.7444823663253698,
"grad_norm": 1.2095221030821062,
"learning_rate": 1.2908737134970364e-06,
"loss": 0.5268,
"step": 409
},
{
"epoch": 0.7463026166097838,
"grad_norm": 1.3840563503977592,
"learning_rate": 1.2735425180709039e-06,
"loss": 0.479,
"step": 410
},
{
"epoch": 0.7481228668941979,
"grad_norm": 1.2789076883026242,
"learning_rate": 1.2563063993116482e-06,
"loss": 0.5503,
"step": 411
},
{
"epoch": 0.7499431171786121,
"grad_norm": 1.283587802005637,
"learning_rate": 1.239165958274933e-06,
"loss": 0.4113,
"step": 412
},
{
"epoch": 0.7517633674630262,
"grad_norm": 1.2909165266250262,
"learning_rate": 1.2221217926799652e-06,
"loss": 0.535,
"step": 413
},
{
"epoch": 0.7535836177474403,
"grad_norm": 1.3531455484884616,
"learning_rate": 1.2051744968886489e-06,
"loss": 0.5052,
"step": 414
},
{
"epoch": 0.7554038680318543,
"grad_norm": 1.2730404093480168,
"learning_rate": 1.1883246618848533e-06,
"loss": 0.4566,
"step": 415
},
{
"epoch": 0.7572241183162685,
"grad_norm": 1.37616764437592,
"learning_rate": 1.1715728752538101e-06,
"loss": 0.566,
"step": 416
},
{
"epoch": 0.7590443686006826,
"grad_norm": 1.1512441975212944,
"learning_rate": 1.1549197211616203e-06,
"loss": 0.5044,
"step": 417
},
{
"epoch": 0.7608646188850967,
"grad_norm": 1.2438970988598956,
"learning_rate": 1.1383657803348835e-06,
"loss": 0.5109,
"step": 418
},
{
"epoch": 0.7626848691695108,
"grad_norm": 1.5233735431446764,
"learning_rate": 1.1219116300404486e-06,
"loss": 0.507,
"step": 419
},
{
"epoch": 0.764505119453925,
"grad_norm": 1.3253161212074762,
"learning_rate": 1.10555784406528e-06,
"loss": 0.5082,
"step": 420
},
{
"epoch": 0.766325369738339,
"grad_norm": 1.1775521474516462,
"learning_rate": 1.089304992696455e-06,
"loss": 0.46,
"step": 421
},
{
"epoch": 0.7681456200227531,
"grad_norm": 1.2462962157301152,
"learning_rate": 1.0731536427012695e-06,
"loss": 0.5253,
"step": 422
},
{
"epoch": 0.7699658703071672,
"grad_norm": 1.3347631673740097,
"learning_rate": 1.0571043573074736e-06,
"loss": 0.4449,
"step": 423
},
{
"epoch": 0.7717861205915814,
"grad_norm": 1.292727758187721,
"learning_rate": 1.041157696183641e-06,
"loss": 0.441,
"step": 424
},
{
"epoch": 0.7736063708759955,
"grad_norm": 1.293278742294603,
"learning_rate": 1.0253142154196415e-06,
"loss": 0.4867,
"step": 425
},
{
"epoch": 0.7754266211604095,
"grad_norm": 1.2102494852297525,
"learning_rate": 1.0095744675072525e-06,
"loss": 0.4898,
"step": 426
},
{
"epoch": 0.7772468714448236,
"grad_norm": 1.224313028246693,
"learning_rate": 9.93939001320895e-07,
"loss": 0.4686,
"step": 427
},
{
"epoch": 0.7790671217292378,
"grad_norm": 1.3632517015375165,
"learning_rate": 9.784083620984884e-07,
"loss": 0.4639,
"step": 428
},
{
"epoch": 0.7808873720136519,
"grad_norm": 1.3987002856426751,
"learning_rate": 9.62983091422446e-07,
"loss": 0.4528,
"step": 429
},
{
"epoch": 0.782707622298066,
"grad_norm": 1.329331750067852,
"learning_rate": 9.476637272007746e-07,
"loss": 0.4562,
"step": 430
},
{
"epoch": 0.7845278725824801,
"grad_norm": 1.4216744583623766,
"learning_rate": 9.324508036483303e-07,
"loss": 0.4622,
"step": 431
},
{
"epoch": 0.7863481228668942,
"grad_norm": 1.3060911776176307,
"learning_rate": 9.173448512681848e-07,
"loss": 0.5405,
"step": 432
},
{
"epoch": 0.7881683731513083,
"grad_norm": 1.3971532684012182,
"learning_rate": 9.023463968331238e-07,
"loss": 0.4642,
"step": 433
},
{
"epoch": 0.7899886234357224,
"grad_norm": 1.351332971443725,
"learning_rate": 8.874559633672754e-07,
"loss": 0.4146,
"step": 434
},
{
"epoch": 0.7918088737201365,
"grad_norm": 1.2506853747891504,
"learning_rate": 8.726740701278808e-07,
"loss": 0.5233,
"step": 435
},
{
"epoch": 0.7936291240045507,
"grad_norm": 1.2588296359051319,
"learning_rate": 8.580012325871773e-07,
"loss": 0.5196,
"step": 436
},
{
"epoch": 0.7954493742889648,
"grad_norm": 1.3656683873360818,
"learning_rate": 8.434379624144261e-07,
"loss": 0.4426,
"step": 437
},
{
"epoch": 0.7972696245733788,
"grad_norm": 1.3415371986074633,
"learning_rate": 8.289847674580702e-07,
"loss": 0.5025,
"step": 438
},
{
"epoch": 0.7990898748577929,
"grad_norm": 1.210310044679145,
"learning_rate": 8.146421517280226e-07,
"loss": 0.4922,
"step": 439
},
{
"epoch": 0.800910125142207,
"grad_norm": 1.675036054936253,
"learning_rate": 8.004106153780967e-07,
"loss": 0.4396,
"step": 440
},
{
"epoch": 0.8027303754266212,
"grad_norm": 1.1849449434556916,
"learning_rate": 7.862906546885559e-07,
"loss": 0.5348,
"step": 441
},
{
"epoch": 0.8045506257110353,
"grad_norm": 1.3294402423567042,
"learning_rate": 7.722827620488108e-07,
"loss": 0.4472,
"step": 442
},
{
"epoch": 0.8063708759954493,
"grad_norm": 1.2341888202472633,
"learning_rate": 7.583874259402545e-07,
"loss": 0.4926,
"step": 443
},
{
"epoch": 0.8081911262798634,
"grad_norm": 1.3727750069417188,
"learning_rate": 7.446051309192203e-07,
"loss": 0.5187,
"step": 444
},
{
"epoch": 0.8100113765642776,
"grad_norm": 1.1665673148184286,
"learning_rate": 7.30936357600088e-07,
"loss": 0.4459,
"step": 445
},
{
"epoch": 0.8118316268486917,
"grad_norm": 1.4461908262228584,
"learning_rate": 7.173815826385246e-07,
"loss": 0.5931,
"step": 446
},
{
"epoch": 0.8136518771331058,
"grad_norm": 1.2164762112018974,
"learning_rate": 7.039412787148586e-07,
"loss": 0.5769,
"step": 447
},
{
"epoch": 0.81547212741752,
"grad_norm": 1.3268169931538385,
"learning_rate": 6.906159145176049e-07,
"loss": 0.4962,
"step": 448
},
{
"epoch": 0.817292377701934,
"grad_norm": 1.3585533527783662,
"learning_rate": 6.774059547271087e-07,
"loss": 0.5011,
"step": 449
},
{
"epoch": 0.8191126279863481,
"grad_norm": 1.2715237655057547,
"learning_rate": 6.643118599993518e-07,
"loss": 0.4591,
"step": 450
},
{
"epoch": 0.8209328782707622,
"grad_norm": 1.1129340141314334,
"learning_rate": 6.513340869498858e-07,
"loss": 0.4818,
"step": 451
},
{
"epoch": 0.8227531285551763,
"grad_norm": 1.1993408265317198,
"learning_rate": 6.384730881379048e-07,
"loss": 0.4826,
"step": 452
},
{
"epoch": 0.8245733788395905,
"grad_norm": 1.305009025174831,
"learning_rate": 6.257293120504692e-07,
"loss": 0.4824,
"step": 453
},
{
"epoch": 0.8263936291240046,
"grad_norm": 1.2620160974509798,
"learning_rate": 6.131032030868635e-07,
"loss": 0.4479,
"step": 454
},
{
"epoch": 0.8282138794084186,
"grad_norm": 1.2693469945741236,
"learning_rate": 6.005952015430993e-07,
"loss": 0.5286,
"step": 455
},
{
"epoch": 0.8300341296928327,
"grad_norm": 1.2953927032105943,
"learning_rate": 5.882057435965619e-07,
"loss": 0.5802,
"step": 456
},
{
"epoch": 0.8318543799772469,
"grad_norm": 1.3055790274997285,
"learning_rate": 5.759352612907999e-07,
"loss": 0.5273,
"step": 457
},
{
"epoch": 0.833674630261661,
"grad_norm": 1.3009913306704852,
"learning_rate": 5.637841825204588e-07,
"loss": 0.4434,
"step": 458
},
{
"epoch": 0.8354948805460751,
"grad_norm": 1.3010149850935786,
"learning_rate": 5.517529310163627e-07,
"loss": 0.5302,
"step": 459
},
{
"epoch": 0.8373151308304891,
"grad_norm": 1.1588504398899486,
"learning_rate": 5.398419263307281e-07,
"loss": 0.4898,
"step": 460
},
{
"epoch": 0.8391353811149033,
"grad_norm": 1.427556447905731,
"learning_rate": 5.280515838225477e-07,
"loss": 0.4583,
"step": 461
},
{
"epoch": 0.8409556313993174,
"grad_norm": 1.3382828189315212,
"learning_rate": 5.163823146430944e-07,
"loss": 0.4544,
"step": 462
},
{
"epoch": 0.8427758816837315,
"grad_norm": 1.3048820751365628,
"learning_rate": 5.048345257215892e-07,
"loss": 0.5348,
"step": 463
},
{
"epoch": 0.8445961319681456,
"grad_norm": 1.3464339683482869,
"learning_rate": 4.934086197510088e-07,
"loss": 0.4866,
"step": 464
},
{
"epoch": 0.8464163822525598,
"grad_norm": 1.3076973707605393,
"learning_rate": 4.821049951740441e-07,
"loss": 0.4374,
"step": 465
},
{
"epoch": 0.8482366325369738,
"grad_norm": 1.207783472984328,
"learning_rate": 4.7092404616920547e-07,
"loss": 0.5268,
"step": 466
},
{
"epoch": 0.8500568828213879,
"grad_norm": 1.3340034898150066,
"learning_rate": 4.59866162637077e-07,
"loss": 0.5163,
"step": 467
},
{
"epoch": 0.851877133105802,
"grad_norm": 1.2793323359204207,
"learning_rate": 4.4893173018671816e-07,
"loss": 0.464,
"step": 468
},
{
"epoch": 0.8536973833902162,
"grad_norm": 1.3875887367624027,
"learning_rate": 4.3812113012222164e-07,
"loss": 0.5605,
"step": 469
},
{
"epoch": 0.8555176336746303,
"grad_norm": 1.2752397131609516,
"learning_rate": 4.2743473942941177e-07,
"loss": 0.5166,
"step": 470
},
{
"epoch": 0.8573378839590444,
"grad_norm": 1.361832548391048,
"learning_rate": 4.168729307626977e-07,
"loss": 0.4494,
"step": 471
},
{
"epoch": 0.8591581342434584,
"grad_norm": 1.3313280628055624,
"learning_rate": 4.0643607243208455e-07,
"loss": 0.4531,
"step": 472
},
{
"epoch": 0.8609783845278726,
"grad_norm": 1.238927541446331,
"learning_rate": 3.9612452839032384e-07,
"loss": 0.4629,
"step": 473
},
{
"epoch": 0.8627986348122867,
"grad_norm": 1.440299941933543,
"learning_rate": 3.859386582202231e-07,
"loss": 0.5238,
"step": 474
},
{
"epoch": 0.8646188850967008,
"grad_norm": 1.2998009060977955,
"learning_rate": 3.758788171221079e-07,
"loss": 0.4126,
"step": 475
},
{
"epoch": 0.8664391353811149,
"grad_norm": 1.2344313543035759,
"learning_rate": 3.659453559014345e-07,
"loss": 0.3997,
"step": 476
},
{
"epoch": 0.868259385665529,
"grad_norm": 1.3123945291493502,
"learning_rate": 3.561386209565582e-07,
"loss": 0.4354,
"step": 477
},
{
"epoch": 0.8700796359499431,
"grad_norm": 1.3385863981096489,
"learning_rate": 3.464589542666485e-07,
"loss": 0.5423,
"step": 478
},
{
"epoch": 0.8718998862343572,
"grad_norm": 1.4693361278099728,
"learning_rate": 3.3690669337976996e-07,
"loss": 0.5439,
"step": 479
},
{
"epoch": 0.8737201365187713,
"grad_norm": 1.219115488818529,
"learning_rate": 3.2748217140111e-07,
"loss": 0.55,
"step": 480
},
{
"epoch": 0.8755403868031855,
"grad_norm": 1.3943875753971013,
"learning_rate": 3.1818571698135976e-07,
"loss": 0.479,
"step": 481
},
{
"epoch": 0.8773606370875996,
"grad_norm": 1.31283354874802,
"learning_rate": 3.0901765430525337e-07,
"loss": 0.4546,
"step": 482
},
{
"epoch": 0.8791808873720136,
"grad_norm": 1.2760697675194013,
"learning_rate": 2.9997830308027003e-07,
"loss": 0.5241,
"step": 483
},
{
"epoch": 0.8810011376564277,
"grad_norm": 1.3057898765814404,
"learning_rate": 2.9106797852547483e-07,
"loss": 0.5045,
"step": 484
},
{
"epoch": 0.8828213879408419,
"grad_norm": 1.228721100779524,
"learning_rate": 2.8228699136053726e-07,
"loss": 0.4588,
"step": 485
},
{
"epoch": 0.884641638225256,
"grad_norm": 1.4327825477254865,
"learning_rate": 2.7363564779488446e-07,
"loss": 0.4911,
"step": 486
},
{
"epoch": 0.8864618885096701,
"grad_norm": 1.1675697744027835,
"learning_rate": 2.6511424951703244e-07,
"loss": 0.4503,
"step": 487
},
{
"epoch": 0.8882821387940842,
"grad_norm": 1.3573793521283821,
"learning_rate": 2.567230936840632e-07,
"loss": 0.5537,
"step": 488
},
{
"epoch": 0.8901023890784983,
"grad_norm": 1.2385857779190943,
"learning_rate": 2.4846247291125897e-07,
"loss": 0.5261,
"step": 489
},
{
"epoch": 0.8919226393629124,
"grad_norm": 1.3747886513978498,
"learning_rate": 2.4033267526190057e-07,
"loss": 0.5116,
"step": 490
},
{
"epoch": 0.8937428896473265,
"grad_norm": 1.3015002806547666,
"learning_rate": 2.323339842372234e-07,
"loss": 0.501,
"step": 491
},
{
"epoch": 0.8955631399317406,
"grad_norm": 1.2282471393147485,
"learning_rate": 2.2446667876652968e-07,
"loss": 0.5615,
"step": 492
},
{
"epoch": 0.8973833902161548,
"grad_norm": 1.2246787300329813,
"learning_rate": 2.1673103319746146e-07,
"loss": 0.5847,
"step": 493
},
{
"epoch": 0.8992036405005688,
"grad_norm": 1.381507003520726,
"learning_rate": 2.0912731728643362e-07,
"loss": 0.4593,
"step": 494
},
{
"epoch": 0.9010238907849829,
"grad_norm": 1.2236872730147548,
"learning_rate": 2.0165579618922757e-07,
"loss": 0.426,
"step": 495
},
{
"epoch": 0.902844141069397,
"grad_norm": 1.5642555968533283,
"learning_rate": 1.943167304517459e-07,
"loss": 0.4669,
"step": 496
},
{
"epoch": 0.9046643913538112,
"grad_norm": 1.4113035349877263,
"learning_rate": 1.871103760009234e-07,
"loss": 0.5189,
"step": 497
},
{
"epoch": 0.9064846416382253,
"grad_norm": 1.2945664446971985,
"learning_rate": 1.8003698413580427e-07,
"loss": 0.5331,
"step": 498
},
{
"epoch": 0.9083048919226394,
"grad_norm": 1.1216172834522593,
"learning_rate": 1.7309680151878126e-07,
"loss": 0.4596,
"step": 499
},
{
"epoch": 0.9101251422070534,
"grad_norm": 1.4490178400997769,
"learning_rate": 1.6629007016698916e-07,
"loss": 0.5719,
"step": 500
},
{
"epoch": 0.9119453924914676,
"grad_norm": 1.4927164965040023,
"learning_rate": 1.5961702744386973e-07,
"loss": 0.4637,
"step": 501
},
{
"epoch": 0.9137656427758817,
"grad_norm": 1.2926779903672145,
"learning_rate": 1.5307790605089045e-07,
"loss": 0.4931,
"step": 502
},
{
"epoch": 0.9155858930602958,
"grad_norm": 1.3434407972538571,
"learning_rate": 1.4667293401943393e-07,
"loss": 0.4843,
"step": 503
},
{
"epoch": 0.9174061433447099,
"grad_norm": 1.2627460036138376,
"learning_rate": 1.404023347028418e-07,
"loss": 0.4628,
"step": 504
},
{
"epoch": 0.919226393629124,
"grad_norm": 1.1980109325087624,
"learning_rate": 1.342663267686297e-07,
"loss": 0.547,
"step": 505
},
{
"epoch": 0.9210466439135381,
"grad_norm": 1.4394748326258473,
"learning_rate": 1.2826512419085922e-07,
"loss": 0.4852,
"step": 506
},
{
"epoch": 0.9228668941979522,
"grad_norm": 1.1712799414971835,
"learning_rate": 1.223989362426785e-07,
"loss": 0.5027,
"step": 507
},
{
"epoch": 0.9246871444823663,
"grad_norm": 1.2917639503148088,
"learning_rate": 1.1666796748902142e-07,
"loss": 0.4318,
"step": 508
},
{
"epoch": 0.9265073947667805,
"grad_norm": 1.407559329871179,
"learning_rate": 1.1107241777947774e-07,
"loss": 0.452,
"step": 509
},
{
"epoch": 0.9283276450511946,
"grad_norm": 1.4395176866301798,
"learning_rate": 1.0561248224132091e-07,
"loss": 0.5792,
"step": 510
},
{
"epoch": 0.9301478953356086,
"grad_norm": 1.3107228117658043,
"learning_rate": 1.0028835127270552e-07,
"loss": 0.523,
"step": 511
},
{
"epoch": 0.9319681456200227,
"grad_norm": 1.319280624009732,
"learning_rate": 9.510021053602679e-08,
"loss": 0.4903,
"step": 512
},
{
"epoch": 0.9337883959044369,
"grad_norm": 1.2825750147020196,
"learning_rate": 9.004824095144581e-08,
"loss": 0.486,
"step": 513
},
{
"epoch": 0.935608646188851,
"grad_norm": 1.3550036994824897,
"learning_rate": 8.513261869058209e-08,
"loss": 0.4342,
"step": 514
},
{
"epoch": 0.9374288964732651,
"grad_norm": 1.2912511428181583,
"learning_rate": 8.035351517036914e-08,
"loss": 0.4975,
"step": 515
},
{
"epoch": 0.9392491467576792,
"grad_norm": 1.2630516224119532,
"learning_rate": 7.571109704707623e-08,
"loss": 0.4942,
"step": 516
},
{
"epoch": 0.9410693970420932,
"grad_norm": 1.2630983628627157,
"learning_rate": 7.120552621049825e-08,
"loss": 0.4581,
"step": 517
},
{
"epoch": 0.9428896473265074,
"grad_norm": 1.184276479260659,
"learning_rate": 6.68369597783096e-08,
"loss": 0.4245,
"step": 518
},
{
"epoch": 0.9447098976109215,
"grad_norm": 1.3479750123046965,
"learning_rate": 6.260555009058288e-08,
"loss": 0.4734,
"step": 519
},
{
"epoch": 0.9465301478953356,
"grad_norm": 1.184265059530281,
"learning_rate": 5.851144470448144e-08,
"loss": 0.5263,
"step": 520
},
{
"epoch": 0.9483503981797498,
"grad_norm": 1.3131542129196199,
"learning_rate": 5.455478638911071e-08,
"loss": 0.369,
"step": 521
},
{
"epoch": 0.9501706484641638,
"grad_norm": 1.3396828056059393,
"learning_rate": 5.073571312053815e-08,
"loss": 0.5098,
"step": 522
},
{
"epoch": 0.9519908987485779,
"grad_norm": 1.3620509437765531,
"learning_rate": 4.705435807698555e-08,
"loss": 0.5595,
"step": 523
},
{
"epoch": 0.953811149032992,
"grad_norm": 1.3476395824069989,
"learning_rate": 4.351084963418117e-08,
"loss": 0.5332,
"step": 524
},
{
"epoch": 0.9556313993174061,
"grad_norm": 1.4056028428746756,
"learning_rate": 4.010531136088691e-08,
"loss": 0.5135,
"step": 525
},
{
"epoch": 0.9574516496018203,
"grad_norm": 1.2931973314368226,
"learning_rate": 3.683786201458439e-08,
"loss": 0.4869,
"step": 526
},
{
"epoch": 0.9592718998862344,
"grad_norm": 1.3728587745363008,
"learning_rate": 3.370861553733784e-08,
"loss": 0.544,
"step": 527
},
{
"epoch": 0.9610921501706484,
"grad_norm": 1.379130929011516,
"learning_rate": 3.071768105181993e-08,
"loss": 0.4312,
"step": 528
},
{
"epoch": 0.9629124004550625,
"grad_norm": 1.4162454724368647,
"learning_rate": 2.786516285750373e-08,
"loss": 0.4464,
"step": 529
},
{
"epoch": 0.9647326507394767,
"grad_norm": 1.310107669303508,
"learning_rate": 2.5151160427029582e-08,
"loss": 0.4641,
"step": 530
},
{
"epoch": 0.9665529010238908,
"grad_norm": 1.3049449814100964,
"learning_rate": 2.2575768402733232e-08,
"loss": 0.5079,
"step": 531
},
{
"epoch": 0.9683731513083049,
"grad_norm": 1.301610299072927,
"learning_rate": 2.013907659334624e-08,
"loss": 0.4798,
"step": 532
},
{
"epoch": 0.970193401592719,
"grad_norm": 1.2912511438851022,
"learning_rate": 1.7841169970866042e-08,
"loss": 0.4962,
"step": 533
},
{
"epoch": 0.9720136518771331,
"grad_norm": 1.3741948558886383,
"learning_rate": 1.5682128667589e-08,
"loss": 0.4556,
"step": 534
},
{
"epoch": 0.9738339021615472,
"grad_norm": 1.4394930710163565,
"learning_rate": 1.3662027973320612e-08,
"loss": 0.4808,
"step": 535
},
{
"epoch": 0.9756541524459613,
"grad_norm": 1.3189189784853037,
"learning_rate": 1.1780938332746515e-08,
"loss": 0.4601,
"step": 536
},
{
"epoch": 0.9774744027303754,
"grad_norm": 1.3675135301050803,
"learning_rate": 1.0038925342977122e-08,
"loss": 0.4862,
"step": 537
},
{
"epoch": 0.9792946530147896,
"grad_norm": 1.2496142970199702,
"learning_rate": 8.43604975126011e-09,
"loss": 0.4972,
"step": 538
},
{
"epoch": 0.9811149032992036,
"grad_norm": 1.2590220357743287,
"learning_rate": 6.972367452863004e-09,
"loss": 0.5048,
"step": 539
},
{
"epoch": 0.9829351535836177,
"grad_norm": 1.4321645089041766,
"learning_rate": 5.647929489122738e-09,
"loss": 0.5688,
"step": 540
},
{
"epoch": 0.9847554038680318,
"grad_norm": 1.3266758203446563,
"learning_rate": 4.462782045664859e-09,
"loss": 0.4745,
"step": 541
},
{
"epoch": 0.986575654152446,
"grad_norm": 1.1860109568892805,
"learning_rate": 3.4169664507959216e-09,
"loss": 0.4616,
"step": 542
},
{
"epoch": 0.9883959044368601,
"grad_norm": 1.2475217679660848,
"learning_rate": 2.5105191740597553e-09,
"loss": 0.6489,
"step": 543
},
{
"epoch": 0.9902161547212742,
"grad_norm": 1.4270596684886099,
"learning_rate": 1.7434718249664803e-09,
"loss": 0.4712,
"step": 544
},
{
"epoch": 0.9920364050056882,
"grad_norm": 1.2999695109285117,
"learning_rate": 1.1158511518902791e-09,
"loss": 0.5143,
"step": 545
},
{
"epoch": 0.9938566552901024,
"grad_norm": 1.5041014788909566,
"learning_rate": 6.276790411372524e-10,
"loss": 0.4971,
"step": 546
},
{
"epoch": 0.9956769055745165,
"grad_norm": 1.151430673100721,
"learning_rate": 2.789725161806977e-10,
"loss": 0.5446,
"step": 547
},
{
"epoch": 0.9974971558589306,
"grad_norm": 1.2513093960410882,
"learning_rate": 6.974373706869486e-11,
"loss": 0.5494,
"step": 548
},
{
"epoch": 0.9993174061433447,
"grad_norm": 1.3143110421818924,
"learning_rate": 0.0,
"loss": 0.4345,
"step": 549
},
{
"epoch": 0.9993174061433447,
"eval_accuracy": 0.8141491715694426,
"eval_accuracy_first_token": 0.7775706741223982,
"eval_accuracy_first_token_<": 0.9648293963254593,
"eval_accuracy_first_token_<_total": 1905,
"eval_accuracy_first_token_<|python_tag|>": 0.9014084507042254,
"eval_accuracy_first_token_<|python_tag|>_total": 994,
"eval_accuracy_first_token_Certainly": 0.743801652892562,
"eval_accuracy_first_token_Certainly_total": 363,
"eval_accuracy_first_token_The": 0.9030402629416598,
"eval_accuracy_first_token_The_total": 2434,
"eval_accuracy_first_token_To": 0.8076923076923077,
"eval_accuracy_first_token_To_total": 936,
"eval_loss": 0.5610479116439819,
"eval_perplexity": 1.110590475782418,
"eval_runtime": 507.5321,
"eval_samples_per_second": 1.373,
"eval_steps_per_second": 0.688,
"eval_total_number_first_token": 9657,
"step": 549
},
{
"epoch": 0.9993174061433447,
"step": 549,
"total_flos": 229846517022720.0,
"train_loss": 0.5347839987994544,
"train_runtime": 35925.1648,
"train_samples_per_second": 0.245,
"train_steps_per_second": 0.015
}
],
"logging_steps": 1.0,
"max_steps": 549,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 5.0,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 229846517022720.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}