NuminaMath-72B-TIR / trainer_state.json
lewtun's picture
lewtun HF staff
Add AI-MO/qwen2-72b-sft-aimo_v03.00 checkpoint
7682df3 verified
raw
history blame
113 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.0,
"eval_steps": 500,
"global_step": 3188,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0012547051442910915,
"grad_norm": 1.3249917794841757,
"learning_rate": 6.269592476489028e-08,
"loss": 0.5553,
"step": 1
},
{
"epoch": 0.006273525721455458,
"grad_norm": 1.3024175063683066,
"learning_rate": 3.134796238244514e-07,
"loss": 0.5542,
"step": 5
},
{
"epoch": 0.012547051442910916,
"grad_norm": 1.1325683997600668,
"learning_rate": 6.269592476489028e-07,
"loss": 0.5626,
"step": 10
},
{
"epoch": 0.018820577164366373,
"grad_norm": 1.1174840712777703,
"learning_rate": 9.404388714733543e-07,
"loss": 0.5642,
"step": 15
},
{
"epoch": 0.025094102885821833,
"grad_norm": 0.8002431481734196,
"learning_rate": 1.2539184952978056e-06,
"loss": 0.5294,
"step": 20
},
{
"epoch": 0.03136762860727729,
"grad_norm": 0.6333668642117685,
"learning_rate": 1.5673981191222572e-06,
"loss": 0.5079,
"step": 25
},
{
"epoch": 0.037641154328732745,
"grad_norm": 0.5771411644874956,
"learning_rate": 1.8808777429467086e-06,
"loss": 0.4655,
"step": 30
},
{
"epoch": 0.043914680050188205,
"grad_norm": 0.49532156556877105,
"learning_rate": 2.1943573667711602e-06,
"loss": 0.4592,
"step": 35
},
{
"epoch": 0.050188205771643665,
"grad_norm": 0.4975545815570396,
"learning_rate": 2.507836990595611e-06,
"loss": 0.4329,
"step": 40
},
{
"epoch": 0.056461731493099125,
"grad_norm": 0.4338284541758203,
"learning_rate": 2.8213166144200626e-06,
"loss": 0.4322,
"step": 45
},
{
"epoch": 0.06273525721455459,
"grad_norm": 0.40887016256028313,
"learning_rate": 3.1347962382445144e-06,
"loss": 0.426,
"step": 50
},
{
"epoch": 0.06900878293601004,
"grad_norm": 0.4045559102283436,
"learning_rate": 3.448275862068966e-06,
"loss": 0.4227,
"step": 55
},
{
"epoch": 0.07528230865746549,
"grad_norm": 0.3889912721227527,
"learning_rate": 3.7617554858934172e-06,
"loss": 0.4189,
"step": 60
},
{
"epoch": 0.08155583437892096,
"grad_norm": 0.3618436855347915,
"learning_rate": 4.075235109717869e-06,
"loss": 0.4177,
"step": 65
},
{
"epoch": 0.08782936010037641,
"grad_norm": 0.36459156314439295,
"learning_rate": 4.3887147335423205e-06,
"loss": 0.4049,
"step": 70
},
{
"epoch": 0.09410288582183186,
"grad_norm": 0.3488164206674813,
"learning_rate": 4.7021943573667714e-06,
"loss": 0.3961,
"step": 75
},
{
"epoch": 0.10037641154328733,
"grad_norm": 0.37238296726059605,
"learning_rate": 5.015673981191222e-06,
"loss": 0.3955,
"step": 80
},
{
"epoch": 0.10664993726474278,
"grad_norm": 0.3605666937163523,
"learning_rate": 5.329153605015674e-06,
"loss": 0.377,
"step": 85
},
{
"epoch": 0.11292346298619825,
"grad_norm": 0.35760877488985304,
"learning_rate": 5.642633228840125e-06,
"loss": 0.3915,
"step": 90
},
{
"epoch": 0.1191969887076537,
"grad_norm": 0.356976698911797,
"learning_rate": 5.956112852664577e-06,
"loss": 0.3938,
"step": 95
},
{
"epoch": 0.12547051442910917,
"grad_norm": 0.3620265985263758,
"learning_rate": 6.269592476489029e-06,
"loss": 0.404,
"step": 100
},
{
"epoch": 0.13174404015056462,
"grad_norm": 0.3710088599948379,
"learning_rate": 6.58307210031348e-06,
"loss": 0.3941,
"step": 105
},
{
"epoch": 0.13801756587202008,
"grad_norm": 0.3831687285367315,
"learning_rate": 6.896551724137932e-06,
"loss": 0.3859,
"step": 110
},
{
"epoch": 0.14429109159347553,
"grad_norm": 0.36915661135239697,
"learning_rate": 7.210031347962383e-06,
"loss": 0.3895,
"step": 115
},
{
"epoch": 0.15056461731493098,
"grad_norm": 0.3573608106603279,
"learning_rate": 7.5235109717868345e-06,
"loss": 0.3857,
"step": 120
},
{
"epoch": 0.15683814303638646,
"grad_norm": 0.3508976740749952,
"learning_rate": 7.836990595611285e-06,
"loss": 0.3915,
"step": 125
},
{
"epoch": 0.16311166875784192,
"grad_norm": 0.36876078420022057,
"learning_rate": 8.150470219435737e-06,
"loss": 0.3989,
"step": 130
},
{
"epoch": 0.16938519447929737,
"grad_norm": 0.38691192833572297,
"learning_rate": 8.463949843260189e-06,
"loss": 0.3874,
"step": 135
},
{
"epoch": 0.17565872020075282,
"grad_norm": 0.3735894147297392,
"learning_rate": 8.777429467084641e-06,
"loss": 0.385,
"step": 140
},
{
"epoch": 0.18193224592220827,
"grad_norm": 0.3710457112042887,
"learning_rate": 9.090909090909091e-06,
"loss": 0.3764,
"step": 145
},
{
"epoch": 0.18820577164366373,
"grad_norm": 0.3708916534975576,
"learning_rate": 9.404388714733543e-06,
"loss": 0.3808,
"step": 150
},
{
"epoch": 0.1944792973651192,
"grad_norm": 0.3552539505765215,
"learning_rate": 9.717868338557995e-06,
"loss": 0.4018,
"step": 155
},
{
"epoch": 0.20075282308657466,
"grad_norm": 0.3666385022952003,
"learning_rate": 1.0031347962382445e-05,
"loss": 0.3839,
"step": 160
},
{
"epoch": 0.20702634880803011,
"grad_norm": 0.3631078993279255,
"learning_rate": 1.0344827586206898e-05,
"loss": 0.3858,
"step": 165
},
{
"epoch": 0.21329987452948557,
"grad_norm": 0.39722641298153644,
"learning_rate": 1.0658307210031348e-05,
"loss": 0.3847,
"step": 170
},
{
"epoch": 0.21957340025094102,
"grad_norm": 0.3691443252830364,
"learning_rate": 1.09717868338558e-05,
"loss": 0.3826,
"step": 175
},
{
"epoch": 0.2258469259723965,
"grad_norm": 0.39755512136650856,
"learning_rate": 1.128526645768025e-05,
"loss": 0.3766,
"step": 180
},
{
"epoch": 0.23212045169385195,
"grad_norm": 0.3627344076525724,
"learning_rate": 1.1598746081504704e-05,
"loss": 0.409,
"step": 185
},
{
"epoch": 0.2383939774153074,
"grad_norm": 0.36755006735418844,
"learning_rate": 1.1912225705329154e-05,
"loss": 0.385,
"step": 190
},
{
"epoch": 0.24466750313676286,
"grad_norm": 0.3491529760844153,
"learning_rate": 1.2225705329153606e-05,
"loss": 0.3944,
"step": 195
},
{
"epoch": 0.25094102885821834,
"grad_norm": 0.38344322242421625,
"learning_rate": 1.2539184952978058e-05,
"loss": 0.3814,
"step": 200
},
{
"epoch": 0.2572145545796738,
"grad_norm": 0.3668326981157007,
"learning_rate": 1.285266457680251e-05,
"loss": 0.3769,
"step": 205
},
{
"epoch": 0.26348808030112925,
"grad_norm": 0.3606059535955065,
"learning_rate": 1.316614420062696e-05,
"loss": 0.3882,
"step": 210
},
{
"epoch": 0.2697616060225847,
"grad_norm": 0.3743059326798353,
"learning_rate": 1.3479623824451411e-05,
"loss": 0.3792,
"step": 215
},
{
"epoch": 0.27603513174404015,
"grad_norm": 0.3987264784021991,
"learning_rate": 1.3793103448275863e-05,
"loss": 0.3742,
"step": 220
},
{
"epoch": 0.2823086574654956,
"grad_norm": 0.41749115581528207,
"learning_rate": 1.4106583072100315e-05,
"loss": 0.3879,
"step": 225
},
{
"epoch": 0.28858218318695106,
"grad_norm": 0.36990941061813976,
"learning_rate": 1.4420062695924765e-05,
"loss": 0.3763,
"step": 230
},
{
"epoch": 0.2948557089084065,
"grad_norm": 0.350857226534173,
"learning_rate": 1.4733542319749217e-05,
"loss": 0.3936,
"step": 235
},
{
"epoch": 0.30112923462986196,
"grad_norm": 0.3699010810723354,
"learning_rate": 1.5047021943573669e-05,
"loss": 0.3827,
"step": 240
},
{
"epoch": 0.3074027603513174,
"grad_norm": 0.383417773018869,
"learning_rate": 1.536050156739812e-05,
"loss": 0.3753,
"step": 245
},
{
"epoch": 0.3136762860727729,
"grad_norm": 0.37418028131825143,
"learning_rate": 1.567398119122257e-05,
"loss": 0.393,
"step": 250
},
{
"epoch": 0.3199498117942284,
"grad_norm": 0.375158938790808,
"learning_rate": 1.598746081504702e-05,
"loss": 0.3803,
"step": 255
},
{
"epoch": 0.32622333751568383,
"grad_norm": 0.35545430825067814,
"learning_rate": 1.6300940438871475e-05,
"loss": 0.3858,
"step": 260
},
{
"epoch": 0.3324968632371393,
"grad_norm": 0.39047195961342007,
"learning_rate": 1.6614420062695925e-05,
"loss": 0.3956,
"step": 265
},
{
"epoch": 0.33877038895859474,
"grad_norm": 0.3763861614570858,
"learning_rate": 1.6927899686520378e-05,
"loss": 0.3863,
"step": 270
},
{
"epoch": 0.3450439146800502,
"grad_norm": 0.33810866516589266,
"learning_rate": 1.7241379310344828e-05,
"loss": 0.3861,
"step": 275
},
{
"epoch": 0.35131744040150564,
"grad_norm": 0.3577598772376036,
"learning_rate": 1.7554858934169282e-05,
"loss": 0.3847,
"step": 280
},
{
"epoch": 0.3575909661229611,
"grad_norm": 0.39952196485063435,
"learning_rate": 1.7868338557993732e-05,
"loss": 0.3803,
"step": 285
},
{
"epoch": 0.36386449184441655,
"grad_norm": 0.3560924326294842,
"learning_rate": 1.8181818181818182e-05,
"loss": 0.3823,
"step": 290
},
{
"epoch": 0.370138017565872,
"grad_norm": 0.36592070219456535,
"learning_rate": 1.8495297805642636e-05,
"loss": 0.3877,
"step": 295
},
{
"epoch": 0.37641154328732745,
"grad_norm": 0.36996538529023604,
"learning_rate": 1.8808777429467086e-05,
"loss": 0.3816,
"step": 300
},
{
"epoch": 0.38268506900878296,
"grad_norm": 0.3944353625018826,
"learning_rate": 1.9122257053291536e-05,
"loss": 0.3885,
"step": 305
},
{
"epoch": 0.3889585947302384,
"grad_norm": 0.3605090389245764,
"learning_rate": 1.943573667711599e-05,
"loss": 0.3746,
"step": 310
},
{
"epoch": 0.39523212045169387,
"grad_norm": 0.3489216159243111,
"learning_rate": 1.9749216300940443e-05,
"loss": 0.3807,
"step": 315
},
{
"epoch": 0.4015056461731493,
"grad_norm": 0.3902867659960602,
"learning_rate": 1.9999994004731887e-05,
"loss": 0.394,
"step": 320
},
{
"epoch": 0.4077791718946048,
"grad_norm": 0.3833754079678544,
"learning_rate": 1.999978417110275e-05,
"loss": 0.3657,
"step": 325
},
{
"epoch": 0.41405269761606023,
"grad_norm": 0.37800625764548235,
"learning_rate": 1.9999274581256576e-05,
"loss": 0.3822,
"step": 330
},
{
"epoch": 0.4203262233375157,
"grad_norm": 0.3548880184290815,
"learning_rate": 1.999846525046898e-05,
"loss": 0.3816,
"step": 335
},
{
"epoch": 0.42659974905897113,
"grad_norm": 0.37094441921423654,
"learning_rate": 1.9997356203000667e-05,
"loss": 0.3853,
"step": 340
},
{
"epoch": 0.4328732747804266,
"grad_norm": 0.3594535908155001,
"learning_rate": 1.9995947472096752e-05,
"loss": 0.3703,
"step": 345
},
{
"epoch": 0.43914680050188204,
"grad_norm": 0.3492383976696482,
"learning_rate": 1.9994239099985727e-05,
"loss": 0.3858,
"step": 350
},
{
"epoch": 0.4454203262233375,
"grad_norm": 0.3545945901168298,
"learning_rate": 1.9992231137878213e-05,
"loss": 0.3723,
"step": 355
},
{
"epoch": 0.451693851944793,
"grad_norm": 0.36080708632510694,
"learning_rate": 1.9989923645965418e-05,
"loss": 0.3952,
"step": 360
},
{
"epoch": 0.45796737766624845,
"grad_norm": 0.5493517294545232,
"learning_rate": 1.998731669341735e-05,
"loss": 0.3723,
"step": 365
},
{
"epoch": 0.4642409033877039,
"grad_norm": 0.34501637664109114,
"learning_rate": 1.998441035838071e-05,
"loss": 0.3787,
"step": 370
},
{
"epoch": 0.47051442910915936,
"grad_norm": 0.36834831423716424,
"learning_rate": 1.9981204727976577e-05,
"loss": 0.3871,
"step": 375
},
{
"epoch": 0.4767879548306148,
"grad_norm": 0.359064185739471,
"learning_rate": 1.9977699898297794e-05,
"loss": 0.4078,
"step": 380
},
{
"epoch": 0.48306148055207027,
"grad_norm": 0.37676963093518556,
"learning_rate": 1.997389597440608e-05,
"loss": 0.3997,
"step": 385
},
{
"epoch": 0.4893350062735257,
"grad_norm": 0.35164360932787275,
"learning_rate": 1.9969793070328872e-05,
"loss": 0.3706,
"step": 390
},
{
"epoch": 0.49560853199498117,
"grad_norm": 0.3499090374786232,
"learning_rate": 1.996539130905593e-05,
"loss": 0.3931,
"step": 395
},
{
"epoch": 0.5018820577164367,
"grad_norm": 0.35516374008901963,
"learning_rate": 1.9960690822535632e-05,
"loss": 0.3917,
"step": 400
},
{
"epoch": 0.5081555834378921,
"grad_norm": 0.3859915521260531,
"learning_rate": 1.995569175167102e-05,
"loss": 0.3862,
"step": 405
},
{
"epoch": 0.5144291091593476,
"grad_norm": 0.3700726542619639,
"learning_rate": 1.9950394246315594e-05,
"loss": 0.3977,
"step": 410
},
{
"epoch": 0.520702634880803,
"grad_norm": 0.36807219365873434,
"learning_rate": 1.994479846526879e-05,
"loss": 0.391,
"step": 415
},
{
"epoch": 0.5269761606022585,
"grad_norm": 0.34834015226308745,
"learning_rate": 1.9938904576271247e-05,
"loss": 0.3947,
"step": 420
},
{
"epoch": 0.533249686323714,
"grad_norm": 0.3475144346489988,
"learning_rate": 1.9932712755999768e-05,
"loss": 0.3797,
"step": 425
},
{
"epoch": 0.5395232120451694,
"grad_norm": 0.3471973955561176,
"learning_rate": 1.9926223190062015e-05,
"loss": 0.3777,
"step": 430
},
{
"epoch": 0.5457967377666249,
"grad_norm": 0.3702513132489804,
"learning_rate": 1.9919436072990967e-05,
"loss": 0.4113,
"step": 435
},
{
"epoch": 0.5520702634880803,
"grad_norm": 0.3480545734033953,
"learning_rate": 1.9912351608239064e-05,
"loss": 0.3717,
"step": 440
},
{
"epoch": 0.5583437892095358,
"grad_norm": 0.3705010339965962,
"learning_rate": 1.9904970008172128e-05,
"loss": 0.3903,
"step": 445
},
{
"epoch": 0.5646173149309912,
"grad_norm": 0.3545524157937066,
"learning_rate": 1.989729149406298e-05,
"loss": 0.3971,
"step": 450
},
{
"epoch": 0.5708908406524467,
"grad_norm": 0.34489507859345525,
"learning_rate": 1.988931629608483e-05,
"loss": 0.3715,
"step": 455
},
{
"epoch": 0.5771643663739021,
"grad_norm": 0.3419133499625096,
"learning_rate": 1.9881044653304347e-05,
"loss": 0.3722,
"step": 460
},
{
"epoch": 0.5834378920953576,
"grad_norm": 0.3544487799327636,
"learning_rate": 1.9872476813674527e-05,
"loss": 0.3854,
"step": 465
},
{
"epoch": 0.589711417816813,
"grad_norm": 0.3526030038053533,
"learning_rate": 1.9863613034027224e-05,
"loss": 0.3854,
"step": 470
},
{
"epoch": 0.5959849435382685,
"grad_norm": 0.35801287965649553,
"learning_rate": 1.9854453580065485e-05,
"loss": 0.3773,
"step": 475
},
{
"epoch": 0.6022584692597239,
"grad_norm": 0.33687873507306343,
"learning_rate": 1.984499872635556e-05,
"loss": 0.3839,
"step": 480
},
{
"epoch": 0.6085319949811794,
"grad_norm": 0.34289605388996236,
"learning_rate": 1.983524875631868e-05,
"loss": 0.3899,
"step": 485
},
{
"epoch": 0.6148055207026348,
"grad_norm": 0.35193056323984084,
"learning_rate": 1.9825203962222573e-05,
"loss": 0.3736,
"step": 490
},
{
"epoch": 0.6210790464240903,
"grad_norm": 0.3603036239481184,
"learning_rate": 1.9814864645172684e-05,
"loss": 0.3927,
"step": 495
},
{
"epoch": 0.6273525721455459,
"grad_norm": 0.3610916261894253,
"learning_rate": 1.9804231115103155e-05,
"loss": 0.3729,
"step": 500
},
{
"epoch": 0.6336260978670013,
"grad_norm": 0.33695606360007346,
"learning_rate": 1.9793303690767543e-05,
"loss": 0.3773,
"step": 505
},
{
"epoch": 0.6398996235884568,
"grad_norm": 0.37708359246844253,
"learning_rate": 1.9782082699729255e-05,
"loss": 0.3915,
"step": 510
},
{
"epoch": 0.6461731493099122,
"grad_norm": 0.37063002060189804,
"learning_rate": 1.9770568478351736e-05,
"loss": 0.384,
"step": 515
},
{
"epoch": 0.6524466750313677,
"grad_norm": 0.347781428498976,
"learning_rate": 1.9758761371788376e-05,
"loss": 0.3912,
"step": 520
},
{
"epoch": 0.6587202007528231,
"grad_norm": 0.37037259025304425,
"learning_rate": 1.974666173397218e-05,
"loss": 0.3795,
"step": 525
},
{
"epoch": 0.6649937264742786,
"grad_norm": 0.35088852677024873,
"learning_rate": 1.9734269927605134e-05,
"loss": 0.3803,
"step": 530
},
{
"epoch": 0.671267252195734,
"grad_norm": 0.3569377164202312,
"learning_rate": 1.972158632414736e-05,
"loss": 0.3775,
"step": 535
},
{
"epoch": 0.6775407779171895,
"grad_norm": 0.3394929535259636,
"learning_rate": 1.970861130380596e-05,
"loss": 0.3741,
"step": 540
},
{
"epoch": 0.6838143036386449,
"grad_norm": 0.3420677685134694,
"learning_rate": 1.9695345255523634e-05,
"loss": 0.3878,
"step": 545
},
{
"epoch": 0.6900878293601004,
"grad_norm": 0.3531313507208391,
"learning_rate": 1.9681788576967004e-05,
"loss": 0.3787,
"step": 550
},
{
"epoch": 0.6963613550815558,
"grad_norm": 0.338822289408876,
"learning_rate": 1.9667941674514712e-05,
"loss": 0.3829,
"step": 555
},
{
"epoch": 0.7026348808030113,
"grad_norm": 0.36931664322135854,
"learning_rate": 1.9653804963245226e-05,
"loss": 0.3893,
"step": 560
},
{
"epoch": 0.7089084065244667,
"grad_norm": 0.35793185710348313,
"learning_rate": 1.9639378866924405e-05,
"loss": 0.3816,
"step": 565
},
{
"epoch": 0.7151819322459222,
"grad_norm": 0.36642001426471366,
"learning_rate": 1.9624663817992783e-05,
"loss": 0.3825,
"step": 570
},
{
"epoch": 0.7214554579673776,
"grad_norm": 0.36184277756969885,
"learning_rate": 1.960966025755262e-05,
"loss": 0.3918,
"step": 575
},
{
"epoch": 0.7277289836888331,
"grad_norm": 0.34191036136824954,
"learning_rate": 1.9594368635354676e-05,
"loss": 0.3812,
"step": 580
},
{
"epoch": 0.7340025094102886,
"grad_norm": 0.4053922705018229,
"learning_rate": 1.9578789409784727e-05,
"loss": 0.3796,
"step": 585
},
{
"epoch": 0.740276035131744,
"grad_norm": 0.3254681140813354,
"learning_rate": 1.9562923047849828e-05,
"loss": 0.381,
"step": 590
},
{
"epoch": 0.7465495608531995,
"grad_norm": 0.3655079090417533,
"learning_rate": 1.9546770025164304e-05,
"loss": 0.3808,
"step": 595
},
{
"epoch": 0.7528230865746549,
"grad_norm": 0.3309145310004598,
"learning_rate": 1.95303308259355e-05,
"loss": 0.3924,
"step": 600
},
{
"epoch": 0.7590966122961104,
"grad_norm": 0.34826624551002194,
"learning_rate": 1.9513605942949277e-05,
"loss": 0.3752,
"step": 605
},
{
"epoch": 0.7653701380175659,
"grad_norm": 0.3463792668326179,
"learning_rate": 1.9496595877555212e-05,
"loss": 0.3922,
"step": 610
},
{
"epoch": 0.7716436637390214,
"grad_norm": 0.3409526655488519,
"learning_rate": 1.94793011396516e-05,
"loss": 0.3868,
"step": 615
},
{
"epoch": 0.7779171894604768,
"grad_norm": 0.34614309257239506,
"learning_rate": 1.946172224767015e-05,
"loss": 0.3816,
"step": 620
},
{
"epoch": 0.7841907151819323,
"grad_norm": 0.36738342533490437,
"learning_rate": 1.9443859728560458e-05,
"loss": 0.3809,
"step": 625
},
{
"epoch": 0.7904642409033877,
"grad_norm": 0.5642673777402893,
"learning_rate": 1.9425714117774183e-05,
"loss": 0.3678,
"step": 630
},
{
"epoch": 0.7967377666248432,
"grad_norm": 0.3734297840619315,
"learning_rate": 1.940728595924904e-05,
"loss": 0.3856,
"step": 635
},
{
"epoch": 0.8030112923462986,
"grad_norm": 0.34992964702526325,
"learning_rate": 1.9388575805392453e-05,
"loss": 0.3937,
"step": 640
},
{
"epoch": 0.8092848180677541,
"grad_norm": 0.34532043284186725,
"learning_rate": 1.9369584217065025e-05,
"loss": 0.3718,
"step": 645
},
{
"epoch": 0.8155583437892095,
"grad_norm": 0.3390467838397306,
"learning_rate": 1.935031176356371e-05,
"loss": 0.3829,
"step": 650
},
{
"epoch": 0.821831869510665,
"grad_norm": 0.3522889151490647,
"learning_rate": 1.933075902260475e-05,
"loss": 0.3876,
"step": 655
},
{
"epoch": 0.8281053952321205,
"grad_norm": 0.38123310023732215,
"learning_rate": 1.9310926580306365e-05,
"loss": 0.3707,
"step": 660
},
{
"epoch": 0.8343789209535759,
"grad_norm": 0.37101425687357464,
"learning_rate": 1.929081503117117e-05,
"loss": 0.3777,
"step": 665
},
{
"epoch": 0.8406524466750314,
"grad_norm": 0.35327596368808917,
"learning_rate": 1.9270424978068368e-05,
"loss": 0.372,
"step": 670
},
{
"epoch": 0.8469259723964868,
"grad_norm": 0.48025305553666425,
"learning_rate": 1.9249757032215674e-05,
"loss": 0.3719,
"step": 675
},
{
"epoch": 0.8531994981179423,
"grad_norm": 1.3954653921865106,
"learning_rate": 1.9228811813160972e-05,
"loss": 0.3774,
"step": 680
},
{
"epoch": 0.8594730238393977,
"grad_norm": 0.3729066926513616,
"learning_rate": 1.920758994876379e-05,
"loss": 0.3845,
"step": 685
},
{
"epoch": 0.8657465495608532,
"grad_norm": 0.3368507345057043,
"learning_rate": 1.918609207517643e-05,
"loss": 0.3674,
"step": 690
},
{
"epoch": 0.8720200752823086,
"grad_norm": 0.3381829661628343,
"learning_rate": 1.9164318836824928e-05,
"loss": 0.3895,
"step": 695
},
{
"epoch": 0.8782936010037641,
"grad_norm": 0.35406742056379137,
"learning_rate": 1.9142270886389726e-05,
"loss": 0.3888,
"step": 700
},
{
"epoch": 0.8845671267252195,
"grad_norm": 0.35267605322252205,
"learning_rate": 1.911994888478611e-05,
"loss": 0.416,
"step": 705
},
{
"epoch": 0.890840652446675,
"grad_norm": 0.36994927588198245,
"learning_rate": 1.9097353501144403e-05,
"loss": 0.3881,
"step": 710
},
{
"epoch": 0.8971141781681304,
"grad_norm": 0.344878616221491,
"learning_rate": 1.9074485412789886e-05,
"loss": 0.3916,
"step": 715
},
{
"epoch": 0.903387703889586,
"grad_norm": 0.3618364967642679,
"learning_rate": 1.9051345305222527e-05,
"loss": 0.3877,
"step": 720
},
{
"epoch": 0.9096612296110415,
"grad_norm": 0.3348128268620291,
"learning_rate": 1.9027933872096403e-05,
"loss": 0.3758,
"step": 725
},
{
"epoch": 0.9159347553324969,
"grad_norm": 0.34960639615133043,
"learning_rate": 1.900425181519893e-05,
"loss": 0.3842,
"step": 730
},
{
"epoch": 0.9222082810539524,
"grad_norm": 0.36222991570969465,
"learning_rate": 1.8980299844429804e-05,
"loss": 0.379,
"step": 735
},
{
"epoch": 0.9284818067754078,
"grad_norm": 0.3499002387438987,
"learning_rate": 1.8956078677779738e-05,
"loss": 0.3715,
"step": 740
},
{
"epoch": 0.9347553324968633,
"grad_norm": 0.33211819623666405,
"learning_rate": 1.8931589041308926e-05,
"loss": 0.3736,
"step": 745
},
{
"epoch": 0.9410288582183187,
"grad_norm": 0.3495894811325005,
"learning_rate": 1.8906831669125293e-05,
"loss": 0.3778,
"step": 750
},
{
"epoch": 0.9473023839397742,
"grad_norm": 0.3315684966056081,
"learning_rate": 1.8881807303362484e-05,
"loss": 0.3776,
"step": 755
},
{
"epoch": 0.9535759096612296,
"grad_norm": 0.3409622621170755,
"learning_rate": 1.885651669415761e-05,
"loss": 0.3873,
"step": 760
},
{
"epoch": 0.9598494353826851,
"grad_norm": 0.34785709596472625,
"learning_rate": 1.883096059962876e-05,
"loss": 0.3861,
"step": 765
},
{
"epoch": 0.9661229611041405,
"grad_norm": 0.3477615933807258,
"learning_rate": 1.8805139785852297e-05,
"loss": 0.3783,
"step": 770
},
{
"epoch": 0.972396486825596,
"grad_norm": 0.343324271874741,
"learning_rate": 1.877905502683987e-05,
"loss": 0.3746,
"step": 775
},
{
"epoch": 0.9786700125470514,
"grad_norm": 0.47223667640674677,
"learning_rate": 1.8752707104515223e-05,
"loss": 0.3793,
"step": 780
},
{
"epoch": 0.9849435382685069,
"grad_norm": 0.3234411234996811,
"learning_rate": 1.8726096808690757e-05,
"loss": 0.3676,
"step": 785
},
{
"epoch": 0.9912170639899623,
"grad_norm": 0.3574211042049705,
"learning_rate": 1.8699224937043846e-05,
"loss": 0.3735,
"step": 790
},
{
"epoch": 0.9974905897114178,
"grad_norm": 0.3811194221638418,
"learning_rate": 1.8672092295092935e-05,
"loss": 0.3939,
"step": 795
},
{
"epoch": 1.0,
"eval_loss": 0.3792824149131775,
"eval_runtime": 2.8922,
"eval_samples_per_second": 12.101,
"eval_steps_per_second": 0.692,
"step": 797
},
{
"epoch": 1.0037641154328734,
"grad_norm": 0.3353538294927158,
"learning_rate": 1.8644699696173393e-05,
"loss": 0.3241,
"step": 800
},
{
"epoch": 1.0100376411543288,
"grad_norm": 0.3828496951284579,
"learning_rate": 1.8617047961413122e-05,
"loss": 0.2601,
"step": 805
},
{
"epoch": 1.0163111668757843,
"grad_norm": 0.33996880676819174,
"learning_rate": 1.858913791970795e-05,
"loss": 0.2689,
"step": 810
},
{
"epoch": 1.0225846925972397,
"grad_norm": 0.3787602779680391,
"learning_rate": 1.8560970407696787e-05,
"loss": 0.2686,
"step": 815
},
{
"epoch": 1.0288582183186952,
"grad_norm": 0.5196175862012924,
"learning_rate": 1.8532546269736546e-05,
"loss": 0.2747,
"step": 820
},
{
"epoch": 1.0351317440401506,
"grad_norm": 0.37332309268508856,
"learning_rate": 1.850386635787682e-05,
"loss": 0.2627,
"step": 825
},
{
"epoch": 1.041405269761606,
"grad_norm": 0.3471559800133524,
"learning_rate": 1.847493153183435e-05,
"loss": 0.2787,
"step": 830
},
{
"epoch": 1.0476787954830615,
"grad_norm": 0.3558613761958236,
"learning_rate": 1.844574265896726e-05,
"loss": 0.268,
"step": 835
},
{
"epoch": 1.053952321204517,
"grad_norm": 0.3816351874674652,
"learning_rate": 1.8416300614249044e-05,
"loss": 0.2668,
"step": 840
},
{
"epoch": 1.0602258469259724,
"grad_norm": 0.34373084787867114,
"learning_rate": 1.8386606280242342e-05,
"loss": 0.2648,
"step": 845
},
{
"epoch": 1.066499372647428,
"grad_norm": 0.34772061503647006,
"learning_rate": 1.8356660547072493e-05,
"loss": 0.2664,
"step": 850
},
{
"epoch": 1.0727728983688833,
"grad_norm": 0.3575051030279196,
"learning_rate": 1.8326464312400835e-05,
"loss": 0.2741,
"step": 855
},
{
"epoch": 1.0790464240903388,
"grad_norm": 0.3496509974703566,
"learning_rate": 1.8296018481397818e-05,
"loss": 0.2583,
"step": 860
},
{
"epoch": 1.0853199498117942,
"grad_norm": 0.3724719074552806,
"learning_rate": 1.826532396671585e-05,
"loss": 0.2694,
"step": 865
},
{
"epoch": 1.0915934755332497,
"grad_norm": 0.3727323786558387,
"learning_rate": 1.8234381688461943e-05,
"loss": 0.2772,
"step": 870
},
{
"epoch": 1.0978670012547052,
"grad_norm": 0.3335956714818362,
"learning_rate": 1.8203192574170154e-05,
"loss": 0.2563,
"step": 875
},
{
"epoch": 1.1041405269761606,
"grad_norm": 0.34610791981119976,
"learning_rate": 1.8171757558773747e-05,
"loss": 0.254,
"step": 880
},
{
"epoch": 1.110414052697616,
"grad_norm": 0.3647153604437883,
"learning_rate": 1.8140077584577193e-05,
"loss": 0.269,
"step": 885
},
{
"epoch": 1.1166875784190715,
"grad_norm": 0.3620894127576232,
"learning_rate": 1.81081536012279e-05,
"loss": 0.2642,
"step": 890
},
{
"epoch": 1.122961104140527,
"grad_norm": 0.3368502269678017,
"learning_rate": 1.8075986565687785e-05,
"loss": 0.2621,
"step": 895
},
{
"epoch": 1.1292346298619824,
"grad_norm": 0.3487017808626127,
"learning_rate": 1.804357744220454e-05,
"loss": 0.2741,
"step": 900
},
{
"epoch": 1.1355081555834379,
"grad_norm": 0.346569504128483,
"learning_rate": 1.8010927202282758e-05,
"loss": 0.2522,
"step": 905
},
{
"epoch": 1.1417816813048933,
"grad_norm": 0.3448349298442695,
"learning_rate": 1.7978036824654806e-05,
"loss": 0.2539,
"step": 910
},
{
"epoch": 1.1480552070263488,
"grad_norm": 0.35055993507582917,
"learning_rate": 1.7944907295251478e-05,
"loss": 0.2716,
"step": 915
},
{
"epoch": 1.1543287327478042,
"grad_norm": 0.3483298654120021,
"learning_rate": 1.7911539607172447e-05,
"loss": 0.2585,
"step": 920
},
{
"epoch": 1.1606022584692597,
"grad_norm": 0.3452159026039351,
"learning_rate": 1.78779347606565e-05,
"loss": 0.2598,
"step": 925
},
{
"epoch": 1.1668757841907151,
"grad_norm": 0.3434243523852655,
"learning_rate": 1.7844093763051543e-05,
"loss": 0.2681,
"step": 930
},
{
"epoch": 1.1731493099121706,
"grad_norm": 0.3459951917334934,
"learning_rate": 1.7810017628784416e-05,
"loss": 0.2567,
"step": 935
},
{
"epoch": 1.179422835633626,
"grad_norm": 0.31955125292674175,
"learning_rate": 1.777570737933047e-05,
"loss": 0.2673,
"step": 940
},
{
"epoch": 1.1856963613550815,
"grad_norm": 0.33191052346485606,
"learning_rate": 1.7741164043182967e-05,
"loss": 0.258,
"step": 945
},
{
"epoch": 1.191969887076537,
"grad_norm": 0.3745301206766351,
"learning_rate": 1.7706388655822223e-05,
"loss": 0.2671,
"step": 950
},
{
"epoch": 1.1982434127979924,
"grad_norm": 0.3435343557827377,
"learning_rate": 1.7671382259684603e-05,
"loss": 0.2711,
"step": 955
},
{
"epoch": 1.2045169385194479,
"grad_norm": 0.3522241286327592,
"learning_rate": 1.7636145904131233e-05,
"loss": 0.2715,
"step": 960
},
{
"epoch": 1.2107904642409033,
"grad_norm": 0.34428731270476376,
"learning_rate": 1.7600680645416583e-05,
"loss": 0.2655,
"step": 965
},
{
"epoch": 1.2170639899623588,
"grad_norm": 0.3534701552438621,
"learning_rate": 1.7564987546656778e-05,
"loss": 0.2601,
"step": 970
},
{
"epoch": 1.2233375156838142,
"grad_norm": 0.34480580113424486,
"learning_rate": 1.7529067677797727e-05,
"loss": 0.2581,
"step": 975
},
{
"epoch": 1.2296110414052697,
"grad_norm": 0.3741756431765012,
"learning_rate": 1.7492922115583077e-05,
"loss": 0.2701,
"step": 980
},
{
"epoch": 1.2358845671267251,
"grad_norm": 0.3555549239846533,
"learning_rate": 1.745655194352191e-05,
"loss": 0.2716,
"step": 985
},
{
"epoch": 1.2421580928481806,
"grad_norm": 0.3327711036535926,
"learning_rate": 1.7419958251856276e-05,
"loss": 0.2577,
"step": 990
},
{
"epoch": 1.248431618569636,
"grad_norm": 0.3661107928778811,
"learning_rate": 1.738314213752851e-05,
"loss": 0.2649,
"step": 995
},
{
"epoch": 1.2547051442910915,
"grad_norm": 0.3407939311803759,
"learning_rate": 1.7346104704148343e-05,
"loss": 0.2506,
"step": 1000
},
{
"epoch": 1.260978670012547,
"grad_norm": 0.33612826697533044,
"learning_rate": 1.730884706195983e-05,
"loss": 0.2645,
"step": 1005
},
{
"epoch": 1.2672521957340024,
"grad_norm": 0.34109027432250294,
"learning_rate": 1.727137032780807e-05,
"loss": 0.2687,
"step": 1010
},
{
"epoch": 1.2735257214554578,
"grad_norm": 0.3424345987544216,
"learning_rate": 1.7233675625105703e-05,
"loss": 0.2659,
"step": 1015
},
{
"epoch": 1.2797992471769133,
"grad_norm": 0.3670304260632612,
"learning_rate": 1.7195764083799277e-05,
"loss": 0.2785,
"step": 1020
},
{
"epoch": 1.286072772898369,
"grad_norm": 0.35224545856472056,
"learning_rate": 1.7157636840335334e-05,
"loss": 0.2736,
"step": 1025
},
{
"epoch": 1.2923462986198244,
"grad_norm": 0.8557471051222927,
"learning_rate": 1.7119295037626366e-05,
"loss": 0.2598,
"step": 1030
},
{
"epoch": 1.2986198243412799,
"grad_norm": 0.3367737302829996,
"learning_rate": 1.708073982501656e-05,
"loss": 0.2612,
"step": 1035
},
{
"epoch": 1.3048933500627353,
"grad_norm": 0.34877985553107826,
"learning_rate": 1.704197235824732e-05,
"loss": 0.2726,
"step": 1040
},
{
"epoch": 1.3111668757841908,
"grad_norm": 0.34440291584591926,
"learning_rate": 1.7002993799422652e-05,
"loss": 0.2618,
"step": 1045
},
{
"epoch": 1.3174404015056462,
"grad_norm": 0.33534085525712676,
"learning_rate": 1.6963805316974303e-05,
"loss": 0.2609,
"step": 1050
},
{
"epoch": 1.3237139272271017,
"grad_norm": 0.3662577678235449,
"learning_rate": 1.6924408085626756e-05,
"loss": 0.2571,
"step": 1055
},
{
"epoch": 1.3299874529485571,
"grad_norm": 0.3516887590807691,
"learning_rate": 1.6884803286362e-05,
"loss": 0.2549,
"step": 1060
},
{
"epoch": 1.3362609786700126,
"grad_norm": 0.35333946604034366,
"learning_rate": 1.684499210638414e-05,
"loss": 0.264,
"step": 1065
},
{
"epoch": 1.342534504391468,
"grad_norm": 0.34307627027241056,
"learning_rate": 1.6804975739083803e-05,
"loss": 0.2503,
"step": 1070
},
{
"epoch": 1.3488080301129235,
"grad_norm": 0.35493913250542247,
"learning_rate": 1.6764755384002372e-05,
"loss": 0.2759,
"step": 1075
},
{
"epoch": 1.355081555834379,
"grad_norm": 0.3540415537021871,
"learning_rate": 1.6724332246796008e-05,
"loss": 0.2697,
"step": 1080
},
{
"epoch": 1.3613550815558344,
"grad_norm": 0.3318252959324338,
"learning_rate": 1.6683707539199538e-05,
"loss": 0.2669,
"step": 1085
},
{
"epoch": 1.3676286072772899,
"grad_norm": 0.33137640825273385,
"learning_rate": 1.6642882478990112e-05,
"loss": 0.2485,
"step": 1090
},
{
"epoch": 1.3739021329987453,
"grad_norm": 0.34017928191383223,
"learning_rate": 1.66018582899507e-05,
"loss": 0.2784,
"step": 1095
},
{
"epoch": 1.3801756587202008,
"grad_norm": 0.34200366455572445,
"learning_rate": 1.6560636201833423e-05,
"loss": 0.2673,
"step": 1100
},
{
"epoch": 1.3864491844416562,
"grad_norm": 0.35384418379159516,
"learning_rate": 1.6519217450322657e-05,
"loss": 0.2713,
"step": 1105
},
{
"epoch": 1.3927227101631117,
"grad_norm": 0.33307624664228463,
"learning_rate": 1.6477603276998037e-05,
"loss": 0.2742,
"step": 1110
},
{
"epoch": 1.3989962358845671,
"grad_norm": 0.356765357000532,
"learning_rate": 1.64357949292972e-05,
"loss": 0.2689,
"step": 1115
},
{
"epoch": 1.4052697616060226,
"grad_norm": 0.3252207580977864,
"learning_rate": 1.6393793660478406e-05,
"loss": 0.2506,
"step": 1120
},
{
"epoch": 1.411543287327478,
"grad_norm": 0.36204092721369197,
"learning_rate": 1.6351600729582977e-05,
"loss": 0.2636,
"step": 1125
},
{
"epoch": 1.4178168130489335,
"grad_norm": 0.3335102584738542,
"learning_rate": 1.630921740139755e-05,
"loss": 0.2616,
"step": 1130
},
{
"epoch": 1.424090338770389,
"grad_norm": 0.34898006017841243,
"learning_rate": 1.6266644946416148e-05,
"loss": 0.2781,
"step": 1135
},
{
"epoch": 1.4303638644918444,
"grad_norm": 0.3487474238464629,
"learning_rate": 1.622388464080213e-05,
"loss": 0.2773,
"step": 1140
},
{
"epoch": 1.4366373902132998,
"grad_norm": 0.35273437538491903,
"learning_rate": 1.61809377663499e-05,
"loss": 0.2682,
"step": 1145
},
{
"epoch": 1.4429109159347553,
"grad_norm": 0.36466350484277693,
"learning_rate": 1.6137805610446508e-05,
"loss": 0.2685,
"step": 1150
},
{
"epoch": 1.4491844416562107,
"grad_norm": 0.35090011471222154,
"learning_rate": 1.609448946603304e-05,
"loss": 0.2657,
"step": 1155
},
{
"epoch": 1.4554579673776662,
"grad_norm": 0.36713922458350784,
"learning_rate": 1.6050990631565894e-05,
"loss": 0.276,
"step": 1160
},
{
"epoch": 1.4617314930991216,
"grad_norm": 0.3552048435134842,
"learning_rate": 1.6007310410977807e-05,
"loss": 0.2796,
"step": 1165
},
{
"epoch": 1.468005018820577,
"grad_norm": 0.3469345931045424,
"learning_rate": 1.5963450113638815e-05,
"loss": 0.2592,
"step": 1170
},
{
"epoch": 1.4742785445420326,
"grad_norm": 2.7973818683619016,
"learning_rate": 1.5919411054316966e-05,
"loss": 0.2727,
"step": 1175
},
{
"epoch": 1.480552070263488,
"grad_norm": 0.3477691166092876,
"learning_rate": 1.5875194553138942e-05,
"loss": 0.2708,
"step": 1180
},
{
"epoch": 1.4868255959849435,
"grad_norm": 0.42047181326852134,
"learning_rate": 1.5830801935550462e-05,
"loss": 0.2583,
"step": 1185
},
{
"epoch": 1.4930991217063991,
"grad_norm": 0.38351145606864295,
"learning_rate": 1.5786234532276555e-05,
"loss": 0.2665,
"step": 1190
},
{
"epoch": 1.4993726474278546,
"grad_norm": 0.33888116221517656,
"learning_rate": 1.574149367928168e-05,
"loss": 0.2788,
"step": 1195
},
{
"epoch": 1.50564617314931,
"grad_norm": 0.3533659596857954,
"learning_rate": 1.5696580717729665e-05,
"loss": 0.2709,
"step": 1200
},
{
"epoch": 1.5119196988707655,
"grad_norm": 0.33797242592368726,
"learning_rate": 1.5651496993943507e-05,
"loss": 0.2552,
"step": 1205
},
{
"epoch": 1.518193224592221,
"grad_norm": 0.3380805624100092,
"learning_rate": 1.5606243859365033e-05,
"loss": 0.2696,
"step": 1210
},
{
"epoch": 1.5244667503136764,
"grad_norm": 0.3589011909382504,
"learning_rate": 1.5560822670514356e-05,
"loss": 0.2729,
"step": 1215
},
{
"epoch": 1.5307402760351319,
"grad_norm": 0.3410908318740735,
"learning_rate": 1.5515234788949238e-05,
"loss": 0.2658,
"step": 1220
},
{
"epoch": 1.5370138017565873,
"grad_norm": 0.4071357427518539,
"learning_rate": 1.5469481581224274e-05,
"loss": 0.2618,
"step": 1225
},
{
"epoch": 1.5432873274780428,
"grad_norm": 0.4050327080992723,
"learning_rate": 1.5423564418849895e-05,
"loss": 0.2707,
"step": 1230
},
{
"epoch": 1.5495608531994982,
"grad_norm": 0.3512739607697075,
"learning_rate": 1.537748467825131e-05,
"loss": 0.2762,
"step": 1235
},
{
"epoch": 1.5558343789209537,
"grad_norm": 0.3548946907435895,
"learning_rate": 1.5331243740727203e-05,
"loss": 0.285,
"step": 1240
},
{
"epoch": 1.5621079046424091,
"grad_norm": 0.3241833124220892,
"learning_rate": 1.5284842992408336e-05,
"loss": 0.2675,
"step": 1245
},
{
"epoch": 1.5683814303638646,
"grad_norm": 0.35085102130243395,
"learning_rate": 1.5238283824216015e-05,
"loss": 0.2681,
"step": 1250
},
{
"epoch": 1.57465495608532,
"grad_norm": 0.37148590291911643,
"learning_rate": 1.5191567631820364e-05,
"loss": 0.27,
"step": 1255
},
{
"epoch": 1.5809284818067755,
"grad_norm": 0.3535604634586656,
"learning_rate": 1.5144695815598529e-05,
"loss": 0.2717,
"step": 1260
},
{
"epoch": 1.587202007528231,
"grad_norm": 0.34794195830922975,
"learning_rate": 1.5097669780592658e-05,
"loss": 0.2633,
"step": 1265
},
{
"epoch": 1.5934755332496864,
"grad_norm": 0.35489225458342305,
"learning_rate": 1.5050490936467814e-05,
"loss": 0.2735,
"step": 1270
},
{
"epoch": 1.5997490589711418,
"grad_norm": 0.3350637434620735,
"learning_rate": 1.5003160697469707e-05,
"loss": 0.2544,
"step": 1275
},
{
"epoch": 1.6060225846925973,
"grad_norm": 0.3381089856973841,
"learning_rate": 1.4955680482382296e-05,
"loss": 0.2564,
"step": 1280
},
{
"epoch": 1.6122961104140527,
"grad_norm": 0.32684634964616277,
"learning_rate": 1.4908051714485266e-05,
"loss": 0.2741,
"step": 1285
},
{
"epoch": 1.6185696361355082,
"grad_norm": 0.34645698891886967,
"learning_rate": 1.4860275821511359e-05,
"loss": 0.2748,
"step": 1290
},
{
"epoch": 1.6248431618569636,
"grad_norm": 0.36961906780329834,
"learning_rate": 1.481235423560358e-05,
"loss": 0.2721,
"step": 1295
},
{
"epoch": 1.631116687578419,
"grad_norm": 0.350579115602874,
"learning_rate": 1.4764288393272258e-05,
"loss": 0.2628,
"step": 1300
},
{
"epoch": 1.6373902132998746,
"grad_norm": 0.3306477114399784,
"learning_rate": 1.4716079735352006e-05,
"loss": 0.2729,
"step": 1305
},
{
"epoch": 1.64366373902133,
"grad_norm": 0.35455700175873195,
"learning_rate": 1.46677297069585e-05,
"loss": 0.2667,
"step": 1310
},
{
"epoch": 1.6499372647427855,
"grad_norm": 0.33847253281006606,
"learning_rate": 1.4619239757445187e-05,
"loss": 0.2706,
"step": 1315
},
{
"epoch": 1.656210790464241,
"grad_norm": 0.34327567130216446,
"learning_rate": 1.4570611340359821e-05,
"loss": 0.266,
"step": 1320
},
{
"epoch": 1.6624843161856964,
"grad_norm": 0.3557992340297897,
"learning_rate": 1.4521845913400891e-05,
"loss": 0.2746,
"step": 1325
},
{
"epoch": 1.6687578419071518,
"grad_norm": 0.35121604621554686,
"learning_rate": 1.4472944938373945e-05,
"loss": 0.2704,
"step": 1330
},
{
"epoch": 1.6750313676286073,
"grad_norm": 0.34165359487510566,
"learning_rate": 1.4423909881147747e-05,
"loss": 0.2692,
"step": 1335
},
{
"epoch": 1.6813048933500627,
"grad_norm": 0.36505578592627197,
"learning_rate": 1.4374742211610345e-05,
"loss": 0.2662,
"step": 1340
},
{
"epoch": 1.6875784190715182,
"grad_norm": 0.34422086687232467,
"learning_rate": 1.4325443403625012e-05,
"loss": 0.275,
"step": 1345
},
{
"epoch": 1.6938519447929736,
"grad_norm": 0.3675735392039838,
"learning_rate": 1.4276014934986064e-05,
"loss": 0.272,
"step": 1350
},
{
"epoch": 1.700125470514429,
"grad_norm": 0.35389447351847136,
"learning_rate": 1.4226458287374555e-05,
"loss": 0.2713,
"step": 1355
},
{
"epoch": 1.7063989962358845,
"grad_norm": 0.3379229270723559,
"learning_rate": 1.4176774946313872e-05,
"loss": 0.2625,
"step": 1360
},
{
"epoch": 1.71267252195734,
"grad_norm": 0.34402002879314064,
"learning_rate": 1.4126966401125189e-05,
"loss": 0.268,
"step": 1365
},
{
"epoch": 1.7189460476787954,
"grad_norm": 0.3532344899080162,
"learning_rate": 1.4077034144882843e-05,
"loss": 0.2632,
"step": 1370
},
{
"epoch": 1.725219573400251,
"grad_norm": 0.3401295622140909,
"learning_rate": 1.4026979674369566e-05,
"loss": 0.2613,
"step": 1375
},
{
"epoch": 1.7314930991217063,
"grad_norm": 0.3391840532185442,
"learning_rate": 1.3976804490031608e-05,
"loss": 0.2719,
"step": 1380
},
{
"epoch": 1.7377666248431618,
"grad_norm": 0.3427490028776178,
"learning_rate": 1.3926510095933781e-05,
"loss": 0.2692,
"step": 1385
},
{
"epoch": 1.7440401505646173,
"grad_norm": 0.3433287489294571,
"learning_rate": 1.387609799971435e-05,
"loss": 0.2649,
"step": 1390
},
{
"epoch": 1.7503136762860727,
"grad_norm": 0.33114909580993174,
"learning_rate": 1.3825569712539864e-05,
"loss": 0.2527,
"step": 1395
},
{
"epoch": 1.7565872020075282,
"grad_norm": 0.337515925683474,
"learning_rate": 1.3774926749059826e-05,
"loss": 0.2556,
"step": 1400
},
{
"epoch": 1.7628607277289836,
"grad_norm": 0.3419996547297016,
"learning_rate": 1.3724170627361323e-05,
"loss": 0.2638,
"step": 1405
},
{
"epoch": 1.769134253450439,
"grad_norm": 0.35073376743994084,
"learning_rate": 1.3673302868923491e-05,
"loss": 0.2704,
"step": 1410
},
{
"epoch": 1.7754077791718945,
"grad_norm": 0.32861646593191174,
"learning_rate": 1.3622324998571928e-05,
"loss": 0.2519,
"step": 1415
},
{
"epoch": 1.78168130489335,
"grad_norm": 0.33312714617584277,
"learning_rate": 1.3571238544432968e-05,
"loss": 0.2664,
"step": 1420
},
{
"epoch": 1.7879548306148054,
"grad_norm": 0.3374614354621205,
"learning_rate": 1.352004503788789e-05,
"loss": 0.2585,
"step": 1425
},
{
"epoch": 1.7942283563362609,
"grad_norm": 0.3554543246581463,
"learning_rate": 1.3468746013527e-05,
"loss": 0.2762,
"step": 1430
},
{
"epoch": 1.8005018820577163,
"grad_norm": 0.3367124035830617,
"learning_rate": 1.3417343009103634e-05,
"loss": 0.261,
"step": 1435
},
{
"epoch": 1.8067754077791718,
"grad_norm": 0.3181211071389625,
"learning_rate": 1.3365837565488065e-05,
"loss": 0.2715,
"step": 1440
},
{
"epoch": 1.8130489335006272,
"grad_norm": 0.3480040032313692,
"learning_rate": 1.3314231226621305e-05,
"loss": 0.2624,
"step": 1445
},
{
"epoch": 1.8193224592220827,
"grad_norm": 0.3416171971995866,
"learning_rate": 1.3262525539468839e-05,
"loss": 0.2642,
"step": 1450
},
{
"epoch": 1.8255959849435381,
"grad_norm": 0.3635819899278629,
"learning_rate": 1.3210722053974233e-05,
"loss": 0.2632,
"step": 1455
},
{
"epoch": 1.8318695106649936,
"grad_norm": 0.3265652400172599,
"learning_rate": 1.315882232301269e-05,
"loss": 0.2612,
"step": 1460
},
{
"epoch": 1.838143036386449,
"grad_norm": 0.3342794882432052,
"learning_rate": 1.3106827902344485e-05,
"loss": 0.2623,
"step": 1465
},
{
"epoch": 1.8444165621079045,
"grad_norm": 0.3347520583742969,
"learning_rate": 1.3054740350568346e-05,
"loss": 0.2741,
"step": 1470
},
{
"epoch": 1.85069008782936,
"grad_norm": 0.34328068285452285,
"learning_rate": 1.3002561229074719e-05,
"loss": 0.2561,
"step": 1475
},
{
"epoch": 1.8569636135508154,
"grad_norm": 0.35131941587904497,
"learning_rate": 1.2950292101998967e-05,
"loss": 0.2747,
"step": 1480
},
{
"epoch": 1.8632371392722709,
"grad_norm": 0.34360731854818805,
"learning_rate": 1.289793453617449e-05,
"loss": 0.2627,
"step": 1485
},
{
"epoch": 1.8695106649937263,
"grad_norm": 0.3498923646707763,
"learning_rate": 1.2845490101085744e-05,
"loss": 0.2562,
"step": 1490
},
{
"epoch": 1.875784190715182,
"grad_norm": 0.34701974294822086,
"learning_rate": 1.2792960368821212e-05,
"loss": 0.265,
"step": 1495
},
{
"epoch": 1.8820577164366374,
"grad_norm": 0.33841922800891855,
"learning_rate": 1.2740346914026258e-05,
"loss": 0.2638,
"step": 1500
},
{
"epoch": 1.888331242158093,
"grad_norm": 0.3375511420369947,
"learning_rate": 1.2687651313855937e-05,
"loss": 0.2589,
"step": 1505
},
{
"epoch": 1.8946047678795483,
"grad_norm": 0.34124342881268466,
"learning_rate": 1.2634875147927726e-05,
"loss": 0.2689,
"step": 1510
},
{
"epoch": 1.9008782936010038,
"grad_norm": 0.3492720225961315,
"learning_rate": 1.2582019998274142e-05,
"loss": 0.2619,
"step": 1515
},
{
"epoch": 1.9071518193224593,
"grad_norm": 0.3424592345393382,
"learning_rate": 1.252908744929536e-05,
"loss": 0.2673,
"step": 1520
},
{
"epoch": 1.9134253450439147,
"grad_norm": 0.3541786605023589,
"learning_rate": 1.2476079087711695e-05,
"loss": 0.2741,
"step": 1525
},
{
"epoch": 1.9196988707653702,
"grad_norm": 0.36100779817450435,
"learning_rate": 1.2422996502516023e-05,
"loss": 0.2708,
"step": 1530
},
{
"epoch": 1.9259723964868256,
"grad_norm": 0.3637900870051361,
"learning_rate": 1.236984128492619e-05,
"loss": 0.2679,
"step": 1535
},
{
"epoch": 1.932245922208281,
"grad_norm": 0.34654925833715405,
"learning_rate": 1.231661502833728e-05,
"loss": 0.2705,
"step": 1540
},
{
"epoch": 1.9385194479297365,
"grad_norm": 0.3555711761052598,
"learning_rate": 1.2263319328273853e-05,
"loss": 0.2732,
"step": 1545
},
{
"epoch": 1.944792973651192,
"grad_norm": 0.4361336719233523,
"learning_rate": 1.220995578234214e-05,
"loss": 0.2818,
"step": 1550
},
{
"epoch": 1.9510664993726474,
"grad_norm": 0.35818262065748885,
"learning_rate": 1.2156525990182132e-05,
"loss": 0.2714,
"step": 1555
},
{
"epoch": 1.9573400250941029,
"grad_norm": 0.34020072227504516,
"learning_rate": 1.2103031553419629e-05,
"loss": 0.2561,
"step": 1560
},
{
"epoch": 1.9636135508155583,
"grad_norm": 0.35424378525712236,
"learning_rate": 1.2049474075618244e-05,
"loss": 0.2817,
"step": 1565
},
{
"epoch": 1.9698870765370138,
"grad_norm": 0.3436811391936569,
"learning_rate": 1.1995855162231323e-05,
"loss": 0.2727,
"step": 1570
},
{
"epoch": 1.9761606022584692,
"grad_norm": 0.35547373104319596,
"learning_rate": 1.1942176420553817e-05,
"loss": 0.279,
"step": 1575
},
{
"epoch": 1.9824341279799247,
"grad_norm": 0.33151574599317196,
"learning_rate": 1.1888439459674107e-05,
"loss": 0.2736,
"step": 1580
},
{
"epoch": 1.9887076537013801,
"grad_norm": 0.3489287650284772,
"learning_rate": 1.1834645890425773e-05,
"loss": 0.2674,
"step": 1585
},
{
"epoch": 1.9949811794228356,
"grad_norm": 0.4674786217571983,
"learning_rate": 1.1780797325339301e-05,
"loss": 0.2618,
"step": 1590
},
{
"epoch": 2.0,
"eval_loss": 0.3876406252384186,
"eval_runtime": 2.3724,
"eval_samples_per_second": 14.753,
"eval_steps_per_second": 0.843,
"step": 1594
},
{
"epoch": 2.0012547051442913,
"grad_norm": 0.41826837808172157,
"learning_rate": 1.1726895378593745e-05,
"loss": 0.2453,
"step": 1595
},
{
"epoch": 2.0075282308657467,
"grad_norm": 0.3896113439351613,
"learning_rate": 1.167294166596834e-05,
"loss": 0.1307,
"step": 1600
},
{
"epoch": 2.013801756587202,
"grad_norm": 0.40723627183822325,
"learning_rate": 1.1618937804794077e-05,
"loss": 0.1253,
"step": 1605
},
{
"epoch": 2.0200752823086576,
"grad_norm": 0.3393249627107008,
"learning_rate": 1.1564885413905205e-05,
"loss": 0.1212,
"step": 1610
},
{
"epoch": 2.026348808030113,
"grad_norm": 0.3969867667512431,
"learning_rate": 1.1510786113590715e-05,
"loss": 0.1213,
"step": 1615
},
{
"epoch": 2.0326223337515685,
"grad_norm": 0.3807098036418188,
"learning_rate": 1.1456641525545768e-05,
"loss": 0.115,
"step": 1620
},
{
"epoch": 2.038895859473024,
"grad_norm": 0.3600283222530161,
"learning_rate": 1.1402453272823086e-05,
"loss": 0.1178,
"step": 1625
},
{
"epoch": 2.0451693851944794,
"grad_norm": 0.37082981681871713,
"learning_rate": 1.1348222979784289e-05,
"loss": 0.1186,
"step": 1630
},
{
"epoch": 2.051442910915935,
"grad_norm": 0.3701722986939684,
"learning_rate": 1.1293952272051217e-05,
"loss": 0.1161,
"step": 1635
},
{
"epoch": 2.0577164366373903,
"grad_norm": 0.39437640184582917,
"learning_rate": 1.1239642776457176e-05,
"loss": 0.112,
"step": 1640
},
{
"epoch": 2.063989962358846,
"grad_norm": 0.35512823472089206,
"learning_rate": 1.1185296120998208e-05,
"loss": 0.1227,
"step": 1645
},
{
"epoch": 2.0702634880803013,
"grad_norm": 0.39699134768151145,
"learning_rate": 1.1130913934784255e-05,
"loss": 0.118,
"step": 1650
},
{
"epoch": 2.0765370138017567,
"grad_norm": 0.361766879756225,
"learning_rate": 1.107649784799034e-05,
"loss": 0.1148,
"step": 1655
},
{
"epoch": 2.082810539523212,
"grad_norm": 0.37983521046428353,
"learning_rate": 1.1022049491807703e-05,
"loss": 0.1105,
"step": 1660
},
{
"epoch": 2.0890840652446676,
"grad_norm": 0.36791365726333974,
"learning_rate": 1.0967570498394895e-05,
"loss": 0.1197,
"step": 1665
},
{
"epoch": 2.095357590966123,
"grad_norm": 0.38778258880907535,
"learning_rate": 1.0913062500828865e-05,
"loss": 0.119,
"step": 1670
},
{
"epoch": 2.1016311166875785,
"grad_norm": 0.3686039497467697,
"learning_rate": 1.0858527133055994e-05,
"loss": 0.1197,
"step": 1675
},
{
"epoch": 2.107904642409034,
"grad_norm": 0.39330229406582323,
"learning_rate": 1.0803966029843114e-05,
"loss": 0.1166,
"step": 1680
},
{
"epoch": 2.1141781681304894,
"grad_norm": 0.3636181831711105,
"learning_rate": 1.0749380826728513e-05,
"loss": 0.1133,
"step": 1685
},
{
"epoch": 2.120451693851945,
"grad_norm": 0.4335397800674325,
"learning_rate": 1.0694773159972912e-05,
"loss": 0.1246,
"step": 1690
},
{
"epoch": 2.1267252195734003,
"grad_norm": 0.37508400928061725,
"learning_rate": 1.0640144666510392e-05,
"loss": 0.1196,
"step": 1695
},
{
"epoch": 2.132998745294856,
"grad_norm": 0.3987416537308343,
"learning_rate": 1.0585496983899361e-05,
"loss": 0.1226,
"step": 1700
},
{
"epoch": 2.1392722710163112,
"grad_norm": 0.38791702283560353,
"learning_rate": 1.0530831750273428e-05,
"loss": 0.1117,
"step": 1705
},
{
"epoch": 2.1455457967377667,
"grad_norm": 0.39592521931999036,
"learning_rate": 1.0476150604292329e-05,
"loss": 0.1198,
"step": 1710
},
{
"epoch": 2.151819322459222,
"grad_norm": 0.6134906074452066,
"learning_rate": 1.0421455185092784e-05,
"loss": 0.1168,
"step": 1715
},
{
"epoch": 2.1580928481806776,
"grad_norm": 0.3936394784460519,
"learning_rate": 1.0366747132239374e-05,
"loss": 0.1137,
"step": 1720
},
{
"epoch": 2.164366373902133,
"grad_norm": 0.38023062505112215,
"learning_rate": 1.0312028085675393e-05,
"loss": 0.1216,
"step": 1725
},
{
"epoch": 2.1706398996235885,
"grad_norm": 0.3879080250933175,
"learning_rate": 1.025729968567368e-05,
"loss": 0.1163,
"step": 1730
},
{
"epoch": 2.176913425345044,
"grad_norm": 0.37494689918032786,
"learning_rate": 1.0202563572787457e-05,
"loss": 0.1155,
"step": 1735
},
{
"epoch": 2.1831869510664994,
"grad_norm": 0.4170219240353852,
"learning_rate": 1.0147821387801154e-05,
"loss": 0.1231,
"step": 1740
},
{
"epoch": 2.189460476787955,
"grad_norm": 0.36447068742427746,
"learning_rate": 1.0093074771681214e-05,
"loss": 0.1173,
"step": 1745
},
{
"epoch": 2.1957340025094103,
"grad_norm": 0.37383691981995226,
"learning_rate": 1.003832536552691e-05,
"loss": 0.1181,
"step": 1750
},
{
"epoch": 2.2020075282308658,
"grad_norm": 0.35361007854482546,
"learning_rate": 9.983574810521151e-06,
"loss": 0.1141,
"step": 1755
},
{
"epoch": 2.208281053952321,
"grad_norm": 0.37593622722746173,
"learning_rate": 9.928824747881286e-06,
"loss": 0.117,
"step": 1760
},
{
"epoch": 2.2145545796737767,
"grad_norm": 0.38527811994324745,
"learning_rate": 9.874076818809903e-06,
"loss": 0.1222,
"step": 1765
},
{
"epoch": 2.220828105395232,
"grad_norm": 0.37756703878021675,
"learning_rate": 9.81933266444563e-06,
"loss": 0.117,
"step": 1770
},
{
"epoch": 2.2271016311166876,
"grad_norm": 0.3981056432095895,
"learning_rate": 9.76459392581395e-06,
"loss": 0.1187,
"step": 1775
},
{
"epoch": 2.233375156838143,
"grad_norm": 0.35779965724307555,
"learning_rate": 9.709862243777998e-06,
"loss": 0.1201,
"step": 1780
},
{
"epoch": 2.2396486825595985,
"grad_norm": 0.39287403460106407,
"learning_rate": 9.655139258989379e-06,
"loss": 0.1173,
"step": 1785
},
{
"epoch": 2.245922208281054,
"grad_norm": 0.39081524580807464,
"learning_rate": 9.60042661183899e-06,
"loss": 0.114,
"step": 1790
},
{
"epoch": 2.2521957340025094,
"grad_norm": 0.5348108468458116,
"learning_rate": 9.54572594240784e-06,
"loss": 0.1145,
"step": 1795
},
{
"epoch": 2.258469259723965,
"grad_norm": 0.35615572604956347,
"learning_rate": 9.491038890417894e-06,
"loss": 0.1128,
"step": 1800
},
{
"epoch": 2.2647427854454203,
"grad_norm": 0.37107476709616843,
"learning_rate": 9.436367095182916e-06,
"loss": 0.1228,
"step": 1805
},
{
"epoch": 2.2710163111668757,
"grad_norm": 0.38323250415960275,
"learning_rate": 9.381712195559324e-06,
"loss": 0.118,
"step": 1810
},
{
"epoch": 2.277289836888331,
"grad_norm": 0.359171122780413,
"learning_rate": 9.327075829897082e-06,
"loss": 0.1191,
"step": 1815
},
{
"epoch": 2.2835633626097867,
"grad_norm": 0.4114654516418914,
"learning_rate": 9.272459635990563e-06,
"loss": 0.1235,
"step": 1820
},
{
"epoch": 2.289836888331242,
"grad_norm": 0.39460304180293915,
"learning_rate": 9.217865251029469e-06,
"loss": 0.1187,
"step": 1825
},
{
"epoch": 2.2961104140526976,
"grad_norm": 0.35839055639361983,
"learning_rate": 9.163294311549753e-06,
"loss": 0.1156,
"step": 1830
},
{
"epoch": 2.302383939774153,
"grad_norm": 0.3482054803738314,
"learning_rate": 9.108748453384559e-06,
"loss": 0.1198,
"step": 1835
},
{
"epoch": 2.3086574654956085,
"grad_norm": 0.3820940219983755,
"learning_rate": 9.054229311615178e-06,
"loss": 0.117,
"step": 1840
},
{
"epoch": 2.314930991217064,
"grad_norm": 0.39037798204086893,
"learning_rate": 8.999738520522065e-06,
"loss": 0.1197,
"step": 1845
},
{
"epoch": 2.3212045169385194,
"grad_norm": 0.3822304512858218,
"learning_rate": 8.945277713535809e-06,
"loss": 0.1152,
"step": 1850
},
{
"epoch": 2.327478042659975,
"grad_norm": 0.3965998438998078,
"learning_rate": 8.890848523188192e-06,
"loss": 0.1243,
"step": 1855
},
{
"epoch": 2.3337515683814303,
"grad_norm": 0.3939416904284715,
"learning_rate": 8.836452581063248e-06,
"loss": 0.1195,
"step": 1860
},
{
"epoch": 2.3400250941028857,
"grad_norm": 0.4276348064874773,
"learning_rate": 8.78209151774835e-06,
"loss": 0.1211,
"step": 1865
},
{
"epoch": 2.346298619824341,
"grad_norm": 0.39140182898284753,
"learning_rate": 8.727766962785344e-06,
"loss": 0.1157,
"step": 1870
},
{
"epoch": 2.3525721455457966,
"grad_norm": 0.3634468198734603,
"learning_rate": 8.673480544621681e-06,
"loss": 0.1129,
"step": 1875
},
{
"epoch": 2.358845671267252,
"grad_norm": 0.39837989743789176,
"learning_rate": 8.61923389056162e-06,
"loss": 0.1198,
"step": 1880
},
{
"epoch": 2.3651191969887075,
"grad_norm": 0.3935841490044898,
"learning_rate": 8.565028626717435e-06,
"loss": 0.1203,
"step": 1885
},
{
"epoch": 2.371392722710163,
"grad_norm": 0.35305962598333074,
"learning_rate": 8.51086637796068e-06,
"loss": 0.1128,
"step": 1890
},
{
"epoch": 2.3776662484316184,
"grad_norm": 0.394308865525823,
"learning_rate": 8.456748767873474e-06,
"loss": 0.1124,
"step": 1895
},
{
"epoch": 2.383939774153074,
"grad_norm": 0.3841761354621664,
"learning_rate": 8.402677418699842e-06,
"loss": 0.1145,
"step": 1900
},
{
"epoch": 2.3902132998745294,
"grad_norm": 0.3893473191716482,
"learning_rate": 8.34865395129707e-06,
"loss": 0.1197,
"step": 1905
},
{
"epoch": 2.396486825595985,
"grad_norm": 0.42534674714643167,
"learning_rate": 8.294679985087137e-06,
"loss": 0.1179,
"step": 1910
},
{
"epoch": 2.4027603513174403,
"grad_norm": 0.40229677875453496,
"learning_rate": 8.240757138008149e-06,
"loss": 0.1236,
"step": 1915
},
{
"epoch": 2.4090338770388957,
"grad_norm": 0.37978448174191587,
"learning_rate": 8.186887026465857e-06,
"loss": 0.1125,
"step": 1920
},
{
"epoch": 2.415307402760351,
"grad_norm": 0.38090023973889275,
"learning_rate": 8.133071265285209e-06,
"loss": 0.1175,
"step": 1925
},
{
"epoch": 2.4215809284818066,
"grad_norm": 0.3709724709852035,
"learning_rate": 8.079311467661912e-06,
"loss": 0.1189,
"step": 1930
},
{
"epoch": 2.427854454203262,
"grad_norm": 0.3968859415543936,
"learning_rate": 8.025609245114107e-06,
"loss": 0.1208,
"step": 1935
},
{
"epoch": 2.4341279799247175,
"grad_norm": 0.38140929315858313,
"learning_rate": 7.971966207434045e-06,
"loss": 0.1167,
"step": 1940
},
{
"epoch": 2.440401505646173,
"grad_norm": 0.3813844728323988,
"learning_rate": 7.918383962639835e-06,
"loss": 0.1186,
"step": 1945
},
{
"epoch": 2.4466750313676284,
"grad_norm": 0.36925458542907064,
"learning_rate": 7.864864116927245e-06,
"loss": 0.1167,
"step": 1950
},
{
"epoch": 2.452948557089084,
"grad_norm": 0.3813928431538188,
"learning_rate": 7.811408274621549e-06,
"loss": 0.1217,
"step": 1955
},
{
"epoch": 2.4592220828105393,
"grad_norm": 0.38045397571366496,
"learning_rate": 7.75801803812944e-06,
"loss": 0.1176,
"step": 1960
},
{
"epoch": 2.4654956085319952,
"grad_norm": 0.409917244408148,
"learning_rate": 7.704695007890988e-06,
"loss": 0.1214,
"step": 1965
},
{
"epoch": 2.4717691342534502,
"grad_norm": 0.39881220216006136,
"learning_rate": 7.651440782331679e-06,
"loss": 0.1176,
"step": 1970
},
{
"epoch": 2.478042659974906,
"grad_norm": 0.3551138970811604,
"learning_rate": 7.598256957814479e-06,
"loss": 0.1156,
"step": 1975
},
{
"epoch": 2.484316185696361,
"grad_norm": 0.38209748278035194,
"learning_rate": 7.545145128592009e-06,
"loss": 0.1128,
"step": 1980
},
{
"epoch": 2.490589711417817,
"grad_norm": 0.38676641594007305,
"learning_rate": 7.49210688675873e-06,
"loss": 0.1175,
"step": 1985
},
{
"epoch": 2.496863237139272,
"grad_norm": 0.36950074837730973,
"learning_rate": 7.4391438222032265e-06,
"loss": 0.1139,
"step": 1990
},
{
"epoch": 2.503136762860728,
"grad_norm": 0.3915583584835345,
"learning_rate": 7.3862575225605535e-06,
"loss": 0.1179,
"step": 1995
},
{
"epoch": 2.509410288582183,
"grad_norm": 0.3866601171887957,
"learning_rate": 7.333449573164634e-06,
"loss": 0.1207,
"step": 2000
},
{
"epoch": 2.515683814303639,
"grad_norm": 0.4110664967201194,
"learning_rate": 7.280721557000759e-06,
"loss": 0.1166,
"step": 2005
},
{
"epoch": 2.521957340025094,
"grad_norm": 0.37778473055073203,
"learning_rate": 7.228075054658096e-06,
"loss": 0.1157,
"step": 2010
},
{
"epoch": 2.5282308657465498,
"grad_norm": 0.37225323631681123,
"learning_rate": 7.175511644282349e-06,
"loss": 0.1156,
"step": 2015
},
{
"epoch": 2.5345043914680048,
"grad_norm": 0.3824521057716352,
"learning_rate": 7.123032901528431e-06,
"loss": 0.1182,
"step": 2020
},
{
"epoch": 2.5407779171894607,
"grad_norm": 0.38115325746292966,
"learning_rate": 7.070640399513232e-06,
"loss": 0.1158,
"step": 2025
},
{
"epoch": 2.5470514429109157,
"grad_norm": 0.3822411406686136,
"learning_rate": 7.018335708768467e-06,
"loss": 0.1177,
"step": 2030
},
{
"epoch": 2.5533249686323716,
"grad_norm": 0.407300076312883,
"learning_rate": 6.966120397193605e-06,
"loss": 0.1152,
"step": 2035
},
{
"epoch": 2.5595984943538266,
"grad_norm": 0.3666367383521563,
"learning_rate": 6.913996030008853e-06,
"loss": 0.1153,
"step": 2040
},
{
"epoch": 2.5658720200752825,
"grad_norm": 0.3632368895081262,
"learning_rate": 6.861964169708245e-06,
"loss": 0.1107,
"step": 2045
},
{
"epoch": 2.572145545796738,
"grad_norm": 0.3649524530723953,
"learning_rate": 6.810026376012808e-06,
"loss": 0.1145,
"step": 2050
},
{
"epoch": 2.5784190715181934,
"grad_norm": 0.39293153907868017,
"learning_rate": 6.758184205823791e-06,
"loss": 0.1162,
"step": 2055
},
{
"epoch": 2.584692597239649,
"grad_norm": 0.3931749546166204,
"learning_rate": 6.706439213176028e-06,
"loss": 0.1132,
"step": 2060
},
{
"epoch": 2.5909661229611043,
"grad_norm": 0.3825377326068989,
"learning_rate": 6.654792949191317e-06,
"loss": 0.1222,
"step": 2065
},
{
"epoch": 2.5972396486825597,
"grad_norm": 0.35192545209688325,
"learning_rate": 6.603246962031942e-06,
"loss": 0.112,
"step": 2070
},
{
"epoch": 2.603513174404015,
"grad_norm": 0.3638182336052127,
"learning_rate": 6.551802796854265e-06,
"loss": 0.1144,
"step": 2075
},
{
"epoch": 2.6097867001254706,
"grad_norm": 0.38711833706267534,
"learning_rate": 6.500461995762402e-06,
"loss": 0.1133,
"step": 2080
},
{
"epoch": 2.616060225846926,
"grad_norm": 0.3897188834137444,
"learning_rate": 6.449226097762e-06,
"loss": 0.123,
"step": 2085
},
{
"epoch": 2.6223337515683816,
"grad_norm": 0.37823736524426615,
"learning_rate": 6.398096638714106e-06,
"loss": 0.1184,
"step": 2090
},
{
"epoch": 2.628607277289837,
"grad_norm": 0.3932949556705336,
"learning_rate": 6.34707515128912e-06,
"loss": 0.1143,
"step": 2095
},
{
"epoch": 2.6348808030112925,
"grad_norm": 0.3884831444296721,
"learning_rate": 6.296163164920858e-06,
"loss": 0.1113,
"step": 2100
},
{
"epoch": 2.641154328732748,
"grad_norm": 0.36382200051340396,
"learning_rate": 6.245362205760703e-06,
"loss": 0.1159,
"step": 2105
},
{
"epoch": 2.6474278544542034,
"grad_norm": 0.3924419492733963,
"learning_rate": 6.194673796631852e-06,
"loss": 0.109,
"step": 2110
},
{
"epoch": 2.653701380175659,
"grad_norm": 0.3947013863716826,
"learning_rate": 6.144099456983681e-06,
"loss": 0.1115,
"step": 2115
},
{
"epoch": 2.6599749058971143,
"grad_norm": 0.3943261219862062,
"learning_rate": 6.093640702846182e-06,
"loss": 0.1122,
"step": 2120
},
{
"epoch": 2.6662484316185697,
"grad_norm": 0.39005774295522977,
"learning_rate": 6.043299046784526e-06,
"loss": 0.1187,
"step": 2125
},
{
"epoch": 2.672521957340025,
"grad_norm": 0.37092235530502005,
"learning_rate": 5.993075997853719e-06,
"loss": 0.1149,
"step": 2130
},
{
"epoch": 2.6787954830614806,
"grad_norm": 0.3807422171419007,
"learning_rate": 5.94297306155337e-06,
"loss": 0.1135,
"step": 2135
},
{
"epoch": 2.685069008782936,
"grad_norm": 0.3792169573613968,
"learning_rate": 5.892991739782557e-06,
"loss": 0.1199,
"step": 2140
},
{
"epoch": 2.6913425345043915,
"grad_norm": 0.3831811436833811,
"learning_rate": 5.843133530794817e-06,
"loss": 0.1096,
"step": 2145
},
{
"epoch": 2.697616060225847,
"grad_norm": 0.37933754782226464,
"learning_rate": 5.793399929153216e-06,
"loss": 0.1106,
"step": 2150
},
{
"epoch": 2.7038895859473024,
"grad_norm": 0.38366361518402914,
"learning_rate": 5.743792425685554e-06,
"loss": 0.1154,
"step": 2155
},
{
"epoch": 2.710163111668758,
"grad_norm": 0.3631788769456335,
"learning_rate": 5.694312507439691e-06,
"loss": 0.1141,
"step": 2160
},
{
"epoch": 2.7164366373902133,
"grad_norm": 0.39467438603923816,
"learning_rate": 5.644961657638942e-06,
"loss": 0.1148,
"step": 2165
},
{
"epoch": 2.722710163111669,
"grad_norm": 0.3643220647785576,
"learning_rate": 5.595741355637645e-06,
"loss": 0.1098,
"step": 2170
},
{
"epoch": 2.7289836888331243,
"grad_norm": 0.39291715101415214,
"learning_rate": 5.5466530768768005e-06,
"loss": 0.1159,
"step": 2175
},
{
"epoch": 2.7352572145545797,
"grad_norm": 0.40345238378734466,
"learning_rate": 5.497698292839835e-06,
"loss": 0.1117,
"step": 2180
},
{
"epoch": 2.741530740276035,
"grad_norm": 0.41110337094460886,
"learning_rate": 5.448878471008513e-06,
"loss": 0.1134,
"step": 2185
},
{
"epoch": 2.7478042659974906,
"grad_norm": 0.4006579187318595,
"learning_rate": 5.400195074818924e-06,
"loss": 0.1228,
"step": 2190
},
{
"epoch": 2.754077791718946,
"grad_norm": 0.4006920463713775,
"learning_rate": 5.351649563617638e-06,
"loss": 0.1157,
"step": 2195
},
{
"epoch": 2.7603513174404015,
"grad_norm": 0.40311348122015783,
"learning_rate": 5.3032433926179395e-06,
"loss": 0.1211,
"step": 2200
},
{
"epoch": 2.766624843161857,
"grad_norm": 0.3833185052390898,
"learning_rate": 5.25497801285622e-06,
"loss": 0.1085,
"step": 2205
},
{
"epoch": 2.7728983688833124,
"grad_norm": 0.40698791566770903,
"learning_rate": 5.206854871148466e-06,
"loss": 0.1145,
"step": 2210
},
{
"epoch": 2.779171894604768,
"grad_norm": 0.37018390800544043,
"learning_rate": 5.158875410046906e-06,
"loss": 0.1148,
"step": 2215
},
{
"epoch": 2.7854454203262233,
"grad_norm": 0.4097654897566281,
"learning_rate": 5.111041067796754e-06,
"loss": 0.1112,
"step": 2220
},
{
"epoch": 2.791718946047679,
"grad_norm": 0.38185615087945834,
"learning_rate": 5.063353278293106e-06,
"loss": 0.1129,
"step": 2225
},
{
"epoch": 2.7979924717691342,
"grad_norm": 0.38860623269143496,
"learning_rate": 5.0158134710379595e-06,
"loss": 0.1157,
"step": 2230
},
{
"epoch": 2.8042659974905897,
"grad_norm": 0.41677106161054384,
"learning_rate": 4.9684230710973394e-06,
"loss": 0.1179,
"step": 2235
},
{
"epoch": 2.810539523212045,
"grad_norm": 0.3659456249492505,
"learning_rate": 4.921183499058615e-06,
"loss": 0.1179,
"step": 2240
},
{
"epoch": 2.8168130489335006,
"grad_norm": 0.390728899484992,
"learning_rate": 4.8740961709878834e-06,
"loss": 0.1118,
"step": 2245
},
{
"epoch": 2.823086574654956,
"grad_norm": 0.41326006522454833,
"learning_rate": 4.827162498387544e-06,
"loss": 0.1123,
"step": 2250
},
{
"epoch": 2.8293601003764115,
"grad_norm": 0.37982873419881247,
"learning_rate": 4.780383888153983e-06,
"loss": 0.1099,
"step": 2255
},
{
"epoch": 2.835633626097867,
"grad_norm": 0.3826347732319729,
"learning_rate": 4.733761742535381e-06,
"loss": 0.1119,
"step": 2260
},
{
"epoch": 2.8419071518193224,
"grad_norm": 0.4070056837379538,
"learning_rate": 4.687297459089708e-06,
"loss": 0.1169,
"step": 2265
},
{
"epoch": 2.848180677540778,
"grad_norm": 0.3907501846052804,
"learning_rate": 4.640992430642801e-06,
"loss": 0.1184,
"step": 2270
},
{
"epoch": 2.8544542032622333,
"grad_norm": 0.3980879608103747,
"learning_rate": 4.594848045246638e-06,
"loss": 0.1143,
"step": 2275
},
{
"epoch": 2.8607277289836888,
"grad_norm": 0.3831724940346972,
"learning_rate": 4.548865686137718e-06,
"loss": 0.1165,
"step": 2280
},
{
"epoch": 2.867001254705144,
"grad_norm": 0.3554795717929537,
"learning_rate": 4.503046731695584e-06,
"loss": 0.1103,
"step": 2285
},
{
"epoch": 2.8732747804265997,
"grad_norm": 0.37945026941432614,
"learning_rate": 4.457392555401531e-06,
"loss": 0.1165,
"step": 2290
},
{
"epoch": 2.879548306148055,
"grad_norm": 0.4040384960083639,
"learning_rate": 4.411904525797408e-06,
"loss": 0.112,
"step": 2295
},
{
"epoch": 2.8858218318695106,
"grad_norm": 0.37374435050142246,
"learning_rate": 4.3665840064446165e-06,
"loss": 0.1152,
"step": 2300
},
{
"epoch": 2.892095357590966,
"grad_norm": 0.38395743236739355,
"learning_rate": 4.321432355883219e-06,
"loss": 0.1158,
"step": 2305
},
{
"epoch": 2.8983688833124215,
"grad_norm": 0.38107246526906685,
"learning_rate": 4.276450927591229e-06,
"loss": 0.1099,
"step": 2310
},
{
"epoch": 2.904642409033877,
"grad_norm": 0.40288534078286153,
"learning_rate": 4.231641069944019e-06,
"loss": 0.1135,
"step": 2315
},
{
"epoch": 2.9109159347553324,
"grad_norm": 0.39407625180559624,
"learning_rate": 4.187004126173928e-06,
"loss": 0.1153,
"step": 2320
},
{
"epoch": 2.917189460476788,
"grad_norm": 0.390694853717115,
"learning_rate": 4.1425414343299734e-06,
"loss": 0.1113,
"step": 2325
},
{
"epoch": 2.9234629861982433,
"grad_norm": 0.37505537186817833,
"learning_rate": 4.098254327237742e-06,
"loss": 0.1107,
"step": 2330
},
{
"epoch": 2.9297365119196987,
"grad_norm": 0.3797569810287486,
"learning_rate": 4.054144132459471e-06,
"loss": 0.1125,
"step": 2335
},
{
"epoch": 2.936010037641154,
"grad_norm": 0.39039171406836526,
"learning_rate": 4.010212172254201e-06,
"loss": 0.1139,
"step": 2340
},
{
"epoch": 2.9422835633626097,
"grad_norm": 0.3735230020383592,
"learning_rate": 3.966459763538179e-06,
"loss": 0.1162,
"step": 2345
},
{
"epoch": 2.948557089084065,
"grad_norm": 0.3799822252359316,
"learning_rate": 3.92288821784536e-06,
"loss": 0.1157,
"step": 2350
},
{
"epoch": 2.9548306148055206,
"grad_norm": 0.37435313221883065,
"learning_rate": 3.879498841288105e-06,
"loss": 0.1109,
"step": 2355
},
{
"epoch": 2.961104140526976,
"grad_norm": 0.37834252836846144,
"learning_rate": 3.836292934518029e-06,
"loss": 0.1124,
"step": 2360
},
{
"epoch": 2.9673776662484315,
"grad_norm": 0.3755100276116184,
"learning_rate": 3.793271792686993e-06,
"loss": 0.1122,
"step": 2365
},
{
"epoch": 2.973651191969887,
"grad_norm": 0.38213190237560674,
"learning_rate": 3.750436705408311e-06,
"loss": 0.1131,
"step": 2370
},
{
"epoch": 2.9799247176913424,
"grad_norm": 0.39502326617269706,
"learning_rate": 3.7077889567180625e-06,
"loss": 0.113,
"step": 2375
},
{
"epoch": 2.9861982434127983,
"grad_norm": 0.3907005868892978,
"learning_rate": 3.6653298250366265e-06,
"loss": 0.1121,
"step": 2380
},
{
"epoch": 2.9924717691342533,
"grad_norm": 0.3757187294889032,
"learning_rate": 3.6230605831303354e-06,
"loss": 0.1138,
"step": 2385
},
{
"epoch": 2.998745294855709,
"grad_norm": 0.36519480053180337,
"learning_rate": 3.5809824980733445e-06,
"loss": 0.1141,
"step": 2390
},
{
"epoch": 3.0,
"eval_loss": 0.43099531531333923,
"eval_runtime": 2.3647,
"eval_samples_per_second": 14.801,
"eval_steps_per_second": 0.846,
"step": 2391
},
{
"epoch": 3.005018820577164,
"grad_norm": 0.2442197751144262,
"learning_rate": 3.5390968312096396e-06,
"loss": 0.0533,
"step": 2395
},
{
"epoch": 3.0112923462986196,
"grad_norm": 0.21528572471833773,
"learning_rate": 3.497404838115219e-06,
"loss": 0.0415,
"step": 2400
},
{
"epoch": 3.017565872020075,
"grad_norm": 0.36503732282533347,
"learning_rate": 3.455907768560477e-06,
"loss": 0.0419,
"step": 2405
},
{
"epoch": 3.0238393977415305,
"grad_norm": 0.34377984007445206,
"learning_rate": 3.414606866472707e-06,
"loss": 0.0402,
"step": 2410
},
{
"epoch": 3.030112923462986,
"grad_norm": 0.3334896061335003,
"learning_rate": 3.373503369898862e-06,
"loss": 0.04,
"step": 2415
},
{
"epoch": 3.0363864491844414,
"grad_norm": 0.2723536621745364,
"learning_rate": 3.3325985109683877e-06,
"loss": 0.0396,
"step": 2420
},
{
"epoch": 3.042659974905897,
"grad_norm": 0.2691974596202031,
"learning_rate": 3.291893515856334e-06,
"loss": 0.0389,
"step": 2425
},
{
"epoch": 3.0489335006273524,
"grad_norm": 0.28187769151055436,
"learning_rate": 3.2513896047465654e-06,
"loss": 0.0379,
"step": 2430
},
{
"epoch": 3.055207026348808,
"grad_norm": 0.26156534288635025,
"learning_rate": 3.211087991795201e-06,
"loss": 0.0385,
"step": 2435
},
{
"epoch": 3.0614805520702637,
"grad_norm": 0.2632608582642457,
"learning_rate": 3.1709898850942234e-06,
"loss": 0.0381,
"step": 2440
},
{
"epoch": 3.067754077791719,
"grad_norm": 0.28572880260972616,
"learning_rate": 3.1310964866352524e-06,
"loss": 0.0389,
"step": 2445
},
{
"epoch": 3.0740276035131746,
"grad_norm": 0.28301745878682383,
"learning_rate": 3.0914089922735215e-06,
"loss": 0.0375,
"step": 2450
},
{
"epoch": 3.08030112923463,
"grad_norm": 0.29208432776276283,
"learning_rate": 3.051928591692017e-06,
"loss": 0.0381,
"step": 2455
},
{
"epoch": 3.0865746549560855,
"grad_norm": 0.2847903910965424,
"learning_rate": 3.012656468365842e-06,
"loss": 0.0368,
"step": 2460
},
{
"epoch": 3.092848180677541,
"grad_norm": 0.2966192251884188,
"learning_rate": 2.9735937995267108e-06,
"loss": 0.0371,
"step": 2465
},
{
"epoch": 3.0991217063989964,
"grad_norm": 0.2796882316242176,
"learning_rate": 2.9347417561276812e-06,
"loss": 0.0376,
"step": 2470
},
{
"epoch": 3.105395232120452,
"grad_norm": 0.3038556064545983,
"learning_rate": 2.8961015028080506e-06,
"loss": 0.0385,
"step": 2475
},
{
"epoch": 3.1116687578419073,
"grad_norm": 0.2800625738008763,
"learning_rate": 2.8576741978584265e-06,
"loss": 0.0386,
"step": 2480
},
{
"epoch": 3.117942283563363,
"grad_norm": 0.29844382477038445,
"learning_rate": 2.819460993186032e-06,
"loss": 0.0388,
"step": 2485
},
{
"epoch": 3.1242158092848182,
"grad_norm": 0.3177106222509896,
"learning_rate": 2.781463034280153e-06,
"loss": 0.0379,
"step": 2490
},
{
"epoch": 3.1304893350062737,
"grad_norm": 0.2995753268233054,
"learning_rate": 2.7436814601778174e-06,
"loss": 0.0387,
"step": 2495
},
{
"epoch": 3.136762860727729,
"grad_norm": 0.30624087534086814,
"learning_rate": 2.7061174034296434e-06,
"loss": 0.0398,
"step": 2500
},
{
"epoch": 3.1430363864491846,
"grad_norm": 0.3087285778258557,
"learning_rate": 2.668771990065884e-06,
"loss": 0.0394,
"step": 2505
},
{
"epoch": 3.14930991217064,
"grad_norm": 0.2941075887579034,
"learning_rate": 2.631646339562689e-06,
"loss": 0.0386,
"step": 2510
},
{
"epoch": 3.1555834378920955,
"grad_norm": 0.3118597830883225,
"learning_rate": 2.594741564808527e-06,
"loss": 0.0397,
"step": 2515
},
{
"epoch": 3.161856963613551,
"grad_norm": 0.3208254308061256,
"learning_rate": 2.558058772070846e-06,
"loss": 0.0385,
"step": 2520
},
{
"epoch": 3.1681304893350064,
"grad_norm": 0.26274648640285136,
"learning_rate": 2.521599060962895e-06,
"loss": 0.039,
"step": 2525
},
{
"epoch": 3.174404015056462,
"grad_norm": 0.3244783368883984,
"learning_rate": 2.4853635244107743e-06,
"loss": 0.0373,
"step": 2530
},
{
"epoch": 3.1806775407779173,
"grad_norm": 0.3321834864243195,
"learning_rate": 2.449353248620657e-06,
"loss": 0.0369,
"step": 2535
},
{
"epoch": 3.1869510664993728,
"grad_norm": 0.298380472148583,
"learning_rate": 2.41356931304625e-06,
"loss": 0.0383,
"step": 2540
},
{
"epoch": 3.193224592220828,
"grad_norm": 0.2887212580399006,
"learning_rate": 2.37801279035642e-06,
"loss": 0.0383,
"step": 2545
},
{
"epoch": 3.1994981179422837,
"grad_norm": 0.2685957992866305,
"learning_rate": 2.342684746403037e-06,
"loss": 0.0382,
"step": 2550
},
{
"epoch": 3.205771643663739,
"grad_norm": 0.3137213818837783,
"learning_rate": 2.307586240189049e-06,
"loss": 0.04,
"step": 2555
},
{
"epoch": 3.2120451693851946,
"grad_norm": 0.29781867047646216,
"learning_rate": 2.272718323836701e-06,
"loss": 0.0379,
"step": 2560
},
{
"epoch": 3.21831869510665,
"grad_norm": 0.28001089505480686,
"learning_rate": 2.238082042556029e-06,
"loss": 0.0382,
"step": 2565
},
{
"epoch": 3.2245922208281055,
"grad_norm": 0.31914940913340023,
"learning_rate": 2.2036784346134976e-06,
"loss": 0.0376,
"step": 2570
},
{
"epoch": 3.230865746549561,
"grad_norm": 0.2728867395710102,
"learning_rate": 2.169508531300908e-06,
"loss": 0.0382,
"step": 2575
},
{
"epoch": 3.2371392722710164,
"grad_norm": 0.30042453420426346,
"learning_rate": 2.1355733569044633e-06,
"loss": 0.0389,
"step": 2580
},
{
"epoch": 3.243412797992472,
"grad_norm": 0.2752139872422629,
"learning_rate": 2.101873928674064e-06,
"loss": 0.0374,
"step": 2585
},
{
"epoch": 3.2496863237139273,
"grad_norm": 0.3662305601638148,
"learning_rate": 2.0684112567928314e-06,
"loss": 0.0369,
"step": 2590
},
{
"epoch": 3.2559598494353827,
"grad_norm": 0.2853508636394266,
"learning_rate": 2.035186344346801e-06,
"loss": 0.0379,
"step": 2595
},
{
"epoch": 3.262233375156838,
"grad_norm": 0.2909774250341541,
"learning_rate": 2.0022001872948814e-06,
"loss": 0.0374,
"step": 2600
},
{
"epoch": 3.2685069008782937,
"grad_norm": 0.303083557347797,
"learning_rate": 1.9694537744389754e-06,
"loss": 0.0372,
"step": 2605
},
{
"epoch": 3.274780426599749,
"grad_norm": 0.2837072081387949,
"learning_rate": 1.9369480873943524e-06,
"loss": 0.037,
"step": 2610
},
{
"epoch": 3.2810539523212046,
"grad_norm": 0.293463671873792,
"learning_rate": 1.9046841005602268e-06,
"loss": 0.0368,
"step": 2615
},
{
"epoch": 3.28732747804266,
"grad_norm": 0.29070250069814846,
"learning_rate": 1.8726627810905284e-06,
"loss": 0.037,
"step": 2620
},
{
"epoch": 3.2936010037641155,
"grad_norm": 0.28990000456159315,
"learning_rate": 1.8408850888649398e-06,
"loss": 0.0368,
"step": 2625
},
{
"epoch": 3.299874529485571,
"grad_norm": 0.2768135817440434,
"learning_rate": 1.8093519764600931e-06,
"loss": 0.0381,
"step": 2630
},
{
"epoch": 3.3061480552070264,
"grad_norm": 0.3106851238797327,
"learning_rate": 1.778064389121048e-06,
"loss": 0.0373,
"step": 2635
},
{
"epoch": 3.312421580928482,
"grad_norm": 0.26724960926250796,
"learning_rate": 1.7470232647329222e-06,
"loss": 0.0382,
"step": 2640
},
{
"epoch": 3.3186951066499373,
"grad_norm": 0.3030133716572175,
"learning_rate": 1.7162295337928036e-06,
"loss": 0.0384,
"step": 2645
},
{
"epoch": 3.3249686323713927,
"grad_norm": 0.31346010178385275,
"learning_rate": 1.685684119381844e-06,
"loss": 0.038,
"step": 2650
},
{
"epoch": 3.331242158092848,
"grad_norm": 0.28444292176954256,
"learning_rate": 1.655387937137589e-06,
"loss": 0.0363,
"step": 2655
},
{
"epoch": 3.3375156838143036,
"grad_norm": 0.29308964877533195,
"learning_rate": 1.6253418952265398e-06,
"loss": 0.0372,
"step": 2660
},
{
"epoch": 3.343789209535759,
"grad_norm": 0.29189420850538267,
"learning_rate": 1.5955468943169217e-06,
"loss": 0.0377,
"step": 2665
},
{
"epoch": 3.3500627352572145,
"grad_norm": 0.3102531079304282,
"learning_rate": 1.5660038275516898e-06,
"loss": 0.0355,
"step": 2670
},
{
"epoch": 3.35633626097867,
"grad_norm": 0.27790750204144,
"learning_rate": 1.536713580521746e-06,
"loss": 0.0372,
"step": 2675
},
{
"epoch": 3.3626097867001254,
"grad_norm": 0.3226354962908228,
"learning_rate": 1.5076770312394096e-06,
"loss": 0.0369,
"step": 2680
},
{
"epoch": 3.368883312421581,
"grad_norm": 0.26898149132297955,
"learning_rate": 1.4788950501120781e-06,
"loss": 0.0382,
"step": 2685
},
{
"epoch": 3.3751568381430364,
"grad_norm": 0.3188502485680553,
"learning_rate": 1.450368499916155e-06,
"loss": 0.0381,
"step": 2690
},
{
"epoch": 3.381430363864492,
"grad_norm": 0.3007960797574723,
"learning_rate": 1.4220982357711743e-06,
"loss": 0.0372,
"step": 2695
},
{
"epoch": 3.3877038895859473,
"grad_norm": 0.29650117453341984,
"learning_rate": 1.3940851051141646e-06,
"loss": 0.0377,
"step": 2700
},
{
"epoch": 3.3939774153074027,
"grad_norm": 0.29709762351304697,
"learning_rate": 1.366329947674263e-06,
"loss": 0.0372,
"step": 2705
},
{
"epoch": 3.400250941028858,
"grad_norm": 0.2951278070367851,
"learning_rate": 1.3388335954475207e-06,
"loss": 0.0375,
"step": 2710
},
{
"epoch": 3.4065244667503136,
"grad_norm": 0.27968586803836637,
"learning_rate": 1.3115968726719819e-06,
"loss": 0.0359,
"step": 2715
},
{
"epoch": 3.412797992471769,
"grad_norm": 0.30665692772003755,
"learning_rate": 1.284620595802969e-06,
"loss": 0.0376,
"step": 2720
},
{
"epoch": 3.4190715181932245,
"grad_norm": 0.3166229935564368,
"learning_rate": 1.2579055734886004e-06,
"loss": 0.0361,
"step": 2725
},
{
"epoch": 3.42534504391468,
"grad_norm": 0.28812682524392,
"learning_rate": 1.2314526065455678e-06,
"loss": 0.0371,
"step": 2730
},
{
"epoch": 3.4316185696361354,
"grad_norm": 0.2962918652302945,
"learning_rate": 1.2052624879351105e-06,
"loss": 0.0375,
"step": 2735
},
{
"epoch": 3.437892095357591,
"grad_norm": 0.2959207573782531,
"learning_rate": 1.179336002739263e-06,
"loss": 0.0368,
"step": 2740
},
{
"epoch": 3.4441656210790463,
"grad_norm": 0.3222288463448185,
"learning_rate": 1.1536739281373122e-06,
"loss": 0.0366,
"step": 2745
},
{
"epoch": 3.450439146800502,
"grad_norm": 0.2924594477361343,
"learning_rate": 1.1282770333825022e-06,
"loss": 0.0375,
"step": 2750
},
{
"epoch": 3.4567126725219572,
"grad_norm": 0.32299277919132047,
"learning_rate": 1.1031460797789718e-06,
"loss": 0.0374,
"step": 2755
},
{
"epoch": 3.4629861982434127,
"grad_norm": 0.3109436060749281,
"learning_rate": 1.0782818206589375e-06,
"loss": 0.0367,
"step": 2760
},
{
"epoch": 3.469259723964868,
"grad_norm": 0.27272406454567366,
"learning_rate": 1.053685001360112e-06,
"loss": 0.0356,
"step": 2765
},
{
"epoch": 3.4755332496863236,
"grad_norm": 0.30006540518061847,
"learning_rate": 1.0293563592033595e-06,
"loss": 0.0366,
"step": 2770
},
{
"epoch": 3.481806775407779,
"grad_norm": 0.28002730472344367,
"learning_rate": 1.0052966234705953e-06,
"loss": 0.0383,
"step": 2775
},
{
"epoch": 3.4880803011292345,
"grad_norm": 0.32173748821041304,
"learning_rate": 9.815065153829195e-07,
"loss": 0.0366,
"step": 2780
},
{
"epoch": 3.49435382685069,
"grad_norm": 0.3042093630820431,
"learning_rate": 9.579867480790061e-07,
"loss": 0.0369,
"step": 2785
},
{
"epoch": 3.5006273525721454,
"grad_norm": 0.2975724581779801,
"learning_rate": 9.347380265937167e-07,
"loss": 0.0361,
"step": 2790
},
{
"epoch": 3.506900878293601,
"grad_norm": 0.2893317769414031,
"learning_rate": 9.117610478369743e-07,
"loss": 0.0383,
"step": 2795
},
{
"epoch": 3.5131744040150563,
"grad_norm": 0.30137847743662044,
"learning_rate": 8.890565005728691e-07,
"loss": 0.0384,
"step": 2800
},
{
"epoch": 3.5194479297365118,
"grad_norm": 0.2713586307285477,
"learning_rate": 8.666250653990071e-07,
"loss": 0.0364,
"step": 2805
},
{
"epoch": 3.5257214554579672,
"grad_norm": 0.2749487801394873,
"learning_rate": 8.44467414726119e-07,
"loss": 0.0353,
"step": 2810
},
{
"epoch": 3.5319949811794227,
"grad_norm": 0.3243183464549784,
"learning_rate": 8.225842127578909e-07,
"loss": 0.0369,
"step": 2815
},
{
"epoch": 3.538268506900878,
"grad_norm": 0.2890309082875547,
"learning_rate": 8.009761154710671e-07,
"loss": 0.0369,
"step": 2820
},
{
"epoch": 3.544542032622334,
"grad_norm": 0.2879952226873715,
"learning_rate": 7.796437705957782e-07,
"loss": 0.0368,
"step": 2825
},
{
"epoch": 3.550815558343789,
"grad_norm": 0.2696320382813176,
"learning_rate": 7.585878175961237e-07,
"loss": 0.0371,
"step": 2830
},
{
"epoch": 3.557089084065245,
"grad_norm": 0.2936735805318314,
"learning_rate": 7.378088876510092e-07,
"loss": 0.0376,
"step": 2835
},
{
"epoch": 3.5633626097867,
"grad_norm": 0.25416101821322773,
"learning_rate": 7.1730760363522e-07,
"loss": 0.0352,
"step": 2840
},
{
"epoch": 3.569636135508156,
"grad_norm": 0.30300674321039756,
"learning_rate": 6.970845801007564e-07,
"loss": 0.0353,
"step": 2845
},
{
"epoch": 3.575909661229611,
"grad_norm": 0.2732998588656073,
"learning_rate": 6.771404232584011e-07,
"loss": 0.0362,
"step": 2850
},
{
"epoch": 3.5821831869510667,
"grad_norm": 0.3795076600221709,
"learning_rate": 6.574757309595636e-07,
"loss": 0.0361,
"step": 2855
},
{
"epoch": 3.5884567126725218,
"grad_norm": 0.28624108454045416,
"learning_rate": 6.380910926783402e-07,
"loss": 0.0363,
"step": 2860
},
{
"epoch": 3.5947302383939777,
"grad_norm": 0.2674910858708607,
"learning_rate": 6.189870894938587e-07,
"loss": 0.0353,
"step": 2865
},
{
"epoch": 3.6010037641154327,
"grad_norm": 0.2789232394059912,
"learning_rate": 6.001642940728503e-07,
"loss": 0.0354,
"step": 2870
},
{
"epoch": 3.6072772898368886,
"grad_norm": 0.30882594178964384,
"learning_rate": 5.816232706524838e-07,
"loss": 0.0366,
"step": 2875
},
{
"epoch": 3.6135508155583436,
"grad_norm": 0.2806952241750321,
"learning_rate": 5.63364575023465e-07,
"loss": 0.0361,
"step": 2880
},
{
"epoch": 3.6198243412797995,
"grad_norm": 0.30164828030350377,
"learning_rate": 5.453887545133563e-07,
"loss": 0.0378,
"step": 2885
},
{
"epoch": 3.6260978670012545,
"grad_norm": 0.2679619622153946,
"learning_rate": 5.276963479701857e-07,
"loss": 0.0358,
"step": 2890
},
{
"epoch": 3.6323713927227104,
"grad_norm": 0.3231146432041487,
"learning_rate": 5.102878857462811e-07,
"loss": 0.0389,
"step": 2895
},
{
"epoch": 3.6386449184441654,
"grad_norm": 0.2779119069062509,
"learning_rate": 4.931638896823876e-07,
"loss": 0.0384,
"step": 2900
},
{
"epoch": 3.6449184441656213,
"grad_norm": 0.2722147852694776,
"learning_rate": 4.763248730920089e-07,
"loss": 0.0358,
"step": 2905
},
{
"epoch": 3.6511919698870763,
"grad_norm": 0.293829432405976,
"learning_rate": 4.5977134074603246e-07,
"loss": 0.037,
"step": 2910
},
{
"epoch": 3.657465495608532,
"grad_norm": 0.3065193645356902,
"learning_rate": 4.4350378885759105e-07,
"loss": 0.0371,
"step": 2915
},
{
"epoch": 3.663739021329987,
"grad_norm": 0.28257044296271877,
"learning_rate": 4.275227050671904e-07,
"loss": 0.0364,
"step": 2920
},
{
"epoch": 3.670012547051443,
"grad_norm": 0.2782812011499114,
"learning_rate": 4.1182856842809204e-07,
"loss": 0.0348,
"step": 2925
},
{
"epoch": 3.676286072772898,
"grad_norm": 0.28872652510766195,
"learning_rate": 3.964218493919525e-07,
"loss": 0.0361,
"step": 2930
},
{
"epoch": 3.682559598494354,
"grad_norm": 0.3053021191496778,
"learning_rate": 3.813030097947212e-07,
"loss": 0.0357,
"step": 2935
},
{
"epoch": 3.6888331242158094,
"grad_norm": 0.3648251833459472,
"learning_rate": 3.6647250284279735e-07,
"loss": 0.0404,
"step": 2940
},
{
"epoch": 3.695106649937265,
"grad_norm": 0.3319781395595929,
"learning_rate": 3.5193077309943923e-07,
"loss": 0.0403,
"step": 2945
},
{
"epoch": 3.7013801756587204,
"grad_norm": 0.3085438905662011,
"learning_rate": 3.376782564714476e-07,
"loss": 0.0369,
"step": 2950
},
{
"epoch": 3.707653701380176,
"grad_norm": 0.2849277889599658,
"learning_rate": 3.237153801960868e-07,
"loss": 0.0362,
"step": 2955
},
{
"epoch": 3.7139272271016313,
"grad_norm": 0.31923000028121545,
"learning_rate": 3.100425628282899e-07,
"loss": 0.0369,
"step": 2960
},
{
"epoch": 3.7202007528230867,
"grad_norm": 0.2929761574906958,
"learning_rate": 2.9666021422810274e-07,
"loss": 0.0369,
"step": 2965
},
{
"epoch": 3.726474278544542,
"grad_norm": 0.2992636928106286,
"learning_rate": 2.8356873554840514e-07,
"loss": 0.0364,
"step": 2970
},
{
"epoch": 3.7327478042659976,
"grad_norm": 0.301134189273727,
"learning_rate": 2.7076851922287704e-07,
"loss": 0.0354,
"step": 2975
},
{
"epoch": 3.739021329987453,
"grad_norm": 0.2943853551390973,
"learning_rate": 2.5825994895424255e-07,
"loss": 0.0346,
"step": 2980
},
{
"epoch": 3.7452948557089085,
"grad_norm": 0.2920091794037564,
"learning_rate": 2.460433997027634e-07,
"loss": 0.0377,
"step": 2985
},
{
"epoch": 3.751568381430364,
"grad_norm": 0.27495401738635517,
"learning_rate": 2.3411923767500455e-07,
"loss": 0.0378,
"step": 2990
},
{
"epoch": 3.7578419071518194,
"grad_norm": 0.39264097272429527,
"learning_rate": 2.224878203128511e-07,
"loss": 0.037,
"step": 2995
},
{
"epoch": 3.764115432873275,
"grad_norm": 0.30250413845451674,
"learning_rate": 2.1114949628279201e-07,
"loss": 0.0366,
"step": 3000
},
{
"epoch": 3.7703889585947303,
"grad_norm": 0.35611796959076564,
"learning_rate": 2.001046054654776e-07,
"loss": 0.0371,
"step": 3005
},
{
"epoch": 3.776662484316186,
"grad_norm": 0.2561175523575295,
"learning_rate": 1.893534789455209e-07,
"loss": 0.0351,
"step": 3010
},
{
"epoch": 3.7829360100376412,
"grad_norm": 0.28295745535344086,
"learning_rate": 1.7889643900158016e-07,
"loss": 0.0361,
"step": 3015
},
{
"epoch": 3.7892095357590967,
"grad_norm": 0.28512419406398154,
"learning_rate": 1.6873379909669307e-07,
"loss": 0.0376,
"step": 3020
},
{
"epoch": 3.795483061480552,
"grad_norm": 0.30485457369447977,
"learning_rate": 1.5886586386888449e-07,
"loss": 0.0377,
"step": 3025
},
{
"epoch": 3.8017565872020076,
"grad_norm": 0.2901818797318041,
"learning_rate": 1.4929292912203354e-07,
"loss": 0.0357,
"step": 3030
},
{
"epoch": 3.808030112923463,
"grad_norm": 0.2667781630207582,
"learning_rate": 1.4001528181700196e-07,
"loss": 0.0382,
"step": 3035
},
{
"epoch": 3.8143036386449185,
"grad_norm": 0.3034166947926825,
"learning_rate": 1.3103320006303766e-07,
"loss": 0.0382,
"step": 3040
},
{
"epoch": 3.820577164366374,
"grad_norm": 0.3119770203364585,
"learning_rate": 1.2234695310944012e-07,
"loss": 0.0367,
"step": 3045
},
{
"epoch": 3.8268506900878294,
"grad_norm": 0.31429783494460234,
"learning_rate": 1.1395680133747811e-07,
"loss": 0.036,
"step": 3050
},
{
"epoch": 3.833124215809285,
"grad_norm": 0.2655885847696681,
"learning_rate": 1.0586299625259699e-07,
"loss": 0.0372,
"step": 3055
},
{
"epoch": 3.8393977415307403,
"grad_norm": 0.37692736629023765,
"learning_rate": 9.806578047687254e-08,
"loss": 0.0401,
"step": 3060
},
{
"epoch": 3.8456712672521958,
"grad_norm": 0.28037362911434216,
"learning_rate": 9.056538774174117e-08,
"loss": 0.0383,
"step": 3065
},
{
"epoch": 3.851944792973651,
"grad_norm": 0.2813649801748803,
"learning_rate": 8.336204288098671e-08,
"loss": 0.0409,
"step": 3070
},
{
"epoch": 3.8582183186951067,
"grad_norm": 0.28158812132522376,
"learning_rate": 7.64559618240146e-08,
"loss": 0.0371,
"step": 3075
},
{
"epoch": 3.864491844416562,
"grad_norm": 0.29334208788094884,
"learning_rate": 6.984735158936384e-08,
"loss": 0.0377,
"step": 3080
},
{
"epoch": 3.8707653701380176,
"grad_norm": 0.2776501474412646,
"learning_rate": 6.353641027850965e-08,
"loss": 0.0368,
"step": 3085
},
{
"epoch": 3.877038895859473,
"grad_norm": 0.27164861879214747,
"learning_rate": 5.7523327069926024e-08,
"loss": 0.0371,
"step": 3090
},
{
"epoch": 3.8833124215809285,
"grad_norm": 0.26782362958737654,
"learning_rate": 5.1808282213410276e-08,
"loss": 0.0366,
"step": 3095
},
{
"epoch": 3.889585947302384,
"grad_norm": 0.31086719768707505,
"learning_rate": 4.63914470246829e-08,
"loss": 0.0361,
"step": 3100
},
{
"epoch": 3.8958594730238394,
"grad_norm": 0.30662847914316993,
"learning_rate": 4.1272983880249476e-08,
"loss": 0.0374,
"step": 3105
},
{
"epoch": 3.902132998745295,
"grad_norm": 0.2915681277032509,
"learning_rate": 3.645304621253787e-08,
"loss": 0.0365,
"step": 3110
},
{
"epoch": 3.9084065244667503,
"grad_norm": 0.31776081537688317,
"learning_rate": 3.193177850529416e-08,
"loss": 0.0352,
"step": 3115
},
{
"epoch": 3.9146800501882058,
"grad_norm": 0.26314408385807814,
"learning_rate": 2.7709316289253885e-08,
"loss": 0.0362,
"step": 3120
},
{
"epoch": 3.920953575909661,
"grad_norm": 0.30353550353186937,
"learning_rate": 2.378578613807969e-08,
"loss": 0.0367,
"step": 3125
},
{
"epoch": 3.9272271016311167,
"grad_norm": 0.3008469541365023,
"learning_rate": 2.0161305664563312e-08,
"loss": 0.0377,
"step": 3130
},
{
"epoch": 3.933500627352572,
"grad_norm": 0.29989474797608723,
"learning_rate": 1.6835983517108357e-08,
"loss": 0.0364,
"step": 3135
},
{
"epoch": 3.9397741530740276,
"grad_norm": 0.3007906674460069,
"learning_rate": 1.3809919376461811e-08,
"loss": 0.0367,
"step": 3140
},
{
"epoch": 3.946047678795483,
"grad_norm": 0.2822359703350314,
"learning_rate": 1.1083203952737543e-08,
"loss": 0.0371,
"step": 3145
},
{
"epoch": 3.9523212045169385,
"grad_norm": 0.32240083725231283,
"learning_rate": 8.655918982689582e-09,
"loss": 0.0367,
"step": 3150
},
{
"epoch": 3.958594730238394,
"grad_norm": 0.27144908135189433,
"learning_rate": 6.528137227262976e-09,
"loss": 0.0368,
"step": 3155
},
{
"epoch": 3.9648682559598494,
"grad_norm": 0.29276670822521234,
"learning_rate": 4.6999224694166405e-09,
"loss": 0.0363,
"step": 3160
},
{
"epoch": 3.971141781681305,
"grad_norm": 0.2920658294802816,
"learning_rate": 3.1713295122071107e-09,
"loss": 0.0352,
"step": 3165
},
{
"epoch": 3.9774153074027603,
"grad_norm": 0.29999904692777374,
"learning_rate": 1.9424041771465286e-09,
"loss": 0.0354,
"step": 3170
},
{
"epoch": 3.9836888331242157,
"grad_norm": 0.28940456234051576,
"learning_rate": 1.013183302832621e-09,
"loss": 0.0379,
"step": 3175
},
{
"epoch": 3.989962358845671,
"grad_norm": 0.28743205163263064,
"learning_rate": 3.8369474383848083e-10,
"loss": 0.0349,
"step": 3180
},
{
"epoch": 3.9962358845671266,
"grad_norm": 0.3004497086648703,
"learning_rate": 5.395736988322853e-11,
"loss": 0.0363,
"step": 3185
},
{
"epoch": 4.0,
"eval_loss": 0.47916728258132935,
"eval_runtime": 2.4435,
"eval_samples_per_second": 14.324,
"eval_steps_per_second": 0.818,
"step": 3188
},
{
"epoch": 4.0,
"step": 3188,
"total_flos": 2680278636036096.0,
"train_loss": 0.20304497943459596,
"train_runtime": 21214.9362,
"train_samples_per_second": 4.806,
"train_steps_per_second": 0.15
}
],
"logging_steps": 5,
"max_steps": 3188,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 319,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2680278636036096.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}