llama3-sudo-3epochs-mask / trainer_state.json
Qin Liu
Model save
63a93d0 verified
raw
history blame
24.9 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 687,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004366812227074236,
"grad_norm": 1.8753604454650958,
"learning_rate": 2.898550724637681e-06,
"loss": 3.7085,
"step": 1
},
{
"epoch": 0.021834061135371178,
"grad_norm": 1.801543636467953,
"learning_rate": 1.4492753623188407e-05,
"loss": 3.628,
"step": 5
},
{
"epoch": 0.043668122270742356,
"grad_norm": 2.2358499812546344,
"learning_rate": 2.8985507246376814e-05,
"loss": 3.5205,
"step": 10
},
{
"epoch": 0.06550218340611354,
"grad_norm": 3.2738653822216937,
"learning_rate": 4.347826086956522e-05,
"loss": 3.3516,
"step": 15
},
{
"epoch": 0.08733624454148471,
"grad_norm": 1.2942851623807274,
"learning_rate": 5.797101449275363e-05,
"loss": 2.7683,
"step": 20
},
{
"epoch": 0.1091703056768559,
"grad_norm": 0.6794530467895241,
"learning_rate": 7.246376811594203e-05,
"loss": 2.4019,
"step": 25
},
{
"epoch": 0.13100436681222707,
"grad_norm": 1.2145036683171,
"learning_rate": 8.695652173913044e-05,
"loss": 2.1421,
"step": 30
},
{
"epoch": 0.15283842794759825,
"grad_norm": 0.8541800089072681,
"learning_rate": 0.00010144927536231885,
"loss": 1.9309,
"step": 35
},
{
"epoch": 0.17467248908296942,
"grad_norm": 0.5114546320288847,
"learning_rate": 0.00011594202898550725,
"loss": 1.9047,
"step": 40
},
{
"epoch": 0.1965065502183406,
"grad_norm": 0.4569825196048275,
"learning_rate": 0.00013043478260869567,
"loss": 1.8231,
"step": 45
},
{
"epoch": 0.2183406113537118,
"grad_norm": 0.5192344724636687,
"learning_rate": 0.00014492753623188405,
"loss": 1.7305,
"step": 50
},
{
"epoch": 0.24017467248908297,
"grad_norm": 0.34840209414620005,
"learning_rate": 0.00015942028985507247,
"loss": 1.7933,
"step": 55
},
{
"epoch": 0.26200873362445415,
"grad_norm": 0.3678371866239178,
"learning_rate": 0.00017391304347826088,
"loss": 1.6143,
"step": 60
},
{
"epoch": 0.2838427947598253,
"grad_norm": 0.35400819586929,
"learning_rate": 0.00018840579710144927,
"loss": 1.7367,
"step": 65
},
{
"epoch": 0.3056768558951965,
"grad_norm": 0.38841557534962395,
"learning_rate": 0.00019999870791268066,
"loss": 1.6669,
"step": 70
},
{
"epoch": 0.32751091703056767,
"grad_norm": 0.4474417485361201,
"learning_rate": 0.00019995348836233516,
"loss": 1.601,
"step": 75
},
{
"epoch": 0.34934497816593885,
"grad_norm": 0.35698828488835516,
"learning_rate": 0.00019984369783193688,
"loss": 1.5357,
"step": 80
},
{
"epoch": 0.37117903930131,
"grad_norm": 0.3585527157477228,
"learning_rate": 0.00019966940724729603,
"loss": 1.6221,
"step": 85
},
{
"epoch": 0.3930131004366812,
"grad_norm": 0.34800578812888927,
"learning_rate": 0.0001994307292019204,
"loss": 1.6741,
"step": 90
},
{
"epoch": 0.4148471615720524,
"grad_norm": 0.35237447031832875,
"learning_rate": 0.0001991278178842786,
"loss": 1.5682,
"step": 95
},
{
"epoch": 0.4366812227074236,
"grad_norm": 0.39076567242886867,
"learning_rate": 0.00019876086897819284,
"loss": 1.5629,
"step": 100
},
{
"epoch": 0.4585152838427948,
"grad_norm": 0.34983112900974217,
"learning_rate": 0.00019833011953642525,
"loss": 1.6151,
"step": 105
},
{
"epoch": 0.48034934497816595,
"grad_norm": 0.34530341844436674,
"learning_rate": 0.00019783584782753918,
"loss": 1.5494,
"step": 110
},
{
"epoch": 0.5021834061135371,
"grad_norm": 0.38496791747153597,
"learning_rate": 0.00019727837315613504,
"loss": 1.5526,
"step": 115
},
{
"epoch": 0.5240174672489083,
"grad_norm": 0.35057371546794386,
"learning_rate": 0.00019665805565657603,
"loss": 1.5933,
"step": 120
},
{
"epoch": 0.5458515283842795,
"grad_norm": 0.35636722375213375,
"learning_rate": 0.00019597529606033782,
"loss": 1.5726,
"step": 125
},
{
"epoch": 0.5676855895196506,
"grad_norm": 0.375075170102756,
"learning_rate": 0.0001952305354371319,
"loss": 1.572,
"step": 130
},
{
"epoch": 0.5895196506550219,
"grad_norm": 0.38062791914729815,
"learning_rate": 0.00019442425490996988,
"loss": 1.5393,
"step": 135
},
{
"epoch": 0.611353711790393,
"grad_norm": 0.3465840806200646,
"learning_rate": 0.0001935569753443532,
"loss": 1.502,
"step": 140
},
{
"epoch": 0.6331877729257642,
"grad_norm": 0.3566623409862706,
"learning_rate": 0.00019262925701178866,
"loss": 1.6075,
"step": 145
},
{
"epoch": 0.6550218340611353,
"grad_norm": 0.3513669246328785,
"learning_rate": 0.00019164169922784716,
"loss": 1.5125,
"step": 150
},
{
"epoch": 0.6768558951965066,
"grad_norm": 0.34755655244144995,
"learning_rate": 0.00019059493996499986,
"loss": 1.6031,
"step": 155
},
{
"epoch": 0.6986899563318777,
"grad_norm": 0.35330261801172397,
"learning_rate": 0.00018948965544048128,
"loss": 1.5862,
"step": 160
},
{
"epoch": 0.7205240174672489,
"grad_norm": 0.3540773871103764,
"learning_rate": 0.00018832655967944607,
"loss": 1.6557,
"step": 165
},
{
"epoch": 0.74235807860262,
"grad_norm": 0.3385589855658203,
"learning_rate": 0.00018710640405370145,
"loss": 1.5771,
"step": 170
},
{
"epoch": 0.7641921397379913,
"grad_norm": 0.3948832031761469,
"learning_rate": 0.00018582997679631315,
"loss": 1.5896,
"step": 175
},
{
"epoch": 0.7860262008733624,
"grad_norm": 0.3414806859938428,
"learning_rate": 0.00018449810249239902,
"loss": 1.5278,
"step": 180
},
{
"epoch": 0.8078602620087336,
"grad_norm": 0.3306470223687427,
"learning_rate": 0.00018311164154643836,
"loss": 1.4916,
"step": 185
},
{
"epoch": 0.8296943231441049,
"grad_norm": 0.37143716375731756,
"learning_rate": 0.00018167148962644193,
"loss": 1.625,
"step": 190
},
{
"epoch": 0.851528384279476,
"grad_norm": 0.3610023951445998,
"learning_rate": 0.00018017857708534107,
"loss": 1.6859,
"step": 195
},
{
"epoch": 0.8733624454148472,
"grad_norm": 0.3969039590867808,
"learning_rate": 0.00017863386835997028,
"loss": 1.6366,
"step": 200
},
{
"epoch": 0.8951965065502183,
"grad_norm": 0.3466218548962297,
"learning_rate": 0.00017703836134803105,
"loss": 1.4699,
"step": 205
},
{
"epoch": 0.9170305676855895,
"grad_norm": 0.35771965160066704,
"learning_rate": 0.00017539308676343973,
"loss": 1.5723,
"step": 210
},
{
"epoch": 0.9388646288209607,
"grad_norm": 0.37691637209544226,
"learning_rate": 0.00017369910747047572,
"loss": 1.584,
"step": 215
},
{
"epoch": 0.9606986899563319,
"grad_norm": 0.40411027091446045,
"learning_rate": 0.00017195751779716027,
"loss": 1.6019,
"step": 220
},
{
"epoch": 0.982532751091703,
"grad_norm": 0.3742848841742168,
"learning_rate": 0.00017016944282830933,
"loss": 1.4947,
"step": 225
},
{
"epoch": 1.0043668122270741,
"grad_norm": 0.38407235492099207,
"learning_rate": 0.00016833603767871713,
"loss": 1.5812,
"step": 230
},
{
"epoch": 1.0262008733624455,
"grad_norm": 0.37035584561109924,
"learning_rate": 0.0001664584867469403,
"loss": 1.5113,
"step": 235
},
{
"epoch": 1.0480349344978166,
"grad_norm": 0.4171769928455508,
"learning_rate": 0.0001645380029501641,
"loss": 1.427,
"step": 240
},
{
"epoch": 1.0698689956331877,
"grad_norm": 0.4014807349448744,
"learning_rate": 0.00016257582694064558,
"loss": 1.488,
"step": 245
},
{
"epoch": 1.091703056768559,
"grad_norm": 0.41603472494317567,
"learning_rate": 0.00016057322630423935,
"loss": 1.4085,
"step": 250
},
{
"epoch": 1.1135371179039302,
"grad_norm": 0.41696638930397173,
"learning_rate": 0.00015853149474152423,
"loss": 1.417,
"step": 255
},
{
"epoch": 1.1353711790393013,
"grad_norm": 0.42289781015455985,
"learning_rate": 0.0001564519512320593,
"loss": 1.4374,
"step": 260
},
{
"epoch": 1.1572052401746724,
"grad_norm": 0.4136181415806817,
"learning_rate": 0.00015433593918230955,
"loss": 1.5384,
"step": 265
},
{
"epoch": 1.1790393013100438,
"grad_norm": 0.4304951924614819,
"learning_rate": 0.00015218482555779165,
"loss": 1.4184,
"step": 270
},
{
"epoch": 1.2008733624454149,
"grad_norm": 0.4567200712003195,
"learning_rate": 0.00015000000000000001,
"loss": 1.5038,
"step": 275
},
{
"epoch": 1.222707423580786,
"grad_norm": 0.418159543142196,
"learning_rate": 0.00014778287392868417,
"loss": 1.4477,
"step": 280
},
{
"epoch": 1.244541484716157,
"grad_norm": 0.43554491426674474,
"learning_rate": 0.0001455348796300571,
"loss": 1.362,
"step": 285
},
{
"epoch": 1.2663755458515285,
"grad_norm": 0.4303389630106061,
"learning_rate": 0.0001432574693315238,
"loss": 1.5283,
"step": 290
},
{
"epoch": 1.2882096069868996,
"grad_norm": 0.4434624782109262,
"learning_rate": 0.0001409521142635272,
"loss": 1.5192,
"step": 295
},
{
"epoch": 1.3100436681222707,
"grad_norm": 0.4160064760695591,
"learning_rate": 0.0001386203037091183,
"loss": 1.5222,
"step": 300
},
{
"epoch": 1.3318777292576418,
"grad_norm": 0.44897734162151265,
"learning_rate": 0.00013626354404186404,
"loss": 1.5612,
"step": 305
},
{
"epoch": 1.3537117903930131,
"grad_norm": 0.4378993469096492,
"learning_rate": 0.00013388335775271467,
"loss": 1.4442,
"step": 310
},
{
"epoch": 1.3755458515283843,
"grad_norm": 0.4403089794683692,
"learning_rate": 0.0001314812824664585,
"loss": 1.5151,
"step": 315
},
{
"epoch": 1.3973799126637554,
"grad_norm": 0.44038010399377847,
"learning_rate": 0.000129058869948401,
"loss": 1.5635,
"step": 320
},
{
"epoch": 1.4192139737991267,
"grad_norm": 0.4640302354831066,
"learning_rate": 0.00012661768510190816,
"loss": 1.4904,
"step": 325
},
{
"epoch": 1.4410480349344978,
"grad_norm": 0.45432337967027536,
"learning_rate": 0.00012415930495746302,
"loss": 1.5045,
"step": 330
},
{
"epoch": 1.462882096069869,
"grad_norm": 0.451657842641418,
"learning_rate": 0.00012168531765388755,
"loss": 1.392,
"step": 335
},
{
"epoch": 1.48471615720524,
"grad_norm": 0.43285118470181244,
"learning_rate": 0.00011919732141238898,
"loss": 1.461,
"step": 340
},
{
"epoch": 1.5065502183406112,
"grad_norm": 0.4819434243733131,
"learning_rate": 0.00011669692350409223,
"loss": 1.6045,
"step": 345
},
{
"epoch": 1.5283842794759825,
"grad_norm": 0.440140827625262,
"learning_rate": 0.00011418573921172635,
"loss": 1.5039,
"step": 350
},
{
"epoch": 1.5502183406113537,
"grad_norm": 0.4475130341158297,
"learning_rate": 0.00011166539078613525,
"loss": 1.5163,
"step": 355
},
{
"epoch": 1.572052401746725,
"grad_norm": 0.4563951417730544,
"learning_rate": 0.00010913750639828711,
"loss": 1.4896,
"step": 360
},
{
"epoch": 1.5938864628820961,
"grad_norm": 0.4570029980279654,
"learning_rate": 0.0001066037190874591,
"loss": 1.4318,
"step": 365
},
{
"epoch": 1.6157205240174672,
"grad_norm": 0.45220265285594413,
"learning_rate": 0.00010406566570627713,
"loss": 1.484,
"step": 370
},
{
"epoch": 1.6375545851528384,
"grad_norm": 0.43217867151367745,
"learning_rate": 0.0001015249858632926,
"loss": 1.3569,
"step": 375
},
{
"epoch": 1.6593886462882095,
"grad_norm": 0.45524805316851796,
"learning_rate": 9.898332086377805e-05,
"loss": 1.5309,
"step": 380
},
{
"epoch": 1.6812227074235808,
"grad_norm": 0.45864967058418454,
"learning_rate": 9.644231264942724e-05,
"loss": 1.5568,
"step": 385
},
{
"epoch": 1.703056768558952,
"grad_norm": 0.46156320050595717,
"learning_rate": 9.390360273764411e-05,
"loss": 1.5541,
"step": 390
},
{
"epoch": 1.7248908296943233,
"grad_norm": 0.4424924245408702,
"learning_rate": 9.136883116110542e-05,
"loss": 1.4779,
"step": 395
},
{
"epoch": 1.7467248908296944,
"grad_norm": 0.4631844693338652,
"learning_rate": 8.88396354082829e-05,
"loss": 1.5041,
"step": 400
},
{
"epoch": 1.7685589519650655,
"grad_norm": 0.49717848531939474,
"learning_rate": 8.6317649365609e-05,
"loss": 1.4976,
"step": 405
},
{
"epoch": 1.7903930131004366,
"grad_norm": 0.44110942790393187,
"learning_rate": 8.380450226196925e-05,
"loss": 1.3881,
"step": 410
},
{
"epoch": 1.8122270742358078,
"grad_norm": 0.5089589581007182,
"learning_rate": 8.130181761620392e-05,
"loss": 1.4779,
"step": 415
},
{
"epoch": 1.8340611353711789,
"grad_norm": 0.4698355464759574,
"learning_rate": 7.881121218829787e-05,
"loss": 1.4198,
"step": 420
},
{
"epoch": 1.8558951965065502,
"grad_norm": 0.4659031789449208,
"learning_rate": 7.63342949349373e-05,
"loss": 1.4861,
"step": 425
},
{
"epoch": 1.8777292576419216,
"grad_norm": 0.44894777329606517,
"learning_rate": 7.387266597010704e-05,
"loss": 1.503,
"step": 430
},
{
"epoch": 1.8995633187772927,
"grad_norm": 0.4475374854045599,
"learning_rate": 7.142791553140045e-05,
"loss": 1.5077,
"step": 435
},
{
"epoch": 1.9213973799126638,
"grad_norm": 0.4750409585386964,
"learning_rate": 6.900162295270968e-05,
"loss": 1.515,
"step": 440
},
{
"epoch": 1.943231441048035,
"grad_norm": 0.4786776337696274,
"learning_rate": 6.659535564395982e-05,
"loss": 1.5167,
"step": 445
},
{
"epoch": 1.965065502183406,
"grad_norm": 0.5012494244267932,
"learning_rate": 6.421066807854584e-05,
"loss": 1.5364,
"step": 450
},
{
"epoch": 1.9868995633187772,
"grad_norm": 0.445678005467061,
"learning_rate": 6.184910078912687e-05,
"loss": 1.4215,
"step": 455
},
{
"epoch": 2.0087336244541483,
"grad_norm": 0.4871967245962684,
"learning_rate": 5.9512179372426325e-05,
"loss": 1.4481,
"step": 460
},
{
"epoch": 2.03056768558952,
"grad_norm": 0.48019280605703346,
"learning_rate": 5.720141350368072e-05,
"loss": 1.45,
"step": 465
},
{
"epoch": 2.052401746724891,
"grad_norm": 0.5224256236632303,
"learning_rate": 5.4918295961373923e-05,
"loss": 1.4061,
"step": 470
},
{
"epoch": 2.074235807860262,
"grad_norm": 0.5108294729537572,
"learning_rate": 5.266430166288705e-05,
"loss": 1.3943,
"step": 475
},
{
"epoch": 2.096069868995633,
"grad_norm": 0.5371496292909286,
"learning_rate": 5.044088671168644e-05,
"loss": 1.3578,
"step": 480
},
{
"epoch": 2.1179039301310043,
"grad_norm": 0.5481243225476391,
"learning_rate": 4.824948745666621e-05,
"loss": 1.3686,
"step": 485
},
{
"epoch": 2.1397379912663754,
"grad_norm": 0.5490865552703573,
"learning_rate": 4.6091519564251793e-05,
"loss": 1.3655,
"step": 490
},
{
"epoch": 2.1615720524017465,
"grad_norm": 0.5341608991319753,
"learning_rate": 4.3968377103865024e-05,
"loss": 1.3681,
"step": 495
},
{
"epoch": 2.183406113537118,
"grad_norm": 0.5539995610447883,
"learning_rate": 4.1881431647341054e-05,
"loss": 1.3703,
"step": 500
},
{
"epoch": 2.2052401746724892,
"grad_norm": 0.5617916414503629,
"learning_rate": 3.9832031382878766e-05,
"loss": 1.3506,
"step": 505
},
{
"epoch": 2.2270742358078603,
"grad_norm": 0.5457137690561883,
"learning_rate": 3.7821500244097274e-05,
"loss": 1.33,
"step": 510
},
{
"epoch": 2.2489082969432315,
"grad_norm": 0.5996161081223403,
"learning_rate": 3.585113705476143e-05,
"loss": 1.389,
"step": 515
},
{
"epoch": 2.2707423580786026,
"grad_norm": 0.5451761149567781,
"learning_rate": 3.392221468972805e-05,
"loss": 1.3908,
"step": 520
},
{
"epoch": 2.2925764192139737,
"grad_norm": 0.6118020705146557,
"learning_rate": 3.203597925265598e-05,
"loss": 1.3559,
"step": 525
},
{
"epoch": 2.314410480349345,
"grad_norm": 0.5671066923036526,
"learning_rate": 3.0193649271010095e-05,
"loss": 1.3478,
"step": 530
},
{
"epoch": 2.3362445414847164,
"grad_norm": 0.6001118016692729,
"learning_rate": 2.8396414908880098e-05,
"loss": 1.3509,
"step": 535
},
{
"epoch": 2.3580786026200875,
"grad_norm": 0.5894837707963337,
"learning_rate": 2.6645437198122502e-05,
"loss": 1.4214,
"step": 540
},
{
"epoch": 2.3799126637554586,
"grad_norm": 0.5845209440156467,
"learning_rate": 2.4941847288321797e-05,
"loss": 1.3788,
"step": 545
},
{
"epoch": 2.4017467248908297,
"grad_norm": 0.5895995554814336,
"learning_rate": 2.328674571605637e-05,
"loss": 1.391,
"step": 550
},
{
"epoch": 2.423580786026201,
"grad_norm": 0.6008227175127105,
"learning_rate": 2.1681201693940668e-05,
"loss": 1.4373,
"step": 555
},
{
"epoch": 2.445414847161572,
"grad_norm": 0.6000811157919661,
"learning_rate": 2.0126252419902614e-05,
"loss": 1.4406,
"step": 560
},
{
"epoch": 2.467248908296943,
"grad_norm": 0.5911207625229586,
"learning_rate": 1.8622902407143394e-05,
"loss": 1.5141,
"step": 565
},
{
"epoch": 2.489082969432314,
"grad_norm": 0.615377574511787,
"learning_rate": 1.7172122835211337e-05,
"loss": 1.3896,
"step": 570
},
{
"epoch": 2.5109170305676853,
"grad_norm": 0.5857047240018323,
"learning_rate": 1.577485092261012e-05,
"loss": 1.4468,
"step": 575
},
{
"epoch": 2.532751091703057,
"grad_norm": 0.6061359574672782,
"learning_rate": 1.4431989321345974e-05,
"loss": 1.4299,
"step": 580
},
{
"epoch": 2.554585152838428,
"grad_norm": 0.6360146523250017,
"learning_rate": 1.3144405533805138e-05,
"loss": 1.4552,
"step": 585
},
{
"epoch": 2.576419213973799,
"grad_norm": 0.6147809358329509,
"learning_rate": 1.191293135233844e-05,
"loss": 1.3359,
"step": 590
},
{
"epoch": 2.5982532751091703,
"grad_norm": 0.6021229536534441,
"learning_rate": 1.0738362321914997e-05,
"loss": 1.3927,
"step": 595
},
{
"epoch": 2.6200873362445414,
"grad_norm": 0.5961109142620634,
"learning_rate": 9.62145722619182e-06,
"loss": 1.4896,
"step": 600
},
{
"epoch": 2.641921397379913,
"grad_norm": 0.578602732768201,
"learning_rate": 8.562937597331899e-06,
"loss": 1.4565,
"step": 605
},
{
"epoch": 2.6637554585152836,
"grad_norm": 0.6458615704409548,
"learning_rate": 7.563487249887024e-06,
"loss": 1.4511,
"step": 610
},
{
"epoch": 2.685589519650655,
"grad_norm": 0.6022868302443363,
"learning_rate": 6.623751839046455e-06,
"loss": 1.3836,
"step": 615
},
{
"epoch": 2.7074235807860263,
"grad_norm": 0.5988174075228108,
"learning_rate": 5.744338443537134e-06,
"loss": 1.4891,
"step": 620
},
{
"epoch": 2.7292576419213974,
"grad_norm": 0.6026020303199385,
"learning_rate": 4.92581517344457e-06,
"loss": 1.3409,
"step": 625
},
{
"epoch": 2.7510917030567685,
"grad_norm": 0.5919990549417811,
"learning_rate": 4.168710803207865e-06,
"loss": 1.4157,
"step": 630
},
{
"epoch": 2.7729257641921397,
"grad_norm": 0.6038824964947008,
"learning_rate": 3.473514430026026e-06,
"loss": 1.4138,
"step": 635
},
{
"epoch": 2.7947598253275108,
"grad_norm": 0.6135553904485512,
"learning_rate": 2.840675157896111e-06,
"loss": 1.411,
"step": 640
},
{
"epoch": 2.816593886462882,
"grad_norm": 0.6332119474956278,
"learning_rate": 2.2706018074875045e-06,
"loss": 1.2871,
"step": 645
},
{
"epoch": 2.8384279475982535,
"grad_norm": 0.607982963030175,
"learning_rate": 1.7636626520395105e-06,
"loss": 1.4341,
"step": 650
},
{
"epoch": 2.8602620087336246,
"grad_norm": 0.6326939787843192,
"learning_rate": 1.3201851794530373e-06,
"loss": 1.3992,
"step": 655
},
{
"epoch": 2.8820960698689957,
"grad_norm": 0.6624316672369768,
"learning_rate": 9.404558807301067e-07,
"loss": 1.424,
"step": 660
},
{
"epoch": 2.903930131004367,
"grad_norm": 0.5890404802190057,
"learning_rate": 6.247200648976991e-07,
"loss": 1.3734,
"step": 665
},
{
"epoch": 2.925764192139738,
"grad_norm": 0.588934601286785,
"learning_rate": 3.7318170053559644e-07,
"loss": 1.413,
"step": 670
},
{
"epoch": 2.947598253275109,
"grad_norm": 0.6171879119635909,
"learning_rate": 1.8600328401061629e-07,
"loss": 1.3424,
"step": 675
},
{
"epoch": 2.96943231441048,
"grad_norm": 0.6328278329175181,
"learning_rate": 6.33057345022281e-08,
"loss": 1.3699,
"step": 680
},
{
"epoch": 2.9912663755458517,
"grad_norm": 0.609229235169678,
"learning_rate": 5.1683158875937e-09,
"loss": 1.3262,
"step": 685
},
{
"epoch": 3.0,
"step": 687,
"total_flos": 1562393690767360.0,
"train_loss": 1.5581738730184898,
"train_runtime": 1339.3383,
"train_samples_per_second": 32.797,
"train_steps_per_second": 0.513
}
],
"logging_steps": 5,
"max_steps": 687,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1562393690767360.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}