{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 1044, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009578544061302681, "grad_norm": 0.6597593171619324, "learning_rate": 9.523809523809525e-08, "loss": 1.1525, "step": 1 }, { "epoch": 0.004789272030651341, "grad_norm": 0.6607550915738474, "learning_rate": 4.7619047619047623e-07, "loss": 1.1632, "step": 5 }, { "epoch": 0.009578544061302681, "grad_norm": 0.6985874969645404, "learning_rate": 9.523809523809525e-07, "loss": 1.1753, "step": 10 }, { "epoch": 0.014367816091954023, "grad_norm": 0.6553768135986224, "learning_rate": 1.4285714285714286e-06, "loss": 1.1488, "step": 15 }, { "epoch": 0.019157088122605363, "grad_norm": 0.6087816631352209, "learning_rate": 1.904761904761905e-06, "loss": 1.1384, "step": 20 }, { "epoch": 0.023946360153256706, "grad_norm": 0.5225466109798246, "learning_rate": 2.380952380952381e-06, "loss": 1.122, "step": 25 }, { "epoch": 0.028735632183908046, "grad_norm": 0.5080052683286845, "learning_rate": 2.8571428571428573e-06, "loss": 1.1505, "step": 30 }, { "epoch": 0.033524904214559385, "grad_norm": 0.4341207873557241, "learning_rate": 3.3333333333333333e-06, "loss": 1.1339, "step": 35 }, { "epoch": 0.038314176245210725, "grad_norm": 0.3985137882747194, "learning_rate": 3.80952380952381e-06, "loss": 1.1282, "step": 40 }, { "epoch": 0.04310344827586207, "grad_norm": 0.36938748261120696, "learning_rate": 4.2857142857142855e-06, "loss": 1.1141, "step": 45 }, { "epoch": 0.04789272030651341, "grad_norm": 0.31186688358132403, "learning_rate": 4.761904761904762e-06, "loss": 1.0914, "step": 50 }, { "epoch": 0.05268199233716475, "grad_norm": 0.3279958963323018, "learning_rate": 5.2380952380952384e-06, "loss": 1.1115, "step": 55 }, { "epoch": 0.05747126436781609, "grad_norm": 0.3107640896000187, "learning_rate": 5.7142857142857145e-06, "loss": 1.0897, "step": 60 }, { "epoch": 0.06226053639846743, "grad_norm": 0.3071723661825106, "learning_rate": 6.1904761904761914e-06, "loss": 1.1045, "step": 65 }, { "epoch": 0.06704980842911877, "grad_norm": 0.2936092764163838, "learning_rate": 6.666666666666667e-06, "loss": 1.0661, "step": 70 }, { "epoch": 0.07183908045977011, "grad_norm": 0.28053978486957076, "learning_rate": 7.1428571428571436e-06, "loss": 1.0887, "step": 75 }, { "epoch": 0.07662835249042145, "grad_norm": 0.2891456438800518, "learning_rate": 7.61904761904762e-06, "loss": 1.0756, "step": 80 }, { "epoch": 0.08141762452107279, "grad_norm": 0.24950135731624706, "learning_rate": 8.095238095238097e-06, "loss": 1.0478, "step": 85 }, { "epoch": 0.08620689655172414, "grad_norm": 0.23670461068088686, "learning_rate": 8.571428571428571e-06, "loss": 1.0389, "step": 90 }, { "epoch": 0.09099616858237548, "grad_norm": 0.2507461508986584, "learning_rate": 9.047619047619049e-06, "loss": 1.0509, "step": 95 }, { "epoch": 0.09578544061302682, "grad_norm": 0.22930855800514297, "learning_rate": 9.523809523809525e-06, "loss": 1.0377, "step": 100 }, { "epoch": 0.10057471264367816, "grad_norm": 0.24445927836597864, "learning_rate": 1e-05, "loss": 1.0506, "step": 105 }, { "epoch": 0.1053639846743295, "grad_norm": 0.22388714842997454, "learning_rate": 9.999300418283908e-06, "loss": 1.0377, "step": 110 }, { "epoch": 0.11015325670498084, "grad_norm": 0.2406057364217601, "learning_rate": 9.997201868901463e-06, "loss": 1.0374, "step": 115 }, { "epoch": 0.11494252873563218, "grad_norm": 0.2275701466092952, "learning_rate": 9.993704939095376e-06, "loss": 1.0663, "step": 120 }, { "epoch": 0.11973180076628352, "grad_norm": 0.22200651495287974, "learning_rate": 9.988810607420912e-06, "loss": 1.0448, "step": 125 }, { "epoch": 0.12452107279693486, "grad_norm": 0.22051254365376602, "learning_rate": 9.982520243472044e-06, "loss": 1.0099, "step": 130 }, { "epoch": 0.12931034482758622, "grad_norm": 0.23279527845715264, "learning_rate": 9.974835607498224e-06, "loss": 1.0212, "step": 135 }, { "epoch": 0.13409961685823754, "grad_norm": 0.22813702069156522, "learning_rate": 9.965758849911774e-06, "loss": 1.023, "step": 140 }, { "epoch": 0.1388888888888889, "grad_norm": 0.23555996785615885, "learning_rate": 9.955292510686156e-06, "loss": 0.9997, "step": 145 }, { "epoch": 0.14367816091954022, "grad_norm": 0.2560040654609538, "learning_rate": 9.943439518645193e-06, "loss": 1.0172, "step": 150 }, { "epoch": 0.14846743295019157, "grad_norm": 0.24096780952464245, "learning_rate": 9.930203190643491e-06, "loss": 0.9876, "step": 155 }, { "epoch": 0.1532567049808429, "grad_norm": 0.24193989881098193, "learning_rate": 9.915587230638269e-06, "loss": 1.0417, "step": 160 }, { "epoch": 0.15804597701149425, "grad_norm": 0.24788380685714667, "learning_rate": 9.899595728652883e-06, "loss": 1.0332, "step": 165 }, { "epoch": 0.16283524904214558, "grad_norm": 0.2570877531631883, "learning_rate": 9.882233159632297e-06, "loss": 1.0129, "step": 170 }, { "epoch": 0.16762452107279693, "grad_norm": 0.26735645578353723, "learning_rate": 9.863504382190838e-06, "loss": 1.0255, "step": 175 }, { "epoch": 0.1724137931034483, "grad_norm": 0.24596191852493296, "learning_rate": 9.843414637252615e-06, "loss": 1.0125, "step": 180 }, { "epoch": 0.1772030651340996, "grad_norm": 0.2510254446995529, "learning_rate": 9.821969546584922e-06, "loss": 1.022, "step": 185 }, { "epoch": 0.18199233716475097, "grad_norm": 0.2626288820781652, "learning_rate": 9.79917511122509e-06, "loss": 1.0016, "step": 190 }, { "epoch": 0.1867816091954023, "grad_norm": 0.2588872211996587, "learning_rate": 9.775037709801206e-06, "loss": 1.0292, "step": 195 }, { "epoch": 0.19157088122605365, "grad_norm": 0.28272889728543066, "learning_rate": 9.749564096747148e-06, "loss": 1.0255, "step": 200 }, { "epoch": 0.19636015325670497, "grad_norm": 0.25691560926098406, "learning_rate": 9.722761400412496e-06, "loss": 1.0205, "step": 205 }, { "epoch": 0.20114942528735633, "grad_norm": 0.31206698507075104, "learning_rate": 9.694637121067764e-06, "loss": 1.0018, "step": 210 }, { "epoch": 0.20593869731800765, "grad_norm": 0.24725309050346184, "learning_rate": 9.6651991288056e-06, "loss": 1.013, "step": 215 }, { "epoch": 0.210727969348659, "grad_norm": 0.2530356632425841, "learning_rate": 9.63445566133846e-06, "loss": 0.9921, "step": 220 }, { "epoch": 0.21551724137931033, "grad_norm": 0.2784122469884546, "learning_rate": 9.602415321693434e-06, "loss": 1.0066, "step": 225 }, { "epoch": 0.22030651340996169, "grad_norm": 0.3098632552557301, "learning_rate": 9.569087075804842e-06, "loss": 1.0062, "step": 230 }, { "epoch": 0.22509578544061304, "grad_norm": 0.26451308039501314, "learning_rate": 9.534480250005263e-06, "loss": 0.9951, "step": 235 }, { "epoch": 0.22988505747126436, "grad_norm": 0.27301629134605937, "learning_rate": 9.498604528415731e-06, "loss": 1.0347, "step": 240 }, { "epoch": 0.23467432950191572, "grad_norm": 0.2891326880180606, "learning_rate": 9.461469950235795e-06, "loss": 1.0114, "step": 245 }, { "epoch": 0.23946360153256704, "grad_norm": 0.26574747916476, "learning_rate": 9.423086906934228e-06, "loss": 1.0272, "step": 250 }, { "epoch": 0.2442528735632184, "grad_norm": 0.26114773270931335, "learning_rate": 9.38346613934115e-06, "loss": 1.0039, "step": 255 }, { "epoch": 0.24904214559386972, "grad_norm": 0.28383740993290196, "learning_rate": 9.342618734642395e-06, "loss": 1.0077, "step": 260 }, { "epoch": 0.25383141762452105, "grad_norm": 0.28127484887890697, "learning_rate": 9.300556123276955e-06, "loss": 1.0306, "step": 265 }, { "epoch": 0.25862068965517243, "grad_norm": 0.2702807344360773, "learning_rate": 9.257290075738365e-06, "loss": 0.9924, "step": 270 }, { "epoch": 0.26340996168582376, "grad_norm": 0.2912366855364206, "learning_rate": 9.212832699280942e-06, "loss": 1.026, "step": 275 }, { "epoch": 0.2681992337164751, "grad_norm": 0.30617702200806085, "learning_rate": 9.16719643453177e-06, "loss": 1.0247, "step": 280 }, { "epoch": 0.27298850574712646, "grad_norm": 0.2590934020046758, "learning_rate": 9.120394052009412e-06, "loss": 1.0211, "step": 285 }, { "epoch": 0.2777777777777778, "grad_norm": 0.27434400868197933, "learning_rate": 9.072438648550304e-06, "loss": 1.0243, "step": 290 }, { "epoch": 0.2825670498084291, "grad_norm": 0.2813695079919259, "learning_rate": 9.023343643643821e-06, "loss": 1.0008, "step": 295 }, { "epoch": 0.28735632183908044, "grad_norm": 0.2764043745947579, "learning_rate": 8.973122775677078e-06, "loss": 1.0066, "step": 300 }, { "epoch": 0.2921455938697318, "grad_norm": 0.2993769190948342, "learning_rate": 8.921790098090477e-06, "loss": 1.015, "step": 305 }, { "epoch": 0.29693486590038315, "grad_norm": 0.28718834695385625, "learning_rate": 8.869359975445085e-06, "loss": 1.0212, "step": 310 }, { "epoch": 0.3017241379310345, "grad_norm": 0.3787907086651271, "learning_rate": 8.815847079402972e-06, "loss": 1.0079, "step": 315 }, { "epoch": 0.3065134099616858, "grad_norm": 0.3012426621088511, "learning_rate": 8.761266384621599e-06, "loss": 1.0245, "step": 320 }, { "epoch": 0.3113026819923372, "grad_norm": 0.545229044151294, "learning_rate": 8.705633164563413e-06, "loss": 1.0014, "step": 325 }, { "epoch": 0.3160919540229885, "grad_norm": 0.3018913245413703, "learning_rate": 8.648962987221837e-06, "loss": 1.0035, "step": 330 }, { "epoch": 0.32088122605363983, "grad_norm": 0.2937330943277994, "learning_rate": 8.591271710764839e-06, "loss": 0.9932, "step": 335 }, { "epoch": 0.32567049808429116, "grad_norm": 0.30415781861083496, "learning_rate": 8.532575479097294e-06, "loss": 0.982, "step": 340 }, { "epoch": 0.33045977011494254, "grad_norm": 0.2726446228184506, "learning_rate": 8.472890717343391e-06, "loss": 0.9992, "step": 345 }, { "epoch": 0.33524904214559387, "grad_norm": 0.2841460417699516, "learning_rate": 8.412234127250353e-06, "loss": 1.0007, "step": 350 }, { "epoch": 0.3400383141762452, "grad_norm": 0.313253280455998, "learning_rate": 8.350622682514735e-06, "loss": 0.9951, "step": 355 }, { "epoch": 0.3448275862068966, "grad_norm": 0.33772538052784323, "learning_rate": 8.288073624032634e-06, "loss": 1.0169, "step": 360 }, { "epoch": 0.3496168582375479, "grad_norm": 0.2810004404857808, "learning_rate": 8.224604455075115e-06, "loss": 1.0086, "step": 365 }, { "epoch": 0.3544061302681992, "grad_norm": 0.2817546952474833, "learning_rate": 8.160232936390239e-06, "loss": 0.9888, "step": 370 }, { "epoch": 0.35919540229885055, "grad_norm": 0.32316453097441, "learning_rate": 8.094977081233006e-06, "loss": 0.997, "step": 375 }, { "epoch": 0.36398467432950193, "grad_norm": 0.2920029553424864, "learning_rate": 8.02885515032467e-06, "loss": 1.0172, "step": 380 }, { "epoch": 0.36877394636015326, "grad_norm": 0.28736749227532504, "learning_rate": 7.961885646742793e-06, "loss": 1.0092, "step": 385 }, { "epoch": 0.3735632183908046, "grad_norm": 0.29552008976856375, "learning_rate": 7.894087310743468e-06, "loss": 0.9952, "step": 390 }, { "epoch": 0.3783524904214559, "grad_norm": 0.28037261402408636, "learning_rate": 7.825479114517197e-06, "loss": 1.0148, "step": 395 }, { "epoch": 0.3831417624521073, "grad_norm": 0.3259544387907705, "learning_rate": 7.756080256879837e-06, "loss": 1.0172, "step": 400 }, { "epoch": 0.3879310344827586, "grad_norm": 0.2866306787997861, "learning_rate": 7.685910157900158e-06, "loss": 0.9969, "step": 405 }, { "epoch": 0.39272030651340994, "grad_norm": 0.27836252924957877, "learning_rate": 7.614988453465469e-06, "loss": 0.9981, "step": 410 }, { "epoch": 0.3975095785440613, "grad_norm": 0.2881265537652179, "learning_rate": 7.5433349897868445e-06, "loss": 1.0075, "step": 415 }, { "epoch": 0.40229885057471265, "grad_norm": 0.286210264814618, "learning_rate": 7.470969817845518e-06, "loss": 1.0025, "step": 420 }, { "epoch": 0.407088122605364, "grad_norm": 0.28784515054514276, "learning_rate": 7.397913187781962e-06, "loss": 0.9918, "step": 425 }, { "epoch": 0.4118773946360153, "grad_norm": 0.28225636678596183, "learning_rate": 7.324185543229226e-06, "loss": 1.0164, "step": 430 }, { "epoch": 0.4166666666666667, "grad_norm": 0.30273570716493303, "learning_rate": 7.249807515592149e-06, "loss": 0.991, "step": 435 }, { "epoch": 0.421455938697318, "grad_norm": 0.29638416230646264, "learning_rate": 7.174799918274018e-06, "loss": 1.0103, "step": 440 }, { "epoch": 0.42624521072796934, "grad_norm": 0.27796285045314706, "learning_rate": 7.099183740852296e-06, "loss": 0.9929, "step": 445 }, { "epoch": 0.43103448275862066, "grad_norm": 0.3024800017009751, "learning_rate": 7.022980143205046e-06, "loss": 0.9945, "step": 450 }, { "epoch": 0.43582375478927204, "grad_norm": 0.308729839004599, "learning_rate": 6.946210449589714e-06, "loss": 1.0131, "step": 455 }, { "epoch": 0.44061302681992337, "grad_norm": 0.29787252415350113, "learning_rate": 6.868896142675903e-06, "loss": 1.0053, "step": 460 }, { "epoch": 0.4454022988505747, "grad_norm": 0.28555796101935293, "learning_rate": 6.791058857533814e-06, "loss": 1.0106, "step": 465 }, { "epoch": 0.4501915708812261, "grad_norm": 0.27754653040108523, "learning_rate": 6.712720375580057e-06, "loss": 1.0127, "step": 470 }, { "epoch": 0.4549808429118774, "grad_norm": 0.3064883256256046, "learning_rate": 6.633902618482484e-06, "loss": 1.0137, "step": 475 }, { "epoch": 0.45977011494252873, "grad_norm": 0.2871757827962217, "learning_rate": 6.554627642025807e-06, "loss": 0.9808, "step": 480 }, { "epoch": 0.46455938697318006, "grad_norm": 0.32016740057261645, "learning_rate": 6.474917629939652e-06, "loss": 1.0154, "step": 485 }, { "epoch": 0.46934865900383144, "grad_norm": 0.28945157272232624, "learning_rate": 6.394794887690838e-06, "loss": 0.987, "step": 490 }, { "epoch": 0.47413793103448276, "grad_norm": 0.45234637623400176, "learning_rate": 6.314281836241573e-06, "loss": 1.0072, "step": 495 }, { "epoch": 0.4789272030651341, "grad_norm": 0.3043561526142259, "learning_rate": 6.233401005775339e-06, "loss": 0.9947, "step": 500 }, { "epoch": 0.4837164750957854, "grad_norm": 0.3021339898048032, "learning_rate": 6.1521750293922035e-06, "loss": 1.0168, "step": 505 }, { "epoch": 0.4885057471264368, "grad_norm": 0.262840671697266, "learning_rate": 6.070626636775349e-06, "loss": 0.9854, "step": 510 }, { "epoch": 0.4932950191570881, "grad_norm": 0.2717384304425024, "learning_rate": 5.988778647830554e-06, "loss": 0.9847, "step": 515 }, { "epoch": 0.49808429118773945, "grad_norm": 0.3005959006799932, "learning_rate": 5.906653966300444e-06, "loss": 1.0007, "step": 520 }, { "epoch": 0.5028735632183908, "grad_norm": 0.2810709946318669, "learning_rate": 5.824275573355278e-06, "loss": 0.9891, "step": 525 }, { "epoch": 0.5076628352490421, "grad_norm": 0.31324328313914135, "learning_rate": 5.741666521162055e-06, "loss": 1.0049, "step": 530 }, { "epoch": 0.5124521072796935, "grad_norm": 0.29241234641844166, "learning_rate": 5.658849926433774e-06, "loss": 1.0019, "step": 535 }, { "epoch": 0.5172413793103449, "grad_norm": 0.29226941844525284, "learning_rate": 5.575848963960621e-06, "loss": 0.9964, "step": 540 }, { "epoch": 0.5220306513409961, "grad_norm": 0.2869562590824223, "learning_rate": 5.4926868601249e-06, "loss": 1.003, "step": 545 }, { "epoch": 0.5268199233716475, "grad_norm": 0.2887086296052449, "learning_rate": 5.4093868864015405e-06, "loss": 0.9911, "step": 550 }, { "epoch": 0.5316091954022989, "grad_norm": 0.2990975733324071, "learning_rate": 5.325972352845965e-06, "loss": 0.9961, "step": 555 }, { "epoch": 0.5363984674329502, "grad_norm": 0.2964344269886325, "learning_rate": 5.24246660157119e-06, "loss": 1.0045, "step": 560 }, { "epoch": 0.5411877394636015, "grad_norm": 0.3080697315686108, "learning_rate": 5.1588930002159255e-06, "loss": 0.9897, "step": 565 }, { "epoch": 0.5459770114942529, "grad_norm": 0.3217970889035437, "learning_rate": 5.075274935405554e-06, "loss": 1.0022, "step": 570 }, { "epoch": 0.5507662835249042, "grad_norm": 0.32311237689799727, "learning_rate": 4.991635806207788e-06, "loss": 0.9918, "step": 575 }, { "epoch": 0.5555555555555556, "grad_norm": 0.3076150767243147, "learning_rate": 4.90799901758484e-06, "loss": 1.0156, "step": 580 }, { "epoch": 0.5603448275862069, "grad_norm": 0.29579117268352634, "learning_rate": 4.824387973843957e-06, "loss": 0.9859, "step": 585 }, { "epoch": 0.5651340996168582, "grad_norm": 0.27808702972070926, "learning_rate": 4.74082607208812e-06, "loss": 0.988, "step": 590 }, { "epoch": 0.5699233716475096, "grad_norm": 0.2772865889207572, "learning_rate": 4.6573366956687885e-06, "loss": 1.0042, "step": 595 }, { "epoch": 0.5747126436781609, "grad_norm": 0.2741988412251713, "learning_rate": 4.573943207642452e-06, "loss": 1.018, "step": 600 }, { "epoch": 0.5795019157088123, "grad_norm": 0.3041843000888798, "learning_rate": 4.4906689442328935e-06, "loss": 1.0095, "step": 605 }, { "epoch": 0.5842911877394636, "grad_norm": 0.3234627396285633, "learning_rate": 4.407537208300957e-06, "loss": 0.9981, "step": 610 }, { "epoch": 0.5890804597701149, "grad_norm": 0.30361837699199173, "learning_rate": 4.3245712628236356e-06, "loss": 0.9945, "step": 615 }, { "epoch": 0.5938697318007663, "grad_norm": 0.3068819956047309, "learning_rate": 4.241794324384334e-06, "loss": 0.9829, "step": 620 }, { "epoch": 0.5986590038314177, "grad_norm": 0.3010857723276443, "learning_rate": 4.159229556676111e-06, "loss": 0.9778, "step": 625 }, { "epoch": 0.603448275862069, "grad_norm": 0.3212249651416007, "learning_rate": 4.076900064019721e-06, "loss": 1.007, "step": 630 }, { "epoch": 0.6082375478927203, "grad_norm": 0.2842660737481042, "learning_rate": 3.994828884898267e-06, "loss": 1.0056, "step": 635 }, { "epoch": 0.6130268199233716, "grad_norm": 0.2887728882458368, "learning_rate": 3.91303898551028e-06, "loss": 1.0131, "step": 640 }, { "epoch": 0.617816091954023, "grad_norm": 0.31374850764975326, "learning_rate": 3.8315532533430285e-06, "loss": 0.9979, "step": 645 }, { "epoch": 0.6226053639846744, "grad_norm": 0.2990812859204113, "learning_rate": 3.7503944907678543e-06, "loss": 0.9979, "step": 650 }, { "epoch": 0.6273946360153256, "grad_norm": 0.31174935012060845, "learning_rate": 3.6695854086593126e-06, "loss": 0.9907, "step": 655 }, { "epoch": 0.632183908045977, "grad_norm": 0.30150077718156976, "learning_rate": 3.5891486200399413e-06, "loss": 0.9937, "step": 660 }, { "epoch": 0.6369731800766284, "grad_norm": 0.304753149818871, "learning_rate": 3.509106633752387e-06, "loss": 1.0164, "step": 665 }, { "epoch": 0.6417624521072797, "grad_norm": 0.2993950641947178, "learning_rate": 3.429481848160702e-06, "loss": 1.0093, "step": 670 }, { "epoch": 0.646551724137931, "grad_norm": 0.3102107207132909, "learning_rate": 3.350296544882543e-06, "loss": 0.969, "step": 675 }, { "epoch": 0.6513409961685823, "grad_norm": 0.27647028696280174, "learning_rate": 3.2715728825540525e-06, "loss": 1.0102, "step": 680 }, { "epoch": 0.6561302681992337, "grad_norm": 0.30329159137172645, "learning_rate": 3.19333289062915e-06, "loss": 0.9992, "step": 685 }, { "epoch": 0.6609195402298851, "grad_norm": 0.3233989260192753, "learning_rate": 3.1155984632149565e-06, "loss": 0.9984, "step": 690 }, { "epoch": 0.6657088122605364, "grad_norm": 0.284592333087901, "learning_rate": 3.0383913529451286e-06, "loss": 1.0097, "step": 695 }, { "epoch": 0.6704980842911877, "grad_norm": 0.284014773245373, "learning_rate": 2.961733164892744e-06, "loss": 1.0048, "step": 700 }, { "epoch": 0.6752873563218391, "grad_norm": 0.29864923395210635, "learning_rate": 2.8856453505245018e-06, "loss": 1.008, "step": 705 }, { "epoch": 0.6800766283524904, "grad_norm": 0.28455877244795186, "learning_rate": 2.8101492016979027e-06, "loss": 1.0082, "step": 710 }, { "epoch": 0.6848659003831418, "grad_norm": 0.29297293441472344, "learning_rate": 2.7352658447030882e-06, "loss": 1.0137, "step": 715 }, { "epoch": 0.6896551724137931, "grad_norm": 0.29774107321754356, "learning_rate": 2.6610162343510183e-06, "loss": 0.9878, "step": 720 }, { "epoch": 0.6944444444444444, "grad_norm": 0.2886426546973218, "learning_rate": 2.587421148109619e-06, "loss": 0.9855, "step": 725 }, { "epoch": 0.6992337164750958, "grad_norm": 0.30340545161406907, "learning_rate": 2.5145011802895835e-06, "loss": 1.004, "step": 730 }, { "epoch": 0.7040229885057471, "grad_norm": 0.28629744556011416, "learning_rate": 2.4422767362814045e-06, "loss": 0.9935, "step": 735 }, { "epoch": 0.7088122605363985, "grad_norm": 0.29480150488982737, "learning_rate": 2.370768026845276e-06, "loss": 1.0013, "step": 740 }, { "epoch": 0.7136015325670498, "grad_norm": 0.30424158130358797, "learning_rate": 2.299995062455459e-06, "loss": 0.9932, "step": 745 }, { "epoch": 0.7183908045977011, "grad_norm": 0.3179663498265742, "learning_rate": 2.2299776477007073e-06, "loss": 1.007, "step": 750 }, { "epoch": 0.7231800766283525, "grad_norm": 0.303515618658323, "learning_rate": 2.16073537574229e-06, "loss": 0.9963, "step": 755 }, { "epoch": 0.7279693486590039, "grad_norm": 0.30307569076630475, "learning_rate": 2.0922876228311833e-06, "loss": 0.9772, "step": 760 }, { "epoch": 0.7327586206896551, "grad_norm": 0.3143565291039063, "learning_rate": 2.0246535428859652e-06, "loss": 0.9899, "step": 765 }, { "epoch": 0.7375478927203065, "grad_norm": 0.28738666111079514, "learning_rate": 1.957852062132924e-06, "loss": 0.9848, "step": 770 }, { "epoch": 0.7423371647509579, "grad_norm": 0.2850271965375348, "learning_rate": 1.8919018738098704e-06, "loss": 1.0076, "step": 775 }, { "epoch": 0.7471264367816092, "grad_norm": 0.29940367717031596, "learning_rate": 1.8268214329351797e-06, "loss": 0.9864, "step": 780 }, { "epoch": 0.7519157088122606, "grad_norm": 0.2957695153675884, "learning_rate": 1.762628951143454e-06, "loss": 0.9972, "step": 785 }, { "epoch": 0.7567049808429118, "grad_norm": 0.2884019240313734, "learning_rate": 1.6993423915893241e-06, "loss": 0.9969, "step": 790 }, { "epoch": 0.7614942528735632, "grad_norm": 0.30882836660635715, "learning_rate": 1.6369794639207626e-06, "loss": 1.0005, "step": 795 }, { "epoch": 0.7662835249042146, "grad_norm": 0.2880996272538283, "learning_rate": 1.575557619323353e-06, "loss": 0.9853, "step": 800 }, { "epoch": 0.7710727969348659, "grad_norm": 0.28083999464104503, "learning_rate": 1.5150940456368784e-06, "loss": 0.9579, "step": 805 }, { "epoch": 0.7758620689655172, "grad_norm": 0.2977795145319976, "learning_rate": 1.4556056625455922e-06, "loss": 0.9944, "step": 810 }, { "epoch": 0.7806513409961686, "grad_norm": 0.3044689990672244, "learning_rate": 1.3971091168435463e-06, "loss": 0.997, "step": 815 }, { "epoch": 0.7854406130268199, "grad_norm": 0.2951506289424605, "learning_rate": 1.3396207777762732e-06, "loss": 1.0116, "step": 820 }, { "epoch": 0.7902298850574713, "grad_norm": 0.284297818009739, "learning_rate": 1.2831567324601325e-06, "loss": 0.9792, "step": 825 }, { "epoch": 0.7950191570881227, "grad_norm": 0.31450502170794314, "learning_rate": 1.2277327813806123e-06, "loss": 0.9927, "step": 830 }, { "epoch": 0.7998084291187739, "grad_norm": 0.2896076523698672, "learning_rate": 1.173364433970835e-06, "loss": 0.9795, "step": 835 }, { "epoch": 0.8045977011494253, "grad_norm": 0.2834770010415917, "learning_rate": 1.1200669042715163e-06, "loss": 0.9966, "step": 840 }, { "epoch": 0.8093869731800766, "grad_norm": 0.34009482243032485, "learning_rate": 1.0678551066735671e-06, "loss": 0.9767, "step": 845 }, { "epoch": 0.814176245210728, "grad_norm": 0.29936688994813515, "learning_rate": 1.0167436517445777e-06, "loss": 1.003, "step": 850 }, { "epoch": 0.8189655172413793, "grad_norm": 0.29389391810900556, "learning_rate": 9.66746842140287e-07, "loss": 0.9888, "step": 855 }, { "epoch": 0.8237547892720306, "grad_norm": 0.29840441886793445, "learning_rate": 9.178786686022417e-07, "loss": 1.0011, "step": 860 }, { "epoch": 0.828544061302682, "grad_norm": 0.3050673773124508, "learning_rate": 8.701528060427194e-07, "loss": 0.9867, "step": 865 }, { "epoch": 0.8333333333333334, "grad_norm": 0.2809885669064277, "learning_rate": 8.235826097180566e-07, "loss": 0.9802, "step": 870 }, { "epoch": 0.8381226053639846, "grad_norm": 0.29288746432276136, "learning_rate": 7.781811114913995e-07, "loss": 0.9965, "step": 875 }, { "epoch": 0.842911877394636, "grad_norm": 0.29110203494522685, "learning_rate": 7.339610161859618e-07, "loss": 0.9809, "step": 880 }, { "epoch": 0.8477011494252874, "grad_norm": 0.30532295542101273, "learning_rate": 6.909346980298093e-07, "loss": 1.0039, "step": 885 }, { "epoch": 0.8524904214559387, "grad_norm": 0.3178843529727934, "learning_rate": 6.49114197193137e-07, "loss": 0.9992, "step": 890 }, { "epoch": 0.85727969348659, "grad_norm": 0.30030807864509074, "learning_rate": 6.085112164190466e-07, "loss": 0.9967, "step": 895 }, { "epoch": 0.8620689655172413, "grad_norm": 0.283276426877559, "learning_rate": 5.691371177487215e-07, "loss": 0.9951, "step": 900 }, { "epoch": 0.8668582375478927, "grad_norm": 0.2771047728402987, "learning_rate": 5.310029193419697e-07, "loss": 0.9823, "step": 905 }, { "epoch": 0.8716475095785441, "grad_norm": 0.30200927801260424, "learning_rate": 4.941192923939769e-07, "loss": 0.9944, "step": 910 }, { "epoch": 0.8764367816091954, "grad_norm": 0.29243200143497705, "learning_rate": 4.5849655814915683e-07, "loss": 0.9923, "step": 915 }, { "epoch": 0.8812260536398467, "grad_norm": 0.38073762499381975, "learning_rate": 4.2414468501293217e-07, "loss": 0.9931, "step": 920 }, { "epoch": 0.8860153256704981, "grad_norm": 0.28457589633356245, "learning_rate": 3.9107328576224736e-07, "loss": 0.9879, "step": 925 }, { "epoch": 0.8908045977011494, "grad_norm": 0.29913915061920765, "learning_rate": 3.5929161485559694e-07, "loss": 1.0269, "step": 930 }, { "epoch": 0.8955938697318008, "grad_norm": 0.28795408054019833, "learning_rate": 3.2880856584333043e-07, "loss": 0.984, "step": 935 }, { "epoch": 0.9003831417624522, "grad_norm": 0.28456219117461123, "learning_rate": 2.9963266887894526e-07, "loss": 1.0007, "step": 940 }, { "epoch": 0.9051724137931034, "grad_norm": 0.30655155945282825, "learning_rate": 2.717720883320685e-07, "loss": 1.0093, "step": 945 }, { "epoch": 0.9099616858237548, "grad_norm": 0.28653578966574017, "learning_rate": 2.4523462050379864e-07, "loss": 0.9861, "step": 950 }, { "epoch": 0.9147509578544061, "grad_norm": 0.30075605446410747, "learning_rate": 2.2002769144504943e-07, "loss": 0.997, "step": 955 }, { "epoch": 0.9195402298850575, "grad_norm": 0.28984184857418177, "learning_rate": 1.9615835487849677e-07, "loss": 0.9772, "step": 960 }, { "epoch": 0.9243295019157088, "grad_norm": 0.2971882152661699, "learning_rate": 1.7363329022471564e-07, "loss": 1.0125, "step": 965 }, { "epoch": 0.9291187739463601, "grad_norm": 0.30026576782404996, "learning_rate": 1.5245880073305963e-07, "loss": 1.0128, "step": 970 }, { "epoch": 0.9339080459770115, "grad_norm": 0.2971224564720333, "learning_rate": 1.3264081171780797e-07, "loss": 1.0114, "step": 975 }, { "epoch": 0.9386973180076629, "grad_norm": 0.278422035094591, "learning_rate": 1.1418486890006574e-07, "loss": 0.982, "step": 980 }, { "epoch": 0.9434865900383141, "grad_norm": 0.29059990973676236, "learning_rate": 9.709613685589314e-08, "loss": 0.998, "step": 985 }, { "epoch": 0.9482758620689655, "grad_norm": 0.29092073403074437, "learning_rate": 8.137939757108526e-08, "loss": 1.011, "step": 990 }, { "epoch": 0.9530651340996169, "grad_norm": 0.2923297366913393, "learning_rate": 6.703904910301929e-08, "loss": 0.9656, "step": 995 }, { "epoch": 0.9578544061302682, "grad_norm": 0.2931478943843858, "learning_rate": 5.4079104349929465e-08, "loss": 1.0036, "step": 1000 }, { "epoch": 0.9626436781609196, "grad_norm": 0.2923376776086664, "learning_rate": 4.250318992797375e-08, "loss": 1.0083, "step": 1005 }, { "epoch": 0.9674329501915708, "grad_norm": 0.2863912260830555, "learning_rate": 3.231454515638221e-08, "loss": 0.9955, "step": 1010 }, { "epoch": 0.9722222222222222, "grad_norm": 0.2974177186982834, "learning_rate": 2.351602115099272e-08, "loss": 0.9865, "step": 1015 }, { "epoch": 0.9770114942528736, "grad_norm": 0.29561321400349233, "learning_rate": 1.6110080026414123e-08, "loss": 1.0083, "step": 1020 }, { "epoch": 0.9818007662835249, "grad_norm": 0.29932685669861786, "learning_rate": 1.0098794207047402e-08, "loss": 1.0118, "step": 1025 }, { "epoch": 0.9865900383141762, "grad_norm": 0.29155813115628504, "learning_rate": 5.483845847151226e-09, "loss": 0.9846, "step": 1030 }, { "epoch": 0.9913793103448276, "grad_norm": 0.2910040393119768, "learning_rate": 2.2665263601240328e-09, "loss": 0.9812, "step": 1035 }, { "epoch": 0.9961685823754789, "grad_norm": 0.28919827046140006, "learning_rate": 4.4773605712089554e-10, "loss": 1.0115, "step": 1040 }, { "epoch": 1.0, "eval_runtime": 6595.9355, "eval_samples_per_second": 3.504, "eval_steps_per_second": 0.876, "step": 1044 }, { "epoch": 1.0, "step": 1044, "total_flos": 1940427569627136.0, "train_loss": 1.0122476654034465, "train_runtime": 20247.7008, "train_samples_per_second": 3.298, "train_steps_per_second": 0.052 } ], "logging_steps": 5, "max_steps": 1044, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1940427569627136.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }