RefalMachine's picture
Upload folder using huggingface_hub
7157875 verified
raw
history blame
192 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9999558635300348,
"eval_steps": 2000,
"global_step": 11328,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 8.827293993026438e-05,
"eval_accuracy": 0.31068875219818615,
"eval_loss": 5.8817362785339355,
"eval_runtime": 6.9767,
"eval_samples_per_second": 45.58,
"eval_steps_per_second": 0.43,
"step": 1
},
{
"epoch": 0.0008827293993026437,
"grad_norm": 7.96875,
"learning_rate": 5e-05,
"loss": 6.0559,
"step": 10
},
{
"epoch": 0.0017654587986052875,
"grad_norm": 2.015625,
"learning_rate": 0.0001,
"loss": 5.1955,
"step": 20
},
{
"epoch": 0.0026481881979079315,
"grad_norm": 1.21875,
"learning_rate": 0.00015,
"loss": 4.4828,
"step": 30
},
{
"epoch": 0.003530917597210575,
"grad_norm": 0.6640625,
"learning_rate": 0.0002,
"loss": 3.9175,
"step": 40
},
{
"epoch": 0.0044136469965132185,
"grad_norm": 0.36328125,
"learning_rate": 0.00025,
"loss": 3.5422,
"step": 50
},
{
"epoch": 0.005296376395815863,
"grad_norm": 0.298828125,
"learning_rate": 0.0003,
"loss": 3.2998,
"step": 60
},
{
"epoch": 0.0061791057951185065,
"grad_norm": 0.283203125,
"learning_rate": 0.00035,
"loss": 3.1526,
"step": 70
},
{
"epoch": 0.00706183519442115,
"grad_norm": 0.26953125,
"learning_rate": 0.0004,
"loss": 3.0186,
"step": 80
},
{
"epoch": 0.007944564593723794,
"grad_norm": 0.33203125,
"learning_rate": 0.00045000000000000004,
"loss": 2.9449,
"step": 90
},
{
"epoch": 0.008827293993026437,
"grad_norm": 0.38671875,
"learning_rate": 0.0005,
"loss": 2.8837,
"step": 100
},
{
"epoch": 0.009710023392329082,
"grad_norm": 0.36328125,
"learning_rate": 0.0004999990214012265,
"loss": 2.8424,
"step": 110
},
{
"epoch": 0.010592752791631726,
"grad_norm": 0.416015625,
"learning_rate": 0.000499996085612567,
"loss": 2.8241,
"step": 120
},
{
"epoch": 0.01147548219093437,
"grad_norm": 0.345703125,
"learning_rate": 0.0004999911926570055,
"loss": 2.8067,
"step": 130
},
{
"epoch": 0.012358211590237013,
"grad_norm": 0.2578125,
"learning_rate": 0.0004999843425728476,
"loss": 2.7826,
"step": 140
},
{
"epoch": 0.013240940989539656,
"grad_norm": 0.28125,
"learning_rate": 0.0004999755354137212,
"loss": 2.7627,
"step": 150
},
{
"epoch": 0.0141236703888423,
"grad_norm": 0.2734375,
"learning_rate": 0.000499964771248576,
"loss": 2.7586,
"step": 160
},
{
"epoch": 0.015006399788144944,
"grad_norm": 0.275390625,
"learning_rate": 0.000499952050161682,
"loss": 2.7426,
"step": 170
},
{
"epoch": 0.015889129187447587,
"grad_norm": 0.32421875,
"learning_rate": 0.0004999373722526303,
"loss": 2.7326,
"step": 180
},
{
"epoch": 0.016771858586750232,
"grad_norm": 0.388671875,
"learning_rate": 0.0004999207376363309,
"loss": 2.7255,
"step": 190
},
{
"epoch": 0.017654587986052874,
"grad_norm": 0.232421875,
"learning_rate": 0.0004999021464430128,
"loss": 2.7191,
"step": 200
},
{
"epoch": 0.01853731738535552,
"grad_norm": 0.25390625,
"learning_rate": 0.0004998815988182225,
"loss": 2.7228,
"step": 210
},
{
"epoch": 0.019420046784658165,
"grad_norm": 0.3046875,
"learning_rate": 0.0004998590949228232,
"loss": 2.6945,
"step": 220
},
{
"epoch": 0.020302776183960806,
"grad_norm": 0.2734375,
"learning_rate": 0.000499834634932993,
"loss": 2.6945,
"step": 230
},
{
"epoch": 0.02118550558326345,
"grad_norm": 0.263671875,
"learning_rate": 0.0004998082190402241,
"loss": 2.6918,
"step": 240
},
{
"epoch": 0.022068234982566094,
"grad_norm": 0.30078125,
"learning_rate": 0.0004997798474513211,
"loss": 2.6832,
"step": 250
},
{
"epoch": 0.02295096438186874,
"grad_norm": 0.240234375,
"learning_rate": 0.000499749520388399,
"loss": 2.6809,
"step": 260
},
{
"epoch": 0.02383369378117138,
"grad_norm": 0.29296875,
"learning_rate": 0.0004997172380888822,
"loss": 2.6734,
"step": 270
},
{
"epoch": 0.024716423180474026,
"grad_norm": 0.28125,
"learning_rate": 0.0004996830008055017,
"loss": 2.6599,
"step": 280
},
{
"epoch": 0.02559915257977667,
"grad_norm": 0.296875,
"learning_rate": 0.0004996468088062946,
"loss": 2.6679,
"step": 290
},
{
"epoch": 0.026481881979079313,
"grad_norm": 0.212890625,
"learning_rate": 0.0004996086623746,
"loss": 2.6578,
"step": 300
},
{
"epoch": 0.027364611378381958,
"grad_norm": 0.2265625,
"learning_rate": 0.0004995685618090584,
"loss": 2.651,
"step": 310
},
{
"epoch": 0.0282473407776846,
"grad_norm": 0.2490234375,
"learning_rate": 0.0004995265074236088,
"loss": 2.661,
"step": 320
},
{
"epoch": 0.029130070176987245,
"grad_norm": 0.2060546875,
"learning_rate": 0.0004994824995474863,
"loss": 2.6544,
"step": 330
},
{
"epoch": 0.030012799576289887,
"grad_norm": 0.2197265625,
"learning_rate": 0.0004994365385252189,
"loss": 2.6704,
"step": 340
},
{
"epoch": 0.030895528975592532,
"grad_norm": 0.2333984375,
"learning_rate": 0.0004993886247166261,
"loss": 2.6713,
"step": 350
},
{
"epoch": 0.031778258374895174,
"grad_norm": 0.228515625,
"learning_rate": 0.000499338758496815,
"loss": 2.6464,
"step": 360
},
{
"epoch": 0.03266098777419782,
"grad_norm": 0.224609375,
"learning_rate": 0.000499286940256178,
"loss": 2.6611,
"step": 370
},
{
"epoch": 0.033543717173500465,
"grad_norm": 0.2099609375,
"learning_rate": 0.0004992331704003889,
"loss": 2.6427,
"step": 380
},
{
"epoch": 0.03442644657280311,
"grad_norm": 0.291015625,
"learning_rate": 0.0004991774493504007,
"loss": 2.6505,
"step": 390
},
{
"epoch": 0.03530917597210575,
"grad_norm": 0.2216796875,
"learning_rate": 0.0004991197775424418,
"loss": 2.6229,
"step": 400
},
{
"epoch": 0.036191905371408394,
"grad_norm": 0.22265625,
"learning_rate": 0.0004990601554280128,
"loss": 2.643,
"step": 410
},
{
"epoch": 0.03707463477071104,
"grad_norm": 0.330078125,
"learning_rate": 0.0004989985834738824,
"loss": 2.6313,
"step": 420
},
{
"epoch": 0.037957364170013684,
"grad_norm": 0.2109375,
"learning_rate": 0.0004989350621620851,
"loss": 2.6317,
"step": 430
},
{
"epoch": 0.03884009356931633,
"grad_norm": 0.205078125,
"learning_rate": 0.0004988695919899154,
"loss": 2.6456,
"step": 440
},
{
"epoch": 0.03972282296861897,
"grad_norm": 0.25390625,
"learning_rate": 0.0004988021734699258,
"loss": 2.6272,
"step": 450
},
{
"epoch": 0.04060555236792161,
"grad_norm": 0.296875,
"learning_rate": 0.0004987328071299217,
"loss": 2.6182,
"step": 460
},
{
"epoch": 0.04148828176722426,
"grad_norm": 0.232421875,
"learning_rate": 0.0004986614935129576,
"loss": 2.6343,
"step": 470
},
{
"epoch": 0.0423710111665269,
"grad_norm": 0.26953125,
"learning_rate": 0.0004985882331773328,
"loss": 2.614,
"step": 480
},
{
"epoch": 0.04325374056582954,
"grad_norm": 0.361328125,
"learning_rate": 0.0004985130266965871,
"loss": 2.6091,
"step": 490
},
{
"epoch": 0.04413646996513219,
"grad_norm": 0.306640625,
"learning_rate": 0.0004984358746594964,
"loss": 2.6036,
"step": 500
},
{
"epoch": 0.04501919936443483,
"grad_norm": 0.21875,
"learning_rate": 0.0004983567776700676,
"loss": 2.6175,
"step": 510
},
{
"epoch": 0.04590192876373748,
"grad_norm": 0.216796875,
"learning_rate": 0.0004982757363475346,
"loss": 2.6096,
"step": 520
},
{
"epoch": 0.04678465816304012,
"grad_norm": 0.2275390625,
"learning_rate": 0.0004981927513263529,
"loss": 2.6205,
"step": 530
},
{
"epoch": 0.04766738756234276,
"grad_norm": 0.2109375,
"learning_rate": 0.0004981078232561947,
"loss": 2.6116,
"step": 540
},
{
"epoch": 0.048550116961645406,
"grad_norm": 0.1962890625,
"learning_rate": 0.0004980209528019441,
"loss": 2.6113,
"step": 550
},
{
"epoch": 0.04943284636094805,
"grad_norm": 0.216796875,
"learning_rate": 0.0004979321406436917,
"loss": 2.6001,
"step": 560
},
{
"epoch": 0.0503155757602507,
"grad_norm": 0.3203125,
"learning_rate": 0.0004978413874767291,
"loss": 2.6143,
"step": 570
},
{
"epoch": 0.05119830515955334,
"grad_norm": 0.193359375,
"learning_rate": 0.0004977486940115441,
"loss": 2.6178,
"step": 580
},
{
"epoch": 0.05208103455885598,
"grad_norm": 0.2470703125,
"learning_rate": 0.0004976540609738143,
"loss": 2.6071,
"step": 590
},
{
"epoch": 0.052963763958158626,
"grad_norm": 0.345703125,
"learning_rate": 0.0004975574891044017,
"loss": 2.6138,
"step": 600
},
{
"epoch": 0.05384649335746127,
"grad_norm": 0.18359375,
"learning_rate": 0.0004974589791593472,
"loss": 2.5974,
"step": 610
},
{
"epoch": 0.054729222756763916,
"grad_norm": 0.2470703125,
"learning_rate": 0.0004973585319098648,
"loss": 2.6022,
"step": 620
},
{
"epoch": 0.055611952156066555,
"grad_norm": 0.20703125,
"learning_rate": 0.0004972561481423346,
"loss": 2.6128,
"step": 630
},
{
"epoch": 0.0564946815553692,
"grad_norm": 0.2412109375,
"learning_rate": 0.0004971518286582979,
"loss": 2.6068,
"step": 640
},
{
"epoch": 0.057377410954671845,
"grad_norm": 0.21484375,
"learning_rate": 0.0004970455742744499,
"loss": 2.5949,
"step": 650
},
{
"epoch": 0.05826014035397449,
"grad_norm": 0.1904296875,
"learning_rate": 0.0004969373858226341,
"loss": 2.5997,
"step": 660
},
{
"epoch": 0.059142869753277136,
"grad_norm": 0.1943359375,
"learning_rate": 0.0004968272641498349,
"loss": 2.5979,
"step": 670
},
{
"epoch": 0.060025599152579774,
"grad_norm": 0.2080078125,
"learning_rate": 0.0004967152101181717,
"loss": 2.5975,
"step": 680
},
{
"epoch": 0.06090832855188242,
"grad_norm": 0.2421875,
"learning_rate": 0.0004966012246048924,
"loss": 2.5956,
"step": 690
},
{
"epoch": 0.061791057951185065,
"grad_norm": 0.20703125,
"learning_rate": 0.0004964853085023653,
"loss": 2.5859,
"step": 700
},
{
"epoch": 0.0626737873504877,
"grad_norm": 0.1845703125,
"learning_rate": 0.0004963674627180735,
"loss": 2.6004,
"step": 710
},
{
"epoch": 0.06355651674979035,
"grad_norm": 0.29296875,
"learning_rate": 0.0004962476881746068,
"loss": 2.5842,
"step": 720
},
{
"epoch": 0.064439246149093,
"grad_norm": 0.205078125,
"learning_rate": 0.000496125985809655,
"loss": 2.576,
"step": 730
},
{
"epoch": 0.06532197554839564,
"grad_norm": 0.2431640625,
"learning_rate": 0.0004960023565760003,
"loss": 2.5892,
"step": 740
},
{
"epoch": 0.06620470494769828,
"grad_norm": 0.1708984375,
"learning_rate": 0.0004958768014415103,
"loss": 2.585,
"step": 750
},
{
"epoch": 0.06708743434700093,
"grad_norm": 0.1904296875,
"learning_rate": 0.0004957493213891295,
"loss": 2.6027,
"step": 760
},
{
"epoch": 0.06797016374630357,
"grad_norm": 0.23046875,
"learning_rate": 0.0004956199174168725,
"loss": 2.5856,
"step": 770
},
{
"epoch": 0.06885289314560622,
"grad_norm": 0.1953125,
"learning_rate": 0.000495488590537816,
"loss": 2.5701,
"step": 780
},
{
"epoch": 0.06973562254490885,
"grad_norm": 0.208984375,
"learning_rate": 0.0004953553417800905,
"loss": 2.5805,
"step": 790
},
{
"epoch": 0.0706183519442115,
"grad_norm": 0.189453125,
"learning_rate": 0.0004952201721868726,
"loss": 2.5825,
"step": 800
},
{
"epoch": 0.07150108134351414,
"grad_norm": 0.1748046875,
"learning_rate": 0.0004950830828163767,
"loss": 2.588,
"step": 810
},
{
"epoch": 0.07238381074281679,
"grad_norm": 0.20703125,
"learning_rate": 0.0004949440747418467,
"loss": 2.589,
"step": 820
},
{
"epoch": 0.07326654014211943,
"grad_norm": 0.1904296875,
"learning_rate": 0.0004948031490515476,
"loss": 2.5865,
"step": 830
},
{
"epoch": 0.07414926954142208,
"grad_norm": 0.2373046875,
"learning_rate": 0.0004946603068487572,
"loss": 2.5766,
"step": 840
},
{
"epoch": 0.07503199894072472,
"grad_norm": 0.232421875,
"learning_rate": 0.0004945155492517569,
"loss": 2.578,
"step": 850
},
{
"epoch": 0.07591472834002737,
"grad_norm": 0.1826171875,
"learning_rate": 0.0004943688773938237,
"loss": 2.5858,
"step": 860
},
{
"epoch": 0.07679745773933001,
"grad_norm": 0.19921875,
"learning_rate": 0.000494220292423221,
"loss": 2.5783,
"step": 870
},
{
"epoch": 0.07768018713863266,
"grad_norm": 0.1962890625,
"learning_rate": 0.000494069795503189,
"loss": 2.581,
"step": 880
},
{
"epoch": 0.07856291653793529,
"grad_norm": 0.20703125,
"learning_rate": 0.0004939173878119366,
"loss": 2.5741,
"step": 890
},
{
"epoch": 0.07944564593723794,
"grad_norm": 0.1630859375,
"learning_rate": 0.0004937630705426318,
"loss": 2.567,
"step": 900
},
{
"epoch": 0.08032837533654058,
"grad_norm": 0.203125,
"learning_rate": 0.000493606844903392,
"loss": 2.5788,
"step": 910
},
{
"epoch": 0.08121110473584323,
"grad_norm": 0.2314453125,
"learning_rate": 0.000493448712117275,
"loss": 2.5789,
"step": 920
},
{
"epoch": 0.08209383413514587,
"grad_norm": 0.193359375,
"learning_rate": 0.0004932886734222693,
"loss": 2.558,
"step": 930
},
{
"epoch": 0.08297656353444852,
"grad_norm": 0.1865234375,
"learning_rate": 0.000493126730071284,
"loss": 2.5654,
"step": 940
},
{
"epoch": 0.08385929293375116,
"grad_norm": 0.2353515625,
"learning_rate": 0.0004929628833321397,
"loss": 2.5772,
"step": 950
},
{
"epoch": 0.0847420223330538,
"grad_norm": 0.1923828125,
"learning_rate": 0.0004927971344875585,
"loss": 2.5739,
"step": 960
},
{
"epoch": 0.08562475173235645,
"grad_norm": 0.2197265625,
"learning_rate": 0.0004926294848351528,
"loss": 2.5725,
"step": 970
},
{
"epoch": 0.08650748113165908,
"grad_norm": 0.1875,
"learning_rate": 0.0004924599356874169,
"loss": 2.5709,
"step": 980
},
{
"epoch": 0.08739021053096173,
"grad_norm": 0.2431640625,
"learning_rate": 0.0004922884883717154,
"loss": 2.557,
"step": 990
},
{
"epoch": 0.08827293993026437,
"grad_norm": 0.1806640625,
"learning_rate": 0.0004921151442302732,
"loss": 2.5722,
"step": 1000
},
{
"epoch": 0.08915566932956702,
"grad_norm": 0.181640625,
"learning_rate": 0.0004919399046201656,
"loss": 2.567,
"step": 1010
},
{
"epoch": 0.09003839872886966,
"grad_norm": 0.2197265625,
"learning_rate": 0.0004917627709133064,
"loss": 2.5629,
"step": 1020
},
{
"epoch": 0.09092112812817231,
"grad_norm": 0.1708984375,
"learning_rate": 0.0004915837444964383,
"loss": 2.5806,
"step": 1030
},
{
"epoch": 0.09180385752747496,
"grad_norm": 0.173828125,
"learning_rate": 0.0004914028267711217,
"loss": 2.5642,
"step": 1040
},
{
"epoch": 0.0926865869267776,
"grad_norm": 0.203125,
"learning_rate": 0.0004912200191537233,
"loss": 2.5789,
"step": 1050
},
{
"epoch": 0.09356931632608025,
"grad_norm": 0.16796875,
"learning_rate": 0.0004910353230754057,
"loss": 2.5657,
"step": 1060
},
{
"epoch": 0.09445204572538288,
"grad_norm": 0.2314453125,
"learning_rate": 0.0004908487399821158,
"loss": 2.5721,
"step": 1070
},
{
"epoch": 0.09533477512468552,
"grad_norm": 0.224609375,
"learning_rate": 0.0004906602713345735,
"loss": 2.5663,
"step": 1080
},
{
"epoch": 0.09621750452398817,
"grad_norm": 0.1865234375,
"learning_rate": 0.0004904699186082602,
"loss": 2.5603,
"step": 1090
},
{
"epoch": 0.09710023392329081,
"grad_norm": 0.1650390625,
"learning_rate": 0.0004902776832934074,
"loss": 2.5647,
"step": 1100
},
{
"epoch": 0.09798296332259346,
"grad_norm": 0.201171875,
"learning_rate": 0.0004900835668949852,
"loss": 2.5559,
"step": 1110
},
{
"epoch": 0.0988656927218961,
"grad_norm": 0.1865234375,
"learning_rate": 0.00048988757093269,
"loss": 2.5599,
"step": 1120
},
{
"epoch": 0.09974842212119875,
"grad_norm": 0.1748046875,
"learning_rate": 0.0004896896969409332,
"loss": 2.5621,
"step": 1130
},
{
"epoch": 0.1006311515205014,
"grad_norm": 0.1708984375,
"learning_rate": 0.0004894899464688287,
"loss": 2.5699,
"step": 1140
},
{
"epoch": 0.10151388091980404,
"grad_norm": 0.1591796875,
"learning_rate": 0.000489288321080181,
"loss": 2.566,
"step": 1150
},
{
"epoch": 0.10239661031910668,
"grad_norm": 0.19921875,
"learning_rate": 0.0004890848223534732,
"loss": 2.5834,
"step": 1160
},
{
"epoch": 0.10327933971840932,
"grad_norm": 0.1796875,
"learning_rate": 0.0004888794518818538,
"loss": 2.5496,
"step": 1170
},
{
"epoch": 0.10416206911771196,
"grad_norm": 0.1572265625,
"learning_rate": 0.0004886722112731253,
"loss": 2.5583,
"step": 1180
},
{
"epoch": 0.1050447985170146,
"grad_norm": 0.181640625,
"learning_rate": 0.000488463102149731,
"loss": 2.564,
"step": 1190
},
{
"epoch": 0.10592752791631725,
"grad_norm": 0.2138671875,
"learning_rate": 0.0004882521261487422,
"loss": 2.5742,
"step": 1200
},
{
"epoch": 0.1068102573156199,
"grad_norm": 0.197265625,
"learning_rate": 0.0004880392849218459,
"loss": 2.5743,
"step": 1210
},
{
"epoch": 0.10769298671492254,
"grad_norm": 0.205078125,
"learning_rate": 0.00048782458013533125,
"loss": 2.5613,
"step": 1220
},
{
"epoch": 0.10857571611422519,
"grad_norm": 0.2197265625,
"learning_rate": 0.00048760801347007716,
"loss": 2.5522,
"step": 1230
},
{
"epoch": 0.10945844551352783,
"grad_norm": 0.1845703125,
"learning_rate": 0.0004873895866215385,
"loss": 2.5649,
"step": 1240
},
{
"epoch": 0.11034117491283048,
"grad_norm": 0.208984375,
"learning_rate": 0.00048716930129973323,
"loss": 2.5571,
"step": 1250
},
{
"epoch": 0.11122390431213311,
"grad_norm": 0.2255859375,
"learning_rate": 0.0004869471592292289,
"loss": 2.5664,
"step": 1260
},
{
"epoch": 0.11210663371143575,
"grad_norm": 0.2119140625,
"learning_rate": 0.0004867231621491293,
"loss": 2.5603,
"step": 1270
},
{
"epoch": 0.1129893631107384,
"grad_norm": 0.16015625,
"learning_rate": 0.00048649731181306047,
"loss": 2.5477,
"step": 1280
},
{
"epoch": 0.11387209251004105,
"grad_norm": 0.2451171875,
"learning_rate": 0.00048626960998915733,
"loss": 2.5609,
"step": 1290
},
{
"epoch": 0.11475482190934369,
"grad_norm": 0.1796875,
"learning_rate": 0.0004860400584600496,
"loss": 2.5668,
"step": 1300
},
{
"epoch": 0.11563755130864634,
"grad_norm": 0.1796875,
"learning_rate": 0.0004858086590228482,
"loss": 2.5504,
"step": 1310
},
{
"epoch": 0.11652028070794898,
"grad_norm": 0.171875,
"learning_rate": 0.0004855754134891307,
"loss": 2.5613,
"step": 1320
},
{
"epoch": 0.11740301010725163,
"grad_norm": 0.208984375,
"learning_rate": 0.0004853403236849274,
"loss": 2.5539,
"step": 1330
},
{
"epoch": 0.11828573950655427,
"grad_norm": 0.197265625,
"learning_rate": 0.0004851033914507071,
"loss": 2.5613,
"step": 1340
},
{
"epoch": 0.1191684689058569,
"grad_norm": 0.1796875,
"learning_rate": 0.00048486461864136253,
"loss": 2.5599,
"step": 1350
},
{
"epoch": 0.12005119830515955,
"grad_norm": 0.1787109375,
"learning_rate": 0.0004846240071261959,
"loss": 2.5392,
"step": 1360
},
{
"epoch": 0.1209339277044622,
"grad_norm": 0.17578125,
"learning_rate": 0.00048438155878890434,
"loss": 2.5409,
"step": 1370
},
{
"epoch": 0.12181665710376484,
"grad_norm": 0.1669921875,
"learning_rate": 0.00048413727552756505,
"loss": 2.5521,
"step": 1380
},
{
"epoch": 0.12269938650306748,
"grad_norm": 0.1572265625,
"learning_rate": 0.00048389115925462025,
"loss": 2.5431,
"step": 1390
},
{
"epoch": 0.12358211590237013,
"grad_norm": 0.2353515625,
"learning_rate": 0.00048364321189686276,
"loss": 2.5519,
"step": 1400
},
{
"epoch": 0.12446484530167277,
"grad_norm": 0.1669921875,
"learning_rate": 0.00048339343539542033,
"loss": 2.5424,
"step": 1410
},
{
"epoch": 0.1253475747009754,
"grad_norm": 0.162109375,
"learning_rate": 0.0004831418317057409,
"loss": 2.5411,
"step": 1420
},
{
"epoch": 0.12623030410027805,
"grad_norm": 0.169921875,
"learning_rate": 0.0004828884027975768,
"loss": 2.5336,
"step": 1430
},
{
"epoch": 0.1271130334995807,
"grad_norm": 0.259765625,
"learning_rate": 0.00048263315065497,
"loss": 2.5506,
"step": 1440
},
{
"epoch": 0.12799576289888334,
"grad_norm": 0.16796875,
"learning_rate": 0.0004823760772762358,
"loss": 2.5433,
"step": 1450
},
{
"epoch": 0.128878492298186,
"grad_norm": 0.1513671875,
"learning_rate": 0.00048211718467394774,
"loss": 2.5515,
"step": 1460
},
{
"epoch": 0.12976122169748863,
"grad_norm": 0.181640625,
"learning_rate": 0.0004818564748749218,
"loss": 2.5384,
"step": 1470
},
{
"epoch": 0.13064395109679128,
"grad_norm": 0.1748046875,
"learning_rate": 0.0004815939499202001,
"loss": 2.553,
"step": 1480
},
{
"epoch": 0.13152668049609392,
"grad_norm": 0.162109375,
"learning_rate": 0.0004813296118650357,
"loss": 2.5587,
"step": 1490
},
{
"epoch": 0.13240940989539657,
"grad_norm": 0.1806640625,
"learning_rate": 0.0004810634627788756,
"loss": 2.543,
"step": 1500
},
{
"epoch": 0.1332921392946992,
"grad_norm": 0.16796875,
"learning_rate": 0.0004807955047453452,
"loss": 2.5504,
"step": 1510
},
{
"epoch": 0.13417486869400186,
"grad_norm": 0.158203125,
"learning_rate": 0.0004805257398622317,
"loss": 2.5478,
"step": 1520
},
{
"epoch": 0.1350575980933045,
"grad_norm": 0.185546875,
"learning_rate": 0.0004802541702414678,
"loss": 2.5459,
"step": 1530
},
{
"epoch": 0.13594032749260715,
"grad_norm": 0.1923828125,
"learning_rate": 0.000479980798009115,
"loss": 2.545,
"step": 1540
},
{
"epoch": 0.1368230568919098,
"grad_norm": 0.1689453125,
"learning_rate": 0.00047970562530534724,
"loss": 2.5507,
"step": 1550
},
{
"epoch": 0.13770578629121244,
"grad_norm": 0.2236328125,
"learning_rate": 0.0004794286542844338,
"loss": 2.5439,
"step": 1560
},
{
"epoch": 0.13858851569051509,
"grad_norm": 0.1728515625,
"learning_rate": 0.00047914988711472283,
"loss": 2.5482,
"step": 1570
},
{
"epoch": 0.1394712450898177,
"grad_norm": 0.19140625,
"learning_rate": 0.00047886932597862396,
"loss": 2.5358,
"step": 1580
},
{
"epoch": 0.14035397448912035,
"grad_norm": 0.1552734375,
"learning_rate": 0.0004785869730725914,
"loss": 2.5476,
"step": 1590
},
{
"epoch": 0.141236703888423,
"grad_norm": 0.1865234375,
"learning_rate": 0.0004783028306071069,
"loss": 2.5427,
"step": 1600
},
{
"epoch": 0.14211943328772564,
"grad_norm": 0.1640625,
"learning_rate": 0.00047801690080666206,
"loss": 2.5343,
"step": 1610
},
{
"epoch": 0.14300216268702828,
"grad_norm": 0.1728515625,
"learning_rate": 0.00047772918590974136,
"loss": 2.5404,
"step": 1620
},
{
"epoch": 0.14388489208633093,
"grad_norm": 0.1640625,
"learning_rate": 0.00047743968816880446,
"loss": 2.5495,
"step": 1630
},
{
"epoch": 0.14476762148563357,
"grad_norm": 0.19140625,
"learning_rate": 0.0004771484098502683,
"loss": 2.5438,
"step": 1640
},
{
"epoch": 0.14565035088493622,
"grad_norm": 0.158203125,
"learning_rate": 0.0004768553532344899,
"loss": 2.5342,
"step": 1650
},
{
"epoch": 0.14653308028423886,
"grad_norm": 0.2021484375,
"learning_rate": 0.0004765605206157478,
"loss": 2.541,
"step": 1660
},
{
"epoch": 0.1474158096835415,
"grad_norm": 0.1826171875,
"learning_rate": 0.0004762639143022248,
"loss": 2.5504,
"step": 1670
},
{
"epoch": 0.14829853908284416,
"grad_norm": 0.169921875,
"learning_rate": 0.00047596553661598956,
"loss": 2.5269,
"step": 1680
},
{
"epoch": 0.1491812684821468,
"grad_norm": 0.1650390625,
"learning_rate": 0.00047566538989297837,
"loss": 2.5442,
"step": 1690
},
{
"epoch": 0.15006399788144945,
"grad_norm": 0.16796875,
"learning_rate": 0.00047536347648297685,
"loss": 2.5452,
"step": 1700
},
{
"epoch": 0.1509467272807521,
"grad_norm": 0.236328125,
"learning_rate": 0.0004750597987496018,
"loss": 2.5453,
"step": 1710
},
{
"epoch": 0.15182945668005474,
"grad_norm": 0.1728515625,
"learning_rate": 0.00047475435907028254,
"loss": 2.5416,
"step": 1720
},
{
"epoch": 0.15271218607935738,
"grad_norm": 0.212890625,
"learning_rate": 0.0004744471598362421,
"loss": 2.5395,
"step": 1730
},
{
"epoch": 0.15359491547866003,
"grad_norm": 0.1806640625,
"learning_rate": 0.0004741382034524789,
"loss": 2.5423,
"step": 1740
},
{
"epoch": 0.15447764487796267,
"grad_norm": 0.1650390625,
"learning_rate": 0.0004738274923377478,
"loss": 2.5322,
"step": 1750
},
{
"epoch": 0.15536037427726532,
"grad_norm": 0.1513671875,
"learning_rate": 0.0004735150289245407,
"loss": 2.5331,
"step": 1760
},
{
"epoch": 0.15624310367656793,
"grad_norm": 0.2041015625,
"learning_rate": 0.00047320081565906813,
"loss": 2.5491,
"step": 1770
},
{
"epoch": 0.15712583307587058,
"grad_norm": 0.171875,
"learning_rate": 0.0004728848550012399,
"loss": 2.5477,
"step": 1780
},
{
"epoch": 0.15800856247517323,
"grad_norm": 0.154296875,
"learning_rate": 0.00047256714942464574,
"loss": 2.5364,
"step": 1790
},
{
"epoch": 0.15889129187447587,
"grad_norm": 0.1630859375,
"learning_rate": 0.0004722477014165358,
"loss": 2.5308,
"step": 1800
},
{
"epoch": 0.15977402127377852,
"grad_norm": 0.154296875,
"learning_rate": 0.0004719265134778017,
"loss": 2.5375,
"step": 1810
},
{
"epoch": 0.16065675067308116,
"grad_norm": 0.1796875,
"learning_rate": 0.00047160358812295633,
"loss": 2.5245,
"step": 1820
},
{
"epoch": 0.1615394800723838,
"grad_norm": 0.20703125,
"learning_rate": 0.0004712789278801145,
"loss": 2.5464,
"step": 1830
},
{
"epoch": 0.16242220947168645,
"grad_norm": 0.1513671875,
"learning_rate": 0.00047095253529097313,
"loss": 2.5394,
"step": 1840
},
{
"epoch": 0.1633049388709891,
"grad_norm": 0.1669921875,
"learning_rate": 0.0004706244129107914,
"loss": 2.5325,
"step": 1850
},
{
"epoch": 0.16418766827029174,
"grad_norm": 0.15234375,
"learning_rate": 0.00047029456330837055,
"loss": 2.5355,
"step": 1860
},
{
"epoch": 0.1650703976695944,
"grad_norm": 0.173828125,
"learning_rate": 0.0004699629890660339,
"loss": 2.5369,
"step": 1870
},
{
"epoch": 0.16595312706889703,
"grad_norm": 0.1669921875,
"learning_rate": 0.00046962969277960663,
"loss": 2.546,
"step": 1880
},
{
"epoch": 0.16683585646819968,
"grad_norm": 0.1865234375,
"learning_rate": 0.00046929467705839544,
"loss": 2.5437,
"step": 1890
},
{
"epoch": 0.16771858586750232,
"grad_norm": 0.1611328125,
"learning_rate": 0.0004689579445251681,
"loss": 2.5414,
"step": 1900
},
{
"epoch": 0.16860131526680497,
"grad_norm": 0.1552734375,
"learning_rate": 0.000468619497816133,
"loss": 2.548,
"step": 1910
},
{
"epoch": 0.1694840446661076,
"grad_norm": 0.2041015625,
"learning_rate": 0.0004682793395809184,
"loss": 2.5408,
"step": 1920
},
{
"epoch": 0.17036677406541026,
"grad_norm": 0.166015625,
"learning_rate": 0.0004679374724825517,
"loss": 2.5349,
"step": 1930
},
{
"epoch": 0.1712495034647129,
"grad_norm": 0.1513671875,
"learning_rate": 0.00046759389919743876,
"loss": 2.5381,
"step": 1940
},
{
"epoch": 0.17213223286401555,
"grad_norm": 0.1640625,
"learning_rate": 0.0004672486224153427,
"loss": 2.5384,
"step": 1950
},
{
"epoch": 0.17301496226331817,
"grad_norm": 0.1533203125,
"learning_rate": 0.0004669016448393631,
"loss": 2.5296,
"step": 1960
},
{
"epoch": 0.1738976916626208,
"grad_norm": 0.1533203125,
"learning_rate": 0.0004665529691859144,
"loss": 2.5334,
"step": 1970
},
{
"epoch": 0.17478042106192346,
"grad_norm": 0.203125,
"learning_rate": 0.00046620259818470536,
"loss": 2.5341,
"step": 1980
},
{
"epoch": 0.1756631504612261,
"grad_norm": 0.185546875,
"learning_rate": 0.0004658505345787169,
"loss": 2.5369,
"step": 1990
},
{
"epoch": 0.17654587986052875,
"grad_norm": 0.1533203125,
"learning_rate": 0.00046549678112418116,
"loss": 2.556,
"step": 2000
},
{
"epoch": 0.17654587986052875,
"eval_accuracy": 0.5045637800354782,
"eval_loss": 2.4191172122955322,
"eval_runtime": 6.9583,
"eval_samples_per_second": 45.701,
"eval_steps_per_second": 0.431,
"step": 2000
},
{
"epoch": 0.1774286092598314,
"grad_norm": 0.181640625,
"learning_rate": 0.0004651413405905597,
"loss": 2.5262,
"step": 2010
},
{
"epoch": 0.17831133865913404,
"grad_norm": 0.1796875,
"learning_rate": 0.00046478421576052196,
"loss": 2.5394,
"step": 2020
},
{
"epoch": 0.17919406805843668,
"grad_norm": 0.2255859375,
"learning_rate": 0.00046442540942992315,
"loss": 2.5327,
"step": 2030
},
{
"epoch": 0.18007679745773933,
"grad_norm": 0.1865234375,
"learning_rate": 0.00046406492440778294,
"loss": 2.5218,
"step": 2040
},
{
"epoch": 0.18095952685704197,
"grad_norm": 0.1806640625,
"learning_rate": 0.0004637027635162627,
"loss": 2.5349,
"step": 2050
},
{
"epoch": 0.18184225625634462,
"grad_norm": 0.1708984375,
"learning_rate": 0.00046333892959064425,
"loss": 2.5356,
"step": 2060
},
{
"epoch": 0.18272498565564727,
"grad_norm": 0.21484375,
"learning_rate": 0.0004629734254793071,
"loss": 2.5306,
"step": 2070
},
{
"epoch": 0.1836077150549499,
"grad_norm": 0.15625,
"learning_rate": 0.00046260625404370606,
"loss": 2.5442,
"step": 2080
},
{
"epoch": 0.18449044445425256,
"grad_norm": 0.154296875,
"learning_rate": 0.0004622374181583494,
"loss": 2.52,
"step": 2090
},
{
"epoch": 0.1853731738535552,
"grad_norm": 0.1650390625,
"learning_rate": 0.00046186692071077586,
"loss": 2.5189,
"step": 2100
},
{
"epoch": 0.18625590325285785,
"grad_norm": 0.1787109375,
"learning_rate": 0.00046149476460153216,
"loss": 2.5297,
"step": 2110
},
{
"epoch": 0.1871386326521605,
"grad_norm": 0.1669921875,
"learning_rate": 0.0004611209527441504,
"loss": 2.5336,
"step": 2120
},
{
"epoch": 0.18802136205146314,
"grad_norm": 0.173828125,
"learning_rate": 0.0004607454880651253,
"loss": 2.5322,
"step": 2130
},
{
"epoch": 0.18890409145076575,
"grad_norm": 0.1396484375,
"learning_rate": 0.0004603683735038909,
"loss": 2.5361,
"step": 2140
},
{
"epoch": 0.1897868208500684,
"grad_norm": 0.16015625,
"learning_rate": 0.00045998961201279814,
"loss": 2.5197,
"step": 2150
},
{
"epoch": 0.19066955024937104,
"grad_norm": 0.173828125,
"learning_rate": 0.00045960920655709113,
"loss": 2.5213,
"step": 2160
},
{
"epoch": 0.1915522796486737,
"grad_norm": 0.1806640625,
"learning_rate": 0.0004592271601148844,
"loss": 2.5096,
"step": 2170
},
{
"epoch": 0.19243500904797634,
"grad_norm": 0.234375,
"learning_rate": 0.00045884347567713945,
"loss": 2.5222,
"step": 2180
},
{
"epoch": 0.19331773844727898,
"grad_norm": 0.1728515625,
"learning_rate": 0.0004584581562476412,
"loss": 2.5463,
"step": 2190
},
{
"epoch": 0.19420046784658163,
"grad_norm": 0.1669921875,
"learning_rate": 0.0004580712048429746,
"loss": 2.5331,
"step": 2200
},
{
"epoch": 0.19508319724588427,
"grad_norm": 0.154296875,
"learning_rate": 0.000457682624492501,
"loss": 2.5172,
"step": 2210
},
{
"epoch": 0.19596592664518692,
"grad_norm": 0.1806640625,
"learning_rate": 0.0004572924182383346,
"loss": 2.5282,
"step": 2220
},
{
"epoch": 0.19684865604448956,
"grad_norm": 0.18359375,
"learning_rate": 0.00045690058913531794,
"loss": 2.5318,
"step": 2230
},
{
"epoch": 0.1977313854437922,
"grad_norm": 0.16015625,
"learning_rate": 0.0004565071402509992,
"loss": 2.5202,
"step": 2240
},
{
"epoch": 0.19861411484309485,
"grad_norm": 0.1669921875,
"learning_rate": 0.000456112074665607,
"loss": 2.5342,
"step": 2250
},
{
"epoch": 0.1994968442423975,
"grad_norm": 0.162109375,
"learning_rate": 0.0004557153954720269,
"loss": 2.5226,
"step": 2260
},
{
"epoch": 0.20037957364170014,
"grad_norm": 0.16796875,
"learning_rate": 0.0004553171057757772,
"loss": 2.535,
"step": 2270
},
{
"epoch": 0.2012623030410028,
"grad_norm": 0.150390625,
"learning_rate": 0.0004549172086949842,
"loss": 2.5182,
"step": 2280
},
{
"epoch": 0.20214503244030543,
"grad_norm": 0.162109375,
"learning_rate": 0.0004545157073603584,
"loss": 2.5335,
"step": 2290
},
{
"epoch": 0.20302776183960808,
"grad_norm": 0.1787109375,
"learning_rate": 0.0004541126049151694,
"loss": 2.545,
"step": 2300
},
{
"epoch": 0.20391049123891072,
"grad_norm": 0.1689453125,
"learning_rate": 0.00045370790451522165,
"loss": 2.5158,
"step": 2310
},
{
"epoch": 0.20479322063821337,
"grad_norm": 0.150390625,
"learning_rate": 0.0004533016093288298,
"loss": 2.5105,
"step": 2320
},
{
"epoch": 0.205675950037516,
"grad_norm": 0.2109375,
"learning_rate": 0.0004528937225367935,
"loss": 2.5305,
"step": 2330
},
{
"epoch": 0.20655867943681863,
"grad_norm": 0.1611328125,
"learning_rate": 0.0004524842473323729,
"loss": 2.5335,
"step": 2340
},
{
"epoch": 0.20744140883612128,
"grad_norm": 0.1396484375,
"learning_rate": 0.0004520731869212634,
"loss": 2.5196,
"step": 2350
},
{
"epoch": 0.20832413823542392,
"grad_norm": 0.1591796875,
"learning_rate": 0.0004516605445215709,
"loss": 2.5201,
"step": 2360
},
{
"epoch": 0.20920686763472657,
"grad_norm": 0.185546875,
"learning_rate": 0.00045124632336378603,
"loss": 2.5195,
"step": 2370
},
{
"epoch": 0.2100895970340292,
"grad_norm": 0.15625,
"learning_rate": 0.00045083052669075936,
"loss": 2.5264,
"step": 2380
},
{
"epoch": 0.21097232643333186,
"grad_norm": 0.197265625,
"learning_rate": 0.0004504131577576758,
"loss": 2.5283,
"step": 2390
},
{
"epoch": 0.2118550558326345,
"grad_norm": 0.18359375,
"learning_rate": 0.00044999421983202905,
"loss": 2.5266,
"step": 2400
},
{
"epoch": 0.21273778523193715,
"grad_norm": 0.1455078125,
"learning_rate": 0.00044957371619359644,
"loss": 2.5356,
"step": 2410
},
{
"epoch": 0.2136205146312398,
"grad_norm": 0.1513671875,
"learning_rate": 0.00044915165013441257,
"loss": 2.5281,
"step": 2420
},
{
"epoch": 0.21450324403054244,
"grad_norm": 0.1533203125,
"learning_rate": 0.0004487280249587441,
"loss": 2.5335,
"step": 2430
},
{
"epoch": 0.21538597342984508,
"grad_norm": 0.1396484375,
"learning_rate": 0.00044830284398306375,
"loss": 2.5303,
"step": 2440
},
{
"epoch": 0.21626870282914773,
"grad_norm": 0.1455078125,
"learning_rate": 0.000447876110536024,
"loss": 2.5297,
"step": 2450
},
{
"epoch": 0.21715143222845038,
"grad_norm": 0.1513671875,
"learning_rate": 0.0004474478279584316,
"loss": 2.5287,
"step": 2460
},
{
"epoch": 0.21803416162775302,
"grad_norm": 0.185546875,
"learning_rate": 0.00044701799960322085,
"loss": 2.525,
"step": 2470
},
{
"epoch": 0.21891689102705567,
"grad_norm": 0.150390625,
"learning_rate": 0.000446586628835428,
"loss": 2.5278,
"step": 2480
},
{
"epoch": 0.2197996204263583,
"grad_norm": 0.138671875,
"learning_rate": 0.00044615371903216407,
"loss": 2.5087,
"step": 2490
},
{
"epoch": 0.22068234982566096,
"grad_norm": 0.140625,
"learning_rate": 0.00044571927358258917,
"loss": 2.5294,
"step": 2500
},
{
"epoch": 0.22156507922496357,
"grad_norm": 0.1767578125,
"learning_rate": 0.0004452832958878856,
"loss": 2.5296,
"step": 2510
},
{
"epoch": 0.22244780862426622,
"grad_norm": 0.16796875,
"learning_rate": 0.0004448457893612311,
"loss": 2.5267,
"step": 2520
},
{
"epoch": 0.22333053802356886,
"grad_norm": 0.1474609375,
"learning_rate": 0.0004444067574277727,
"loss": 2.5221,
"step": 2530
},
{
"epoch": 0.2242132674228715,
"grad_norm": 0.154296875,
"learning_rate": 0.00044396620352459915,
"loss": 2.5178,
"step": 2540
},
{
"epoch": 0.22509599682217415,
"grad_norm": 0.1494140625,
"learning_rate": 0.00044352413110071453,
"loss": 2.5104,
"step": 2550
},
{
"epoch": 0.2259787262214768,
"grad_norm": 0.1845703125,
"learning_rate": 0.0004430805436170111,
"loss": 2.5262,
"step": 2560
},
{
"epoch": 0.22686145562077945,
"grad_norm": 0.1650390625,
"learning_rate": 0.00044263544454624224,
"loss": 2.5204,
"step": 2570
},
{
"epoch": 0.2277441850200821,
"grad_norm": 0.1533203125,
"learning_rate": 0.00044218883737299526,
"loss": 2.5159,
"step": 2580
},
{
"epoch": 0.22862691441938474,
"grad_norm": 0.173828125,
"learning_rate": 0.00044174072559366386,
"loss": 2.5133,
"step": 2590
},
{
"epoch": 0.22950964381868738,
"grad_norm": 0.173828125,
"learning_rate": 0.00044129111271642117,
"loss": 2.5293,
"step": 2600
},
{
"epoch": 0.23039237321799003,
"grad_norm": 0.1748046875,
"learning_rate": 0.0004408400022611921,
"loss": 2.511,
"step": 2610
},
{
"epoch": 0.23127510261729267,
"grad_norm": 0.158203125,
"learning_rate": 0.00044038739775962584,
"loss": 2.5088,
"step": 2620
},
{
"epoch": 0.23215783201659532,
"grad_norm": 0.1494140625,
"learning_rate": 0.0004399333027550679,
"loss": 2.5066,
"step": 2630
},
{
"epoch": 0.23304056141589796,
"grad_norm": 0.16015625,
"learning_rate": 0.000439477720802533,
"loss": 2.5231,
"step": 2640
},
{
"epoch": 0.2339232908152006,
"grad_norm": 0.16796875,
"learning_rate": 0.00043902065546867655,
"loss": 2.5164,
"step": 2650
},
{
"epoch": 0.23480602021450325,
"grad_norm": 0.140625,
"learning_rate": 0.0004385621103317671,
"loss": 2.511,
"step": 2660
},
{
"epoch": 0.2356887496138059,
"grad_norm": 0.1484375,
"learning_rate": 0.00043810208898165836,
"loss": 2.5049,
"step": 2670
},
{
"epoch": 0.23657147901310854,
"grad_norm": 0.1611328125,
"learning_rate": 0.000437640595019761,
"loss": 2.5262,
"step": 2680
},
{
"epoch": 0.2374542084124112,
"grad_norm": 0.181640625,
"learning_rate": 0.00043717763205901436,
"loss": 2.5205,
"step": 2690
},
{
"epoch": 0.2383369378117138,
"grad_norm": 0.1376953125,
"learning_rate": 0.00043671320372385834,
"loss": 2.5136,
"step": 2700
},
{
"epoch": 0.23921966721101645,
"grad_norm": 0.1435546875,
"learning_rate": 0.00043624731365020505,
"loss": 2.5185,
"step": 2710
},
{
"epoch": 0.2401023966103191,
"grad_norm": 0.158203125,
"learning_rate": 0.00043577996548541,
"loss": 2.5148,
"step": 2720
},
{
"epoch": 0.24098512600962174,
"grad_norm": 0.1484375,
"learning_rate": 0.00043531116288824393,
"loss": 2.5225,
"step": 2730
},
{
"epoch": 0.2418678554089244,
"grad_norm": 0.1689453125,
"learning_rate": 0.00043484090952886404,
"loss": 2.524,
"step": 2740
},
{
"epoch": 0.24275058480822703,
"grad_norm": 0.162109375,
"learning_rate": 0.0004343692090887852,
"loss": 2.5035,
"step": 2750
},
{
"epoch": 0.24363331420752968,
"grad_norm": 0.1396484375,
"learning_rate": 0.0004338960652608511,
"loss": 2.5136,
"step": 2760
},
{
"epoch": 0.24451604360683232,
"grad_norm": 0.1416015625,
"learning_rate": 0.0004334214817492057,
"loss": 2.4963,
"step": 2770
},
{
"epoch": 0.24539877300613497,
"grad_norm": 0.1494140625,
"learning_rate": 0.0004329454622692636,
"loss": 2.5087,
"step": 2780
},
{
"epoch": 0.2462815024054376,
"grad_norm": 0.1572265625,
"learning_rate": 0.00043246801054768147,
"loss": 2.5188,
"step": 2790
},
{
"epoch": 0.24716423180474026,
"grad_norm": 0.1416015625,
"learning_rate": 0.0004319891303223287,
"loss": 2.5059,
"step": 2800
},
{
"epoch": 0.2480469612040429,
"grad_norm": 0.166015625,
"learning_rate": 0.000431508825342258,
"loss": 2.5227,
"step": 2810
},
{
"epoch": 0.24892969060334555,
"grad_norm": 0.1416015625,
"learning_rate": 0.0004310270993676764,
"loss": 2.5215,
"step": 2820
},
{
"epoch": 0.2498124200026482,
"grad_norm": 0.1484375,
"learning_rate": 0.00043054395616991535,
"loss": 2.513,
"step": 2830
},
{
"epoch": 0.2506951494019508,
"grad_norm": 0.150390625,
"learning_rate": 0.0004300593995314017,
"loss": 2.5106,
"step": 2840
},
{
"epoch": 0.2515778788012535,
"grad_norm": 0.146484375,
"learning_rate": 0.0004295734332456277,
"loss": 2.4932,
"step": 2850
},
{
"epoch": 0.2524606082005561,
"grad_norm": 0.1591796875,
"learning_rate": 0.00042908606111712136,
"loss": 2.5118,
"step": 2860
},
{
"epoch": 0.2533433375998588,
"grad_norm": 0.197265625,
"learning_rate": 0.0004285972869614169,
"loss": 2.5157,
"step": 2870
},
{
"epoch": 0.2542260669991614,
"grad_norm": 0.1572265625,
"learning_rate": 0.00042810711460502447,
"loss": 2.5073,
"step": 2880
},
{
"epoch": 0.25510879639846407,
"grad_norm": 0.158203125,
"learning_rate": 0.00042761554788540084,
"loss": 2.5368,
"step": 2890
},
{
"epoch": 0.2559915257977667,
"grad_norm": 0.1708984375,
"learning_rate": 0.0004271225906509186,
"loss": 2.5143,
"step": 2900
},
{
"epoch": 0.25687425519706936,
"grad_norm": 0.1494140625,
"learning_rate": 0.0004266282467608365,
"loss": 2.5078,
"step": 2910
},
{
"epoch": 0.257756984596372,
"grad_norm": 0.1630859375,
"learning_rate": 0.00042613252008526914,
"loss": 2.5283,
"step": 2920
},
{
"epoch": 0.25863971399567465,
"grad_norm": 0.15234375,
"learning_rate": 0.0004256354145051567,
"loss": 2.5008,
"step": 2930
},
{
"epoch": 0.25952244339497726,
"grad_norm": 0.1572265625,
"learning_rate": 0.0004251369339122344,
"loss": 2.5261,
"step": 2940
},
{
"epoch": 0.26040517279427994,
"grad_norm": 0.1669921875,
"learning_rate": 0.00042463708220900225,
"loss": 2.5284,
"step": 2950
},
{
"epoch": 0.26128790219358256,
"grad_norm": 0.1689453125,
"learning_rate": 0.00042413586330869446,
"loss": 2.5355,
"step": 2960
},
{
"epoch": 0.26217063159288523,
"grad_norm": 0.1484375,
"learning_rate": 0.00042363328113524846,
"loss": 2.5201,
"step": 2970
},
{
"epoch": 0.26305336099218785,
"grad_norm": 0.150390625,
"learning_rate": 0.0004231293396232747,
"loss": 2.5249,
"step": 2980
},
{
"epoch": 0.26393609039149046,
"grad_norm": 0.1591796875,
"learning_rate": 0.00042262404271802565,
"loss": 2.5161,
"step": 2990
},
{
"epoch": 0.26481881979079314,
"grad_norm": 0.1953125,
"learning_rate": 0.00042211739437536457,
"loss": 2.522,
"step": 3000
},
{
"epoch": 0.26570154919009575,
"grad_norm": 0.1435546875,
"learning_rate": 0.0004216093985617352,
"loss": 2.5132,
"step": 3010
},
{
"epoch": 0.2665842785893984,
"grad_norm": 0.1474609375,
"learning_rate": 0.0004211000592541301,
"loss": 2.5156,
"step": 3020
},
{
"epoch": 0.26746700798870104,
"grad_norm": 0.1474609375,
"learning_rate": 0.0004205893804400599,
"loss": 2.5121,
"step": 3030
},
{
"epoch": 0.2683497373880037,
"grad_norm": 0.150390625,
"learning_rate": 0.0004200773661175219,
"loss": 2.5054,
"step": 3040
},
{
"epoch": 0.26923246678730633,
"grad_norm": 0.158203125,
"learning_rate": 0.0004195640202949687,
"loss": 2.5012,
"step": 3050
},
{
"epoch": 0.270115196186609,
"grad_norm": 0.1513671875,
"learning_rate": 0.00041904934699127713,
"loss": 2.5139,
"step": 3060
},
{
"epoch": 0.2709979255859116,
"grad_norm": 0.1650390625,
"learning_rate": 0.0004185333502357164,
"loss": 2.501,
"step": 3070
},
{
"epoch": 0.2718806549852143,
"grad_norm": 0.1494140625,
"learning_rate": 0.000418016034067917,
"loss": 2.5065,
"step": 3080
},
{
"epoch": 0.2727633843845169,
"grad_norm": 0.150390625,
"learning_rate": 0.00041749740253783853,
"loss": 2.5107,
"step": 3090
},
{
"epoch": 0.2736461137838196,
"grad_norm": 0.15234375,
"learning_rate": 0.00041697745970573855,
"loss": 2.5222,
"step": 3100
},
{
"epoch": 0.2745288431831222,
"grad_norm": 0.14453125,
"learning_rate": 0.00041645620964214023,
"loss": 2.5137,
"step": 3110
},
{
"epoch": 0.2754115725824249,
"grad_norm": 0.1806640625,
"learning_rate": 0.0004159336564278012,
"loss": 2.5347,
"step": 3120
},
{
"epoch": 0.2762943019817275,
"grad_norm": 0.1748046875,
"learning_rate": 0.0004154098041536807,
"loss": 2.5243,
"step": 3130
},
{
"epoch": 0.27717703138103017,
"grad_norm": 0.1513671875,
"learning_rate": 0.00041488465692090837,
"loss": 2.5265,
"step": 3140
},
{
"epoch": 0.2780597607803328,
"grad_norm": 0.1513671875,
"learning_rate": 0.00041435821884075176,
"loss": 2.5145,
"step": 3150
},
{
"epoch": 0.2789424901796354,
"grad_norm": 0.1640625,
"learning_rate": 0.00041383049403458403,
"loss": 2.5187,
"step": 3160
},
{
"epoch": 0.2798252195789381,
"grad_norm": 0.1533203125,
"learning_rate": 0.0004133014866338521,
"loss": 2.5217,
"step": 3170
},
{
"epoch": 0.2807079489782407,
"grad_norm": 0.1728515625,
"learning_rate": 0.00041277120078004383,
"loss": 2.4996,
"step": 3180
},
{
"epoch": 0.28159067837754337,
"grad_norm": 0.150390625,
"learning_rate": 0.0004122396406246559,
"loss": 2.5199,
"step": 3190
},
{
"epoch": 0.282473407776846,
"grad_norm": 0.16015625,
"learning_rate": 0.0004117068103291614,
"loss": 2.516,
"step": 3200
},
{
"epoch": 0.28335613717614866,
"grad_norm": 0.142578125,
"learning_rate": 0.00041117271406497665,
"loss": 2.5019,
"step": 3210
},
{
"epoch": 0.2842388665754513,
"grad_norm": 0.146484375,
"learning_rate": 0.00041063735601342934,
"loss": 2.5114,
"step": 3220
},
{
"epoch": 0.28512159597475395,
"grad_norm": 0.1416015625,
"learning_rate": 0.0004101007403657255,
"loss": 2.5148,
"step": 3230
},
{
"epoch": 0.28600432537405657,
"grad_norm": 0.1552734375,
"learning_rate": 0.00040956287132291625,
"loss": 2.5008,
"step": 3240
},
{
"epoch": 0.28688705477335924,
"grad_norm": 0.1796875,
"learning_rate": 0.00040902375309586557,
"loss": 2.5153,
"step": 3250
},
{
"epoch": 0.28776978417266186,
"grad_norm": 0.134765625,
"learning_rate": 0.00040848338990521696,
"loss": 2.5147,
"step": 3260
},
{
"epoch": 0.28865251357196453,
"grad_norm": 0.1337890625,
"learning_rate": 0.00040794178598136033,
"loss": 2.5062,
"step": 3270
},
{
"epoch": 0.28953524297126715,
"grad_norm": 0.17578125,
"learning_rate": 0.0004073989455643994,
"loss": 2.526,
"step": 3280
},
{
"epoch": 0.2904179723705698,
"grad_norm": 0.150390625,
"learning_rate": 0.00040685487290411765,
"loss": 2.5167,
"step": 3290
},
{
"epoch": 0.29130070176987244,
"grad_norm": 0.1884765625,
"learning_rate": 0.0004063095722599459,
"loss": 2.5083,
"step": 3300
},
{
"epoch": 0.2921834311691751,
"grad_norm": 0.1640625,
"learning_rate": 0.00040576304790092857,
"loss": 2.5069,
"step": 3310
},
{
"epoch": 0.29306616056847773,
"grad_norm": 0.1611328125,
"learning_rate": 0.00040521530410569007,
"loss": 2.5289,
"step": 3320
},
{
"epoch": 0.2939488899677804,
"grad_norm": 0.158203125,
"learning_rate": 0.0004046663451624016,
"loss": 2.5135,
"step": 3330
},
{
"epoch": 0.294831619367083,
"grad_norm": 0.171875,
"learning_rate": 0.0004041161753687478,
"loss": 2.4992,
"step": 3340
},
{
"epoch": 0.29571434876638564,
"grad_norm": 0.1484375,
"learning_rate": 0.00040356479903189233,
"loss": 2.5231,
"step": 3350
},
{
"epoch": 0.2965970781656883,
"grad_norm": 0.150390625,
"learning_rate": 0.0004030122204684449,
"loss": 2.509,
"step": 3360
},
{
"epoch": 0.29747980756499093,
"grad_norm": 0.1328125,
"learning_rate": 0.0004024584440044271,
"loss": 2.5044,
"step": 3370
},
{
"epoch": 0.2983625369642936,
"grad_norm": 0.140625,
"learning_rate": 0.00040190347397523873,
"loss": 2.5097,
"step": 3380
},
{
"epoch": 0.2992452663635962,
"grad_norm": 0.1396484375,
"learning_rate": 0.0004013473147256238,
"loss": 2.5068,
"step": 3390
},
{
"epoch": 0.3001279957628989,
"grad_norm": 0.1396484375,
"learning_rate": 0.0004007899706096363,
"loss": 2.5048,
"step": 3400
},
{
"epoch": 0.3010107251622015,
"grad_norm": 0.185546875,
"learning_rate": 0.00040023144599060623,
"loss": 2.4943,
"step": 3410
},
{
"epoch": 0.3018934545615042,
"grad_norm": 0.1806640625,
"learning_rate": 0.00039967174524110596,
"loss": 2.5269,
"step": 3420
},
{
"epoch": 0.3027761839608068,
"grad_norm": 0.16015625,
"learning_rate": 0.000399110872742915,
"loss": 2.5052,
"step": 3430
},
{
"epoch": 0.3036589133601095,
"grad_norm": 0.1650390625,
"learning_rate": 0.0003985488328869865,
"loss": 2.5229,
"step": 3440
},
{
"epoch": 0.3045416427594121,
"grad_norm": 0.1396484375,
"learning_rate": 0.0003979856300734126,
"loss": 2.5035,
"step": 3450
},
{
"epoch": 0.30542437215871476,
"grad_norm": 0.158203125,
"learning_rate": 0.00039742126871138996,
"loss": 2.5109,
"step": 3460
},
{
"epoch": 0.3063071015580174,
"grad_norm": 0.16796875,
"learning_rate": 0.0003968557532191852,
"loss": 2.5181,
"step": 3470
},
{
"epoch": 0.30718983095732005,
"grad_norm": 0.173828125,
"learning_rate": 0.00039628908802410057,
"loss": 2.5159,
"step": 3480
},
{
"epoch": 0.30807256035662267,
"grad_norm": 0.15625,
"learning_rate": 0.00039572127756243904,
"loss": 2.508,
"step": 3490
},
{
"epoch": 0.30895528975592534,
"grad_norm": 0.1455078125,
"learning_rate": 0.0003951523262794693,
"loss": 2.5088,
"step": 3500
},
{
"epoch": 0.30983801915522796,
"grad_norm": 0.1474609375,
"learning_rate": 0.00039458223862939184,
"loss": 2.5191,
"step": 3510
},
{
"epoch": 0.31072074855453063,
"grad_norm": 0.1396484375,
"learning_rate": 0.00039401101907530323,
"loss": 2.5113,
"step": 3520
},
{
"epoch": 0.31160347795383325,
"grad_norm": 0.1494140625,
"learning_rate": 0.0003934386720891614,
"loss": 2.5097,
"step": 3530
},
{
"epoch": 0.31248620735313587,
"grad_norm": 0.1533203125,
"learning_rate": 0.00039286520215175085,
"loss": 2.4938,
"step": 3540
},
{
"epoch": 0.31336893675243854,
"grad_norm": 0.14453125,
"learning_rate": 0.0003922906137526474,
"loss": 2.5175,
"step": 3550
},
{
"epoch": 0.31425166615174116,
"grad_norm": 0.189453125,
"learning_rate": 0.00039171491139018325,
"loss": 2.5129,
"step": 3560
},
{
"epoch": 0.31513439555104383,
"grad_norm": 0.1767578125,
"learning_rate": 0.0003911380995714111,
"loss": 2.5287,
"step": 3570
},
{
"epoch": 0.31601712495034645,
"grad_norm": 0.18359375,
"learning_rate": 0.0003905601828120698,
"loss": 2.5026,
"step": 3580
},
{
"epoch": 0.3168998543496491,
"grad_norm": 0.140625,
"learning_rate": 0.0003899811656365485,
"loss": 2.5138,
"step": 3590
},
{
"epoch": 0.31778258374895174,
"grad_norm": 0.15234375,
"learning_rate": 0.0003894010525778511,
"loss": 2.5213,
"step": 3600
},
{
"epoch": 0.3186653131482544,
"grad_norm": 0.1640625,
"learning_rate": 0.000388819848177561,
"loss": 2.515,
"step": 3610
},
{
"epoch": 0.31954804254755703,
"grad_norm": 0.1767578125,
"learning_rate": 0.00038823755698580545,
"loss": 2.5041,
"step": 3620
},
{
"epoch": 0.3204307719468597,
"grad_norm": 0.1640625,
"learning_rate": 0.0003876541835612202,
"loss": 2.5209,
"step": 3630
},
{
"epoch": 0.3213135013461623,
"grad_norm": 0.1708984375,
"learning_rate": 0.0003870697324709132,
"loss": 2.5182,
"step": 3640
},
{
"epoch": 0.322196230745465,
"grad_norm": 0.1689453125,
"learning_rate": 0.00038648420829042954,
"loss": 2.5174,
"step": 3650
},
{
"epoch": 0.3230789601447676,
"grad_norm": 0.1376953125,
"learning_rate": 0.00038589761560371515,
"loss": 2.5152,
"step": 3660
},
{
"epoch": 0.3239616895440703,
"grad_norm": 0.1533203125,
"learning_rate": 0.00038530995900308107,
"loss": 2.4956,
"step": 3670
},
{
"epoch": 0.3248444189433729,
"grad_norm": 0.1591796875,
"learning_rate": 0.00038472124308916753,
"loss": 2.5054,
"step": 3680
},
{
"epoch": 0.3257271483426756,
"grad_norm": 0.1357421875,
"learning_rate": 0.00038413147247090795,
"loss": 2.5215,
"step": 3690
},
{
"epoch": 0.3266098777419782,
"grad_norm": 0.140625,
"learning_rate": 0.00038354065176549274,
"loss": 2.5235,
"step": 3700
},
{
"epoch": 0.32749260714128087,
"grad_norm": 0.1435546875,
"learning_rate": 0.00038294878559833317,
"loss": 2.5044,
"step": 3710
},
{
"epoch": 0.3283753365405835,
"grad_norm": 0.15625,
"learning_rate": 0.0003823558786030255,
"loss": 2.5099,
"step": 3720
},
{
"epoch": 0.3292580659398861,
"grad_norm": 0.1728515625,
"learning_rate": 0.00038176193542131386,
"loss": 2.5144,
"step": 3730
},
{
"epoch": 0.3301407953391888,
"grad_norm": 0.1611328125,
"learning_rate": 0.00038116696070305503,
"loss": 2.5195,
"step": 3740
},
{
"epoch": 0.3310235247384914,
"grad_norm": 0.14453125,
"learning_rate": 0.00038057095910618125,
"loss": 2.5066,
"step": 3750
},
{
"epoch": 0.33190625413779407,
"grad_norm": 0.1611328125,
"learning_rate": 0.00037997393529666393,
"loss": 2.5161,
"step": 3760
},
{
"epoch": 0.3327889835370967,
"grad_norm": 0.1318359375,
"learning_rate": 0.00037937589394847714,
"loss": 2.4976,
"step": 3770
},
{
"epoch": 0.33367171293639936,
"grad_norm": 0.142578125,
"learning_rate": 0.00037877683974356114,
"loss": 2.5082,
"step": 3780
},
{
"epoch": 0.334554442335702,
"grad_norm": 0.138671875,
"learning_rate": 0.0003781767773717857,
"loss": 2.5061,
"step": 3790
},
{
"epoch": 0.33543717173500465,
"grad_norm": 0.15234375,
"learning_rate": 0.00037757571153091324,
"loss": 2.5106,
"step": 3800
},
{
"epoch": 0.33631990113430726,
"grad_norm": 0.154296875,
"learning_rate": 0.000376973646926562,
"loss": 2.509,
"step": 3810
},
{
"epoch": 0.33720263053360994,
"grad_norm": 0.1494140625,
"learning_rate": 0.00037637058827216964,
"loss": 2.4963,
"step": 3820
},
{
"epoch": 0.33808535993291255,
"grad_norm": 0.142578125,
"learning_rate": 0.00037576654028895554,
"loss": 2.5123,
"step": 3830
},
{
"epoch": 0.3389680893322152,
"grad_norm": 0.134765625,
"learning_rate": 0.00037516150770588487,
"loss": 2.4981,
"step": 3840
},
{
"epoch": 0.33985081873151785,
"grad_norm": 0.146484375,
"learning_rate": 0.00037455549525963066,
"loss": 2.5043,
"step": 3850
},
{
"epoch": 0.3407335481308205,
"grad_norm": 0.1513671875,
"learning_rate": 0.0003739485076945373,
"loss": 2.5044,
"step": 3860
},
{
"epoch": 0.34161627753012314,
"grad_norm": 0.146484375,
"learning_rate": 0.000373340549762583,
"loss": 2.483,
"step": 3870
},
{
"epoch": 0.3424990069294258,
"grad_norm": 0.1484375,
"learning_rate": 0.0003727316262233429,
"loss": 2.5101,
"step": 3880
},
{
"epoch": 0.3433817363287284,
"grad_norm": 0.142578125,
"learning_rate": 0.0003721217418439516,
"loss": 2.4966,
"step": 3890
},
{
"epoch": 0.3442644657280311,
"grad_norm": 0.162109375,
"learning_rate": 0.00037151090139906593,
"loss": 2.505,
"step": 3900
},
{
"epoch": 0.3451471951273337,
"grad_norm": 0.13671875,
"learning_rate": 0.00037089910967082765,
"loss": 2.5099,
"step": 3910
},
{
"epoch": 0.34602992452663633,
"grad_norm": 0.150390625,
"learning_rate": 0.0003702863714488257,
"loss": 2.5158,
"step": 3920
},
{
"epoch": 0.346912653925939,
"grad_norm": 0.16015625,
"learning_rate": 0.0003696726915300592,
"loss": 2.5127,
"step": 3930
},
{
"epoch": 0.3477953833252416,
"grad_norm": 0.15234375,
"learning_rate": 0.0003690580747188995,
"loss": 2.5141,
"step": 3940
},
{
"epoch": 0.3486781127245443,
"grad_norm": 0.1875,
"learning_rate": 0.00036844252582705244,
"loss": 2.4919,
"step": 3950
},
{
"epoch": 0.3495608421238469,
"grad_norm": 0.1484375,
"learning_rate": 0.0003678260496735214,
"loss": 2.5091,
"step": 3960
},
{
"epoch": 0.3504435715231496,
"grad_norm": 0.15234375,
"learning_rate": 0.0003672086510845687,
"loss": 2.504,
"step": 3970
},
{
"epoch": 0.3513263009224522,
"grad_norm": 0.1455078125,
"learning_rate": 0.00036659033489367835,
"loss": 2.5042,
"step": 3980
},
{
"epoch": 0.3522090303217549,
"grad_norm": 0.1494140625,
"learning_rate": 0.0003659711059415182,
"loss": 2.5083,
"step": 3990
},
{
"epoch": 0.3530917597210575,
"grad_norm": 0.1767578125,
"learning_rate": 0.0003653509690759016,
"loss": 2.5196,
"step": 4000
},
{
"epoch": 0.3530917597210575,
"eval_accuracy": 0.5083680819529877,
"eval_loss": 2.3899683952331543,
"eval_runtime": 6.95,
"eval_samples_per_second": 45.756,
"eval_steps_per_second": 0.432,
"step": 4000
},
{
"epoch": 0.35397448912036017,
"grad_norm": 0.1513671875,
"learning_rate": 0.00036472992915175017,
"loss": 2.4973,
"step": 4010
},
{
"epoch": 0.3548572185196628,
"grad_norm": 0.134765625,
"learning_rate": 0.00036410799103105503,
"loss": 2.5225,
"step": 4020
},
{
"epoch": 0.35573994791896546,
"grad_norm": 0.1875,
"learning_rate": 0.0003634851595828393,
"loss": 2.5052,
"step": 4030
},
{
"epoch": 0.3566226773182681,
"grad_norm": 0.1640625,
"learning_rate": 0.00036286143968311963,
"loss": 2.5045,
"step": 4040
},
{
"epoch": 0.35750540671757075,
"grad_norm": 0.134765625,
"learning_rate": 0.00036223683621486845,
"loss": 2.5086,
"step": 4050
},
{
"epoch": 0.35838813611687337,
"grad_norm": 0.15234375,
"learning_rate": 0.00036161135406797504,
"loss": 2.4969,
"step": 4060
},
{
"epoch": 0.35927086551617604,
"grad_norm": 0.1982421875,
"learning_rate": 0.0003609849981392079,
"loss": 2.4986,
"step": 4070
},
{
"epoch": 0.36015359491547866,
"grad_norm": 0.1396484375,
"learning_rate": 0.0003603577733321764,
"loss": 2.4938,
"step": 4080
},
{
"epoch": 0.3610363243147813,
"grad_norm": 0.1376953125,
"learning_rate": 0.0003597296845572917,
"loss": 2.4959,
"step": 4090
},
{
"epoch": 0.36191905371408395,
"grad_norm": 0.1435546875,
"learning_rate": 0.00035910073673172933,
"loss": 2.5011,
"step": 4100
},
{
"epoch": 0.36280178311338657,
"grad_norm": 0.146484375,
"learning_rate": 0.00035847093477938953,
"loss": 2.4938,
"step": 4110
},
{
"epoch": 0.36368451251268924,
"grad_norm": 0.146484375,
"learning_rate": 0.00035784028363085985,
"loss": 2.4949,
"step": 4120
},
{
"epoch": 0.36456724191199186,
"grad_norm": 0.154296875,
"learning_rate": 0.00035720878822337576,
"loss": 2.4885,
"step": 4130
},
{
"epoch": 0.36544997131129453,
"grad_norm": 0.1494140625,
"learning_rate": 0.00035657645350078233,
"loss": 2.5217,
"step": 4140
},
{
"epoch": 0.36633270071059715,
"grad_norm": 0.162109375,
"learning_rate": 0.0003559432844134954,
"loss": 2.5107,
"step": 4150
},
{
"epoch": 0.3672154301098998,
"grad_norm": 0.1669921875,
"learning_rate": 0.0003553092859184629,
"loss": 2.5016,
"step": 4160
},
{
"epoch": 0.36809815950920244,
"grad_norm": 0.1591796875,
"learning_rate": 0.0003546744629791261,
"loss": 2.4954,
"step": 4170
},
{
"epoch": 0.3689808889085051,
"grad_norm": 0.177734375,
"learning_rate": 0.00035403882056538044,
"loss": 2.4903,
"step": 4180
},
{
"epoch": 0.36986361830780773,
"grad_norm": 0.17578125,
"learning_rate": 0.00035340236365353724,
"loss": 2.5008,
"step": 4190
},
{
"epoch": 0.3707463477071104,
"grad_norm": 0.1630859375,
"learning_rate": 0.000352765097226284,
"loss": 2.5107,
"step": 4200
},
{
"epoch": 0.371629077106413,
"grad_norm": 0.150390625,
"learning_rate": 0.0003521270262726458,
"loss": 2.5058,
"step": 4210
},
{
"epoch": 0.3725118065057157,
"grad_norm": 0.1708984375,
"learning_rate": 0.00035148815578794635,
"loss": 2.4966,
"step": 4220
},
{
"epoch": 0.3733945359050183,
"grad_norm": 0.146484375,
"learning_rate": 0.0003508484907737687,
"loss": 2.4942,
"step": 4230
},
{
"epoch": 0.374277265304321,
"grad_norm": 0.1796875,
"learning_rate": 0.0003502080362379159,
"loss": 2.5097,
"step": 4240
},
{
"epoch": 0.3751599947036236,
"grad_norm": 0.15234375,
"learning_rate": 0.00034956679719437225,
"loss": 2.5197,
"step": 4250
},
{
"epoch": 0.3760427241029263,
"grad_norm": 0.142578125,
"learning_rate": 0.00034892477866326356,
"loss": 2.4971,
"step": 4260
},
{
"epoch": 0.3769254535022289,
"grad_norm": 0.1806640625,
"learning_rate": 0.0003482819856708183,
"loss": 2.4922,
"step": 4270
},
{
"epoch": 0.3778081829015315,
"grad_norm": 0.1494140625,
"learning_rate": 0.00034763842324932794,
"loss": 2.5172,
"step": 4280
},
{
"epoch": 0.3786909123008342,
"grad_norm": 0.1630859375,
"learning_rate": 0.00034699409643710764,
"loss": 2.51,
"step": 4290
},
{
"epoch": 0.3795736417001368,
"grad_norm": 0.13671875,
"learning_rate": 0.00034634901027845677,
"loss": 2.5022,
"step": 4300
},
{
"epoch": 0.38045637109943947,
"grad_norm": 0.140625,
"learning_rate": 0.0003457031698236196,
"loss": 2.4982,
"step": 4310
},
{
"epoch": 0.3813391004987421,
"grad_norm": 0.146484375,
"learning_rate": 0.00034505658012874544,
"loss": 2.5115,
"step": 4320
},
{
"epoch": 0.38222182989804476,
"grad_norm": 0.1337890625,
"learning_rate": 0.00034440924625584954,
"loss": 2.4998,
"step": 4330
},
{
"epoch": 0.3831045592973474,
"grad_norm": 0.138671875,
"learning_rate": 0.0003437611732727728,
"loss": 2.4927,
"step": 4340
},
{
"epoch": 0.38398728869665005,
"grad_norm": 0.1748046875,
"learning_rate": 0.0003431123662531427,
"loss": 2.505,
"step": 4350
},
{
"epoch": 0.38487001809595267,
"grad_norm": 0.1396484375,
"learning_rate": 0.0003424628302763332,
"loss": 2.4918,
"step": 4360
},
{
"epoch": 0.38575274749525534,
"grad_norm": 0.1484375,
"learning_rate": 0.0003418125704274252,
"loss": 2.4941,
"step": 4370
},
{
"epoch": 0.38663547689455796,
"grad_norm": 0.1748046875,
"learning_rate": 0.00034116159179716675,
"loss": 2.5237,
"step": 4380
},
{
"epoch": 0.38751820629386063,
"grad_norm": 0.1611328125,
"learning_rate": 0.0003405098994819329,
"loss": 2.5107,
"step": 4390
},
{
"epoch": 0.38840093569316325,
"grad_norm": 0.1298828125,
"learning_rate": 0.00033985749858368605,
"loss": 2.4944,
"step": 4400
},
{
"epoch": 0.3892836650924659,
"grad_norm": 0.138671875,
"learning_rate": 0.0003392043942099358,
"loss": 2.5099,
"step": 4410
},
{
"epoch": 0.39016639449176854,
"grad_norm": 0.1357421875,
"learning_rate": 0.0003385505914736994,
"loss": 2.5048,
"step": 4420
},
{
"epoch": 0.3910491238910712,
"grad_norm": 0.150390625,
"learning_rate": 0.00033789609549346146,
"loss": 2.4965,
"step": 4430
},
{
"epoch": 0.39193185329037383,
"grad_norm": 0.15625,
"learning_rate": 0.0003372409113931334,
"loss": 2.4925,
"step": 4440
},
{
"epoch": 0.3928145826896765,
"grad_norm": 0.146484375,
"learning_rate": 0.0003365850443020142,
"loss": 2.4915,
"step": 4450
},
{
"epoch": 0.3936973120889791,
"grad_norm": 0.1376953125,
"learning_rate": 0.00033592849935474965,
"loss": 2.5076,
"step": 4460
},
{
"epoch": 0.39458004148828174,
"grad_norm": 0.1513671875,
"learning_rate": 0.0003352712816912925,
"loss": 2.5088,
"step": 4470
},
{
"epoch": 0.3954627708875844,
"grad_norm": 0.1552734375,
"learning_rate": 0.00033461339645686196,
"loss": 2.5017,
"step": 4480
},
{
"epoch": 0.39634550028688703,
"grad_norm": 0.140625,
"learning_rate": 0.0003339548488019033,
"loss": 2.496,
"step": 4490
},
{
"epoch": 0.3972282296861897,
"grad_norm": 0.1396484375,
"learning_rate": 0.00033329564388204816,
"loss": 2.4897,
"step": 4500
},
{
"epoch": 0.3981109590854923,
"grad_norm": 0.158203125,
"learning_rate": 0.0003326357868580734,
"loss": 2.4998,
"step": 4510
},
{
"epoch": 0.398993688484795,
"grad_norm": 0.1669921875,
"learning_rate": 0.0003319752828958613,
"loss": 2.5061,
"step": 4520
},
{
"epoch": 0.3998764178840976,
"grad_norm": 0.1455078125,
"learning_rate": 0.0003313141371663587,
"loss": 2.514,
"step": 4530
},
{
"epoch": 0.4007591472834003,
"grad_norm": 0.1435546875,
"learning_rate": 0.000330652354845537,
"loss": 2.5237,
"step": 4540
},
{
"epoch": 0.4016418766827029,
"grad_norm": 0.224609375,
"learning_rate": 0.0003299899411143509,
"loss": 2.5028,
"step": 4550
},
{
"epoch": 0.4025246060820056,
"grad_norm": 0.1630859375,
"learning_rate": 0.0003293269011586986,
"loss": 2.4957,
"step": 4560
},
{
"epoch": 0.4034073354813082,
"grad_norm": 0.1484375,
"learning_rate": 0.00032866324016938095,
"loss": 2.4829,
"step": 4570
},
{
"epoch": 0.40429006488061087,
"grad_norm": 0.154296875,
"learning_rate": 0.00032799896334206045,
"loss": 2.5011,
"step": 4580
},
{
"epoch": 0.4051727942799135,
"grad_norm": 0.138671875,
"learning_rate": 0.000327334075877221,
"loss": 2.5175,
"step": 4590
},
{
"epoch": 0.40605552367921616,
"grad_norm": 0.142578125,
"learning_rate": 0.000326668582980127,
"loss": 2.5149,
"step": 4600
},
{
"epoch": 0.4069382530785188,
"grad_norm": 0.1357421875,
"learning_rate": 0.00032600248986078295,
"loss": 2.5004,
"step": 4610
},
{
"epoch": 0.40782098247782145,
"grad_norm": 0.185546875,
"learning_rate": 0.00032533580173389195,
"loss": 2.488,
"step": 4620
},
{
"epoch": 0.40870371187712407,
"grad_norm": 0.1796875,
"learning_rate": 0.0003246685238188154,
"loss": 2.4982,
"step": 4630
},
{
"epoch": 0.40958644127642674,
"grad_norm": 0.138671875,
"learning_rate": 0.00032400066133953225,
"loss": 2.5085,
"step": 4640
},
{
"epoch": 0.41046917067572936,
"grad_norm": 0.1318359375,
"learning_rate": 0.0003233322195245977,
"loss": 2.506,
"step": 4650
},
{
"epoch": 0.411351900075032,
"grad_norm": 0.1337890625,
"learning_rate": 0.00032266320360710237,
"loss": 2.5029,
"step": 4660
},
{
"epoch": 0.41223462947433465,
"grad_norm": 0.138671875,
"learning_rate": 0.0003219936188246317,
"loss": 2.5053,
"step": 4670
},
{
"epoch": 0.41311735887363726,
"grad_norm": 0.142578125,
"learning_rate": 0.0003213234704192243,
"loss": 2.5006,
"step": 4680
},
{
"epoch": 0.41400008827293994,
"grad_norm": 0.1591796875,
"learning_rate": 0.00032065276363733137,
"loss": 2.4967,
"step": 4690
},
{
"epoch": 0.41488281767224255,
"grad_norm": 0.1494140625,
"learning_rate": 0.00031998150372977577,
"loss": 2.5042,
"step": 4700
},
{
"epoch": 0.4157655470715452,
"grad_norm": 0.14453125,
"learning_rate": 0.0003193096959517103,
"loss": 2.4926,
"step": 4710
},
{
"epoch": 0.41664827647084784,
"grad_norm": 0.138671875,
"learning_rate": 0.0003186373455625774,
"loss": 2.4898,
"step": 4720
},
{
"epoch": 0.4175310058701505,
"grad_norm": 0.13671875,
"learning_rate": 0.0003179644578260669,
"loss": 2.4932,
"step": 4730
},
{
"epoch": 0.41841373526945314,
"grad_norm": 0.142578125,
"learning_rate": 0.00031729103801007575,
"loss": 2.493,
"step": 4740
},
{
"epoch": 0.4192964646687558,
"grad_norm": 0.1943359375,
"learning_rate": 0.0003166170913866665,
"loss": 2.5047,
"step": 4750
},
{
"epoch": 0.4201791940680584,
"grad_norm": 0.142578125,
"learning_rate": 0.00031594262323202577,
"loss": 2.4976,
"step": 4760
},
{
"epoch": 0.4210619234673611,
"grad_norm": 0.138671875,
"learning_rate": 0.0003152676388264234,
"loss": 2.4955,
"step": 4770
},
{
"epoch": 0.4219446528666637,
"grad_norm": 0.1396484375,
"learning_rate": 0.00031459214345417046,
"loss": 2.4742,
"step": 4780
},
{
"epoch": 0.4228273822659664,
"grad_norm": 0.21484375,
"learning_rate": 0.00031391614240357864,
"loss": 2.4919,
"step": 4790
},
{
"epoch": 0.423710111665269,
"grad_norm": 0.1552734375,
"learning_rate": 0.00031323964096691825,
"loss": 2.5034,
"step": 4800
},
{
"epoch": 0.4245928410645717,
"grad_norm": 0.158203125,
"learning_rate": 0.0003125626444403772,
"loss": 2.4846,
"step": 4810
},
{
"epoch": 0.4254755704638743,
"grad_norm": 0.1318359375,
"learning_rate": 0.00031188515812401917,
"loss": 2.5016,
"step": 4820
},
{
"epoch": 0.42635829986317697,
"grad_norm": 0.142578125,
"learning_rate": 0.00031120718732174235,
"loss": 2.4959,
"step": 4830
},
{
"epoch": 0.4272410292624796,
"grad_norm": 0.142578125,
"learning_rate": 0.000310528737341238,
"loss": 2.4716,
"step": 4840
},
{
"epoch": 0.4281237586617822,
"grad_norm": 0.150390625,
"learning_rate": 0.00030984981349394864,
"loss": 2.4994,
"step": 4850
},
{
"epoch": 0.4290064880610849,
"grad_norm": 0.1552734375,
"learning_rate": 0.00030917042109502663,
"loss": 2.499,
"step": 4860
},
{
"epoch": 0.4298892174603875,
"grad_norm": 0.1396484375,
"learning_rate": 0.00030849056546329253,
"loss": 2.4879,
"step": 4870
},
{
"epoch": 0.43077194685969017,
"grad_norm": 0.13671875,
"learning_rate": 0.0003078102519211933,
"loss": 2.4751,
"step": 4880
},
{
"epoch": 0.4316546762589928,
"grad_norm": 0.154296875,
"learning_rate": 0.0003071294857947612,
"loss": 2.5005,
"step": 4890
},
{
"epoch": 0.43253740565829546,
"grad_norm": 0.1533203125,
"learning_rate": 0.0003064482724135711,
"loss": 2.5128,
"step": 4900
},
{
"epoch": 0.4334201350575981,
"grad_norm": 0.1337890625,
"learning_rate": 0.00030576661711069985,
"loss": 2.4908,
"step": 4910
},
{
"epoch": 0.43430286445690075,
"grad_norm": 0.1435546875,
"learning_rate": 0.0003050845252226837,
"loss": 2.5086,
"step": 4920
},
{
"epoch": 0.43518559385620337,
"grad_norm": 0.1376953125,
"learning_rate": 0.0003044020020894769,
"loss": 2.498,
"step": 4930
},
{
"epoch": 0.43606832325550604,
"grad_norm": 0.1669921875,
"learning_rate": 0.00030371905305441,
"loss": 2.4986,
"step": 4940
},
{
"epoch": 0.43695105265480866,
"grad_norm": 0.146484375,
"learning_rate": 0.0003030356834641476,
"loss": 2.4886,
"step": 4950
},
{
"epoch": 0.43783378205411133,
"grad_norm": 0.1435546875,
"learning_rate": 0.0003023518986686469,
"loss": 2.4948,
"step": 4960
},
{
"epoch": 0.43871651145341395,
"grad_norm": 0.14453125,
"learning_rate": 0.0003016677040211154,
"loss": 2.5024,
"step": 4970
},
{
"epoch": 0.4395992408527166,
"grad_norm": 0.1435546875,
"learning_rate": 0.00030098310487796965,
"loss": 2.4908,
"step": 4980
},
{
"epoch": 0.44048197025201924,
"grad_norm": 0.1484375,
"learning_rate": 0.00030029810659879273,
"loss": 2.4913,
"step": 4990
},
{
"epoch": 0.4413646996513219,
"grad_norm": 0.1953125,
"learning_rate": 0.00029961271454629235,
"loss": 2.502,
"step": 5000
},
{
"epoch": 0.44224742905062453,
"grad_norm": 0.1318359375,
"learning_rate": 0.0002989269340862591,
"loss": 2.4902,
"step": 5010
},
{
"epoch": 0.44313015844992715,
"grad_norm": 0.1484375,
"learning_rate": 0.0002982407705875243,
"loss": 2.5009,
"step": 5020
},
{
"epoch": 0.4440128878492298,
"grad_norm": 0.1318359375,
"learning_rate": 0.00029755422942191805,
"loss": 2.489,
"step": 5030
},
{
"epoch": 0.44489561724853244,
"grad_norm": 0.16015625,
"learning_rate": 0.0002968673159642271,
"loss": 2.5023,
"step": 5040
},
{
"epoch": 0.4457783466478351,
"grad_norm": 0.13671875,
"learning_rate": 0.00029618003559215276,
"loss": 2.5067,
"step": 5050
},
{
"epoch": 0.44666107604713773,
"grad_norm": 0.134765625,
"learning_rate": 0.0002954923936862689,
"loss": 2.4927,
"step": 5060
},
{
"epoch": 0.4475438054464404,
"grad_norm": 0.146484375,
"learning_rate": 0.00029480439562997964,
"loss": 2.5039,
"step": 5070
},
{
"epoch": 0.448426534845743,
"grad_norm": 0.13671875,
"learning_rate": 0.00029411604680947755,
"loss": 2.49,
"step": 5080
},
{
"epoch": 0.4493092642450457,
"grad_norm": 0.1435546875,
"learning_rate": 0.00029342735261370095,
"loss": 2.4923,
"step": 5090
},
{
"epoch": 0.4501919936443483,
"grad_norm": 0.138671875,
"learning_rate": 0.0002927383184342924,
"loss": 2.487,
"step": 5100
},
{
"epoch": 0.451074723043651,
"grad_norm": 0.1484375,
"learning_rate": 0.00029204894966555577,
"loss": 2.5052,
"step": 5110
},
{
"epoch": 0.4519574524429536,
"grad_norm": 0.1337890625,
"learning_rate": 0.00029135925170441457,
"loss": 2.5055,
"step": 5120
},
{
"epoch": 0.4528401818422563,
"grad_norm": 0.138671875,
"learning_rate": 0.0002906692299503694,
"loss": 2.5038,
"step": 5130
},
{
"epoch": 0.4537229112415589,
"grad_norm": 0.1474609375,
"learning_rate": 0.00028997888980545586,
"loss": 2.4906,
"step": 5140
},
{
"epoch": 0.45460564064086156,
"grad_norm": 0.138671875,
"learning_rate": 0.00028928823667420206,
"loss": 2.4871,
"step": 5150
},
{
"epoch": 0.4554883700401642,
"grad_norm": 0.154296875,
"learning_rate": 0.00028859727596358643,
"loss": 2.4991,
"step": 5160
},
{
"epoch": 0.45637109943946685,
"grad_norm": 0.1708984375,
"learning_rate": 0.00028790601308299545,
"loss": 2.4943,
"step": 5170
},
{
"epoch": 0.45725382883876947,
"grad_norm": 0.1650390625,
"learning_rate": 0.0002872144534441812,
"loss": 2.4936,
"step": 5180
},
{
"epoch": 0.45813655823807214,
"grad_norm": 0.1376953125,
"learning_rate": 0.0002865226024612189,
"loss": 2.5053,
"step": 5190
},
{
"epoch": 0.45901928763737476,
"grad_norm": 0.1474609375,
"learning_rate": 0.00028583046555046487,
"loss": 2.4854,
"step": 5200
},
{
"epoch": 0.4599020170366774,
"grad_norm": 0.1357421875,
"learning_rate": 0.0002851380481305136,
"loss": 2.4905,
"step": 5210
},
{
"epoch": 0.46078474643598005,
"grad_norm": 0.1328125,
"learning_rate": 0.00028444535562215594,
"loss": 2.4897,
"step": 5220
},
{
"epoch": 0.46166747583528267,
"grad_norm": 0.150390625,
"learning_rate": 0.00028375239344833616,
"loss": 2.49,
"step": 5230
},
{
"epoch": 0.46255020523458534,
"grad_norm": 0.142578125,
"learning_rate": 0.00028305916703410974,
"loss": 2.5025,
"step": 5240
},
{
"epoch": 0.46343293463388796,
"grad_norm": 0.181640625,
"learning_rate": 0.00028236568180660073,
"loss": 2.4854,
"step": 5250
},
{
"epoch": 0.46431566403319063,
"grad_norm": 0.1376953125,
"learning_rate": 0.0002816719431949596,
"loss": 2.4993,
"step": 5260
},
{
"epoch": 0.46519839343249325,
"grad_norm": 0.12890625,
"learning_rate": 0.0002809779566303203,
"loss": 2.507,
"step": 5270
},
{
"epoch": 0.4660811228317959,
"grad_norm": 0.16796875,
"learning_rate": 0.00028028372754575805,
"loss": 2.5041,
"step": 5280
},
{
"epoch": 0.46696385223109854,
"grad_norm": 0.138671875,
"learning_rate": 0.0002795892613762467,
"loss": 2.4892,
"step": 5290
},
{
"epoch": 0.4678465816304012,
"grad_norm": 0.1455078125,
"learning_rate": 0.00027889456355861635,
"loss": 2.5044,
"step": 5300
},
{
"epoch": 0.46872931102970383,
"grad_norm": 0.16015625,
"learning_rate": 0.00027819963953151024,
"loss": 2.4867,
"step": 5310
},
{
"epoch": 0.4696120404290065,
"grad_norm": 0.12890625,
"learning_rate": 0.0002775044947353428,
"loss": 2.5044,
"step": 5320
},
{
"epoch": 0.4704947698283091,
"grad_norm": 0.138671875,
"learning_rate": 0.0002768091346122569,
"loss": 2.4989,
"step": 5330
},
{
"epoch": 0.4713774992276118,
"grad_norm": 0.1357421875,
"learning_rate": 0.000276113564606081,
"loss": 2.4902,
"step": 5340
},
{
"epoch": 0.4722602286269144,
"grad_norm": 0.146484375,
"learning_rate": 0.00027541779016228664,
"loss": 2.4804,
"step": 5350
},
{
"epoch": 0.4731429580262171,
"grad_norm": 0.1376953125,
"learning_rate": 0.0002747218167279461,
"loss": 2.5,
"step": 5360
},
{
"epoch": 0.4740256874255197,
"grad_norm": 0.1279296875,
"learning_rate": 0.00027402564975168925,
"loss": 2.484,
"step": 5370
},
{
"epoch": 0.4749084168248224,
"grad_norm": 0.13671875,
"learning_rate": 0.0002733292946836615,
"loss": 2.4866,
"step": 5380
},
{
"epoch": 0.475791146224125,
"grad_norm": 0.146484375,
"learning_rate": 0.0002726327569754803,
"loss": 2.4957,
"step": 5390
},
{
"epoch": 0.4766738756234276,
"grad_norm": 0.1357421875,
"learning_rate": 0.00027193604208019346,
"loss": 2.5038,
"step": 5400
},
{
"epoch": 0.4775566050227303,
"grad_norm": 0.14453125,
"learning_rate": 0.0002712391554522355,
"loss": 2.4937,
"step": 5410
},
{
"epoch": 0.4784393344220329,
"grad_norm": 0.1435546875,
"learning_rate": 0.0002705421025473857,
"loss": 2.4954,
"step": 5420
},
{
"epoch": 0.4793220638213356,
"grad_norm": 0.14453125,
"learning_rate": 0.0002698448888227251,
"loss": 2.4861,
"step": 5430
},
{
"epoch": 0.4802047932206382,
"grad_norm": 0.1474609375,
"learning_rate": 0.0002691475197365936,
"loss": 2.4765,
"step": 5440
},
{
"epoch": 0.48108752261994087,
"grad_norm": 0.16015625,
"learning_rate": 0.00026845000074854754,
"loss": 2.5035,
"step": 5450
},
{
"epoch": 0.4819702520192435,
"grad_norm": 0.1455078125,
"learning_rate": 0.0002677523373193165,
"loss": 2.4953,
"step": 5460
},
{
"epoch": 0.48285298141854616,
"grad_norm": 0.130859375,
"learning_rate": 0.00026705453491076127,
"loss": 2.4888,
"step": 5470
},
{
"epoch": 0.4837357108178488,
"grad_norm": 0.18359375,
"learning_rate": 0.00026635659898583043,
"loss": 2.4894,
"step": 5480
},
{
"epoch": 0.48461844021715145,
"grad_norm": 0.12890625,
"learning_rate": 0.000265658535008518,
"loss": 2.5047,
"step": 5490
},
{
"epoch": 0.48550116961645406,
"grad_norm": 0.1494140625,
"learning_rate": 0.00026496034844382036,
"loss": 2.4942,
"step": 5500
},
{
"epoch": 0.48638389901575674,
"grad_norm": 0.134765625,
"learning_rate": 0.0002642620447576935,
"loss": 2.4822,
"step": 5510
},
{
"epoch": 0.48726662841505936,
"grad_norm": 0.146484375,
"learning_rate": 0.0002635636294170106,
"loss": 2.4992,
"step": 5520
},
{
"epoch": 0.48814935781436203,
"grad_norm": 0.138671875,
"learning_rate": 0.00026286510788951886,
"loss": 2.4969,
"step": 5530
},
{
"epoch": 0.48903208721366465,
"grad_norm": 0.1484375,
"learning_rate": 0.0002621664856437967,
"loss": 2.4898,
"step": 5540
},
{
"epoch": 0.4899148166129673,
"grad_norm": 0.154296875,
"learning_rate": 0.00026146776814921105,
"loss": 2.5002,
"step": 5550
},
{
"epoch": 0.49079754601226994,
"grad_norm": 0.15234375,
"learning_rate": 0.0002607689608758746,
"loss": 2.5127,
"step": 5560
},
{
"epoch": 0.4916802754115726,
"grad_norm": 0.134765625,
"learning_rate": 0.000260070069294603,
"loss": 2.4692,
"step": 5570
},
{
"epoch": 0.4925630048108752,
"grad_norm": 0.142578125,
"learning_rate": 0.00025937109887687164,
"loss": 2.495,
"step": 5580
},
{
"epoch": 0.49344573421017784,
"grad_norm": 0.134765625,
"learning_rate": 0.00025867205509477335,
"loss": 2.4888,
"step": 5590
},
{
"epoch": 0.4943284636094805,
"grad_norm": 0.1376953125,
"learning_rate": 0.0002579729434209752,
"loss": 2.4946,
"step": 5600
},
{
"epoch": 0.49521119300878313,
"grad_norm": 0.1318359375,
"learning_rate": 0.00025727376932867593,
"loss": 2.499,
"step": 5610
},
{
"epoch": 0.4960939224080858,
"grad_norm": 0.130859375,
"learning_rate": 0.00025657453829156256,
"loss": 2.4931,
"step": 5620
},
{
"epoch": 0.4969766518073884,
"grad_norm": 0.13671875,
"learning_rate": 0.00025587525578376843,
"loss": 2.4886,
"step": 5630
},
{
"epoch": 0.4978593812066911,
"grad_norm": 0.142578125,
"learning_rate": 0.0002551759272798295,
"loss": 2.4854,
"step": 5640
},
{
"epoch": 0.4987421106059937,
"grad_norm": 0.130859375,
"learning_rate": 0.00025447655825464174,
"loss": 2.5096,
"step": 5650
},
{
"epoch": 0.4996248400052964,
"grad_norm": 0.1328125,
"learning_rate": 0.0002537771541834187,
"loss": 2.4855,
"step": 5660
},
{
"epoch": 0.5005075694045991,
"grad_norm": 0.1318359375,
"learning_rate": 0.00025307772054164804,
"loss": 2.5022,
"step": 5670
},
{
"epoch": 0.5013902988039016,
"grad_norm": 0.12890625,
"learning_rate": 0.000252378262805049,
"loss": 2.4872,
"step": 5680
},
{
"epoch": 0.5022730282032043,
"grad_norm": 0.1376953125,
"learning_rate": 0.0002516787864495294,
"loss": 2.4977,
"step": 5690
},
{
"epoch": 0.503155757602507,
"grad_norm": 0.166015625,
"learning_rate": 0.00025097929695114295,
"loss": 2.4897,
"step": 5700
},
{
"epoch": 0.5040384870018096,
"grad_norm": 0.181640625,
"learning_rate": 0.00025027979978604615,
"loss": 2.4892,
"step": 5710
},
{
"epoch": 0.5049212164011122,
"grad_norm": 0.1455078125,
"learning_rate": 0.0002495803004304556,
"loss": 2.4854,
"step": 5720
},
{
"epoch": 0.5058039458004149,
"grad_norm": 0.1357421875,
"learning_rate": 0.0002488808043606048,
"loss": 2.4952,
"step": 5730
},
{
"epoch": 0.5066866751997176,
"grad_norm": 0.1416015625,
"learning_rate": 0.0002481813170527019,
"loss": 2.4973,
"step": 5740
},
{
"epoch": 0.5075694045990202,
"grad_norm": 0.1474609375,
"learning_rate": 0.0002474818439828862,
"loss": 2.4893,
"step": 5750
},
{
"epoch": 0.5084521339983228,
"grad_norm": 0.166015625,
"learning_rate": 0.0002467823906271856,
"loss": 2.4936,
"step": 5760
},
{
"epoch": 0.5093348633976255,
"grad_norm": 0.1357421875,
"learning_rate": 0.00024608296246147375,
"loss": 2.4944,
"step": 5770
},
{
"epoch": 0.5102175927969281,
"grad_norm": 0.15234375,
"learning_rate": 0.00024538356496142693,
"loss": 2.4862,
"step": 5780
},
{
"epoch": 0.5111003221962307,
"grad_norm": 0.1337890625,
"learning_rate": 0.00024468420360248145,
"loss": 2.4958,
"step": 5790
},
{
"epoch": 0.5119830515955334,
"grad_norm": 0.154296875,
"learning_rate": 0.00024398488385979055,
"loss": 2.4893,
"step": 5800
},
{
"epoch": 0.512865780994836,
"grad_norm": 0.138671875,
"learning_rate": 0.00024328561120818195,
"loss": 2.4969,
"step": 5810
},
{
"epoch": 0.5137485103941387,
"grad_norm": 0.1494140625,
"learning_rate": 0.00024258639112211453,
"loss": 2.5057,
"step": 5820
},
{
"epoch": 0.5146312397934413,
"grad_norm": 0.13671875,
"learning_rate": 0.00024188722907563537,
"loss": 2.4883,
"step": 5830
},
{
"epoch": 0.515513969192744,
"grad_norm": 0.138671875,
"learning_rate": 0.00024118813054233774,
"loss": 2.4837,
"step": 5840
},
{
"epoch": 0.5163966985920466,
"grad_norm": 0.142578125,
"learning_rate": 0.00024048910099531726,
"loss": 2.4996,
"step": 5850
},
{
"epoch": 0.5172794279913493,
"grad_norm": 0.1572265625,
"learning_rate": 0.00023979014590712962,
"loss": 2.4796,
"step": 5860
},
{
"epoch": 0.5181621573906519,
"grad_norm": 0.16015625,
"learning_rate": 0.00023909127074974744,
"loss": 2.4937,
"step": 5870
},
{
"epoch": 0.5190448867899545,
"grad_norm": 0.1318359375,
"learning_rate": 0.00023839248099451782,
"loss": 2.4882,
"step": 5880
},
{
"epoch": 0.5199276161892572,
"grad_norm": 0.13671875,
"learning_rate": 0.00023769378211211916,
"loss": 2.4752,
"step": 5890
},
{
"epoch": 0.5208103455885599,
"grad_norm": 0.130859375,
"learning_rate": 0.00023699517957251825,
"loss": 2.4822,
"step": 5900
},
{
"epoch": 0.5216930749878624,
"grad_norm": 0.142578125,
"learning_rate": 0.00023629667884492799,
"loss": 2.4913,
"step": 5910
},
{
"epoch": 0.5225758043871651,
"grad_norm": 0.1396484375,
"learning_rate": 0.00023559828539776394,
"loss": 2.4878,
"step": 5920
},
{
"epoch": 0.5234585337864678,
"grad_norm": 0.158203125,
"learning_rate": 0.00023490000469860185,
"loss": 2.487,
"step": 5930
},
{
"epoch": 0.5243412631857705,
"grad_norm": 0.125,
"learning_rate": 0.0002342018422141347,
"loss": 2.4839,
"step": 5940
},
{
"epoch": 0.525223992585073,
"grad_norm": 0.1376953125,
"learning_rate": 0.00023350380341013034,
"loss": 2.5015,
"step": 5950
},
{
"epoch": 0.5261067219843757,
"grad_norm": 0.1435546875,
"learning_rate": 0.000232805893751388,
"loss": 2.5031,
"step": 5960
},
{
"epoch": 0.5269894513836784,
"grad_norm": 0.1337890625,
"learning_rate": 0.0002321081187016959,
"loss": 2.4887,
"step": 5970
},
{
"epoch": 0.5278721807829809,
"grad_norm": 0.1982421875,
"learning_rate": 0.00023141048372378863,
"loss": 2.4911,
"step": 5980
},
{
"epoch": 0.5287549101822836,
"grad_norm": 0.1298828125,
"learning_rate": 0.00023071299427930396,
"loss": 2.4885,
"step": 5990
},
{
"epoch": 0.5296376395815863,
"grad_norm": 0.134765625,
"learning_rate": 0.00023001565582874046,
"loss": 2.4919,
"step": 6000
},
{
"epoch": 0.5296376395815863,
"eval_accuracy": 0.5102518027046329,
"eval_loss": 2.375392198562622,
"eval_runtime": 6.985,
"eval_samples_per_second": 45.526,
"eval_steps_per_second": 0.429,
"step": 6000
},
{
"epoch": 0.530520368980889,
"grad_norm": 0.1435546875,
"learning_rate": 0.00022931847383141446,
"loss": 2.4796,
"step": 6010
},
{
"epoch": 0.5314030983801915,
"grad_norm": 0.1337890625,
"learning_rate": 0.00022862145374541768,
"loss": 2.4886,
"step": 6020
},
{
"epoch": 0.5322858277794942,
"grad_norm": 0.138671875,
"learning_rate": 0.00022792460102757407,
"loss": 2.4902,
"step": 6030
},
{
"epoch": 0.5331685571787969,
"grad_norm": 0.1494140625,
"learning_rate": 0.00022722792113339722,
"loss": 2.4898,
"step": 6040
},
{
"epoch": 0.5340512865780995,
"grad_norm": 0.146484375,
"learning_rate": 0.0002265314195170481,
"loss": 2.5015,
"step": 6050
},
{
"epoch": 0.5349340159774021,
"grad_norm": 0.166015625,
"learning_rate": 0.00022583510163129162,
"loss": 2.4765,
"step": 6060
},
{
"epoch": 0.5358167453767048,
"grad_norm": 0.1416015625,
"learning_rate": 0.00022513897292745434,
"loss": 2.5056,
"step": 6070
},
{
"epoch": 0.5366994747760074,
"grad_norm": 0.125,
"learning_rate": 0.00022444303885538178,
"loss": 2.4953,
"step": 6080
},
{
"epoch": 0.5375822041753101,
"grad_norm": 0.1357421875,
"learning_rate": 0.000223747304863396,
"loss": 2.4893,
"step": 6090
},
{
"epoch": 0.5384649335746127,
"grad_norm": 0.1357421875,
"learning_rate": 0.0002230517763982523,
"loss": 2.5002,
"step": 6100
},
{
"epoch": 0.5393476629739153,
"grad_norm": 0.142578125,
"learning_rate": 0.0002223564589050971,
"loss": 2.4939,
"step": 6110
},
{
"epoch": 0.540230392373218,
"grad_norm": 0.1396484375,
"learning_rate": 0.00022166135782742525,
"loss": 2.4857,
"step": 6120
},
{
"epoch": 0.5411131217725207,
"grad_norm": 0.1455078125,
"learning_rate": 0.0002209664786070372,
"loss": 2.4857,
"step": 6130
},
{
"epoch": 0.5419958511718233,
"grad_norm": 0.1396484375,
"learning_rate": 0.00022027182668399653,
"loss": 2.4873,
"step": 6140
},
{
"epoch": 0.5428785805711259,
"grad_norm": 0.142578125,
"learning_rate": 0.0002195774074965874,
"loss": 2.486,
"step": 6150
},
{
"epoch": 0.5437613099704286,
"grad_norm": 0.1748046875,
"learning_rate": 0.00021888322648127206,
"loss": 2.4984,
"step": 6160
},
{
"epoch": 0.5446440393697312,
"grad_norm": 0.14453125,
"learning_rate": 0.0002181892890726479,
"loss": 2.4945,
"step": 6170
},
{
"epoch": 0.5455267687690338,
"grad_norm": 0.13671875,
"learning_rate": 0.00021749560070340534,
"loss": 2.4869,
"step": 6180
},
{
"epoch": 0.5464094981683365,
"grad_norm": 0.1318359375,
"learning_rate": 0.000216802166804285,
"loss": 2.4863,
"step": 6190
},
{
"epoch": 0.5472922275676392,
"grad_norm": 0.1357421875,
"learning_rate": 0.00021610899280403555,
"loss": 2.495,
"step": 6200
},
{
"epoch": 0.5481749569669417,
"grad_norm": 0.140625,
"learning_rate": 0.00021541608412937075,
"loss": 2.4794,
"step": 6210
},
{
"epoch": 0.5490576863662444,
"grad_norm": 0.142578125,
"learning_rate": 0.000214723446204927,
"loss": 2.4992,
"step": 6220
},
{
"epoch": 0.5499404157655471,
"grad_norm": 0.12890625,
"learning_rate": 0.00021403108445322168,
"loss": 2.4956,
"step": 6230
},
{
"epoch": 0.5508231451648498,
"grad_norm": 0.1513671875,
"learning_rate": 0.0002133390042946094,
"loss": 2.4832,
"step": 6240
},
{
"epoch": 0.5517058745641523,
"grad_norm": 0.14453125,
"learning_rate": 0.00021264721114724064,
"loss": 2.4871,
"step": 6250
},
{
"epoch": 0.552588603963455,
"grad_norm": 0.1357421875,
"learning_rate": 0.0002119557104270187,
"loss": 2.4968,
"step": 6260
},
{
"epoch": 0.5534713333627577,
"grad_norm": 0.15234375,
"learning_rate": 0.00021126450754755774,
"loss": 2.484,
"step": 6270
},
{
"epoch": 0.5543540627620603,
"grad_norm": 0.13671875,
"learning_rate": 0.00021057360792014004,
"loss": 2.4831,
"step": 6280
},
{
"epoch": 0.5552367921613629,
"grad_norm": 0.138671875,
"learning_rate": 0.0002098830169536738,
"loss": 2.4831,
"step": 6290
},
{
"epoch": 0.5561195215606656,
"grad_norm": 0.1298828125,
"learning_rate": 0.00020919274005465083,
"loss": 2.4874,
"step": 6300
},
{
"epoch": 0.5570022509599682,
"grad_norm": 0.1337890625,
"learning_rate": 0.00020850278262710416,
"loss": 2.492,
"step": 6310
},
{
"epoch": 0.5578849803592708,
"grad_norm": 0.146484375,
"learning_rate": 0.0002078131500725657,
"loss": 2.4912,
"step": 6320
},
{
"epoch": 0.5587677097585735,
"grad_norm": 0.158203125,
"learning_rate": 0.00020712384779002392,
"loss": 2.4868,
"step": 6330
},
{
"epoch": 0.5596504391578762,
"grad_norm": 0.1337890625,
"learning_rate": 0.00020643488117588199,
"loss": 2.486,
"step": 6340
},
{
"epoch": 0.5605331685571788,
"grad_norm": 0.138671875,
"learning_rate": 0.00020574625562391494,
"loss": 2.4893,
"step": 6350
},
{
"epoch": 0.5614158979564814,
"grad_norm": 0.12890625,
"learning_rate": 0.00020505797652522751,
"loss": 2.4902,
"step": 6360
},
{
"epoch": 0.5622986273557841,
"grad_norm": 0.140625,
"learning_rate": 0.00020437004926821255,
"loss": 2.4928,
"step": 6370
},
{
"epoch": 0.5631813567550867,
"grad_norm": 0.162109375,
"learning_rate": 0.00020368247923850826,
"loss": 2.4896,
"step": 6380
},
{
"epoch": 0.5640640861543894,
"grad_norm": 0.130859375,
"learning_rate": 0.00020299527181895602,
"loss": 2.4764,
"step": 6390
},
{
"epoch": 0.564946815553692,
"grad_norm": 0.14453125,
"learning_rate": 0.00020230843238955854,
"loss": 2.4788,
"step": 6400
},
{
"epoch": 0.5658295449529946,
"grad_norm": 0.1337890625,
"learning_rate": 0.0002016219663274377,
"loss": 2.4956,
"step": 6410
},
{
"epoch": 0.5667122743522973,
"grad_norm": 0.142578125,
"learning_rate": 0.00020093587900679217,
"loss": 2.4832,
"step": 6420
},
{
"epoch": 0.5675950037516,
"grad_norm": 0.1298828125,
"learning_rate": 0.00020025017579885563,
"loss": 2.499,
"step": 6430
},
{
"epoch": 0.5684777331509026,
"grad_norm": 0.1376953125,
"learning_rate": 0.00019956486207185477,
"loss": 2.488,
"step": 6440
},
{
"epoch": 0.5693604625502052,
"grad_norm": 0.126953125,
"learning_rate": 0.0001988799431909668,
"loss": 2.4964,
"step": 6450
},
{
"epoch": 0.5702431919495079,
"grad_norm": 0.1494140625,
"learning_rate": 0.00019819542451827808,
"loss": 2.4901,
"step": 6460
},
{
"epoch": 0.5711259213488106,
"grad_norm": 0.12890625,
"learning_rate": 0.00019751131141274147,
"loss": 2.4834,
"step": 6470
},
{
"epoch": 0.5720086507481131,
"grad_norm": 0.1416015625,
"learning_rate": 0.0001968276092301352,
"loss": 2.4859,
"step": 6480
},
{
"epoch": 0.5728913801474158,
"grad_norm": 0.12890625,
"learning_rate": 0.00019614432332302006,
"loss": 2.4835,
"step": 6490
},
{
"epoch": 0.5737741095467185,
"grad_norm": 0.1318359375,
"learning_rate": 0.00019546145904069808,
"loss": 2.4865,
"step": 6500
},
{
"epoch": 0.574656838946021,
"grad_norm": 0.1279296875,
"learning_rate": 0.00019477902172917045,
"loss": 2.4848,
"step": 6510
},
{
"epoch": 0.5755395683453237,
"grad_norm": 0.1416015625,
"learning_rate": 0.0001940970167310957,
"loss": 2.502,
"step": 6520
},
{
"epoch": 0.5764222977446264,
"grad_norm": 0.140625,
"learning_rate": 0.0001934154493857479,
"loss": 2.4871,
"step": 6530
},
{
"epoch": 0.5773050271439291,
"grad_norm": 0.138671875,
"learning_rate": 0.0001927343250289747,
"loss": 2.5027,
"step": 6540
},
{
"epoch": 0.5781877565432316,
"grad_norm": 0.1396484375,
"learning_rate": 0.00019205364899315593,
"loss": 2.4757,
"step": 6550
},
{
"epoch": 0.5790704859425343,
"grad_norm": 0.1416015625,
"learning_rate": 0.00019137342660716133,
"loss": 2.4887,
"step": 6560
},
{
"epoch": 0.579953215341837,
"grad_norm": 0.1552734375,
"learning_rate": 0.00019069366319630923,
"loss": 2.488,
"step": 6570
},
{
"epoch": 0.5808359447411396,
"grad_norm": 0.13671875,
"learning_rate": 0.00019001436408232496,
"loss": 2.4843,
"step": 6580
},
{
"epoch": 0.5817186741404422,
"grad_norm": 0.1298828125,
"learning_rate": 0.00018933553458329856,
"loss": 2.4844,
"step": 6590
},
{
"epoch": 0.5826014035397449,
"grad_norm": 0.126953125,
"learning_rate": 0.00018865718001364375,
"loss": 2.477,
"step": 6600
},
{
"epoch": 0.5834841329390476,
"grad_norm": 0.138671875,
"learning_rate": 0.00018797930568405612,
"loss": 2.4859,
"step": 6610
},
{
"epoch": 0.5843668623383502,
"grad_norm": 0.1357421875,
"learning_rate": 0.00018730191690147176,
"loss": 2.4824,
"step": 6620
},
{
"epoch": 0.5852495917376528,
"grad_norm": 0.1259765625,
"learning_rate": 0.00018662501896902519,
"loss": 2.4692,
"step": 6630
},
{
"epoch": 0.5861323211369555,
"grad_norm": 0.1298828125,
"learning_rate": 0.0001859486171860082,
"loss": 2.4747,
"step": 6640
},
{
"epoch": 0.5870150505362581,
"grad_norm": 0.1357421875,
"learning_rate": 0.00018527271684782865,
"loss": 2.4858,
"step": 6650
},
{
"epoch": 0.5878977799355608,
"grad_norm": 0.13671875,
"learning_rate": 0.00018459732324596834,
"loss": 2.4894,
"step": 6660
},
{
"epoch": 0.5887805093348634,
"grad_norm": 0.12890625,
"learning_rate": 0.0001839224416679421,
"loss": 2.5035,
"step": 6670
},
{
"epoch": 0.589663238734166,
"grad_norm": 0.1396484375,
"learning_rate": 0.00018324807739725614,
"loss": 2.4819,
"step": 6680
},
{
"epoch": 0.5905459681334687,
"grad_norm": 0.1357421875,
"learning_rate": 0.000182574235713367,
"loss": 2.4958,
"step": 6690
},
{
"epoch": 0.5914286975327713,
"grad_norm": 0.140625,
"learning_rate": 0.00018190092189163974,
"loss": 2.5124,
"step": 6700
},
{
"epoch": 0.592311426932074,
"grad_norm": 0.14453125,
"learning_rate": 0.00018122814120330688,
"loss": 2.4775,
"step": 6710
},
{
"epoch": 0.5931941563313766,
"grad_norm": 0.146484375,
"learning_rate": 0.00018055589891542758,
"loss": 2.4865,
"step": 6720
},
{
"epoch": 0.5940768857306793,
"grad_norm": 0.13671875,
"learning_rate": 0.00017988420029084551,
"loss": 2.4782,
"step": 6730
},
{
"epoch": 0.5949596151299819,
"grad_norm": 0.130859375,
"learning_rate": 0.00017921305058814818,
"loss": 2.4903,
"step": 6740
},
{
"epoch": 0.5958423445292845,
"grad_norm": 0.1328125,
"learning_rate": 0.00017854245506162582,
"loss": 2.4786,
"step": 6750
},
{
"epoch": 0.5967250739285872,
"grad_norm": 0.1298828125,
"learning_rate": 0.00017787241896123024,
"loss": 2.4919,
"step": 6760
},
{
"epoch": 0.5976078033278899,
"grad_norm": 0.130859375,
"learning_rate": 0.00017720294753253345,
"loss": 2.4921,
"step": 6770
},
{
"epoch": 0.5984905327271924,
"grad_norm": 0.154296875,
"learning_rate": 0.00017653404601668666,
"loss": 2.4775,
"step": 6780
},
{
"epoch": 0.5993732621264951,
"grad_norm": 0.15625,
"learning_rate": 0.00017586571965037966,
"loss": 2.491,
"step": 6790
},
{
"epoch": 0.6002559915257978,
"grad_norm": 0.1455078125,
"learning_rate": 0.0001751979736657993,
"loss": 2.4803,
"step": 6800
},
{
"epoch": 0.6011387209251005,
"grad_norm": 0.130859375,
"learning_rate": 0.00017453081329058882,
"loss": 2.4797,
"step": 6810
},
{
"epoch": 0.602021450324403,
"grad_norm": 0.138671875,
"learning_rate": 0.0001738642437478067,
"loss": 2.4761,
"step": 6820
},
{
"epoch": 0.6029041797237057,
"grad_norm": 0.1328125,
"learning_rate": 0.00017319827025588614,
"loss": 2.4584,
"step": 6830
},
{
"epoch": 0.6037869091230084,
"grad_norm": 0.1337890625,
"learning_rate": 0.0001725328980285939,
"loss": 2.4875,
"step": 6840
},
{
"epoch": 0.604669638522311,
"grad_norm": 0.1337890625,
"learning_rate": 0.00017186813227498937,
"loss": 2.4854,
"step": 6850
},
{
"epoch": 0.6055523679216136,
"grad_norm": 0.1279296875,
"learning_rate": 0.0001712039781993844,
"loss": 2.4811,
"step": 6860
},
{
"epoch": 0.6064350973209163,
"grad_norm": 0.150390625,
"learning_rate": 0.00017054044100130178,
"loss": 2.4804,
"step": 6870
},
{
"epoch": 0.607317826720219,
"grad_norm": 0.1298828125,
"learning_rate": 0.0001698775258754351,
"loss": 2.4854,
"step": 6880
},
{
"epoch": 0.6082005561195215,
"grad_norm": 0.1337890625,
"learning_rate": 0.00016921523801160756,
"loss": 2.4895,
"step": 6890
},
{
"epoch": 0.6090832855188242,
"grad_norm": 0.1357421875,
"learning_rate": 0.00016855358259473217,
"loss": 2.4823,
"step": 6900
},
{
"epoch": 0.6099660149181269,
"grad_norm": 0.1494140625,
"learning_rate": 0.00016789256480477023,
"loss": 2.4752,
"step": 6910
},
{
"epoch": 0.6108487443174295,
"grad_norm": 0.1552734375,
"learning_rate": 0.00016723218981669127,
"loss": 2.4758,
"step": 6920
},
{
"epoch": 0.6117314737167321,
"grad_norm": 0.1201171875,
"learning_rate": 0.00016657246280043266,
"loss": 2.4934,
"step": 6930
},
{
"epoch": 0.6126142031160348,
"grad_norm": 0.1376953125,
"learning_rate": 0.00016591338892085874,
"loss": 2.488,
"step": 6940
},
{
"epoch": 0.6134969325153374,
"grad_norm": 0.1376953125,
"learning_rate": 0.0001652549733377206,
"loss": 2.4797,
"step": 6950
},
{
"epoch": 0.6143796619146401,
"grad_norm": 0.1259765625,
"learning_rate": 0.00016459722120561567,
"loss": 2.467,
"step": 6960
},
{
"epoch": 0.6152623913139427,
"grad_norm": 0.1337890625,
"learning_rate": 0.0001639401376739475,
"loss": 2.4972,
"step": 6970
},
{
"epoch": 0.6161451207132453,
"grad_norm": 0.1318359375,
"learning_rate": 0.0001632837278868851,
"loss": 2.4728,
"step": 6980
},
{
"epoch": 0.617027850112548,
"grad_norm": 0.142578125,
"learning_rate": 0.00016262799698332292,
"loss": 2.4723,
"step": 6990
},
{
"epoch": 0.6179105795118507,
"grad_norm": 0.1337890625,
"learning_rate": 0.00016197295009684077,
"loss": 2.4765,
"step": 7000
},
{
"epoch": 0.6187933089111533,
"grad_norm": 0.130859375,
"learning_rate": 0.00016131859235566325,
"loss": 2.476,
"step": 7010
},
{
"epoch": 0.6196760383104559,
"grad_norm": 0.12890625,
"learning_rate": 0.00016066492888261983,
"loss": 2.4959,
"step": 7020
},
{
"epoch": 0.6205587677097586,
"grad_norm": 0.1259765625,
"learning_rate": 0.00016001196479510448,
"loss": 2.4946,
"step": 7030
},
{
"epoch": 0.6214414971090613,
"grad_norm": 0.1259765625,
"learning_rate": 0.00015935970520503638,
"loss": 2.4895,
"step": 7040
},
{
"epoch": 0.6223242265083638,
"grad_norm": 0.1240234375,
"learning_rate": 0.0001587081552188188,
"loss": 2.4848,
"step": 7050
},
{
"epoch": 0.6232069559076665,
"grad_norm": 0.125,
"learning_rate": 0.0001580573199372999,
"loss": 2.4822,
"step": 7060
},
{
"epoch": 0.6240896853069692,
"grad_norm": 0.126953125,
"learning_rate": 0.00015740720445573262,
"loss": 2.4835,
"step": 7070
},
{
"epoch": 0.6249724147062717,
"grad_norm": 0.15234375,
"learning_rate": 0.00015675781386373462,
"loss": 2.4818,
"step": 7080
},
{
"epoch": 0.6258551441055744,
"grad_norm": 0.1435546875,
"learning_rate": 0.0001561091532452486,
"loss": 2.4918,
"step": 7090
},
{
"epoch": 0.6267378735048771,
"grad_norm": 0.134765625,
"learning_rate": 0.00015546122767850232,
"loss": 2.4882,
"step": 7100
},
{
"epoch": 0.6276206029041798,
"grad_norm": 0.126953125,
"learning_rate": 0.00015481404223596939,
"loss": 2.4933,
"step": 7110
},
{
"epoch": 0.6285033323034823,
"grad_norm": 0.1357421875,
"learning_rate": 0.0001541676019843286,
"loss": 2.4844,
"step": 7120
},
{
"epoch": 0.629386061702785,
"grad_norm": 0.1259765625,
"learning_rate": 0.00015352191198442507,
"loss": 2.4719,
"step": 7130
},
{
"epoch": 0.6302687911020877,
"grad_norm": 0.1572265625,
"learning_rate": 0.00015287697729123045,
"loss": 2.4803,
"step": 7140
},
{
"epoch": 0.6311515205013903,
"grad_norm": 0.130859375,
"learning_rate": 0.0001522328029538031,
"loss": 2.4894,
"step": 7150
},
{
"epoch": 0.6320342499006929,
"grad_norm": 0.130859375,
"learning_rate": 0.00015158939401524877,
"loss": 2.491,
"step": 7160
},
{
"epoch": 0.6329169792999956,
"grad_norm": 0.16015625,
"learning_rate": 0.00015094675551268096,
"loss": 2.487,
"step": 7170
},
{
"epoch": 0.6337997086992982,
"grad_norm": 0.134765625,
"learning_rate": 0.00015030489247718173,
"loss": 2.4762,
"step": 7180
},
{
"epoch": 0.6346824380986009,
"grad_norm": 0.134765625,
"learning_rate": 0.00014966380993376217,
"loss": 2.4862,
"step": 7190
},
{
"epoch": 0.6355651674979035,
"grad_norm": 0.1484375,
"learning_rate": 0.0001490235129013228,
"loss": 2.4853,
"step": 7200
},
{
"epoch": 0.6364478968972062,
"grad_norm": 0.1416015625,
"learning_rate": 0.00014838400639261503,
"loss": 2.4968,
"step": 7210
},
{
"epoch": 0.6373306262965088,
"grad_norm": 0.1572265625,
"learning_rate": 0.000147745295414201,
"loss": 2.4892,
"step": 7220
},
{
"epoch": 0.6382133556958115,
"grad_norm": 0.1337890625,
"learning_rate": 0.00014710738496641492,
"loss": 2.4631,
"step": 7230
},
{
"epoch": 0.6390960850951141,
"grad_norm": 0.1318359375,
"learning_rate": 0.0001464702800433238,
"loss": 2.467,
"step": 7240
},
{
"epoch": 0.6399788144944167,
"grad_norm": 0.12890625,
"learning_rate": 0.00014583398563268858,
"loss": 2.4861,
"step": 7250
},
{
"epoch": 0.6408615438937194,
"grad_norm": 0.1396484375,
"learning_rate": 0.00014519850671592467,
"loss": 2.493,
"step": 7260
},
{
"epoch": 0.641744273293022,
"grad_norm": 0.1513671875,
"learning_rate": 0.000144563848268063,
"loss": 2.4982,
"step": 7270
},
{
"epoch": 0.6426270026923246,
"grad_norm": 0.1318359375,
"learning_rate": 0.00014393001525771153,
"loss": 2.4848,
"step": 7280
},
{
"epoch": 0.6435097320916273,
"grad_norm": 0.1513671875,
"learning_rate": 0.00014329701264701597,
"loss": 2.4833,
"step": 7290
},
{
"epoch": 0.64439246149093,
"grad_norm": 0.1298828125,
"learning_rate": 0.0001426648453916208,
"loss": 2.4801,
"step": 7300
},
{
"epoch": 0.6452751908902326,
"grad_norm": 0.1298828125,
"learning_rate": 0.00014203351844063088,
"loss": 2.4711,
"step": 7310
},
{
"epoch": 0.6461579202895352,
"grad_norm": 0.1435546875,
"learning_rate": 0.0001414030367365725,
"loss": 2.4779,
"step": 7320
},
{
"epoch": 0.6470406496888379,
"grad_norm": 0.1328125,
"learning_rate": 0.00014077340521535472,
"loss": 2.4893,
"step": 7330
},
{
"epoch": 0.6479233790881406,
"grad_norm": 0.1328125,
"learning_rate": 0.00014014462880623042,
"loss": 2.4746,
"step": 7340
},
{
"epoch": 0.6488061084874431,
"grad_norm": 0.1318359375,
"learning_rate": 0.00013951671243175824,
"loss": 2.4794,
"step": 7350
},
{
"epoch": 0.6496888378867458,
"grad_norm": 0.1279296875,
"learning_rate": 0.00013888966100776386,
"loss": 2.4845,
"step": 7360
},
{
"epoch": 0.6505715672860485,
"grad_norm": 0.1376953125,
"learning_rate": 0.00013826347944330116,
"loss": 2.4637,
"step": 7370
},
{
"epoch": 0.6514542966853512,
"grad_norm": 0.1328125,
"learning_rate": 0.00013763817264061425,
"loss": 2.493,
"step": 7380
},
{
"epoch": 0.6523370260846537,
"grad_norm": 0.1416015625,
"learning_rate": 0.00013701374549509899,
"loss": 2.4888,
"step": 7390
},
{
"epoch": 0.6532197554839564,
"grad_norm": 0.1396484375,
"learning_rate": 0.00013639020289526438,
"loss": 2.4966,
"step": 7400
},
{
"epoch": 0.6541024848832591,
"grad_norm": 0.126953125,
"learning_rate": 0.00013576754972269463,
"loss": 2.4908,
"step": 7410
},
{
"epoch": 0.6549852142825617,
"grad_norm": 0.130859375,
"learning_rate": 0.0001351457908520109,
"loss": 2.4792,
"step": 7420
},
{
"epoch": 0.6558679436818643,
"grad_norm": 0.13671875,
"learning_rate": 0.0001345249311508328,
"loss": 2.4821,
"step": 7430
},
{
"epoch": 0.656750673081167,
"grad_norm": 0.140625,
"learning_rate": 0.00013390497547974078,
"loss": 2.4825,
"step": 7440
},
{
"epoch": 0.6576334024804696,
"grad_norm": 0.1328125,
"learning_rate": 0.00013328592869223747,
"loss": 2.4828,
"step": 7450
},
{
"epoch": 0.6585161318797722,
"grad_norm": 0.1259765625,
"learning_rate": 0.00013266779563471064,
"loss": 2.4791,
"step": 7460
},
{
"epoch": 0.6593988612790749,
"grad_norm": 0.12890625,
"learning_rate": 0.00013205058114639407,
"loss": 2.486,
"step": 7470
},
{
"epoch": 0.6602815906783776,
"grad_norm": 0.150390625,
"learning_rate": 0.00013143429005933052,
"loss": 2.484,
"step": 7480
},
{
"epoch": 0.6611643200776802,
"grad_norm": 0.125,
"learning_rate": 0.00013081892719833378,
"loss": 2.4683,
"step": 7490
},
{
"epoch": 0.6620470494769828,
"grad_norm": 0.12451171875,
"learning_rate": 0.0001302044973809503,
"loss": 2.484,
"step": 7500
},
{
"epoch": 0.6629297788762855,
"grad_norm": 0.134765625,
"learning_rate": 0.00012959100541742248,
"loss": 2.4898,
"step": 7510
},
{
"epoch": 0.6638125082755881,
"grad_norm": 0.13671875,
"learning_rate": 0.0001289784561106499,
"loss": 2.4868,
"step": 7520
},
{
"epoch": 0.6646952376748908,
"grad_norm": 0.140625,
"learning_rate": 0.00012836685425615275,
"loss": 2.4984,
"step": 7530
},
{
"epoch": 0.6655779670741934,
"grad_norm": 0.130859375,
"learning_rate": 0.00012775620464203365,
"loss": 2.4814,
"step": 7540
},
{
"epoch": 0.666460696473496,
"grad_norm": 0.1279296875,
"learning_rate": 0.0001271465120489401,
"loss": 2.4742,
"step": 7550
},
{
"epoch": 0.6673434258727987,
"grad_norm": 0.126953125,
"learning_rate": 0.0001265377812500278,
"loss": 2.4828,
"step": 7560
},
{
"epoch": 0.6682261552721014,
"grad_norm": 0.12890625,
"learning_rate": 0.00012593001701092233,
"loss": 2.482,
"step": 7570
},
{
"epoch": 0.669108884671404,
"grad_norm": 0.1279296875,
"learning_rate": 0.00012532322408968221,
"loss": 2.4777,
"step": 7580
},
{
"epoch": 0.6699916140707066,
"grad_norm": 0.1376953125,
"learning_rate": 0.00012471740723676213,
"loss": 2.4852,
"step": 7590
},
{
"epoch": 0.6708743434700093,
"grad_norm": 0.1240234375,
"learning_rate": 0.000124112571194975,
"loss": 2.4819,
"step": 7600
},
{
"epoch": 0.671757072869312,
"grad_norm": 0.1337890625,
"learning_rate": 0.00012350872069945547,
"loss": 2.4826,
"step": 7610
},
{
"epoch": 0.6726398022686145,
"grad_norm": 0.13671875,
"learning_rate": 0.00012290586047762216,
"loss": 2.4811,
"step": 7620
},
{
"epoch": 0.6735225316679172,
"grad_norm": 0.12890625,
"learning_rate": 0.00012230399524914136,
"loss": 2.472,
"step": 7630
},
{
"epoch": 0.6744052610672199,
"grad_norm": 0.1298828125,
"learning_rate": 0.00012170312972588974,
"loss": 2.469,
"step": 7640
},
{
"epoch": 0.6752879904665224,
"grad_norm": 0.1240234375,
"learning_rate": 0.00012110326861191722,
"loss": 2.4757,
"step": 7650
},
{
"epoch": 0.6761707198658251,
"grad_norm": 0.1357421875,
"learning_rate": 0.00012050441660341074,
"loss": 2.4811,
"step": 7660
},
{
"epoch": 0.6770534492651278,
"grad_norm": 0.138671875,
"learning_rate": 0.00011990657838865706,
"loss": 2.4748,
"step": 7670
},
{
"epoch": 0.6779361786644305,
"grad_norm": 0.126953125,
"learning_rate": 0.00011930975864800603,
"loss": 2.478,
"step": 7680
},
{
"epoch": 0.678818908063733,
"grad_norm": 0.1318359375,
"learning_rate": 0.0001187139620538342,
"loss": 2.4918,
"step": 7690
},
{
"epoch": 0.6797016374630357,
"grad_norm": 0.1376953125,
"learning_rate": 0.0001181191932705081,
"loss": 2.4842,
"step": 7700
},
{
"epoch": 0.6805843668623384,
"grad_norm": 0.12890625,
"learning_rate": 0.00011752545695434788,
"loss": 2.491,
"step": 7710
},
{
"epoch": 0.681467096261641,
"grad_norm": 0.12890625,
"learning_rate": 0.00011693275775359049,
"loss": 2.4993,
"step": 7720
},
{
"epoch": 0.6823498256609436,
"grad_norm": 0.1259765625,
"learning_rate": 0.00011634110030835341,
"loss": 2.4743,
"step": 7730
},
{
"epoch": 0.6832325550602463,
"grad_norm": 0.12890625,
"learning_rate": 0.000115750489250599,
"loss": 2.4776,
"step": 7740
},
{
"epoch": 0.684115284459549,
"grad_norm": 0.12890625,
"learning_rate": 0.00011516092920409706,
"loss": 2.4867,
"step": 7750
},
{
"epoch": 0.6849980138588516,
"grad_norm": 0.1298828125,
"learning_rate": 0.00011457242478438962,
"loss": 2.4774,
"step": 7760
},
{
"epoch": 0.6858807432581542,
"grad_norm": 0.1376953125,
"learning_rate": 0.00011398498059875434,
"loss": 2.4809,
"step": 7770
},
{
"epoch": 0.6867634726574569,
"grad_norm": 0.1337890625,
"learning_rate": 0.00011339860124616833,
"loss": 2.4625,
"step": 7780
},
{
"epoch": 0.6876462020567595,
"grad_norm": 0.1259765625,
"learning_rate": 0.00011281329131727272,
"loss": 2.4775,
"step": 7790
},
{
"epoch": 0.6885289314560622,
"grad_norm": 0.12890625,
"learning_rate": 0.00011222905539433593,
"loss": 2.4737,
"step": 7800
},
{
"epoch": 0.6894116608553648,
"grad_norm": 0.1259765625,
"learning_rate": 0.00011164589805121852,
"loss": 2.4744,
"step": 7810
},
{
"epoch": 0.6902943902546674,
"grad_norm": 0.13671875,
"learning_rate": 0.00011106382385333708,
"loss": 2.4642,
"step": 7820
},
{
"epoch": 0.6911771196539701,
"grad_norm": 0.1337890625,
"learning_rate": 0.00011048283735762806,
"loss": 2.4921,
"step": 7830
},
{
"epoch": 0.6920598490532727,
"grad_norm": 0.123046875,
"learning_rate": 0.00010990294311251328,
"loss": 2.4849,
"step": 7840
},
{
"epoch": 0.6929425784525753,
"grad_norm": 0.1376953125,
"learning_rate": 0.00010932414565786286,
"loss": 2.4835,
"step": 7850
},
{
"epoch": 0.693825307851878,
"grad_norm": 0.1240234375,
"learning_rate": 0.0001087464495249606,
"loss": 2.49,
"step": 7860
},
{
"epoch": 0.6947080372511807,
"grad_norm": 0.1201171875,
"learning_rate": 0.00010816985923646838,
"loss": 2.48,
"step": 7870
},
{
"epoch": 0.6955907666504832,
"grad_norm": 0.1552734375,
"learning_rate": 0.00010759437930639058,
"loss": 2.4782,
"step": 7880
},
{
"epoch": 0.6964734960497859,
"grad_norm": 0.12255859375,
"learning_rate": 0.00010702001424003896,
"loss": 2.4723,
"step": 7890
},
{
"epoch": 0.6973562254490886,
"grad_norm": 0.12353515625,
"learning_rate": 0.00010644676853399688,
"loss": 2.4662,
"step": 7900
},
{
"epoch": 0.6982389548483913,
"grad_norm": 0.1328125,
"learning_rate": 0.00010587464667608484,
"loss": 2.493,
"step": 7910
},
{
"epoch": 0.6991216842476938,
"grad_norm": 0.13671875,
"learning_rate": 0.00010530365314532488,
"loss": 2.4957,
"step": 7920
},
{
"epoch": 0.7000044136469965,
"grad_norm": 0.158203125,
"learning_rate": 0.00010473379241190542,
"loss": 2.4864,
"step": 7930
},
{
"epoch": 0.7008871430462992,
"grad_norm": 0.1318359375,
"learning_rate": 0.00010416506893714662,
"loss": 2.48,
"step": 7940
},
{
"epoch": 0.7017698724456018,
"grad_norm": 0.1357421875,
"learning_rate": 0.00010359748717346534,
"loss": 2.4745,
"step": 7950
},
{
"epoch": 0.7026526018449044,
"grad_norm": 0.12890625,
"learning_rate": 0.00010303105156433998,
"loss": 2.4918,
"step": 7960
},
{
"epoch": 0.7035353312442071,
"grad_norm": 0.12353515625,
"learning_rate": 0.00010246576654427611,
"loss": 2.4878,
"step": 7970
},
{
"epoch": 0.7044180606435098,
"grad_norm": 0.1259765625,
"learning_rate": 0.0001019016365387716,
"loss": 2.4767,
"step": 7980
},
{
"epoch": 0.7053007900428124,
"grad_norm": 0.1240234375,
"learning_rate": 0.00010133866596428196,
"loss": 2.4814,
"step": 7990
},
{
"epoch": 0.706183519442115,
"grad_norm": 0.12353515625,
"learning_rate": 0.0001007768592281856,
"loss": 2.4884,
"step": 8000
},
{
"epoch": 0.706183519442115,
"eval_accuracy": 0.5112470338885433,
"eval_loss": 2.367577075958252,
"eval_runtime": 6.9898,
"eval_samples_per_second": 45.495,
"eval_steps_per_second": 0.429,
"step": 8000
},
{
"epoch": 0.7070662488414177,
"grad_norm": 0.138671875,
"learning_rate": 0.00010021622072874948,
"loss": 2.4866,
"step": 8010
},
{
"epoch": 0.7079489782407203,
"grad_norm": 0.1318359375,
"learning_rate": 9.965675485509504e-05,
"loss": 2.4817,
"step": 8020
},
{
"epoch": 0.7088317076400229,
"grad_norm": 0.13671875,
"learning_rate": 9.909846598716302e-05,
"loss": 2.4782,
"step": 8030
},
{
"epoch": 0.7097144370393256,
"grad_norm": 0.1240234375,
"learning_rate": 9.854135849567988e-05,
"loss": 2.4829,
"step": 8040
},
{
"epoch": 0.7105971664386282,
"grad_norm": 0.1328125,
"learning_rate": 9.79854367421234e-05,
"loss": 2.48,
"step": 8050
},
{
"epoch": 0.7114798958379309,
"grad_norm": 0.1279296875,
"learning_rate": 9.743070507868818e-05,
"loss": 2.4849,
"step": 8060
},
{
"epoch": 0.7123626252372335,
"grad_norm": 0.126953125,
"learning_rate": 9.687716784825218e-05,
"loss": 2.485,
"step": 8070
},
{
"epoch": 0.7132453546365362,
"grad_norm": 0.1494140625,
"learning_rate": 9.632482938434197e-05,
"loss": 2.477,
"step": 8080
},
{
"epoch": 0.7141280840358388,
"grad_norm": 0.1279296875,
"learning_rate": 9.577369401109987e-05,
"loss": 2.4836,
"step": 8090
},
{
"epoch": 0.7150108134351415,
"grad_norm": 0.12353515625,
"learning_rate": 9.522376604324889e-05,
"loss": 2.4868,
"step": 8100
},
{
"epoch": 0.7158935428344441,
"grad_norm": 0.1298828125,
"learning_rate": 9.467504978605956e-05,
"loss": 2.4857,
"step": 8110
},
{
"epoch": 0.7167762722337467,
"grad_norm": 0.126953125,
"learning_rate": 9.412754953531663e-05,
"loss": 2.4781,
"step": 8120
},
{
"epoch": 0.7176590016330494,
"grad_norm": 0.12255859375,
"learning_rate": 9.35812695772845e-05,
"loss": 2.4731,
"step": 8130
},
{
"epoch": 0.7185417310323521,
"grad_norm": 0.130859375,
"learning_rate": 9.303621418867444e-05,
"loss": 2.4815,
"step": 8140
},
{
"epoch": 0.7194244604316546,
"grad_norm": 0.12353515625,
"learning_rate": 9.24923876366106e-05,
"loss": 2.4876,
"step": 8150
},
{
"epoch": 0.7203071898309573,
"grad_norm": 0.125,
"learning_rate": 9.194979417859705e-05,
"loss": 2.4689,
"step": 8160
},
{
"epoch": 0.72118991923026,
"grad_norm": 0.12255859375,
"learning_rate": 9.14084380624842e-05,
"loss": 2.4709,
"step": 8170
},
{
"epoch": 0.7220726486295626,
"grad_norm": 0.1220703125,
"learning_rate": 9.086832352643535e-05,
"loss": 2.4813,
"step": 8180
},
{
"epoch": 0.7229553780288652,
"grad_norm": 0.130859375,
"learning_rate": 9.032945479889391e-05,
"loss": 2.4797,
"step": 8190
},
{
"epoch": 0.7238381074281679,
"grad_norm": 0.125,
"learning_rate": 8.979183609855024e-05,
"loss": 2.4909,
"step": 8200
},
{
"epoch": 0.7247208368274706,
"grad_norm": 0.1259765625,
"learning_rate": 8.925547163430812e-05,
"loss": 2.4756,
"step": 8210
},
{
"epoch": 0.7256035662267731,
"grad_norm": 0.12890625,
"learning_rate": 8.872036560525254e-05,
"loss": 2.466,
"step": 8220
},
{
"epoch": 0.7264862956260758,
"grad_norm": 0.1298828125,
"learning_rate": 8.818652220061638e-05,
"loss": 2.4651,
"step": 8230
},
{
"epoch": 0.7273690250253785,
"grad_norm": 0.13671875,
"learning_rate": 8.76539455997475e-05,
"loss": 2.482,
"step": 8240
},
{
"epoch": 0.7282517544246812,
"grad_norm": 0.123046875,
"learning_rate": 8.71226399720764e-05,
"loss": 2.4888,
"step": 8250
},
{
"epoch": 0.7291344838239837,
"grad_norm": 0.1279296875,
"learning_rate": 8.659260947708344e-05,
"loss": 2.4901,
"step": 8260
},
{
"epoch": 0.7300172132232864,
"grad_norm": 0.123046875,
"learning_rate": 8.606385826426621e-05,
"loss": 2.4829,
"step": 8270
},
{
"epoch": 0.7308999426225891,
"grad_norm": 0.1279296875,
"learning_rate": 8.553639047310685e-05,
"loss": 2.4875,
"step": 8280
},
{
"epoch": 0.7317826720218917,
"grad_norm": 0.1259765625,
"learning_rate": 8.50102102330401e-05,
"loss": 2.4883,
"step": 8290
},
{
"epoch": 0.7326654014211943,
"grad_norm": 0.12890625,
"learning_rate": 8.448532166342077e-05,
"loss": 2.4676,
"step": 8300
},
{
"epoch": 0.733548130820497,
"grad_norm": 0.125,
"learning_rate": 8.396172887349115e-05,
"loss": 2.4799,
"step": 8310
},
{
"epoch": 0.7344308602197996,
"grad_norm": 0.125,
"learning_rate": 8.343943596234943e-05,
"loss": 2.4862,
"step": 8320
},
{
"epoch": 0.7353135896191023,
"grad_norm": 0.1279296875,
"learning_rate": 8.291844701891732e-05,
"loss": 2.4746,
"step": 8330
},
{
"epoch": 0.7361963190184049,
"grad_norm": 0.12060546875,
"learning_rate": 8.239876612190778e-05,
"loss": 2.4751,
"step": 8340
},
{
"epoch": 0.7370790484177075,
"grad_norm": 0.12158203125,
"learning_rate": 8.188039733979366e-05,
"loss": 2.4868,
"step": 8350
},
{
"epoch": 0.7379617778170102,
"grad_norm": 0.125,
"learning_rate": 8.136334473077519e-05,
"loss": 2.4861,
"step": 8360
},
{
"epoch": 0.7388445072163128,
"grad_norm": 0.12255859375,
"learning_rate": 8.084761234274906e-05,
"loss": 2.4648,
"step": 8370
},
{
"epoch": 0.7397272366156155,
"grad_norm": 0.1328125,
"learning_rate": 8.033320421327578e-05,
"loss": 2.4733,
"step": 8380
},
{
"epoch": 0.7406099660149181,
"grad_norm": 0.125,
"learning_rate": 7.982012436954849e-05,
"loss": 2.4632,
"step": 8390
},
{
"epoch": 0.7414926954142208,
"grad_norm": 0.1318359375,
"learning_rate": 7.930837682836195e-05,
"loss": 2.4826,
"step": 8400
},
{
"epoch": 0.7423754248135234,
"grad_norm": 0.1279296875,
"learning_rate": 7.87979655960801e-05,
"loss": 2.4837,
"step": 8410
},
{
"epoch": 0.743258154212826,
"grad_norm": 0.12353515625,
"learning_rate": 7.828889466860551e-05,
"loss": 2.4816,
"step": 8420
},
{
"epoch": 0.7441408836121287,
"grad_norm": 0.1279296875,
"learning_rate": 7.77811680313475e-05,
"loss": 2.4895,
"step": 8430
},
{
"epoch": 0.7450236130114314,
"grad_norm": 0.1298828125,
"learning_rate": 7.727478965919144e-05,
"loss": 2.4837,
"step": 8440
},
{
"epoch": 0.745906342410734,
"grad_norm": 0.12890625,
"learning_rate": 7.67697635164675e-05,
"loss": 2.4751,
"step": 8450
},
{
"epoch": 0.7467890718100366,
"grad_norm": 0.1435546875,
"learning_rate": 7.626609355691922e-05,
"loss": 2.4792,
"step": 8460
},
{
"epoch": 0.7476718012093393,
"grad_norm": 0.12451171875,
"learning_rate": 7.576378372367306e-05,
"loss": 2.4755,
"step": 8470
},
{
"epoch": 0.748554530608642,
"grad_norm": 0.12451171875,
"learning_rate": 7.52628379492075e-05,
"loss": 2.4761,
"step": 8480
},
{
"epoch": 0.7494372600079445,
"grad_norm": 0.11962890625,
"learning_rate": 7.476326015532162e-05,
"loss": 2.4775,
"step": 8490
},
{
"epoch": 0.7503199894072472,
"grad_norm": 0.1259765625,
"learning_rate": 7.426505425310531e-05,
"loss": 2.4919,
"step": 8500
},
{
"epoch": 0.7512027188065499,
"grad_norm": 0.12158203125,
"learning_rate": 7.376822414290804e-05,
"loss": 2.483,
"step": 8510
},
{
"epoch": 0.7520854482058525,
"grad_norm": 0.1259765625,
"learning_rate": 7.327277371430858e-05,
"loss": 2.4805,
"step": 8520
},
{
"epoch": 0.7529681776051551,
"grad_norm": 0.125,
"learning_rate": 7.27787068460842e-05,
"loss": 2.4868,
"step": 8530
},
{
"epoch": 0.7538509070044578,
"grad_norm": 0.125,
"learning_rate": 7.228602740618085e-05,
"loss": 2.4842,
"step": 8540
},
{
"epoch": 0.7547336364037605,
"grad_norm": 0.1279296875,
"learning_rate": 7.179473925168256e-05,
"loss": 2.4829,
"step": 8550
},
{
"epoch": 0.755616365803063,
"grad_norm": 0.130859375,
"learning_rate": 7.130484622878108e-05,
"loss": 2.4933,
"step": 8560
},
{
"epoch": 0.7564990952023657,
"grad_norm": 0.125,
"learning_rate": 7.081635217274617e-05,
"loss": 2.4848,
"step": 8570
},
{
"epoch": 0.7573818246016684,
"grad_norm": 0.125,
"learning_rate": 7.032926090789537e-05,
"loss": 2.4788,
"step": 8580
},
{
"epoch": 0.758264554000971,
"grad_norm": 0.123046875,
"learning_rate": 6.984357624756388e-05,
"loss": 2.4789,
"step": 8590
},
{
"epoch": 0.7591472834002736,
"grad_norm": 0.12109375,
"learning_rate": 6.935930199407501e-05,
"loss": 2.4821,
"step": 8600
},
{
"epoch": 0.7600300127995763,
"grad_norm": 0.12890625,
"learning_rate": 6.887644193871042e-05,
"loss": 2.4781,
"step": 8610
},
{
"epoch": 0.7609127421988789,
"grad_norm": 0.1298828125,
"learning_rate": 6.839499986167999e-05,
"loss": 2.4975,
"step": 8620
},
{
"epoch": 0.7617954715981816,
"grad_norm": 0.12060546875,
"learning_rate": 6.791497953209289e-05,
"loss": 2.4711,
"step": 8630
},
{
"epoch": 0.7626782009974842,
"grad_norm": 0.1181640625,
"learning_rate": 6.743638470792735e-05,
"loss": 2.4693,
"step": 8640
},
{
"epoch": 0.7635609303967869,
"grad_norm": 0.125,
"learning_rate": 6.695921913600212e-05,
"loss": 2.4798,
"step": 8650
},
{
"epoch": 0.7644436597960895,
"grad_norm": 0.1240234375,
"learning_rate": 6.648348655194613e-05,
"loss": 2.4843,
"step": 8660
},
{
"epoch": 0.7653263891953922,
"grad_norm": 0.1171875,
"learning_rate": 6.600919068017006e-05,
"loss": 2.4709,
"step": 8670
},
{
"epoch": 0.7662091185946948,
"grad_norm": 0.12060546875,
"learning_rate": 6.553633523383682e-05,
"loss": 2.4822,
"step": 8680
},
{
"epoch": 0.7670918479939974,
"grad_norm": 0.12158203125,
"learning_rate": 6.506492391483232e-05,
"loss": 2.4729,
"step": 8690
},
{
"epoch": 0.7679745773933001,
"grad_norm": 0.12109375,
"learning_rate": 6.459496041373708e-05,
"loss": 2.4769,
"step": 8700
},
{
"epoch": 0.7688573067926028,
"grad_norm": 0.1201171875,
"learning_rate": 6.412644840979656e-05,
"loss": 2.4866,
"step": 8710
},
{
"epoch": 0.7697400361919053,
"grad_norm": 0.12451171875,
"learning_rate": 6.365939157089304e-05,
"loss": 2.4759,
"step": 8720
},
{
"epoch": 0.770622765591208,
"grad_norm": 0.1298828125,
"learning_rate": 6.319379355351653e-05,
"loss": 2.4632,
"step": 8730
},
{
"epoch": 0.7715054949905107,
"grad_norm": 0.1259765625,
"learning_rate": 6.272965800273608e-05,
"loss": 2.4707,
"step": 8740
},
{
"epoch": 0.7723882243898132,
"grad_norm": 0.123046875,
"learning_rate": 6.226698855217178e-05,
"loss": 2.4834,
"step": 8750
},
{
"epoch": 0.7732709537891159,
"grad_norm": 0.134765625,
"learning_rate": 6.180578882396556e-05,
"loss": 2.485,
"step": 8760
},
{
"epoch": 0.7741536831884186,
"grad_norm": 0.12255859375,
"learning_rate": 6.134606242875324e-05,
"loss": 2.4726,
"step": 8770
},
{
"epoch": 0.7750364125877213,
"grad_norm": 0.12158203125,
"learning_rate": 6.088781296563636e-05,
"loss": 2.4861,
"step": 8780
},
{
"epoch": 0.7759191419870238,
"grad_norm": 0.1328125,
"learning_rate": 6.043104402215388e-05,
"loss": 2.4927,
"step": 8790
},
{
"epoch": 0.7768018713863265,
"grad_norm": 0.1279296875,
"learning_rate": 5.9975759174254075e-05,
"loss": 2.4853,
"step": 8800
},
{
"epoch": 0.7776846007856292,
"grad_norm": 0.1279296875,
"learning_rate": 5.952196198626633e-05,
"loss": 2.4986,
"step": 8810
},
{
"epoch": 0.7785673301849318,
"grad_norm": 0.125,
"learning_rate": 5.906965601087369e-05,
"loss": 2.4872,
"step": 8820
},
{
"epoch": 0.7794500595842344,
"grad_norm": 0.1220703125,
"learning_rate": 5.861884478908483e-05,
"loss": 2.475,
"step": 8830
},
{
"epoch": 0.7803327889835371,
"grad_norm": 0.12109375,
"learning_rate": 5.816953185020607e-05,
"loss": 2.482,
"step": 8840
},
{
"epoch": 0.7812155183828398,
"grad_norm": 0.1240234375,
"learning_rate": 5.7721720711814195e-05,
"loss": 2.4814,
"step": 8850
},
{
"epoch": 0.7820982477821424,
"grad_norm": 0.12109375,
"learning_rate": 5.727541487972876e-05,
"loss": 2.4722,
"step": 8860
},
{
"epoch": 0.782980977181445,
"grad_norm": 0.1259765625,
"learning_rate": 5.68306178479843e-05,
"loss": 2.4738,
"step": 8870
},
{
"epoch": 0.7838637065807477,
"grad_norm": 0.125,
"learning_rate": 5.638733309880353e-05,
"loss": 2.4844,
"step": 8880
},
{
"epoch": 0.7847464359800503,
"grad_norm": 0.12890625,
"learning_rate": 5.5945564102569764e-05,
"loss": 2.4867,
"step": 8890
},
{
"epoch": 0.785629165379353,
"grad_norm": 0.12109375,
"learning_rate": 5.550531431779984e-05,
"loss": 2.4711,
"step": 8900
},
{
"epoch": 0.7865118947786556,
"grad_norm": 0.12890625,
"learning_rate": 5.50665871911169e-05,
"loss": 2.4816,
"step": 8910
},
{
"epoch": 0.7873946241779582,
"grad_norm": 0.1318359375,
"learning_rate": 5.4629386157223434e-05,
"loss": 2.4665,
"step": 8920
},
{
"epoch": 0.7882773535772609,
"grad_norm": 0.13671875,
"learning_rate": 5.4193714638874845e-05,
"loss": 2.4881,
"step": 8930
},
{
"epoch": 0.7891600829765635,
"grad_norm": 0.130859375,
"learning_rate": 5.375957604685186e-05,
"loss": 2.4593,
"step": 8940
},
{
"epoch": 0.7900428123758662,
"grad_norm": 0.1240234375,
"learning_rate": 5.3326973779934506e-05,
"loss": 2.4863,
"step": 8950
},
{
"epoch": 0.7909255417751688,
"grad_norm": 0.123046875,
"learning_rate": 5.289591122487522e-05,
"loss": 2.4839,
"step": 8960
},
{
"epoch": 0.7918082711744715,
"grad_norm": 0.130859375,
"learning_rate": 5.246639175637216e-05,
"loss": 2.4887,
"step": 8970
},
{
"epoch": 0.7926910005737741,
"grad_norm": 0.130859375,
"learning_rate": 5.203841873704329e-05,
"loss": 2.4873,
"step": 8980
},
{
"epoch": 0.7935737299730767,
"grad_norm": 0.11962890625,
"learning_rate": 5.161199551739942e-05,
"loss": 2.4577,
"step": 8990
},
{
"epoch": 0.7944564593723794,
"grad_norm": 0.1201171875,
"learning_rate": 5.1187125435818575e-05,
"loss": 2.4905,
"step": 9000
},
{
"epoch": 0.7953391887716821,
"grad_norm": 0.1279296875,
"learning_rate": 5.0763811818519494e-05,
"loss": 2.4824,
"step": 9010
},
{
"epoch": 0.7962219181709846,
"grad_norm": 0.119140625,
"learning_rate": 5.0342057979535507e-05,
"loss": 2.487,
"step": 9020
},
{
"epoch": 0.7971046475702873,
"grad_norm": 0.123046875,
"learning_rate": 4.99218672206892e-05,
"loss": 2.4845,
"step": 9030
},
{
"epoch": 0.79798737696959,
"grad_norm": 0.1201171875,
"learning_rate": 4.950324283156562e-05,
"loss": 2.4858,
"step": 9040
},
{
"epoch": 0.7988701063688927,
"grad_norm": 0.11962890625,
"learning_rate": 4.908618808948748e-05,
"loss": 2.4718,
"step": 9050
},
{
"epoch": 0.7997528357681952,
"grad_norm": 0.119140625,
"learning_rate": 4.867070625948866e-05,
"loss": 2.4967,
"step": 9060
},
{
"epoch": 0.8006355651674979,
"grad_norm": 0.12451171875,
"learning_rate": 4.825680059428933e-05,
"loss": 2.4717,
"step": 9070
},
{
"epoch": 0.8015182945668006,
"grad_norm": 0.12060546875,
"learning_rate": 4.784447433427016e-05,
"loss": 2.4791,
"step": 9080
},
{
"epoch": 0.8024010239661032,
"grad_norm": 0.123046875,
"learning_rate": 4.7433730707446805e-05,
"loss": 2.483,
"step": 9090
},
{
"epoch": 0.8032837533654058,
"grad_norm": 0.12158203125,
"learning_rate": 4.702457292944498e-05,
"loss": 2.4791,
"step": 9100
},
{
"epoch": 0.8041664827647085,
"grad_norm": 0.125,
"learning_rate": 4.661700420347517e-05,
"loss": 2.4739,
"step": 9110
},
{
"epoch": 0.8050492121640112,
"grad_norm": 0.12890625,
"learning_rate": 4.62110277203073e-05,
"loss": 2.4814,
"step": 9120
},
{
"epoch": 0.8059319415633137,
"grad_norm": 0.123046875,
"learning_rate": 4.5806646658246104e-05,
"loss": 2.4909,
"step": 9130
},
{
"epoch": 0.8068146709626164,
"grad_norm": 0.126953125,
"learning_rate": 4.5403864183106184e-05,
"loss": 2.4878,
"step": 9140
},
{
"epoch": 0.8076974003619191,
"grad_norm": 0.12060546875,
"learning_rate": 4.5002683448186866e-05,
"loss": 2.4955,
"step": 9150
},
{
"epoch": 0.8085801297612217,
"grad_norm": 0.12353515625,
"learning_rate": 4.460310759424802e-05,
"loss": 2.4799,
"step": 9160
},
{
"epoch": 0.8094628591605243,
"grad_norm": 0.126953125,
"learning_rate": 4.420513974948517e-05,
"loss": 2.4743,
"step": 9170
},
{
"epoch": 0.810345588559827,
"grad_norm": 0.1240234375,
"learning_rate": 4.3808783029505166e-05,
"loss": 2.4726,
"step": 9180
},
{
"epoch": 0.8112283179591296,
"grad_norm": 0.126953125,
"learning_rate": 4.341404053730147e-05,
"loss": 2.485,
"step": 9190
},
{
"epoch": 0.8121110473584323,
"grad_norm": 0.123046875,
"learning_rate": 4.3020915363230274e-05,
"loss": 2.4825,
"step": 9200
},
{
"epoch": 0.8129937767577349,
"grad_norm": 0.1240234375,
"learning_rate": 4.262941058498615e-05,
"loss": 2.4725,
"step": 9210
},
{
"epoch": 0.8138765061570375,
"grad_norm": 0.12451171875,
"learning_rate": 4.2239529267577736e-05,
"loss": 2.4801,
"step": 9220
},
{
"epoch": 0.8147592355563402,
"grad_norm": 0.1201171875,
"learning_rate": 4.1851274463304165e-05,
"loss": 2.4842,
"step": 9230
},
{
"epoch": 0.8156419649556429,
"grad_norm": 0.12451171875,
"learning_rate": 4.146464921173088e-05,
"loss": 2.4747,
"step": 9240
},
{
"epoch": 0.8165246943549455,
"grad_norm": 0.11962890625,
"learning_rate": 4.1079656539665696e-05,
"loss": 2.4856,
"step": 9250
},
{
"epoch": 0.8174074237542481,
"grad_norm": 0.1201171875,
"learning_rate": 4.069629946113565e-05,
"loss": 2.4736,
"step": 9260
},
{
"epoch": 0.8182901531535508,
"grad_norm": 0.1220703125,
"learning_rate": 4.0314580977362655e-05,
"loss": 2.4811,
"step": 9270
},
{
"epoch": 0.8191728825528535,
"grad_norm": 0.12255859375,
"learning_rate": 3.99345040767409e-05,
"loss": 2.4788,
"step": 9280
},
{
"epoch": 0.820055611952156,
"grad_norm": 0.1181640625,
"learning_rate": 3.955607173481254e-05,
"loss": 2.4807,
"step": 9290
},
{
"epoch": 0.8209383413514587,
"grad_norm": 0.12353515625,
"learning_rate": 3.9179286914244884e-05,
"loss": 2.4758,
"step": 9300
},
{
"epoch": 0.8218210707507614,
"grad_norm": 0.1240234375,
"learning_rate": 3.880415256480749e-05,
"loss": 2.4897,
"step": 9310
},
{
"epoch": 0.822703800150064,
"grad_norm": 0.11767578125,
"learning_rate": 3.843067162334826e-05,
"loss": 2.4591,
"step": 9320
},
{
"epoch": 0.8235865295493666,
"grad_norm": 0.12255859375,
"learning_rate": 3.805884701377127e-05,
"loss": 2.4757,
"step": 9330
},
{
"epoch": 0.8244692589486693,
"grad_norm": 0.1259765625,
"learning_rate": 3.768868164701325e-05,
"loss": 2.4784,
"step": 9340
},
{
"epoch": 0.825351988347972,
"grad_norm": 0.1259765625,
"learning_rate": 3.732017842102126e-05,
"loss": 2.5042,
"step": 9350
},
{
"epoch": 0.8262347177472745,
"grad_norm": 0.1201171875,
"learning_rate": 3.695334022072977e-05,
"loss": 2.4785,
"step": 9360
},
{
"epoch": 0.8271174471465772,
"grad_norm": 0.12353515625,
"learning_rate": 3.658816991803798e-05,
"loss": 2.4855,
"step": 9370
},
{
"epoch": 0.8280001765458799,
"grad_norm": 0.12109375,
"learning_rate": 3.622467037178765e-05,
"loss": 2.4773,
"step": 9380
},
{
"epoch": 0.8288829059451825,
"grad_norm": 0.12353515625,
"learning_rate": 3.586284442774049e-05,
"loss": 2.4639,
"step": 9390
},
{
"epoch": 0.8297656353444851,
"grad_norm": 0.11767578125,
"learning_rate": 3.550269491855579e-05,
"loss": 2.4763,
"step": 9400
},
{
"epoch": 0.8306483647437878,
"grad_norm": 0.11962890625,
"learning_rate": 3.514422466376857e-05,
"loss": 2.4841,
"step": 9410
},
{
"epoch": 0.8315310941430905,
"grad_norm": 0.1201171875,
"learning_rate": 3.478743646976726e-05,
"loss": 2.4838,
"step": 9420
},
{
"epoch": 0.8324138235423931,
"grad_norm": 0.1220703125,
"learning_rate": 3.443233312977176e-05,
"loss": 2.4814,
"step": 9430
},
{
"epoch": 0.8332965529416957,
"grad_norm": 0.11767578125,
"learning_rate": 3.4078917423811556e-05,
"loss": 2.4671,
"step": 9440
},
{
"epoch": 0.8341792823409984,
"grad_norm": 0.1259765625,
"learning_rate": 3.372719211870412e-05,
"loss": 2.4649,
"step": 9450
},
{
"epoch": 0.835062011740301,
"grad_norm": 0.12109375,
"learning_rate": 3.3377159968033085e-05,
"loss": 2.4915,
"step": 9460
},
{
"epoch": 0.8359447411396037,
"grad_norm": 0.12890625,
"learning_rate": 3.302882371212665e-05,
"loss": 2.4789,
"step": 9470
},
{
"epoch": 0.8368274705389063,
"grad_norm": 0.123046875,
"learning_rate": 3.2682186078036304e-05,
"loss": 2.4878,
"step": 9480
},
{
"epoch": 0.8377101999382089,
"grad_norm": 0.11767578125,
"learning_rate": 3.2337249779515436e-05,
"loss": 2.4843,
"step": 9490
},
{
"epoch": 0.8385929293375116,
"grad_norm": 0.12158203125,
"learning_rate": 3.199401751699782e-05,
"loss": 2.4744,
"step": 9500
},
{
"epoch": 0.8394756587368142,
"grad_norm": 0.1279296875,
"learning_rate": 3.1652491977576883e-05,
"loss": 2.481,
"step": 9510
},
{
"epoch": 0.8403583881361169,
"grad_norm": 0.12353515625,
"learning_rate": 3.131267583498448e-05,
"loss": 2.4857,
"step": 9520
},
{
"epoch": 0.8412411175354195,
"grad_norm": 0.1201171875,
"learning_rate": 3.097457174956977e-05,
"loss": 2.489,
"step": 9530
},
{
"epoch": 0.8421238469347222,
"grad_norm": 0.123046875,
"learning_rate": 3.063818236827884e-05,
"loss": 2.4832,
"step": 9540
},
{
"epoch": 0.8430065763340248,
"grad_norm": 0.12060546875,
"learning_rate": 3.030351032463341e-05,
"loss": 2.491,
"step": 9550
},
{
"epoch": 0.8438893057333274,
"grad_norm": 0.1201171875,
"learning_rate": 2.9970558238710865e-05,
"loss": 2.4858,
"step": 9560
},
{
"epoch": 0.8447720351326301,
"grad_norm": 0.11865234375,
"learning_rate": 2.9639328717123104e-05,
"loss": 2.4705,
"step": 9570
},
{
"epoch": 0.8456547645319328,
"grad_norm": 0.1201171875,
"learning_rate": 2.9309824352996618e-05,
"loss": 2.4779,
"step": 9580
},
{
"epoch": 0.8465374939312353,
"grad_norm": 0.1220703125,
"learning_rate": 2.898204772595195e-05,
"loss": 2.4793,
"step": 9590
},
{
"epoch": 0.847420223330538,
"grad_norm": 0.12255859375,
"learning_rate": 2.865600140208349e-05,
"loss": 2.4618,
"step": 9600
},
{
"epoch": 0.8483029527298407,
"grad_norm": 0.126953125,
"learning_rate": 2.833168793393956e-05,
"loss": 2.4848,
"step": 9610
},
{
"epoch": 0.8491856821291434,
"grad_norm": 0.12158203125,
"learning_rate": 2.8009109860502174e-05,
"loss": 2.4771,
"step": 9620
},
{
"epoch": 0.8500684115284459,
"grad_norm": 0.1201171875,
"learning_rate": 2.768826970716745e-05,
"loss": 2.4828,
"step": 9630
},
{
"epoch": 0.8509511409277486,
"grad_norm": 0.126953125,
"learning_rate": 2.736916998572567e-05,
"loss": 2.487,
"step": 9640
},
{
"epoch": 0.8518338703270513,
"grad_norm": 0.11669921875,
"learning_rate": 2.705181319434144e-05,
"loss": 2.4868,
"step": 9650
},
{
"epoch": 0.8527165997263539,
"grad_norm": 0.11962890625,
"learning_rate": 2.6736201817534696e-05,
"loss": 2.4807,
"step": 9660
},
{
"epoch": 0.8535993291256565,
"grad_norm": 0.12158203125,
"learning_rate": 2.6422338326160618e-05,
"loss": 2.4829,
"step": 9670
},
{
"epoch": 0.8544820585249592,
"grad_norm": 0.11767578125,
"learning_rate": 2.6110225177390534e-05,
"loss": 2.4849,
"step": 9680
},
{
"epoch": 0.8553647879242618,
"grad_norm": 0.1279296875,
"learning_rate": 2.5799864814692902e-05,
"loss": 2.4787,
"step": 9690
},
{
"epoch": 0.8562475173235644,
"grad_norm": 0.12060546875,
"learning_rate": 2.549125966781385e-05,
"loss": 2.4753,
"step": 9700
},
{
"epoch": 0.8571302467228671,
"grad_norm": 0.12158203125,
"learning_rate": 2.518441215275838e-05,
"loss": 2.4752,
"step": 9710
},
{
"epoch": 0.8580129761221698,
"grad_norm": 0.12158203125,
"learning_rate": 2.48793246717712e-05,
"loss": 2.4789,
"step": 9720
},
{
"epoch": 0.8588957055214724,
"grad_norm": 0.12255859375,
"learning_rate": 2.4575999613318245e-05,
"loss": 2.4878,
"step": 9730
},
{
"epoch": 0.859778434920775,
"grad_norm": 0.125,
"learning_rate": 2.4274439352067828e-05,
"loss": 2.4793,
"step": 9740
},
{
"epoch": 0.8606611643200777,
"grad_norm": 0.1220703125,
"learning_rate": 2.3974646248871827e-05,
"loss": 2.4803,
"step": 9750
},
{
"epoch": 0.8615438937193803,
"grad_norm": 0.11767578125,
"learning_rate": 2.3676622650747603e-05,
"loss": 2.4748,
"step": 9760
},
{
"epoch": 0.862426623118683,
"grad_norm": 0.12158203125,
"learning_rate": 2.3380370890859454e-05,
"loss": 2.4787,
"step": 9770
},
{
"epoch": 0.8633093525179856,
"grad_norm": 0.12158203125,
"learning_rate": 2.3085893288500136e-05,
"loss": 2.4785,
"step": 9780
},
{
"epoch": 0.8641920819172882,
"grad_norm": 0.126953125,
"learning_rate": 2.279319214907305e-05,
"loss": 2.4603,
"step": 9790
},
{
"epoch": 0.8650748113165909,
"grad_norm": 0.1171875,
"learning_rate": 2.2502269764074017e-05,
"loss": 2.4593,
"step": 9800
},
{
"epoch": 0.8659575407158936,
"grad_norm": 0.12255859375,
"learning_rate": 2.2213128411073396e-05,
"loss": 2.4895,
"step": 9810
},
{
"epoch": 0.8668402701151962,
"grad_norm": 0.11962890625,
"learning_rate": 2.1925770353698137e-05,
"loss": 2.4858,
"step": 9820
},
{
"epoch": 0.8677229995144988,
"grad_norm": 0.12060546875,
"learning_rate": 2.1640197841614083e-05,
"loss": 2.4799,
"step": 9830
},
{
"epoch": 0.8686057289138015,
"grad_norm": 0.1181640625,
"learning_rate": 2.1356413110508675e-05,
"loss": 2.4731,
"step": 9840
},
{
"epoch": 0.8694884583131042,
"grad_norm": 0.11962890625,
"learning_rate": 2.1074418382072912e-05,
"loss": 2.4795,
"step": 9850
},
{
"epoch": 0.8703711877124067,
"grad_norm": 0.12158203125,
"learning_rate": 2.0794215863984417e-05,
"loss": 2.4697,
"step": 9860
},
{
"epoch": 0.8712539171117094,
"grad_norm": 0.125,
"learning_rate": 2.0515807749889954e-05,
"loss": 2.4764,
"step": 9870
},
{
"epoch": 0.8721366465110121,
"grad_norm": 0.12255859375,
"learning_rate": 2.0239196219388133e-05,
"loss": 2.4901,
"step": 9880
},
{
"epoch": 0.8730193759103146,
"grad_norm": 0.1201171875,
"learning_rate": 1.9964383438012685e-05,
"loss": 2.4936,
"step": 9890
},
{
"epoch": 0.8739021053096173,
"grad_norm": 0.1181640625,
"learning_rate": 1.969137155721509e-05,
"loss": 2.4794,
"step": 9900
},
{
"epoch": 0.87478483470892,
"grad_norm": 0.119140625,
"learning_rate": 1.942016271434821e-05,
"loss": 2.4851,
"step": 9910
},
{
"epoch": 0.8756675641082227,
"grad_norm": 0.1220703125,
"learning_rate": 1.915075903264915e-05,
"loss": 2.4784,
"step": 9920
},
{
"epoch": 0.8765502935075252,
"grad_norm": 0.119140625,
"learning_rate": 1.8883162621222693e-05,
"loss": 2.4947,
"step": 9930
},
{
"epoch": 0.8774330229068279,
"grad_norm": 0.12060546875,
"learning_rate": 1.8617375575025186e-05,
"loss": 2.4925,
"step": 9940
},
{
"epoch": 0.8783157523061306,
"grad_norm": 0.1201171875,
"learning_rate": 1.835339997484753e-05,
"loss": 2.4929,
"step": 9950
},
{
"epoch": 0.8791984817054332,
"grad_norm": 0.12451171875,
"learning_rate": 1.8091237887299357e-05,
"loss": 2.4804,
"step": 9960
},
{
"epoch": 0.8800812111047358,
"grad_norm": 0.11865234375,
"learning_rate": 1.783089136479257e-05,
"loss": 2.4877,
"step": 9970
},
{
"epoch": 0.8809639405040385,
"grad_norm": 0.126953125,
"learning_rate": 1.757236244552557e-05,
"loss": 2.4866,
"step": 9980
},
{
"epoch": 0.8818466699033412,
"grad_norm": 0.119140625,
"learning_rate": 1.7315653153466977e-05,
"loss": 2.4786,
"step": 9990
},
{
"epoch": 0.8827293993026438,
"grad_norm": 0.12060546875,
"learning_rate": 1.7060765498339958e-05,
"loss": 2.4865,
"step": 10000
},
{
"epoch": 0.8827293993026438,
"eval_accuracy": 0.5115012171615945,
"eval_loss": 2.3658125400543213,
"eval_runtime": 6.9822,
"eval_samples_per_second": 45.544,
"eval_steps_per_second": 0.43,
"step": 10000
},
{
"epoch": 0.8836121287019464,
"grad_norm": 0.11767578125,
"learning_rate": 1.6807701475606534e-05,
"loss": 2.4905,
"step": 10010
},
{
"epoch": 0.8844948581012491,
"grad_norm": 0.1201171875,
"learning_rate": 1.6556463066451837e-05,
"loss": 2.4761,
"step": 10020
},
{
"epoch": 0.8853775875005517,
"grad_norm": 0.123046875,
"learning_rate": 1.63070522377686e-05,
"loss": 2.4894,
"step": 10030
},
{
"epoch": 0.8862603168998543,
"grad_norm": 0.119140625,
"learning_rate": 1.6059470942141912e-05,
"loss": 2.4742,
"step": 10040
},
{
"epoch": 0.887143046299157,
"grad_norm": 0.1201171875,
"learning_rate": 1.5813721117833828e-05,
"loss": 2.4906,
"step": 10050
},
{
"epoch": 0.8880257756984596,
"grad_norm": 0.11767578125,
"learning_rate": 1.5569804688768092e-05,
"loss": 2.4644,
"step": 10060
},
{
"epoch": 0.8889085050977623,
"grad_norm": 0.1298828125,
"learning_rate": 1.532772356451531e-05,
"loss": 2.4756,
"step": 10070
},
{
"epoch": 0.8897912344970649,
"grad_norm": 0.12255859375,
"learning_rate": 1.5087479640277763e-05,
"loss": 2.4801,
"step": 10080
},
{
"epoch": 0.8906739638963675,
"grad_norm": 0.11767578125,
"learning_rate": 1.4849074796874779e-05,
"loss": 2.493,
"step": 10090
},
{
"epoch": 0.8915566932956702,
"grad_norm": 0.11962890625,
"learning_rate": 1.4612510900727794e-05,
"loss": 2.478,
"step": 10100
},
{
"epoch": 0.8924394226949729,
"grad_norm": 0.12255859375,
"learning_rate": 1.4377789803845964e-05,
"loss": 2.4824,
"step": 10110
},
{
"epoch": 0.8933221520942755,
"grad_norm": 0.11767578125,
"learning_rate": 1.4144913343811544e-05,
"loss": 2.4745,
"step": 10120
},
{
"epoch": 0.8942048814935781,
"grad_norm": 0.1171875,
"learning_rate": 1.3913883343765394e-05,
"loss": 2.4786,
"step": 10130
},
{
"epoch": 0.8950876108928808,
"grad_norm": 0.11962890625,
"learning_rate": 1.3684701612392963e-05,
"loss": 2.4768,
"step": 10140
},
{
"epoch": 0.8959703402921835,
"grad_norm": 0.12060546875,
"learning_rate": 1.345736994390992e-05,
"loss": 2.4689,
"step": 10150
},
{
"epoch": 0.896853069691486,
"grad_norm": 0.11962890625,
"learning_rate": 1.3231890118048179e-05,
"loss": 2.4817,
"step": 10160
},
{
"epoch": 0.8977357990907887,
"grad_norm": 0.11572265625,
"learning_rate": 1.300826390004209e-05,
"loss": 2.49,
"step": 10170
},
{
"epoch": 0.8986185284900914,
"grad_norm": 0.12060546875,
"learning_rate": 1.2786493040614245e-05,
"loss": 2.4965,
"step": 10180
},
{
"epoch": 0.8995012578893941,
"grad_norm": 0.11962890625,
"learning_rate": 1.2566579275962303e-05,
"loss": 2.4725,
"step": 10190
},
{
"epoch": 0.9003839872886966,
"grad_norm": 0.12060546875,
"learning_rate": 1.2348524327744943e-05,
"loss": 2.4706,
"step": 10200
},
{
"epoch": 0.9012667166879993,
"grad_norm": 0.11767578125,
"learning_rate": 1.2132329903068563e-05,
"loss": 2.4772,
"step": 10210
},
{
"epoch": 0.902149446087302,
"grad_norm": 0.12060546875,
"learning_rate": 1.1917997694473992e-05,
"loss": 2.4823,
"step": 10220
},
{
"epoch": 0.9030321754866045,
"grad_norm": 0.123046875,
"learning_rate": 1.1705529379923085e-05,
"loss": 2.4674,
"step": 10230
},
{
"epoch": 0.9039149048859072,
"grad_norm": 0.1220703125,
"learning_rate": 1.1494926622785811e-05,
"loss": 2.4772,
"step": 10240
},
{
"epoch": 0.9047976342852099,
"grad_norm": 0.1201171875,
"learning_rate": 1.1286191071826823e-05,
"loss": 2.472,
"step": 10250
},
{
"epoch": 0.9056803636845125,
"grad_norm": 0.1240234375,
"learning_rate": 1.1079324361193022e-05,
"loss": 2.5008,
"step": 10260
},
{
"epoch": 0.9065630930838151,
"grad_norm": 0.119140625,
"learning_rate": 1.0874328110400511e-05,
"loss": 2.4843,
"step": 10270
},
{
"epoch": 0.9074458224831178,
"grad_norm": 0.11962890625,
"learning_rate": 1.0671203924321887e-05,
"loss": 2.4851,
"step": 10280
},
{
"epoch": 0.9083285518824205,
"grad_norm": 0.1171875,
"learning_rate": 1.0469953393173776e-05,
"loss": 2.4739,
"step": 10290
},
{
"epoch": 0.9092112812817231,
"grad_norm": 0.12109375,
"learning_rate": 1.0270578092504396e-05,
"loss": 2.4763,
"step": 10300
},
{
"epoch": 0.9100940106810257,
"grad_norm": 0.12060546875,
"learning_rate": 1.0073079583181126e-05,
"loss": 2.4805,
"step": 10310
},
{
"epoch": 0.9109767400803284,
"grad_norm": 0.12060546875,
"learning_rate": 9.877459411378325e-06,
"loss": 2.4858,
"step": 10320
},
{
"epoch": 0.911859469479631,
"grad_norm": 0.1279296875,
"learning_rate": 9.683719108565331e-06,
"loss": 2.4798,
"step": 10330
},
{
"epoch": 0.9127421988789337,
"grad_norm": 0.11767578125,
"learning_rate": 9.49186019149434e-06,
"loss": 2.4887,
"step": 10340
},
{
"epoch": 0.9136249282782363,
"grad_norm": 0.119140625,
"learning_rate": 9.301884162188496e-06,
"loss": 2.4803,
"step": 10350
},
{
"epoch": 0.9145076576775389,
"grad_norm": 0.12158203125,
"learning_rate": 9.113792507930263e-06,
"loss": 2.4804,
"step": 10360
},
{
"epoch": 0.9153903870768416,
"grad_norm": 0.12353515625,
"learning_rate": 8.927586701249852e-06,
"loss": 2.4766,
"step": 10370
},
{
"epoch": 0.9162731164761443,
"grad_norm": 0.1220703125,
"learning_rate": 8.743268199913307e-06,
"loss": 2.4674,
"step": 10380
},
{
"epoch": 0.9171558458754469,
"grad_norm": 0.11962890625,
"learning_rate": 8.560838446911607e-06,
"loss": 2.473,
"step": 10390
},
{
"epoch": 0.9180385752747495,
"grad_norm": 0.1201171875,
"learning_rate": 8.380298870449e-06,
"loss": 2.4637,
"step": 10400
},
{
"epoch": 0.9189213046740522,
"grad_norm": 0.1220703125,
"learning_rate": 8.201650883931904e-06,
"loss": 2.4808,
"step": 10410
},
{
"epoch": 0.9198040340733548,
"grad_norm": 0.1201171875,
"learning_rate": 8.024895885957978e-06,
"loss": 2.4655,
"step": 10420
},
{
"epoch": 0.9206867634726574,
"grad_norm": 0.12158203125,
"learning_rate": 7.85003526030495e-06,
"loss": 2.4755,
"step": 10430
},
{
"epoch": 0.9215694928719601,
"grad_norm": 0.119140625,
"learning_rate": 7.677070375920026e-06,
"loss": 2.4738,
"step": 10440
},
{
"epoch": 0.9224522222712628,
"grad_norm": 0.1181640625,
"learning_rate": 7.506002586909006e-06,
"loss": 2.4909,
"step": 10450
},
{
"epoch": 0.9233349516705653,
"grad_norm": 0.1318359375,
"learning_rate": 7.336833232525625e-06,
"loss": 2.4758,
"step": 10460
},
{
"epoch": 0.924217681069868,
"grad_norm": 0.12060546875,
"learning_rate": 7.169563637161397e-06,
"loss": 2.4819,
"step": 10470
},
{
"epoch": 0.9251004104691707,
"grad_norm": 0.11962890625,
"learning_rate": 7.004195110334788e-06,
"loss": 2.4731,
"step": 10480
},
{
"epoch": 0.9259831398684734,
"grad_norm": 0.1201171875,
"learning_rate": 6.840728946681363e-06,
"loss": 2.4937,
"step": 10490
},
{
"epoch": 0.9268658692677759,
"grad_norm": 0.11962890625,
"learning_rate": 6.679166425943351e-06,
"loss": 2.473,
"step": 10500
},
{
"epoch": 0.9277485986670786,
"grad_norm": 0.11962890625,
"learning_rate": 6.519508812959873e-06,
"loss": 2.4801,
"step": 10510
},
{
"epoch": 0.9286313280663813,
"grad_norm": 0.12109375,
"learning_rate": 6.3617573576569274e-06,
"loss": 2.479,
"step": 10520
},
{
"epoch": 0.9295140574656839,
"grad_norm": 0.12109375,
"learning_rate": 6.205913295037474e-06,
"loss": 2.4722,
"step": 10530
},
{
"epoch": 0.9303967868649865,
"grad_norm": 0.11865234375,
"learning_rate": 6.051977845172002e-06,
"loss": 2.4912,
"step": 10540
},
{
"epoch": 0.9312795162642892,
"grad_norm": 0.130859375,
"learning_rate": 5.899952213188897e-06,
"loss": 2.4677,
"step": 10550
},
{
"epoch": 0.9321622456635918,
"grad_norm": 0.11962890625,
"learning_rate": 5.749837589264895e-06,
"loss": 2.4811,
"step": 10560
},
{
"epoch": 0.9330449750628945,
"grad_norm": 0.12109375,
"learning_rate": 5.601635148615891e-06,
"loss": 2.4728,
"step": 10570
},
{
"epoch": 0.9339277044621971,
"grad_norm": 0.11962890625,
"learning_rate": 5.4553460514877304e-06,
"loss": 2.49,
"step": 10580
},
{
"epoch": 0.9348104338614998,
"grad_norm": 0.1201171875,
"learning_rate": 5.3109714431470165e-06,
"loss": 2.4939,
"step": 10590
},
{
"epoch": 0.9356931632608024,
"grad_norm": 0.11572265625,
"learning_rate": 5.168512453872287e-06,
"loss": 2.4792,
"step": 10600
},
{
"epoch": 0.936575892660105,
"grad_norm": 0.12109375,
"learning_rate": 5.027970198945076e-06,
"loss": 2.4792,
"step": 10610
},
{
"epoch": 0.9374586220594077,
"grad_norm": 0.119140625,
"learning_rate": 4.889345778641252e-06,
"loss": 2.4766,
"step": 10620
},
{
"epoch": 0.9383413514587103,
"grad_norm": 0.12109375,
"learning_rate": 4.752640278222254e-06,
"loss": 2.4851,
"step": 10630
},
{
"epoch": 0.939224080858013,
"grad_norm": 0.11865234375,
"learning_rate": 4.617854767926782e-06,
"loss": 2.4703,
"step": 10640
},
{
"epoch": 0.9401068102573156,
"grad_norm": 0.12158203125,
"learning_rate": 4.484990302962344e-06,
"loss": 2.4966,
"step": 10650
},
{
"epoch": 0.9409895396566182,
"grad_norm": 0.12060546875,
"learning_rate": 4.354047923496917e-06,
"loss": 2.4759,
"step": 10660
},
{
"epoch": 0.9418722690559209,
"grad_norm": 0.11669921875,
"learning_rate": 4.2250286546509365e-06,
"loss": 2.4697,
"step": 10670
},
{
"epoch": 0.9427549984552236,
"grad_norm": 0.1181640625,
"learning_rate": 4.09793350648921e-06,
"loss": 2.4771,
"step": 10680
},
{
"epoch": 0.9436377278545262,
"grad_norm": 0.11767578125,
"learning_rate": 3.9727634740129585e-06,
"loss": 2.486,
"step": 10690
},
{
"epoch": 0.9445204572538288,
"grad_norm": 0.11962890625,
"learning_rate": 3.849519537152124e-06,
"loss": 2.4853,
"step": 10700
},
{
"epoch": 0.9454031866531315,
"grad_norm": 0.11767578125,
"learning_rate": 3.7282026607576016e-06,
"loss": 2.48,
"step": 10710
},
{
"epoch": 0.9462859160524342,
"grad_norm": 0.12109375,
"learning_rate": 3.608813794593796e-06,
"loss": 2.4858,
"step": 10720
},
{
"epoch": 0.9471686454517367,
"grad_norm": 0.12158203125,
"learning_rate": 3.491353873331077e-06,
"loss": 2.4776,
"step": 10730
},
{
"epoch": 0.9480513748510394,
"grad_norm": 0.1201171875,
"learning_rate": 3.3758238165384757e-06,
"loss": 2.474,
"step": 10740
},
{
"epoch": 0.9489341042503421,
"grad_norm": 0.1201171875,
"learning_rate": 3.262224528676666e-06,
"loss": 2.4638,
"step": 10750
},
{
"epoch": 0.9498168336496448,
"grad_norm": 0.12060546875,
"learning_rate": 3.1505568990905787e-06,
"loss": 2.4872,
"step": 10760
},
{
"epoch": 0.9506995630489473,
"grad_norm": 0.1171875,
"learning_rate": 3.040821802002658e-06,
"loss": 2.4675,
"step": 10770
},
{
"epoch": 0.95158229244825,
"grad_norm": 0.11669921875,
"learning_rate": 2.9330200965059507e-06,
"loss": 2.4689,
"step": 10780
},
{
"epoch": 0.9524650218475527,
"grad_norm": 0.11962890625,
"learning_rate": 2.827152626557389e-06,
"loss": 2.4866,
"step": 10790
},
{
"epoch": 0.9533477512468552,
"grad_norm": 0.11865234375,
"learning_rate": 2.72322022097124e-06,
"loss": 2.4686,
"step": 10800
},
{
"epoch": 0.9542304806461579,
"grad_norm": 0.11962890625,
"learning_rate": 2.621223693412417e-06,
"loss": 2.482,
"step": 10810
},
{
"epoch": 0.9551132100454606,
"grad_norm": 0.12109375,
"learning_rate": 2.5211638423903725e-06,
"loss": 2.4855,
"step": 10820
},
{
"epoch": 0.9559959394447632,
"grad_norm": 0.12158203125,
"learning_rate": 2.4230414512527166e-06,
"loss": 2.4815,
"step": 10830
},
{
"epoch": 0.9568786688440658,
"grad_norm": 0.11767578125,
"learning_rate": 2.326857288178996e-06,
"loss": 2.477,
"step": 10840
},
{
"epoch": 0.9577613982433685,
"grad_norm": 0.125,
"learning_rate": 2.232612106174897e-06,
"loss": 2.4788,
"step": 10850
},
{
"epoch": 0.9586441276426712,
"grad_norm": 0.12353515625,
"learning_rate": 2.1403066430661644e-06,
"loss": 2.4843,
"step": 10860
},
{
"epoch": 0.9595268570419738,
"grad_norm": 0.1220703125,
"learning_rate": 2.0499416214928844e-06,
"loss": 2.4884,
"step": 10870
},
{
"epoch": 0.9604095864412764,
"grad_norm": 0.1201171875,
"learning_rate": 1.9615177489038792e-06,
"loss": 2.4683,
"step": 10880
},
{
"epoch": 0.9612923158405791,
"grad_norm": 0.12109375,
"learning_rate": 1.8750357175510435e-06,
"loss": 2.478,
"step": 10890
},
{
"epoch": 0.9621750452398817,
"grad_norm": 0.11962890625,
"learning_rate": 1.7904962044841266e-06,
"loss": 2.493,
"step": 10900
},
{
"epoch": 0.9630577746391844,
"grad_norm": 0.11865234375,
"learning_rate": 1.70789987154521e-06,
"loss": 2.4737,
"step": 10910
},
{
"epoch": 0.963940504038487,
"grad_norm": 0.1181640625,
"learning_rate": 1.6272473653636266e-06,
"loss": 2.4757,
"step": 10920
},
{
"epoch": 0.9648232334377896,
"grad_norm": 0.119140625,
"learning_rate": 1.5485393173509388e-06,
"loss": 2.4702,
"step": 10930
},
{
"epoch": 0.9657059628370923,
"grad_norm": 0.12060546875,
"learning_rate": 1.4717763436959685e-06,
"loss": 2.4828,
"step": 10940
},
{
"epoch": 0.966588692236395,
"grad_norm": 0.1171875,
"learning_rate": 1.3969590453598858e-06,
"loss": 2.467,
"step": 10950
},
{
"epoch": 0.9674714216356975,
"grad_norm": 0.12255859375,
"learning_rate": 1.3240880080716832e-06,
"loss": 2.4732,
"step": 10960
},
{
"epoch": 0.9683541510350002,
"grad_norm": 0.119140625,
"learning_rate": 1.2531638023233761e-06,
"loss": 2.4722,
"step": 10970
},
{
"epoch": 0.9692368804343029,
"grad_norm": 0.119140625,
"learning_rate": 1.1841869833656981e-06,
"loss": 2.5025,
"step": 10980
},
{
"epoch": 0.9701196098336055,
"grad_norm": 0.119140625,
"learning_rate": 1.1171580912036627e-06,
"loss": 2.463,
"step": 10990
},
{
"epoch": 0.9710023392329081,
"grad_norm": 0.119140625,
"learning_rate": 1.0520776505924812e-06,
"loss": 2.4808,
"step": 11000
},
{
"epoch": 0.9718850686322108,
"grad_norm": 0.12451171875,
"learning_rate": 9.889461710332059e-07,
"loss": 2.4864,
"step": 11010
},
{
"epoch": 0.9727677980315135,
"grad_norm": 0.119140625,
"learning_rate": 9.277641467689279e-07,
"loss": 2.477,
"step": 11020
},
{
"epoch": 0.973650527430816,
"grad_norm": 0.12255859375,
"learning_rate": 8.685320567809741e-07,
"loss": 2.4772,
"step": 11030
},
{
"epoch": 0.9745332568301187,
"grad_norm": 0.12158203125,
"learning_rate": 8.112503647848546e-07,
"loss": 2.4618,
"step": 11040
},
{
"epoch": 0.9754159862294214,
"grad_norm": 0.1181640625,
"learning_rate": 7.559195192269608e-07,
"loss": 2.4775,
"step": 11050
},
{
"epoch": 0.9762987156287241,
"grad_norm": 0.12158203125,
"learning_rate": 7.025399532808452e-07,
"loss": 2.4819,
"step": 11060
},
{
"epoch": 0.9771814450280266,
"grad_norm": 0.12158203125,
"learning_rate": 6.511120848439467e-07,
"loss": 2.4891,
"step": 11070
},
{
"epoch": 0.9780641744273293,
"grad_norm": 0.12109375,
"learning_rate": 6.016363165342875e-07,
"loss": 2.4727,
"step": 11080
},
{
"epoch": 0.978946903826632,
"grad_norm": 0.1201171875,
"learning_rate": 5.54113035687226e-07,
"loss": 2.4755,
"step": 11090
},
{
"epoch": 0.9798296332259346,
"grad_norm": 0.12060546875,
"learning_rate": 5.085426143525695e-07,
"loss": 2.4663,
"step": 11100
},
{
"epoch": 0.9807123626252372,
"grad_norm": 0.1181640625,
"learning_rate": 4.649254092916333e-07,
"loss": 2.4814,
"step": 11110
},
{
"epoch": 0.9815950920245399,
"grad_norm": 0.1279296875,
"learning_rate": 4.2326176197429735e-07,
"loss": 2.4864,
"step": 11120
},
{
"epoch": 0.9824778214238425,
"grad_norm": 0.1220703125,
"learning_rate": 3.835519985765368e-07,
"loss": 2.4657,
"step": 11130
},
{
"epoch": 0.9833605508231452,
"grad_norm": 0.126953125,
"learning_rate": 3.457964299777849e-07,
"loss": 2.4785,
"step": 11140
},
{
"epoch": 0.9842432802224478,
"grad_norm": 0.126953125,
"learning_rate": 3.099953517584353e-07,
"loss": 2.4743,
"step": 11150
},
{
"epoch": 0.9851260096217505,
"grad_norm": 0.123046875,
"learning_rate": 2.761490441976211e-07,
"loss": 2.4793,
"step": 11160
},
{
"epoch": 0.9860087390210531,
"grad_norm": 0.1201171875,
"learning_rate": 2.4425777227102265e-07,
"loss": 2.4827,
"step": 11170
},
{
"epoch": 0.9868914684203557,
"grad_norm": 0.12060546875,
"learning_rate": 2.1432178564867455e-07,
"loss": 2.4843,
"step": 11180
},
{
"epoch": 0.9877741978196584,
"grad_norm": 0.12060546875,
"learning_rate": 1.8634131869313397e-07,
"loss": 2.4731,
"step": 11190
},
{
"epoch": 0.988656927218961,
"grad_norm": 0.12158203125,
"learning_rate": 1.6031659045759318e-07,
"loss": 2.4696,
"step": 11200
},
{
"epoch": 0.9895396566182637,
"grad_norm": 0.11865234375,
"learning_rate": 1.3624780468424192e-07,
"loss": 2.4817,
"step": 11210
},
{
"epoch": 0.9904223860175663,
"grad_norm": 0.12158203125,
"learning_rate": 1.1413514980254669e-07,
"loss": 2.4791,
"step": 11220
},
{
"epoch": 0.9913051154168689,
"grad_norm": 0.12060546875,
"learning_rate": 9.397879892777961e-08,
"loss": 2.4808,
"step": 11230
},
{
"epoch": 0.9921878448161716,
"grad_norm": 0.1220703125,
"learning_rate": 7.577890985985269e-08,
"loss": 2.4776,
"step": 11240
},
{
"epoch": 0.9930705742154743,
"grad_norm": 0.12158203125,
"learning_rate": 5.953562508184684e-08,
"loss": 2.4801,
"step": 11250
},
{
"epoch": 0.9939533036147769,
"grad_norm": 0.11865234375,
"learning_rate": 4.524907175904036e-08,
"loss": 2.4766,
"step": 11260
},
{
"epoch": 0.9948360330140795,
"grad_norm": 0.11962890625,
"learning_rate": 3.2919361737854256e-08,
"loss": 2.4889,
"step": 11270
},
{
"epoch": 0.9957187624133822,
"grad_norm": 0.11962890625,
"learning_rate": 2.2546591544991833e-08,
"loss": 2.4668,
"step": 11280
},
{
"epoch": 0.9966014918126849,
"grad_norm": 0.119140625,
"learning_rate": 1.4130842386717025e-08,
"loss": 2.4816,
"step": 11290
},
{
"epoch": 0.9974842212119874,
"grad_norm": 0.12109375,
"learning_rate": 7.672180148132757e-09,
"loss": 2.4718,
"step": 11300
},
{
"epoch": 0.9983669506112901,
"grad_norm": 0.12353515625,
"learning_rate": 3.1706553927923763e-09,
"loss": 2.4761,
"step": 11310
},
{
"epoch": 0.9992496800105928,
"grad_norm": 0.11767578125,
"learning_rate": 6.263033621722869e-10,
"loss": 2.4773,
"step": 11320
},
{
"epoch": 0.9999558635300348,
"step": 11328,
"total_flos": 2.202782600225295e+20,
"train_loss": 2.520355888510828,
"train_runtime": 24592.232,
"train_samples_per_second": 117.927,
"train_steps_per_second": 0.461
}
],
"logging_steps": 10,
"max_steps": 11328,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.202782600225295e+20,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}