zephyr-2b-gemma-sft-qlora / trainer_state.json
Ritvik19's picture
Upload 12 files
bd70703 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9997655182451062,
"eval_steps": 500,
"global_step": 966,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0010349539526346855,
"grad_norm": 1.046875,
"learning_rate": 2.061855670103093e-06,
"loss": 1.7433,
"step": 1
},
{
"epoch": 0.005174769763173427,
"grad_norm": 0.59765625,
"learning_rate": 1.0309278350515464e-05,
"loss": 1.7444,
"step": 5
},
{
"epoch": 0.010349539526346855,
"grad_norm": 1.0,
"learning_rate": 2.0618556701030927e-05,
"loss": 1.7363,
"step": 10
},
{
"epoch": 0.015524309289520283,
"grad_norm": 0.6484375,
"learning_rate": 3.0927835051546395e-05,
"loss": 1.7131,
"step": 15
},
{
"epoch": 0.02069907905269371,
"grad_norm": 0.89453125,
"learning_rate": 4.1237113402061855e-05,
"loss": 1.7097,
"step": 20
},
{
"epoch": 0.02587384881586714,
"grad_norm": 3.3125,
"learning_rate": 5.1546391752577315e-05,
"loss": 1.6876,
"step": 25
},
{
"epoch": 0.031048618579040566,
"grad_norm": 0.72265625,
"learning_rate": 6.185567010309279e-05,
"loss": 1.552,
"step": 30
},
{
"epoch": 0.036223388342213995,
"grad_norm": 0.353515625,
"learning_rate": 7.216494845360825e-05,
"loss": 1.5049,
"step": 35
},
{
"epoch": 0.04139815810538742,
"grad_norm": 1.578125,
"learning_rate": 8.247422680412371e-05,
"loss": 1.4914,
"step": 40
},
{
"epoch": 0.04657292786856085,
"grad_norm": 0.408203125,
"learning_rate": 9.278350515463918e-05,
"loss": 1.4412,
"step": 45
},
{
"epoch": 0.05174769763173428,
"grad_norm": 0.197265625,
"learning_rate": 0.00010309278350515463,
"loss": 1.4186,
"step": 50
},
{
"epoch": 0.0569224673949077,
"grad_norm": 0.26171875,
"learning_rate": 0.0001134020618556701,
"loss": 1.4002,
"step": 55
},
{
"epoch": 0.06209723715808113,
"grad_norm": 0.1552734375,
"learning_rate": 0.00012371134020618558,
"loss": 1.3963,
"step": 60
},
{
"epoch": 0.06727200692125455,
"grad_norm": 0.16015625,
"learning_rate": 0.00013402061855670103,
"loss": 1.3791,
"step": 65
},
{
"epoch": 0.07244677668442799,
"grad_norm": 0.1640625,
"learning_rate": 0.0001443298969072165,
"loss": 1.3596,
"step": 70
},
{
"epoch": 0.07762154644760141,
"grad_norm": 0.1484375,
"learning_rate": 0.00015463917525773197,
"loss": 1.3376,
"step": 75
},
{
"epoch": 0.08279631621077484,
"grad_norm": 0.08935546875,
"learning_rate": 0.00016494845360824742,
"loss": 1.3465,
"step": 80
},
{
"epoch": 0.08797108597394827,
"grad_norm": 0.09521484375,
"learning_rate": 0.0001752577319587629,
"loss": 1.3416,
"step": 85
},
{
"epoch": 0.0931458557371217,
"grad_norm": 0.076171875,
"learning_rate": 0.00018556701030927837,
"loss": 1.329,
"step": 90
},
{
"epoch": 0.09832062550029512,
"grad_norm": 0.0869140625,
"learning_rate": 0.00019587628865979381,
"loss": 1.3077,
"step": 95
},
{
"epoch": 0.10349539526346856,
"grad_norm": 0.0869140625,
"learning_rate": 0.00019999411876613578,
"loss": 1.3115,
"step": 100
},
{
"epoch": 0.10867016502664198,
"grad_norm": 0.07568359375,
"learning_rate": 0.00019995818039769561,
"loss": 1.3185,
"step": 105
},
{
"epoch": 0.1138449347898154,
"grad_norm": 0.10302734375,
"learning_rate": 0.00019988958274085367,
"loss": 1.2811,
"step": 110
},
{
"epoch": 0.11901970455298884,
"grad_norm": 0.09375,
"learning_rate": 0.0001997883482084895,
"loss": 1.3017,
"step": 115
},
{
"epoch": 0.12419447431616226,
"grad_norm": 0.0771484375,
"learning_rate": 0.0001996545098769134,
"loss": 1.2927,
"step": 120
},
{
"epoch": 0.1293692440793357,
"grad_norm": 0.091796875,
"learning_rate": 0.0001994881114750593,
"loss": 1.3015,
"step": 125
},
{
"epoch": 0.1345440138425091,
"grad_norm": 0.09228515625,
"learning_rate": 0.00019928920737019733,
"loss": 1.2854,
"step": 130
},
{
"epoch": 0.13971878360568254,
"grad_norm": 0.126953125,
"learning_rate": 0.00019905786255017044,
"loss": 1.2995,
"step": 135
},
{
"epoch": 0.14489355336885598,
"grad_norm": 0.1728515625,
"learning_rate": 0.00019879415260216088,
"loss": 1.2739,
"step": 140
},
{
"epoch": 0.1500683231320294,
"grad_norm": 0.11328125,
"learning_rate": 0.00019849816368799355,
"loss": 1.2802,
"step": 145
},
{
"epoch": 0.15524309289520283,
"grad_norm": 0.08740234375,
"learning_rate": 0.0001981699925159847,
"loss": 1.3024,
"step": 150
},
{
"epoch": 0.16041786265837626,
"grad_norm": 0.07421875,
"learning_rate": 0.0001978097463093441,
"loss": 1.2642,
"step": 155
},
{
"epoch": 0.16559263242154967,
"grad_norm": 0.06982421875,
"learning_rate": 0.00019741754277114232,
"loss": 1.2516,
"step": 160
},
{
"epoch": 0.1707674021847231,
"grad_norm": 0.072265625,
"learning_rate": 0.00019699351004585352,
"loss": 1.2675,
"step": 165
},
{
"epoch": 0.17594217194789655,
"grad_norm": 0.07958984375,
"learning_rate": 0.00019653778667748695,
"loss": 1.2646,
"step": 170
},
{
"epoch": 0.18111694171106996,
"grad_norm": 0.095703125,
"learning_rate": 0.00019605052156432043,
"loss": 1.2773,
"step": 175
},
{
"epoch": 0.1862917114742434,
"grad_norm": 0.07763671875,
"learning_rate": 0.000195531873910251,
"loss": 1.2653,
"step": 180
},
{
"epoch": 0.19146648123741683,
"grad_norm": 0.10595703125,
"learning_rate": 0.00019498201317277828,
"loss": 1.2572,
"step": 185
},
{
"epoch": 0.19664125100059024,
"grad_norm": 0.0654296875,
"learning_rate": 0.0001944011190076376,
"loss": 1.2737,
"step": 190
},
{
"epoch": 0.20181602076376368,
"grad_norm": 0.0869140625,
"learning_rate": 0.00019378938121010128,
"loss": 1.2848,
"step": 195
},
{
"epoch": 0.2069907905269371,
"grad_norm": 0.08154296875,
"learning_rate": 0.00019314699965296696,
"loss": 1.2814,
"step": 200
},
{
"epoch": 0.21216556029011052,
"grad_norm": 0.337890625,
"learning_rate": 0.000192474184221253,
"loss": 1.2685,
"step": 205
},
{
"epoch": 0.21734033005328396,
"grad_norm": 0.07275390625,
"learning_rate": 0.00019177115474362322,
"loss": 1.2611,
"step": 210
},
{
"epoch": 0.2225150998164574,
"grad_norm": 0.06982421875,
"learning_rate": 0.000191038140920562,
"loss": 1.2541,
"step": 215
},
{
"epoch": 0.2276898695796308,
"grad_norm": 0.11083984375,
"learning_rate": 0.0001902753822493248,
"loss": 1.2657,
"step": 220
},
{
"epoch": 0.23286463934280424,
"grad_norm": 0.09912109375,
"learning_rate": 0.00018948312794568673,
"loss": 1.2687,
"step": 225
},
{
"epoch": 0.23803940910597768,
"grad_norm": 0.08349609375,
"learning_rate": 0.0001886616368625171,
"loss": 1.2779,
"step": 230
},
{
"epoch": 0.2432141788691511,
"grad_norm": 0.08935546875,
"learning_rate": 0.00018781117740520386,
"loss": 1.2647,
"step": 235
},
{
"epoch": 0.24838894863232452,
"grad_norm": 0.0810546875,
"learning_rate": 0.00018693202744395827,
"loss": 1.244,
"step": 240
},
{
"epoch": 0.25356371839549796,
"grad_norm": 0.12451171875,
"learning_rate": 0.00018602447422302597,
"loss": 1.2408,
"step": 245
},
{
"epoch": 0.2587384881586714,
"grad_norm": 0.1318359375,
"learning_rate": 0.00018508881426683614,
"loss": 1.2875,
"step": 250
},
{
"epoch": 0.2639132579218448,
"grad_norm": 0.076171875,
"learning_rate": 0.00018412535328311814,
"loss": 1.259,
"step": 255
},
{
"epoch": 0.2690880276850182,
"grad_norm": 0.09033203125,
"learning_rate": 0.00018313440606301763,
"loss": 1.2646,
"step": 260
},
{
"epoch": 0.27426279744819165,
"grad_norm": 0.083984375,
"learning_rate": 0.00018211629637824515,
"loss": 1.2452,
"step": 265
},
{
"epoch": 0.2794375672113651,
"grad_norm": 0.08740234375,
"learning_rate": 0.00018107135687529043,
"loss": 1.2608,
"step": 270
},
{
"epoch": 0.2846123369745385,
"grad_norm": 0.09765625,
"learning_rate": 0.00017999992896673665,
"loss": 1.2508,
"step": 275
},
{
"epoch": 0.28978710673771196,
"grad_norm": 0.11962890625,
"learning_rate": 0.00017890236271971122,
"loss": 1.2413,
"step": 280
},
{
"epoch": 0.29496187650088534,
"grad_norm": 0.1015625,
"learning_rate": 0.00017777901674150828,
"loss": 1.269,
"step": 285
},
{
"epoch": 0.3001366462640588,
"grad_norm": 0.0859375,
"learning_rate": 0.00017663025806242123,
"loss": 1.2453,
"step": 290
},
{
"epoch": 0.3053114160272322,
"grad_norm": 0.07421875,
"learning_rate": 0.00017545646201582303,
"loss": 1.2591,
"step": 295
},
{
"epoch": 0.31048618579040566,
"grad_norm": 0.08935546875,
"learning_rate": 0.00017425801211553373,
"loss": 1.2555,
"step": 300
},
{
"epoch": 0.3156609555535791,
"grad_norm": 0.07080078125,
"learning_rate": 0.0001730352999305152,
"loss": 1.2656,
"step": 305
},
{
"epoch": 0.32083572531675253,
"grad_norm": 0.154296875,
"learning_rate": 0.00017178872495693398,
"loss": 1.2608,
"step": 310
},
{
"epoch": 0.3260104950799259,
"grad_norm": 0.076171875,
"learning_rate": 0.00017051869448763408,
"loss": 1.2787,
"step": 315
},
{
"epoch": 0.33118526484309935,
"grad_norm": 0.1513671875,
"learning_rate": 0.00016922562347906238,
"loss": 1.2887,
"step": 320
},
{
"epoch": 0.3363600346062728,
"grad_norm": 0.11962890625,
"learning_rate": 0.00016790993441569022,
"loss": 1.2624,
"step": 325
},
{
"epoch": 0.3415348043694462,
"grad_norm": 0.07470703125,
"learning_rate": 0.00016657205717197495,
"loss": 1.2616,
"step": 330
},
{
"epoch": 0.34670957413261966,
"grad_norm": 0.1162109375,
"learning_rate": 0.00016521242887190764,
"loss": 1.2643,
"step": 335
},
{
"epoch": 0.3518843438957931,
"grad_norm": 0.123046875,
"learning_rate": 0.0001638314937461915,
"loss": 1.2767,
"step": 340
},
{
"epoch": 0.3570591136589665,
"grad_norm": 0.0986328125,
"learning_rate": 0.00016242970298709867,
"loss": 1.2648,
"step": 345
},
{
"epoch": 0.3622338834221399,
"grad_norm": 0.12109375,
"learning_rate": 0.00016100751460105243,
"loss": 1.2409,
"step": 350
},
{
"epoch": 0.36740865318531335,
"grad_norm": 0.0908203125,
"learning_rate": 0.000159565393258983,
"loss": 1.2392,
"step": 355
},
{
"epoch": 0.3725834229484868,
"grad_norm": 0.0849609375,
"learning_rate": 0.00015810381014450556,
"loss": 1.2709,
"step": 360
},
{
"epoch": 0.3777581927116602,
"grad_norm": 0.0908203125,
"learning_rate": 0.00015662324279997115,
"loss": 1.2556,
"step": 365
},
{
"epoch": 0.38293296247483366,
"grad_norm": 0.10498046875,
"learning_rate": 0.000155124174970439,
"loss": 1.2623,
"step": 370
},
{
"epoch": 0.38810773223800704,
"grad_norm": 0.08935546875,
"learning_rate": 0.00015360709644562297,
"loss": 1.2489,
"step": 375
},
{
"epoch": 0.3932825020011805,
"grad_norm": 0.07861328125,
"learning_rate": 0.0001520725028998629,
"loss": 1.2771,
"step": 380
},
{
"epoch": 0.3984572717643539,
"grad_norm": 0.10693359375,
"learning_rate": 0.00015052089573017283,
"loss": 1.2523,
"step": 385
},
{
"epoch": 0.40363204152752735,
"grad_norm": 0.0869140625,
"learning_rate": 0.00014895278189242016,
"loss": 1.241,
"step": 390
},
{
"epoch": 0.4088068112907008,
"grad_norm": 0.10302734375,
"learning_rate": 0.00014736867373568794,
"loss": 1.2578,
"step": 395
},
{
"epoch": 0.4139815810538742,
"grad_norm": 0.0810546875,
"learning_rate": 0.00014576908883487548,
"loss": 1.2625,
"step": 400
},
{
"epoch": 0.4191563508170476,
"grad_norm": 0.08935546875,
"learning_rate": 0.0001441545498215912,
"loss": 1.2606,
"step": 405
},
{
"epoch": 0.42433112058022104,
"grad_norm": 0.10009765625,
"learning_rate": 0.00014252558421339354,
"loss": 1.2321,
"step": 410
},
{
"epoch": 0.4295058903433945,
"grad_norm": 0.0810546875,
"learning_rate": 0.00014088272424143546,
"loss": 1.2494,
"step": 415
},
{
"epoch": 0.4346806601065679,
"grad_norm": 0.0771484375,
"learning_rate": 0.0001392265066765686,
"loss": 1.2588,
"step": 420
},
{
"epoch": 0.43985542986974135,
"grad_norm": 0.07958984375,
"learning_rate": 0.00013755747265396467,
"loss": 1.2534,
"step": 425
},
{
"epoch": 0.4450301996329148,
"grad_norm": 0.091796875,
"learning_rate": 0.00013587616749631036,
"loss": 1.2393,
"step": 430
},
{
"epoch": 0.4502049693960882,
"grad_norm": 0.0849609375,
"learning_rate": 0.00013418314053563455,
"loss": 1.2526,
"step": 435
},
{
"epoch": 0.4553797391592616,
"grad_norm": 0.078125,
"learning_rate": 0.00013247894493382506,
"loss": 1.2459,
"step": 440
},
{
"epoch": 0.46055450892243505,
"grad_norm": 0.080078125,
"learning_rate": 0.00013076413750189467,
"loss": 1.2582,
"step": 445
},
{
"epoch": 0.4657292786856085,
"grad_norm": 0.10888671875,
"learning_rate": 0.00012903927851805414,
"loss": 1.2537,
"step": 450
},
{
"epoch": 0.4709040484487819,
"grad_norm": 0.08447265625,
"learning_rate": 0.00012730493154465311,
"loss": 1.2601,
"step": 455
},
{
"epoch": 0.47607881821195536,
"grad_norm": 0.0869140625,
"learning_rate": 0.0001255616632440475,
"loss": 1.2612,
"step": 460
},
{
"epoch": 0.48125358797512874,
"grad_norm": 0.08740234375,
"learning_rate": 0.00012381004319345402,
"loss": 1.2555,
"step": 465
},
{
"epoch": 0.4864283577383022,
"grad_norm": 0.17578125,
"learning_rate": 0.00012205064369885291,
"loss": 1.247,
"step": 470
},
{
"epoch": 0.4916031275014756,
"grad_norm": 0.09228515625,
"learning_rate": 0.00012028403960799821,
"loss": 1.2467,
"step": 475
},
{
"epoch": 0.49677789726464905,
"grad_norm": 0.0791015625,
"learning_rate": 0.00011851080812259838,
"loss": 1.2369,
"step": 480
},
{
"epoch": 0.5019526670278225,
"grad_norm": 0.08203125,
"learning_rate": 0.00011673152860972724,
"loss": 1.2644,
"step": 485
},
{
"epoch": 0.5071274367909959,
"grad_norm": 0.0966796875,
"learning_rate": 0.00011494678241252781,
"loss": 1.2345,
"step": 490
},
{
"epoch": 0.5123022065541694,
"grad_norm": 0.30078125,
"learning_rate": 0.00011315715266027014,
"loss": 1.2556,
"step": 495
},
{
"epoch": 0.5174769763173428,
"grad_norm": 0.0810546875,
"learning_rate": 0.00011136322407782602,
"loss": 1.2403,
"step": 500
},
{
"epoch": 0.5226517460805162,
"grad_norm": 0.0751953125,
"learning_rate": 0.00010956558279462187,
"loss": 1.2345,
"step": 505
},
{
"epoch": 0.5278265158436896,
"grad_norm": 0.099609375,
"learning_rate": 0.00010776481615313315,
"loss": 1.2298,
"step": 510
},
{
"epoch": 0.533001285606863,
"grad_norm": 0.087890625,
"learning_rate": 0.00010596151251698199,
"loss": 1.2475,
"step": 515
},
{
"epoch": 0.5381760553700364,
"grad_norm": 0.08056640625,
"learning_rate": 0.0001041562610787017,
"loss": 1.2542,
"step": 520
},
{
"epoch": 0.5433508251332099,
"grad_norm": 0.1025390625,
"learning_rate": 0.00010234965166723003,
"loss": 1.2534,
"step": 525
},
{
"epoch": 0.5485255948963833,
"grad_norm": 0.1474609375,
"learning_rate": 0.00010054227455519458,
"loss": 1.2598,
"step": 530
},
{
"epoch": 0.5537003646595567,
"grad_norm": 0.1953125,
"learning_rate": 9.873472026605346e-05,
"loss": 1.2551,
"step": 535
},
{
"epoch": 0.5588751344227302,
"grad_norm": 0.0771484375,
"learning_rate": 9.692757938115376e-05,
"loss": 1.2553,
"step": 540
},
{
"epoch": 0.5640499041859036,
"grad_norm": 0.10888671875,
"learning_rate": 9.512144234677116e-05,
"loss": 1.2584,
"step": 545
},
{
"epoch": 0.569224673949077,
"grad_norm": 0.08642578125,
"learning_rate": 9.331689928119381e-05,
"loss": 1.2492,
"step": 550
},
{
"epoch": 0.5743994437122505,
"grad_norm": 0.07177734375,
"learning_rate": 9.151453978191311e-05,
"loss": 1.2315,
"step": 555
},
{
"epoch": 0.5795742134754239,
"grad_norm": 0.0830078125,
"learning_rate": 8.9714952732985e-05,
"loss": 1.2255,
"step": 560
},
{
"epoch": 0.5847489832385974,
"grad_norm": 0.08984375,
"learning_rate": 8.791872611262393e-05,
"loss": 1.2269,
"step": 565
},
{
"epoch": 0.5899237530017707,
"grad_norm": 0.0791015625,
"learning_rate": 8.612644680109319e-05,
"loss": 1.2481,
"step": 570
},
{
"epoch": 0.5950985227649441,
"grad_norm": 0.07958984375,
"learning_rate": 8.433870038895377e-05,
"loss": 1.2304,
"step": 575
},
{
"epoch": 0.6002732925281176,
"grad_norm": 0.09716796875,
"learning_rate": 8.255607098573454e-05,
"loss": 1.2507,
"step": 580
},
{
"epoch": 0.605448062291291,
"grad_norm": 0.1162109375,
"learning_rate": 8.077914102908661e-05,
"loss": 1.2652,
"step": 585
},
{
"epoch": 0.6106228320544644,
"grad_norm": 0.07470703125,
"learning_rate": 7.90084910944835e-05,
"loss": 1.2474,
"step": 590
},
{
"epoch": 0.6157976018176379,
"grad_norm": 0.1806640625,
"learning_rate": 7.724469970553027e-05,
"loss": 1.2387,
"step": 595
},
{
"epoch": 0.6209723715808113,
"grad_norm": 0.0712890625,
"learning_rate": 7.548834314494255e-05,
"loss": 1.236,
"step": 600
},
{
"epoch": 0.6261471413439847,
"grad_norm": 0.09423828125,
"learning_rate": 7.37399952662581e-05,
"loss": 1.2304,
"step": 605
},
{
"epoch": 0.6313219111071582,
"grad_norm": 0.08154296875,
"learning_rate": 7.200022730634187e-05,
"loss": 1.229,
"step": 610
},
{
"epoch": 0.6364966808703316,
"grad_norm": 0.1552734375,
"learning_rate": 7.026960769874605e-05,
"loss": 1.2483,
"step": 615
},
{
"epoch": 0.6416714506335051,
"grad_norm": 0.07080078125,
"learning_rate": 6.85487018879861e-05,
"loss": 1.2515,
"step": 620
},
{
"epoch": 0.6468462203966785,
"grad_norm": 0.140625,
"learning_rate": 6.683807214479323e-05,
"loss": 1.2388,
"step": 625
},
{
"epoch": 0.6520209901598518,
"grad_norm": 0.10009765625,
"learning_rate": 6.513827738240407e-05,
"loss": 1.2316,
"step": 630
},
{
"epoch": 0.6571957599230253,
"grad_norm": 0.08203125,
"learning_rate": 6.344987297394713e-05,
"loss": 1.2409,
"step": 635
},
{
"epoch": 0.6623705296861987,
"grad_norm": 0.11767578125,
"learning_rate": 6.177341057098607e-05,
"loss": 1.2661,
"step": 640
},
{
"epoch": 0.6675452994493721,
"grad_norm": 0.08056640625,
"learning_rate": 6.010943792327875e-05,
"loss": 1.2272,
"step": 645
},
{
"epoch": 0.6727200692125456,
"grad_norm": 0.1142578125,
"learning_rate": 5.845849869981137e-05,
"loss": 1.2615,
"step": 650
},
{
"epoch": 0.677894838975719,
"grad_norm": 0.07275390625,
"learning_rate": 5.682113231116557e-05,
"loss": 1.2653,
"step": 655
},
{
"epoch": 0.6830696087388924,
"grad_norm": 0.0751953125,
"learning_rate": 5.519787373327725e-05,
"loss": 1.2355,
"step": 660
},
{
"epoch": 0.6882443785020659,
"grad_norm": 0.0859375,
"learning_rate": 5.358925333264403e-05,
"loss": 1.2322,
"step": 665
},
{
"epoch": 0.6934191482652393,
"grad_norm": 0.08349609375,
"learning_rate": 5.1995796693038844e-05,
"loss": 1.2549,
"step": 670
},
{
"epoch": 0.6985939180284128,
"grad_norm": 0.0927734375,
"learning_rate": 5.0418024443786395e-05,
"loss": 1.2155,
"step": 675
},
{
"epoch": 0.7037686877915862,
"grad_norm": 0.078125,
"learning_rate": 4.885645208965779e-05,
"loss": 1.2497,
"step": 680
},
{
"epoch": 0.7089434575547596,
"grad_norm": 0.072265625,
"learning_rate": 4.7311589842440415e-05,
"loss": 1.2692,
"step": 685
},
{
"epoch": 0.714118227317933,
"grad_norm": 0.07275390625,
"learning_rate": 4.578394245423626e-05,
"loss": 1.2593,
"step": 690
},
{
"epoch": 0.7192929970811064,
"grad_norm": 0.0712890625,
"learning_rate": 4.427400905254483e-05,
"loss": 1.2458,
"step": 695
},
{
"epoch": 0.7244677668442798,
"grad_norm": 0.09033203125,
"learning_rate": 4.278228297718307e-05,
"loss": 1.2285,
"step": 700
},
{
"epoch": 0.7296425366074533,
"grad_norm": 0.09423828125,
"learning_rate": 4.1309251619097154e-05,
"loss": 1.243,
"step": 705
},
{
"epoch": 0.7348173063706267,
"grad_norm": 0.0791015625,
"learning_rate": 3.985539626111708e-05,
"loss": 1.2492,
"step": 710
},
{
"epoch": 0.7399920761338001,
"grad_norm": 0.10693359375,
"learning_rate": 3.842119192070762e-05,
"loss": 1.2401,
"step": 715
},
{
"epoch": 0.7451668458969736,
"grad_norm": 0.12158203125,
"learning_rate": 3.700710719476594e-05,
"loss": 1.2563,
"step": 720
},
{
"epoch": 0.750341615660147,
"grad_norm": 0.08740234375,
"learning_rate": 3.561360410651713e-05,
"loss": 1.2479,
"step": 725
},
{
"epoch": 0.7555163854233204,
"grad_norm": 0.07373046875,
"learning_rate": 3.424113795455779e-05,
"loss": 1.2366,
"step": 730
},
{
"epoch": 0.7606911551864939,
"grad_norm": 0.12255859375,
"learning_rate": 3.289015716409631e-05,
"loss": 1.2184,
"step": 735
},
{
"epoch": 0.7658659249496673,
"grad_norm": 0.0693359375,
"learning_rate": 3.156110314043933e-05,
"loss": 1.2454,
"step": 740
},
{
"epoch": 0.7710406947128408,
"grad_norm": 0.095703125,
"learning_rate": 3.025441012477126e-05,
"loss": 1.2323,
"step": 745
},
{
"epoch": 0.7762154644760141,
"grad_norm": 0.1806640625,
"learning_rate": 2.89705050522753e-05,
"loss": 1.2344,
"step": 750
},
{
"epoch": 0.7813902342391875,
"grad_norm": 0.06689453125,
"learning_rate": 2.7709807412640743e-05,
"loss": 1.2428,
"step": 755
},
{
"epoch": 0.786565004002361,
"grad_norm": 0.1279296875,
"learning_rate": 2.6472729113003615e-05,
"loss": 1.2585,
"step": 760
},
{
"epoch": 0.7917397737655344,
"grad_norm": 0.06884765625,
"learning_rate": 2.5259674343364104e-05,
"loss": 1.2397,
"step": 765
},
{
"epoch": 0.7969145435287078,
"grad_norm": 0.15625,
"learning_rate": 2.4071039444526046e-05,
"loss": 1.2581,
"step": 770
},
{
"epoch": 0.8020893132918813,
"grad_norm": 0.07177734375,
"learning_rate": 2.290721277860064e-05,
"loss": 1.2582,
"step": 775
},
{
"epoch": 0.8072640830550547,
"grad_norm": 0.09716796875,
"learning_rate": 2.176857460211693e-05,
"loss": 1.2389,
"step": 780
},
{
"epoch": 0.8124388528182281,
"grad_norm": 0.09716796875,
"learning_rate": 2.0655496941780994e-05,
"loss": 1.245,
"step": 785
},
{
"epoch": 0.8176136225814016,
"grad_norm": 0.091796875,
"learning_rate": 1.9568343472923524e-05,
"loss": 1.2398,
"step": 790
},
{
"epoch": 0.822788392344575,
"grad_norm": 0.10205078125,
"learning_rate": 1.8507469400676735e-05,
"loss": 1.2408,
"step": 795
},
{
"epoch": 0.8279631621077485,
"grad_norm": 0.076171875,
"learning_rate": 1.7473221343918055e-05,
"loss": 1.2355,
"step": 800
},
{
"epoch": 0.8331379318709219,
"grad_norm": 0.1826171875,
"learning_rate": 1.6465937222019745e-05,
"loss": 1.2552,
"step": 805
},
{
"epoch": 0.8383127016340952,
"grad_norm": 0.08544921875,
"learning_rate": 1.5485946144440354e-05,
"loss": 1.2461,
"step": 810
},
{
"epoch": 0.8434874713972687,
"grad_norm": 0.07861328125,
"learning_rate": 1.4533568303195333e-05,
"loss": 1.2449,
"step": 815
},
{
"epoch": 0.8486622411604421,
"grad_norm": 0.06591796875,
"learning_rate": 1.3609114868240536e-05,
"loss": 1.2424,
"step": 820
},
{
"epoch": 0.8538370109236155,
"grad_norm": 0.0986328125,
"learning_rate": 1.2712887885803937e-05,
"loss": 1.2218,
"step": 825
},
{
"epoch": 0.859011780686789,
"grad_norm": 0.06982421875,
"learning_rate": 1.184518017969829e-05,
"loss": 1.2556,
"step": 830
},
{
"epoch": 0.8641865504499624,
"grad_norm": 0.10498046875,
"learning_rate": 1.1006275255646681e-05,
"loss": 1.2493,
"step": 835
},
{
"epoch": 0.8693613202131358,
"grad_norm": 0.0751953125,
"learning_rate": 1.0196447208653193e-05,
"loss": 1.2509,
"step": 840
},
{
"epoch": 0.8745360899763093,
"grad_norm": 0.10009765625,
"learning_rate": 9.415960633447674e-06,
"loss": 1.2322,
"step": 845
},
{
"epoch": 0.8797108597394827,
"grad_norm": 0.0908203125,
"learning_rate": 8.665070538035036e-06,
"loss": 1.2422,
"step": 850
},
{
"epoch": 0.8848856295026561,
"grad_norm": 0.109375,
"learning_rate": 7.944022260376416e-06,
"loss": 1.238,
"step": 855
},
{
"epoch": 0.8900603992658296,
"grad_norm": 0.10205078125,
"learning_rate": 7.253051388230248e-06,
"loss": 1.25,
"step": 860
},
{
"epoch": 0.895235169029003,
"grad_norm": 0.07177734375,
"learning_rate": 6.592383682178593e-06,
"loss": 1.2485,
"step": 865
},
{
"epoch": 0.9004099387921763,
"grad_norm": 0.06689453125,
"learning_rate": 5.96223500186458e-06,
"loss": 1.2503,
"step": 870
},
{
"epoch": 0.9055847085553498,
"grad_norm": 0.10400390625,
"learning_rate": 5.362811235464726e-06,
"loss": 1.2552,
"step": 875
},
{
"epoch": 0.9107594783185232,
"grad_norm": 0.07373046875,
"learning_rate": 4.794308232419065e-06,
"loss": 1.2271,
"step": 880
},
{
"epoch": 0.9159342480816967,
"grad_norm": 0.11865234375,
"learning_rate": 4.256911739441505e-06,
"loss": 1.2295,
"step": 885
},
{
"epoch": 0.9211090178448701,
"grad_norm": 0.068359375,
"learning_rate": 3.7507973398307584e-06,
"loss": 1.2511,
"step": 890
},
{
"epoch": 0.9262837876080435,
"grad_norm": 0.07177734375,
"learning_rate": 3.276130396102217e-06,
"loss": 1.2527,
"step": 895
},
{
"epoch": 0.931458557371217,
"grad_norm": 0.2080078125,
"learning_rate": 2.8330659959589946e-06,
"loss": 1.2444,
"step": 900
},
{
"epoch": 0.9366333271343904,
"grad_norm": 0.06640625,
"learning_rate": 2.421748901620369e-06,
"loss": 1.2495,
"step": 905
},
{
"epoch": 0.9418080968975638,
"grad_norm": 0.06884765625,
"learning_rate": 2.0423135025235294e-06,
"loss": 1.2631,
"step": 910
},
{
"epoch": 0.9469828666607373,
"grad_norm": 0.08154296875,
"learning_rate": 1.6948837714146793e-06,
"loss": 1.2336,
"step": 915
},
{
"epoch": 0.9521576364239107,
"grad_norm": 0.0712890625,
"learning_rate": 1.3795732238434556e-06,
"loss": 1.28,
"step": 920
},
{
"epoch": 0.9573324061870842,
"grad_norm": 0.107421875,
"learning_rate": 1.0964848810740113e-06,
"loss": 1.2429,
"step": 925
},
{
"epoch": 0.9625071759502575,
"grad_norm": 0.07958984375,
"learning_rate": 8.457112364250064e-07,
"loss": 1.2278,
"step": 930
},
{
"epoch": 0.9676819457134309,
"grad_norm": 0.10302734375,
"learning_rate": 6.273342250492254e-07,
"loss": 1.2675,
"step": 935
},
{
"epoch": 0.9728567154766044,
"grad_norm": 0.09326171875,
"learning_rate": 4.4142519716300344e-07,
"loss": 1.2405,
"step": 940
},
{
"epoch": 0.9780314852397778,
"grad_norm": 0.072265625,
"learning_rate": 2.8804489473395203e-07,
"loss": 1.2512,
"step": 945
},
{
"epoch": 0.9832062550029512,
"grad_norm": 0.0859375,
"learning_rate": 1.6724343163483368e-07,
"loss": 1.2416,
"step": 950
},
{
"epoch": 0.9883810247661247,
"grad_norm": 0.1171875,
"learning_rate": 7.906027726981568e-08,
"loss": 1.2292,
"step": 955
},
{
"epoch": 0.9935557945292981,
"grad_norm": 0.08935546875,
"learning_rate": 2.3524243678685332e-08,
"loss": 1.2187,
"step": 960
},
{
"epoch": 0.9987305642924715,
"grad_norm": 0.07080078125,
"learning_rate": 6.534761230914122e-10,
"loss": 1.2557,
"step": 965
},
{
"epoch": 0.9997655182451062,
"eval_loss": 1.2492531538009644,
"eval_runtime": 2450.1313,
"eval_samples_per_second": 5.585,
"eval_steps_per_second": 5.585,
"step": 966
},
{
"epoch": 0.9997655182451062,
"step": 966,
"total_flos": 3.041046347630248e+18,
"train_loss": 1.1233504201067654,
"train_runtime": 77534.6649,
"train_samples_per_second": 1.595,
"train_steps_per_second": 0.012
}
],
"logging_steps": 5,
"max_steps": 966,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 20,
"total_flos": 3.041046347630248e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}