|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.0, |
|
"eval_steps": 500, |
|
"global_step": 3188, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0012547051442910915, |
|
"grad_norm": 1.3249917794841757, |
|
"learning_rate": 6.269592476489028e-08, |
|
"loss": 0.5553, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.006273525721455458, |
|
"grad_norm": 1.3024175063683066, |
|
"learning_rate": 3.134796238244514e-07, |
|
"loss": 0.5542, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.012547051442910916, |
|
"grad_norm": 1.1325683997600668, |
|
"learning_rate": 6.269592476489028e-07, |
|
"loss": 0.5626, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.018820577164366373, |
|
"grad_norm": 1.1174840712777703, |
|
"learning_rate": 9.404388714733543e-07, |
|
"loss": 0.5642, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.025094102885821833, |
|
"grad_norm": 0.8002431481734196, |
|
"learning_rate": 1.2539184952978056e-06, |
|
"loss": 0.5294, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03136762860727729, |
|
"grad_norm": 0.6333668642117685, |
|
"learning_rate": 1.5673981191222572e-06, |
|
"loss": 0.5079, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.037641154328732745, |
|
"grad_norm": 0.5771411644874956, |
|
"learning_rate": 1.8808777429467086e-06, |
|
"loss": 0.4655, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.043914680050188205, |
|
"grad_norm": 0.49532156556877105, |
|
"learning_rate": 2.1943573667711602e-06, |
|
"loss": 0.4592, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.050188205771643665, |
|
"grad_norm": 0.4975545815570396, |
|
"learning_rate": 2.507836990595611e-06, |
|
"loss": 0.4329, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.056461731493099125, |
|
"grad_norm": 0.4338284541758203, |
|
"learning_rate": 2.8213166144200626e-06, |
|
"loss": 0.4322, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.06273525721455459, |
|
"grad_norm": 0.40887016256028313, |
|
"learning_rate": 3.1347962382445144e-06, |
|
"loss": 0.426, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.06900878293601004, |
|
"grad_norm": 0.4045559102283436, |
|
"learning_rate": 3.448275862068966e-06, |
|
"loss": 0.4227, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.07528230865746549, |
|
"grad_norm": 0.3889912721227527, |
|
"learning_rate": 3.7617554858934172e-06, |
|
"loss": 0.4189, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.08155583437892096, |
|
"grad_norm": 0.3618436855347915, |
|
"learning_rate": 4.075235109717869e-06, |
|
"loss": 0.4177, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.08782936010037641, |
|
"grad_norm": 0.36459156314439295, |
|
"learning_rate": 4.3887147335423205e-06, |
|
"loss": 0.4049, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.09410288582183186, |
|
"grad_norm": 0.3488164206674813, |
|
"learning_rate": 4.7021943573667714e-06, |
|
"loss": 0.3961, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.10037641154328733, |
|
"grad_norm": 0.37238296726059605, |
|
"learning_rate": 5.015673981191222e-06, |
|
"loss": 0.3955, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.10664993726474278, |
|
"grad_norm": 0.3605666937163523, |
|
"learning_rate": 5.329153605015674e-06, |
|
"loss": 0.377, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.11292346298619825, |
|
"grad_norm": 0.35760877488985304, |
|
"learning_rate": 5.642633228840125e-06, |
|
"loss": 0.3915, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.1191969887076537, |
|
"grad_norm": 0.356976698911797, |
|
"learning_rate": 5.956112852664577e-06, |
|
"loss": 0.3938, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.12547051442910917, |
|
"grad_norm": 0.3620265985263758, |
|
"learning_rate": 6.269592476489029e-06, |
|
"loss": 0.404, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.13174404015056462, |
|
"grad_norm": 0.3710088599948379, |
|
"learning_rate": 6.58307210031348e-06, |
|
"loss": 0.3941, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.13801756587202008, |
|
"grad_norm": 0.3831687285367315, |
|
"learning_rate": 6.896551724137932e-06, |
|
"loss": 0.3859, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.14429109159347553, |
|
"grad_norm": 0.36915661135239697, |
|
"learning_rate": 7.210031347962383e-06, |
|
"loss": 0.3895, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.15056461731493098, |
|
"grad_norm": 0.3573608106603279, |
|
"learning_rate": 7.5235109717868345e-06, |
|
"loss": 0.3857, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.15683814303638646, |
|
"grad_norm": 0.3508976740749952, |
|
"learning_rate": 7.836990595611285e-06, |
|
"loss": 0.3915, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.16311166875784192, |
|
"grad_norm": 0.36876078420022057, |
|
"learning_rate": 8.150470219435737e-06, |
|
"loss": 0.3989, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.16938519447929737, |
|
"grad_norm": 0.38691192833572297, |
|
"learning_rate": 8.463949843260189e-06, |
|
"loss": 0.3874, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.17565872020075282, |
|
"grad_norm": 0.3735894147297392, |
|
"learning_rate": 8.777429467084641e-06, |
|
"loss": 0.385, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.18193224592220827, |
|
"grad_norm": 0.3710457112042887, |
|
"learning_rate": 9.090909090909091e-06, |
|
"loss": 0.3764, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.18820577164366373, |
|
"grad_norm": 0.3708916534975576, |
|
"learning_rate": 9.404388714733543e-06, |
|
"loss": 0.3808, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1944792973651192, |
|
"grad_norm": 0.3552539505765215, |
|
"learning_rate": 9.717868338557995e-06, |
|
"loss": 0.4018, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.20075282308657466, |
|
"grad_norm": 0.3666385022952003, |
|
"learning_rate": 1.0031347962382445e-05, |
|
"loss": 0.3839, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.20702634880803011, |
|
"grad_norm": 0.3631078993279255, |
|
"learning_rate": 1.0344827586206898e-05, |
|
"loss": 0.3858, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.21329987452948557, |
|
"grad_norm": 0.39722641298153644, |
|
"learning_rate": 1.0658307210031348e-05, |
|
"loss": 0.3847, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.21957340025094102, |
|
"grad_norm": 0.3691443252830364, |
|
"learning_rate": 1.09717868338558e-05, |
|
"loss": 0.3826, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.2258469259723965, |
|
"grad_norm": 0.39755512136650856, |
|
"learning_rate": 1.128526645768025e-05, |
|
"loss": 0.3766, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.23212045169385195, |
|
"grad_norm": 0.3627344076525724, |
|
"learning_rate": 1.1598746081504704e-05, |
|
"loss": 0.409, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.2383939774153074, |
|
"grad_norm": 0.36755006735418844, |
|
"learning_rate": 1.1912225705329154e-05, |
|
"loss": 0.385, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.24466750313676286, |
|
"grad_norm": 0.3491529760844153, |
|
"learning_rate": 1.2225705329153606e-05, |
|
"loss": 0.3944, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.25094102885821834, |
|
"grad_norm": 0.38344322242421625, |
|
"learning_rate": 1.2539184952978058e-05, |
|
"loss": 0.3814, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.2572145545796738, |
|
"grad_norm": 0.3668326981157007, |
|
"learning_rate": 1.285266457680251e-05, |
|
"loss": 0.3769, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.26348808030112925, |
|
"grad_norm": 0.3606059535955065, |
|
"learning_rate": 1.316614420062696e-05, |
|
"loss": 0.3882, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.2697616060225847, |
|
"grad_norm": 0.3743059326798353, |
|
"learning_rate": 1.3479623824451411e-05, |
|
"loss": 0.3792, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.27603513174404015, |
|
"grad_norm": 0.3987264784021991, |
|
"learning_rate": 1.3793103448275863e-05, |
|
"loss": 0.3742, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.2823086574654956, |
|
"grad_norm": 0.41749115581528207, |
|
"learning_rate": 1.4106583072100315e-05, |
|
"loss": 0.3879, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.28858218318695106, |
|
"grad_norm": 0.36990941061813976, |
|
"learning_rate": 1.4420062695924765e-05, |
|
"loss": 0.3763, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.2948557089084065, |
|
"grad_norm": 0.350857226534173, |
|
"learning_rate": 1.4733542319749217e-05, |
|
"loss": 0.3936, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.30112923462986196, |
|
"grad_norm": 0.3699010810723354, |
|
"learning_rate": 1.5047021943573669e-05, |
|
"loss": 0.3827, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.3074027603513174, |
|
"grad_norm": 0.383417773018869, |
|
"learning_rate": 1.536050156739812e-05, |
|
"loss": 0.3753, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.3136762860727729, |
|
"grad_norm": 0.37418028131825143, |
|
"learning_rate": 1.567398119122257e-05, |
|
"loss": 0.393, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.3199498117942284, |
|
"grad_norm": 0.375158938790808, |
|
"learning_rate": 1.598746081504702e-05, |
|
"loss": 0.3803, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.32622333751568383, |
|
"grad_norm": 0.35545430825067814, |
|
"learning_rate": 1.6300940438871475e-05, |
|
"loss": 0.3858, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.3324968632371393, |
|
"grad_norm": 0.39047195961342007, |
|
"learning_rate": 1.6614420062695925e-05, |
|
"loss": 0.3956, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.33877038895859474, |
|
"grad_norm": 0.3763861614570858, |
|
"learning_rate": 1.6927899686520378e-05, |
|
"loss": 0.3863, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.3450439146800502, |
|
"grad_norm": 0.33810866516589266, |
|
"learning_rate": 1.7241379310344828e-05, |
|
"loss": 0.3861, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.35131744040150564, |
|
"grad_norm": 0.3577598772376036, |
|
"learning_rate": 1.7554858934169282e-05, |
|
"loss": 0.3847, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.3575909661229611, |
|
"grad_norm": 0.39952196485063435, |
|
"learning_rate": 1.7868338557993732e-05, |
|
"loss": 0.3803, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.36386449184441655, |
|
"grad_norm": 0.3560924326294842, |
|
"learning_rate": 1.8181818181818182e-05, |
|
"loss": 0.3823, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.370138017565872, |
|
"grad_norm": 0.36592070219456535, |
|
"learning_rate": 1.8495297805642636e-05, |
|
"loss": 0.3877, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.37641154328732745, |
|
"grad_norm": 0.36996538529023604, |
|
"learning_rate": 1.8808777429467086e-05, |
|
"loss": 0.3816, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.38268506900878296, |
|
"grad_norm": 0.3944353625018826, |
|
"learning_rate": 1.9122257053291536e-05, |
|
"loss": 0.3885, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.3889585947302384, |
|
"grad_norm": 0.3605090389245764, |
|
"learning_rate": 1.943573667711599e-05, |
|
"loss": 0.3746, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.39523212045169387, |
|
"grad_norm": 0.3489216159243111, |
|
"learning_rate": 1.9749216300940443e-05, |
|
"loss": 0.3807, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.4015056461731493, |
|
"grad_norm": 0.3902867659960602, |
|
"learning_rate": 1.9999994004731887e-05, |
|
"loss": 0.394, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.4077791718946048, |
|
"grad_norm": 0.3833754079678544, |
|
"learning_rate": 1.999978417110275e-05, |
|
"loss": 0.3657, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.41405269761606023, |
|
"grad_norm": 0.37800625764548235, |
|
"learning_rate": 1.9999274581256576e-05, |
|
"loss": 0.3822, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.4203262233375157, |
|
"grad_norm": 0.3548880184290815, |
|
"learning_rate": 1.999846525046898e-05, |
|
"loss": 0.3816, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.42659974905897113, |
|
"grad_norm": 0.37094441921423654, |
|
"learning_rate": 1.9997356203000667e-05, |
|
"loss": 0.3853, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.4328732747804266, |
|
"grad_norm": 0.3594535908155001, |
|
"learning_rate": 1.9995947472096752e-05, |
|
"loss": 0.3703, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.43914680050188204, |
|
"grad_norm": 0.3492383976696482, |
|
"learning_rate": 1.9994239099985727e-05, |
|
"loss": 0.3858, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.4454203262233375, |
|
"grad_norm": 0.3545945901168298, |
|
"learning_rate": 1.9992231137878213e-05, |
|
"loss": 0.3723, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.451693851944793, |
|
"grad_norm": 0.36080708632510694, |
|
"learning_rate": 1.9989923645965418e-05, |
|
"loss": 0.3952, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.45796737766624845, |
|
"grad_norm": 0.5493517294545232, |
|
"learning_rate": 1.998731669341735e-05, |
|
"loss": 0.3723, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.4642409033877039, |
|
"grad_norm": 0.34501637664109114, |
|
"learning_rate": 1.998441035838071e-05, |
|
"loss": 0.3787, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.47051442910915936, |
|
"grad_norm": 0.36834831423716424, |
|
"learning_rate": 1.9981204727976577e-05, |
|
"loss": 0.3871, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.4767879548306148, |
|
"grad_norm": 0.359064185739471, |
|
"learning_rate": 1.9977699898297794e-05, |
|
"loss": 0.4078, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.48306148055207027, |
|
"grad_norm": 0.37676963093518556, |
|
"learning_rate": 1.997389597440608e-05, |
|
"loss": 0.3997, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.4893350062735257, |
|
"grad_norm": 0.35164360932787275, |
|
"learning_rate": 1.9969793070328872e-05, |
|
"loss": 0.3706, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.49560853199498117, |
|
"grad_norm": 0.3499090374786232, |
|
"learning_rate": 1.996539130905593e-05, |
|
"loss": 0.3931, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.5018820577164367, |
|
"grad_norm": 0.35516374008901963, |
|
"learning_rate": 1.9960690822535632e-05, |
|
"loss": 0.3917, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.5081555834378921, |
|
"grad_norm": 0.3859915521260531, |
|
"learning_rate": 1.995569175167102e-05, |
|
"loss": 0.3862, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.5144291091593476, |
|
"grad_norm": 0.3700726542619639, |
|
"learning_rate": 1.9950394246315594e-05, |
|
"loss": 0.3977, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.520702634880803, |
|
"grad_norm": 0.36807219365873434, |
|
"learning_rate": 1.994479846526879e-05, |
|
"loss": 0.391, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.5269761606022585, |
|
"grad_norm": 0.34834015226308745, |
|
"learning_rate": 1.9938904576271247e-05, |
|
"loss": 0.3947, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.533249686323714, |
|
"grad_norm": 0.3475144346489988, |
|
"learning_rate": 1.9932712755999768e-05, |
|
"loss": 0.3797, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.5395232120451694, |
|
"grad_norm": 0.3471973955561176, |
|
"learning_rate": 1.9926223190062015e-05, |
|
"loss": 0.3777, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.5457967377666249, |
|
"grad_norm": 0.3702513132489804, |
|
"learning_rate": 1.9919436072990967e-05, |
|
"loss": 0.4113, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.5520702634880803, |
|
"grad_norm": 0.3480545734033953, |
|
"learning_rate": 1.9912351608239064e-05, |
|
"loss": 0.3717, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.5583437892095358, |
|
"grad_norm": 0.3705010339965962, |
|
"learning_rate": 1.9904970008172128e-05, |
|
"loss": 0.3903, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.5646173149309912, |
|
"grad_norm": 0.3545524157937066, |
|
"learning_rate": 1.989729149406298e-05, |
|
"loss": 0.3971, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.5708908406524467, |
|
"grad_norm": 0.34489507859345525, |
|
"learning_rate": 1.988931629608483e-05, |
|
"loss": 0.3715, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.5771643663739021, |
|
"grad_norm": 0.3419133499625096, |
|
"learning_rate": 1.9881044653304347e-05, |
|
"loss": 0.3722, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.5834378920953576, |
|
"grad_norm": 0.3544487799327636, |
|
"learning_rate": 1.9872476813674527e-05, |
|
"loss": 0.3854, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.589711417816813, |
|
"grad_norm": 0.3526030038053533, |
|
"learning_rate": 1.9863613034027224e-05, |
|
"loss": 0.3854, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.5959849435382685, |
|
"grad_norm": 0.35801287965649553, |
|
"learning_rate": 1.9854453580065485e-05, |
|
"loss": 0.3773, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.6022584692597239, |
|
"grad_norm": 0.33687873507306343, |
|
"learning_rate": 1.984499872635556e-05, |
|
"loss": 0.3839, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.6085319949811794, |
|
"grad_norm": 0.34289605388996236, |
|
"learning_rate": 1.983524875631868e-05, |
|
"loss": 0.3899, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.6148055207026348, |
|
"grad_norm": 0.35193056323984084, |
|
"learning_rate": 1.9825203962222573e-05, |
|
"loss": 0.3736, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.6210790464240903, |
|
"grad_norm": 0.3603036239481184, |
|
"learning_rate": 1.9814864645172684e-05, |
|
"loss": 0.3927, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.6273525721455459, |
|
"grad_norm": 0.3610916261894253, |
|
"learning_rate": 1.9804231115103155e-05, |
|
"loss": 0.3729, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.6336260978670013, |
|
"grad_norm": 0.33695606360007346, |
|
"learning_rate": 1.9793303690767543e-05, |
|
"loss": 0.3773, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.6398996235884568, |
|
"grad_norm": 0.37708359246844253, |
|
"learning_rate": 1.9782082699729255e-05, |
|
"loss": 0.3915, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.6461731493099122, |
|
"grad_norm": 0.37063002060189804, |
|
"learning_rate": 1.9770568478351736e-05, |
|
"loss": 0.384, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.6524466750313677, |
|
"grad_norm": 0.347781428498976, |
|
"learning_rate": 1.9758761371788376e-05, |
|
"loss": 0.3912, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.6587202007528231, |
|
"grad_norm": 0.37037259025304425, |
|
"learning_rate": 1.974666173397218e-05, |
|
"loss": 0.3795, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.6649937264742786, |
|
"grad_norm": 0.35088852677024873, |
|
"learning_rate": 1.9734269927605134e-05, |
|
"loss": 0.3803, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.671267252195734, |
|
"grad_norm": 0.3569377164202312, |
|
"learning_rate": 1.972158632414736e-05, |
|
"loss": 0.3775, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.6775407779171895, |
|
"grad_norm": 0.3394929535259636, |
|
"learning_rate": 1.970861130380596e-05, |
|
"loss": 0.3741, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.6838143036386449, |
|
"grad_norm": 0.3420677685134694, |
|
"learning_rate": 1.9695345255523634e-05, |
|
"loss": 0.3878, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.6900878293601004, |
|
"grad_norm": 0.3531313507208391, |
|
"learning_rate": 1.9681788576967004e-05, |
|
"loss": 0.3787, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.6963613550815558, |
|
"grad_norm": 0.338822289408876, |
|
"learning_rate": 1.9667941674514712e-05, |
|
"loss": 0.3829, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.7026348808030113, |
|
"grad_norm": 0.36931664322135854, |
|
"learning_rate": 1.9653804963245226e-05, |
|
"loss": 0.3893, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.7089084065244667, |
|
"grad_norm": 0.35793185710348313, |
|
"learning_rate": 1.9639378866924405e-05, |
|
"loss": 0.3816, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.7151819322459222, |
|
"grad_norm": 0.36642001426471366, |
|
"learning_rate": 1.9624663817992783e-05, |
|
"loss": 0.3825, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.7214554579673776, |
|
"grad_norm": 0.36184277756969885, |
|
"learning_rate": 1.960966025755262e-05, |
|
"loss": 0.3918, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.7277289836888331, |
|
"grad_norm": 0.34191036136824954, |
|
"learning_rate": 1.9594368635354676e-05, |
|
"loss": 0.3812, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.7340025094102886, |
|
"grad_norm": 0.4053922705018229, |
|
"learning_rate": 1.9578789409784727e-05, |
|
"loss": 0.3796, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.740276035131744, |
|
"grad_norm": 0.3254681140813354, |
|
"learning_rate": 1.9562923047849828e-05, |
|
"loss": 0.381, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.7465495608531995, |
|
"grad_norm": 0.3655079090417533, |
|
"learning_rate": 1.9546770025164304e-05, |
|
"loss": 0.3808, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.7528230865746549, |
|
"grad_norm": 0.3309145310004598, |
|
"learning_rate": 1.95303308259355e-05, |
|
"loss": 0.3924, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.7590966122961104, |
|
"grad_norm": 0.34826624551002194, |
|
"learning_rate": 1.9513605942949277e-05, |
|
"loss": 0.3752, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.7653701380175659, |
|
"grad_norm": 0.3463792668326179, |
|
"learning_rate": 1.9496595877555212e-05, |
|
"loss": 0.3922, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.7716436637390214, |
|
"grad_norm": 0.3409526655488519, |
|
"learning_rate": 1.94793011396516e-05, |
|
"loss": 0.3868, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.7779171894604768, |
|
"grad_norm": 0.34614309257239506, |
|
"learning_rate": 1.946172224767015e-05, |
|
"loss": 0.3816, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.7841907151819323, |
|
"grad_norm": 0.36738342533490437, |
|
"learning_rate": 1.9443859728560458e-05, |
|
"loss": 0.3809, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.7904642409033877, |
|
"grad_norm": 0.5642673777402893, |
|
"learning_rate": 1.9425714117774183e-05, |
|
"loss": 0.3678, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.7967377666248432, |
|
"grad_norm": 0.3734297840619315, |
|
"learning_rate": 1.940728595924904e-05, |
|
"loss": 0.3856, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.8030112923462986, |
|
"grad_norm": 0.34992964702526325, |
|
"learning_rate": 1.9388575805392453e-05, |
|
"loss": 0.3937, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.8092848180677541, |
|
"grad_norm": 0.34532043284186725, |
|
"learning_rate": 1.9369584217065025e-05, |
|
"loss": 0.3718, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.8155583437892095, |
|
"grad_norm": 0.3390467838397306, |
|
"learning_rate": 1.935031176356371e-05, |
|
"loss": 0.3829, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.821831869510665, |
|
"grad_norm": 0.3522889151490647, |
|
"learning_rate": 1.933075902260475e-05, |
|
"loss": 0.3876, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.8281053952321205, |
|
"grad_norm": 0.38123310023732215, |
|
"learning_rate": 1.9310926580306365e-05, |
|
"loss": 0.3707, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.8343789209535759, |
|
"grad_norm": 0.37101425687357464, |
|
"learning_rate": 1.929081503117117e-05, |
|
"loss": 0.3777, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.8406524466750314, |
|
"grad_norm": 0.35327596368808917, |
|
"learning_rate": 1.9270424978068368e-05, |
|
"loss": 0.372, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.8469259723964868, |
|
"grad_norm": 0.48025305553666425, |
|
"learning_rate": 1.9249757032215674e-05, |
|
"loss": 0.3719, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.8531994981179423, |
|
"grad_norm": 1.3954653921865106, |
|
"learning_rate": 1.9228811813160972e-05, |
|
"loss": 0.3774, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.8594730238393977, |
|
"grad_norm": 0.3729066926513616, |
|
"learning_rate": 1.920758994876379e-05, |
|
"loss": 0.3845, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.8657465495608532, |
|
"grad_norm": 0.3368507345057043, |
|
"learning_rate": 1.918609207517643e-05, |
|
"loss": 0.3674, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.8720200752823086, |
|
"grad_norm": 0.3381829661628343, |
|
"learning_rate": 1.9164318836824928e-05, |
|
"loss": 0.3895, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.8782936010037641, |
|
"grad_norm": 0.35406742056379137, |
|
"learning_rate": 1.9142270886389726e-05, |
|
"loss": 0.3888, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.8845671267252195, |
|
"grad_norm": 0.35267605322252205, |
|
"learning_rate": 1.911994888478611e-05, |
|
"loss": 0.416, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.890840652446675, |
|
"grad_norm": 0.36994927588198245, |
|
"learning_rate": 1.9097353501144403e-05, |
|
"loss": 0.3881, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.8971141781681304, |
|
"grad_norm": 0.344878616221491, |
|
"learning_rate": 1.9074485412789886e-05, |
|
"loss": 0.3916, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.903387703889586, |
|
"grad_norm": 0.3618364967642679, |
|
"learning_rate": 1.9051345305222527e-05, |
|
"loss": 0.3877, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.9096612296110415, |
|
"grad_norm": 0.3348128268620291, |
|
"learning_rate": 1.9027933872096403e-05, |
|
"loss": 0.3758, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.9159347553324969, |
|
"grad_norm": 0.34960639615133043, |
|
"learning_rate": 1.900425181519893e-05, |
|
"loss": 0.3842, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.9222082810539524, |
|
"grad_norm": 0.36222991570969465, |
|
"learning_rate": 1.8980299844429804e-05, |
|
"loss": 0.379, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.9284818067754078, |
|
"grad_norm": 0.3499002387438987, |
|
"learning_rate": 1.8956078677779738e-05, |
|
"loss": 0.3715, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.9347553324968633, |
|
"grad_norm": 0.33211819623666405, |
|
"learning_rate": 1.8931589041308926e-05, |
|
"loss": 0.3736, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.9410288582183187, |
|
"grad_norm": 0.3495894811325005, |
|
"learning_rate": 1.8906831669125293e-05, |
|
"loss": 0.3778, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.9473023839397742, |
|
"grad_norm": 0.3315684966056081, |
|
"learning_rate": 1.8881807303362484e-05, |
|
"loss": 0.3776, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.9535759096612296, |
|
"grad_norm": 0.3409622621170755, |
|
"learning_rate": 1.885651669415761e-05, |
|
"loss": 0.3873, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.9598494353826851, |
|
"grad_norm": 0.34785709596472625, |
|
"learning_rate": 1.883096059962876e-05, |
|
"loss": 0.3861, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.9661229611041405, |
|
"grad_norm": 0.3477615933807258, |
|
"learning_rate": 1.8805139785852297e-05, |
|
"loss": 0.3783, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.972396486825596, |
|
"grad_norm": 0.343324271874741, |
|
"learning_rate": 1.877905502683987e-05, |
|
"loss": 0.3746, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.9786700125470514, |
|
"grad_norm": 0.47223667640674677, |
|
"learning_rate": 1.8752707104515223e-05, |
|
"loss": 0.3793, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.9849435382685069, |
|
"grad_norm": 0.3234411234996811, |
|
"learning_rate": 1.8726096808690757e-05, |
|
"loss": 0.3676, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.9912170639899623, |
|
"grad_norm": 0.3574211042049705, |
|
"learning_rate": 1.8699224937043846e-05, |
|
"loss": 0.3735, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.9974905897114178, |
|
"grad_norm": 0.3811194221638418, |
|
"learning_rate": 1.8672092295092935e-05, |
|
"loss": 0.3939, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.3792824149131775, |
|
"eval_runtime": 2.8922, |
|
"eval_samples_per_second": 12.101, |
|
"eval_steps_per_second": 0.692, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 1.0037641154328734, |
|
"grad_norm": 0.3353538294927158, |
|
"learning_rate": 1.8644699696173393e-05, |
|
"loss": 0.3241, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.0100376411543288, |
|
"grad_norm": 0.3828496951284579, |
|
"learning_rate": 1.8617047961413122e-05, |
|
"loss": 0.2601, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 1.0163111668757843, |
|
"grad_norm": 0.33996880676819174, |
|
"learning_rate": 1.858913791970795e-05, |
|
"loss": 0.2689, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.0225846925972397, |
|
"grad_norm": 0.3787602779680391, |
|
"learning_rate": 1.8560970407696787e-05, |
|
"loss": 0.2686, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 1.0288582183186952, |
|
"grad_norm": 0.5196175862012924, |
|
"learning_rate": 1.8532546269736546e-05, |
|
"loss": 0.2747, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.0351317440401506, |
|
"grad_norm": 0.37332309268508856, |
|
"learning_rate": 1.850386635787682e-05, |
|
"loss": 0.2627, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 1.041405269761606, |
|
"grad_norm": 0.3471559800133524, |
|
"learning_rate": 1.847493153183435e-05, |
|
"loss": 0.2787, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.0476787954830615, |
|
"grad_norm": 0.3558613761958236, |
|
"learning_rate": 1.844574265896726e-05, |
|
"loss": 0.268, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 1.053952321204517, |
|
"grad_norm": 0.3816351874674652, |
|
"learning_rate": 1.8416300614249044e-05, |
|
"loss": 0.2668, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.0602258469259724, |
|
"grad_norm": 0.34373084787867114, |
|
"learning_rate": 1.8386606280242342e-05, |
|
"loss": 0.2648, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 1.066499372647428, |
|
"grad_norm": 0.34772061503647006, |
|
"learning_rate": 1.8356660547072493e-05, |
|
"loss": 0.2664, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.0727728983688833, |
|
"grad_norm": 0.3575051030279196, |
|
"learning_rate": 1.8326464312400835e-05, |
|
"loss": 0.2741, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 1.0790464240903388, |
|
"grad_norm": 0.3496509974703566, |
|
"learning_rate": 1.8296018481397818e-05, |
|
"loss": 0.2583, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.0853199498117942, |
|
"grad_norm": 0.3724719074552806, |
|
"learning_rate": 1.826532396671585e-05, |
|
"loss": 0.2694, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 1.0915934755332497, |
|
"grad_norm": 0.3727323786558387, |
|
"learning_rate": 1.8234381688461943e-05, |
|
"loss": 0.2772, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.0978670012547052, |
|
"grad_norm": 0.3335956714818362, |
|
"learning_rate": 1.8203192574170154e-05, |
|
"loss": 0.2563, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 1.1041405269761606, |
|
"grad_norm": 0.34610791981119976, |
|
"learning_rate": 1.8171757558773747e-05, |
|
"loss": 0.254, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.110414052697616, |
|
"grad_norm": 0.3647153604437883, |
|
"learning_rate": 1.8140077584577193e-05, |
|
"loss": 0.269, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 1.1166875784190715, |
|
"grad_norm": 0.3620894127576232, |
|
"learning_rate": 1.81081536012279e-05, |
|
"loss": 0.2642, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.122961104140527, |
|
"grad_norm": 0.3368502269678017, |
|
"learning_rate": 1.8075986565687785e-05, |
|
"loss": 0.2621, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 1.1292346298619824, |
|
"grad_norm": 0.3487017808626127, |
|
"learning_rate": 1.804357744220454e-05, |
|
"loss": 0.2741, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.1355081555834379, |
|
"grad_norm": 0.346569504128483, |
|
"learning_rate": 1.8010927202282758e-05, |
|
"loss": 0.2522, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 1.1417816813048933, |
|
"grad_norm": 0.3448349298442695, |
|
"learning_rate": 1.7978036824654806e-05, |
|
"loss": 0.2539, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.1480552070263488, |
|
"grad_norm": 0.35055993507582917, |
|
"learning_rate": 1.7944907295251478e-05, |
|
"loss": 0.2716, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 1.1543287327478042, |
|
"grad_norm": 0.3483298654120021, |
|
"learning_rate": 1.7911539607172447e-05, |
|
"loss": 0.2585, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.1606022584692597, |
|
"grad_norm": 0.3452159026039351, |
|
"learning_rate": 1.78779347606565e-05, |
|
"loss": 0.2598, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 1.1668757841907151, |
|
"grad_norm": 0.3434243523852655, |
|
"learning_rate": 1.7844093763051543e-05, |
|
"loss": 0.2681, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.1731493099121706, |
|
"grad_norm": 0.3459951917334934, |
|
"learning_rate": 1.7810017628784416e-05, |
|
"loss": 0.2567, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 1.179422835633626, |
|
"grad_norm": 0.31955125292674175, |
|
"learning_rate": 1.777570737933047e-05, |
|
"loss": 0.2673, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.1856963613550815, |
|
"grad_norm": 0.33191052346485606, |
|
"learning_rate": 1.7741164043182967e-05, |
|
"loss": 0.258, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 1.191969887076537, |
|
"grad_norm": 0.3745301206766351, |
|
"learning_rate": 1.7706388655822223e-05, |
|
"loss": 0.2671, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.1982434127979924, |
|
"grad_norm": 0.3435343557827377, |
|
"learning_rate": 1.7671382259684603e-05, |
|
"loss": 0.2711, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 1.2045169385194479, |
|
"grad_norm": 0.3522241286327592, |
|
"learning_rate": 1.7636145904131233e-05, |
|
"loss": 0.2715, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.2107904642409033, |
|
"grad_norm": 0.34428731270476376, |
|
"learning_rate": 1.7600680645416583e-05, |
|
"loss": 0.2655, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 1.2170639899623588, |
|
"grad_norm": 0.3534701552438621, |
|
"learning_rate": 1.7564987546656778e-05, |
|
"loss": 0.2601, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.2233375156838142, |
|
"grad_norm": 0.34480580113424486, |
|
"learning_rate": 1.7529067677797727e-05, |
|
"loss": 0.2581, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 1.2296110414052697, |
|
"grad_norm": 0.3741756431765012, |
|
"learning_rate": 1.7492922115583077e-05, |
|
"loss": 0.2701, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.2358845671267251, |
|
"grad_norm": 0.3555549239846533, |
|
"learning_rate": 1.745655194352191e-05, |
|
"loss": 0.2716, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 1.2421580928481806, |
|
"grad_norm": 0.3327711036535926, |
|
"learning_rate": 1.7419958251856276e-05, |
|
"loss": 0.2577, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.248431618569636, |
|
"grad_norm": 0.3661107928778811, |
|
"learning_rate": 1.738314213752851e-05, |
|
"loss": 0.2649, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 1.2547051442910915, |
|
"grad_norm": 0.3407939311803759, |
|
"learning_rate": 1.7346104704148343e-05, |
|
"loss": 0.2506, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.260978670012547, |
|
"grad_norm": 0.33612826697533044, |
|
"learning_rate": 1.730884706195983e-05, |
|
"loss": 0.2645, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 1.2672521957340024, |
|
"grad_norm": 0.34109027432250294, |
|
"learning_rate": 1.727137032780807e-05, |
|
"loss": 0.2687, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.2735257214554578, |
|
"grad_norm": 0.3424345987544216, |
|
"learning_rate": 1.7233675625105703e-05, |
|
"loss": 0.2659, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 1.2797992471769133, |
|
"grad_norm": 0.3670304260632612, |
|
"learning_rate": 1.7195764083799277e-05, |
|
"loss": 0.2785, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.286072772898369, |
|
"grad_norm": 0.35224545856472056, |
|
"learning_rate": 1.7157636840335334e-05, |
|
"loss": 0.2736, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 1.2923462986198244, |
|
"grad_norm": 0.8557471051222927, |
|
"learning_rate": 1.7119295037626366e-05, |
|
"loss": 0.2598, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.2986198243412799, |
|
"grad_norm": 0.3367737302829996, |
|
"learning_rate": 1.708073982501656e-05, |
|
"loss": 0.2612, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 1.3048933500627353, |
|
"grad_norm": 0.34877985553107826, |
|
"learning_rate": 1.704197235824732e-05, |
|
"loss": 0.2726, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.3111668757841908, |
|
"grad_norm": 0.34440291584591926, |
|
"learning_rate": 1.7002993799422652e-05, |
|
"loss": 0.2618, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 1.3174404015056462, |
|
"grad_norm": 0.33534085525712676, |
|
"learning_rate": 1.6963805316974303e-05, |
|
"loss": 0.2609, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.3237139272271017, |
|
"grad_norm": 0.3662577678235449, |
|
"learning_rate": 1.6924408085626756e-05, |
|
"loss": 0.2571, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 1.3299874529485571, |
|
"grad_norm": 0.3516887590807691, |
|
"learning_rate": 1.6884803286362e-05, |
|
"loss": 0.2549, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.3362609786700126, |
|
"grad_norm": 0.35333946604034366, |
|
"learning_rate": 1.684499210638414e-05, |
|
"loss": 0.264, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 1.342534504391468, |
|
"grad_norm": 0.34307627027241056, |
|
"learning_rate": 1.6804975739083803e-05, |
|
"loss": 0.2503, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.3488080301129235, |
|
"grad_norm": 0.35493913250542247, |
|
"learning_rate": 1.6764755384002372e-05, |
|
"loss": 0.2759, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 1.355081555834379, |
|
"grad_norm": 0.3540415537021871, |
|
"learning_rate": 1.6724332246796008e-05, |
|
"loss": 0.2697, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.3613550815558344, |
|
"grad_norm": 0.3318252959324338, |
|
"learning_rate": 1.6683707539199538e-05, |
|
"loss": 0.2669, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 1.3676286072772899, |
|
"grad_norm": 0.33137640825273385, |
|
"learning_rate": 1.6642882478990112e-05, |
|
"loss": 0.2485, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.3739021329987453, |
|
"grad_norm": 0.34017928191383223, |
|
"learning_rate": 1.66018582899507e-05, |
|
"loss": 0.2784, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 1.3801756587202008, |
|
"grad_norm": 0.34200366455572445, |
|
"learning_rate": 1.6560636201833423e-05, |
|
"loss": 0.2673, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.3864491844416562, |
|
"grad_norm": 0.35384418379159516, |
|
"learning_rate": 1.6519217450322657e-05, |
|
"loss": 0.2713, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 1.3927227101631117, |
|
"grad_norm": 0.33307624664228463, |
|
"learning_rate": 1.6477603276998037e-05, |
|
"loss": 0.2742, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.3989962358845671, |
|
"grad_norm": 0.356765357000532, |
|
"learning_rate": 1.64357949292972e-05, |
|
"loss": 0.2689, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 1.4052697616060226, |
|
"grad_norm": 0.3252207580977864, |
|
"learning_rate": 1.6393793660478406e-05, |
|
"loss": 0.2506, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.411543287327478, |
|
"grad_norm": 0.36204092721369197, |
|
"learning_rate": 1.6351600729582977e-05, |
|
"loss": 0.2636, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 1.4178168130489335, |
|
"grad_norm": 0.3335102584738542, |
|
"learning_rate": 1.630921740139755e-05, |
|
"loss": 0.2616, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.424090338770389, |
|
"grad_norm": 0.34898006017841243, |
|
"learning_rate": 1.6266644946416148e-05, |
|
"loss": 0.2781, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 1.4303638644918444, |
|
"grad_norm": 0.3487474238464629, |
|
"learning_rate": 1.622388464080213e-05, |
|
"loss": 0.2773, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.4366373902132998, |
|
"grad_norm": 0.35273437538491903, |
|
"learning_rate": 1.61809377663499e-05, |
|
"loss": 0.2682, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 1.4429109159347553, |
|
"grad_norm": 0.36466350484277693, |
|
"learning_rate": 1.6137805610446508e-05, |
|
"loss": 0.2685, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.4491844416562107, |
|
"grad_norm": 0.35090011471222154, |
|
"learning_rate": 1.609448946603304e-05, |
|
"loss": 0.2657, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 1.4554579673776662, |
|
"grad_norm": 0.36713922458350784, |
|
"learning_rate": 1.6050990631565894e-05, |
|
"loss": 0.276, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.4617314930991216, |
|
"grad_norm": 0.3552048435134842, |
|
"learning_rate": 1.6007310410977807e-05, |
|
"loss": 0.2796, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 1.468005018820577, |
|
"grad_norm": 0.3469345931045424, |
|
"learning_rate": 1.5963450113638815e-05, |
|
"loss": 0.2592, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.4742785445420326, |
|
"grad_norm": 2.7973818683619016, |
|
"learning_rate": 1.5919411054316966e-05, |
|
"loss": 0.2727, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 1.480552070263488, |
|
"grad_norm": 0.3477691166092876, |
|
"learning_rate": 1.5875194553138942e-05, |
|
"loss": 0.2708, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.4868255959849435, |
|
"grad_norm": 0.42047181326852134, |
|
"learning_rate": 1.5830801935550462e-05, |
|
"loss": 0.2583, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 1.4930991217063991, |
|
"grad_norm": 0.38351145606864295, |
|
"learning_rate": 1.5786234532276555e-05, |
|
"loss": 0.2665, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.4993726474278546, |
|
"grad_norm": 0.33888116221517656, |
|
"learning_rate": 1.574149367928168e-05, |
|
"loss": 0.2788, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 1.50564617314931, |
|
"grad_norm": 0.3533659596857954, |
|
"learning_rate": 1.5696580717729665e-05, |
|
"loss": 0.2709, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.5119196988707655, |
|
"grad_norm": 0.33797242592368726, |
|
"learning_rate": 1.5651496993943507e-05, |
|
"loss": 0.2552, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 1.518193224592221, |
|
"grad_norm": 0.3380805624100092, |
|
"learning_rate": 1.5606243859365033e-05, |
|
"loss": 0.2696, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.5244667503136764, |
|
"grad_norm": 0.3589011909382504, |
|
"learning_rate": 1.5560822670514356e-05, |
|
"loss": 0.2729, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 1.5307402760351319, |
|
"grad_norm": 0.3410908318740735, |
|
"learning_rate": 1.5515234788949238e-05, |
|
"loss": 0.2658, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.5370138017565873, |
|
"grad_norm": 0.4071357427518539, |
|
"learning_rate": 1.5469481581224274e-05, |
|
"loss": 0.2618, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.5432873274780428, |
|
"grad_norm": 0.4050327080992723, |
|
"learning_rate": 1.5423564418849895e-05, |
|
"loss": 0.2707, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.5495608531994982, |
|
"grad_norm": 0.3512739607697075, |
|
"learning_rate": 1.537748467825131e-05, |
|
"loss": 0.2762, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 1.5558343789209537, |
|
"grad_norm": 0.3548946907435895, |
|
"learning_rate": 1.5331243740727203e-05, |
|
"loss": 0.285, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.5621079046424091, |
|
"grad_norm": 0.3241833124220892, |
|
"learning_rate": 1.5284842992408336e-05, |
|
"loss": 0.2675, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 1.5683814303638646, |
|
"grad_norm": 0.35085102130243395, |
|
"learning_rate": 1.5238283824216015e-05, |
|
"loss": 0.2681, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.57465495608532, |
|
"grad_norm": 0.37148590291911643, |
|
"learning_rate": 1.5191567631820364e-05, |
|
"loss": 0.27, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 1.5809284818067755, |
|
"grad_norm": 0.3535604634586656, |
|
"learning_rate": 1.5144695815598529e-05, |
|
"loss": 0.2717, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.587202007528231, |
|
"grad_norm": 0.34794195830922975, |
|
"learning_rate": 1.5097669780592658e-05, |
|
"loss": 0.2633, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 1.5934755332496864, |
|
"grad_norm": 0.35489225458342305, |
|
"learning_rate": 1.5050490936467814e-05, |
|
"loss": 0.2735, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.5997490589711418, |
|
"grad_norm": 0.3350637434620735, |
|
"learning_rate": 1.5003160697469707e-05, |
|
"loss": 0.2544, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.6060225846925973, |
|
"grad_norm": 0.3381089856973841, |
|
"learning_rate": 1.4955680482382296e-05, |
|
"loss": 0.2564, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.6122961104140527, |
|
"grad_norm": 0.32684634964616277, |
|
"learning_rate": 1.4908051714485266e-05, |
|
"loss": 0.2741, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 1.6185696361355082, |
|
"grad_norm": 0.34645698891886967, |
|
"learning_rate": 1.4860275821511359e-05, |
|
"loss": 0.2748, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.6248431618569636, |
|
"grad_norm": 0.36961906780329834, |
|
"learning_rate": 1.481235423560358e-05, |
|
"loss": 0.2721, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 1.631116687578419, |
|
"grad_norm": 0.350579115602874, |
|
"learning_rate": 1.4764288393272258e-05, |
|
"loss": 0.2628, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.6373902132998746, |
|
"grad_norm": 0.3306477114399784, |
|
"learning_rate": 1.4716079735352006e-05, |
|
"loss": 0.2729, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 1.64366373902133, |
|
"grad_norm": 0.35455700175873195, |
|
"learning_rate": 1.46677297069585e-05, |
|
"loss": 0.2667, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.6499372647427855, |
|
"grad_norm": 0.33847253281006606, |
|
"learning_rate": 1.4619239757445187e-05, |
|
"loss": 0.2706, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 1.656210790464241, |
|
"grad_norm": 0.34327567130216446, |
|
"learning_rate": 1.4570611340359821e-05, |
|
"loss": 0.266, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.6624843161856964, |
|
"grad_norm": 0.3557992340297897, |
|
"learning_rate": 1.4521845913400891e-05, |
|
"loss": 0.2746, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.6687578419071518, |
|
"grad_norm": 0.35121604621554686, |
|
"learning_rate": 1.4472944938373945e-05, |
|
"loss": 0.2704, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.6750313676286073, |
|
"grad_norm": 0.34165359487510566, |
|
"learning_rate": 1.4423909881147747e-05, |
|
"loss": 0.2692, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 1.6813048933500627, |
|
"grad_norm": 0.36505578592627197, |
|
"learning_rate": 1.4374742211610345e-05, |
|
"loss": 0.2662, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.6875784190715182, |
|
"grad_norm": 0.34422086687232467, |
|
"learning_rate": 1.4325443403625012e-05, |
|
"loss": 0.275, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 1.6938519447929736, |
|
"grad_norm": 0.3675735392039838, |
|
"learning_rate": 1.4276014934986064e-05, |
|
"loss": 0.272, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.700125470514429, |
|
"grad_norm": 0.35389447351847136, |
|
"learning_rate": 1.4226458287374555e-05, |
|
"loss": 0.2713, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 1.7063989962358845, |
|
"grad_norm": 0.3379229270723559, |
|
"learning_rate": 1.4176774946313872e-05, |
|
"loss": 0.2625, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.71267252195734, |
|
"grad_norm": 0.34402002879314064, |
|
"learning_rate": 1.4126966401125189e-05, |
|
"loss": 0.268, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 1.7189460476787954, |
|
"grad_norm": 0.3532344899080162, |
|
"learning_rate": 1.4077034144882843e-05, |
|
"loss": 0.2632, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.725219573400251, |
|
"grad_norm": 0.3401295622140909, |
|
"learning_rate": 1.4026979674369566e-05, |
|
"loss": 0.2613, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.7314930991217063, |
|
"grad_norm": 0.3391840532185442, |
|
"learning_rate": 1.3976804490031608e-05, |
|
"loss": 0.2719, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.7377666248431618, |
|
"grad_norm": 0.3427490028776178, |
|
"learning_rate": 1.3926510095933781e-05, |
|
"loss": 0.2692, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 1.7440401505646173, |
|
"grad_norm": 0.3433287489294571, |
|
"learning_rate": 1.387609799971435e-05, |
|
"loss": 0.2649, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.7503136762860727, |
|
"grad_norm": 0.33114909580993174, |
|
"learning_rate": 1.3825569712539864e-05, |
|
"loss": 0.2527, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 1.7565872020075282, |
|
"grad_norm": 0.337515925683474, |
|
"learning_rate": 1.3774926749059826e-05, |
|
"loss": 0.2556, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.7628607277289836, |
|
"grad_norm": 0.3419996547297016, |
|
"learning_rate": 1.3724170627361323e-05, |
|
"loss": 0.2638, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 1.769134253450439, |
|
"grad_norm": 0.35073376743994084, |
|
"learning_rate": 1.3673302868923491e-05, |
|
"loss": 0.2704, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.7754077791718945, |
|
"grad_norm": 0.32861646593191174, |
|
"learning_rate": 1.3622324998571928e-05, |
|
"loss": 0.2519, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 1.78168130489335, |
|
"grad_norm": 0.33312714617584277, |
|
"learning_rate": 1.3571238544432968e-05, |
|
"loss": 0.2664, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.7879548306148054, |
|
"grad_norm": 0.3374614354621205, |
|
"learning_rate": 1.352004503788789e-05, |
|
"loss": 0.2585, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.7942283563362609, |
|
"grad_norm": 0.3554543246581463, |
|
"learning_rate": 1.3468746013527e-05, |
|
"loss": 0.2762, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.8005018820577163, |
|
"grad_norm": 0.3367124035830617, |
|
"learning_rate": 1.3417343009103634e-05, |
|
"loss": 0.261, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 1.8067754077791718, |
|
"grad_norm": 0.3181211071389625, |
|
"learning_rate": 1.3365837565488065e-05, |
|
"loss": 0.2715, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.8130489335006272, |
|
"grad_norm": 0.3480040032313692, |
|
"learning_rate": 1.3314231226621305e-05, |
|
"loss": 0.2624, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 1.8193224592220827, |
|
"grad_norm": 0.3416171971995866, |
|
"learning_rate": 1.3262525539468839e-05, |
|
"loss": 0.2642, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.8255959849435381, |
|
"grad_norm": 0.3635819899278629, |
|
"learning_rate": 1.3210722053974233e-05, |
|
"loss": 0.2632, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 1.8318695106649936, |
|
"grad_norm": 0.3265652400172599, |
|
"learning_rate": 1.315882232301269e-05, |
|
"loss": 0.2612, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.838143036386449, |
|
"grad_norm": 0.3342794882432052, |
|
"learning_rate": 1.3106827902344485e-05, |
|
"loss": 0.2623, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 1.8444165621079045, |
|
"grad_norm": 0.3347520583742969, |
|
"learning_rate": 1.3054740350568346e-05, |
|
"loss": 0.2741, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.85069008782936, |
|
"grad_norm": 0.34328068285452285, |
|
"learning_rate": 1.3002561229074719e-05, |
|
"loss": 0.2561, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.8569636135508154, |
|
"grad_norm": 0.35131941587904497, |
|
"learning_rate": 1.2950292101998967e-05, |
|
"loss": 0.2747, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.8632371392722709, |
|
"grad_norm": 0.34360731854818805, |
|
"learning_rate": 1.289793453617449e-05, |
|
"loss": 0.2627, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 1.8695106649937263, |
|
"grad_norm": 0.3498923646707763, |
|
"learning_rate": 1.2845490101085744e-05, |
|
"loss": 0.2562, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.875784190715182, |
|
"grad_norm": 0.34701974294822086, |
|
"learning_rate": 1.2792960368821212e-05, |
|
"loss": 0.265, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 1.8820577164366374, |
|
"grad_norm": 0.33841922800891855, |
|
"learning_rate": 1.2740346914026258e-05, |
|
"loss": 0.2638, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.888331242158093, |
|
"grad_norm": 0.3375511420369947, |
|
"learning_rate": 1.2687651313855937e-05, |
|
"loss": 0.2589, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 1.8946047678795483, |
|
"grad_norm": 0.34124342881268466, |
|
"learning_rate": 1.2634875147927726e-05, |
|
"loss": 0.2689, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.9008782936010038, |
|
"grad_norm": 0.3492720225961315, |
|
"learning_rate": 1.2582019998274142e-05, |
|
"loss": 0.2619, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 1.9071518193224593, |
|
"grad_norm": 0.3424592345393382, |
|
"learning_rate": 1.252908744929536e-05, |
|
"loss": 0.2673, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.9134253450439147, |
|
"grad_norm": 0.3541786605023589, |
|
"learning_rate": 1.2476079087711695e-05, |
|
"loss": 0.2741, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.9196988707653702, |
|
"grad_norm": 0.36100779817450435, |
|
"learning_rate": 1.2422996502516023e-05, |
|
"loss": 0.2708, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.9259723964868256, |
|
"grad_norm": 0.3637900870051361, |
|
"learning_rate": 1.236984128492619e-05, |
|
"loss": 0.2679, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 1.932245922208281, |
|
"grad_norm": 0.34654925833715405, |
|
"learning_rate": 1.231661502833728e-05, |
|
"loss": 0.2705, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.9385194479297365, |
|
"grad_norm": 0.3555711761052598, |
|
"learning_rate": 1.2263319328273853e-05, |
|
"loss": 0.2732, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 1.944792973651192, |
|
"grad_norm": 0.4361336719233523, |
|
"learning_rate": 1.220995578234214e-05, |
|
"loss": 0.2818, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.9510664993726474, |
|
"grad_norm": 0.35818262065748885, |
|
"learning_rate": 1.2156525990182132e-05, |
|
"loss": 0.2714, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 1.9573400250941029, |
|
"grad_norm": 0.34020072227504516, |
|
"learning_rate": 1.2103031553419629e-05, |
|
"loss": 0.2561, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.9636135508155583, |
|
"grad_norm": 0.35424378525712236, |
|
"learning_rate": 1.2049474075618244e-05, |
|
"loss": 0.2817, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 1.9698870765370138, |
|
"grad_norm": 0.3436811391936569, |
|
"learning_rate": 1.1995855162231323e-05, |
|
"loss": 0.2727, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.9761606022584692, |
|
"grad_norm": 0.35547373104319596, |
|
"learning_rate": 1.1942176420553817e-05, |
|
"loss": 0.279, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.9824341279799247, |
|
"grad_norm": 0.33151574599317196, |
|
"learning_rate": 1.1888439459674107e-05, |
|
"loss": 0.2736, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.9887076537013801, |
|
"grad_norm": 0.3489287650284772, |
|
"learning_rate": 1.1834645890425773e-05, |
|
"loss": 0.2674, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 1.9949811794228356, |
|
"grad_norm": 0.4674786217571983, |
|
"learning_rate": 1.1780797325339301e-05, |
|
"loss": 0.2618, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.3876406252384186, |
|
"eval_runtime": 2.3724, |
|
"eval_samples_per_second": 14.753, |
|
"eval_steps_per_second": 0.843, |
|
"step": 1594 |
|
}, |
|
{ |
|
"epoch": 2.0012547051442913, |
|
"grad_norm": 0.41826837808172157, |
|
"learning_rate": 1.1726895378593745e-05, |
|
"loss": 0.2453, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 2.0075282308657467, |
|
"grad_norm": 0.3896113439351613, |
|
"learning_rate": 1.167294166596834e-05, |
|
"loss": 0.1307, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.013801756587202, |
|
"grad_norm": 0.40723627183822325, |
|
"learning_rate": 1.1618937804794077e-05, |
|
"loss": 0.1253, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 2.0200752823086576, |
|
"grad_norm": 0.3393249627107008, |
|
"learning_rate": 1.1564885413905205e-05, |
|
"loss": 0.1212, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.026348808030113, |
|
"grad_norm": 0.3969867667512431, |
|
"learning_rate": 1.1510786113590715e-05, |
|
"loss": 0.1213, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 2.0326223337515685, |
|
"grad_norm": 0.3807098036418188, |
|
"learning_rate": 1.1456641525545768e-05, |
|
"loss": 0.115, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.038895859473024, |
|
"grad_norm": 0.3600283222530161, |
|
"learning_rate": 1.1402453272823086e-05, |
|
"loss": 0.1178, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 2.0451693851944794, |
|
"grad_norm": 0.37082981681871713, |
|
"learning_rate": 1.1348222979784289e-05, |
|
"loss": 0.1186, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.051442910915935, |
|
"grad_norm": 0.3701722986939684, |
|
"learning_rate": 1.1293952272051217e-05, |
|
"loss": 0.1161, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 2.0577164366373903, |
|
"grad_norm": 0.39437640184582917, |
|
"learning_rate": 1.1239642776457176e-05, |
|
"loss": 0.112, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.063989962358846, |
|
"grad_norm": 0.35512823472089206, |
|
"learning_rate": 1.1185296120998208e-05, |
|
"loss": 0.1227, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 2.0702634880803013, |
|
"grad_norm": 0.39699134768151145, |
|
"learning_rate": 1.1130913934784255e-05, |
|
"loss": 0.118, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.0765370138017567, |
|
"grad_norm": 0.361766879756225, |
|
"learning_rate": 1.107649784799034e-05, |
|
"loss": 0.1148, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 2.082810539523212, |
|
"grad_norm": 0.37983521046428353, |
|
"learning_rate": 1.1022049491807703e-05, |
|
"loss": 0.1105, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.0890840652446676, |
|
"grad_norm": 0.36791365726333974, |
|
"learning_rate": 1.0967570498394895e-05, |
|
"loss": 0.1197, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 2.095357590966123, |
|
"grad_norm": 0.38778258880907535, |
|
"learning_rate": 1.0913062500828865e-05, |
|
"loss": 0.119, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.1016311166875785, |
|
"grad_norm": 0.3686039497467697, |
|
"learning_rate": 1.0858527133055994e-05, |
|
"loss": 0.1197, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 2.107904642409034, |
|
"grad_norm": 0.39330229406582323, |
|
"learning_rate": 1.0803966029843114e-05, |
|
"loss": 0.1166, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.1141781681304894, |
|
"grad_norm": 0.3636181831711105, |
|
"learning_rate": 1.0749380826728513e-05, |
|
"loss": 0.1133, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 2.120451693851945, |
|
"grad_norm": 0.4335397800674325, |
|
"learning_rate": 1.0694773159972912e-05, |
|
"loss": 0.1246, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.1267252195734003, |
|
"grad_norm": 0.37508400928061725, |
|
"learning_rate": 1.0640144666510392e-05, |
|
"loss": 0.1196, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 2.132998745294856, |
|
"grad_norm": 0.3987416537308343, |
|
"learning_rate": 1.0585496983899361e-05, |
|
"loss": 0.1226, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.1392722710163112, |
|
"grad_norm": 0.38791702283560353, |
|
"learning_rate": 1.0530831750273428e-05, |
|
"loss": 0.1117, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 2.1455457967377667, |
|
"grad_norm": 0.39592521931999036, |
|
"learning_rate": 1.0476150604292329e-05, |
|
"loss": 0.1198, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.151819322459222, |
|
"grad_norm": 0.6134906074452066, |
|
"learning_rate": 1.0421455185092784e-05, |
|
"loss": 0.1168, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 2.1580928481806776, |
|
"grad_norm": 0.3936394784460519, |
|
"learning_rate": 1.0366747132239374e-05, |
|
"loss": 0.1137, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.164366373902133, |
|
"grad_norm": 0.38023062505112215, |
|
"learning_rate": 1.0312028085675393e-05, |
|
"loss": 0.1216, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 2.1706398996235885, |
|
"grad_norm": 0.3879080250933175, |
|
"learning_rate": 1.025729968567368e-05, |
|
"loss": 0.1163, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.176913425345044, |
|
"grad_norm": 0.37494689918032786, |
|
"learning_rate": 1.0202563572787457e-05, |
|
"loss": 0.1155, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 2.1831869510664994, |
|
"grad_norm": 0.4170219240353852, |
|
"learning_rate": 1.0147821387801154e-05, |
|
"loss": 0.1231, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.189460476787955, |
|
"grad_norm": 0.36447068742427746, |
|
"learning_rate": 1.0093074771681214e-05, |
|
"loss": 0.1173, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 2.1957340025094103, |
|
"grad_norm": 0.37383691981995226, |
|
"learning_rate": 1.003832536552691e-05, |
|
"loss": 0.1181, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.2020075282308658, |
|
"grad_norm": 0.35361007854482546, |
|
"learning_rate": 9.983574810521151e-06, |
|
"loss": 0.1141, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 2.208281053952321, |
|
"grad_norm": 0.37593622722746173, |
|
"learning_rate": 9.928824747881286e-06, |
|
"loss": 0.117, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.2145545796737767, |
|
"grad_norm": 0.38527811994324745, |
|
"learning_rate": 9.874076818809903e-06, |
|
"loss": 0.1222, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 2.220828105395232, |
|
"grad_norm": 0.37756703878021675, |
|
"learning_rate": 9.81933266444563e-06, |
|
"loss": 0.117, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 2.2271016311166876, |
|
"grad_norm": 0.3981056432095895, |
|
"learning_rate": 9.76459392581395e-06, |
|
"loss": 0.1187, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 2.233375156838143, |
|
"grad_norm": 0.35779965724307555, |
|
"learning_rate": 9.709862243777998e-06, |
|
"loss": 0.1201, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 2.2396486825595985, |
|
"grad_norm": 0.39287403460106407, |
|
"learning_rate": 9.655139258989379e-06, |
|
"loss": 0.1173, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 2.245922208281054, |
|
"grad_norm": 0.39081524580807464, |
|
"learning_rate": 9.60042661183899e-06, |
|
"loss": 0.114, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 2.2521957340025094, |
|
"grad_norm": 0.5348108468458116, |
|
"learning_rate": 9.54572594240784e-06, |
|
"loss": 0.1145, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 2.258469259723965, |
|
"grad_norm": 0.35615572604956347, |
|
"learning_rate": 9.491038890417894e-06, |
|
"loss": 0.1128, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.2647427854454203, |
|
"grad_norm": 0.37107476709616843, |
|
"learning_rate": 9.436367095182916e-06, |
|
"loss": 0.1228, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 2.2710163111668757, |
|
"grad_norm": 0.38323250415960275, |
|
"learning_rate": 9.381712195559324e-06, |
|
"loss": 0.118, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 2.277289836888331, |
|
"grad_norm": 0.359171122780413, |
|
"learning_rate": 9.327075829897082e-06, |
|
"loss": 0.1191, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 2.2835633626097867, |
|
"grad_norm": 0.4114654516418914, |
|
"learning_rate": 9.272459635990563e-06, |
|
"loss": 0.1235, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 2.289836888331242, |
|
"grad_norm": 0.39460304180293915, |
|
"learning_rate": 9.217865251029469e-06, |
|
"loss": 0.1187, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 2.2961104140526976, |
|
"grad_norm": 0.35839055639361983, |
|
"learning_rate": 9.163294311549753e-06, |
|
"loss": 0.1156, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 2.302383939774153, |
|
"grad_norm": 0.3482054803738314, |
|
"learning_rate": 9.108748453384559e-06, |
|
"loss": 0.1198, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 2.3086574654956085, |
|
"grad_norm": 0.3820940219983755, |
|
"learning_rate": 9.054229311615178e-06, |
|
"loss": 0.117, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.314930991217064, |
|
"grad_norm": 0.39037798204086893, |
|
"learning_rate": 8.999738520522065e-06, |
|
"loss": 0.1197, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 2.3212045169385194, |
|
"grad_norm": 0.3822304512858218, |
|
"learning_rate": 8.945277713535809e-06, |
|
"loss": 0.1152, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 2.327478042659975, |
|
"grad_norm": 0.3965998438998078, |
|
"learning_rate": 8.890848523188192e-06, |
|
"loss": 0.1243, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 2.3337515683814303, |
|
"grad_norm": 0.3939416904284715, |
|
"learning_rate": 8.836452581063248e-06, |
|
"loss": 0.1195, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 2.3400250941028857, |
|
"grad_norm": 0.4276348064874773, |
|
"learning_rate": 8.78209151774835e-06, |
|
"loss": 0.1211, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 2.346298619824341, |
|
"grad_norm": 0.39140182898284753, |
|
"learning_rate": 8.727766962785344e-06, |
|
"loss": 0.1157, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 2.3525721455457966, |
|
"grad_norm": 0.3634468198734603, |
|
"learning_rate": 8.673480544621681e-06, |
|
"loss": 0.1129, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 2.358845671267252, |
|
"grad_norm": 0.39837989743789176, |
|
"learning_rate": 8.61923389056162e-06, |
|
"loss": 0.1198, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 2.3651191969887075, |
|
"grad_norm": 0.3935841490044898, |
|
"learning_rate": 8.565028626717435e-06, |
|
"loss": 0.1203, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 2.371392722710163, |
|
"grad_norm": 0.35305962598333074, |
|
"learning_rate": 8.51086637796068e-06, |
|
"loss": 0.1128, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 2.3776662484316184, |
|
"grad_norm": 0.394308865525823, |
|
"learning_rate": 8.456748767873474e-06, |
|
"loss": 0.1124, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 2.383939774153074, |
|
"grad_norm": 0.3841761354621664, |
|
"learning_rate": 8.402677418699842e-06, |
|
"loss": 0.1145, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 2.3902132998745294, |
|
"grad_norm": 0.3893473191716482, |
|
"learning_rate": 8.34865395129707e-06, |
|
"loss": 0.1197, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 2.396486825595985, |
|
"grad_norm": 0.42534674714643167, |
|
"learning_rate": 8.294679985087137e-06, |
|
"loss": 0.1179, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 2.4027603513174403, |
|
"grad_norm": 0.40229677875453496, |
|
"learning_rate": 8.240757138008149e-06, |
|
"loss": 0.1236, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 2.4090338770388957, |
|
"grad_norm": 0.37978448174191587, |
|
"learning_rate": 8.186887026465857e-06, |
|
"loss": 0.1125, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 2.415307402760351, |
|
"grad_norm": 0.38090023973889275, |
|
"learning_rate": 8.133071265285209e-06, |
|
"loss": 0.1175, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 2.4215809284818066, |
|
"grad_norm": 0.3709724709852035, |
|
"learning_rate": 8.079311467661912e-06, |
|
"loss": 0.1189, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 2.427854454203262, |
|
"grad_norm": 0.3968859415543936, |
|
"learning_rate": 8.025609245114107e-06, |
|
"loss": 0.1208, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 2.4341279799247175, |
|
"grad_norm": 0.38140929315858313, |
|
"learning_rate": 7.971966207434045e-06, |
|
"loss": 0.1167, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 2.440401505646173, |
|
"grad_norm": 0.3813844728323988, |
|
"learning_rate": 7.918383962639835e-06, |
|
"loss": 0.1186, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 2.4466750313676284, |
|
"grad_norm": 0.36925458542907064, |
|
"learning_rate": 7.864864116927245e-06, |
|
"loss": 0.1167, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 2.452948557089084, |
|
"grad_norm": 0.3813928431538188, |
|
"learning_rate": 7.811408274621549e-06, |
|
"loss": 0.1217, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 2.4592220828105393, |
|
"grad_norm": 0.38045397571366496, |
|
"learning_rate": 7.75801803812944e-06, |
|
"loss": 0.1176, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 2.4654956085319952, |
|
"grad_norm": 0.409917244408148, |
|
"learning_rate": 7.704695007890988e-06, |
|
"loss": 0.1214, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 2.4717691342534502, |
|
"grad_norm": 0.39881220216006136, |
|
"learning_rate": 7.651440782331679e-06, |
|
"loss": 0.1176, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 2.478042659974906, |
|
"grad_norm": 0.3551138970811604, |
|
"learning_rate": 7.598256957814479e-06, |
|
"loss": 0.1156, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 2.484316185696361, |
|
"grad_norm": 0.38209748278035194, |
|
"learning_rate": 7.545145128592009e-06, |
|
"loss": 0.1128, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 2.490589711417817, |
|
"grad_norm": 0.38676641594007305, |
|
"learning_rate": 7.49210688675873e-06, |
|
"loss": 0.1175, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 2.496863237139272, |
|
"grad_norm": 0.36950074837730973, |
|
"learning_rate": 7.4391438222032265e-06, |
|
"loss": 0.1139, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 2.503136762860728, |
|
"grad_norm": 0.3915583584835345, |
|
"learning_rate": 7.3862575225605535e-06, |
|
"loss": 0.1179, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 2.509410288582183, |
|
"grad_norm": 0.3866601171887957, |
|
"learning_rate": 7.333449573164634e-06, |
|
"loss": 0.1207, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.515683814303639, |
|
"grad_norm": 0.4110664967201194, |
|
"learning_rate": 7.280721557000759e-06, |
|
"loss": 0.1166, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 2.521957340025094, |
|
"grad_norm": 0.37778473055073203, |
|
"learning_rate": 7.228075054658096e-06, |
|
"loss": 0.1157, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 2.5282308657465498, |
|
"grad_norm": 0.37225323631681123, |
|
"learning_rate": 7.175511644282349e-06, |
|
"loss": 0.1156, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 2.5345043914680048, |
|
"grad_norm": 0.3824521057716352, |
|
"learning_rate": 7.123032901528431e-06, |
|
"loss": 0.1182, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 2.5407779171894607, |
|
"grad_norm": 0.38115325746292966, |
|
"learning_rate": 7.070640399513232e-06, |
|
"loss": 0.1158, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 2.5470514429109157, |
|
"grad_norm": 0.3822411406686136, |
|
"learning_rate": 7.018335708768467e-06, |
|
"loss": 0.1177, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 2.5533249686323716, |
|
"grad_norm": 0.407300076312883, |
|
"learning_rate": 6.966120397193605e-06, |
|
"loss": 0.1152, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 2.5595984943538266, |
|
"grad_norm": 0.3666367383521563, |
|
"learning_rate": 6.913996030008853e-06, |
|
"loss": 0.1153, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 2.5658720200752825, |
|
"grad_norm": 0.3632368895081262, |
|
"learning_rate": 6.861964169708245e-06, |
|
"loss": 0.1107, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 2.572145545796738, |
|
"grad_norm": 0.3649524530723953, |
|
"learning_rate": 6.810026376012808e-06, |
|
"loss": 0.1145, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 2.5784190715181934, |
|
"grad_norm": 0.39293153907868017, |
|
"learning_rate": 6.758184205823791e-06, |
|
"loss": 0.1162, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 2.584692597239649, |
|
"grad_norm": 0.3931749546166204, |
|
"learning_rate": 6.706439213176028e-06, |
|
"loss": 0.1132, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 2.5909661229611043, |
|
"grad_norm": 0.3825377326068989, |
|
"learning_rate": 6.654792949191317e-06, |
|
"loss": 0.1222, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 2.5972396486825597, |
|
"grad_norm": 0.35192545209688325, |
|
"learning_rate": 6.603246962031942e-06, |
|
"loss": 0.112, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 2.603513174404015, |
|
"grad_norm": 0.3638182336052127, |
|
"learning_rate": 6.551802796854265e-06, |
|
"loss": 0.1144, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 2.6097867001254706, |
|
"grad_norm": 0.38711833706267534, |
|
"learning_rate": 6.500461995762402e-06, |
|
"loss": 0.1133, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 2.616060225846926, |
|
"grad_norm": 0.3897188834137444, |
|
"learning_rate": 6.449226097762e-06, |
|
"loss": 0.123, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 2.6223337515683816, |
|
"grad_norm": 0.37823736524426615, |
|
"learning_rate": 6.398096638714106e-06, |
|
"loss": 0.1184, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 2.628607277289837, |
|
"grad_norm": 0.3932949556705336, |
|
"learning_rate": 6.34707515128912e-06, |
|
"loss": 0.1143, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 2.6348808030112925, |
|
"grad_norm": 0.3884831444296721, |
|
"learning_rate": 6.296163164920858e-06, |
|
"loss": 0.1113, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 2.641154328732748, |
|
"grad_norm": 0.36382200051340396, |
|
"learning_rate": 6.245362205760703e-06, |
|
"loss": 0.1159, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 2.6474278544542034, |
|
"grad_norm": 0.3924419492733963, |
|
"learning_rate": 6.194673796631852e-06, |
|
"loss": 0.109, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 2.653701380175659, |
|
"grad_norm": 0.3947013863716826, |
|
"learning_rate": 6.144099456983681e-06, |
|
"loss": 0.1115, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 2.6599749058971143, |
|
"grad_norm": 0.3943261219862062, |
|
"learning_rate": 6.093640702846182e-06, |
|
"loss": 0.1122, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 2.6662484316185697, |
|
"grad_norm": 0.39005774295522977, |
|
"learning_rate": 6.043299046784526e-06, |
|
"loss": 0.1187, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 2.672521957340025, |
|
"grad_norm": 0.37092235530502005, |
|
"learning_rate": 5.993075997853719e-06, |
|
"loss": 0.1149, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 2.6787954830614806, |
|
"grad_norm": 0.3807422171419007, |
|
"learning_rate": 5.94297306155337e-06, |
|
"loss": 0.1135, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 2.685069008782936, |
|
"grad_norm": 0.3792169573613968, |
|
"learning_rate": 5.892991739782557e-06, |
|
"loss": 0.1199, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 2.6913425345043915, |
|
"grad_norm": 0.3831811436833811, |
|
"learning_rate": 5.843133530794817e-06, |
|
"loss": 0.1096, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 2.697616060225847, |
|
"grad_norm": 0.37933754782226464, |
|
"learning_rate": 5.793399929153216e-06, |
|
"loss": 0.1106, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 2.7038895859473024, |
|
"grad_norm": 0.38366361518402914, |
|
"learning_rate": 5.743792425685554e-06, |
|
"loss": 0.1154, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 2.710163111668758, |
|
"grad_norm": 0.3631788769456335, |
|
"learning_rate": 5.694312507439691e-06, |
|
"loss": 0.1141, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 2.7164366373902133, |
|
"grad_norm": 0.39467438603923816, |
|
"learning_rate": 5.644961657638942e-06, |
|
"loss": 0.1148, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 2.722710163111669, |
|
"grad_norm": 0.3643220647785576, |
|
"learning_rate": 5.595741355637645e-06, |
|
"loss": 0.1098, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 2.7289836888331243, |
|
"grad_norm": 0.39291715101415214, |
|
"learning_rate": 5.5466530768768005e-06, |
|
"loss": 0.1159, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 2.7352572145545797, |
|
"grad_norm": 0.40345238378734466, |
|
"learning_rate": 5.497698292839835e-06, |
|
"loss": 0.1117, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 2.741530740276035, |
|
"grad_norm": 0.41110337094460886, |
|
"learning_rate": 5.448878471008513e-06, |
|
"loss": 0.1134, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 2.7478042659974906, |
|
"grad_norm": 0.4006579187318595, |
|
"learning_rate": 5.400195074818924e-06, |
|
"loss": 0.1228, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 2.754077791718946, |
|
"grad_norm": 0.4006920463713775, |
|
"learning_rate": 5.351649563617638e-06, |
|
"loss": 0.1157, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 2.7603513174404015, |
|
"grad_norm": 0.40311348122015783, |
|
"learning_rate": 5.3032433926179395e-06, |
|
"loss": 0.1211, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.766624843161857, |
|
"grad_norm": 0.3833185052390898, |
|
"learning_rate": 5.25497801285622e-06, |
|
"loss": 0.1085, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 2.7728983688833124, |
|
"grad_norm": 0.40698791566770903, |
|
"learning_rate": 5.206854871148466e-06, |
|
"loss": 0.1145, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 2.779171894604768, |
|
"grad_norm": 0.37018390800544043, |
|
"learning_rate": 5.158875410046906e-06, |
|
"loss": 0.1148, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 2.7854454203262233, |
|
"grad_norm": 0.4097654897566281, |
|
"learning_rate": 5.111041067796754e-06, |
|
"loss": 0.1112, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 2.791718946047679, |
|
"grad_norm": 0.38185615087945834, |
|
"learning_rate": 5.063353278293106e-06, |
|
"loss": 0.1129, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 2.7979924717691342, |
|
"grad_norm": 0.38860623269143496, |
|
"learning_rate": 5.0158134710379595e-06, |
|
"loss": 0.1157, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 2.8042659974905897, |
|
"grad_norm": 0.41677106161054384, |
|
"learning_rate": 4.9684230710973394e-06, |
|
"loss": 0.1179, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 2.810539523212045, |
|
"grad_norm": 0.3659456249492505, |
|
"learning_rate": 4.921183499058615e-06, |
|
"loss": 0.1179, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 2.8168130489335006, |
|
"grad_norm": 0.390728899484992, |
|
"learning_rate": 4.8740961709878834e-06, |
|
"loss": 0.1118, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 2.823086574654956, |
|
"grad_norm": 0.41326006522454833, |
|
"learning_rate": 4.827162498387544e-06, |
|
"loss": 0.1123, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 2.8293601003764115, |
|
"grad_norm": 0.37982873419881247, |
|
"learning_rate": 4.780383888153983e-06, |
|
"loss": 0.1099, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 2.835633626097867, |
|
"grad_norm": 0.3826347732319729, |
|
"learning_rate": 4.733761742535381e-06, |
|
"loss": 0.1119, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 2.8419071518193224, |
|
"grad_norm": 0.4070056837379538, |
|
"learning_rate": 4.687297459089708e-06, |
|
"loss": 0.1169, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 2.848180677540778, |
|
"grad_norm": 0.3907501846052804, |
|
"learning_rate": 4.640992430642801e-06, |
|
"loss": 0.1184, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 2.8544542032622333, |
|
"grad_norm": 0.3980879608103747, |
|
"learning_rate": 4.594848045246638e-06, |
|
"loss": 0.1143, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 2.8607277289836888, |
|
"grad_norm": 0.3831724940346972, |
|
"learning_rate": 4.548865686137718e-06, |
|
"loss": 0.1165, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 2.867001254705144, |
|
"grad_norm": 0.3554795717929537, |
|
"learning_rate": 4.503046731695584e-06, |
|
"loss": 0.1103, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 2.8732747804265997, |
|
"grad_norm": 0.37945026941432614, |
|
"learning_rate": 4.457392555401531e-06, |
|
"loss": 0.1165, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 2.879548306148055, |
|
"grad_norm": 0.4040384960083639, |
|
"learning_rate": 4.411904525797408e-06, |
|
"loss": 0.112, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 2.8858218318695106, |
|
"grad_norm": 0.37374435050142246, |
|
"learning_rate": 4.3665840064446165e-06, |
|
"loss": 0.1152, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.892095357590966, |
|
"grad_norm": 0.38395743236739355, |
|
"learning_rate": 4.321432355883219e-06, |
|
"loss": 0.1158, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 2.8983688833124215, |
|
"grad_norm": 0.38107246526906685, |
|
"learning_rate": 4.276450927591229e-06, |
|
"loss": 0.1099, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 2.904642409033877, |
|
"grad_norm": 0.40288534078286153, |
|
"learning_rate": 4.231641069944019e-06, |
|
"loss": 0.1135, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 2.9109159347553324, |
|
"grad_norm": 0.39407625180559624, |
|
"learning_rate": 4.187004126173928e-06, |
|
"loss": 0.1153, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 2.917189460476788, |
|
"grad_norm": 0.390694853717115, |
|
"learning_rate": 4.1425414343299734e-06, |
|
"loss": 0.1113, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 2.9234629861982433, |
|
"grad_norm": 0.37505537186817833, |
|
"learning_rate": 4.098254327237742e-06, |
|
"loss": 0.1107, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 2.9297365119196987, |
|
"grad_norm": 0.3797569810287486, |
|
"learning_rate": 4.054144132459471e-06, |
|
"loss": 0.1125, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 2.936010037641154, |
|
"grad_norm": 0.39039171406836526, |
|
"learning_rate": 4.010212172254201e-06, |
|
"loss": 0.1139, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 2.9422835633626097, |
|
"grad_norm": 0.3735230020383592, |
|
"learning_rate": 3.966459763538179e-06, |
|
"loss": 0.1162, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 2.948557089084065, |
|
"grad_norm": 0.3799822252359316, |
|
"learning_rate": 3.92288821784536e-06, |
|
"loss": 0.1157, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.9548306148055206, |
|
"grad_norm": 0.37435313221883065, |
|
"learning_rate": 3.879498841288105e-06, |
|
"loss": 0.1109, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 2.961104140526976, |
|
"grad_norm": 0.37834252836846144, |
|
"learning_rate": 3.836292934518029e-06, |
|
"loss": 0.1124, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 2.9673776662484315, |
|
"grad_norm": 0.3755100276116184, |
|
"learning_rate": 3.793271792686993e-06, |
|
"loss": 0.1122, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 2.973651191969887, |
|
"grad_norm": 0.38213190237560674, |
|
"learning_rate": 3.750436705408311e-06, |
|
"loss": 0.1131, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 2.9799247176913424, |
|
"grad_norm": 0.39502326617269706, |
|
"learning_rate": 3.7077889567180625e-06, |
|
"loss": 0.113, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 2.9861982434127983, |
|
"grad_norm": 0.3907005868892978, |
|
"learning_rate": 3.6653298250366265e-06, |
|
"loss": 0.1121, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 2.9924717691342533, |
|
"grad_norm": 0.3757187294889032, |
|
"learning_rate": 3.6230605831303354e-06, |
|
"loss": 0.1138, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 2.998745294855709, |
|
"grad_norm": 0.36519480053180337, |
|
"learning_rate": 3.5809824980733445e-06, |
|
"loss": 0.1141, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.43099531531333923, |
|
"eval_runtime": 2.3647, |
|
"eval_samples_per_second": 14.801, |
|
"eval_steps_per_second": 0.846, |
|
"step": 2391 |
|
}, |
|
{ |
|
"epoch": 3.005018820577164, |
|
"grad_norm": 0.2442197751144262, |
|
"learning_rate": 3.5390968312096396e-06, |
|
"loss": 0.0533, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 3.0112923462986196, |
|
"grad_norm": 0.21528572471833773, |
|
"learning_rate": 3.497404838115219e-06, |
|
"loss": 0.0415, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 3.017565872020075, |
|
"grad_norm": 0.36503732282533347, |
|
"learning_rate": 3.455907768560477e-06, |
|
"loss": 0.0419, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 3.0238393977415305, |
|
"grad_norm": 0.34377984007445206, |
|
"learning_rate": 3.414606866472707e-06, |
|
"loss": 0.0402, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 3.030112923462986, |
|
"grad_norm": 0.3334896061335003, |
|
"learning_rate": 3.373503369898862e-06, |
|
"loss": 0.04, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 3.0363864491844414, |
|
"grad_norm": 0.2723536621745364, |
|
"learning_rate": 3.3325985109683877e-06, |
|
"loss": 0.0396, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 3.042659974905897, |
|
"grad_norm": 0.2691974596202031, |
|
"learning_rate": 3.291893515856334e-06, |
|
"loss": 0.0389, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 3.0489335006273524, |
|
"grad_norm": 0.28187769151055436, |
|
"learning_rate": 3.2513896047465654e-06, |
|
"loss": 0.0379, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 3.055207026348808, |
|
"grad_norm": 0.26156534288635025, |
|
"learning_rate": 3.211087991795201e-06, |
|
"loss": 0.0385, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 3.0614805520702637, |
|
"grad_norm": 0.2632608582642457, |
|
"learning_rate": 3.1709898850942234e-06, |
|
"loss": 0.0381, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 3.067754077791719, |
|
"grad_norm": 0.28572880260972616, |
|
"learning_rate": 3.1310964866352524e-06, |
|
"loss": 0.0389, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 3.0740276035131746, |
|
"grad_norm": 0.28301745878682383, |
|
"learning_rate": 3.0914089922735215e-06, |
|
"loss": 0.0375, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 3.08030112923463, |
|
"grad_norm": 0.29208432776276283, |
|
"learning_rate": 3.051928591692017e-06, |
|
"loss": 0.0381, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 3.0865746549560855, |
|
"grad_norm": 0.2847903910965424, |
|
"learning_rate": 3.012656468365842e-06, |
|
"loss": 0.0368, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 3.092848180677541, |
|
"grad_norm": 0.2966192251884188, |
|
"learning_rate": 2.9735937995267108e-06, |
|
"loss": 0.0371, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 3.0991217063989964, |
|
"grad_norm": 0.2796882316242176, |
|
"learning_rate": 2.9347417561276812e-06, |
|
"loss": 0.0376, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 3.105395232120452, |
|
"grad_norm": 0.3038556064545983, |
|
"learning_rate": 2.8961015028080506e-06, |
|
"loss": 0.0385, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 3.1116687578419073, |
|
"grad_norm": 0.2800625738008763, |
|
"learning_rate": 2.8576741978584265e-06, |
|
"loss": 0.0386, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 3.117942283563363, |
|
"grad_norm": 0.29844382477038445, |
|
"learning_rate": 2.819460993186032e-06, |
|
"loss": 0.0388, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 3.1242158092848182, |
|
"grad_norm": 0.3177106222509896, |
|
"learning_rate": 2.781463034280153e-06, |
|
"loss": 0.0379, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 3.1304893350062737, |
|
"grad_norm": 0.2995753268233054, |
|
"learning_rate": 2.7436814601778174e-06, |
|
"loss": 0.0387, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 3.136762860727729, |
|
"grad_norm": 0.30624087534086814, |
|
"learning_rate": 2.7061174034296434e-06, |
|
"loss": 0.0398, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 3.1430363864491846, |
|
"grad_norm": 0.3087285778258557, |
|
"learning_rate": 2.668771990065884e-06, |
|
"loss": 0.0394, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 3.14930991217064, |
|
"grad_norm": 0.2941075887579034, |
|
"learning_rate": 2.631646339562689e-06, |
|
"loss": 0.0386, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 3.1555834378920955, |
|
"grad_norm": 0.3118597830883225, |
|
"learning_rate": 2.594741564808527e-06, |
|
"loss": 0.0397, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 3.161856963613551, |
|
"grad_norm": 0.3208254308061256, |
|
"learning_rate": 2.558058772070846e-06, |
|
"loss": 0.0385, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 3.1681304893350064, |
|
"grad_norm": 0.26274648640285136, |
|
"learning_rate": 2.521599060962895e-06, |
|
"loss": 0.039, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 3.174404015056462, |
|
"grad_norm": 0.3244783368883984, |
|
"learning_rate": 2.4853635244107743e-06, |
|
"loss": 0.0373, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 3.1806775407779173, |
|
"grad_norm": 0.3321834864243195, |
|
"learning_rate": 2.449353248620657e-06, |
|
"loss": 0.0369, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 3.1869510664993728, |
|
"grad_norm": 0.298380472148583, |
|
"learning_rate": 2.41356931304625e-06, |
|
"loss": 0.0383, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 3.193224592220828, |
|
"grad_norm": 0.2887212580399006, |
|
"learning_rate": 2.37801279035642e-06, |
|
"loss": 0.0383, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 3.1994981179422837, |
|
"grad_norm": 0.2685957992866305, |
|
"learning_rate": 2.342684746403037e-06, |
|
"loss": 0.0382, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 3.205771643663739, |
|
"grad_norm": 0.3137213818837783, |
|
"learning_rate": 2.307586240189049e-06, |
|
"loss": 0.04, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 3.2120451693851946, |
|
"grad_norm": 0.29781867047646216, |
|
"learning_rate": 2.272718323836701e-06, |
|
"loss": 0.0379, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 3.21831869510665, |
|
"grad_norm": 0.28001089505480686, |
|
"learning_rate": 2.238082042556029e-06, |
|
"loss": 0.0382, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 3.2245922208281055, |
|
"grad_norm": 0.31914940913340023, |
|
"learning_rate": 2.2036784346134976e-06, |
|
"loss": 0.0376, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 3.230865746549561, |
|
"grad_norm": 0.2728867395710102, |
|
"learning_rate": 2.169508531300908e-06, |
|
"loss": 0.0382, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 3.2371392722710164, |
|
"grad_norm": 0.30042453420426346, |
|
"learning_rate": 2.1355733569044633e-06, |
|
"loss": 0.0389, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 3.243412797992472, |
|
"grad_norm": 0.2752139872422629, |
|
"learning_rate": 2.101873928674064e-06, |
|
"loss": 0.0374, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 3.2496863237139273, |
|
"grad_norm": 0.3662305601638148, |
|
"learning_rate": 2.0684112567928314e-06, |
|
"loss": 0.0369, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 3.2559598494353827, |
|
"grad_norm": 0.2853508636394266, |
|
"learning_rate": 2.035186344346801e-06, |
|
"loss": 0.0379, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 3.262233375156838, |
|
"grad_norm": 0.2909774250341541, |
|
"learning_rate": 2.0022001872948814e-06, |
|
"loss": 0.0374, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 3.2685069008782937, |
|
"grad_norm": 0.303083557347797, |
|
"learning_rate": 1.9694537744389754e-06, |
|
"loss": 0.0372, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 3.274780426599749, |
|
"grad_norm": 0.2837072081387949, |
|
"learning_rate": 1.9369480873943524e-06, |
|
"loss": 0.037, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 3.2810539523212046, |
|
"grad_norm": 0.293463671873792, |
|
"learning_rate": 1.9046841005602268e-06, |
|
"loss": 0.0368, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 3.28732747804266, |
|
"grad_norm": 0.29070250069814846, |
|
"learning_rate": 1.8726627810905284e-06, |
|
"loss": 0.037, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 3.2936010037641155, |
|
"grad_norm": 0.28990000456159315, |
|
"learning_rate": 1.8408850888649398e-06, |
|
"loss": 0.0368, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 3.299874529485571, |
|
"grad_norm": 0.2768135817440434, |
|
"learning_rate": 1.8093519764600931e-06, |
|
"loss": 0.0381, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 3.3061480552070264, |
|
"grad_norm": 0.3106851238797327, |
|
"learning_rate": 1.778064389121048e-06, |
|
"loss": 0.0373, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 3.312421580928482, |
|
"grad_norm": 0.26724960926250796, |
|
"learning_rate": 1.7470232647329222e-06, |
|
"loss": 0.0382, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 3.3186951066499373, |
|
"grad_norm": 0.3030133716572175, |
|
"learning_rate": 1.7162295337928036e-06, |
|
"loss": 0.0384, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 3.3249686323713927, |
|
"grad_norm": 0.31346010178385275, |
|
"learning_rate": 1.685684119381844e-06, |
|
"loss": 0.038, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 3.331242158092848, |
|
"grad_norm": 0.28444292176954256, |
|
"learning_rate": 1.655387937137589e-06, |
|
"loss": 0.0363, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 3.3375156838143036, |
|
"grad_norm": 0.29308964877533195, |
|
"learning_rate": 1.6253418952265398e-06, |
|
"loss": 0.0372, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 3.343789209535759, |
|
"grad_norm": 0.29189420850538267, |
|
"learning_rate": 1.5955468943169217e-06, |
|
"loss": 0.0377, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 3.3500627352572145, |
|
"grad_norm": 0.3102531079304282, |
|
"learning_rate": 1.5660038275516898e-06, |
|
"loss": 0.0355, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 3.35633626097867, |
|
"grad_norm": 0.27790750204144, |
|
"learning_rate": 1.536713580521746e-06, |
|
"loss": 0.0372, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 3.3626097867001254, |
|
"grad_norm": 0.3226354962908228, |
|
"learning_rate": 1.5076770312394096e-06, |
|
"loss": 0.0369, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 3.368883312421581, |
|
"grad_norm": 0.26898149132297955, |
|
"learning_rate": 1.4788950501120781e-06, |
|
"loss": 0.0382, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 3.3751568381430364, |
|
"grad_norm": 0.3188502485680553, |
|
"learning_rate": 1.450368499916155e-06, |
|
"loss": 0.0381, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 3.381430363864492, |
|
"grad_norm": 0.3007960797574723, |
|
"learning_rate": 1.4220982357711743e-06, |
|
"loss": 0.0372, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 3.3877038895859473, |
|
"grad_norm": 0.29650117453341984, |
|
"learning_rate": 1.3940851051141646e-06, |
|
"loss": 0.0377, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 3.3939774153074027, |
|
"grad_norm": 0.29709762351304697, |
|
"learning_rate": 1.366329947674263e-06, |
|
"loss": 0.0372, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 3.400250941028858, |
|
"grad_norm": 0.2951278070367851, |
|
"learning_rate": 1.3388335954475207e-06, |
|
"loss": 0.0375, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 3.4065244667503136, |
|
"grad_norm": 0.27968586803836637, |
|
"learning_rate": 1.3115968726719819e-06, |
|
"loss": 0.0359, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 3.412797992471769, |
|
"grad_norm": 0.30665692772003755, |
|
"learning_rate": 1.284620595802969e-06, |
|
"loss": 0.0376, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 3.4190715181932245, |
|
"grad_norm": 0.3166229935564368, |
|
"learning_rate": 1.2579055734886004e-06, |
|
"loss": 0.0361, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 3.42534504391468, |
|
"grad_norm": 0.28812682524392, |
|
"learning_rate": 1.2314526065455678e-06, |
|
"loss": 0.0371, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 3.4316185696361354, |
|
"grad_norm": 0.2962918652302945, |
|
"learning_rate": 1.2052624879351105e-06, |
|
"loss": 0.0375, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 3.437892095357591, |
|
"grad_norm": 0.2959207573782531, |
|
"learning_rate": 1.179336002739263e-06, |
|
"loss": 0.0368, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 3.4441656210790463, |
|
"grad_norm": 0.3222288463448185, |
|
"learning_rate": 1.1536739281373122e-06, |
|
"loss": 0.0366, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 3.450439146800502, |
|
"grad_norm": 0.2924594477361343, |
|
"learning_rate": 1.1282770333825022e-06, |
|
"loss": 0.0375, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 3.4567126725219572, |
|
"grad_norm": 0.32299277919132047, |
|
"learning_rate": 1.1031460797789718e-06, |
|
"loss": 0.0374, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 3.4629861982434127, |
|
"grad_norm": 0.3109436060749281, |
|
"learning_rate": 1.0782818206589375e-06, |
|
"loss": 0.0367, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 3.469259723964868, |
|
"grad_norm": 0.27272406454567366, |
|
"learning_rate": 1.053685001360112e-06, |
|
"loss": 0.0356, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 3.4755332496863236, |
|
"grad_norm": 0.30006540518061847, |
|
"learning_rate": 1.0293563592033595e-06, |
|
"loss": 0.0366, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 3.481806775407779, |
|
"grad_norm": 0.28002730472344367, |
|
"learning_rate": 1.0052966234705953e-06, |
|
"loss": 0.0383, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 3.4880803011292345, |
|
"grad_norm": 0.32173748821041304, |
|
"learning_rate": 9.815065153829195e-07, |
|
"loss": 0.0366, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 3.49435382685069, |
|
"grad_norm": 0.3042093630820431, |
|
"learning_rate": 9.579867480790061e-07, |
|
"loss": 0.0369, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 3.5006273525721454, |
|
"grad_norm": 0.2975724581779801, |
|
"learning_rate": 9.347380265937167e-07, |
|
"loss": 0.0361, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 3.506900878293601, |
|
"grad_norm": 0.2893317769414031, |
|
"learning_rate": 9.117610478369743e-07, |
|
"loss": 0.0383, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 3.5131744040150563, |
|
"grad_norm": 0.30137847743662044, |
|
"learning_rate": 8.890565005728691e-07, |
|
"loss": 0.0384, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 3.5194479297365118, |
|
"grad_norm": 0.2713586307285477, |
|
"learning_rate": 8.666250653990071e-07, |
|
"loss": 0.0364, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 3.5257214554579672, |
|
"grad_norm": 0.2749487801394873, |
|
"learning_rate": 8.44467414726119e-07, |
|
"loss": 0.0353, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 3.5319949811794227, |
|
"grad_norm": 0.3243183464549784, |
|
"learning_rate": 8.225842127578909e-07, |
|
"loss": 0.0369, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 3.538268506900878, |
|
"grad_norm": 0.2890309082875547, |
|
"learning_rate": 8.009761154710671e-07, |
|
"loss": 0.0369, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 3.544542032622334, |
|
"grad_norm": 0.2879952226873715, |
|
"learning_rate": 7.796437705957782e-07, |
|
"loss": 0.0368, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 3.550815558343789, |
|
"grad_norm": 0.2696320382813176, |
|
"learning_rate": 7.585878175961237e-07, |
|
"loss": 0.0371, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 3.557089084065245, |
|
"grad_norm": 0.2936735805318314, |
|
"learning_rate": 7.378088876510092e-07, |
|
"loss": 0.0376, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 3.5633626097867, |
|
"grad_norm": 0.25416101821322773, |
|
"learning_rate": 7.1730760363522e-07, |
|
"loss": 0.0352, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 3.569636135508156, |
|
"grad_norm": 0.30300674321039756, |
|
"learning_rate": 6.970845801007564e-07, |
|
"loss": 0.0353, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 3.575909661229611, |
|
"grad_norm": 0.2732998588656073, |
|
"learning_rate": 6.771404232584011e-07, |
|
"loss": 0.0362, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 3.5821831869510667, |
|
"grad_norm": 0.3795076600221709, |
|
"learning_rate": 6.574757309595636e-07, |
|
"loss": 0.0361, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 3.5884567126725218, |
|
"grad_norm": 0.28624108454045416, |
|
"learning_rate": 6.380910926783402e-07, |
|
"loss": 0.0363, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 3.5947302383939777, |
|
"grad_norm": 0.2674910858708607, |
|
"learning_rate": 6.189870894938587e-07, |
|
"loss": 0.0353, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 3.6010037641154327, |
|
"grad_norm": 0.2789232394059912, |
|
"learning_rate": 6.001642940728503e-07, |
|
"loss": 0.0354, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 3.6072772898368886, |
|
"grad_norm": 0.30882594178964384, |
|
"learning_rate": 5.816232706524838e-07, |
|
"loss": 0.0366, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 3.6135508155583436, |
|
"grad_norm": 0.2806952241750321, |
|
"learning_rate": 5.63364575023465e-07, |
|
"loss": 0.0361, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 3.6198243412797995, |
|
"grad_norm": 0.30164828030350377, |
|
"learning_rate": 5.453887545133563e-07, |
|
"loss": 0.0378, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 3.6260978670012545, |
|
"grad_norm": 0.2679619622153946, |
|
"learning_rate": 5.276963479701857e-07, |
|
"loss": 0.0358, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 3.6323713927227104, |
|
"grad_norm": 0.3231146432041487, |
|
"learning_rate": 5.102878857462811e-07, |
|
"loss": 0.0389, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 3.6386449184441654, |
|
"grad_norm": 0.2779119069062509, |
|
"learning_rate": 4.931638896823876e-07, |
|
"loss": 0.0384, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 3.6449184441656213, |
|
"grad_norm": 0.2722147852694776, |
|
"learning_rate": 4.763248730920089e-07, |
|
"loss": 0.0358, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 3.6511919698870763, |
|
"grad_norm": 0.293829432405976, |
|
"learning_rate": 4.5977134074603246e-07, |
|
"loss": 0.037, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 3.657465495608532, |
|
"grad_norm": 0.3065193645356902, |
|
"learning_rate": 4.4350378885759105e-07, |
|
"loss": 0.0371, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 3.663739021329987, |
|
"grad_norm": 0.28257044296271877, |
|
"learning_rate": 4.275227050671904e-07, |
|
"loss": 0.0364, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 3.670012547051443, |
|
"grad_norm": 0.2782812011499114, |
|
"learning_rate": 4.1182856842809204e-07, |
|
"loss": 0.0348, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 3.676286072772898, |
|
"grad_norm": 0.28872652510766195, |
|
"learning_rate": 3.964218493919525e-07, |
|
"loss": 0.0361, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 3.682559598494354, |
|
"grad_norm": 0.3053021191496778, |
|
"learning_rate": 3.813030097947212e-07, |
|
"loss": 0.0357, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 3.6888331242158094, |
|
"grad_norm": 0.3648251833459472, |
|
"learning_rate": 3.6647250284279735e-07, |
|
"loss": 0.0404, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 3.695106649937265, |
|
"grad_norm": 0.3319781395595929, |
|
"learning_rate": 3.5193077309943923e-07, |
|
"loss": 0.0403, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 3.7013801756587204, |
|
"grad_norm": 0.3085438905662011, |
|
"learning_rate": 3.376782564714476e-07, |
|
"loss": 0.0369, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 3.707653701380176, |
|
"grad_norm": 0.2849277889599658, |
|
"learning_rate": 3.237153801960868e-07, |
|
"loss": 0.0362, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 3.7139272271016313, |
|
"grad_norm": 0.31923000028121545, |
|
"learning_rate": 3.100425628282899e-07, |
|
"loss": 0.0369, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 3.7202007528230867, |
|
"grad_norm": 0.2929761574906958, |
|
"learning_rate": 2.9666021422810274e-07, |
|
"loss": 0.0369, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 3.726474278544542, |
|
"grad_norm": 0.2992636928106286, |
|
"learning_rate": 2.8356873554840514e-07, |
|
"loss": 0.0364, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 3.7327478042659976, |
|
"grad_norm": 0.301134189273727, |
|
"learning_rate": 2.7076851922287704e-07, |
|
"loss": 0.0354, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 3.739021329987453, |
|
"grad_norm": 0.2943853551390973, |
|
"learning_rate": 2.5825994895424255e-07, |
|
"loss": 0.0346, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 3.7452948557089085, |
|
"grad_norm": 0.2920091794037564, |
|
"learning_rate": 2.460433997027634e-07, |
|
"loss": 0.0377, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 3.751568381430364, |
|
"grad_norm": 0.27495401738635517, |
|
"learning_rate": 2.3411923767500455e-07, |
|
"loss": 0.0378, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 3.7578419071518194, |
|
"grad_norm": 0.39264097272429527, |
|
"learning_rate": 2.224878203128511e-07, |
|
"loss": 0.037, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 3.764115432873275, |
|
"grad_norm": 0.30250413845451674, |
|
"learning_rate": 2.1114949628279201e-07, |
|
"loss": 0.0366, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 3.7703889585947303, |
|
"grad_norm": 0.35611796959076564, |
|
"learning_rate": 2.001046054654776e-07, |
|
"loss": 0.0371, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 3.776662484316186, |
|
"grad_norm": 0.2561175523575295, |
|
"learning_rate": 1.893534789455209e-07, |
|
"loss": 0.0351, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 3.7829360100376412, |
|
"grad_norm": 0.28295745535344086, |
|
"learning_rate": 1.7889643900158016e-07, |
|
"loss": 0.0361, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 3.7892095357590967, |
|
"grad_norm": 0.28512419406398154, |
|
"learning_rate": 1.6873379909669307e-07, |
|
"loss": 0.0376, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 3.795483061480552, |
|
"grad_norm": 0.30485457369447977, |
|
"learning_rate": 1.5886586386888449e-07, |
|
"loss": 0.0377, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 3.8017565872020076, |
|
"grad_norm": 0.2901818797318041, |
|
"learning_rate": 1.4929292912203354e-07, |
|
"loss": 0.0357, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 3.808030112923463, |
|
"grad_norm": 0.2667781630207582, |
|
"learning_rate": 1.4001528181700196e-07, |
|
"loss": 0.0382, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 3.8143036386449185, |
|
"grad_norm": 0.3034166947926825, |
|
"learning_rate": 1.3103320006303766e-07, |
|
"loss": 0.0382, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 3.820577164366374, |
|
"grad_norm": 0.3119770203364585, |
|
"learning_rate": 1.2234695310944012e-07, |
|
"loss": 0.0367, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 3.8268506900878294, |
|
"grad_norm": 0.31429783494460234, |
|
"learning_rate": 1.1395680133747811e-07, |
|
"loss": 0.036, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 3.833124215809285, |
|
"grad_norm": 0.2655885847696681, |
|
"learning_rate": 1.0586299625259699e-07, |
|
"loss": 0.0372, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 3.8393977415307403, |
|
"grad_norm": 0.37692736629023765, |
|
"learning_rate": 9.806578047687254e-08, |
|
"loss": 0.0401, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 3.8456712672521958, |
|
"grad_norm": 0.28037362911434216, |
|
"learning_rate": 9.056538774174117e-08, |
|
"loss": 0.0383, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 3.851944792973651, |
|
"grad_norm": 0.2813649801748803, |
|
"learning_rate": 8.336204288098671e-08, |
|
"loss": 0.0409, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 3.8582183186951067, |
|
"grad_norm": 0.28158812132522376, |
|
"learning_rate": 7.64559618240146e-08, |
|
"loss": 0.0371, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 3.864491844416562, |
|
"grad_norm": 0.29334208788094884, |
|
"learning_rate": 6.984735158936384e-08, |
|
"loss": 0.0377, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 3.8707653701380176, |
|
"grad_norm": 0.2776501474412646, |
|
"learning_rate": 6.353641027850965e-08, |
|
"loss": 0.0368, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 3.877038895859473, |
|
"grad_norm": 0.27164861879214747, |
|
"learning_rate": 5.7523327069926024e-08, |
|
"loss": 0.0371, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 3.8833124215809285, |
|
"grad_norm": 0.26782362958737654, |
|
"learning_rate": 5.1808282213410276e-08, |
|
"loss": 0.0366, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 3.889585947302384, |
|
"grad_norm": 0.31086719768707505, |
|
"learning_rate": 4.63914470246829e-08, |
|
"loss": 0.0361, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 3.8958594730238394, |
|
"grad_norm": 0.30662847914316993, |
|
"learning_rate": 4.1272983880249476e-08, |
|
"loss": 0.0374, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 3.902132998745295, |
|
"grad_norm": 0.2915681277032509, |
|
"learning_rate": 3.645304621253787e-08, |
|
"loss": 0.0365, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 3.9084065244667503, |
|
"grad_norm": 0.31776081537688317, |
|
"learning_rate": 3.193177850529416e-08, |
|
"loss": 0.0352, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 3.9146800501882058, |
|
"grad_norm": 0.26314408385807814, |
|
"learning_rate": 2.7709316289253885e-08, |
|
"loss": 0.0362, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 3.920953575909661, |
|
"grad_norm": 0.30353550353186937, |
|
"learning_rate": 2.378578613807969e-08, |
|
"loss": 0.0367, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 3.9272271016311167, |
|
"grad_norm": 0.3008469541365023, |
|
"learning_rate": 2.0161305664563312e-08, |
|
"loss": 0.0377, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 3.933500627352572, |
|
"grad_norm": 0.29989474797608723, |
|
"learning_rate": 1.6835983517108357e-08, |
|
"loss": 0.0364, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 3.9397741530740276, |
|
"grad_norm": 0.3007906674460069, |
|
"learning_rate": 1.3809919376461811e-08, |
|
"loss": 0.0367, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 3.946047678795483, |
|
"grad_norm": 0.2822359703350314, |
|
"learning_rate": 1.1083203952737543e-08, |
|
"loss": 0.0371, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 3.9523212045169385, |
|
"grad_norm": 0.32240083725231283, |
|
"learning_rate": 8.655918982689582e-09, |
|
"loss": 0.0367, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 3.958594730238394, |
|
"grad_norm": 0.27144908135189433, |
|
"learning_rate": 6.528137227262976e-09, |
|
"loss": 0.0368, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 3.9648682559598494, |
|
"grad_norm": 0.29276670822521234, |
|
"learning_rate": 4.6999224694166405e-09, |
|
"loss": 0.0363, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 3.971141781681305, |
|
"grad_norm": 0.2920658294802816, |
|
"learning_rate": 3.1713295122071107e-09, |
|
"loss": 0.0352, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 3.9774153074027603, |
|
"grad_norm": 0.29999904692777374, |
|
"learning_rate": 1.9424041771465286e-09, |
|
"loss": 0.0354, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 3.9836888331242157, |
|
"grad_norm": 0.28940456234051576, |
|
"learning_rate": 1.013183302832621e-09, |
|
"loss": 0.0379, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 3.989962358845671, |
|
"grad_norm": 0.28743205163263064, |
|
"learning_rate": 3.8369474383848083e-10, |
|
"loss": 0.0349, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 3.9962358845671266, |
|
"grad_norm": 0.3004497086648703, |
|
"learning_rate": 5.395736988322853e-11, |
|
"loss": 0.0363, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.47916728258132935, |
|
"eval_runtime": 2.4435, |
|
"eval_samples_per_second": 14.324, |
|
"eval_steps_per_second": 0.818, |
|
"step": 3188 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"step": 3188, |
|
"total_flos": 2680278636036096.0, |
|
"train_loss": 0.20304497943459596, |
|
"train_runtime": 21214.9362, |
|
"train_samples_per_second": 4.806, |
|
"train_steps_per_second": 0.15 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 3188, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 319, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2680278636036096.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|