{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 1000, "global_step": 100000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01, "grad_norm": 3.153648614883423, "learning_rate": 5.9999999999999995e-05, "loss": 1.9475, "step": 100 }, { "epoch": 0.01, "grad_norm": 1.318032145500183, "learning_rate": 0.00011999999999999999, "loss": 1.7581, "step": 200 }, { "epoch": 0.01, "grad_norm": 1.7753959894180298, "learning_rate": 0.00017999999999999998, "loss": 1.7353, "step": 300 }, { "epoch": 0.02, "grad_norm": 1.262397289276123, "learning_rate": 0.00023999999999999998, "loss": 1.7634, "step": 400 }, { "epoch": 0.03, "grad_norm": 1.793628454208374, "learning_rate": 0.0003, "loss": 1.7828, "step": 500 }, { "epoch": 0.03, "grad_norm": 1.29915452003479, "learning_rate": 0.00029969849246231153, "loss": 1.7973, "step": 600 }, { "epoch": 0.04, "grad_norm": 1.3753058910369873, "learning_rate": 0.0002993969849246231, "loss": 1.7677, "step": 700 }, { "epoch": 0.04, "grad_norm": 3.032518148422241, "learning_rate": 0.00029909547738693465, "loss": 1.7775, "step": 800 }, { "epoch": 0.04, "grad_norm": 1.4150569438934326, "learning_rate": 0.0002987939698492462, "loss": 1.7872, "step": 900 }, { "epoch": 0.05, "grad_norm": 1.6522352695465088, "learning_rate": 0.00029849246231155777, "loss": 1.7433, "step": 1000 }, { "epoch": 0.05, "eval_loss": 1.7443219423294067, "eval_runtime": 37.3589, "eval_samples_per_second": 26.767, "eval_steps_per_second": 3.346, "step": 1000 }, { "epoch": 0.06, "grad_norm": 1.8125498294830322, "learning_rate": 0.00029819095477386933, "loss": 1.7885, "step": 1100 }, { "epoch": 0.06, "grad_norm": 1.4740030765533447, "learning_rate": 0.0002978894472361809, "loss": 1.7616, "step": 1200 }, { "epoch": 0.07, "grad_norm": 1.670320749282837, "learning_rate": 0.00029758793969849245, "loss": 1.7545, "step": 1300 }, { "epoch": 0.07, "grad_norm": 1.096781611442566, "learning_rate": 0.000297286432160804, "loss": 1.7072, "step": 1400 }, { "epoch": 0.07, "grad_norm": 1.8927794694900513, "learning_rate": 0.0002969849246231155, "loss": 1.7243, "step": 1500 }, { "epoch": 0.08, "grad_norm": 2.244074583053589, "learning_rate": 0.00029668341708542713, "loss": 1.7369, "step": 1600 }, { "epoch": 0.09, "grad_norm": 1.6228864192962646, "learning_rate": 0.0002963819095477387, "loss": 1.7541, "step": 1700 }, { "epoch": 0.09, "grad_norm": 0.9201287627220154, "learning_rate": 0.00029608040201005025, "loss": 1.7236, "step": 1800 }, { "epoch": 0.1, "grad_norm": 1.308199405670166, "learning_rate": 0.00029577889447236175, "loss": 1.7345, "step": 1900 }, { "epoch": 0.1, "grad_norm": 1.5356643199920654, "learning_rate": 0.00029548040201005023, "loss": 1.7256, "step": 2000 }, { "epoch": 0.1, "eval_loss": 1.7340322732925415, "eval_runtime": 37.5614, "eval_samples_per_second": 26.623, "eval_steps_per_second": 3.328, "step": 2000 }, { "epoch": 0.1, "grad_norm": 2.5252068042755127, "learning_rate": 0.0002951788944723618, "loss": 1.7454, "step": 2100 }, { "epoch": 0.11, "grad_norm": 1.5635749101638794, "learning_rate": 0.00029487738693467335, "loss": 1.7461, "step": 2200 }, { "epoch": 0.12, "grad_norm": 1.3426711559295654, "learning_rate": 0.0002945758793969849, "loss": 1.75, "step": 2300 }, { "epoch": 0.12, "grad_norm": 1.3829035758972168, "learning_rate": 0.00029427437185929647, "loss": 1.7554, "step": 2400 }, { "epoch": 0.12, "grad_norm": 1.2779866456985474, "learning_rate": 0.00029397286432160803, "loss": 1.7262, "step": 2500 }, { "epoch": 0.13, "grad_norm": 1.4913188219070435, "learning_rate": 0.0002936713567839196, "loss": 1.7333, "step": 2600 }, { "epoch": 0.14, "grad_norm": 1.5596439838409424, "learning_rate": 0.00029336984924623115, "loss": 1.728, "step": 2700 }, { "epoch": 0.14, "grad_norm": 1.1473088264465332, "learning_rate": 0.0002930683417085427, "loss": 1.7063, "step": 2800 }, { "epoch": 0.14, "grad_norm": 1.4872281551361084, "learning_rate": 0.0002927668341708542, "loss": 1.7417, "step": 2900 }, { "epoch": 0.15, "grad_norm": 1.3485779762268066, "learning_rate": 0.0002924653266331658, "loss": 1.7314, "step": 3000 }, { "epoch": 0.15, "eval_loss": 1.680450677871704, "eval_runtime": 38.1024, "eval_samples_per_second": 26.245, "eval_steps_per_second": 3.281, "step": 3000 }, { "epoch": 0.15, "grad_norm": 1.3290046453475952, "learning_rate": 0.0002921638190954774, "loss": 1.7101, "step": 3100 }, { "epoch": 0.16, "grad_norm": 0.9259174466133118, "learning_rate": 0.00029186231155778895, "loss": 1.75, "step": 3200 }, { "epoch": 0.17, "grad_norm": 1.4094349145889282, "learning_rate": 0.00029156080402010045, "loss": 1.7219, "step": 3300 }, { "epoch": 0.17, "grad_norm": 1.2119574546813965, "learning_rate": 0.000291259296482412, "loss": 1.741, "step": 3400 }, { "epoch": 0.17, "grad_norm": 1.1941887140274048, "learning_rate": 0.0002909577889447236, "loss": 1.7005, "step": 3500 }, { "epoch": 0.18, "grad_norm": 1.1339538097381592, "learning_rate": 0.00029065628140703513, "loss": 1.6965, "step": 3600 }, { "epoch": 0.18, "grad_norm": 2.2265231609344482, "learning_rate": 0.0002903547738693467, "loss": 1.701, "step": 3700 }, { "epoch": 0.19, "grad_norm": 2.011225938796997, "learning_rate": 0.00029005326633165825, "loss": 1.7241, "step": 3800 }, { "epoch": 0.2, "grad_norm": 1.3544070720672607, "learning_rate": 0.00028975175879396986, "loss": 1.7039, "step": 3900 }, { "epoch": 0.2, "grad_norm": 1.909501314163208, "learning_rate": 0.00028945025125628137, "loss": 1.676, "step": 4000 }, { "epoch": 0.2, "eval_loss": 1.6920135021209717, "eval_runtime": 37.5936, "eval_samples_per_second": 26.6, "eval_steps_per_second": 3.325, "step": 4000 }, { "epoch": 0.2, "grad_norm": 1.7186845541000366, "learning_rate": 0.0002891487437185929, "loss": 1.7, "step": 4100 }, { "epoch": 0.21, "grad_norm": 0.9834026098251343, "learning_rate": 0.0002888502512562814, "loss": 1.6639, "step": 4200 }, { "epoch": 0.21, "grad_norm": 0.9373641014099121, "learning_rate": 0.0002885487437185929, "loss": 1.6786, "step": 4300 }, { "epoch": 0.22, "grad_norm": 1.4471676349639893, "learning_rate": 0.00028824723618090447, "loss": 1.7199, "step": 4400 }, { "epoch": 0.23, "grad_norm": 0.9259161949157715, "learning_rate": 0.0002879457286432161, "loss": 1.7013, "step": 4500 }, { "epoch": 0.23, "grad_norm": 1.3199846744537354, "learning_rate": 0.0002876442211055276, "loss": 1.6865, "step": 4600 }, { "epoch": 0.23, "grad_norm": 1.6591229438781738, "learning_rate": 0.00028734271356783915, "loss": 1.6976, "step": 4700 }, { "epoch": 0.24, "grad_norm": 1.0676679611206055, "learning_rate": 0.0002870412060301507, "loss": 1.6921, "step": 4800 }, { "epoch": 0.24, "grad_norm": 1.3471635580062866, "learning_rate": 0.0002867396984924623, "loss": 1.6951, "step": 4900 }, { "epoch": 0.25, "grad_norm": 1.6372334957122803, "learning_rate": 0.00028643819095477383, "loss": 1.6951, "step": 5000 }, { "epoch": 0.25, "eval_loss": 1.6992712020874023, "eval_runtime": 37.5723, "eval_samples_per_second": 26.615, "eval_steps_per_second": 3.327, "step": 5000 }, { "epoch": 0.26, "grad_norm": 2.9456429481506348, "learning_rate": 0.0002861366834170854, "loss": 1.6708, "step": 5100 }, { "epoch": 0.26, "grad_norm": 1.8768843412399292, "learning_rate": 0.00028583517587939695, "loss": 1.701, "step": 5200 }, { "epoch": 0.27, "grad_norm": 1.4709163904190063, "learning_rate": 0.0002855336683417085, "loss": 1.7119, "step": 5300 }, { "epoch": 0.27, "grad_norm": 1.15412175655365, "learning_rate": 0.00028523216080402007, "loss": 1.6864, "step": 5400 }, { "epoch": 0.28, "grad_norm": 1.9388611316680908, "learning_rate": 0.0002849306532663316, "loss": 1.6781, "step": 5500 }, { "epoch": 0.28, "grad_norm": 2.746967315673828, "learning_rate": 0.0002846291457286432, "loss": 1.7134, "step": 5600 }, { "epoch": 0.28, "grad_norm": 1.7656490802764893, "learning_rate": 0.00028432763819095474, "loss": 1.6441, "step": 5700 }, { "epoch": 0.29, "grad_norm": 1.4275192022323608, "learning_rate": 0.0002840261306532663, "loss": 1.6893, "step": 5800 }, { "epoch": 0.29, "grad_norm": 1.8908779621124268, "learning_rate": 0.00028372462311557786, "loss": 1.6814, "step": 5900 }, { "epoch": 0.3, "grad_norm": 2.0805633068084717, "learning_rate": 0.0002834231155778894, "loss": 1.6765, "step": 6000 }, { "epoch": 0.3, "eval_loss": 1.6724690198898315, "eval_runtime": 37.9367, "eval_samples_per_second": 26.36, "eval_steps_per_second": 3.295, "step": 6000 }, { "epoch": 0.3, "grad_norm": 1.8758082389831543, "learning_rate": 0.000283121608040201, "loss": 1.6462, "step": 6100 }, { "epoch": 0.31, "grad_norm": 1.4686168432235718, "learning_rate": 0.00028282010050251254, "loss": 1.7117, "step": 6200 }, { "epoch": 0.32, "grad_norm": 1.857920527458191, "learning_rate": 0.0002825185929648241, "loss": 1.648, "step": 6300 }, { "epoch": 0.32, "grad_norm": 1.852232575416565, "learning_rate": 0.00028221708542713566, "loss": 1.7096, "step": 6400 }, { "epoch": 0.33, "grad_norm": 1.6206820011138916, "learning_rate": 0.0002819155778894472, "loss": 1.6379, "step": 6500 }, { "epoch": 0.33, "grad_norm": 0.9906002879142761, "learning_rate": 0.0002816140703517588, "loss": 1.6905, "step": 6600 }, { "epoch": 0.34, "grad_norm": 1.8640550374984741, "learning_rate": 0.00028131256281407034, "loss": 1.6842, "step": 6700 }, { "epoch": 0.34, "grad_norm": 1.5478594303131104, "learning_rate": 0.00028101407035175876, "loss": 1.6554, "step": 6800 }, { "epoch": 0.34, "grad_norm": 1.2689837217330933, "learning_rate": 0.0002807125628140703, "loss": 1.6775, "step": 6900 }, { "epoch": 0.35, "grad_norm": 1.8730539083480835, "learning_rate": 0.0002804110552763819, "loss": 1.6603, "step": 7000 }, { "epoch": 0.35, "eval_loss": 1.6557646989822388, "eval_runtime": 37.4609, "eval_samples_per_second": 26.694, "eval_steps_per_second": 3.337, "step": 7000 }, { "epoch": 0.35, "grad_norm": 1.455672025680542, "learning_rate": 0.00028010954773869344, "loss": 1.6635, "step": 7100 }, { "epoch": 0.36, "grad_norm": 1.6935358047485352, "learning_rate": 0.000279808040201005, "loss": 1.6796, "step": 7200 }, { "epoch": 0.36, "grad_norm": 2.0117626190185547, "learning_rate": 0.00027950653266331656, "loss": 1.6972, "step": 7300 }, { "epoch": 0.37, "grad_norm": 0.9937806129455566, "learning_rate": 0.0002792050251256281, "loss": 1.6666, "step": 7400 }, { "epoch": 0.38, "grad_norm": 1.1981546878814697, "learning_rate": 0.0002789035175879397, "loss": 1.665, "step": 7500 }, { "epoch": 0.38, "grad_norm": 2.0641427040100098, "learning_rate": 0.00027860201005025124, "loss": 1.6535, "step": 7600 }, { "epoch": 0.39, "grad_norm": 1.7236799001693726, "learning_rate": 0.0002783005025125628, "loss": 1.676, "step": 7700 }, { "epoch": 0.39, "grad_norm": 3.5143849849700928, "learning_rate": 0.00027799899497487436, "loss": 1.6521, "step": 7800 }, { "epoch": 0.4, "grad_norm": 1.7787047624588013, "learning_rate": 0.0002776974874371859, "loss": 1.6637, "step": 7900 }, { "epoch": 0.4, "grad_norm": 1.713392972946167, "learning_rate": 0.0002773959798994975, "loss": 1.6704, "step": 8000 }, { "epoch": 0.4, "eval_loss": 1.656675100326538, "eval_runtime": 37.9526, "eval_samples_per_second": 26.349, "eval_steps_per_second": 3.294, "step": 8000 }, { "epoch": 0.41, "grad_norm": 1.1325870752334595, "learning_rate": 0.00027709447236180904, "loss": 1.6118, "step": 8100 }, { "epoch": 0.41, "grad_norm": 1.746856689453125, "learning_rate": 0.0002767929648241206, "loss": 1.6718, "step": 8200 }, { "epoch": 0.41, "grad_norm": 1.4181280136108398, "learning_rate": 0.0002764914572864321, "loss": 1.6957, "step": 8300 }, { "epoch": 0.42, "grad_norm": 2.078387975692749, "learning_rate": 0.0002761899497487437, "loss": 1.6546, "step": 8400 }, { "epoch": 0.42, "grad_norm": 2.694249153137207, "learning_rate": 0.0002758884422110553, "loss": 1.6875, "step": 8500 }, { "epoch": 0.43, "grad_norm": 2.530956745147705, "learning_rate": 0.00027558693467336683, "loss": 1.6112, "step": 8600 }, { "epoch": 0.43, "grad_norm": 1.4322137832641602, "learning_rate": 0.00027528542713567834, "loss": 1.6572, "step": 8700 }, { "epoch": 0.44, "grad_norm": 1.996591329574585, "learning_rate": 0.0002749839195979899, "loss": 1.6442, "step": 8800 }, { "epoch": 0.45, "grad_norm": 1.6497749090194702, "learning_rate": 0.0002746824120603015, "loss": 1.6404, "step": 8900 }, { "epoch": 0.45, "grad_norm": 2.429196357727051, "learning_rate": 0.000274380904522613, "loss": 1.656, "step": 9000 }, { "epoch": 0.45, "eval_loss": 1.6714575290679932, "eval_runtime": 37.5615, "eval_samples_per_second": 26.623, "eval_steps_per_second": 3.328, "step": 9000 }, { "epoch": 0.46, "grad_norm": 1.9687163829803467, "learning_rate": 0.0002740824120603015, "loss": 1.6644, "step": 9100 }, { "epoch": 0.46, "grad_norm": 1.4858875274658203, "learning_rate": 0.00027378090452261306, "loss": 1.6434, "step": 9200 }, { "epoch": 0.47, "grad_norm": 1.8358262777328491, "learning_rate": 0.0002734793969849246, "loss": 1.6565, "step": 9300 }, { "epoch": 0.47, "grad_norm": 1.3822225332260132, "learning_rate": 0.0002731778894472361, "loss": 1.6506, "step": 9400 }, { "epoch": 0.47, "grad_norm": 1.9319818019866943, "learning_rate": 0.00027287638190954774, "loss": 1.6256, "step": 9500 }, { "epoch": 0.48, "grad_norm": 2.601515293121338, "learning_rate": 0.0002725748743718593, "loss": 1.6243, "step": 9600 }, { "epoch": 0.48, "grad_norm": 1.610561490058899, "learning_rate": 0.0002722733668341708, "loss": 1.6259, "step": 9700 }, { "epoch": 0.49, "grad_norm": 2.060863494873047, "learning_rate": 0.00027197185929648236, "loss": 1.655, "step": 9800 }, { "epoch": 0.49, "grad_norm": 2.7591397762298584, "learning_rate": 0.000271670351758794, "loss": 1.6342, "step": 9900 }, { "epoch": 0.5, "grad_norm": 1.2489566802978516, "learning_rate": 0.00027136884422110553, "loss": 1.6431, "step": 10000 }, { "epoch": 0.5, "eval_loss": 1.6347763538360596, "eval_runtime": 37.5252, "eval_samples_per_second": 26.649, "eval_steps_per_second": 3.331, "step": 10000 }, { "epoch": 0.51, "grad_norm": 1.2953132390975952, "learning_rate": 0.00027106733668341704, "loss": 1.6872, "step": 10100 }, { "epoch": 0.51, "grad_norm": 1.9919564723968506, "learning_rate": 0.0002707658291457286, "loss": 1.6576, "step": 10200 }, { "epoch": 0.52, "grad_norm": 1.3081834316253662, "learning_rate": 0.0002704643216080402, "loss": 1.6354, "step": 10300 }, { "epoch": 0.52, "grad_norm": 1.605245590209961, "learning_rate": 0.0002701628140703517, "loss": 1.6687, "step": 10400 }, { "epoch": 0.53, "grad_norm": 2.1541988849639893, "learning_rate": 0.0002698613065326633, "loss": 1.6631, "step": 10500 }, { "epoch": 0.53, "grad_norm": 1.8258408308029175, "learning_rate": 0.00026955979899497484, "loss": 1.6633, "step": 10600 }, { "epoch": 0.54, "grad_norm": 1.3377407789230347, "learning_rate": 0.00026925829145728645, "loss": 1.6544, "step": 10700 }, { "epoch": 0.54, "grad_norm": 0.998458206653595, "learning_rate": 0.00026895678391959795, "loss": 1.6126, "step": 10800 }, { "epoch": 0.55, "grad_norm": 1.3561229705810547, "learning_rate": 0.0002686552763819095, "loss": 1.6614, "step": 10900 }, { "epoch": 0.55, "grad_norm": 2.4729514122009277, "learning_rate": 0.0002683537688442211, "loss": 1.6439, "step": 11000 }, { "epoch": 0.55, "eval_loss": 1.648992657661438, "eval_runtime": 37.818, "eval_samples_per_second": 26.442, "eval_steps_per_second": 3.305, "step": 11000 }, { "epoch": 0.56, "grad_norm": 1.2856806516647339, "learning_rate": 0.00026805226130653263, "loss": 1.605, "step": 11100 }, { "epoch": 0.56, "grad_norm": 1.7708286046981812, "learning_rate": 0.0002677507537688442, "loss": 1.6257, "step": 11200 }, { "epoch": 0.56, "grad_norm": 2.085149049758911, "learning_rate": 0.00026744924623115575, "loss": 1.6347, "step": 11300 }, { "epoch": 0.57, "grad_norm": 0.9750702977180481, "learning_rate": 0.0002671477386934673, "loss": 1.6496, "step": 11400 }, { "epoch": 0.57, "grad_norm": 1.9253658056259155, "learning_rate": 0.00026684623115577887, "loss": 1.6395, "step": 11500 }, { "epoch": 0.58, "grad_norm": 2.441312313079834, "learning_rate": 0.00026654472361809043, "loss": 1.6444, "step": 11600 }, { "epoch": 0.58, "grad_norm": 1.259020447731018, "learning_rate": 0.000266243216080402, "loss": 1.6114, "step": 11700 }, { "epoch": 0.59, "grad_norm": 1.3337092399597168, "learning_rate": 0.00026594170854271355, "loss": 1.6243, "step": 11800 }, { "epoch": 0.59, "grad_norm": 1.423687219619751, "learning_rate": 0.0002656402010050251, "loss": 1.623, "step": 11900 }, { "epoch": 0.6, "grad_norm": 1.2547069787979126, "learning_rate": 0.00026533869346733667, "loss": 1.631, "step": 12000 }, { "epoch": 0.6, "eval_loss": 1.6530547142028809, "eval_runtime": 37.4898, "eval_samples_per_second": 26.674, "eval_steps_per_second": 3.334, "step": 12000 }, { "epoch": 0.6, "grad_norm": 1.1078051328659058, "learning_rate": 0.00026503718592964823, "loss": 1.6229, "step": 12100 }, { "epoch": 0.61, "grad_norm": 1.2142789363861084, "learning_rate": 0.0002647356783919598, "loss": 1.6251, "step": 12200 }, { "epoch": 0.61, "grad_norm": 1.5933152437210083, "learning_rate": 0.00026443417085427135, "loss": 1.627, "step": 12300 }, { "epoch": 0.62, "grad_norm": 1.6759315729141235, "learning_rate": 0.0002641326633165829, "loss": 1.6294, "step": 12400 }, { "epoch": 0.62, "grad_norm": 1.487029790878296, "learning_rate": 0.00026383115577889447, "loss": 1.6406, "step": 12500 }, { "epoch": 0.63, "grad_norm": 2.004643201828003, "learning_rate": 0.000263529648241206, "loss": 1.6342, "step": 12600 }, { "epoch": 0.64, "grad_norm": 1.94633948802948, "learning_rate": 0.00026322814070351753, "loss": 1.6279, "step": 12700 }, { "epoch": 0.64, "grad_norm": 1.5213886499404907, "learning_rate": 0.00026292663316582914, "loss": 1.6023, "step": 12800 }, { "epoch": 0.65, "grad_norm": 1.4710832834243774, "learning_rate": 0.0002626251256281407, "loss": 1.6202, "step": 12900 }, { "epoch": 0.65, "grad_norm": 2.367037296295166, "learning_rate": 0.00026232361809045226, "loss": 1.6463, "step": 13000 }, { "epoch": 0.65, "eval_loss": 1.6226933002471924, "eval_runtime": 37.4429, "eval_samples_per_second": 26.707, "eval_steps_per_second": 3.338, "step": 13000 }, { "epoch": 0.66, "grad_norm": 2.803264856338501, "learning_rate": 0.0002620251256281407, "loss": 1.6219, "step": 13100 }, { "epoch": 0.66, "grad_norm": 1.5915918350219727, "learning_rate": 0.00026172361809045225, "loss": 1.6252, "step": 13200 }, { "epoch": 0.67, "grad_norm": 1.5102565288543701, "learning_rate": 0.0002614221105527638, "loss": 1.6346, "step": 13300 }, { "epoch": 0.67, "grad_norm": 1.8278179168701172, "learning_rate": 0.00026112060301507537, "loss": 1.6386, "step": 13400 }, { "epoch": 0.68, "grad_norm": 1.6756057739257812, "learning_rate": 0.0002608190954773869, "loss": 1.6135, "step": 13500 }, { "epoch": 0.68, "grad_norm": 1.4984675645828247, "learning_rate": 0.0002605175879396985, "loss": 1.6316, "step": 13600 }, { "epoch": 0.69, "grad_norm": 2.2536373138427734, "learning_rate": 0.00026021608040201005, "loss": 1.6317, "step": 13700 }, { "epoch": 0.69, "grad_norm": 1.3149595260620117, "learning_rate": 0.0002599145728643216, "loss": 1.6311, "step": 13800 }, { "epoch": 0.69, "grad_norm": 1.662287712097168, "learning_rate": 0.00025961306532663316, "loss": 1.6229, "step": 13900 }, { "epoch": 0.7, "grad_norm": 1.2096275091171265, "learning_rate": 0.0002593115577889447, "loss": 1.6305, "step": 14000 }, { "epoch": 0.7, "eval_loss": 1.632125735282898, "eval_runtime": 37.4575, "eval_samples_per_second": 26.697, "eval_steps_per_second": 3.337, "step": 14000 }, { "epoch": 0.7, "grad_norm": 2.01643705368042, "learning_rate": 0.00025901005025125623, "loss": 1.6271, "step": 14100 }, { "epoch": 0.71, "grad_norm": 1.2067662477493286, "learning_rate": 0.0002587085427135678, "loss": 1.6082, "step": 14200 }, { "epoch": 0.71, "grad_norm": 2.0471389293670654, "learning_rate": 0.0002584070351758794, "loss": 1.6346, "step": 14300 }, { "epoch": 0.72, "grad_norm": 1.1723861694335938, "learning_rate": 0.0002581055276381909, "loss": 1.6449, "step": 14400 }, { "epoch": 0.72, "grad_norm": 1.3024895191192627, "learning_rate": 0.00025780402010050247, "loss": 1.6159, "step": 14500 }, { "epoch": 0.73, "grad_norm": 1.6715235710144043, "learning_rate": 0.000257502512562814, "loss": 1.6096, "step": 14600 }, { "epoch": 0.73, "grad_norm": 2.116154432296753, "learning_rate": 0.00025720100502512564, "loss": 1.5984, "step": 14700 }, { "epoch": 0.74, "grad_norm": 1.2046904563903809, "learning_rate": 0.00025689949748743714, "loss": 1.6125, "step": 14800 }, { "epoch": 0.74, "grad_norm": 1.5058480501174927, "learning_rate": 0.0002565979899497487, "loss": 1.6164, "step": 14900 }, { "epoch": 0.75, "grad_norm": 2.2814691066741943, "learning_rate": 0.00025629648241206026, "loss": 1.6312, "step": 15000 }, { "epoch": 0.75, "eval_loss": 1.6935200691223145, "eval_runtime": 37.4213, "eval_samples_per_second": 26.723, "eval_steps_per_second": 3.34, "step": 15000 }, { "epoch": 0.76, "grad_norm": 1.5283032655715942, "learning_rate": 0.0002559979899497487, "loss": 1.6201, "step": 15100 }, { "epoch": 0.76, "grad_norm": 2.6960535049438477, "learning_rate": 0.00025569648241206025, "loss": 1.6062, "step": 15200 }, { "epoch": 0.77, "grad_norm": 1.080701470375061, "learning_rate": 0.00025539798994974873, "loss": 1.622, "step": 15300 }, { "epoch": 0.77, "grad_norm": 1.6446688175201416, "learning_rate": 0.0002550964824120603, "loss": 1.6075, "step": 15400 }, { "epoch": 0.78, "grad_norm": 1.9143463373184204, "learning_rate": 0.00025479497487437185, "loss": 1.6363, "step": 15500 }, { "epoch": 0.78, "grad_norm": 1.6148111820220947, "learning_rate": 0.0002544934673366834, "loss": 1.6367, "step": 15600 }, { "epoch": 0.79, "grad_norm": 1.7216590642929077, "learning_rate": 0.00025419195979899497, "loss": 1.5997, "step": 15700 }, { "epoch": 0.79, "grad_norm": 1.442865014076233, "learning_rate": 0.00025389045226130647, "loss": 1.5776, "step": 15800 }, { "epoch": 0.8, "grad_norm": 0.9140738844871521, "learning_rate": 0.0002535889447236181, "loss": 1.6275, "step": 15900 }, { "epoch": 0.8, "grad_norm": 2.099306583404541, "learning_rate": 0.00025328743718592965, "loss": 1.6072, "step": 16000 }, { "epoch": 0.8, "eval_loss": 1.5524722337722778, "eval_runtime": 37.4031, "eval_samples_per_second": 26.736, "eval_steps_per_second": 3.342, "step": 16000 }, { "epoch": 0.81, "grad_norm": 1.560035228729248, "learning_rate": 0.0002529859296482412, "loss": 1.5983, "step": 16100 }, { "epoch": 0.81, "grad_norm": 1.8614755868911743, "learning_rate": 0.0002526844221105527, "loss": 1.6271, "step": 16200 }, { "epoch": 0.81, "grad_norm": 1.7034022808074951, "learning_rate": 0.0002523829145728643, "loss": 1.6238, "step": 16300 }, { "epoch": 0.82, "grad_norm": 1.086572527885437, "learning_rate": 0.0002520814070351759, "loss": 1.6241, "step": 16400 }, { "epoch": 0.82, "grad_norm": 0.8860014081001282, "learning_rate": 0.0002517798994974874, "loss": 1.562, "step": 16500 }, { "epoch": 0.83, "grad_norm": 1.7819429636001587, "learning_rate": 0.00025147839195979895, "loss": 1.5995, "step": 16600 }, { "epoch": 0.83, "grad_norm": 3.512892246246338, "learning_rate": 0.00025117688442211056, "loss": 1.5866, "step": 16700 }, { "epoch": 0.84, "grad_norm": 1.5962600708007812, "learning_rate": 0.0002508753768844221, "loss": 1.6358, "step": 16800 }, { "epoch": 0.84, "grad_norm": 1.7714247703552246, "learning_rate": 0.0002505738693467336, "loss": 1.6043, "step": 16900 }, { "epoch": 0.85, "grad_norm": 1.3199384212493896, "learning_rate": 0.0002502723618090452, "loss": 1.6102, "step": 17000 }, { "epoch": 0.85, "eval_loss": 1.577386736869812, "eval_runtime": 37.4173, "eval_samples_per_second": 26.726, "eval_steps_per_second": 3.341, "step": 17000 }, { "epoch": 0.85, "grad_norm": 1.777269721031189, "learning_rate": 0.00024997085427135675, "loss": 1.6064, "step": 17100 }, { "epoch": 0.86, "grad_norm": 2.2964231967926025, "learning_rate": 0.0002496693467336683, "loss": 1.5964, "step": 17200 }, { "epoch": 0.86, "grad_norm": 1.7127012014389038, "learning_rate": 0.00024936783919597986, "loss": 1.6334, "step": 17300 }, { "epoch": 0.87, "grad_norm": 1.442181944847107, "learning_rate": 0.0002490663316582914, "loss": 1.5679, "step": 17400 }, { "epoch": 0.88, "grad_norm": 1.7092599868774414, "learning_rate": 0.000248764824120603, "loss": 1.6125, "step": 17500 }, { "epoch": 0.88, "grad_norm": 1.8017587661743164, "learning_rate": 0.00024846331658291454, "loss": 1.6386, "step": 17600 }, { "epoch": 0.89, "grad_norm": 3.3435380458831787, "learning_rate": 0.0002481618090452261, "loss": 1.5632, "step": 17700 }, { "epoch": 0.89, "grad_norm": 1.2036772966384888, "learning_rate": 0.00024786030150753766, "loss": 1.5972, "step": 17800 }, { "epoch": 0.9, "grad_norm": 2.212369203567505, "learning_rate": 0.0002475587939698492, "loss": 1.5777, "step": 17900 }, { "epoch": 0.9, "grad_norm": 1.9149020910263062, "learning_rate": 0.0002472572864321608, "loss": 1.601, "step": 18000 }, { "epoch": 0.9, "eval_loss": 1.5726821422576904, "eval_runtime": 37.4487, "eval_samples_per_second": 26.703, "eval_steps_per_second": 3.338, "step": 18000 }, { "epoch": 0.91, "grad_norm": 2.2173101902008057, "learning_rate": 0.00024695577889447234, "loss": 1.5975, "step": 18100 }, { "epoch": 0.91, "grad_norm": 2.0151243209838867, "learning_rate": 0.0002466542713567839, "loss": 1.5975, "step": 18200 }, { "epoch": 0.92, "grad_norm": 1.6275769472122192, "learning_rate": 0.00024635276381909546, "loss": 1.5909, "step": 18300 }, { "epoch": 0.92, "grad_norm": 0.8248587846755981, "learning_rate": 0.000246051256281407, "loss": 1.5954, "step": 18400 }, { "epoch": 0.93, "grad_norm": 2.119255542755127, "learning_rate": 0.0002457497487437186, "loss": 1.5651, "step": 18500 }, { "epoch": 0.93, "grad_norm": 1.9526431560516357, "learning_rate": 0.00024544824120603014, "loss": 1.6184, "step": 18600 }, { "epoch": 0.94, "grad_norm": 2.407723903656006, "learning_rate": 0.0002451467336683417, "loss": 1.6181, "step": 18700 }, { "epoch": 0.94, "grad_norm": 1.526853084564209, "learning_rate": 0.00024484522613065326, "loss": 1.6172, "step": 18800 }, { "epoch": 0.94, "grad_norm": 1.4161404371261597, "learning_rate": 0.0002445437185929648, "loss": 1.6154, "step": 18900 }, { "epoch": 0.95, "grad_norm": 1.3028178215026855, "learning_rate": 0.0002442422110552764, "loss": 1.6053, "step": 19000 }, { "epoch": 0.95, "eval_loss": 1.5818778276443481, "eval_runtime": 37.4111, "eval_samples_per_second": 26.73, "eval_steps_per_second": 3.341, "step": 19000 }, { "epoch": 0.95, "grad_norm": 1.6454105377197266, "learning_rate": 0.0002439407035175879, "loss": 1.6149, "step": 19100 }, { "epoch": 0.96, "grad_norm": 1.664665699005127, "learning_rate": 0.00024364221105527636, "loss": 1.6178, "step": 19200 }, { "epoch": 0.96, "grad_norm": 1.53481125831604, "learning_rate": 0.00024334070351758792, "loss": 1.5858, "step": 19300 }, { "epoch": 0.97, "grad_norm": 6.695281982421875, "learning_rate": 0.00024303919597989948, "loss": 1.5717, "step": 19400 }, { "epoch": 0.97, "grad_norm": 1.6920980215072632, "learning_rate": 0.00024273768844221104, "loss": 1.5754, "step": 19500 }, { "epoch": 0.98, "grad_norm": 2.0135791301727295, "learning_rate": 0.0002424361809045226, "loss": 1.6219, "step": 19600 }, { "epoch": 0.98, "grad_norm": 1.3980337381362915, "learning_rate": 0.00024213467336683413, "loss": 1.5913, "step": 19700 }, { "epoch": 0.99, "grad_norm": 1.8416180610656738, "learning_rate": 0.0002418331658291457, "loss": 1.5691, "step": 19800 }, { "epoch": 0.99, "grad_norm": 1.366356372833252, "learning_rate": 0.00024153165829145728, "loss": 1.5804, "step": 19900 }, { "epoch": 1.0, "grad_norm": 1.3517309427261353, "learning_rate": 0.00024123015075376884, "loss": 1.5603, "step": 20000 }, { "epoch": 1.0, "eval_loss": 1.6314821243286133, "eval_runtime": 37.4553, "eval_samples_per_second": 26.698, "eval_steps_per_second": 3.337, "step": 20000 }, { "epoch": 1.0, "grad_norm": 1.8194776773452759, "learning_rate": 0.00024092864321608037, "loss": 1.5964, "step": 20100 }, { "epoch": 1.01, "grad_norm": 3.187936305999756, "learning_rate": 0.00024062713567839193, "loss": 1.5926, "step": 20200 }, { "epoch": 1.01, "grad_norm": 1.2052127122879028, "learning_rate": 0.00024032562814070351, "loss": 1.5727, "step": 20300 }, { "epoch": 1.02, "grad_norm": 1.6609078645706177, "learning_rate": 0.00024002412060301505, "loss": 1.5967, "step": 20400 }, { "epoch": 1.02, "grad_norm": 1.502246379852295, "learning_rate": 0.0002397226130653266, "loss": 1.5708, "step": 20500 }, { "epoch": 1.03, "grad_norm": 2.180826425552368, "learning_rate": 0.00023942110552763817, "loss": 1.5688, "step": 20600 }, { "epoch": 1.03, "grad_norm": 2.257434606552124, "learning_rate": 0.00023911959798994975, "loss": 1.5263, "step": 20700 }, { "epoch": 1.04, "grad_norm": 1.9630309343338013, "learning_rate": 0.00023881809045226128, "loss": 1.5776, "step": 20800 }, { "epoch": 1.04, "grad_norm": 2.248621940612793, "learning_rate": 0.00023851658291457284, "loss": 1.6, "step": 20900 }, { "epoch": 1.05, "grad_norm": 1.0489450693130493, "learning_rate": 0.0002382150753768844, "loss": 1.5699, "step": 21000 }, { "epoch": 1.05, "eval_loss": 1.5522246360778809, "eval_runtime": 37.4533, "eval_samples_per_second": 26.7, "eval_steps_per_second": 3.337, "step": 21000 }, { "epoch": 1.05, "grad_norm": 1.457306981086731, "learning_rate": 0.00023791356783919596, "loss": 1.5953, "step": 21100 }, { "epoch": 1.06, "grad_norm": 2.088200807571411, "learning_rate": 0.00023761206030150752, "loss": 1.5881, "step": 21200 }, { "epoch": 1.06, "grad_norm": 1.4078480005264282, "learning_rate": 0.00023731356783919598, "loss": 1.5746, "step": 21300 }, { "epoch": 1.07, "grad_norm": 1.5498270988464355, "learning_rate": 0.00023701206030150753, "loss": 1.5878, "step": 21400 }, { "epoch": 1.07, "grad_norm": 1.8573285341262817, "learning_rate": 0.00023671055276381907, "loss": 1.574, "step": 21500 }, { "epoch": 1.08, "grad_norm": 1.3603477478027344, "learning_rate": 0.00023640904522613063, "loss": 1.5994, "step": 21600 }, { "epoch": 1.08, "grad_norm": 1.6878479719161987, "learning_rate": 0.0002361075376884422, "loss": 1.63, "step": 21700 }, { "epoch": 1.09, "grad_norm": 1.879296064376831, "learning_rate": 0.00023580603015075375, "loss": 1.6313, "step": 21800 }, { "epoch": 1.09, "grad_norm": 1.695983648300171, "learning_rate": 0.0002355045226130653, "loss": 1.5837, "step": 21900 }, { "epoch": 1.1, "grad_norm": 2.6149425506591797, "learning_rate": 0.00023520301507537686, "loss": 1.5967, "step": 22000 }, { "epoch": 1.1, "eval_loss": 1.6031874418258667, "eval_runtime": 37.522, "eval_samples_per_second": 26.651, "eval_steps_per_second": 3.331, "step": 22000 }, { "epoch": 1.1, "grad_norm": 2.0208663940429688, "learning_rate": 0.00023490150753768845, "loss": 1.5691, "step": 22100 }, { "epoch": 1.11, "grad_norm": 2.202256202697754, "learning_rate": 0.00023459999999999998, "loss": 1.5815, "step": 22200 }, { "epoch": 1.11, "grad_norm": 1.9692652225494385, "learning_rate": 0.00023429849246231154, "loss": 1.5854, "step": 22300 }, { "epoch": 1.12, "grad_norm": 1.9030089378356934, "learning_rate": 0.00023399698492462308, "loss": 1.5937, "step": 22400 }, { "epoch": 1.12, "grad_norm": 2.265805244445801, "learning_rate": 0.00023369547738693463, "loss": 1.5409, "step": 22500 }, { "epoch": 1.13, "grad_norm": 1.9824811220169067, "learning_rate": 0.00023339396984924622, "loss": 1.5977, "step": 22600 }, { "epoch": 1.14, "grad_norm": 1.3028334379196167, "learning_rate": 0.00023309246231155778, "loss": 1.573, "step": 22700 }, { "epoch": 1.14, "grad_norm": 2.472731590270996, "learning_rate": 0.0002327909547738693, "loss": 1.5962, "step": 22800 }, { "epoch": 1.15, "grad_norm": 1.1267619132995605, "learning_rate": 0.00023248944723618087, "loss": 1.5871, "step": 22900 }, { "epoch": 1.15, "grad_norm": 1.7546107769012451, "learning_rate": 0.00023218793969849246, "loss": 1.6053, "step": 23000 }, { "epoch": 1.15, "eval_loss": 1.5985630750656128, "eval_runtime": 37.3462, "eval_samples_per_second": 26.777, "eval_steps_per_second": 3.347, "step": 23000 }, { "epoch": 1.16, "grad_norm": 1.7325044870376587, "learning_rate": 0.000231886432160804, "loss": 1.557, "step": 23100 }, { "epoch": 1.16, "grad_norm": 1.771462321281433, "learning_rate": 0.00023158492462311555, "loss": 1.6102, "step": 23200 }, { "epoch": 1.17, "grad_norm": 1.077540397644043, "learning_rate": 0.0002312834170854271, "loss": 1.5607, "step": 23300 }, { "epoch": 1.17, "grad_norm": 1.2537496089935303, "learning_rate": 0.0002309819095477387, "loss": 1.5801, "step": 23400 }, { "epoch": 1.18, "grad_norm": 1.5025357007980347, "learning_rate": 0.00023068040201005023, "loss": 1.583, "step": 23500 }, { "epoch": 1.18, "grad_norm": 1.2420939207077026, "learning_rate": 0.0002303788944723618, "loss": 1.569, "step": 23600 }, { "epoch": 1.19, "grad_norm": 2.900120735168457, "learning_rate": 0.00023007738693467335, "loss": 1.5561, "step": 23700 }, { "epoch": 1.19, "grad_norm": 1.6182540655136108, "learning_rate": 0.0002297758793969849, "loss": 1.5673, "step": 23800 }, { "epoch": 1.2, "grad_norm": 1.5824536085128784, "learning_rate": 0.00022947437185929647, "loss": 1.5603, "step": 23900 }, { "epoch": 1.2, "grad_norm": 1.1365340948104858, "learning_rate": 0.00022917286432160803, "loss": 1.571, "step": 24000 }, { "epoch": 1.2, "eval_loss": 1.5782357454299927, "eval_runtime": 37.4619, "eval_samples_per_second": 26.694, "eval_steps_per_second": 3.337, "step": 24000 }, { "epoch": 1.21, "grad_norm": 1.8765733242034912, "learning_rate": 0.00022887135678391956, "loss": 1.5628, "step": 24100 }, { "epoch": 1.21, "grad_norm": 2.7541840076446533, "learning_rate": 0.00022856984924623115, "loss": 1.5967, "step": 24200 }, { "epoch": 1.22, "grad_norm": 1.2741538286209106, "learning_rate": 0.0002282683417085427, "loss": 1.5142, "step": 24300 }, { "epoch": 1.22, "grad_norm": 1.5907307863235474, "learning_rate": 0.00022796683417085426, "loss": 1.5685, "step": 24400 }, { "epoch": 1.23, "grad_norm": 2.187331438064575, "learning_rate": 0.0002276653266331658, "loss": 1.5788, "step": 24500 }, { "epoch": 1.23, "grad_norm": 2.0679118633270264, "learning_rate": 0.00022736381909547736, "loss": 1.5585, "step": 24600 }, { "epoch": 1.23, "grad_norm": 1.2467107772827148, "learning_rate": 0.00022706231155778894, "loss": 1.5531, "step": 24700 }, { "epoch": 1.24, "grad_norm": 1.334625005722046, "learning_rate": 0.00022676080402010047, "loss": 1.5831, "step": 24800 }, { "epoch": 1.25, "grad_norm": 1.466030240058899, "learning_rate": 0.00022645929648241203, "loss": 1.5753, "step": 24900 }, { "epoch": 1.25, "grad_norm": 2.8223164081573486, "learning_rate": 0.0002261577889447236, "loss": 1.6131, "step": 25000 }, { "epoch": 1.25, "eval_loss": 1.5768604278564453, "eval_runtime": 37.5229, "eval_samples_per_second": 26.65, "eval_steps_per_second": 3.331, "step": 25000 }, { "epoch": 1.25, "grad_norm": 1.1590194702148438, "learning_rate": 0.00022585628140703518, "loss": 1.5831, "step": 25100 }, { "epoch": 1.26, "grad_norm": 1.1502777338027954, "learning_rate": 0.00022555778894472358, "loss": 1.5673, "step": 25200 }, { "epoch": 1.27, "grad_norm": 1.8671566247940063, "learning_rate": 0.00022525628140703517, "loss": 1.5632, "step": 25300 }, { "epoch": 1.27, "grad_norm": 2.4084360599517822, "learning_rate": 0.00022495477386934673, "loss": 1.5727, "step": 25400 }, { "epoch": 1.27, "grad_norm": 1.3032163381576538, "learning_rate": 0.00022465326633165826, "loss": 1.5754, "step": 25500 }, { "epoch": 1.28, "grad_norm": 2.2420716285705566, "learning_rate": 0.00022435175879396982, "loss": 1.5802, "step": 25600 }, { "epoch": 1.28, "grad_norm": 1.3116205930709839, "learning_rate": 0.0002240502512562814, "loss": 1.5744, "step": 25700 }, { "epoch": 1.29, "grad_norm": 1.0008848905563354, "learning_rate": 0.00022374874371859294, "loss": 1.5423, "step": 25800 }, { "epoch": 1.29, "grad_norm": 2.07833194732666, "learning_rate": 0.0002234472361809045, "loss": 1.5657, "step": 25900 }, { "epoch": 1.3, "grad_norm": 1.894939661026001, "learning_rate": 0.00022314572864321605, "loss": 1.5464, "step": 26000 }, { "epoch": 1.3, "eval_loss": 1.568109154701233, "eval_runtime": 37.5055, "eval_samples_per_second": 26.663, "eval_steps_per_second": 3.333, "step": 26000 }, { "epoch": 1.3, "grad_norm": 2.3820385932922363, "learning_rate": 0.00022284422110552764, "loss": 1.562, "step": 26100 }, { "epoch": 1.31, "grad_norm": 1.6219903230667114, "learning_rate": 0.00022254271356783917, "loss": 1.5683, "step": 26200 }, { "epoch": 1.31, "grad_norm": 1.721933126449585, "learning_rate": 0.00022224120603015073, "loss": 1.5646, "step": 26300 }, { "epoch": 1.32, "grad_norm": 1.8133726119995117, "learning_rate": 0.0002219396984924623, "loss": 1.5574, "step": 26400 }, { "epoch": 1.32, "grad_norm": 1.871902585029602, "learning_rate": 0.00022163819095477385, "loss": 1.5728, "step": 26500 }, { "epoch": 1.33, "grad_norm": 2.7765748500823975, "learning_rate": 0.0002213366834170854, "loss": 1.5768, "step": 26600 }, { "epoch": 1.33, "grad_norm": 1.848578691482544, "learning_rate": 0.00022103819095477386, "loss": 1.5721, "step": 26700 }, { "epoch": 1.34, "grad_norm": 1.944535732269287, "learning_rate": 0.00022073668341708542, "loss": 1.5722, "step": 26800 }, { "epoch": 1.34, "grad_norm": 1.5262278318405151, "learning_rate": 0.00022043517587939696, "loss": 1.5642, "step": 26900 }, { "epoch": 1.35, "grad_norm": 2.425851821899414, "learning_rate": 0.00022013366834170852, "loss": 1.5441, "step": 27000 }, { "epoch": 1.35, "eval_loss": 1.5537927150726318, "eval_runtime": 37.516, "eval_samples_per_second": 26.655, "eval_steps_per_second": 3.332, "step": 27000 }, { "epoch": 1.35, "grad_norm": 1.2145837545394897, "learning_rate": 0.0002198321608040201, "loss": 1.5994, "step": 27100 }, { "epoch": 1.36, "grad_norm": 2.361107349395752, "learning_rate": 0.00021953065326633163, "loss": 1.5551, "step": 27200 }, { "epoch": 1.36, "grad_norm": 1.1575865745544434, "learning_rate": 0.0002192291457286432, "loss": 1.5749, "step": 27300 }, { "epoch": 1.37, "grad_norm": 1.6274192333221436, "learning_rate": 0.00021892763819095475, "loss": 1.5729, "step": 27400 }, { "epoch": 1.38, "grad_norm": 1.660609245300293, "learning_rate": 0.00021862613065326629, "loss": 1.577, "step": 27500 }, { "epoch": 1.38, "grad_norm": 2.8537380695343018, "learning_rate": 0.00021832462311557787, "loss": 1.5962, "step": 27600 }, { "epoch": 1.39, "grad_norm": 7.63838529586792, "learning_rate": 0.00021802311557788943, "loss": 1.5462, "step": 27700 }, { "epoch": 1.39, "grad_norm": 3.0265047550201416, "learning_rate": 0.000217721608040201, "loss": 1.6263, "step": 27800 }, { "epoch": 1.4, "grad_norm": 1.552452802658081, "learning_rate": 0.00021742010050251252, "loss": 1.5683, "step": 27900 }, { "epoch": 1.4, "grad_norm": 1.2861206531524658, "learning_rate": 0.0002171185929648241, "loss": 1.5776, "step": 28000 }, { "epoch": 1.4, "eval_loss": 1.5543726682662964, "eval_runtime": 37.5463, "eval_samples_per_second": 26.634, "eval_steps_per_second": 3.329, "step": 28000 }, { "epoch": 1.41, "grad_norm": 2.410630941390991, "learning_rate": 0.00021681708542713567, "loss": 1.5513, "step": 28100 }, { "epoch": 1.41, "grad_norm": 1.8043034076690674, "learning_rate": 0.0002165155778894472, "loss": 1.594, "step": 28200 }, { "epoch": 1.42, "grad_norm": 1.8414337635040283, "learning_rate": 0.00021621407035175876, "loss": 1.5547, "step": 28300 }, { "epoch": 1.42, "grad_norm": 1.9315450191497803, "learning_rate": 0.00021591256281407035, "loss": 1.5387, "step": 28400 }, { "epoch": 1.43, "grad_norm": 2.3810365200042725, "learning_rate": 0.0002156110552763819, "loss": 1.564, "step": 28500 }, { "epoch": 1.43, "grad_norm": 1.4747709035873413, "learning_rate": 0.00021530954773869344, "loss": 1.5325, "step": 28600 }, { "epoch": 1.44, "grad_norm": 1.264954924583435, "learning_rate": 0.000215008040201005, "loss": 1.5716, "step": 28700 }, { "epoch": 1.44, "grad_norm": 2.2243235111236572, "learning_rate": 0.00021470653266331659, "loss": 1.5388, "step": 28800 }, { "epoch": 1.45, "grad_norm": 4.794454097747803, "learning_rate": 0.00021440502512562812, "loss": 1.5404, "step": 28900 }, { "epoch": 1.45, "grad_norm": 2.5027246475219727, "learning_rate": 0.00021410351758793968, "loss": 1.5592, "step": 29000 }, { "epoch": 1.45, "eval_loss": 1.569115400314331, "eval_runtime": 37.5972, "eval_samples_per_second": 26.598, "eval_steps_per_second": 3.325, "step": 29000 }, { "epoch": 1.46, "grad_norm": 2.8168818950653076, "learning_rate": 0.00021380201005025124, "loss": 1.5781, "step": 29100 }, { "epoch": 1.46, "grad_norm": 1.4402812719345093, "learning_rate": 0.0002135005025125628, "loss": 1.5626, "step": 29200 }, { "epoch": 1.47, "grad_norm": 1.5837584733963013, "learning_rate": 0.00021319899497487436, "loss": 1.6064, "step": 29300 }, { "epoch": 1.47, "grad_norm": 1.435542106628418, "learning_rate": 0.00021289748743718592, "loss": 1.5621, "step": 29400 }, { "epoch": 1.48, "grad_norm": 1.8227218389511108, "learning_rate": 0.00021259597989949745, "loss": 1.5481, "step": 29500 }, { "epoch": 1.48, "grad_norm": 1.9331644773483276, "learning_rate": 0.00021229447236180903, "loss": 1.5811, "step": 29600 }, { "epoch": 1.48, "grad_norm": 2.971740484237671, "learning_rate": 0.0002119929648241206, "loss": 1.5412, "step": 29700 }, { "epoch": 1.49, "grad_norm": 3.569145917892456, "learning_rate": 0.00021169145728643215, "loss": 1.5508, "step": 29800 }, { "epoch": 1.5, "grad_norm": 2.1709325313568115, "learning_rate": 0.00021138994974874369, "loss": 1.5802, "step": 29900 }, { "epoch": 1.5, "grad_norm": 1.8602783679962158, "learning_rate": 0.00021108844221105524, "loss": 1.5607, "step": 30000 }, { "epoch": 1.5, "eval_loss": 1.5550180673599243, "eval_runtime": 37.5792, "eval_samples_per_second": 26.61, "eval_steps_per_second": 3.326, "step": 30000 }, { "epoch": 1.5, "grad_norm": 1.3990256786346436, "learning_rate": 0.00021078693467336683, "loss": 1.5611, "step": 30100 }, { "epoch": 1.51, "grad_norm": 4.223874568939209, "learning_rate": 0.00021048542713567836, "loss": 1.6039, "step": 30200 }, { "epoch": 1.52, "grad_norm": 1.763484001159668, "learning_rate": 0.00021018391959798992, "loss": 1.5938, "step": 30300 }, { "epoch": 1.52, "grad_norm": 2.2013938426971436, "learning_rate": 0.00020988241206030148, "loss": 1.5883, "step": 30400 }, { "epoch": 1.52, "grad_norm": 2.446477174758911, "learning_rate": 0.00020958391959798994, "loss": 1.5289, "step": 30500 }, { "epoch": 1.53, "grad_norm": 1.6682789325714111, "learning_rate": 0.00020928241206030147, "loss": 1.5689, "step": 30600 }, { "epoch": 1.54, "grad_norm": 1.8999930620193481, "learning_rate": 0.00020898090452261305, "loss": 1.568, "step": 30700 }, { "epoch": 1.54, "grad_norm": 2.189378261566162, "learning_rate": 0.00020867939698492461, "loss": 1.5363, "step": 30800 }, { "epoch": 1.54, "grad_norm": 2.609349489212036, "learning_rate": 0.00020837788944723615, "loss": 1.5523, "step": 30900 }, { "epoch": 1.55, "grad_norm": 2.2627103328704834, "learning_rate": 0.0002080763819095477, "loss": 1.574, "step": 31000 }, { "epoch": 1.55, "eval_loss": 1.5256458520889282, "eval_runtime": 37.5514, "eval_samples_per_second": 26.63, "eval_steps_per_second": 3.329, "step": 31000 }, { "epoch": 1.56, "grad_norm": 1.9266563653945923, "learning_rate": 0.0002077748743718593, "loss": 1.5387, "step": 31100 }, { "epoch": 1.56, "grad_norm": 1.8447223901748657, "learning_rate": 0.00020747336683417085, "loss": 1.5632, "step": 31200 }, { "epoch": 1.56, "grad_norm": 1.375827670097351, "learning_rate": 0.00020717185929648238, "loss": 1.5494, "step": 31300 }, { "epoch": 1.57, "grad_norm": 2.164782762527466, "learning_rate": 0.00020687035175879394, "loss": 1.5468, "step": 31400 }, { "epoch": 1.57, "grad_norm": 1.186018943786621, "learning_rate": 0.00020656884422110553, "loss": 1.584, "step": 31500 }, { "epoch": 1.58, "grad_norm": 5.939981460571289, "learning_rate": 0.00020626733668341706, "loss": 1.5602, "step": 31600 }, { "epoch": 1.58, "grad_norm": 2.185728073120117, "learning_rate": 0.00020596582914572862, "loss": 1.5301, "step": 31700 }, { "epoch": 1.59, "grad_norm": 1.4321199655532837, "learning_rate": 0.00020566432160804018, "loss": 1.5308, "step": 31800 }, { "epoch": 1.59, "grad_norm": 1.8272013664245605, "learning_rate": 0.00020536281407035177, "loss": 1.5828, "step": 31900 }, { "epoch": 1.6, "grad_norm": 1.3199920654296875, "learning_rate": 0.0002050613065326633, "loss": 1.5704, "step": 32000 }, { "epoch": 1.6, "eval_loss": 1.536350131034851, "eval_runtime": 37.4489, "eval_samples_per_second": 26.703, "eval_steps_per_second": 3.338, "step": 32000 }, { "epoch": 1.6, "grad_norm": 1.796789526939392, "learning_rate": 0.00020475979899497486, "loss": 1.56, "step": 32100 }, { "epoch": 1.61, "grad_norm": 1.4835096597671509, "learning_rate": 0.0002044582914572864, "loss": 1.5409, "step": 32200 }, { "epoch": 1.61, "grad_norm": 1.225799798965454, "learning_rate": 0.00020415678391959798, "loss": 1.5109, "step": 32300 }, { "epoch": 1.62, "grad_norm": 1.5552102327346802, "learning_rate": 0.00020385527638190954, "loss": 1.5717, "step": 32400 }, { "epoch": 1.62, "grad_norm": 1.3638191223144531, "learning_rate": 0.0002035537688442211, "loss": 1.5763, "step": 32500 }, { "epoch": 1.63, "grad_norm": 2.177093505859375, "learning_rate": 0.00020325226130653263, "loss": 1.5173, "step": 32600 }, { "epoch": 1.64, "grad_norm": 3.643524169921875, "learning_rate": 0.0002029507537688442, "loss": 1.5596, "step": 32700 }, { "epoch": 1.64, "grad_norm": 1.2808345556259155, "learning_rate": 0.00020264924623115578, "loss": 1.5812, "step": 32800 }, { "epoch": 1.65, "grad_norm": 2.262430191040039, "learning_rate": 0.0002023477386934673, "loss": 1.5424, "step": 32900 }, { "epoch": 1.65, "grad_norm": 1.8705729246139526, "learning_rate": 0.00020204623115577887, "loss": 1.5552, "step": 33000 }, { "epoch": 1.65, "eval_loss": 1.5897144079208374, "eval_runtime": 37.4808, "eval_samples_per_second": 26.68, "eval_steps_per_second": 3.335, "step": 33000 }, { "epoch": 1.66, "grad_norm": 1.52475106716156, "learning_rate": 0.00020174472361809043, "loss": 1.5737, "step": 33100 }, { "epoch": 1.66, "grad_norm": 1.8177305459976196, "learning_rate": 0.00020144321608040201, "loss": 1.5638, "step": 33200 }, { "epoch": 1.67, "grad_norm": 1.8988004922866821, "learning_rate": 0.00020114170854271355, "loss": 1.54, "step": 33300 }, { "epoch": 1.67, "grad_norm": 1.61077880859375, "learning_rate": 0.0002008402010050251, "loss": 1.5463, "step": 33400 }, { "epoch": 1.68, "grad_norm": 1.3590441942214966, "learning_rate": 0.00020053869346733667, "loss": 1.5646, "step": 33500 }, { "epoch": 1.68, "grad_norm": 1.5618160963058472, "learning_rate": 0.00020023718592964822, "loss": 1.5316, "step": 33600 }, { "epoch": 1.69, "grad_norm": 1.7087860107421875, "learning_rate": 0.00019993567839195978, "loss": 1.5429, "step": 33700 }, { "epoch": 1.69, "grad_norm": 2.558692693710327, "learning_rate": 0.00019963417085427134, "loss": 1.558, "step": 33800 }, { "epoch": 1.69, "grad_norm": 2.3594534397125244, "learning_rate": 0.00019933266331658288, "loss": 1.5453, "step": 33900 }, { "epoch": 1.7, "grad_norm": 1.57821786403656, "learning_rate": 0.00019903115577889446, "loss": 1.5541, "step": 34000 }, { "epoch": 1.7, "eval_loss": 1.5565224885940552, "eval_runtime": 37.4415, "eval_samples_per_second": 26.708, "eval_steps_per_second": 3.339, "step": 34000 }, { "epoch": 1.71, "grad_norm": 1.9368255138397217, "learning_rate": 0.00019872964824120602, "loss": 1.5808, "step": 34100 }, { "epoch": 1.71, "grad_norm": 3.0767080783843994, "learning_rate": 0.00019842814070351758, "loss": 1.5613, "step": 34200 }, { "epoch": 1.71, "grad_norm": 1.8630317449569702, "learning_rate": 0.00019812663316582911, "loss": 1.5638, "step": 34300 }, { "epoch": 1.72, "grad_norm": 3.0488193035125732, "learning_rate": 0.0001978251256281407, "loss": 1.5615, "step": 34400 }, { "epoch": 1.73, "grad_norm": 3.8991503715515137, "learning_rate": 0.00019752361809045226, "loss": 1.4857, "step": 34500 }, { "epoch": 1.73, "grad_norm": 2.0849859714508057, "learning_rate": 0.0001972221105527638, "loss": 1.535, "step": 34600 }, { "epoch": 1.73, "grad_norm": 1.825913906097412, "learning_rate": 0.00019692060301507535, "loss": 1.5664, "step": 34700 }, { "epoch": 1.74, "grad_norm": 5.08195686340332, "learning_rate": 0.0001966190954773869, "loss": 1.5252, "step": 34800 }, { "epoch": 1.75, "grad_norm": 1.785659909248352, "learning_rate": 0.0001963175879396985, "loss": 1.5391, "step": 34900 }, { "epoch": 1.75, "grad_norm": 1.4543670415878296, "learning_rate": 0.00019601909547738692, "loss": 1.5484, "step": 35000 }, { "epoch": 1.75, "eval_loss": 1.573486566543579, "eval_runtime": 37.5549, "eval_samples_per_second": 26.628, "eval_steps_per_second": 3.328, "step": 35000 }, { "epoch": 1.75, "grad_norm": 1.6768901348114014, "learning_rate": 0.00019571758793969848, "loss": 1.5479, "step": 35100 }, { "epoch": 1.76, "grad_norm": 1.9926207065582275, "learning_rate": 0.00019541608040201004, "loss": 1.5712, "step": 35200 }, { "epoch": 1.77, "grad_norm": 1.9285818338394165, "learning_rate": 0.00019511457286432157, "loss": 1.567, "step": 35300 }, { "epoch": 1.77, "grad_norm": 4.1938252449035645, "learning_rate": 0.00019481306532663313, "loss": 1.5538, "step": 35400 }, { "epoch": 1.77, "grad_norm": 3.3867177963256836, "learning_rate": 0.00019451155778894472, "loss": 1.532, "step": 35500 }, { "epoch": 1.78, "grad_norm": 2.6039962768554688, "learning_rate": 0.00019421005025125625, "loss": 1.5914, "step": 35600 }, { "epoch": 1.79, "grad_norm": 1.900150179862976, "learning_rate": 0.0001939085427135678, "loss": 1.5435, "step": 35700 }, { "epoch": 1.79, "grad_norm": 6.110165596008301, "learning_rate": 0.00019360703517587937, "loss": 1.5532, "step": 35800 }, { "epoch": 1.79, "grad_norm": 2.2953763008117676, "learning_rate": 0.00019330552763819096, "loss": 1.5573, "step": 35900 }, { "epoch": 1.8, "grad_norm": 1.9452694654464722, "learning_rate": 0.0001930040201005025, "loss": 1.5602, "step": 36000 }, { "epoch": 1.8, "eval_loss": 1.5185086727142334, "eval_runtime": 37.4604, "eval_samples_per_second": 26.695, "eval_steps_per_second": 3.337, "step": 36000 }, { "epoch": 1.81, "grad_norm": 1.729576587677002, "learning_rate": 0.00019270251256281405, "loss": 1.5778, "step": 36100 }, { "epoch": 1.81, "grad_norm": 2.98600172996521, "learning_rate": 0.0001924010050251256, "loss": 1.5318, "step": 36200 }, { "epoch": 1.81, "grad_norm": 2.2745935916900635, "learning_rate": 0.00019209949748743717, "loss": 1.5475, "step": 36300 }, { "epoch": 1.82, "grad_norm": 1.7677953243255615, "learning_rate": 0.00019179798994974873, "loss": 1.5174, "step": 36400 }, { "epoch": 1.82, "grad_norm": 3.6853749752044678, "learning_rate": 0.0001914964824120603, "loss": 1.5709, "step": 36500 }, { "epoch": 1.83, "grad_norm": 2.8981404304504395, "learning_rate": 0.00019119497487437182, "loss": 1.5519, "step": 36600 }, { "epoch": 1.83, "grad_norm": 1.9862598180770874, "learning_rate": 0.0001908934673366834, "loss": 1.5547, "step": 36700 }, { "epoch": 1.84, "grad_norm": 2.030618667602539, "learning_rate": 0.00019059195979899497, "loss": 1.5388, "step": 36800 }, { "epoch": 1.84, "grad_norm": 1.609573483467102, "learning_rate": 0.00019029045226130653, "loss": 1.5727, "step": 36900 }, { "epoch": 1.85, "grad_norm": 2.0508673191070557, "learning_rate": 0.00018998894472361806, "loss": 1.5522, "step": 37000 }, { "epoch": 1.85, "eval_loss": 1.4889146089553833, "eval_runtime": 37.4776, "eval_samples_per_second": 26.683, "eval_steps_per_second": 3.335, "step": 37000 }, { "epoch": 1.85, "grad_norm": 1.5507237911224365, "learning_rate": 0.00018968743718592964, "loss": 1.5196, "step": 37100 }, { "epoch": 1.86, "grad_norm": 2.1241345405578613, "learning_rate": 0.0001893859296482412, "loss": 1.553, "step": 37200 }, { "epoch": 1.86, "grad_norm": 1.6903210878372192, "learning_rate": 0.00018908442211055274, "loss": 1.5292, "step": 37300 }, { "epoch": 1.87, "grad_norm": 1.684809923171997, "learning_rate": 0.0001887829145728643, "loss": 1.5597, "step": 37400 }, { "epoch": 1.88, "grad_norm": 2.072158098220825, "learning_rate": 0.00018848442211055275, "loss": 1.4914, "step": 37500 }, { "epoch": 1.88, "grad_norm": 2.545748472213745, "learning_rate": 0.0001881829145728643, "loss": 1.54, "step": 37600 }, { "epoch": 1.89, "grad_norm": 2.4186930656433105, "learning_rate": 0.00018788140703517587, "loss": 1.5298, "step": 37700 }, { "epoch": 1.89, "grad_norm": 1.8362512588500977, "learning_rate": 0.00018757989949748743, "loss": 1.5372, "step": 37800 }, { "epoch": 1.9, "grad_norm": 2.376615047454834, "learning_rate": 0.000187278391959799, "loss": 1.5343, "step": 37900 }, { "epoch": 1.9, "grad_norm": 2.72920823097229, "learning_rate": 0.00018697688442211052, "loss": 1.4955, "step": 38000 }, { "epoch": 1.9, "eval_loss": 1.5501998662948608, "eval_runtime": 37.4829, "eval_samples_per_second": 26.679, "eval_steps_per_second": 3.335, "step": 38000 }, { "epoch": 1.91, "grad_norm": 2.2700889110565186, "learning_rate": 0.00018667537688442208, "loss": 1.5149, "step": 38100 }, { "epoch": 1.91, "grad_norm": 2.9030048847198486, "learning_rate": 0.00018637386934673367, "loss": 1.5451, "step": 38200 }, { "epoch": 1.92, "grad_norm": 3.4646873474121094, "learning_rate": 0.00018607236180904522, "loss": 1.5477, "step": 38300 }, { "epoch": 1.92, "grad_norm": 2.5595543384552, "learning_rate": 0.00018577085427135676, "loss": 1.5013, "step": 38400 }, { "epoch": 1.93, "grad_norm": 1.3651905059814453, "learning_rate": 0.00018546934673366832, "loss": 1.5199, "step": 38500 }, { "epoch": 1.93, "grad_norm": 2.045830011367798, "learning_rate": 0.0001851678391959799, "loss": 1.5677, "step": 38600 }, { "epoch": 1.94, "grad_norm": 1.8075040578842163, "learning_rate": 0.00018486633165829144, "loss": 1.5564, "step": 38700 }, { "epoch": 1.94, "grad_norm": 1.4847893714904785, "learning_rate": 0.000184564824120603, "loss": 1.5178, "step": 38800 }, { "epoch": 1.94, "grad_norm": 3.2269062995910645, "learning_rate": 0.00018426331658291455, "loss": 1.513, "step": 38900 }, { "epoch": 1.95, "grad_norm": 1.8105851411819458, "learning_rate": 0.00018396180904522614, "loss": 1.5007, "step": 39000 }, { "epoch": 1.95, "eval_loss": 1.5438071489334106, "eval_runtime": 37.4427, "eval_samples_per_second": 26.707, "eval_steps_per_second": 3.338, "step": 39000 }, { "epoch": 1.96, "grad_norm": 1.8713701963424683, "learning_rate": 0.00018366030150753767, "loss": 1.5266, "step": 39100 }, { "epoch": 1.96, "grad_norm": 1.4177159070968628, "learning_rate": 0.00018335879396984923, "loss": 1.5345, "step": 39200 }, { "epoch": 1.96, "grad_norm": 2.068962574005127, "learning_rate": 0.00018305728643216076, "loss": 1.5136, "step": 39300 }, { "epoch": 1.97, "grad_norm": 2.0991291999816895, "learning_rate": 0.00018275577889447235, "loss": 1.5509, "step": 39400 }, { "epoch": 1.98, "grad_norm": 2.322105884552002, "learning_rate": 0.0001824542713567839, "loss": 1.5222, "step": 39500 }, { "epoch": 1.98, "grad_norm": 2.1995983123779297, "learning_rate": 0.00018215276381909547, "loss": 1.544, "step": 39600 }, { "epoch": 1.98, "grad_norm": 1.4002470970153809, "learning_rate": 0.000181851256281407, "loss": 1.5184, "step": 39700 }, { "epoch": 1.99, "grad_norm": 2.0827932357788086, "learning_rate": 0.0001815497487437186, "loss": 1.5388, "step": 39800 }, { "epoch": 2.0, "grad_norm": 3.9671342372894287, "learning_rate": 0.00018124824120603015, "loss": 1.5331, "step": 39900 }, { "epoch": 2.0, "grad_norm": 2.979598045349121, "learning_rate": 0.00018094673366834168, "loss": 1.5058, "step": 40000 }, { "epoch": 2.0, "eval_loss": 1.5161113739013672, "eval_runtime": 37.4278, "eval_samples_per_second": 26.718, "eval_steps_per_second": 3.34, "step": 40000 }, { "epoch": 2.0, "grad_norm": 2.0998573303222656, "learning_rate": 0.00018064824120603013, "loss": 1.4831, "step": 40100 }, { "epoch": 2.01, "grad_norm": 2.1450562477111816, "learning_rate": 0.0001803467336683417, "loss": 1.5567, "step": 40200 }, { "epoch": 2.02, "grad_norm": 2.080376148223877, "learning_rate": 0.00018004522613065325, "loss": 1.5088, "step": 40300 }, { "epoch": 2.02, "grad_norm": 1.682288408279419, "learning_rate": 0.0001797437185929648, "loss": 1.5461, "step": 40400 }, { "epoch": 2.02, "grad_norm": 1.6313551664352417, "learning_rate": 0.00017944221105527637, "loss": 1.4776, "step": 40500 }, { "epoch": 2.03, "grad_norm": 3.1746320724487305, "learning_rate": 0.00017914070351758793, "loss": 1.534, "step": 40600 }, { "epoch": 2.04, "grad_norm": 1.7039170265197754, "learning_rate": 0.00017883919597989946, "loss": 1.565, "step": 40700 }, { "epoch": 2.04, "grad_norm": 1.738771677017212, "learning_rate": 0.00017853768844221102, "loss": 1.5391, "step": 40800 }, { "epoch": 2.04, "grad_norm": 4.0762715339660645, "learning_rate": 0.0001782361809045226, "loss": 1.5153, "step": 40900 }, { "epoch": 2.05, "grad_norm": 3.6206607818603516, "learning_rate": 0.00017793467336683417, "loss": 1.5254, "step": 41000 }, { "epoch": 2.05, "eval_loss": 1.5652326345443726, "eval_runtime": 37.621, "eval_samples_per_second": 26.581, "eval_steps_per_second": 3.323, "step": 41000 }, { "epoch": 2.06, "grad_norm": 1.2320189476013184, "learning_rate": 0.0001776331658291457, "loss": 1.5153, "step": 41100 }, { "epoch": 2.06, "grad_norm": 2.275785446166992, "learning_rate": 0.00017733165829145726, "loss": 1.5192, "step": 41200 }, { "epoch": 2.06, "grad_norm": 1.8659756183624268, "learning_rate": 0.00017703015075376885, "loss": 1.5563, "step": 41300 }, { "epoch": 2.07, "grad_norm": 1.6176475286483765, "learning_rate": 0.00017672864321608038, "loss": 1.5251, "step": 41400 }, { "epoch": 2.08, "grad_norm": 1.575243592262268, "learning_rate": 0.00017642713567839194, "loss": 1.5178, "step": 41500 }, { "epoch": 2.08, "grad_norm": 1.6838304996490479, "learning_rate": 0.0001761256281407035, "loss": 1.5255, "step": 41600 }, { "epoch": 2.08, "grad_norm": 5.561213493347168, "learning_rate": 0.00017582412060301509, "loss": 1.4885, "step": 41700 }, { "epoch": 2.09, "grad_norm": 2.0810487270355225, "learning_rate": 0.00017552261306532662, "loss": 1.491, "step": 41800 }, { "epoch": 2.1, "grad_norm": 1.894136667251587, "learning_rate": 0.00017522110552763818, "loss": 1.5072, "step": 41900 }, { "epoch": 2.1, "grad_norm": 4.531317710876465, "learning_rate": 0.00017492261306532663, "loss": 1.5108, "step": 42000 }, { "epoch": 2.1, "eval_loss": 1.5298963785171509, "eval_runtime": 37.5318, "eval_samples_per_second": 26.644, "eval_steps_per_second": 3.331, "step": 42000 }, { "epoch": 2.1, "grad_norm": 1.9595602750778198, "learning_rate": 0.00017462110552763816, "loss": 1.5178, "step": 42100 }, { "epoch": 2.11, "grad_norm": 2.1969220638275146, "learning_rate": 0.00017431959798994972, "loss": 1.508, "step": 42200 }, { "epoch": 2.12, "grad_norm": 3.0334370136260986, "learning_rate": 0.0001740180904522613, "loss": 1.5255, "step": 42300 }, { "epoch": 2.12, "grad_norm": 1.8148131370544434, "learning_rate": 0.00017371658291457287, "loss": 1.4686, "step": 42400 }, { "epoch": 2.12, "grad_norm": 5.726377964019775, "learning_rate": 0.0001734150753768844, "loss": 1.5355, "step": 42500 }, { "epoch": 2.13, "grad_norm": 3.15856671333313, "learning_rate": 0.00017311356783919596, "loss": 1.534, "step": 42600 }, { "epoch": 2.13, "grad_norm": 6.238559246063232, "learning_rate": 0.00017281206030150755, "loss": 1.5073, "step": 42700 }, { "epoch": 2.14, "grad_norm": 2.281912326812744, "learning_rate": 0.00017251055276381908, "loss": 1.5341, "step": 42800 }, { "epoch": 2.15, "grad_norm": 3.874361991882324, "learning_rate": 0.00017220904522613064, "loss": 1.5204, "step": 42900 }, { "epoch": 2.15, "grad_norm": 2.5632574558258057, "learning_rate": 0.0001719075376884422, "loss": 1.5028, "step": 43000 }, { "epoch": 2.15, "eval_loss": 1.5213427543640137, "eval_runtime": 37.5103, "eval_samples_per_second": 26.659, "eval_steps_per_second": 3.332, "step": 43000 }, { "epoch": 2.15, "grad_norm": 4.4505133628845215, "learning_rate": 0.00017160603015075373, "loss": 1.5318, "step": 43100 }, { "epoch": 2.16, "grad_norm": 1.6099460124969482, "learning_rate": 0.00017130452261306532, "loss": 1.5093, "step": 43200 }, { "epoch": 2.17, "grad_norm": 4.269103527069092, "learning_rate": 0.00017100301507537688, "loss": 1.4839, "step": 43300 }, { "epoch": 2.17, "grad_norm": 10.070590019226074, "learning_rate": 0.0001707015075376884, "loss": 1.5407, "step": 43400 }, { "epoch": 2.17, "grad_norm": 17.754796981811523, "learning_rate": 0.00017039999999999997, "loss": 1.5102, "step": 43500 }, { "epoch": 2.18, "grad_norm": 2.8812096118927, "learning_rate": 0.00017009849246231155, "loss": 1.5227, "step": 43600 }, { "epoch": 2.19, "grad_norm": 2.861893653869629, "learning_rate": 0.00016979698492462311, "loss": 1.544, "step": 43700 }, { "epoch": 2.19, "grad_norm": 1.9935343265533447, "learning_rate": 0.00016949547738693465, "loss": 1.5209, "step": 43800 }, { "epoch": 2.19, "grad_norm": 2.8261117935180664, "learning_rate": 0.0001691939698492462, "loss": 1.5404, "step": 43900 }, { "epoch": 2.2, "grad_norm": 2.6541242599487305, "learning_rate": 0.0001688924623115578, "loss": 1.5048, "step": 44000 }, { "epoch": 2.2, "eval_loss": 1.5309374332427979, "eval_runtime": 65.7795, "eval_samples_per_second": 15.202, "eval_steps_per_second": 1.9, "step": 44000 }, { "epoch": 2.21, "grad_norm": 1.1707357168197632, "learning_rate": 0.00016859095477386932, "loss": 1.5479, "step": 44100 }, { "epoch": 2.21, "grad_norm": 2.058530807495117, "learning_rate": 0.00016828944723618088, "loss": 1.5423, "step": 44200 }, { "epoch": 2.21, "grad_norm": 2.841702699661255, "learning_rate": 0.00016798793969849244, "loss": 1.5196, "step": 44300 }, { "epoch": 2.22, "grad_norm": 1.6679179668426514, "learning_rate": 0.00016768643216080403, "loss": 1.5085, "step": 44400 }, { "epoch": 2.23, "grad_norm": 2.1461362838745117, "learning_rate": 0.00016738492462311556, "loss": 1.538, "step": 44500 }, { "epoch": 2.23, "grad_norm": 2.351161241531372, "learning_rate": 0.00016708341708542712, "loss": 1.4874, "step": 44600 }, { "epoch": 2.23, "grad_norm": 2.474755048751831, "learning_rate": 0.00016678190954773868, "loss": 1.5646, "step": 44700 }, { "epoch": 2.24, "grad_norm": 2.6611087322235107, "learning_rate": 0.00016648040201005024, "loss": 1.5373, "step": 44800 }, { "epoch": 2.25, "grad_norm": 2.518184185028076, "learning_rate": 0.0001661788944723618, "loss": 1.4969, "step": 44900 }, { "epoch": 2.25, "grad_norm": 4.304960250854492, "learning_rate": 0.00016587738693467336, "loss": 1.4827, "step": 45000 }, { "epoch": 2.25, "eval_loss": 1.5653032064437866, "eval_runtime": 39.6553, "eval_samples_per_second": 25.217, "eval_steps_per_second": 3.152, "step": 45000 }, { "epoch": 2.25, "grad_norm": 1.1882766485214233, "learning_rate": 0.0001655758793969849, "loss": 1.5062, "step": 45100 }, { "epoch": 2.26, "grad_norm": 2.196139335632324, "learning_rate": 0.00016527437185929648, "loss": 1.5026, "step": 45200 }, { "epoch": 2.27, "grad_norm": 1.8797615766525269, "learning_rate": 0.00016497286432160804, "loss": 1.4947, "step": 45300 }, { "epoch": 2.27, "grad_norm": 1.396849274635315, "learning_rate": 0.00016467135678391957, "loss": 1.5002, "step": 45400 }, { "epoch": 2.27, "grad_norm": 2.432685375213623, "learning_rate": 0.00016436984924623113, "loss": 1.5021, "step": 45500 }, { "epoch": 2.28, "grad_norm": 2.2086715698242188, "learning_rate": 0.0001640683417085427, "loss": 1.5067, "step": 45600 }, { "epoch": 2.29, "grad_norm": 4.262127876281738, "learning_rate": 0.00016376683417085428, "loss": 1.519, "step": 45700 }, { "epoch": 2.29, "grad_norm": 2.92459774017334, "learning_rate": 0.0001634653266331658, "loss": 1.5138, "step": 45800 }, { "epoch": 2.29, "grad_norm": 1.3408390283584595, "learning_rate": 0.00016316381909547737, "loss": 1.5296, "step": 45900 }, { "epoch": 2.3, "grad_norm": 1.9604805707931519, "learning_rate": 0.00016286231155778893, "loss": 1.5207, "step": 46000 }, { "epoch": 2.3, "eval_loss": 1.5254641771316528, "eval_runtime": 42.9693, "eval_samples_per_second": 23.272, "eval_steps_per_second": 2.909, "step": 46000 }, { "epoch": 2.31, "grad_norm": 2.5174312591552734, "learning_rate": 0.00016256080402010049, "loss": 1.5487, "step": 46100 }, { "epoch": 2.31, "grad_norm": 1.4481089115142822, "learning_rate": 0.00016225929648241205, "loss": 1.5167, "step": 46200 }, { "epoch": 2.31, "grad_norm": 1.5994378328323364, "learning_rate": 0.0001619577889447236, "loss": 1.4549, "step": 46300 }, { "epoch": 2.32, "grad_norm": 2.4599764347076416, "learning_rate": 0.00016165628140703514, "loss": 1.5128, "step": 46400 }, { "epoch": 2.33, "grad_norm": 1.8390551805496216, "learning_rate": 0.00016135477386934672, "loss": 1.5248, "step": 46500 }, { "epoch": 2.33, "grad_norm": 2.679804801940918, "learning_rate": 0.00016105326633165828, "loss": 1.5105, "step": 46600 }, { "epoch": 2.33, "grad_norm": 1.7645354270935059, "learning_rate": 0.00016075477386934674, "loss": 1.4787, "step": 46700 }, { "epoch": 2.34, "grad_norm": 1.035243034362793, "learning_rate": 0.00016045326633165827, "loss": 1.5257, "step": 46800 }, { "epoch": 2.34, "grad_norm": 11.433294296264648, "learning_rate": 0.00016015175879396983, "loss": 1.4615, "step": 46900 }, { "epoch": 2.35, "grad_norm": 1.600629448890686, "learning_rate": 0.0001598502512562814, "loss": 1.4973, "step": 47000 }, { "epoch": 2.35, "eval_loss": 1.5015385150909424, "eval_runtime": 41.9852, "eval_samples_per_second": 23.818, "eval_steps_per_second": 2.977, "step": 47000 }, { "epoch": 2.35, "grad_norm": 2.1351780891418457, "learning_rate": 0.00015954874371859297, "loss": 1.5163, "step": 47100 }, { "epoch": 2.36, "grad_norm": 2.6234447956085205, "learning_rate": 0.0001592472361809045, "loss": 1.5147, "step": 47200 }, { "epoch": 2.37, "grad_norm": 2.0315120220184326, "learning_rate": 0.00015894874371859296, "loss": 1.5044, "step": 47300 }, { "epoch": 2.37, "grad_norm": 2.7582693099975586, "learning_rate": 0.00015864723618090452, "loss": 1.5033, "step": 47400 }, { "epoch": 2.38, "grad_norm": 2.5383968353271484, "learning_rate": 0.00015834572864321605, "loss": 1.5198, "step": 47500 }, { "epoch": 2.38, "grad_norm": 3.1361851692199707, "learning_rate": 0.0001580442211055276, "loss": 1.5006, "step": 47600 }, { "epoch": 2.38, "grad_norm": 1.669693946838379, "learning_rate": 0.0001577427135678392, "loss": 1.5203, "step": 47700 }, { "epoch": 2.39, "grad_norm": 3.662080764770508, "learning_rate": 0.00015744120603015076, "loss": 1.5427, "step": 47800 }, { "epoch": 2.4, "grad_norm": 2.101423740386963, "learning_rate": 0.0001571396984924623, "loss": 1.489, "step": 47900 }, { "epoch": 2.4, "grad_norm": 8.889337539672852, "learning_rate": 0.00015684120603015074, "loss": 1.5198, "step": 48000 }, { "epoch": 2.4, "eval_loss": 1.5436657667160034, "eval_runtime": 42.227, "eval_samples_per_second": 23.682, "eval_steps_per_second": 2.96, "step": 48000 }, { "epoch": 2.41, "grad_norm": 9.11814022064209, "learning_rate": 0.0001565396984924623, "loss": 1.5623, "step": 48100 }, { "epoch": 2.41, "grad_norm": 11.104165077209473, "learning_rate": 0.00015623819095477383, "loss": 1.5019, "step": 48200 }, { "epoch": 2.42, "grad_norm": 4.448848724365234, "learning_rate": 0.00015593668341708542, "loss": 1.5464, "step": 48300 }, { "epoch": 2.42, "grad_norm": 2.207139253616333, "learning_rate": 0.00015563517587939698, "loss": 1.5105, "step": 48400 }, { "epoch": 2.42, "grad_norm": 2.692350149154663, "learning_rate": 0.00015533366834170854, "loss": 1.5351, "step": 48500 }, { "epoch": 2.43, "grad_norm": 1.8738429546356201, "learning_rate": 0.00015503216080402007, "loss": 1.4882, "step": 48600 }, { "epoch": 2.44, "grad_norm": 2.8214309215545654, "learning_rate": 0.00015473065326633166, "loss": 1.4932, "step": 48700 }, { "epoch": 2.44, "grad_norm": 1.4316879510879517, "learning_rate": 0.00015442914572864322, "loss": 1.498, "step": 48800 }, { "epoch": 2.44, "grad_norm": 1.438391923904419, "learning_rate": 0.00015412763819095475, "loss": 1.4975, "step": 48900 }, { "epoch": 2.45, "grad_norm": 2.7039265632629395, "learning_rate": 0.0001538261306532663, "loss": 1.511, "step": 49000 }, { "epoch": 2.45, "eval_loss": 1.5066561698913574, "eval_runtime": 42.052, "eval_samples_per_second": 23.78, "eval_steps_per_second": 2.973, "step": 49000 }, { "epoch": 2.46, "grad_norm": 2.8025197982788086, "learning_rate": 0.00015352462311557787, "loss": 1.5082, "step": 49100 }, { "epoch": 2.46, "grad_norm": 4.876307964324951, "learning_rate": 0.00015322311557788946, "loss": 1.4825, "step": 49200 }, { "epoch": 2.46, "grad_norm": 4.34946346282959, "learning_rate": 0.000152921608040201, "loss": 1.5117, "step": 49300 }, { "epoch": 2.47, "grad_norm": 2.5481321811676025, "learning_rate": 0.00015262010050251255, "loss": 1.5361, "step": 49400 }, { "epoch": 2.48, "grad_norm": 1.5703433752059937, "learning_rate": 0.00015231859296482408, "loss": 1.5113, "step": 49500 }, { "epoch": 2.48, "grad_norm": 1.8963671922683716, "learning_rate": 0.00015201708542713567, "loss": 1.5095, "step": 49600 }, { "epoch": 2.48, "grad_norm": 2.296262264251709, "learning_rate": 0.00015171557788944723, "loss": 1.5479, "step": 49700 }, { "epoch": 2.49, "grad_norm": 2.796741247177124, "learning_rate": 0.00015141407035175879, "loss": 1.4946, "step": 49800 }, { "epoch": 2.5, "grad_norm": 2.633420705795288, "learning_rate": 0.00015111256281407032, "loss": 1.4917, "step": 49900 }, { "epoch": 2.5, "grad_norm": 3.0930418968200684, "learning_rate": 0.0001508110552763819, "loss": 1.5058, "step": 50000 }, { "epoch": 2.5, "eval_loss": 1.539115071296692, "eval_runtime": 41.8844, "eval_samples_per_second": 23.875, "eval_steps_per_second": 2.984, "step": 50000 }, { "epoch": 2.5, "grad_norm": 2.2942283153533936, "learning_rate": 0.00015050954773869346, "loss": 1.4981, "step": 50100 }, { "epoch": 2.51, "grad_norm": 2.622481346130371, "learning_rate": 0.000150208040201005, "loss": 1.4829, "step": 50200 }, { "epoch": 2.52, "grad_norm": 2.5901362895965576, "learning_rate": 0.00014990653266331658, "loss": 1.495, "step": 50300 }, { "epoch": 2.52, "grad_norm": 3.8583192825317383, "learning_rate": 0.00014960502512562812, "loss": 1.494, "step": 50400 }, { "epoch": 2.52, "grad_norm": 2.841306209564209, "learning_rate": 0.00014930653266331657, "loss": 1.4806, "step": 50500 }, { "epoch": 2.53, "grad_norm": 2.646027088165283, "learning_rate": 0.00014900502512562813, "loss": 1.4907, "step": 50600 }, { "epoch": 2.54, "grad_norm": 2.384547472000122, "learning_rate": 0.0001487035175879397, "loss": 1.4809, "step": 50700 }, { "epoch": 2.54, "grad_norm": 2.698951005935669, "learning_rate": 0.00014840201005025125, "loss": 1.5145, "step": 50800 }, { "epoch": 2.54, "grad_norm": 4.710977077484131, "learning_rate": 0.0001481035175879397, "loss": 1.4805, "step": 50900 }, { "epoch": 2.55, "grad_norm": 5.228128433227539, "learning_rate": 0.00014780201005025123, "loss": 1.512, "step": 51000 }, { "epoch": 2.55, "eval_loss": 1.5644181966781616, "eval_runtime": 37.1723, "eval_samples_per_second": 26.902, "eval_steps_per_second": 3.363, "step": 51000 }, { "epoch": 2.56, "grad_norm": 5.459705829620361, "learning_rate": 0.00014750050251256282, "loss": 1.4861, "step": 51100 }, { "epoch": 2.56, "grad_norm": 1.9696797132492065, "learning_rate": 0.00014719899497487435, "loss": 1.5177, "step": 51200 }, { "epoch": 2.56, "grad_norm": 2.262742757797241, "learning_rate": 0.0001468974874371859, "loss": 1.5074, "step": 51300 }, { "epoch": 2.57, "grad_norm": 3.0286054611206055, "learning_rate": 0.00014659597989949747, "loss": 1.5233, "step": 51400 }, { "epoch": 2.58, "grad_norm": 2.805699110031128, "learning_rate": 0.00014629447236180903, "loss": 1.5002, "step": 51500 }, { "epoch": 2.58, "grad_norm": 7.009899616241455, "learning_rate": 0.0001459929648241206, "loss": 1.5224, "step": 51600 }, { "epoch": 2.58, "grad_norm": 2.203697443008423, "learning_rate": 0.00014569145728643215, "loss": 1.5264, "step": 51700 }, { "epoch": 2.59, "grad_norm": 38.17327117919922, "learning_rate": 0.0001453899497487437, "loss": 1.5233, "step": 51800 }, { "epoch": 2.59, "grad_norm": 3.463594436645508, "learning_rate": 0.00014508844221105527, "loss": 1.5476, "step": 51900 }, { "epoch": 2.6, "grad_norm": 4.695711135864258, "learning_rate": 0.00014478693467336683, "loss": 1.5092, "step": 52000 }, { "epoch": 2.6, "eval_loss": 1.524036169052124, "eval_runtime": 41.6949, "eval_samples_per_second": 23.984, "eval_steps_per_second": 2.998, "step": 52000 }, { "epoch": 2.6, "grad_norm": 3.0102827548980713, "learning_rate": 0.00014448542713567836, "loss": 1.4978, "step": 52100 }, { "epoch": 2.61, "grad_norm": 4.087903022766113, "learning_rate": 0.00014418391959798995, "loss": 1.4772, "step": 52200 }, { "epoch": 2.62, "grad_norm": 2.3847885131835938, "learning_rate": 0.00014388241206030148, "loss": 1.5173, "step": 52300 }, { "epoch": 2.62, "grad_norm": 2.907341241836548, "learning_rate": 0.00014358090452261306, "loss": 1.5287, "step": 52400 }, { "epoch": 2.62, "grad_norm": 3.0708484649658203, "learning_rate": 0.0001432793969849246, "loss": 1.524, "step": 52500 }, { "epoch": 2.63, "grad_norm": 2.70914888381958, "learning_rate": 0.00014297788944723618, "loss": 1.5115, "step": 52600 }, { "epoch": 2.63, "grad_norm": 3.0721583366394043, "learning_rate": 0.00014267638190954772, "loss": 1.5273, "step": 52700 }, { "epoch": 2.64, "grad_norm": 4.067818641662598, "learning_rate": 0.00014237487437185927, "loss": 1.5237, "step": 52800 }, { "epoch": 2.65, "grad_norm": 22.84881591796875, "learning_rate": 0.00014207336683417083, "loss": 1.489, "step": 52900 }, { "epoch": 2.65, "grad_norm": 4.021462440490723, "learning_rate": 0.0001417718592964824, "loss": 1.5284, "step": 53000 }, { "epoch": 2.65, "eval_loss": 1.5171101093292236, "eval_runtime": 41.2537, "eval_samples_per_second": 24.24, "eval_steps_per_second": 3.03, "step": 53000 }, { "epoch": 2.66, "grad_norm": 14.866168975830078, "learning_rate": 0.00014147035175879395, "loss": 1.5114, "step": 53100 }, { "epoch": 2.66, "grad_norm": 8.153103828430176, "learning_rate": 0.0001411688442211055, "loss": 1.4384, "step": 53200 }, { "epoch": 2.67, "grad_norm": 3.6942877769470215, "learning_rate": 0.00014086733668341707, "loss": 1.5189, "step": 53300 }, { "epoch": 2.67, "grad_norm": 1.7417048215866089, "learning_rate": 0.00014056582914572863, "loss": 1.5491, "step": 53400 }, { "epoch": 2.67, "grad_norm": 3.291808605194092, "learning_rate": 0.0001402643216080402, "loss": 1.5085, "step": 53500 }, { "epoch": 2.68, "grad_norm": 4.061951637268066, "learning_rate": 0.00013996281407035175, "loss": 1.5328, "step": 53600 }, { "epoch": 2.69, "grad_norm": 2.119297742843628, "learning_rate": 0.0001396613065326633, "loss": 1.4872, "step": 53700 }, { "epoch": 2.69, "grad_norm": 3.23767352104187, "learning_rate": 0.00013935979899497487, "loss": 1.5208, "step": 53800 }, { "epoch": 2.69, "grad_norm": 4.745997428894043, "learning_rate": 0.00013905829145728643, "loss": 1.5251, "step": 53900 }, { "epoch": 2.7, "grad_norm": 1.6231697797775269, "learning_rate": 0.000138756783919598, "loss": 1.5151, "step": 54000 }, { "epoch": 2.7, "eval_loss": 1.4974777698516846, "eval_runtime": 37.2945, "eval_samples_per_second": 26.814, "eval_steps_per_second": 3.352, "step": 54000 }, { "epoch": 2.71, "grad_norm": 2.7751882076263428, "learning_rate": 0.00013845527638190955, "loss": 1.4891, "step": 54100 }, { "epoch": 2.71, "grad_norm": 3.2454652786254883, "learning_rate": 0.00013815376884422108, "loss": 1.4929, "step": 54200 }, { "epoch": 2.71, "grad_norm": 2.5808818340301514, "learning_rate": 0.00013785226130653264, "loss": 1.5728, "step": 54300 }, { "epoch": 2.72, "grad_norm": 3.754495859146118, "learning_rate": 0.0001375507537688442, "loss": 1.5073, "step": 54400 }, { "epoch": 2.73, "grad_norm": 2.158400535583496, "learning_rate": 0.00013724924623115576, "loss": 1.5014, "step": 54500 }, { "epoch": 2.73, "grad_norm": 2.022975206375122, "learning_rate": 0.00013694773869346732, "loss": 1.5147, "step": 54600 }, { "epoch": 2.73, "grad_norm": 2.061910629272461, "learning_rate": 0.00013664623115577888, "loss": 1.4701, "step": 54700 }, { "epoch": 2.74, "grad_norm": 2.8262743949890137, "learning_rate": 0.00013634472361809044, "loss": 1.4988, "step": 54800 }, { "epoch": 2.75, "grad_norm": 2.389618158340454, "learning_rate": 0.000136043216080402, "loss": 1.5138, "step": 54900 }, { "epoch": 2.75, "grad_norm": 2.31430983543396, "learning_rate": 0.00013574170854271356, "loss": 1.5234, "step": 55000 }, { "epoch": 2.75, "eval_loss": 1.4553519487380981, "eval_runtime": 43.1448, "eval_samples_per_second": 23.178, "eval_steps_per_second": 2.897, "step": 55000 }, { "epoch": 2.75, "grad_norm": 2.14939022064209, "learning_rate": 0.00013544020100502512, "loss": 1.4971, "step": 55100 }, { "epoch": 2.76, "grad_norm": 2.3362553119659424, "learning_rate": 0.00013513869346733667, "loss": 1.4899, "step": 55200 }, { "epoch": 2.77, "grad_norm": 1.8975802659988403, "learning_rate": 0.00013483718592964823, "loss": 1.4855, "step": 55300 }, { "epoch": 2.77, "grad_norm": 1.6080349683761597, "learning_rate": 0.0001345356783919598, "loss": 1.4789, "step": 55400 }, { "epoch": 2.77, "grad_norm": 2.4170632362365723, "learning_rate": 0.00013423417085427135, "loss": 1.5052, "step": 55500 }, { "epoch": 2.78, "grad_norm": 3.0807037353515625, "learning_rate": 0.0001339326633165829, "loss": 1.4848, "step": 55600 }, { "epoch": 2.79, "grad_norm": 2.7296719551086426, "learning_rate": 0.00013363115577889447, "loss": 1.4695, "step": 55700 }, { "epoch": 2.79, "grad_norm": 2.392073154449463, "learning_rate": 0.000133329648241206, "loss": 1.4757, "step": 55800 }, { "epoch": 2.79, "grad_norm": 1.3958042860031128, "learning_rate": 0.0001330281407035176, "loss": 1.5152, "step": 55900 }, { "epoch": 2.8, "grad_norm": 5.350344657897949, "learning_rate": 0.00013272663316582912, "loss": 1.4918, "step": 56000 }, { "epoch": 2.8, "eval_loss": 1.4755498170852661, "eval_runtime": 55.3712, "eval_samples_per_second": 18.06, "eval_steps_per_second": 2.257, "step": 56000 }, { "epoch": 2.81, "grad_norm": 2.9437155723571777, "learning_rate": 0.0001324251256281407, "loss": 1.4673, "step": 56100 }, { "epoch": 2.81, "grad_norm": 2.0883982181549072, "learning_rate": 0.00013212361809045224, "loss": 1.5435, "step": 56200 }, { "epoch": 2.81, "grad_norm": 1.475618600845337, "learning_rate": 0.00013182211055276383, "loss": 1.4895, "step": 56300 }, { "epoch": 2.82, "grad_norm": 3.1873602867126465, "learning_rate": 0.00013152060301507536, "loss": 1.4753, "step": 56400 }, { "epoch": 2.83, "grad_norm": 1.1994943618774414, "learning_rate": 0.00013121909547738692, "loss": 1.4627, "step": 56500 }, { "epoch": 2.83, "grad_norm": 2.3503193855285645, "learning_rate": 0.00013091758793969848, "loss": 1.5312, "step": 56600 }, { "epoch": 2.83, "grad_norm": 1.5025432109832764, "learning_rate": 0.00013061608040201004, "loss": 1.4981, "step": 56700 }, { "epoch": 2.84, "grad_norm": 1.2746458053588867, "learning_rate": 0.0001303145728643216, "loss": 1.4644, "step": 56800 }, { "epoch": 2.84, "grad_norm": 2.1974568367004395, "learning_rate": 0.00013001306532663316, "loss": 1.4958, "step": 56900 }, { "epoch": 2.85, "grad_norm": 2.137925863265991, "learning_rate": 0.00012971155778894472, "loss": 1.4808, "step": 57000 }, { "epoch": 2.85, "eval_loss": 1.4832957983016968, "eval_runtime": 58.8442, "eval_samples_per_second": 16.994, "eval_steps_per_second": 2.124, "step": 57000 }, { "epoch": 2.85, "grad_norm": 1.9320988655090332, "learning_rate": 0.00012941005025125628, "loss": 1.4718, "step": 57100 }, { "epoch": 2.86, "grad_norm": 2.317089080810547, "learning_rate": 0.00012910854271356784, "loss": 1.436, "step": 57200 }, { "epoch": 2.87, "grad_norm": 2.3655266761779785, "learning_rate": 0.00012880703517587937, "loss": 1.4967, "step": 57300 }, { "epoch": 2.87, "grad_norm": 2.1567230224609375, "learning_rate": 0.00012850552763819096, "loss": 1.4319, "step": 57400 }, { "epoch": 2.88, "grad_norm": 1.759560227394104, "learning_rate": 0.0001282040201005025, "loss": 1.4719, "step": 57500 }, { "epoch": 2.88, "grad_norm": 3.5496578216552734, "learning_rate": 0.00012790251256281407, "loss": 1.513, "step": 57600 }, { "epoch": 2.88, "grad_norm": 1.4334951639175415, "learning_rate": 0.0001276010050251256, "loss": 1.5037, "step": 57700 }, { "epoch": 2.89, "grad_norm": 1.5903607606887817, "learning_rate": 0.0001272994974874372, "loss": 1.4386, "step": 57800 }, { "epoch": 2.9, "grad_norm": 1.8693747520446777, "learning_rate": 0.00012699798994974873, "loss": 1.4691, "step": 57900 }, { "epoch": 2.9, "grad_norm": 1.4653220176696777, "learning_rate": 0.00012669648241206029, "loss": 1.4531, "step": 58000 }, { "epoch": 2.9, "eval_loss": 1.4672846794128418, "eval_runtime": 70.1651, "eval_samples_per_second": 14.252, "eval_steps_per_second": 1.782, "step": 58000 }, { "epoch": 2.91, "grad_norm": 1.9329131841659546, "learning_rate": 0.00012639497487437184, "loss": 1.4538, "step": 58100 }, { "epoch": 2.91, "grad_norm": 1.952959418296814, "learning_rate": 0.0001260934673366834, "loss": 1.4872, "step": 58200 }, { "epoch": 2.92, "grad_norm": 1.1974529027938843, "learning_rate": 0.00012579195979899496, "loss": 1.4487, "step": 58300 }, { "epoch": 2.92, "grad_norm": 1.7099848985671997, "learning_rate": 0.00012549045226130652, "loss": 1.4708, "step": 58400 }, { "epoch": 2.92, "grad_norm": 1.2982145547866821, "learning_rate": 0.00012518894472361808, "loss": 1.4524, "step": 58500 }, { "epoch": 2.93, "grad_norm": 1.275212049484253, "learning_rate": 0.00012488743718592964, "loss": 1.4771, "step": 58600 }, { "epoch": 2.94, "grad_norm": 1.4571611881256104, "learning_rate": 0.0001245859296482412, "loss": 1.4454, "step": 58700 }, { "epoch": 2.94, "grad_norm": 1.9355239868164062, "learning_rate": 0.00012428442211055273, "loss": 1.4873, "step": 58800 }, { "epoch": 2.94, "grad_norm": 1.2806047201156616, "learning_rate": 0.00012398291457286432, "loss": 1.4566, "step": 58900 }, { "epoch": 2.95, "grad_norm": 3.054337739944458, "learning_rate": 0.00012368442211055277, "loss": 1.5002, "step": 59000 }, { "epoch": 2.95, "eval_loss": 1.4360119104385376, "eval_runtime": 40.7087, "eval_samples_per_second": 24.565, "eval_steps_per_second": 3.071, "step": 59000 }, { "epoch": 2.96, "grad_norm": 3.0794198513031006, "learning_rate": 0.0001233829145728643, "loss": 1.4463, "step": 59100 }, { "epoch": 2.96, "grad_norm": 2.48854923248291, "learning_rate": 0.00012308140703517586, "loss": 1.4211, "step": 59200 }, { "epoch": 2.96, "grad_norm": 1.163191556930542, "learning_rate": 0.00012277989949748742, "loss": 1.4596, "step": 59300 }, { "epoch": 2.97, "grad_norm": 2.0918381214141846, "learning_rate": 0.00012247839195979898, "loss": 1.4713, "step": 59400 }, { "epoch": 2.98, "grad_norm": 2.207432746887207, "learning_rate": 0.00012217688442211054, "loss": 1.4747, "step": 59500 }, { "epoch": 2.98, "grad_norm": 2.485342502593994, "learning_rate": 0.00012187537688442209, "loss": 1.4531, "step": 59600 }, { "epoch": 2.98, "grad_norm": 1.9726412296295166, "learning_rate": 0.00012157386934673366, "loss": 1.4685, "step": 59700 }, { "epoch": 2.99, "grad_norm": 2.211527109146118, "learning_rate": 0.00012127236180904521, "loss": 1.488, "step": 59800 }, { "epoch": 3.0, "grad_norm": 2.2727925777435303, "learning_rate": 0.00012097085427135678, "loss": 1.4919, "step": 59900 }, { "epoch": 3.0, "grad_norm": 1.9899802207946777, "learning_rate": 0.00012067236180904521, "loss": 1.496, "step": 60000 }, { "epoch": 3.0, "eval_loss": 1.480151891708374, "eval_runtime": 37.1123, "eval_samples_per_second": 26.945, "eval_steps_per_second": 3.368, "step": 60000 }, { "epoch": 3.0, "grad_norm": 2.080214023590088, "learning_rate": 0.00012037085427135678, "loss": 1.4381, "step": 60100 }, { "epoch": 3.01, "grad_norm": 2.6185152530670166, "learning_rate": 0.00012006934673366833, "loss": 1.4701, "step": 60200 }, { "epoch": 3.02, "grad_norm": 2.33237361907959, "learning_rate": 0.0001197678391959799, "loss": 1.4754, "step": 60300 }, { "epoch": 3.02, "grad_norm": 1.6651071310043335, "learning_rate": 0.00011946633165829144, "loss": 1.4264, "step": 60400 }, { "epoch": 3.02, "grad_norm": 2.55290150642395, "learning_rate": 0.000119164824120603, "loss": 1.4643, "step": 60500 }, { "epoch": 3.03, "grad_norm": 2.1360106468200684, "learning_rate": 0.00011886331658291456, "loss": 1.4485, "step": 60600 }, { "epoch": 3.04, "grad_norm": 2.012885808944702, "learning_rate": 0.00011856180904522612, "loss": 1.4613, "step": 60700 }, { "epoch": 3.04, "grad_norm": 2.1006858348846436, "learning_rate": 0.00011826030150753767, "loss": 1.4536, "step": 60800 }, { "epoch": 3.04, "grad_norm": 1.3656415939331055, "learning_rate": 0.00011795879396984924, "loss": 1.452, "step": 60900 }, { "epoch": 3.05, "grad_norm": 2.107713460922241, "learning_rate": 0.00011765728643216079, "loss": 1.4506, "step": 61000 }, { "epoch": 3.05, "eval_loss": 1.4423640966415405, "eval_runtime": 36.6897, "eval_samples_per_second": 27.256, "eval_steps_per_second": 3.407, "step": 61000 }, { "epoch": 3.06, "grad_norm": 1.7318499088287354, "learning_rate": 0.00011735577889447236, "loss": 1.4439, "step": 61100 }, { "epoch": 3.06, "grad_norm": 2.1133487224578857, "learning_rate": 0.00011705427135678391, "loss": 1.4401, "step": 61200 }, { "epoch": 3.06, "grad_norm": 1.763225793838501, "learning_rate": 0.00011675276381909548, "loss": 1.4196, "step": 61300 }, { "epoch": 3.07, "grad_norm": 3.682509660720825, "learning_rate": 0.00011645125628140703, "loss": 1.4355, "step": 61400 }, { "epoch": 3.08, "grad_norm": 1.6236958503723145, "learning_rate": 0.00011614974874371859, "loss": 1.4288, "step": 61500 }, { "epoch": 3.08, "grad_norm": 2.428751230239868, "learning_rate": 0.00011584824120603014, "loss": 1.4445, "step": 61600 }, { "epoch": 3.08, "grad_norm": 1.9624316692352295, "learning_rate": 0.0001155467336683417, "loss": 1.4509, "step": 61700 }, { "epoch": 3.09, "grad_norm": 1.2408591508865356, "learning_rate": 0.00011524522613065325, "loss": 1.4604, "step": 61800 }, { "epoch": 3.1, "grad_norm": 1.9674237966537476, "learning_rate": 0.00011494371859296481, "loss": 1.4187, "step": 61900 }, { "epoch": 3.1, "grad_norm": 1.1599769592285156, "learning_rate": 0.00011464221105527637, "loss": 1.4393, "step": 62000 }, { "epoch": 3.1, "eval_loss": 1.4237370491027832, "eval_runtime": 36.8952, "eval_samples_per_second": 27.104, "eval_steps_per_second": 3.388, "step": 62000 }, { "epoch": 3.1, "grad_norm": 1.2391512393951416, "learning_rate": 0.00011434070351758793, "loss": 1.4673, "step": 62100 }, { "epoch": 3.11, "grad_norm": 1.8557316064834595, "learning_rate": 0.00011403919597989949, "loss": 1.423, "step": 62200 }, { "epoch": 3.12, "grad_norm": 2.8055520057678223, "learning_rate": 0.00011373768844221103, "loss": 1.4722, "step": 62300 }, { "epoch": 3.12, "grad_norm": 3.723037004470825, "learning_rate": 0.0001134361809045226, "loss": 1.4313, "step": 62400 }, { "epoch": 3.12, "grad_norm": 1.5366125106811523, "learning_rate": 0.00011313467336683415, "loss": 1.4678, "step": 62500 }, { "epoch": 3.13, "grad_norm": 1.5289653539657593, "learning_rate": 0.00011283316582914573, "loss": 1.4313, "step": 62600 }, { "epoch": 3.13, "grad_norm": 1.496334195137024, "learning_rate": 0.00011253165829145727, "loss": 1.4249, "step": 62700 }, { "epoch": 3.14, "grad_norm": 2.485732078552246, "learning_rate": 0.00011223015075376884, "loss": 1.4071, "step": 62800 }, { "epoch": 3.15, "grad_norm": 1.784316897392273, "learning_rate": 0.00011192864321608039, "loss": 1.458, "step": 62900 }, { "epoch": 3.15, "grad_norm": 1.126514196395874, "learning_rate": 0.00011162713567839195, "loss": 1.4606, "step": 63000 }, { "epoch": 3.15, "eval_loss": 1.4603298902511597, "eval_runtime": 36.816, "eval_samples_per_second": 27.162, "eval_steps_per_second": 3.395, "step": 63000 }, { "epoch": 3.15, "grad_norm": 1.4732190370559692, "learning_rate": 0.00011132562814070351, "loss": 1.4118, "step": 63100 }, { "epoch": 3.16, "grad_norm": 2.9530584812164307, "learning_rate": 0.00011102412060301507, "loss": 1.4413, "step": 63200 }, { "epoch": 3.17, "grad_norm": 1.5375559329986572, "learning_rate": 0.00011072261306532661, "loss": 1.4301, "step": 63300 }, { "epoch": 3.17, "grad_norm": 3.584829568862915, "learning_rate": 0.00011042110552763819, "loss": 1.4563, "step": 63400 }, { "epoch": 3.17, "grad_norm": 3.209752082824707, "learning_rate": 0.00011011959798994973, "loss": 1.4456, "step": 63500 }, { "epoch": 3.18, "grad_norm": 1.9197559356689453, "learning_rate": 0.0001098180904522613, "loss": 1.4324, "step": 63600 }, { "epoch": 3.19, "grad_norm": 1.6928168535232544, "learning_rate": 0.00010951658291457285, "loss": 1.4237, "step": 63700 }, { "epoch": 3.19, "grad_norm": 2.0929553508758545, "learning_rate": 0.00010921507537688443, "loss": 1.4348, "step": 63800 }, { "epoch": 3.19, "grad_norm": 1.1685271263122559, "learning_rate": 0.00010891356783919597, "loss": 1.4297, "step": 63900 }, { "epoch": 3.2, "grad_norm": 1.703434944152832, "learning_rate": 0.00010861206030150753, "loss": 1.4324, "step": 64000 }, { "epoch": 3.2, "eval_loss": 1.458383560180664, "eval_runtime": 37.3628, "eval_samples_per_second": 26.765, "eval_steps_per_second": 3.346, "step": 64000 }, { "epoch": 3.21, "grad_norm": 2.0976409912109375, "learning_rate": 0.00010831055276381909, "loss": 1.4132, "step": 64100 }, { "epoch": 3.21, "grad_norm": 2.9562056064605713, "learning_rate": 0.00010801206030150753, "loss": 1.4216, "step": 64200 }, { "epoch": 3.21, "grad_norm": 3.074629545211792, "learning_rate": 0.00010771055276381909, "loss": 1.4099, "step": 64300 }, { "epoch": 3.22, "grad_norm": 2.264169454574585, "learning_rate": 0.00010740904522613064, "loss": 1.4065, "step": 64400 }, { "epoch": 3.23, "grad_norm": 1.6872237920761108, "learning_rate": 0.00010710753768844221, "loss": 1.4264, "step": 64500 }, { "epoch": 3.23, "grad_norm": 1.8734939098358154, "learning_rate": 0.00010680603015075375, "loss": 1.4346, "step": 64600 }, { "epoch": 3.23, "grad_norm": 2.0195159912109375, "learning_rate": 0.00010650452261306531, "loss": 1.4557, "step": 64700 }, { "epoch": 3.24, "grad_norm": 1.7375565767288208, "learning_rate": 0.00010620301507537687, "loss": 1.4237, "step": 64800 }, { "epoch": 3.25, "grad_norm": 3.346041202545166, "learning_rate": 0.00010590452261306531, "loss": 1.4393, "step": 64900 }, { "epoch": 3.25, "grad_norm": 1.8216383457183838, "learning_rate": 0.00010560301507537687, "loss": 1.4029, "step": 65000 }, { "epoch": 3.25, "eval_loss": 1.4565457105636597, "eval_runtime": 36.7819, "eval_samples_per_second": 27.187, "eval_steps_per_second": 3.398, "step": 65000 }, { "epoch": 3.25, "grad_norm": 2.628929853439331, "learning_rate": 0.00010530150753768843, "loss": 1.3928, "step": 65100 }, { "epoch": 3.26, "grad_norm": 2.476813554763794, "learning_rate": 0.00010499999999999999, "loss": 1.4377, "step": 65200 }, { "epoch": 3.27, "grad_norm": 4.745122909545898, "learning_rate": 0.00010469849246231155, "loss": 1.4599, "step": 65300 }, { "epoch": 3.27, "grad_norm": 4.166439533233643, "learning_rate": 0.0001043969849246231, "loss": 1.4135, "step": 65400 }, { "epoch": 3.27, "grad_norm": 1.3662863969802856, "learning_rate": 0.00010409547738693467, "loss": 1.4398, "step": 65500 }, { "epoch": 3.28, "grad_norm": 1.9009519815444946, "learning_rate": 0.00010379396984924622, "loss": 1.4354, "step": 65600 }, { "epoch": 3.29, "grad_norm": 1.9361233711242676, "learning_rate": 0.00010349246231155779, "loss": 1.382, "step": 65700 }, { "epoch": 3.29, "grad_norm": 2.0850670337677, "learning_rate": 0.00010319095477386933, "loss": 1.3983, "step": 65800 }, { "epoch": 3.29, "grad_norm": 3.42114520072937, "learning_rate": 0.0001028894472361809, "loss": 1.414, "step": 65900 }, { "epoch": 3.3, "grad_norm": 1.9612314701080322, "learning_rate": 0.00010258793969849245, "loss": 1.4494, "step": 66000 }, { "epoch": 3.3, "eval_loss": 1.4201833009719849, "eval_runtime": 36.8374, "eval_samples_per_second": 27.146, "eval_steps_per_second": 3.393, "step": 66000 }, { "epoch": 3.31, "grad_norm": 2.3793907165527344, "learning_rate": 0.00010228643216080401, "loss": 1.4567, "step": 66100 }, { "epoch": 3.31, "grad_norm": 2.0631420612335205, "learning_rate": 0.00010198492462311557, "loss": 1.4246, "step": 66200 }, { "epoch": 3.31, "grad_norm": 1.973575234413147, "learning_rate": 0.00010168643216080401, "loss": 1.4315, "step": 66300 }, { "epoch": 3.32, "grad_norm": 4.41493034362793, "learning_rate": 0.00010138492462311557, "loss": 1.3869, "step": 66400 }, { "epoch": 3.33, "grad_norm": 4.229668617248535, "learning_rate": 0.00010108341708542713, "loss": 1.4064, "step": 66500 }, { "epoch": 3.33, "grad_norm": 2.8293817043304443, "learning_rate": 0.00010078190954773868, "loss": 1.4204, "step": 66600 }, { "epoch": 3.33, "grad_norm": 2.0787456035614014, "learning_rate": 0.00010048040201005025, "loss": 1.4011, "step": 66700 }, { "epoch": 3.34, "grad_norm": 2.429809093475342, "learning_rate": 0.0001001788944723618, "loss": 1.4205, "step": 66800 }, { "epoch": 3.34, "grad_norm": 1.428712248802185, "learning_rate": 9.987738693467337e-05, "loss": 1.4363, "step": 66900 }, { "epoch": 3.35, "grad_norm": 1.6250827312469482, "learning_rate": 9.957587939698491e-05, "loss": 1.4225, "step": 67000 }, { "epoch": 3.35, "eval_loss": 1.4025810956954956, "eval_runtime": 37.6269, "eval_samples_per_second": 26.577, "eval_steps_per_second": 3.322, "step": 67000 }, { "epoch": 3.35, "grad_norm": 16.908023834228516, "learning_rate": 9.927437185929649e-05, "loss": 1.4019, "step": 67100 }, { "epoch": 3.36, "grad_norm": 2.529090166091919, "learning_rate": 9.897286432160803e-05, "loss": 1.4536, "step": 67200 }, { "epoch": 3.37, "grad_norm": 2.6016106605529785, "learning_rate": 9.867135678391959e-05, "loss": 1.4144, "step": 67300 }, { "epoch": 3.37, "grad_norm": 6.646603584289551, "learning_rate": 9.836984924623115e-05, "loss": 1.4585, "step": 67400 }, { "epoch": 3.38, "grad_norm": 2.8519158363342285, "learning_rate": 9.80683417085427e-05, "loss": 1.4281, "step": 67500 }, { "epoch": 3.38, "grad_norm": 2.310377836227417, "learning_rate": 9.776683417085426e-05, "loss": 1.42, "step": 67600 }, { "epoch": 3.38, "grad_norm": 1.6958341598510742, "learning_rate": 9.746532663316582e-05, "loss": 1.4229, "step": 67700 }, { "epoch": 3.39, "grad_norm": 3.763411045074463, "learning_rate": 9.716381909547738e-05, "loss": 1.4212, "step": 67800 }, { "epoch": 3.4, "grad_norm": 3.6792852878570557, "learning_rate": 9.686231155778894e-05, "loss": 1.4438, "step": 67900 }, { "epoch": 3.4, "grad_norm": 1.9338295459747314, "learning_rate": 9.65608040201005e-05, "loss": 1.4285, "step": 68000 }, { "epoch": 3.4, "eval_loss": 1.4275307655334473, "eval_runtime": 36.7735, "eval_samples_per_second": 27.193, "eval_steps_per_second": 3.399, "step": 68000 }, { "epoch": 3.41, "grad_norm": 2.4784669876098633, "learning_rate": 9.625929648241204e-05, "loss": 1.426, "step": 68100 }, { "epoch": 3.41, "grad_norm": 2.1132450103759766, "learning_rate": 9.595778894472361e-05, "loss": 1.4243, "step": 68200 }, { "epoch": 3.42, "grad_norm": 1.0781810283660889, "learning_rate": 9.565628140703516e-05, "loss": 1.3787, "step": 68300 }, { "epoch": 3.42, "grad_norm": 1.9404023885726929, "learning_rate": 9.535477386934673e-05, "loss": 1.4554, "step": 68400 }, { "epoch": 3.42, "grad_norm": 2.5888288021087646, "learning_rate": 9.505326633165828e-05, "loss": 1.4224, "step": 68500 }, { "epoch": 3.43, "grad_norm": 1.2692792415618896, "learning_rate": 9.475175879396985e-05, "loss": 1.4262, "step": 68600 }, { "epoch": 3.44, "grad_norm": 3.69474196434021, "learning_rate": 9.44502512562814e-05, "loss": 1.4348, "step": 68700 }, { "epoch": 3.44, "grad_norm": 2.908108949661255, "learning_rate": 9.414874371859296e-05, "loss": 1.3915, "step": 68800 }, { "epoch": 3.44, "grad_norm": 2.764848470687866, "learning_rate": 9.384723618090452e-05, "loss": 1.4007, "step": 68900 }, { "epoch": 3.45, "grad_norm": 2.2583723068237305, "learning_rate": 9.354572864321608e-05, "loss": 1.4274, "step": 69000 }, { "epoch": 3.45, "eval_loss": 1.4281996488571167, "eval_runtime": 36.7261, "eval_samples_per_second": 27.229, "eval_steps_per_second": 3.404, "step": 69000 }, { "epoch": 3.46, "grad_norm": 2.0189828872680664, "learning_rate": 9.324422110552762e-05, "loss": 1.3972, "step": 69100 }, { "epoch": 3.46, "grad_norm": 2.173034191131592, "learning_rate": 9.29427135678392e-05, "loss": 1.431, "step": 69200 }, { "epoch": 3.46, "grad_norm": 6.977257251739502, "learning_rate": 9.264120603015074e-05, "loss": 1.422, "step": 69300 }, { "epoch": 3.47, "grad_norm": 1.9481549263000488, "learning_rate": 9.233969849246231e-05, "loss": 1.424, "step": 69400 }, { "epoch": 3.48, "grad_norm": 1.5347398519515991, "learning_rate": 9.203819095477386e-05, "loss": 1.4017, "step": 69500 }, { "epoch": 3.48, "grad_norm": 4.210050582885742, "learning_rate": 9.173668341708543e-05, "loss": 1.4088, "step": 69600 }, { "epoch": 3.48, "grad_norm": 1.4653583765029907, "learning_rate": 9.143517587939698e-05, "loss": 1.4136, "step": 69700 }, { "epoch": 3.49, "grad_norm": 1.9992530345916748, "learning_rate": 9.113366834170852e-05, "loss": 1.4389, "step": 69800 }, { "epoch": 3.5, "grad_norm": 2.986983060836792, "learning_rate": 9.08321608040201e-05, "loss": 1.4207, "step": 69900 }, { "epoch": 3.5, "grad_norm": 3.8553383350372314, "learning_rate": 9.053065326633164e-05, "loss": 1.4158, "step": 70000 }, { "epoch": 3.5, "eval_loss": 1.4086616039276123, "eval_runtime": 44.1292, "eval_samples_per_second": 22.661, "eval_steps_per_second": 2.833, "step": 70000 }, { "epoch": 3.5, "grad_norm": 1.801664113998413, "learning_rate": 9.022914572864322e-05, "loss": 1.4129, "step": 70100 }, { "epoch": 3.51, "grad_norm": 4.6173529624938965, "learning_rate": 8.992763819095476e-05, "loss": 1.4052, "step": 70200 }, { "epoch": 3.52, "grad_norm": 2.9409828186035156, "learning_rate": 8.962914572864322e-05, "loss": 1.3987, "step": 70300 }, { "epoch": 3.52, "grad_norm": 1.4580490589141846, "learning_rate": 8.932763819095476e-05, "loss": 1.4278, "step": 70400 }, { "epoch": 3.52, "grad_norm": 3.432305335998535, "learning_rate": 8.902613065326632e-05, "loss": 1.4045, "step": 70500 }, { "epoch": 3.53, "grad_norm": 1.5630279779434204, "learning_rate": 8.872763819095476e-05, "loss": 1.4617, "step": 70600 }, { "epoch": 3.54, "grad_norm": 2.1527063846588135, "learning_rate": 8.842613065326632e-05, "loss": 1.4445, "step": 70700 }, { "epoch": 3.54, "grad_norm": 4.876186847686768, "learning_rate": 8.812462311557788e-05, "loss": 1.4025, "step": 70800 }, { "epoch": 3.54, "grad_norm": 2.359768867492676, "learning_rate": 8.782311557788944e-05, "loss": 1.3849, "step": 70900 }, { "epoch": 3.55, "grad_norm": 1.0750428438186646, "learning_rate": 8.7521608040201e-05, "loss": 1.3715, "step": 71000 }, { "epoch": 3.55, "eval_loss": 1.4002715349197388, "eval_runtime": 44.6209, "eval_samples_per_second": 22.411, "eval_steps_per_second": 2.801, "step": 71000 }, { "epoch": 3.56, "grad_norm": 3.729515552520752, "learning_rate": 8.722010050251256e-05, "loss": 1.4194, "step": 71100 }, { "epoch": 3.56, "grad_norm": 3.288299560546875, "learning_rate": 8.69185929648241e-05, "loss": 1.4096, "step": 71200 }, { "epoch": 3.56, "grad_norm": 1.2813684940338135, "learning_rate": 8.661708542713568e-05, "loss": 1.4256, "step": 71300 }, { "epoch": 3.57, "grad_norm": 1.7768347263336182, "learning_rate": 8.631557788944722e-05, "loss": 1.4021, "step": 71400 }, { "epoch": 3.58, "grad_norm": 1.5083990097045898, "learning_rate": 8.60140703517588e-05, "loss": 1.3925, "step": 71500 }, { "epoch": 3.58, "grad_norm": 2.056110382080078, "learning_rate": 8.571256281407034e-05, "loss": 1.3997, "step": 71600 }, { "epoch": 3.58, "grad_norm": 1.591378927230835, "learning_rate": 8.54110552763819e-05, "loss": 1.382, "step": 71700 }, { "epoch": 3.59, "grad_norm": 2.8046460151672363, "learning_rate": 8.510954773869346e-05, "loss": 1.4041, "step": 71800 }, { "epoch": 3.59, "grad_norm": 3.3818118572235107, "learning_rate": 8.480804020100502e-05, "loss": 1.3774, "step": 71900 }, { "epoch": 3.6, "grad_norm": 1.9918705224990845, "learning_rate": 8.450653266331658e-05, "loss": 1.4071, "step": 72000 }, { "epoch": 3.6, "eval_loss": 1.3878278732299805, "eval_runtime": 44.5957, "eval_samples_per_second": 22.424, "eval_steps_per_second": 2.803, "step": 72000 }, { "epoch": 3.6, "grad_norm": 2.911811351776123, "learning_rate": 8.420502512562814e-05, "loss": 1.3809, "step": 72100 }, { "epoch": 3.61, "grad_norm": 1.6014336347579956, "learning_rate": 8.390351758793968e-05, "loss": 1.434, "step": 72200 }, { "epoch": 3.62, "grad_norm": 2.2394659519195557, "learning_rate": 8.360201005025126e-05, "loss": 1.4076, "step": 72300 }, { "epoch": 3.62, "grad_norm": 2.3014962673187256, "learning_rate": 8.33005025125628e-05, "loss": 1.4367, "step": 72400 }, { "epoch": 3.62, "grad_norm": 3.4550154209136963, "learning_rate": 8.299899497487438e-05, "loss": 1.3938, "step": 72500 }, { "epoch": 3.63, "grad_norm": 2.599987268447876, "learning_rate": 8.269748743718592e-05, "loss": 1.3873, "step": 72600 }, { "epoch": 3.63, "grad_norm": 2.6119203567504883, "learning_rate": 8.23959798994975e-05, "loss": 1.4069, "step": 72700 }, { "epoch": 3.64, "grad_norm": 1.702276349067688, "learning_rate": 8.209447236180904e-05, "loss": 1.3923, "step": 72800 }, { "epoch": 3.65, "grad_norm": 2.805104970932007, "learning_rate": 8.179296482412059e-05, "loss": 1.386, "step": 72900 }, { "epoch": 3.65, "grad_norm": 2.306410312652588, "learning_rate": 8.149145728643216e-05, "loss": 1.4118, "step": 73000 }, { "epoch": 3.65, "eval_loss": 1.387115716934204, "eval_runtime": 37.4562, "eval_samples_per_second": 26.698, "eval_steps_per_second": 3.337, "step": 73000 }, { "epoch": 3.66, "grad_norm": 2.5508077144622803, "learning_rate": 8.11899497487437e-05, "loss": 1.3963, "step": 73100 }, { "epoch": 3.66, "grad_norm": 1.1488244533538818, "learning_rate": 8.088844221105527e-05, "loss": 1.4107, "step": 73200 }, { "epoch": 3.67, "grad_norm": 3.1272361278533936, "learning_rate": 8.058693467336682e-05, "loss": 1.3444, "step": 73300 }, { "epoch": 3.67, "grad_norm": 2.3045852184295654, "learning_rate": 8.028542713567838e-05, "loss": 1.3874, "step": 73400 }, { "epoch": 3.67, "grad_norm": 2.2844760417938232, "learning_rate": 7.998391959798994e-05, "loss": 1.3748, "step": 73500 }, { "epoch": 3.68, "grad_norm": 2.9295897483825684, "learning_rate": 7.96824120603015e-05, "loss": 1.3786, "step": 73600 }, { "epoch": 3.69, "grad_norm": 1.5064588785171509, "learning_rate": 7.938090452261305e-05, "loss": 1.4079, "step": 73700 }, { "epoch": 3.69, "grad_norm": 1.8818271160125732, "learning_rate": 7.907939698492462e-05, "loss": 1.3954, "step": 73800 }, { "epoch": 3.69, "grad_norm": 1.1885104179382324, "learning_rate": 7.877788944723617e-05, "loss": 1.3836, "step": 73900 }, { "epoch": 3.7, "grad_norm": 3.0773730278015137, "learning_rate": 7.847638190954774e-05, "loss": 1.3724, "step": 74000 }, { "epoch": 3.7, "eval_loss": 1.3931760787963867, "eval_runtime": 37.2623, "eval_samples_per_second": 26.837, "eval_steps_per_second": 3.355, "step": 74000 }, { "epoch": 3.71, "grad_norm": 1.4827876091003418, "learning_rate": 7.817487437185929e-05, "loss": 1.4122, "step": 74100 }, { "epoch": 3.71, "grad_norm": 1.8637442588806152, "learning_rate": 7.787336683417086e-05, "loss": 1.3466, "step": 74200 }, { "epoch": 3.71, "grad_norm": 2.286088228225708, "learning_rate": 7.75718592964824e-05, "loss": 1.3815, "step": 74300 }, { "epoch": 3.72, "grad_norm": 1.8894226551055908, "learning_rate": 7.727035175879396e-05, "loss": 1.4096, "step": 74400 }, { "epoch": 3.73, "grad_norm": 1.9993146657943726, "learning_rate": 7.696884422110552e-05, "loss": 1.4196, "step": 74500 }, { "epoch": 3.73, "grad_norm": 2.170727491378784, "learning_rate": 7.666733668341708e-05, "loss": 1.4034, "step": 74600 }, { "epoch": 3.73, "grad_norm": 1.9965826272964478, "learning_rate": 7.636582914572863e-05, "loss": 1.3971, "step": 74700 }, { "epoch": 3.74, "grad_norm": 2.5607786178588867, "learning_rate": 7.60643216080402e-05, "loss": 1.4116, "step": 74800 }, { "epoch": 3.75, "grad_norm": 1.5363885164260864, "learning_rate": 7.576281407035175e-05, "loss": 1.3932, "step": 74900 }, { "epoch": 3.75, "grad_norm": 1.9062122106552124, "learning_rate": 7.546130653266332e-05, "loss": 1.3625, "step": 75000 }, { "epoch": 3.75, "eval_loss": 1.3875294923782349, "eval_runtime": 37.1595, "eval_samples_per_second": 26.911, "eval_steps_per_second": 3.364, "step": 75000 }, { "epoch": 3.75, "grad_norm": 2.8305203914642334, "learning_rate": 7.516281407035175e-05, "loss": 1.3907, "step": 75100 }, { "epoch": 3.76, "grad_norm": 1.6526367664337158, "learning_rate": 7.486130653266331e-05, "loss": 1.3605, "step": 75200 }, { "epoch": 3.77, "grad_norm": 3.5784194469451904, "learning_rate": 7.455979899497487e-05, "loss": 1.3833, "step": 75300 }, { "epoch": 3.77, "grad_norm": 0.9500262141227722, "learning_rate": 7.425829145728643e-05, "loss": 1.3548, "step": 75400 }, { "epoch": 3.77, "grad_norm": 2.8620994091033936, "learning_rate": 7.395678391959799e-05, "loss": 1.3829, "step": 75500 }, { "epoch": 3.78, "grad_norm": 1.9316253662109375, "learning_rate": 7.365527638190954e-05, "loss": 1.3801, "step": 75600 }, { "epoch": 3.79, "grad_norm": 1.5958226919174194, "learning_rate": 7.33537688442211e-05, "loss": 1.3837, "step": 75700 }, { "epoch": 3.79, "grad_norm": 1.995013952255249, "learning_rate": 7.305226130653266e-05, "loss": 1.3933, "step": 75800 }, { "epoch": 3.79, "grad_norm": 1.4954237937927246, "learning_rate": 7.275075376884422e-05, "loss": 1.4223, "step": 75900 }, { "epoch": 3.8, "grad_norm": 1.363682508468628, "learning_rate": 7.244924623115577e-05, "loss": 1.3811, "step": 76000 }, { "epoch": 3.8, "eval_loss": 1.3870151042938232, "eval_runtime": 37.1537, "eval_samples_per_second": 26.915, "eval_steps_per_second": 3.364, "step": 76000 }, { "epoch": 3.81, "grad_norm": 1.6989458799362183, "learning_rate": 7.214773869346733e-05, "loss": 1.3755, "step": 76100 }, { "epoch": 3.81, "grad_norm": 2.042025327682495, "learning_rate": 7.184623115577889e-05, "loss": 1.3641, "step": 76200 }, { "epoch": 3.81, "grad_norm": 1.4355963468551636, "learning_rate": 7.154472361809045e-05, "loss": 1.3675, "step": 76300 }, { "epoch": 3.82, "grad_norm": 1.0404947996139526, "learning_rate": 7.124321608040201e-05, "loss": 1.3755, "step": 76400 }, { "epoch": 3.83, "grad_norm": 2.588839530944824, "learning_rate": 7.094170854271357e-05, "loss": 1.369, "step": 76500 }, { "epoch": 3.83, "grad_norm": 2.2656760215759277, "learning_rate": 7.064020100502511e-05, "loss": 1.4254, "step": 76600 }, { "epoch": 3.83, "grad_norm": 2.1572184562683105, "learning_rate": 7.033869346733667e-05, "loss": 1.3928, "step": 76700 }, { "epoch": 3.84, "grad_norm": 1.9283751249313354, "learning_rate": 7.003718592964823e-05, "loss": 1.3502, "step": 76800 }, { "epoch": 3.84, "grad_norm": 2.215160846710205, "learning_rate": 6.973567839195979e-05, "loss": 1.3701, "step": 76900 }, { "epoch": 3.85, "grad_norm": 1.9962304830551147, "learning_rate": 6.943417085427135e-05, "loss": 1.3353, "step": 77000 }, { "epoch": 3.85, "eval_loss": 1.3705061674118042, "eval_runtime": 36.8621, "eval_samples_per_second": 27.128, "eval_steps_per_second": 3.391, "step": 77000 }, { "epoch": 3.85, "grad_norm": 1.818708062171936, "learning_rate": 6.913266331658291e-05, "loss": 1.4337, "step": 77100 }, { "epoch": 3.86, "grad_norm": 2.5512075424194336, "learning_rate": 6.883115577889447e-05, "loss": 1.3923, "step": 77200 }, { "epoch": 3.87, "grad_norm": 2.3595049381256104, "learning_rate": 6.852964824120603e-05, "loss": 1.3626, "step": 77300 }, { "epoch": 3.87, "grad_norm": 2.033395767211914, "learning_rate": 6.822814070351757e-05, "loss": 1.4061, "step": 77400 }, { "epoch": 3.88, "grad_norm": 1.8597971200942993, "learning_rate": 6.792663316582913e-05, "loss": 1.3672, "step": 77500 }, { "epoch": 3.88, "grad_norm": 6.214807987213135, "learning_rate": 6.762512562814069e-05, "loss": 1.3817, "step": 77600 }, { "epoch": 3.88, "grad_norm": 2.384474515914917, "learning_rate": 6.732361809045225e-05, "loss": 1.3811, "step": 77700 }, { "epoch": 3.89, "grad_norm": 2.6242575645446777, "learning_rate": 6.702211055276381e-05, "loss": 1.3883, "step": 77800 }, { "epoch": 3.9, "grad_norm": 1.4827545881271362, "learning_rate": 6.672060301507537e-05, "loss": 1.3755, "step": 77900 }, { "epoch": 3.9, "grad_norm": 2.150613307952881, "learning_rate": 6.641909547738693e-05, "loss": 1.3937, "step": 78000 }, { "epoch": 3.9, "eval_loss": 1.379770278930664, "eval_runtime": 37.0122, "eval_samples_per_second": 27.018, "eval_steps_per_second": 3.377, "step": 78000 }, { "epoch": 3.91, "grad_norm": 1.4805549383163452, "learning_rate": 6.611758793969849e-05, "loss": 1.365, "step": 78100 }, { "epoch": 3.91, "grad_norm": 1.699484944343567, "learning_rate": 6.581608040201005e-05, "loss": 1.3893, "step": 78200 }, { "epoch": 3.92, "grad_norm": 2.039008140563965, "learning_rate": 6.551758793969849e-05, "loss": 1.3601, "step": 78300 }, { "epoch": 3.92, "grad_norm": 1.3309062719345093, "learning_rate": 6.521608040201005e-05, "loss": 1.4056, "step": 78400 }, { "epoch": 3.92, "grad_norm": 2.382688045501709, "learning_rate": 6.491758793969849e-05, "loss": 1.3758, "step": 78500 }, { "epoch": 3.93, "grad_norm": 3.1065239906311035, "learning_rate": 6.461608040201005e-05, "loss": 1.3725, "step": 78600 }, { "epoch": 3.94, "grad_norm": 2.193957805633545, "learning_rate": 6.431457286432161e-05, "loss": 1.3347, "step": 78700 }, { "epoch": 3.94, "grad_norm": 1.8022537231445312, "learning_rate": 6.401306532663317e-05, "loss": 1.3568, "step": 78800 }, { "epoch": 3.94, "grad_norm": 1.7433044910430908, "learning_rate": 6.371155778894473e-05, "loss": 1.3643, "step": 78900 }, { "epoch": 3.95, "grad_norm": 3.883768081665039, "learning_rate": 6.341005025125627e-05, "loss": 1.3158, "step": 79000 }, { "epoch": 3.95, "eval_loss": 1.37774658203125, "eval_runtime": 36.9703, "eval_samples_per_second": 27.049, "eval_steps_per_second": 3.381, "step": 79000 }, { "epoch": 3.96, "grad_norm": 2.9996466636657715, "learning_rate": 6.310854271356783e-05, "loss": 1.3295, "step": 79100 }, { "epoch": 3.96, "grad_norm": 1.2524511814117432, "learning_rate": 6.280703517587939e-05, "loss": 1.3362, "step": 79200 }, { "epoch": 3.96, "grad_norm": 1.8024784326553345, "learning_rate": 6.250552763819095e-05, "loss": 1.3857, "step": 79300 }, { "epoch": 3.97, "grad_norm": 1.3988703489303589, "learning_rate": 6.220402010050251e-05, "loss": 1.3463, "step": 79400 }, { "epoch": 3.98, "grad_norm": 2.499194622039795, "learning_rate": 6.190251256281407e-05, "loss": 1.3761, "step": 79500 }, { "epoch": 3.98, "grad_norm": 2.799959897994995, "learning_rate": 6.160100502512562e-05, "loss": 1.3995, "step": 79600 }, { "epoch": 3.98, "grad_norm": 1.2367918491363525, "learning_rate": 6.129949748743717e-05, "loss": 1.3789, "step": 79700 }, { "epoch": 3.99, "grad_norm": 1.7651324272155762, "learning_rate": 6.0997989949748734e-05, "loss": 1.3691, "step": 79800 }, { "epoch": 4.0, "grad_norm": 1.8847932815551758, "learning_rate": 6.0696482412060293e-05, "loss": 1.353, "step": 79900 }, { "epoch": 4.0, "grad_norm": 1.8860230445861816, "learning_rate": 6.039497487437185e-05, "loss": 1.3168, "step": 80000 }, { "epoch": 4.0, "eval_loss": 1.3387900590896606, "eval_runtime": 37.0136, "eval_samples_per_second": 27.017, "eval_steps_per_second": 3.377, "step": 80000 }, { "epoch": 4.0, "grad_norm": 2.109632968902588, "learning_rate": 6.009346733668341e-05, "loss": 1.3518, "step": 80100 }, { "epoch": 4.01, "grad_norm": 3.0571937561035156, "learning_rate": 5.9791959798994965e-05, "loss": 1.3664, "step": 80200 }, { "epoch": 4.01, "grad_norm": 1.8600945472717285, "learning_rate": 5.9490452261306525e-05, "loss": 1.3252, "step": 80300 }, { "epoch": 4.02, "grad_norm": 2.461089611053467, "learning_rate": 5.9188944723618084e-05, "loss": 1.3507, "step": 80400 }, { "epoch": 4.03, "grad_norm": 3.1424946784973145, "learning_rate": 5.8887437185929643e-05, "loss": 1.3414, "step": 80500 }, { "epoch": 4.03, "grad_norm": 1.701357126235962, "learning_rate": 5.85859296482412e-05, "loss": 1.3047, "step": 80600 }, { "epoch": 4.04, "grad_norm": 2.331779956817627, "learning_rate": 5.8284422110552756e-05, "loss": 1.3309, "step": 80700 }, { "epoch": 4.04, "grad_norm": 5.718899726867676, "learning_rate": 5.7982914572864315e-05, "loss": 1.3796, "step": 80800 }, { "epoch": 4.04, "grad_norm": 1.1299536228179932, "learning_rate": 5.7681407035175874e-05, "loss": 1.3517, "step": 80900 }, { "epoch": 4.05, "grad_norm": 2.256490468978882, "learning_rate": 5.7379899497487434e-05, "loss": 1.35, "step": 81000 }, { "epoch": 4.05, "eval_loss": 1.3527089357376099, "eval_runtime": 37.0846, "eval_samples_per_second": 26.965, "eval_steps_per_second": 3.371, "step": 81000 }, { "epoch": 4.05, "grad_norm": 2.3526551723480225, "learning_rate": 5.707839195979899e-05, "loss": 1.3327, "step": 81100 }, { "epoch": 4.06, "grad_norm": 2.987255573272705, "learning_rate": 5.677688442211055e-05, "loss": 1.3369, "step": 81200 }, { "epoch": 4.07, "grad_norm": 3.35562801361084, "learning_rate": 5.6475376884422105e-05, "loss": 1.3583, "step": 81300 }, { "epoch": 4.07, "grad_norm": 2.192720651626587, "learning_rate": 5.6173869346733665e-05, "loss": 1.3405, "step": 81400 }, { "epoch": 4.08, "grad_norm": 1.91267728805542, "learning_rate": 5.5872361809045224e-05, "loss": 1.3281, "step": 81500 }, { "epoch": 4.08, "grad_norm": 2.5109171867370605, "learning_rate": 5.5570854271356784e-05, "loss": 1.3126, "step": 81600 }, { "epoch": 4.08, "grad_norm": 1.4430352449417114, "learning_rate": 5.526934673366834e-05, "loss": 1.2983, "step": 81700 }, { "epoch": 4.09, "grad_norm": 1.873565912246704, "learning_rate": 5.4967839195979896e-05, "loss": 1.292, "step": 81800 }, { "epoch": 4.09, "grad_norm": 1.6436150074005127, "learning_rate": 5.466633165829145e-05, "loss": 1.3268, "step": 81900 }, { "epoch": 4.1, "grad_norm": 3.9100093841552734, "learning_rate": 5.436482412060301e-05, "loss": 1.3388, "step": 82000 }, { "epoch": 4.1, "eval_loss": 1.3313664197921753, "eval_runtime": 37.2377, "eval_samples_per_second": 26.855, "eval_steps_per_second": 3.357, "step": 82000 }, { "epoch": 4.11, "grad_norm": 3.0072224140167236, "learning_rate": 5.406331658291457e-05, "loss": 1.2984, "step": 82100 }, { "epoch": 4.11, "grad_norm": 1.3772695064544678, "learning_rate": 5.376180904522612e-05, "loss": 1.3182, "step": 82200 }, { "epoch": 4.12, "grad_norm": 3.444744110107422, "learning_rate": 5.346030150753768e-05, "loss": 1.3237, "step": 82300 }, { "epoch": 4.12, "grad_norm": 2.719045400619507, "learning_rate": 5.315879396984924e-05, "loss": 1.349, "step": 82400 }, { "epoch": 4.12, "grad_norm": 4.052174091339111, "learning_rate": 5.28572864321608e-05, "loss": 1.3504, "step": 82500 }, { "epoch": 4.13, "grad_norm": 1.3267349004745483, "learning_rate": 5.255577889447236e-05, "loss": 1.3038, "step": 82600 }, { "epoch": 4.13, "grad_norm": 2.1242218017578125, "learning_rate": 5.225427135678391e-05, "loss": 1.3087, "step": 82700 }, { "epoch": 4.14, "grad_norm": 3.2896034717559814, "learning_rate": 5.195276381909547e-05, "loss": 1.31, "step": 82800 }, { "epoch": 4.14, "grad_norm": 2.3824546337127686, "learning_rate": 5.165125628140703e-05, "loss": 1.2972, "step": 82900 }, { "epoch": 4.15, "grad_norm": 2.0106704235076904, "learning_rate": 5.134974874371859e-05, "loss": 1.3578, "step": 83000 }, { "epoch": 4.15, "eval_loss": 1.3431867361068726, "eval_runtime": 36.9495, "eval_samples_per_second": 27.064, "eval_steps_per_second": 3.383, "step": 83000 }, { "epoch": 4.16, "grad_norm": 2.094095230102539, "learning_rate": 5.104824120603015e-05, "loss": 1.3172, "step": 83100 }, { "epoch": 4.16, "grad_norm": 2.0634231567382812, "learning_rate": 5.074673366834171e-05, "loss": 1.3381, "step": 83200 }, { "epoch": 4.17, "grad_norm": 4.8768486976623535, "learning_rate": 5.044522613065326e-05, "loss": 1.3148, "step": 83300 }, { "epoch": 4.17, "grad_norm": 4.0618486404418945, "learning_rate": 5.014371859296482e-05, "loss": 1.3635, "step": 83400 }, { "epoch": 4.17, "grad_norm": 1.6911975145339966, "learning_rate": 4.984221105527638e-05, "loss": 1.3333, "step": 83500 }, { "epoch": 4.18, "grad_norm": 2.164687156677246, "learning_rate": 4.954070351758794e-05, "loss": 1.3391, "step": 83600 }, { "epoch": 4.18, "grad_norm": 1.3353967666625977, "learning_rate": 4.92391959798995e-05, "loss": 1.2989, "step": 83700 }, { "epoch": 4.19, "grad_norm": 3.4308252334594727, "learning_rate": 4.893768844221105e-05, "loss": 1.2966, "step": 83800 }, { "epoch": 4.2, "grad_norm": 2.58341121673584, "learning_rate": 4.863618090452261e-05, "loss": 1.3036, "step": 83900 }, { "epoch": 4.2, "grad_norm": 2.0035691261291504, "learning_rate": 4.833467336683417e-05, "loss": 1.3028, "step": 84000 }, { "epoch": 4.2, "eval_loss": 1.3128948211669922, "eval_runtime": 36.9125, "eval_samples_per_second": 27.091, "eval_steps_per_second": 3.386, "step": 84000 }, { "epoch": 4.21, "grad_norm": 3.858201503753662, "learning_rate": 4.803316582914573e-05, "loss": 1.3459, "step": 84100 }, { "epoch": 4.21, "grad_norm": 1.9417260885238647, "learning_rate": 4.7731658291457275e-05, "loss": 1.3243, "step": 84200 }, { "epoch": 4.21, "grad_norm": 3.000342607498169, "learning_rate": 4.7430150753768835e-05, "loss": 1.3175, "step": 84300 }, { "epoch": 4.22, "grad_norm": 5.205733776092529, "learning_rate": 4.7128643216080394e-05, "loss": 1.3486, "step": 84400 }, { "epoch": 4.22, "grad_norm": 3.2263059616088867, "learning_rate": 4.683015075376885e-05, "loss": 1.3103, "step": 84500 }, { "epoch": 4.23, "grad_norm": 2.210460662841797, "learning_rate": 4.652864321608039e-05, "loss": 1.339, "step": 84600 }, { "epoch": 4.24, "grad_norm": 3.195037364959717, "learning_rate": 4.622713567839195e-05, "loss": 1.3472, "step": 84700 }, { "epoch": 4.24, "grad_norm": 2.5944504737854004, "learning_rate": 4.592562814070351e-05, "loss": 1.3159, "step": 84800 }, { "epoch": 4.25, "grad_norm": 2.6246960163116455, "learning_rate": 4.562412060301507e-05, "loss": 1.2896, "step": 84900 }, { "epoch": 4.25, "grad_norm": 2.556218385696411, "learning_rate": 4.5322613065326624e-05, "loss": 1.3187, "step": 85000 }, { "epoch": 4.25, "eval_loss": 1.324548363685608, "eval_runtime": 36.9422, "eval_samples_per_second": 27.069, "eval_steps_per_second": 3.384, "step": 85000 }, { "epoch": 4.25, "grad_norm": 2.348661184310913, "learning_rate": 4.5021105527638184e-05, "loss": 1.3004, "step": 85100 }, { "epoch": 4.26, "grad_norm": 2.0715219974517822, "learning_rate": 4.471959798994974e-05, "loss": 1.3265, "step": 85200 }, { "epoch": 4.26, "grad_norm": 2.5301461219787598, "learning_rate": 4.44180904522613e-05, "loss": 1.3187, "step": 85300 }, { "epoch": 4.27, "grad_norm": 5.115157604217529, "learning_rate": 4.411658291457286e-05, "loss": 1.3179, "step": 85400 }, { "epoch": 4.28, "grad_norm": 2.671520471572876, "learning_rate": 4.3815075376884415e-05, "loss": 1.2988, "step": 85500 }, { "epoch": 4.28, "grad_norm": 2.0337090492248535, "learning_rate": 4.3513567839195974e-05, "loss": 1.2781, "step": 85600 }, { "epoch": 4.29, "grad_norm": 2.735685348510742, "learning_rate": 4.3212060301507534e-05, "loss": 1.3167, "step": 85700 }, { "epoch": 4.29, "grad_norm": 2.573694944381714, "learning_rate": 4.291055276381909e-05, "loss": 1.3057, "step": 85800 }, { "epoch": 4.29, "grad_norm": 1.5959097146987915, "learning_rate": 4.260904522613065e-05, "loss": 1.327, "step": 85900 }, { "epoch": 4.3, "grad_norm": 2.1494078636169434, "learning_rate": 4.2307537688442205e-05, "loss": 1.3291, "step": 86000 }, { "epoch": 4.3, "eval_loss": 1.3199845552444458, "eval_runtime": 36.8676, "eval_samples_per_second": 27.124, "eval_steps_per_second": 3.391, "step": 86000 }, { "epoch": 4.3, "grad_norm": 1.7684657573699951, "learning_rate": 4.2006030150753765e-05, "loss": 1.3341, "step": 86100 }, { "epoch": 4.31, "grad_norm": 2.656754732131958, "learning_rate": 4.1704522613065324e-05, "loss": 1.3275, "step": 86200 }, { "epoch": 4.32, "grad_norm": 1.330156683921814, "learning_rate": 4.1403015075376883e-05, "loss": 1.3365, "step": 86300 }, { "epoch": 4.32, "grad_norm": 3.0051920413970947, "learning_rate": 4.110150753768844e-05, "loss": 1.303, "step": 86400 }, { "epoch": 4.33, "grad_norm": 1.494343638420105, "learning_rate": 4.08e-05, "loss": 1.2839, "step": 86500 }, { "epoch": 4.33, "grad_norm": 3.456125020980835, "learning_rate": 4.0498492462311555e-05, "loss": 1.2964, "step": 86600 }, { "epoch": 4.33, "grad_norm": 2.114023447036743, "learning_rate": 4.0196984924623115e-05, "loss": 1.3397, "step": 86700 }, { "epoch": 4.34, "grad_norm": 2.5307486057281494, "learning_rate": 3.9895477386934674e-05, "loss": 1.2841, "step": 86800 }, { "epoch": 4.34, "grad_norm": 1.866807222366333, "learning_rate": 3.9596984924623113e-05, "loss": 1.3218, "step": 86900 }, { "epoch": 4.35, "grad_norm": 1.9067096710205078, "learning_rate": 3.929547738693467e-05, "loss": 1.3257, "step": 87000 }, { "epoch": 4.35, "eval_loss": 1.3586480617523193, "eval_runtime": 37.2768, "eval_samples_per_second": 26.826, "eval_steps_per_second": 3.353, "step": 87000 }, { "epoch": 4.36, "grad_norm": 3.3964035511016846, "learning_rate": 3.899396984924623e-05, "loss": 1.3385, "step": 87100 }, { "epoch": 4.36, "grad_norm": 3.567638874053955, "learning_rate": 3.869246231155778e-05, "loss": 1.3281, "step": 87200 }, { "epoch": 4.37, "grad_norm": 2.170928955078125, "learning_rate": 3.839095477386934e-05, "loss": 1.3159, "step": 87300 }, { "epoch": 4.37, "grad_norm": 2.326664924621582, "learning_rate": 3.80894472361809e-05, "loss": 1.3499, "step": 87400 }, { "epoch": 4.38, "grad_norm": 2.8397111892700195, "learning_rate": 3.778793969849246e-05, "loss": 1.349, "step": 87500 }, { "epoch": 4.38, "grad_norm": 2.2665247917175293, "learning_rate": 3.7486432160804016e-05, "loss": 1.3172, "step": 87600 }, { "epoch": 4.38, "grad_norm": 4.748266696929932, "learning_rate": 3.718793969849246e-05, "loss": 1.2857, "step": 87700 }, { "epoch": 4.39, "grad_norm": 1.4008738994598389, "learning_rate": 3.6886432160804015e-05, "loss": 1.276, "step": 87800 }, { "epoch": 4.39, "grad_norm": 3.0103044509887695, "learning_rate": 3.6584924623115574e-05, "loss": 1.3114, "step": 87900 }, { "epoch": 4.4, "grad_norm": 1.6098419427871704, "learning_rate": 3.6283417085427134e-05, "loss": 1.2995, "step": 88000 }, { "epoch": 4.4, "eval_loss": 1.2803822755813599, "eval_runtime": 37.2497, "eval_samples_per_second": 26.846, "eval_steps_per_second": 3.356, "step": 88000 }, { "epoch": 4.41, "grad_norm": 1.6589020490646362, "learning_rate": 3.5981909547738693e-05, "loss": 1.3214, "step": 88100 }, { "epoch": 4.41, "grad_norm": 3.2466797828674316, "learning_rate": 3.5680402010050246e-05, "loss": 1.2685, "step": 88200 }, { "epoch": 4.42, "grad_norm": 3.273249626159668, "learning_rate": 3.5378894472361806e-05, "loss": 1.2773, "step": 88300 }, { "epoch": 4.42, "grad_norm": 2.712141513824463, "learning_rate": 3.5077386934673365e-05, "loss": 1.3107, "step": 88400 }, { "epoch": 4.42, "grad_norm": 2.671769857406616, "learning_rate": 3.477587939698492e-05, "loss": 1.3149, "step": 88500 }, { "epoch": 4.43, "grad_norm": 1.6911038160324097, "learning_rate": 3.447437185929648e-05, "loss": 1.312, "step": 88600 }, { "epoch": 4.43, "grad_norm": 2.3690972328186035, "learning_rate": 3.4172864321608037e-05, "loss": 1.2826, "step": 88700 }, { "epoch": 4.44, "grad_norm": 2.1747915744781494, "learning_rate": 3.3871356783919596e-05, "loss": 1.3187, "step": 88800 }, { "epoch": 4.45, "grad_norm": 1.7691452503204346, "learning_rate": 3.3569849246231155e-05, "loss": 1.3356, "step": 88900 }, { "epoch": 4.45, "grad_norm": 2.9473211765289307, "learning_rate": 3.3268341708542715e-05, "loss": 1.2721, "step": 89000 }, { "epoch": 4.45, "eval_loss": 1.3034113645553589, "eval_runtime": 37.3017, "eval_samples_per_second": 26.808, "eval_steps_per_second": 3.351, "step": 89000 }, { "epoch": 4.46, "grad_norm": 2.840188503265381, "learning_rate": 3.296683417085427e-05, "loss": 1.2842, "step": 89100 }, { "epoch": 4.46, "grad_norm": 1.5584784746170044, "learning_rate": 3.266532663316583e-05, "loss": 1.2961, "step": 89200 }, { "epoch": 4.46, "grad_norm": 1.889328956604004, "learning_rate": 3.236381909547738e-05, "loss": 1.3121, "step": 89300 }, { "epoch": 4.47, "grad_norm": 1.627302646636963, "learning_rate": 3.206231155778894e-05, "loss": 1.3169, "step": 89400 }, { "epoch": 4.47, "grad_norm": 3.445068359375, "learning_rate": 3.17608040201005e-05, "loss": 1.2691, "step": 89500 }, { "epoch": 4.48, "grad_norm": 3.2616968154907227, "learning_rate": 3.145929648241206e-05, "loss": 1.2818, "step": 89600 }, { "epoch": 4.49, "grad_norm": 2.2444956302642822, "learning_rate": 3.115778894472362e-05, "loss": 1.2836, "step": 89700 }, { "epoch": 4.49, "grad_norm": 2.341099500656128, "learning_rate": 3.085628140703517e-05, "loss": 1.2712, "step": 89800 }, { "epoch": 4.5, "grad_norm": 1.882042407989502, "learning_rate": 3.055477386934673e-05, "loss": 1.3316, "step": 89900 }, { "epoch": 4.5, "grad_norm": 3.1667628288269043, "learning_rate": 3.025326633165829e-05, "loss": 1.312, "step": 90000 }, { "epoch": 4.5, "eval_loss": 1.309124231338501, "eval_runtime": 37.1306, "eval_samples_per_second": 26.932, "eval_steps_per_second": 3.366, "step": 90000 }, { "epoch": 4.5, "grad_norm": 2.1243770122528076, "learning_rate": 2.9954773869346732e-05, "loss": 1.2924, "step": 90100 }, { "epoch": 4.51, "grad_norm": 10.278544425964355, "learning_rate": 2.9653266331658288e-05, "loss": 1.3081, "step": 90200 }, { "epoch": 4.51, "grad_norm": 8.060081481933594, "learning_rate": 2.9351758793969847e-05, "loss": 1.3317, "step": 90300 }, { "epoch": 4.52, "grad_norm": 1.364973783493042, "learning_rate": 2.9050251256281404e-05, "loss": 1.2882, "step": 90400 }, { "epoch": 4.53, "grad_norm": 1.1993753910064697, "learning_rate": 2.8748743718592963e-05, "loss": 1.292, "step": 90500 }, { "epoch": 4.53, "grad_norm": 2.897918462753296, "learning_rate": 2.8447236180904522e-05, "loss": 1.2804, "step": 90600 }, { "epoch": 4.54, "grad_norm": 2.7222065925598145, "learning_rate": 2.814572864321608e-05, "loss": 1.2985, "step": 90700 }, { "epoch": 4.54, "grad_norm": 5.0719099044799805, "learning_rate": 2.7844221105527635e-05, "loss": 1.2768, "step": 90800 }, { "epoch": 4.54, "grad_norm": 2.1225016117095947, "learning_rate": 2.754271356783919e-05, "loss": 1.2774, "step": 90900 }, { "epoch": 4.55, "grad_norm": 1.314208984375, "learning_rate": 2.724120603015075e-05, "loss": 1.3064, "step": 91000 }, { "epoch": 4.55, "eval_loss": 1.2666804790496826, "eval_runtime": 37.0231, "eval_samples_per_second": 27.01, "eval_steps_per_second": 3.376, "step": 91000 }, { "epoch": 4.55, "grad_norm": 1.5543466806411743, "learning_rate": 2.693969849246231e-05, "loss": 1.2905, "step": 91100 }, { "epoch": 4.56, "grad_norm": 1.3213764429092407, "learning_rate": 2.6638190954773866e-05, "loss": 1.2541, "step": 91200 }, { "epoch": 4.56, "grad_norm": 2.523181200027466, "learning_rate": 2.6336683417085425e-05, "loss": 1.2743, "step": 91300 }, { "epoch": 4.57, "grad_norm": 3.6385650634765625, "learning_rate": 2.6035175879396984e-05, "loss": 1.2936, "step": 91400 }, { "epoch": 4.58, "grad_norm": 2.2603909969329834, "learning_rate": 2.573366834170854e-05, "loss": 1.252, "step": 91500 }, { "epoch": 4.58, "grad_norm": 3.178767442703247, "learning_rate": 2.54321608040201e-05, "loss": 1.2637, "step": 91600 }, { "epoch": 4.58, "grad_norm": 1.8341214656829834, "learning_rate": 2.5130653266331656e-05, "loss": 1.2692, "step": 91700 }, { "epoch": 4.59, "grad_norm": 3.1009461879730225, "learning_rate": 2.4829145728643216e-05, "loss": 1.2821, "step": 91800 }, { "epoch": 4.59, "grad_norm": 2.75738263130188, "learning_rate": 2.4527638190954775e-05, "loss": 1.284, "step": 91900 }, { "epoch": 4.6, "grad_norm": 3.137352466583252, "learning_rate": 2.4226130653266328e-05, "loss": 1.2846, "step": 92000 }, { "epoch": 4.6, "eval_loss": 1.244607925415039, "eval_runtime": 37.127, "eval_samples_per_second": 26.935, "eval_steps_per_second": 3.367, "step": 92000 }, { "epoch": 4.61, "grad_norm": 2.5030505657196045, "learning_rate": 2.3924623115577887e-05, "loss": 1.2793, "step": 92100 }, { "epoch": 4.61, "grad_norm": 2.625671625137329, "learning_rate": 2.3623115577889443e-05, "loss": 1.2575, "step": 92200 }, { "epoch": 4.62, "grad_norm": 1.8129239082336426, "learning_rate": 2.3321608040201003e-05, "loss": 1.265, "step": 92300 }, { "epoch": 4.62, "grad_norm": 2.747807502746582, "learning_rate": 2.3020100502512562e-05, "loss": 1.2781, "step": 92400 }, { "epoch": 4.62, "grad_norm": 3.083634853363037, "learning_rate": 2.2718592964824118e-05, "loss": 1.2667, "step": 92500 }, { "epoch": 4.63, "grad_norm": 3.475771427154541, "learning_rate": 2.2417085427135678e-05, "loss": 1.2278, "step": 92600 }, { "epoch": 4.63, "grad_norm": 4.056103706359863, "learning_rate": 2.2115577889447234e-05, "loss": 1.276, "step": 92700 }, { "epoch": 4.64, "grad_norm": 3.8437623977661133, "learning_rate": 2.1814070351758793e-05, "loss": 1.2799, "step": 92800 }, { "epoch": 4.64, "grad_norm": 2.0102827548980713, "learning_rate": 2.1512562814070353e-05, "loss": 1.2589, "step": 92900 }, { "epoch": 4.65, "grad_norm": 3.917171001434326, "learning_rate": 2.121105527638191e-05, "loss": 1.245, "step": 93000 }, { "epoch": 4.65, "eval_loss": 1.2793523073196411, "eval_runtime": 37.5859, "eval_samples_per_second": 26.606, "eval_steps_per_second": 3.326, "step": 93000 }, { "epoch": 4.66, "grad_norm": 1.847579836845398, "learning_rate": 2.0909547738693465e-05, "loss": 1.2172, "step": 93100 }, { "epoch": 4.66, "grad_norm": 1.8271080255508423, "learning_rate": 2.060804020100502e-05, "loss": 1.2781, "step": 93200 }, { "epoch": 4.67, "grad_norm": 3.5897600650787354, "learning_rate": 2.030653266331658e-05, "loss": 1.2587, "step": 93300 }, { "epoch": 4.67, "grad_norm": 2.1579065322875977, "learning_rate": 2.0008040201005026e-05, "loss": 1.2742, "step": 93400 }, { "epoch": 4.67, "grad_norm": 2.0196428298950195, "learning_rate": 1.970653266331658e-05, "loss": 1.24, "step": 93500 }, { "epoch": 4.68, "grad_norm": 2.4959633350372314, "learning_rate": 1.940502512562814e-05, "loss": 1.28, "step": 93600 }, { "epoch": 4.69, "grad_norm": 2.4107792377471924, "learning_rate": 1.9103517587939695e-05, "loss": 1.326, "step": 93700 }, { "epoch": 4.69, "grad_norm": 3.1433868408203125, "learning_rate": 1.8802010050251254e-05, "loss": 1.2423, "step": 93800 }, { "epoch": 4.7, "grad_norm": 3.2713465690612793, "learning_rate": 1.8500502512562814e-05, "loss": 1.3136, "step": 93900 }, { "epoch": 4.7, "grad_norm": 2.5864298343658447, "learning_rate": 1.819899497487437e-05, "loss": 1.3029, "step": 94000 }, { "epoch": 4.7, "eval_loss": 1.3026132583618164, "eval_runtime": 37.3025, "eval_samples_per_second": 26.808, "eval_steps_per_second": 3.351, "step": 94000 }, { "epoch": 4.71, "grad_norm": 3.219913959503174, "learning_rate": 1.789748743718593e-05, "loss": 1.3279, "step": 94100 }, { "epoch": 4.71, "grad_norm": 2.7392513751983643, "learning_rate": 1.7595979899497485e-05, "loss": 1.2637, "step": 94200 }, { "epoch": 4.71, "grad_norm": 3.506613254547119, "learning_rate": 1.7294472361809045e-05, "loss": 1.2762, "step": 94300 }, { "epoch": 4.72, "grad_norm": 1.6417380571365356, "learning_rate": 1.69929648241206e-05, "loss": 1.2733, "step": 94400 }, { "epoch": 4.72, "grad_norm": 3.572312355041504, "learning_rate": 1.669145728643216e-05, "loss": 1.2524, "step": 94500 }, { "epoch": 4.73, "grad_norm": 2.3081557750701904, "learning_rate": 1.6389949748743716e-05, "loss": 1.2388, "step": 94600 }, { "epoch": 4.74, "grad_norm": 3.257410764694214, "learning_rate": 1.6088442211055276e-05, "loss": 1.2395, "step": 94700 }, { "epoch": 4.74, "grad_norm": 3.3409955501556396, "learning_rate": 1.5786934673366835e-05, "loss": 1.2464, "step": 94800 }, { "epoch": 4.75, "grad_norm": 2.011337995529175, "learning_rate": 1.548542713567839e-05, "loss": 1.2799, "step": 94900 }, { "epoch": 4.75, "grad_norm": 3.929819107055664, "learning_rate": 1.5183919597989947e-05, "loss": 1.2408, "step": 95000 }, { "epoch": 4.75, "eval_loss": 1.2593971490859985, "eval_runtime": 37.2434, "eval_samples_per_second": 26.85, "eval_steps_per_second": 3.356, "step": 95000 }, { "epoch": 4.75, "grad_norm": 3.053884983062744, "learning_rate": 1.4882412060301507e-05, "loss": 1.2509, "step": 95100 }, { "epoch": 4.76, "grad_norm": 2.4242303371429443, "learning_rate": 1.4580904522613064e-05, "loss": 1.2419, "step": 95200 }, { "epoch": 4.76, "grad_norm": 5.195686340332031, "learning_rate": 1.4279396984924622e-05, "loss": 1.2542, "step": 95300 }, { "epoch": 4.77, "grad_norm": 2.4507782459259033, "learning_rate": 1.397788944723618e-05, "loss": 1.2388, "step": 95400 }, { "epoch": 4.78, "grad_norm": 3.998619318008423, "learning_rate": 1.3676381909547736e-05, "loss": 1.225, "step": 95500 }, { "epoch": 4.78, "grad_norm": 6.8137664794921875, "learning_rate": 1.3374874371859295e-05, "loss": 1.2379, "step": 95600 }, { "epoch": 4.79, "grad_norm": 2.8627781867980957, "learning_rate": 1.3073366834170853e-05, "loss": 1.2752, "step": 95700 }, { "epoch": 4.79, "grad_norm": 2.2454240322113037, "learning_rate": 1.2771859296482411e-05, "loss": 1.243, "step": 95800 }, { "epoch": 4.79, "grad_norm": 1.6306633949279785, "learning_rate": 1.2470351758793969e-05, "loss": 1.2593, "step": 95900 }, { "epoch": 4.8, "grad_norm": 2.288721799850464, "learning_rate": 1.217185929648241e-05, "loss": 1.246, "step": 96000 }, { "epoch": 4.8, "eval_loss": 1.245192527770996, "eval_runtime": 37.2658, "eval_samples_per_second": 26.834, "eval_steps_per_second": 3.354, "step": 96000 }, { "epoch": 4.8, "grad_norm": 3.1359307765960693, "learning_rate": 1.187035175879397e-05, "loss": 1.2531, "step": 96100 }, { "epoch": 4.81, "grad_norm": 3.468085527420044, "learning_rate": 1.1568844221105527e-05, "loss": 1.2602, "step": 96200 }, { "epoch": 4.81, "grad_norm": 2.3539419174194336, "learning_rate": 1.1267336683417085e-05, "loss": 1.236, "step": 96300 }, { "epoch": 4.82, "grad_norm": 2.901660442352295, "learning_rate": 1.0965829145728641e-05, "loss": 1.2401, "step": 96400 }, { "epoch": 4.83, "grad_norm": 2.2228598594665527, "learning_rate": 1.0664321608040199e-05, "loss": 1.2718, "step": 96500 }, { "epoch": 4.83, "grad_norm": 2.1923272609710693, "learning_rate": 1.0362814070351758e-05, "loss": 1.2642, "step": 96600 }, { "epoch": 4.83, "grad_norm": 3.3871023654937744, "learning_rate": 1.0061306532663316e-05, "loss": 1.2247, "step": 96700 }, { "epoch": 4.84, "grad_norm": 4.223761081695557, "learning_rate": 9.759798994974874e-06, "loss": 1.2761, "step": 96800 }, { "epoch": 4.84, "grad_norm": 4.008758068084717, "learning_rate": 9.458291457286431e-06, "loss": 1.2895, "step": 96900 }, { "epoch": 4.85, "grad_norm": 3.956162214279175, "learning_rate": 9.156783919597989e-06, "loss": 1.2573, "step": 97000 }, { "epoch": 4.85, "eval_loss": 1.2709555625915527, "eval_runtime": 37.1777, "eval_samples_per_second": 26.898, "eval_steps_per_second": 3.362, "step": 97000 }, { "epoch": 4.86, "grad_norm": 1.0692745447158813, "learning_rate": 8.855276381909547e-06, "loss": 1.242, "step": 97100 }, { "epoch": 4.86, "grad_norm": 2.2347793579101562, "learning_rate": 8.553768844221105e-06, "loss": 1.2211, "step": 97200 }, { "epoch": 4.87, "grad_norm": 3.8911612033843994, "learning_rate": 8.252261306532662e-06, "loss": 1.2532, "step": 97300 }, { "epoch": 4.87, "grad_norm": 3.3509294986724854, "learning_rate": 7.95075376884422e-06, "loss": 1.2599, "step": 97400 }, { "epoch": 4.88, "grad_norm": 2.3662168979644775, "learning_rate": 7.649246231155778e-06, "loss": 1.2768, "step": 97500 }, { "epoch": 4.88, "grad_norm": 3.0100605487823486, "learning_rate": 7.3477386934673365e-06, "loss": 1.266, "step": 97600 }, { "epoch": 4.88, "grad_norm": 3.673358201980591, "learning_rate": 7.046231155778893e-06, "loss": 1.2288, "step": 97700 }, { "epoch": 4.89, "grad_norm": 3.1406137943267822, "learning_rate": 6.744723618090451e-06, "loss": 1.2193, "step": 97800 }, { "epoch": 4.89, "grad_norm": 2.9519011974334717, "learning_rate": 6.44321608040201e-06, "loss": 1.2347, "step": 97900 }, { "epoch": 4.9, "grad_norm": 2.01814866065979, "learning_rate": 6.141708542713567e-06, "loss": 1.242, "step": 98000 }, { "epoch": 4.9, "eval_loss": 1.2466384172439575, "eval_runtime": 37.1499, "eval_samples_per_second": 26.918, "eval_steps_per_second": 3.365, "step": 98000 }, { "epoch": 4.91, "grad_norm": 6.0040059089660645, "learning_rate": 5.840201005025125e-06, "loss": 1.2014, "step": 98100 }, { "epoch": 4.91, "grad_norm": 5.865699291229248, "learning_rate": 5.538693467336683e-06, "loss": 1.2544, "step": 98200 }, { "epoch": 4.92, "grad_norm": 3.115818500518799, "learning_rate": 5.23718592964824e-06, "loss": 1.2369, "step": 98300 }, { "epoch": 4.92, "grad_norm": 2.521693468093872, "learning_rate": 4.9356783919597985e-06, "loss": 1.2166, "step": 98400 }, { "epoch": 4.92, "grad_norm": 3.7152583599090576, "learning_rate": 4.634170854271356e-06, "loss": 1.2458, "step": 98500 }, { "epoch": 4.93, "grad_norm": 8.62978458404541, "learning_rate": 4.332663316582914e-06, "loss": 1.2525, "step": 98600 }, { "epoch": 4.94, "grad_norm": 3.633636951446533, "learning_rate": 4.031155778894472e-06, "loss": 1.2316, "step": 98700 }, { "epoch": 4.94, "grad_norm": 4.357114791870117, "learning_rate": 3.7296482412060296e-06, "loss": 1.2195, "step": 98800 }, { "epoch": 4.95, "grad_norm": 1.2212806940078735, "learning_rate": 3.4281407035175877e-06, "loss": 1.1938, "step": 98900 }, { "epoch": 4.95, "grad_norm": 3.4345273971557617, "learning_rate": 3.1266331658291455e-06, "loss": 1.2312, "step": 99000 }, { "epoch": 4.95, "eval_loss": 1.2566660642623901, "eval_runtime": 37.0254, "eval_samples_per_second": 27.009, "eval_steps_per_second": 3.376, "step": 99000 }, { "epoch": 4.96, "grad_norm": 4.7800188064575195, "learning_rate": 2.8251256281407033e-06, "loss": 1.1972, "step": 99100 }, { "epoch": 4.96, "grad_norm": 2.3637685775756836, "learning_rate": 2.5236180904522614e-06, "loss": 1.2475, "step": 99200 }, { "epoch": 4.96, "grad_norm": 3.591388463973999, "learning_rate": 2.2221105527638188e-06, "loss": 1.2505, "step": 99300 }, { "epoch": 4.97, "grad_norm": 3.078840970993042, "learning_rate": 1.9236180904522612e-06, "loss": 1.2421, "step": 99400 }, { "epoch": 4.97, "grad_norm": 3.289644479751587, "learning_rate": 1.622110552763819e-06, "loss": 1.257, "step": 99500 }, { "epoch": 4.98, "grad_norm": 4.775394439697266, "learning_rate": 1.3206030150753765e-06, "loss": 1.256, "step": 99600 }, { "epoch": 4.99, "grad_norm": 4.110093116760254, "learning_rate": 1.0190954773869345e-06, "loss": 1.1982, "step": 99700 }, { "epoch": 4.99, "grad_norm": 1.86680006980896, "learning_rate": 7.175879396984924e-07, "loss": 1.2571, "step": 99800 }, { "epoch": 5.0, "grad_norm": 3.5120298862457275, "learning_rate": 4.160804020100502e-07, "loss": 1.2196, "step": 99900 }, { "epoch": 5.0, "grad_norm": 2.356171131134033, "learning_rate": 1.1457286432160803e-07, "loss": 1.2164, "step": 100000 }, { "epoch": 5.0, "eval_loss": 1.2736848592758179, "eval_runtime": 37.4852, "eval_samples_per_second": 26.677, "eval_steps_per_second": 3.335, "step": 100000 } ], "logging_steps": 100, "max_steps": 100000, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 1000, "total_flos": 1.2076594495488e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }