|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9958960328317374, |
|
"eval_steps": 250, |
|
"global_step": 1095, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0027359781121751026, |
|
"grad_norm": 0.24075667560100555, |
|
"learning_rate": 2.2727272727272728e-06, |
|
"loss": 1.9078, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.005471956224350205, |
|
"grad_norm": 0.2584983706474304, |
|
"learning_rate": 4.5454545454545455e-06, |
|
"loss": 2.1152, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.008207934336525308, |
|
"grad_norm": 0.24978317320346832, |
|
"learning_rate": 6.818181818181818e-06, |
|
"loss": 1.9649, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01094391244870041, |
|
"grad_norm": 0.2477930635213852, |
|
"learning_rate": 9.090909090909091e-06, |
|
"loss": 1.9163, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.013679890560875513, |
|
"grad_norm": 0.2425757646560669, |
|
"learning_rate": 1.1363636363636365e-05, |
|
"loss": 2.1254, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.016415868673050615, |
|
"grad_norm": 0.251596599817276, |
|
"learning_rate": 1.3636363636363637e-05, |
|
"loss": 2.0832, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.019151846785225718, |
|
"grad_norm": 0.2359585464000702, |
|
"learning_rate": 1.590909090909091e-05, |
|
"loss": 1.8204, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.02188782489740082, |
|
"grad_norm": 0.25108030438423157, |
|
"learning_rate": 1.8181818181818182e-05, |
|
"loss": 1.9224, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.024623803009575923, |
|
"grad_norm": 0.25240400433540344, |
|
"learning_rate": 2.0454545454545457e-05, |
|
"loss": 1.8371, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.027359781121751026, |
|
"grad_norm": 0.22698192298412323, |
|
"learning_rate": 2.272727272727273e-05, |
|
"loss": 1.9962, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.030095759233926128, |
|
"grad_norm": 0.24213118851184845, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.9918, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.03283173734610123, |
|
"grad_norm": 0.20904198288917542, |
|
"learning_rate": 2.7272727272727273e-05, |
|
"loss": 1.9463, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.03556771545827633, |
|
"grad_norm": 0.21374110877513885, |
|
"learning_rate": 2.954545454545455e-05, |
|
"loss": 1.8628, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.038303693570451436, |
|
"grad_norm": 0.27983948588371277, |
|
"learning_rate": 3.181818181818182e-05, |
|
"loss": 2.1597, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.04103967168262654, |
|
"grad_norm": 0.2309175282716751, |
|
"learning_rate": 3.409090909090909e-05, |
|
"loss": 2.0305, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.04377564979480164, |
|
"grad_norm": 0.21376170217990875, |
|
"learning_rate": 3.6363636363636364e-05, |
|
"loss": 1.8866, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.046511627906976744, |
|
"grad_norm": 0.23382064700126648, |
|
"learning_rate": 3.8636363636363636e-05, |
|
"loss": 1.9912, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.049247606019151846, |
|
"grad_norm": 0.20445913076400757, |
|
"learning_rate": 4.0909090909090915e-05, |
|
"loss": 1.9773, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.05198358413132695, |
|
"grad_norm": 0.21313798427581787, |
|
"learning_rate": 4.318181818181819e-05, |
|
"loss": 1.9965, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.05471956224350205, |
|
"grad_norm": 0.22150328755378723, |
|
"learning_rate": 4.545454545454546e-05, |
|
"loss": 1.9538, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.057455540355677154, |
|
"grad_norm": 0.17602485418319702, |
|
"learning_rate": 4.772727272727273e-05, |
|
"loss": 1.9113, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.060191518467852256, |
|
"grad_norm": 0.20136825740337372, |
|
"learning_rate": 5e-05, |
|
"loss": 1.9837, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.06292749658002736, |
|
"grad_norm": 0.1977069228887558, |
|
"learning_rate": 4.999989284560115e-05, |
|
"loss": 1.8121, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.06566347469220246, |
|
"grad_norm": 0.18849150836467743, |
|
"learning_rate": 4.9999571383323136e-05, |
|
"loss": 1.8013, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.06839945280437756, |
|
"grad_norm": 0.17495211958885193, |
|
"learning_rate": 4.9999035615921664e-05, |
|
"loss": 1.8375, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.07113543091655267, |
|
"grad_norm": 0.28524237871170044, |
|
"learning_rate": 4.999828554798952e-05, |
|
"loss": 2.026, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.07387140902872777, |
|
"grad_norm": 0.2135252058506012, |
|
"learning_rate": 4.999732118595654e-05, |
|
"loss": 1.7874, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.07660738714090287, |
|
"grad_norm": 0.213677778840065, |
|
"learning_rate": 4.999614253808959e-05, |
|
"loss": 1.7844, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.07934336525307797, |
|
"grad_norm": 0.20930789411067963, |
|
"learning_rate": 4.999474961449243e-05, |
|
"loss": 1.857, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.08207934336525308, |
|
"grad_norm": 0.18232876062393188, |
|
"learning_rate": 4.999314242710572e-05, |
|
"loss": 1.7145, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.08481532147742818, |
|
"grad_norm": 0.2061523050069809, |
|
"learning_rate": 4.9991320989706816e-05, |
|
"loss": 1.8137, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.08755129958960328, |
|
"grad_norm": 0.19810044765472412, |
|
"learning_rate": 4.9989285317909725e-05, |
|
"loss": 1.8154, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.09028727770177838, |
|
"grad_norm": 0.19258971512317657, |
|
"learning_rate": 4.9987035429164954e-05, |
|
"loss": 1.8612, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.09302325581395349, |
|
"grad_norm": 0.21273574233055115, |
|
"learning_rate": 4.9984571342759326e-05, |
|
"loss": 1.6282, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.09575923392612859, |
|
"grad_norm": 0.18505199253559113, |
|
"learning_rate": 4.998189307981586e-05, |
|
"loss": 1.7134, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.09849521203830369, |
|
"grad_norm": 0.20874658226966858, |
|
"learning_rate": 4.997900066329357e-05, |
|
"loss": 1.7648, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.1012311901504788, |
|
"grad_norm": 0.18423666059970856, |
|
"learning_rate": 4.9975894117987265e-05, |
|
"loss": 1.7784, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.1039671682626539, |
|
"grad_norm": 0.16870878636837006, |
|
"learning_rate": 4.9972573470527354e-05, |
|
"loss": 1.7352, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.106703146374829, |
|
"grad_norm": 0.20023159682750702, |
|
"learning_rate": 4.996903874937959e-05, |
|
"loss": 1.7784, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.1094391244870041, |
|
"grad_norm": 0.22696524858474731, |
|
"learning_rate": 4.9965289984844846e-05, |
|
"loss": 1.7722, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.1121751025991792, |
|
"grad_norm": 0.18610844016075134, |
|
"learning_rate": 4.9961327209058844e-05, |
|
"loss": 1.761, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.11491108071135431, |
|
"grad_norm": 0.17688068747520447, |
|
"learning_rate": 4.99571504559919e-05, |
|
"loss": 1.6316, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.11764705882352941, |
|
"grad_norm": 0.20090024173259735, |
|
"learning_rate": 4.995275976144862e-05, |
|
"loss": 1.8064, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.12038303693570451, |
|
"grad_norm": 0.17762398719787598, |
|
"learning_rate": 4.9948155163067565e-05, |
|
"loss": 1.8417, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.12311901504787962, |
|
"grad_norm": 0.1874111294746399, |
|
"learning_rate": 4.9943336700320985e-05, |
|
"loss": 1.7245, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.12585499316005472, |
|
"grad_norm": 0.16139264404773712, |
|
"learning_rate": 4.993830441451444e-05, |
|
"loss": 1.8852, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.12859097127222982, |
|
"grad_norm": 0.16358217597007751, |
|
"learning_rate": 4.993305834878643e-05, |
|
"loss": 1.7314, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.13132694938440492, |
|
"grad_norm": 0.16465303301811218, |
|
"learning_rate": 4.992759854810811e-05, |
|
"loss": 1.6362, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.13406292749658003, |
|
"grad_norm": 0.1597519814968109, |
|
"learning_rate": 4.99219250592828e-05, |
|
"loss": 1.5936, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.13679890560875513, |
|
"grad_norm": 0.16433416306972504, |
|
"learning_rate": 4.991603793094564e-05, |
|
"loss": 1.5236, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.13953488372093023, |
|
"grad_norm": 0.16490432620048523, |
|
"learning_rate": 4.9909937213563165e-05, |
|
"loss": 1.7191, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.14227086183310533, |
|
"grad_norm": 0.1576634645462036, |
|
"learning_rate": 4.990362295943287e-05, |
|
"loss": 1.5944, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.14500683994528044, |
|
"grad_norm": 0.16687043011188507, |
|
"learning_rate": 4.989709522268278e-05, |
|
"loss": 1.7204, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.14774281805745554, |
|
"grad_norm": 0.18680374324321747, |
|
"learning_rate": 4.989035405927093e-05, |
|
"loss": 1.8231, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.15047879616963064, |
|
"grad_norm": 0.16919435560703278, |
|
"learning_rate": 4.9883399526984944e-05, |
|
"loss": 1.7527, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.15321477428180574, |
|
"grad_norm": 0.16811755299568176, |
|
"learning_rate": 4.9876231685441545e-05, |
|
"loss": 1.6824, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.15595075239398085, |
|
"grad_norm": 0.15195395052433014, |
|
"learning_rate": 4.986885059608597e-05, |
|
"loss": 1.6515, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.15868673050615595, |
|
"grad_norm": 0.17854413390159607, |
|
"learning_rate": 4.986125632219152e-05, |
|
"loss": 1.613, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.16142270861833105, |
|
"grad_norm": 0.168907031416893, |
|
"learning_rate": 4.985344892885899e-05, |
|
"loss": 1.798, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.16415868673050615, |
|
"grad_norm": 0.173956036567688, |
|
"learning_rate": 4.98454284830161e-05, |
|
"loss": 1.4941, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.16689466484268126, |
|
"grad_norm": 0.19879266619682312, |
|
"learning_rate": 4.983719505341693e-05, |
|
"loss": 1.7941, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.16963064295485636, |
|
"grad_norm": 0.16308054327964783, |
|
"learning_rate": 4.982874871064134e-05, |
|
"loss": 1.6984, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.17236662106703146, |
|
"grad_norm": 0.16789625585079193, |
|
"learning_rate": 4.982008952709435e-05, |
|
"loss": 1.6375, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.17510259917920656, |
|
"grad_norm": 0.16414505243301392, |
|
"learning_rate": 4.9811217577005533e-05, |
|
"loss": 1.5667, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.17783857729138167, |
|
"grad_norm": 0.16586610674858093, |
|
"learning_rate": 4.980213293642837e-05, |
|
"loss": 1.6688, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.18057455540355677, |
|
"grad_norm": 0.21007002890110016, |
|
"learning_rate": 4.979283568323958e-05, |
|
"loss": 1.7249, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.18331053351573187, |
|
"grad_norm": 0.15806783735752106, |
|
"learning_rate": 4.97833258971385e-05, |
|
"loss": 1.6918, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.18604651162790697, |
|
"grad_norm": 0.17978119850158691, |
|
"learning_rate": 4.977360365964637e-05, |
|
"loss": 1.6815, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.18878248974008208, |
|
"grad_norm": 0.16976381838321686, |
|
"learning_rate": 4.976366905410562e-05, |
|
"loss": 1.7212, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.19151846785225718, |
|
"grad_norm": 0.18810401856899261, |
|
"learning_rate": 4.9753522165679195e-05, |
|
"loss": 1.707, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.19425444596443228, |
|
"grad_norm": 0.17066176235675812, |
|
"learning_rate": 4.974316308134978e-05, |
|
"loss": 1.7765, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.19699042407660738, |
|
"grad_norm": 0.17143549025058746, |
|
"learning_rate": 4.9732591889919114e-05, |
|
"loss": 1.7317, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.1997264021887825, |
|
"grad_norm": 0.15364663302898407, |
|
"learning_rate": 4.972180868200714e-05, |
|
"loss": 1.6406, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.2024623803009576, |
|
"grad_norm": 0.1750328093767166, |
|
"learning_rate": 4.971081355005133e-05, |
|
"loss": 1.8737, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.2051983584131327, |
|
"grad_norm": 0.17950010299682617, |
|
"learning_rate": 4.969960658830582e-05, |
|
"loss": 1.7941, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.2079343365253078, |
|
"grad_norm": 0.1876641809940338, |
|
"learning_rate": 4.968818789284063e-05, |
|
"loss": 1.5792, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.2106703146374829, |
|
"grad_norm": 0.17048972845077515, |
|
"learning_rate": 4.9676557561540845e-05, |
|
"loss": 1.607, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.213406292749658, |
|
"grad_norm": 0.18965713679790497, |
|
"learning_rate": 4.966471569410574e-05, |
|
"loss": 1.6142, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.2161422708618331, |
|
"grad_norm": 0.16774089634418488, |
|
"learning_rate": 4.965266239204798e-05, |
|
"loss": 1.6446, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.2188782489740082, |
|
"grad_norm": 0.16337555646896362, |
|
"learning_rate": 4.9640397758692715e-05, |
|
"loss": 1.6319, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.2216142270861833, |
|
"grad_norm": 0.17532555758953094, |
|
"learning_rate": 4.9627921899176686e-05, |
|
"loss": 1.554, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.2243502051983584, |
|
"grad_norm": 0.16369390487670898, |
|
"learning_rate": 4.9615234920447365e-05, |
|
"loss": 1.7499, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.2270861833105335, |
|
"grad_norm": 0.170010045170784, |
|
"learning_rate": 4.9602336931261995e-05, |
|
"loss": 1.7786, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.22982216142270862, |
|
"grad_norm": 0.18399612605571747, |
|
"learning_rate": 4.958922804218667e-05, |
|
"loss": 1.8338, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.23255813953488372, |
|
"grad_norm": 0.16553553938865662, |
|
"learning_rate": 4.957590836559541e-05, |
|
"loss": 1.8186, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.23529411764705882, |
|
"grad_norm": 0.17777277529239655, |
|
"learning_rate": 4.9562378015669166e-05, |
|
"loss": 1.7326, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.23803009575923392, |
|
"grad_norm": 0.18170584738254547, |
|
"learning_rate": 4.954863710839485e-05, |
|
"loss": 1.7293, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.24076607387140903, |
|
"grad_norm": 0.17840476334095, |
|
"learning_rate": 4.953468576156437e-05, |
|
"loss": 1.6941, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.24350205198358413, |
|
"grad_norm": 0.16968044638633728, |
|
"learning_rate": 4.9520524094773575e-05, |
|
"loss": 1.7643, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.24623803009575923, |
|
"grad_norm": 0.17756614089012146, |
|
"learning_rate": 4.950615222942125e-05, |
|
"loss": 1.7599, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.24897400820793433, |
|
"grad_norm": 0.162063866853714, |
|
"learning_rate": 4.94915702887081e-05, |
|
"loss": 1.7365, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.25170998632010944, |
|
"grad_norm": 0.18559984862804413, |
|
"learning_rate": 4.947677839763563e-05, |
|
"loss": 1.4424, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.25444596443228457, |
|
"grad_norm": 0.17151783406734467, |
|
"learning_rate": 4.946177668300515e-05, |
|
"loss": 1.5231, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.25718194254445964, |
|
"grad_norm": 0.20623093843460083, |
|
"learning_rate": 4.944656527341663e-05, |
|
"loss": 1.8024, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.25991792065663477, |
|
"grad_norm": 0.19318287074565887, |
|
"learning_rate": 4.943114429926763e-05, |
|
"loss": 1.6366, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.26265389876880985, |
|
"grad_norm": 0.1877240240573883, |
|
"learning_rate": 4.941551389275217e-05, |
|
"loss": 1.5424, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.265389876880985, |
|
"grad_norm": 0.17889100313186646, |
|
"learning_rate": 4.939967418785959e-05, |
|
"loss": 1.6603, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.26812585499316005, |
|
"grad_norm": 0.18533948063850403, |
|
"learning_rate": 4.938362532037341e-05, |
|
"loss": 1.7546, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.2708618331053352, |
|
"grad_norm": 0.19442051649093628, |
|
"learning_rate": 4.936736742787018e-05, |
|
"loss": 1.6938, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.27359781121751026, |
|
"grad_norm": 0.20610813796520233, |
|
"learning_rate": 4.935090064971827e-05, |
|
"loss": 1.7104, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.2763337893296854, |
|
"grad_norm": 0.19741714000701904, |
|
"learning_rate": 4.933422512707671e-05, |
|
"loss": 1.6205, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.27906976744186046, |
|
"grad_norm": 0.20613451302051544, |
|
"learning_rate": 4.931734100289393e-05, |
|
"loss": 1.7549, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.2818057455540356, |
|
"grad_norm": 0.19265608489513397, |
|
"learning_rate": 4.930024842190658e-05, |
|
"loss": 1.8493, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.28454172366621067, |
|
"grad_norm": 0.1976325511932373, |
|
"learning_rate": 4.9282947530638295e-05, |
|
"loss": 1.8059, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.2872777017783858, |
|
"grad_norm": 0.20444822311401367, |
|
"learning_rate": 4.926543847739841e-05, |
|
"loss": 1.5852, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.29001367989056087, |
|
"grad_norm": 0.18967504799365997, |
|
"learning_rate": 4.9247721412280667e-05, |
|
"loss": 1.6282, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.292749658002736, |
|
"grad_norm": 0.206316277384758, |
|
"learning_rate": 4.9229796487162e-05, |
|
"loss": 1.6186, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.2954856361149111, |
|
"grad_norm": 0.17524588108062744, |
|
"learning_rate": 4.9211663855701165e-05, |
|
"loss": 1.6383, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.2982216142270862, |
|
"grad_norm": 0.2032768577337265, |
|
"learning_rate": 4.9193323673337476e-05, |
|
"loss": 1.785, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.3009575923392613, |
|
"grad_norm": 0.17396964132785797, |
|
"learning_rate": 4.917477609728941e-05, |
|
"loss": 1.4461, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.3036935704514364, |
|
"grad_norm": 0.19011397659778595, |
|
"learning_rate": 4.915602128655333e-05, |
|
"loss": 1.7313, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.3064295485636115, |
|
"grad_norm": 0.19239680469036102, |
|
"learning_rate": 4.913705940190207e-05, |
|
"loss": 1.6015, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.3091655266757866, |
|
"grad_norm": 0.19917477667331696, |
|
"learning_rate": 4.9117890605883574e-05, |
|
"loss": 1.6411, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.3119015047879617, |
|
"grad_norm": 0.19318480789661407, |
|
"learning_rate": 4.909851506281952e-05, |
|
"loss": 1.6737, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.3146374829001368, |
|
"grad_norm": 0.1994866579771042, |
|
"learning_rate": 4.907893293880387e-05, |
|
"loss": 1.518, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.3173734610123119, |
|
"grad_norm": 0.1941988468170166, |
|
"learning_rate": 4.905914440170147e-05, |
|
"loss": 1.7589, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.320109439124487, |
|
"grad_norm": 0.20112977921962738, |
|
"learning_rate": 4.903914962114665e-05, |
|
"loss": 1.6671, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.3228454172366621, |
|
"grad_norm": 0.21141856908798218, |
|
"learning_rate": 4.9018948768541694e-05, |
|
"loss": 1.6731, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.32558139534883723, |
|
"grad_norm": 0.18585005402565002, |
|
"learning_rate": 4.899854201705542e-05, |
|
"loss": 1.6276, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.3283173734610123, |
|
"grad_norm": 0.1975102424621582, |
|
"learning_rate": 4.8977929541621676e-05, |
|
"loss": 1.6103, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.33105335157318744, |
|
"grad_norm": 0.19365696609020233, |
|
"learning_rate": 4.895711151893786e-05, |
|
"loss": 1.5401, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.3337893296853625, |
|
"grad_norm": 0.20823922753334045, |
|
"learning_rate": 4.8936088127463396e-05, |
|
"loss": 1.5811, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.33652530779753764, |
|
"grad_norm": 0.2523277997970581, |
|
"learning_rate": 4.8914859547418186e-05, |
|
"loss": 1.5561, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.3392612859097127, |
|
"grad_norm": 0.19290074706077576, |
|
"learning_rate": 4.889342596078109e-05, |
|
"loss": 1.6078, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.34199726402188785, |
|
"grad_norm": 0.17994068562984467, |
|
"learning_rate": 4.887178755128835e-05, |
|
"loss": 1.7232, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.3447332421340629, |
|
"grad_norm": 0.18888920545578003, |
|
"learning_rate": 4.8849944504432035e-05, |
|
"loss": 1.5538, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.34746922024623805, |
|
"grad_norm": 0.18798014521598816, |
|
"learning_rate": 4.882789700745843e-05, |
|
"loss": 1.6081, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.35020519835841313, |
|
"grad_norm": 0.19080984592437744, |
|
"learning_rate": 4.880564524936643e-05, |
|
"loss": 1.4809, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.35294117647058826, |
|
"grad_norm": 0.19421504437923431, |
|
"learning_rate": 4.878318942090594e-05, |
|
"loss": 1.6523, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.35567715458276333, |
|
"grad_norm": 0.1982857584953308, |
|
"learning_rate": 4.876052971457623e-05, |
|
"loss": 1.6948, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.35841313269493846, |
|
"grad_norm": 0.20162495970726013, |
|
"learning_rate": 4.8737666324624265e-05, |
|
"loss": 1.7504, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.36114911080711354, |
|
"grad_norm": 0.20931164920330048, |
|
"learning_rate": 4.8714599447043085e-05, |
|
"loss": 1.4829, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.36388508891928867, |
|
"grad_norm": 0.21997657418251038, |
|
"learning_rate": 4.869132927957007e-05, |
|
"loss": 1.5546, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.36662106703146374, |
|
"grad_norm": 0.20648987591266632, |
|
"learning_rate": 4.866785602168528e-05, |
|
"loss": 1.6716, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.3693570451436389, |
|
"grad_norm": 0.20821620523929596, |
|
"learning_rate": 4.864417987460975e-05, |
|
"loss": 1.6197, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.37209302325581395, |
|
"grad_norm": 0.21084287762641907, |
|
"learning_rate": 4.862030104130375e-05, |
|
"loss": 1.6454, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.3748290013679891, |
|
"grad_norm": 0.21337808668613434, |
|
"learning_rate": 4.859621972646503e-05, |
|
"loss": 1.5733, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.37756497948016415, |
|
"grad_norm": 0.21093016862869263, |
|
"learning_rate": 4.857193613652711e-05, |
|
"loss": 1.6716, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.3803009575923393, |
|
"grad_norm": 0.20863549411296844, |
|
"learning_rate": 4.854745047965745e-05, |
|
"loss": 1.6117, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.38303693570451436, |
|
"grad_norm": 0.20271432399749756, |
|
"learning_rate": 4.852276296575573e-05, |
|
"loss": 1.5765, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.3857729138166895, |
|
"grad_norm": 0.20297355949878693, |
|
"learning_rate": 4.849787380645201e-05, |
|
"loss": 1.7296, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.38850889192886456, |
|
"grad_norm": 0.20244932174682617, |
|
"learning_rate": 4.847278321510491e-05, |
|
"loss": 1.7651, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.3912448700410397, |
|
"grad_norm": 0.20252294838428497, |
|
"learning_rate": 4.844749140679982e-05, |
|
"loss": 1.6427, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.39398084815321477, |
|
"grad_norm": 0.20163066685199738, |
|
"learning_rate": 4.842199859834702e-05, |
|
"loss": 1.6398, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.3967168262653899, |
|
"grad_norm": 0.21187250316143036, |
|
"learning_rate": 4.839630500827982e-05, |
|
"loss": 1.6173, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.399452804377565, |
|
"grad_norm": 0.18655481934547424, |
|
"learning_rate": 4.837041085685273e-05, |
|
"loss": 1.5184, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.4021887824897401, |
|
"grad_norm": 0.21056880056858063, |
|
"learning_rate": 4.834431636603953e-05, |
|
"loss": 1.5687, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.4049247606019152, |
|
"grad_norm": 0.2020365595817566, |
|
"learning_rate": 4.831802175953138e-05, |
|
"loss": 1.6445, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.4076607387140903, |
|
"grad_norm": 0.18952035903930664, |
|
"learning_rate": 4.8291527262734886e-05, |
|
"loss": 1.7945, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.4103967168262654, |
|
"grad_norm": 0.1896515041589737, |
|
"learning_rate": 4.826483310277021e-05, |
|
"loss": 1.546, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.4131326949384405, |
|
"grad_norm": 0.20328114926815033, |
|
"learning_rate": 4.8237939508469075e-05, |
|
"loss": 1.5621, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.4158686730506156, |
|
"grad_norm": 0.18790675699710846, |
|
"learning_rate": 4.821084671037285e-05, |
|
"loss": 1.6609, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.4186046511627907, |
|
"grad_norm": 0.20515787601470947, |
|
"learning_rate": 4.8183554940730524e-05, |
|
"loss": 1.671, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.4213406292749658, |
|
"grad_norm": 0.19960254430770874, |
|
"learning_rate": 4.815606443349675e-05, |
|
"loss": 1.4732, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.4240766073871409, |
|
"grad_norm": 0.21313494443893433, |
|
"learning_rate": 4.812837542432984e-05, |
|
"loss": 1.7094, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.426812585499316, |
|
"grad_norm": 0.1988086998462677, |
|
"learning_rate": 4.8100488150589704e-05, |
|
"loss": 1.6241, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.42954856361149113, |
|
"grad_norm": 0.1885240077972412, |
|
"learning_rate": 4.80724028513359e-05, |
|
"loss": 1.64, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.4322845417236662, |
|
"grad_norm": 0.19793474674224854, |
|
"learning_rate": 4.8044119767325455e-05, |
|
"loss": 1.6615, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.43502051983584133, |
|
"grad_norm": 0.2065788209438324, |
|
"learning_rate": 4.801563914101096e-05, |
|
"loss": 1.5414, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.4377564979480164, |
|
"grad_norm": 0.19855208694934845, |
|
"learning_rate": 4.798696121653833e-05, |
|
"loss": 1.7804, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.44049247606019154, |
|
"grad_norm": 0.21048115193843842, |
|
"learning_rate": 4.795808623974485e-05, |
|
"loss": 1.6356, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.4432284541723666, |
|
"grad_norm": 0.18900255858898163, |
|
"learning_rate": 4.792901445815698e-05, |
|
"loss": 1.7154, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.44596443228454175, |
|
"grad_norm": 0.21627336740493774, |
|
"learning_rate": 4.7899746120988245e-05, |
|
"loss": 1.5644, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.4487004103967168, |
|
"grad_norm": 0.22913017868995667, |
|
"learning_rate": 4.7870281479137146e-05, |
|
"loss": 1.5764, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.45143638850889195, |
|
"grad_norm": 0.1932600438594818, |
|
"learning_rate": 4.784062078518496e-05, |
|
"loss": 1.7484, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.454172366621067, |
|
"grad_norm": 0.2302398383617401, |
|
"learning_rate": 4.781076429339359e-05, |
|
"loss": 1.5941, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.45690834473324216, |
|
"grad_norm": 0.2317684292793274, |
|
"learning_rate": 4.77807122597034e-05, |
|
"loss": 1.6844, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.45964432284541723, |
|
"grad_norm": 0.2009151577949524, |
|
"learning_rate": 4.775046494173098e-05, |
|
"loss": 1.735, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.46238030095759236, |
|
"grad_norm": 0.23902402818202972, |
|
"learning_rate": 4.7720022598767e-05, |
|
"loss": 1.6606, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.46511627906976744, |
|
"grad_norm": 0.22881416976451874, |
|
"learning_rate": 4.768938549177393e-05, |
|
"loss": 1.6456, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.46785225718194257, |
|
"grad_norm": 0.1795547604560852, |
|
"learning_rate": 4.765855388338383e-05, |
|
"loss": 1.6853, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.47058823529411764, |
|
"grad_norm": 0.1893799602985382, |
|
"learning_rate": 4.7627528037896106e-05, |
|
"loss": 1.5587, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.47332421340629277, |
|
"grad_norm": 0.21993231773376465, |
|
"learning_rate": 4.759630822127522e-05, |
|
"loss": 1.8, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.47606019151846785, |
|
"grad_norm": 0.2114606648683548, |
|
"learning_rate": 4.756489470114842e-05, |
|
"loss": 1.543, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.478796169630643, |
|
"grad_norm": 0.24682380259037018, |
|
"learning_rate": 4.753328774680347e-05, |
|
"loss": 1.6537, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.48153214774281805, |
|
"grad_norm": 0.21504569053649902, |
|
"learning_rate": 4.750148762918629e-05, |
|
"loss": 1.7841, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.4842681258549932, |
|
"grad_norm": 0.2285577356815338, |
|
"learning_rate": 4.746949462089868e-05, |
|
"loss": 1.5911, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.48700410396716826, |
|
"grad_norm": 0.20859524607658386, |
|
"learning_rate": 4.743730899619598e-05, |
|
"loss": 1.5816, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.4897400820793434, |
|
"grad_norm": 0.2234918773174286, |
|
"learning_rate": 4.740493103098468e-05, |
|
"loss": 1.5763, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.49247606019151846, |
|
"grad_norm": 0.20745696127414703, |
|
"learning_rate": 4.7372361002820085e-05, |
|
"loss": 1.4441, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.4952120383036936, |
|
"grad_norm": 0.18486277759075165, |
|
"learning_rate": 4.733959919090396e-05, |
|
"loss": 1.6772, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.49794801641586867, |
|
"grad_norm": 0.21360483765602112, |
|
"learning_rate": 4.7306645876082066e-05, |
|
"loss": 1.5955, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.5006839945280438, |
|
"grad_norm": 0.2167551964521408, |
|
"learning_rate": 4.727350134084182e-05, |
|
"loss": 1.6222, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.5034199726402189, |
|
"grad_norm": 0.2134450525045395, |
|
"learning_rate": 4.7240165869309846e-05, |
|
"loss": 1.6059, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.506155950752394, |
|
"grad_norm": 0.21351881325244904, |
|
"learning_rate": 4.720663974724953e-05, |
|
"loss": 1.7258, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.5088919288645691, |
|
"grad_norm": 0.1952323615550995, |
|
"learning_rate": 4.71729232620586e-05, |
|
"loss": 1.4591, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.5116279069767442, |
|
"grad_norm": 0.2462124228477478, |
|
"learning_rate": 4.7139016702766615e-05, |
|
"loss": 1.5708, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.5143638850889193, |
|
"grad_norm": 0.24391762912273407, |
|
"learning_rate": 4.7104920360032545e-05, |
|
"loss": 1.7813, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.5170998632010944, |
|
"grad_norm": 0.21117964386940002, |
|
"learning_rate": 4.707063452614224e-05, |
|
"loss": 1.6948, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.5198358413132695, |
|
"grad_norm": 0.20722931623458862, |
|
"learning_rate": 4.703615949500593e-05, |
|
"loss": 1.5925, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.5225718194254446, |
|
"grad_norm": 0.22132059931755066, |
|
"learning_rate": 4.700149556215571e-05, |
|
"loss": 1.6448, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.5253077975376197, |
|
"grad_norm": 0.22742822766304016, |
|
"learning_rate": 4.696664302474302e-05, |
|
"loss": 1.5866, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.5280437756497948, |
|
"grad_norm": 0.2043454349040985, |
|
"learning_rate": 4.693160218153607e-05, |
|
"loss": 1.704, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.53077975376197, |
|
"grad_norm": 0.21721500158309937, |
|
"learning_rate": 4.68963733329173e-05, |
|
"loss": 1.6768, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.533515731874145, |
|
"grad_norm": 0.2336382120847702, |
|
"learning_rate": 4.6860956780880796e-05, |
|
"loss": 1.5906, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.5362517099863201, |
|
"grad_norm": 0.2269752472639084, |
|
"learning_rate": 4.6825352829029705e-05, |
|
"loss": 1.5308, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.5389876880984952, |
|
"grad_norm": 0.21256719529628754, |
|
"learning_rate": 4.6789561782573635e-05, |
|
"loss": 1.4821, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.5417236662106704, |
|
"grad_norm": 0.21226170659065247, |
|
"learning_rate": 4.675358394832603e-05, |
|
"loss": 1.7818, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.5444596443228454, |
|
"grad_norm": 0.21833762526512146, |
|
"learning_rate": 4.671741963470155e-05, |
|
"loss": 1.6556, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.5471956224350205, |
|
"grad_norm": 0.21411390602588654, |
|
"learning_rate": 4.668106915171341e-05, |
|
"loss": 1.5892, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.5499316005471956, |
|
"grad_norm": 0.22407633066177368, |
|
"learning_rate": 4.664453281097075e-05, |
|
"loss": 1.6843, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.5526675786593708, |
|
"grad_norm": 0.20786328613758087, |
|
"learning_rate": 4.660781092567593e-05, |
|
"loss": 1.6561, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.5554035567715458, |
|
"grad_norm": 0.2268221080303192, |
|
"learning_rate": 4.657090381062189e-05, |
|
"loss": 1.585, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.5581395348837209, |
|
"grad_norm": 0.2261151373386383, |
|
"learning_rate": 4.6533811782189385e-05, |
|
"loss": 1.6033, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.560875512995896, |
|
"grad_norm": 0.19799037277698517, |
|
"learning_rate": 4.649653515834436e-05, |
|
"loss": 1.6294, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.5636114911080712, |
|
"grad_norm": 0.20902524888515472, |
|
"learning_rate": 4.645907425863514e-05, |
|
"loss": 1.7012, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.5663474692202463, |
|
"grad_norm": 0.2055850476026535, |
|
"learning_rate": 4.642142940418973e-05, |
|
"loss": 1.5587, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.5690834473324213, |
|
"grad_norm": 0.2205619066953659, |
|
"learning_rate": 4.638360091771309e-05, |
|
"loss": 1.7727, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.5718194254445964, |
|
"grad_norm": 0.25723960995674133, |
|
"learning_rate": 4.6345589123484314e-05, |
|
"loss": 1.6037, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.5745554035567716, |
|
"grad_norm": 0.24420645833015442, |
|
"learning_rate": 4.630739434735387e-05, |
|
"loss": 1.7479, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.5772913816689467, |
|
"grad_norm": 0.24823623895645142, |
|
"learning_rate": 4.626901691674083e-05, |
|
"loss": 1.7085, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.5800273597811217, |
|
"grad_norm": 0.1966015100479126, |
|
"learning_rate": 4.623045716063002e-05, |
|
"loss": 1.544, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.5827633378932968, |
|
"grad_norm": 0.22418932616710663, |
|
"learning_rate": 4.6191715409569244e-05, |
|
"loss": 1.6949, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.585499316005472, |
|
"grad_norm": 0.24894970655441284, |
|
"learning_rate": 4.6152791995666445e-05, |
|
"loss": 1.5838, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.5882352941176471, |
|
"grad_norm": 0.230922132730484, |
|
"learning_rate": 4.61136872525868e-05, |
|
"loss": 1.6267, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.5909712722298222, |
|
"grad_norm": 0.22752411663532257, |
|
"learning_rate": 4.6074401515549934e-05, |
|
"loss": 1.4752, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.5937072503419972, |
|
"grad_norm": 0.22325444221496582, |
|
"learning_rate": 4.6034935121327025e-05, |
|
"loss": 1.7513, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.5964432284541724, |
|
"grad_norm": 0.21255116164684296, |
|
"learning_rate": 4.599528840823787e-05, |
|
"loss": 1.642, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.5991792065663475, |
|
"grad_norm": 0.24686329066753387, |
|
"learning_rate": 4.5955461716148065e-05, |
|
"loss": 1.6138, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.6019151846785226, |
|
"grad_norm": 0.20993772149085999, |
|
"learning_rate": 4.5915455386466014e-05, |
|
"loss": 1.7554, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.6046511627906976, |
|
"grad_norm": 0.2285137176513672, |
|
"learning_rate": 4.587526976214006e-05, |
|
"loss": 1.6073, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.6073871409028728, |
|
"grad_norm": 0.22348423302173615, |
|
"learning_rate": 4.5834905187655526e-05, |
|
"loss": 1.6707, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.6101231190150479, |
|
"grad_norm": 0.23014281690120697, |
|
"learning_rate": 4.5794362009031735e-05, |
|
"loss": 1.4968, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.612859097127223, |
|
"grad_norm": 0.23900607228279114, |
|
"learning_rate": 4.575364057381909e-05, |
|
"loss": 1.6317, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.615595075239398, |
|
"grad_norm": 0.24715475738048553, |
|
"learning_rate": 4.571274123109606e-05, |
|
"loss": 1.5805, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.6183310533515732, |
|
"grad_norm": 0.22451570630073547, |
|
"learning_rate": 4.5671664331466205e-05, |
|
"loss": 1.5956, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.6210670314637483, |
|
"grad_norm": 0.22505944967269897, |
|
"learning_rate": 4.5630410227055154e-05, |
|
"loss": 1.64, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.6238030095759234, |
|
"grad_norm": 0.24184450507164001, |
|
"learning_rate": 4.5588979271507625e-05, |
|
"loss": 1.6055, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.6265389876880985, |
|
"grad_norm": 0.22158478200435638, |
|
"learning_rate": 4.5547371819984344e-05, |
|
"loss": 1.5691, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.6292749658002736, |
|
"grad_norm": 0.23034054040908813, |
|
"learning_rate": 4.5505588229159025e-05, |
|
"loss": 1.6157, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.6320109439124487, |
|
"grad_norm": 0.240454763174057, |
|
"learning_rate": 4.5463628857215306e-05, |
|
"loss": 1.6337, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.6347469220246238, |
|
"grad_norm": 0.2200327068567276, |
|
"learning_rate": 4.5421494063843695e-05, |
|
"loss": 1.4941, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.6374829001367989, |
|
"grad_norm": 0.23234665393829346, |
|
"learning_rate": 4.537918421023848e-05, |
|
"loss": 1.5992, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.640218878248974, |
|
"grad_norm": 0.23742978274822235, |
|
"learning_rate": 4.53366996590946e-05, |
|
"loss": 1.481, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.6429548563611491, |
|
"grad_norm": 0.22323279082775116, |
|
"learning_rate": 4.5294040774604576e-05, |
|
"loss": 1.5291, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.6456908344733242, |
|
"grad_norm": 0.21182994544506073, |
|
"learning_rate": 4.525120792245538e-05, |
|
"loss": 1.5151, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.6484268125854993, |
|
"grad_norm": 0.24394573271274567, |
|
"learning_rate": 4.5208201469825304e-05, |
|
"loss": 1.6732, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.6511627906976745, |
|
"grad_norm": 0.20500163733959198, |
|
"learning_rate": 4.516502178538079e-05, |
|
"loss": 1.7066, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.6538987688098495, |
|
"grad_norm": 0.2329857349395752, |
|
"learning_rate": 4.512166923927329e-05, |
|
"loss": 1.4977, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.6566347469220246, |
|
"grad_norm": 0.23349729180335999, |
|
"learning_rate": 4.507814420313608e-05, |
|
"loss": 1.5837, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.6593707250341997, |
|
"grad_norm": 0.24870243668556213, |
|
"learning_rate": 4.503444705008107e-05, |
|
"loss": 1.6073, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.6621067031463749, |
|
"grad_norm": 0.2594373822212219, |
|
"learning_rate": 4.4990578154695676e-05, |
|
"loss": 1.6752, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.66484268125855, |
|
"grad_norm": 0.2545240819454193, |
|
"learning_rate": 4.4946537893039476e-05, |
|
"loss": 1.5185, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.667578659370725, |
|
"grad_norm": 0.24223731458187103, |
|
"learning_rate": 4.4902326642641095e-05, |
|
"loss": 1.7599, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.6703146374829001, |
|
"grad_norm": 0.23293475806713104, |
|
"learning_rate": 4.485794478249493e-05, |
|
"loss": 1.6919, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.6730506155950753, |
|
"grad_norm": 0.2452557235956192, |
|
"learning_rate": 4.481339269305792e-05, |
|
"loss": 1.4946, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.6757865937072504, |
|
"grad_norm": 0.23280148208141327, |
|
"learning_rate": 4.4768670756246237e-05, |
|
"loss": 1.548, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.6785225718194254, |
|
"grad_norm": 0.254730224609375, |
|
"learning_rate": 4.4723779355432056e-05, |
|
"loss": 1.6227, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.6812585499316005, |
|
"grad_norm": 0.25781765580177307, |
|
"learning_rate": 4.4678718875440276e-05, |
|
"loss": 1.6403, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.6839945280437757, |
|
"grad_norm": 0.2513294816017151, |
|
"learning_rate": 4.463348970254518e-05, |
|
"loss": 1.5416, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.6839945280437757, |
|
"eval_loss": 1.6065112352371216, |
|
"eval_runtime": 13.3947, |
|
"eval_samples_per_second": 11.497, |
|
"eval_steps_per_second": 11.497, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.6867305061559508, |
|
"grad_norm": 0.26413506269454956, |
|
"learning_rate": 4.458809222446717e-05, |
|
"loss": 1.4294, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.6894664842681258, |
|
"grad_norm": 0.24414490163326263, |
|
"learning_rate": 4.454252683036939e-05, |
|
"loss": 1.6371, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.6922024623803009, |
|
"grad_norm": 0.23543910682201385, |
|
"learning_rate": 4.449679391085444e-05, |
|
"loss": 1.6226, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.6949384404924761, |
|
"grad_norm": 0.24758242070674896, |
|
"learning_rate": 4.445089385796099e-05, |
|
"loss": 1.7229, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.6976744186046512, |
|
"grad_norm": 0.2550957500934601, |
|
"learning_rate": 4.4404827065160455e-05, |
|
"loss": 1.6636, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.7004103967168263, |
|
"grad_norm": 0.22393615543842316, |
|
"learning_rate": 4.4358593927353585e-05, |
|
"loss": 1.5795, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.7031463748290013, |
|
"grad_norm": 0.22478732466697693, |
|
"learning_rate": 4.431219484086712e-05, |
|
"loss": 1.6236, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.7058823529411765, |
|
"grad_norm": 0.21901032328605652, |
|
"learning_rate": 4.4265630203450355e-05, |
|
"loss": 1.5929, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.7086183310533516, |
|
"grad_norm": 0.2183765321969986, |
|
"learning_rate": 4.421890041427174e-05, |
|
"loss": 1.6372, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.7113543091655267, |
|
"grad_norm": 0.23780032992362976, |
|
"learning_rate": 4.417200587391547e-05, |
|
"loss": 1.4413, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.7140902872777017, |
|
"grad_norm": 0.22141356766223907, |
|
"learning_rate": 4.412494698437806e-05, |
|
"loss": 1.6594, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.7168262653898769, |
|
"grad_norm": 0.2270331084728241, |
|
"learning_rate": 4.407772414906487e-05, |
|
"loss": 1.6701, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.719562243502052, |
|
"grad_norm": 0.24480143189430237, |
|
"learning_rate": 4.403033777278666e-05, |
|
"loss": 1.4914, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.7222982216142271, |
|
"grad_norm": 0.23042482137680054, |
|
"learning_rate": 4.3982788261756114e-05, |
|
"loss": 1.6845, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.7250341997264022, |
|
"grad_norm": 0.23995645344257355, |
|
"learning_rate": 4.393507602358438e-05, |
|
"loss": 1.658, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.7277701778385773, |
|
"grad_norm": 0.25735434889793396, |
|
"learning_rate": 4.388720146727756e-05, |
|
"loss": 1.4613, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.7305061559507524, |
|
"grad_norm": 0.27036193013191223, |
|
"learning_rate": 4.3839165003233195e-05, |
|
"loss": 1.5341, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.7332421340629275, |
|
"grad_norm": 0.2285039722919464, |
|
"learning_rate": 4.3790967043236754e-05, |
|
"loss": 1.453, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.7359781121751026, |
|
"grad_norm": 0.2619383931159973, |
|
"learning_rate": 4.374260800045812e-05, |
|
"loss": 1.4931, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.7387140902872777, |
|
"grad_norm": 0.22855600714683533, |
|
"learning_rate": 4.369408828944801e-05, |
|
"loss": 1.7458, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.7414500683994528, |
|
"grad_norm": 0.23299537599086761, |
|
"learning_rate": 4.364540832613449e-05, |
|
"loss": 1.5653, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.7441860465116279, |
|
"grad_norm": 0.25914594531059265, |
|
"learning_rate": 4.35965685278193e-05, |
|
"loss": 1.6876, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.746922024623803, |
|
"grad_norm": 0.2430303692817688, |
|
"learning_rate": 4.3547569313174404e-05, |
|
"loss": 1.5761, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.7496580027359782, |
|
"grad_norm": 0.23968341946601868, |
|
"learning_rate": 4.34984111022383e-05, |
|
"loss": 1.5577, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.7523939808481532, |
|
"grad_norm": 0.25374993681907654, |
|
"learning_rate": 4.3449094316412485e-05, |
|
"loss": 1.6936, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.7551299589603283, |
|
"grad_norm": 0.25317367911338806, |
|
"learning_rate": 4.339961937845779e-05, |
|
"loss": 1.6957, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.7578659370725034, |
|
"grad_norm": 0.231050044298172, |
|
"learning_rate": 4.3349986712490795e-05, |
|
"loss": 1.5386, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.7606019151846786, |
|
"grad_norm": 0.23644736409187317, |
|
"learning_rate": 4.330019674398019e-05, |
|
"loss": 1.6485, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.7633378932968536, |
|
"grad_norm": 0.24971966445446014, |
|
"learning_rate": 4.32502498997431e-05, |
|
"loss": 1.6196, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.7660738714090287, |
|
"grad_norm": 0.22572147846221924, |
|
"learning_rate": 4.3200146607941444e-05, |
|
"loss": 1.6654, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.7688098495212038, |
|
"grad_norm": 0.256597101688385, |
|
"learning_rate": 4.3149887298078276e-05, |
|
"loss": 1.6552, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.771545827633379, |
|
"grad_norm": 0.226451575756073, |
|
"learning_rate": 4.3099472400994084e-05, |
|
"loss": 1.6164, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.774281805745554, |
|
"grad_norm": 0.2507132291793823, |
|
"learning_rate": 4.3048902348863116e-05, |
|
"loss": 1.555, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.7770177838577291, |
|
"grad_norm": 0.23009182512760162, |
|
"learning_rate": 4.2998177575189644e-05, |
|
"loss": 1.4741, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.7797537619699042, |
|
"grad_norm": 0.24451418220996857, |
|
"learning_rate": 4.294729851480429e-05, |
|
"loss": 1.5704, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.7824897400820794, |
|
"grad_norm": 0.2464141994714737, |
|
"learning_rate": 4.2896265603860246e-05, |
|
"loss": 1.6357, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.7852257181942545, |
|
"grad_norm": 0.24296295642852783, |
|
"learning_rate": 4.28450792798296e-05, |
|
"loss": 1.5492, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.7879616963064295, |
|
"grad_norm": 0.2681036591529846, |
|
"learning_rate": 4.279373998149954e-05, |
|
"loss": 1.6973, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.7906976744186046, |
|
"grad_norm": 0.23149123787879944, |
|
"learning_rate": 4.2742248148968576e-05, |
|
"loss": 1.6626, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.7934336525307798, |
|
"grad_norm": 0.22642168402671814, |
|
"learning_rate": 4.269060422364284e-05, |
|
"loss": 1.7018, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.7961696306429549, |
|
"grad_norm": 0.23424457013607025, |
|
"learning_rate": 4.263880864823221e-05, |
|
"loss": 1.6973, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.79890560875513, |
|
"grad_norm": 0.24797867238521576, |
|
"learning_rate": 4.2586861866746606e-05, |
|
"loss": 1.6631, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.801641586867305, |
|
"grad_norm": 0.24017377197742462, |
|
"learning_rate": 4.2534764324492115e-05, |
|
"loss": 1.692, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.8043775649794802, |
|
"grad_norm": 0.22952371835708618, |
|
"learning_rate": 4.248251646806719e-05, |
|
"loss": 1.3692, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.8071135430916553, |
|
"grad_norm": 0.2534586787223816, |
|
"learning_rate": 4.243011874535886e-05, |
|
"loss": 1.7223, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.8098495212038304, |
|
"grad_norm": 0.23179998993873596, |
|
"learning_rate": 4.237757160553883e-05, |
|
"loss": 1.6688, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.8125854993160054, |
|
"grad_norm": 0.21562454104423523, |
|
"learning_rate": 4.2324875499059693e-05, |
|
"loss": 1.5706, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.8153214774281806, |
|
"grad_norm": 0.30047473311424255, |
|
"learning_rate": 4.227203087765099e-05, |
|
"loss": 1.5275, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.8180574555403557, |
|
"grad_norm": 0.23342545330524445, |
|
"learning_rate": 4.221903819431543e-05, |
|
"loss": 1.6569, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.8207934336525308, |
|
"grad_norm": 0.2696247100830078, |
|
"learning_rate": 4.216589790332495e-05, |
|
"loss": 1.5803, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.8235294117647058, |
|
"grad_norm": 0.24961510300636292, |
|
"learning_rate": 4.21126104602168e-05, |
|
"loss": 1.4535, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.826265389876881, |
|
"grad_norm": 0.24784626066684723, |
|
"learning_rate": 4.205917632178972e-05, |
|
"loss": 1.6035, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.8290013679890561, |
|
"grad_norm": 0.22941957414150238, |
|
"learning_rate": 4.200559594609994e-05, |
|
"loss": 1.6502, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.8317373461012312, |
|
"grad_norm": 0.23505181074142456, |
|
"learning_rate": 4.195186979245728e-05, |
|
"loss": 1.7577, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.8344733242134063, |
|
"grad_norm": 0.2516261041164398, |
|
"learning_rate": 4.189799832142126e-05, |
|
"loss": 1.4947, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.8372093023255814, |
|
"grad_norm": 0.28713610768318176, |
|
"learning_rate": 4.1843981994797075e-05, |
|
"loss": 1.8651, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.8399452804377565, |
|
"grad_norm": 0.2537023425102234, |
|
"learning_rate": 4.178982127563169e-05, |
|
"loss": 1.3971, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.8426812585499316, |
|
"grad_norm": 0.2575843930244446, |
|
"learning_rate": 4.173551662820985e-05, |
|
"loss": 1.7023, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.8454172366621067, |
|
"grad_norm": 0.2560933828353882, |
|
"learning_rate": 4.168106851805009e-05, |
|
"loss": 1.5333, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.8481532147742818, |
|
"grad_norm": 0.22800202667713165, |
|
"learning_rate": 4.16264774119008e-05, |
|
"loss": 1.5548, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.8508891928864569, |
|
"grad_norm": 0.2175893485546112, |
|
"learning_rate": 4.1571743777736116e-05, |
|
"loss": 1.7173, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.853625170998632, |
|
"grad_norm": 0.23807293176651, |
|
"learning_rate": 4.151686808475204e-05, |
|
"loss": 1.5548, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.8563611491108071, |
|
"grad_norm": 0.26670193672180176, |
|
"learning_rate": 4.1461850803362314e-05, |
|
"loss": 1.5273, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.8590971272229823, |
|
"grad_norm": 0.2579667866230011, |
|
"learning_rate": 4.140669240519442e-05, |
|
"loss": 1.6407, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.8618331053351573, |
|
"grad_norm": 0.23136496543884277, |
|
"learning_rate": 4.135139336308559e-05, |
|
"loss": 1.5134, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.8645690834473324, |
|
"grad_norm": 0.2494489699602127, |
|
"learning_rate": 4.129595415107864e-05, |
|
"loss": 1.5429, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.8673050615595075, |
|
"grad_norm": 0.2552199363708496, |
|
"learning_rate": 4.124037524441803e-05, |
|
"loss": 1.5348, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.8700410396716827, |
|
"grad_norm": 0.2449236959218979, |
|
"learning_rate": 4.118465711954569e-05, |
|
"loss": 1.7689, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.8727770177838577, |
|
"grad_norm": 0.2517363429069519, |
|
"learning_rate": 4.112880025409701e-05, |
|
"loss": 1.6258, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.8755129958960328, |
|
"grad_norm": 0.27004796266555786, |
|
"learning_rate": 4.107280512689668e-05, |
|
"loss": 1.5855, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.8782489740082079, |
|
"grad_norm": 0.27461642026901245, |
|
"learning_rate": 4.101667221795465e-05, |
|
"loss": 1.57, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.8809849521203831, |
|
"grad_norm": 0.2513730525970459, |
|
"learning_rate": 4.0960402008461955e-05, |
|
"loss": 1.5509, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.8837209302325582, |
|
"grad_norm": 0.2727186977863312, |
|
"learning_rate": 4.090399498078664e-05, |
|
"loss": 1.5905, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.8864569083447332, |
|
"grad_norm": 0.24590528011322021, |
|
"learning_rate": 4.084745161846961e-05, |
|
"loss": 1.5459, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.8891928864569083, |
|
"grad_norm": 0.22626182436943054, |
|
"learning_rate": 4.079077240622043e-05, |
|
"loss": 1.6127, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.8919288645690835, |
|
"grad_norm": 0.24243368208408356, |
|
"learning_rate": 4.0733957829913296e-05, |
|
"loss": 1.5687, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.8946648426812586, |
|
"grad_norm": 0.2647876739501953, |
|
"learning_rate": 4.0677008376582716e-05, |
|
"loss": 1.6761, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.8974008207934336, |
|
"grad_norm": 0.2398100048303604, |
|
"learning_rate": 4.061992453441946e-05, |
|
"loss": 1.5666, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.9001367989056087, |
|
"grad_norm": 0.2463439702987671, |
|
"learning_rate": 4.056270679276631e-05, |
|
"loss": 1.7682, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.9028727770177839, |
|
"grad_norm": 0.23695454001426697, |
|
"learning_rate": 4.050535564211388e-05, |
|
"loss": 1.4589, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.905608755129959, |
|
"grad_norm": 0.24376732110977173, |
|
"learning_rate": 4.044787157409642e-05, |
|
"loss": 1.4797, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.908344733242134, |
|
"grad_norm": 0.24585182964801788, |
|
"learning_rate": 4.039025508148758e-05, |
|
"loss": 1.5843, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.9110807113543091, |
|
"grad_norm": 0.266337126493454, |
|
"learning_rate": 4.0332506658196225e-05, |
|
"loss": 1.7075, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.9138166894664843, |
|
"grad_norm": 0.2664279043674469, |
|
"learning_rate": 4.027462679926215e-05, |
|
"loss": 1.6063, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.9165526675786594, |
|
"grad_norm": 0.2796095013618469, |
|
"learning_rate": 4.021661600085189e-05, |
|
"loss": 1.7072, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.9192886456908345, |
|
"grad_norm": 0.26439887285232544, |
|
"learning_rate": 4.01584747602544e-05, |
|
"loss": 1.5579, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.9220246238030095, |
|
"grad_norm": 0.2440110445022583, |
|
"learning_rate": 4.010020357587687e-05, |
|
"loss": 1.5581, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.9247606019151847, |
|
"grad_norm": 0.2567618787288666, |
|
"learning_rate": 4.00418029472404e-05, |
|
"loss": 1.7241, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.9274965800273598, |
|
"grad_norm": 0.26454323530197144, |
|
"learning_rate": 3.9983273374975726e-05, |
|
"loss": 1.5839, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.9302325581395349, |
|
"grad_norm": 0.2620810866355896, |
|
"learning_rate": 3.9924615360818934e-05, |
|
"loss": 1.7599, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.93296853625171, |
|
"grad_norm": 0.2683962285518646, |
|
"learning_rate": 3.986582940760717e-05, |
|
"loss": 1.6423, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.9357045143638851, |
|
"grad_norm": 0.22277259826660156, |
|
"learning_rate": 3.9806916019274306e-05, |
|
"loss": 1.546, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.9384404924760602, |
|
"grad_norm": 0.26338738203048706, |
|
"learning_rate": 3.9747875700846646e-05, |
|
"loss": 1.5739, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.9411764705882353, |
|
"grad_norm": 0.27578940987586975, |
|
"learning_rate": 3.968870895843858e-05, |
|
"loss": 1.6281, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.9439124487004104, |
|
"grad_norm": 0.24065442383289337, |
|
"learning_rate": 3.962941629924824e-05, |
|
"loss": 1.4456, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.9466484268125855, |
|
"grad_norm": 0.25176844000816345, |
|
"learning_rate": 3.956999823155315e-05, |
|
"loss": 1.6577, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.9493844049247606, |
|
"grad_norm": 0.2668481767177582, |
|
"learning_rate": 3.951045526470592e-05, |
|
"loss": 1.5013, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.9521203830369357, |
|
"grad_norm": 0.27450481057167053, |
|
"learning_rate": 3.94507879091298e-05, |
|
"loss": 1.5217, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.9548563611491108, |
|
"grad_norm": 0.2651226818561554, |
|
"learning_rate": 3.939099667631438e-05, |
|
"loss": 1.5877, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.957592339261286, |
|
"grad_norm": 0.27575191855430603, |
|
"learning_rate": 3.933108207881112e-05, |
|
"loss": 1.5836, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.960328317373461, |
|
"grad_norm": 0.248700350522995, |
|
"learning_rate": 3.927104463022906e-05, |
|
"loss": 1.6258, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.9630642954856361, |
|
"grad_norm": 0.27306076884269714, |
|
"learning_rate": 3.921088484523032e-05, |
|
"loss": 1.5255, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.9658002735978112, |
|
"grad_norm": 0.2685610055923462, |
|
"learning_rate": 3.9150603239525765e-05, |
|
"loss": 1.607, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.9685362517099864, |
|
"grad_norm": 0.2934316396713257, |
|
"learning_rate": 3.909020032987051e-05, |
|
"loss": 1.5252, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.9712722298221614, |
|
"grad_norm": 0.2768978774547577, |
|
"learning_rate": 3.902967663405956e-05, |
|
"loss": 1.6291, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.9740082079343365, |
|
"grad_norm": 0.29954272508621216, |
|
"learning_rate": 3.8969032670923346e-05, |
|
"loss": 1.481, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.9767441860465116, |
|
"grad_norm": 0.26432037353515625, |
|
"learning_rate": 3.890826896032326e-05, |
|
"loss": 1.8236, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.9794801641586868, |
|
"grad_norm": 0.2719402611255646, |
|
"learning_rate": 3.884738602314719e-05, |
|
"loss": 1.626, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.9822161422708618, |
|
"grad_norm": 0.2573629915714264, |
|
"learning_rate": 3.8786384381305123e-05, |
|
"loss": 1.4863, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.9849521203830369, |
|
"grad_norm": 0.27962571382522583, |
|
"learning_rate": 3.87252645577246e-05, |
|
"loss": 1.567, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.987688098495212, |
|
"grad_norm": 0.3235297203063965, |
|
"learning_rate": 3.866402707634624e-05, |
|
"loss": 1.5815, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.9904240766073872, |
|
"grad_norm": 0.24766239523887634, |
|
"learning_rate": 3.8602672462119294e-05, |
|
"loss": 1.6327, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.9931600547195623, |
|
"grad_norm": 0.2470933198928833, |
|
"learning_rate": 3.85412012409971e-05, |
|
"loss": 1.626, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.9958960328317373, |
|
"grad_norm": 0.2527347207069397, |
|
"learning_rate": 3.847961393993261e-05, |
|
"loss": 1.5373, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.9986320109439124, |
|
"grad_norm": 0.24633722007274628, |
|
"learning_rate": 3.8417911086873834e-05, |
|
"loss": 1.5994, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.0013679890560876, |
|
"grad_norm": 0.2634814381599426, |
|
"learning_rate": 3.835609321075934e-05, |
|
"loss": 1.5801, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.0041039671682626, |
|
"grad_norm": 0.2554261386394501, |
|
"learning_rate": 3.8294160841513715e-05, |
|
"loss": 1.6, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.0068399452804377, |
|
"grad_norm": 0.23322933912277222, |
|
"learning_rate": 3.823211451004304e-05, |
|
"loss": 1.4199, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.009575923392613, |
|
"grad_norm": 0.23709669709205627, |
|
"learning_rate": 3.816995474823028e-05, |
|
"loss": 1.5409, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.012311901504788, |
|
"grad_norm": 0.2459351122379303, |
|
"learning_rate": 3.8107682088930794e-05, |
|
"loss": 1.5137, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.015047879616963, |
|
"grad_norm": 0.2492617666721344, |
|
"learning_rate": 3.8045297065967744e-05, |
|
"loss": 1.6054, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.0177838577291383, |
|
"grad_norm": 0.2558859884738922, |
|
"learning_rate": 3.798280021412749e-05, |
|
"loss": 1.6189, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.0205198358413132, |
|
"grad_norm": 0.23855186998844147, |
|
"learning_rate": 3.792019206915504e-05, |
|
"loss": 1.7069, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.0232558139534884, |
|
"grad_norm": 0.2665524482727051, |
|
"learning_rate": 3.7857473167749435e-05, |
|
"loss": 1.5505, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.0259917920656634, |
|
"grad_norm": 0.24079938232898712, |
|
"learning_rate": 3.779464404755919e-05, |
|
"loss": 1.4712, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.0287277701778386, |
|
"grad_norm": 0.2564024329185486, |
|
"learning_rate": 3.773170524717763e-05, |
|
"loss": 1.5464, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.0314637482900137, |
|
"grad_norm": 0.23460035026073456, |
|
"learning_rate": 3.766865730613828e-05, |
|
"loss": 1.6146, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.0341997264021887, |
|
"grad_norm": 0.24386173486709595, |
|
"learning_rate": 3.760550076491031e-05, |
|
"loss": 1.4695, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.036935704514364, |
|
"grad_norm": 0.29760661721229553, |
|
"learning_rate": 3.754223616489379e-05, |
|
"loss": 1.3848, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.039671682626539, |
|
"grad_norm": 0.243236243724823, |
|
"learning_rate": 3.7478864048415136e-05, |
|
"loss": 1.4215, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.042407660738714, |
|
"grad_norm": 0.259069561958313, |
|
"learning_rate": 3.7415384958722445e-05, |
|
"loss": 1.5352, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.0451436388508892, |
|
"grad_norm": 0.281246542930603, |
|
"learning_rate": 3.735179943998081e-05, |
|
"loss": 1.6266, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.0478796169630642, |
|
"grad_norm": 0.2552565634250641, |
|
"learning_rate": 3.728810803726767e-05, |
|
"loss": 1.6849, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.0506155950752394, |
|
"grad_norm": 0.25851351022720337, |
|
"learning_rate": 3.7224311296568134e-05, |
|
"loss": 1.4284, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.0533515731874146, |
|
"grad_norm": 0.2506335973739624, |
|
"learning_rate": 3.716040976477033e-05, |
|
"loss": 1.4498, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.0560875512995895, |
|
"grad_norm": 0.3301680088043213, |
|
"learning_rate": 3.709640398966067e-05, |
|
"loss": 1.5649, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.0588235294117647, |
|
"grad_norm": 0.2851655185222626, |
|
"learning_rate": 3.703229451991918e-05, |
|
"loss": 1.6324, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.06155950752394, |
|
"grad_norm": 0.27497461438179016, |
|
"learning_rate": 3.69680819051148e-05, |
|
"loss": 1.5628, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.0642954856361149, |
|
"grad_norm": 0.2582625448703766, |
|
"learning_rate": 3.6903766695700656e-05, |
|
"loss": 1.5979, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.06703146374829, |
|
"grad_norm": 0.2679811418056488, |
|
"learning_rate": 3.6839349443009364e-05, |
|
"loss": 1.51, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.069767441860465, |
|
"grad_norm": 0.2523844242095947, |
|
"learning_rate": 3.677483069924827e-05, |
|
"loss": 1.6217, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.0725034199726402, |
|
"grad_norm": 0.23916080594062805, |
|
"learning_rate": 3.671021101749476e-05, |
|
"loss": 1.6128, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.0752393980848154, |
|
"grad_norm": 0.2704644203186035, |
|
"learning_rate": 3.664549095169148e-05, |
|
"loss": 1.5524, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.0779753761969904, |
|
"grad_norm": 0.27870067954063416, |
|
"learning_rate": 3.6580671056641616e-05, |
|
"loss": 1.6693, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.0807113543091655, |
|
"grad_norm": 0.2483988106250763, |
|
"learning_rate": 3.6515751888004113e-05, |
|
"loss": 1.4771, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.0834473324213407, |
|
"grad_norm": 0.2584916949272156, |
|
"learning_rate": 3.645073400228895e-05, |
|
"loss": 1.5099, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.0861833105335157, |
|
"grad_norm": 0.24778081476688385, |
|
"learning_rate": 3.6385617956852286e-05, |
|
"loss": 1.6118, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.0889192886456909, |
|
"grad_norm": 0.2260492891073227, |
|
"learning_rate": 3.632040430989181e-05, |
|
"loss": 1.5512, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.0916552667578658, |
|
"grad_norm": 0.2751753330230713, |
|
"learning_rate": 3.6255093620441834e-05, |
|
"loss": 1.559, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.094391244870041, |
|
"grad_norm": 0.2648756206035614, |
|
"learning_rate": 3.618968644836859e-05, |
|
"loss": 1.5666, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.0971272229822162, |
|
"grad_norm": 0.24056974053382874, |
|
"learning_rate": 3.612418335436536e-05, |
|
"loss": 1.5169, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.0998632010943912, |
|
"grad_norm": 0.23621565103530884, |
|
"learning_rate": 3.605858489994771e-05, |
|
"loss": 1.7854, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.1025991792065664, |
|
"grad_norm": 0.25893279910087585, |
|
"learning_rate": 3.5992891647448696e-05, |
|
"loss": 1.6353, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.1053351573187415, |
|
"grad_norm": 0.2806542217731476, |
|
"learning_rate": 3.592710416001398e-05, |
|
"loss": 1.7623, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 1.1080711354309165, |
|
"grad_norm": 0.29038006067276, |
|
"learning_rate": 3.586122300159707e-05, |
|
"loss": 1.4457, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.1108071135430917, |
|
"grad_norm": 0.27859950065612793, |
|
"learning_rate": 3.5795248736954426e-05, |
|
"loss": 1.4407, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 1.1135430916552667, |
|
"grad_norm": 0.2583613991737366, |
|
"learning_rate": 3.5729181931640674e-05, |
|
"loss": 1.7205, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 1.1162790697674418, |
|
"grad_norm": 0.25556913018226624, |
|
"learning_rate": 3.5663023152003705e-05, |
|
"loss": 1.5662, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.119015047879617, |
|
"grad_norm": 0.26569557189941406, |
|
"learning_rate": 3.559677296517987e-05, |
|
"loss": 1.5788, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 1.121751025991792, |
|
"grad_norm": 0.2778777778148651, |
|
"learning_rate": 3.5530431939089084e-05, |
|
"loss": 1.747, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.1244870041039672, |
|
"grad_norm": 0.29031357169151306, |
|
"learning_rate": 3.546400064242997e-05, |
|
"loss": 1.6342, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.1272229822161424, |
|
"grad_norm": 0.26415035128593445, |
|
"learning_rate": 3.5397479644674964e-05, |
|
"loss": 1.5624, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.1299589603283173, |
|
"grad_norm": 0.2662270963191986, |
|
"learning_rate": 3.533086951606549e-05, |
|
"loss": 1.6745, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 1.1326949384404925, |
|
"grad_norm": 0.2567162811756134, |
|
"learning_rate": 3.5264170827607004e-05, |
|
"loss": 1.6727, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.1354309165526675, |
|
"grad_norm": 0.28177884221076965, |
|
"learning_rate": 3.519738415106413e-05, |
|
"loss": 1.4554, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.1381668946648427, |
|
"grad_norm": 0.277425616979599, |
|
"learning_rate": 3.513051005895576e-05, |
|
"loss": 1.5096, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 1.1409028727770179, |
|
"grad_norm": 0.2886553108692169, |
|
"learning_rate": 3.506354912455016e-05, |
|
"loss": 1.6723, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.1436388508891928, |
|
"grad_norm": 0.245221808552742, |
|
"learning_rate": 3.499650192186001e-05, |
|
"loss": 1.4508, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 1.146374829001368, |
|
"grad_norm": 0.31830281019210815, |
|
"learning_rate": 3.492936902563754e-05, |
|
"loss": 1.4235, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 1.1491108071135432, |
|
"grad_norm": 0.2477853149175644, |
|
"learning_rate": 3.486215101136954e-05, |
|
"loss": 1.4991, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.1518467852257182, |
|
"grad_norm": 0.2735631465911865, |
|
"learning_rate": 3.47948484552725e-05, |
|
"loss": 1.5141, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.1545827633378933, |
|
"grad_norm": 0.263138085603714, |
|
"learning_rate": 3.47274619342876e-05, |
|
"loss": 1.7006, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.1573187414500685, |
|
"grad_norm": 0.2716856598854065, |
|
"learning_rate": 3.465999202607583e-05, |
|
"loss": 1.6627, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.1600547195622435, |
|
"grad_norm": 0.26602792739868164, |
|
"learning_rate": 3.459243930901297e-05, |
|
"loss": 1.5501, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.1627906976744187, |
|
"grad_norm": 0.27663204073905945, |
|
"learning_rate": 3.452480436218471e-05, |
|
"loss": 1.7229, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.1655266757865936, |
|
"grad_norm": 0.3033023178577423, |
|
"learning_rate": 3.4457087765381584e-05, |
|
"loss": 1.5789, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.1682626538987688, |
|
"grad_norm": 0.2735162675380707, |
|
"learning_rate": 3.43892900990941e-05, |
|
"loss": 1.6559, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.170998632010944, |
|
"grad_norm": 0.26795685291290283, |
|
"learning_rate": 3.432141194450772e-05, |
|
"loss": 1.4425, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.173734610123119, |
|
"grad_norm": 0.2699303925037384, |
|
"learning_rate": 3.425345388349786e-05, |
|
"loss": 1.5429, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.1764705882352942, |
|
"grad_norm": 0.2614208459854126, |
|
"learning_rate": 3.418541649862494e-05, |
|
"loss": 1.5916, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.1792065663474691, |
|
"grad_norm": 0.30162835121154785, |
|
"learning_rate": 3.4117300373129376e-05, |
|
"loss": 1.4801, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.1819425444596443, |
|
"grad_norm": 0.2973942458629608, |
|
"learning_rate": 3.404910609092655e-05, |
|
"loss": 1.5476, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.1846785225718195, |
|
"grad_norm": 0.26698416471481323, |
|
"learning_rate": 3.3980834236601853e-05, |
|
"loss": 1.6382, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.1874145006839945, |
|
"grad_norm": 0.3085285723209381, |
|
"learning_rate": 3.391248539540565e-05, |
|
"loss": 1.5059, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.1901504787961696, |
|
"grad_norm": 0.25455278158187866, |
|
"learning_rate": 3.384406015324826e-05, |
|
"loss": 1.6612, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.1928864569083448, |
|
"grad_norm": 0.2725888192653656, |
|
"learning_rate": 3.3775559096694933e-05, |
|
"loss": 1.5598, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.1956224350205198, |
|
"grad_norm": 0.34251880645751953, |
|
"learning_rate": 3.370698281296083e-05, |
|
"loss": 1.5973, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.198358413132695, |
|
"grad_norm": 0.2710118889808655, |
|
"learning_rate": 3.363833188990599e-05, |
|
"loss": 1.4116, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.2010943912448702, |
|
"grad_norm": 0.2622835636138916, |
|
"learning_rate": 3.3569606916030294e-05, |
|
"loss": 1.4546, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.2038303693570451, |
|
"grad_norm": 0.2685026228427887, |
|
"learning_rate": 3.350080848046839e-05, |
|
"loss": 1.5789, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.2065663474692203, |
|
"grad_norm": 0.2792280912399292, |
|
"learning_rate": 3.343193717298469e-05, |
|
"loss": 1.5287, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.2093023255813953, |
|
"grad_norm": 0.3190619647502899, |
|
"learning_rate": 3.3362993583968264e-05, |
|
"loss": 1.5803, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.2120383036935705, |
|
"grad_norm": 0.27128350734710693, |
|
"learning_rate": 3.329397830442784e-05, |
|
"loss": 1.5371, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.2147742818057456, |
|
"grad_norm": 0.2853277325630188, |
|
"learning_rate": 3.322489192598665e-05, |
|
"loss": 1.5739, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.2175102599179206, |
|
"grad_norm": 0.2518289387226105, |
|
"learning_rate": 3.3155735040877465e-05, |
|
"loss": 1.6408, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.2202462380300958, |
|
"grad_norm": 0.297750324010849, |
|
"learning_rate": 3.308650824193744e-05, |
|
"loss": 1.6255, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.2229822161422708, |
|
"grad_norm": 0.2988694906234741, |
|
"learning_rate": 3.301721212260306e-05, |
|
"loss": 1.5215, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.225718194254446, |
|
"grad_norm": 0.2906099557876587, |
|
"learning_rate": 3.294784727690503e-05, |
|
"loss": 1.5124, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.2284541723666211, |
|
"grad_norm": 0.28081414103507996, |
|
"learning_rate": 3.2878414299463225e-05, |
|
"loss": 1.6691, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.231190150478796, |
|
"grad_norm": 0.26118987798690796, |
|
"learning_rate": 3.280891378548156e-05, |
|
"loss": 1.524, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.2339261285909713, |
|
"grad_norm": 0.2828025817871094, |
|
"learning_rate": 3.273934633074291e-05, |
|
"loss": 1.7315, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 1.2366621067031465, |
|
"grad_norm": 0.27713146805763245, |
|
"learning_rate": 3.2669712531603966e-05, |
|
"loss": 1.534, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.2393980848153214, |
|
"grad_norm": 0.30562645196914673, |
|
"learning_rate": 3.2600012984990165e-05, |
|
"loss": 1.4404, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.2421340629274966, |
|
"grad_norm": 0.30407342314720154, |
|
"learning_rate": 3.2530248288390555e-05, |
|
"loss": 1.5327, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.2448700410396718, |
|
"grad_norm": 0.3026122450828552, |
|
"learning_rate": 3.246041903985264e-05, |
|
"loss": 1.4441, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.2476060191518468, |
|
"grad_norm": 0.3250129520893097, |
|
"learning_rate": 3.2390525837977334e-05, |
|
"loss": 1.661, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.250341997264022, |
|
"grad_norm": 0.29971036314964294, |
|
"learning_rate": 3.232056928191376e-05, |
|
"loss": 1.5989, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 1.253077975376197, |
|
"grad_norm": 0.296903133392334, |
|
"learning_rate": 3.225054997135413e-05, |
|
"loss": 1.5724, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 1.255813953488372, |
|
"grad_norm": 0.31808045506477356, |
|
"learning_rate": 3.218046850652862e-05, |
|
"loss": 1.6791, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.2585499316005473, |
|
"grad_norm": 0.2797161340713501, |
|
"learning_rate": 3.211032548820019e-05, |
|
"loss": 1.4345, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.2612859097127223, |
|
"grad_norm": 0.3660871386528015, |
|
"learning_rate": 3.20401215176595e-05, |
|
"loss": 1.5669, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 1.2640218878248974, |
|
"grad_norm": 0.30877435207366943, |
|
"learning_rate": 3.196985719671968e-05, |
|
"loss": 1.5406, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.2667578659370724, |
|
"grad_norm": 0.29524242877960205, |
|
"learning_rate": 3.1899533127711215e-05, |
|
"loss": 1.3333, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 1.2694938440492476, |
|
"grad_norm": 0.3187881112098694, |
|
"learning_rate": 3.182914991347677e-05, |
|
"loss": 1.5237, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.2722298221614228, |
|
"grad_norm": 0.2810283303260803, |
|
"learning_rate": 3.1758708157366036e-05, |
|
"loss": 1.4334, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.2749658002735977, |
|
"grad_norm": 0.27640587091445923, |
|
"learning_rate": 3.168820846323053e-05, |
|
"loss": 1.5078, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 1.277701778385773, |
|
"grad_norm": 0.2678714692592621, |
|
"learning_rate": 3.161765143541843e-05, |
|
"loss": 1.439, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 1.2804377564979479, |
|
"grad_norm": 0.315186470746994, |
|
"learning_rate": 3.154703767876942e-05, |
|
"loss": 1.4901, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.283173734610123, |
|
"grad_norm": 0.32562458515167236, |
|
"learning_rate": 3.1476367798609475e-05, |
|
"loss": 1.4822, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 1.2859097127222983, |
|
"grad_norm": 0.31798624992370605, |
|
"learning_rate": 3.1405642400745664e-05, |
|
"loss": 1.5145, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.2886456908344734, |
|
"grad_norm": 0.29922667145729065, |
|
"learning_rate": 3.133486209146099e-05, |
|
"loss": 1.4984, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.2913816689466484, |
|
"grad_norm": 0.3259302079677582, |
|
"learning_rate": 3.12640274775092e-05, |
|
"loss": 1.5048, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.2941176470588236, |
|
"grad_norm": 0.2870180010795593, |
|
"learning_rate": 3.119313916610948e-05, |
|
"loss": 1.52, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 1.2968536251709986, |
|
"grad_norm": 0.3135444223880768, |
|
"learning_rate": 3.112219776494142e-05, |
|
"loss": 1.494, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.2995896032831737, |
|
"grad_norm": 0.29159173369407654, |
|
"learning_rate": 3.105120388213966e-05, |
|
"loss": 1.7019, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.302325581395349, |
|
"grad_norm": 0.31548362970352173, |
|
"learning_rate": 3.098015812628875e-05, |
|
"loss": 1.6465, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.305061559507524, |
|
"grad_norm": 0.2820277214050293, |
|
"learning_rate": 3.090906110641791e-05, |
|
"loss": 1.5705, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.307797537619699, |
|
"grad_norm": 0.2773056924343109, |
|
"learning_rate": 3.083791343199582e-05, |
|
"loss": 1.785, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.310533515731874, |
|
"grad_norm": 0.2650894522666931, |
|
"learning_rate": 3.0766715712925384e-05, |
|
"loss": 1.6026, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.3132694938440492, |
|
"grad_norm": 0.2996635138988495, |
|
"learning_rate": 3.06954685595385e-05, |
|
"loss": 1.489, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.3160054719562244, |
|
"grad_norm": 0.3386262059211731, |
|
"learning_rate": 3.062417258259084e-05, |
|
"loss": 1.538, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.3187414500683994, |
|
"grad_norm": 0.30783751606941223, |
|
"learning_rate": 3.055282839325661e-05, |
|
"loss": 1.42, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.3214774281805746, |
|
"grad_norm": 0.28174683451652527, |
|
"learning_rate": 3.0481436603123292e-05, |
|
"loss": 1.5068, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.3242134062927495, |
|
"grad_norm": 0.2959563732147217, |
|
"learning_rate": 3.0409997824186453e-05, |
|
"loss": 1.4343, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.3269493844049247, |
|
"grad_norm": 0.24625587463378906, |
|
"learning_rate": 3.0338512668844443e-05, |
|
"loss": 1.5942, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.3296853625171, |
|
"grad_norm": 0.3001210689544678, |
|
"learning_rate": 3.0266981749893157e-05, |
|
"loss": 1.5715, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.332421340629275, |
|
"grad_norm": 0.2976323962211609, |
|
"learning_rate": 3.0195405680520828e-05, |
|
"loss": 1.4471, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.33515731874145, |
|
"grad_norm": 0.3303026854991913, |
|
"learning_rate": 3.012378507430269e-05, |
|
"loss": 1.6357, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.3378932968536252, |
|
"grad_norm": 0.308095246553421, |
|
"learning_rate": 3.005212054519579e-05, |
|
"loss": 1.6527, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.3406292749658002, |
|
"grad_norm": 0.2990988790988922, |
|
"learning_rate": 2.99804127075337e-05, |
|
"loss": 1.5795, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.3433652530779754, |
|
"grad_norm": 0.29910552501678467, |
|
"learning_rate": 2.9908662176021225e-05, |
|
"loss": 1.6597, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.3461012311901506, |
|
"grad_norm": 0.28705233335494995, |
|
"learning_rate": 2.9836869565729176e-05, |
|
"loss": 1.4978, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.3488372093023255, |
|
"grad_norm": 0.33388686180114746, |
|
"learning_rate": 2.9765035492089072e-05, |
|
"loss": 1.4049, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.3515731874145007, |
|
"grad_norm": 0.28467613458633423, |
|
"learning_rate": 2.9693160570887873e-05, |
|
"loss": 1.4809, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.3543091655266757, |
|
"grad_norm": 0.2524206340312958, |
|
"learning_rate": 2.9621245418262694e-05, |
|
"loss": 1.5561, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.3570451436388509, |
|
"grad_norm": 0.27135151624679565, |
|
"learning_rate": 2.954929065069554e-05, |
|
"loss": 1.5686, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.359781121751026, |
|
"grad_norm": 0.31753405928611755, |
|
"learning_rate": 2.9477296885007988e-05, |
|
"loss": 1.4994, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.362517099863201, |
|
"grad_norm": 0.274442583322525, |
|
"learning_rate": 2.9405264738355946e-05, |
|
"loss": 1.5835, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.3652530779753762, |
|
"grad_norm": 0.3164355754852295, |
|
"learning_rate": 2.9333194828224316e-05, |
|
"loss": 1.4077, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.3679890560875512, |
|
"grad_norm": 0.3447605073451996, |
|
"learning_rate": 2.926108777242172e-05, |
|
"loss": 1.4768, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.3679890560875512, |
|
"eval_loss": 1.573856234550476, |
|
"eval_runtime": 13.7722, |
|
"eval_samples_per_second": 11.182, |
|
"eval_steps_per_second": 11.182, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.3707250341997264, |
|
"grad_norm": 0.31595364212989807, |
|
"learning_rate": 2.9188944189075235e-05, |
|
"loss": 1.6381, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.3734610123119015, |
|
"grad_norm": 0.28893882036209106, |
|
"learning_rate": 2.9116764696625033e-05, |
|
"loss": 1.6677, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.3761969904240767, |
|
"grad_norm": 0.27313435077667236, |
|
"learning_rate": 2.9044549913819124e-05, |
|
"loss": 1.5746, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.3789329685362517, |
|
"grad_norm": 0.2701267600059509, |
|
"learning_rate": 2.897230045970804e-05, |
|
"loss": 1.2618, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.3816689466484269, |
|
"grad_norm": 0.30766165256500244, |
|
"learning_rate": 2.890001695363953e-05, |
|
"loss": 1.4354, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.3844049247606018, |
|
"grad_norm": 0.34688156843185425, |
|
"learning_rate": 2.8827700015253246e-05, |
|
"loss": 1.4171, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.387140902872777, |
|
"grad_norm": 0.28550222516059875, |
|
"learning_rate": 2.875535026447543e-05, |
|
"loss": 1.6484, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.3898768809849522, |
|
"grad_norm": 0.2944300174713135, |
|
"learning_rate": 2.868296832151361e-05, |
|
"loss": 1.6701, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.3926128590971272, |
|
"grad_norm": 0.2670783996582031, |
|
"learning_rate": 2.8610554806851264e-05, |
|
"loss": 1.5159, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.3953488372093024, |
|
"grad_norm": 0.2730109989643097, |
|
"learning_rate": 2.853811034124253e-05, |
|
"loss": 1.463, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.3980848153214773, |
|
"grad_norm": 0.2924908995628357, |
|
"learning_rate": 2.8465635545706858e-05, |
|
"loss": 1.6673, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.4008207934336525, |
|
"grad_norm": 0.2832358777523041, |
|
"learning_rate": 2.8393131041523702e-05, |
|
"loss": 1.4224, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.4035567715458277, |
|
"grad_norm": 0.3116619884967804, |
|
"learning_rate": 2.8320597450227186e-05, |
|
"loss": 1.4651, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.4062927496580027, |
|
"grad_norm": 0.2656640410423279, |
|
"learning_rate": 2.824803539360078e-05, |
|
"loss": 1.5117, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.4090287277701778, |
|
"grad_norm": 0.321814626455307, |
|
"learning_rate": 2.8175445493671972e-05, |
|
"loss": 1.5859, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.4117647058823528, |
|
"grad_norm": 0.3401627540588379, |
|
"learning_rate": 2.8102828372706926e-05, |
|
"loss": 1.6612, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.414500683994528, |
|
"grad_norm": 0.3001568019390106, |
|
"learning_rate": 2.803018465320515e-05, |
|
"loss": 1.655, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.4172366621067032, |
|
"grad_norm": 0.27258241176605225, |
|
"learning_rate": 2.795751495789418e-05, |
|
"loss": 1.6612, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.4199726402188784, |
|
"grad_norm": 0.2985587418079376, |
|
"learning_rate": 2.7884819909724224e-05, |
|
"loss": 1.452, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.4227086183310533, |
|
"grad_norm": 0.2565891146659851, |
|
"learning_rate": 2.78121001318628e-05, |
|
"loss": 1.6304, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.4254445964432285, |
|
"grad_norm": 0.29893383383750916, |
|
"learning_rate": 2.7739356247689446e-05, |
|
"loss": 1.4655, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.4281805745554035, |
|
"grad_norm": 0.28050941228866577, |
|
"learning_rate": 2.7666588880790335e-05, |
|
"loss": 1.5134, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.4309165526675787, |
|
"grad_norm": 0.29267653822898865, |
|
"learning_rate": 2.759379865495294e-05, |
|
"loss": 1.4355, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.4336525307797539, |
|
"grad_norm": 0.279461145401001, |
|
"learning_rate": 2.752098619416069e-05, |
|
"loss": 1.5606, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.4363885088919288, |
|
"grad_norm": 0.260345995426178, |
|
"learning_rate": 2.7448152122587634e-05, |
|
"loss": 1.702, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.439124487004104, |
|
"grad_norm": 0.2932511270046234, |
|
"learning_rate": 2.7375297064593063e-05, |
|
"loss": 1.3368, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.441860465116279, |
|
"grad_norm": 0.3288111686706543, |
|
"learning_rate": 2.730242164471616e-05, |
|
"loss": 1.4919, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.4445964432284542, |
|
"grad_norm": 0.27260783314704895, |
|
"learning_rate": 2.7229526487670676e-05, |
|
"loss": 1.5306, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.4473324213406293, |
|
"grad_norm": 0.2748279273509979, |
|
"learning_rate": 2.7156612218339544e-05, |
|
"loss": 1.5238, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.4500683994528043, |
|
"grad_norm": 0.26095664501190186, |
|
"learning_rate": 2.708367946176956e-05, |
|
"loss": 1.5456, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.4528043775649795, |
|
"grad_norm": 0.3169284164905548, |
|
"learning_rate": 2.701072884316595e-05, |
|
"loss": 1.6298, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.4555403556771545, |
|
"grad_norm": 0.2977512776851654, |
|
"learning_rate": 2.6937760987887112e-05, |
|
"loss": 1.4394, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.4582763337893296, |
|
"grad_norm": 0.2777084708213806, |
|
"learning_rate": 2.6864776521439166e-05, |
|
"loss": 1.5062, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.4610123119015048, |
|
"grad_norm": 0.29830804467201233, |
|
"learning_rate": 2.6791776069470658e-05, |
|
"loss": 1.5056, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.46374829001368, |
|
"grad_norm": 0.35224923491477966, |
|
"learning_rate": 2.6718760257767135e-05, |
|
"loss": 1.4127, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.466484268125855, |
|
"grad_norm": 0.31385019421577454, |
|
"learning_rate": 2.6645729712245832e-05, |
|
"loss": 1.6469, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.4692202462380302, |
|
"grad_norm": 0.3666173815727234, |
|
"learning_rate": 2.6572685058950298e-05, |
|
"loss": 1.6698, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.4719562243502051, |
|
"grad_norm": 0.29228919744491577, |
|
"learning_rate": 2.649962692404499e-05, |
|
"loss": 1.5971, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.4746922024623803, |
|
"grad_norm": 0.2931617796421051, |
|
"learning_rate": 2.6426555933809954e-05, |
|
"loss": 1.4914, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.4774281805745555, |
|
"grad_norm": 0.2743561863899231, |
|
"learning_rate": 2.635347271463544e-05, |
|
"loss": 1.5475, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.4801641586867305, |
|
"grad_norm": 0.3290240466594696, |
|
"learning_rate": 2.6280377893016516e-05, |
|
"loss": 1.4155, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.4829001367989056, |
|
"grad_norm": 0.30161550641059875, |
|
"learning_rate": 2.6207272095547718e-05, |
|
"loss": 1.4564, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.4856361149110806, |
|
"grad_norm": 0.3175734877586365, |
|
"learning_rate": 2.613415594891767e-05, |
|
"loss": 1.5312, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.4883720930232558, |
|
"grad_norm": 0.3023424744606018, |
|
"learning_rate": 2.606103007990371e-05, |
|
"loss": 1.4688, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.491108071135431, |
|
"grad_norm": 0.3181714415550232, |
|
"learning_rate": 2.5987895115366516e-05, |
|
"loss": 1.738, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.493844049247606, |
|
"grad_norm": 0.2678375244140625, |
|
"learning_rate": 2.5914751682244748e-05, |
|
"loss": 1.5491, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.4965800273597811, |
|
"grad_norm": 0.3041532039642334, |
|
"learning_rate": 2.5841600407549642e-05, |
|
"loss": 1.6654, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.499316005471956, |
|
"grad_norm": 0.30127188563346863, |
|
"learning_rate": 2.5768441918359692e-05, |
|
"loss": 1.5671, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.5020519835841313, |
|
"grad_norm": 0.34145042300224304, |
|
"learning_rate": 2.5695276841815186e-05, |
|
"loss": 1.6131, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.5047879616963065, |
|
"grad_norm": 0.2628389000892639, |
|
"learning_rate": 2.562210580511291e-05, |
|
"loss": 1.7636, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.5075239398084817, |
|
"grad_norm": 0.2784722149372101, |
|
"learning_rate": 2.5548929435500758e-05, |
|
"loss": 1.4967, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.5102599179206566, |
|
"grad_norm": 0.29418709874153137, |
|
"learning_rate": 2.547574836027231e-05, |
|
"loss": 1.5771, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.5129958960328316, |
|
"grad_norm": 0.29850900173187256, |
|
"learning_rate": 2.54025632067615e-05, |
|
"loss": 1.4134, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.5157318741450068, |
|
"grad_norm": 0.29349079728126526, |
|
"learning_rate": 2.5329374602337215e-05, |
|
"loss": 1.4498, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.518467852257182, |
|
"grad_norm": 0.27603679895401, |
|
"learning_rate": 2.525618317439793e-05, |
|
"loss": 1.5103, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.5212038303693571, |
|
"grad_norm": 0.29642152786254883, |
|
"learning_rate": 2.518298955036632e-05, |
|
"loss": 1.5597, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.523939808481532, |
|
"grad_norm": 0.3032752573490143, |
|
"learning_rate": 2.5109794357683885e-05, |
|
"loss": 1.5697, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.5266757865937073, |
|
"grad_norm": 0.3295765519142151, |
|
"learning_rate": 2.503659822380558e-05, |
|
"loss": 1.6343, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.5294117647058822, |
|
"grad_norm": 0.28324607014656067, |
|
"learning_rate": 2.496340177619442e-05, |
|
"loss": 1.551, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.5321477428180574, |
|
"grad_norm": 0.2895617187023163, |
|
"learning_rate": 2.489020564231612e-05, |
|
"loss": 1.6563, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.5348837209302326, |
|
"grad_norm": 0.3268161714076996, |
|
"learning_rate": 2.4817010449633688e-05, |
|
"loss": 1.6045, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.5376196990424078, |
|
"grad_norm": 0.27146685123443604, |
|
"learning_rate": 2.474381682560208e-05, |
|
"loss": 1.5727, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.5403556771545828, |
|
"grad_norm": 0.27629193663597107, |
|
"learning_rate": 2.4670625397662787e-05, |
|
"loss": 1.4352, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.5430916552667577, |
|
"grad_norm": 0.27017953991889954, |
|
"learning_rate": 2.4597436793238506e-05, |
|
"loss": 1.646, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.545827633378933, |
|
"grad_norm": 0.289787232875824, |
|
"learning_rate": 2.45242516397277e-05, |
|
"loss": 1.7107, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.548563611491108, |
|
"grad_norm": 0.2762577533721924, |
|
"learning_rate": 2.4451070564499245e-05, |
|
"loss": 1.5694, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.5512995896032833, |
|
"grad_norm": 0.2959858179092407, |
|
"learning_rate": 2.4377894194887095e-05, |
|
"loss": 1.639, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.5540355677154583, |
|
"grad_norm": 0.28595441579818726, |
|
"learning_rate": 2.4304723158184827e-05, |
|
"loss": 1.3091, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.5567715458276332, |
|
"grad_norm": 0.3006284236907959, |
|
"learning_rate": 2.4231558081640314e-05, |
|
"loss": 1.569, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.5595075239398084, |
|
"grad_norm": 0.2800249457359314, |
|
"learning_rate": 2.415839959245036e-05, |
|
"loss": 1.5124, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.5622435020519836, |
|
"grad_norm": 0.3035907745361328, |
|
"learning_rate": 2.4085248317755254e-05, |
|
"loss": 1.5785, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.5649794801641588, |
|
"grad_norm": 0.3111293315887451, |
|
"learning_rate": 2.4012104884633486e-05, |
|
"loss": 1.5544, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.5677154582763337, |
|
"grad_norm": 0.31877052783966064, |
|
"learning_rate": 2.39389699200963e-05, |
|
"loss": 1.482, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.570451436388509, |
|
"grad_norm": 0.3056713342666626, |
|
"learning_rate": 2.386584405108233e-05, |
|
"loss": 1.5234, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.573187414500684, |
|
"grad_norm": 0.289650022983551, |
|
"learning_rate": 2.3792727904452285e-05, |
|
"loss": 1.398, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.575923392612859, |
|
"grad_norm": 0.28617268800735474, |
|
"learning_rate": 2.3719622106983486e-05, |
|
"loss": 1.5721, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.5786593707250343, |
|
"grad_norm": 0.29906001687049866, |
|
"learning_rate": 2.3646527285364565e-05, |
|
"loss": 1.3543, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.5813953488372094, |
|
"grad_norm": 0.2876405417919159, |
|
"learning_rate": 2.3573444066190052e-05, |
|
"loss": 1.5365, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.5841313269493844, |
|
"grad_norm": 0.3173394799232483, |
|
"learning_rate": 2.3500373075955022e-05, |
|
"loss": 1.4437, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.5868673050615594, |
|
"grad_norm": 0.2820620834827423, |
|
"learning_rate": 2.342731494104971e-05, |
|
"loss": 1.5992, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.5896032831737346, |
|
"grad_norm": 0.3694467842578888, |
|
"learning_rate": 2.3354270287754174e-05, |
|
"loss": 1.5443, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.5923392612859097, |
|
"grad_norm": 0.2960888147354126, |
|
"learning_rate": 2.328123974223288e-05, |
|
"loss": 1.5173, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.595075239398085, |
|
"grad_norm": 0.3080427348613739, |
|
"learning_rate": 2.3208223930529347e-05, |
|
"loss": 1.749, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.59781121751026, |
|
"grad_norm": 0.29521963000297546, |
|
"learning_rate": 2.313522347856084e-05, |
|
"loss": 1.522, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.6005471956224349, |
|
"grad_norm": 0.32432520389556885, |
|
"learning_rate": 2.306223901211289e-05, |
|
"loss": 1.6394, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.60328317373461, |
|
"grad_norm": 0.326926052570343, |
|
"learning_rate": 2.2989271156834057e-05, |
|
"loss": 1.5092, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.6060191518467852, |
|
"grad_norm": 0.2793523967266083, |
|
"learning_rate": 2.291632053823045e-05, |
|
"loss": 1.5266, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.6087551299589604, |
|
"grad_norm": 0.3044006824493408, |
|
"learning_rate": 2.2843387781660452e-05, |
|
"loss": 1.7137, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.6114911080711354, |
|
"grad_norm": 0.33133581280708313, |
|
"learning_rate": 2.2770473512329333e-05, |
|
"loss": 1.5494, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.6142270861833106, |
|
"grad_norm": 0.30066078901290894, |
|
"learning_rate": 2.269757835528385e-05, |
|
"loss": 1.387, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.6169630642954855, |
|
"grad_norm": 0.29674389958381653, |
|
"learning_rate": 2.2624702935406943e-05, |
|
"loss": 1.6985, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.6196990424076607, |
|
"grad_norm": 0.2950797975063324, |
|
"learning_rate": 2.255184787741237e-05, |
|
"loss": 1.558, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.622435020519836, |
|
"grad_norm": 0.32414302229881287, |
|
"learning_rate": 2.2479013805839318e-05, |
|
"loss": 1.4547, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.625170998632011, |
|
"grad_norm": 0.2893849015235901, |
|
"learning_rate": 2.240620134504707e-05, |
|
"loss": 1.6205, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.627906976744186, |
|
"grad_norm": 0.30795803666114807, |
|
"learning_rate": 2.233341111920967e-05, |
|
"loss": 1.4725, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.630642954856361, |
|
"grad_norm": 0.28829425573349, |
|
"learning_rate": 2.226064375231056e-05, |
|
"loss": 1.5676, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.6333789329685362, |
|
"grad_norm": 0.32631853222846985, |
|
"learning_rate": 2.2187899868137206e-05, |
|
"loss": 1.447, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.6361149110807114, |
|
"grad_norm": 0.2999882996082306, |
|
"learning_rate": 2.211518009027579e-05, |
|
"loss": 1.4184, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.6388508891928866, |
|
"grad_norm": 0.302869975566864, |
|
"learning_rate": 2.204248504210582e-05, |
|
"loss": 1.5356, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.6415868673050615, |
|
"grad_norm": 0.32066816091537476, |
|
"learning_rate": 2.1969815346794857e-05, |
|
"loss": 1.507, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.6443228454172365, |
|
"grad_norm": 0.26394781470298767, |
|
"learning_rate": 2.189717162729309e-05, |
|
"loss": 1.519, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.6470588235294117, |
|
"grad_norm": 0.28778427839279175, |
|
"learning_rate": 2.182455450632803e-05, |
|
"loss": 1.6243, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.6497948016415869, |
|
"grad_norm": 0.28757721185684204, |
|
"learning_rate": 2.1751964606399224e-05, |
|
"loss": 1.6119, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.652530779753762, |
|
"grad_norm": 0.32660529017448425, |
|
"learning_rate": 2.167940254977282e-05, |
|
"loss": 1.481, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.655266757865937, |
|
"grad_norm": 0.2857491075992584, |
|
"learning_rate": 2.1606868958476304e-05, |
|
"loss": 1.4287, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.6580027359781122, |
|
"grad_norm": 0.3346758782863617, |
|
"learning_rate": 2.1534364454293148e-05, |
|
"loss": 1.6915, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 1.6607387140902872, |
|
"grad_norm": 0.2858891189098358, |
|
"learning_rate": 2.146188965875747e-05, |
|
"loss": 1.4857, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 1.6634746922024624, |
|
"grad_norm": 0.2926919758319855, |
|
"learning_rate": 2.1389445193148742e-05, |
|
"loss": 1.5348, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.6662106703146375, |
|
"grad_norm": 0.3735974133014679, |
|
"learning_rate": 2.1317031678486402e-05, |
|
"loss": 1.697, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 1.6689466484268127, |
|
"grad_norm": 0.3130210340023041, |
|
"learning_rate": 2.124464973552457e-05, |
|
"loss": 1.6426, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.6716826265389877, |
|
"grad_norm": 0.36122313141822815, |
|
"learning_rate": 2.117229998474676e-05, |
|
"loss": 1.7516, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 1.6744186046511627, |
|
"grad_norm": 0.2955617904663086, |
|
"learning_rate": 2.109998304636048e-05, |
|
"loss": 1.4681, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.6771545827633378, |
|
"grad_norm": 0.3331190347671509, |
|
"learning_rate": 2.1027699540291965e-05, |
|
"loss": 1.5334, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 1.679890560875513, |
|
"grad_norm": 0.39551764726638794, |
|
"learning_rate": 2.0955450086180882e-05, |
|
"loss": 1.497, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 1.6826265389876882, |
|
"grad_norm": 0.2910989224910736, |
|
"learning_rate": 2.088323530337498e-05, |
|
"loss": 1.4593, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.6853625170998632, |
|
"grad_norm": 0.2927062511444092, |
|
"learning_rate": 2.0811055810924768e-05, |
|
"loss": 1.5019, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.6880984952120381, |
|
"grad_norm": 0.2986646592617035, |
|
"learning_rate": 2.0738912227578283e-05, |
|
"loss": 1.5167, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 1.6908344733242133, |
|
"grad_norm": 0.2710409462451935, |
|
"learning_rate": 2.0666805171775687e-05, |
|
"loss": 1.5491, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 1.6935704514363885, |
|
"grad_norm": 0.32555902004241943, |
|
"learning_rate": 2.0594735261644056e-05, |
|
"loss": 1.5945, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 1.6963064295485637, |
|
"grad_norm": 0.3129582703113556, |
|
"learning_rate": 2.0522703114992018e-05, |
|
"loss": 1.5388, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.6990424076607387, |
|
"grad_norm": 0.29060521721839905, |
|
"learning_rate": 2.0450709349304463e-05, |
|
"loss": 1.5335, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 1.7017783857729138, |
|
"grad_norm": 0.2991231381893158, |
|
"learning_rate": 2.0378754581737308e-05, |
|
"loss": 1.5857, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 1.7045143638850888, |
|
"grad_norm": 0.3089081346988678, |
|
"learning_rate": 2.0306839429112136e-05, |
|
"loss": 1.509, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 1.707250341997264, |
|
"grad_norm": 0.30709123611450195, |
|
"learning_rate": 2.023496450791093e-05, |
|
"loss": 1.5935, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.7099863201094392, |
|
"grad_norm": 0.29340359568595886, |
|
"learning_rate": 2.0163130434270833e-05, |
|
"loss": 1.547, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.7127222982216144, |
|
"grad_norm": 0.3518681824207306, |
|
"learning_rate": 2.009133782397879e-05, |
|
"loss": 1.513, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 1.7154582763337893, |
|
"grad_norm": 0.31759360432624817, |
|
"learning_rate": 2.0019587292466306e-05, |
|
"loss": 1.5252, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 1.7181942544459643, |
|
"grad_norm": 0.2931315302848816, |
|
"learning_rate": 1.9947879454804216e-05, |
|
"loss": 1.7602, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.7209302325581395, |
|
"grad_norm": 0.2938210368156433, |
|
"learning_rate": 1.9876214925697323e-05, |
|
"loss": 1.5154, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 1.7236662106703147, |
|
"grad_norm": 0.28195831179618835, |
|
"learning_rate": 1.980459431947918e-05, |
|
"loss": 1.4129, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.7264021887824899, |
|
"grad_norm": 0.31024137139320374, |
|
"learning_rate": 1.973301825010685e-05, |
|
"loss": 1.5008, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 1.7291381668946648, |
|
"grad_norm": 0.3060285747051239, |
|
"learning_rate": 1.9661487331155563e-05, |
|
"loss": 1.5469, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 1.7318741450068398, |
|
"grad_norm": 0.38896477222442627, |
|
"learning_rate": 1.9590002175813553e-05, |
|
"loss": 1.6018, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 1.734610123119015, |
|
"grad_norm": 0.32051512598991394, |
|
"learning_rate": 1.9518563396876717e-05, |
|
"loss": 1.5078, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 1.7373461012311902, |
|
"grad_norm": 0.2726178467273712, |
|
"learning_rate": 1.9447171606743398e-05, |
|
"loss": 1.4586, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.7400820793433653, |
|
"grad_norm": 0.2772533595561981, |
|
"learning_rate": 1.9375827417409165e-05, |
|
"loss": 1.5022, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.7428180574555403, |
|
"grad_norm": 0.30202561616897583, |
|
"learning_rate": 1.9304531440461506e-05, |
|
"loss": 1.5036, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 1.7455540355677155, |
|
"grad_norm": 0.2932077944278717, |
|
"learning_rate": 1.923328428707461e-05, |
|
"loss": 1.5541, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 1.7482900136798905, |
|
"grad_norm": 0.3096188008785248, |
|
"learning_rate": 1.916208656800418e-05, |
|
"loss": 1.3744, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 1.7510259917920656, |
|
"grad_norm": 0.282226026058197, |
|
"learning_rate": 1.9090938893582088e-05, |
|
"loss": 1.6306, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.7537619699042408, |
|
"grad_norm": 0.27213621139526367, |
|
"learning_rate": 1.9019841873711255e-05, |
|
"loss": 1.4953, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 1.756497948016416, |
|
"grad_norm": 0.30166521668434143, |
|
"learning_rate": 1.8948796117860348e-05, |
|
"loss": 1.5158, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 1.759233926128591, |
|
"grad_norm": 0.27089768648147583, |
|
"learning_rate": 1.8877802235058585e-05, |
|
"loss": 1.4954, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 1.761969904240766, |
|
"grad_norm": 0.31477153301239014, |
|
"learning_rate": 1.8806860833890528e-05, |
|
"loss": 1.5449, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 1.7647058823529411, |
|
"grad_norm": 0.2903262674808502, |
|
"learning_rate": 1.8735972522490818e-05, |
|
"loss": 1.4129, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.7674418604651163, |
|
"grad_norm": 0.3140583038330078, |
|
"learning_rate": 1.8665137908539004e-05, |
|
"loss": 1.6374, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 1.7701778385772915, |
|
"grad_norm": 0.29897943139076233, |
|
"learning_rate": 1.859435759925434e-05, |
|
"loss": 1.605, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 1.7729138166894665, |
|
"grad_norm": 0.2993355989456177, |
|
"learning_rate": 1.8523632201390537e-05, |
|
"loss": 1.495, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.7756497948016414, |
|
"grad_norm": 0.3301627039909363, |
|
"learning_rate": 1.845296232123058e-05, |
|
"loss": 1.6064, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.7783857729138166, |
|
"grad_norm": 0.30024030804634094, |
|
"learning_rate": 1.8382348564581574e-05, |
|
"loss": 1.3258, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.7811217510259918, |
|
"grad_norm": 0.3279036283493042, |
|
"learning_rate": 1.8311791536769483e-05, |
|
"loss": 1.4348, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 1.783857729138167, |
|
"grad_norm": 0.3231453001499176, |
|
"learning_rate": 1.8241291842633966e-05, |
|
"loss": 1.4327, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.786593707250342, |
|
"grad_norm": 0.3033272624015808, |
|
"learning_rate": 1.817085008652324e-05, |
|
"loss": 1.4668, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 1.7893296853625171, |
|
"grad_norm": 0.28523799777030945, |
|
"learning_rate": 1.810046687228879e-05, |
|
"loss": 1.5058, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 1.792065663474692, |
|
"grad_norm": 0.35666465759277344, |
|
"learning_rate": 1.803014280328033e-05, |
|
"loss": 1.4551, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.7948016415868673, |
|
"grad_norm": 0.33736082911491394, |
|
"learning_rate": 1.795987848234051e-05, |
|
"loss": 1.2991, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.7975376196990425, |
|
"grad_norm": 0.3194032311439514, |
|
"learning_rate": 1.7889674511799812e-05, |
|
"loss": 1.4764, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 1.8002735978112177, |
|
"grad_norm": 0.28460192680358887, |
|
"learning_rate": 1.7819531493471392e-05, |
|
"loss": 1.4962, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 1.8030095759233926, |
|
"grad_norm": 0.3380921483039856, |
|
"learning_rate": 1.7749450028645875e-05, |
|
"loss": 1.4772, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 1.8057455540355676, |
|
"grad_norm": 0.32514598965644836, |
|
"learning_rate": 1.7679430718086243e-05, |
|
"loss": 1.6421, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.8084815321477428, |
|
"grad_norm": 0.293314129114151, |
|
"learning_rate": 1.7609474162022665e-05, |
|
"loss": 1.5151, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 1.811217510259918, |
|
"grad_norm": 0.3001095950603485, |
|
"learning_rate": 1.753958096014737e-05, |
|
"loss": 1.4422, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 1.8139534883720931, |
|
"grad_norm": 0.3403547704219818, |
|
"learning_rate": 1.7469751711609454e-05, |
|
"loss": 1.6122, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 1.816689466484268, |
|
"grad_norm": 0.3033615052700043, |
|
"learning_rate": 1.739998701500984e-05, |
|
"loss": 1.596, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.819425444596443, |
|
"grad_norm": 0.34662926197052, |
|
"learning_rate": 1.7330287468396033e-05, |
|
"loss": 1.5458, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.8221614227086183, |
|
"grad_norm": 0.32165685296058655, |
|
"learning_rate": 1.7260653669257093e-05, |
|
"loss": 1.5433, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 1.8248974008207934, |
|
"grad_norm": 0.32157987356185913, |
|
"learning_rate": 1.7191086214518447e-05, |
|
"loss": 1.6236, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 1.8276333789329686, |
|
"grad_norm": 0.3016352355480194, |
|
"learning_rate": 1.712158570053678e-05, |
|
"loss": 1.3893, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.8303693570451436, |
|
"grad_norm": 0.40459707379341125, |
|
"learning_rate": 1.7052152723094976e-05, |
|
"loss": 1.3767, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 1.8331053351573188, |
|
"grad_norm": 0.34724652767181396, |
|
"learning_rate": 1.698278787739695e-05, |
|
"loss": 1.6067, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.8358413132694937, |
|
"grad_norm": 0.3703734874725342, |
|
"learning_rate": 1.6913491758062557e-05, |
|
"loss": 1.4538, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 1.838577291381669, |
|
"grad_norm": 0.31886354088783264, |
|
"learning_rate": 1.6844264959122534e-05, |
|
"loss": 1.6776, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.841313269493844, |
|
"grad_norm": 0.32411274313926697, |
|
"learning_rate": 1.6775108074013356e-05, |
|
"loss": 1.5405, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 1.8440492476060193, |
|
"grad_norm": 0.32398101687431335, |
|
"learning_rate": 1.670602169557217e-05, |
|
"loss": 1.5429, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 1.8467852257181943, |
|
"grad_norm": 0.3220042288303375, |
|
"learning_rate": 1.663700641603174e-05, |
|
"loss": 1.5339, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.8495212038303692, |
|
"grad_norm": 0.2694465219974518, |
|
"learning_rate": 1.6568062827015317e-05, |
|
"loss": 1.5641, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 1.8522571819425444, |
|
"grad_norm": 0.3089315593242645, |
|
"learning_rate": 1.6499191519531614e-05, |
|
"loss": 1.3319, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 1.8549931600547196, |
|
"grad_norm": 0.3302357792854309, |
|
"learning_rate": 1.643039308396971e-05, |
|
"loss": 1.478, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 1.8577291381668948, |
|
"grad_norm": 0.30908071994781494, |
|
"learning_rate": 1.6361668110094007e-05, |
|
"loss": 1.5557, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 1.8604651162790697, |
|
"grad_norm": 0.2804611027240753, |
|
"learning_rate": 1.6293017187039174e-05, |
|
"loss": 1.6756, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.8632010943912447, |
|
"grad_norm": 0.3693443238735199, |
|
"learning_rate": 1.6224440903305076e-05, |
|
"loss": 1.5825, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 1.86593707250342, |
|
"grad_norm": 0.33380305767059326, |
|
"learning_rate": 1.615593984675174e-05, |
|
"loss": 1.5312, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 1.868673050615595, |
|
"grad_norm": 0.3250941038131714, |
|
"learning_rate": 1.6087514604594353e-05, |
|
"loss": 1.5642, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 1.8714090287277703, |
|
"grad_norm": 0.3749421238899231, |
|
"learning_rate": 1.6019165763398152e-05, |
|
"loss": 1.6141, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 1.8741450068399452, |
|
"grad_norm": 0.32209932804107666, |
|
"learning_rate": 1.5950893909073453e-05, |
|
"loss": 1.4791, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.8768809849521204, |
|
"grad_norm": 0.34071722626686096, |
|
"learning_rate": 1.5882699626870633e-05, |
|
"loss": 1.6126, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 1.8796169630642954, |
|
"grad_norm": 0.2984413802623749, |
|
"learning_rate": 1.5814583501375064e-05, |
|
"loss": 1.5191, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 1.8823529411764706, |
|
"grad_norm": 0.33248278498649597, |
|
"learning_rate": 1.574654611650214e-05, |
|
"loss": 1.6346, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 1.8850889192886457, |
|
"grad_norm": 0.308098703622818, |
|
"learning_rate": 1.567858805549229e-05, |
|
"loss": 1.5521, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 1.887824897400821, |
|
"grad_norm": 0.2899650037288666, |
|
"learning_rate": 1.56107099009059e-05, |
|
"loss": 1.3675, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.890560875512996, |
|
"grad_norm": 0.33524468541145325, |
|
"learning_rate": 1.5542912234618422e-05, |
|
"loss": 1.6501, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 1.8932968536251709, |
|
"grad_norm": 0.3770296275615692, |
|
"learning_rate": 1.54751956378153e-05, |
|
"loss": 1.5518, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 1.896032831737346, |
|
"grad_norm": 0.3348696529865265, |
|
"learning_rate": 1.540756069098702e-05, |
|
"loss": 1.4388, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 1.8987688098495212, |
|
"grad_norm": 0.37193214893341064, |
|
"learning_rate": 1.5340007973924176e-05, |
|
"loss": 1.499, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 1.9015047879616964, |
|
"grad_norm": 0.33301231265068054, |
|
"learning_rate": 1.5272538065712403e-05, |
|
"loss": 1.6687, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.9042407660738714, |
|
"grad_norm": 0.33597227931022644, |
|
"learning_rate": 1.5205151544727509e-05, |
|
"loss": 1.589, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 1.9069767441860463, |
|
"grad_norm": 0.3002086281776428, |
|
"learning_rate": 1.5137848988630465e-05, |
|
"loss": 1.7462, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 1.9097127222982215, |
|
"grad_norm": 0.30558449029922485, |
|
"learning_rate": 1.5070630974362473e-05, |
|
"loss": 1.6179, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 1.9124487004103967, |
|
"grad_norm": 0.314813494682312, |
|
"learning_rate": 1.5003498078139988e-05, |
|
"loss": 1.6029, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 1.915184678522572, |
|
"grad_norm": 0.31106236577033997, |
|
"learning_rate": 1.4936450875449845e-05, |
|
"loss": 1.5462, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.9179206566347469, |
|
"grad_norm": 0.32419320940971375, |
|
"learning_rate": 1.4869489941044235e-05, |
|
"loss": 1.632, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 1.920656634746922, |
|
"grad_norm": 0.28728801012039185, |
|
"learning_rate": 1.4802615848935875e-05, |
|
"loss": 1.5625, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 1.923392612859097, |
|
"grad_norm": 0.35819682478904724, |
|
"learning_rate": 1.4735829172393007e-05, |
|
"loss": 1.416, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 1.9261285909712722, |
|
"grad_norm": 0.3097285032272339, |
|
"learning_rate": 1.4669130483934512e-05, |
|
"loss": 1.4333, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 1.9288645690834474, |
|
"grad_norm": 0.31058430671691895, |
|
"learning_rate": 1.4602520355325039e-05, |
|
"loss": 1.5175, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.9316005471956226, |
|
"grad_norm": 0.3093342185020447, |
|
"learning_rate": 1.4535999357570046e-05, |
|
"loss": 1.4105, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 1.9343365253077975, |
|
"grad_norm": 0.3064606189727783, |
|
"learning_rate": 1.4469568060910915e-05, |
|
"loss": 1.6593, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 1.9370725034199725, |
|
"grad_norm": 0.3192756772041321, |
|
"learning_rate": 1.4403227034820139e-05, |
|
"loss": 1.5919, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 1.9398084815321477, |
|
"grad_norm": 0.30623388290405273, |
|
"learning_rate": 1.4336976847996303e-05, |
|
"loss": 1.5053, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 1.9425444596443229, |
|
"grad_norm": 0.3014221787452698, |
|
"learning_rate": 1.4270818068359336e-05, |
|
"loss": 1.4876, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.945280437756498, |
|
"grad_norm": 0.34191763401031494, |
|
"learning_rate": 1.420475126304558e-05, |
|
"loss": 1.435, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 1.948016415868673, |
|
"grad_norm": 0.33173930644989014, |
|
"learning_rate": 1.4138776998402927e-05, |
|
"loss": 1.5467, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 1.950752393980848, |
|
"grad_norm": 0.3360328674316406, |
|
"learning_rate": 1.4072895839986023e-05, |
|
"loss": 1.4525, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 1.9534883720930232, |
|
"grad_norm": 0.2906378209590912, |
|
"learning_rate": 1.4007108352551313e-05, |
|
"loss": 1.5788, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 1.9562243502051984, |
|
"grad_norm": 0.3077907860279083, |
|
"learning_rate": 1.3941415100052293e-05, |
|
"loss": 1.4087, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.9589603283173735, |
|
"grad_norm": 0.2951776385307312, |
|
"learning_rate": 1.387581664563465e-05, |
|
"loss": 1.4699, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 1.9616963064295485, |
|
"grad_norm": 0.29453444480895996, |
|
"learning_rate": 1.3810313551631426e-05, |
|
"loss": 1.6203, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 1.9644322845417237, |
|
"grad_norm": 0.3803625702857971, |
|
"learning_rate": 1.3744906379558165e-05, |
|
"loss": 1.6514, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 1.9671682626538987, |
|
"grad_norm": 0.3245275318622589, |
|
"learning_rate": 1.3679595690108193e-05, |
|
"loss": 1.5904, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 1.9699042407660738, |
|
"grad_norm": 0.3239131271839142, |
|
"learning_rate": 1.3614382043147725e-05, |
|
"loss": 1.5117, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.972640218878249, |
|
"grad_norm": 0.3001211881637573, |
|
"learning_rate": 1.3549265997711057e-05, |
|
"loss": 1.5985, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 1.9753761969904242, |
|
"grad_norm": 0.29681622982025146, |
|
"learning_rate": 1.3484248111995892e-05, |
|
"loss": 1.3762, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 1.9781121751025992, |
|
"grad_norm": 0.38813284039497375, |
|
"learning_rate": 1.3419328943358392e-05, |
|
"loss": 1.548, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 1.9808481532147741, |
|
"grad_norm": 0.29337620735168457, |
|
"learning_rate": 1.3354509048308527e-05, |
|
"loss": 1.4466, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 1.9835841313269493, |
|
"grad_norm": 0.3134080171585083, |
|
"learning_rate": 1.328978898250525e-05, |
|
"loss": 1.523, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.9863201094391245, |
|
"grad_norm": 0.32457372546195984, |
|
"learning_rate": 1.3225169300751738e-05, |
|
"loss": 1.4018, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 1.9890560875512997, |
|
"grad_norm": 0.30359819531440735, |
|
"learning_rate": 1.3160650556990644e-05, |
|
"loss": 1.4209, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 1.9917920656634747, |
|
"grad_norm": 0.3376854360103607, |
|
"learning_rate": 1.3096233304299346e-05, |
|
"loss": 1.4914, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 1.9945280437756496, |
|
"grad_norm": 0.3038989305496216, |
|
"learning_rate": 1.30319180948852e-05, |
|
"loss": 1.4995, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 1.9972640218878248, |
|
"grad_norm": 0.2987303137779236, |
|
"learning_rate": 1.2967705480080819e-05, |
|
"loss": 1.5415, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.36076483130455017, |
|
"learning_rate": 1.2903596010339338e-05, |
|
"loss": 1.3103, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 2.002735978112175, |
|
"grad_norm": 0.329522043466568, |
|
"learning_rate": 1.2839590235229668e-05, |
|
"loss": 1.7937, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 2.0054719562243504, |
|
"grad_norm": 0.33667078614234924, |
|
"learning_rate": 1.2775688703431871e-05, |
|
"loss": 1.333, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 2.008207934336525, |
|
"grad_norm": 0.30493631958961487, |
|
"learning_rate": 1.2711891962732342e-05, |
|
"loss": 1.3868, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 2.0109439124487003, |
|
"grad_norm": 0.3139844536781311, |
|
"learning_rate": 1.26482005600192e-05, |
|
"loss": 1.4361, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 2.0136798905608755, |
|
"grad_norm": 0.30529195070266724, |
|
"learning_rate": 1.258461504127756e-05, |
|
"loss": 1.5679, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 2.0164158686730507, |
|
"grad_norm": 0.3701570928096771, |
|
"learning_rate": 1.252113595158487e-05, |
|
"loss": 1.7087, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 2.019151846785226, |
|
"grad_norm": 0.3016231060028076, |
|
"learning_rate": 1.245776383510622e-05, |
|
"loss": 1.5592, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 2.0218878248974006, |
|
"grad_norm": 0.29768821597099304, |
|
"learning_rate": 1.2394499235089699e-05, |
|
"loss": 1.4926, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 2.024623803009576, |
|
"grad_norm": 0.3342791497707367, |
|
"learning_rate": 1.2331342693861716e-05, |
|
"loss": 1.4944, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.027359781121751, |
|
"grad_norm": 0.30851346254348755, |
|
"learning_rate": 1.2268294752822376e-05, |
|
"loss": 1.7049, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 2.030095759233926, |
|
"grad_norm": 0.3023451864719391, |
|
"learning_rate": 1.2205355952440817e-05, |
|
"loss": 1.5705, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 2.0328317373461013, |
|
"grad_norm": 0.3253107964992523, |
|
"learning_rate": 1.2142526832250561e-05, |
|
"loss": 1.4404, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 2.0355677154582765, |
|
"grad_norm": 0.33139538764953613, |
|
"learning_rate": 1.2079807930844977e-05, |
|
"loss": 1.4355, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 2.0383036935704513, |
|
"grad_norm": 0.3191182613372803, |
|
"learning_rate": 1.2017199785872523e-05, |
|
"loss": 1.428, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 2.0410396716826265, |
|
"grad_norm": 0.33677810430526733, |
|
"learning_rate": 1.1954702934032267e-05, |
|
"loss": 1.4214, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 2.0437756497948016, |
|
"grad_norm": 0.31797662377357483, |
|
"learning_rate": 1.1892317911069212e-05, |
|
"loss": 1.5024, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 2.046511627906977, |
|
"grad_norm": 0.30894604325294495, |
|
"learning_rate": 1.183004525176973e-05, |
|
"loss": 1.5084, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 2.049247606019152, |
|
"grad_norm": 0.30327004194259644, |
|
"learning_rate": 1.176788548995697e-05, |
|
"loss": 1.5586, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 2.0519835841313268, |
|
"grad_norm": 0.31576216220855713, |
|
"learning_rate": 1.1705839158486284e-05, |
|
"loss": 1.5315, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.0519835841313268, |
|
"eval_loss": 1.5589618682861328, |
|
"eval_runtime": 12.9752, |
|
"eval_samples_per_second": 11.869, |
|
"eval_steps_per_second": 11.869, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.054719562243502, |
|
"grad_norm": 0.3079953193664551, |
|
"learning_rate": 1.1643906789240664e-05, |
|
"loss": 1.6127, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 2.057455540355677, |
|
"grad_norm": 0.30269429087638855, |
|
"learning_rate": 1.1582088913126172e-05, |
|
"loss": 1.5077, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 2.0601915184678523, |
|
"grad_norm": 0.34395742416381836, |
|
"learning_rate": 1.1520386060067401e-05, |
|
"loss": 1.4675, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 2.0629274965800275, |
|
"grad_norm": 0.3478207588195801, |
|
"learning_rate": 1.1458798759002897e-05, |
|
"loss": 1.5474, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 2.0656634746922027, |
|
"grad_norm": 0.3087264895439148, |
|
"learning_rate": 1.139732753788072e-05, |
|
"loss": 1.4078, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 2.0683994528043774, |
|
"grad_norm": 0.2971005141735077, |
|
"learning_rate": 1.1335972923653774e-05, |
|
"loss": 1.6057, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 2.0711354309165526, |
|
"grad_norm": 0.30974239110946655, |
|
"learning_rate": 1.1274735442275402e-05, |
|
"loss": 1.4899, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 2.073871409028728, |
|
"grad_norm": 0.3504452705383301, |
|
"learning_rate": 1.121361561869488e-05, |
|
"loss": 1.5413, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 2.076607387140903, |
|
"grad_norm": 0.3584803640842438, |
|
"learning_rate": 1.1152613976852804e-05, |
|
"loss": 1.5634, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 2.079343365253078, |
|
"grad_norm": 0.3279834985733032, |
|
"learning_rate": 1.1091731039676754e-05, |
|
"loss": 1.4428, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.082079343365253, |
|
"grad_norm": 0.2989148199558258, |
|
"learning_rate": 1.1030967329076658e-05, |
|
"loss": 1.3392, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 2.084815321477428, |
|
"grad_norm": 0.3270657956600189, |
|
"learning_rate": 1.0970323365940444e-05, |
|
"loss": 1.4114, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 2.0875512995896033, |
|
"grad_norm": 0.31774088740348816, |
|
"learning_rate": 1.0909799670129497e-05, |
|
"loss": 1.4251, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 2.0902872777017785, |
|
"grad_norm": 0.3060950040817261, |
|
"learning_rate": 1.0849396760474246e-05, |
|
"loss": 1.4517, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 2.0930232558139537, |
|
"grad_norm": 0.2948378026485443, |
|
"learning_rate": 1.078911515476968e-05, |
|
"loss": 1.4872, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 2.0957592339261284, |
|
"grad_norm": 0.2836145758628845, |
|
"learning_rate": 1.0728955369770941e-05, |
|
"loss": 1.4987, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 2.0984952120383036, |
|
"grad_norm": 0.33373308181762695, |
|
"learning_rate": 1.0668917921188885e-05, |
|
"loss": 1.4806, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 2.1012311901504788, |
|
"grad_norm": 0.29727426171302795, |
|
"learning_rate": 1.060900332368562e-05, |
|
"loss": 1.506, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 2.103967168262654, |
|
"grad_norm": 0.3088816702365875, |
|
"learning_rate": 1.0549212090870203e-05, |
|
"loss": 1.7262, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 2.106703146374829, |
|
"grad_norm": 0.37396734952926636, |
|
"learning_rate": 1.0489544735294088e-05, |
|
"loss": 1.4311, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.109439124487004, |
|
"grad_norm": 0.3265590965747833, |
|
"learning_rate": 1.0430001768446856e-05, |
|
"loss": 1.3106, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 2.112175102599179, |
|
"grad_norm": 0.3493903577327728, |
|
"learning_rate": 1.0370583700751774e-05, |
|
"loss": 1.6644, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 2.1149110807113543, |
|
"grad_norm": 0.32931625843048096, |
|
"learning_rate": 1.0311291041561428e-05, |
|
"loss": 1.448, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 2.1176470588235294, |
|
"grad_norm": 0.3126463294029236, |
|
"learning_rate": 1.0252124299153353e-05, |
|
"loss": 1.5439, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 2.1203830369357046, |
|
"grad_norm": 0.3198404014110565, |
|
"learning_rate": 1.0193083980725696e-05, |
|
"loss": 1.4068, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 2.12311901504788, |
|
"grad_norm": 0.3446153998374939, |
|
"learning_rate": 1.0134170592392836e-05, |
|
"loss": 1.4597, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 2.1258549931600546, |
|
"grad_norm": 0.34225985407829285, |
|
"learning_rate": 1.007538463918107e-05, |
|
"loss": 1.5849, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 2.1285909712722297, |
|
"grad_norm": 0.3094581663608551, |
|
"learning_rate": 1.0016726625024287e-05, |
|
"loss": 1.5853, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 2.131326949384405, |
|
"grad_norm": 0.27853038907051086, |
|
"learning_rate": 9.9581970527596e-06, |
|
"loss": 1.5799, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 2.13406292749658, |
|
"grad_norm": 0.33034032583236694, |
|
"learning_rate": 9.899796424123136e-06, |
|
"loss": 1.456, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.1367989056087553, |
|
"grad_norm": 0.3268153965473175, |
|
"learning_rate": 9.841525239745605e-06, |
|
"loss": 1.3883, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 2.13953488372093, |
|
"grad_norm": 0.32485464215278625, |
|
"learning_rate": 9.783383999148118e-06, |
|
"loss": 1.4159, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 2.1422708618331052, |
|
"grad_norm": 0.29737138748168945, |
|
"learning_rate": 9.72537320073785e-06, |
|
"loss": 1.5254, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 2.1450068399452804, |
|
"grad_norm": 0.2851496934890747, |
|
"learning_rate": 9.667493341803777e-06, |
|
"loss": 1.3903, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 2.1477428180574556, |
|
"grad_norm": 0.32795658707618713, |
|
"learning_rate": 9.60974491851242e-06, |
|
"loss": 1.5147, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 2.150478796169631, |
|
"grad_norm": 0.33503639698028564, |
|
"learning_rate": 9.552128425903586e-06, |
|
"loss": 1.4495, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 2.153214774281806, |
|
"grad_norm": 0.31696316599845886, |
|
"learning_rate": 9.494644357886124e-06, |
|
"loss": 1.5787, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 2.1559507523939807, |
|
"grad_norm": 0.321809321641922, |
|
"learning_rate": 9.437293207233695e-06, |
|
"loss": 1.5817, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 2.158686730506156, |
|
"grad_norm": 0.29343757033348083, |
|
"learning_rate": 9.380075465580552e-06, |
|
"loss": 1.5604, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 2.161422708618331, |
|
"grad_norm": 0.4285193383693695, |
|
"learning_rate": 9.322991623417285e-06, |
|
"loss": 1.4387, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.1641586867305063, |
|
"grad_norm": 0.29187867045402527, |
|
"learning_rate": 9.266042170086717e-06, |
|
"loss": 1.4871, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 2.1668946648426815, |
|
"grad_norm": 0.2987869679927826, |
|
"learning_rate": 9.209227593779573e-06, |
|
"loss": 1.3972, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 2.169630642954856, |
|
"grad_norm": 0.29556939005851746, |
|
"learning_rate": 9.152548381530405e-06, |
|
"loss": 1.4057, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 2.1723666210670314, |
|
"grad_norm": 0.32454240322113037, |
|
"learning_rate": 9.096005019213363e-06, |
|
"loss": 1.5509, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 2.1751025991792066, |
|
"grad_norm": 0.30358558893203735, |
|
"learning_rate": 9.039597991538043e-06, |
|
"loss": 1.5466, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 2.1778385772913817, |
|
"grad_norm": 0.3079793155193329, |
|
"learning_rate": 8.983327782045359e-06, |
|
"loss": 1.8167, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 2.180574555403557, |
|
"grad_norm": 0.3283805847167969, |
|
"learning_rate": 8.927194873103322e-06, |
|
"loss": 1.5464, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 2.1833105335157317, |
|
"grad_norm": 0.3089551031589508, |
|
"learning_rate": 8.871199745902997e-06, |
|
"loss": 1.4067, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 2.186046511627907, |
|
"grad_norm": 0.3624851405620575, |
|
"learning_rate": 8.815342880454311e-06, |
|
"loss": 1.4652, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 2.188782489740082, |
|
"grad_norm": 0.31375181674957275, |
|
"learning_rate": 8.75962475558198e-06, |
|
"loss": 1.4584, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.1915184678522572, |
|
"grad_norm": 0.36298835277557373, |
|
"learning_rate": 8.704045848921358e-06, |
|
"loss": 1.6755, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 2.1942544459644324, |
|
"grad_norm": 0.3472815155982971, |
|
"learning_rate": 8.648606636914416e-06, |
|
"loss": 1.4749, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 2.196990424076607, |
|
"grad_norm": 0.30575889348983765, |
|
"learning_rate": 8.593307594805586e-06, |
|
"loss": 1.616, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 2.1997264021887823, |
|
"grad_norm": 0.33051928877830505, |
|
"learning_rate": 8.53814919663769e-06, |
|
"loss": 1.4458, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 2.2024623803009575, |
|
"grad_norm": 0.31675076484680176, |
|
"learning_rate": 8.483131915247968e-06, |
|
"loss": 1.4981, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 2.2051983584131327, |
|
"grad_norm": 0.33154037594795227, |
|
"learning_rate": 8.428256222263888e-06, |
|
"loss": 1.5108, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 2.207934336525308, |
|
"grad_norm": 0.3389985263347626, |
|
"learning_rate": 8.373522588099211e-06, |
|
"loss": 1.4522, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 2.210670314637483, |
|
"grad_norm": 0.35865381360054016, |
|
"learning_rate": 8.318931481949906e-06, |
|
"loss": 1.5059, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 2.213406292749658, |
|
"grad_norm": 0.3100070357322693, |
|
"learning_rate": 8.264483371790156e-06, |
|
"loss": 1.441, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 2.216142270861833, |
|
"grad_norm": 0.3514600694179535, |
|
"learning_rate": 8.210178724368312e-06, |
|
"loss": 1.6657, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.218878248974008, |
|
"grad_norm": 0.2982495427131653, |
|
"learning_rate": 8.156018005202926e-06, |
|
"loss": 1.5945, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 2.2216142270861834, |
|
"grad_norm": 0.2904994487762451, |
|
"learning_rate": 8.10200167857874e-06, |
|
"loss": 1.4298, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 2.2243502051983586, |
|
"grad_norm": 0.3352917432785034, |
|
"learning_rate": 8.04813020754272e-06, |
|
"loss": 1.5512, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 2.2270861833105333, |
|
"grad_norm": 0.31548169255256653, |
|
"learning_rate": 7.994404053900076e-06, |
|
"loss": 1.5735, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 2.2298221614227085, |
|
"grad_norm": 0.31222277879714966, |
|
"learning_rate": 7.940823678210282e-06, |
|
"loss": 1.4862, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 2.2325581395348837, |
|
"grad_norm": 0.3492494225502014, |
|
"learning_rate": 7.887389539783208e-06, |
|
"loss": 1.6087, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 2.235294117647059, |
|
"grad_norm": 0.32157906889915466, |
|
"learning_rate": 7.834102096675064e-06, |
|
"loss": 1.5771, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 2.238030095759234, |
|
"grad_norm": 0.3391493260860443, |
|
"learning_rate": 7.780961805684575e-06, |
|
"loss": 1.5289, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 2.2407660738714092, |
|
"grad_norm": 0.32275089621543884, |
|
"learning_rate": 7.727969122349017e-06, |
|
"loss": 1.407, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 2.243502051983584, |
|
"grad_norm": 0.34808090329170227, |
|
"learning_rate": 7.675124500940317e-06, |
|
"loss": 1.4914, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.246238030095759, |
|
"grad_norm": 0.35853490233421326, |
|
"learning_rate": 7.62242839446117e-06, |
|
"loss": 1.4587, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 2.2489740082079344, |
|
"grad_norm": 0.298554390668869, |
|
"learning_rate": 7.569881254641142e-06, |
|
"loss": 1.52, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 2.2517099863201095, |
|
"grad_norm": 0.3177081048488617, |
|
"learning_rate": 7.517483531932809e-06, |
|
"loss": 1.3685, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 2.2544459644322847, |
|
"grad_norm": 0.30928516387939453, |
|
"learning_rate": 7.4652356755078895e-06, |
|
"loss": 1.4636, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 2.2571819425444595, |
|
"grad_norm": 0.3332304060459137, |
|
"learning_rate": 7.4131381332534e-06, |
|
"loss": 1.4894, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.2599179206566347, |
|
"grad_norm": 0.331177681684494, |
|
"learning_rate": 7.3611913517677875e-06, |
|
"loss": 1.5236, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 2.26265389876881, |
|
"grad_norm": 0.3152174949645996, |
|
"learning_rate": 7.30939577635717e-06, |
|
"loss": 1.4508, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 2.265389876880985, |
|
"grad_norm": 0.33167538046836853, |
|
"learning_rate": 7.257751851031428e-06, |
|
"loss": 1.4507, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 2.26812585499316, |
|
"grad_norm": 0.26186057925224304, |
|
"learning_rate": 7.206260018500469e-06, |
|
"loss": 1.581, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 2.270861833105335, |
|
"grad_norm": 0.31686386466026306, |
|
"learning_rate": 7.154920720170399e-06, |
|
"loss": 1.5677, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.27359781121751, |
|
"grad_norm": 0.32669878005981445, |
|
"learning_rate": 7.1037343961397495e-06, |
|
"loss": 1.5512, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 2.2763337893296853, |
|
"grad_norm": 0.3603561520576477, |
|
"learning_rate": 7.05270148519572e-06, |
|
"loss": 1.6478, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 2.2790697674418605, |
|
"grad_norm": 0.3097962737083435, |
|
"learning_rate": 7.001822424810359e-06, |
|
"loss": 1.5819, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 2.2818057455540357, |
|
"grad_norm": 0.32818931341171265, |
|
"learning_rate": 6.951097651136889e-06, |
|
"loss": 1.5053, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 2.2845417236662104, |
|
"grad_norm": 0.41288748383522034, |
|
"learning_rate": 6.900527599005918e-06, |
|
"loss": 1.5391, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 2.2872777017783856, |
|
"grad_norm": 0.3428272008895874, |
|
"learning_rate": 6.8501127019217346e-06, |
|
"loss": 1.4097, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 2.290013679890561, |
|
"grad_norm": 0.34903377294540405, |
|
"learning_rate": 6.799853392058561e-06, |
|
"loss": 1.3627, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 2.292749658002736, |
|
"grad_norm": 0.3197937309741974, |
|
"learning_rate": 6.749750100256902e-06, |
|
"loss": 1.5362, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 2.295485636114911, |
|
"grad_norm": 0.3300299644470215, |
|
"learning_rate": 6.699803256019815e-06, |
|
"loss": 1.6251, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 2.2982216142270864, |
|
"grad_norm": 0.33497533202171326, |
|
"learning_rate": 6.650013287509199e-06, |
|
"loss": 1.5388, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.300957592339261, |
|
"grad_norm": 0.36092567443847656, |
|
"learning_rate": 6.600380621542216e-06, |
|
"loss": 1.4969, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 2.3036935704514363, |
|
"grad_norm": 0.3416774868965149, |
|
"learning_rate": 6.550905683587513e-06, |
|
"loss": 1.5842, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 2.3064295485636115, |
|
"grad_norm": 0.30258703231811523, |
|
"learning_rate": 6.5015888977617016e-06, |
|
"loss": 1.3653, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 2.3091655266757867, |
|
"grad_norm": 0.2986971139907837, |
|
"learning_rate": 6.452430686825603e-06, |
|
"loss": 1.5223, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 2.311901504787962, |
|
"grad_norm": 0.32431551814079285, |
|
"learning_rate": 6.403431472180707e-06, |
|
"loss": 1.4931, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 2.314637482900137, |
|
"grad_norm": 0.32962799072265625, |
|
"learning_rate": 6.354591673865523e-06, |
|
"loss": 1.4833, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 2.317373461012312, |
|
"grad_norm": 0.36730074882507324, |
|
"learning_rate": 6.30591171055199e-06, |
|
"loss": 1.4419, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 2.320109439124487, |
|
"grad_norm": 0.3103657066822052, |
|
"learning_rate": 6.257391999541887e-06, |
|
"loss": 1.4914, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 2.322845417236662, |
|
"grad_norm": 0.3909322917461395, |
|
"learning_rate": 6.209032956763247e-06, |
|
"loss": 1.4929, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 2.3255813953488373, |
|
"grad_norm": 0.346003919839859, |
|
"learning_rate": 6.160834996766815e-06, |
|
"loss": 1.3601, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.3283173734610125, |
|
"grad_norm": 0.3200609087944031, |
|
"learning_rate": 6.112798532722438e-06, |
|
"loss": 1.6096, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 2.3310533515731873, |
|
"grad_norm": 0.31736400723457336, |
|
"learning_rate": 6.064923976415626e-06, |
|
"loss": 1.4462, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 2.3337893296853625, |
|
"grad_norm": 0.2982243299484253, |
|
"learning_rate": 6.0172117382438944e-06, |
|
"loss": 1.6742, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 2.3365253077975376, |
|
"grad_norm": 0.36678165197372437, |
|
"learning_rate": 5.969662227213352e-06, |
|
"loss": 1.429, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 2.339261285909713, |
|
"grad_norm": 0.29900214076042175, |
|
"learning_rate": 5.922275850935136e-06, |
|
"loss": 1.5513, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 2.341997264021888, |
|
"grad_norm": 0.3283601403236389, |
|
"learning_rate": 5.875053015621943e-06, |
|
"loss": 1.7077, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 2.3447332421340628, |
|
"grad_norm": 0.33298107981681824, |
|
"learning_rate": 5.827994126084532e-06, |
|
"loss": 1.5916, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 2.347469220246238, |
|
"grad_norm": 0.3546065390110016, |
|
"learning_rate": 5.781099585728267e-06, |
|
"loss": 1.4382, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 2.350205198358413, |
|
"grad_norm": 0.30142465233802795, |
|
"learning_rate": 5.734369796549652e-06, |
|
"loss": 1.4578, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 2.3529411764705883, |
|
"grad_norm": 0.3024732768535614, |
|
"learning_rate": 5.687805159132881e-06, |
|
"loss": 1.4102, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.3556771545827635, |
|
"grad_norm": 0.3300219774246216, |
|
"learning_rate": 5.641406072646418e-06, |
|
"loss": 1.5036, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 2.3584131326949382, |
|
"grad_norm": 0.37552839517593384, |
|
"learning_rate": 5.595172934839546e-06, |
|
"loss": 1.4401, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 2.3611491108071134, |
|
"grad_norm": 0.3124082386493683, |
|
"learning_rate": 5.549106142039018e-06, |
|
"loss": 1.3749, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 2.3638850889192886, |
|
"grad_norm": 0.34462398290634155, |
|
"learning_rate": 5.503206089145568e-06, |
|
"loss": 1.578, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 2.366621067031464, |
|
"grad_norm": 0.3301493227481842, |
|
"learning_rate": 5.4574731696306146e-06, |
|
"loss": 1.7064, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 2.369357045143639, |
|
"grad_norm": 0.33227112889289856, |
|
"learning_rate": 5.411907775532832e-06, |
|
"loss": 1.4069, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 2.3720930232558137, |
|
"grad_norm": 0.3299727439880371, |
|
"learning_rate": 5.366510297454816e-06, |
|
"loss": 1.4481, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 2.374829001367989, |
|
"grad_norm": 0.3437407314777374, |
|
"learning_rate": 5.321281124559727e-06, |
|
"loss": 1.3406, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 2.377564979480164, |
|
"grad_norm": 0.3354208469390869, |
|
"learning_rate": 5.276220644567948e-06, |
|
"loss": 1.3835, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 2.3803009575923393, |
|
"grad_norm": 0.29964345693588257, |
|
"learning_rate": 5.231329243753772e-06, |
|
"loss": 1.3945, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.3830369357045145, |
|
"grad_norm": 0.3642790615558624, |
|
"learning_rate": 5.186607306942085e-06, |
|
"loss": 1.3748, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 2.3857729138166897, |
|
"grad_norm": 0.29215675592422485, |
|
"learning_rate": 5.142055217505074e-06, |
|
"loss": 1.5209, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 2.3885088919288644, |
|
"grad_norm": 0.35326606035232544, |
|
"learning_rate": 5.097673357358907e-06, |
|
"loss": 1.5554, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 2.3912448700410396, |
|
"grad_norm": 0.3523205816745758, |
|
"learning_rate": 5.053462106960532e-06, |
|
"loss": 1.5203, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 2.3939808481532148, |
|
"grad_norm": 0.3213190734386444, |
|
"learning_rate": 5.00942184530433e-06, |
|
"loss": 1.4171, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.39671682626539, |
|
"grad_norm": 0.3184296488761902, |
|
"learning_rate": 4.96555294991892e-06, |
|
"loss": 1.3045, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 2.399452804377565, |
|
"grad_norm": 0.3270394802093506, |
|
"learning_rate": 4.921855796863933e-06, |
|
"loss": 1.5365, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 2.4021887824897403, |
|
"grad_norm": 0.36887189745903015, |
|
"learning_rate": 4.878330760726713e-06, |
|
"loss": 1.2823, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 2.404924760601915, |
|
"grad_norm": 0.3077680766582489, |
|
"learning_rate": 4.834978214619215e-06, |
|
"loss": 1.5222, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 2.4076607387140903, |
|
"grad_norm": 0.3177037835121155, |
|
"learning_rate": 4.791798530174699e-06, |
|
"loss": 1.4489, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.4103967168262654, |
|
"grad_norm": 0.3479926586151123, |
|
"learning_rate": 4.748792077544623e-06, |
|
"loss": 1.5376, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 2.4131326949384406, |
|
"grad_norm": 0.32515963912010193, |
|
"learning_rate": 4.7059592253954315e-06, |
|
"loss": 1.5803, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 2.415868673050616, |
|
"grad_norm": 0.33568742871284485, |
|
"learning_rate": 4.66330034090541e-06, |
|
"loss": 1.5066, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 2.4186046511627906, |
|
"grad_norm": 0.328945130109787, |
|
"learning_rate": 4.620815789761526e-06, |
|
"loss": 1.4279, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 2.4213406292749657, |
|
"grad_norm": 0.3277706503868103, |
|
"learning_rate": 4.578505936156302e-06, |
|
"loss": 1.5832, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 2.424076607387141, |
|
"grad_norm": 0.35384321212768555, |
|
"learning_rate": 4.5363711427847015e-06, |
|
"loss": 1.4504, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 2.426812585499316, |
|
"grad_norm": 0.3523906469345093, |
|
"learning_rate": 4.494411770840978e-06, |
|
"loss": 1.5453, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 2.4295485636114913, |
|
"grad_norm": 0.30951249599456787, |
|
"learning_rate": 4.452628180015664e-06, |
|
"loss": 1.5488, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 2.432284541723666, |
|
"grad_norm": 0.3258694112300873, |
|
"learning_rate": 4.411020728492374e-06, |
|
"loss": 1.4878, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 2.4350205198358412, |
|
"grad_norm": 0.33756595849990845, |
|
"learning_rate": 4.3695897729448485e-06, |
|
"loss": 1.4415, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.4377564979480164, |
|
"grad_norm": 0.3134118318557739, |
|
"learning_rate": 4.328335668533806e-06, |
|
"loss": 1.3717, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 2.4404924760601916, |
|
"grad_norm": 0.2975947856903076, |
|
"learning_rate": 4.2872587689039484e-06, |
|
"loss": 1.5299, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 2.443228454172367, |
|
"grad_norm": 0.3447892963886261, |
|
"learning_rate": 4.246359426180918e-06, |
|
"loss": 1.4769, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 2.4459644322845415, |
|
"grad_norm": 0.32196173071861267, |
|
"learning_rate": 4.20563799096827e-06, |
|
"loss": 1.4507, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 2.4487004103967167, |
|
"grad_norm": 0.30560824275016785, |
|
"learning_rate": 4.165094812344478e-06, |
|
"loss": 1.6074, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 2.451436388508892, |
|
"grad_norm": 0.3304429352283478, |
|
"learning_rate": 4.124730237859939e-06, |
|
"loss": 1.5368, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 2.454172366621067, |
|
"grad_norm": 0.33104217052459717, |
|
"learning_rate": 4.0845446135339945e-06, |
|
"loss": 1.4162, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 2.4569083447332423, |
|
"grad_norm": 0.3179687261581421, |
|
"learning_rate": 4.0445382838519365e-06, |
|
"loss": 1.6278, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 2.459644322845417, |
|
"grad_norm": 0.3108776807785034, |
|
"learning_rate": 4.004711591762133e-06, |
|
"loss": 1.6218, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 2.462380300957592, |
|
"grad_norm": 0.3062944710254669, |
|
"learning_rate": 3.965064878672983e-06, |
|
"loss": 1.4816, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.4651162790697674, |
|
"grad_norm": 0.3266476094722748, |
|
"learning_rate": 3.925598484450066e-06, |
|
"loss": 1.846, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 2.4678522571819426, |
|
"grad_norm": 0.344200998544693, |
|
"learning_rate": 3.886312747413204e-06, |
|
"loss": 1.3894, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 2.4705882352941178, |
|
"grad_norm": 0.315949410200119, |
|
"learning_rate": 3.847208004333561e-06, |
|
"loss": 1.4174, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 2.473324213406293, |
|
"grad_norm": 0.3364209234714508, |
|
"learning_rate": 3.8082845904307525e-06, |
|
"loss": 1.4093, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 2.4760601915184677, |
|
"grad_norm": 0.3172418475151062, |
|
"learning_rate": 3.7695428393699854e-06, |
|
"loss": 1.3446, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 2.478796169630643, |
|
"grad_norm": 0.3134661316871643, |
|
"learning_rate": 3.730983083259179e-06, |
|
"loss": 1.507, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 2.481532147742818, |
|
"grad_norm": 0.3210635185241699, |
|
"learning_rate": 3.6926056526461334e-06, |
|
"loss": 1.3344, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 2.4842681258549932, |
|
"grad_norm": 0.32122674584388733, |
|
"learning_rate": 3.6544108765156933e-06, |
|
"loss": 1.5514, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 2.4870041039671684, |
|
"grad_norm": 0.36177971959114075, |
|
"learning_rate": 3.6163990822869088e-06, |
|
"loss": 1.5981, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 2.4897400820793436, |
|
"grad_norm": 0.3207845389842987, |
|
"learning_rate": 3.578570595810274e-06, |
|
"loss": 1.5989, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.4924760601915183, |
|
"grad_norm": 0.32466524839401245, |
|
"learning_rate": 3.540925741364873e-06, |
|
"loss": 1.483, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 2.4952120383036935, |
|
"grad_norm": 0.31727364659309387, |
|
"learning_rate": 3.5034648416556486e-06, |
|
"loss": 1.5308, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 2.4979480164158687, |
|
"grad_norm": 0.3401485085487366, |
|
"learning_rate": 3.4661882178106176e-06, |
|
"loss": 1.6316, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 2.500683994528044, |
|
"grad_norm": 0.3355841338634491, |
|
"learning_rate": 3.429096189378114e-06, |
|
"loss": 1.5952, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 2.503419972640219, |
|
"grad_norm": 0.3085014522075653, |
|
"learning_rate": 3.392189074324073e-06, |
|
"loss": 1.6774, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 2.506155950752394, |
|
"grad_norm": 0.3420393466949463, |
|
"learning_rate": 3.355467189029257e-06, |
|
"loss": 1.4805, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 2.508891928864569, |
|
"grad_norm": 0.3142971098423004, |
|
"learning_rate": 3.3189308482865917e-06, |
|
"loss": 1.4958, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 2.511627906976744, |
|
"grad_norm": 0.37307271361351013, |
|
"learning_rate": 3.2825803652984516e-06, |
|
"loss": 1.4541, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 2.5143638850889194, |
|
"grad_norm": 0.33256828784942627, |
|
"learning_rate": 3.2464160516739755e-06, |
|
"loss": 1.4565, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 2.5170998632010946, |
|
"grad_norm": 0.29588747024536133, |
|
"learning_rate": 3.210438217426365e-06, |
|
"loss": 1.4852, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.5198358413132693, |
|
"grad_norm": 0.314485102891922, |
|
"learning_rate": 3.1746471709702964e-06, |
|
"loss": 1.3962, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 2.5225718194254445, |
|
"grad_norm": 0.30730655789375305, |
|
"learning_rate": 3.1390432191192115e-06, |
|
"loss": 1.5221, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 2.5253077975376197, |
|
"grad_norm": 0.34515947103500366, |
|
"learning_rate": 3.1036266670827014e-06, |
|
"loss": 1.4745, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 2.528043775649795, |
|
"grad_norm": 0.3266810476779938, |
|
"learning_rate": 3.068397818463936e-06, |
|
"loss": 1.4541, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 2.53077975376197, |
|
"grad_norm": 0.2986440062522888, |
|
"learning_rate": 3.033356975256979e-06, |
|
"loss": 1.5795, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 2.533515731874145, |
|
"grad_norm": 0.337341845035553, |
|
"learning_rate": 2.9985044378442933e-06, |
|
"loss": 1.5815, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 2.53625170998632, |
|
"grad_norm": 0.3632700741291046, |
|
"learning_rate": 2.963840504994075e-06, |
|
"loss": 1.54, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 2.538987688098495, |
|
"grad_norm": 0.3714698553085327, |
|
"learning_rate": 2.9293654738577647e-06, |
|
"loss": 1.5019, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 2.5417236662106704, |
|
"grad_norm": 0.3200991153717041, |
|
"learning_rate": 2.8950796399674573e-06, |
|
"loss": 1.4587, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 2.5444596443228455, |
|
"grad_norm": 0.3523624539375305, |
|
"learning_rate": 2.860983297233388e-06, |
|
"loss": 1.4111, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.5471956224350203, |
|
"grad_norm": 0.3558140993118286, |
|
"learning_rate": 2.8270767379414087e-06, |
|
"loss": 1.3241, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 2.5499316005471955, |
|
"grad_norm": 0.3104929029941559, |
|
"learning_rate": 2.7933602527504738e-06, |
|
"loss": 1.6253, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 2.5526675786593707, |
|
"grad_norm": 0.33675897121429443, |
|
"learning_rate": 2.7598341306901643e-06, |
|
"loss": 1.5365, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 2.555403556771546, |
|
"grad_norm": 0.39364495873451233, |
|
"learning_rate": 2.726498659158183e-06, |
|
"loss": 1.5522, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 2.558139534883721, |
|
"grad_norm": 0.31920477747917175, |
|
"learning_rate": 2.693354123917943e-06, |
|
"loss": 1.5633, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 2.5608755129958958, |
|
"grad_norm": 0.37531107664108276, |
|
"learning_rate": 2.660400809096045e-06, |
|
"loss": 1.3603, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 2.5636114911080714, |
|
"grad_norm": 0.31020471453666687, |
|
"learning_rate": 2.6276389971799153e-06, |
|
"loss": 1.3862, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 2.566347469220246, |
|
"grad_norm": 0.3524358570575714, |
|
"learning_rate": 2.595068969015327e-06, |
|
"loss": 1.5923, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 2.5690834473324213, |
|
"grad_norm": 0.3156653940677643, |
|
"learning_rate": 2.562691003804024e-06, |
|
"loss": 1.6537, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 2.5718194254445965, |
|
"grad_norm": 0.3592538833618164, |
|
"learning_rate": 2.5305053791013194e-06, |
|
"loss": 1.3821, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 2.5745554035567717, |
|
"grad_norm": 0.3512822091579437, |
|
"learning_rate": 2.498512370813716e-06, |
|
"loss": 1.5147, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 2.577291381668947, |
|
"grad_norm": 0.31111329793930054, |
|
"learning_rate": 2.466712253196535e-06, |
|
"loss": 1.443, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 2.5800273597811216, |
|
"grad_norm": 0.31163808703422546, |
|
"learning_rate": 2.4351052988515783e-06, |
|
"loss": 1.5938, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 2.582763337893297, |
|
"grad_norm": 0.34021854400634766, |
|
"learning_rate": 2.403691778724787e-06, |
|
"loss": 1.4879, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 2.585499316005472, |
|
"grad_norm": 0.30090874433517456, |
|
"learning_rate": 2.3724719621038923e-06, |
|
"loss": 1.5188, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 2.588235294117647, |
|
"grad_norm": 0.2972038984298706, |
|
"learning_rate": 2.341446116616172e-06, |
|
"loss": 1.4422, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 2.5909712722298224, |
|
"grad_norm": 0.360759973526001, |
|
"learning_rate": 2.310614508226078e-06, |
|
"loss": 1.4662, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 2.593707250341997, |
|
"grad_norm": 0.33597201108932495, |
|
"learning_rate": 2.2799774012330076e-06, |
|
"loss": 1.4008, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 2.5964432284541723, |
|
"grad_norm": 0.3385523855686188, |
|
"learning_rate": 2.2495350582690254e-06, |
|
"loss": 1.4652, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 2.5991792065663475, |
|
"grad_norm": 0.3732840418815613, |
|
"learning_rate": 2.219287740296605e-06, |
|
"loss": 1.3399, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.6019151846785227, |
|
"grad_norm": 0.37210366129875183, |
|
"learning_rate": 2.1892357066064128e-06, |
|
"loss": 1.3226, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 2.604651162790698, |
|
"grad_norm": 0.33239394426345825, |
|
"learning_rate": 2.1593792148150437e-06, |
|
"loss": 1.3563, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 2.6073871409028726, |
|
"grad_norm": 0.3127799332141876, |
|
"learning_rate": 2.1297185208628585e-06, |
|
"loss": 1.5005, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 2.610123119015048, |
|
"grad_norm": 0.3832361698150635, |
|
"learning_rate": 2.10025387901176e-06, |
|
"loss": 1.3859, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 2.612859097127223, |
|
"grad_norm": 0.3399605453014374, |
|
"learning_rate": 2.0709855418430317e-06, |
|
"loss": 1.559, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 2.615595075239398, |
|
"grad_norm": 0.36757224798202515, |
|
"learning_rate": 2.0419137602551516e-06, |
|
"loss": 1.475, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 2.6183310533515733, |
|
"grad_norm": 0.3125869929790497, |
|
"learning_rate": 2.013038783461674e-06, |
|
"loss": 1.4586, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 2.621067031463748, |
|
"grad_norm": 0.31474483013153076, |
|
"learning_rate": 1.9843608589890513e-06, |
|
"loss": 1.4757, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 2.6238030095759233, |
|
"grad_norm": 0.3160763382911682, |
|
"learning_rate": 1.9558802326745424e-06, |
|
"loss": 1.5655, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 2.6265389876880985, |
|
"grad_norm": 0.41808021068573, |
|
"learning_rate": 1.927597148664112e-06, |
|
"loss": 1.4522, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 2.6292749658002736, |
|
"grad_norm": 0.38478416204452515, |
|
"learning_rate": 1.8995118494102903e-06, |
|
"loss": 1.5056, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 2.632010943912449, |
|
"grad_norm": 0.33798423409461975, |
|
"learning_rate": 1.8716245756701694e-06, |
|
"loss": 1.5478, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 2.6347469220246236, |
|
"grad_norm": 0.34396031498908997, |
|
"learning_rate": 1.843935566503252e-06, |
|
"loss": 1.438, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 2.6374829001367988, |
|
"grad_norm": 0.317111998796463, |
|
"learning_rate": 1.816445059269481e-06, |
|
"loss": 1.4886, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 2.640218878248974, |
|
"grad_norm": 0.314860075712204, |
|
"learning_rate": 1.7891532896271547e-06, |
|
"loss": 1.5923, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 2.642954856361149, |
|
"grad_norm": 0.29904037714004517, |
|
"learning_rate": 1.7620604915309257e-06, |
|
"loss": 1.3853, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 2.6456908344733243, |
|
"grad_norm": 0.34991443157196045, |
|
"learning_rate": 1.7351668972297924e-06, |
|
"loss": 1.5588, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 2.648426812585499, |
|
"grad_norm": 0.3276471197605133, |
|
"learning_rate": 1.7084727372651155e-06, |
|
"loss": 1.4885, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 2.6511627906976747, |
|
"grad_norm": 0.3197418749332428, |
|
"learning_rate": 1.6819782404686263e-06, |
|
"loss": 1.5063, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 2.6538987688098494, |
|
"grad_norm": 0.3408433198928833, |
|
"learning_rate": 1.655683633960467e-06, |
|
"loss": 1.4564, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 2.6566347469220246, |
|
"grad_norm": 0.33302435278892517, |
|
"learning_rate": 1.62958914314727e-06, |
|
"loss": 1.3688, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 2.6593707250342, |
|
"grad_norm": 0.34161651134490967, |
|
"learning_rate": 1.6036949917201783e-06, |
|
"loss": 1.5104, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 2.662106703146375, |
|
"grad_norm": 0.33324992656707764, |
|
"learning_rate": 1.578001401652987e-06, |
|
"loss": 1.4552, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 2.66484268125855, |
|
"grad_norm": 0.30797359347343445, |
|
"learning_rate": 1.5525085932001832e-06, |
|
"loss": 1.5157, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 2.667578659370725, |
|
"grad_norm": 0.3368433117866516, |
|
"learning_rate": 1.5272167848950913e-06, |
|
"loss": 1.435, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 2.6703146374829, |
|
"grad_norm": 0.3299857974052429, |
|
"learning_rate": 1.5021261935479924e-06, |
|
"loss": 1.5031, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 2.6730506155950753, |
|
"grad_norm": 0.29794323444366455, |
|
"learning_rate": 1.4772370342442694e-06, |
|
"loss": 1.4679, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 2.6757865937072505, |
|
"grad_norm": 0.31903737783432007, |
|
"learning_rate": 1.4525495203425498e-06, |
|
"loss": 1.6308, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 2.6785225718194257, |
|
"grad_norm": 0.3376205861568451, |
|
"learning_rate": 1.428063863472895e-06, |
|
"loss": 1.5566, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 2.6812585499316004, |
|
"grad_norm": 0.3032618761062622, |
|
"learning_rate": 1.4037802735349743e-06, |
|
"loss": 1.4632, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 2.6839945280437756, |
|
"grad_norm": 0.33116862177848816, |
|
"learning_rate": 1.3796989586962544e-06, |
|
"loss": 1.5811, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 2.6867305061559508, |
|
"grad_norm": 0.34618982672691345, |
|
"learning_rate": 1.3558201253902531e-06, |
|
"loss": 1.3734, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 2.689466484268126, |
|
"grad_norm": 0.3325488567352295, |
|
"learning_rate": 1.3321439783147233e-06, |
|
"loss": 1.4946, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 2.692202462380301, |
|
"grad_norm": 0.32722416520118713, |
|
"learning_rate": 1.3086707204299414e-06, |
|
"loss": 1.5425, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 2.694938440492476, |
|
"grad_norm": 0.3376059830188751, |
|
"learning_rate": 1.2854005529569224e-06, |
|
"loss": 1.4031, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 2.697674418604651, |
|
"grad_norm": 0.3256185054779053, |
|
"learning_rate": 1.2623336753757348e-06, |
|
"loss": 1.493, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 2.7004103967168263, |
|
"grad_norm": 0.3338906168937683, |
|
"learning_rate": 1.2394702854237744e-06, |
|
"loss": 1.5289, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 2.7031463748290014, |
|
"grad_norm": 0.31372472643852234, |
|
"learning_rate": 1.2168105790940603e-06, |
|
"loss": 1.5584, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 2.7058823529411766, |
|
"grad_norm": 0.3929591774940491, |
|
"learning_rate": 1.1943547506335723e-06, |
|
"loss": 1.482, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 2.7086183310533514, |
|
"grad_norm": 0.3206530809402466, |
|
"learning_rate": 1.1721029925415738e-06, |
|
"loss": 1.5406, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 2.7113543091655266, |
|
"grad_norm": 0.32885223627090454, |
|
"learning_rate": 1.15005549556797e-06, |
|
"loss": 1.3643, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 2.7140902872777017, |
|
"grad_norm": 0.3618469536304474, |
|
"learning_rate": 1.1282124487116518e-06, |
|
"loss": 1.5104, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 2.716826265389877, |
|
"grad_norm": 0.31617462635040283, |
|
"learning_rate": 1.106574039218919e-06, |
|
"loss": 1.4661, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 2.719562243502052, |
|
"grad_norm": 0.31909239292144775, |
|
"learning_rate": 1.0851404525818204e-06, |
|
"loss": 1.5708, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 2.722298221614227, |
|
"grad_norm": 0.3482041358947754, |
|
"learning_rate": 1.0639118725366049e-06, |
|
"loss": 1.4782, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 2.725034199726402, |
|
"grad_norm": 0.3372204601764679, |
|
"learning_rate": 1.0428884810621403e-06, |
|
"loss": 1.4833, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 2.7277701778385772, |
|
"grad_norm": 0.32383978366851807, |
|
"learning_rate": 1.0220704583783242e-06, |
|
"loss": 1.5989, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 2.7305061559507524, |
|
"grad_norm": 0.3657885491847992, |
|
"learning_rate": 1.0014579829445842e-06, |
|
"loss": 1.6592, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 2.7332421340629276, |
|
"grad_norm": 0.3399549722671509, |
|
"learning_rate": 9.810512314583075e-07, |
|
"loss": 1.3902, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 2.7359781121751023, |
|
"grad_norm": 0.3272568881511688, |
|
"learning_rate": 9.60850378853348e-07, |
|
"loss": 1.3684, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.7359781121751023, |
|
"eval_loss": 1.5569868087768555, |
|
"eval_runtime": 13.8259, |
|
"eval_samples_per_second": 11.138, |
|
"eval_steps_per_second": 11.138, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.738714090287278, |
|
"grad_norm": 0.3006269037723541, |
|
"learning_rate": 9.408555982985262e-07, |
|
"loss": 1.4081, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 2.7414500683994527, |
|
"grad_norm": 0.32424142956733704, |
|
"learning_rate": 9.210670611961375e-07, |
|
"loss": 1.47, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 2.744186046511628, |
|
"grad_norm": 0.3788749873638153, |
|
"learning_rate": 9.014849371804829e-07, |
|
"loss": 1.4683, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 2.746922024623803, |
|
"grad_norm": 0.3551234006881714, |
|
"learning_rate": 8.821093941164233e-07, |
|
"loss": 1.6791, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 2.7496580027359783, |
|
"grad_norm": 0.38576632738113403, |
|
"learning_rate": 8.629405980979332e-07, |
|
"loss": 1.49, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 2.7523939808481535, |
|
"grad_norm": 0.3681055009365082, |
|
"learning_rate": 8.439787134466692e-07, |
|
"loss": 1.4691, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 2.755129958960328, |
|
"grad_norm": 0.3119199872016907, |
|
"learning_rate": 8.252239027105924e-07, |
|
"loss": 1.5351, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 2.7578659370725034, |
|
"grad_norm": 0.31998804211616516, |
|
"learning_rate": 8.066763266625282e-07, |
|
"loss": 1.468, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 2.7606019151846786, |
|
"grad_norm": 0.3693805932998657, |
|
"learning_rate": 7.883361442988374e-07, |
|
"loss": 1.5278, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 2.7633378932968538, |
|
"grad_norm": 0.3538669943809509, |
|
"learning_rate": 7.702035128380053e-07, |
|
"loss": 1.5411, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 2.766073871409029, |
|
"grad_norm": 0.3684946596622467, |
|
"learning_rate": 7.522785877193378e-07, |
|
"loss": 1.4696, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 2.7688098495212037, |
|
"grad_norm": 0.28718748688697815, |
|
"learning_rate": 7.345615226016011e-07, |
|
"loss": 1.5301, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 2.771545827633379, |
|
"grad_norm": 0.33110928535461426, |
|
"learning_rate": 7.170524693617064e-07, |
|
"loss": 1.6021, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 2.774281805745554, |
|
"grad_norm": 0.31040212512016296, |
|
"learning_rate": 6.997515780934244e-07, |
|
"loss": 1.4569, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 2.7770177838577292, |
|
"grad_norm": 0.3069436848163605, |
|
"learning_rate": 6.826589971060837e-07, |
|
"loss": 1.4963, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 2.7797537619699044, |
|
"grad_norm": 0.3085753321647644, |
|
"learning_rate": 6.657748729233026e-07, |
|
"loss": 1.6278, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 2.782489740082079, |
|
"grad_norm": 0.31628352403640747, |
|
"learning_rate": 6.490993502817289e-07, |
|
"loss": 1.4166, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 2.7852257181942544, |
|
"grad_norm": 0.32792040705680847, |
|
"learning_rate": 6.32632572129821e-07, |
|
"loss": 1.4757, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 2.7879616963064295, |
|
"grad_norm": 0.34112995862960815, |
|
"learning_rate": 6.163746796265885e-07, |
|
"loss": 1.4505, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 2.7906976744186047, |
|
"grad_norm": 0.3526972830295563, |
|
"learning_rate": 6.003258121404148e-07, |
|
"loss": 1.4417, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 2.79343365253078, |
|
"grad_norm": 0.3196633458137512, |
|
"learning_rate": 5.844861072478336e-07, |
|
"loss": 1.5024, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 2.7961696306429547, |
|
"grad_norm": 0.3368687033653259, |
|
"learning_rate": 5.688557007323708e-07, |
|
"loss": 1.6036, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 2.79890560875513, |
|
"grad_norm": 0.33089619874954224, |
|
"learning_rate": 5.534347265833711e-07, |
|
"loss": 1.5716, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 2.801641586867305, |
|
"grad_norm": 0.35482943058013916, |
|
"learning_rate": 5.382233169948541e-07, |
|
"loss": 1.5263, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 2.80437756497948, |
|
"grad_norm": 0.3278489112854004, |
|
"learning_rate": 5.232216023643711e-07, |
|
"loss": 1.4273, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 2.8071135430916554, |
|
"grad_norm": 0.3070161044597626, |
|
"learning_rate": 5.084297112919051e-07, |
|
"loss": 1.7237, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 2.80984952120383, |
|
"grad_norm": 0.3581262528896332, |
|
"learning_rate": 4.938477705787508e-07, |
|
"loss": 1.4228, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 2.8125854993160053, |
|
"grad_norm": 0.3500869572162628, |
|
"learning_rate": 4.794759052264259e-07, |
|
"loss": 1.6744, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 2.8153214774281805, |
|
"grad_norm": 0.3327549397945404, |
|
"learning_rate": 4.653142384356324e-07, |
|
"loss": 1.3299, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 2.8180574555403557, |
|
"grad_norm": 0.32055315375328064, |
|
"learning_rate": 4.513628916051532e-07, |
|
"loss": 1.5822, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 2.820793433652531, |
|
"grad_norm": 0.35483717918395996, |
|
"learning_rate": 4.376219843308438e-07, |
|
"loss": 1.4478, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 2.8235294117647056, |
|
"grad_norm": 0.3212404251098633, |
|
"learning_rate": 4.2409163440459697e-07, |
|
"loss": 1.6226, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 2.8262653898768813, |
|
"grad_norm": 0.3329022526741028, |
|
"learning_rate": 4.107719578133301e-07, |
|
"loss": 1.5192, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 2.829001367989056, |
|
"grad_norm": 0.3651573061943054, |
|
"learning_rate": 3.9766306873801064e-07, |
|
"loss": 1.5169, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 2.831737346101231, |
|
"grad_norm": 0.3816221356391907, |
|
"learning_rate": 3.8476507955263475e-07, |
|
"loss": 1.39, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 2.8344733242134064, |
|
"grad_norm": 0.35796451568603516, |
|
"learning_rate": 3.720781008233143e-07, |
|
"loss": 1.4573, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 2.8372093023255816, |
|
"grad_norm": 0.3537936210632324, |
|
"learning_rate": 3.5960224130728857e-07, |
|
"loss": 1.5344, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 2.8399452804377567, |
|
"grad_norm": 0.35659074783325195, |
|
"learning_rate": 3.473376079520224e-07, |
|
"loss": 1.3814, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 2.8426812585499315, |
|
"grad_norm": 0.31420376896858215, |
|
"learning_rate": 3.352843058942623e-07, |
|
"loss": 1.4956, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 2.8454172366621067, |
|
"grad_norm": 0.31859591603279114, |
|
"learning_rate": 3.2344243845915945e-07, |
|
"loss": 1.6171, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 2.848153214774282, |
|
"grad_norm": 0.30240893363952637, |
|
"learning_rate": 3.118121071593677e-07, |
|
"loss": 1.4719, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 2.850889192886457, |
|
"grad_norm": 0.3180094361305237, |
|
"learning_rate": 3.003934116941776e-07, |
|
"loss": 1.4596, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 2.853625170998632, |
|
"grad_norm": 0.3070809841156006, |
|
"learning_rate": 2.891864499486724e-07, |
|
"loss": 1.5858, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 2.856361149110807, |
|
"grad_norm": 0.3325824737548828, |
|
"learning_rate": 2.7819131799285967e-07, |
|
"loss": 1.5769, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 2.859097127222982, |
|
"grad_norm": 0.3298075199127197, |
|
"learning_rate": 2.6740811008089393e-07, |
|
"loss": 1.5655, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 2.8618331053351573, |
|
"grad_norm": 0.35839757323265076, |
|
"learning_rate": 2.568369186502162e-07, |
|
"loss": 1.3463, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 2.8645690834473325, |
|
"grad_norm": 0.3779064118862152, |
|
"learning_rate": 2.464778343208074e-07, |
|
"loss": 1.4612, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 2.8673050615595077, |
|
"grad_norm": 0.32933345437049866, |
|
"learning_rate": 2.36330945894378e-07, |
|
"loss": 1.4131, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 2.8700410396716824, |
|
"grad_norm": 0.3202325999736786, |
|
"learning_rate": 2.2639634035363234e-07, |
|
"loss": 1.5189, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 2.8727770177838576, |
|
"grad_norm": 0.3435043394565582, |
|
"learning_rate": 2.1667410286149992e-07, |
|
"loss": 1.4026, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.875512995896033, |
|
"grad_norm": 0.33565640449523926, |
|
"learning_rate": 2.071643167604248e-07, |
|
"loss": 1.4982, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 2.878248974008208, |
|
"grad_norm": 0.31919005513191223, |
|
"learning_rate": 1.9786706357163842e-07, |
|
"loss": 1.6241, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 2.880984952120383, |
|
"grad_norm": 0.33048737049102783, |
|
"learning_rate": 1.8878242299446848e-07, |
|
"loss": 1.586, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 2.883720930232558, |
|
"grad_norm": 0.31295284628868103, |
|
"learning_rate": 1.7991047290565056e-07, |
|
"loss": 1.5296, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 2.886456908344733, |
|
"grad_norm": 0.30597612261772156, |
|
"learning_rate": 1.712512893586593e-07, |
|
"loss": 1.4865, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 2.8891928864569083, |
|
"grad_norm": 0.3064374327659607, |
|
"learning_rate": 1.6280494658307e-07, |
|
"loss": 1.5211, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 2.8919288645690835, |
|
"grad_norm": 0.3949945867061615, |
|
"learning_rate": 1.5457151698390347e-07, |
|
"loss": 1.5316, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 2.8946648426812587, |
|
"grad_norm": 0.4148292541503906, |
|
"learning_rate": 1.4655107114101007e-07, |
|
"loss": 1.3986, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 2.8974008207934334, |
|
"grad_norm": 0.32118478417396545, |
|
"learning_rate": 1.387436778084783e-07, |
|
"loss": 1.7486, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 2.9001367989056086, |
|
"grad_norm": 0.34837549924850464, |
|
"learning_rate": 1.3114940391403263e-07, |
|
"loss": 1.3545, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 2.902872777017784, |
|
"grad_norm": 0.331442266702652, |
|
"learning_rate": 1.2376831455845893e-07, |
|
"loss": 1.5355, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 2.905608755129959, |
|
"grad_norm": 0.3617335855960846, |
|
"learning_rate": 1.1660047301505495e-07, |
|
"loss": 1.5907, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 2.908344733242134, |
|
"grad_norm": 0.32971206307411194, |
|
"learning_rate": 1.0964594072907786e-07, |
|
"loss": 1.6133, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 2.911080711354309, |
|
"grad_norm": 0.34654051065444946, |
|
"learning_rate": 1.0290477731722814e-07, |
|
"loss": 1.7007, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 2.9138166894664845, |
|
"grad_norm": 0.3612649738788605, |
|
"learning_rate": 9.637704056713049e-08, |
|
"loss": 1.2852, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 2.9165526675786593, |
|
"grad_norm": 0.2997490465641022, |
|
"learning_rate": 9.006278643683696e-08, |
|
"loss": 1.5689, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 2.9192886456908345, |
|
"grad_norm": 0.3305412828922272, |
|
"learning_rate": 8.396206905436355e-08, |
|
"loss": 1.5476, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 2.9220246238030096, |
|
"grad_norm": 0.36031877994537354, |
|
"learning_rate": 7.807494071720156e-08, |
|
"loss": 1.6777, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 2.924760601915185, |
|
"grad_norm": 0.3095710873603821, |
|
"learning_rate": 7.24014518918903e-08, |
|
"loss": 1.589, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 2.92749658002736, |
|
"grad_norm": 0.30806005001068115, |
|
"learning_rate": 6.694165121356732e-08, |
|
"loss": 1.4103, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 2.9302325581395348, |
|
"grad_norm": 0.36173051595687866, |
|
"learning_rate": 6.169558548556887e-08, |
|
"loss": 1.3676, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 2.93296853625171, |
|
"grad_norm": 0.2915642559528351, |
|
"learning_rate": 5.6663299679018955e-08, |
|
"loss": 1.4523, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 2.935704514363885, |
|
"grad_norm": 0.3511941730976105, |
|
"learning_rate": 5.1844836932438134e-08, |
|
"loss": 1.5194, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 2.9384404924760603, |
|
"grad_norm": 0.3125559389591217, |
|
"learning_rate": 4.7240238551385376e-08, |
|
"loss": 1.4604, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 2.9411764705882355, |
|
"grad_norm": 0.3312918245792389, |
|
"learning_rate": 4.284954400810004e-08, |
|
"loss": 1.4807, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 2.9439124487004102, |
|
"grad_norm": 0.3593026101589203, |
|
"learning_rate": 3.86727909411605e-08, |
|
"loss": 1.7197, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 2.9466484268125854, |
|
"grad_norm": 0.3443048298358917, |
|
"learning_rate": 3.471001515516215e-08, |
|
"loss": 1.6237, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 2.9493844049247606, |
|
"grad_norm": 0.32331955432891846, |
|
"learning_rate": 3.096125062041488e-08, |
|
"loss": 1.5434, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 2.952120383036936, |
|
"grad_norm": 0.3328414261341095, |
|
"learning_rate": 2.7426529472648898e-08, |
|
"loss": 1.6007, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 2.954856361149111, |
|
"grad_norm": 0.31277886033058167, |
|
"learning_rate": 2.4105882012734336e-08, |
|
"loss": 1.6117, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 2.9575923392612857, |
|
"grad_norm": 0.32368844747543335, |
|
"learning_rate": 2.0999336706434282e-08, |
|
"loss": 1.4795, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 2.960328317373461, |
|
"grad_norm": 0.3160129487514496, |
|
"learning_rate": 1.810692018414384e-08, |
|
"loss": 1.5325, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 2.963064295485636, |
|
"grad_norm": 0.3517950773239136, |
|
"learning_rate": 1.5428657240676437e-08, |
|
"loss": 1.4103, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 2.9658002735978113, |
|
"grad_norm": 0.36211562156677246, |
|
"learning_rate": 1.2964570835047318e-08, |
|
"loss": 1.6729, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 2.9685362517099865, |
|
"grad_norm": 0.3121832609176636, |
|
"learning_rate": 1.0714682090270933e-08, |
|
"loss": 1.3272, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 2.971272229822161, |
|
"grad_norm": 0.33123496174812317, |
|
"learning_rate": 8.679010293183298e-09, |
|
"loss": 1.506, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 2.9740082079343364, |
|
"grad_norm": 0.32456517219543457, |
|
"learning_rate": 6.857572894281017e-09, |
|
"loss": 1.5943, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 2.9767441860465116, |
|
"grad_norm": 0.3651670813560486, |
|
"learning_rate": 5.250385507568623e-09, |
|
"loss": 1.4488, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 2.9794801641586868, |
|
"grad_norm": 0.3096288740634918, |
|
"learning_rate": 3.8574619104170265e-09, |
|
"loss": 1.4775, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 2.982216142270862, |
|
"grad_norm": 0.33430325984954834, |
|
"learning_rate": 2.67881404346082e-09, |
|
"loss": 1.5276, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 2.9849521203830367, |
|
"grad_norm": 0.2988739609718323, |
|
"learning_rate": 1.7144520104844798e-09, |
|
"loss": 1.4893, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 2.987688098495212, |
|
"grad_norm": 0.35142961144447327, |
|
"learning_rate": 9.643840783363224e-10, |
|
"loss": 1.4864, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 2.990424076607387, |
|
"grad_norm": 0.2971309721469879, |
|
"learning_rate": 4.2861667686744377e-10, |
|
"loss": 1.6171, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 2.9931600547195623, |
|
"grad_norm": 0.36940792202949524, |
|
"learning_rate": 1.0715439885677913e-10, |
|
"loss": 1.5264, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 2.9958960328317374, |
|
"grad_norm": 0.31959155201911926, |
|
"learning_rate": 0.0, |
|
"loss": 1.585, |
|
"step": 1095 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 1095, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 250, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.0313245522312704e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|