|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 21.0, |
|
"global_step": 80997, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5e-09, |
|
"loss": 10.5291, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 2.5e-06, |
|
"loss": 9.5195, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 5e-06, |
|
"loss": 7.8918, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 7.5e-06, |
|
"loss": 6.8556, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1e-05, |
|
"loss": 6.5569, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.25e-05, |
|
"loss": 6.3942, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.5e-05, |
|
"loss": 6.286, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.75e-05, |
|
"loss": 6.1963, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2e-05, |
|
"loss": 6.1271, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 2.25e-05, |
|
"loss": 6.0662, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 2.5e-05, |
|
"loss": 6.0132, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.7500000000000004e-05, |
|
"loss": 5.9636, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 3e-05, |
|
"loss": 5.9214, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.2500000000000004e-05, |
|
"loss": 5.893, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.5e-05, |
|
"loss": 5.8604, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 5.8285, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4e-05, |
|
"loss": 5.8056, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.2495e-05, |
|
"loss": 5.7834, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.4995000000000005e-05, |
|
"loss": 5.7641, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.7495e-05, |
|
"loss": 5.7448, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.9995000000000005e-05, |
|
"loss": 5.7285, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.998375521920668e-05, |
|
"loss": 5.7139, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 4.996744519832985e-05, |
|
"loss": 5.6956, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 4.995113517745303e-05, |
|
"loss": 5.6856, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 4.9934825156576205e-05, |
|
"loss": 5.6724, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 4.991854775574113e-05, |
|
"loss": 5.6621, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 4.9902237734864304e-05, |
|
"loss": 5.6517, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 4.988592771398747e-05, |
|
"loss": 5.643, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 4.986961769311065e-05, |
|
"loss": 5.6321, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 4.9853307672233826e-05, |
|
"loss": 5.6268, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 4.983703027139875e-05, |
|
"loss": 5.6193, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 4.9820752870563676e-05, |
|
"loss": 5.6084, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 4.980444284968685e-05, |
|
"loss": 5.6034, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 4.978813282881002e-05, |
|
"loss": 5.597, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 4.977182280793319e-05, |
|
"loss": 5.5927, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 4.975551278705637e-05, |
|
"loss": 5.5841, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 4.9739202766179545e-05, |
|
"loss": 5.5778, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 4.972289274530272e-05, |
|
"loss": 5.5748, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 4.970658272442589e-05, |
|
"loss": 5.5692, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 4.969030532359082e-05, |
|
"loss": 5.5608, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 4.967399530271399e-05, |
|
"loss": 5.5594, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 4.965771790187892e-05, |
|
"loss": 5.5563, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 4.964140788100209e-05, |
|
"loss": 5.5528, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 4.962509786012526e-05, |
|
"loss": 5.5461, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 4.960878783924844e-05, |
|
"loss": 5.5413, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 4.959251043841336e-05, |
|
"loss": 5.5347, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 4.957623303757829e-05, |
|
"loss": 5.535, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 4.9559923016701466e-05, |
|
"loss": 5.5296, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 4.954361299582464e-05, |
|
"loss": 5.5252, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 4.952730297494781e-05, |
|
"loss": 5.5228, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 6.48, |
|
"learning_rate": 4.951099295407098e-05, |
|
"loss": 5.5196, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 6.61, |
|
"learning_rate": 4.949468293319416e-05, |
|
"loss": 5.5208, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 4.947837291231733e-05, |
|
"loss": 5.5133, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 4.9462062891440504e-05, |
|
"loss": 5.5112, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 4.9445785490605426e-05, |
|
"loss": 5.5082, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 4.94294754697286e-05, |
|
"loss": 5.4993, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 7.26, |
|
"learning_rate": 4.941316544885178e-05, |
|
"loss": 5.5017, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 4.9396855427974955e-05, |
|
"loss": 5.4975, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 4.9380545407098125e-05, |
|
"loss": 5.4939, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 4.9364268006263046e-05, |
|
"loss": 5.4887, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 4.934795798538622e-05, |
|
"loss": 5.4916, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 4.93316479645094e-05, |
|
"loss": 5.4877, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 4.9315337943632576e-05, |
|
"loss": 5.4827, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"learning_rate": 4.929902792275574e-05, |
|
"loss": 5.4813, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 4.9282750521920674e-05, |
|
"loss": 5.4788, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 4.926644050104384e-05, |
|
"loss": 5.4751, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 4.925013048016702e-05, |
|
"loss": 5.4706, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 4.923382045929019e-05, |
|
"loss": 5.4727, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 4.921754305845512e-05, |
|
"loss": 5.4667, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 4.9201233037578294e-05, |
|
"loss": 5.3861, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 4.9184923016701464e-05, |
|
"loss": 5.1493, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 4.9168612995824633e-05, |
|
"loss": 4.972, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 4.915230297494781e-05, |
|
"loss": 4.7871, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 4.9135992954070986e-05, |
|
"loss": 4.5924, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"learning_rate": 4.9119715553235915e-05, |
|
"loss": 4.4234, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 4.9103405532359084e-05, |
|
"loss": 4.2503, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 9.85, |
|
"learning_rate": 4.9087095511482254e-05, |
|
"loss": 4.0878, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 9.98, |
|
"learning_rate": 4.907078549060543e-05, |
|
"loss": 3.8941, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 10.11, |
|
"learning_rate": 4.905447546972861e-05, |
|
"loss": 3.4534, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 10.24, |
|
"learning_rate": 4.9038165448851776e-05, |
|
"loss": 2.8029, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 10.37, |
|
"learning_rate": 4.902185542797495e-05, |
|
"loss": 2.3458, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 4.900554540709812e-05, |
|
"loss": 2.1352, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 10.63, |
|
"learning_rate": 4.89893006263048e-05, |
|
"loss": 2.0089, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 10.76, |
|
"learning_rate": 4.897299060542798e-05, |
|
"loss": 1.9194, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 10.89, |
|
"learning_rate": 4.895668058455115e-05, |
|
"loss": 1.8505, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 4.8940370563674325e-05, |
|
"loss": 1.7901, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 4.8924060542797495e-05, |
|
"loss": 1.7405, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 11.28, |
|
"learning_rate": 4.890775052192067e-05, |
|
"loss": 1.6962, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 11.41, |
|
"learning_rate": 4.889144050104385e-05, |
|
"loss": 1.6585, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 11.54, |
|
"learning_rate": 4.887516310020877e-05, |
|
"loss": 1.6184, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 11.67, |
|
"learning_rate": 4.8858853079331946e-05, |
|
"loss": 1.5853, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 11.8, |
|
"learning_rate": 4.8842543058455116e-05, |
|
"loss": 1.5588, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 11.93, |
|
"learning_rate": 4.882623303757829e-05, |
|
"loss": 1.5296, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 12.06, |
|
"learning_rate": 4.880992301670146e-05, |
|
"loss": 1.504, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 12.19, |
|
"learning_rate": 4.879364561586639e-05, |
|
"loss": 1.4818, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 12.32, |
|
"learning_rate": 4.8777335594989566e-05, |
|
"loss": 1.4608, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 12.44, |
|
"learning_rate": 4.876102557411274e-05, |
|
"loss": 1.4441, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 12.57, |
|
"learning_rate": 4.8744715553235906e-05, |
|
"loss": 1.4253, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 12.7, |
|
"learning_rate": 4.872840553235908e-05, |
|
"loss": 1.4099, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 12.83, |
|
"learning_rate": 4.871209551148226e-05, |
|
"loss": 1.3957, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 12.96, |
|
"learning_rate": 4.869578549060543e-05, |
|
"loss": 1.3802, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 13.09, |
|
"learning_rate": 4.8679475469728604e-05, |
|
"loss": 1.3644, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 13.22, |
|
"learning_rate": 4.8663198068893526e-05, |
|
"loss": 1.3536, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 13.35, |
|
"learning_rate": 4.86468880480167e-05, |
|
"loss": 1.3423, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 13.48, |
|
"learning_rate": 4.863061064718163e-05, |
|
"loss": 1.328, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 13.61, |
|
"learning_rate": 4.86143006263048e-05, |
|
"loss": 1.3191, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 13.74, |
|
"learning_rate": 4.859799060542798e-05, |
|
"loss": 1.3084, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 13.87, |
|
"learning_rate": 4.858168058455115e-05, |
|
"loss": 1.2963, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 4.856540318371608e-05, |
|
"loss": 1.2866, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 14.13, |
|
"learning_rate": 4.854909316283925e-05, |
|
"loss": 1.2748, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 14.26, |
|
"learning_rate": 4.853278314196242e-05, |
|
"loss": 1.2662, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 14.39, |
|
"learning_rate": 4.85164731210856e-05, |
|
"loss": 1.2603, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 14.52, |
|
"learning_rate": 4.8500163100208774e-05, |
|
"loss": 1.2479, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 14.65, |
|
"learning_rate": 4.8483885699373696e-05, |
|
"loss": 1.2429, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 14.78, |
|
"learning_rate": 4.8467575678496865e-05, |
|
"loss": 1.2346, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 14.91, |
|
"learning_rate": 4.845126565762004e-05, |
|
"loss": 1.227, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 4.843495563674322e-05, |
|
"loss": 1.2201, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 15.17, |
|
"learning_rate": 4.8418645615866394e-05, |
|
"loss": 1.2107, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 15.3, |
|
"learning_rate": 4.8402368215031316e-05, |
|
"loss": 1.2042, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 15.43, |
|
"learning_rate": 4.838605819415449e-05, |
|
"loss": 1.1966, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 15.56, |
|
"learning_rate": 4.836974817327766e-05, |
|
"loss": 1.1893, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 15.69, |
|
"learning_rate": 4.835343815240084e-05, |
|
"loss": 1.1849, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 15.82, |
|
"learning_rate": 4.833716075156576e-05, |
|
"loss": 1.1777, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 15.95, |
|
"learning_rate": 4.832085073068894e-05, |
|
"loss": 1.1731, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 16.07, |
|
"learning_rate": 4.830454070981211e-05, |
|
"loss": 1.1659, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 16.2, |
|
"learning_rate": 4.828823068893528e-05, |
|
"loss": 1.159, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 16.33, |
|
"learning_rate": 4.827192066805846e-05, |
|
"loss": 1.1537, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 16.46, |
|
"learning_rate": 4.825564326722338e-05, |
|
"loss": 1.1486, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 16.59, |
|
"learning_rate": 4.823936586638831e-05, |
|
"loss": 1.1415, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 16.72, |
|
"learning_rate": 4.8223055845511486e-05, |
|
"loss": 1.138, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 16.85, |
|
"learning_rate": 4.8206745824634655e-05, |
|
"loss": 1.1307, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 16.98, |
|
"learning_rate": 4.819043580375783e-05, |
|
"loss": 1.127, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 17.11, |
|
"learning_rate": 4.817415840292276e-05, |
|
"loss": 1.1194, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 17.24, |
|
"learning_rate": 4.815784838204594e-05, |
|
"loss": 1.1152, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 17.37, |
|
"learning_rate": 4.8141538361169106e-05, |
|
"loss": 1.1109, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 4.8125228340292276e-05, |
|
"loss": 1.1096, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 17.63, |
|
"learning_rate": 4.8108950939457204e-05, |
|
"loss": 1.1036, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 17.76, |
|
"learning_rate": 4.809264091858038e-05, |
|
"loss": 1.0995, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 17.89, |
|
"learning_rate": 4.807633089770355e-05, |
|
"loss": 1.0941, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 18.02, |
|
"learning_rate": 4.806002087682672e-05, |
|
"loss": 1.09, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 18.15, |
|
"learning_rate": 4.8043710855949896e-05, |
|
"loss": 1.0841, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 18.28, |
|
"learning_rate": 4.802740083507307e-05, |
|
"loss": 1.0802, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 18.41, |
|
"learning_rate": 4.801109081419625e-05, |
|
"loss": 1.0769, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 18.54, |
|
"learning_rate": 4.799481341336117e-05, |
|
"loss": 1.0729, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 18.67, |
|
"learning_rate": 4.79785360125261e-05, |
|
"loss": 1.0696, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 18.8, |
|
"learning_rate": 4.7962225991649276e-05, |
|
"loss": 1.0645, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 18.93, |
|
"learning_rate": 4.7945915970772445e-05, |
|
"loss": 1.06, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 19.06, |
|
"learning_rate": 4.7929605949895615e-05, |
|
"loss": 1.0546, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 19.19, |
|
"learning_rate": 4.791329592901879e-05, |
|
"loss": 1.0523, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 19.32, |
|
"learning_rate": 4.789698590814197e-05, |
|
"loss": 1.0483, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 19.45, |
|
"learning_rate": 4.788067588726514e-05, |
|
"loss": 1.0445, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 19.57, |
|
"learning_rate": 4.7864365866388314e-05, |
|
"loss": 1.0402, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 19.7, |
|
"learning_rate": 4.7848088465553235e-05, |
|
"loss": 1.0354, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 19.83, |
|
"learning_rate": 4.783177844467641e-05, |
|
"loss": 1.0329, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 19.96, |
|
"learning_rate": 4.781546842379959e-05, |
|
"loss": 1.0308, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 20.09, |
|
"learning_rate": 4.779919102296451e-05, |
|
"loss": 1.0272, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 20.22, |
|
"learning_rate": 4.7782881002087686e-05, |
|
"loss": 1.0218, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 20.35, |
|
"learning_rate": 4.7766570981210856e-05, |
|
"loss": 1.0197, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 20.48, |
|
"learning_rate": 4.775026096033403e-05, |
|
"loss": 1.0147, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 20.61, |
|
"learning_rate": 4.773395093945721e-05, |
|
"loss": 1.0115, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 20.74, |
|
"learning_rate": 4.771764091858038e-05, |
|
"loss": 1.0095, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 20.87, |
|
"learning_rate": 4.770136351774531e-05, |
|
"loss": 1.0087, |
|
"step": 80500 |
|
} |
|
], |
|
"max_steps": 1542800, |
|
"num_train_epochs": 400, |
|
"total_flos": 2.1827467373904396e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|