|
{ |
|
"best_metric": 0.9312475247524753, |
|
"best_model_checkpoint": "swin-base-patch4-window7-224-in22k-food101-24-12/checkpoint-9468", |
|
"epoch": 11.996198923028192, |
|
"eval_steps": 500, |
|
"global_step": 9468, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.279831045406548e-07, |
|
"loss": 4.7106, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.0559662090813095e-06, |
|
"loss": 4.6906, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.583949313621964e-06, |
|
"loss": 4.7116, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 2.111932418162619e-06, |
|
"loss": 4.6605, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 2.6399155227032734e-06, |
|
"loss": 4.6749, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 3.167898627243928e-06, |
|
"loss": 4.6545, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 3.6958817317845833e-06, |
|
"loss": 4.6486, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.223864836325238e-06, |
|
"loss": 4.6175, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.751847940865893e-06, |
|
"loss": 4.6091, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 5.279831045406547e-06, |
|
"loss": 4.5804, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 5.807814149947202e-06, |
|
"loss": 4.553, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 6.335797254487856e-06, |
|
"loss": 4.5225, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 6.863780359028511e-06, |
|
"loss": 4.4715, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.3917634635691666e-06, |
|
"loss": 4.4484, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.91974656810982e-06, |
|
"loss": 4.412, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.447729672650476e-06, |
|
"loss": 4.3148, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.97571277719113e-06, |
|
"loss": 4.2547, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.503695881731786e-06, |
|
"loss": 4.1471, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.003167898627244e-05, |
|
"loss": 4.0151, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.0559662090813093e-05, |
|
"loss": 3.8575, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.108764519535375e-05, |
|
"loss": 3.6421, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.1615628299894405e-05, |
|
"loss": 3.4716, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.2143611404435058e-05, |
|
"loss": 3.2361, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.2671594508975712e-05, |
|
"loss": 3.0604, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3199577613516367e-05, |
|
"loss": 2.8739, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3727560718057022e-05, |
|
"loss": 2.6947, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.4255543822597678e-05, |
|
"loss": 2.4712, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.4783526927138333e-05, |
|
"loss": 2.3397, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.531151003167899e-05, |
|
"loss": 2.2093, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.583949313621964e-05, |
|
"loss": 1.9769, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.6367476240760296e-05, |
|
"loss": 1.8977, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.6895459345300952e-05, |
|
"loss": 1.87, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.7423442449841605e-05, |
|
"loss": 1.7175, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.795142555438226e-05, |
|
"loss": 1.5784, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.8479408658922915e-05, |
|
"loss": 1.5497, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.900739176346357e-05, |
|
"loss": 1.4654, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9535374868004224e-05, |
|
"loss": 1.3991, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.006335797254488e-05, |
|
"loss": 1.3978, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 2.0591341077085534e-05, |
|
"loss": 1.3576, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 2.1119324181626187e-05, |
|
"loss": 1.2948, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 2.1647307286166843e-05, |
|
"loss": 1.2813, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 2.21752903907075e-05, |
|
"loss": 1.2638, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 2.2703273495248153e-05, |
|
"loss": 1.232, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 2.323125659978881e-05, |
|
"loss": 1.2045, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 2.3759239704329462e-05, |
|
"loss": 1.2264, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 2.4287222808870115e-05, |
|
"loss": 1.1378, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 2.4815205913410772e-05, |
|
"loss": 1.1677, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.5343189017951425e-05, |
|
"loss": 1.0539, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.587117212249208e-05, |
|
"loss": 1.1444, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.6399155227032734e-05, |
|
"loss": 1.1713, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.6927138331573394e-05, |
|
"loss": 1.0818, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.7455121436114044e-05, |
|
"loss": 1.0298, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.7983104540654697e-05, |
|
"loss": 1.1905, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.8511087645195357e-05, |
|
"loss": 1.0584, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.903907074973601e-05, |
|
"loss": 1.0452, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.9567053854276666e-05, |
|
"loss": 1.0811, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.009503695881732e-05, |
|
"loss": 1.0017, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.062302006335798e-05, |
|
"loss": 0.9961, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.115100316789863e-05, |
|
"loss": 1.0452, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.167898627243928e-05, |
|
"loss": 1.0892, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.220696937697994e-05, |
|
"loss": 1.0172, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.273495248152059e-05, |
|
"loss": 0.9657, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.326293558606125e-05, |
|
"loss": 1.0045, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.3790918690601904e-05, |
|
"loss": 0.967, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.4318901795142554e-05, |
|
"loss": 1.1282, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.484688489968321e-05, |
|
"loss": 1.0085, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.537486800422387e-05, |
|
"loss": 0.9861, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.590285110876452e-05, |
|
"loss": 0.9449, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.643083421330517e-05, |
|
"loss": 0.9347, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.695881731784583e-05, |
|
"loss": 0.9065, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.7486800422386486e-05, |
|
"loss": 0.9231, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.801478352692714e-05, |
|
"loss": 0.9832, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.85427666314678e-05, |
|
"loss": 0.8982, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 3.907074973600845e-05, |
|
"loss": 0.9783, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 3.9598732840549105e-05, |
|
"loss": 0.9307, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.012671594508976e-05, |
|
"loss": 0.8962, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.065469904963041e-05, |
|
"loss": 0.912, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.118268215417107e-05, |
|
"loss": 0.8657, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.8662574257425742, |
|
"eval_loss": 0.46975982189178467, |
|
"eval_runtime": 210.5111, |
|
"eval_samples_per_second": 119.946, |
|
"eval_steps_per_second": 5.002, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.1710665258711724e-05, |
|
"loss": 0.9095, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.2238648363252374e-05, |
|
"loss": 0.8309, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 4.276663146779304e-05, |
|
"loss": 0.8223, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.3294614572333687e-05, |
|
"loss": 0.8348, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.382259767687434e-05, |
|
"loss": 0.8674, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.4350580781415e-05, |
|
"loss": 0.8729, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.487856388595565e-05, |
|
"loss": 0.8924, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 4.5406546990496306e-05, |
|
"loss": 0.7832, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 4.593453009503696e-05, |
|
"loss": 0.8358, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.646251319957762e-05, |
|
"loss": 0.8367, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 4.699049630411827e-05, |
|
"loss": 0.8348, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 4.7518479408658925e-05, |
|
"loss": 0.8652, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 4.804646251319958e-05, |
|
"loss": 0.8556, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 4.857444561774023e-05, |
|
"loss": 0.8999, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 4.9102428722280894e-05, |
|
"loss": 0.7832, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 4.9630411826821544e-05, |
|
"loss": 0.8093, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 4.998239643234363e-05, |
|
"loss": 0.7716, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.992371787348903e-05, |
|
"loss": 0.842, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 4.986503931463444e-05, |
|
"loss": 0.7367, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 4.980636075577984e-05, |
|
"loss": 0.8929, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 4.974768219692524e-05, |
|
"loss": 0.903, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 4.968900363807065e-05, |
|
"loss": 0.8327, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 4.963032507921605e-05, |
|
"loss": 0.7298, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 4.957164652036146e-05, |
|
"loss": 0.8142, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 4.951296796150687e-05, |
|
"loss": 0.7885, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 4.945428940265227e-05, |
|
"loss": 0.8675, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 4.939561084379768e-05, |
|
"loss": 0.8437, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 4.933693228494308e-05, |
|
"loss": 0.8231, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 4.927825372608849e-05, |
|
"loss": 0.7956, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 4.921957516723389e-05, |
|
"loss": 0.8997, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 4.91608966083793e-05, |
|
"loss": 0.7776, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.91022180495247e-05, |
|
"loss": 0.8387, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 4.904353949067011e-05, |
|
"loss": 0.7962, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 4.898486093181552e-05, |
|
"loss": 0.7851, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 4.892618237296092e-05, |
|
"loss": 0.8689, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 4.886750381410633e-05, |
|
"loss": 0.7802, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 4.880882525525173e-05, |
|
"loss": 0.8475, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 4.8750146696397135e-05, |
|
"loss": 0.7887, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 4.8691468137542544e-05, |
|
"loss": 0.8627, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 4.8632789578687945e-05, |
|
"loss": 0.7858, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.8574111019833354e-05, |
|
"loss": 0.7821, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.851543246097876e-05, |
|
"loss": 0.7806, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.8456753902124165e-05, |
|
"loss": 0.8233, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.839807534326957e-05, |
|
"loss": 0.7659, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.8339396784414975e-05, |
|
"loss": 0.7935, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.8280718225560384e-05, |
|
"loss": 0.7032, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.8222039666705786e-05, |
|
"loss": 0.7737, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.8163361107851194e-05, |
|
"loss": 0.7596, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.8104682548996596e-05, |
|
"loss": 0.7665, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.8046003990142005e-05, |
|
"loss": 0.8028, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 4.7987325431287414e-05, |
|
"loss": 0.7909, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.7928646872432815e-05, |
|
"loss": 0.6995, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.7869968313578224e-05, |
|
"loss": 0.7288, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.7811289754723626e-05, |
|
"loss": 0.8578, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.775261119586903e-05, |
|
"loss": 0.8708, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.7693932637014437e-05, |
|
"loss": 0.7869, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.763525407815984e-05, |
|
"loss": 0.7231, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.757657551930525e-05, |
|
"loss": 0.7614, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 4.7517896960450656e-05, |
|
"loss": 0.7119, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 4.7459218401596064e-05, |
|
"loss": 0.7015, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.7400539842741466e-05, |
|
"loss": 0.7512, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 4.734186128388687e-05, |
|
"loss": 0.7182, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 4.728318272503228e-05, |
|
"loss": 0.7416, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.722450416617768e-05, |
|
"loss": 0.8222, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.716582560732309e-05, |
|
"loss": 0.8414, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.710714704846849e-05, |
|
"loss": 0.7486, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 4.70484684896139e-05, |
|
"loss": 0.6933, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 4.6989789930759307e-05, |
|
"loss": 0.7664, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.693111137190471e-05, |
|
"loss": 0.7418, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.687243281305012e-05, |
|
"loss": 0.7018, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.681375425419552e-05, |
|
"loss": 0.7502, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.675507569534092e-05, |
|
"loss": 0.7117, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.669639713648633e-05, |
|
"loss": 0.7302, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.663771857763173e-05, |
|
"loss": 0.7288, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.657904001877714e-05, |
|
"loss": 0.7861, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.652036145992255e-05, |
|
"loss": 0.6804, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.646168290106796e-05, |
|
"loss": 0.7793, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.640300434221336e-05, |
|
"loss": 0.7224, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.634432578335876e-05, |
|
"loss": 0.7506, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9005940594059406, |
|
"eval_loss": 0.34191781282424927, |
|
"eval_runtime": 209.1051, |
|
"eval_samples_per_second": 120.753, |
|
"eval_steps_per_second": 5.036, |
|
"step": 1578 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 4.628564722450417e-05, |
|
"loss": 0.6433, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.622696866564957e-05, |
|
"loss": 0.5974, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.616829010679498e-05, |
|
"loss": 0.6022, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.610961154794038e-05, |
|
"loss": 0.561, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.605093298908579e-05, |
|
"loss": 0.6672, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.59922544302312e-05, |
|
"loss": 0.6367, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.59335758713766e-05, |
|
"loss": 0.6287, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.587489731252201e-05, |
|
"loss": 0.6689, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.581621875366741e-05, |
|
"loss": 0.7079, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.575754019481282e-05, |
|
"loss": 0.6441, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.569886163595822e-05, |
|
"loss": 0.6438, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.5640183077103624e-05, |
|
"loss": 0.5514, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.558150451824903e-05, |
|
"loss": 0.6582, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 4.552282595939444e-05, |
|
"loss": 0.6572, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 4.546414740053985e-05, |
|
"loss": 0.6319, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 4.540546884168525e-05, |
|
"loss": 0.5865, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.5346790282830654e-05, |
|
"loss": 0.671, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.528811172397606e-05, |
|
"loss": 0.6589, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 4.5229433165121465e-05, |
|
"loss": 0.6898, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 4.517075460626687e-05, |
|
"loss": 0.5984, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 4.5112076047412275e-05, |
|
"loss": 0.6316, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 4.5053397488557684e-05, |
|
"loss": 0.6466, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.499471892970309e-05, |
|
"loss": 0.6146, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.4936040370848494e-05, |
|
"loss": 0.6227, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.48773618119939e-05, |
|
"loss": 0.6742, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.4818683253139305e-05, |
|
"loss": 0.6185, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.4760004694284714e-05, |
|
"loss": 0.6783, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.4701326135430115e-05, |
|
"loss": 0.5918, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.464264757657552e-05, |
|
"loss": 0.6669, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.4583969017720926e-05, |
|
"loss": 0.7018, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.452529045886633e-05, |
|
"loss": 0.687, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.4466611900011736e-05, |
|
"loss": 0.6701, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.4407933341157145e-05, |
|
"loss": 0.6083, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.434925478230255e-05, |
|
"loss": 0.6167, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 4.4290576223447956e-05, |
|
"loss": 0.6917, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 4.423189766459336e-05, |
|
"loss": 0.5455, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.4173219105738766e-05, |
|
"loss": 0.6248, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 4.411454054688417e-05, |
|
"loss": 0.6861, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 4.405586198802958e-05, |
|
"loss": 0.5801, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.399718342917498e-05, |
|
"loss": 0.6926, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 4.393850487032039e-05, |
|
"loss": 0.6494, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 4.3879826311465796e-05, |
|
"loss": 0.6579, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 4.38211477526112e-05, |
|
"loss": 0.6515, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 4.3762469193756606e-05, |
|
"loss": 0.6195, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 4.370379063490201e-05, |
|
"loss": 0.6173, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.364511207604741e-05, |
|
"loss": 0.6473, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 4.358643351719282e-05, |
|
"loss": 0.6032, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.352775495833822e-05, |
|
"loss": 0.5736, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.346907639948363e-05, |
|
"loss": 0.638, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 4.341039784062904e-05, |
|
"loss": 0.643, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 4.335171928177445e-05, |
|
"loss": 0.6095, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 4.329304072291985e-05, |
|
"loss": 0.5665, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 4.323436216406525e-05, |
|
"loss": 0.6305, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 4.317568360521066e-05, |
|
"loss": 0.6017, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 4.311700504635606e-05, |
|
"loss": 0.6791, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 4.305832648750147e-05, |
|
"loss": 0.5995, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.299964792864687e-05, |
|
"loss": 0.59, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.294096936979228e-05, |
|
"loss": 0.6056, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.288229081093769e-05, |
|
"loss": 0.651, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.282361225208309e-05, |
|
"loss": 0.6775, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.27649336932285e-05, |
|
"loss": 0.6268, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 4.27062551343739e-05, |
|
"loss": 0.5738, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 4.26475765755193e-05, |
|
"loss": 0.5926, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 4.258889801666471e-05, |
|
"loss": 0.634, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 4.2530219457810114e-05, |
|
"loss": 0.6561, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 4.247154089895552e-05, |
|
"loss": 0.6498, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 4.241286234010093e-05, |
|
"loss": 0.6192, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 4.235418378124634e-05, |
|
"loss": 0.6191, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 4.229550522239174e-05, |
|
"loss": 0.5939, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 4.2236826663537143e-05, |
|
"loss": 0.5603, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 4.217814810468255e-05, |
|
"loss": 0.5979, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 4.2119469545827954e-05, |
|
"loss": 0.6472, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 4.206079098697336e-05, |
|
"loss": 0.5945, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 4.2002112428118765e-05, |
|
"loss": 0.7002, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.194343386926417e-05, |
|
"loss": 0.548, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 4.188475531040958e-05, |
|
"loss": 0.6557, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 4.1826076751554984e-05, |
|
"loss": 0.6296, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 4.176739819270039e-05, |
|
"loss": 0.5829, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 4.1708719633845794e-05, |
|
"loss": 0.6379, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.9115643564356436, |
|
"eval_loss": 0.30606213212013245, |
|
"eval_runtime": 205.5557, |
|
"eval_samples_per_second": 122.838, |
|
"eval_steps_per_second": 5.123, |
|
"step": 2367 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 4.16500410749912e-05, |
|
"loss": 0.4852, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 4.1591362516136605e-05, |
|
"loss": 0.5154, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 4.153268395728201e-05, |
|
"loss": 0.5472, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 4.1474005398427415e-05, |
|
"loss": 0.4925, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 4.1415326839572824e-05, |
|
"loss": 0.5924, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 4.135664828071823e-05, |
|
"loss": 0.5517, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 4.1297969721863635e-05, |
|
"loss": 0.5663, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 4.1239291163009036e-05, |
|
"loss": 0.511, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 4.1180612604154445e-05, |
|
"loss": 0.5845, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 4.112193404529985e-05, |
|
"loss": 0.5001, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 4.1063255486445256e-05, |
|
"loss": 0.4782, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 4.100457692759066e-05, |
|
"loss": 0.4579, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 4.0945898368736066e-05, |
|
"loss": 0.5431, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 4.0887219809881475e-05, |
|
"loss": 0.5234, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 4.082854125102688e-05, |
|
"loss": 0.5162, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 4.0769862692172285e-05, |
|
"loss": 0.5387, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 4.071118413331769e-05, |
|
"loss": 0.5256, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 4.0652505574463096e-05, |
|
"loss": 0.4907, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 4.05938270156085e-05, |
|
"loss": 0.4827, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 4.05351484567539e-05, |
|
"loss": 0.5738, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 4.047646989789931e-05, |
|
"loss": 0.5928, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 4.041779133904472e-05, |
|
"loss": 0.5682, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 4.0359112780190126e-05, |
|
"loss": 0.5793, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 4.030043422133553e-05, |
|
"loss": 0.5124, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 4.024175566248093e-05, |
|
"loss": 0.553, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 4.018307710362634e-05, |
|
"loss": 0.5148, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 4.012439854477174e-05, |
|
"loss": 0.5012, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 4.006571998591715e-05, |
|
"loss": 0.5206, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 4.000704142706255e-05, |
|
"loss": 0.5447, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 3.994836286820796e-05, |
|
"loss": 0.5857, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 3.988968430935337e-05, |
|
"loss": 0.5603, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 3.983100575049877e-05, |
|
"loss": 0.6013, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 3.977232719164418e-05, |
|
"loss": 0.51, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 3.971364863278958e-05, |
|
"loss": 0.5851, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 3.965497007393499e-05, |
|
"loss": 0.5382, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 3.959629151508039e-05, |
|
"loss": 0.5677, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 3.953761295622579e-05, |
|
"loss": 0.536, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 3.94789343973712e-05, |
|
"loss": 0.5584, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 3.942025583851661e-05, |
|
"loss": 0.5447, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 3.936157727966202e-05, |
|
"loss": 0.5571, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 3.930289872080742e-05, |
|
"loss": 0.5712, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 3.924422016195282e-05, |
|
"loss": 0.612, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 3.918554160309823e-05, |
|
"loss": 0.5277, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 3.912686304424363e-05, |
|
"loss": 0.516, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 3.906818448538904e-05, |
|
"loss": 0.5403, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 3.9009505926534443e-05, |
|
"loss": 0.5671, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 3.895082736767985e-05, |
|
"loss": 0.523, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 3.8892148808825254e-05, |
|
"loss": 0.4644, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 3.883347024997066e-05, |
|
"loss": 0.5075, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 3.877479169111607e-05, |
|
"loss": 0.4995, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 3.871611313226147e-05, |
|
"loss": 0.4651, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 3.865743457340688e-05, |
|
"loss": 0.5649, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 3.8598756014552284e-05, |
|
"loss": 0.5433, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 3.8540077455697686e-05, |
|
"loss": 0.5607, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 3.8481398896843094e-05, |
|
"loss": 0.5778, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 3.8422720337988496e-05, |
|
"loss": 0.5377, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 3.8364041779133905e-05, |
|
"loss": 0.5723, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 3.830536322027931e-05, |
|
"loss": 0.6007, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 3.824668466142472e-05, |
|
"loss": 0.4302, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 3.8188006102570124e-05, |
|
"loss": 0.5, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 3.8129327543715526e-05, |
|
"loss": 0.5277, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 3.8070648984860934e-05, |
|
"loss": 0.6177, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 3.8011970426006336e-05, |
|
"loss": 0.546, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 3.7953291867151745e-05, |
|
"loss": 0.5487, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 3.789461330829715e-05, |
|
"loss": 0.4998, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 3.7835934749442556e-05, |
|
"loss": 0.508, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 3.7777256190587964e-05, |
|
"loss": 0.4856, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 3.7718577631733366e-05, |
|
"loss": 0.5739, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 3.7659899072878775e-05, |
|
"loss": 0.5409, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 3.7601220514024177e-05, |
|
"loss": 0.5291, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 3.754254195516958e-05, |
|
"loss": 0.562, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 3.748386339631499e-05, |
|
"loss": 0.4756, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 3.742518483746039e-05, |
|
"loss": 0.4598, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 3.73665062786058e-05, |
|
"loss": 0.5634, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 3.7307827719751206e-05, |
|
"loss": 0.446, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 3.7249149160896615e-05, |
|
"loss": 0.5313, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 3.719047060204202e-05, |
|
"loss": 0.5393, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 3.713179204318742e-05, |
|
"loss": 0.543, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 3.707311348433283e-05, |
|
"loss": 0.5223, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.9148910891089109, |
|
"eval_loss": 0.29058367013931274, |
|
"eval_runtime": 205.3408, |
|
"eval_samples_per_second": 122.966, |
|
"eval_steps_per_second": 5.128, |
|
"step": 3157 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 3.701443492547823e-05, |
|
"loss": 0.5372, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 3.695575636662364e-05, |
|
"loss": 0.4666, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 3.689707780776904e-05, |
|
"loss": 0.4141, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 3.683839924891445e-05, |
|
"loss": 0.4977, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 3.677972069005986e-05, |
|
"loss": 0.4279, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 3.672104213120526e-05, |
|
"loss": 0.4818, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 3.666236357235067e-05, |
|
"loss": 0.3827, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 3.660368501349607e-05, |
|
"loss": 0.5068, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 3.654500645464148e-05, |
|
"loss": 0.4627, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 3.648632789578688e-05, |
|
"loss": 0.4714, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 3.642764933693228e-05, |
|
"loss": 0.4057, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 3.636897077807769e-05, |
|
"loss": 0.4258, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 3.63102922192231e-05, |
|
"loss": 0.4713, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 3.625161366036851e-05, |
|
"loss": 0.4278, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 3.619293510151391e-05, |
|
"loss": 0.4187, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 3.613425654265931e-05, |
|
"loss": 0.4299, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 3.607557798380472e-05, |
|
"loss": 0.466, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 3.601689942495012e-05, |
|
"loss": 0.4718, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 3.595822086609553e-05, |
|
"loss": 0.4739, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 3.589954230724093e-05, |
|
"loss": 0.4473, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 3.584086374838634e-05, |
|
"loss": 0.4114, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 3.578218518953175e-05, |
|
"loss": 0.5411, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 3.572350663067715e-05, |
|
"loss": 0.5255, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 3.566482807182256e-05, |
|
"loss": 0.3926, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 3.560614951296796e-05, |
|
"loss": 0.5723, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 3.554747095411337e-05, |
|
"loss": 0.4964, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 3.548879239525877e-05, |
|
"loss": 0.4697, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 3.5430113836404175e-05, |
|
"loss": 0.4456, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 3.5371435277549584e-05, |
|
"loss": 0.4168, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 3.531275671869499e-05, |
|
"loss": 0.4628, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 3.52540781598404e-05, |
|
"loss": 0.4812, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 3.51953996009858e-05, |
|
"loss": 0.4786, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 3.5136721042131205e-05, |
|
"loss": 0.4806, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 3.507804248327661e-05, |
|
"loss": 0.5063, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 3.5019363924422015e-05, |
|
"loss": 0.533, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 3.4960685365567424e-05, |
|
"loss": 0.4398, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 3.4902006806712826e-05, |
|
"loss": 0.4345, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 3.4843328247858234e-05, |
|
"loss": 0.5306, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 3.478464968900364e-05, |
|
"loss": 0.4646, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 3.4725971130149045e-05, |
|
"loss": 0.4922, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 3.4667292571294454e-05, |
|
"loss": 0.4363, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 3.4608614012439855e-05, |
|
"loss": 0.5095, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 3.4549935453585264e-05, |
|
"loss": 0.5099, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 3.4491256894730666e-05, |
|
"loss": 0.4374, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 3.443257833587607e-05, |
|
"loss": 0.4809, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 3.4373899777021477e-05, |
|
"loss": 0.4665, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 3.4315221218166885e-05, |
|
"loss": 0.4992, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 3.4256542659312294e-05, |
|
"loss": 0.4232, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 3.4197864100457696e-05, |
|
"loss": 0.3981, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 3.4139185541603104e-05, |
|
"loss": 0.4732, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 3.4080506982748506e-05, |
|
"loss": 0.4602, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 3.402182842389391e-05, |
|
"loss": 0.4717, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 3.396314986503932e-05, |
|
"loss": 0.4408, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 3.390447130618472e-05, |
|
"loss": 0.4377, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 3.384579274733013e-05, |
|
"loss": 0.3973, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 3.3787114188475536e-05, |
|
"loss": 0.4642, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 3.372843562962094e-05, |
|
"loss": 0.5274, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 3.3669757070766347e-05, |
|
"loss": 0.4784, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 3.361107851191175e-05, |
|
"loss": 0.4363, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 3.355239995305716e-05, |
|
"loss": 0.4744, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 3.349372139420256e-05, |
|
"loss": 0.4682, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 3.343504283534796e-05, |
|
"loss": 0.5083, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 3.337636427649337e-05, |
|
"loss": 0.4881, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 3.331768571763877e-05, |
|
"loss": 0.5167, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 3.325900715878418e-05, |
|
"loss": 0.4809, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 3.320032859992959e-05, |
|
"loss": 0.477, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 3.3141650041075e-05, |
|
"loss": 0.4748, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 3.30829714822204e-05, |
|
"loss": 0.4468, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 3.30242929233658e-05, |
|
"loss": 0.4747, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 3.296561436451121e-05, |
|
"loss": 0.5463, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 3.290693580565661e-05, |
|
"loss": 0.4443, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 3.284825724680202e-05, |
|
"loss": 0.4492, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 3.278957868794742e-05, |
|
"loss": 0.4608, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 3.273090012909283e-05, |
|
"loss": 0.4599, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 3.267222157023824e-05, |
|
"loss": 0.5169, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 3.261354301138364e-05, |
|
"loss": 0.5056, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 3.255486445252905e-05, |
|
"loss": 0.3813, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 3.249618589367445e-05, |
|
"loss": 0.463, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 3.243750733481986e-05, |
|
"loss": 0.4989, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.9204752475247525, |
|
"eval_loss": 0.2782570719718933, |
|
"eval_runtime": 206.4182, |
|
"eval_samples_per_second": 122.325, |
|
"eval_steps_per_second": 5.101, |
|
"step": 3946 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 3.237882877596526e-05, |
|
"loss": 0.4458, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 3.2320150217110664e-05, |
|
"loss": 0.4076, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 3.226147165825607e-05, |
|
"loss": 0.4453, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 3.220279309940148e-05, |
|
"loss": 0.395, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 3.214411454054689e-05, |
|
"loss": 0.4829, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 3.208543598169229e-05, |
|
"loss": 0.4169, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 3.2026757422837694e-05, |
|
"loss": 0.4055, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 3.19680788639831e-05, |
|
"loss": 0.423, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 3.1909400305128505e-05, |
|
"loss": 0.3491, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 3.185072174627391e-05, |
|
"loss": 0.3507, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 3.1792043187419315e-05, |
|
"loss": 0.3939, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 3.1733364628564724e-05, |
|
"loss": 0.4037, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 3.167468606971013e-05, |
|
"loss": 0.4346, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 3.1616007510855534e-05, |
|
"loss": 0.36, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 3.155732895200094e-05, |
|
"loss": 0.4178, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 3.1498650393146345e-05, |
|
"loss": 0.3803, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 3.1439971834291754e-05, |
|
"loss": 0.4111, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 3.1381293275437155e-05, |
|
"loss": 0.381, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 3.132261471658256e-05, |
|
"loss": 0.4, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 3.1263936157727966e-05, |
|
"loss": 0.3908, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 3.1205257598873375e-05, |
|
"loss": 0.4138, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"learning_rate": 3.114657904001878e-05, |
|
"loss": 0.3986, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 3.1087900481164185e-05, |
|
"loss": 0.4471, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 3.102922192230959e-05, |
|
"loss": 0.3875, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 3.0970543363454996e-05, |
|
"loss": 0.3798, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 3.09118648046004e-05, |
|
"loss": 0.4232, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 3.0853186245745806e-05, |
|
"loss": 0.428, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 3.079450768689121e-05, |
|
"loss": 0.4497, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 3.073582912803662e-05, |
|
"loss": 0.4457, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 3.0677150569182025e-05, |
|
"loss": 0.422, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 3.061847201032743e-05, |
|
"loss": 0.3987, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 3.0559793451472836e-05, |
|
"loss": 0.4106, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 3.0501114892618238e-05, |
|
"loss": 0.42, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 3.0442436333763646e-05, |
|
"loss": 0.4512, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 3.038375777490905e-05, |
|
"loss": 0.4611, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 3.0325079216054454e-05, |
|
"loss": 0.3822, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 3.0266400657199862e-05, |
|
"loss": 0.5147, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 3.0207722098345264e-05, |
|
"loss": 0.3723, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 3.0149043539490673e-05, |
|
"loss": 0.3271, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 3.0090364980636078e-05, |
|
"loss": 0.416, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 3.003168642178148e-05, |
|
"loss": 0.3915, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"learning_rate": 2.997300786292689e-05, |
|
"loss": 0.4107, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 5.54, |
|
"learning_rate": 2.991432930407229e-05, |
|
"loss": 0.3801, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 2.98556507452177e-05, |
|
"loss": 0.3642, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 2.9796972186363104e-05, |
|
"loss": 0.3844, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 2.9738293627508513e-05, |
|
"loss": 0.4704, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 2.9679615068653915e-05, |
|
"loss": 0.391, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 2.962093650979932e-05, |
|
"loss": 0.4143, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 2.956225795094473e-05, |
|
"loss": 0.4215, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 2.950357939209013e-05, |
|
"loss": 0.4399, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 2.944490083323554e-05, |
|
"loss": 0.3714, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 5.65, |
|
"learning_rate": 2.938622227438094e-05, |
|
"loss": 0.4223, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 2.9327543715526347e-05, |
|
"loss": 0.4409, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 5.68, |
|
"learning_rate": 2.9268865156671755e-05, |
|
"loss": 0.396, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 2.9210186597817157e-05, |
|
"loss": 0.4342, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 2.9151508038962566e-05, |
|
"loss": 0.4026, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 2.9092829480107968e-05, |
|
"loss": 0.4271, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 2.903415092125338e-05, |
|
"loss": 0.4658, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 2.897547236239878e-05, |
|
"loss": 0.406, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 2.8916793803544183e-05, |
|
"loss": 0.4104, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 2.8858115244689592e-05, |
|
"loss": 0.438, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 2.8799436685834997e-05, |
|
"loss": 0.429, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 2.8740758126980406e-05, |
|
"loss": 0.3924, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 2.8682079568125808e-05, |
|
"loss": 0.4626, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 2.862340100927121e-05, |
|
"loss": 0.37, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 2.856472245041662e-05, |
|
"loss": 0.4609, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 2.8506043891562024e-05, |
|
"loss": 0.4181, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 2.8447365332707432e-05, |
|
"loss": 0.3355, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 5.87, |
|
"learning_rate": 2.8388686773852834e-05, |
|
"loss": 0.4343, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 2.8330008214998243e-05, |
|
"loss": 0.3644, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 2.8271329656143648e-05, |
|
"loss": 0.4337, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 2.821265109728905e-05, |
|
"loss": 0.3836, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 2.815397253843446e-05, |
|
"loss": 0.5148, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 2.809529397957986e-05, |
|
"loss": 0.3483, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 2.803661542072527e-05, |
|
"loss": 0.4047, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 2.7977936861870675e-05, |
|
"loss": 0.4604, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 2.7919258303016076e-05, |
|
"loss": 0.4343, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 2.7860579744161485e-05, |
|
"loss": 0.4236, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"learning_rate": 2.780190118530689e-05, |
|
"loss": 0.4163, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.9224950495049505, |
|
"eval_loss": 0.273190975189209, |
|
"eval_runtime": 205.7757, |
|
"eval_samples_per_second": 122.706, |
|
"eval_steps_per_second": 5.117, |
|
"step": 4735 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 2.77432226264523e-05, |
|
"loss": 0.4122, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 2.76845440675977e-05, |
|
"loss": 0.3402, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 2.7625865508743103e-05, |
|
"loss": 0.3125, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 2.756718694988851e-05, |
|
"loss": 0.4288, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 2.7508508391033917e-05, |
|
"loss": 0.3229, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 2.7449829832179325e-05, |
|
"loss": 0.445, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 6.08, |
|
"learning_rate": 2.7391151273324727e-05, |
|
"loss": 0.3602, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 2.7332472714470136e-05, |
|
"loss": 0.3676, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 2.727379415561554e-05, |
|
"loss": 0.3627, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 2.7215115596760943e-05, |
|
"loss": 0.344, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 2.715643703790635e-05, |
|
"loss": 0.3592, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"learning_rate": 2.7097758479051754e-05, |
|
"loss": 0.4404, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 2.7039079920197162e-05, |
|
"loss": 0.3617, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 2.6980401361342567e-05, |
|
"loss": 0.4269, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 2.692172280248797e-05, |
|
"loss": 0.337, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 2.6863044243633378e-05, |
|
"loss": 0.402, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 2.6804365684778783e-05, |
|
"loss": 0.3494, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 2.6745687125924192e-05, |
|
"loss": 0.3566, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 2.6687008567069594e-05, |
|
"loss": 0.3723, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 2.6628330008215002e-05, |
|
"loss": 0.3863, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 2.6569651449360404e-05, |
|
"loss": 0.3916, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"learning_rate": 2.651097289050581e-05, |
|
"loss": 0.3515, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 2.6452294331651218e-05, |
|
"loss": 0.3799, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 2.639361577279662e-05, |
|
"loss": 0.3983, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 2.633493721394203e-05, |
|
"loss": 0.3482, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 2.627625865508743e-05, |
|
"loss": 0.321, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 2.6217580096232836e-05, |
|
"loss": 0.4025, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 2.6158901537378245e-05, |
|
"loss": 0.3869, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 2.6100222978523646e-05, |
|
"loss": 0.325, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 6.37, |
|
"learning_rate": 2.6041544419669055e-05, |
|
"loss": 0.3657, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 2.598286586081446e-05, |
|
"loss": 0.3795, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 2.5924187301959862e-05, |
|
"loss": 0.3795, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 2.586550874310527e-05, |
|
"loss": 0.3538, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 2.5806830184250673e-05, |
|
"loss": 0.3107, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 2.574815162539608e-05, |
|
"loss": 0.3997, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 2.5689473066541487e-05, |
|
"loss": 0.4113, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 2.5630794507686895e-05, |
|
"loss": 0.3955, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 2.5572115948832297e-05, |
|
"loss": 0.378, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 2.5513437389977703e-05, |
|
"loss": 0.3988, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 2.545475883112311e-05, |
|
"loss": 0.3992, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 2.5396080272268513e-05, |
|
"loss": 0.3651, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 2.5337401713413922e-05, |
|
"loss": 0.3354, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 2.5278723154559324e-05, |
|
"loss": 0.4064, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 6.55, |
|
"learning_rate": 2.522004459570473e-05, |
|
"loss": 0.3831, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 2.5161366036850138e-05, |
|
"loss": 0.3761, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 2.510268747799554e-05, |
|
"loss": 0.3983, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"learning_rate": 2.5044008919140948e-05, |
|
"loss": 0.3185, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 2.4985330360286353e-05, |
|
"loss": 0.3569, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 6.61, |
|
"learning_rate": 2.492665180143176e-05, |
|
"loss": 0.3662, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 2.4867973242577164e-05, |
|
"loss": 0.3299, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 2.480929468372257e-05, |
|
"loss": 0.3473, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 2.4750616124867974e-05, |
|
"loss": 0.398, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 2.469193756601338e-05, |
|
"loss": 0.3666, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 6.68, |
|
"learning_rate": 2.4633259007158785e-05, |
|
"loss": 0.39, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 2.457458044830419e-05, |
|
"loss": 0.3492, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 2.4515901889449596e-05, |
|
"loss": 0.3733, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 6.72, |
|
"learning_rate": 2.4457223330595004e-05, |
|
"loss": 0.3381, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 2.4398544771740406e-05, |
|
"loss": 0.4035, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 2.433986621288581e-05, |
|
"loss": 0.4103, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 2.4281187654031217e-05, |
|
"loss": 0.354, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 2.4222509095176625e-05, |
|
"loss": 0.4118, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 2.416383053632203e-05, |
|
"loss": 0.3443, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 2.4105151977467436e-05, |
|
"loss": 0.3777, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 2.4046473418612838e-05, |
|
"loss": 0.3126, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 6.82, |
|
"learning_rate": 2.3987794859758246e-05, |
|
"loss": 0.3591, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 2.392911630090365e-05, |
|
"loss": 0.4017, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 2.3870437742049057e-05, |
|
"loss": 0.3405, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 2.3811759183194462e-05, |
|
"loss": 0.3906, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 2.3753080624339867e-05, |
|
"loss": 0.4004, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 2.3694402065485273e-05, |
|
"loss": 0.3569, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 2.3635723506630678e-05, |
|
"loss": 0.3667, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 2.3577044947776083e-05, |
|
"loss": 0.3715, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"learning_rate": 2.351836638892149e-05, |
|
"loss": 0.3527, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 2.3459687830066894e-05, |
|
"loss": 0.4153, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 2.3401009271212302e-05, |
|
"loss": 0.3581, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 2.3342330712357704e-05, |
|
"loss": 0.3371, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 2.328365215350311e-05, |
|
"loss": 0.3589, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 2.3224973594648515e-05, |
|
"loss": 0.3827, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"learning_rate": 2.3166295035793923e-05, |
|
"loss": 0.3954, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.9254653465346535, |
|
"eval_loss": 0.2674882709980011, |
|
"eval_runtime": 207.0381, |
|
"eval_samples_per_second": 121.958, |
|
"eval_steps_per_second": 5.086, |
|
"step": 5524 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 2.310761647693933e-05, |
|
"loss": 0.3321, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 2.3048937918084734e-05, |
|
"loss": 0.3251, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 2.2990259359230136e-05, |
|
"loss": 0.3579, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 2.2931580800375545e-05, |
|
"loss": 0.2955, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 2.287290224152095e-05, |
|
"loss": 0.345, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 2.2814223682666355e-05, |
|
"loss": 0.3264, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 2.275554512381176e-05, |
|
"loss": 0.3116, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 2.2696866564957166e-05, |
|
"loss": 0.3086, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 2.263818800610257e-05, |
|
"loss": 0.3855, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 2.2579509447247976e-05, |
|
"loss": 0.369, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 2.252083088839338e-05, |
|
"loss": 0.309, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 2.2462152329538787e-05, |
|
"loss": 0.323, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 2.2403473770684195e-05, |
|
"loss": 0.2983, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"learning_rate": 2.2344795211829597e-05, |
|
"loss": 0.3396, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 2.2286116652975002e-05, |
|
"loss": 0.3742, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 2.2227438094120408e-05, |
|
"loss": 0.3238, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 2.2168759535265816e-05, |
|
"loss": 0.338, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 2.211008097641122e-05, |
|
"loss": 0.3677, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 2.2051402417556627e-05, |
|
"loss": 0.3427, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 2.199272385870203e-05, |
|
"loss": 0.3292, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 7.26, |
|
"learning_rate": 2.1934045299847437e-05, |
|
"loss": 0.3758, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 2.1875366740992843e-05, |
|
"loss": 0.3846, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 2.1816688182138248e-05, |
|
"loss": 0.358, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 2.1758009623283653e-05, |
|
"loss": 0.3291, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 2.169933106442906e-05, |
|
"loss": 0.3478, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 7.32, |
|
"learning_rate": 2.1640652505574464e-05, |
|
"loss": 0.3128, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 2.158197394671987e-05, |
|
"loss": 0.3389, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"learning_rate": 2.1523295387865274e-05, |
|
"loss": 0.2623, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 2.146461682901068e-05, |
|
"loss": 0.3143, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 2.1405938270156088e-05, |
|
"loss": 0.3198, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 2.1347259711301494e-05, |
|
"loss": 0.3274, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"learning_rate": 2.1288581152446895e-05, |
|
"loss": 0.2896, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 2.12299025935923e-05, |
|
"loss": 0.3262, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 2.117122403473771e-05, |
|
"loss": 0.3018, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 7.44, |
|
"learning_rate": 2.1112545475883115e-05, |
|
"loss": 0.3591, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 2.105386691702852e-05, |
|
"loss": 0.3262, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 2.0995188358173925e-05, |
|
"loss": 0.3759, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 7.48, |
|
"learning_rate": 2.093650979931933e-05, |
|
"loss": 0.284, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 2.0877831240464736e-05, |
|
"loss": 0.2967, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 2.081915268161014e-05, |
|
"loss": 0.3008, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 7.51, |
|
"learning_rate": 2.0760474122755546e-05, |
|
"loss": 0.3084, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 2.070179556390095e-05, |
|
"loss": 0.3124, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 2.0643117005046357e-05, |
|
"loss": 0.3706, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 2.0584438446191762e-05, |
|
"loss": 0.3276, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 2.0525759887337167e-05, |
|
"loss": 0.281, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 2.0467081328482573e-05, |
|
"loss": 0.3318, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 7.59, |
|
"learning_rate": 2.0408402769627978e-05, |
|
"loss": 0.3004, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 2.0349724210773387e-05, |
|
"loss": 0.3474, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 2.029104565191879e-05, |
|
"loss": 0.3286, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 2.0232367093064194e-05, |
|
"loss": 0.2868, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 2.01736885342096e-05, |
|
"loss": 0.3196, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 2.0115009975355008e-05, |
|
"loss": 0.3361, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"learning_rate": 2.0056331416500413e-05, |
|
"loss": 0.3154, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 1.9997652857645818e-05, |
|
"loss": 0.3525, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 1.993897429879122e-05, |
|
"loss": 0.3494, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 1.988029573993663e-05, |
|
"loss": 0.4165, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 1.9821617181082034e-05, |
|
"loss": 0.3688, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 7.73, |
|
"learning_rate": 1.976293862222744e-05, |
|
"loss": 0.3128, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 1.9704260063372844e-05, |
|
"loss": 0.3196, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 1.964558150451825e-05, |
|
"loss": 0.3302, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 7.77, |
|
"learning_rate": 1.9586902945663655e-05, |
|
"loss": 0.3313, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 1.952822438680906e-05, |
|
"loss": 0.2861, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 1.9469545827954466e-05, |
|
"loss": 0.275, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"learning_rate": 1.941086726909987e-05, |
|
"loss": 0.3218, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 1.935218871024528e-05, |
|
"loss": 0.3825, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 1.9293510151390685e-05, |
|
"loss": 0.3579, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 1.9234831592536087e-05, |
|
"loss": 0.293, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 1.9176153033681492e-05, |
|
"loss": 0.3922, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 7.87, |
|
"learning_rate": 1.91174744748269e-05, |
|
"loss": 0.3546, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 1.9058795915972306e-05, |
|
"loss": 0.3781, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 1.900011735711771e-05, |
|
"loss": 0.3082, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 1.8941438798263113e-05, |
|
"loss": 0.3528, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 1.888276023940852e-05, |
|
"loss": 0.3292, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 1.8824081680553927e-05, |
|
"loss": 0.3307, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 1.8765403121699332e-05, |
|
"loss": 0.3652, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"learning_rate": 1.8706724562844737e-05, |
|
"loss": 0.282, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 7.97, |
|
"learning_rate": 1.8648046003990143e-05, |
|
"loss": 0.2981, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 1.8589367445135548e-05, |
|
"loss": 0.3432, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 1.8530688886280953e-05, |
|
"loss": 0.3466, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.9240396039603961, |
|
"eval_loss": 0.270953506231308, |
|
"eval_runtime": 209.3532, |
|
"eval_samples_per_second": 120.61, |
|
"eval_steps_per_second": 5.03, |
|
"step": 6314 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 1.847201032742636e-05, |
|
"loss": 0.292, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 1.8413331768571764e-05, |
|
"loss": 0.301, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 1.8354653209717172e-05, |
|
"loss": 0.3619, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 1.8295974650862578e-05, |
|
"loss": 0.2991, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 1.823729609200798e-05, |
|
"loss": 0.3055, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 1.8178617533153385e-05, |
|
"loss": 0.2974, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"learning_rate": 1.8119938974298793e-05, |
|
"loss": 0.3067, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"learning_rate": 1.80612604154442e-05, |
|
"loss": 0.3132, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 1.8002581856589604e-05, |
|
"loss": 0.3126, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 1.794390329773501e-05, |
|
"loss": 0.3019, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 1.788522473888041e-05, |
|
"loss": 0.2615, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 8.15, |
|
"learning_rate": 1.782654618002582e-05, |
|
"loss": 0.3004, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 1.7767867621171225e-05, |
|
"loss": 0.3107, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"learning_rate": 1.770918906231663e-05, |
|
"loss": 0.3638, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 1.7650510503462036e-05, |
|
"loss": 0.2754, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 1.759183194460744e-05, |
|
"loss": 0.3072, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"learning_rate": 1.7533153385752846e-05, |
|
"loss": 0.294, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 8.22, |
|
"learning_rate": 1.747447482689825e-05, |
|
"loss": 0.2757, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 1.7415796268043657e-05, |
|
"loss": 0.2728, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 1.7357117709189062e-05, |
|
"loss": 0.2816, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 1.729843915033447e-05, |
|
"loss": 0.3329, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 1.7239760591479876e-05, |
|
"loss": 0.3035, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 1.7181082032625278e-05, |
|
"loss": 0.2773, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 1.7122403473770683e-05, |
|
"loss": 0.2845, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 1.7063724914916092e-05, |
|
"loss": 0.2949, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 1.7005046356061497e-05, |
|
"loss": 0.3024, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 1.6946367797206902e-05, |
|
"loss": 0.3123, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 1.6887689238352304e-05, |
|
"loss": 0.311, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 1.6829010679497713e-05, |
|
"loss": 0.2862, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 1.6770332120643118e-05, |
|
"loss": 0.261, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 1.6711653561788523e-05, |
|
"loss": 0.2658, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 1.665297500293393e-05, |
|
"loss": 0.3366, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 1.6594296444079334e-05, |
|
"loss": 0.3008, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 1.653561788522474e-05, |
|
"loss": 0.293, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"learning_rate": 1.6476939326370144e-05, |
|
"loss": 0.2639, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 8.45, |
|
"learning_rate": 1.641826076751555e-05, |
|
"loss": 0.2626, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 1.6359582208660955e-05, |
|
"loss": 0.2745, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 1.6300903649806364e-05, |
|
"loss": 0.3025, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 8.49, |
|
"learning_rate": 1.624222509095177e-05, |
|
"loss": 0.2824, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 1.618354653209717e-05, |
|
"loss": 0.2928, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 1.6124867973242576e-05, |
|
"loss": 0.3257, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 1.6066189414387985e-05, |
|
"loss": 0.2867, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 1.600751085553339e-05, |
|
"loss": 0.341, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 8.55, |
|
"learning_rate": 1.5948832296678795e-05, |
|
"loss": 0.2635, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 1.58901537378242e-05, |
|
"loss": 0.2683, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 8.58, |
|
"learning_rate": 1.5831475178969606e-05, |
|
"loss": 0.2594, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 8.59, |
|
"learning_rate": 1.577279662011501e-05, |
|
"loss": 0.2814, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 1.5714118061260416e-05, |
|
"loss": 0.3668, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 1.565543950240582e-05, |
|
"loss": 0.2786, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 1.5596760943551227e-05, |
|
"loss": 0.3057, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 1.5538082384696635e-05, |
|
"loss": 0.3034, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 1.5479403825842037e-05, |
|
"loss": 0.2695, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"learning_rate": 1.5420725266987443e-05, |
|
"loss": 0.2943, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 1.5362046708132848e-05, |
|
"loss": 0.3393, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 1.5303368149278257e-05, |
|
"loss": 0.2923, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"learning_rate": 1.524468959042366e-05, |
|
"loss": 0.298, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 1.5186011031569064e-05, |
|
"loss": 0.3213, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 8.73, |
|
"learning_rate": 1.512733247271447e-05, |
|
"loss": 0.3212, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 1.5068653913859876e-05, |
|
"loss": 0.2647, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 8.76, |
|
"learning_rate": 1.5009975355005281e-05, |
|
"loss": 0.2903, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 8.77, |
|
"learning_rate": 1.4951296796150688e-05, |
|
"loss": 0.2543, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 1.4892618237296093e-05, |
|
"loss": 0.2796, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 1.4833939678441497e-05, |
|
"loss": 0.3065, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 8.81, |
|
"learning_rate": 1.4775261119586902e-05, |
|
"loss": 0.3442, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 1.471658256073231e-05, |
|
"loss": 0.2613, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 1.4657904001877714e-05, |
|
"loss": 0.2693, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 1.4599225443023121e-05, |
|
"loss": 0.2915, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 1.4540546884168527e-05, |
|
"loss": 0.2919, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 1.448186832531393e-05, |
|
"loss": 0.2804, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 1.4423189766459336e-05, |
|
"loss": 0.2705, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 1.436451120760474e-05, |
|
"loss": 0.3024, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 1.4305832648750148e-05, |
|
"loss": 0.3135, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"learning_rate": 1.4247154089895553e-05, |
|
"loss": 0.3437, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 1.418847553104096e-05, |
|
"loss": 0.3407, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"learning_rate": 1.4129796972186362e-05, |
|
"loss": 0.3183, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"learning_rate": 1.4071118413331769e-05, |
|
"loss": 0.3212, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 1.4012439854477174e-05, |
|
"loss": 0.33, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 8.98, |
|
"learning_rate": 1.3953761295622581e-05, |
|
"loss": 0.3157, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 1.3895082736767986e-05, |
|
"loss": 0.3666, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_accuracy": 0.9274851485148515, |
|
"eval_loss": 0.26253506541252136, |
|
"eval_runtime": 207.4896, |
|
"eval_samples_per_second": 121.693, |
|
"eval_steps_per_second": 5.075, |
|
"step": 7103 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 1.3836404177913392e-05, |
|
"loss": 0.2964, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 1.3777725619058795e-05, |
|
"loss": 0.2888, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 1.3719047060204202e-05, |
|
"loss": 0.2829, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 9.05, |
|
"learning_rate": 1.3660368501349607e-05, |
|
"loss": 0.3137, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 1.3601689942495013e-05, |
|
"loss": 0.3306, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 1.354301138364042e-05, |
|
"loss": 0.2745, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 9.08, |
|
"learning_rate": 1.3484332824785825e-05, |
|
"loss": 0.2985, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 1.3425654265931229e-05, |
|
"loss": 0.2479, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 1.3366975707076634e-05, |
|
"loss": 0.2755, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 1.330829714822204e-05, |
|
"loss": 0.2782, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 1.3249618589367446e-05, |
|
"loss": 0.3187, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 1.3190940030512853e-05, |
|
"loss": 0.2405, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 9.16, |
|
"learning_rate": 1.3132261471658255e-05, |
|
"loss": 0.2707, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 1.3073582912803662e-05, |
|
"loss": 0.2439, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"learning_rate": 1.3014904353949067e-05, |
|
"loss": 0.2318, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 1.2956225795094474e-05, |
|
"loss": 0.2652, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 1.289754723623988e-05, |
|
"loss": 0.2634, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 9.22, |
|
"learning_rate": 1.2838868677385285e-05, |
|
"loss": 0.2985, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 1.2780190118530688e-05, |
|
"loss": 0.2556, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 1.2721511559676093e-05, |
|
"loss": 0.3017, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 1.26628330008215e-05, |
|
"loss": 0.2944, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 1.2604154441966906e-05, |
|
"loss": 0.2883, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"learning_rate": 1.2545475883112313e-05, |
|
"loss": 0.3068, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 1.2486797324257716e-05, |
|
"loss": 0.2383, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 1.2428118765403123e-05, |
|
"loss": 0.2251, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 1.2369440206548527e-05, |
|
"loss": 0.2213, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 1.2310761647693934e-05, |
|
"loss": 0.2663, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 9.35, |
|
"learning_rate": 1.2252083088839339e-05, |
|
"loss": 0.2656, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 1.2193404529984744e-05, |
|
"loss": 0.2949, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 1.213472597113015e-05, |
|
"loss": 0.2768, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"learning_rate": 1.2076047412275555e-05, |
|
"loss": 0.2462, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 1.201736885342096e-05, |
|
"loss": 0.2908, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 1.1958690294566365e-05, |
|
"loss": 0.2452, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 1.1900011735711772e-05, |
|
"loss": 0.3567, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 1.1841333176857176e-05, |
|
"loss": 0.2481, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 9.45, |
|
"learning_rate": 1.1782654618002583e-05, |
|
"loss": 0.2726, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 1.1723976059147988e-05, |
|
"loss": 0.2593, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 9.48, |
|
"learning_rate": 1.1665297500293393e-05, |
|
"loss": 0.2714, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 9.49, |
|
"learning_rate": 1.1606618941438799e-05, |
|
"loss": 0.2607, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 1.1547940382584206e-05, |
|
"loss": 0.2482, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 1.1489261823729609e-05, |
|
"loss": 0.2851, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 1.1430583264875014e-05, |
|
"loss": 0.2778, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 9.54, |
|
"learning_rate": 1.1371904706020421e-05, |
|
"loss": 0.2334, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 1.1313226147165825e-05, |
|
"loss": 0.3403, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 1.1254547588311232e-05, |
|
"loss": 0.2872, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 1.1195869029456637e-05, |
|
"loss": 0.2489, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"learning_rate": 1.1137190470602042e-05, |
|
"loss": 0.2794, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 1.1078511911747448e-05, |
|
"loss": 0.2952, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 1.1019833352892855e-05, |
|
"loss": 0.2569, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 1.0961154794038258e-05, |
|
"loss": 0.2513, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 1.0902476235183665e-05, |
|
"loss": 0.2315, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 9.65, |
|
"learning_rate": 1.084379767632907e-05, |
|
"loss": 0.3262, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 9.67, |
|
"learning_rate": 1.0785119117474476e-05, |
|
"loss": 0.2678, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 1.0726440558619881e-05, |
|
"loss": 0.2782, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 1.0667761999765286e-05, |
|
"loss": 0.2481, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 1.0609083440910692e-05, |
|
"loss": 0.2002, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 1.0550404882056097e-05, |
|
"loss": 0.2472, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 1.0491726323201502e-05, |
|
"loss": 0.2776, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 1.0433047764346907e-05, |
|
"loss": 0.2371, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 9.76, |
|
"learning_rate": 1.0374369205492314e-05, |
|
"loss": 0.2373, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"learning_rate": 1.0315690646637718e-05, |
|
"loss": 0.2826, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 1.0257012087783125e-05, |
|
"loss": 0.2087, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 9.79, |
|
"learning_rate": 1.019833352892853e-05, |
|
"loss": 0.2847, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 1.0139654970073935e-05, |
|
"loss": 0.2585, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"learning_rate": 1.008097641121934e-05, |
|
"loss": 0.2737, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 9.83, |
|
"learning_rate": 1.0022297852364746e-05, |
|
"loss": 0.3026, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 9.963619293510151e-06, |
|
"loss": 0.2519, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 9.904940734655556e-06, |
|
"loss": 0.26, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 9.846262175800963e-06, |
|
"loss": 0.2244, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 9.787583616946367e-06, |
|
"loss": 0.2982, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 9.728905058091774e-06, |
|
"loss": 0.3107, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 9.91, |
|
"learning_rate": 9.67022649923718e-06, |
|
"loss": 0.2674, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 9.92, |
|
"learning_rate": 9.611547940382585e-06, |
|
"loss": 0.2689, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 9.55286938152799e-06, |
|
"loss": 0.2581, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 9.494190822673397e-06, |
|
"loss": 0.251, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 9.4355122638188e-06, |
|
"loss": 0.2765, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 9.376833704964207e-06, |
|
"loss": 0.3042, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 9.98, |
|
"learning_rate": 9.318155146109613e-06, |
|
"loss": 0.2992, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 9.259476587255018e-06, |
|
"loss": 0.2085, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_accuracy": 0.9295445544554456, |
|
"eval_loss": 0.2578248977661133, |
|
"eval_runtime": 209.3581, |
|
"eval_samples_per_second": 120.607, |
|
"eval_steps_per_second": 5.03, |
|
"step": 7892 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 9.200798028400423e-06, |
|
"loss": 0.2411, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 9.142119469545828e-06, |
|
"loss": 0.258, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 9.083440910691234e-06, |
|
"loss": 0.2462, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 10.05, |
|
"learning_rate": 9.024762351836639e-06, |
|
"loss": 0.2661, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 10.06, |
|
"learning_rate": 8.966083792982046e-06, |
|
"loss": 0.2645, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 8.90740523412745e-06, |
|
"loss": 0.2493, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 10.09, |
|
"learning_rate": 8.848726675272856e-06, |
|
"loss": 0.2406, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 10.1, |
|
"learning_rate": 8.790048116418262e-06, |
|
"loss": 0.2708, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 10.11, |
|
"learning_rate": 8.731369557563667e-06, |
|
"loss": 0.2987, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 10.12, |
|
"learning_rate": 8.672690998709072e-06, |
|
"loss": 0.3072, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 10.14, |
|
"learning_rate": 8.614012439854477e-06, |
|
"loss": 0.2898, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 10.15, |
|
"learning_rate": 8.555333880999883e-06, |
|
"loss": 0.237, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 10.16, |
|
"learning_rate": 8.496655322145288e-06, |
|
"loss": 0.234, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 10.17, |
|
"learning_rate": 8.437976763290693e-06, |
|
"loss": 0.2357, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 10.19, |
|
"learning_rate": 8.379298204436099e-06, |
|
"loss": 0.2409, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 10.2, |
|
"learning_rate": 8.320619645581505e-06, |
|
"loss": 0.2736, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 10.21, |
|
"learning_rate": 8.261941086726909e-06, |
|
"loss": 0.1918, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 10.22, |
|
"learning_rate": 8.203262527872316e-06, |
|
"loss": 0.2389, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 10.24, |
|
"learning_rate": 8.144583969017721e-06, |
|
"loss": 0.264, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 10.25, |
|
"learning_rate": 8.085905410163127e-06, |
|
"loss": 0.2499, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"learning_rate": 8.027226851308532e-06, |
|
"loss": 0.2617, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 10.28, |
|
"learning_rate": 7.968548292453939e-06, |
|
"loss": 0.3009, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 7.909869733599342e-06, |
|
"loss": 0.2762, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 10.3, |
|
"learning_rate": 7.85119117474475e-06, |
|
"loss": 0.2551, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 10.31, |
|
"learning_rate": 7.792512615890155e-06, |
|
"loss": 0.2181, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 10.33, |
|
"learning_rate": 7.73383405703556e-06, |
|
"loss": 0.2627, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 10.34, |
|
"learning_rate": 7.675155498180965e-06, |
|
"loss": 0.2635, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 10.35, |
|
"learning_rate": 7.616476939326371e-06, |
|
"loss": 0.217, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 10.36, |
|
"learning_rate": 7.557798380471776e-06, |
|
"loss": 0.2041, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 10.38, |
|
"learning_rate": 7.499119821617182e-06, |
|
"loss": 0.2333, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 10.39, |
|
"learning_rate": 7.440441262762587e-06, |
|
"loss": 0.2979, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 10.4, |
|
"learning_rate": 7.381762703907992e-06, |
|
"loss": 0.2427, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 10.41, |
|
"learning_rate": 7.323084145053398e-06, |
|
"loss": 0.2956, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 10.43, |
|
"learning_rate": 7.264405586198804e-06, |
|
"loss": 0.2991, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 10.44, |
|
"learning_rate": 7.205727027344208e-06, |
|
"loss": 0.2213, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 10.45, |
|
"learning_rate": 7.147048468489614e-06, |
|
"loss": 0.2513, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 10.47, |
|
"learning_rate": 7.08836990963502e-06, |
|
"loss": 0.28, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 7.029691350780425e-06, |
|
"loss": 0.2634, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 10.49, |
|
"learning_rate": 6.971012791925831e-06, |
|
"loss": 0.2564, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 6.912334233071237e-06, |
|
"loss": 0.2362, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 10.52, |
|
"learning_rate": 6.8536556742166414e-06, |
|
"loss": 0.2572, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 10.53, |
|
"learning_rate": 6.7949771153620476e-06, |
|
"loss": 0.2329, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 10.54, |
|
"learning_rate": 6.736298556507452e-06, |
|
"loss": 0.3466, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 10.55, |
|
"learning_rate": 6.677619997652858e-06, |
|
"loss": 0.2416, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"learning_rate": 6.618941438798263e-06, |
|
"loss": 0.2992, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 10.58, |
|
"learning_rate": 6.560262879943669e-06, |
|
"loss": 0.2459, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 10.59, |
|
"learning_rate": 6.501584321089074e-06, |
|
"loss": 0.3164, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 10.61, |
|
"learning_rate": 6.44290576223448e-06, |
|
"loss": 0.2305, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 10.62, |
|
"learning_rate": 6.3842272033798844e-06, |
|
"loss": 0.2364, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 10.63, |
|
"learning_rate": 6.3255486445252906e-06, |
|
"loss": 0.2439, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 10.64, |
|
"learning_rate": 6.266870085670697e-06, |
|
"loss": 0.2197, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 6.208191526816102e-06, |
|
"loss": 0.2131, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 10.67, |
|
"learning_rate": 6.149512967961507e-06, |
|
"loss": 0.2498, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 10.68, |
|
"learning_rate": 6.0908344091069125e-06, |
|
"loss": 0.2443, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 10.69, |
|
"learning_rate": 6.032155850252319e-06, |
|
"loss": 0.2288, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"learning_rate": 5.973477291397724e-06, |
|
"loss": 0.2426, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 10.72, |
|
"learning_rate": 5.914798732543129e-06, |
|
"loss": 0.2364, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 10.73, |
|
"learning_rate": 5.856120173688534e-06, |
|
"loss": 0.2468, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 10.74, |
|
"learning_rate": 5.79744161483394e-06, |
|
"loss": 0.2676, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 10.76, |
|
"learning_rate": 5.738763055979345e-06, |
|
"loss": 0.2589, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"learning_rate": 5.680084497124751e-06, |
|
"loss": 0.239, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 10.78, |
|
"learning_rate": 5.621405938270156e-06, |
|
"loss": 0.2736, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 10.8, |
|
"learning_rate": 5.562727379415562e-06, |
|
"loss": 0.2545, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 10.81, |
|
"learning_rate": 5.504048820560968e-06, |
|
"loss": 0.2182, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 10.82, |
|
"learning_rate": 5.445370261706373e-06, |
|
"loss": 0.2499, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 10.83, |
|
"learning_rate": 5.386691702851778e-06, |
|
"loss": 0.2321, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 10.85, |
|
"learning_rate": 5.3280131439971835e-06, |
|
"loss": 0.2763, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 10.86, |
|
"learning_rate": 5.26933458514259e-06, |
|
"loss": 0.278, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 10.87, |
|
"learning_rate": 5.210656026287995e-06, |
|
"loss": 0.2391, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 10.88, |
|
"learning_rate": 5.1519774674334e-06, |
|
"loss": 0.2454, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 10.9, |
|
"learning_rate": 5.093298908578805e-06, |
|
"loss": 0.2306, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 10.91, |
|
"learning_rate": 5.034620349724211e-06, |
|
"loss": 0.2483, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 10.92, |
|
"learning_rate": 4.975941790869616e-06, |
|
"loss": 0.2101, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 10.93, |
|
"learning_rate": 4.917263232015022e-06, |
|
"loss": 0.2836, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 10.95, |
|
"learning_rate": 4.858584673160427e-06, |
|
"loss": 0.228, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 10.96, |
|
"learning_rate": 4.799906114305833e-06, |
|
"loss": 0.2362, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 4.741227555451239e-06, |
|
"loss": 0.2404, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 10.99, |
|
"learning_rate": 4.682548996596644e-06, |
|
"loss": 0.2467, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 4.623870437742049e-06, |
|
"loss": 0.263, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_accuracy": 0.9301782178217822, |
|
"eval_loss": 0.25631117820739746, |
|
"eval_runtime": 207.4622, |
|
"eval_samples_per_second": 121.709, |
|
"eval_steps_per_second": 5.076, |
|
"step": 8681 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 4.565191878887455e-06, |
|
"loss": 0.2362, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 4.506513320032861e-06, |
|
"loss": 0.2589, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.447834761178266e-06, |
|
"loss": 0.243, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 11.05, |
|
"learning_rate": 4.389156202323671e-06, |
|
"loss": 0.2477, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"learning_rate": 4.3304776434690765e-06, |
|
"loss": 0.2724, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 11.07, |
|
"learning_rate": 4.271799084614482e-06, |
|
"loss": 0.2871, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 11.09, |
|
"learning_rate": 4.213120525759887e-06, |
|
"loss": 0.2383, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 11.1, |
|
"learning_rate": 4.154441966905293e-06, |
|
"loss": 0.2428, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 11.11, |
|
"learning_rate": 4.095763408050698e-06, |
|
"loss": 0.304, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 11.12, |
|
"learning_rate": 4.037084849196104e-06, |
|
"loss": 0.2288, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 11.14, |
|
"learning_rate": 3.97840629034151e-06, |
|
"loss": 0.2147, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 3.919727731486915e-06, |
|
"loss": 0.2121, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 11.16, |
|
"learning_rate": 3.86104917263232e-06, |
|
"loss": 0.2593, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 11.18, |
|
"learning_rate": 3.802370613777726e-06, |
|
"loss": 0.2018, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 11.19, |
|
"learning_rate": 3.7436920549231313e-06, |
|
"loss": 0.2225, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 11.2, |
|
"learning_rate": 3.6850134960685365e-06, |
|
"loss": 0.2875, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 11.21, |
|
"learning_rate": 3.6263349372139426e-06, |
|
"loss": 0.274, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 11.23, |
|
"learning_rate": 3.567656378359348e-06, |
|
"loss": 0.1874, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 11.24, |
|
"learning_rate": 3.508977819504753e-06, |
|
"loss": 0.2055, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 3.4502992606501584e-06, |
|
"loss": 0.2593, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 11.26, |
|
"learning_rate": 3.391620701795564e-06, |
|
"loss": 0.2235, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 11.28, |
|
"learning_rate": 3.3329421429409694e-06, |
|
"loss": 0.2122, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 11.29, |
|
"learning_rate": 3.2742635840863747e-06, |
|
"loss": 0.2728, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 11.3, |
|
"learning_rate": 3.2155850252317808e-06, |
|
"loss": 0.2228, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 11.31, |
|
"learning_rate": 3.156906466377186e-06, |
|
"loss": 0.2213, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 11.33, |
|
"learning_rate": 3.0982279075225913e-06, |
|
"loss": 0.2376, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 11.34, |
|
"learning_rate": 3.0395493486679966e-06, |
|
"loss": 0.2463, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 11.35, |
|
"learning_rate": 2.9808707898134023e-06, |
|
"loss": 0.2346, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 11.37, |
|
"learning_rate": 2.9221922309588076e-06, |
|
"loss": 0.2405, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 11.38, |
|
"learning_rate": 2.8635136721042132e-06, |
|
"loss": 0.2342, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 11.39, |
|
"learning_rate": 2.804835113249619e-06, |
|
"loss": 0.2382, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 11.4, |
|
"learning_rate": 2.746156554395024e-06, |
|
"loss": 0.2257, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 11.42, |
|
"learning_rate": 2.6874779955404295e-06, |
|
"loss": 0.2055, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 11.43, |
|
"learning_rate": 2.628799436685835e-06, |
|
"loss": 0.2524, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 11.44, |
|
"learning_rate": 2.5701208778312404e-06, |
|
"loss": 0.2421, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 11.45, |
|
"learning_rate": 2.511442318976646e-06, |
|
"loss": 0.1833, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 11.47, |
|
"learning_rate": 2.4527637601220514e-06, |
|
"loss": 0.2737, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 11.48, |
|
"learning_rate": 2.394085201267457e-06, |
|
"loss": 0.2364, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 11.49, |
|
"learning_rate": 2.3354066424128624e-06, |
|
"loss": 0.2459, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 11.5, |
|
"learning_rate": 2.2767280835582676e-06, |
|
"loss": 0.2531, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 11.52, |
|
"learning_rate": 2.2180495247036733e-06, |
|
"loss": 0.2078, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 11.53, |
|
"learning_rate": 2.159370965849079e-06, |
|
"loss": 0.2672, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 11.54, |
|
"learning_rate": 2.1006924069944843e-06, |
|
"loss": 0.2578, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 11.56, |
|
"learning_rate": 2.04201384813989e-06, |
|
"loss": 0.2364, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 11.57, |
|
"learning_rate": 1.9833352892852952e-06, |
|
"loss": 0.2721, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 11.58, |
|
"learning_rate": 1.9246567304307005e-06, |
|
"loss": 0.2355, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 11.59, |
|
"learning_rate": 1.8659781715761062e-06, |
|
"loss": 0.2513, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 11.61, |
|
"learning_rate": 1.8072996127215115e-06, |
|
"loss": 0.2187, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 11.62, |
|
"learning_rate": 1.7486210538669172e-06, |
|
"loss": 0.2145, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 11.63, |
|
"learning_rate": 1.6899424950123226e-06, |
|
"loss": 0.2395, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 11.64, |
|
"learning_rate": 1.631263936157728e-06, |
|
"loss": 0.2992, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 11.66, |
|
"learning_rate": 1.5725853773031336e-06, |
|
"loss": 0.2019, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 11.67, |
|
"learning_rate": 1.513906818448539e-06, |
|
"loss": 0.2465, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 11.68, |
|
"learning_rate": 1.4552282595939443e-06, |
|
"loss": 0.2272, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 11.69, |
|
"learning_rate": 1.3965497007393498e-06, |
|
"loss": 0.2824, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 11.71, |
|
"learning_rate": 1.3378711418847555e-06, |
|
"loss": 0.3012, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 11.72, |
|
"learning_rate": 1.2791925830301608e-06, |
|
"loss": 0.2647, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 11.73, |
|
"learning_rate": 1.2205140241755663e-06, |
|
"loss": 0.2813, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 11.75, |
|
"learning_rate": 1.1618354653209717e-06, |
|
"loss": 0.2313, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 11.76, |
|
"learning_rate": 1.1031569064663772e-06, |
|
"loss": 0.216, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 11.77, |
|
"learning_rate": 1.0444783476117827e-06, |
|
"loss": 0.263, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 11.78, |
|
"learning_rate": 9.857997887571882e-07, |
|
"loss": 0.2233, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 11.8, |
|
"learning_rate": 9.271212299025936e-07, |
|
"loss": 0.231, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 11.81, |
|
"learning_rate": 8.684426710479991e-07, |
|
"loss": 0.2347, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 11.82, |
|
"learning_rate": 8.097641121934046e-07, |
|
"loss": 0.2086, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 11.83, |
|
"learning_rate": 7.5108555333881e-07, |
|
"loss": 0.2396, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 11.85, |
|
"learning_rate": 6.924069944842155e-07, |
|
"loss": 0.2432, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 11.86, |
|
"learning_rate": 6.33728435629621e-07, |
|
"loss": 0.2528, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 11.87, |
|
"learning_rate": 5.750498767750264e-07, |
|
"loss": 0.2686, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 11.88, |
|
"learning_rate": 5.163713179204318e-07, |
|
"loss": 0.2161, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 11.9, |
|
"learning_rate": 4.5769275906583735e-07, |
|
"loss": 0.2449, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 11.91, |
|
"learning_rate": 3.990142002112428e-07, |
|
"loss": 0.2246, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 11.92, |
|
"learning_rate": 3.403356413566483e-07, |
|
"loss": 0.2258, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 11.94, |
|
"learning_rate": 2.816570825020538e-07, |
|
"loss": 0.2467, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 11.95, |
|
"learning_rate": 2.2297852364745924e-07, |
|
"loss": 0.2385, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 11.96, |
|
"learning_rate": 1.642999647928647e-07, |
|
"loss": 0.2556, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 11.97, |
|
"learning_rate": 1.0562140593827016e-07, |
|
"loss": 0.2383, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 11.99, |
|
"learning_rate": 4.6942847083675625e-08, |
|
"loss": 0.2171, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.9312475247524753, |
|
"eval_loss": 0.252435564994812, |
|
"eval_runtime": 211.7134, |
|
"eval_samples_per_second": 119.265, |
|
"eval_steps_per_second": 4.974, |
|
"step": 9468 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"step": 9468, |
|
"total_flos": 7.1278240936914076e+19, |
|
"train_loss": 0.5668512925499343, |
|
"train_runtime": 16210.628, |
|
"train_samples_per_second": 56.074, |
|
"train_steps_per_second": 0.584 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 9468, |
|
"num_train_epochs": 12, |
|
"save_steps": 500, |
|
"total_flos": 7.1278240936914076e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|