|
[ |
|
{ |
|
"loss": 2.1632102966308593, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"epoch": 0.2127659574468085, |
|
"total_flos": 4181675520000, |
|
"step": 10 |
|
}, |
|
{ |
|
"loss": 1.9698543548583984, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"epoch": 0.425531914893617, |
|
"total_flos": 8363351040000, |
|
"step": 20 |
|
}, |
|
{ |
|
"loss": 1.5943157196044921, |
|
"learning_rate": 3e-06, |
|
"epoch": 0.6382978723404256, |
|
"total_flos": 12545026560000, |
|
"step": 30 |
|
}, |
|
{ |
|
"loss": 1.1324241638183594, |
|
"learning_rate": 4.000000000000001e-06, |
|
"epoch": 0.851063829787234, |
|
"total_flos": 16726702080000, |
|
"step": 40 |
|
}, |
|
{ |
|
"loss": 0.7799140930175781, |
|
"learning_rate": 5e-06, |
|
"epoch": 1.0638297872340425, |
|
"total_flos": 20620887408000, |
|
"step": 50 |
|
}, |
|
{ |
|
"loss": 0.5984169006347656, |
|
"learning_rate": 6e-06, |
|
"epoch": 1.2765957446808511, |
|
"total_flos": 24802562928000, |
|
"step": 60 |
|
}, |
|
{ |
|
"loss": 0.4769012451171875, |
|
"learning_rate": 7.000000000000001e-06, |
|
"epoch": 1.4893617021276595, |
|
"total_flos": 28984238448000, |
|
"step": 70 |
|
}, |
|
{ |
|
"loss": 0.3644599914550781, |
|
"learning_rate": 8.000000000000001e-06, |
|
"epoch": 1.702127659574468, |
|
"total_flos": 33165913968000, |
|
"step": 80 |
|
}, |
|
{ |
|
"loss": 0.29639739990234376, |
|
"learning_rate": 9e-06, |
|
"epoch": 1.9148936170212765, |
|
"total_flos": 37347589488000, |
|
"step": 90 |
|
}, |
|
{ |
|
"loss": 0.26825637817382814, |
|
"learning_rate": 1e-05, |
|
"epoch": 2.127659574468085, |
|
"total_flos": 41241774816000, |
|
"step": 100 |
|
}, |
|
{ |
|
"loss": 0.19923171997070313, |
|
"learning_rate": 1.1000000000000001e-05, |
|
"epoch": 2.3404255319148937, |
|
"total_flos": 45423450336000, |
|
"step": 110 |
|
}, |
|
{ |
|
"loss": 0.19310073852539061, |
|
"learning_rate": 1.2e-05, |
|
"epoch": 2.5531914893617023, |
|
"total_flos": 49605125856000, |
|
"step": 120 |
|
}, |
|
{ |
|
"loss": 0.15804977416992189, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"epoch": 2.7659574468085104, |
|
"total_flos": 53786801376000, |
|
"step": 130 |
|
}, |
|
{ |
|
"loss": 0.17403030395507812, |
|
"learning_rate": 1.4000000000000001e-05, |
|
"epoch": 2.978723404255319, |
|
"total_flos": 57968476896000, |
|
"step": 140 |
|
}, |
|
{ |
|
"loss": 0.11231231689453125, |
|
"learning_rate": 1.5e-05, |
|
"epoch": 3.1914893617021276, |
|
"total_flos": 61862662224000, |
|
"step": 150 |
|
}, |
|
{ |
|
"loss": 0.0972991943359375, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"epoch": 3.404255319148936, |
|
"total_flos": 66044337744000, |
|
"step": 160 |
|
}, |
|
{ |
|
"loss": 0.09285507202148438, |
|
"learning_rate": 1.7000000000000003e-05, |
|
"epoch": 3.617021276595745, |
|
"total_flos": 70226013264000, |
|
"step": 170 |
|
}, |
|
{ |
|
"loss": 0.10490646362304687, |
|
"learning_rate": 1.8e-05, |
|
"epoch": 3.829787234042553, |
|
"total_flos": 74407688784000, |
|
"step": 180 |
|
}, |
|
{ |
|
"loss": 0.08623809814453125, |
|
"learning_rate": 1.9e-05, |
|
"epoch": 4.042553191489362, |
|
"total_flos": 78301874112000, |
|
"step": 190 |
|
}, |
|
{ |
|
"loss": 0.05904541015625, |
|
"learning_rate": 2e-05, |
|
"epoch": 4.25531914893617, |
|
"total_flos": 82483549632000, |
|
"step": 200 |
|
}, |
|
{ |
|
"loss": 0.052008056640625, |
|
"learning_rate": 2.1e-05, |
|
"epoch": 4.468085106382979, |
|
"total_flos": 86665225152000, |
|
"step": 210 |
|
}, |
|
{ |
|
"loss": 0.06222076416015625, |
|
"learning_rate": 2.2000000000000003e-05, |
|
"epoch": 4.680851063829787, |
|
"total_flos": 90846900672000, |
|
"step": 220 |
|
}, |
|
{ |
|
"loss": 0.05680923461914063, |
|
"learning_rate": 2.3000000000000003e-05, |
|
"epoch": 4.8936170212765955, |
|
"total_flos": 95028576192000, |
|
"step": 230 |
|
}, |
|
{ |
|
"eval_loss": 0.0899211324525602, |
|
"eval_accuracy_score": 0.974121282348397, |
|
"eval_precision": 0.9776031434184675, |
|
"eval_recall": 0.9749216300940439, |
|
"eval_f1": 0.9762605454188739, |
|
"epoch": 5.0, |
|
"total_flos": 96831923760000, |
|
"step": 235 |
|
} |
|
] |