{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.998691442030882, "eval_steps": 500, "global_step": 477, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.010468463752944255, "grad_norm": 12.869433218842474, "learning_rate": 5.208333333333333e-08, "logits/chosen": -2.7303454875946045, "logits/rejected": -2.6748926639556885, "logps/chosen": -1.0383304357528687, "logps/rejected": -1.093923807144165, "loss": 2.0262, "rewards/accuracies": 0.518750011920929, "rewards/chosen": -1.0383304357528687, "rewards/margins": 0.05559344217181206, "rewards/rejected": -1.093923807144165, "step": 5 }, { "epoch": 0.02093692750588851, "grad_norm": 12.963970831519292, "learning_rate": 1.0416666666666667e-07, "logits/chosen": -2.8506417274475098, "logits/rejected": -2.7654807567596436, "logps/chosen": -1.025138020515442, "logps/rejected": -1.1677943468093872, "loss": 2.0292, "rewards/accuracies": 0.581250011920929, "rewards/chosen": -1.025138020515442, "rewards/margins": 0.1426563411951065, "rewards/rejected": -1.1677943468093872, "step": 10 }, { "epoch": 0.031405391258832765, "grad_norm": 15.099715830730236, "learning_rate": 1.5624999999999999e-07, "logits/chosen": -2.77595853805542, "logits/rejected": -2.7310924530029297, "logps/chosen": -1.0620777606964111, "logps/rejected": -1.1130949258804321, "loss": 2.0243, "rewards/accuracies": 0.5375000238418579, "rewards/chosen": -1.0620777606964111, "rewards/margins": 0.051017045974731445, "rewards/rejected": -1.1130949258804321, "step": 15 }, { "epoch": 0.04187385501177702, "grad_norm": 14.125874639614404, "learning_rate": 2.0833333333333333e-07, "logits/chosen": -2.798926830291748, "logits/rejected": -2.6963000297546387, "logps/chosen": -1.026979684829712, "logps/rejected": -1.1289321184158325, "loss": 2.0224, "rewards/accuracies": 0.5562499761581421, "rewards/chosen": -1.026979684829712, "rewards/margins": 0.1019524484872818, "rewards/rejected": -1.1289321184158325, "step": 20 }, { "epoch": 0.05234231876472128, "grad_norm": 12.739023934080043, "learning_rate": 2.604166666666667e-07, "logits/chosen": -2.797764301300049, "logits/rejected": -2.6895253658294678, "logps/chosen": -0.9821299314498901, "logps/rejected": -1.1730941534042358, "loss": 1.9952, "rewards/accuracies": 0.5687500238418579, "rewards/chosen": -0.9821299314498901, "rewards/margins": 0.19096432626247406, "rewards/rejected": -1.1730941534042358, "step": 25 }, { "epoch": 0.06281078251766553, "grad_norm": 18.287014857226495, "learning_rate": 3.1249999999999997e-07, "logits/chosen": -2.7723119258880615, "logits/rejected": -2.7337851524353027, "logps/chosen": -1.0000275373458862, "logps/rejected": -1.0346484184265137, "loss": 1.9952, "rewards/accuracies": 0.5625, "rewards/chosen": -1.0000275373458862, "rewards/margins": 0.034620799124240875, "rewards/rejected": -1.0346484184265137, "step": 30 }, { "epoch": 0.07327924627060979, "grad_norm": 17.8360996219737, "learning_rate": 3.645833333333333e-07, "logits/chosen": -2.7883362770080566, "logits/rejected": -2.715696334838867, "logps/chosen": -1.0128945112228394, "logps/rejected": -1.225642442703247, "loss": 1.9748, "rewards/accuracies": 0.59375, "rewards/chosen": -1.0128945112228394, "rewards/margins": 0.2127479612827301, "rewards/rejected": -1.225642442703247, "step": 35 }, { "epoch": 0.08374771002355404, "grad_norm": 18.333373463486794, "learning_rate": 4.1666666666666667e-07, "logits/chosen": -2.7703633308410645, "logits/rejected": -2.6918962001800537, "logps/chosen": -1.0058094263076782, "logps/rejected": -1.131239891052246, "loss": 1.9619, "rewards/accuracies": 0.59375, "rewards/chosen": -1.0058094263076782, "rewards/margins": 0.12543055415153503, "rewards/rejected": -1.131239891052246, "step": 40 }, { "epoch": 0.0942161737764983, "grad_norm": 15.971248778309374, "learning_rate": 4.6874999999999996e-07, "logits/chosen": -2.732913017272949, "logits/rejected": -2.7122368812561035, "logps/chosen": -0.9598785638809204, "logps/rejected": -1.1564022302627563, "loss": 1.9542, "rewards/accuracies": 0.6187499761581421, "rewards/chosen": -0.9598785638809204, "rewards/margins": 0.196523517370224, "rewards/rejected": -1.1564022302627563, "step": 45 }, { "epoch": 0.10468463752944256, "grad_norm": 23.366392169452176, "learning_rate": 4.999731868769026e-07, "logits/chosen": -2.672335624694824, "logits/rejected": -2.6771888732910156, "logps/chosen": -1.0208399295806885, "logps/rejected": -1.2945979833602905, "loss": 1.9581, "rewards/accuracies": 0.612500011920929, "rewards/chosen": -1.0208399295806885, "rewards/margins": 0.2737579941749573, "rewards/rejected": -1.2945979833602905, "step": 50 }, { "epoch": 0.11515310128238682, "grad_norm": 19.929116348471684, "learning_rate": 4.996716052911017e-07, "logits/chosen": -2.7723307609558105, "logits/rejected": -2.747865676879883, "logps/chosen": -1.0856590270996094, "logps/rejected": -1.219327449798584, "loss": 1.9648, "rewards/accuracies": 0.574999988079071, "rewards/chosen": -1.0856590270996094, "rewards/margins": 0.13366840779781342, "rewards/rejected": -1.219327449798584, "step": 55 }, { "epoch": 0.12562156503533106, "grad_norm": 23.73282923950602, "learning_rate": 4.990353313429303e-07, "logits/chosen": -2.7040231227874756, "logits/rejected": -2.6466102600097656, "logps/chosen": -1.0688912868499756, "logps/rejected": -1.260387659072876, "loss": 1.9501, "rewards/accuracies": 0.5562499761581421, "rewards/chosen": -1.0688912868499756, "rewards/margins": 0.19149653613567352, "rewards/rejected": -1.260387659072876, "step": 60 }, { "epoch": 0.1360900287882753, "grad_norm": 18.942833820850367, "learning_rate": 4.980652179769217e-07, "logits/chosen": -2.716078519821167, "logits/rejected": -2.7220449447631836, "logps/chosen": -1.027010440826416, "logps/rejected": -1.2260963916778564, "loss": 1.9536, "rewards/accuracies": 0.612500011920929, "rewards/chosen": -1.027010440826416, "rewards/margins": 0.199085995554924, "rewards/rejected": -1.2260963916778564, "step": 65 }, { "epoch": 0.14655849254121958, "grad_norm": 22.28146945050057, "learning_rate": 4.967625656594781e-07, "logits/chosen": -2.7145752906799316, "logits/rejected": -2.6811351776123047, "logps/chosen": -1.074029564857483, "logps/rejected": -1.3044393062591553, "loss": 1.9325, "rewards/accuracies": 0.574999988079071, "rewards/chosen": -1.074029564857483, "rewards/margins": 0.2304096519947052, "rewards/rejected": -1.3044393062591553, "step": 70 }, { "epoch": 0.15702695629416383, "grad_norm": 30.956539188115524, "learning_rate": 4.951291206355559e-07, "logits/chosen": -2.640056610107422, "logits/rejected": -2.5216145515441895, "logps/chosen": -1.0524249076843262, "logps/rejected": -1.358999490737915, "loss": 1.9439, "rewards/accuracies": 0.6187499761581421, "rewards/chosen": -1.0524249076843262, "rewards/margins": 0.3065745234489441, "rewards/rejected": -1.358999490737915, "step": 75 }, { "epoch": 0.16749542004710807, "grad_norm": 19.173789722918567, "learning_rate": 4.93167072587771e-07, "logits/chosen": -2.5955734252929688, "logits/rejected": -2.5555953979492188, "logps/chosen": -1.0017749071121216, "logps/rejected": -1.4020613431930542, "loss": 1.9082, "rewards/accuracies": 0.6875, "rewards/chosen": -1.0017749071121216, "rewards/margins": 0.4002866744995117, "rewards/rejected": -1.4020613431930542, "step": 80 }, { "epoch": 0.17796388380005235, "grad_norm": 26.77907962647508, "learning_rate": 4.908790517010636e-07, "logits/chosen": -2.730196475982666, "logits/rejected": -2.6316986083984375, "logps/chosen": -1.0233409404754639, "logps/rejected": -1.4025558233261108, "loss": 1.9293, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.0233409404754639, "rewards/margins": 0.3792147934436798, "rewards/rejected": -1.4025558233261108, "step": 85 }, { "epoch": 0.1884323475529966, "grad_norm": 20.94451147733862, "learning_rate": 4.882681251368548e-07, "logits/chosen": -2.7333931922912598, "logits/rejected": -2.594924211502075, "logps/chosen": -1.0183618068695068, "logps/rejected": -1.4752748012542725, "loss": 1.903, "rewards/accuracies": 0.637499988079071, "rewards/chosen": -1.0183618068695068, "rewards/margins": 0.4569129943847656, "rewards/rejected": -1.4752748012542725, "step": 90 }, { "epoch": 0.19890081130594087, "grad_norm": 27.23027264774861, "learning_rate": 4.853377929214243e-07, "logits/chosen": -2.6479899883270264, "logits/rejected": -2.612231492996216, "logps/chosen": -0.9796420335769653, "logps/rejected": -1.2683651447296143, "loss": 1.9133, "rewards/accuracies": 0.612500011920929, "rewards/chosen": -0.9796420335769653, "rewards/margins": 0.28872328996658325, "rewards/rejected": -1.2683651447296143, "step": 95 }, { "epoch": 0.2093692750588851, "grad_norm": 20.647765632695478, "learning_rate": 4.820919832540181e-07, "logits/chosen": -2.7929205894470215, "logits/rejected": -2.730536460876465, "logps/chosen": -1.0671679973602295, "logps/rejected": -1.3122754096984863, "loss": 1.9562, "rewards/accuracies": 0.5249999761581421, "rewards/chosen": -1.0671679973602295, "rewards/margins": 0.24510732293128967, "rewards/rejected": -1.3122754096984863, "step": 100 }, { "epoch": 0.21983773881182936, "grad_norm": 19.882099174165575, "learning_rate": 4.785350472409791e-07, "logits/chosen": -2.7290585041046143, "logits/rejected": -2.661957263946533, "logps/chosen": -1.1218279600143433, "logps/rejected": -1.4676848649978638, "loss": 1.8476, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.1218279600143433, "rewards/margins": 0.3458569347858429, "rewards/rejected": -1.4676848649978638, "step": 105 }, { "epoch": 0.23030620256477363, "grad_norm": 29.356563953547646, "learning_rate": 4.7467175306295647e-07, "logits/chosen": -2.677428722381592, "logits/rejected": -2.587128162384033, "logps/chosen": -1.0471330881118774, "logps/rejected": -1.3834892511367798, "loss": 1.8785, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -1.0471330881118774, "rewards/margins": 0.33635613322257996, "rewards/rejected": -1.3834892511367798, "step": 110 }, { "epoch": 0.24077466631771788, "grad_norm": 22.70599557212694, "learning_rate": 4.70507279583015e-07, "logits/chosen": -2.699206590652466, "logits/rejected": -2.6752054691314697, "logps/chosen": -1.037027359008789, "logps/rejected": -1.2504864931106567, "loss": 1.8781, "rewards/accuracies": 0.625, "rewards/chosen": -1.037027359008789, "rewards/margins": 0.21345916390419006, "rewards/rejected": -1.2504864931106567, "step": 115 }, { "epoch": 0.2512431300706621, "grad_norm": 22.718720973130083, "learning_rate": 4.6604720940421207e-07, "logits/chosen": -2.7028520107269287, "logits/rejected": -2.581420660018921, "logps/chosen": -0.958320140838623, "logps/rejected": -1.4017829895019531, "loss": 1.8894, "rewards/accuracies": 0.6312500238418579, "rewards/chosen": -0.958320140838623, "rewards/margins": 0.44346290826797485, "rewards/rejected": -1.4017829895019531, "step": 120 }, { "epoch": 0.26171159382360637, "grad_norm": 27.839637111582167, "learning_rate": 4.612975213859487e-07, "logits/chosen": -2.5790719985961914, "logits/rejected": -2.534557342529297, "logps/chosen": -1.0435621738433838, "logps/rejected": -1.4774357080459595, "loss": 1.9045, "rewards/accuracies": 0.668749988079071, "rewards/chosen": -1.0435621738433838, "rewards/margins": 0.43387335538864136, "rewards/rejected": -1.4774357080459595, "step": 125 }, { "epoch": 0.2721800575765506, "grad_norm": 24.64049193676725, "learning_rate": 4.5626458262912735e-07, "logits/chosen": -2.6262755393981934, "logits/rejected": -2.5886640548706055, "logps/chosen": -1.1033378839492798, "logps/rejected": -1.5101052522659302, "loss": 1.8774, "rewards/accuracies": 0.668749988079071, "rewards/chosen": -1.1033378839492798, "rewards/margins": 0.40676751732826233, "rewards/rejected": -1.5101052522659302, "step": 130 }, { "epoch": 0.2826485213294949, "grad_norm": 24.054739646210816, "learning_rate": 4.5095513994085974e-07, "logits/chosen": -2.5852103233337402, "logits/rejected": -2.5114357471466064, "logps/chosen": -1.036170244216919, "logps/rejected": -1.5170302391052246, "loss": 1.8964, "rewards/accuracies": 0.6625000238418579, "rewards/chosen": -1.036170244216919, "rewards/margins": 0.4808600842952728, "rewards/rejected": -1.5170302391052246, "step": 135 }, { "epoch": 0.29311698508243916, "grad_norm": 32.17420726348548, "learning_rate": 4.453763107901675e-07, "logits/chosen": -2.5846521854400635, "logits/rejected": -2.540843963623047, "logps/chosen": -1.1467993259429932, "logps/rejected": -1.3851490020751953, "loss": 1.958, "rewards/accuracies": 0.5375000238418579, "rewards/chosen": -1.1467993259429932, "rewards/margins": 0.23834970593452454, "rewards/rejected": -1.3851490020751953, "step": 140 }, { "epoch": 0.3035854488353834, "grad_norm": 31.293592588837402, "learning_rate": 4.395355737667985e-07, "logits/chosen": -2.5842862129211426, "logits/rejected": -2.481231451034546, "logps/chosen": -1.019995927810669, "logps/rejected": -1.450305700302124, "loss": 1.8621, "rewards/accuracies": 0.625, "rewards/chosen": -1.019995927810669, "rewards/margins": 0.43030983209609985, "rewards/rejected": -1.450305700302124, "step": 145 }, { "epoch": 0.31405391258832765, "grad_norm": 31.395757438350795, "learning_rate": 4.3344075855595097e-07, "logits/chosen": -2.6442532539367676, "logits/rejected": -2.5464065074920654, "logps/chosen": -1.066561222076416, "logps/rejected": -1.4769625663757324, "loss": 1.8856, "rewards/accuracies": 0.581250011920929, "rewards/chosen": -1.066561222076416, "rewards/margins": 0.4104014039039612, "rewards/rejected": -1.4769625663757324, "step": 150 }, { "epoch": 0.3245223763412719, "grad_norm": 54.99577539115471, "learning_rate": 4.271000354423425e-07, "logits/chosen": -2.584909200668335, "logits/rejected": -2.5346925258636475, "logps/chosen": -1.1342883110046387, "logps/rejected": -1.709241509437561, "loss": 1.8796, "rewards/accuracies": 0.7124999761581421, "rewards/chosen": -1.1342883110046387, "rewards/margins": 0.5749532580375671, "rewards/rejected": -1.709241509437561, "step": 155 }, { "epoch": 0.33499084009421615, "grad_norm": 62.43588869532461, "learning_rate": 4.2052190435769554e-07, "logits/chosen": -2.5957446098327637, "logits/rejected": -2.523685932159424, "logps/chosen": -1.048396348953247, "logps/rejected": -1.6725311279296875, "loss": 1.8954, "rewards/accuracies": 0.6812499761581421, "rewards/chosen": -1.048396348953247, "rewards/margins": 0.6241349577903748, "rewards/rejected": -1.6725311279296875, "step": 160 }, { "epoch": 0.34545930384716045, "grad_norm": 33.35503026194792, "learning_rate": 4.137151834863213e-07, "logits/chosen": -2.730977773666382, "logits/rejected": -2.645996570587158, "logps/chosen": -1.0159308910369873, "logps/rejected": -1.359212040901184, "loss": 1.8352, "rewards/accuracies": 0.612500011920929, "rewards/chosen": -1.0159308910369873, "rewards/margins": 0.343281090259552, "rewards/rejected": -1.359212040901184, "step": 165 }, { "epoch": 0.3559277676001047, "grad_norm": 32.04170211845891, "learning_rate": 4.0668899744407567e-07, "logits/chosen": -2.7265422344207764, "logits/rejected": -2.597689151763916, "logps/chosen": -0.9698517918586731, "logps/rejected": -1.3197566270828247, "loss": 1.8688, "rewards/accuracies": 0.643750011920929, "rewards/chosen": -0.9698517918586731, "rewards/margins": 0.349904865026474, "rewards/rejected": -1.3197566270828247, "step": 170 }, { "epoch": 0.36639623135304894, "grad_norm": 27.34421984947454, "learning_rate": 3.994527650465352e-07, "logits/chosen": -2.652785062789917, "logits/rejected": -2.5916075706481934, "logps/chosen": -0.9570480585098267, "logps/rejected": -1.3661720752716064, "loss": 1.8595, "rewards/accuracies": 0.6875, "rewards/chosen": -0.9570480585098267, "rewards/margins": 0.409123957157135, "rewards/rejected": -1.3661720752716064, "step": 175 }, { "epoch": 0.3768646951059932, "grad_norm": 35.429346147889426, "learning_rate": 3.920161866827889e-07, "logits/chosen": -2.6718568801879883, "logits/rejected": -2.613348960876465, "logps/chosen": -1.037642240524292, "logps/rejected": -1.553224802017212, "loss": 1.8512, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.037642240524292, "rewards/margins": 0.5155824422836304, "rewards/rejected": -1.553224802017212, "step": 180 }, { "epoch": 0.38733315885893743, "grad_norm": 21.700575234978658, "learning_rate": 3.8438923131177237e-07, "logits/chosen": -2.6914803981781006, "logits/rejected": -2.6440558433532715, "logps/chosen": -0.9652795791625977, "logps/rejected": -1.3787884712219238, "loss": 1.8051, "rewards/accuracies": 0.6625000238418579, "rewards/chosen": -0.9652795791625977, "rewards/margins": 0.41350871324539185, "rewards/rejected": -1.3787884712219238, "step": 185 }, { "epoch": 0.39780162261188173, "grad_norm": 32.670099112476585, "learning_rate": 3.765821230985757e-07, "logits/chosen": -2.6735928058624268, "logits/rejected": -2.640225648880005, "logps/chosen": -1.0144473314285278, "logps/rejected": -1.4563645124435425, "loss": 1.8273, "rewards/accuracies": 0.6625000238418579, "rewards/chosen": -1.0144473314285278, "rewards/margins": 0.4419172406196594, "rewards/rejected": -1.4563645124435425, "step": 190 }, { "epoch": 0.408270086364826, "grad_norm": 25.238718639520766, "learning_rate": 3.6860532770864005e-07, "logits/chosen": -2.705747365951538, "logits/rejected": -2.6104602813720703, "logps/chosen": -1.0011558532714844, "logps/rejected": -1.5988446474075317, "loss": 1.8926, "rewards/accuracies": 0.71875, "rewards/chosen": -1.0011558532714844, "rewards/margins": 0.5976886749267578, "rewards/rejected": -1.5988446474075317, "step": 195 }, { "epoch": 0.4187385501177702, "grad_norm": 23.97929885445044, "learning_rate": 3.604695382782159e-07, "logits/chosen": -2.644472122192383, "logits/rejected": -2.5559470653533936, "logps/chosen": -1.0497673749923706, "logps/rejected": -1.3750804662704468, "loss": 1.853, "rewards/accuracies": 0.6187499761581421, "rewards/chosen": -1.0497673749923706, "rewards/margins": 0.32531291246414185, "rewards/rejected": -1.3750804662704468, "step": 200 }, { "epoch": 0.42920701387071447, "grad_norm": 24.184854862916072, "learning_rate": 3.5218566107988867e-07, "logits/chosen": -2.730355739593506, "logits/rejected": -2.683979034423828, "logps/chosen": -0.9922051429748535, "logps/rejected": -1.5398856401443481, "loss": 1.8724, "rewards/accuracies": 0.6312500238418579, "rewards/chosen": -0.9922051429748535, "rewards/margins": 0.5476802587509155, "rewards/rejected": -1.5398856401443481, "step": 205 }, { "epoch": 0.4396754776236587, "grad_norm": 26.69782735208681, "learning_rate": 3.4376480090239047e-07, "logits/chosen": -2.5600266456604004, "logits/rejected": -2.5641884803771973, "logps/chosen": -1.0444707870483398, "logps/rejected": -1.4003455638885498, "loss": 1.8764, "rewards/accuracies": 0.637499988079071, "rewards/chosen": -1.0444707870483398, "rewards/margins": 0.35587459802627563, "rewards/rejected": -1.4003455638885498, "step": 210 }, { "epoch": 0.45014394137660296, "grad_norm": 39.9600540211152, "learning_rate": 3.3521824616429284e-07, "logits/chosen": -2.594709873199463, "logits/rejected": -2.5683107376098633, "logps/chosen": -1.126425862312317, "logps/rejected": -1.614018440246582, "loss": 1.8882, "rewards/accuracies": 0.6812499761581421, "rewards/chosen": -1.126425862312317, "rewards/margins": 0.4875926077365875, "rewards/rejected": -1.614018440246582, "step": 215 }, { "epoch": 0.46061240512954726, "grad_norm": 35.61682268118621, "learning_rate": 3.265574537815398e-07, "logits/chosen": -2.5961153507232666, "logits/rejected": -2.541783094406128, "logps/chosen": -1.102477788925171, "logps/rejected": -1.5613858699798584, "loss": 1.8517, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.102477788925171, "rewards/margins": 0.45890823006629944, "rewards/rejected": -1.5613858699798584, "step": 220 }, { "epoch": 0.4710808688824915, "grad_norm": 28.356766220400576, "learning_rate": 3.1779403380910425e-07, "logits/chosen": -2.6027655601501465, "logits/rejected": -2.4865376949310303, "logps/chosen": -1.0470962524414062, "logps/rejected": -1.5330448150634766, "loss": 1.8686, "rewards/accuracies": 0.71875, "rewards/chosen": -1.0470962524414062, "rewards/margins": 0.4859485626220703, "rewards/rejected": -1.5330448150634766, "step": 225 }, { "epoch": 0.48154933263543576, "grad_norm": 46.96042544932364, "learning_rate": 3.0893973387735683e-07, "logits/chosen": -2.6151013374328613, "logits/rejected": -2.59317684173584, "logps/chosen": -1.0420851707458496, "logps/rejected": -1.637976050376892, "loss": 1.8322, "rewards/accuracies": 0.706250011920929, "rewards/chosen": -1.0420851707458496, "rewards/margins": 0.5958911180496216, "rewards/rejected": -1.637976050376892, "step": 230 }, { "epoch": 0.49201779638838, "grad_norm": 26.100391626514853, "learning_rate": 3.000064234440111e-07, "logits/chosen": -2.5428245067596436, "logits/rejected": -2.4851417541503906, "logps/chosen": -1.0481144189834595, "logps/rejected": -1.4148355722427368, "loss": 1.8385, "rewards/accuracies": 0.65625, "rewards/chosen": -1.0481144189834595, "rewards/margins": 0.3667212724685669, "rewards/rejected": -1.4148355722427368, "step": 235 }, { "epoch": 0.5024862601413242, "grad_norm": 24.14989026973158, "learning_rate": 2.910060778827554e-07, "logits/chosen": -2.537334680557251, "logits/rejected": -2.472512722015381, "logps/chosen": -1.0776469707489014, "logps/rejected": -1.6333630084991455, "loss": 1.8508, "rewards/accuracies": 0.6937500238418579, "rewards/chosen": -1.0776469707489014, "rewards/margins": 0.5557159185409546, "rewards/rejected": -1.6333630084991455, "step": 240 }, { "epoch": 0.5129547238942685, "grad_norm": 32.38845341862983, "learning_rate": 2.8195076242990116e-07, "logits/chosen": -2.6101131439208984, "logits/rejected": -2.5712387561798096, "logps/chosen": -0.9908210635185242, "logps/rejected": -1.6324228048324585, "loss": 1.8075, "rewards/accuracies": 0.675000011920929, "rewards/chosen": -0.9908210635185242, "rewards/margins": 0.6416017413139343, "rewards/rejected": -1.6324228048324585, "step": 245 }, { "epoch": 0.5234231876472127, "grad_norm": 24.243213200791057, "learning_rate": 2.7285261601056697e-07, "logits/chosen": -2.559380292892456, "logits/rejected": -2.4537248611450195, "logps/chosen": -1.0512077808380127, "logps/rejected": -1.6259485483169556, "loss": 1.8615, "rewards/accuracies": 0.6812499761581421, "rewards/chosen": -1.0512077808380127, "rewards/margins": 0.5747405290603638, "rewards/rejected": -1.6259485483169556, "step": 250 }, { "epoch": 0.533891651400157, "grad_norm": 24.228357677990427, "learning_rate": 2.6372383496608186e-07, "logits/chosen": -2.654383420944214, "logits/rejected": -2.566788911819458, "logps/chosen": -1.0131515264511108, "logps/rejected": -1.4951848983764648, "loss": 1.8526, "rewards/accuracies": 0.574999988079071, "rewards/chosen": -1.0131515264511108, "rewards/margins": 0.4820333421230316, "rewards/rejected": -1.4951848983764648, "step": 255 }, { "epoch": 0.5443601151531012, "grad_norm": 26.47263727780241, "learning_rate": 2.5457665670441937e-07, "logits/chosen": -2.597144603729248, "logits/rejected": -2.572925090789795, "logps/chosen": -0.9740694761276245, "logps/rejected": -1.3979696035385132, "loss": 1.8422, "rewards/accuracies": 0.6625000238418579, "rewards/chosen": -0.9740694761276245, "rewards/margins": 0.4239000678062439, "rewards/rejected": -1.3979696035385132, "step": 260 }, { "epoch": 0.5548285789060455, "grad_norm": 22.019445712488015, "learning_rate": 2.454233432955807e-07, "logits/chosen": -2.612497329711914, "logits/rejected": -2.5739381313323975, "logps/chosen": -1.0217043161392212, "logps/rejected": -1.4334819316864014, "loss": 1.8467, "rewards/accuracies": 0.6937500238418579, "rewards/chosen": -1.0217043161392212, "rewards/margins": 0.4117775857448578, "rewards/rejected": -1.4334819316864014, "step": 265 }, { "epoch": 0.5652970426589898, "grad_norm": 31.0373082982515, "learning_rate": 2.3627616503391812e-07, "logits/chosen": -2.673020839691162, "logits/rejected": -2.6832852363586426, "logps/chosen": -1.1109861135482788, "logps/rejected": -1.4625297784805298, "loss": 1.8524, "rewards/accuracies": 0.6312500238418579, "rewards/chosen": -1.1109861135482788, "rewards/margins": 0.3515436351299286, "rewards/rejected": -1.4625297784805298, "step": 270 }, { "epoch": 0.575765506411934, "grad_norm": 29.601053617457204, "learning_rate": 2.2714738398943308e-07, "logits/chosen": -2.6691536903381348, "logits/rejected": -2.587501049041748, "logps/chosen": -1.0023844242095947, "logps/rejected": -1.6990854740142822, "loss": 1.7818, "rewards/accuracies": 0.6812499761581421, "rewards/chosen": -1.0023844242095947, "rewards/margins": 0.6967008113861084, "rewards/rejected": -1.6990854740142822, "step": 275 }, { "epoch": 0.5862339701648783, "grad_norm": 23.304378897117243, "learning_rate": 2.1804923757009882e-07, "logits/chosen": -2.7478485107421875, "logits/rejected": -2.6739115715026855, "logps/chosen": -0.967931866645813, "logps/rejected": -1.6212472915649414, "loss": 1.8289, "rewards/accuracies": 0.6312500238418579, "rewards/chosen": -0.967931866645813, "rewards/margins": 0.6533153653144836, "rewards/rejected": -1.6212472915649414, "step": 280 }, { "epoch": 0.5967024339178225, "grad_norm": 30.765932632186015, "learning_rate": 2.089939221172446e-07, "logits/chosen": -2.683306932449341, "logits/rejected": -2.599849224090576, "logps/chosen": -1.079810380935669, "logps/rejected": -1.5674939155578613, "loss": 1.8595, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.079810380935669, "rewards/margins": 0.4876834750175476, "rewards/rejected": -1.5674939155578613, "step": 285 }, { "epoch": 0.6071708976707668, "grad_norm": 29.814903147144946, "learning_rate": 1.9999357655598891e-07, "logits/chosen": -2.6533594131469727, "logits/rejected": -2.5845093727111816, "logps/chosen": -1.0454188585281372, "logps/rejected": -1.3878004550933838, "loss": 1.824, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.0454188585281372, "rewards/margins": 0.3423817753791809, "rewards/rejected": -1.3878004550933838, "step": 290 }, { "epoch": 0.6176393614237111, "grad_norm": 27.45666647767728, "learning_rate": 1.9106026612264315e-07, "logits/chosen": -2.6601853370666504, "logits/rejected": -2.6058852672576904, "logps/chosen": -1.038104772567749, "logps/rejected": -1.3897347450256348, "loss": 1.8319, "rewards/accuracies": 0.643750011920929, "rewards/chosen": -1.038104772567749, "rewards/margins": 0.3516300320625305, "rewards/rejected": -1.3897347450256348, "step": 295 }, { "epoch": 0.6281078251766553, "grad_norm": 30.25718786399205, "learning_rate": 1.8220596619089573e-07, "logits/chosen": -2.5553667545318604, "logits/rejected": -2.451190233230591, "logps/chosen": -1.108663558959961, "logps/rejected": -1.5200892686843872, "loss": 1.8653, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.108663558959961, "rewards/margins": 0.4114256501197815, "rewards/rejected": -1.5200892686843872, "step": 300 }, { "epoch": 0.6385762889295996, "grad_norm": 30.020985917091327, "learning_rate": 1.7344254621846017e-07, "logits/chosen": -2.640211343765259, "logits/rejected": -2.6043474674224854, "logps/chosen": -1.0029397010803223, "logps/rejected": -1.265366792678833, "loss": 1.8019, "rewards/accuracies": 0.606249988079071, "rewards/chosen": -1.0029397010803223, "rewards/margins": 0.2624271512031555, "rewards/rejected": -1.265366792678833, "step": 305 }, { "epoch": 0.6490447526825438, "grad_norm": 35.00385114807507, "learning_rate": 1.647817538357072e-07, "logits/chosen": -2.629307270050049, "logits/rejected": -2.4974796772003174, "logps/chosen": -1.0108293294906616, "logps/rejected": -1.5987032651901245, "loss": 1.825, "rewards/accuracies": 0.6625000238418579, "rewards/chosen": -1.0108293294906616, "rewards/margins": 0.5878738164901733, "rewards/rejected": -1.5987032651901245, "step": 310 }, { "epoch": 0.6595132164354881, "grad_norm": 33.95334568069759, "learning_rate": 1.562351990976095e-07, "logits/chosen": -2.6276791095733643, "logits/rejected": -2.503952980041504, "logps/chosen": -1.0764460563659668, "logps/rejected": -1.5906479358673096, "loss": 1.8359, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.0764460563659668, "rewards/margins": 0.514201819896698, "rewards/rejected": -1.5906479358673096, "step": 315 }, { "epoch": 0.6699816801884323, "grad_norm": 29.685480077784828, "learning_rate": 1.478143389201113e-07, "logits/chosen": -2.592771291732788, "logits/rejected": -2.5038256645202637, "logps/chosen": -1.034784197807312, "logps/rejected": -1.6445497274398804, "loss": 1.7974, "rewards/accuracies": 0.6312500238418579, "rewards/chosen": -1.034784197807312, "rewards/margins": 0.6097655892372131, "rewards/rejected": -1.6445497274398804, "step": 320 }, { "epoch": 0.6804501439413766, "grad_norm": 45.93159466769048, "learning_rate": 1.3953046172178413e-07, "logits/chosen": -2.5624806880950928, "logits/rejected": -2.4602789878845215, "logps/chosen": -1.1025031805038452, "logps/rejected": -1.7038898468017578, "loss": 1.8729, "rewards/accuracies": 0.6875, "rewards/chosen": -1.1025031805038452, "rewards/margins": 0.6013867855072021, "rewards/rejected": -1.7038898468017578, "step": 325 }, { "epoch": 0.6909186076943209, "grad_norm": 31.678205608236812, "learning_rate": 1.3139467229135998e-07, "logits/chosen": -2.599463939666748, "logits/rejected": -2.5258090496063232, "logps/chosen": -0.9680983424186707, "logps/rejected": -1.6296237707138062, "loss": 1.8012, "rewards/accuracies": 0.706250011920929, "rewards/chosen": -0.9680983424186707, "rewards/margins": 0.6615251302719116, "rewards/rejected": -1.6296237707138062, "step": 330 }, { "epoch": 0.7013870714472651, "grad_norm": 37.038121368656675, "learning_rate": 1.2341787690142435e-07, "logits/chosen": -2.5534796714782715, "logits/rejected": -2.517007827758789, "logps/chosen": -1.1220486164093018, "logps/rejected": -1.6694402694702148, "loss": 1.8584, "rewards/accuracies": 0.737500011920929, "rewards/chosen": -1.1220486164093018, "rewards/margins": 0.5473915338516235, "rewards/rejected": -1.6694402694702148, "step": 335 }, { "epoch": 0.7118555352002094, "grad_norm": 32.99732563960218, "learning_rate": 1.1561076868822755e-07, "logits/chosen": -2.58585524559021, "logits/rejected": -2.4757778644561768, "logps/chosen": -1.0188167095184326, "logps/rejected": -1.71088445186615, "loss": 1.8167, "rewards/accuracies": 0.6875, "rewards/chosen": -1.0188167095184326, "rewards/margins": 0.6920677423477173, "rewards/rejected": -1.71088445186615, "step": 340 }, { "epoch": 0.7223239989531536, "grad_norm": 31.150959005223893, "learning_rate": 1.0798381331721107e-07, "logits/chosen": -2.595191478729248, "logits/rejected": -2.491788387298584, "logps/chosen": -1.0228018760681152, "logps/rejected": -1.5643987655639648, "loss": 1.8471, "rewards/accuracies": 0.731249988079071, "rewards/chosen": -1.0228018760681152, "rewards/margins": 0.5415968298912048, "rewards/rejected": -1.5643987655639648, "step": 345 }, { "epoch": 0.7327924627060979, "grad_norm": 32.974367073048874, "learning_rate": 1.0054723495346482e-07, "logits/chosen": -2.4937729835510254, "logits/rejected": -2.3299617767333984, "logps/chosen": -1.100921630859375, "logps/rejected": -1.6467349529266357, "loss": 1.8361, "rewards/accuracies": 0.6312500238418579, "rewards/chosen": -1.100921630859375, "rewards/margins": 0.5458132028579712, "rewards/rejected": -1.6467349529266357, "step": 350 }, { "epoch": 0.7432609264590422, "grad_norm": 28.926067639567258, "learning_rate": 9.331100255592436e-08, "logits/chosen": -2.4389522075653076, "logits/rejected": -2.374527931213379, "logps/chosen": -1.0656179189682007, "logps/rejected": -1.67294442653656, "loss": 1.8606, "rewards/accuracies": 0.675000011920929, "rewards/chosen": -1.0656179189682007, "rewards/margins": 0.6073265075683594, "rewards/rejected": -1.67294442653656, "step": 355 }, { "epoch": 0.7537293902119864, "grad_norm": 41.902021950847704, "learning_rate": 8.628481651367875e-08, "logits/chosen": -2.497872829437256, "logits/rejected": -2.4256420135498047, "logps/chosen": -1.0714236497879028, "logps/rejected": -1.6986573934555054, "loss": 1.8331, "rewards/accuracies": 0.7124999761581421, "rewards/chosen": -1.0714236497879028, "rewards/margins": 0.6272337436676025, "rewards/rejected": -1.6986573934555054, "step": 360 }, { "epoch": 0.7641978539649307, "grad_norm": 31.435500480488756, "learning_rate": 7.947809564230445e-08, "logits/chosen": -2.532473087310791, "logits/rejected": -2.366628646850586, "logps/chosen": -1.0359852313995361, "logps/rejected": -1.7668260335922241, "loss": 1.7899, "rewards/accuracies": 0.6812499761581421, "rewards/chosen": -1.0359852313995361, "rewards/margins": 0.730840802192688, "rewards/rejected": -1.7668260335922241, "step": 365 }, { "epoch": 0.7746663177178749, "grad_norm": 39.54146493100524, "learning_rate": 7.289996455765748e-08, "logits/chosen": -2.484086513519287, "logits/rejected": -2.349870204925537, "logps/chosen": -1.0333083868026733, "logps/rejected": -1.6121408939361572, "loss": 1.7669, "rewards/accuracies": 0.7124999761581421, "rewards/chosen": -1.0333083868026733, "rewards/margins": 0.5788324475288391, "rewards/rejected": -1.6121408939361572, "step": 370 }, { "epoch": 0.7851347814708192, "grad_norm": 44.26258075929706, "learning_rate": 6.655924144404906e-08, "logits/chosen": -2.413093328475952, "logits/rejected": -2.3522191047668457, "logps/chosen": -0.994648277759552, "logps/rejected": -1.8310960531234741, "loss": 1.8098, "rewards/accuracies": 0.706250011920929, "rewards/chosen": -0.994648277759552, "rewards/margins": 0.8364478945732117, "rewards/rejected": -1.8310960531234741, "step": 375 }, { "epoch": 0.7956032452237635, "grad_norm": 42.81392142197046, "learning_rate": 6.046442623320145e-08, "logits/chosen": -2.5220797061920166, "logits/rejected": -2.5181660652160645, "logps/chosen": -1.0257781744003296, "logps/rejected": -1.565146803855896, "loss": 1.848, "rewards/accuracies": 0.668749988079071, "rewards/chosen": -1.0257781744003296, "rewards/margins": 0.5393685102462769, "rewards/rejected": -1.565146803855896, "step": 380 }, { "epoch": 0.8060717089767077, "grad_norm": 29.963279293888323, "learning_rate": 5.4623689209832484e-08, "logits/chosen": -2.4084603786468506, "logits/rejected": -2.3918216228485107, "logps/chosen": -1.0294010639190674, "logps/rejected": -1.5665457248687744, "loss": 1.8117, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.0294010639190674, "rewards/margins": 0.537144660949707, "rewards/rejected": -1.5665457248687744, "step": 385 }, { "epoch": 0.816540172729652, "grad_norm": 33.04796978561497, "learning_rate": 4.904486005914027e-08, "logits/chosen": -2.4773898124694824, "logits/rejected": -2.334437608718872, "logps/chosen": -1.0476093292236328, "logps/rejected": -1.5581729412078857, "loss": 1.8541, "rewards/accuracies": 0.612500011920929, "rewards/chosen": -1.0476093292236328, "rewards/margins": 0.5105637311935425, "rewards/rejected": -1.5581729412078857, "step": 390 }, { "epoch": 0.8270086364825961, "grad_norm": 36.23962004411558, "learning_rate": 4.373541737087263e-08, "logits/chosen": -2.4543843269348145, "logits/rejected": -2.3622539043426514, "logps/chosen": -1.0004545450210571, "logps/rejected": -1.7118451595306396, "loss": 1.822, "rewards/accuracies": 0.6875, "rewards/chosen": -1.0004545450210571, "rewards/margins": 0.7113906145095825, "rewards/rejected": -1.7118451595306396, "step": 395 }, { "epoch": 0.8374771002355405, "grad_norm": 35.25243486313715, "learning_rate": 3.8702478614051345e-08, "logits/chosen": -2.4065768718719482, "logits/rejected": -2.2744433879852295, "logps/chosen": -1.054172396659851, "logps/rejected": -1.5296390056610107, "loss": 1.8213, "rewards/accuracies": 0.643750011920929, "rewards/chosen": -1.054172396659851, "rewards/margins": 0.4754667282104492, "rewards/rejected": -1.5296390056610107, "step": 400 }, { "epoch": 0.8479455639884846, "grad_norm": 37.565064056547584, "learning_rate": 3.3952790595787986e-08, "logits/chosen": -2.449446439743042, "logits/rejected": -2.3016552925109863, "logps/chosen": -1.0504591464996338, "logps/rejected": -1.7570114135742188, "loss": 1.8237, "rewards/accuracies": 0.675000011920929, "rewards/chosen": -1.0504591464996338, "rewards/margins": 0.7065523266792297, "rewards/rejected": -1.7570114135742188, "step": 405 }, { "epoch": 0.8584140277414289, "grad_norm": 32.90354747950104, "learning_rate": 2.9492720416985e-08, "logits/chosen": -2.3721275329589844, "logits/rejected": -2.268225908279419, "logps/chosen": -1.1106208562850952, "logps/rejected": -1.9150855541229248, "loss": 1.8043, "rewards/accuracies": 0.668749988079071, "rewards/chosen": -1.1106208562850952, "rewards/margins": 0.8044648170471191, "rewards/rejected": -1.9150855541229248, "step": 410 }, { "epoch": 0.8688824914943732, "grad_norm": 56.846827175134244, "learning_rate": 2.5328246937043525e-08, "logits/chosen": -2.4376819133758545, "logits/rejected": -2.3333914279937744, "logps/chosen": -1.1099225282669067, "logps/rejected": -1.6782783269882202, "loss": 1.8252, "rewards/accuracies": 0.612500011920929, "rewards/chosen": -1.1099225282669067, "rewards/margins": 0.568355917930603, "rewards/rejected": -1.6782783269882202, "step": 415 }, { "epoch": 0.8793509552473174, "grad_norm": 43.77757640842041, "learning_rate": 2.1464952759020856e-08, "logits/chosen": -2.4453635215759277, "logits/rejected": -2.2761971950531006, "logps/chosen": -1.050205111503601, "logps/rejected": -1.8707281351089478, "loss": 1.7931, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -1.050205111503601, "rewards/margins": 0.8205229640007019, "rewards/rejected": -1.8707281351089478, "step": 420 }, { "epoch": 0.8898194190002617, "grad_norm": 37.41494353520166, "learning_rate": 1.7908016745981856e-08, "logits/chosen": -2.448782444000244, "logits/rejected": -2.2901499271392822, "logps/chosen": -1.043718695640564, "logps/rejected": -1.739383339881897, "loss": 1.8267, "rewards/accuracies": 0.6875, "rewards/chosen": -1.043718695640564, "rewards/margins": 0.695664644241333, "rewards/rejected": -1.739383339881897, "step": 425 }, { "epoch": 0.9002878827532059, "grad_norm": 30.693244043593708, "learning_rate": 1.4662207078575684e-08, "logits/chosen": -2.4449446201324463, "logits/rejected": -2.294093608856201, "logps/chosen": -1.0893933773040771, "logps/rejected": -1.642295241355896, "loss": 1.7605, "rewards/accuracies": 0.6937500238418579, "rewards/chosen": -1.0893933773040771, "rewards/margins": 0.5529020428657532, "rewards/rejected": -1.642295241355896, "step": 430 }, { "epoch": 0.9107563465061502, "grad_norm": 33.12531830225382, "learning_rate": 1.1731874863145142e-08, "logits/chosen": -2.3656153678894043, "logits/rejected": -2.352541208267212, "logps/chosen": -0.9648414850234985, "logps/rejected": -1.5097295045852661, "loss": 1.7984, "rewards/accuracies": 0.65625, "rewards/chosen": -0.9648414850234985, "rewards/margins": 0.5448881983757019, "rewards/rejected": -1.5097295045852661, "step": 435 }, { "epoch": 0.9212248102590945, "grad_norm": 40.86500194765816, "learning_rate": 9.12094829893642e-09, "logits/chosen": -2.48162579536438, "logits/rejected": -2.3563125133514404, "logps/chosen": -1.0542184114456177, "logps/rejected": -1.5571963787078857, "loss": 1.8252, "rewards/accuracies": 0.6499999761581421, "rewards/chosen": -1.0542184114456177, "rewards/margins": 0.5029779672622681, "rewards/rejected": -1.5571963787078857, "step": 440 }, { "epoch": 0.9316932740120387, "grad_norm": 30.792748106701044, "learning_rate": 6.832927412229017e-09, "logits/chosen": -2.399744987487793, "logits/rejected": -2.232243061065674, "logps/chosen": -1.0769870281219482, "logps/rejected": -1.728468656539917, "loss": 1.7915, "rewards/accuracies": 0.675000011920929, "rewards/chosen": -1.0769870281219482, "rewards/margins": 0.6514816284179688, "rewards/rejected": -1.728468656539917, "step": 445 }, { "epoch": 0.942161737764983, "grad_norm": 29.999766652539517, "learning_rate": 4.8708793644441086e-09, "logits/chosen": -2.459473133087158, "logits/rejected": -2.2956230640411377, "logps/chosen": -1.0055310726165771, "logps/rejected": -1.5959385633468628, "loss": 1.8126, "rewards/accuracies": 0.668749988079071, "rewards/chosen": -1.0055310726165771, "rewards/margins": 0.5904075503349304, "rewards/rejected": -1.5959385633468628, "step": 450 }, { "epoch": 0.9526302015179272, "grad_norm": 36.992525127734154, "learning_rate": 3.2374343405217884e-09, "logits/chosen": -2.42592191696167, "logits/rejected": -2.3259921073913574, "logps/chosen": -1.0747435092926025, "logps/rejected": -1.6361202001571655, "loss": 1.8922, "rewards/accuracies": 0.6812499761581421, "rewards/chosen": -1.0747435092926025, "rewards/margins": 0.5613764524459839, "rewards/rejected": -1.6361202001571655, "step": 455 }, { "epoch": 0.9630986652708715, "grad_norm": 43.04775872660601, "learning_rate": 1.9347820230782295e-09, "logits/chosen": -2.488020896911621, "logits/rejected": -2.239854097366333, "logps/chosen": -0.9606490135192871, "logps/rejected": -1.82891845703125, "loss": 1.8115, "rewards/accuracies": 0.7749999761581421, "rewards/chosen": -0.9606490135192871, "rewards/margins": 0.8682695627212524, "rewards/rejected": -1.82891845703125, "step": 460 }, { "epoch": 0.9735671290238157, "grad_norm": 25.076712480507634, "learning_rate": 9.64668657069706e-10, "logits/chosen": -2.4506993293762207, "logits/rejected": -2.252009868621826, "logps/chosen": -1.0200697183609009, "logps/rejected": -1.8124902248382568, "loss": 1.8012, "rewards/accuracies": 0.7250000238418579, "rewards/chosen": -1.0200697183609009, "rewards/margins": 0.7924206852912903, "rewards/rejected": -1.8124902248382568, "step": 465 }, { "epoch": 0.98403559277676, "grad_norm": 26.74704043050654, "learning_rate": 3.2839470889836627e-10, "logits/chosen": -2.3772401809692383, "logits/rejected": -2.2120604515075684, "logps/chosen": -0.9293681979179382, "logps/rejected": -1.7934061288833618, "loss": 1.7671, "rewards/accuracies": 0.6875, "rewards/chosen": -0.9293681979179382, "rewards/margins": 0.8640381097793579, "rewards/rejected": -1.7934061288833618, "step": 470 }, { "epoch": 0.9945040565297043, "grad_norm": 50.411989213964986, "learning_rate": 2.6813123097352287e-11, "logits/chosen": -2.4256701469421387, "logits/rejected": -2.3270206451416016, "logps/chosen": -0.9691774249076843, "logps/rejected": -1.615407943725586, "loss": 1.797, "rewards/accuracies": 0.637499988079071, "rewards/chosen": -0.9691774249076843, "rewards/margins": 0.6462305784225464, "rewards/rejected": -1.615407943725586, "step": 475 }, { "epoch": 0.998691442030882, "step": 477, "total_flos": 0.0, "train_loss": 1.8663299063716545, "train_runtime": 8277.1533, "train_samples_per_second": 7.386, "train_steps_per_second": 0.058 } ], "logging_steps": 5, "max_steps": 477, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }