diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,4233 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.007874009289690544, + "eval_steps": 500, + "global_step": 600, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 1.3123348816150905e-05, + "grad_norm": 7.072159767150879, + "learning_rate": 5e-06, + "loss": 2.0831, + "step": 1 + }, + { + "epoch": 2.624669763230181e-05, + "grad_norm": 6.047312259674072, + "learning_rate": 1e-05, + "loss": 1.888, + "step": 2 + }, + { + "epoch": 3.9370046448452714e-05, + "grad_norm": 4.877658843994141, + "learning_rate": 1.5e-05, + "loss": 1.8215, + "step": 3 + }, + { + "epoch": 5.249339526460362e-05, + "grad_norm": 5.105004787445068, + "learning_rate": 2e-05, + "loss": 1.7382, + "step": 4 + }, + { + "epoch": 6.561674408075452e-05, + "grad_norm": 4.398859024047852, + "learning_rate": 2.5e-05, + "loss": 1.519, + "step": 5 + }, + { + "epoch": 7.874009289690543e-05, + "grad_norm": 4.532390117645264, + "learning_rate": 3e-05, + "loss": 1.6656, + "step": 6 + }, + { + "epoch": 9.186344171305633e-05, + "grad_norm": 5.0148749351501465, + "learning_rate": 3.5e-05, + "loss": 1.6811, + "step": 7 + }, + { + "epoch": 0.00010498679052920724, + "grad_norm": 4.4031877517700195, + "learning_rate": 4e-05, + "loss": 1.7468, + "step": 8 + }, + { + "epoch": 0.00011811013934535815, + "grad_norm": 5.005381107330322, + "learning_rate": 4.5e-05, + "loss": 1.3947, + "step": 9 + }, + { + "epoch": 0.00013123348816150904, + "grad_norm": 3.713514566421509, + "learning_rate": 5e-05, + "loss": 1.405, + "step": 10 + }, + { + "epoch": 0.00014435683697765995, + "grad_norm": 9.202157974243164, + "learning_rate": 4.995798319327731e-05, + "loss": 1.5283, + "step": 11 + }, + { + "epoch": 0.00015748018579381085, + "grad_norm": 4.239948749542236, + "learning_rate": 4.991596638655463e-05, + "loss": 1.3864, + "step": 12 + }, + { + "epoch": 0.00017060353460996176, + "grad_norm": 2.7827541828155518, + "learning_rate": 4.9873949579831936e-05, + "loss": 1.4856, + "step": 13 + }, + { + "epoch": 0.00018372688342611267, + "grad_norm": 3.2062172889709473, + "learning_rate": 4.9831932773109245e-05, + "loss": 1.1977, + "step": 14 + }, + { + "epoch": 0.00019685023224226357, + "grad_norm": 3.271573066711426, + "learning_rate": 4.978991596638656e-05, + "loss": 1.2708, + "step": 15 + }, + { + "epoch": 0.00020997358105841448, + "grad_norm": 2.543915033340454, + "learning_rate": 4.974789915966387e-05, + "loss": 1.4229, + "step": 16 + }, + { + "epoch": 0.0002230969298745654, + "grad_norm": 3.8253746032714844, + "learning_rate": 4.970588235294118e-05, + "loss": 1.1775, + "step": 17 + }, + { + "epoch": 0.0002362202786907163, + "grad_norm": 3.0684823989868164, + "learning_rate": 4.966386554621849e-05, + "loss": 1.1546, + "step": 18 + }, + { + "epoch": 0.00024934362750686723, + "grad_norm": 3.0834896564483643, + "learning_rate": 4.9621848739495804e-05, + "loss": 1.3198, + "step": 19 + }, + { + "epoch": 0.0002624669763230181, + "grad_norm": 3.402059555053711, + "learning_rate": 4.957983193277311e-05, + "loss": 1.2536, + "step": 20 + }, + { + "epoch": 0.000275590325139169, + "grad_norm": 2.9694066047668457, + "learning_rate": 4.953781512605042e-05, + "loss": 1.2723, + "step": 21 + }, + { + "epoch": 0.0002887136739553199, + "grad_norm": 2.3450403213500977, + "learning_rate": 4.949579831932774e-05, + "loss": 1.4902, + "step": 22 + }, + { + "epoch": 0.0003018370227714708, + "grad_norm": 3.2001914978027344, + "learning_rate": 4.9453781512605046e-05, + "loss": 1.3031, + "step": 23 + }, + { + "epoch": 0.0003149603715876217, + "grad_norm": 2.4799821376800537, + "learning_rate": 4.9411764705882355e-05, + "loss": 1.3414, + "step": 24 + }, + { + "epoch": 0.0003280837204037726, + "grad_norm": 2.330723285675049, + "learning_rate": 4.936974789915967e-05, + "loss": 1.133, + "step": 25 + }, + { + "epoch": 0.0003412070692199235, + "grad_norm": 2.308879852294922, + "learning_rate": 4.932773109243698e-05, + "loss": 1.4592, + "step": 26 + }, + { + "epoch": 0.00035433041803607443, + "grad_norm": 3.835871934890747, + "learning_rate": 4.928571428571429e-05, + "loss": 0.8886, + "step": 27 + }, + { + "epoch": 0.00036745376685222534, + "grad_norm": 2.9454944133758545, + "learning_rate": 4.9243697478991605e-05, + "loss": 1.1211, + "step": 28 + }, + { + "epoch": 0.00038057711566837624, + "grad_norm": 2.6786818504333496, + "learning_rate": 4.920168067226891e-05, + "loss": 1.1148, + "step": 29 + }, + { + "epoch": 0.00039370046448452715, + "grad_norm": 2.5377731323242188, + "learning_rate": 4.9159663865546216e-05, + "loss": 1.1519, + "step": 30 + }, + { + "epoch": 0.00040682381330067806, + "grad_norm": 3.0859804153442383, + "learning_rate": 4.911764705882353e-05, + "loss": 1.2233, + "step": 31 + }, + { + "epoch": 0.00041994716211682896, + "grad_norm": 2.6658570766448975, + "learning_rate": 4.907563025210084e-05, + "loss": 1.2685, + "step": 32 + }, + { + "epoch": 0.00043307051093297987, + "grad_norm": 2.3701725006103516, + "learning_rate": 4.903361344537815e-05, + "loss": 1.274, + "step": 33 + }, + { + "epoch": 0.0004461938597491308, + "grad_norm": 2.379979133605957, + "learning_rate": 4.8991596638655466e-05, + "loss": 1.4761, + "step": 34 + }, + { + "epoch": 0.0004593172085652817, + "grad_norm": 2.8867926597595215, + "learning_rate": 4.8949579831932775e-05, + "loss": 1.0537, + "step": 35 + }, + { + "epoch": 0.0004724405573814326, + "grad_norm": 2.3834433555603027, + "learning_rate": 4.8907563025210084e-05, + "loss": 0.9892, + "step": 36 + }, + { + "epoch": 0.0004855639061975835, + "grad_norm": 2.4166579246520996, + "learning_rate": 4.886554621848739e-05, + "loss": 1.3343, + "step": 37 + }, + { + "epoch": 0.0004986872550137345, + "grad_norm": 2.471297264099121, + "learning_rate": 4.882352941176471e-05, + "loss": 0.8938, + "step": 38 + }, + { + "epoch": 0.0005118106038298853, + "grad_norm": 2.4426910877227783, + "learning_rate": 4.878151260504202e-05, + "loss": 0.8963, + "step": 39 + }, + { + "epoch": 0.0005249339526460362, + "grad_norm": 2.60628342628479, + "learning_rate": 4.8739495798319326e-05, + "loss": 1.1995, + "step": 40 + }, + { + "epoch": 0.0005380573014621871, + "grad_norm": 2.7314293384552, + "learning_rate": 4.869747899159664e-05, + "loss": 1.3207, + "step": 41 + }, + { + "epoch": 0.000551180650278338, + "grad_norm": 2.0626285076141357, + "learning_rate": 4.865546218487395e-05, + "loss": 1.0551, + "step": 42 + }, + { + "epoch": 0.0005643039990944889, + "grad_norm": 2.429570436477661, + "learning_rate": 4.861344537815126e-05, + "loss": 1.163, + "step": 43 + }, + { + "epoch": 0.0005774273479106398, + "grad_norm": 2.1089115142822266, + "learning_rate": 4.8571428571428576e-05, + "loss": 1.1919, + "step": 44 + }, + { + "epoch": 0.0005905506967267908, + "grad_norm": 2.2090256214141846, + "learning_rate": 4.8529411764705885e-05, + "loss": 1.0066, + "step": 45 + }, + { + "epoch": 0.0006036740455429416, + "grad_norm": 2.6260828971862793, + "learning_rate": 4.8487394957983194e-05, + "loss": 1.1277, + "step": 46 + }, + { + "epoch": 0.0006167973943590926, + "grad_norm": 2.6351325511932373, + "learning_rate": 4.844537815126051e-05, + "loss": 1.1907, + "step": 47 + }, + { + "epoch": 0.0006299207431752434, + "grad_norm": 2.4064207077026367, + "learning_rate": 4.840336134453782e-05, + "loss": 1.1826, + "step": 48 + }, + { + "epoch": 0.0006430440919913944, + "grad_norm": 2.2323639392852783, + "learning_rate": 4.836134453781513e-05, + "loss": 1.2202, + "step": 49 + }, + { + "epoch": 0.0006561674408075452, + "grad_norm": 2.712174654006958, + "learning_rate": 4.831932773109244e-05, + "loss": 1.0955, + "step": 50 + }, + { + "epoch": 0.0006692907896236962, + "grad_norm": 1.9196995496749878, + "learning_rate": 4.827731092436975e-05, + "loss": 0.9956, + "step": 51 + }, + { + "epoch": 0.000682414138439847, + "grad_norm": 2.415881395339966, + "learning_rate": 4.823529411764706e-05, + "loss": 1.1644, + "step": 52 + }, + { + "epoch": 0.000695537487255998, + "grad_norm": 2.6141293048858643, + "learning_rate": 4.819327731092437e-05, + "loss": 1.2137, + "step": 53 + }, + { + "epoch": 0.0007086608360721489, + "grad_norm": 2.353332757949829, + "learning_rate": 4.8151260504201686e-05, + "loss": 1.0817, + "step": 54 + }, + { + "epoch": 0.0007217841848882998, + "grad_norm": 2.222496271133423, + "learning_rate": 4.8109243697478995e-05, + "loss": 0.9575, + "step": 55 + }, + { + "epoch": 0.0007349075337044507, + "grad_norm": 2.616074323654175, + "learning_rate": 4.8067226890756304e-05, + "loss": 1.1872, + "step": 56 + }, + { + "epoch": 0.0007480308825206016, + "grad_norm": 2.496767520904541, + "learning_rate": 4.802521008403362e-05, + "loss": 1.1774, + "step": 57 + }, + { + "epoch": 0.0007611542313367525, + "grad_norm": 2.3791749477386475, + "learning_rate": 4.798319327731093e-05, + "loss": 1.0514, + "step": 58 + }, + { + "epoch": 0.0007742775801529034, + "grad_norm": 2.317368745803833, + "learning_rate": 4.794117647058824e-05, + "loss": 1.1349, + "step": 59 + }, + { + "epoch": 0.0007874009289690543, + "grad_norm": 2.3613386154174805, + "learning_rate": 4.7899159663865554e-05, + "loss": 0.9665, + "step": 60 + }, + { + "epoch": 0.0008005242777852053, + "grad_norm": 2.117526054382324, + "learning_rate": 4.785714285714286e-05, + "loss": 1.1252, + "step": 61 + }, + { + "epoch": 0.0008136476266013561, + "grad_norm": 2.248126745223999, + "learning_rate": 4.781512605042017e-05, + "loss": 1.0003, + "step": 62 + }, + { + "epoch": 0.0008267709754175071, + "grad_norm": 2.05367374420166, + "learning_rate": 4.777310924369748e-05, + "loss": 1.4639, + "step": 63 + }, + { + "epoch": 0.0008398943242336579, + "grad_norm": 2.026202440261841, + "learning_rate": 4.7731092436974796e-05, + "loss": 1.1478, + "step": 64 + }, + { + "epoch": 0.0008530176730498089, + "grad_norm": 2.1997697353363037, + "learning_rate": 4.7689075630252105e-05, + "loss": 1.2204, + "step": 65 + }, + { + "epoch": 0.0008661410218659597, + "grad_norm": 1.8945584297180176, + "learning_rate": 4.7647058823529414e-05, + "loss": 1.0535, + "step": 66 + }, + { + "epoch": 0.0008792643706821107, + "grad_norm": 2.010007619857788, + "learning_rate": 4.760504201680672e-05, + "loss": 1.0941, + "step": 67 + }, + { + "epoch": 0.0008923877194982616, + "grad_norm": 2.593773365020752, + "learning_rate": 4.756302521008403e-05, + "loss": 1.154, + "step": 68 + }, + { + "epoch": 0.0009055110683144125, + "grad_norm": 2.3330702781677246, + "learning_rate": 4.752100840336134e-05, + "loss": 1.1325, + "step": 69 + }, + { + "epoch": 0.0009186344171305634, + "grad_norm": 2.130695343017578, + "learning_rate": 4.747899159663866e-05, + "loss": 0.916, + "step": 70 + }, + { + "epoch": 0.0009317577659467143, + "grad_norm": 2.6290881633758545, + "learning_rate": 4.7436974789915966e-05, + "loss": 1.295, + "step": 71 + }, + { + "epoch": 0.0009448811147628652, + "grad_norm": 2.073453187942505, + "learning_rate": 4.7394957983193275e-05, + "loss": 1.2464, + "step": 72 + }, + { + "epoch": 0.0009580044635790161, + "grad_norm": 2.0779201984405518, + "learning_rate": 4.735294117647059e-05, + "loss": 1.3151, + "step": 73 + }, + { + "epoch": 0.000971127812395167, + "grad_norm": 2.802860975265503, + "learning_rate": 4.73109243697479e-05, + "loss": 0.872, + "step": 74 + }, + { + "epoch": 0.000984251161211318, + "grad_norm": 2.034449815750122, + "learning_rate": 4.726890756302521e-05, + "loss": 1.1338, + "step": 75 + }, + { + "epoch": 0.000997374510027469, + "grad_norm": 1.8365809917449951, + "learning_rate": 4.7226890756302525e-05, + "loss": 1.268, + "step": 76 + }, + { + "epoch": 0.0010104978588436197, + "grad_norm": 2.7444915771484375, + "learning_rate": 4.7184873949579834e-05, + "loss": 1.1891, + "step": 77 + }, + { + "epoch": 0.0010236212076597706, + "grad_norm": 2.1176328659057617, + "learning_rate": 4.714285714285714e-05, + "loss": 1.3602, + "step": 78 + }, + { + "epoch": 0.0010367445564759216, + "grad_norm": 2.3768794536590576, + "learning_rate": 4.710084033613446e-05, + "loss": 1.2641, + "step": 79 + }, + { + "epoch": 0.0010498679052920723, + "grad_norm": 2.1677494049072266, + "learning_rate": 4.705882352941177e-05, + "loss": 1.1084, + "step": 80 + }, + { + "epoch": 0.0010629912541082233, + "grad_norm": 2.5021719932556152, + "learning_rate": 4.7016806722689076e-05, + "loss": 1.0985, + "step": 81 + }, + { + "epoch": 0.0010761146029243742, + "grad_norm": 2.0837371349334717, + "learning_rate": 4.6974789915966385e-05, + "loss": 1.3129, + "step": 82 + }, + { + "epoch": 0.0010892379517405252, + "grad_norm": 2.4068455696105957, + "learning_rate": 4.69327731092437e-05, + "loss": 1.0889, + "step": 83 + }, + { + "epoch": 0.001102361300556676, + "grad_norm": 2.3378102779388428, + "learning_rate": 4.689075630252101e-05, + "loss": 1.1516, + "step": 84 + }, + { + "epoch": 0.001115484649372827, + "grad_norm": 2.165783166885376, + "learning_rate": 4.684873949579832e-05, + "loss": 1.2695, + "step": 85 + }, + { + "epoch": 0.0011286079981889779, + "grad_norm": 2.403777837753296, + "learning_rate": 4.6806722689075635e-05, + "loss": 1.0025, + "step": 86 + }, + { + "epoch": 0.0011417313470051288, + "grad_norm": 2.1567893028259277, + "learning_rate": 4.6764705882352944e-05, + "loss": 1.0176, + "step": 87 + }, + { + "epoch": 0.0011548546958212796, + "grad_norm": 2.3139235973358154, + "learning_rate": 4.672268907563025e-05, + "loss": 0.9044, + "step": 88 + }, + { + "epoch": 0.0011679780446374305, + "grad_norm": 2.56440806388855, + "learning_rate": 4.668067226890757e-05, + "loss": 0.7672, + "step": 89 + }, + { + "epoch": 0.0011811013934535815, + "grad_norm": 2.718919515609741, + "learning_rate": 4.663865546218488e-05, + "loss": 1.1918, + "step": 90 + }, + { + "epoch": 0.0011942247422697325, + "grad_norm": 1.8234260082244873, + "learning_rate": 4.6596638655462187e-05, + "loss": 0.9833, + "step": 91 + }, + { + "epoch": 0.0012073480910858832, + "grad_norm": 2.540635585784912, + "learning_rate": 4.65546218487395e-05, + "loss": 0.7035, + "step": 92 + }, + { + "epoch": 0.0012204714399020342, + "grad_norm": 2.1362297534942627, + "learning_rate": 4.651260504201681e-05, + "loss": 0.9999, + "step": 93 + }, + { + "epoch": 0.0012335947887181851, + "grad_norm": 2.085761547088623, + "learning_rate": 4.647058823529412e-05, + "loss": 1.1759, + "step": 94 + }, + { + "epoch": 0.001246718137534336, + "grad_norm": 2.096830129623413, + "learning_rate": 4.642857142857143e-05, + "loss": 1.0324, + "step": 95 + }, + { + "epoch": 0.0012598414863504868, + "grad_norm": 2.817042112350464, + "learning_rate": 4.6386554621848745e-05, + "loss": 1.1897, + "step": 96 + }, + { + "epoch": 0.0012729648351666378, + "grad_norm": 2.062160015106201, + "learning_rate": 4.6344537815126054e-05, + "loss": 0.8634, + "step": 97 + }, + { + "epoch": 0.0012860881839827888, + "grad_norm": 1.8099182844161987, + "learning_rate": 4.630252100840336e-05, + "loss": 1.1785, + "step": 98 + }, + { + "epoch": 0.0012992115327989397, + "grad_norm": 2.1783230304718018, + "learning_rate": 4.626050420168068e-05, + "loss": 1.2024, + "step": 99 + }, + { + "epoch": 0.0013123348816150905, + "grad_norm": 1.9956905841827393, + "learning_rate": 4.621848739495799e-05, + "loss": 0.9637, + "step": 100 + }, + { + "epoch": 0.0013254582304312414, + "grad_norm": 2.4066596031188965, + "learning_rate": 4.61764705882353e-05, + "loss": 1.2583, + "step": 101 + }, + { + "epoch": 0.0013385815792473924, + "grad_norm": 2.0673747062683105, + "learning_rate": 4.613445378151261e-05, + "loss": 1.1186, + "step": 102 + }, + { + "epoch": 0.0013517049280635433, + "grad_norm": 1.8033102750778198, + "learning_rate": 4.6092436974789915e-05, + "loss": 1.1626, + "step": 103 + }, + { + "epoch": 0.001364828276879694, + "grad_norm": 2.032924175262451, + "learning_rate": 4.6050420168067224e-05, + "loss": 1.0937, + "step": 104 + }, + { + "epoch": 0.001377951625695845, + "grad_norm": 2.554765462875366, + "learning_rate": 4.600840336134454e-05, + "loss": 1.0015, + "step": 105 + }, + { + "epoch": 0.001391074974511996, + "grad_norm": 2.19063401222229, + "learning_rate": 4.596638655462185e-05, + "loss": 0.9031, + "step": 106 + }, + { + "epoch": 0.001404198323328147, + "grad_norm": 2.0013623237609863, + "learning_rate": 4.592436974789916e-05, + "loss": 1.1653, + "step": 107 + }, + { + "epoch": 0.0014173216721442977, + "grad_norm": 1.965734601020813, + "learning_rate": 4.588235294117647e-05, + "loss": 0.9685, + "step": 108 + }, + { + "epoch": 0.0014304450209604487, + "grad_norm": 2.040010690689087, + "learning_rate": 4.584033613445378e-05, + "loss": 1.1904, + "step": 109 + }, + { + "epoch": 0.0014435683697765996, + "grad_norm": 2.6654715538024902, + "learning_rate": 4.579831932773109e-05, + "loss": 1.2973, + "step": 110 + }, + { + "epoch": 0.0014566917185927506, + "grad_norm": 2.138746500015259, + "learning_rate": 4.575630252100841e-05, + "loss": 1.0583, + "step": 111 + }, + { + "epoch": 0.0014698150674089013, + "grad_norm": 2.018465995788574, + "learning_rate": 4.5714285714285716e-05, + "loss": 1.2435, + "step": 112 + }, + { + "epoch": 0.0014829384162250523, + "grad_norm": 1.9513568878173828, + "learning_rate": 4.5672268907563025e-05, + "loss": 1.0105, + "step": 113 + }, + { + "epoch": 0.0014960617650412033, + "grad_norm": 2.152390718460083, + "learning_rate": 4.5630252100840334e-05, + "loss": 1.279, + "step": 114 + }, + { + "epoch": 0.0015091851138573542, + "grad_norm": 2.2886104583740234, + "learning_rate": 4.558823529411765e-05, + "loss": 1.106, + "step": 115 + }, + { + "epoch": 0.001522308462673505, + "grad_norm": 2.221177816390991, + "learning_rate": 4.554621848739496e-05, + "loss": 0.8511, + "step": 116 + }, + { + "epoch": 0.001535431811489656, + "grad_norm": 1.791222095489502, + "learning_rate": 4.550420168067227e-05, + "loss": 1.082, + "step": 117 + }, + { + "epoch": 0.001548555160305807, + "grad_norm": 2.0436720848083496, + "learning_rate": 4.5462184873949584e-05, + "loss": 1.4781, + "step": 118 + }, + { + "epoch": 0.0015616785091219576, + "grad_norm": 2.416624069213867, + "learning_rate": 4.542016806722689e-05, + "loss": 1.0889, + "step": 119 + }, + { + "epoch": 0.0015748018579381086, + "grad_norm": 2.185920476913452, + "learning_rate": 4.53781512605042e-05, + "loss": 1.0695, + "step": 120 + }, + { + "epoch": 0.0015879252067542596, + "grad_norm": 2.1533713340759277, + "learning_rate": 4.533613445378152e-05, + "loss": 0.9034, + "step": 121 + }, + { + "epoch": 0.0016010485555704105, + "grad_norm": 2.1179723739624023, + "learning_rate": 4.5294117647058826e-05, + "loss": 1.094, + "step": 122 + }, + { + "epoch": 0.0016141719043865613, + "grad_norm": 2.0470693111419678, + "learning_rate": 4.5252100840336135e-05, + "loss": 1.0756, + "step": 123 + }, + { + "epoch": 0.0016272952532027122, + "grad_norm": 1.9948326349258423, + "learning_rate": 4.521008403361345e-05, + "loss": 1.1651, + "step": 124 + }, + { + "epoch": 0.0016404186020188632, + "grad_norm": 1.8133087158203125, + "learning_rate": 4.516806722689076e-05, + "loss": 1.356, + "step": 125 + }, + { + "epoch": 0.0016535419508350141, + "grad_norm": 2.027836322784424, + "learning_rate": 4.512605042016807e-05, + "loss": 1.0432, + "step": 126 + }, + { + "epoch": 0.001666665299651165, + "grad_norm": 2.365647792816162, + "learning_rate": 4.508403361344538e-05, + "loss": 1.0608, + "step": 127 + }, + { + "epoch": 0.0016797886484673159, + "grad_norm": 2.0502471923828125, + "learning_rate": 4.5042016806722694e-05, + "loss": 0.9935, + "step": 128 + }, + { + "epoch": 0.0016929119972834668, + "grad_norm": 2.469956398010254, + "learning_rate": 4.5e-05, + "loss": 1.0578, + "step": 129 + }, + { + "epoch": 0.0017060353460996178, + "grad_norm": 2.224716901779175, + "learning_rate": 4.495798319327731e-05, + "loss": 1.0054, + "step": 130 + }, + { + "epoch": 0.0017191586949157685, + "grad_norm": 2.0795905590057373, + "learning_rate": 4.491596638655463e-05, + "loss": 0.9304, + "step": 131 + }, + { + "epoch": 0.0017322820437319195, + "grad_norm": 2.5419461727142334, + "learning_rate": 4.4873949579831937e-05, + "loss": 1.0532, + "step": 132 + }, + { + "epoch": 0.0017454053925480704, + "grad_norm": 1.9789624214172363, + "learning_rate": 4.4831932773109246e-05, + "loss": 0.7064, + "step": 133 + }, + { + "epoch": 0.0017585287413642214, + "grad_norm": 2.256998062133789, + "learning_rate": 4.478991596638656e-05, + "loss": 1.0151, + "step": 134 + }, + { + "epoch": 0.0017716520901803721, + "grad_norm": 1.9684431552886963, + "learning_rate": 4.474789915966387e-05, + "loss": 1.3969, + "step": 135 + }, + { + "epoch": 0.001784775438996523, + "grad_norm": 1.9599086046218872, + "learning_rate": 4.470588235294118e-05, + "loss": 0.9254, + "step": 136 + }, + { + "epoch": 0.001797898787812674, + "grad_norm": 2.0463593006134033, + "learning_rate": 4.4663865546218495e-05, + "loss": 1.1675, + "step": 137 + }, + { + "epoch": 0.001811022136628825, + "grad_norm": 2.3988492488861084, + "learning_rate": 4.4621848739495804e-05, + "loss": 1.1021, + "step": 138 + }, + { + "epoch": 0.0018241454854449758, + "grad_norm": 2.1077089309692383, + "learning_rate": 4.457983193277311e-05, + "loss": 1.0832, + "step": 139 + }, + { + "epoch": 0.0018372688342611267, + "grad_norm": 2.027130365371704, + "learning_rate": 4.453781512605042e-05, + "loss": 1.1428, + "step": 140 + }, + { + "epoch": 0.0018503921830772777, + "grad_norm": 1.8151514530181885, + "learning_rate": 4.449579831932773e-05, + "loss": 1.6232, + "step": 141 + }, + { + "epoch": 0.0018635155318934287, + "grad_norm": 2.0072662830352783, + "learning_rate": 4.445378151260504e-05, + "loss": 0.9918, + "step": 142 + }, + { + "epoch": 0.0018766388807095794, + "grad_norm": 2.0975875854492188, + "learning_rate": 4.4411764705882356e-05, + "loss": 1.1645, + "step": 143 + }, + { + "epoch": 0.0018897622295257304, + "grad_norm": 1.8917436599731445, + "learning_rate": 4.4369747899159665e-05, + "loss": 0.9891, + "step": 144 + }, + { + "epoch": 0.0019028855783418813, + "grad_norm": 2.3437271118164062, + "learning_rate": 4.4327731092436974e-05, + "loss": 0.9545, + "step": 145 + }, + { + "epoch": 0.0019160089271580323, + "grad_norm": 1.8604134321212769, + "learning_rate": 4.428571428571428e-05, + "loss": 1.0553, + "step": 146 + }, + { + "epoch": 0.001929132275974183, + "grad_norm": 2.1565375328063965, + "learning_rate": 4.42436974789916e-05, + "loss": 1.411, + "step": 147 + }, + { + "epoch": 0.001942255624790334, + "grad_norm": 2.3755760192871094, + "learning_rate": 4.420168067226891e-05, + "loss": 0.9843, + "step": 148 + }, + { + "epoch": 0.0019553789736064847, + "grad_norm": 2.2564103603363037, + "learning_rate": 4.4159663865546217e-05, + "loss": 0.9633, + "step": 149 + }, + { + "epoch": 0.001968502322422636, + "grad_norm": 2.1097161769866943, + "learning_rate": 4.411764705882353e-05, + "loss": 0.935, + "step": 150 + }, + { + "epoch": 0.0019816256712387867, + "grad_norm": 1.8181250095367432, + "learning_rate": 4.407563025210084e-05, + "loss": 1.0492, + "step": 151 + }, + { + "epoch": 0.001994749020054938, + "grad_norm": 1.8607770204544067, + "learning_rate": 4.403361344537815e-05, + "loss": 0.9848, + "step": 152 + }, + { + "epoch": 0.0020078723688710886, + "grad_norm": 1.7969127893447876, + "learning_rate": 4.3991596638655466e-05, + "loss": 1.0657, + "step": 153 + }, + { + "epoch": 0.0020209957176872393, + "grad_norm": 2.0169637203216553, + "learning_rate": 4.3949579831932775e-05, + "loss": 1.0059, + "step": 154 + }, + { + "epoch": 0.0020341190665033905, + "grad_norm": 1.8811042308807373, + "learning_rate": 4.3907563025210084e-05, + "loss": 1.0386, + "step": 155 + }, + { + "epoch": 0.0020472424153195412, + "grad_norm": 1.9799407720565796, + "learning_rate": 4.38655462184874e-05, + "loss": 0.8287, + "step": 156 + }, + { + "epoch": 0.002060365764135692, + "grad_norm": 1.5626941919326782, + "learning_rate": 4.382352941176471e-05, + "loss": 0.9688, + "step": 157 + }, + { + "epoch": 0.002073489112951843, + "grad_norm": 2.183609962463379, + "learning_rate": 4.378151260504202e-05, + "loss": 1.3839, + "step": 158 + }, + { + "epoch": 0.002086612461767994, + "grad_norm": 1.854836106300354, + "learning_rate": 4.373949579831933e-05, + "loss": 1.2027, + "step": 159 + }, + { + "epoch": 0.0020997358105841447, + "grad_norm": 2.2002053260803223, + "learning_rate": 4.369747899159664e-05, + "loss": 1.078, + "step": 160 + }, + { + "epoch": 0.002112859159400296, + "grad_norm": 1.8938093185424805, + "learning_rate": 4.365546218487395e-05, + "loss": 1.2866, + "step": 161 + }, + { + "epoch": 0.0021259825082164466, + "grad_norm": 2.015660524368286, + "learning_rate": 4.361344537815126e-05, + "loss": 1.0272, + "step": 162 + }, + { + "epoch": 0.0021391058570325978, + "grad_norm": 1.8368442058563232, + "learning_rate": 4.3571428571428576e-05, + "loss": 1.2469, + "step": 163 + }, + { + "epoch": 0.0021522292058487485, + "grad_norm": 2.056487560272217, + "learning_rate": 4.3529411764705885e-05, + "loss": 1.1621, + "step": 164 + }, + { + "epoch": 0.0021653525546648992, + "grad_norm": 1.8423100709915161, + "learning_rate": 4.3487394957983194e-05, + "loss": 1.0864, + "step": 165 + }, + { + "epoch": 0.0021784759034810504, + "grad_norm": 1.7012152671813965, + "learning_rate": 4.344537815126051e-05, + "loss": 0.7503, + "step": 166 + }, + { + "epoch": 0.002191599252297201, + "grad_norm": 2.687711000442505, + "learning_rate": 4.340336134453782e-05, + "loss": 1.2613, + "step": 167 + }, + { + "epoch": 0.002204722601113352, + "grad_norm": 1.9029107093811035, + "learning_rate": 4.336134453781513e-05, + "loss": 1.0015, + "step": 168 + }, + { + "epoch": 0.002217845949929503, + "grad_norm": 1.7663654088974, + "learning_rate": 4.3319327731092444e-05, + "loss": 0.9883, + "step": 169 + }, + { + "epoch": 0.002230969298745654, + "grad_norm": 1.8921157121658325, + "learning_rate": 4.327731092436975e-05, + "loss": 1.1788, + "step": 170 + }, + { + "epoch": 0.002244092647561805, + "grad_norm": 1.870092511177063, + "learning_rate": 4.323529411764706e-05, + "loss": 0.8684, + "step": 171 + }, + { + "epoch": 0.0022572159963779558, + "grad_norm": 1.918101191520691, + "learning_rate": 4.319327731092437e-05, + "loss": 0.7238, + "step": 172 + }, + { + "epoch": 0.0022703393451941065, + "grad_norm": 2.038949966430664, + "learning_rate": 4.3151260504201687e-05, + "loss": 1.1109, + "step": 173 + }, + { + "epoch": 0.0022834626940102577, + "grad_norm": 2.1337692737579346, + "learning_rate": 4.3109243697478996e-05, + "loss": 1.0716, + "step": 174 + }, + { + "epoch": 0.0022965860428264084, + "grad_norm": 1.7539408206939697, + "learning_rate": 4.3067226890756305e-05, + "loss": 1.1325, + "step": 175 + }, + { + "epoch": 0.002309709391642559, + "grad_norm": 2.0798778533935547, + "learning_rate": 4.302521008403362e-05, + "loss": 0.9844, + "step": 176 + }, + { + "epoch": 0.0023228327404587103, + "grad_norm": 2.239736318588257, + "learning_rate": 4.298319327731093e-05, + "loss": 0.9153, + "step": 177 + }, + { + "epoch": 0.002335956089274861, + "grad_norm": 1.7551506757736206, + "learning_rate": 4.294117647058823e-05, + "loss": 0.7961, + "step": 178 + }, + { + "epoch": 0.0023490794380910123, + "grad_norm": 1.6866077184677124, + "learning_rate": 4.289915966386555e-05, + "loss": 0.8045, + "step": 179 + }, + { + "epoch": 0.002362202786907163, + "grad_norm": 2.037832260131836, + "learning_rate": 4.2857142857142856e-05, + "loss": 1.0526, + "step": 180 + }, + { + "epoch": 0.0023753261357233138, + "grad_norm": 1.9151948690414429, + "learning_rate": 4.2815126050420165e-05, + "loss": 1.0393, + "step": 181 + }, + { + "epoch": 0.002388449484539465, + "grad_norm": 1.807278037071228, + "learning_rate": 4.277310924369748e-05, + "loss": 0.7957, + "step": 182 + }, + { + "epoch": 0.0024015728333556157, + "grad_norm": 1.9136124849319458, + "learning_rate": 4.273109243697479e-05, + "loss": 1.0114, + "step": 183 + }, + { + "epoch": 0.0024146961821717664, + "grad_norm": 2.3868467807769775, + "learning_rate": 4.26890756302521e-05, + "loss": 1.1214, + "step": 184 + }, + { + "epoch": 0.0024278195309879176, + "grad_norm": 2.0041465759277344, + "learning_rate": 4.2647058823529415e-05, + "loss": 1.2091, + "step": 185 + }, + { + "epoch": 0.0024409428798040683, + "grad_norm": 1.787702202796936, + "learning_rate": 4.2605042016806724e-05, + "loss": 1.2995, + "step": 186 + }, + { + "epoch": 0.0024540662286202195, + "grad_norm": 1.82904052734375, + "learning_rate": 4.256302521008403e-05, + "loss": 1.129, + "step": 187 + }, + { + "epoch": 0.0024671895774363703, + "grad_norm": 2.173250675201416, + "learning_rate": 4.252100840336135e-05, + "loss": 1.0799, + "step": 188 + }, + { + "epoch": 0.002480312926252521, + "grad_norm": 1.7813307046890259, + "learning_rate": 4.247899159663866e-05, + "loss": 1.0202, + "step": 189 + }, + { + "epoch": 0.002493436275068672, + "grad_norm": 2.2016477584838867, + "learning_rate": 4.2436974789915967e-05, + "loss": 0.718, + "step": 190 + }, + { + "epoch": 0.002506559623884823, + "grad_norm": 2.0901100635528564, + "learning_rate": 4.2394957983193276e-05, + "loss": 1.0118, + "step": 191 + }, + { + "epoch": 0.0025196829727009737, + "grad_norm": 2.0191895961761475, + "learning_rate": 4.235294117647059e-05, + "loss": 1.1978, + "step": 192 + }, + { + "epoch": 0.002532806321517125, + "grad_norm": 1.896297812461853, + "learning_rate": 4.23109243697479e-05, + "loss": 0.9719, + "step": 193 + }, + { + "epoch": 0.0025459296703332756, + "grad_norm": 1.894870400428772, + "learning_rate": 4.226890756302521e-05, + "loss": 1.229, + "step": 194 + }, + { + "epoch": 0.0025590530191494263, + "grad_norm": 1.902392029762268, + "learning_rate": 4.2226890756302525e-05, + "loss": 1.3133, + "step": 195 + }, + { + "epoch": 0.0025721763679655775, + "grad_norm": 2.120286464691162, + "learning_rate": 4.2184873949579834e-05, + "loss": 1.1485, + "step": 196 + }, + { + "epoch": 0.0025852997167817283, + "grad_norm": 2.0686750411987305, + "learning_rate": 4.214285714285714e-05, + "loss": 1.0732, + "step": 197 + }, + { + "epoch": 0.0025984230655978794, + "grad_norm": 1.6315714120864868, + "learning_rate": 4.210084033613446e-05, + "loss": 0.6762, + "step": 198 + }, + { + "epoch": 0.00261154641441403, + "grad_norm": 1.7602286338806152, + "learning_rate": 4.205882352941177e-05, + "loss": 0.8897, + "step": 199 + }, + { + "epoch": 0.002624669763230181, + "grad_norm": 2.136775493621826, + "learning_rate": 4.201680672268908e-05, + "loss": 1.1179, + "step": 200 + }, + { + "epoch": 0.002637793112046332, + "grad_norm": 1.9759247303009033, + "learning_rate": 4.197478991596639e-05, + "loss": 1.078, + "step": 201 + }, + { + "epoch": 0.002650916460862483, + "grad_norm": 2.1391897201538086, + "learning_rate": 4.19327731092437e-05, + "loss": 1.0673, + "step": 202 + }, + { + "epoch": 0.0026640398096786336, + "grad_norm": 2.3699793815612793, + "learning_rate": 4.189075630252101e-05, + "loss": 1.0816, + "step": 203 + }, + { + "epoch": 0.0026771631584947848, + "grad_norm": 1.7057411670684814, + "learning_rate": 4.184873949579832e-05, + "loss": 1.1088, + "step": 204 + }, + { + "epoch": 0.0026902865073109355, + "grad_norm": 1.6120874881744385, + "learning_rate": 4.1806722689075635e-05, + "loss": 1.2588, + "step": 205 + }, + { + "epoch": 0.0027034098561270867, + "grad_norm": 2.3143081665039062, + "learning_rate": 4.1764705882352944e-05, + "loss": 1.3962, + "step": 206 + }, + { + "epoch": 0.0027165332049432374, + "grad_norm": 2.304960250854492, + "learning_rate": 4.172268907563025e-05, + "loss": 1.0249, + "step": 207 + }, + { + "epoch": 0.002729656553759388, + "grad_norm": 1.8955740928649902, + "learning_rate": 4.168067226890757e-05, + "loss": 0.7611, + "step": 208 + }, + { + "epoch": 0.0027427799025755394, + "grad_norm": 2.069223403930664, + "learning_rate": 4.163865546218488e-05, + "loss": 0.9099, + "step": 209 + }, + { + "epoch": 0.00275590325139169, + "grad_norm": 1.9223281145095825, + "learning_rate": 4.159663865546219e-05, + "loss": 1.3042, + "step": 210 + }, + { + "epoch": 0.002769026600207841, + "grad_norm": 1.9752320051193237, + "learning_rate": 4.15546218487395e-05, + "loss": 1.0339, + "step": 211 + }, + { + "epoch": 0.002782149949023992, + "grad_norm": 1.8929498195648193, + "learning_rate": 4.151260504201681e-05, + "loss": 0.9886, + "step": 212 + }, + { + "epoch": 0.0027952732978401428, + "grad_norm": 2.2697014808654785, + "learning_rate": 4.147058823529412e-05, + "loss": 1.1485, + "step": 213 + }, + { + "epoch": 0.002808396646656294, + "grad_norm": 1.9803391695022583, + "learning_rate": 4.1428571428571437e-05, + "loss": 1.1344, + "step": 214 + }, + { + "epoch": 0.0028215199954724447, + "grad_norm": 2.1848607063293457, + "learning_rate": 4.138655462184874e-05, + "loss": 1.1566, + "step": 215 + }, + { + "epoch": 0.0028346433442885954, + "grad_norm": 1.7310750484466553, + "learning_rate": 4.134453781512605e-05, + "loss": 0.7865, + "step": 216 + }, + { + "epoch": 0.0028477666931047466, + "grad_norm": 1.832837462425232, + "learning_rate": 4.1302521008403364e-05, + "loss": 1.0452, + "step": 217 + }, + { + "epoch": 0.0028608900419208974, + "grad_norm": 2.0909769535064697, + "learning_rate": 4.126050420168067e-05, + "loss": 0.965, + "step": 218 + }, + { + "epoch": 0.002874013390737048, + "grad_norm": 2.217707872390747, + "learning_rate": 4.121848739495798e-05, + "loss": 1.0826, + "step": 219 + }, + { + "epoch": 0.0028871367395531993, + "grad_norm": 1.8040285110473633, + "learning_rate": 4.11764705882353e-05, + "loss": 0.8233, + "step": 220 + }, + { + "epoch": 0.00290026008836935, + "grad_norm": 1.7424750328063965, + "learning_rate": 4.1134453781512606e-05, + "loss": 0.9798, + "step": 221 + }, + { + "epoch": 0.002913383437185501, + "grad_norm": 2.0254833698272705, + "learning_rate": 4.1092436974789915e-05, + "loss": 0.9843, + "step": 222 + }, + { + "epoch": 0.002926506786001652, + "grad_norm": 2.2267329692840576, + "learning_rate": 4.1050420168067224e-05, + "loss": 0.8196, + "step": 223 + }, + { + "epoch": 0.0029396301348178027, + "grad_norm": 2.4124982357025146, + "learning_rate": 4.100840336134454e-05, + "loss": 1.3434, + "step": 224 + }, + { + "epoch": 0.002952753483633954, + "grad_norm": 1.988885760307312, + "learning_rate": 4.096638655462185e-05, + "loss": 1.1987, + "step": 225 + }, + { + "epoch": 0.0029658768324501046, + "grad_norm": 1.7969712018966675, + "learning_rate": 4.092436974789916e-05, + "loss": 1.2787, + "step": 226 + }, + { + "epoch": 0.0029790001812662554, + "grad_norm": 1.963279128074646, + "learning_rate": 4.0882352941176474e-05, + "loss": 1.0341, + "step": 227 + }, + { + "epoch": 0.0029921235300824065, + "grad_norm": 1.8143019676208496, + "learning_rate": 4.084033613445378e-05, + "loss": 1.3237, + "step": 228 + }, + { + "epoch": 0.0030052468788985573, + "grad_norm": 2.0594189167022705, + "learning_rate": 4.079831932773109e-05, + "loss": 0.8073, + "step": 229 + }, + { + "epoch": 0.0030183702277147085, + "grad_norm": 1.9188764095306396, + "learning_rate": 4.075630252100841e-05, + "loss": 1.1112, + "step": 230 + }, + { + "epoch": 0.003031493576530859, + "grad_norm": 1.802120327949524, + "learning_rate": 4.0714285714285717e-05, + "loss": 1.1549, + "step": 231 + }, + { + "epoch": 0.00304461692534701, + "grad_norm": 1.954909324645996, + "learning_rate": 4.0672268907563026e-05, + "loss": 1.1407, + "step": 232 + }, + { + "epoch": 0.003057740274163161, + "grad_norm": 1.897289752960205, + "learning_rate": 4.063025210084034e-05, + "loss": 0.9183, + "step": 233 + }, + { + "epoch": 0.003070863622979312, + "grad_norm": 2.0453414916992188, + "learning_rate": 4.058823529411765e-05, + "loss": 1.2498, + "step": 234 + }, + { + "epoch": 0.0030839869717954626, + "grad_norm": 1.9474778175354004, + "learning_rate": 4.054621848739496e-05, + "loss": 1.018, + "step": 235 + }, + { + "epoch": 0.003097110320611614, + "grad_norm": 1.6720625162124634, + "learning_rate": 4.050420168067227e-05, + "loss": 1.0404, + "step": 236 + }, + { + "epoch": 0.0031102336694277645, + "grad_norm": 1.9552992582321167, + "learning_rate": 4.0462184873949584e-05, + "loss": 0.9688, + "step": 237 + }, + { + "epoch": 0.0031233570182439153, + "grad_norm": 2.513448476791382, + "learning_rate": 4.042016806722689e-05, + "loss": 1.0062, + "step": 238 + }, + { + "epoch": 0.0031364803670600665, + "grad_norm": 2.00137996673584, + "learning_rate": 4.03781512605042e-05, + "loss": 0.9615, + "step": 239 + }, + { + "epoch": 0.003149603715876217, + "grad_norm": 2.3891799449920654, + "learning_rate": 4.033613445378152e-05, + "loss": 0.8978, + "step": 240 + }, + { + "epoch": 0.0031627270646923684, + "grad_norm": 2.0855042934417725, + "learning_rate": 4.029411764705883e-05, + "loss": 0.852, + "step": 241 + }, + { + "epoch": 0.003175850413508519, + "grad_norm": 2.079238176345825, + "learning_rate": 4.0252100840336136e-05, + "loss": 0.9779, + "step": 242 + }, + { + "epoch": 0.00318897376232467, + "grad_norm": 1.7385294437408447, + "learning_rate": 4.021008403361345e-05, + "loss": 0.9277, + "step": 243 + }, + { + "epoch": 0.003202097111140821, + "grad_norm": 1.709777593612671, + "learning_rate": 4.016806722689076e-05, + "loss": 1.1463, + "step": 244 + }, + { + "epoch": 0.003215220459956972, + "grad_norm": 1.9025888442993164, + "learning_rate": 4.012605042016807e-05, + "loss": 1.1001, + "step": 245 + }, + { + "epoch": 0.0032283438087731225, + "grad_norm": 1.8409806489944458, + "learning_rate": 4.0084033613445385e-05, + "loss": 1.0485, + "step": 246 + }, + { + "epoch": 0.0032414671575892737, + "grad_norm": 1.6577681303024292, + "learning_rate": 4.0042016806722694e-05, + "loss": 0.92, + "step": 247 + }, + { + "epoch": 0.0032545905064054245, + "grad_norm": 1.6596322059631348, + "learning_rate": 4e-05, + "loss": 1.1192, + "step": 248 + }, + { + "epoch": 0.0032677138552215756, + "grad_norm": 2.784494161605835, + "learning_rate": 3.995798319327731e-05, + "loss": 1.0173, + "step": 249 + }, + { + "epoch": 0.0032808372040377264, + "grad_norm": 2.500913381576538, + "learning_rate": 3.991596638655463e-05, + "loss": 1.1406, + "step": 250 + }, + { + "epoch": 0.003293960552853877, + "grad_norm": 1.7924182415008545, + "learning_rate": 3.987394957983194e-05, + "loss": 1.0221, + "step": 251 + }, + { + "epoch": 0.0033070839016700283, + "grad_norm": 2.1434240341186523, + "learning_rate": 3.9831932773109246e-05, + "loss": 1.1815, + "step": 252 + }, + { + "epoch": 0.003320207250486179, + "grad_norm": 1.7933728694915771, + "learning_rate": 3.9789915966386555e-05, + "loss": 1.2489, + "step": 253 + }, + { + "epoch": 0.00333333059930233, + "grad_norm": 1.8706622123718262, + "learning_rate": 3.9747899159663864e-05, + "loss": 1.1719, + "step": 254 + }, + { + "epoch": 0.003346453948118481, + "grad_norm": 1.8657987117767334, + "learning_rate": 3.970588235294117e-05, + "loss": 1.0179, + "step": 255 + }, + { + "epoch": 0.0033595772969346317, + "grad_norm": 1.7262569665908813, + "learning_rate": 3.966386554621849e-05, + "loss": 0.975, + "step": 256 + }, + { + "epoch": 0.003372700645750783, + "grad_norm": 1.865169882774353, + "learning_rate": 3.96218487394958e-05, + "loss": 0.8847, + "step": 257 + }, + { + "epoch": 0.0033858239945669336, + "grad_norm": 1.642320156097412, + "learning_rate": 3.957983193277311e-05, + "loss": 1.0513, + "step": 258 + }, + { + "epoch": 0.0033989473433830844, + "grad_norm": 2.1787827014923096, + "learning_rate": 3.953781512605042e-05, + "loss": 0.9024, + "step": 259 + }, + { + "epoch": 0.0034120706921992356, + "grad_norm": 1.8304857015609741, + "learning_rate": 3.949579831932773e-05, + "loss": 1.2173, + "step": 260 + }, + { + "epoch": 0.0034251940410153863, + "grad_norm": 1.6914318799972534, + "learning_rate": 3.945378151260504e-05, + "loss": 1.2612, + "step": 261 + }, + { + "epoch": 0.003438317389831537, + "grad_norm": 2.179232597351074, + "learning_rate": 3.9411764705882356e-05, + "loss": 0.9205, + "step": 262 + }, + { + "epoch": 0.003451440738647688, + "grad_norm": 1.780931830406189, + "learning_rate": 3.9369747899159665e-05, + "loss": 1.0604, + "step": 263 + }, + { + "epoch": 0.003464564087463839, + "grad_norm": 2.5096867084503174, + "learning_rate": 3.9327731092436974e-05, + "loss": 1.3598, + "step": 264 + }, + { + "epoch": 0.00347768743627999, + "grad_norm": 1.8979699611663818, + "learning_rate": 3.928571428571429e-05, + "loss": 1.1095, + "step": 265 + }, + { + "epoch": 0.003490810785096141, + "grad_norm": 1.8904502391815186, + "learning_rate": 3.92436974789916e-05, + "loss": 0.883, + "step": 266 + }, + { + "epoch": 0.0035039341339122916, + "grad_norm": 1.938879132270813, + "learning_rate": 3.920168067226891e-05, + "loss": 0.9752, + "step": 267 + }, + { + "epoch": 0.003517057482728443, + "grad_norm": 2.0486090183258057, + "learning_rate": 3.915966386554622e-05, + "loss": 1.2316, + "step": 268 + }, + { + "epoch": 0.0035301808315445935, + "grad_norm": 1.877327561378479, + "learning_rate": 3.911764705882353e-05, + "loss": 1.117, + "step": 269 + }, + { + "epoch": 0.0035433041803607443, + "grad_norm": 1.7931269407272339, + "learning_rate": 3.907563025210084e-05, + "loss": 1.2002, + "step": 270 + }, + { + "epoch": 0.0035564275291768955, + "grad_norm": 1.9284062385559082, + "learning_rate": 3.903361344537815e-05, + "loss": 1.3203, + "step": 271 + }, + { + "epoch": 0.003569550877993046, + "grad_norm": 2.175030469894409, + "learning_rate": 3.8991596638655467e-05, + "loss": 1.1273, + "step": 272 + }, + { + "epoch": 0.003582674226809197, + "grad_norm": 1.7806055545806885, + "learning_rate": 3.8949579831932776e-05, + "loss": 1.1313, + "step": 273 + }, + { + "epoch": 0.003595797575625348, + "grad_norm": 1.6851775646209717, + "learning_rate": 3.8907563025210084e-05, + "loss": 1.2696, + "step": 274 + }, + { + "epoch": 0.003608920924441499, + "grad_norm": 2.279461622238159, + "learning_rate": 3.88655462184874e-05, + "loss": 1.0396, + "step": 275 + }, + { + "epoch": 0.00362204427325765, + "grad_norm": 2.158630847930908, + "learning_rate": 3.882352941176471e-05, + "loss": 0.8833, + "step": 276 + }, + { + "epoch": 0.003635167622073801, + "grad_norm": 2.1691272258758545, + "learning_rate": 3.878151260504202e-05, + "loss": 1.0293, + "step": 277 + }, + { + "epoch": 0.0036482909708899515, + "grad_norm": 1.891581654548645, + "learning_rate": 3.8739495798319334e-05, + "loss": 0.8519, + "step": 278 + }, + { + "epoch": 0.0036614143197061027, + "grad_norm": 1.8272262811660767, + "learning_rate": 3.869747899159664e-05, + "loss": 0.8203, + "step": 279 + }, + { + "epoch": 0.0036745376685222535, + "grad_norm": 1.7315893173217773, + "learning_rate": 3.865546218487395e-05, + "loss": 1.2325, + "step": 280 + }, + { + "epoch": 0.003687661017338404, + "grad_norm": 1.6825525760650635, + "learning_rate": 3.861344537815126e-05, + "loss": 1.2273, + "step": 281 + }, + { + "epoch": 0.0037007843661545554, + "grad_norm": 2.0799367427825928, + "learning_rate": 3.857142857142858e-05, + "loss": 1.2272, + "step": 282 + }, + { + "epoch": 0.003713907714970706, + "grad_norm": 2.270346164703369, + "learning_rate": 3.8529411764705886e-05, + "loss": 0.998, + "step": 283 + }, + { + "epoch": 0.0037270310637868573, + "grad_norm": 1.8360576629638672, + "learning_rate": 3.8487394957983195e-05, + "loss": 1.044, + "step": 284 + }, + { + "epoch": 0.003740154412603008, + "grad_norm": 1.993119716644287, + "learning_rate": 3.844537815126051e-05, + "loss": 1.1743, + "step": 285 + }, + { + "epoch": 0.003753277761419159, + "grad_norm": 1.6042628288269043, + "learning_rate": 3.840336134453782e-05, + "loss": 1.2144, + "step": 286 + }, + { + "epoch": 0.00376640111023531, + "grad_norm": 1.707850694656372, + "learning_rate": 3.836134453781513e-05, + "loss": 1.1868, + "step": 287 + }, + { + "epoch": 0.0037795244590514607, + "grad_norm": 1.8242610692977905, + "learning_rate": 3.8319327731092444e-05, + "loss": 1.228, + "step": 288 + }, + { + "epoch": 0.0037926478078676115, + "grad_norm": 2.129991054534912, + "learning_rate": 3.8277310924369746e-05, + "loss": 0.9623, + "step": 289 + }, + { + "epoch": 0.0038057711566837626, + "grad_norm": 1.8809386491775513, + "learning_rate": 3.8235294117647055e-05, + "loss": 1.0932, + "step": 290 + }, + { + "epoch": 0.0038188945054999134, + "grad_norm": 1.7827450037002563, + "learning_rate": 3.819327731092437e-05, + "loss": 0.6651, + "step": 291 + }, + { + "epoch": 0.0038320178543160646, + "grad_norm": 1.965325951576233, + "learning_rate": 3.815126050420168e-05, + "loss": 1.0463, + "step": 292 + }, + { + "epoch": 0.0038451412031322153, + "grad_norm": 1.9733631610870361, + "learning_rate": 3.810924369747899e-05, + "loss": 1.1021, + "step": 293 + }, + { + "epoch": 0.003858264551948366, + "grad_norm": 1.8928642272949219, + "learning_rate": 3.8067226890756305e-05, + "loss": 1.0295, + "step": 294 + }, + { + "epoch": 0.0038713879007645172, + "grad_norm": 1.9470313787460327, + "learning_rate": 3.8025210084033614e-05, + "loss": 0.9477, + "step": 295 + }, + { + "epoch": 0.003884511249580668, + "grad_norm": 1.9431153535842896, + "learning_rate": 3.798319327731092e-05, + "loss": 1.2317, + "step": 296 + }, + { + "epoch": 0.0038976345983968187, + "grad_norm": 1.8935089111328125, + "learning_rate": 3.794117647058824e-05, + "loss": 0.7785, + "step": 297 + }, + { + "epoch": 0.0039107579472129695, + "grad_norm": 1.6975841522216797, + "learning_rate": 3.789915966386555e-05, + "loss": 0.7581, + "step": 298 + }, + { + "epoch": 0.003923881296029121, + "grad_norm": 1.9243444204330444, + "learning_rate": 3.785714285714286e-05, + "loss": 0.9032, + "step": 299 + }, + { + "epoch": 0.003937004644845272, + "grad_norm": 1.9851335287094116, + "learning_rate": 3.7815126050420166e-05, + "loss": 1.055, + "step": 300 + }, + { + "epoch": 0.003950127993661423, + "grad_norm": 1.655465841293335, + "learning_rate": 3.777310924369748e-05, + "loss": 0.9511, + "step": 301 + }, + { + "epoch": 0.003963251342477573, + "grad_norm": 1.923627495765686, + "learning_rate": 3.773109243697479e-05, + "loss": 1.3182, + "step": 302 + }, + { + "epoch": 0.003976374691293724, + "grad_norm": 1.884634017944336, + "learning_rate": 3.76890756302521e-05, + "loss": 1.2622, + "step": 303 + }, + { + "epoch": 0.003989498040109876, + "grad_norm": 2.0541749000549316, + "learning_rate": 3.7647058823529415e-05, + "loss": 1.2305, + "step": 304 + }, + { + "epoch": 0.004002621388926026, + "grad_norm": 2.058009147644043, + "learning_rate": 3.7605042016806724e-05, + "loss": 1.0203, + "step": 305 + }, + { + "epoch": 0.004015744737742177, + "grad_norm": 2.1403768062591553, + "learning_rate": 3.756302521008403e-05, + "loss": 1.1616, + "step": 306 + }, + { + "epoch": 0.004028868086558328, + "grad_norm": 2.3530690670013428, + "learning_rate": 3.752100840336135e-05, + "loss": 0.8852, + "step": 307 + }, + { + "epoch": 0.004041991435374479, + "grad_norm": 1.6284990310668945, + "learning_rate": 3.747899159663866e-05, + "loss": 1.0244, + "step": 308 + }, + { + "epoch": 0.004055114784190629, + "grad_norm": 2.03271746635437, + "learning_rate": 3.743697478991597e-05, + "loss": 1.0321, + "step": 309 + }, + { + "epoch": 0.004068238133006781, + "grad_norm": 2.220259666442871, + "learning_rate": 3.739495798319328e-05, + "loss": 0.9091, + "step": 310 + }, + { + "epoch": 0.004081361481822932, + "grad_norm": 1.762771487236023, + "learning_rate": 3.735294117647059e-05, + "loss": 1.0977, + "step": 311 + }, + { + "epoch": 0.0040944848306390825, + "grad_norm": 1.9179171323776245, + "learning_rate": 3.73109243697479e-05, + "loss": 1.238, + "step": 312 + }, + { + "epoch": 0.004107608179455233, + "grad_norm": 1.933432936668396, + "learning_rate": 3.726890756302521e-05, + "loss": 0.9684, + "step": 313 + }, + { + "epoch": 0.004120731528271384, + "grad_norm": 1.7434195280075073, + "learning_rate": 3.7226890756302525e-05, + "loss": 1.0966, + "step": 314 + }, + { + "epoch": 0.004133854877087536, + "grad_norm": 1.692446231842041, + "learning_rate": 3.7184873949579834e-05, + "loss": 0.8348, + "step": 315 + }, + { + "epoch": 0.004146978225903686, + "grad_norm": 2.00734281539917, + "learning_rate": 3.7142857142857143e-05, + "loss": 0.9317, + "step": 316 + }, + { + "epoch": 0.004160101574719837, + "grad_norm": 2.176311492919922, + "learning_rate": 3.710084033613446e-05, + "loss": 1.1143, + "step": 317 + }, + { + "epoch": 0.004173224923535988, + "grad_norm": 1.5961421728134155, + "learning_rate": 3.705882352941177e-05, + "loss": 1.1549, + "step": 318 + }, + { + "epoch": 0.0041863482723521386, + "grad_norm": 1.9222772121429443, + "learning_rate": 3.701680672268908e-05, + "loss": 1.0957, + "step": 319 + }, + { + "epoch": 0.004199471621168289, + "grad_norm": 1.6824790239334106, + "learning_rate": 3.697478991596639e-05, + "loss": 0.9843, + "step": 320 + }, + { + "epoch": 0.004212594969984441, + "grad_norm": 2.092710256576538, + "learning_rate": 3.69327731092437e-05, + "loss": 0.9426, + "step": 321 + }, + { + "epoch": 0.004225718318800592, + "grad_norm": 1.9568061828613281, + "learning_rate": 3.689075630252101e-05, + "loss": 1.1585, + "step": 322 + }, + { + "epoch": 0.004238841667616742, + "grad_norm": 1.711948275566101, + "learning_rate": 3.684873949579833e-05, + "loss": 1.0928, + "step": 323 + }, + { + "epoch": 0.004251965016432893, + "grad_norm": 2.0425591468811035, + "learning_rate": 3.6806722689075636e-05, + "loss": 1.0866, + "step": 324 + }, + { + "epoch": 0.004265088365249044, + "grad_norm": 1.9059627056121826, + "learning_rate": 3.6764705882352945e-05, + "loss": 1.1211, + "step": 325 + }, + { + "epoch": 0.0042782117140651955, + "grad_norm": 1.7678598165512085, + "learning_rate": 3.6722689075630254e-05, + "loss": 0.9953, + "step": 326 + }, + { + "epoch": 0.004291335062881346, + "grad_norm": 1.829349160194397, + "learning_rate": 3.668067226890756e-05, + "loss": 1.2049, + "step": 327 + }, + { + "epoch": 0.004304458411697497, + "grad_norm": 1.9673535823822021, + "learning_rate": 3.663865546218487e-05, + "loss": 1.1124, + "step": 328 + }, + { + "epoch": 0.004317581760513648, + "grad_norm": 1.9089158773422241, + "learning_rate": 3.659663865546219e-05, + "loss": 0.9055, + "step": 329 + }, + { + "epoch": 0.0043307051093297985, + "grad_norm": 1.805672287940979, + "learning_rate": 3.6554621848739496e-05, + "loss": 0.815, + "step": 330 + }, + { + "epoch": 0.00434382845814595, + "grad_norm": 2.055886745452881, + "learning_rate": 3.6512605042016805e-05, + "loss": 0.923, + "step": 331 + }, + { + "epoch": 0.004356951806962101, + "grad_norm": 2.1708016395568848, + "learning_rate": 3.6470588235294114e-05, + "loss": 1.1073, + "step": 332 + }, + { + "epoch": 0.004370075155778252, + "grad_norm": 1.728820562362671, + "learning_rate": 3.642857142857143e-05, + "loss": 0.9458, + "step": 333 + }, + { + "epoch": 0.004383198504594402, + "grad_norm": 1.920433521270752, + "learning_rate": 3.638655462184874e-05, + "loss": 1.3087, + "step": 334 + }, + { + "epoch": 0.004396321853410553, + "grad_norm": 2.3520448207855225, + "learning_rate": 3.634453781512605e-05, + "loss": 1.0046, + "step": 335 + }, + { + "epoch": 0.004409445202226704, + "grad_norm": 1.5997459888458252, + "learning_rate": 3.6302521008403364e-05, + "loss": 1.0329, + "step": 336 + }, + { + "epoch": 0.004422568551042855, + "grad_norm": 1.7262095212936401, + "learning_rate": 3.626050420168067e-05, + "loss": 0.7392, + "step": 337 + }, + { + "epoch": 0.004435691899859006, + "grad_norm": 1.6960225105285645, + "learning_rate": 3.621848739495798e-05, + "loss": 0.8531, + "step": 338 + }, + { + "epoch": 0.004448815248675157, + "grad_norm": 2.0684216022491455, + "learning_rate": 3.61764705882353e-05, + "loss": 1.1234, + "step": 339 + }, + { + "epoch": 0.004461938597491308, + "grad_norm": 2.0957415103912354, + "learning_rate": 3.613445378151261e-05, + "loss": 1.0791, + "step": 340 + }, + { + "epoch": 0.004475061946307458, + "grad_norm": 1.9028066396713257, + "learning_rate": 3.6092436974789916e-05, + "loss": 1.1277, + "step": 341 + }, + { + "epoch": 0.00448818529512361, + "grad_norm": 1.9382456541061401, + "learning_rate": 3.605042016806723e-05, + "loss": 1.0221, + "step": 342 + }, + { + "epoch": 0.004501308643939761, + "grad_norm": 1.6922097206115723, + "learning_rate": 3.600840336134454e-05, + "loss": 1.0441, + "step": 343 + }, + { + "epoch": 0.0045144319927559115, + "grad_norm": 1.9319812059402466, + "learning_rate": 3.596638655462185e-05, + "loss": 1.1206, + "step": 344 + }, + { + "epoch": 0.004527555341572062, + "grad_norm": 1.809410572052002, + "learning_rate": 3.592436974789916e-05, + "loss": 1.3128, + "step": 345 + }, + { + "epoch": 0.004540678690388213, + "grad_norm": 2.202211380004883, + "learning_rate": 3.5882352941176474e-05, + "loss": 1.2144, + "step": 346 + }, + { + "epoch": 0.004553802039204364, + "grad_norm": 1.9447896480560303, + "learning_rate": 3.584033613445378e-05, + "loss": 0.931, + "step": 347 + }, + { + "epoch": 0.004566925388020515, + "grad_norm": 1.7502014636993408, + "learning_rate": 3.579831932773109e-05, + "loss": 1.0937, + "step": 348 + }, + { + "epoch": 0.004580048736836666, + "grad_norm": 2.021968364715576, + "learning_rate": 3.575630252100841e-05, + "loss": 1.1214, + "step": 349 + }, + { + "epoch": 0.004593172085652817, + "grad_norm": 1.980770230293274, + "learning_rate": 3.571428571428572e-05, + "loss": 0.8333, + "step": 350 + }, + { + "epoch": 0.004606295434468968, + "grad_norm": 1.9700994491577148, + "learning_rate": 3.5672268907563026e-05, + "loss": 1.2036, + "step": 351 + }, + { + "epoch": 0.004619418783285118, + "grad_norm": 1.9649808406829834, + "learning_rate": 3.563025210084034e-05, + "loss": 0.8695, + "step": 352 + }, + { + "epoch": 0.00463254213210127, + "grad_norm": 1.7544773817062378, + "learning_rate": 3.558823529411765e-05, + "loss": 0.9898, + "step": 353 + }, + { + "epoch": 0.004645665480917421, + "grad_norm": 1.497639536857605, + "learning_rate": 3.554621848739496e-05, + "loss": 1.0418, + "step": 354 + }, + { + "epoch": 0.004658788829733571, + "grad_norm": 1.8401079177856445, + "learning_rate": 3.5504201680672275e-05, + "loss": 0.7553, + "step": 355 + }, + { + "epoch": 0.004671912178549722, + "grad_norm": 1.866067886352539, + "learning_rate": 3.5462184873949584e-05, + "loss": 1.23, + "step": 356 + }, + { + "epoch": 0.004685035527365873, + "grad_norm": 2.1467809677124023, + "learning_rate": 3.5420168067226893e-05, + "loss": 1.021, + "step": 357 + }, + { + "epoch": 0.0046981588761820245, + "grad_norm": 1.55268132686615, + "learning_rate": 3.53781512605042e-05, + "loss": 0.7216, + "step": 358 + }, + { + "epoch": 0.004711282224998175, + "grad_norm": 2.081575632095337, + "learning_rate": 3.533613445378152e-05, + "loss": 1.0118, + "step": 359 + }, + { + "epoch": 0.004724405573814326, + "grad_norm": 2.164594888687134, + "learning_rate": 3.529411764705883e-05, + "loss": 1.1738, + "step": 360 + }, + { + "epoch": 0.004737528922630477, + "grad_norm": 2.3375766277313232, + "learning_rate": 3.5252100840336136e-05, + "loss": 1.2937, + "step": 361 + }, + { + "epoch": 0.0047506522714466275, + "grad_norm": 1.8707530498504639, + "learning_rate": 3.521008403361345e-05, + "loss": 0.7568, + "step": 362 + }, + { + "epoch": 0.004763775620262778, + "grad_norm": 1.7898566722869873, + "learning_rate": 3.516806722689076e-05, + "loss": 1.0675, + "step": 363 + }, + { + "epoch": 0.00477689896907893, + "grad_norm": 2.018583059310913, + "learning_rate": 3.512605042016806e-05, + "loss": 1.2629, + "step": 364 + }, + { + "epoch": 0.004790022317895081, + "grad_norm": 1.7101985216140747, + "learning_rate": 3.508403361344538e-05, + "loss": 1.2771, + "step": 365 + }, + { + "epoch": 0.004803145666711231, + "grad_norm": 1.6018425226211548, + "learning_rate": 3.504201680672269e-05, + "loss": 1.3823, + "step": 366 + }, + { + "epoch": 0.004816269015527382, + "grad_norm": 1.9749130010604858, + "learning_rate": 3.5e-05, + "loss": 0.8286, + "step": 367 + }, + { + "epoch": 0.004829392364343533, + "grad_norm": 1.8984951972961426, + "learning_rate": 3.495798319327731e-05, + "loss": 0.8364, + "step": 368 + }, + { + "epoch": 0.0048425157131596844, + "grad_norm": 1.8086698055267334, + "learning_rate": 3.491596638655462e-05, + "loss": 0.9319, + "step": 369 + }, + { + "epoch": 0.004855639061975835, + "grad_norm": 1.8288997411727905, + "learning_rate": 3.487394957983193e-05, + "loss": 1.1326, + "step": 370 + }, + { + "epoch": 0.004868762410791986, + "grad_norm": 2.247119903564453, + "learning_rate": 3.4831932773109246e-05, + "loss": 0.9275, + "step": 371 + }, + { + "epoch": 0.004881885759608137, + "grad_norm": 1.9033137559890747, + "learning_rate": 3.4789915966386555e-05, + "loss": 1.0881, + "step": 372 + }, + { + "epoch": 0.004895009108424287, + "grad_norm": 2.0219953060150146, + "learning_rate": 3.4747899159663864e-05, + "loss": 0.787, + "step": 373 + }, + { + "epoch": 0.004908132457240439, + "grad_norm": 1.8668859004974365, + "learning_rate": 3.470588235294118e-05, + "loss": 0.7476, + "step": 374 + }, + { + "epoch": 0.00492125580605659, + "grad_norm": 1.9940170049667358, + "learning_rate": 3.466386554621849e-05, + "loss": 1.0816, + "step": 375 + }, + { + "epoch": 0.0049343791548727405, + "grad_norm": 1.749637246131897, + "learning_rate": 3.46218487394958e-05, + "loss": 1.0829, + "step": 376 + }, + { + "epoch": 0.004947502503688891, + "grad_norm": 1.9126695394515991, + "learning_rate": 3.457983193277311e-05, + "loss": 0.9706, + "step": 377 + }, + { + "epoch": 0.004960625852505042, + "grad_norm": 2.0018599033355713, + "learning_rate": 3.453781512605042e-05, + "loss": 1.0444, + "step": 378 + }, + { + "epoch": 0.004973749201321193, + "grad_norm": 1.6699880361557007, + "learning_rate": 3.449579831932773e-05, + "loss": 0.9605, + "step": 379 + }, + { + "epoch": 0.004986872550137344, + "grad_norm": 2.0638937950134277, + "learning_rate": 3.445378151260504e-05, + "loss": 0.8426, + "step": 380 + }, + { + "epoch": 0.004999995898953495, + "grad_norm": 1.8554656505584717, + "learning_rate": 3.441176470588236e-05, + "loss": 1.2654, + "step": 381 + }, + { + "epoch": 0.005013119247769646, + "grad_norm": 2.101221799850464, + "learning_rate": 3.4369747899159666e-05, + "loss": 0.9724, + "step": 382 + }, + { + "epoch": 0.005026242596585797, + "grad_norm": 2.00722336769104, + "learning_rate": 3.4327731092436975e-05, + "loss": 0.9005, + "step": 383 + }, + { + "epoch": 0.005039365945401947, + "grad_norm": 1.6129575967788696, + "learning_rate": 3.428571428571429e-05, + "loss": 1.0246, + "step": 384 + }, + { + "epoch": 0.005052489294218099, + "grad_norm": 1.8390120267868042, + "learning_rate": 3.42436974789916e-05, + "loss": 1.2598, + "step": 385 + }, + { + "epoch": 0.00506561264303425, + "grad_norm": 2.0762507915496826, + "learning_rate": 3.420168067226891e-05, + "loss": 1.3111, + "step": 386 + }, + { + "epoch": 0.0050787359918504004, + "grad_norm": 1.7343683242797852, + "learning_rate": 3.4159663865546224e-05, + "loss": 0.8649, + "step": 387 + }, + { + "epoch": 0.005091859340666551, + "grad_norm": 2.0712454319000244, + "learning_rate": 3.411764705882353e-05, + "loss": 1.2333, + "step": 388 + }, + { + "epoch": 0.005104982689482702, + "grad_norm": 1.7495871782302856, + "learning_rate": 3.407563025210084e-05, + "loss": 1.2728, + "step": 389 + }, + { + "epoch": 0.005118106038298853, + "grad_norm": 1.8208390474319458, + "learning_rate": 3.403361344537815e-05, + "loss": 0.873, + "step": 390 + }, + { + "epoch": 0.005131229387115004, + "grad_norm": 1.7520668506622314, + "learning_rate": 3.399159663865547e-05, + "loss": 0.8759, + "step": 391 + }, + { + "epoch": 0.005144352735931155, + "grad_norm": 1.6596717834472656, + "learning_rate": 3.3949579831932776e-05, + "loss": 1.3544, + "step": 392 + }, + { + "epoch": 0.005157476084747306, + "grad_norm": 1.497276782989502, + "learning_rate": 3.3907563025210085e-05, + "loss": 1.13, + "step": 393 + }, + { + "epoch": 0.0051705994335634565, + "grad_norm": 2.0411555767059326, + "learning_rate": 3.38655462184874e-05, + "loss": 0.9337, + "step": 394 + }, + { + "epoch": 0.005183722782379607, + "grad_norm": 2.373502016067505, + "learning_rate": 3.382352941176471e-05, + "loss": 1.022, + "step": 395 + }, + { + "epoch": 0.005196846131195759, + "grad_norm": 1.9800091981887817, + "learning_rate": 3.378151260504202e-05, + "loss": 1.0296, + "step": 396 + }, + { + "epoch": 0.00520996948001191, + "grad_norm": 2.000948667526245, + "learning_rate": 3.3739495798319334e-05, + "loss": 0.9769, + "step": 397 + }, + { + "epoch": 0.00522309282882806, + "grad_norm": 1.7864891290664673, + "learning_rate": 3.3697478991596643e-05, + "loss": 0.9527, + "step": 398 + }, + { + "epoch": 0.005236216177644211, + "grad_norm": 1.7589308023452759, + "learning_rate": 3.365546218487395e-05, + "loss": 0.8136, + "step": 399 + }, + { + "epoch": 0.005249339526460362, + "grad_norm": 1.9042621850967407, + "learning_rate": 3.361344537815127e-05, + "loss": 1.371, + "step": 400 + }, + { + "epoch": 0.0052624628752765135, + "grad_norm": 2.0386345386505127, + "learning_rate": 3.357142857142857e-05, + "loss": 1.282, + "step": 401 + }, + { + "epoch": 0.005275586224092664, + "grad_norm": 1.938333511352539, + "learning_rate": 3.352941176470588e-05, + "loss": 0.9357, + "step": 402 + }, + { + "epoch": 0.005288709572908815, + "grad_norm": 2.111893892288208, + "learning_rate": 3.3487394957983195e-05, + "loss": 0.9913, + "step": 403 + }, + { + "epoch": 0.005301832921724966, + "grad_norm": 1.8991844654083252, + "learning_rate": 3.3445378151260504e-05, + "loss": 1.0779, + "step": 404 + }, + { + "epoch": 0.0053149562705411164, + "grad_norm": 2.0768351554870605, + "learning_rate": 3.340336134453781e-05, + "loss": 1.2548, + "step": 405 + }, + { + "epoch": 0.005328079619357267, + "grad_norm": 1.9904977083206177, + "learning_rate": 3.336134453781513e-05, + "loss": 0.9451, + "step": 406 + }, + { + "epoch": 0.005341202968173419, + "grad_norm": 1.7511869668960571, + "learning_rate": 3.331932773109244e-05, + "loss": 1.3958, + "step": 407 + }, + { + "epoch": 0.0053543263169895695, + "grad_norm": 1.9327552318572998, + "learning_rate": 3.327731092436975e-05, + "loss": 1.1969, + "step": 408 + }, + { + "epoch": 0.00536744966580572, + "grad_norm": 1.969369888305664, + "learning_rate": 3.3235294117647056e-05, + "loss": 0.9569, + "step": 409 + }, + { + "epoch": 0.005380573014621871, + "grad_norm": 1.5960437059402466, + "learning_rate": 3.319327731092437e-05, + "loss": 1.0643, + "step": 410 + }, + { + "epoch": 0.005393696363438022, + "grad_norm": 2.1960508823394775, + "learning_rate": 3.315126050420168e-05, + "loss": 0.9549, + "step": 411 + }, + { + "epoch": 0.005406819712254173, + "grad_norm": 2.2554352283477783, + "learning_rate": 3.310924369747899e-05, + "loss": 0.868, + "step": 412 + }, + { + "epoch": 0.005419943061070324, + "grad_norm": 1.931188941001892, + "learning_rate": 3.3067226890756305e-05, + "loss": 0.9063, + "step": 413 + }, + { + "epoch": 0.005433066409886475, + "grad_norm": 2.1585495471954346, + "learning_rate": 3.3025210084033614e-05, + "loss": 0.8968, + "step": 414 + }, + { + "epoch": 0.005446189758702626, + "grad_norm": 2.0901172161102295, + "learning_rate": 3.2983193277310923e-05, + "loss": 1.1077, + "step": 415 + }, + { + "epoch": 0.005459313107518776, + "grad_norm": 1.8528449535369873, + "learning_rate": 3.294117647058824e-05, + "loss": 1.0224, + "step": 416 + }, + { + "epoch": 0.005472436456334928, + "grad_norm": 2.231144428253174, + "learning_rate": 3.289915966386555e-05, + "loss": 1.0864, + "step": 417 + }, + { + "epoch": 0.005485559805151079, + "grad_norm": 1.7245928049087524, + "learning_rate": 3.285714285714286e-05, + "loss": 1.0887, + "step": 418 + }, + { + "epoch": 0.0054986831539672295, + "grad_norm": 1.8599853515625, + "learning_rate": 3.281512605042017e-05, + "loss": 0.7638, + "step": 419 + }, + { + "epoch": 0.00551180650278338, + "grad_norm": 1.6251364946365356, + "learning_rate": 3.277310924369748e-05, + "loss": 1.0903, + "step": 420 + }, + { + "epoch": 0.005524929851599531, + "grad_norm": 1.7949484586715698, + "learning_rate": 3.273109243697479e-05, + "loss": 1.454, + "step": 421 + }, + { + "epoch": 0.005538053200415682, + "grad_norm": 1.6007757186889648, + "learning_rate": 3.26890756302521e-05, + "loss": 1.0825, + "step": 422 + }, + { + "epoch": 0.005551176549231833, + "grad_norm": 1.8195713758468628, + "learning_rate": 3.2647058823529416e-05, + "loss": 0.7635, + "step": 423 + }, + { + "epoch": 0.005564299898047984, + "grad_norm": 2.4276785850524902, + "learning_rate": 3.2605042016806725e-05, + "loss": 1.28, + "step": 424 + }, + { + "epoch": 0.005577423246864135, + "grad_norm": 1.8293254375457764, + "learning_rate": 3.2563025210084034e-05, + "loss": 1.1156, + "step": 425 + }, + { + "epoch": 0.0055905465956802855, + "grad_norm": 1.8416686058044434, + "learning_rate": 3.252100840336135e-05, + "loss": 1.0777, + "step": 426 + }, + { + "epoch": 0.005603669944496436, + "grad_norm": 1.7739605903625488, + "learning_rate": 3.247899159663866e-05, + "loss": 0.7509, + "step": 427 + }, + { + "epoch": 0.005616793293312588, + "grad_norm": 1.3967030048370361, + "learning_rate": 3.243697478991597e-05, + "loss": 1.0822, + "step": 428 + }, + { + "epoch": 0.005629916642128739, + "grad_norm": 1.643563985824585, + "learning_rate": 3.239495798319328e-05, + "loss": 1.0556, + "step": 429 + }, + { + "epoch": 0.005643039990944889, + "grad_norm": 1.8893026113510132, + "learning_rate": 3.235294117647059e-05, + "loss": 0.9656, + "step": 430 + }, + { + "epoch": 0.00565616333976104, + "grad_norm": 2.050086736679077, + "learning_rate": 3.23109243697479e-05, + "loss": 1.2009, + "step": 431 + }, + { + "epoch": 0.005669286688577191, + "grad_norm": 1.813232660293579, + "learning_rate": 3.226890756302522e-05, + "loss": 0.9385, + "step": 432 + }, + { + "epoch": 0.005682410037393342, + "grad_norm": 2.043942928314209, + "learning_rate": 3.2226890756302526e-05, + "loss": 1.2035, + "step": 433 + }, + { + "epoch": 0.005695533386209493, + "grad_norm": 2.143899917602539, + "learning_rate": 3.2184873949579835e-05, + "loss": 0.9407, + "step": 434 + }, + { + "epoch": 0.005708656735025644, + "grad_norm": 2.0677599906921387, + "learning_rate": 3.2142857142857144e-05, + "loss": 0.9974, + "step": 435 + }, + { + "epoch": 0.005721780083841795, + "grad_norm": 1.8135740756988525, + "learning_rate": 3.210084033613446e-05, + "loss": 0.9685, + "step": 436 + }, + { + "epoch": 0.0057349034326579455, + "grad_norm": 1.9901634454727173, + "learning_rate": 3.205882352941177e-05, + "loss": 0.7587, + "step": 437 + }, + { + "epoch": 0.005748026781474096, + "grad_norm": 2.103733777999878, + "learning_rate": 3.201680672268908e-05, + "loss": 1.0126, + "step": 438 + }, + { + "epoch": 0.005761150130290248, + "grad_norm": 2.0034232139587402, + "learning_rate": 3.197478991596639e-05, + "loss": 1.0483, + "step": 439 + }, + { + "epoch": 0.0057742734791063986, + "grad_norm": 2.2869062423706055, + "learning_rate": 3.1932773109243696e-05, + "loss": 0.6886, + "step": 440 + }, + { + "epoch": 0.005787396827922549, + "grad_norm": 1.5798542499542236, + "learning_rate": 3.1890756302521005e-05, + "loss": 0.8474, + "step": 441 + }, + { + "epoch": 0.0058005201767387, + "grad_norm": 1.913794755935669, + "learning_rate": 3.184873949579832e-05, + "loss": 0.8935, + "step": 442 + }, + { + "epoch": 0.005813643525554851, + "grad_norm": 1.8879010677337646, + "learning_rate": 3.180672268907563e-05, + "loss": 1.1235, + "step": 443 + }, + { + "epoch": 0.005826766874371002, + "grad_norm": 2.1588127613067627, + "learning_rate": 3.176470588235294e-05, + "loss": 1.1129, + "step": 444 + }, + { + "epoch": 0.005839890223187153, + "grad_norm": 1.6797983646392822, + "learning_rate": 3.1722689075630254e-05, + "loss": 1.2736, + "step": 445 + }, + { + "epoch": 0.005853013572003304, + "grad_norm": 1.5610647201538086, + "learning_rate": 3.168067226890756e-05, + "loss": 1.0033, + "step": 446 + }, + { + "epoch": 0.005866136920819455, + "grad_norm": 1.567574381828308, + "learning_rate": 3.163865546218487e-05, + "loss": 1.0611, + "step": 447 + }, + { + "epoch": 0.005879260269635605, + "grad_norm": 1.9480379819869995, + "learning_rate": 3.159663865546219e-05, + "loss": 1.3461, + "step": 448 + }, + { + "epoch": 0.005892383618451756, + "grad_norm": 1.6626474857330322, + "learning_rate": 3.15546218487395e-05, + "loss": 0.934, + "step": 449 + }, + { + "epoch": 0.005905506967267908, + "grad_norm": 1.9532502889633179, + "learning_rate": 3.1512605042016806e-05, + "loss": 1.2752, + "step": 450 + }, + { + "epoch": 0.0059186303160840585, + "grad_norm": 1.858870267868042, + "learning_rate": 3.147058823529412e-05, + "loss": 0.8752, + "step": 451 + }, + { + "epoch": 0.005931753664900209, + "grad_norm": 1.7586513757705688, + "learning_rate": 3.142857142857143e-05, + "loss": 1.3883, + "step": 452 + }, + { + "epoch": 0.00594487701371636, + "grad_norm": 1.7400894165039062, + "learning_rate": 3.138655462184874e-05, + "loss": 1.1679, + "step": 453 + }, + { + "epoch": 0.005958000362532511, + "grad_norm": 1.8386495113372803, + "learning_rate": 3.134453781512605e-05, + "loss": 1.1425, + "step": 454 + }, + { + "epoch": 0.005971123711348662, + "grad_norm": 1.9906026124954224, + "learning_rate": 3.1302521008403364e-05, + "loss": 0.9552, + "step": 455 + }, + { + "epoch": 0.005984247060164813, + "grad_norm": 2.0481302738189697, + "learning_rate": 3.1260504201680673e-05, + "loss": 1.0198, + "step": 456 + }, + { + "epoch": 0.005997370408980964, + "grad_norm": 1.8666774034500122, + "learning_rate": 3.121848739495798e-05, + "loss": 0.9435, + "step": 457 + }, + { + "epoch": 0.0060104937577971146, + "grad_norm": 1.7503705024719238, + "learning_rate": 3.11764705882353e-05, + "loss": 1.238, + "step": 458 + }, + { + "epoch": 0.006023617106613265, + "grad_norm": 1.548405647277832, + "learning_rate": 3.113445378151261e-05, + "loss": 0.9299, + "step": 459 + }, + { + "epoch": 0.006036740455429417, + "grad_norm": 1.5533720254898071, + "learning_rate": 3.1092436974789916e-05, + "loss": 1.0623, + "step": 460 + }, + { + "epoch": 0.006049863804245568, + "grad_norm": 1.8287532329559326, + "learning_rate": 3.105042016806723e-05, + "loss": 1.0802, + "step": 461 + }, + { + "epoch": 0.006062987153061718, + "grad_norm": 1.733933448791504, + "learning_rate": 3.100840336134454e-05, + "loss": 1.527, + "step": 462 + }, + { + "epoch": 0.006076110501877869, + "grad_norm": 1.5526071786880493, + "learning_rate": 3.096638655462185e-05, + "loss": 1.0319, + "step": 463 + }, + { + "epoch": 0.00608923385069402, + "grad_norm": 1.6806490421295166, + "learning_rate": 3.0924369747899166e-05, + "loss": 1.5237, + "step": 464 + }, + { + "epoch": 0.006102357199510171, + "grad_norm": 3.8694965839385986, + "learning_rate": 3.0882352941176475e-05, + "loss": 1.1718, + "step": 465 + }, + { + "epoch": 0.006115480548326322, + "grad_norm": 1.8779933452606201, + "learning_rate": 3.0840336134453784e-05, + "loss": 0.6304, + "step": 466 + }, + { + "epoch": 0.006128603897142473, + "grad_norm": 1.7891600131988525, + "learning_rate": 3.079831932773109e-05, + "loss": 1.1527, + "step": 467 + }, + { + "epoch": 0.006141727245958624, + "grad_norm": 1.9199869632720947, + "learning_rate": 3.075630252100841e-05, + "loss": 1.1132, + "step": 468 + }, + { + "epoch": 0.0061548505947747745, + "grad_norm": 1.7727890014648438, + "learning_rate": 3.071428571428572e-05, + "loss": 0.976, + "step": 469 + }, + { + "epoch": 0.006167973943590925, + "grad_norm": 2.1416878700256348, + "learning_rate": 3.0672268907563026e-05, + "loss": 1.0768, + "step": 470 + }, + { + "epoch": 0.006181097292407077, + "grad_norm": 2.037363290786743, + "learning_rate": 3.063025210084034e-05, + "loss": 1.2424, + "step": 471 + }, + { + "epoch": 0.006194220641223228, + "grad_norm": 1.7760509252548218, + "learning_rate": 3.058823529411765e-05, + "loss": 0.881, + "step": 472 + }, + { + "epoch": 0.006207343990039378, + "grad_norm": 1.8409169912338257, + "learning_rate": 3.054621848739496e-05, + "loss": 0.9362, + "step": 473 + }, + { + "epoch": 0.006220467338855529, + "grad_norm": 1.7716394662857056, + "learning_rate": 3.0504201680672273e-05, + "loss": 0.9846, + "step": 474 + }, + { + "epoch": 0.00623359068767168, + "grad_norm": 1.9643959999084473, + "learning_rate": 3.0462184873949578e-05, + "loss": 1.2064, + "step": 475 + }, + { + "epoch": 0.0062467140364878306, + "grad_norm": 1.6986589431762695, + "learning_rate": 3.042016806722689e-05, + "loss": 0.7699, + "step": 476 + }, + { + "epoch": 0.006259837385303982, + "grad_norm": 1.9324336051940918, + "learning_rate": 3.03781512605042e-05, + "loss": 1.0145, + "step": 477 + }, + { + "epoch": 0.006272960734120133, + "grad_norm": 1.7184754610061646, + "learning_rate": 3.0336134453781512e-05, + "loss": 1.2643, + "step": 478 + }, + { + "epoch": 0.006286084082936284, + "grad_norm": 1.7461143732070923, + "learning_rate": 3.0294117647058824e-05, + "loss": 1.1703, + "step": 479 + }, + { + "epoch": 0.006299207431752434, + "grad_norm": 1.843173861503601, + "learning_rate": 3.0252100840336133e-05, + "loss": 0.9474, + "step": 480 + }, + { + "epoch": 0.006312330780568585, + "grad_norm": 2.223266839981079, + "learning_rate": 3.0210084033613446e-05, + "loss": 1.1058, + "step": 481 + }, + { + "epoch": 0.006325454129384737, + "grad_norm": 2.2394039630889893, + "learning_rate": 3.0168067226890755e-05, + "loss": 0.9657, + "step": 482 + }, + { + "epoch": 0.0063385774782008875, + "grad_norm": 1.9419907331466675, + "learning_rate": 3.0126050420168067e-05, + "loss": 1.1787, + "step": 483 + }, + { + "epoch": 0.006351700827017038, + "grad_norm": 1.90862238407135, + "learning_rate": 3.008403361344538e-05, + "loss": 1.5107, + "step": 484 + }, + { + "epoch": 0.006364824175833189, + "grad_norm": 1.8972816467285156, + "learning_rate": 3.004201680672269e-05, + "loss": 0.9348, + "step": 485 + }, + { + "epoch": 0.00637794752464934, + "grad_norm": 1.6643794775009155, + "learning_rate": 3e-05, + "loss": 0.5877, + "step": 486 + }, + { + "epoch": 0.006391070873465491, + "grad_norm": 2.072737216949463, + "learning_rate": 2.995798319327731e-05, + "loss": 1.0851, + "step": 487 + }, + { + "epoch": 0.006404194222281642, + "grad_norm": 2.223226308822632, + "learning_rate": 2.9915966386554622e-05, + "loss": 1.1913, + "step": 488 + }, + { + "epoch": 0.006417317571097793, + "grad_norm": 1.7678091526031494, + "learning_rate": 2.9873949579831935e-05, + "loss": 0.9093, + "step": 489 + }, + { + "epoch": 0.006430440919913944, + "grad_norm": 1.7829846143722534, + "learning_rate": 2.9831932773109244e-05, + "loss": 1.1833, + "step": 490 + }, + { + "epoch": 0.006443564268730094, + "grad_norm": 1.6032719612121582, + "learning_rate": 2.9789915966386556e-05, + "loss": 0.6613, + "step": 491 + }, + { + "epoch": 0.006456687617546245, + "grad_norm": 1.7531033754348755, + "learning_rate": 2.9747899159663868e-05, + "loss": 1.0188, + "step": 492 + }, + { + "epoch": 0.006469810966362397, + "grad_norm": 2.0264689922332764, + "learning_rate": 2.9705882352941177e-05, + "loss": 0.9985, + "step": 493 + }, + { + "epoch": 0.006482934315178547, + "grad_norm": 1.5329761505126953, + "learning_rate": 2.966386554621849e-05, + "loss": 1.1659, + "step": 494 + }, + { + "epoch": 0.006496057663994698, + "grad_norm": 2.2568962574005127, + "learning_rate": 2.96218487394958e-05, + "loss": 1.2176, + "step": 495 + }, + { + "epoch": 0.006509181012810849, + "grad_norm": 1.817195177078247, + "learning_rate": 2.957983193277311e-05, + "loss": 0.704, + "step": 496 + }, + { + "epoch": 0.006522304361627, + "grad_norm": 1.8540563583374023, + "learning_rate": 2.9537815126050423e-05, + "loss": 0.9465, + "step": 497 + }, + { + "epoch": 0.006535427710443151, + "grad_norm": 1.8276430368423462, + "learning_rate": 2.9495798319327732e-05, + "loss": 1.1567, + "step": 498 + }, + { + "epoch": 0.006548551059259302, + "grad_norm": 1.963805913925171, + "learning_rate": 2.9453781512605045e-05, + "loss": 0.9897, + "step": 499 + }, + { + "epoch": 0.006561674408075453, + "grad_norm": 1.851785659790039, + "learning_rate": 2.9411764705882354e-05, + "loss": 1.0783, + "step": 500 + }, + { + "epoch": 0.0065747977568916035, + "grad_norm": 1.7229440212249756, + "learning_rate": 2.9369747899159666e-05, + "loss": 0.884, + "step": 501 + }, + { + "epoch": 0.006587921105707754, + "grad_norm": 1.7793407440185547, + "learning_rate": 2.932773109243698e-05, + "loss": 1.2513, + "step": 502 + }, + { + "epoch": 0.006601044454523905, + "grad_norm": 1.5506298542022705, + "learning_rate": 2.9285714285714288e-05, + "loss": 0.9357, + "step": 503 + }, + { + "epoch": 0.006614167803340057, + "grad_norm": 1.7655850648880005, + "learning_rate": 2.92436974789916e-05, + "loss": 0.9378, + "step": 504 + }, + { + "epoch": 0.006627291152156207, + "grad_norm": 1.7562974691390991, + "learning_rate": 2.9201680672268912e-05, + "loss": 1.1196, + "step": 505 + }, + { + "epoch": 0.006640414500972358, + "grad_norm": 1.629323959350586, + "learning_rate": 2.915966386554622e-05, + "loss": 1.1883, + "step": 506 + }, + { + "epoch": 0.006653537849788509, + "grad_norm": 1.8579301834106445, + "learning_rate": 2.9117647058823534e-05, + "loss": 1.2, + "step": 507 + }, + { + "epoch": 0.00666666119860466, + "grad_norm": 1.8027002811431885, + "learning_rate": 2.9075630252100843e-05, + "loss": 0.9524, + "step": 508 + }, + { + "epoch": 0.006679784547420811, + "grad_norm": 1.908895492553711, + "learning_rate": 2.9033613445378155e-05, + "loss": 0.993, + "step": 509 + }, + { + "epoch": 0.006692907896236962, + "grad_norm": 1.8436781167984009, + "learning_rate": 2.8991596638655467e-05, + "loss": 0.8254, + "step": 510 + }, + { + "epoch": 0.006706031245053113, + "grad_norm": 2.0006208419799805, + "learning_rate": 2.8949579831932776e-05, + "loss": 1.2799, + "step": 511 + }, + { + "epoch": 0.006719154593869263, + "grad_norm": 1.581337571144104, + "learning_rate": 2.890756302521009e-05, + "loss": 1.0694, + "step": 512 + }, + { + "epoch": 0.006732277942685414, + "grad_norm": 2.022268056869507, + "learning_rate": 2.8865546218487394e-05, + "loss": 0.7551, + "step": 513 + }, + { + "epoch": 0.006745401291501566, + "grad_norm": 1.7129675149917603, + "learning_rate": 2.8823529411764703e-05, + "loss": 1.1265, + "step": 514 + }, + { + "epoch": 0.0067585246403177165, + "grad_norm": 1.994723916053772, + "learning_rate": 2.8781512605042016e-05, + "loss": 1.0374, + "step": 515 + }, + { + "epoch": 0.006771647989133867, + "grad_norm": 1.7831419706344604, + "learning_rate": 2.8739495798319328e-05, + "loss": 0.8946, + "step": 516 + }, + { + "epoch": 0.006784771337950018, + "grad_norm": 1.705159306526184, + "learning_rate": 2.8697478991596637e-05, + "loss": 1.0248, + "step": 517 + }, + { + "epoch": 0.006797894686766169, + "grad_norm": 1.675284504890442, + "learning_rate": 2.865546218487395e-05, + "loss": 1.0446, + "step": 518 + }, + { + "epoch": 0.0068110180355823195, + "grad_norm": 1.8878024816513062, + "learning_rate": 2.861344537815126e-05, + "loss": 1.1455, + "step": 519 + }, + { + "epoch": 0.006824141384398471, + "grad_norm": 1.783176302909851, + "learning_rate": 2.857142857142857e-05, + "loss": 1.1697, + "step": 520 + }, + { + "epoch": 0.006837264733214622, + "grad_norm": 1.6017670631408691, + "learning_rate": 2.8529411764705883e-05, + "loss": 0.883, + "step": 521 + }, + { + "epoch": 0.006850388082030773, + "grad_norm": 2.264923572540283, + "learning_rate": 2.8487394957983192e-05, + "loss": 1.2535, + "step": 522 + }, + { + "epoch": 0.006863511430846923, + "grad_norm": 1.7666747570037842, + "learning_rate": 2.8445378151260505e-05, + "loss": 1.0606, + "step": 523 + }, + { + "epoch": 0.006876634779663074, + "grad_norm": 1.9199743270874023, + "learning_rate": 2.8403361344537817e-05, + "loss": 1.0026, + "step": 524 + }, + { + "epoch": 0.006889758128479226, + "grad_norm": 1.6176460981369019, + "learning_rate": 2.8361344537815126e-05, + "loss": 1.2288, + "step": 525 + }, + { + "epoch": 0.006902881477295376, + "grad_norm": 2.1176071166992188, + "learning_rate": 2.831932773109244e-05, + "loss": 0.8439, + "step": 526 + }, + { + "epoch": 0.006916004826111527, + "grad_norm": 1.7792637348175049, + "learning_rate": 2.8277310924369747e-05, + "loss": 0.8614, + "step": 527 + }, + { + "epoch": 0.006929128174927678, + "grad_norm": 2.24349045753479, + "learning_rate": 2.823529411764706e-05, + "loss": 1.1758, + "step": 528 + }, + { + "epoch": 0.006942251523743829, + "grad_norm": 1.985783576965332, + "learning_rate": 2.8193277310924372e-05, + "loss": 0.9556, + "step": 529 + }, + { + "epoch": 0.00695537487255998, + "grad_norm": 1.7994742393493652, + "learning_rate": 2.815126050420168e-05, + "loss": 0.8336, + "step": 530 + }, + { + "epoch": 0.006968498221376131, + "grad_norm": 1.6311463117599487, + "learning_rate": 2.8109243697478993e-05, + "loss": 0.9338, + "step": 531 + }, + { + "epoch": 0.006981621570192282, + "grad_norm": 1.8948392868041992, + "learning_rate": 2.8067226890756302e-05, + "loss": 1.2195, + "step": 532 + }, + { + "epoch": 0.0069947449190084325, + "grad_norm": 1.5576555728912354, + "learning_rate": 2.8025210084033615e-05, + "loss": 1.2998, + "step": 533 + }, + { + "epoch": 0.007007868267824583, + "grad_norm": 1.9670222997665405, + "learning_rate": 2.7983193277310927e-05, + "loss": 1.1826, + "step": 534 + }, + { + "epoch": 0.007020991616640734, + "grad_norm": 1.8046157360076904, + "learning_rate": 2.7941176470588236e-05, + "loss": 1.1026, + "step": 535 + }, + { + "epoch": 0.007034114965456886, + "grad_norm": 2.0909459590911865, + "learning_rate": 2.789915966386555e-05, + "loss": 0.9153, + "step": 536 + }, + { + "epoch": 0.007047238314273036, + "grad_norm": 1.9456409215927124, + "learning_rate": 2.785714285714286e-05, + "loss": 1.1347, + "step": 537 + }, + { + "epoch": 0.007060361663089187, + "grad_norm": 1.6101740598678589, + "learning_rate": 2.781512605042017e-05, + "loss": 0.8582, + "step": 538 + }, + { + "epoch": 0.007073485011905338, + "grad_norm": 1.9965813159942627, + "learning_rate": 2.7773109243697482e-05, + "loss": 1.0083, + "step": 539 + }, + { + "epoch": 0.007086608360721489, + "grad_norm": 1.5864198207855225, + "learning_rate": 2.773109243697479e-05, + "loss": 0.8058, + "step": 540 + }, + { + "epoch": 0.00709973170953764, + "grad_norm": 1.828796625137329, + "learning_rate": 2.7689075630252104e-05, + "loss": 0.9877, + "step": 541 + }, + { + "epoch": 0.007112855058353791, + "grad_norm": 2.0490660667419434, + "learning_rate": 2.7647058823529416e-05, + "loss": 0.8834, + "step": 542 + }, + { + "epoch": 0.007125978407169942, + "grad_norm": 1.8640308380126953, + "learning_rate": 2.7605042016806725e-05, + "loss": 1.1126, + "step": 543 + }, + { + "epoch": 0.007139101755986092, + "grad_norm": 1.790665864944458, + "learning_rate": 2.7563025210084037e-05, + "loss": 1.0966, + "step": 544 + }, + { + "epoch": 0.007152225104802243, + "grad_norm": 1.8603692054748535, + "learning_rate": 2.7521008403361346e-05, + "loss": 0.9278, + "step": 545 + }, + { + "epoch": 0.007165348453618394, + "grad_norm": 1.5675926208496094, + "learning_rate": 2.747899159663866e-05, + "loss": 0.9246, + "step": 546 + }, + { + "epoch": 0.0071784718024345455, + "grad_norm": 2.200451612472534, + "learning_rate": 2.743697478991597e-05, + "loss": 1.3844, + "step": 547 + }, + { + "epoch": 0.007191595151250696, + "grad_norm": 1.7244689464569092, + "learning_rate": 2.739495798319328e-05, + "loss": 1.0398, + "step": 548 + }, + { + "epoch": 0.007204718500066847, + "grad_norm": 1.7680357694625854, + "learning_rate": 2.7352941176470593e-05, + "loss": 0.9782, + "step": 549 + }, + { + "epoch": 0.007217841848882998, + "grad_norm": 1.7877306938171387, + "learning_rate": 2.7310924369747898e-05, + "loss": 0.9214, + "step": 550 + }, + { + "epoch": 0.0072309651976991485, + "grad_norm": 1.6443262100219727, + "learning_rate": 2.7268907563025207e-05, + "loss": 1.002, + "step": 551 + }, + { + "epoch": 0.0072440885465153, + "grad_norm": 1.9577383995056152, + "learning_rate": 2.722689075630252e-05, + "loss": 1.4069, + "step": 552 + }, + { + "epoch": 0.007257211895331451, + "grad_norm": 1.518309235572815, + "learning_rate": 2.7184873949579832e-05, + "loss": 0.7467, + "step": 553 + }, + { + "epoch": 0.007270335244147602, + "grad_norm": 1.8443019390106201, + "learning_rate": 2.714285714285714e-05, + "loss": 0.8807, + "step": 554 + }, + { + "epoch": 0.007283458592963752, + "grad_norm": 1.7123879194259644, + "learning_rate": 2.7100840336134453e-05, + "loss": 0.8683, + "step": 555 + }, + { + "epoch": 0.007296581941779903, + "grad_norm": 1.9878864288330078, + "learning_rate": 2.7058823529411766e-05, + "loss": 1.2636, + "step": 556 + }, + { + "epoch": 0.007309705290596055, + "grad_norm": 1.8165278434753418, + "learning_rate": 2.7016806722689075e-05, + "loss": 1.2438, + "step": 557 + }, + { + "epoch": 0.0073228286394122055, + "grad_norm": 2.156190872192383, + "learning_rate": 2.6974789915966387e-05, + "loss": 1.0709, + "step": 558 + }, + { + "epoch": 0.007335951988228356, + "grad_norm": 1.7182382345199585, + "learning_rate": 2.6932773109243696e-05, + "loss": 0.7785, + "step": 559 + }, + { + "epoch": 0.007349075337044507, + "grad_norm": 2.2051150798797607, + "learning_rate": 2.689075630252101e-05, + "loss": 1.0183, + "step": 560 + }, + { + "epoch": 0.007362198685860658, + "grad_norm": 1.5881297588348389, + "learning_rate": 2.684873949579832e-05, + "loss": 0.8281, + "step": 561 + }, + { + "epoch": 0.007375322034676808, + "grad_norm": 1.5579569339752197, + "learning_rate": 2.680672268907563e-05, + "loss": 1.1472, + "step": 562 + }, + { + "epoch": 0.00738844538349296, + "grad_norm": 1.930672287940979, + "learning_rate": 2.6764705882352942e-05, + "loss": 0.9618, + "step": 563 + }, + { + "epoch": 0.007401568732309111, + "grad_norm": 1.9797165393829346, + "learning_rate": 2.672268907563025e-05, + "loss": 0.7419, + "step": 564 + }, + { + "epoch": 0.0074146920811252615, + "grad_norm": 1.7777879238128662, + "learning_rate": 2.6680672268907564e-05, + "loss": 1.2476, + "step": 565 + }, + { + "epoch": 0.007427815429941412, + "grad_norm": 1.7834422588348389, + "learning_rate": 2.6638655462184876e-05, + "loss": 0.9808, + "step": 566 + }, + { + "epoch": 0.007440938778757563, + "grad_norm": 4.512395858764648, + "learning_rate": 2.6596638655462185e-05, + "loss": 1.2391, + "step": 567 + }, + { + "epoch": 0.007454062127573715, + "grad_norm": 1.4711298942565918, + "learning_rate": 2.6554621848739497e-05, + "loss": 0.9466, + "step": 568 + }, + { + "epoch": 0.007467185476389865, + "grad_norm": 2.1497201919555664, + "learning_rate": 2.651260504201681e-05, + "loss": 1.0423, + "step": 569 + }, + { + "epoch": 0.007480308825206016, + "grad_norm": 2.1284451484680176, + "learning_rate": 2.647058823529412e-05, + "loss": 1.4471, + "step": 570 + }, + { + "epoch": 0.007493432174022167, + "grad_norm": 2.2171754837036133, + "learning_rate": 2.642857142857143e-05, + "loss": 1.0436, + "step": 571 + }, + { + "epoch": 0.007506555522838318, + "grad_norm": 2.088338613510132, + "learning_rate": 2.638655462184874e-05, + "loss": 1.0471, + "step": 572 + }, + { + "epoch": 0.007519678871654469, + "grad_norm": 1.9278613328933716, + "learning_rate": 2.6344537815126052e-05, + "loss": 1.0237, + "step": 573 + }, + { + "epoch": 0.00753280222047062, + "grad_norm": 1.7411606311798096, + "learning_rate": 2.6302521008403365e-05, + "loss": 1.2358, + "step": 574 + }, + { + "epoch": 0.007545925569286771, + "grad_norm": 1.9148597717285156, + "learning_rate": 2.6260504201680674e-05, + "loss": 1.0746, + "step": 575 + }, + { + "epoch": 0.0075590489181029214, + "grad_norm": 1.8952248096466064, + "learning_rate": 2.6218487394957986e-05, + "loss": 0.9406, + "step": 576 + }, + { + "epoch": 0.007572172266919072, + "grad_norm": 1.5231568813323975, + "learning_rate": 2.6176470588235295e-05, + "loss": 1.0948, + "step": 577 + }, + { + "epoch": 0.007585295615735223, + "grad_norm": 1.7733741998672485, + "learning_rate": 2.6134453781512608e-05, + "loss": 0.7948, + "step": 578 + }, + { + "epoch": 0.0075984189645513745, + "grad_norm": 2.0340850353240967, + "learning_rate": 2.609243697478992e-05, + "loss": 0.8924, + "step": 579 + }, + { + "epoch": 0.007611542313367525, + "grad_norm": 2.275388240814209, + "learning_rate": 2.605042016806723e-05, + "loss": 1.245, + "step": 580 + }, + { + "epoch": 0.007624665662183676, + "grad_norm": 1.734357237815857, + "learning_rate": 2.600840336134454e-05, + "loss": 1.1312, + "step": 581 + }, + { + "epoch": 0.007637789010999827, + "grad_norm": 1.541772723197937, + "learning_rate": 2.5966386554621854e-05, + "loss": 0.7925, + "step": 582 + }, + { + "epoch": 0.0076509123598159775, + "grad_norm": 1.9468810558319092, + "learning_rate": 2.5924369747899163e-05, + "loss": 0.868, + "step": 583 + }, + { + "epoch": 0.007664035708632129, + "grad_norm": 1.7712815999984741, + "learning_rate": 2.5882352941176475e-05, + "loss": 0.8882, + "step": 584 + }, + { + "epoch": 0.00767715905744828, + "grad_norm": 2.1997227668762207, + "learning_rate": 2.5840336134453784e-05, + "loss": 0.8745, + "step": 585 + }, + { + "epoch": 0.007690282406264431, + "grad_norm": 1.7818772792816162, + "learning_rate": 2.5798319327731096e-05, + "loss": 1.1102, + "step": 586 + }, + { + "epoch": 0.007703405755080581, + "grad_norm": 1.9181782007217407, + "learning_rate": 2.5756302521008402e-05, + "loss": 0.9866, + "step": 587 + }, + { + "epoch": 0.007716529103896732, + "grad_norm": 1.757861614227295, + "learning_rate": 2.5714285714285714e-05, + "loss": 0.9548, + "step": 588 + }, + { + "epoch": 0.007729652452712883, + "grad_norm": 2.0078554153442383, + "learning_rate": 2.5672268907563023e-05, + "loss": 1.0659, + "step": 589 + }, + { + "epoch": 0.0077427758015290345, + "grad_norm": 2.041961669921875, + "learning_rate": 2.5630252100840336e-05, + "loss": 1.1867, + "step": 590 + }, + { + "epoch": 0.007755899150345185, + "grad_norm": 1.8703291416168213, + "learning_rate": 2.5588235294117645e-05, + "loss": 0.9826, + "step": 591 + }, + { + "epoch": 0.007769022499161336, + "grad_norm": 1.6957659721374512, + "learning_rate": 2.5546218487394957e-05, + "loss": 0.9584, + "step": 592 + }, + { + "epoch": 0.007782145847977487, + "grad_norm": 2.2199056148529053, + "learning_rate": 2.550420168067227e-05, + "loss": 1.2266, + "step": 593 + }, + { + "epoch": 0.0077952691967936374, + "grad_norm": 1.8995167016983032, + "learning_rate": 2.546218487394958e-05, + "loss": 1.0272, + "step": 594 + }, + { + "epoch": 0.007808392545609789, + "grad_norm": 1.8795480728149414, + "learning_rate": 2.542016806722689e-05, + "loss": 1.2015, + "step": 595 + }, + { + "epoch": 0.007821515894425939, + "grad_norm": 1.541920781135559, + "learning_rate": 2.53781512605042e-05, + "loss": 0.9476, + "step": 596 + }, + { + "epoch": 0.00783463924324209, + "grad_norm": 1.7051913738250732, + "learning_rate": 2.5336134453781512e-05, + "loss": 1.0941, + "step": 597 + }, + { + "epoch": 0.007847762592058242, + "grad_norm": 1.6097373962402344, + "learning_rate": 2.5294117647058825e-05, + "loss": 0.9476, + "step": 598 + }, + { + "epoch": 0.007860885940874392, + "grad_norm": 1.8319956064224243, + "learning_rate": 2.5252100840336134e-05, + "loss": 1.1252, + "step": 599 + }, + { + "epoch": 0.007874009289690544, + "grad_norm": 1.8047981262207031, + "learning_rate": 2.5210084033613446e-05, + "loss": 0.9398, + "step": 600 + } + ], + "logging_steps": 1, + "max_steps": 1200, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 200, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.644383561996206e+17, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}