sfulay commited on
Commit
af12e30
1 Parent(s): d6da671

Model save

Browse files
README.md CHANGED
@@ -2,15 +2,10 @@
2
  license: apache-2.0
3
  base_model: alignment-handbook/zephyr-7b-sft-full
4
  tags:
5
- - alignment-handbook
6
- - trl
7
- - dpo
8
- - generated_from_trainer
9
  - trl
10
  - dpo
 
11
  - generated_from_trainer
12
- datasets:
13
- - HuggingFaceH4/ultrafeedback_binarized
14
  model-index:
15
  - name: zephyr-7b-dpo-full-ultrabin-high-bleu
16
  results: []
@@ -21,17 +16,17 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # zephyr-7b-dpo-full-ultrabin-high-bleu
23
 
24
- This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on the HuggingFaceH4/ultrafeedback_binarized dataset.
25
  It achieves the following results on the evaluation set:
26
- - Loss: 0.6282
27
- - Rewards/chosen: -0.2999
28
- - Rewards/rejected: -0.5424
29
- - Rewards/accuracies: 0.7031
30
- - Rewards/margins: 0.2425
31
- - Logps/rejected: -316.9008
32
- - Logps/chosen: -292.6180
33
- - Logits/rejected: -2.2318
34
- - Logits/chosen: -2.2813
35
 
36
  ## Model description
37
 
@@ -68,8 +63,8 @@ The following hyperparameters were used during training:
68
 
69
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
70
  |:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
71
- | 0.66 | 0.3484 | 50 | 0.6513 | -0.1132 | -0.2374 | 0.6641 | 0.1242 | -286.3991 | -273.9513 | -2.4980 | -2.5369 |
72
- | 0.6376 | 0.6969 | 100 | 0.6329 | -0.2639 | -0.4887 | 0.6875 | 0.2248 | -311.5294 | -289.0153 | -2.2959 | -2.3456 |
73
 
74
 
75
  ### Framework versions
 
2
  license: apache-2.0
3
  base_model: alignment-handbook/zephyr-7b-sft-full
4
  tags:
 
 
 
 
5
  - trl
6
  - dpo
7
+ - alignment-handbook
8
  - generated_from_trainer
 
 
9
  model-index:
10
  - name: zephyr-7b-dpo-full-ultrabin-high-bleu
11
  results: []
 
16
 
17
  # zephyr-7b-dpo-full-ultrabin-high-bleu
18
 
19
+ This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.6311
22
+ - Rewards/chosen: -0.3272
23
+ - Rewards/rejected: -0.5547
24
+ - Rewards/accuracies: 0.6875
25
+ - Rewards/margins: 0.2276
26
+ - Logps/rejected: -318.1367
27
+ - Logps/chosen: -295.3459
28
+ - Logits/rejected: -2.3484
29
+ - Logits/chosen: -2.3932
30
 
31
  ## Model description
32
 
 
63
 
64
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
65
  |:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
66
+ | 0.6597 | 0.3484 | 50 | 0.6513 | -0.1047 | -0.2299 | 0.6641 | 0.1252 | -285.6537 | -273.1035 | -2.4969 | -2.5355 |
67
+ | 0.6383 | 0.6969 | 100 | 0.6311 | -0.3272 | -0.5547 | 0.6875 | 0.2276 | -318.1367 | -295.3459 | -2.3484 | -2.3932 |
68
 
69
 
70
  ### Framework versions
all_results.json CHANGED
@@ -14,9 +14,9 @@
14
  "eval_samples_per_second": 19.065,
15
  "eval_steps_per_second": 0.305,
16
  "total_flos": 0.0,
17
- "train_loss": 0.6509068312344851,
18
- "train_runtime": 3880.8302,
19
  "train_samples": 18340,
20
- "train_samples_per_second": 4.726,
21
  "train_steps_per_second": 0.037
22
  }
 
14
  "eval_samples_per_second": 19.065,
15
  "eval_steps_per_second": 0.305,
16
  "total_flos": 0.0,
17
+ "train_loss": 0.6505647914392965,
18
+ "train_runtime": 3852.398,
19
  "train_samples": 18340,
20
+ "train_samples_per_second": 4.761,
21
  "train_steps_per_second": 0.037
22
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b2b9259176ea21cd446935814aeeb2cbd2c26189b316fc1111f2c292a1ada18
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9378636ec91b3e02e5687373b6eec44210ca5fc9ce51802efdecb5468d62118e
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54164d9bfaf5d4fa59c2cdff47ac3d0b8ceba43b2b18c0ec87bc3225001a9c42
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18e62239c73ac4e9d825ec302bfc5d88815699688cc69adf69aeec4c380376a4
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5581f68d1158abd35ff5b749a7461714f51a7895e6f7ab6de27643ccc9c44e02
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81d7e8997d64033b8232ad6f56b4654c23d7c466f9b3ca9b049e64261e4af8e6
3
  size 4540516344
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9965156794425087,
3
  "total_flos": 0.0,
4
- "train_loss": 0.6509068312344851,
5
- "train_runtime": 3880.8302,
6
  "train_samples": 18340,
7
- "train_samples_per_second": 4.726,
8
  "train_steps_per_second": 0.037
9
  }
 
1
  {
2
  "epoch": 0.9965156794425087,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.6505647914392965,
5
+ "train_runtime": 3852.398,
6
  "train_samples": 18340,
7
+ "train_samples_per_second": 4.761,
8
  "train_steps_per_second": 0.037
9
  }
trainer_state.json CHANGED
@@ -10,253 +10,253 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.06968641114982578,
13
- "grad_norm": 7.974906667324644,
14
  "learning_rate": 3.333333333333333e-07,
15
- "logits/chosen": -2.747501850128174,
16
- "logits/rejected": -2.7370193004608154,
17
- "logps/chosen": -291.8797302246094,
18
- "logps/rejected": -281.50189208984375,
19
  "loss": 0.693,
20
- "rewards/accuracies": 0.3812499940395355,
21
- "rewards/chosen": -0.00011452056060079485,
22
- "rewards/margins": -3.700270826811902e-05,
23
- "rewards/rejected": -7.751785597065464e-05,
24
  "step": 10
25
  },
26
  {
27
  "epoch": 0.13937282229965156,
28
- "grad_norm": 8.1980266894883,
29
  "learning_rate": 4.981198836496775e-07,
30
- "logits/chosen": -2.7630066871643066,
31
- "logits/rejected": -2.748826026916504,
32
- "logps/chosen": -298.94158935546875,
33
- "logps/rejected": -320.3427429199219,
34
- "loss": 0.6902,
35
- "rewards/accuracies": 0.637499988079071,
36
- "rewards/chosen": 0.008970660157501698,
37
- "rewards/margins": 0.006611389108002186,
38
- "rewards/rejected": 0.002359271515160799,
39
  "step": 20
40
  },
41
  {
42
  "epoch": 0.20905923344947736,
43
- "grad_norm": 8.031029605438988,
44
  "learning_rate": 4.832481997086846e-07,
45
- "logits/chosen": -2.7137038707733154,
46
- "logits/rejected": -2.7142727375030518,
47
- "logps/chosen": -296.2449951171875,
48
- "logps/rejected": -289.28814697265625,
49
  "loss": 0.681,
50
- "rewards/accuracies": 0.6875,
51
- "rewards/chosen": 0.040586382150650024,
52
- "rewards/margins": 0.03953887149691582,
53
- "rewards/rejected": 0.0010475128656253219,
54
  "step": 30
55
  },
56
  {
57
  "epoch": 0.2787456445993031,
58
- "grad_norm": 9.014305613217058,
59
  "learning_rate": 4.543962032878959e-07,
60
- "logits/chosen": -2.679992198944092,
61
- "logits/rejected": -2.675081253051758,
62
- "logps/chosen": -298.83929443359375,
63
- "logps/rejected": -318.2642822265625,
64
  "loss": 0.6637,
65
- "rewards/accuracies": 0.6625000238418579,
66
- "rewards/chosen": -0.003969565033912659,
67
- "rewards/margins": 0.06901098042726517,
68
- "rewards/rejected": -0.07298054546117783,
69
  "step": 40
70
  },
71
  {
72
  "epoch": 0.34843205574912894,
73
- "grad_norm": 10.923357206738944,
74
  "learning_rate": 4.1329321073844413e-07,
75
- "logits/chosen": -2.547337770462036,
76
- "logits/rejected": -2.5558876991271973,
77
- "logps/chosen": -272.10382080078125,
78
- "logps/rejected": -269.55987548828125,
79
- "loss": 0.66,
80
  "rewards/accuracies": 0.59375,
81
- "rewards/chosen": -0.13082024455070496,
82
- "rewards/margins": 0.06251106411218643,
83
- "rewards/rejected": -0.1933313012123108,
84
  "step": 50
85
  },
86
  {
87
  "epoch": 0.34843205574912894,
88
- "eval_logits/chosen": -2.536876678466797,
89
- "eval_logits/rejected": -2.4980242252349854,
90
- "eval_logps/chosen": -273.9512634277344,
91
- "eval_logps/rejected": -286.3990783691406,
92
- "eval_loss": 0.6512594819068909,
93
  "eval_rewards/accuracies": 0.6640625,
94
- "eval_rewards/chosen": -0.11321306228637695,
95
- "eval_rewards/margins": 0.12415696680545807,
96
- "eval_rewards/rejected": -0.23737002909183502,
97
- "eval_runtime": 103.5252,
98
- "eval_samples_per_second": 19.319,
99
- "eval_steps_per_second": 0.309,
100
  "step": 50
101
  },
102
  {
103
  "epoch": 0.4181184668989547,
104
- "grad_norm": 11.450114313746024,
105
  "learning_rate": 3.624028324136517e-07,
106
- "logits/chosen": -2.5856661796569824,
107
- "logits/rejected": -2.5894103050231934,
108
- "logps/chosen": -315.35540771484375,
109
- "logps/rejected": -320.2020263671875,
110
- "loss": 0.6507,
111
- "rewards/accuracies": 0.6937500238418579,
112
- "rewards/chosen": -0.12617941200733185,
113
- "rewards/margins": 0.15071702003479004,
114
- "rewards/rejected": -0.2768964171409607,
115
  "step": 60
116
  },
117
  {
118
  "epoch": 0.4878048780487805,
119
- "grad_norm": 10.762410651532795,
120
  "learning_rate": 3.047753100392174e-07,
121
- "logits/chosen": -2.531461238861084,
122
- "logits/rejected": -2.5228896141052246,
123
- "logps/chosen": -301.96588134765625,
124
- "logps/rejected": -300.63177490234375,
125
- "loss": 0.6458,
126
  "rewards/accuracies": 0.699999988079071,
127
- "rewards/chosen": -0.23531746864318848,
128
- "rewards/margins": 0.1557210385799408,
129
- "rewards/rejected": -0.3910384774208069,
130
  "step": 70
131
  },
132
  {
133
  "epoch": 0.5574912891986062,
134
- "grad_norm": 10.243677934319797,
135
  "learning_rate": 2.4386469286927194e-07,
136
- "logits/chosen": -2.531421184539795,
137
- "logits/rejected": -2.5143864154815674,
138
- "logps/chosen": -289.48651123046875,
139
- "logps/rejected": -301.8067932128906,
140
- "loss": 0.6511,
141
- "rewards/accuracies": 0.6499999761581421,
142
- "rewards/chosen": -0.187635600566864,
143
- "rewards/margins": 0.1388576477766037,
144
- "rewards/rejected": -0.3264932334423065,
145
  "step": 80
146
  },
147
  {
148
  "epoch": 0.627177700348432,
149
- "grad_norm": 10.605767289766804,
150
  "learning_rate": 1.8332181063127542e-07,
151
- "logits/chosen": -2.370906114578247,
152
- "logits/rejected": -2.396063804626465,
153
- "logps/chosen": -266.50262451171875,
154
- "logps/rejected": -298.68292236328125,
155
- "loss": 0.634,
156
- "rewards/accuracies": 0.6312500238418579,
157
- "rewards/chosen": -0.13985350728034973,
158
- "rewards/margins": 0.18151375651359558,
159
- "rewards/rejected": -0.3213672339916229,
160
  "step": 90
161
  },
162
  {
163
  "epoch": 0.6968641114982579,
164
- "grad_norm": 11.921015517224783,
165
  "learning_rate": 1.26775451942554e-07,
166
- "logits/chosen": -2.410521984100342,
167
- "logits/rejected": -2.4090819358825684,
168
- "logps/chosen": -325.18603515625,
169
- "logps/rejected": -335.88739013671875,
170
- "loss": 0.6376,
171
- "rewards/accuracies": 0.6937500238418579,
172
- "rewards/chosen": -0.2180793583393097,
173
- "rewards/margins": 0.21515274047851562,
174
- "rewards/rejected": -0.4332321286201477,
175
  "step": 100
176
  },
177
  {
178
  "epoch": 0.6968641114982579,
179
- "eval_logits/chosen": -2.345639705657959,
180
- "eval_logits/rejected": -2.2959439754486084,
181
- "eval_logps/chosen": -289.0152893066406,
182
- "eval_logps/rejected": -311.52935791015625,
183
- "eval_loss": 0.6328787803649902,
184
  "eval_rewards/accuracies": 0.6875,
185
- "eval_rewards/chosen": -0.26385369896888733,
186
- "eval_rewards/margins": 0.22481901943683624,
187
- "eval_rewards/rejected": -0.4886727035045624,
188
- "eval_runtime": 105.3794,
189
- "eval_samples_per_second": 18.979,
190
- "eval_steps_per_second": 0.304,
191
  "step": 100
192
  },
193
  {
194
  "epoch": 0.7665505226480837,
195
- "grad_norm": 10.917825147199611,
196
  "learning_rate": 7.761486381573326e-08,
197
- "logits/chosen": -2.425245761871338,
198
- "logits/rejected": -2.4225878715515137,
199
- "logps/chosen": -340.6777038574219,
200
- "logps/rejected": -376.40008544921875,
201
- "loss": 0.6263,
202
  "rewards/accuracies": 0.6937500238418579,
203
- "rewards/chosen": -0.29727238416671753,
204
- "rewards/margins": 0.22700247168540955,
205
- "rewards/rejected": -0.5242748856544495,
206
  "step": 110
207
  },
208
  {
209
  "epoch": 0.8362369337979094,
210
- "grad_norm": 11.561681770035124,
211
  "learning_rate": 3.878660868757322e-08,
212
- "logits/chosen": -2.329202175140381,
213
- "logits/rejected": -2.3225998878479004,
214
- "logps/chosen": -323.3353271484375,
215
- "logps/rejected": -341.76959228515625,
216
- "loss": 0.6324,
217
- "rewards/accuracies": 0.6625000238418579,
218
- "rewards/chosen": -0.2884915769100189,
219
- "rewards/margins": 0.22677147388458252,
220
- "rewards/rejected": -0.5152631402015686,
221
  "step": 120
222
  },
223
  {
224
  "epoch": 0.9059233449477352,
225
- "grad_norm": 11.514501964616796,
226
  "learning_rate": 1.261795485174083e-08,
227
- "logits/chosen": -2.403071165084839,
228
- "logits/rejected": -2.402184009552002,
229
- "logps/chosen": -327.9507141113281,
230
- "logps/rejected": -337.8323059082031,
231
- "loss": 0.6248,
232
- "rewards/accuracies": 0.6875,
233
- "rewards/chosen": -0.27428099513053894,
234
- "rewards/margins": 0.22150692343711853,
235
- "rewards/rejected": -0.49578791856765747,
236
  "step": 130
237
  },
238
  {
239
  "epoch": 0.975609756097561,
240
- "grad_norm": 11.032500717539952,
241
  "learning_rate": 6.773858303274482e-10,
242
- "logits/chosen": -2.3575150966644287,
243
- "logits/rejected": -2.346174716949463,
244
- "logps/chosen": -298.3825988769531,
245
- "logps/rejected": -313.13848876953125,
246
- "loss": 0.6258,
247
- "rewards/accuracies": 0.6312500238418579,
248
- "rewards/chosen": -0.3439212739467621,
249
- "rewards/margins": 0.12675735354423523,
250
- "rewards/rejected": -0.4706786274909973,
251
  "step": 140
252
  },
253
  {
254
  "epoch": 0.9965156794425087,
255
  "step": 143,
256
  "total_flos": 0.0,
257
- "train_loss": 0.6509068312344851,
258
- "train_runtime": 3880.8302,
259
- "train_samples_per_second": 4.726,
260
  "train_steps_per_second": 0.037
261
  }
262
  ],
 
10
  "log_history": [
11
  {
12
  "epoch": 0.06968641114982578,
13
+ "grad_norm": 7.942402444587964,
14
  "learning_rate": 3.333333333333333e-07,
15
+ "logits/chosen": -2.747418165206909,
16
+ "logits/rejected": -2.7369940280914307,
17
+ "logps/chosen": -291.8316650390625,
18
+ "logps/rejected": -281.5037536621094,
19
  "loss": 0.693,
20
+ "rewards/accuracies": 0.45625001192092896,
21
+ "rewards/chosen": 0.0003662299714051187,
22
+ "rewards/margins": 0.0004625393485184759,
23
+ "rewards/rejected": -9.630931162973866e-05,
24
  "step": 10
25
  },
26
  {
27
  "epoch": 0.13937282229965156,
28
+ "grad_norm": 8.199051257817597,
29
  "learning_rate": 4.981198836496775e-07,
30
+ "logits/chosen": -2.7628424167633057,
31
+ "logits/rejected": -2.7486491203308105,
32
+ "logps/chosen": -298.9639587402344,
33
+ "logps/rejected": -320.4031066894531,
34
+ "loss": 0.6901,
35
+ "rewards/accuracies": 0.550000011920929,
36
+ "rewards/chosen": 0.008746958337724209,
37
+ "rewards/margins": 0.006991321686655283,
38
+ "rewards/rejected": 0.0017556389793753624,
39
  "step": 20
40
  },
41
  {
42
  "epoch": 0.20905923344947736,
43
+ "grad_norm": 8.009396280118287,
44
  "learning_rate": 4.832481997086846e-07,
45
+ "logits/chosen": -2.713383436203003,
46
+ "logits/rejected": -2.714205503463745,
47
+ "logps/chosen": -296.2840881347656,
48
+ "logps/rejected": -289.38421630859375,
49
  "loss": 0.681,
50
+ "rewards/accuracies": 0.6937500238418579,
51
+ "rewards/chosen": 0.04019603878259659,
52
+ "rewards/margins": 0.04010898619890213,
53
+ "rewards/rejected": 8.704829815542325e-05,
54
  "step": 30
55
  },
56
  {
57
  "epoch": 0.2787456445993031,
58
+ "grad_norm": 8.916455655557288,
59
  "learning_rate": 4.543962032878959e-07,
60
+ "logits/chosen": -2.679121732711792,
61
+ "logits/rejected": -2.674286365509033,
62
+ "logps/chosen": -298.8094177246094,
63
+ "logps/rejected": -318.2736511230469,
64
  "loss": 0.6637,
65
+ "rewards/accuracies": 0.668749988079071,
66
+ "rewards/chosen": -0.003670867532491684,
67
+ "rewards/margins": 0.06940315663814545,
68
+ "rewards/rejected": -0.07307400554418564,
69
  "step": 40
70
  },
71
  {
72
  "epoch": 0.34843205574912894,
73
+ "grad_norm": 11.007582392023577,
74
  "learning_rate": 4.1329321073844413e-07,
75
+ "logits/chosen": -2.5453474521636963,
76
+ "logits/rejected": -2.553844451904297,
77
+ "logps/chosen": -272.39227294921875,
78
+ "logps/rejected": -269.97125244140625,
79
+ "loss": 0.6597,
80
  "rewards/accuracies": 0.59375,
81
+ "rewards/chosen": -0.1337047815322876,
82
+ "rewards/margins": 0.06374003738164902,
83
+ "rewards/rejected": -0.1974448263645172,
84
  "step": 50
85
  },
86
  {
87
  "epoch": 0.34843205574912894,
88
+ "eval_logits/chosen": -2.5355470180511475,
89
+ "eval_logits/rejected": -2.496910572052002,
90
+ "eval_logps/chosen": -273.103515625,
91
+ "eval_logps/rejected": -285.6536865234375,
92
+ "eval_loss": 0.6512949466705322,
93
  "eval_rewards/accuracies": 0.6640625,
94
+ "eval_rewards/chosen": -0.1047358363866806,
95
+ "eval_rewards/margins": 0.1251804083585739,
96
+ "eval_rewards/rejected": -0.22991624474525452,
97
+ "eval_runtime": 105.224,
98
+ "eval_samples_per_second": 19.007,
99
+ "eval_steps_per_second": 0.304,
100
  "step": 50
101
  },
102
  {
103
  "epoch": 0.4181184668989547,
104
+ "grad_norm": 11.130762557977343,
105
  "learning_rate": 3.624028324136517e-07,
106
+ "logits/chosen": -2.5859858989715576,
107
+ "logits/rejected": -2.590031147003174,
108
+ "logps/chosen": -312.9039611816406,
109
+ "logps/rejected": -316.5061950683594,
110
+ "loss": 0.6515,
111
+ "rewards/accuracies": 0.6499999761581421,
112
+ "rewards/chosen": -0.10166505724191666,
113
+ "rewards/margins": 0.13827314972877502,
114
+ "rewards/rejected": -0.23993822932243347,
115
  "step": 60
116
  },
117
  {
118
  "epoch": 0.4878048780487805,
119
+ "grad_norm": 10.367706676784401,
120
  "learning_rate": 3.047753100392174e-07,
121
+ "logits/chosen": -2.515353202819824,
122
+ "logits/rejected": -2.5070269107818604,
123
+ "logps/chosen": -303.077392578125,
124
+ "logps/rejected": -301.082763671875,
125
+ "loss": 0.6453,
126
  "rewards/accuracies": 0.699999988079071,
127
+ "rewards/chosen": -0.2464326173067093,
128
+ "rewards/margins": 0.14911559224128723,
129
+ "rewards/rejected": -0.3955482840538025,
130
  "step": 70
131
  },
132
  {
133
  "epoch": 0.5574912891986062,
134
+ "grad_norm": 10.328785377431698,
135
  "learning_rate": 2.4386469286927194e-07,
136
+ "logits/chosen": -2.534677028656006,
137
+ "logits/rejected": -2.515333652496338,
138
+ "logps/chosen": -305.86944580078125,
139
+ "logps/rejected": -315.22369384765625,
140
+ "loss": 0.6525,
141
+ "rewards/accuracies": 0.612500011920929,
142
+ "rewards/chosen": -0.3514646887779236,
143
+ "rewards/margins": 0.10919757932424545,
144
+ "rewards/rejected": -0.460662305355072,
145
  "step": 80
146
  },
147
  {
148
  "epoch": 0.627177700348432,
149
+ "grad_norm": 10.256439047308396,
150
  "learning_rate": 1.8332181063127542e-07,
151
+ "logits/chosen": -2.4058992862701416,
152
+ "logits/rejected": -2.4308979511260986,
153
+ "logps/chosen": -272.29986572265625,
154
+ "logps/rejected": -304.98779296875,
155
+ "loss": 0.6315,
156
+ "rewards/accuracies": 0.625,
157
+ "rewards/chosen": -0.19782552123069763,
158
+ "rewards/margins": 0.18659016489982605,
159
+ "rewards/rejected": -0.38441571593284607,
160
  "step": 90
161
  },
162
  {
163
  "epoch": 0.6968641114982579,
164
+ "grad_norm": 11.456138890836337,
165
  "learning_rate": 1.26775451942554e-07,
166
+ "logits/chosen": -2.4451889991760254,
167
+ "logits/rejected": -2.447258472442627,
168
+ "logps/chosen": -332.0220031738281,
169
+ "logps/rejected": -343.72979736328125,
170
+ "loss": 0.6383,
171
+ "rewards/accuracies": 0.675000011920929,
172
+ "rewards/chosen": -0.2864390015602112,
173
+ "rewards/margins": 0.22521762549877167,
174
+ "rewards/rejected": -0.511656641960144,
175
  "step": 100
176
  },
177
  {
178
  "epoch": 0.6968641114982579,
179
+ "eval_logits/chosen": -2.393218517303467,
180
+ "eval_logits/rejected": -2.3483967781066895,
181
+ "eval_logps/chosen": -295.345947265625,
182
+ "eval_logps/rejected": -318.13665771484375,
183
+ "eval_loss": 0.6311340928077698,
184
  "eval_rewards/accuracies": 0.6875,
185
+ "eval_rewards/chosen": -0.3271602392196655,
186
+ "eval_rewards/margins": 0.22758512198925018,
187
+ "eval_rewards/rejected": -0.5547453165054321,
188
+ "eval_runtime": 103.8391,
189
+ "eval_samples_per_second": 19.261,
190
+ "eval_steps_per_second": 0.308,
191
  "step": 100
192
  },
193
  {
194
  "epoch": 0.7665505226480837,
195
+ "grad_norm": 15.049705548114918,
196
  "learning_rate": 7.761486381573326e-08,
197
+ "logits/chosen": -2.472832679748535,
198
+ "logits/rejected": -2.4680256843566895,
199
+ "logps/chosen": -347.2950439453125,
200
+ "logps/rejected": -383.56903076171875,
201
+ "loss": 0.6223,
202
  "rewards/accuracies": 0.6937500238418579,
203
+ "rewards/chosen": -0.36344587802886963,
204
+ "rewards/margins": 0.232518270611763,
205
+ "rewards/rejected": -0.5959641933441162,
206
  "step": 110
207
  },
208
  {
209
  "epoch": 0.8362369337979094,
210
+ "grad_norm": 11.65420047853757,
211
  "learning_rate": 3.878660868757322e-08,
212
+ "logits/chosen": -2.3938543796539307,
213
+ "logits/rejected": -2.3905398845672607,
214
+ "logps/chosen": -329.33978271484375,
215
+ "logps/rejected": -346.9275817871094,
216
+ "loss": 0.6335,
217
+ "rewards/accuracies": 0.6187499761581421,
218
+ "rewards/chosen": -0.34853675961494446,
219
+ "rewards/margins": 0.21830621361732483,
220
+ "rewards/rejected": -0.5668429136276245,
221
  "step": 120
222
  },
223
  {
224
  "epoch": 0.9059233449477352,
225
+ "grad_norm": 11.07371443957527,
226
  "learning_rate": 1.261795485174083e-08,
227
+ "logits/chosen": -2.4500081539154053,
228
+ "logits/rejected": -2.451570510864258,
229
+ "logps/chosen": -330.7276306152344,
230
+ "logps/rejected": -342.2732849121094,
231
+ "loss": 0.626,
232
+ "rewards/accuracies": 0.699999988079071,
233
+ "rewards/chosen": -0.3020505905151367,
234
+ "rewards/margins": 0.23814749717712402,
235
+ "rewards/rejected": -0.540198028087616,
236
  "step": 130
237
  },
238
  {
239
  "epoch": 0.975609756097561,
240
+ "grad_norm": 11.70940252198527,
241
  "learning_rate": 6.773858303274482e-10,
242
+ "logits/chosen": -2.421135425567627,
243
+ "logits/rejected": -2.4089722633361816,
244
+ "logps/chosen": -300.1723937988281,
245
+ "logps/rejected": -315.6776123046875,
246
+ "loss": 0.6255,
247
+ "rewards/accuracies": 0.5874999761581421,
248
+ "rewards/chosen": -0.36181899905204773,
249
+ "rewards/margins": 0.13425084948539734,
250
+ "rewards/rejected": -0.49606984853744507,
251
  "step": 140
252
  },
253
  {
254
  "epoch": 0.9965156794425087,
255
  "step": 143,
256
  "total_flos": 0.0,
257
+ "train_loss": 0.6505647914392965,
258
+ "train_runtime": 3852.398,
259
+ "train_samples_per_second": 4.761,
260
  "train_steps_per_second": 0.037
261
  }
262
  ],