communityai commited on
Commit
c06dbd8
1 Parent(s): bdce25c

Model save

Browse files
README.md CHANGED
@@ -4,18 +4,18 @@ base_model: 01-ai/Yi-6B
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
- - name: apt-chat-yi-34b-sft-full
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- # apt-chat-yi-34b-sft-full
15
 
16
  This model is a fine-tuned version of [01-ai/Yi-6B](https://huggingface.co/01-ai/Yi-6B) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 1.3422
19
 
20
  ## Model description
21
 
@@ -39,26 +39,25 @@ The following hyperparameters were used during training:
39
  - eval_batch_size: 1
40
  - seed: 42
41
  - distributed_type: multi-GPU
42
- - num_devices: 6
43
  - gradient_accumulation_steps: 4
44
- - total_train_batch_size: 24
45
- - total_eval_batch_size: 6
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: cosine
48
- - num_epochs: 3
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
- | 1.1029 | 0.15 | 135 | 1.0746 |
55
- | 0.8271 | 1.15 | 270 | 1.1266 |
56
- | 0.5131 | 2.15 | 405 | 1.3422 |
57
 
58
 
59
  ### Framework versions
60
 
61
  - Transformers 4.35.0
62
- - Pytorch 2.1.0.dev20230605+cu121
63
  - Datasets 2.14.6
64
  - Tokenizers 0.14.1
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
+ - name: apt-chat-yi-6B-sft-full
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ # apt-chat-yi-6B-sft-full
15
 
16
  This model is a fine-tuned version of [01-ai/Yi-6B](https://huggingface.co/01-ai/Yi-6B) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 1.0677
19
 
20
  ## Model description
21
 
 
39
  - eval_batch_size: 1
40
  - seed: 42
41
  - distributed_type: multi-GPU
42
+ - num_devices: 8
43
  - gradient_accumulation_steps: 4
44
+ - total_train_batch_size: 32
45
+ - total_eval_batch_size: 8
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: cosine
48
+ - num_epochs: 2
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
+ | 1.0548 | 0.15 | 1368 | 1.0247 |
55
+ | 0.9254 | 1.15 | 2736 | 1.0677 |
 
56
 
57
 
58
  ### Framework versions
59
 
60
  - Transformers 4.35.0
61
+ - Pytorch 2.1.0+cu118
62
  - Datasets 2.14.6
63
  - Tokenizers 0.14.1
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
- "epoch": 2.15,
3
- "eval_loss": 1.3422094583511353,
4
- "eval_runtime": 4.6035,
5
  "eval_samples": 500,
6
- "eval_samples_per_second": 108.613,
7
- "eval_steps_per_second": 18.247,
8
- "train_loss": 0.8341983215308484,
9
- "train_runtime": 1136.3762,
10
- "train_samples": 21407,
11
- "train_samples_per_second": 56.514,
12
- "train_steps_per_second": 2.355
13
  }
 
1
  {
2
+ "epoch": 1.15,
3
+ "eval_loss": 1.0676991939544678,
4
+ "eval_runtime": 4.4863,
5
  "eval_samples": 500,
6
+ "eval_samples_per_second": 111.451,
7
+ "eval_steps_per_second": 14.043,
8
+ "train_loss": 0.9719247023264567,
9
+ "train_runtime": 13352.0365,
10
+ "train_samples": 285436,
11
+ "train_samples_per_second": 42.755,
12
+ "train_steps_per_second": 1.336
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 2.15,
3
- "eval_loss": 1.3422094583511353,
4
- "eval_runtime": 4.6035,
5
  "eval_samples": 500,
6
- "eval_samples_per_second": 108.613,
7
- "eval_steps_per_second": 18.247
8
  }
 
1
  {
2
+ "epoch": 1.15,
3
+ "eval_loss": 1.0676991939544678,
4
+ "eval_runtime": 4.4863,
5
  "eval_samples": 500,
6
+ "eval_samples_per_second": 111.451,
7
+ "eval_steps_per_second": 14.043
8
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38adde306149008ebde6bd1c47df7c569461f342c6e44c5662babc9a9aa8fcba
3
  size 4932711032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:374370d80ea8bab03b1355f3dbd5b5dbd44c7043e0798d6f44b6dc4c9c18a45b
3
  size 4932711032
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae8166c4e72f2bb5633bbea473189925f9f8190892eedebee22ed7f4a13da880
3
  size 4976802304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7645743c9ee76cd989bc5ec83312837f22fb36dd590496f415fb778ddd1d8707
3
  size 4976802304
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8c3783f503806d1a8e7a383780717e1c9c2772e63f37f86e4423c6c8bda6cd4
3
  size 2212590400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dc189809d22dc30a9a7d52f17b178e73925e2e40e63d22c874158e212f2a6ab
3
  size 2212590400
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 2.15,
3
- "train_loss": 0.8341983215308484,
4
- "train_runtime": 1136.3762,
5
- "train_samples": 21407,
6
- "train_samples_per_second": 56.514,
7
- "train_steps_per_second": 2.355
8
  }
 
1
  {
2
+ "epoch": 1.15,
3
+ "train_loss": 0.9719247023264567,
4
+ "train_runtime": 13352.0365,
5
+ "train_samples": 285436,
6
+ "train_samples_per_second": 42.755,
7
+ "train_steps_per_second": 1.336
8
  }
trainer_state.json CHANGED
@@ -1,106 +1,374 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.151345291479821,
5
  "eval_steps": 500,
6
- "global_step": 405,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.0,
13
- "learning_rate": 1.9999993108758315e-05,
14
- "loss": 1.5531,
15
  "step": 1
16
  },
17
  {
18
- "epoch": 0.06,
19
- "learning_rate": 1.9982776840027333e-05,
20
- "loss": 1.177,
21
  "step": 50
22
  },
23
  {
24
- "epoch": 0.11,
25
- "learning_rate": 1.993116668755721e-05,
26
- "loss": 1.1029,
27
  "step": 100
28
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  {
30
  "epoch": 0.15,
31
- "eval_loss": 1.0746039152145386,
32
- "eval_runtime": 3.9638,
33
- "eval_samples_per_second": 126.14,
34
- "eval_steps_per_second": 21.192,
35
- "step": 135
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  },
37
  {
38
  "epoch": 1.02,
39
- "learning_rate": 1.984534732057208e-05,
40
- "loss": 1.036,
41
- "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  },
43
  {
44
  "epoch": 1.07,
45
- "learning_rate": 1.9725614355209207e-05,
46
- "loss": 0.8394,
47
- "step": 200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
  },
49
  {
50
  "epoch": 1.13,
51
- "learning_rate": 1.957238022747188e-05,
52
- "loss": 0.8271,
53
- "step": 250
54
  },
55
  {
56
- "epoch": 1.15,
57
- "eval_loss": 1.1265954971313477,
58
- "eval_runtime": 3.6594,
59
- "eval_samples_per_second": 136.636,
60
- "eval_steps_per_second": 22.955,
61
- "step": 270
62
  },
63
  {
64
- "epoch": 2.03,
65
- "learning_rate": 1.9386172772539162e-05,
66
- "loss": 0.667,
67
- "step": 300
68
  },
69
  {
70
- "epoch": 2.09,
71
- "learning_rate": 1.916763340656793e-05,
72
- "loss": 0.5326,
73
- "step": 350
74
  },
75
  {
76
- "epoch": 2.15,
77
- "learning_rate": 1.8917514917250276e-05,
78
- "loss": 0.5131,
79
- "step": 400
80
  },
81
  {
82
- "epoch": 2.15,
83
- "eval_loss": 1.3422094583511353,
84
- "eval_runtime": 4.9335,
85
- "eval_samples_per_second": 101.349,
86
- "eval_steps_per_second": 17.027,
87
- "step": 405
88
  },
89
  {
90
- "epoch": 2.15,
91
- "step": 405,
92
- "total_flos": 63559811727360.0,
93
- "train_loss": 0.8341983215308484,
94
- "train_runtime": 1136.3762,
95
- "train_samples_per_second": 56.514,
96
- "train_steps_per_second": 2.355
97
  }
98
  ],
99
  "logging_steps": 50,
100
- "max_steps": 2676,
101
- "num_train_epochs": 3,
102
  "save_steps": 500,
103
- "total_flos": 63559811727360.0,
104
  "trial_name": null,
105
  "trial_params": null
106
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.1533632286995517,
5
  "eval_steps": 500,
6
+ "global_step": 2736,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.0,
13
+ "learning_rate": 1.9999999844947046e-05,
14
+ "loss": 1.7024,
15
  "step": 1
16
  },
17
  {
18
+ "epoch": 0.01,
19
+ "learning_rate": 1.999961237011484e-05,
20
+ "loss": 1.1507,
21
  "step": 50
22
  },
23
  {
24
+ "epoch": 0.01,
25
+ "learning_rate": 1.9998449510510744e-05,
26
+ "loss": 1.0928,
27
  "step": 100
28
  },
29
+ {
30
+ "epoch": 0.02,
31
+ "learning_rate": 1.999651151133954e-05,
32
+ "loss": 1.0793,
33
+ "step": 150
34
+ },
35
+ {
36
+ "epoch": 0.02,
37
+ "learning_rate": 1.999379852284651e-05,
38
+ "loss": 1.0867,
39
+ "step": 200
40
+ },
41
+ {
42
+ "epoch": 0.03,
43
+ "learning_rate": 1.999031075535873e-05,
44
+ "loss": 1.0857,
45
+ "step": 250
46
+ },
47
+ {
48
+ "epoch": 0.03,
49
+ "learning_rate": 1.9986048479268788e-05,
50
+ "loss": 1.0721,
51
+ "step": 300
52
+ },
53
+ {
54
+ "epoch": 0.04,
55
+ "learning_rate": 1.99810120250138e-05,
56
+ "loss": 1.0923,
57
+ "step": 350
58
+ },
59
+ {
60
+ "epoch": 0.04,
61
+ "learning_rate": 1.9975201783049804e-05,
62
+ "loss": 1.0836,
63
+ "step": 400
64
+ },
65
+ {
66
+ "epoch": 0.05,
67
+ "learning_rate": 1.9968618203821487e-05,
68
+ "loss": 1.0769,
69
+ "step": 450
70
+ },
71
+ {
72
+ "epoch": 0.06,
73
+ "learning_rate": 1.9961261797727256e-05,
74
+ "loss": 1.0574,
75
+ "step": 500
76
+ },
77
+ {
78
+ "epoch": 0.06,
79
+ "learning_rate": 1.9953133135079686e-05,
80
+ "loss": 1.042,
81
+ "step": 550
82
+ },
83
+ {
84
+ "epoch": 0.07,
85
+ "learning_rate": 1.9944232846061284e-05,
86
+ "loss": 1.0554,
87
+ "step": 600
88
+ },
89
+ {
90
+ "epoch": 0.07,
91
+ "learning_rate": 1.993456162067566e-05,
92
+ "loss": 1.0735,
93
+ "step": 650
94
+ },
95
+ {
96
+ "epoch": 0.08,
97
+ "learning_rate": 1.992412020869401e-05,
98
+ "loss": 1.0785,
99
+ "step": 700
100
+ },
101
+ {
102
+ "epoch": 0.08,
103
+ "learning_rate": 1.9912909419596993e-05,
104
+ "loss": 1.0654,
105
+ "step": 750
106
+ },
107
+ {
108
+ "epoch": 0.09,
109
+ "learning_rate": 1.9900930122511993e-05,
110
+ "loss": 1.0606,
111
+ "step": 800
112
+ },
113
+ {
114
+ "epoch": 0.1,
115
+ "learning_rate": 1.988818324614572e-05,
116
+ "loss": 1.0664,
117
+ "step": 850
118
+ },
119
+ {
120
+ "epoch": 0.1,
121
+ "learning_rate": 1.9874669778712215e-05,
122
+ "loss": 1.0604,
123
+ "step": 900
124
+ },
125
+ {
126
+ "epoch": 0.11,
127
+ "learning_rate": 1.9860390767856244e-05,
128
+ "loss": 1.0674,
129
+ "step": 950
130
+ },
131
+ {
132
+ "epoch": 0.11,
133
+ "learning_rate": 1.984534732057208e-05,
134
+ "loss": 1.042,
135
+ "step": 1000
136
+ },
137
+ {
138
+ "epoch": 0.12,
139
+ "learning_rate": 1.9829540603117667e-05,
140
+ "loss": 1.0452,
141
+ "step": 1050
142
+ },
143
+ {
144
+ "epoch": 0.12,
145
+ "learning_rate": 1.9812971840924222e-05,
146
+ "loss": 1.0577,
147
+ "step": 1100
148
+ },
149
+ {
150
+ "epoch": 0.13,
151
+ "learning_rate": 1.979564231850122e-05,
152
+ "loss": 1.0471,
153
+ "step": 1150
154
+ },
155
+ {
156
+ "epoch": 0.13,
157
+ "learning_rate": 1.977755337933682e-05,
158
+ "loss": 1.0704,
159
+ "step": 1200
160
+ },
161
+ {
162
+ "epoch": 0.14,
163
+ "learning_rate": 1.9758706425793702e-05,
164
+ "loss": 1.0282,
165
+ "step": 1250
166
+ },
167
+ {
168
+ "epoch": 0.15,
169
+ "learning_rate": 1.973910291900036e-05,
170
+ "loss": 1.0515,
171
+ "step": 1300
172
+ },
173
+ {
174
+ "epoch": 0.15,
175
+ "learning_rate": 1.97187443787378e-05,
176
+ "loss": 1.0548,
177
+ "step": 1350
178
+ },
179
  {
180
  "epoch": 0.15,
181
+ "eval_loss": 1.0247304439544678,
182
+ "eval_runtime": 4.5889,
183
+ "eval_samples_per_second": 108.959,
184
+ "eval_steps_per_second": 13.729,
185
+ "step": 1368
186
+ },
187
+ {
188
+ "epoch": 1.0,
189
+ "learning_rate": 1.9697632383321755e-05,
190
+ "loss": 0.9636,
191
+ "step": 1400
192
+ },
193
+ {
194
+ "epoch": 1.01,
195
+ "learning_rate": 1.96757685694803e-05,
196
+ "loss": 0.9026,
197
+ "step": 1450
198
+ },
199
+ {
200
+ "epoch": 1.01,
201
+ "learning_rate": 1.965315463222695e-05,
202
+ "loss": 0.8808,
203
+ "step": 1500
204
  },
205
  {
206
  "epoch": 1.02,
207
+ "learning_rate": 1.9629792324729302e-05,
208
+ "loss": 0.8712,
209
+ "step": 1550
210
+ },
211
+ {
212
+ "epoch": 1.03,
213
+ "learning_rate": 1.960568345817306e-05,
214
+ "loss": 0.8967,
215
+ "step": 1600
216
+ },
217
+ {
218
+ "epoch": 1.03,
219
+ "learning_rate": 1.9580829901621666e-05,
220
+ "loss": 0.8676,
221
+ "step": 1650
222
+ },
223
+ {
224
+ "epoch": 1.04,
225
+ "learning_rate": 1.9555233581871366e-05,
226
+ "loss": 0.8723,
227
+ "step": 1700
228
+ },
229
+ {
230
+ "epoch": 1.04,
231
+ "learning_rate": 1.9528896483301866e-05,
232
+ "loss": 0.9122,
233
+ "step": 1750
234
+ },
235
+ {
236
+ "epoch": 1.05,
237
+ "learning_rate": 1.9501820647722458e-05,
238
+ "loss": 0.8687,
239
+ "step": 1800
240
+ },
241
+ {
242
+ "epoch": 1.05,
243
+ "learning_rate": 1.947400817421375e-05,
244
+ "loss": 0.8726,
245
+ "step": 1850
246
+ },
247
+ {
248
+ "epoch": 1.06,
249
+ "learning_rate": 1.944546121896493e-05,
250
+ "loss": 0.8505,
251
+ "step": 1900
252
  },
253
  {
254
  "epoch": 1.07,
255
+ "learning_rate": 1.9416181995106585e-05,
256
+ "loss": 0.8458,
257
+ "step": 1950
258
+ },
259
+ {
260
+ "epoch": 1.07,
261
+ "learning_rate": 1.9386172772539162e-05,
262
+ "loss": 0.8721,
263
+ "step": 2000
264
+ },
265
+ {
266
+ "epoch": 1.08,
267
+ "learning_rate": 1.9355435877756957e-05,
268
+ "loss": 0.8676,
269
+ "step": 2050
270
+ },
271
+ {
272
+ "epoch": 1.08,
273
+ "learning_rate": 1.9323973693667762e-05,
274
+ "loss": 0.8826,
275
+ "step": 2100
276
+ },
277
+ {
278
+ "epoch": 1.09,
279
+ "learning_rate": 1.929178865940815e-05,
280
+ "loss": 0.8607,
281
+ "step": 2150
282
+ },
283
+ {
284
+ "epoch": 1.09,
285
+ "learning_rate": 1.925888327015434e-05,
286
+ "loss": 0.8561,
287
+ "step": 2200
288
+ },
289
+ {
290
+ "epoch": 1.1,
291
+ "learning_rate": 1.9225260076928783e-05,
292
+ "loss": 0.8687,
293
+ "step": 2250
294
+ },
295
+ {
296
+ "epoch": 1.1,
297
+ "learning_rate": 1.919092168640239e-05,
298
+ "loss": 0.874,
299
+ "step": 2300
300
+ },
301
+ {
302
+ "epoch": 1.11,
303
+ "learning_rate": 1.915587076069243e-05,
304
+ "loss": 0.8563,
305
+ "step": 2350
306
+ },
307
+ {
308
+ "epoch": 1.12,
309
+ "learning_rate": 1.9120110017156172e-05,
310
+ "loss": 0.8445,
311
+ "step": 2400
312
+ },
313
+ {
314
+ "epoch": 1.12,
315
+ "learning_rate": 1.908364222818019e-05,
316
+ "loss": 0.8646,
317
+ "step": 2450
318
  },
319
  {
320
  "epoch": 1.13,
321
+ "learning_rate": 1.9046470220965457e-05,
322
+ "loss": 0.8479,
323
+ "step": 2500
324
  },
325
  {
326
+ "epoch": 1.13,
327
+ "learning_rate": 1.9008596877308157e-05,
328
+ "loss": 0.8788,
329
+ "step": 2550
 
 
330
  },
331
  {
332
+ "epoch": 1.14,
333
+ "learning_rate": 1.8970025133376252e-05,
334
+ "loss": 0.9,
335
+ "step": 2600
336
  },
337
  {
338
+ "epoch": 1.14,
339
+ "learning_rate": 1.893075797948188e-05,
340
+ "loss": 0.8791,
341
+ "step": 2650
342
  },
343
  {
344
+ "epoch": 1.15,
345
+ "learning_rate": 1.889079845984951e-05,
346
+ "loss": 0.9254,
347
+ "step": 2700
348
  },
349
  {
350
+ "epoch": 1.15,
351
+ "eval_loss": 1.0676991939544678,
352
+ "eval_runtime": 4.5191,
353
+ "eval_samples_per_second": 110.641,
354
+ "eval_steps_per_second": 13.941,
355
+ "step": 2736
356
  },
357
  {
358
+ "epoch": 1.15,
359
+ "step": 2736,
360
+ "total_flos": 572810393026560.0,
361
+ "train_loss": 0.9719247023264567,
362
+ "train_runtime": 13352.0365,
363
+ "train_samples_per_second": 42.755,
364
+ "train_steps_per_second": 1.336
365
  }
366
  ],
367
  "logging_steps": 50,
368
+ "max_steps": 17840,
369
+ "num_train_epochs": 2,
370
  "save_steps": 500,
371
+ "total_flos": 572810393026560.0,
372
  "trial_name": null,
373
  "trial_params": null
374
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38fa2e8b442d3c15b05cc3aad3c4f2d90f76d575a77420975deb6bc91e346800
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f08e31bceb220d21152bd52b1b3723abfb215236126fc2fed6cf222adf0775bd
3
  size 5624