abdouaziiz commited on
Commit
4771d36
1 Parent(s): 1bd18d9

End of training

Browse files
README.md CHANGED
@@ -3,11 +3,23 @@ license: apache-2.0
3
  base_model: openai/whisper-medium
4
  tags:
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - wer
8
  model-index:
9
  - name: whisper-m-wo
10
- results: []
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,10 +27,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # whisper-m-wo
17
 
18
- This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.4883
21
- - Wer: 0.2429
22
 
23
  ## Model description
24
 
 
3
  base_model: openai/whisper-medium
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - abdouaziiz/wolof_lam_asr
8
  metrics:
9
  - wer
10
  model-index:
11
  - name: whisper-m-wo
12
+ results:
13
+ - task:
14
+ name: Automatic Speech Recognition
15
+ type: automatic-speech-recognition
16
+ dataset:
17
+ name: abdouaziiz/wolof_lam_asr
18
+ type: abdouaziiz/wolof_lam_asr
19
+ metrics:
20
+ - name: Wer
21
+ type: wer
22
+ value: 0.2595195074616877
23
  ---
24
 
25
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
27
 
28
  # whisper-m-wo
29
 
30
+ This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the abdouaziiz/wolof_lam_asr dataset.
31
  It achieves the following results on the evaluation set:
32
+ - Loss: 0.4811
33
+ - Wer: 0.2595
34
 
35
  ## Model description
36
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.912363067292645,
3
+ "eval_loss": 0.4810967445373535,
4
+ "eval_runtime": 961.9713,
5
+ "eval_samples": 1010,
6
+ "eval_samples_per_second": 1.05,
7
+ "eval_steps_per_second": 0.132,
8
+ "eval_wer": 0.2595195074616877,
9
+ "total_flos": 8.16483926016e+19,
10
+ "train_loss": 0.35271793479919433,
11
+ "train_runtime": 106272.5194,
12
+ "train_samples": 20448,
13
+ "train_samples_per_second": 1.204,
14
+ "train_steps_per_second": 0.151
15
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.912363067292645,
3
+ "eval_loss": 0.4810967445373535,
4
+ "eval_runtime": 961.9713,
5
+ "eval_samples": 1010,
6
+ "eval_samples_per_second": 1.05,
7
+ "eval_steps_per_second": 0.132,
8
+ "eval_wer": 0.2595195074616877
9
+ }
runs/Oct07_16-10-05_bambaraspeechtotext/events.out.tfevents.1728425299.bambaraspeechtotext.21209.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53ff1274614fe6af53c78ae1f1b8d2888f367c9ee788eaae2e0fa0e4eb716516
3
+ size 406
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.912363067292645,
3
+ "total_flos": 8.16483926016e+19,
4
+ "train_loss": 0.35271793479919433,
5
+ "train_runtime": 106272.5194,
6
+ "train_samples": 20448,
7
+ "train_samples_per_second": 1.204,
8
+ "train_steps_per_second": 0.151
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,281 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.4810967445373535,
3
+ "best_model_checkpoint": "whisper-m-wo/checkpoint-7000",
4
+ "epoch": 3.912363067292645,
5
+ "eval_steps": 1000,
6
+ "global_step": 10000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.19561815336463223,
13
+ "grad_norm": 14.263213157653809,
14
+ "learning_rate": 9.714285714285715e-06,
15
+ "loss": 1.2317,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.39123630672926446,
20
+ "grad_norm": 11.0503568649292,
21
+ "learning_rate": 9.401002506265666e-06,
22
+ "loss": 0.7816,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.39123630672926446,
27
+ "eval_loss": 0.7274152040481567,
28
+ "eval_runtime": 1154.7126,
29
+ "eval_samples_per_second": 0.875,
30
+ "eval_steps_per_second": 0.11,
31
+ "eval_wer": 0.6368868366459212,
32
+ "step": 1000
33
+ },
34
+ {
35
+ "epoch": 0.5868544600938967,
36
+ "grad_norm": 9.666540145874023,
37
+ "learning_rate": 9.087719298245615e-06,
38
+ "loss": 0.7047,
39
+ "step": 1500
40
+ },
41
+ {
42
+ "epoch": 0.7824726134585289,
43
+ "grad_norm": 9.399227142333984,
44
+ "learning_rate": 8.774436090225564e-06,
45
+ "loss": 0.6368,
46
+ "step": 2000
47
+ },
48
+ {
49
+ "epoch": 0.7824726134585289,
50
+ "eval_loss": 0.6092658042907715,
51
+ "eval_runtime": 1128.5812,
52
+ "eval_samples_per_second": 0.895,
53
+ "eval_steps_per_second": 0.113,
54
+ "eval_wer": 0.5042494813625109,
55
+ "step": 2000
56
+ },
57
+ {
58
+ "epoch": 0.9780907668231612,
59
+ "grad_norm": 9.21610164642334,
60
+ "learning_rate": 8.461152882205514e-06,
61
+ "loss": 0.5824,
62
+ "step": 2500
63
+ },
64
+ {
65
+ "epoch": 1.1737089201877935,
66
+ "grad_norm": 7.924953460693359,
67
+ "learning_rate": 8.147869674185465e-06,
68
+ "loss": 0.3921,
69
+ "step": 3000
70
+ },
71
+ {
72
+ "epoch": 1.1737089201877935,
73
+ "eval_loss": 0.5506494641304016,
74
+ "eval_runtime": 1090.6405,
75
+ "eval_samples_per_second": 0.926,
76
+ "eval_steps_per_second": 0.116,
77
+ "eval_wer": 0.4280265007026701,
78
+ "step": 3000
79
+ },
80
+ {
81
+ "epoch": 1.3693270735524257,
82
+ "grad_norm": 5.910419464111328,
83
+ "learning_rate": 7.834586466165414e-06,
84
+ "loss": 0.3708,
85
+ "step": 3500
86
+ },
87
+ {
88
+ "epoch": 1.5649452269170578,
89
+ "grad_norm": 7.41471529006958,
90
+ "learning_rate": 7.521303258145364e-06,
91
+ "loss": 0.3494,
92
+ "step": 4000
93
+ },
94
+ {
95
+ "epoch": 1.5649452269170578,
96
+ "eval_loss": 0.5246758460998535,
97
+ "eval_runtime": 999.9485,
98
+ "eval_samples_per_second": 1.01,
99
+ "eval_steps_per_second": 0.127,
100
+ "eval_wer": 0.3114501773405608,
101
+ "step": 4000
102
+ },
103
+ {
104
+ "epoch": 1.76056338028169,
105
+ "grad_norm": 4.512611389160156,
106
+ "learning_rate": 7.208646616541354e-06,
107
+ "loss": 0.3506,
108
+ "step": 4500
109
+ },
110
+ {
111
+ "epoch": 1.9561815336463224,
112
+ "grad_norm": 7.675483226776123,
113
+ "learning_rate": 6.895363408521304e-06,
114
+ "loss": 0.3264,
115
+ "step": 5000
116
+ },
117
+ {
118
+ "epoch": 1.9561815336463224,
119
+ "eval_loss": 0.4906909167766571,
120
+ "eval_runtime": 1031.2068,
121
+ "eval_samples_per_second": 0.979,
122
+ "eval_steps_per_second": 0.123,
123
+ "eval_wer": 0.32925115438666935,
124
+ "step": 5000
125
+ },
126
+ {
127
+ "epoch": 2.151799687010955,
128
+ "grad_norm": 7.257579326629639,
129
+ "learning_rate": 6.582080200501254e-06,
130
+ "loss": 0.204,
131
+ "step": 5500
132
+ },
133
+ {
134
+ "epoch": 2.347417840375587,
135
+ "grad_norm": 6.650781154632568,
136
+ "learning_rate": 6.269423558897244e-06,
137
+ "loss": 0.1734,
138
+ "step": 6000
139
+ },
140
+ {
141
+ "epoch": 2.347417840375587,
142
+ "eval_loss": 0.49680861830711365,
143
+ "eval_runtime": 997.6386,
144
+ "eval_samples_per_second": 1.012,
145
+ "eval_steps_per_second": 0.127,
146
+ "eval_wer": 0.2973298534430837,
147
+ "step": 6000
148
+ },
149
+ {
150
+ "epoch": 2.543035993740219,
151
+ "grad_norm": 8.005096435546875,
152
+ "learning_rate": 5.956140350877193e-06,
153
+ "loss": 0.1718,
154
+ "step": 6500
155
+ },
156
+ {
157
+ "epoch": 2.7386541471048513,
158
+ "grad_norm": 6.306284427642822,
159
+ "learning_rate": 5.6428571428571435e-06,
160
+ "loss": 0.1808,
161
+ "step": 7000
162
+ },
163
+ {
164
+ "epoch": 2.7386541471048513,
165
+ "eval_loss": 0.4810967445373535,
166
+ "eval_runtime": 972.537,
167
+ "eval_samples_per_second": 1.039,
168
+ "eval_steps_per_second": 0.131,
169
+ "eval_wer": 0.2595195074616877,
170
+ "step": 7000
171
+ },
172
+ {
173
+ "epoch": 2.9342723004694835,
174
+ "grad_norm": 10.663891792297363,
175
+ "learning_rate": 5.329573934837093e-06,
176
+ "loss": 0.1747,
177
+ "step": 7500
178
+ },
179
+ {
180
+ "epoch": 3.1298904538341157,
181
+ "grad_norm": 5.245405673980713,
182
+ "learning_rate": 5.016290726817044e-06,
183
+ "loss": 0.1064,
184
+ "step": 8000
185
+ },
186
+ {
187
+ "epoch": 3.1298904538341157,
188
+ "eval_loss": 0.4989220201969147,
189
+ "eval_runtime": 978.6774,
190
+ "eval_samples_per_second": 1.032,
191
+ "eval_steps_per_second": 0.13,
192
+ "eval_wer": 0.2490129157465034,
193
+ "step": 8000
194
+ },
195
+ {
196
+ "epoch": 3.325508607198748,
197
+ "grad_norm": 3.174140691757202,
198
+ "learning_rate": 4.703007518796993e-06,
199
+ "loss": 0.0804,
200
+ "step": 8500
201
+ },
202
+ {
203
+ "epoch": 3.52112676056338,
204
+ "grad_norm": 0.9650879502296448,
205
+ "learning_rate": 4.389724310776943e-06,
206
+ "loss": 0.0802,
207
+ "step": 9000
208
+ },
209
+ {
210
+ "epoch": 3.52112676056338,
211
+ "eval_loss": 0.4974755346775055,
212
+ "eval_runtime": 945.4502,
213
+ "eval_samples_per_second": 1.068,
214
+ "eval_steps_per_second": 0.134,
215
+ "eval_wer": 0.22753128555176336,
216
+ "step": 9000
217
+ },
218
+ {
219
+ "epoch": 3.7167449139280127,
220
+ "grad_norm": 6.479980945587158,
221
+ "learning_rate": 4.076441102756892e-06,
222
+ "loss": 0.0817,
223
+ "step": 9500
224
+ },
225
+ {
226
+ "epoch": 3.912363067292645,
227
+ "grad_norm": 4.768033981323242,
228
+ "learning_rate": 3.7631578947368426e-06,
229
+ "loss": 0.0745,
230
+ "step": 10000
231
+ },
232
+ {
233
+ "epoch": 3.912363067292645,
234
+ "eval_loss": 0.48827001452445984,
235
+ "eval_runtime": 978.3234,
236
+ "eval_samples_per_second": 1.032,
237
+ "eval_steps_per_second": 0.13,
238
+ "eval_wer": 0.24292310780967677,
239
+ "step": 10000
240
+ },
241
+ {
242
+ "epoch": 3.912363067292645,
243
+ "step": 10000,
244
+ "total_flos": 8.16483926016e+19,
245
+ "train_loss": 0.35271793479919433,
246
+ "train_runtime": 106272.5194,
247
+ "train_samples_per_second": 1.204,
248
+ "train_steps_per_second": 0.151
249
+ }
250
+ ],
251
+ "logging_steps": 500,
252
+ "max_steps": 16000,
253
+ "num_input_tokens_seen": 0,
254
+ "num_train_epochs": 7,
255
+ "save_steps": 1000,
256
+ "stateful_callbacks": {
257
+ "EarlyStoppingCallback": {
258
+ "args": {
259
+ "early_stopping_patience": 3,
260
+ "early_stopping_threshold": 0.0
261
+ },
262
+ "attributes": {
263
+ "early_stopping_patience_counter": 0
264
+ }
265
+ },
266
+ "TrainerControl": {
267
+ "args": {
268
+ "should_epoch_stop": false,
269
+ "should_evaluate": false,
270
+ "should_log": false,
271
+ "should_save": true,
272
+ "should_training_stop": true
273
+ },
274
+ "attributes": {}
275
+ }
276
+ },
277
+ "total_flos": 8.16483926016e+19,
278
+ "train_batch_size": 8,
279
+ "trial_name": null,
280
+ "trial_params": null
281
+ }