yaygomii commited on
Commit
3ddb29c
1 Parent(s): 3766542

End of training

Browse files
README.md CHANGED
@@ -1,7 +1,11 @@
1
  ---
 
 
2
  license: apache-2.0
3
  base_model: facebook/wav2vec2-large-xlsr-53
4
  tags:
 
 
5
  - generated_from_trainer
6
  datasets:
7
  - common_voice_6_1
@@ -14,15 +18,15 @@ model-index:
14
  name: Automatic Speech Recognition
15
  type: automatic-speech-recognition
16
  dataset:
17
- name: common_voice_6_1
18
  type: common_voice_6_1
19
  config: ta
20
  split: test
21
- args: ta
22
  metrics:
23
  - name: Wer
24
  type: wer
25
- value: 0.7094281298299846
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -30,10 +34,10 @@ should probably proofread and complete it, then remove this comment. -->
30
 
31
  # wav2vec2-common_voice-ta
32
 
33
- This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the common_voice_6_1 dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.6599
36
- - Wer: 0.7094
37
 
38
  ## Model description
39
 
 
1
  ---
2
+ language:
3
+ - ta
4
  license: apache-2.0
5
  base_model: facebook/wav2vec2-large-xlsr-53
6
  tags:
7
+ - automatic-speech-recognition
8
+ - mozilla-foundation/common_voice_6_1
9
  - generated_from_trainer
10
  datasets:
11
  - common_voice_6_1
 
18
  name: Automatic Speech Recognition
19
  type: automatic-speech-recognition
20
  dataset:
21
+ name: MOZILLA-FOUNDATION/COMMON_VOICE_6_1 - TA
22
  type: common_voice_6_1
23
  config: ta
24
  split: test
25
+ args: 'Config: ta, Training split: train+validation, Eval split: test'
26
  metrics:
27
  - name: Wer
28
  type: wer
29
+ value: 0.7095686384712659
30
  ---
31
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
34
 
35
  # wav2vec2-common_voice-ta
36
 
37
+ This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the MOZILLA-FOUNDATION/COMMON_VOICE_6_1 - TA dataset.
38
  It achieves the following results on the evaluation set:
39
+ - Loss: 0.6563
40
+ - Wer: 0.7096
41
 
42
  ## Model description
43
 
all_results.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 14.94,
3
+ "eval_loss": 0.656301736831665,
4
+ "eval_runtime": 121.3421,
5
+ "eval_samples": 1781,
6
+ "eval_samples_per_second": 14.678,
7
+ "eval_steps_per_second": 1.838,
8
+ "eval_wer": 0.7095686384712659,
9
+ "train_loss": 1.3876537419981876,
10
+ "train_runtime": 7307.918,
11
+ "train_samples": 3788,
12
+ "train_samples_per_second": 7.775,
13
+ "train_steps_per_second": 0.242
14
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 14.94,
3
+ "eval_loss": 0.656301736831665,
4
+ "eval_runtime": 121.3421,
5
+ "eval_samples": 1781,
6
+ "eval_samples_per_second": 14.678,
7
+ "eval_steps_per_second": 1.838,
8
+ "eval_wer": 0.7095686384712659
9
+ }
runs/Feb09_15-12-11_96d880c2e067/events.out.tfevents.1707499153.96d880c2e067.6291.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b94465293a77c750fcf9d623b8c32f534ec288d22707af7470d905e26c5a3235
3
+ size 406
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 14.94,
3
+ "train_loss": 1.3876537419981876,
4
+ "train_runtime": 7307.918,
5
+ "train_samples": 3788,
6
+ "train_samples_per_second": 7.775,
7
+ "train_steps_per_second": 0.242
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 14.936708860759493,
5
+ "eval_steps": 100,
6
+ "global_step": 1770,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.84,
13
+ "eval_loss": 4.394069194793701,
14
+ "eval_runtime": 118.8089,
15
+ "eval_samples_per_second": 14.99,
16
+ "eval_steps_per_second": 1.877,
17
+ "eval_wer": 1.0,
18
+ "step": 100
19
+ },
20
+ {
21
+ "epoch": 1.69,
22
+ "eval_loss": 3.200528383255005,
23
+ "eval_runtime": 116.974,
24
+ "eval_samples_per_second": 15.226,
25
+ "eval_steps_per_second": 1.906,
26
+ "eval_wer": 1.0,
27
+ "step": 200
28
+ },
29
+ {
30
+ "epoch": 2.53,
31
+ "eval_loss": 2.7843868732452393,
32
+ "eval_runtime": 118.9835,
33
+ "eval_samples_per_second": 14.968,
34
+ "eval_steps_per_second": 1.874,
35
+ "eval_wer": 1.0144723900519883,
36
+ "step": 300
37
+ },
38
+ {
39
+ "epoch": 3.38,
40
+ "eval_loss": 0.8691409230232239,
41
+ "eval_runtime": 118.7268,
42
+ "eval_samples_per_second": 15.001,
43
+ "eval_steps_per_second": 1.878,
44
+ "eval_wer": 1.000281017282563,
45
+ "step": 400
46
+ },
47
+ {
48
+ "epoch": 4.22,
49
+ "learning_rate": 0.0002988,
50
+ "loss": 4.317,
51
+ "step": 500
52
+ },
53
+ {
54
+ "epoch": 4.22,
55
+ "eval_loss": 0.684583306312561,
56
+ "eval_runtime": 119.9053,
57
+ "eval_samples_per_second": 14.853,
58
+ "eval_steps_per_second": 1.86,
59
+ "eval_wer": 0.8393986230153154,
60
+ "step": 500
61
+ },
62
+ {
63
+ "epoch": 5.06,
64
+ "eval_loss": 0.6270079612731934,
65
+ "eval_runtime": 120.1998,
66
+ "eval_samples_per_second": 14.817,
67
+ "eval_steps_per_second": 1.855,
68
+ "eval_wer": 0.7789799072642968,
69
+ "step": 600
70
+ },
71
+ {
72
+ "epoch": 5.91,
73
+ "eval_loss": 0.593498945236206,
74
+ "eval_runtime": 120.001,
75
+ "eval_samples_per_second": 14.842,
76
+ "eval_steps_per_second": 1.858,
77
+ "eval_wer": 0.7802444850358297,
78
+ "step": 700
79
+ },
80
+ {
81
+ "epoch": 6.75,
82
+ "eval_loss": 0.570054829120636,
83
+ "eval_runtime": 120.3712,
84
+ "eval_samples_per_second": 14.796,
85
+ "eval_steps_per_second": 1.853,
86
+ "eval_wer": 0.7812280455247997,
87
+ "step": 800
88
+ },
89
+ {
90
+ "epoch": 7.59,
91
+ "eval_loss": 0.564914345741272,
92
+ "eval_runtime": 122.0077,
93
+ "eval_samples_per_second": 14.597,
94
+ "eval_steps_per_second": 1.828,
95
+ "eval_wer": 0.7890965294365604,
96
+ "step": 900
97
+ },
98
+ {
99
+ "epoch": 8.44,
100
+ "learning_rate": 0.00018236220472440942,
101
+ "loss": 0.3656,
102
+ "step": 1000
103
+ },
104
+ {
105
+ "epoch": 8.44,
106
+ "eval_loss": 0.6092184782028198,
107
+ "eval_runtime": 120.3224,
108
+ "eval_samples_per_second": 14.802,
109
+ "eval_steps_per_second": 1.853,
110
+ "eval_wer": 0.8177602922579739,
111
+ "step": 1000
112
+ },
113
+ {
114
+ "epoch": 9.28,
115
+ "eval_loss": 0.6092739701271057,
116
+ "eval_runtime": 120.225,
117
+ "eval_samples_per_second": 14.814,
118
+ "eval_steps_per_second": 1.855,
119
+ "eval_wer": 0.7720949838415062,
120
+ "step": 1100
121
+ },
122
+ {
123
+ "epoch": 10.13,
124
+ "eval_loss": 0.6154232621192932,
125
+ "eval_runtime": 120.1186,
126
+ "eval_samples_per_second": 14.827,
127
+ "eval_steps_per_second": 1.856,
128
+ "eval_wer": 0.7286778136855416,
129
+ "step": 1200
130
+ },
131
+ {
132
+ "epoch": 10.97,
133
+ "eval_loss": 0.6283690333366394,
134
+ "eval_runtime": 122.0867,
135
+ "eval_samples_per_second": 14.588,
136
+ "eval_steps_per_second": 1.827,
137
+ "eval_wer": 0.7407615568357454,
138
+ "step": 1300
139
+ },
140
+ {
141
+ "epoch": 11.81,
142
+ "eval_loss": 0.6343453526496887,
143
+ "eval_runtime": 120.4977,
144
+ "eval_samples_per_second": 14.78,
145
+ "eval_steps_per_second": 1.851,
146
+ "eval_wer": 0.7143459322748349,
147
+ "step": 1400
148
+ },
149
+ {
150
+ "epoch": 12.66,
151
+ "learning_rate": 6.425196850393701e-05,
152
+ "loss": 0.1681,
153
+ "step": 1500
154
+ },
155
+ {
156
+ "epoch": 12.66,
157
+ "eval_loss": 0.6523498296737671,
158
+ "eval_runtime": 119.5948,
159
+ "eval_samples_per_second": 14.892,
160
+ "eval_steps_per_second": 1.865,
161
+ "eval_wer": 0.7362652803147394,
162
+ "step": 1500
163
+ },
164
+ {
165
+ "epoch": 13.5,
166
+ "eval_loss": 0.6542894244194031,
167
+ "eval_runtime": 120.5721,
168
+ "eval_samples_per_second": 14.771,
169
+ "eval_steps_per_second": 1.85,
170
+ "eval_wer": 0.7139244063509906,
171
+ "step": 1600
172
+ },
173
+ {
174
+ "epoch": 14.35,
175
+ "eval_loss": 0.6598775386810303,
176
+ "eval_runtime": 124.634,
177
+ "eval_samples_per_second": 14.29,
178
+ "eval_steps_per_second": 1.789,
179
+ "eval_wer": 0.7094281298299846,
180
+ "step": 1700
181
+ },
182
+ {
183
+ "epoch": 14.94,
184
+ "step": 1770,
185
+ "total_flos": 7.546311277696558e+18,
186
+ "train_loss": 1.3876537419981876,
187
+ "train_runtime": 7307.918,
188
+ "train_samples_per_second": 7.775,
189
+ "train_steps_per_second": 0.242
190
+ }
191
+ ],
192
+ "logging_steps": 500,
193
+ "max_steps": 1770,
194
+ "num_input_tokens_seen": 0,
195
+ "num_train_epochs": 15,
196
+ "save_steps": 400,
197
+ "total_flos": 7.546311277696558e+18,
198
+ "train_batch_size": 16,
199
+ "trial_name": null,
200
+ "trial_params": null
201
+ }