samitizerxu commited on
Commit
391a169
1 Parent(s): 24fa7b0

Added model files

Browse files
README.md ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - eo
4
+ license: apache-2.0
5
+ tags:
6
+ - automatic-speech-recognition
7
+ - robust-speech-event
8
+ - common_voice
9
+ - generated_from_trainer
10
+ datasets:
11
+ - common_voice
12
+ model-index:
13
+ - name: wav2vec2-xls-r-300m-eo
14
+ results: []
15
+ ---
16
+
17
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
+ should probably proofread and complete it, then remove this comment. -->
19
+
20
+ # wav2vec2-xls-r-300m-eo
21
+
22
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the COMMON_VOICE - EO dataset.
23
+ It achieves the following results on the evaluation set:
24
+ - Loss: 0.2584
25
+ - Wer: 0.3114
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 0.0003
45
+ - train_batch_size: 16
46
+ - eval_batch_size: 8
47
+ - seed: 42
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - num_epochs: 20.0
51
+ - mixed_precision_training: Native AMP
52
+
53
+ ### Training results
54
+
55
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
56
+ |:-------------:|:-----:|:-----:|:---------------:|:------:|
57
+ | 3.1701 | 0.8 | 500 | 2.8105 | 1.0 |
58
+ | 1.9143 | 1.6 | 1000 | 0.5977 | 0.7002 |
59
+ | 1.1259 | 2.4 | 1500 | 0.5063 | 0.6157 |
60
+ | 0.9732 | 3.2 | 2000 | 0.4264 | 0.5673 |
61
+ | 0.8983 | 4.0 | 2500 | 0.4249 | 0.4902 |
62
+ | 0.8507 | 4.8 | 3000 | 0.3811 | 0.4536 |
63
+ | 0.8064 | 5.6 | 3500 | 0.3643 | 0.4467 |
64
+ | 0.7866 | 6.4 | 4000 | 0.3600 | 0.4453 |
65
+ | 0.7773 | 7.2 | 4500 | 0.3724 | 0.4470 |
66
+ | 0.747 | 8.0 | 5000 | 0.3501 | 0.4189 |
67
+ | 0.7279 | 8.8 | 5500 | 0.3500 | 0.4261 |
68
+ | 0.7153 | 9.6 | 6000 | 0.3328 | 0.3966 |
69
+ | 0.7 | 10.4 | 6500 | 0.3314 | 0.3869 |
70
+ | 0.6784 | 11.2 | 7000 | 0.3396 | 0.4051 |
71
+ | 0.6582 | 12.0 | 7500 | 0.3236 | 0.3899 |
72
+ | 0.6478 | 12.8 | 8000 | 0.3263 | 0.3832 |
73
+ | 0.6277 | 13.6 | 8500 | 0.3139 | 0.3769 |
74
+ | 0.6053 | 14.4 | 9000 | 0.2955 | 0.3536 |
75
+ | 0.5777 | 15.2 | 9500 | 0.2793 | 0.3413 |
76
+ | 0.5631 | 16.0 | 10000 | 0.2789 | 0.3353 |
77
+ | 0.5446 | 16.8 | 10500 | 0.2709 | 0.3264 |
78
+ | 0.528 | 17.6 | 11000 | 0.2693 | 0.3234 |
79
+ | 0.5169 | 18.4 | 11500 | 0.2656 | 0.3193 |
80
+ | 0.5041 | 19.2 | 12000 | 0.2575 | 0.3102 |
81
+ | 0.4971 | 20.0 | 12500 | 0.2584 | 0.3114 |
82
+
83
+
84
+ ### Framework versions
85
+
86
+ - Transformers 4.17.0.dev0
87
+ - Pytorch 1.10.2+cu102
88
+ - Datasets 1.18.2.dev0
89
+ - Tokenizers 0.11.0
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<s>": 55, "</s>": 56}
all_results.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "eval_loss": 0.2584246098995209,
4
+ "eval_runtime": 235.5674,
5
+ "eval_samples": 5000,
6
+ "eval_samples_per_second": 21.225,
7
+ "eval_steps_per_second": 2.653,
8
+ "eval_wer": 0.31141699083267593,
9
+ "train_loss": 0.8456691796875,
10
+ "train_runtime": 20148.7553,
11
+ "train_samples": 10000,
12
+ "train_samples_per_second": 9.926,
13
+ "train_steps_per_second": 0.62
14
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "eval_loss": 0.2584246098995209,
4
+ "eval_runtime": 235.5674,
5
+ "eval_samples": 5000,
6
+ "eval_samples_per_second": 21.225,
7
+ "eval_steps_per_second": 2.653,
8
+ "eval_wer": 0.31141699083267593
9
+ }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2bf3aed7169f1b2cc746039e640bd88b4d53c287920faa6743254e924058776d
3
  size 1262157361
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f3326278c8c0c1ab021b5c56c7e23a0a60336147e6585104e036ef2b55b33e7
3
  size 1262157361
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "./wav2vec2-xls-r-300m-eo", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "train_loss": 0.8456691796875,
4
+ "train_runtime": 20148.7553,
5
+ "train_samples": 10000,
6
+ "train_samples_per_second": 9.926,
7
+ "train_steps_per_second": 0.62
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,400 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 20.0,
5
+ "global_step": 12500,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.8,
12
+ "learning_rate": 0.00028807199999999995,
13
+ "loss": 3.1701,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.8,
18
+ "eval_loss": 2.8104631900787354,
19
+ "eval_runtime": 249.0567,
20
+ "eval_samples_per_second": 20.076,
21
+ "eval_steps_per_second": 2.509,
22
+ "eval_wer": 1.0,
23
+ "step": 500
24
+ },
25
+ {
26
+ "epoch": 1.6,
27
+ "learning_rate": 0.000276072,
28
+ "loss": 1.9143,
29
+ "step": 1000
30
+ },
31
+ {
32
+ "epoch": 1.6,
33
+ "eval_loss": 0.5976669192314148,
34
+ "eval_runtime": 247.4077,
35
+ "eval_samples_per_second": 20.21,
36
+ "eval_steps_per_second": 2.526,
37
+ "eval_wer": 0.7001713772748293,
38
+ "step": 1000
39
+ },
40
+ {
41
+ "epoch": 2.4,
42
+ "learning_rate": 0.00026407199999999996,
43
+ "loss": 1.1259,
44
+ "step": 1500
45
+ },
46
+ {
47
+ "epoch": 2.4,
48
+ "eval_loss": 0.5062595009803772,
49
+ "eval_runtime": 241.7474,
50
+ "eval_samples_per_second": 20.683,
51
+ "eval_steps_per_second": 2.585,
52
+ "eval_wer": 0.6156796605097794,
53
+ "step": 1500
54
+ },
55
+ {
56
+ "epoch": 3.2,
57
+ "learning_rate": 0.000252072,
58
+ "loss": 0.9732,
59
+ "step": 2000
60
+ },
61
+ {
62
+ "epoch": 3.2,
63
+ "eval_loss": 0.42641571164131165,
64
+ "eval_runtime": 234.1368,
65
+ "eval_samples_per_second": 21.355,
66
+ "eval_steps_per_second": 2.669,
67
+ "eval_wer": 0.5673403879111014,
68
+ "step": 2000
69
+ },
70
+ {
71
+ "epoch": 4.0,
72
+ "learning_rate": 0.000240096,
73
+ "loss": 0.8983,
74
+ "step": 2500
75
+ },
76
+ {
77
+ "epoch": 4.0,
78
+ "eval_loss": 0.42490196228027344,
79
+ "eval_runtime": 237.618,
80
+ "eval_samples_per_second": 21.042,
81
+ "eval_steps_per_second": 2.63,
82
+ "eval_wer": 0.49024781697995157,
83
+ "step": 2500
84
+ },
85
+ {
86
+ "epoch": 4.8,
87
+ "learning_rate": 0.00022809599999999998,
88
+ "loss": 0.8507,
89
+ "step": 3000
90
+ },
91
+ {
92
+ "epoch": 4.8,
93
+ "eval_loss": 0.38109204173088074,
94
+ "eval_runtime": 243.3726,
95
+ "eval_samples_per_second": 20.545,
96
+ "eval_steps_per_second": 2.568,
97
+ "eval_wer": 0.4535785207148881,
98
+ "step": 3000
99
+ },
100
+ {
101
+ "epoch": 5.6,
102
+ "learning_rate": 0.00021609599999999996,
103
+ "loss": 0.8064,
104
+ "step": 3500
105
+ },
106
+ {
107
+ "epoch": 5.6,
108
+ "eval_loss": 0.3643375337123871,
109
+ "eval_runtime": 241.9721,
110
+ "eval_samples_per_second": 20.664,
111
+ "eval_steps_per_second": 2.583,
112
+ "eval_wer": 0.44669622697967953,
113
+ "step": 3500
114
+ },
115
+ {
116
+ "epoch": 6.4,
117
+ "learning_rate": 0.000204096,
118
+ "loss": 0.7866,
119
+ "step": 4000
120
+ },
121
+ {
122
+ "epoch": 6.4,
123
+ "eval_loss": 0.36003848910331726,
124
+ "eval_runtime": 240.8042,
125
+ "eval_samples_per_second": 20.764,
126
+ "eval_steps_per_second": 2.595,
127
+ "eval_wer": 0.44533608987785966,
128
+ "step": 4000
129
+ },
130
+ {
131
+ "epoch": 7.2,
132
+ "learning_rate": 0.00019209599999999997,
133
+ "loss": 0.7773,
134
+ "step": 4500
135
+ },
136
+ {
137
+ "epoch": 7.2,
138
+ "eval_loss": 0.37241309881210327,
139
+ "eval_runtime": 234.1677,
140
+ "eval_samples_per_second": 21.352,
141
+ "eval_steps_per_second": 2.669,
142
+ "eval_wer": 0.4470226598841163,
143
+ "step": 4500
144
+ },
145
+ {
146
+ "epoch": 8.0,
147
+ "learning_rate": 0.00018009599999999998,
148
+ "loss": 0.747,
149
+ "step": 5000
150
+ },
151
+ {
152
+ "epoch": 8.0,
153
+ "eval_loss": 0.3500821888446808,
154
+ "eval_runtime": 239.967,
155
+ "eval_samples_per_second": 20.836,
156
+ "eval_steps_per_second": 2.605,
157
+ "eval_wer": 0.41892222736051793,
158
+ "step": 5000
159
+ },
160
+ {
161
+ "epoch": 8.8,
162
+ "learning_rate": 0.00016809599999999998,
163
+ "loss": 0.7279,
164
+ "step": 5500
165
+ },
166
+ {
167
+ "epoch": 8.8,
168
+ "eval_loss": 0.3500120937824249,
169
+ "eval_runtime": 236.3527,
170
+ "eval_samples_per_second": 21.155,
171
+ "eval_steps_per_second": 2.644,
172
+ "eval_wer": 0.42607654851609045,
173
+ "step": 5500
174
+ },
175
+ {
176
+ "epoch": 9.6,
177
+ "learning_rate": 0.00015612,
178
+ "loss": 0.7153,
179
+ "step": 6000
180
+ },
181
+ {
182
+ "epoch": 9.6,
183
+ "eval_loss": 0.33280453085899353,
184
+ "eval_runtime": 237.5821,
185
+ "eval_samples_per_second": 21.045,
186
+ "eval_steps_per_second": 2.631,
187
+ "eval_wer": 0.3965887761486358,
188
+ "step": 6000
189
+ },
190
+ {
191
+ "epoch": 10.4,
192
+ "learning_rate": 0.000144144,
193
+ "loss": 0.7,
194
+ "step": 6500
195
+ },
196
+ {
197
+ "epoch": 10.4,
198
+ "eval_loss": 0.33142638206481934,
199
+ "eval_runtime": 238.2506,
200
+ "eval_samples_per_second": 20.986,
201
+ "eval_steps_per_second": 2.623,
202
+ "eval_wer": 0.3869045999836784,
203
+ "step": 6500
204
+ },
205
+ {
206
+ "epoch": 11.2,
207
+ "learning_rate": 0.000132168,
208
+ "loss": 0.6784,
209
+ "step": 7000
210
+ },
211
+ {
212
+ "epoch": 11.2,
213
+ "eval_loss": 0.33956480026245117,
214
+ "eval_runtime": 241.1503,
215
+ "eval_samples_per_second": 20.734,
216
+ "eval_steps_per_second": 2.592,
217
+ "eval_wer": 0.40510323440602813,
218
+ "step": 7000
219
+ },
220
+ {
221
+ "epoch": 12.0,
222
+ "learning_rate": 0.00012016799999999999,
223
+ "loss": 0.6582,
224
+ "step": 7500
225
+ },
226
+ {
227
+ "epoch": 12.0,
228
+ "eval_loss": 0.3236238658428192,
229
+ "eval_runtime": 237.147,
230
+ "eval_samples_per_second": 21.084,
231
+ "eval_steps_per_second": 2.635,
232
+ "eval_wer": 0.38992410434971847,
233
+ "step": 7500
234
+ },
235
+ {
236
+ "epoch": 12.8,
237
+ "learning_rate": 0.00010816799999999998,
238
+ "loss": 0.6478,
239
+ "step": 8000
240
+ },
241
+ {
242
+ "epoch": 12.8,
243
+ "eval_loss": 0.3263108730316162,
244
+ "eval_runtime": 239.702,
245
+ "eval_samples_per_second": 20.859,
246
+ "eval_steps_per_second": 2.607,
247
+ "eval_wer": 0.3832050270667283,
248
+ "step": 8000
249
+ },
250
+ {
251
+ "epoch": 13.6,
252
+ "learning_rate": 9.616799999999999e-05,
253
+ "loss": 0.6277,
254
+ "step": 8500
255
+ },
256
+ {
257
+ "epoch": 13.6,
258
+ "eval_loss": 0.3138865828514099,
259
+ "eval_runtime": 241.2261,
260
+ "eval_samples_per_second": 20.727,
261
+ "eval_steps_per_second": 2.591,
262
+ "eval_wer": 0.37692119365632054,
263
+ "step": 8500
264
+ },
265
+ {
266
+ "epoch": 14.4,
267
+ "learning_rate": 8.416799999999998e-05,
268
+ "loss": 0.6053,
269
+ "step": 9000
270
+ },
271
+ {
272
+ "epoch": 14.4,
273
+ "eval_loss": 0.2955300807952881,
274
+ "eval_runtime": 239.5196,
275
+ "eval_samples_per_second": 20.875,
276
+ "eval_steps_per_second": 2.609,
277
+ "eval_wer": 0.3535812409890917,
278
+ "step": 9000
279
+ },
280
+ {
281
+ "epoch": 15.2,
282
+ "learning_rate": 7.216799999999999e-05,
283
+ "loss": 0.5777,
284
+ "step": 9500
285
+ },
286
+ {
287
+ "epoch": 15.2,
288
+ "eval_loss": 0.27926793694496155,
289
+ "eval_runtime": 243.0206,
290
+ "eval_samples_per_second": 20.574,
291
+ "eval_steps_per_second": 2.572,
292
+ "eval_wer": 0.3413128043306765,
293
+ "step": 9500
294
+ },
295
+ {
296
+ "epoch": 16.0,
297
+ "learning_rate": 6.016799999999999e-05,
298
+ "loss": 0.5631,
299
+ "step": 10000
300
+ },
301
+ {
302
+ "epoch": 16.0,
303
+ "eval_loss": 0.2789314091205597,
304
+ "eval_runtime": 235.0589,
305
+ "eval_samples_per_second": 21.271,
306
+ "eval_steps_per_second": 2.659,
307
+ "eval_wer": 0.33532820108266914,
308
+ "step": 10000
309
+ },
310
+ {
311
+ "epoch": 16.8,
312
+ "learning_rate": 4.8168e-05,
313
+ "loss": 0.5446,
314
+ "step": 10500
315
+ },
316
+ {
317
+ "epoch": 16.8,
318
+ "eval_loss": 0.27094796299934387,
319
+ "eval_runtime": 244.8594,
320
+ "eval_samples_per_second": 20.42,
321
+ "eval_steps_per_second": 2.552,
322
+ "eval_wer": 0.32637849895269444,
323
+ "step": 10500
324
+ },
325
+ {
326
+ "epoch": 17.6,
327
+ "learning_rate": 3.6167999999999997e-05,
328
+ "loss": 0.528,
329
+ "step": 11000
330
+ },
331
+ {
332
+ "epoch": 17.6,
333
+ "eval_loss": 0.26928168535232544,
334
+ "eval_runtime": 237.8703,
335
+ "eval_samples_per_second": 21.02,
336
+ "eval_steps_per_second": 2.627,
337
+ "eval_wer": 0.3233861973286907,
338
+ "step": 11000
339
+ },
340
+ {
341
+ "epoch": 18.4,
342
+ "learning_rate": 2.4191999999999998e-05,
343
+ "loss": 0.5169,
344
+ "step": 11500
345
+ },
346
+ {
347
+ "epoch": 18.4,
348
+ "eval_loss": 0.265558123588562,
349
+ "eval_runtime": 240.7332,
350
+ "eval_samples_per_second": 20.77,
351
+ "eval_steps_per_second": 2.596,
352
+ "eval_wer": 0.31925138053915836,
353
+ "step": 11500
354
+ },
355
+ {
356
+ "epoch": 19.2,
357
+ "learning_rate": 1.2215999999999998e-05,
358
+ "loss": 0.5041,
359
+ "step": 12000
360
+ },
361
+ {
362
+ "epoch": 19.2,
363
+ "eval_loss": 0.2575398087501526,
364
+ "eval_runtime": 242.0688,
365
+ "eval_samples_per_second": 20.655,
366
+ "eval_steps_per_second": 2.582,
367
+ "eval_wer": 0.31024727292511084,
368
+ "step": 12000
369
+ },
370
+ {
371
+ "epoch": 20.0,
372
+ "learning_rate": 2.16e-07,
373
+ "loss": 0.4971,
374
+ "step": 12500
375
+ },
376
+ {
377
+ "epoch": 20.0,
378
+ "eval_loss": 0.2584246098995209,
379
+ "eval_runtime": 247.3331,
380
+ "eval_samples_per_second": 20.216,
381
+ "eval_steps_per_second": 2.527,
382
+ "eval_wer": 0.31141699083267593,
383
+ "step": 12500
384
+ },
385
+ {
386
+ "epoch": 20.0,
387
+ "step": 12500,
388
+ "total_flos": 3.751048304986411e+19,
389
+ "train_loss": 0.8456691796875,
390
+ "train_runtime": 20148.7553,
391
+ "train_samples_per_second": 9.926,
392
+ "train_steps_per_second": 0.62
393
+ }
394
+ ],
395
+ "max_steps": 12500,
396
+ "num_train_epochs": 20,
397
+ "total_flos": 3.751048304986411e+19,
398
+ "trial_name": null,
399
+ "trial_params": null
400
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:312fb5b03ffb7643c483c67d3f9c080ba092705560e886015616251b77d99ba2
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c6b5ee3512f0a11b714e02a1d75e8b09bbdb1f199638928581165457d0f25fc
3
  size 3055
vocab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"!": 1, "\"": 2, "'": 3, "(": 4, ")": 5, ",": 6, "-": 7, ".": 8, ":": 9, ";": 10, "?": 11, "a": 12, "b": 13, "c": 14, "d": 15, "e": 16, "f": 17, "g": 18, "h": 19, "i": 20, "j": 21, "k": 22, "l": 23, "m": 24, "n": 25, "o": 26, "p": 27, "r": 28, "s": 29, "t": 30, "u": 31, "v": 32, "y": 33, "z": 34, "«": 35, "»": 36, "é": 37, "ñ": 38, "ĉ": 39, "ĝ": 40, "ĥ": 41, "ĵ": 42, "ŝ": 43, "ŭ": 44, "–": 45, "—": 46, "‘": 47, "’": 48, "“": 49, "”": 50, "„": 51, "fi": 52, "|": 0, "[UNK]": 53, "[PAD]": 54}