EzraWilliam commited on
Commit
ee33e8b
1 Parent(s): 2b8aa70

End of training

Browse files
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
  license: apache-2.0
 
3
  tags:
4
  - generated_from_trainer
5
- base_model: facebook/wav2vec2-large-xlsr-53
6
  datasets:
7
  - xtreme_s
8
  metrics:
@@ -11,8 +11,8 @@ model-index:
11
  - name: wav2vec2-XLS-R-Fleurs-demo-google-colab-Ezra_William_Prod8
12
  results:
13
  - task:
14
- type: automatic-speech-recognition
15
  name: Automatic Speech Recognition
 
16
  dataset:
17
  name: xtreme_s
18
  type: xtreme_s
@@ -20,9 +20,9 @@ model-index:
20
  split: test
21
  args: fleurs.id_id
22
  metrics:
23
- - type: wer
24
- value: 0.42321508756174225
25
- name: Wer
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +32,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the xtreme_s dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.8564
36
- - Wer: 0.4232
37
 
38
  ## Model description
39
 
@@ -53,11 +53,11 @@ More information needed
53
 
54
  The following hyperparameters were used during training:
55
  - learning_rate: 0.001
56
- - train_batch_size: 16
57
  - eval_batch_size: 8
58
  - seed: 42
59
  - gradient_accumulation_steps: 4
60
- - total_train_batch_size: 64
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_steps: 600
@@ -68,16 +68,22 @@ The following hyperparameters were used during training:
68
 
69
  | Training Loss | Epoch | Step | Validation Loss | Wer |
70
  |:-------------:|:------:|:----:|:---------------:|:------:|
71
- | 4.801 | 30.77 | 300 | 2.8357 | 1.0 |
72
- | 1.041 | 61.54 | 600 | 0.8673 | 0.5433 |
73
- | 0.1141 | 92.31 | 900 | 0.8976 | 0.4801 |
74
- | 0.0568 | 123.08 | 1200 | 0.8556 | 0.4427 |
75
- | 0.035 | 153.85 | 1500 | 0.8564 | 0.4232 |
 
 
 
 
 
 
76
 
77
 
78
  ### Framework versions
79
 
80
- - Transformers 4.39.0
81
  - Pytorch 2.2.1+cu121
82
  - Datasets 2.18.0
83
  - Tokenizers 0.15.2
 
1
  ---
2
  license: apache-2.0
3
+ base_model: facebook/wav2vec2-large-xlsr-53
4
  tags:
5
  - generated_from_trainer
 
6
  datasets:
7
  - xtreme_s
8
  metrics:
 
11
  - name: wav2vec2-XLS-R-Fleurs-demo-google-colab-Ezra_William_Prod8
12
  results:
13
  - task:
 
14
  name: Automatic Speech Recognition
15
+ type: automatic-speech-recognition
16
  dataset:
17
  name: xtreme_s
18
  type: xtreme_s
 
20
  split: test
21
  args: fleurs.id_id
22
  metrics:
23
+ - name: Wer
24
+ type: wer
25
+ value: 0.42037120191588084
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
32
 
33
  This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the xtreme_s dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 0.9283
36
+ - Wer: 0.4204
37
 
38
  ## Model description
39
 
 
53
 
54
  The following hyperparameters were used during training:
55
  - learning_rate: 0.001
56
+ - train_batch_size: 8
57
  - eval_batch_size: 8
58
  - seed: 42
59
  - gradient_accumulation_steps: 4
60
+ - total_train_batch_size: 32
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_steps: 600
 
68
 
69
  | Training Loss | Epoch | Step | Validation Loss | Wer |
70
  |:-------------:|:------:|:----:|:---------------:|:------:|
71
+ | 4.7974 | 15.58 | 300 | 2.8438 | 1.0 |
72
+ | 1.0661 | 31.17 | 600 | 0.7520 | 0.5781 |
73
+ | 0.1697 | 46.75 | 900 | 0.8342 | 0.5181 |
74
+ | 0.1081 | 62.34 | 1200 | 0.8508 | 0.5010 |
75
+ | 0.0833 | 77.92 | 1500 | 0.9216 | 0.5014 |
76
+ | 0.0616 | 93.51 | 1800 | 0.9659 | 0.4859 |
77
+ | 0.0463 | 109.09 | 2100 | 0.9606 | 0.4646 |
78
+ | 0.0356 | 124.68 | 2400 | 0.9255 | 0.4578 |
79
+ | 0.0291 | 140.26 | 2700 | 0.9889 | 0.4503 |
80
+ | 0.0218 | 155.84 | 3000 | 0.9336 | 0.4371 |
81
+ | 0.0182 | 171.43 | 3300 | 0.9283 | 0.4204 |
82
 
83
 
84
  ### Framework versions
85
 
86
+ - Transformers 4.39.1
87
  - Pytorch 2.2.1+cu121
88
  - Datasets 2.18.0
89
  - Tokenizers 0.15.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41df5b28ae1ce1d1ea69e4bcb1777f74e798dd26dd8f165b4c22dd6a04778504
3
  size 1262073980
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a5d5fc4df8aa288b4d3ae3baf8310203378fd071fd9832368fd3f5d2422d51f
3
  size 1262073980
runs/Mar23_13-55-36_1deffeda88bf/events.out.tfevents.1711202170.1deffeda88bf.1671.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a75e4bf299ee8337486025be8d235856b622a7bbf35e0a853ea7d203ebe3584
3
- size 12207
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab1bd86b3f093c88e3fb154eacbe949e4e0907f0cf39c78d83423e54ff4ad5e8
3
+ size 12561