anushaporwal commited on
Commit
1a69170
1 Parent(s): e670d83

End of training

Browse files
Files changed (5) hide show
  1. README.md +10 -6
  2. all_results.json +15 -0
  3. eval_results.json +9 -0
  4. train_results.json +9 -0
  5. trainer_state.json +94 -0
README.md CHANGED
@@ -1,7 +1,11 @@
1
  ---
 
 
2
  license: apache-2.0
3
  base_model: facebook/wav2vec2-large-xlsr-53
4
  tags:
 
 
5
  - generated_from_trainer
6
  datasets:
7
  - common_voice_16_0
@@ -14,15 +18,15 @@ model-index:
14
  name: Automatic Speech Recognition
15
  type: automatic-speech-recognition
16
  dataset:
17
- name: common_voice_16_0
18
  type: common_voice_16_0
19
  config: tr
20
  split: test[0:250]
21
- args: tr
22
  metrics:
23
  - name: Wer
24
  type: wer
25
- value: 0.9729938271604939
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -30,10 +34,10 @@ should probably proofread and complete it, then remove this comment. -->
30
 
31
  # wav2vec2-common_voice-tr-demo-mini
32
 
33
- This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the common_voice_16_0 dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 1.2465
36
- - Wer: 0.9730
37
 
38
  ## Model description
39
 
 
1
  ---
2
+ language:
3
+ - tr
4
  license: apache-2.0
5
  base_model: facebook/wav2vec2-large-xlsr-53
6
  tags:
7
+ - automatic-speech-recognition
8
+ - mozilla-foundation/common_voice_16_0
9
  - generated_from_trainer
10
  datasets:
11
  - common_voice_16_0
 
18
  name: Automatic Speech Recognition
19
  type: automatic-speech-recognition
20
  dataset:
21
+ name: MOZILLA-FOUNDATION/COMMON_VOICE_16_0 - TR
22
  type: common_voice_16_0
23
  config: tr
24
  split: test[0:250]
25
+ args: 'Config: tr, Training split: train[0:3000], Eval split: test[0:250]'
26
  metrics:
27
  - name: Wer
28
  type: wer
29
+ value: 0.9382716049382716
30
  ---
31
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
34
 
35
  # wav2vec2-common_voice-tr-demo-mini
36
 
37
+ This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the MOZILLA-FOUNDATION/COMMON_VOICE_16_0 - TR dataset.
38
  It achieves the following results on the evaluation set:
39
+ - Loss: 0.9823
40
+ - Wer: 0.9383
41
 
42
  ## Model description
43
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.992,
3
+ "eval_loss": 0.9822820425033569,
4
+ "eval_runtime": 8.381,
5
+ "eval_samples": 250,
6
+ "eval_samples_per_second": 29.829,
7
+ "eval_steps_per_second": 3.818,
8
+ "eval_wer": 0.9382716049382716,
9
+ "total_flos": 1.0050115218562714e+18,
10
+ "train_loss": 4.179523862406957,
11
+ "train_runtime": 1022.6563,
12
+ "train_samples": 3000,
13
+ "train_samples_per_second": 8.801,
14
+ "train_steps_per_second": 0.549
15
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.992,
3
+ "eval_loss": 0.9822820425033569,
4
+ "eval_runtime": 8.381,
5
+ "eval_samples": 250,
6
+ "eval_samples_per_second": 29.829,
7
+ "eval_steps_per_second": 3.818,
8
+ "eval_wer": 0.9382716049382716
9
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.992,
3
+ "total_flos": 1.0050115218562714e+18,
4
+ "train_loss": 4.179523862406957,
5
+ "train_runtime": 1022.6563,
6
+ "train_samples": 3000,
7
+ "train_samples_per_second": 8.801,
8
+ "train_steps_per_second": 0.549
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.992,
5
+ "eval_steps": 100,
6
+ "global_step": 561,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.5333333333333333,
13
+ "eval_loss": 4.023805141448975,
14
+ "eval_runtime": 9.119,
15
+ "eval_samples_per_second": 27.415,
16
+ "eval_steps_per_second": 3.509,
17
+ "eval_wer": 1.0,
18
+ "step": 100
19
+ },
20
+ {
21
+ "epoch": 1.0666666666666667,
22
+ "eval_loss": 3.2450873851776123,
23
+ "eval_runtime": 8.4082,
24
+ "eval_samples_per_second": 29.733,
25
+ "eval_steps_per_second": 3.806,
26
+ "eval_wer": 1.0,
27
+ "step": 200
28
+ },
29
+ {
30
+ "epoch": 1.6,
31
+ "eval_loss": 2.9997410774230957,
32
+ "eval_runtime": 8.4129,
33
+ "eval_samples_per_second": 29.716,
34
+ "eval_steps_per_second": 3.804,
35
+ "eval_wer": 1.0,
36
+ "step": 300
37
+ },
38
+ {
39
+ "epoch": 2.1333333333333333,
40
+ "eval_loss": 1.425586223602295,
41
+ "eval_runtime": 8.5669,
42
+ "eval_samples_per_second": 29.182,
43
+ "eval_steps_per_second": 3.735,
44
+ "eval_wer": 1.0054012345679013,
45
+ "step": 400
46
+ },
47
+ {
48
+ "epoch": 2.6666666666666665,
49
+ "grad_norm": 1.884314775466919,
50
+ "learning_rate": 0.0002982,
51
+ "loss": 4.5926,
52
+ "step": 500
53
+ },
54
+ {
55
+ "epoch": 2.6666666666666665,
56
+ "eval_loss": 1.2465466260910034,
57
+ "eval_runtime": 610.4372,
58
+ "eval_samples_per_second": 0.41,
59
+ "eval_steps_per_second": 0.052,
60
+ "eval_wer": 0.9729938271604939,
61
+ "step": 500
62
+ },
63
+ {
64
+ "epoch": 2.992,
65
+ "step": 561,
66
+ "total_flos": 1.0050115218562714e+18,
67
+ "train_loss": 4.179523862406957,
68
+ "train_runtime": 1022.6563,
69
+ "train_samples_per_second": 8.801,
70
+ "train_steps_per_second": 0.549
71
+ }
72
+ ],
73
+ "logging_steps": 500,
74
+ "max_steps": 561,
75
+ "num_input_tokens_seen": 0,
76
+ "num_train_epochs": 3,
77
+ "save_steps": 400,
78
+ "stateful_callbacks": {
79
+ "TrainerControl": {
80
+ "args": {
81
+ "should_epoch_stop": false,
82
+ "should_evaluate": false,
83
+ "should_log": false,
84
+ "should_save": true,
85
+ "should_training_stop": true
86
+ },
87
+ "attributes": {}
88
+ }
89
+ },
90
+ "total_flos": 1.0050115218562714e+18,
91
+ "train_batch_size": 8,
92
+ "trial_name": null,
93
+ "trial_params": null
94
+ }