vgaraujov commited on
Commit
2a3e357
1 Parent(s): f004577

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +61 -0
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - es
5
+ - en
6
+ base_model: vgaraujov/bart-base-spanish
7
+ tags:
8
+ - generated_from_trainer
9
+ datasets:
10
+ - vgaraujov/wmt13
11
+ metrics:
12
+ - bleu
13
+ model-index:
14
+ - name: bart-base-translation-es-en
15
+ results:
16
+ - task:
17
+ name: Translation
18
+ type: translation
19
+ dataset:
20
+ name: vgaraujov/wmt13 es-en
21
+ type: vgaraujov/wmt13
22
+ config: es-en
23
+ split: validation
24
+ args: es-en
25
+ metrics:
26
+ - name: Bleu
27
+ type: bleu
28
+ value: 30.222
29
+ ---
30
+
31
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
32
+ should probably proofread and complete it, then remove this comment. -->
33
+
34
+ # BARTO (base-sized model) for es-en translation
35
+
36
+ This model is a fine-tuned version of [BARTO](https://huggingface.co/vgaraujov/bart-base-spanish) on a small portion of [WMT13](https://huggingface.co/datasets/vgaraujov/wmt13) es-en dataset.
37
+ It achieves the following results on the evaluation set:
38
+ - Loss: 1.4562
39
+ - Bleu: 30.222
40
+ - Gen Len: 42.0952
41
+
42
+ ### Training hyperparameters
43
+
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 0.005
46
+ - train_batch_size: 96
47
+ - eval_batch_size: 96
48
+ - seed: 42
49
+ - gradient_accumulation_steps: 4
50
+ - total_train_batch_size: 384
51
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
+ - lr_scheduler_type: linear
53
+ - lr_scheduler_warmup_steps: 40000
54
+ - training_steps: 5000
55
+
56
+ ### Framework versions
57
+
58
+ - Transformers 4.33.0.dev0
59
+ - Pytorch 2.0.1+cu117
60
+ - Datasets 2.14.4
61
+ - Tokenizers 0.13.3