mt5-simplification-spanish-clara-med
This model is a fine-tuned version of oskrmiguel/mt5-simplification-spanish on the None dataset. It achieves the following results on the evaluation set:
- Loss: 1.9610
- Rouge1: 33.7922
- Rouge2: 19.5758
- Rougel: 31.3737
- Rougelsum: 31.3428
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5.6e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 30
Training results
Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
---|---|---|---|---|---|---|---|
No log | 1.0 | 190 | 2.6876 | 32.236 | 18.2352 | 29.7852 | 29.7539 |
No log | 2.0 | 380 | 2.4617 | 32.8521 | 18.9712 | 30.4958 | 30.4635 |
3.3018 | 3.0 | 570 | 2.3487 | 33.2554 | 19.3441 | 30.9036 | 30.8525 |
3.3018 | 4.0 | 760 | 2.2711 | 33.0105 | 19.01 | 30.6851 | 30.5767 |
2.7431 | 5.0 | 950 | 2.2254 | 33.1301 | 18.9618 | 30.6744 | 30.6284 |
2.7431 | 6.0 | 1140 | 2.1847 | 33.3701 | 19.1884 | 30.9138 | 30.8611 |
2.7431 | 7.0 | 1330 | 2.1443 | 33.3158 | 19.101 | 30.8317 | 30.7747 |
2.5154 | 8.0 | 1520 | 2.1072 | 33.1638 | 19.0139 | 30.7295 | 30.7162 |
2.5154 | 9.0 | 1710 | 2.0989 | 33.4925 | 19.2107 | 31.0253 | 30.9908 |
2.3763 | 10.0 | 1900 | 2.0709 | 33.3007 | 18.9519 | 30.847 | 30.8018 |
2.3763 | 11.0 | 2090 | 2.0631 | 33.4689 | 19.1995 | 31.0712 | 31.0327 |
2.3763 | 12.0 | 2280 | 2.0418 | 33.2536 | 19.027 | 30.898 | 30.8695 |
2.2811 | 13.0 | 2470 | 2.0345 | 33.5097 | 19.2219 | 31.1057 | 31.0683 |
2.2811 | 14.0 | 2660 | 2.0185 | 33.3544 | 19.1241 | 30.913 | 30.8873 |
2.2173 | 15.0 | 2850 | 2.0138 | 33.3856 | 19.2065 | 31.0173 | 30.9447 |
2.2173 | 16.0 | 3040 | 2.0019 | 33.4035 | 19.1803 | 31.0154 | 30.981 |
2.2173 | 17.0 | 3230 | 1.9977 | 33.4059 | 19.3078 | 31.1196 | 31.0692 |
2.1612 | 18.0 | 3420 | 1.9883 | 33.5097 | 19.3637 | 31.0966 | 31.0554 |
2.1612 | 19.0 | 3610 | 1.9828 | 33.4965 | 19.2754 | 31.1267 | 31.1021 |
2.1115 | 20.0 | 3800 | 1.9834 | 33.7514 | 19.5325 | 31.2833 | 31.2418 |
2.1115 | 21.0 | 3990 | 1.9754 | 33.6193 | 19.429 | 31.2721 | 31.2267 |
2.1115 | 22.0 | 4180 | 1.9716 | 33.5212 | 19.3637 | 31.1326 | 31.1162 |
2.0824 | 23.0 | 4370 | 1.9667 | 33.5156 | 19.3223 | 31.1023 | 31.0709 |
2.0824 | 24.0 | 4560 | 1.9735 | 33.6089 | 19.3842 | 31.1539 | 31.1419 |
2.0657 | 25.0 | 4750 | 1.9674 | 33.6317 | 19.4044 | 31.2361 | 31.2222 |
2.0657 | 26.0 | 4940 | 1.9617 | 33.745 | 19.5099 | 31.3061 | 31.2643 |
2.0657 | 27.0 | 5130 | 1.9613 | 33.7798 | 19.5496 | 31.3761 | 31.3356 |
2.0511 | 28.0 | 5320 | 1.9635 | 33.8568 | 19.594 | 31.4454 | 31.4141 |
2.0511 | 29.0 | 5510 | 1.9609 | 33.805 | 19.5962 | 31.393 | 31.3493 |
2.0377 | 30.0 | 5700 | 1.9610 | 33.7922 | 19.5758 | 31.3737 | 31.3428 |
Framework versions
- Transformers 4.25.1
- Pytorch 1.13.0
- Datasets 2.8.0
- Tokenizers 0.12.1
- Downloads last month
- 71
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.