llama-7b-finnish-instruct-v0.2_En__CMP_TR_size_304_epochs_10_2024-06-23_06-24-07_3558633
This model is a fine-tuned version of Finnish-NLP/llama-7b-finnish-instruct-v0.2 on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.6367
- Accuracy: 0.767
- Chrf: 0.443
- Bleu: 0.322
- Sacrebleu: 0.3
- Rouge1: 0.47
- Rouge2: 0.284
- Rougel: 0.46
- Rougelsum: 0.457
- Meteor: 0.535
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 1
- eval_batch_size: 1
- seed: 3407
- distributed_type: multi-GPU
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 304
- training_steps: 3040
Training results
Training Loss | Epoch | Step | Validation Loss | Accuracy | Chrf | Bleu | Sacrebleu | Rouge1 | Rouge2 | Rougel | Rougelsum | Meteor |
---|---|---|---|---|---|---|---|---|---|---|---|---|
6.6762 | 1.0 | 304 | 1.4150 | 0.78 | 0.177 | 0.096 | 0.1 | 0.304 | 0.171 | 0.297 | 0.299 | 0.381 |
0.0558 | 2.0 | 608 | 1.5036 | 0.775 | 0.085 | 0.0 | 0.0 | 0.091 | 0.03 | 0.091 | 0.091 | 0.176 |
0.0838 | 3.0 | 912 | 1.9965 | 0.766 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
0.0242 | 4.0 | 1216 | 1.8893 | 0.788 | 0.058 | 0.018 | 0.0 | 0.087 | 0.037 | 0.076 | 0.076 | 0.151 |
0.8769 | 5.0 | 1520 | 1.3394 | 0.773 | 0.15 | 0.08 | 0.1 | 0.29 | 0.179 | 0.285 | 0.287 | 0.366 |
0.0351 | 6.0 | 1824 | 1.1324 | 0.774 | 0.198 | 0.093 | 0.1 | 0.293 | 0.208 | 0.291 | 0.293 | 0.396 |
0.0262 | 7.0 | 2128 | 1.2192 | 0.776 | 0.237 | 0.105 | 0.1 | 0.336 | 0.222 | 0.333 | 0.332 | 0.425 |
0.0429 | 8.0 | 2432 | 0.8477 | 0.765 | 0.359 | 0.218 | 0.2 | 0.382 | 0.21 | 0.375 | 0.372 | 0.467 |
0.006 | 9.0 | 2736 | 0.6887 | 0.767 | 0.411 | 0.292 | 0.3 | 0.439 | 0.261 | 0.424 | 0.43 | 0.516 |
0.0203 | 10.0 | 3040 | 0.6367 | 0.767 | 0.443 | 0.322 | 0.3 | 0.47 | 0.284 | 0.46 | 0.457 | 0.535 |
Framework versions
- PEFT 0.7.1
- Transformers 4.37.0
- Pytorch 2.2.1+cu121
- Datasets 2.20.0
- Tokenizers 0.15.2
- Downloads last month
- 5
Model tree for vdavidr/llama-7b-finnish-instruct-v0.2_En__CMP_TR_size_304_epochs_10_2024-06-23_06-24-07_3558633
Base model
Finnish-NLP/llama-7b-finnish-instruct-v0.2