dit-base_tobacco-tiny_tobacco3482_hint
This model is a fine-tuned version of WinKawaks/vit-tiny-patch16-224 on the None dataset. It achieves the following results on the evaluation set:
- Loss: 2.1196
- Accuracy: 0.805
- Brier Loss: 0.3299
- Nll: 1.3687
- F1 Micro: 0.805
- F1 Macro: 0.7917
- Ece: 0.1606
- Aurc: 0.0741
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 100
Training results
Training Loss | Epoch | Step | Validation Loss | Accuracy | Brier Loss | Nll | F1 Micro | F1 Macro | Ece | Aurc |
---|---|---|---|---|---|---|---|---|---|---|
No log | 1.0 | 7 | 3.7183 | 0.225 | 0.9191 | 7.9210 | 0.225 | 0.1037 | 0.3197 | 0.7539 |
No log | 2.0 | 14 | 3.2602 | 0.365 | 0.7808 | 4.4817 | 0.3650 | 0.2558 | 0.2908 | 0.5116 |
No log | 3.0 | 21 | 2.8475 | 0.53 | 0.6305 | 2.7026 | 0.53 | 0.4263 | 0.2496 | 0.2812 |
No log | 4.0 | 28 | 2.6118 | 0.575 | 0.5382 | 1.8007 | 0.575 | 0.5478 | 0.2435 | 0.2022 |
No log | 5.0 | 35 | 2.4289 | 0.66 | 0.4404 | 1.6436 | 0.66 | 0.6317 | 0.2124 | 0.1236 |
No log | 6.0 | 42 | 2.4007 | 0.69 | 0.4123 | 1.9255 | 0.69 | 0.6896 | 0.2038 | 0.1195 |
No log | 7.0 | 49 | 2.3239 | 0.745 | 0.3775 | 1.9428 | 0.745 | 0.7365 | 0.1838 | 0.0957 |
No log | 8.0 | 56 | 2.3109 | 0.735 | 0.3660 | 1.6472 | 0.735 | 0.7370 | 0.1729 | 0.0842 |
No log | 9.0 | 63 | 2.3263 | 0.74 | 0.3867 | 1.5935 | 0.74 | 0.7422 | 0.1803 | 0.0944 |
No log | 10.0 | 70 | 2.3878 | 0.735 | 0.4037 | 1.6604 | 0.735 | 0.7039 | 0.2057 | 0.0909 |
No log | 11.0 | 77 | 2.5009 | 0.715 | 0.4220 | 1.8526 | 0.715 | 0.6918 | 0.2184 | 0.1017 |
No log | 12.0 | 84 | 2.5428 | 0.72 | 0.4169 | 1.8799 | 0.72 | 0.6894 | 0.2000 | 0.0979 |
No log | 13.0 | 91 | 2.5197 | 0.765 | 0.3884 | 1.8802 | 0.765 | 0.7268 | 0.1895 | 0.0933 |
No log | 14.0 | 98 | 2.4928 | 0.725 | 0.4147 | 1.6096 | 0.7250 | 0.6923 | 0.2078 | 0.0940 |
No log | 15.0 | 105 | 2.4172 | 0.765 | 0.3748 | 1.6754 | 0.765 | 0.7317 | 0.1874 | 0.0824 |
No log | 16.0 | 112 | 2.4116 | 0.76 | 0.3873 | 1.4155 | 0.76 | 0.7280 | 0.1972 | 0.0890 |
No log | 17.0 | 119 | 2.9196 | 0.73 | 0.4500 | 1.8493 | 0.7300 | 0.7220 | 0.2290 | 0.1657 |
No log | 18.0 | 126 | 2.5815 | 0.765 | 0.4175 | 2.0326 | 0.765 | 0.7467 | 0.2125 | 0.0990 |
No log | 19.0 | 133 | 2.7076 | 0.735 | 0.4475 | 1.8526 | 0.735 | 0.7152 | 0.2062 | 0.1049 |
No log | 20.0 | 140 | 2.6951 | 0.71 | 0.4709 | 2.0521 | 0.7100 | 0.7258 | 0.2380 | 0.1188 |
No log | 21.0 | 147 | 2.4037 | 0.765 | 0.4013 | 1.8740 | 0.765 | 0.7691 | 0.2009 | 0.0949 |
No log | 22.0 | 154 | 2.6585 | 0.73 | 0.4303 | 2.0299 | 0.7300 | 0.7170 | 0.2110 | 0.1004 |
No log | 23.0 | 161 | 2.4320 | 0.75 | 0.3950 | 1.8720 | 0.75 | 0.7340 | 0.2004 | 0.0895 |
No log | 24.0 | 168 | 2.4891 | 0.74 | 0.4199 | 1.6458 | 0.74 | 0.7458 | 0.2227 | 0.1128 |
No log | 25.0 | 175 | 2.6550 | 0.705 | 0.4833 | 1.7755 | 0.705 | 0.7042 | 0.2478 | 0.1145 |
No log | 26.0 | 182 | 2.3191 | 0.765 | 0.3965 | 1.6941 | 0.765 | 0.7415 | 0.1954 | 0.0731 |
No log | 27.0 | 189 | 2.3000 | 0.785 | 0.3763 | 1.3143 | 0.785 | 0.7416 | 0.2026 | 0.0712 |
No log | 28.0 | 196 | 2.2047 | 0.78 | 0.3409 | 1.4818 | 0.78 | 0.7694 | 0.1759 | 0.0688 |
No log | 29.0 | 203 | 2.3587 | 0.77 | 0.3781 | 1.6779 | 0.7700 | 0.7571 | 0.1937 | 0.0766 |
No log | 30.0 | 210 | 2.5027 | 0.75 | 0.4400 | 1.8454 | 0.75 | 0.7338 | 0.2232 | 0.0817 |
No log | 31.0 | 217 | 2.4092 | 0.77 | 0.3899 | 1.5010 | 0.7700 | 0.7498 | 0.1987 | 0.0710 |
No log | 32.0 | 224 | 2.7655 | 0.74 | 0.4520 | 2.0720 | 0.74 | 0.7177 | 0.2266 | 0.0855 |
No log | 33.0 | 231 | 2.3814 | 0.76 | 0.3979 | 1.4053 | 0.76 | 0.7352 | 0.1982 | 0.0754 |
No log | 34.0 | 238 | 2.3946 | 0.775 | 0.3790 | 1.6969 | 0.775 | 0.7387 | 0.1908 | 0.0876 |
No log | 35.0 | 245 | 2.5158 | 0.775 | 0.4064 | 1.4329 | 0.775 | 0.7428 | 0.2068 | 0.0929 |
No log | 36.0 | 252 | 2.4920 | 0.75 | 0.4281 | 1.5724 | 0.75 | 0.7470 | 0.2161 | 0.1028 |
No log | 37.0 | 259 | 2.4541 | 0.765 | 0.3842 | 1.5272 | 0.765 | 0.7163 | 0.1918 | 0.0853 |
No log | 38.0 | 266 | 2.3785 | 0.82 | 0.3319 | 1.6360 | 0.82 | 0.7944 | 0.1732 | 0.0811 |
No log | 39.0 | 273 | 2.3721 | 0.77 | 0.3831 | 1.2724 | 0.7700 | 0.7649 | 0.1880 | 0.0818 |
No log | 40.0 | 280 | 2.5684 | 0.795 | 0.3494 | 1.7291 | 0.795 | 0.7724 | 0.1771 | 0.1126 |
No log | 41.0 | 287 | 2.4835 | 0.78 | 0.3799 | 1.7700 | 0.78 | 0.7594 | 0.2024 | 0.0821 |
No log | 42.0 | 294 | 2.4690 | 0.795 | 0.3678 | 1.6379 | 0.795 | 0.7804 | 0.1836 | 0.0840 |
No log | 43.0 | 301 | 2.3069 | 0.77 | 0.3809 | 1.4970 | 0.7700 | 0.7621 | 0.1807 | 0.0784 |
No log | 44.0 | 308 | 2.4424 | 0.795 | 0.3558 | 1.7162 | 0.795 | 0.7857 | 0.1879 | 0.0667 |
No log | 45.0 | 315 | 2.0018 | 0.86 | 0.2387 | 1.4924 | 0.8600 | 0.8558 | 0.1180 | 0.0526 |
No log | 46.0 | 322 | 2.4174 | 0.785 | 0.3593 | 1.7688 | 0.785 | 0.7662 | 0.1859 | 0.0773 |
No log | 47.0 | 329 | 2.1816 | 0.84 | 0.2862 | 1.5093 | 0.8400 | 0.8230 | 0.1459 | 0.0679 |
No log | 48.0 | 336 | 2.3006 | 0.78 | 0.3601 | 1.7675 | 0.78 | 0.7496 | 0.1847 | 0.0752 |
No log | 49.0 | 343 | 2.2952 | 0.81 | 0.3231 | 1.4851 | 0.81 | 0.7865 | 0.1691 | 0.0656 |
No log | 50.0 | 350 | 2.1346 | 0.8 | 0.3132 | 1.4625 | 0.8000 | 0.7937 | 0.1660 | 0.0789 |
No log | 51.0 | 357 | 2.2935 | 0.81 | 0.3383 | 1.5764 | 0.81 | 0.7914 | 0.1770 | 0.0717 |
No log | 52.0 | 364 | 2.1792 | 0.825 | 0.3116 | 1.5652 | 0.825 | 0.8163 | 0.1454 | 0.0654 |
No log | 53.0 | 371 | 2.1231 | 0.81 | 0.3066 | 1.3012 | 0.81 | 0.8062 | 0.1552 | 0.0604 |
No log | 54.0 | 378 | 1.9712 | 0.825 | 0.2854 | 1.2891 | 0.825 | 0.8137 | 0.1430 | 0.0521 |
No log | 55.0 | 385 | 2.0133 | 0.825 | 0.2839 | 1.3994 | 0.825 | 0.8086 | 0.1433 | 0.0560 |
No log | 56.0 | 392 | 1.9978 | 0.835 | 0.2800 | 1.4348 | 0.835 | 0.8232 | 0.1415 | 0.0573 |
No log | 57.0 | 399 | 1.9847 | 0.83 | 0.2825 | 1.3907 | 0.83 | 0.8153 | 0.1421 | 0.0560 |
No log | 58.0 | 406 | 1.9892 | 0.83 | 0.2832 | 1.4502 | 0.83 | 0.8153 | 0.1503 | 0.0566 |
No log | 59.0 | 413 | 1.9848 | 0.83 | 0.2851 | 1.4506 | 0.83 | 0.8156 | 0.1462 | 0.0560 |
No log | 60.0 | 420 | 1.9871 | 0.835 | 0.2910 | 1.4527 | 0.835 | 0.8191 | 0.1608 | 0.0566 |
No log | 61.0 | 427 | 1.9914 | 0.825 | 0.2932 | 1.4490 | 0.825 | 0.8095 | 0.1464 | 0.0587 |
No log | 62.0 | 434 | 1.9908 | 0.825 | 0.2958 | 1.4459 | 0.825 | 0.8095 | 0.1493 | 0.0597 |
No log | 63.0 | 441 | 1.9954 | 0.825 | 0.3012 | 1.4480 | 0.825 | 0.8095 | 0.1469 | 0.0606 |
No log | 64.0 | 448 | 2.0111 | 0.82 | 0.3050 | 1.4487 | 0.82 | 0.8026 | 0.1507 | 0.0619 |
No log | 65.0 | 455 | 2.0212 | 0.82 | 0.3046 | 1.4469 | 0.82 | 0.8026 | 0.1604 | 0.0634 |
No log | 66.0 | 462 | 2.0170 | 0.82 | 0.3059 | 1.4443 | 0.82 | 0.8040 | 0.1539 | 0.0639 |
No log | 67.0 | 469 | 2.0170 | 0.815 | 0.3056 | 1.4496 | 0.815 | 0.8019 | 0.1534 | 0.0643 |
No log | 68.0 | 476 | 2.0316 | 0.82 | 0.3115 | 1.4522 | 0.82 | 0.8026 | 0.1606 | 0.0645 |
No log | 69.0 | 483 | 2.0335 | 0.805 | 0.3132 | 1.3831 | 0.805 | 0.7855 | 0.1607 | 0.0654 |
No log | 70.0 | 490 | 2.0362 | 0.815 | 0.3106 | 1.3834 | 0.815 | 0.7989 | 0.1614 | 0.0655 |
No log | 71.0 | 497 | 2.0318 | 0.815 | 0.3105 | 1.3893 | 0.815 | 0.7947 | 0.1541 | 0.0662 |
1.3661 | 72.0 | 504 | 2.0434 | 0.815 | 0.3135 | 1.4473 | 0.815 | 0.7955 | 0.1579 | 0.0653 |
1.3661 | 73.0 | 511 | 2.0517 | 0.81 | 0.3139 | 1.3838 | 0.81 | 0.7917 | 0.1564 | 0.0680 |
1.3661 | 74.0 | 518 | 2.0594 | 0.82 | 0.3162 | 1.3783 | 0.82 | 0.7975 | 0.1626 | 0.0681 |
1.3661 | 75.0 | 525 | 2.0628 | 0.815 | 0.3210 | 1.3752 | 0.815 | 0.7944 | 0.1598 | 0.0706 |
1.3661 | 76.0 | 532 | 2.0605 | 0.81 | 0.3158 | 1.3711 | 0.81 | 0.7886 | 0.1639 | 0.0684 |
1.3661 | 77.0 | 539 | 2.0718 | 0.815 | 0.3187 | 1.3860 | 0.815 | 0.7944 | 0.1710 | 0.0705 |
1.3661 | 78.0 | 546 | 2.0749 | 0.815 | 0.3168 | 1.3658 | 0.815 | 0.7958 | 0.1569 | 0.0713 |
1.3661 | 79.0 | 553 | 2.0796 | 0.83 | 0.3188 | 1.3016 | 0.83 | 0.8147 | 0.1646 | 0.0722 |
1.3661 | 80.0 | 560 | 2.0746 | 0.81 | 0.3210 | 1.3758 | 0.81 | 0.7916 | 0.1580 | 0.0729 |
1.3661 | 81.0 | 567 | 2.0819 | 0.815 | 0.3194 | 1.3686 | 0.815 | 0.7913 | 0.1576 | 0.0722 |
1.3661 | 82.0 | 574 | 2.0866 | 0.825 | 0.3182 | 1.3627 | 0.825 | 0.8085 | 0.1602 | 0.0718 |
1.3661 | 83.0 | 581 | 2.0942 | 0.815 | 0.3246 | 1.3249 | 0.815 | 0.8008 | 0.1591 | 0.0727 |
1.3661 | 84.0 | 588 | 2.0938 | 0.815 | 0.3246 | 1.3680 | 0.815 | 0.7984 | 0.1848 | 0.0727 |
1.3661 | 85.0 | 595 | 2.0912 | 0.82 | 0.3222 | 1.3662 | 0.82 | 0.8012 | 0.1594 | 0.0702 |
1.3661 | 86.0 | 602 | 2.0941 | 0.82 | 0.3234 | 1.3764 | 0.82 | 0.8012 | 0.1576 | 0.0738 |
1.3661 | 87.0 | 609 | 2.1037 | 0.8 | 0.3304 | 1.3821 | 0.8000 | 0.7821 | 0.1599 | 0.0740 |
1.3661 | 88.0 | 616 | 2.1098 | 0.805 | 0.3288 | 1.3587 | 0.805 | 0.7932 | 0.1678 | 0.0718 |
1.3661 | 89.0 | 623 | 2.1119 | 0.81 | 0.3276 | 1.3636 | 0.81 | 0.7945 | 0.1622 | 0.0728 |
1.3661 | 90.0 | 630 | 2.1078 | 0.805 | 0.3279 | 1.3641 | 0.805 | 0.7914 | 0.1734 | 0.0737 |
1.3661 | 91.0 | 637 | 2.1110 | 0.8 | 0.3296 | 1.3686 | 0.8000 | 0.7879 | 0.1636 | 0.0744 |
1.3661 | 92.0 | 644 | 2.1150 | 0.8 | 0.3317 | 1.3685 | 0.8000 | 0.7879 | 0.1730 | 0.0742 |
1.3661 | 93.0 | 651 | 2.1146 | 0.8 | 0.3303 | 1.3693 | 0.8000 | 0.7881 | 0.1631 | 0.0742 |
1.3661 | 94.0 | 658 | 2.1153 | 0.805 | 0.3292 | 1.3657 | 0.805 | 0.7917 | 0.1676 | 0.0734 |
1.3661 | 95.0 | 665 | 2.1188 | 0.805 | 0.3298 | 1.3683 | 0.805 | 0.7917 | 0.1690 | 0.0735 |
1.3661 | 96.0 | 672 | 2.1183 | 0.805 | 0.3291 | 1.3691 | 0.805 | 0.7914 | 0.1687 | 0.0742 |
1.3661 | 97.0 | 679 | 2.1155 | 0.81 | 0.3271 | 1.3664 | 0.81 | 0.7942 | 0.1599 | 0.0743 |
1.3661 | 98.0 | 686 | 2.1183 | 0.805 | 0.3285 | 1.3673 | 0.805 | 0.7914 | 0.1638 | 0.0740 |
1.3661 | 99.0 | 693 | 2.1179 | 0.805 | 0.3297 | 1.3686 | 0.805 | 0.7917 | 0.1613 | 0.0741 |
1.3661 | 100.0 | 700 | 2.1196 | 0.805 | 0.3299 | 1.3687 | 0.805 | 0.7917 | 0.1606 | 0.0741 |
Framework versions
- Transformers 4.36.0.dev0
- Pytorch 2.2.0.dev20231112+cu118
- Datasets 2.14.5
- Tokenizers 0.14.1
- Downloads last month
- 6
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for jordyvl/dit-base_tobacco-tiny_tobacco3482_hint
Base model
WinKawaks/vit-tiny-patch16-224