EduardoPacheco commited on
Commit
aec47c8
1 Parent(s): 133f765

End of training

Browse files
README.md ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: IDEA-Research/grounding-dino-tiny
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: grounding-dino-tiny-aquarium-fine-tune
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # grounding-dino-tiny-aquarium-fine-tune
15
+
16
+ This model is a fine-tuned version of [IDEA-Research/grounding-dino-tiny](https://huggingface.co/IDEA-Research/grounding-dino-tiny) on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 18.3797
19
+ - Map: 0.1008
20
+ - Map 50: 0.1716
21
+ - Map 75: 0.0971
22
+ - Map Small: -1.0
23
+ - Map Medium: 0.1054
24
+ - Map Large: 0.1024
25
+ - Mar 1: 0.1063
26
+ - Mar 10: 0.1758
27
+ - Mar 100: 0.2843
28
+ - Mar Small: -1.0
29
+ - Mar Medium: 0.1848
30
+ - Mar Large: 0.2881
31
+ - Map Fish: 0.0827
32
+ - Mar 100 Fish: 0.411
33
+ - Map Jellyfish: 0.1291
34
+ - Mar 100 Jellyfish: 0.4026
35
+ - Map Penguins: 0.0963
36
+ - Mar 100 Penguins: 0.426
37
+ - Map Sharks: 0.0336
38
+ - Mar 100 Sharks: 0.0561
39
+ - Map Puffins: 0.04
40
+ - Mar 100 Puffins: 0.1
41
+ - Map Stingrays: 0.2407
42
+ - Mar 100 Stingrays: 0.4042
43
+ - Map Starfish: 0.0829
44
+ - Mar 100 Starfish: 0.1906
45
+
46
+ ## Model description
47
+
48
+ More information needed
49
+
50
+ ## Intended uses & limitations
51
+
52
+ More information needed
53
+
54
+ ## Training and evaluation data
55
+
56
+ More information needed
57
+
58
+ ## Training procedure
59
+
60
+ ### Training hyperparameters
61
+
62
+ The following hyperparameters were used during training:
63
+ - learning_rate: 5e-05
64
+ - train_batch_size: 4
65
+ - eval_batch_size: 4
66
+ - seed: 42
67
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
68
+ - lr_scheduler_type: linear
69
+ - lr_scheduler_warmup_steps: 300
70
+ - num_epochs: 10
71
+
72
+ ### Training results
73
+
74
+ | Training Loss | Epoch | Step | Validation Loss | Map | Map 50 | Map 75 | Map Small | Map Medium | Map Large | Mar 1 | Mar 10 | Mar 100 | Mar Small | Mar Medium | Mar Large | Map Fish | Mar 100 Fish | Map Jellyfish | Mar 100 Jellyfish | Map Penguins | Mar 100 Penguins | Map Sharks | Mar 100 Sharks | Map Puffins | Mar 100 Puffins | Map Stingrays | Mar 100 Stingrays | Map Starfish | Mar 100 Starfish |
75
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:----------:|:---------:|:------:|:------:|:-------:|:---------:|:----------:|:---------:|:--------:|:------------:|:-------------:|:-----------------:|:------------:|:----------------:|:----------:|:--------------:|:-----------:|:---------------:|:-------------:|:-----------------:|:------------:|:----------------:|
76
+ | 47786.9107 | 1.0 | 112 | 47816.1406 | 0.0006 | 0.0019 | 0.0003 | -1.0 | 0.0012 | 0.0008 | 0.0005 | 0.0071 | 0.0116 | -1.0 | 0.0455 | 0.0107 | 0.0028 | 0.0491 | 0.0017 | 0.0323 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
77
+ | 36235.5402 | 2.0 | 224 | 26297.4707 | 0.0036 | 0.0117 | 0.0021 | -1.0 | 0.0002 | 0.0039 | 0.0014 | 0.0092 | 0.0326 | -1.0 | 0.0119 | 0.0341 | 0.0182 | 0.1514 | 0.0069 | 0.0768 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
78
+ | 13131.6674 | 3.0 | 336 | 4944.3315 | 0.0022 | 0.0039 | 0.0016 | -1.0 | 0.0005 | 0.0022 | 0.0017 | 0.005 | 0.0085 | -1.0 | 0.0107 | 0.0087 | 0.0139 | 0.0546 | 0.0013 | 0.0052 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
79
+ | 1436.2895 | 4.0 | 448 | 160.7001 | 0.0065 | 0.0175 | 0.004 | -1.0 | 0.0074 | 0.0068 | 0.0044 | 0.0129 | 0.0322 | -1.0 | 0.0119 | 0.0336 | 0.0202 | 0.1621 | 0.0254 | 0.0632 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
80
+ | 48.6727 | 5.0 | 560 | 19.5311 | 0.0183 | 0.0474 | 0.0105 | -1.0 | 0.0086 | 0.0197 | 0.0081 | 0.0338 | 0.0768 | -1.0 | 0.0399 | 0.0798 | 0.0332 | 0.2379 | 0.0663 | 0.211 | 0.0288 | 0.0885 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
81
+ | 19.4237 | 6.0 | 672 | 18.9543 | 0.0649 | 0.12 | 0.0633 | -1.0 | 0.1113 | 0.0658 | 0.0817 | 0.1352 | 0.2257 | -1.0 | 0.1698 | 0.2284 | 0.0611 | 0.3724 | 0.0829 | 0.3542 | 0.0444 | 0.3154 | 0.0 | 0.0 | 0.0113 | 0.0574 | 0.202 | 0.3583 | 0.0529 | 0.1219 |
82
+ | 18.6905 | 7.0 | 784 | 18.5298 | 0.0855 | 0.1458 | 0.0805 | -1.0 | 0.0869 | 0.087 | 0.109 | 0.1788 | 0.2767 | -1.0 | 0.165 | 0.2801 | 0.0615 | 0.3822 | 0.1007 | 0.3826 | 0.0772 | 0.3538 | 0.0107 | 0.0258 | 0.02 | 0.0852 | 0.2416 | 0.4792 | 0.0871 | 0.2281 |
83
+ | 18.3457 | 8.0 | 896 | 18.6190 | 0.0808 | 0.1485 | 0.0724 | -1.0 | 0.1232 | 0.0818 | 0.1071 | 0.1821 | 0.2959 | -1.0 | 0.201 | 0.2993 | 0.0644 | 0.4116 | 0.1096 | 0.4142 | 0.0761 | 0.4067 | 0.0208 | 0.0833 | 0.0386 | 0.1574 | 0.2184 | 0.4542 | 0.0379 | 0.1437 |
84
+ | 18.2089 | 9.0 | 1008 | 18.3819 | 0.0997 | 0.1662 | 0.0979 | -1.0 | 0.0888 | 0.1015 | 0.098 | 0.1657 | 0.2626 | -1.0 | 0.1938 | 0.2653 | 0.0726 | 0.3886 | 0.1283 | 0.3903 | 0.0993 | 0.3904 | 0.0207 | 0.05 | 0.0236 | 0.0593 | 0.2596 | 0.425 | 0.094 | 0.1344 |
85
+ | 17.7603 | 10.0 | 1120 | 18.3797 | 0.1008 | 0.1716 | 0.0971 | -1.0 | 0.1054 | 0.1024 | 0.1063 | 0.1758 | 0.2843 | -1.0 | 0.1848 | 0.2881 | 0.0827 | 0.411 | 0.1291 | 0.4026 | 0.0963 | 0.426 | 0.0336 | 0.0561 | 0.04 | 0.1 | 0.2407 | 0.4042 | 0.0829 | 0.1906 |
86
+
87
+
88
+ ### Framework versions
89
+
90
+ - Transformers 4.45.0.dev0
91
+ - Pytorch 2.1.2
92
+ - Datasets 2.20.0
93
+ - Tokenizers 0.19.1
config.json ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "IDEA-Research/grounding-dino-tiny",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "relu",
5
+ "architectures": [
6
+ "GroundingDinoForObjectDetection"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "auxiliary_loss": true,
10
+ "backbone": null,
11
+ "backbone_config": {
12
+ "model_type": "swin",
13
+ "out_features": [
14
+ "stage2",
15
+ "stage3",
16
+ "stage4"
17
+ ],
18
+ "out_indices": [
19
+ 2,
20
+ 3,
21
+ 4
22
+ ]
23
+ },
24
+ "backbone_kwargs": null,
25
+ "bbox_cost": 5.0,
26
+ "bbox_loss_coefficient": 5.0,
27
+ "class_cost": 1.0,
28
+ "class_loss_coefficient": 2.0,
29
+ "class_loss_reduction": "sum",
30
+ "d_model": 256,
31
+ "decoder_attention_heads": 8,
32
+ "decoder_bbox_embed_share": true,
33
+ "decoder_ffn_dim": 2048,
34
+ "decoder_layers": 6,
35
+ "decoder_n_points": 4,
36
+ "disable_custom_kernels": false,
37
+ "dropout": 0.1,
38
+ "embedding_init_target": true,
39
+ "encoder_attention_heads": 8,
40
+ "encoder_ffn_dim": 2048,
41
+ "encoder_layers": 6,
42
+ "encoder_n_points": 4,
43
+ "focal_alpha": 0.25,
44
+ "fusion_dropout": 0.0,
45
+ "fusion_droppath": 0.1,
46
+ "giou_cost": 2.0,
47
+ "giou_loss_coefficient": 2.0,
48
+ "id2label": {
49
+ "0": "fish",
50
+ "1": "jellyfish",
51
+ "2": "penguins",
52
+ "3": "sharks",
53
+ "4": "puffins",
54
+ "5": "stingrays",
55
+ "6": "starfish"
56
+ },
57
+ "init_std": 0.02,
58
+ "is_encoder_decoder": true,
59
+ "label2id": {
60
+ "fish": 0,
61
+ "jellyfish": 1,
62
+ "penguins": 2,
63
+ "puffins": 4,
64
+ "sharks": 3,
65
+ "starfish": 6,
66
+ "stingrays": 5
67
+ },
68
+ "layer_norm_eps": 1e-05,
69
+ "max_text_len": 256,
70
+ "model_type": "grounding-dino",
71
+ "num_feature_levels": 4,
72
+ "num_queries": 900,
73
+ "position_embedding_type": "sine",
74
+ "positional_embedding_temperature": 20,
75
+ "query_dim": 4,
76
+ "text_config": {
77
+ "model_type": "bert"
78
+ },
79
+ "text_enhancer_dropout": 0.0,
80
+ "torch_dtype": "float32",
81
+ "transformers_version": "4.45.0.dev0",
82
+ "two_stage": true,
83
+ "two_stage_bbox_embed_share": false,
84
+ "use_pretrained_backbone": false,
85
+ "use_timm_backbone": false
86
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2717c605d31b30024fa748f716766b72b5bbbcab312367aadf7f688128f6dad
3
+ size 689359096
preprocessor_config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_annotations": true,
3
+ "do_normalize": true,
4
+ "do_pad": true,
5
+ "do_rescale": true,
6
+ "do_resize": true,
7
+ "format": "coco_detection",
8
+ "image_mean": [
9
+ 0.485,
10
+ 0.456,
11
+ 0.406
12
+ ],
13
+ "image_processor_type": "GroundingDinoImageProcessor",
14
+ "image_std": [
15
+ 0.229,
16
+ 0.224,
17
+ 0.225
18
+ ],
19
+ "pad_size": null,
20
+ "processor_class": "GroundingDinoProcessor",
21
+ "resample": 2,
22
+ "rescale_factor": 0.00392156862745098,
23
+ "size": {
24
+ "longest_edge": 1333,
25
+ "shortest_edge": 800
26
+ }
27
+ }
runs/Aug13_21-27-20_a1a3b3b6dcfb/events.out.tfevents.1723584556.a1a3b3b6dcfb.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f1a6bd2229bcb02528a25b2fb453f45e5c58b662ac365990b39223aa128bec7
3
+ size 6219
runs/Aug13_21-35-54_a1a3b3b6dcfb/events.out.tfevents.1723584958.a1a3b3b6dcfb.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25c1eaf8f365e7a13855dbd05ce06fa1012769f3bb6ac60e9b1e021709647f8a
3
+ size 23318
runs/Aug13_23-11-55_a1a3b3b6dcfb/events.out.tfevents.1723590732.a1a3b3b6dcfb.34.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd34ca0f5318290b4e4959d1bb15c32fb46aa39af61137b0ae89c09c1bd9fdb7
3
+ size 6222
runs/Aug13_23-15-44_a1a3b3b6dcfb/events.out.tfevents.1723590960.a1a3b3b6dcfb.4406.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0371ed8c08ae4a7fdb6c697a9879fcf92a695c8200af9c01e81911fd0ac291b4
3
+ size 25548
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "processor_class": "GroundingDinoProcessor",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:429d5616e0ff7181687c8ca6643d3332ac75c75b1c9ab0a64f99b1bf8a9e7ce3
3
+ size 5240
vocab.txt ADDED
The diff for this file is too large to render. See raw diff