lewtun HF staff lewtun HF staff commited on
Commit
46578b5
0 Parent(s):

Duplicate from autoevaluate/entity-extraction

Browse files

Co-authored-by: Lewis Tunstall <[email protected]>

.gitattributes ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ftz filter=lfs diff=lfs merge=lfs -text
6
+ *.gz filter=lfs diff=lfs merge=lfs -text
7
+ *.h5 filter=lfs diff=lfs merge=lfs -text
8
+ *.joblib filter=lfs diff=lfs merge=lfs -text
9
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
+ *.model filter=lfs diff=lfs merge=lfs -text
11
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
12
+ *.onnx filter=lfs diff=lfs merge=lfs -text
13
+ *.ot filter=lfs diff=lfs merge=lfs -text
14
+ *.parquet filter=lfs diff=lfs merge=lfs -text
15
+ *.pb filter=lfs diff=lfs merge=lfs -text
16
+ *.pt filter=lfs diff=lfs merge=lfs -text
17
+ *.pth filter=lfs diff=lfs merge=lfs -text
18
+ *.rar filter=lfs diff=lfs merge=lfs -text
19
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
20
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
21
+ *.tflite filter=lfs diff=lfs merge=lfs -text
22
+ *.tgz filter=lfs diff=lfs merge=lfs -text
23
+ *.wasm filter=lfs diff=lfs merge=lfs -text
24
+ *.xz filter=lfs diff=lfs merge=lfs -text
25
+ *.zip filter=lfs diff=lfs merge=lfs -text
26
+ *.zstandard filter=lfs diff=lfs merge=lfs -text
27
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
README.md ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ datasets:
6
+ - conll2003
7
+ - autoevaluate/conll2003-sample
8
+ metrics:
9
+ - precision
10
+ - recall
11
+ - f1
12
+ - accuracy
13
+ model-index:
14
+ - name: entity-extraction
15
+ results:
16
+ - task:
17
+ name: Token Classification
18
+ type: token-classification
19
+ dataset:
20
+ name: conll2003
21
+ type: conll2003
22
+ args: conll2003
23
+ metrics:
24
+ - name: Precision
25
+ type: precision
26
+ value: 0.8862817854414493
27
+ - name: Recall
28
+ type: recall
29
+ value: 0.9084908826490659
30
+ - name: F1
31
+ type: f1
32
+ value: 0.8972489227709645
33
+ - name: Accuracy
34
+ type: accuracy
35
+ value: 0.9774889986814304
36
+ duplicated_from: autoevaluate/entity-extraction
37
+ ---
38
+
39
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
40
+ should probably proofread and complete it, then remove this comment. -->
41
+
42
+ # entity-extraction
43
+
44
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the conll2003 dataset.
45
+ It achieves the following results on the evaluation set:
46
+ - Loss: 0.0808
47
+ - Precision: 0.8863
48
+ - Recall: 0.9085
49
+ - F1: 0.8972
50
+ - Accuracy: 0.9775
51
+
52
+ ## Model description
53
+
54
+ More information needed
55
+
56
+ ## Intended uses & limitations
57
+
58
+ More information needed
59
+
60
+ ## Training and evaluation data
61
+
62
+ More information needed
63
+
64
+ ## Training procedure
65
+
66
+ ### Training hyperparameters
67
+
68
+ The following hyperparameters were used during training:
69
+ - learning_rate: 2e-05
70
+ - train_batch_size: 16
71
+ - eval_batch_size: 16
72
+ - seed: 42
73
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
74
+ - lr_scheduler_type: linear
75
+ - num_epochs: 1
76
+
77
+ ### Training results
78
+
79
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
80
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
81
+ | 0.2552 | 1.0 | 878 | 0.0808 | 0.8863 | 0.9085 | 0.8972 | 0.9775 |
82
+
83
+
84
+ ### Framework versions
85
+
86
+ - Transformers 4.19.2
87
+ - Pytorch 1.11.0+cu113
88
+ - Datasets 2.2.2
89
+ - Tokenizers 0.12.1
config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForTokenClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "O",
13
+ "1": "B-PER",
14
+ "2": "I-PER",
15
+ "3": "B-ORG",
16
+ "4": "I-ORG",
17
+ "5": "B-LOC",
18
+ "6": "I-LOC",
19
+ "7": "B-MISC",
20
+ "8": "I-MISC"
21
+ },
22
+ "initializer_range": 0.02,
23
+ "label2id": {
24
+ "B-LOC": 5,
25
+ "B-MISC": 7,
26
+ "B-ORG": 3,
27
+ "B-PER": 1,
28
+ "I-LOC": 6,
29
+ "I-MISC": 8,
30
+ "I-ORG": 4,
31
+ "I-PER": 2,
32
+ "O": 0
33
+ },
34
+ "max_position_embeddings": 512,
35
+ "model_type": "distilbert",
36
+ "n_heads": 12,
37
+ "n_layers": 6,
38
+ "pad_token_id": 0,
39
+ "qa_dropout": 0.1,
40
+ "seq_classif_dropout": 0.2,
41
+ "sinusoidal_pos_embds": false,
42
+ "tie_weights_": true,
43
+ "torch_dtype": "float32",
44
+ "transformers_version": "4.19.2",
45
+ "vocab_size": 30522
46
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48ebb7a89b19b62d0edab1f2814f8202987030f65333d0fea670a8dc00a18e0e
3
+ size 265512821
runs/May25_11-08-01_925bb1d0e2e5/1653476912.8511903/events.out.tfevents.1653476912.925bb1d0e2e5.75.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13cae1981a1b5298628095c63c9d2383bc661bca0692db261e37d77368151771
3
+ size 5186
runs/May25_11-08-01_925bb1d0e2e5/events.out.tfevents.1653476912.925bb1d0e2e5.75.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dbc9e4d4efc519d9807a41e01db74502eee6e2e377b6141f05aa01cd763e3e4
3
+ size 4747
runs/May25_11-08-01_925bb1d0e2e5/events.out.tfevents.1653477011.925bb1d0e2e5.75.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9481e744833a66ea8e764dfa1f1472918511f81c13de676cf6e47d06ff1bfdb9
3
+ size 512
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-uncased", "tokenizer_class": "DistilBertTokenizer"}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d1c6ac821fc97102a539ed0dc7e508dd3cb472466b36d81affda3e3476950e3
3
+ size 3247
vocab.txt ADDED
The diff for this file is too large to render. See raw diff