heBERT
Browse files- config.json +41 -0
- log_history.json +173 -0
- pytorch_model.bin +3 -0
- training_args.bin +3 -0
- vocab.txt +0 -0
config.json
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"BertForTokenClassification"
|
4 |
+
],
|
5 |
+
"attention_probs_dropout_prob": 0.1,
|
6 |
+
"gradient_checkpointing": false,
|
7 |
+
"hidden_act": "gelu",
|
8 |
+
"hidden_dropout_prob": 0.1,
|
9 |
+
"hidden_size": 768,
|
10 |
+
"id2label": {
|
11 |
+
"0": "B_TIME",
|
12 |
+
"1": "B_PERS",
|
13 |
+
"2": "B_MONEY",
|
14 |
+
"3": "B_ORG",
|
15 |
+
"4": "B_PERCENT",
|
16 |
+
"5": "O",
|
17 |
+
"6": "B_LOC",
|
18 |
+
"7": "B_DATE"
|
19 |
+
},
|
20 |
+
"initializer_range": 0.02,
|
21 |
+
"intermediate_size": 3072,
|
22 |
+
"label2id": {
|
23 |
+
"B_DATE": 7,
|
24 |
+
"B_LOC": 6,
|
25 |
+
"B_MONEY": 2,
|
26 |
+
"B_ORG": 3,
|
27 |
+
"B_PERCENT": 4,
|
28 |
+
"B_PERS": 1,
|
29 |
+
"B_TIME": 0,
|
30 |
+
"O": 5
|
31 |
+
},
|
32 |
+
"layer_norm_eps": 1e-12,
|
33 |
+
"max_position_embeddings": 512,
|
34 |
+
"model_type": "bert",
|
35 |
+
"num_attention_heads": 12,
|
36 |
+
"num_hidden_layers": 12,
|
37 |
+
"pad_token_id": 0,
|
38 |
+
"total_flos": 6997313242916978688,
|
39 |
+
"type_vocab_size": 2,
|
40 |
+
"vocab_size": 30522
|
41 |
+
}
|
log_history.json
ADDED
@@ -0,0 +1,173 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"loss": 2.1632102966308593,
|
4 |
+
"learning_rate": 1.0000000000000002e-06,
|
5 |
+
"epoch": 0.2127659574468085,
|
6 |
+
"total_flos": 4181675520000,
|
7 |
+
"step": 10
|
8 |
+
},
|
9 |
+
{
|
10 |
+
"loss": 1.9698543548583984,
|
11 |
+
"learning_rate": 2.0000000000000003e-06,
|
12 |
+
"epoch": 0.425531914893617,
|
13 |
+
"total_flos": 8363351040000,
|
14 |
+
"step": 20
|
15 |
+
},
|
16 |
+
{
|
17 |
+
"loss": 1.5943157196044921,
|
18 |
+
"learning_rate": 3e-06,
|
19 |
+
"epoch": 0.6382978723404256,
|
20 |
+
"total_flos": 12545026560000,
|
21 |
+
"step": 30
|
22 |
+
},
|
23 |
+
{
|
24 |
+
"loss": 1.1324241638183594,
|
25 |
+
"learning_rate": 4.000000000000001e-06,
|
26 |
+
"epoch": 0.851063829787234,
|
27 |
+
"total_flos": 16726702080000,
|
28 |
+
"step": 40
|
29 |
+
},
|
30 |
+
{
|
31 |
+
"loss": 0.7799140930175781,
|
32 |
+
"learning_rate": 5e-06,
|
33 |
+
"epoch": 1.0638297872340425,
|
34 |
+
"total_flos": 20620887408000,
|
35 |
+
"step": 50
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"loss": 0.5984169006347656,
|
39 |
+
"learning_rate": 6e-06,
|
40 |
+
"epoch": 1.2765957446808511,
|
41 |
+
"total_flos": 24802562928000,
|
42 |
+
"step": 60
|
43 |
+
},
|
44 |
+
{
|
45 |
+
"loss": 0.4769012451171875,
|
46 |
+
"learning_rate": 7.000000000000001e-06,
|
47 |
+
"epoch": 1.4893617021276595,
|
48 |
+
"total_flos": 28984238448000,
|
49 |
+
"step": 70
|
50 |
+
},
|
51 |
+
{
|
52 |
+
"loss": 0.3644599914550781,
|
53 |
+
"learning_rate": 8.000000000000001e-06,
|
54 |
+
"epoch": 1.702127659574468,
|
55 |
+
"total_flos": 33165913968000,
|
56 |
+
"step": 80
|
57 |
+
},
|
58 |
+
{
|
59 |
+
"loss": 0.29639739990234376,
|
60 |
+
"learning_rate": 9e-06,
|
61 |
+
"epoch": 1.9148936170212765,
|
62 |
+
"total_flos": 37347589488000,
|
63 |
+
"step": 90
|
64 |
+
},
|
65 |
+
{
|
66 |
+
"loss": 0.26825637817382814,
|
67 |
+
"learning_rate": 1e-05,
|
68 |
+
"epoch": 2.127659574468085,
|
69 |
+
"total_flos": 41241774816000,
|
70 |
+
"step": 100
|
71 |
+
},
|
72 |
+
{
|
73 |
+
"loss": 0.19923171997070313,
|
74 |
+
"learning_rate": 1.1000000000000001e-05,
|
75 |
+
"epoch": 2.3404255319148937,
|
76 |
+
"total_flos": 45423450336000,
|
77 |
+
"step": 110
|
78 |
+
},
|
79 |
+
{
|
80 |
+
"loss": 0.19310073852539061,
|
81 |
+
"learning_rate": 1.2e-05,
|
82 |
+
"epoch": 2.5531914893617023,
|
83 |
+
"total_flos": 49605125856000,
|
84 |
+
"step": 120
|
85 |
+
},
|
86 |
+
{
|
87 |
+
"loss": 0.15804977416992189,
|
88 |
+
"learning_rate": 1.3000000000000001e-05,
|
89 |
+
"epoch": 2.7659574468085104,
|
90 |
+
"total_flos": 53786801376000,
|
91 |
+
"step": 130
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"loss": 0.17403030395507812,
|
95 |
+
"learning_rate": 1.4000000000000001e-05,
|
96 |
+
"epoch": 2.978723404255319,
|
97 |
+
"total_flos": 57968476896000,
|
98 |
+
"step": 140
|
99 |
+
},
|
100 |
+
{
|
101 |
+
"loss": 0.11231231689453125,
|
102 |
+
"learning_rate": 1.5e-05,
|
103 |
+
"epoch": 3.1914893617021276,
|
104 |
+
"total_flos": 61862662224000,
|
105 |
+
"step": 150
|
106 |
+
},
|
107 |
+
{
|
108 |
+
"loss": 0.0972991943359375,
|
109 |
+
"learning_rate": 1.6000000000000003e-05,
|
110 |
+
"epoch": 3.404255319148936,
|
111 |
+
"total_flos": 66044337744000,
|
112 |
+
"step": 160
|
113 |
+
},
|
114 |
+
{
|
115 |
+
"loss": 0.09285507202148438,
|
116 |
+
"learning_rate": 1.7000000000000003e-05,
|
117 |
+
"epoch": 3.617021276595745,
|
118 |
+
"total_flos": 70226013264000,
|
119 |
+
"step": 170
|
120 |
+
},
|
121 |
+
{
|
122 |
+
"loss": 0.10490646362304687,
|
123 |
+
"learning_rate": 1.8e-05,
|
124 |
+
"epoch": 3.829787234042553,
|
125 |
+
"total_flos": 74407688784000,
|
126 |
+
"step": 180
|
127 |
+
},
|
128 |
+
{
|
129 |
+
"loss": 0.08623809814453125,
|
130 |
+
"learning_rate": 1.9e-05,
|
131 |
+
"epoch": 4.042553191489362,
|
132 |
+
"total_flos": 78301874112000,
|
133 |
+
"step": 190
|
134 |
+
},
|
135 |
+
{
|
136 |
+
"loss": 0.05904541015625,
|
137 |
+
"learning_rate": 2e-05,
|
138 |
+
"epoch": 4.25531914893617,
|
139 |
+
"total_flos": 82483549632000,
|
140 |
+
"step": 200
|
141 |
+
},
|
142 |
+
{
|
143 |
+
"loss": 0.052008056640625,
|
144 |
+
"learning_rate": 2.1e-05,
|
145 |
+
"epoch": 4.468085106382979,
|
146 |
+
"total_flos": 86665225152000,
|
147 |
+
"step": 210
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"loss": 0.06222076416015625,
|
151 |
+
"learning_rate": 2.2000000000000003e-05,
|
152 |
+
"epoch": 4.680851063829787,
|
153 |
+
"total_flos": 90846900672000,
|
154 |
+
"step": 220
|
155 |
+
},
|
156 |
+
{
|
157 |
+
"loss": 0.05680923461914063,
|
158 |
+
"learning_rate": 2.3000000000000003e-05,
|
159 |
+
"epoch": 4.8936170212765955,
|
160 |
+
"total_flos": 95028576192000,
|
161 |
+
"step": 230
|
162 |
+
},
|
163 |
+
{
|
164 |
+
"eval_loss": 0.0899211324525602,
|
165 |
+
"eval_accuracy_score": 0.974121282348397,
|
166 |
+
"eval_precision": 0.9776031434184675,
|
167 |
+
"eval_recall": 0.9749216300940439,
|
168 |
+
"eval_f1": 0.9762605454188739,
|
169 |
+
"epoch": 5.0,
|
170 |
+
"total_flos": 96831923760000,
|
171 |
+
"step": 235
|
172 |
+
}
|
173 |
+
]
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d385d4e8a6fd0c04dff5ab0ef9b091d66ad239a1d17ef4c6fee85d860bb6efee
|
3 |
+
size 435679343
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d683997fff536a8cb258bffa83dc61ab5f5b2167c2bb0fbd8408703c1e902c4
|
3 |
+
size 1775
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|