Upload ./training.log with huggingface_hub
Browse files- training.log +244 -0
training.log
ADDED
@@ -0,0 +1,244 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2023-10-27 14:30:09,020 ----------------------------------------------------------------------------------------------------
|
2 |
+
2023-10-27 14:30:09,022 Model: "SequenceTagger(
|
3 |
+
(embeddings): TransformerWordEmbeddings(
|
4 |
+
(model): XLMRobertaModel(
|
5 |
+
(embeddings): XLMRobertaEmbeddings(
|
6 |
+
(word_embeddings): Embedding(250003, 1024)
|
7 |
+
(position_embeddings): Embedding(514, 1024, padding_idx=1)
|
8 |
+
(token_type_embeddings): Embedding(1, 1024)
|
9 |
+
(LayerNorm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
|
10 |
+
(dropout): Dropout(p=0.1, inplace=False)
|
11 |
+
)
|
12 |
+
(encoder): XLMRobertaEncoder(
|
13 |
+
(layer): ModuleList(
|
14 |
+
(0-23): 24 x XLMRobertaLayer(
|
15 |
+
(attention): XLMRobertaAttention(
|
16 |
+
(self): XLMRobertaSelfAttention(
|
17 |
+
(query): Linear(in_features=1024, out_features=1024, bias=True)
|
18 |
+
(key): Linear(in_features=1024, out_features=1024, bias=True)
|
19 |
+
(value): Linear(in_features=1024, out_features=1024, bias=True)
|
20 |
+
(dropout): Dropout(p=0.1, inplace=False)
|
21 |
+
)
|
22 |
+
(output): XLMRobertaSelfOutput(
|
23 |
+
(dense): Linear(in_features=1024, out_features=1024, bias=True)
|
24 |
+
(LayerNorm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
|
25 |
+
(dropout): Dropout(p=0.1, inplace=False)
|
26 |
+
)
|
27 |
+
)
|
28 |
+
(intermediate): XLMRobertaIntermediate(
|
29 |
+
(dense): Linear(in_features=1024, out_features=4096, bias=True)
|
30 |
+
(intermediate_act_fn): GELUActivation()
|
31 |
+
)
|
32 |
+
(output): XLMRobertaOutput(
|
33 |
+
(dense): Linear(in_features=4096, out_features=1024, bias=True)
|
34 |
+
(LayerNorm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
|
35 |
+
(dropout): Dropout(p=0.1, inplace=False)
|
36 |
+
)
|
37 |
+
)
|
38 |
+
)
|
39 |
+
)
|
40 |
+
(pooler): XLMRobertaPooler(
|
41 |
+
(dense): Linear(in_features=1024, out_features=1024, bias=True)
|
42 |
+
(activation): Tanh()
|
43 |
+
)
|
44 |
+
)
|
45 |
+
)
|
46 |
+
(locked_dropout): LockedDropout(p=0.5)
|
47 |
+
(linear): Linear(in_features=1024, out_features=17, bias=True)
|
48 |
+
(loss_function): CrossEntropyLoss()
|
49 |
+
)"
|
50 |
+
2023-10-27 14:30:09,022 ----------------------------------------------------------------------------------------------------
|
51 |
+
2023-10-27 14:30:09,022 Corpus: 14903 train + 3449 dev + 3658 test sentences
|
52 |
+
2023-10-27 14:30:09,022 ----------------------------------------------------------------------------------------------------
|
53 |
+
2023-10-27 14:30:09,022 Train: 14903 sentences
|
54 |
+
2023-10-27 14:30:09,022 (train_with_dev=False, train_with_test=False)
|
55 |
+
2023-10-27 14:30:09,022 ----------------------------------------------------------------------------------------------------
|
56 |
+
2023-10-27 14:30:09,022 Training Params:
|
57 |
+
2023-10-27 14:30:09,022 - learning_rate: "5e-06"
|
58 |
+
2023-10-27 14:30:09,022 - mini_batch_size: "4"
|
59 |
+
2023-10-27 14:30:09,022 - max_epochs: "10"
|
60 |
+
2023-10-27 14:30:09,022 - shuffle: "True"
|
61 |
+
2023-10-27 14:30:09,022 ----------------------------------------------------------------------------------------------------
|
62 |
+
2023-10-27 14:30:09,022 Plugins:
|
63 |
+
2023-10-27 14:30:09,022 - TensorboardLogger
|
64 |
+
2023-10-27 14:30:09,022 - LinearScheduler | warmup_fraction: '0.1'
|
65 |
+
2023-10-27 14:30:09,022 ----------------------------------------------------------------------------------------------------
|
66 |
+
2023-10-27 14:30:09,022 Final evaluation on model from best epoch (best-model.pt)
|
67 |
+
2023-10-27 14:30:09,023 - metric: "('micro avg', 'f1-score')"
|
68 |
+
2023-10-27 14:30:09,023 ----------------------------------------------------------------------------------------------------
|
69 |
+
2023-10-27 14:30:09,023 Computation:
|
70 |
+
2023-10-27 14:30:09,023 - compute on device: cuda:0
|
71 |
+
2023-10-27 14:30:09,023 - embedding storage: none
|
72 |
+
2023-10-27 14:30:09,023 ----------------------------------------------------------------------------------------------------
|
73 |
+
2023-10-27 14:30:09,023 Model training base path: "flair-clean-conll-lr5e-06-bs4-1"
|
74 |
+
2023-10-27 14:30:09,023 ----------------------------------------------------------------------------------------------------
|
75 |
+
2023-10-27 14:30:09,023 ----------------------------------------------------------------------------------------------------
|
76 |
+
2023-10-27 14:30:09,023 Logging anything other than scalars to TensorBoard is currently not supported.
|
77 |
+
2023-10-27 14:31:00,447 epoch 1 - iter 372/3726 - loss 3.59658936 - time (sec): 51.42 - samples/sec: 413.36 - lr: 0.000000 - momentum: 0.000000
|
78 |
+
2023-10-27 14:31:51,041 epoch 1 - iter 744/3726 - loss 2.26388470 - time (sec): 102.02 - samples/sec: 413.01 - lr: 0.000001 - momentum: 0.000000
|
79 |
+
2023-10-27 14:32:42,079 epoch 1 - iter 1116/3726 - loss 1.71280298 - time (sec): 153.05 - samples/sec: 407.18 - lr: 0.000001 - momentum: 0.000000
|
80 |
+
2023-10-27 14:33:32,757 epoch 1 - iter 1488/3726 - loss 1.39694066 - time (sec): 203.73 - samples/sec: 405.85 - lr: 0.000002 - momentum: 0.000000
|
81 |
+
2023-10-27 14:34:23,303 epoch 1 - iter 1860/3726 - loss 1.17731325 - time (sec): 254.28 - samples/sec: 406.61 - lr: 0.000002 - momentum: 0.000000
|
82 |
+
2023-10-27 14:35:13,791 epoch 1 - iter 2232/3726 - loss 1.01805377 - time (sec): 304.77 - samples/sec: 405.77 - lr: 0.000003 - momentum: 0.000000
|
83 |
+
2023-10-27 14:36:04,555 epoch 1 - iter 2604/3726 - loss 0.89618095 - time (sec): 355.53 - samples/sec: 404.90 - lr: 0.000003 - momentum: 0.000000
|
84 |
+
2023-10-27 14:36:55,602 epoch 1 - iter 2976/3726 - loss 0.80113668 - time (sec): 406.58 - samples/sec: 403.81 - lr: 0.000004 - momentum: 0.000000
|
85 |
+
2023-10-27 14:37:46,433 epoch 1 - iter 3348/3726 - loss 0.72868926 - time (sec): 457.41 - samples/sec: 401.79 - lr: 0.000004 - momentum: 0.000000
|
86 |
+
2023-10-27 14:38:36,663 epoch 1 - iter 3720/3726 - loss 0.66657776 - time (sec): 507.64 - samples/sec: 402.50 - lr: 0.000005 - momentum: 0.000000
|
87 |
+
2023-10-27 14:38:37,488 ----------------------------------------------------------------------------------------------------
|
88 |
+
2023-10-27 14:38:37,488 EPOCH 1 done: loss 0.6659 - lr: 0.000005
|
89 |
+
2023-10-27 14:39:02,791 DEV : loss 0.0933869257569313 - f1-score (micro avg) 0.9262
|
90 |
+
2023-10-27 14:39:02,853 saving best model
|
91 |
+
2023-10-27 14:39:05,885 ----------------------------------------------------------------------------------------------------
|
92 |
+
2023-10-27 14:39:56,330 epoch 2 - iter 372/3726 - loss 0.09283601 - time (sec): 50.44 - samples/sec: 393.59 - lr: 0.000005 - momentum: 0.000000
|
93 |
+
2023-10-27 14:40:47,261 epoch 2 - iter 744/3726 - loss 0.08756607 - time (sec): 101.37 - samples/sec: 396.30 - lr: 0.000005 - momentum: 0.000000
|
94 |
+
2023-10-27 14:41:38,098 epoch 2 - iter 1116/3726 - loss 0.08583083 - time (sec): 152.21 - samples/sec: 397.99 - lr: 0.000005 - momentum: 0.000000
|
95 |
+
2023-10-27 14:42:28,471 epoch 2 - iter 1488/3726 - loss 0.08370769 - time (sec): 202.58 - samples/sec: 400.87 - lr: 0.000005 - momentum: 0.000000
|
96 |
+
2023-10-27 14:43:19,871 epoch 2 - iter 1860/3726 - loss 0.08404411 - time (sec): 253.98 - samples/sec: 399.49 - lr: 0.000005 - momentum: 0.000000
|
97 |
+
2023-10-27 14:44:10,168 epoch 2 - iter 2232/3726 - loss 0.08124313 - time (sec): 304.28 - samples/sec: 397.73 - lr: 0.000005 - momentum: 0.000000
|
98 |
+
2023-10-27 14:45:01,496 epoch 2 - iter 2604/3726 - loss 0.08228873 - time (sec): 355.61 - samples/sec: 397.75 - lr: 0.000005 - momentum: 0.000000
|
99 |
+
2023-10-27 14:45:52,335 epoch 2 - iter 2976/3726 - loss 0.08159191 - time (sec): 406.45 - samples/sec: 399.36 - lr: 0.000005 - momentum: 0.000000
|
100 |
+
2023-10-27 14:46:43,142 epoch 2 - iter 3348/3726 - loss 0.08066519 - time (sec): 457.26 - samples/sec: 401.12 - lr: 0.000005 - momentum: 0.000000
|
101 |
+
2023-10-27 14:47:34,049 epoch 2 - iter 3720/3726 - loss 0.08043171 - time (sec): 508.16 - samples/sec: 402.15 - lr: 0.000004 - momentum: 0.000000
|
102 |
+
2023-10-27 14:47:34,866 ----------------------------------------------------------------------------------------------------
|
103 |
+
2023-10-27 14:47:34,866 EPOCH 2 done: loss 0.0805 - lr: 0.000004
|
104 |
+
2023-10-27 14:48:01,712 DEV : loss 0.06152888387441635 - f1-score (micro avg) 0.9481
|
105 |
+
2023-10-27 14:48:01,789 saving best model
|
106 |
+
2023-10-27 14:48:05,294 ----------------------------------------------------------------------------------------------------
|
107 |
+
2023-10-27 14:48:55,650 epoch 3 - iter 372/3726 - loss 0.06085669 - time (sec): 50.35 - samples/sec: 406.39 - lr: 0.000004 - momentum: 0.000000
|
108 |
+
2023-10-27 14:49:46,492 epoch 3 - iter 744/3726 - loss 0.05743897 - time (sec): 101.20 - samples/sec: 400.38 - lr: 0.000004 - momentum: 0.000000
|
109 |
+
2023-10-27 14:50:36,994 epoch 3 - iter 1116/3726 - loss 0.05779768 - time (sec): 151.70 - samples/sec: 398.09 - lr: 0.000004 - momentum: 0.000000
|
110 |
+
2023-10-27 14:51:26,995 epoch 3 - iter 1488/3726 - loss 0.05644504 - time (sec): 201.70 - samples/sec: 401.46 - lr: 0.000004 - momentum: 0.000000
|
111 |
+
2023-10-27 14:52:17,584 epoch 3 - iter 1860/3726 - loss 0.05572416 - time (sec): 252.29 - samples/sec: 403.62 - lr: 0.000004 - momentum: 0.000000
|
112 |
+
2023-10-27 14:53:08,570 epoch 3 - iter 2232/3726 - loss 0.05352112 - time (sec): 303.27 - samples/sec: 403.61 - lr: 0.000004 - momentum: 0.000000
|
113 |
+
2023-10-27 14:53:59,036 epoch 3 - iter 2604/3726 - loss 0.05461713 - time (sec): 353.74 - samples/sec: 404.61 - lr: 0.000004 - momentum: 0.000000
|
114 |
+
2023-10-27 14:54:49,877 epoch 3 - iter 2976/3726 - loss 0.05390525 - time (sec): 404.58 - samples/sec: 405.62 - lr: 0.000004 - momentum: 0.000000
|
115 |
+
2023-10-27 14:55:41,535 epoch 3 - iter 3348/3726 - loss 0.05442763 - time (sec): 456.24 - samples/sec: 402.73 - lr: 0.000004 - momentum: 0.000000
|
116 |
+
2023-10-27 14:56:32,799 epoch 3 - iter 3720/3726 - loss 0.05458475 - time (sec): 507.50 - samples/sec: 402.71 - lr: 0.000004 - momentum: 0.000000
|
117 |
+
2023-10-27 14:56:33,616 ----------------------------------------------------------------------------------------------------
|
118 |
+
2023-10-27 14:56:33,617 EPOCH 3 done: loss 0.0545 - lr: 0.000004
|
119 |
+
2023-10-27 14:57:00,329 DEV : loss 0.06314758211374283 - f1-score (micro avg) 0.9612
|
120 |
+
2023-10-27 14:57:00,411 saving best model
|
121 |
+
2023-10-27 14:57:03,863 ----------------------------------------------------------------------------------------------------
|
122 |
+
2023-10-27 14:57:54,121 epoch 4 - iter 372/3726 - loss 0.03226694 - time (sec): 50.26 - samples/sec: 411.62 - lr: 0.000004 - momentum: 0.000000
|
123 |
+
2023-10-27 14:58:44,919 epoch 4 - iter 744/3726 - loss 0.03936813 - time (sec): 101.05 - samples/sec: 405.68 - lr: 0.000004 - momentum: 0.000000
|
124 |
+
2023-10-27 14:59:35,509 epoch 4 - iter 1116/3726 - loss 0.03970607 - time (sec): 151.64 - samples/sec: 406.37 - lr: 0.000004 - momentum: 0.000000
|
125 |
+
2023-10-27 15:00:23,759 epoch 4 - iter 1488/3726 - loss 0.04029682 - time (sec): 199.89 - samples/sec: 408.57 - lr: 0.000004 - momentum: 0.000000
|
126 |
+
2023-10-27 15:01:12,178 epoch 4 - iter 1860/3726 - loss 0.03993881 - time (sec): 248.31 - samples/sec: 412.03 - lr: 0.000004 - momentum: 0.000000
|
127 |
+
2023-10-27 15:01:59,669 epoch 4 - iter 2232/3726 - loss 0.03825023 - time (sec): 295.80 - samples/sec: 415.38 - lr: 0.000004 - momentum: 0.000000
|
128 |
+
2023-10-27 15:02:46,739 epoch 4 - iter 2604/3726 - loss 0.03860270 - time (sec): 342.87 - samples/sec: 417.88 - lr: 0.000004 - momentum: 0.000000
|
129 |
+
2023-10-27 15:03:33,155 epoch 4 - iter 2976/3726 - loss 0.03883180 - time (sec): 389.29 - samples/sec: 420.36 - lr: 0.000003 - momentum: 0.000000
|
130 |
+
2023-10-27 15:04:18,906 epoch 4 - iter 3348/3726 - loss 0.03843582 - time (sec): 435.04 - samples/sec: 424.15 - lr: 0.000003 - momentum: 0.000000
|
131 |
+
2023-10-27 15:05:06,069 epoch 4 - iter 3720/3726 - loss 0.03900868 - time (sec): 482.20 - samples/sec: 423.76 - lr: 0.000003 - momentum: 0.000000
|
132 |
+
2023-10-27 15:05:06,804 ----------------------------------------------------------------------------------------------------
|
133 |
+
2023-10-27 15:05:06,805 EPOCH 4 done: loss 0.0390 - lr: 0.000003
|
134 |
+
2023-10-27 15:05:31,408 DEV : loss 0.04959763213992119 - f1-score (micro avg) 0.9678
|
135 |
+
2023-10-27 15:05:31,469 saving best model
|
136 |
+
2023-10-27 15:05:34,206 ----------------------------------------------------------------------------------------------------
|
137 |
+
2023-10-27 15:06:22,854 epoch 5 - iter 372/3726 - loss 0.02632753 - time (sec): 48.65 - samples/sec: 413.87 - lr: 0.000003 - momentum: 0.000000
|
138 |
+
2023-10-27 15:07:09,362 epoch 5 - iter 744/3726 - loss 0.02847207 - time (sec): 95.15 - samples/sec: 431.29 - lr: 0.000003 - momentum: 0.000000
|
139 |
+
2023-10-27 15:07:56,415 epoch 5 - iter 1116/3726 - loss 0.02792443 - time (sec): 142.21 - samples/sec: 432.32 - lr: 0.000003 - momentum: 0.000000
|
140 |
+
2023-10-27 15:08:43,809 epoch 5 - iter 1488/3726 - loss 0.02808337 - time (sec): 189.60 - samples/sec: 432.27 - lr: 0.000003 - momentum: 0.000000
|
141 |
+
2023-10-27 15:09:32,174 epoch 5 - iter 1860/3726 - loss 0.02852601 - time (sec): 237.97 - samples/sec: 426.16 - lr: 0.000003 - momentum: 0.000000
|
142 |
+
2023-10-27 15:10:19,545 epoch 5 - iter 2232/3726 - loss 0.02859382 - time (sec): 285.34 - samples/sec: 427.84 - lr: 0.000003 - momentum: 0.000000
|
143 |
+
2023-10-27 15:11:08,221 epoch 5 - iter 2604/3726 - loss 0.02812313 - time (sec): 334.01 - samples/sec: 426.57 - lr: 0.000003 - momentum: 0.000000
|
144 |
+
2023-10-27 15:11:55,520 epoch 5 - iter 2976/3726 - loss 0.02819121 - time (sec): 381.31 - samples/sec: 427.48 - lr: 0.000003 - momentum: 0.000000
|
145 |
+
2023-10-27 15:12:43,398 epoch 5 - iter 3348/3726 - loss 0.02832524 - time (sec): 429.19 - samples/sec: 426.98 - lr: 0.000003 - momentum: 0.000000
|
146 |
+
2023-10-27 15:13:30,169 epoch 5 - iter 3720/3726 - loss 0.02864353 - time (sec): 475.96 - samples/sec: 429.13 - lr: 0.000003 - momentum: 0.000000
|
147 |
+
2023-10-27 15:13:30,908 ----------------------------------------------------------------------------------------------------
|
148 |
+
2023-10-27 15:13:30,909 EPOCH 5 done: loss 0.0286 - lr: 0.000003
|
149 |
+
2023-10-27 15:13:55,872 DEV : loss 0.051205169409513474 - f1-score (micro avg) 0.969
|
150 |
+
2023-10-27 15:13:55,926 saving best model
|
151 |
+
2023-10-27 15:13:58,971 ----------------------------------------------------------------------------------------------------
|
152 |
+
2023-10-27 15:14:46,203 epoch 6 - iter 372/3726 - loss 0.02831335 - time (sec): 47.23 - samples/sec: 428.83 - lr: 0.000003 - momentum: 0.000000
|
153 |
+
2023-10-27 15:15:35,236 epoch 6 - iter 744/3726 - loss 0.02273400 - time (sec): 96.26 - samples/sec: 426.78 - lr: 0.000003 - momentum: 0.000000
|
154 |
+
2023-10-27 15:16:23,193 epoch 6 - iter 1116/3726 - loss 0.02109002 - time (sec): 144.22 - samples/sec: 423.66 - lr: 0.000003 - momentum: 0.000000
|
155 |
+
2023-10-27 15:17:10,302 epoch 6 - iter 1488/3726 - loss 0.02166135 - time (sec): 191.33 - samples/sec: 429.37 - lr: 0.000003 - momentum: 0.000000
|
156 |
+
2023-10-27 15:17:58,889 epoch 6 - iter 1860/3726 - loss 0.02127214 - time (sec): 239.91 - samples/sec: 429.35 - lr: 0.000003 - momentum: 0.000000
|
157 |
+
2023-10-27 15:18:45,994 epoch 6 - iter 2232/3726 - loss 0.02116817 - time (sec): 287.02 - samples/sec: 428.54 - lr: 0.000002 - momentum: 0.000000
|
158 |
+
2023-10-27 15:19:34,532 epoch 6 - iter 2604/3726 - loss 0.02139499 - time (sec): 335.56 - samples/sec: 427.19 - lr: 0.000002 - momentum: 0.000000
|
159 |
+
2023-10-27 15:20:22,599 epoch 6 - iter 2976/3726 - loss 0.02130084 - time (sec): 383.62 - samples/sec: 425.85 - lr: 0.000002 - momentum: 0.000000
|
160 |
+
2023-10-27 15:21:10,024 epoch 6 - iter 3348/3726 - loss 0.02121915 - time (sec): 431.05 - samples/sec: 425.94 - lr: 0.000002 - momentum: 0.000000
|
161 |
+
2023-10-27 15:21:57,177 epoch 6 - iter 3720/3726 - loss 0.02100853 - time (sec): 478.20 - samples/sec: 427.07 - lr: 0.000002 - momentum: 0.000000
|
162 |
+
2023-10-27 15:21:57,964 ----------------------------------------------------------------------------------------------------
|
163 |
+
2023-10-27 15:21:57,964 EPOCH 6 done: loss 0.0210 - lr: 0.000002
|
164 |
+
2023-10-27 15:22:23,206 DEV : loss 0.05652967095375061 - f1-score (micro avg) 0.9686
|
165 |
+
2023-10-27 15:22:23,264 ----------------------------------------------------------------------------------------------------
|
166 |
+
2023-10-27 15:23:11,169 epoch 7 - iter 372/3726 - loss 0.01511217 - time (sec): 47.90 - samples/sec: 438.07 - lr: 0.000002 - momentum: 0.000000
|
167 |
+
2023-10-27 15:23:58,577 epoch 7 - iter 744/3726 - loss 0.01678792 - time (sec): 95.31 - samples/sec: 441.01 - lr: 0.000002 - momentum: 0.000000
|
168 |
+
2023-10-27 15:24:46,315 epoch 7 - iter 1116/3726 - loss 0.01767805 - time (sec): 143.05 - samples/sec: 436.03 - lr: 0.000002 - momentum: 0.000000
|
169 |
+
2023-10-27 15:25:34,119 epoch 7 - iter 1488/3726 - loss 0.01634669 - time (sec): 190.85 - samples/sec: 434.79 - lr: 0.000002 - momentum: 0.000000
|
170 |
+
2023-10-27 15:26:21,351 epoch 7 - iter 1860/3726 - loss 0.02015737 - time (sec): 238.08 - samples/sec: 437.30 - lr: 0.000002 - momentum: 0.000000
|
171 |
+
2023-10-27 15:27:08,458 epoch 7 - iter 2232/3726 - loss 0.01895408 - time (sec): 285.19 - samples/sec: 435.98 - lr: 0.000002 - momentum: 0.000000
|
172 |
+
2023-10-27 15:27:56,799 epoch 7 - iter 2604/3726 - loss 0.01784524 - time (sec): 333.53 - samples/sec: 433.67 - lr: 0.000002 - momentum: 0.000000
|
173 |
+
2023-10-27 15:28:45,164 epoch 7 - iter 2976/3726 - loss 0.01736500 - time (sec): 381.90 - samples/sec: 431.20 - lr: 0.000002 - momentum: 0.000000
|
174 |
+
2023-10-27 15:29:33,002 epoch 7 - iter 3348/3726 - loss 0.01709685 - time (sec): 429.74 - samples/sec: 429.19 - lr: 0.000002 - momentum: 0.000000
|
175 |
+
2023-10-27 15:30:19,962 epoch 7 - iter 3720/3726 - loss 0.01680664 - time (sec): 476.70 - samples/sec: 428.73 - lr: 0.000002 - momentum: 0.000000
|
176 |
+
2023-10-27 15:30:20,674 ----------------------------------------------------------------------------------------------------
|
177 |
+
2023-10-27 15:30:20,674 EPOCH 7 done: loss 0.0168 - lr: 0.000002
|
178 |
+
2023-10-27 15:30:43,202 DEV : loss 0.052628789097070694 - f1-score (micro avg) 0.9722
|
179 |
+
2023-10-27 15:30:43,260 saving best model
|
180 |
+
2023-10-27 15:30:46,516 ----------------------------------------------------------------------------------------------------
|
181 |
+
2023-10-27 15:31:32,754 epoch 8 - iter 372/3726 - loss 0.01025399 - time (sec): 46.24 - samples/sec: 450.08 - lr: 0.000002 - momentum: 0.000000
|
182 |
+
2023-10-27 15:32:18,558 epoch 8 - iter 744/3726 - loss 0.01126091 - time (sec): 92.04 - samples/sec: 447.34 - lr: 0.000002 - momentum: 0.000000
|
183 |
+
2023-10-27 15:33:03,647 epoch 8 - iter 1116/3726 - loss 0.01172293 - time (sec): 137.13 - samples/sec: 445.99 - lr: 0.000002 - momentum: 0.000000
|
184 |
+
2023-10-27 15:33:49,381 epoch 8 - iter 1488/3726 - loss 0.01219846 - time (sec): 182.86 - samples/sec: 440.90 - lr: 0.000001 - momentum: 0.000000
|
185 |
+
2023-10-27 15:34:35,063 epoch 8 - iter 1860/3726 - loss 0.01190833 - time (sec): 228.55 - samples/sec: 443.89 - lr: 0.000001 - momentum: 0.000000
|
186 |
+
2023-10-27 15:35:20,762 epoch 8 - iter 2232/3726 - loss 0.01209883 - time (sec): 274.24 - samples/sec: 442.06 - lr: 0.000001 - momentum: 0.000000
|
187 |
+
2023-10-27 15:36:06,906 epoch 8 - iter 2604/3726 - loss 0.01227890 - time (sec): 320.39 - samples/sec: 442.88 - lr: 0.000001 - momentum: 0.000000
|
188 |
+
2023-10-27 15:36:52,771 epoch 8 - iter 2976/3726 - loss 0.01220756 - time (sec): 366.25 - samples/sec: 444.61 - lr: 0.000001 - momentum: 0.000000
|
189 |
+
2023-10-27 15:37:38,958 epoch 8 - iter 3348/3726 - loss 0.01181082 - time (sec): 412.44 - samples/sec: 446.10 - lr: 0.000001 - momentum: 0.000000
|
190 |
+
2023-10-27 15:38:24,879 epoch 8 - iter 3720/3726 - loss 0.01141249 - time (sec): 458.36 - samples/sec: 445.64 - lr: 0.000001 - momentum: 0.000000
|
191 |
+
2023-10-27 15:38:25,618 ----------------------------------------------------------------------------------------------------
|
192 |
+
2023-10-27 15:38:25,619 EPOCH 8 done: loss 0.0114 - lr: 0.000001
|
193 |
+
2023-10-27 15:38:49,028 DEV : loss 0.053019002079963684 - f1-score (micro avg) 0.9716
|
194 |
+
2023-10-27 15:38:49,088 ----------------------------------------------------------------------------------------------------
|
195 |
+
2023-10-27 15:39:35,150 epoch 9 - iter 372/3726 - loss 0.00801185 - time (sec): 46.06 - samples/sec: 438.23 - lr: 0.000001 - momentum: 0.000000
|
196 |
+
2023-10-27 15:40:20,869 epoch 9 - iter 744/3726 - loss 0.00671566 - time (sec): 91.78 - samples/sec: 435.43 - lr: 0.000001 - momentum: 0.000000
|
197 |
+
2023-10-27 15:41:07,081 epoch 9 - iter 1116/3726 - loss 0.00667937 - time (sec): 137.99 - samples/sec: 438.20 - lr: 0.000001 - momentum: 0.000000
|
198 |
+
2023-10-27 15:41:53,955 epoch 9 - iter 1488/3726 - loss 0.00681524 - time (sec): 184.87 - samples/sec: 439.41 - lr: 0.000001 - momentum: 0.000000
|
199 |
+
2023-10-27 15:42:41,635 epoch 9 - iter 1860/3726 - loss 0.00721784 - time (sec): 232.55 - samples/sec: 437.37 - lr: 0.000001 - momentum: 0.000000
|
200 |
+
2023-10-27 15:43:28,277 epoch 9 - iter 2232/3726 - loss 0.00697985 - time (sec): 279.19 - samples/sec: 440.81 - lr: 0.000001 - momentum: 0.000000
|
201 |
+
2023-10-27 15:44:14,754 epoch 9 - iter 2604/3726 - loss 0.00793074 - time (sec): 325.66 - samples/sec: 440.67 - lr: 0.000001 - momentum: 0.000000
|
202 |
+
2023-10-27 15:45:01,296 epoch 9 - iter 2976/3726 - loss 0.00864270 - time (sec): 372.21 - samples/sec: 439.58 - lr: 0.000001 - momentum: 0.000000
|
203 |
+
2023-10-27 15:45:47,999 epoch 9 - iter 3348/3726 - loss 0.00830012 - time (sec): 418.91 - samples/sec: 438.33 - lr: 0.000001 - momentum: 0.000000
|
204 |
+
2023-10-27 15:46:34,892 epoch 9 - iter 3720/3726 - loss 0.00824128 - time (sec): 465.80 - samples/sec: 438.38 - lr: 0.000001 - momentum: 0.000000
|
205 |
+
2023-10-27 15:46:35,624 ----------------------------------------------------------------------------------------------------
|
206 |
+
2023-10-27 15:46:35,624 EPOCH 9 done: loss 0.0083 - lr: 0.000001
|
207 |
+
2023-10-27 15:46:59,614 DEV : loss 0.05303654819726944 - f1-score (micro avg) 0.9734
|
208 |
+
2023-10-27 15:46:59,675 saving best model
|
209 |
+
2023-10-27 15:47:02,455 ----------------------------------------------------------------------------------------------------
|
210 |
+
2023-10-27 15:47:49,146 epoch 10 - iter 372/3726 - loss 0.00384825 - time (sec): 46.69 - samples/sec: 434.52 - lr: 0.000001 - momentum: 0.000000
|
211 |
+
2023-10-27 15:48:35,604 epoch 10 - iter 744/3726 - loss 0.00345501 - time (sec): 93.15 - samples/sec: 434.11 - lr: 0.000000 - momentum: 0.000000
|
212 |
+
2023-10-27 15:49:22,080 epoch 10 - iter 1116/3726 - loss 0.00418854 - time (sec): 139.62 - samples/sec: 435.29 - lr: 0.000000 - momentum: 0.000000
|
213 |
+
2023-10-27 15:50:08,505 epoch 10 - iter 1488/3726 - loss 0.00522497 - time (sec): 186.05 - samples/sec: 430.51 - lr: 0.000000 - momentum: 0.000000
|
214 |
+
2023-10-27 15:50:54,906 epoch 10 - iter 1860/3726 - loss 0.00509842 - time (sec): 232.45 - samples/sec: 435.12 - lr: 0.000000 - momentum: 0.000000
|
215 |
+
2023-10-27 15:51:41,148 epoch 10 - iter 2232/3726 - loss 0.00554209 - time (sec): 278.69 - samples/sec: 437.63 - lr: 0.000000 - momentum: 0.000000
|
216 |
+
2023-10-27 15:52:26,664 epoch 10 - iter 2604/3726 - loss 0.00584885 - time (sec): 324.21 - samples/sec: 441.07 - lr: 0.000000 - momentum: 0.000000
|
217 |
+
2023-10-27 15:53:13,641 epoch 10 - iter 2976/3726 - loss 0.00599832 - time (sec): 371.18 - samples/sec: 440.13 - lr: 0.000000 - momentum: 0.000000
|
218 |
+
2023-10-27 15:53:59,425 epoch 10 - iter 3348/3726 - loss 0.00600259 - time (sec): 416.97 - samples/sec: 442.42 - lr: 0.000000 - momentum: 0.000000
|
219 |
+
2023-10-27 15:54:45,541 epoch 10 - iter 3720/3726 - loss 0.00576228 - time (sec): 463.08 - samples/sec: 441.09 - lr: 0.000000 - momentum: 0.000000
|
220 |
+
2023-10-27 15:54:46,272 ----------------------------------------------------------------------------------------------------
|
221 |
+
2023-10-27 15:54:46,273 EPOCH 10 done: loss 0.0058 - lr: 0.000000
|
222 |
+
2023-10-27 15:55:09,835 DEV : loss 0.05434899777173996 - f1-score (micro avg) 0.9727
|
223 |
+
2023-10-27 15:55:12,069 ----------------------------------------------------------------------------------------------------
|
224 |
+
2023-10-27 15:55:12,071 Loading model from best epoch ...
|
225 |
+
2023-10-27 15:55:20,258 SequenceTagger predicts: Dictionary with 17 tags: O, S-ORG, B-ORG, E-ORG, I-ORG, S-PER, B-PER, E-PER, I-PER, S-LOC, B-LOC, E-LOC, I-LOC, S-MISC, B-MISC, E-MISC, I-MISC
|
226 |
+
2023-10-27 15:55:43,551
|
227 |
+
Results:
|
228 |
+
- F-score (micro) 0.97
|
229 |
+
- F-score (macro) 0.9642
|
230 |
+
- Accuracy 0.9559
|
231 |
+
|
232 |
+
By class:
|
233 |
+
precision recall f1-score support
|
234 |
+
|
235 |
+
ORG 0.9700 0.9665 0.9682 1909
|
236 |
+
PER 0.9944 0.9956 0.9950 1591
|
237 |
+
LOC 0.9684 0.9745 0.9714 1413
|
238 |
+
MISC 0.9245 0.9200 0.9222 812
|
239 |
+
|
240 |
+
micro avg 0.9700 0.9700 0.9700 5725
|
241 |
+
macro avg 0.9643 0.9641 0.9642 5725
|
242 |
+
weighted avg 0.9699 0.9700 0.9699 5725
|
243 |
+
|
244 |
+
2023-10-27 15:55:43,551 ----------------------------------------------------------------------------------------------------
|