model update
Browse files
README.md
CHANGED
@@ -27,21 +27,6 @@ model-index:
|
|
27 |
type: default
|
28 |
args: default
|
29 |
metrics:
|
30 |
-
- name: BLEU4 (Question & Answer Generation)
|
31 |
-
type: bleu4_question_answer_generation
|
32 |
-
value: 0.0
|
33 |
-
- name: ROUGE-L (Question & Answer Generation)
|
34 |
-
type: rouge_l_question_answer_generation
|
35 |
-
value: 7.56
|
36 |
-
- name: METEOR (Question & Answer Generation)
|
37 |
-
type: meteor_question_answer_generation
|
38 |
-
value: 5.02
|
39 |
-
- name: BERTScore (Question & Answer Generation)
|
40 |
-
type: bertscore_question_answer_generation
|
41 |
-
value: 53.64
|
42 |
-
- name: MoverScore (Question & Answer Generation)
|
43 |
-
type: moverscore_question_answer_generation
|
44 |
-
value: 46.71
|
45 |
- name: QAAlignedF1Score-BERTScore (Question & Answer Generation)
|
46 |
type: qa_aligned_f1_score_bertscore_question_answer_generation
|
47 |
value: 0.1
|
@@ -103,20 +88,12 @@ output = pipe("Empfangs- und Sendeantenne sollen in ihrer Polarisation übereins
|
|
103 |
|
104 |
| | Score | Type | Dataset |
|
105 |
|:--------------------------------|--------:|:--------|:-------------------------------------------------------------------|
|
106 |
-
| BERTScore
|
107 |
-
|
|
108 |
-
|
|
109 |
-
|
|
110 |
-
|
|
111 |
-
|
|
112 |
-
| MoverScore | 46.71 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
113 |
-
| QAAlignedF1Score (BERTScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
114 |
-
| QAAlignedF1Score (MoverScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
115 |
-
| QAAlignedPrecision (BERTScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
116 |
-
| QAAlignedPrecision (MoverScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
117 |
-
| QAAlignedRecall (BERTScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
118 |
-
| QAAlignedRecall (MoverScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
119 |
-
| ROUGE_L | 7.56 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
120 |
|
121 |
|
122 |
|
|
|
27 |
type: default
|
28 |
args: default
|
29 |
metrics:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
- name: QAAlignedF1Score-BERTScore (Question & Answer Generation)
|
31 |
type: qa_aligned_f1_score_bertscore_question_answer_generation
|
32 |
value: 0.1
|
|
|
88 |
|
89 |
| | Score | Type | Dataset |
|
90 |
|:--------------------------------|--------:|:--------|:-------------------------------------------------------------------|
|
91 |
+
| QAAlignedF1Score (BERTScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
92 |
+
| QAAlignedF1Score (MoverScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
93 |
+
| QAAlignedPrecision (BERTScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
94 |
+
| QAAlignedPrecision (MoverScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
95 |
+
| QAAlignedRecall (BERTScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
96 |
+
| QAAlignedRecall (MoverScore) | 0.1 | default | [lmqg/qag_dequad](https://huggingface.co/datasets/lmqg/qag_dequad) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
97 |
|
98 |
|
99 |
|