dongxiaoqun
commited on
Commit
•
928a2cf
1
Parent(s):
1176351
Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ tags:
|
|
5 |
inference: False
|
6 |
---
|
7 |
|
8 |
-
IDEA-CCNL/Randeng_Pegasus_523M_Summary_Chinese model (Chinese) has 523M million parameter, pretrained on 180G Chinese data with GSG task which is stochastically sample important sentences with sampled gap sentence ratios by 25%. The pretraining task just same as the paper PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization mentioned.
|
9 |
|
10 |
Different from the English version of pegasus, considering that the Chinese sentence piece is unstable, we use jieba and Bertokenizer as the tokenizer in chinese pegasus model.
|
11 |
|
|
|
5 |
inference: False
|
6 |
---
|
7 |
|
8 |
+
IDEA-CCNL/Randeng_Pegasus_523M_Summary_Chinese model (Chinese) has 523M million parameter, pretrained on 180G Chinese data with GSG task which is stochastically sample important sentences with sampled gap sentence ratios by 25%. The pretraining task just as same as the paper PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization mentioned.
|
9 |
|
10 |
Different from the English version of pegasus, considering that the Chinese sentence piece is unstable, we use jieba and Bertokenizer as the tokenizer in chinese pegasus model.
|
11 |
|