File size: 815 Bytes
9b18a17 3a3b6d8 9b18a17 3a3b6d8 9b18a17 3a3b6d8 ab423af 3a3b6d8 8110d48 ab423af 3a3b6d8 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 |
---
language: ko
license: mit
tags:
- bart
- grammar
---
# kogrammar-distil
Dataset: 국립국어원 맞춤법 교정 말뭉치
<br>
<br>
**Backbone Model**:
- [kobart-base-v2](https://huggingface.co/gogamza/kobart-base-v2/blob/main/README.md)
**Contributors**:
- ChangGeon Kang: [NeverAsking](https://github.com/NeverAsking/)
- Subin Park: [SOL1archive](https://github.com/SOL1archive/)
**GitHub Repo**:
- [SOL1archive/KoGrammar](https://github.com/SOL1archive/KoGrammar)
## Train Method
전체 데이터셋 중 약 67.5%를 학습데이터로 활용하여 학습함.
<br>
SFT Distillation을 이용해 [kogrammar-base](https://huggingface.co/theSOL1/kogrammar-base) 모델의 Decoder Layer를 6개에서 3개로 줄여 다시 학습시킴.
## Metric
|BLEU-2|ROUGE-2 F1|
|-|-|
|77.8 %|55.0 %|
|