Automatic Translation Alignment of Ancient Greek Texts
GRC-ALIGNMENT model is an XLM-RoBERTa-based model, fine-tuned for automatic multilingual text alignment at the word level. The model is trained on 12 million monolingual ancient Greek tokens with Masked Language Model (MLM) training objective. Further, the model is fine-tuned on 45k parallel sentences, mainly in ancient Greek-English, Greek-Latin, and Greek-Georgian.
Multilingual Training Dataset
Languages | Sentences | Source |
---|---|---|
GRC-ENG | 32.500 | Perseus Digital Library (Iliad, Odyssey, Xenophon, New Testament) |
GRC-LAT | 8.200 | Digital Fragmenta Historicorum Graecorum project |
GRC-KAT GRC-ENG GRC-LAT GRC-ITA GRC-POR |
4.000 | UGARIT Translation Alignment Editor |
Model Performance
Languages | Alignment Error Rate |
---|---|
GRC-ENG | 19.73% (IterMax) |
GRC-POR | 23.91% (IterMax) |
GRC-LAT | 10.60% (ArgMax) |
The gold standard datasets are available on Github.
If you use this model, please cite our papers:
@InProceedings{yousef-EtAl:2022:LREC, author = {Yousef, Tariq and Palladino, Chiara and Shamsian, Farnoosh and d’Orange Ferreira, Anise and Ferreira dos Reis, Michel}, title = {An automatic model and Gold Standard for translation alignment of Ancient Greek}, booktitle = {Proceedings of the Language Resources and Evaluation Conference}, month = {June}, year = {2022}, address = {Marseille, France}, publisher = {European Language Resources Association}, pages = {5894--5905}, url = {https://aclanthology.org/2022.lrec-1.634} } @InProceedings{yousef-EtAl:2022:LT4HALA2022, author = {Yousef, Tariq and Palladino, Chiara and Wright, David J. and Berti, Monica}, title = {Automatic Translation Alignment for Ancient Greek and Latin}, booktitle = {Proceedings of the Second Workshop on Language Technologies for Historical and Ancient Languages}, month = {June}, year = {2022}, address = {Marseille, France}, publisher = {European Language Resources Association}, pages = {101--107}, url = {https://aclanthology.org/2022.lt4hala2022-1.14} }
- Downloads last month
- 27
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.