模型:
MLRS/BERTu
使用BERT(base)架构从头开始在Korpus Malti v4.0上对马耳他语进行单语预训练的模型。
该作品受 Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License 许可。此许可范围以外的权限可能在 https://mlrs.research.um.edu.mt/ 处提供。
这项工作首次在 Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for Maltese 中提出。请引用如下:
@inproceedings{BERTu,
title = "Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and {BERT} Models for {M}altese",
author = "Micallef, Kurt and
Gatt, Albert and
Tanti, Marc and
van der Plas, Lonneke and
Borg, Claudia",
booktitle = "Proceedings of the Third Workshop on Deep Learning for Low-Resource Natural Language Processing",
month = jul,
year = "2022",
address = "Hybrid",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.deeplo-1.10",
doi = "10.18653/v1/2022.deeplo-1.10",
pages = "90--101",
}