--- language: pl license: apache-2.0 ---

polish-roberta-large-v1

An encoder model based on the RoBERTa architecture, pre-trained on a large corpus of Polish texts. More information can be found in our [GitHub repository](https://github.com/sdadas/polish-roberta) and in the publication [Pre-training polish transformer-based language models at scale](https://arxiv.org/pdf/2006.04229). ## Citation ```bibtex @inproceedings{dadas2020pre, title={Pre-training polish transformer-based language models at scale}, author={Dadas, S{\l}awomir and Pere{\l}kiewicz, Micha{\l} and Po{\'s}wiata, Rafa{\l}}, booktitle={International Conference on Artificial Intelligence and Soft Computing}, pages={301--314}, year={2020}, organization={Springer} } ```