Pre-training and Evaluating Transformer-based Language Models for Icelandic

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

In this paper, we evaluate several Transformer-based language models for Icelandic on four downstream tasks: Part-of-Speech tagging, Named Entity Recognition. Dependency Parsing, and Automatic Text Summarization. We pre-train four types of monolingual ELECTRA and ConvBERT models and compare our results to a previously trained monolingual RoBERTa model and the multilingual mBERT model. We find that the Transformer models obtain better results, often by a large margin, compared to previous state-of-the-art models. Furthermore, our results indicate that pre-training larger language models results in a significant reduction in error rates in comparison to smaller models. Finally, our results show that the monolingual models for Icelandic outperform a comparably sized multilingual model.

Original languageEnglish
Title of host publication2022 Language Resources and Evaluation Conference, LREC 2022
EditorsNicoletta Calzolari, Frederic Bechet, Philippe Blache, Khalid Choukri, Christopher Cieri, Thierry Declerck, Sara Goggi, Hitoshi Isahara, Bente Maegaard, Joseph Mariani, Helene Mazo, Jan Odijk, Stelios Piperidis
PublisherEuropean Language Resources Association (ELRA)
Pages7386-7391
Number of pages6
ISBN (Electronic)9791095546726
Publication statusPublished - 2022
Event13th International Conference on Language Resources and Evaluation Conference, LREC 2022 - Marseille, France
Duration: 20 Jun 202225 Jun 2022

Publication series

Name2022 Language Resources and Evaluation Conference, LREC 2022

Conference

Conference13th International Conference on Language Resources and Evaluation Conference, LREC 2022
Country/TerritoryFrance
CityMarseille
Period20/06/2225/06/22

Bibliographical note

Funding Information: This project was funded by the Language Technology Programme for Icelandic 2019-2023 (Nikulásdóttir et Funding Information: This project was funded by the Language Technology Programme for Icelandic 2019-2023 (Nikulásdóttir et al., 2020). The programme, which is managed and coordinated by Almannarómur10, is funded by the Icelandic Ministry of Education, Science and Culture. This research was also supported with Cloud TPUs from Google's TPU Research Cloud (TRC). Publisher Copyright: © European Language Resources Association (ELRA), licensed under CC-BY-NC-4.0.

Other keywords

  • Evaluation
  • Icelandic
  • Language Models
  • Transformer

Fingerprint

Dive into the research topics of 'Pre-training and Evaluating Transformer-based Language Models for Icelandic'. Together they form a unique fingerprint.

Cite this