XLM-T: Multilingual Language Models in Twitter for Sentiment Analysis and Beyond

Language models are ubiquitous in current NLP, and their multilingual capacity has recently attracted considerable attention. However, current analyses have almost exclusively focused on (multilingual variants of) standard benchmarks, and have relied on clean pre-training and task-specific corpora as multilingual signals. In this paper, we introduce XLM-T, a model to train and evaluate multilingual language models in Twitter. In this paper we provide: (1) a new strong multilingual baseline consisting of an XLM-R (Conneau et al. 2020) model pre-trained on millions of tweets in over thirty languages, alongside starter code to subsequently fine-tune on a target task; and (2) a set of unified sentiment analysis Twitter datasets in eight different languages and a XLM-T model fine-tuned on them.

PDF Abstract LREC 2022 PDF LREC 2022 Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Sentiment Analysis TweetEval RoB-RT Emoji 31.4 # 2
Emotion 79.5 # 1
Hate 52.3 # 2
Irony 61.7 # 5
Offensive 80.5 # 1
Sentiment 72.6 # 2
Stance 69.3 # 2
ALL 65.2 # 2

Methods