Pretrained Multilingual Language Models

13 papers with code • 0 benchmarks • 1 datasets

This task has no description! Would you like to contribute one?

Datasets


Most implemented papers

How Good is Your Tokenizer? On the Monolingual Performance of Multilingual Language Models

Adapter-Hub/hgiyt ACL 2021

In this work, we provide a systematic and comprehensive empirical comparison of pretrained multilingual language models versus their monolingual counterparts with regard to their monolingual task performance.

Investigating Math Word Problems using Pretrained Multilingual Language Models

VisualJoyce/andushu 19 May 2021

In this paper, we revisit math word problems~(MWPs) from the cross-lingual and multilingual perspective.

Specializing Multilingual Language Models: An Empirical Study

ethch18/specializing-multilingual EMNLP (MRL) 2021

Pretrained multilingual language models have become a common tool in transferring NLP capabilities to low-resource languages, often with adaptations.

Improving Word Translation via Two-Stage Contrastive Learning

cambridgeltl/contrastivebli ACL ARR November 2021

As Stage C1, we propose to refine standard cross-lingual linear maps between static word embeddings (WEs) via a contrastive learning objective; we also show how to integrate it into the self-learning procedure for even more refined cross-lingual maps.

Improving Word Translation via Two-Stage Contrastive Learning

cambridgeltl/contrastivebli ACL 2022

At Stage C1, we propose to refine standard cross-lingual linear maps between static word embeddings (WEs) via a contrastive learning objective; we also show how to integrate it into the self-learning procedure for even more refined cross-lingual maps.

To Adapt or to Fine-tune: A Case Study on Abstractive Summarization

zsquaredz/adapt_vs_finetune CCL 2022

Recent advances in the field of abstractive summarization leverage pre-trained language models rather than train a model from scratch.

Robustification of Multilingual Language Models to Real-world Noise in Crosslingual Zero-shot Settings with Robust Contrastive Pretraining

amazon-science/multilingual-robust-contrastive-pretraining 10 Oct 2022

To benchmark the performance of pretrained multilingual language models, we construct noisy datasets covering five languages and four NLP tasks and observe a clear gap in the performance between clean and noisy data in the zero-shot cross-lingual setting.

Are Pretrained Multilingual Models Equally Fair Across Languages?

coastalcph/mozart COLING 2022

Pretrained multilingual language models can help bridge the digital language divide, enabling high-quality NLP models for lower resourced languages.

Language Agnostic Multilingual Information Retrieval with Contrastive Learning

xiyanghu/multilingualir 12 Oct 2022

Multilingual information retrieval (IR) is challenging since annotated training data is costly to obtain in many languages.

Improving Bilingual Lexicon Induction with Cross-Encoder Reranking

cambridgeltl/BLICEr 30 Oct 2022

This crucial step is done via 1) creating a word similarity dataset, comprising positive word pairs (i. e., true translations) and hard negative pairs induced from the original CLWE space, and then 2) fine-tuning an mPLM (e. g., mBERT or XLM-R) in a cross-encoder manner to predict the similarity scores.