Cross-Lingual Transfer

291 papers with code • 1 benchmarks • 16 datasets

Cross-lingual transfer refers to transfer learning using data and models available for one language for which ample such resources are available (e.g., English) to solve tasks in another, commonly more low-resource, language.

Libraries

Use these libraries to find Cross-Lingual Transfer models and implementations
2 papers
401
2 papers
128

IRCoder: Intermediate Representations Make Language Models Robust Multilingual Code Generators

UKPLab/arxiv2024-IRCoder 6 Mar 2024

In particular, most mainstream Code-LMs have been pre-trained on source code files alone.

5
06 Mar 2024

Cross-Lingual Learning vs. Low-Resource Fine-Tuning: A Case Study with Fact-Checking in Turkish

firatcekinel/fctr 1 Mar 2024

While misinformation is prevalent in other languages, the majority of research in this field has concentrated on the English language.

0
01 Mar 2024

AdaMergeX: Cross-Lingual Transfer with Large Language Models via Adaptive Adapter Merging

damo-nlp-sg/adamergex 29 Feb 2024

In this paper, we acknowledge the mutual reliance between task ability and language ability and direct our attention toward the gap between the target language and the source language on tasks.

4
29 Feb 2024

Exploring Multilingual Concepts of Human Value in Large Language Models: Is Value Alignment Consistent, Transferable and Controllable across Languages?

shaoyangxu/Multilingual-Human-Value-Concepts 28 Feb 2024

Drawing from our findings on multilingual value alignment, we prudently provide suggestions on the composition of multilingual data for LLMs pre-training: including a limited number of dominant languages for cross-lingual alignment transfer while avoiding their excessive prevalence, and keeping a balanced distribution of non-dominant languages.

1
28 Feb 2024

Investigating Cultural Alignment of Large Language Models

bkhmsi/cultural-trends 20 Feb 2024

The intricate relationship between language and culture has long been a subject of exploration within the realm of linguistic anthropology.

3
20 Feb 2024

LEIA: Facilitating Cross-Lingual Knowledge Transfer in Language Models with Entity-based Data Augmentation

studio-ousia/leia 18 Feb 2024

In this study, we introduce LEIA, a language adaptation tuning method that utilizes Wikipedia entity names aligned across languages.

16
18 Feb 2024

UltraLink: An Open-Source Knowledge-Enhanced Multilingual Supervised Fine-tuning Dataset

openbmb/ultralink 7 Feb 2024

Different from previous works that simply translate English instructions, we consider both the language-specific and language-agnostic abilities of LLMs.

13
07 Feb 2024

Soft Prompt Tuning for Cross-Lingual Transfer: When Less is More

fredxlpy/cross_lingual_prompt_tuning 6 Feb 2024

Soft Prompt Tuning (SPT) is a parameter-efficient method for adapting pre-trained language models (PLMs) to specific tasks by inserting learnable embeddings, or soft prompts, at the input layer of the PLM, without modifying its parameters.

0
06 Feb 2024

Constrained Decoding for Cross-lingual Label Projection

duonglm38/codec 5 Feb 2024

Therefore, it is common to exploit translation and label projection to further improve the performance by (1) translating training data that is available in a high-resource language (e. g., English) together with the gold labels into low-resource languages, and/or (2) translating test data in low-resource languages to a high-source language to run inference on, then projecting the predicted span-level labels back onto the original test data.

1
05 Feb 2024

Translation Errors Significantly Impact Low-Resource Languages in Cross-Lingual Learning

csalt-research/translation-errors-crosslingual-learning 3 Feb 2024

Popular benchmarks (e. g., XNLI) used to evaluate cross-lingual language understanding consist of parallel versions of English evaluation sets in multiple target languages created with the help of professional translators.

0
03 Feb 2024