no code implementations • 3 May 2024 • Yaoyiran Li, Xiang Zhai, Moustafa Alzantot, Keyi Yu, Ivan Vulić, Anna Korhonen, Mohamed Hammad
Building upon the success of Large Language Models (LLMs) in a variety of tasks, researchers have recently explored using LLMs that are pretrained on vast corpora of text for sequential recommendation.
1 code implementation • 15 Feb 2024 • Yaoyiran Li, Anna Korhonen, Ivan Vulić
Recent work has shown that, while large language models (LLMs) demonstrate strong word translation or bilingual lexicon induction (BLI) capabilities in few-shot setups, they still cannot match the performance of 'traditional' mapping-based approaches in the unsupervised scenario where no seed translation pairs are available, especially for lower-resource languages.
Bilingual Lexicon Induction Cross-Lingual Word Embeddings +9
1 code implementation • 21 Oct 2023 • Yaoyiran Li, Anna Korhonen, Ivan Vulić
Bilingual Lexicon Induction (BLI) is a core task in multilingual NLP that still, to a large extent, relies on calculating cross-lingual word representations.
Bilingual Lexicon Induction Cross-Lingual Word Embeddings +8
no code implementations • 30 May 2023 • Yaoyiran Li, Ching-Yun Chang, Stephen Rawls, Ivan Vulić, Anna Korhonen
Research on text-to-image generation (TTI) still predominantly focuses on the English language due to the lack of annotated image-caption data in other languages; in the long run, this might widen inequitable access to TTI technology.
Cross-lingual Text-to-Image Generation Crosslingual Text-to-Image Generation +6
1 code implementation • 30 Oct 2022 • Yaoyiran Li, Fangyu Liu, Ivan Vulić, Anna Korhonen
This crucial step is done via 1) creating a word similarity dataset, comprising positive word pairs (i. e., true translations) and hard negative pairs induced from the original CLWE space, and then 2) fine-tuning an mPLM (e. g., mBERT or XLM-R) in a cross-encoder manner to predict the similarity scores.
Bilingual Lexicon Induction Cross-Lingual Word Embeddings +7
1 code implementation • ACL 2022 • Yaoyiran Li, Fangyu Liu, Nigel Collier, Anna Korhonen, Ivan Vulić
At Stage C1, we propose to refine standard cross-lingual linear maps between static word embeddings (WEs) via a contrastive learning objective; we also show how to integrate it into the self-learning procedure for even more refined cross-lingual maps.
1 code implementation • COLING 2020 • Yaoyiran Li, Edoardo M. Ponti, Ivan Vulić, Anna Korhonen
On the other hand, this also provides an extrinsic evaluation protocol to probe the properties of emergent languages ex vitro.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Yaoyiran Li, Jing Jiang
This paper presents some preliminary investigations of a new co-attention mechanism in neural transduction models.