Search Results for author: Peiqin Lin

Found 13 papers, 9 papers with code

Understanding In-Context Machine Translation for Low-Resource Languages: A Case Study on Manchu

1 code implementation17 Feb 2025 Renhao Pei, Yihong Liu, Peiqin Lin, François Yvon, Hinrich Schütze

In-context machine translation (MT) with large language models (LLMs) is a promising approach for low-resource MT, as it can readily take advantage of linguistic resources such as grammar books and dictionaries.

Data Augmentation In-Context Learning +2

SSMLoRA: Enhancing Low-Rank Adaptation with State Space Model

1 code implementation7 Feb 2025 Jiayang Yu, Yihang Zhang, Bin Wang, Peiqin Lin, Yongkang Liu, Shi Feng

To this end, we propose SSMLoRA (State Space Model Low-Rank Adaptation), an extension of LoRA that incorporates a State Space Model (SSM) to interconnect low-rank matrices.

parameter-efficient fine-tuning

EMMA-500: Enhancing Massively Multilingual Adaptation of Large Language Models

2 code implementations26 Sep 2024 Shaoxiong Ji, Zihao Li, Indraneil Paul, Jaakko Paavola, Peiqin Lin, Pinzhen Chen, Dayyán O'Brien, Hengyu Luo, Hinrich Schütze, Jörg Tiedemann, Barry Haddow

In this work, we introduce EMMA-500, a large-scale multilingual language model continue-trained on texts across 546 languages designed for enhanced multilingual performance, focusing on improving language coverage for low-resource languages.

Cross-Lingual Transfer Language Modeling +1

A Recipe of Parallel Corpora Exploitation for Multilingual Large Language Models

no code implementations29 Jun 2024 Peiqin Lin, André F. T. Martins, Hinrich Schütze

Recent studies have highlighted the potential of exploiting parallel corpora to enhance multilingual large language models, improving performance in both bilingual tasks, e. g., machine translation, and general-purpose tasks, e. g., text classification.

Language Identification Machine Translation +4

XAMPLER: Learning to Retrieve Cross-Lingual In-Context Examples

1 code implementation8 May 2024 Peiqin Lin, André F. T. Martins, Hinrich Schütze

Thus, we introduce XAMPLER: Cross-Lingual Example Retrieval, a method tailored to tackle the challenge of cross-lingual in-context learning using only annotated English data.

In-Context Learning Language Modeling +7

OFA: A Framework of Initializing Unseen Subword Embeddings for Efficient Large-scale Multilingual Continued Pretraining

1 code implementation15 Nov 2023 Yihong Liu, Peiqin Lin, Mingyang Wang, Hinrich Schütze

Instead of pretraining multilingual language models from scratch, a more efficient method is to adapt existing pretrained language models (PLMs) to new languages via vocabulary extension and continued pretraining.

Language Modelling Multilingual Word Embeddings

mPLM-Sim: Better Cross-Lingual Similarity and Transfer in Multilingual Pretrained Language Models

1 code implementation23 May 2023 Peiqin Lin, Chengzhi Hu, Zheyu Zhang, André F. T. Martins, Hinrich Schütze

Recent multilingual pretrained language models (mPLMs) have been shown to encode strong language-specific signals, which are not explicitly provided during pretraining.

Open-Ended Question Answering Zero-Shot Cross-Lingual Transfer

Modeling Content-Emotion Duality via Disentanglement for Empathetic Conversation

1 code implementation26 Sep 2022 Peiqin Lin, Jiashuo Wang, Hinrich Schütze, Wenjie Li

To solve the task, it is essential to model the content-emotion duality of a dialogue, which is composed of the content view (i. e., what personal experiences are described) and the emotion view (i. e., the feelings of the speaker on these experiences).

Disentanglement Empathetic Response Generation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.