Search Results for author: Minghao Wu

Found 22 papers, 8 papers with code

Bridging the Language Gaps in Large Language Models with Inference-Time Cross-Lingual Intervention

1 code implementation16 Oct 2024 Weixuan Wang, Minghao Wu, Barry Haddow, Alexandra Birch

Large Language Models (LLMs) have shown remarkable capabilities in natural language processing but exhibit significant performance gaps among different languages.

The Best of Both Worlds: Bridging Quality and Diversity in Data Selection with Bipartite Graph

no code implementations16 Oct 2024 Minghao Wu, Thuy-Trang Vu, Lizhen Qu, Gholamreza Haffari

In this paper, we introduce GraphFilter, a novel method that represents the dataset as a bipartite graph, linking sentences to their constituent n-grams.

Computational Efficiency Diversity

Mixture-of-Skills: Learning to Optimize Data Usage for Fine-Tuning Large Language Models

no code implementations13 Jun 2024 Minghao Wu, Thuy-Trang Vu, Lizhen Qu, Gholamreza Haffari

In this work, we propose a general, model-agnostic, reinforcement learning framework, Mixture-of-Skills (MoS), that learns to optimize data usage automatically during the fine-tuning process.

Sharing Matters: Analysing Neurons Across Languages and Tasks in LLMs

1 code implementation13 Jun 2024 Weixuan Wang, Barry Haddow, Minghao Wu, Wei Peng, Alexandra Birch

In this study, we aim to fill the research gap by examining how neuron activation is shared across tasks and languages.

Beyond Probabilities: Unveiling the Misalignment in Evaluating Large Language Models

no code implementations21 Feb 2024 Chenyang Lyu, Minghao Wu, Alham Fikri Aji

Large Language Models (LLMs) have demonstrated remarkable capabilities across various applications, fundamentally reshaping the landscape of natural language processing (NLP) research.

Multiple-choice

Importance-Aware Data Augmentation for Document-Level Neural Machine Translation

no code implementations27 Jan 2024 Minghao Wu, YuFei Wang, George Foster, Lizhen Qu, Gholamreza Haffari

Document-level neural machine translation (DocNMT) aims to generate translations that are both coherent and cohesive, in contrast to its sentence-level counterpart.

Data Augmentation Machine Translation +2

Adapting Large Language Models for Document-Level Machine Translation

no code implementations12 Jan 2024 Minghao Wu, Thuy-Trang Vu, Lizhen Qu, George Foster, Gholamreza Haffari

We provide an in-depth analysis of these LLMs tailored for DocMT, examining translation errors, discourse phenomena, strategies for training and inference, the data efficiency of parallel documents, recent test set evaluations, and zero-shot crosslingual transfer.

Document Level Machine Translation Domain Generalization +2

GPT4Video: A Unified Multimodal Large Language Model for lnstruction-Followed Understanding and Safety-Aware Generation

no code implementations25 Nov 2023 Zhanyu Wang, Longyue Wang, Zhen Zhao, Minghao Wu, Chenyang Lyu, Huayang Li, Deng Cai, Luping Zhou, Shuming Shi, Zhaopeng Tu

While the recent advances in Multimodal Large Language Models (MLLMs) constitute a significant leap forward in the field, these models are predominantly confined to the realm of input-side multimodal comprehension, lacking the capacity for multimodal content generation.

Instruction Following Language Modelling +8

Style Over Substance: Evaluation Biases for Large Language Models

no code implementations6 Jul 2023 Minghao Wu, Alham Fikri Aji

This study investigates the behavior of crowd-sourced and expert annotators, as well as LLMs, when comparing outputs from different models.

Text Generation

Macaw-LLM: Multi-Modal Language Modeling with Image, Audio, Video, and Text Integration

1 code implementation15 Jun 2023 Chenyang Lyu, Minghao Wu, Longyue Wang, Xinting Huang, Bingshuai Liu, Zefeng Du, Shuming Shi, Zhaopeng Tu

Although instruction-tuned large language models (LLMs) have exhibited remarkable capabilities across various NLP tasks, their effectiveness on other data modalities beyond text has not been fully studied.

Language Modelling

Bactrian-X: Multilingual Replicable Instruction-Following Models with Low-Rank Adaptation

1 code implementation24 May 2023 Haonan Li, Fajri Koto, Minghao Wu, Alham Fikri Aji, Timothy Baldwin

However, research on multilingual instruction tuning has been limited due to the scarcity of high-quality instruction-response datasets across different languages.

Instruction Following

A Paradigm Shift: The Future of Machine Translation Lies with Large Language Models

no code implementations2 May 2023 Chenyang Lyu, Zefeng Du, Jitao Xu, Yitao Duan, Minghao Wu, Teresa Lynn, Alham Fikri Aji, Derek F. Wong, Siyou Liu, Longyue Wang

We conclude by emphasizing the critical role of LLMs in guiding the future evolution of MT and offer a roadmap for future exploration in the sector.

Document Translation Machine Translation +2

Document Flattening: Beyond Concatenating Context for Document-Level Neural Machine Translation

no code implementations16 Feb 2023 Minghao Wu, George Foster, Lizhen Qu, Gholamreza Haffari

Existing work in document-level neural machine translation commonly concatenates several consecutive sentences as a pseudo-document, and then learns inter-sentential dependencies.

Machine Translation Translation

Uncertainty-Aware Balancing for Multilingual and Multi-Domain Neural Machine Translation Training

no code implementations EMNLP 2021 Minghao Wu, Yitong Li, Meng Zhang, Liangyou Li, Gholamreza Haffari, Qun Liu

In this work, we propose an approach, MultiUAT, that dynamically adjusts the training data usage based on the model's uncertainty on a small set of trusted clean data for multi-corpus machine translation.

Machine Translation Translation

Evaluating the Utility of Hand-crafted Features in Sequence Labelling

1 code implementation EMNLP 2018 Minghao Wu, Fei Liu, Trevor Cohn

Conventional wisdom is that hand-crafted features are redundant for deep learning models, as they already learn adequate representations of text automatically from corpora.

named-entity-recognition Named Entity Recognition +1

Cannot find the paper you are looking for? You can Submit a new open access paper.