Search Results for author: Lingyong Yan

Found 23 papers, 13 papers with code

Reasoning-to-Defend: Safety-Aware Reasoning Can Defend Large Language Models from Jailbreaking

1 code implementation18 Feb 2025 Junda Zhu, Lingyong Yan, Shuaiqiang Wang, Dawei Yin, Lei Sha

The reasoning abilities of Large Language Models (LLMs) have demonstrated remarkable advancement and exceptional performance across diverse domains.

Improving Retrieval-Augmented Generation through Multi-Agent Reinforcement Learning

1 code implementation25 Jan 2025 Yiqun Chen, Lingyong Yan, Weiwei Sun, Xinyu Ma, Yi Zhang, Shuaiqiang Wang, Dawei Yin, Yiming Yang, Jiaxin Mao

Specifically, we present MMOA-RAG, a Multi-Module joint Optimization Algorithm for RAG, which employs multi-agent reinforcement learning to harmonize all agents' goals towards a unified reward, such as the F1 score of the final answer.

Answer Generation Multi-agent Reinforcement Learning +6

Divide-Then-Aggregate: An Efficient Tool Learning Method via Parallel Tool Invocation

no code implementations21 Jan 2025 Dongsheng Zhu, Weixian Shi, Zhengliang Shi, Zhaochun Ren, Shuaiqiang Wang, Lingyong Yan, Dawei Yin

First, we transform traditional tree-based tool search paths into Directed Acyclic Graph (DAG) structure, generating a high-quality parallel tool invocation dataset.

Task Planning

PA-RAG: RAG Alignment via Multi-Perspective Preference Optimization

1 code implementation19 Dec 2024 Jiayi Wu, Hengyi Cai, Lingyong Yan, Hao Sun, Xiang Li, Shuaiqiang Wang, Dawei Yin, Ming Gao

The emergence of Retrieval-augmented generation (RAG) has alleviated the issues of outdated and hallucinatory content in the generation of large language models (LLMs), yet it still reveals numerous limitations.

Informativeness RAG +1

MAIR: A Massive Benchmark for Evaluating Instructed Retrieval

1 code implementation14 Oct 2024 Weiwei Sun, Zhengliang Shi, Jiulong Wu, Lingyong Yan, Xinyu Ma, Yiding Liu, Min Cao, Dawei Yin, Zhaochun Ren

Recent information retrieval (IR) models are pre-trained and instruction-tuned on massive datasets and tasks, enabling them to perform well on a wide range of tasks and potentially generalize to unseen tasks with instructions.

Information Retrieval Re-Ranking +1

MACPO: Weak-to-Strong Alignment via Multi-Agent Contrastive Preference Optimization

no code implementations10 Oct 2024 Yougang Lyu, Lingyong Yan, Zihan Wang, Dawei Yin, Pengjie Ren, Maarten de Rijke, Zhaochun Ren

As large language models (LLMs) are rapidly advancing and achieving near-human capabilities, aligning them with human values is becoming more urgent.

Retrieving, Rethinking and Revising: The Chain-of-Verification Can Improve Retrieval Augmented Generation

no code implementations8 Oct 2024 Bolei He, Nuo Chen, Xinran He, Lingyong Yan, Zhenkai Wei, Jinchang Luo, Zhen-Hua Ling

To address these issues, we propose the chain-of-verification (CoV-RAG) to enhance the external retrieval correctness and internal generation consistency.

Language Modeling Language Modelling +2

ATM: Adversarial Tuning Multi-agent System Makes a Robust Retrieval-Augmented Generator

1 code implementation28 May 2024 Junda Zhu, Lingyong Yan, Haibo Shi, Dawei Yin, Lei Sha

Large language models (LLMs) are proven to benefit a lot from retrieval-augmented generation (RAG) in alleviating hallucinations confronted with knowledge-intensive questions.

Information Retrieval Language Modelling +3

Chain of Tools: Large Language Model is an Automatic Multi-tool Learner

no code implementations26 May 2024 Zhengliang Shi, Shen Gao, Xiuyi Chen, Yue Feng, Lingyong Yan, Haibo Shi, Dawei Yin, Zhumin Chen, Suzan Verberne, Zhaochun Ren

Augmenting large language models (LLMs) with external tools has emerged as a promising approach to extend their utility, empowering them to solve practical tasks.

Language Modeling Language Modelling +1

The Real, the Better: Aligning Large Language Models with Online Human Behaviors

no code implementations1 May 2024 Guanying Jiang, Lingyong Yan, Haibo Shi, Dawei Yin

Large language model alignment is widely used and studied to avoid LLM producing unhelpful and harmful responses.

Language Modeling Language Modelling +1

Improving the Robustness of Large Language Models via Consistency Alignment

no code implementations21 Mar 2024 Yukun Zhao, Lingyong Yan, Weiwei Sun, Guoliang Xing, Shuaiqiang Wang, Chong Meng, Zhicong Cheng, Zhaochun Ren, Dawei Yin

The training process is accomplished by self-rewards inferred from the trained model at the first stage without referring to external human preference resources.

Diversity Instruction Following +1

Learning to Use Tools via Cooperative and Interactive Agents

2 code implementations5 Mar 2024 Zhengliang Shi, Shen Gao, Xiuyi Chen, Yue Feng, Lingyong Yan, Haibo Shi, Dawei Yin, Pengjie Ren, Suzan Verberne, Zhaochun Ren

To mitigate these problems, we propose ConAgents, a Cooperative and interactive Agents framework, which coordinates three specialized agents for tool selection, tool execution, and action calibration separately.

KnowTuning: Knowledge-aware Fine-tuning for Large Language Models

2 code implementations17 Feb 2024 Yougang Lyu, Lingyong Yan, Shuaiqiang Wang, Haibo Shi, Dawei Yin, Pengjie Ren, Zhumin Chen, Maarten de Rijke, Zhaochun Ren

To address these problems, we propose a knowledge-aware fine-tuning (KnowTuning) method to improve fine-grained and coarse-grained knowledge awareness of LLMs.

Question Answering

Instruction Distillation Makes Large Language Models Efficient Zero-shot Rankers

1 code implementation2 Nov 2023 Weiwei Sun, Zheng Chen, Xinyu Ma, Lingyong Yan, Shuaiqiang Wang, Pengjie Ren, Zhumin Chen, Dawei Yin, Zhaochun Ren

Furthermore, our approach surpasses the performance of existing supervised methods like monoT5 and is on par with the state-of-the-art zero-shot methods.

Prompt Engineering

Knowing What LLMs DO NOT Know: A Simple Yet Effective Self-Detection Method

no code implementations27 Oct 2023 Yukun Zhao, Lingyong Yan, Weiwei Sun, Guoliang Xing, Chong Meng, Shuaiqiang Wang, Zhicong Cheng, Zhaochun Ren, Dawei Yin

In this paper, we propose a novel self-detection method to detect which questions that a LLM does not know that are prone to generate nonfactual results.

Element Intervention for Open Relation Extraction

no code implementations ACL 2021 Fangchao Liu, Lingyong Yan, Hongyu Lin, Xianpei Han, Le Sun

Open relation extraction aims to cluster relation instances referring to the same underlying relation, which is a critical step for general relation extraction.

Relation Relation Extraction

Knowledgeable or Educated Guess? Revisiting Language Models as Knowledge Bases

1 code implementation ACL 2021 Boxi Cao, Hongyu Lin, Xianpei Han, Le Sun, Lingyong Yan, Meng Liao, Tong Xue, Jin Xu

Previous literatures show that pre-trained masked language models (MLMs) such as BERT can achieve competitive factual knowledge extraction performance on some datasets, indicating that MLMs can potentially be a reliable knowledge source.

From Bag of Sentences to Document: Distantly Supervised Relation Extraction via Machine Reading Comprehension

1 code implementation8 Dec 2020 Lingyong Yan, Xianpei Han, Le Sun, Fangchao Liu, Ning Bian

By re-organizing all sentences about an entity as a document and extracting relations via querying the document with relation-specific questions, the document-based DS paradigm can simultaneously encode and exploit all sentence-level, inter-sentence-level, and entity-level evidence.

Denoising Machine Reading Comprehension +3

Global Bootstrapping Neural Network for Entity Set Expansion

1 code implementation Findings of the Association for Computational Linguistics 2020 Lingyong Yan, Xianpei Han, Ben He, Le Sun

Bootstrapping for entity set expansion (ESE) has been studied for a long period, which expands new entities using only a few seed entities as supervision.

Decoder

Learning to Bootstrap for Entity Set Expansion

no code implementations IJCNLP 2019 Lingyong Yan, Xianpei Han, Le Sun, Ben He

Bootstrapping for Entity Set Expansion (ESE) aims at iteratively acquiring new instances of a specific target category.

Cannot find the paper you are looking for? You can Submit a new open access paper.