1 code implementation • 18 Feb 2025 • Junda Zhu, Lingyong Yan, Shuaiqiang Wang, Dawei Yin, Lei Sha
The reasoning abilities of Large Language Models (LLMs) have demonstrated remarkable advancement and exceptional performance across diverse domains.
1 code implementation • 25 Jan 2025 • Yiqun Chen, Lingyong Yan, Weiwei Sun, Xinyu Ma, Yi Zhang, Shuaiqiang Wang, Dawei Yin, Yiming Yang, Jiaxin Mao
Specifically, we present MMOA-RAG, a Multi-Module joint Optimization Algorithm for RAG, which employs multi-agent reinforcement learning to harmonize all agents' goals towards a unified reward, such as the F1 score of the final answer.
no code implementations • 21 Jan 2025 • Dongsheng Zhu, Weixian Shi, Zhengliang Shi, Zhaochun Ren, Shuaiqiang Wang, Lingyong Yan, Dawei Yin
First, we transform traditional tree-based tool search paths into Directed Acyclic Graph (DAG) structure, generating a high-quality parallel tool invocation dataset.
1 code implementation • 19 Dec 2024 • Jiayi Wu, Hengyi Cai, Lingyong Yan, Hao Sun, Xiang Li, Shuaiqiang Wang, Dawei Yin, Ming Gao
The emergence of Retrieval-augmented generation (RAG) has alleviated the issues of outdated and hallucinatory content in the generation of large language models (LLMs), yet it still reveals numerous limitations.
1 code implementation • 14 Oct 2024 • Weiwei Sun, Zhengliang Shi, Jiulong Wu, Lingyong Yan, Xinyu Ma, Yiding Liu, Min Cao, Dawei Yin, Zhaochun Ren
Recent information retrieval (IR) models are pre-trained and instruction-tuned on massive datasets and tasks, enabling them to perform well on a wide range of tasks and potentially generalize to unseen tasks with instructions.
no code implementations • 10 Oct 2024 • Yougang Lyu, Lingyong Yan, Zihan Wang, Dawei Yin, Pengjie Ren, Maarten de Rijke, Zhaochun Ren
As large language models (LLMs) are rapidly advancing and achieving near-human capabilities, aligning them with human values is becoming more urgent.
no code implementations • 8 Oct 2024 • Bolei He, Nuo Chen, Xinran He, Lingyong Yan, Zhenkai Wei, Jinchang Luo, Zhen-Hua Ling
To address these issues, we propose the chain-of-verification (CoV-RAG) to enhance the external retrieval correctness and internal generation consistency.
1 code implementation • 28 May 2024 • Junda Zhu, Lingyong Yan, Haibo Shi, Dawei Yin, Lei Sha
Large language models (LLMs) are proven to benefit a lot from retrieval-augmented generation (RAG) in alleviating hallucinations confronted with knowledge-intensive questions.
no code implementations • 26 May 2024 • Zhengliang Shi, Shen Gao, Xiuyi Chen, Yue Feng, Lingyong Yan, Haibo Shi, Dawei Yin, Zhumin Chen, Suzan Verberne, Zhaochun Ren
Augmenting large language models (LLMs) with external tools has emerged as a promising approach to extend their utility, empowering them to solve practical tasks.
no code implementations • 1 May 2024 • Guanying Jiang, Lingyong Yan, Haibo Shi, Dawei Yin
Large language model alignment is widely used and studied to avoid LLM producing unhelpful and harmful responses.
no code implementations • 21 Mar 2024 • Yukun Zhao, Lingyong Yan, Weiwei Sun, Guoliang Xing, Shuaiqiang Wang, Chong Meng, Zhicong Cheng, Zhaochun Ren, Dawei Yin
The training process is accomplished by self-rewards inferred from the trained model at the first stage without referring to external human preference resources.
2 code implementations • 5 Mar 2024 • Zhengliang Shi, Shen Gao, Xiuyi Chen, Yue Feng, Lingyong Yan, Haibo Shi, Dawei Yin, Pengjie Ren, Suzan Verberne, Zhaochun Ren
To mitigate these problems, we propose ConAgents, a Cooperative and interactive Agents framework, which coordinates three specialized agents for tool selection, tool execution, and action calibration separately.
2 code implementations • 17 Feb 2024 • Yougang Lyu, Lingyong Yan, Shuaiqiang Wang, Haibo Shi, Dawei Yin, Pengjie Ren, Zhumin Chen, Maarten de Rijke, Zhaochun Ren
To address these problems, we propose a knowledge-aware fine-tuning (KnowTuning) method to improve fine-grained and coarse-grained knowledge awareness of LLMs.
1 code implementation • 2 Nov 2023 • Weiwei Sun, Zheng Chen, Xinyu Ma, Lingyong Yan, Shuaiqiang Wang, Pengjie Ren, Zhumin Chen, Dawei Yin, Zhaochun Ren
Furthermore, our approach surpasses the performance of existing supervised methods like monoT5 and is on par with the state-of-the-art zero-shot methods.
no code implementations • 27 Oct 2023 • Yukun Zhao, Lingyong Yan, Weiwei Sun, Guoliang Xing, Chong Meng, Shuaiqiang Wang, Zhicong Cheng, Zhaochun Ren, Dawei Yin
In this paper, we propose a novel self-detection method to detect which questions that a LLM does not know that are prone to generate nonfactual results.
no code implementations • 25 Oct 2023 • Yukun Zhao, Lingyong Yan, Weiwei Sun, Chong Meng, Shuaiqiang Wang, Zhicong Cheng, Zhaochun Ren, Dawei Yin
Dialogue assessment plays a critical role in the development of open-domain dialogue systems.
1 code implementation • 19 Apr 2023 • Weiwei Sun, Lingyong Yan, Xinyu Ma, Shuaiqiang Wang, Pengjie Ren, Zhumin Chen, Dawei Yin, Zhaochun Ren
In this paper, we first investigate generative LLMs such as ChatGPT and GPT-4 for relevance ranking in IR.
1 code implementation • EMNLP 2021 • Lingyong Yan, Xianpei Han, Le Sun
Bootstrapping has become the mainstream method for entity set expansion.
no code implementations • ACL 2021 • Fangchao Liu, Lingyong Yan, Hongyu Lin, Xianpei Han, Le Sun
Open relation extraction aims to cluster relation instances referring to the same underlying relation, which is a critical step for general relation extraction.
1 code implementation • ACL 2021 • Boxi Cao, Hongyu Lin, Xianpei Han, Le Sun, Lingyong Yan, Meng Liao, Tong Xue, Jin Xu
Previous literatures show that pre-trained masked language models (MLMs) such as BERT can achieve competitive factual knowledge extraction performance on some datasets, indicating that MLMs can potentially be a reliable knowledge source.
1 code implementation • 8 Dec 2020 • Lingyong Yan, Xianpei Han, Le Sun, Fangchao Liu, Ning Bian
By re-organizing all sentences about an entity as a document and extracting relations via querying the document with relation-specific questions, the document-based DS paradigm can simultaneously encode and exploit all sentence-level, inter-sentence-level, and entity-level evidence.
Ranked #1 on
Relationship Extraction (Distant Supervised)
on NYT
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Lingyong Yan, Xianpei Han, Ben He, Le Sun
Bootstrapping for entity set expansion (ESE) has been studied for a long period, which expands new entities using only a few seed entities as supervision.
no code implementations • IJCNLP 2019 • Lingyong Yan, Xianpei Han, Le Sun, Ben He
Bootstrapping for Entity Set Expansion (ESE) aims at iteratively acquiring new instances of a specific target category.