no code implementations • 8 Apr 2024 • Shen Gao, Hao Li, Zhengliang Shi, Chengrui Huang, Quan Tu, Zhiliang Tian, Minlie Huang, Shuo Shang
The framework employs a novel 360{\deg} performance assessment method for multi-perspective performance evaluation with fine-grained assessment.
no code implementations • 5 Mar 2024 • Zhengliang Shi, Shen Gao, Xiuyi Chen, Lingyong Yan, Haibo Shi, Dawei Yin, Zhumin Chen, Pengjie Ren, Suzan Verberne, Zhaochun Ren
Tool learning empowers large language models (LLMs) as agents to use external tools to extend their capability.
no code implementations • 15 Sep 2023 • Zhengliang Shi, Weiwei Sun, Shuo Zhang, Zhen Zhang, Pengjie Ren, Zhaochun Ren
To this end, we propose the Reference-Assisted Dialogue Evaluation (RADE) approach under the multi-task learning framework, which leverages the pre-created utterance as reference other than the gold response to relief the one-to-many problem.
1 code implementation • 27 Aug 2023 • Shen Gao, Zhengliang Shi, Minghang Zhu, Bowen Fang, Xin Xin, Pengjie Ren, Zhumin Chen, Jun Ma, Zhaochun Ren
Augmenting large language models (LLMs) with external tools has emerged as a promising approach to extending the capability of LLMs.
1 code implementation • 20 Dec 2022 • Weiwei Sun, Zhengliang Shi, Shen Gao, Pengjie Ren, Maarten de Rijke, Zhaochun Ren
MixCL effectively reduces the hallucination of LMs in conversations and achieves the highest performance among LM-based dialogue agents in terms of relevancy and factuality.