Search Results for author: Xunliang Cai

Found 14 papers, 3 papers with code

Not All Contexts Are Equal: Teaching LLMs Credibility-aware Generation

2 code implementations10 Apr 2024 Ruotong Pan, Boxi Cao, Hongyu Lin, Xianpei Han, Jia Zheng, Sirui Wang, Xunliang Cai, Le Sun

In this paper, we propose Credibility-aware Generation (CAG), a universally applicable framework designed to mitigate the impact of flawed information in RAG.

Retrieval

What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation

no code implementations11 Mar 2024 Zhuocheng Gong, Jiahao Liu, Jingang Wang, Xunliang Cai, Dongyan Zhao, Rui Yan

Our findings reveal several connections between the properties of perturbations and LLM performance, providing insights into the failure cases of uniform quantization and suggesting potential solutions to improve the robustness of LLM quantization.

Computational Efficiency Quantization

Unraveling the Mystery of Scaling Laws: Part I

no code implementations11 Mar 2024 Hui Su, Zhi Tian, Xiaoyu Shen, Xunliang Cai

However, the original scaling law paper by OpenAI did not disclose the complete details necessary to derive the precise scaling law formulas, and their conclusions are only based on models containing up to 1. 5 billion parameters.

Beyond the Known: Investigating LLMs Performance on Out-of-Domain Intent Detection

no code implementations27 Feb 2024 Pei Wang, Keqing He, Yejie Wang, Xiaoshuai Song, Yutao Mou, Jingang Wang, Yunsen Xian, Xunliang Cai, Weiran Xu

Out-of-domain (OOD) intent detection aims to examine whether the user's query falls outside the predefined domain of the system, which is crucial for the proper functioning of task-oriented dialogue (TOD) systems.

Intent Detection Transfer Learning

Improving Input-label Mapping with Demonstration Replay for In-context Learning

no code implementations30 Oct 2023 Zhuocheng Gong, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Rui Yan

The effectiveness of ICL can be attributed to the strong language modeling capabilities of large language models (LLMs), which enable them to learn the mapping between input and labels based on in-context demonstrations.

In-Context Learning Language Modelling

Retrieval-based Knowledge Transfer: An Effective Approach for Extreme Large Language Model Compression

no code implementations24 Oct 2023 Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Ran Lucien Wang, Rui Yan

In particular, our approach extracts knowledge from LLMs to construct a knowledge store, from which the small-scale model can retrieve relevant information and leverage it for effective inference.

Language Modelling Large Language Model +3

APP: Adaptive Prototypical Pseudo-Labeling for Few-shot OOD Detection

no code implementations20 Oct 2023 Pei Wang, Keqing He, Yutao Mou, Xiaoshuai Song, Yanan Wu, Jingang Wang, Yunsen Xian, Xunliang Cai, Weiran Xu

Detecting out-of-domain (OOD) intents from user queries is essential for a task-oriented dialogue system.

Large Language Models Meet Open-World Intent Discovery and Recognition: An Evaluation of ChatGPT

1 code implementation16 Oct 2023 Xiaoshuai Song, Keqing He, Pei Wang, Guanting Dong, Yutao Mou, Jingang Wang, Yunsen Xian, Xunliang Cai, Weiran Xu

The tasks of out-of-domain (OOD) intent discovery and generalized intent discovery (GID) aim to extend a closed intent classifier to open-world intent sets, which is crucial to task-oriented dialogue (TOD) systems.

In-Context Learning Intent Discovery

Confidence Calibration for Intent Detection via Hyperspherical Space and Rebalanced Accuracy-Uncertainty Loss

no code implementations17 Mar 2022 Yantao Gong, Cao Liu, Fan Yang, Xunliang Cai, Guanglu Wan, Jiansong Chen, Weipeng Zhang, Houfeng Wang

Experiments on the open datasets verify that our model outperforms the existing calibration methods and achieves a significant improvement on the calibration metric.

Intent Detection

From Paraphrasing to Semantic Parsing: Unsupervised Semantic Parsing via Synchronous Semantic Decoding

no code implementations ACL 2021 Shan Wu, Bo Chen, Chunlei Xin, Xianpei Han, Le Sun, Weipeng Zhang, Jiansong Chen, Fan Yang, Xunliang Cai

During synchronous decoding: the utterance paraphrasing is constrained by the structure of the logical form, therefore the canonical utterance can be paraphrased controlledly; the semantic decoding is guided by the semantics of the canonical utterance, therefore its logical form can be generated unsupervisedly.

Unsupervised semantic parsing

Cannot find the paper you are looking for? You can Submit a new open access paper.