Search Results for author: Kaixin Wu

Found 7 papers, 2 papers with code

Speeding up Transformer Decoding via an Attention Refinement Network

1 code implementation COLING 2022 Kaixin Wu, Yue Zhang, Bojie Hu, Tong Zhang

Extensive experiments on ten WMT machine translation tasks show that the proposed model yields an average of 1. 35x faster (with almost no decrease in BLEU) over the state-of-the-art inference implementation.

Machine Translation NMT +1

Alleviating LLM-based Generative Retrieval Hallucination in Alipay Search

no code implementations27 Mar 2025 Yedan Shen, Kaixin Wu, Yuechen Ding, Jingyuan Wen, Hong Liu, Mingjie Zhong, Zhouhan Lin, Jia Xu, Linjian Mo

Generative retrieval (GR) has revolutionized document retrieval with the advent of large language models (LLMs), and LLM-based GR is gradually being adopted by the industry.

Hallucination Knowledge Distillation +1

Test-time Computing: from System-1 Thinking to System-2 Thinking

1 code implementation5 Jan 2025 Yixin Ji, Juntao Li, Hai Ye, Kaixin Wu, Jia Xu, Linjian Mo, Min Zhang

In System-2 models, it enhances the model's reasoning ability to solve complex problems through repeated sampling, self-correction, and tree search.

Boosting LLM-based Relevance Modeling with Distribution-Aware Robust Learning

no code implementations17 Dec 2024 Hong Liu, Saisai Gong, Yixin Ji, Kaixin Wu, Jia Xu, Jinjie Gu

In this paper, we propose a novel Distribution-Aware Robust Learning framework (DaRL) for relevance modeling in Alipay Search.

Out of Distribution (OOD) Detection

CPRM: A LLM-based Continual Pre-training Framework for Relevance Modeling in Commercial Search

no code implementations2 Dec 2024 Kaixin Wu, Yixin Ji, Zeyuan Chen, Qiang Wang, Cunxiang Wang, Hong Liu, Baijun Ji, Jia Xu, Zhongyi Liu, Jinjie Gu, Yuan Zhou, Linjian Mo

Our CPRM framework includes three modules: 1) employing both queries and multi-field item to jointly pre-train for enhancing domain knowledge, 2) applying in-context pre-training, a novel approach where LLMs are pre-trained on a sequence of related queries or items, and 3) conducting reading comprehension on items to produce associated domain knowledge and background information (e. g., generating summaries and corresponding queries) to further strengthen LLMs.

In-Context Learning Reading Comprehension

Towards Boosting LLMs-driven Relevance Modeling with Progressive Retrieved Behavior-augmented Prompting

no code implementations18 Aug 2024 Zeyuan Chen, Haiyan Wu, Kaixin Wu, Wei Chen, Mingjie Zhong, Jia Xu, Zhongyi Liu, Wei zhang

In response, we propose ProRBP, a novel Progressive Retrieved Behavior-augmented Prompting framework for integrating search scenario-oriented knowledge with LLMs effectively.

Cannot find the paper you are looking for? You can Submit a new open access paper.