1 code implementation • ACL 2022 • Zichu Fei, Qi Zhang, Tao Gui, Di Liang, Sirui Wang, Wei Wu, Xuanjing Huang
CQG employs a simple method to generate the multi-hop questions that contain key entities in multi-hop reasoning chains, which ensure the complexity and quality of the questions.
no code implementations • EMNLP 2020 • Hongzhi Zhang, Yingyao Wang, Sirui Wang, Xuezhi Cao, Fuzheng Zhang, Zhongyuan Wang
Verifying fact on semi-structured evidence like tables requires the ability to encode structural information and perform symbolic reasoning.
1 code implementation • EMNLP 2021 • Yuanmeng Yan, Rumei Li, Sirui Wang, Hongzhi Zhang, Zan Daoguang, Fuzheng Zhang, Wei Wu, Weiran Xu
The key challenge of question answering over knowledge bases (KBQA) is the inconsistency between the natural language questions and the reasoning paths in the knowledge base (KB).
1 code implementation • 22 Apr 2024 • Keheng Wang, Feiyu Duan, Peiguang Li, Sirui Wang, Xunliang Cai
Retrieval-Augmented Generation (RAG) demonstrates great value in alleviating outdated knowledge or hallucination by supplying LLMs with updated and relevant knowledge.
1 code implementation • 10 Apr 2024 • Ruotong Pan, Boxi Cao, Hongyu Lin, Xianpei Han, Jia Zheng, Sirui Wang, Xunliang Cai, Le Sun
In this paper, we propose Credibility-aware Generation (CAG), a universally applicable framework designed to mitigate the impact of flawed information in RAG.
no code implementations • 8 Apr 2024 • Sirui Wang, Peiguang Li, Yunsen Xian, Hongzhi Zhang
The sequential recommendation task aims to predict the item that user is interested in according to his/her historical action sequence.
1 code implementation • 21 Sep 2023 • Shanglin Lei, Guanting Dong, XiaoPing Wang, Keheng Wang, Runqi Qiao, Sirui Wang
The field of emotion recognition of conversation (ERC) has been focusing on separating sentence feature encoding and context modeling, lacking exploration in generative paradigms based on unified designs.
Ranked #3 on Emotion Recognition in Conversation on EmoryNLP
1 code implementation • 28 Aug 2023 • Guanting Dong, Rumei Li, Sirui Wang, Yupeng Zhang, Yunsen Xian, Weiran Xu
Knowledge Base Question Answering (KBQA) aims to answer natural language questions with factual information such as entities and relations in KBs.
Ranked #3 on Knowledge Base Question Answering on WebQuestionsSP
1 code implementation • 25 Aug 2023 • Keheng Wang, Feiyu Duan, Sirui Wang, Peiguang Li, Yunsen Xian, Chuantao Yin, Wenge Rong, Zhang Xiong
Equipped with Chain-of-Thought (CoT), Large language models (LLMs) have shown impressive reasoning ability in various downstream tasks.
1 code implementation • 14 Jun 2023 • Yuntao Li, Zhenpeng Su, Yutian Li, Hanchu Zhang, Sirui Wang, Wei Wu, Yan Zhang
Translating natural language queries into SQLs in a seq2seq manner has attracted much attention recently.
Ranked #11 on Text-To-SQL on spider
no code implementations • 24 Feb 2023 • Chao Xue, Di Liang, Sirui Wang, Wei Wu, Jing Zhang
To alleviate this problem, we propose a novel Dual Path Modeling Framework to enhance the model's ability to perceive subtle differences in sentence pairs by separately modeling affinity and difference semantics.
no code implementations • 24 Feb 2023 • Yonghao Liu, Di Liang, Fang Fang, Sirui Wang, Wei Wu, Rui Jiang
For each given question, TMA first extracts the relevant concepts from the KG, and then feeds them into a multiway adaptive module to produce a \emph{temporal-specific} representation of the question.
2 code implementations • ACL 2022 • Rui Zheng, Rong Bao, Yuhao Zhou, Di Liang, Sirui Wang, Wei Wu, Tao Gui, Qi Zhang, Xuanjing Huang
Recent works on Lottery Ticket Hypothesis have shown that pre-trained language models (PLMs) contain smaller matching subnetworks(winning tickets) which are capable of reaching accuracy comparable to the original models.
no code implementations • 22 Oct 2022 • Yupeng Zhang, Hongzhi Zhang, Sirui Wang, Wei Wu, Zhoujun Li
A wide range of NLP tasks benefit from the fine-tuning of pretrained language models (PLMs).
no code implementations • 16 Oct 2022 • Jian Song, Di Liang, Rumei Li, Yuntao Li, Sirui Wang, Minlong Peng, Wei Wu, Yongxin Yu
Transformer-based pre-trained models like BERT have achieved great progress on Semantic Sentence Matching.
no code implementations • COLING 2022 • Sirui Wang, Di Liang, Jian Song, Yuntao Li, Wei Wu
To alleviate this problem, we propose a novel Dual Attention Enhanced BERT (DABERT) to enhance the ability of BERT to capture fine-grained differences in sentence pairs.
no code implementations • 31 Aug 2022 • Sirui Wang, Kaiwen Wei, Hongzhi Zhang, Yuntao Li, Wei Wu
Inspired by the human learning process, in this paper, we introduce Imitation DEMOnstration Learning (Imitation-Demo) to strengthen demonstration learning via explicitly imitating human review behaviour, which includes: (1) contrastive learning mechanism to concentrate on the similar demonstrations.
1 code implementation • 7 Jun 2022 • Ruotian Ma, Yiding Tan, Xin Zhou, Xuanting Chen, Di Liang, Sirui Wang, Wei Wu, Tao Gui, Qi Zhang
Input distribution shift is one of the vital problems in unsupervised domain adaptation (UDA).
1 code implementation • 8 Mar 2022 • LiWen Wang, Rumei Li, Yang Yan, Yuanmeng Yan, Sirui Wang, Wei Wu, Weiran Xu
Recently, prompt-based methods have achieved significant performance in few-shot learning scenarios by bridging the gap between language model pre-training and fine-tuning for downstream tasks.
1 code implementation • 16 Dec 2021 • Yuntao Li, Hanchu Zhang, Yutian Li, Sirui Wang, Wei Wu, Yan Zhang
Conversational text-to-SQL aims at converting multi-turn natural language queries into their corresponding SQL (Structured Query Language) representations.
Ranked #2 on Text-To-SQL on SParC
1 code implementation • EMNLP 2021 • Kun Zhou, Wayne Xin Zhao, Sirui Wang, Fuzheng Zhang, Wei Wu, Ji-Rong Wen
To solve this issue, various data augmentation techniques are proposed to improve the robustness of PLMs.
1 code implementation • ACL 2021 • Yuanmeng Yan, Rumei Li, Sirui Wang, Fuzheng Zhang, Wei Wu, Weiran Xu
Learning high-quality sentence representations benefits a wide range of natural language processing tasks.
no code implementations • COLING 2020 • Xuemiao Zhang, Kun Zhou, Sirui Wang, Fuzheng Zhang, Zhongyuan Wang, Junfei Liu
Weakly supervised machine reading comprehension (MRC) task is practical and promising for its easily available and massive training data, but inevitablely introduces noise.
no code implementations • 19 Aug 2020 • Kun Zhou, Wayne Xin Zhao, Hui Wang, Sirui Wang, Fuzheng Zhang, Zhongyuan Wang, Ji-Rong Wen
Most of the existing CRS methods focus on learning effective preference representations for users from conversation data alone.
2 code implementations • 18 Aug 2020 • Kun Zhou, Hui Wang, Wayne Xin Zhao, Yutao Zhu, Sirui Wang, Fuzheng Zhang, Zhongyuan Wang, Ji-Rong Wen
To tackle this problem, we propose the model S^3-Rec, which stands for Self-Supervised learning for Sequential Recommendation, based on the self-attentive neural architecture.
no code implementations • 5 Aug 2020 • Qiong Wu, Adam Hare, Sirui Wang, Yuwei Tu, Zhenming Liu, Christopher G. Brinton, Yanhua Li
In this work, we reexamine the inter-related problems of "topic identification" and "text segmentation" for sparse document learning, when there is a single new text of interest.