no code implementations • Findings (ACL) 2022 • Zhuocheng Gong, Di He, Yelong Shen, Tie-Yan Liu, Weizhu Chen, Dongyan Zhao, Ji-Rong Wen, Rui Yan
Empirically, we show that (a) the dominant winning ticket can achieve performance that is comparable with that of the full-parameter model, (b) the dominant winning ticket is transferable across different tasks, (c) and the dominant winning ticket has a natural structure within each parameter matrix.
no code implementations • Findings (EMNLP) 2021 • Qingqing Zhu, Xiuying Chen, Pengfei Wu, Junfei Liu, Dongyan Zhao
Hence, in this paper, we introduce a combination of curriculum learning and knowledge distillation for efficient dialogue generation models, where curriculum learning can help knowledge distillation from data and model aspects.
1 code implementation • ACL 2022 • Chang Liu, Chongyang Tao, Jiazhan Feng, Dongyan Zhao
Transferring the knowledge to a small model through distillation has raised great interest in recent years.
no code implementations • EMNLP 2020 • Zhangming Chan, Yuchi Zhang, Xiuying Chen, Shen Gao, Zhiqiang Zhang, Dongyan Zhao, Rui Yan
(2) generate a post including selected products via the MGenNet (Multi-Generator Network).
no code implementations • EMNLP 2020 • Jizhi Tang, Yansong Feng, Dongyan Zhao
Recent efforts have made great progress to track multiple entities in a procedural text, but usually treat each entity separately and ignore the fact that there are often multiple entities interacting with each other during one process, some of which are even explicitly mentioned.
no code implementations • ACL 2022 • Chang Liu, Xu Tan, Chongyang Tao, Zhenxin Fu, Dongyan Zhao, Tie-Yan Liu, Rui Yan
To enable the chatbot to foresee the dialogue future, we design a beam-search-like roll-out strategy for dialogue future simulation using a typical dialogue generation model and a dialogue selector.
no code implementations • CCL 2020 • Lin Wang, Juntao Li, Rui Yan, Dongyan Zhao
Story generation is a challenging task of automatically creating natural languages to describe a sequence of events, which requires outputting text with not only a consistent topic but also novel wordings.
no code implementations • COLING 2022 • Jiazhan Feng, Chongyang Tao, Zhen Li, Chang Liu, Tao Shen, Dongyan Zhao
In this paper, we propose a reciprocal learning approach to jointly optimize a knowledge retriever and a response ranker for knowledge-grounded response retrieval without ground-truth knowledge labels.
no code implementations • CCL 2020 • Shaoling Jing, Shibo Hong, Dongyan Zhao, Haihua Xie, Zhi Tang
Multi-turn conversational Question Answering (ConvQA) is a practical task that requires the understanding of conversation history, such as previous QA pairs, the passage context, and current question.
no code implementations • CCL 2020 • Nuo Xu, Haihua Xie, Dongyan Zhao
The task of argument extraction is converted to event relation triple extraction.
1 code implementation • 20 Aug 2023 • Quan Tu, Chuanqi Chen, Jinpeng Li, Yanran Li, Shuo Shang, Dongyan Zhao, Ran Wang, Rui Yan
In our modern, fast-paced, and interconnected world, the importance of mental well-being has grown into a matter of great urgency.
no code implementations • 5 Jul 2023 • Hejing Cao, Dongyan Zhao
Experiments on the BEA-2019 shared task and the CoNLL-2014 shared task have shown that AMR-GEC performs comparably to a set of strong baselines with a large number of synthetic data.
1 code implementation • 22 Jun 2023 • Yijia Shao, Yiduo Guo, Dongyan Zhao, Bing Liu
Despite the great success of pre-trained language models, it is still a challenge to use these models for continual learning, especially for the class-incremental learning (CIL) setting due to catastrophic forgetting (CF).
1 code implementation • 7 Jun 2023 • Zhibin Chen, Yansong Feng, Dongyan Zhao
Entailment Graphs (EGs) have been constructed based on extracted corpora as a strong and explainable form to indicate context-independent entailment relations in natural languages.
no code implementations • 4 Jun 2023 • Jianghui Wang, Yuxuan Wang, Dongyan Zhao, Zilong Zheng
We introduce MoviePuzzle, a novel challenge that targets visual narrative reasoning and holistic movie understanding.
1 code implementation • 1 Jun 2023 • Chen Zhang, Jiuheng Lin, Xiao Liu, Yuxuan Lai, Yansong Feng, Dongyan Zhao
We further analyze how well different paradigms of current multi-answer MRC models deal with different types of multi-answer instances.
no code implementations • 30 May 2023 • Zhuocheng Gong, Jiahao Liu, Qifan Wang, Yang Yang, Jingang Wang, Wei Wu, Yunsen Xian, Dongyan Zhao, Rui Yan
While transformer-based pre-trained language models (PLMs) have dominated a number of NLP applications, these models are heavy to deploy and expensive to use.
1 code implementation • 30 May 2023 • Yuxuan Wang, Zilong Zheng, Xueliang Zhao, Jinpeng Li, Yueqian Wang, Dongyan Zhao
Video-grounded dialogue understanding is a challenging problem that requires machine to perceive, parse and reason over situated semantics extracted from weakly aligned video and dialogues.
1 code implementation • 30 May 2023 • Xiao Liu, Da Yin, Chen Zhang, Yansong Feng, Dongyan Zhao
Causal reasoning, the ability to identify cause-and-effect relationship, is crucial in human thinking.
1 code implementation • 30 May 2023 • Yuxuan Wang, Jianghui Wang, Dongyan Zhao, Zilong Zheng
We introduce CDBERT, a new learning paradigm that enhances the semantics understanding ability of the Chinese PLMs with dictionary knowledge and structure of Chinese characters.
no code implementations • 28 May 2023 • Quzhe Huang, Yutong Hu, Shengqi Zhu, Yansong Feng, Chang Liu, Dongyan Zhao
After examining the relation definitions in various ETRE tasks, we observe that all relations can be interpreted using the start and end time points of events.
1 code implementation • 26 May 2023 • Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Wei Wu, Yunsen Xian, Dongyan Zhao, Kai Chen, Rui Yan
In this paper, we propose a novel approach, RankCSE, for unsupervised sentence representation learning, which incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.
1 code implementation • CVPR 2023 • Yiduo Guo, Bing Liu, Dongyan Zhao
A novel optimization objective with a gradient-based adaptive method is proposed to dynamically deal with the problem in the online CL process.
1 code implementation • 19 May 2023 • Xin Cheng, Yankai Lin, Xiuying Chen, Dongyan Zhao, Rui Yan
The key intuition is to decouple the knowledge storage from model parameters with an editable and scalable key-value memory and leverage knowledge in an explainable manner by knowledge retrieval in the DPM.
no code implementations • 19 May 2023 • Yiduo Guo, Yaobo Liang, Dongyan Zhao, Bing Liu, Duan Nan
Existing research has shown that a multilingual pre-trained language model fine-tuned with one (source) language also performs well on downstream tasks for non-source languages, even though no fine-tuning is done on these languages.
1 code implementation • 17 May 2023 • Chenshuo Wang, Shaoguang Mao, Tao Ge, Wenshan Wu, Xun Wang, Yan Xia, Jonathan Tien, Dongyan Zhao
The training dataset comprises over 3. 7 million sentences and 12. 7 million suggestions generated through rules.
1 code implementation • 12 May 2023 • Jiazhan Feng, Chongyang Tao, Xiubo Geng, Tao Shen, Can Xu, Guodong Long, Dongyan Zhao, Daxin Jiang
Information retrieval (IR) plays a crucial role in locating relevant resources from vast amounts of data, and its applications have evolved from traditional knowledge bases to modern search engines (SEs).
no code implementations • 8 May 2023 • Mingxu Tao, Yansong Feng, Dongyan Zhao
Since the embeddings of rear positions are updated fewer times than the front position embeddings, the rear ones may not be properly trained.
1 code implementation • 3 May 2023 • Xin Cheng, Di Luo, Xiuying Chen, Lemao Liu, Dongyan Zhao, Rui Yan
In this paper, by exploring the duality of the primal problem: better generation also prompts better memory, we propose a novel framework, selfmem, which addresses this limitation by iteratively employing a retrieval-augmented generator to create an unbounded memory pool and using a memory selector to choose one output as memory for the subsequent generation round.
no code implementations • 20 Apr 2023 • Yiduo Guo, Yaobo Liang, Chenfei Wu, Wenshan Wu, Dongyan Zhao, Nan Duan
Therefore, we further propose the Learning to Program (\text{LP}) method to ask LLMs themselves to learn the natural language program based on the training dataset of the complex task first and then use the learned program to guide the inference.
1 code implementation • 2 Mar 2023 • Mingxu Tao, Yansong Feng, Dongyan Zhao
Large pre-trained language models help to achieve state of the art on a variety of natural language processing (NLP) tasks, nevertheless, they still suffer from forgetting when incrementally learning a sequence of tasks.
1 code implementation • 26 Feb 2023 • Chen Zhang, Yuxuan Lai, Yansong Feng, Xingyu Shen, Haowei Du, Dongyan Zhao
We convert KB subgraphs into passages to narrow the gap between KB schemas and questions, which enables our model to benefit from recent advances in multilingual pre-trained language models (MPLMs) and cross-lingual machine reading comprehension (xMRC).
Cross-Lingual Question Answering
Machine Reading Comprehension
no code implementations • 27 Jan 2023 • Xin Cheng, Shen Gao, Yuchi Zhang, Yongliang Wang, Xiuying Chen, Mingzhe Li, Dongyan Zhao, Rui Yan
Review summarization is a non-trivial task that aims to summarize the main idea of the product review in the E-commerce website.
no code implementations • 3 Jan 2023 • Mingzhe Li, Xiuying Chen, Weiheng Liao, Yang song, Tao Zhang, Dongyan Zhao, Rui Yan
The key idea is to reduce the number of parameters that rely on interview dialogs by disentangling the knowledge selector and dialog generator so that most parameters can be trained with ungrounded dialogs as well as the resume data that are not low-resource.
1 code implementation • 2 Jan 2023 • Xiuying Chen, Mingzhe Li, Shen Gao, Zhangming Chan, Dongyan Zhao, Xin Gao, Xiangliang Zhang, Rui Yan
Nowadays, time-stamped web documents related to a general news query floods spread throughout the Internet, and timeline summarization targets concisely summarizing the evolution trajectory of events along the timeline.
no code implementations • 20 Dec 2022 • Chang Liu, Chongyang Tao, Xiubo Geng, Tao Shen, Dongyan Zhao, Can Xu, Binxing Jiao, Daxin Jiang
Different from previous works that only rely on one positive and hard negatives as candidate passages, we create dark examples that all have moderate relevance to the query through mixing-up and masking in discrete space.
1 code implementation • 6 Dec 2022 • Xin Cheng, Shen Gao, Lemao Liu, Dongyan Zhao, Rui Yan
Retrieval-augmented Neural Machine Translation models have been successful in many translation scenarios.
1 code implementation • 10 Nov 2022 • Jiazhan Feng, Qingfeng Sun, Can Xu, Pu Zhao, Yaming Yang, Chongyang Tao, Dongyan Zhao, QIngwei Lin
First, it is the largest multi-modal conversation dataset by the number of dialogues by 88x.
Ranked #2 on
Multimodal Intent Recognition
on MMDialog
1 code implementation • 31 Oct 2022 • Zhenwei An, Quzhe Huang, Cong Jiang, Yansong Feng, Dongyan Zhao
The charge prediction task aims to predict the charge for a case given its fact description.
no code implementations • 22 Oct 2022 • Xueliang Zhao, Yuxuan Wang, Chongyang Tao, Chenshuo Wang, Dongyan Zhao
We study video-grounded dialogue generation, where a response is generated based on the dialogue context and the associated video.
no code implementations • 22 Oct 2022 • Xueliang Zhao, Lemao Liu, Tingchen Fu, Shuming Shi, Dongyan Zhao, Rui Yan
With the availability of massive general-domain dialogue data, pre-trained dialogue generation appears to be super appealing to transfer knowledge from the general domain to downstream applications.
1 code implementation • 20 Oct 2022 • Xiao Liu, Yansong Feng, Jizhi Tang, Chengang Hu, Dongyan Zhao
Although pretrained language models can generate fluent recipe texts, they fail to truly learn and use the culinary knowledge in a compositional way.
no code implementations • 7 Sep 2022 • Haowei Du, Quzhe Huang, Chen Zhang, Dongyan Zhao
Multi-hop Knowledge Base Question Answering(KBQA) aims to find the answer entity in a knowledge base which is several hops from the topic entity mentioned in the question.
no code implementations • ACL 2022 • Mingzhe Li, Xiexiong Lin, Xiuying Chen, Jinxiong Chang, Qishen Zhang, Feng Wang, Taifeng Wang, Zhongyi Liu, Wei Chu, Dongyan Zhao, Rui Yan
Contrastive learning has achieved impressive success in generation tasks to militate the "exposure bias" problem and discriminatively exploit the different quality of references.
no code implementations • 18 Apr 2022 • Jiduan Liu, Jiahao Liu, Yang Yang, Jingang Wang, Wei Wu, Dongyan Zhao, Rui Yan
To enhance the performance of dense retrieval models without loss of efficiency, we propose a GNN-encoder model in which query (passage) information is fused into passage (query) representations via graph neural networks that are constructed by queries and their top retrieved passages.
1 code implementation • ACL 2022 • Quzhe Huang, Shibo Hao, Yuan Ye, Shengqi Zhu, Yansong Feng, Dongyan Zhao
DocRED is a widely used dataset for document-level relation extraction.
no code implementations • NAACL 2022 • Xueliang Zhao, Tingchen Fu, Chongyang Tao, Wei Wu, Dongyan Zhao, Rui Yan
Grounding dialogue generation by extra knowledge has shown great potentials towards building a system capable of replying with knowledgeable and engaging responses.
1 code implementation • ACL 2022 • Zhibin Chen, Yansong Feng, Dongyan Zhao
Typed entailment graphs try to learn the entailment relations between predicates from text and model them as edges between predicate nodes.
1 code implementation • ACL 2022 • Xiao Liu, Da Yin, Yansong Feng, Dongyan Zhao
We probe PLMs and models with visual signals, including vision-language pretrained models and image synthesis models, on this benchmark, and find that image synthesis models are more capable of learning accurate and consistent spatial knowledge than other models.
1 code implementation • 27 Dec 2021 • Shen Gao, Yuchi Zhang, Yongliang Wang, Yang Dong, Xiuying Chen, Dongyan Zhao, Rui Yan
Most of the CQA methods only incorporate articles or Wikipedia to extract knowledge and answer the user's question.
1 code implementation • NeurIPS 2021 • Jinpeng Li, Yingce Xia, Rui Yan, Hongda Sun, Dongyan Zhao, Tie-Yan Liu
Considering there is no parallel data between the contexts and the responses of target style S1, existing works mainly use back translation to generate stylized synthetic data for training, where the data about context, target style S1 and an intermediate style S0 is used.
no code implementations • NeurIPS 2021 • Qi Qin, Wenpeng Hu, Han Peng, Dongyan Zhao, Bing Liu
Continual learning (CL) of a sequence of tasks is often accompanied with the catastrophic forgetting(CF) problem.
no code implementations • 29 Sep 2021 • Yiduo Guo, Dongyan Zhao, Bing Liu
Most existing techniques for online continual learning are based on experience-replay.
1 code implementation • Findings (EMNLP) 2021 • Chen Zhang, Yuxuan Lai, Yansong Feng, Dongyan Zhao
In this paper, we present a new verification style reading comprehension dataset named VGaokao from Chinese Language tests of Gaokao.
1 code implementation • ACM Transactions on Information Systems 2021 • Ruijian Xu, Chongyang Tao, Jiazhan Feng, Wei Wu, Rui Yan, Dongyan Zhao
To tackle these challenges, we propose a representation[K]-interaction[L]-matching framework that explores multiple types of deep interactive representations to build context-response matching models for response selection.
1 code implementation • ACL 2021 • Xiuying Chen, Hind Alamro, Mingzhe Li, Shen Gao, Xiangliang Zhang, Dongyan Zhao, Rui Yan
Hence, in this paper, we propose a Relation-aware Related work Generator (RRG), which generates an abstractive related work from the given multiple scientific papers in the same research area.
1 code implementation • ACL 2021 • Quzhe Huang, Shengqi Zhu, Yansong Feng, Yuan Ye, Yuxuan Lai, Dongyan Zhao
Document-level Relation Extraction (RE) is a more challenging task than sentence RE as it often requires reasoning over multiple sentences.
Ranked #48 on
Relation Extraction
on DocRED
no code implementations • ACL 2021 • Quzhe Huang, Shengqi Zhu, Yansong Feng, Dongyan Zhao
Recent studies strive to incorporate various human rationales into neural networks to improve model performance, but few pay attention to the quality of the rationales.
1 code implementation • Findings (ACL) 2021 • Yuxuan Lai, Chen Zhang, Yansong Feng, Quzhe Huang, Dongyan Zhao
A thorough empirical analysis shows that MRC models tend to learn shortcut questions earlier than challenging questions, and the high proportions of shortcut questions in training sets hinder models from exploring the sophisticated reasoning skills in the later stage of training.
no code implementations • NAACL 2021 • Chongyang Tao, Shen Gao, Juntao Li, Yansong Feng, Dongyan Zhao, Rui Yan
Sequential information, a. k. a., orders, is assumed to be essential for processing a sequence with recurrent neural network or convolutional neural network based encoders.
1 code implementation • NAACL 2021 • Xiao Liu, Da Yin, Yansong Feng, Yuting Wu, Dongyan Zhao
Causal inference is the process of capturing cause-effect relationship among variables.
2 code implementations • NAACL 2021 • Yuxuan Lai, Yijia Liu, Yansong Feng, Songfang Huang, Dongyan Zhao
Further analysis shows that Lattice-BERT can harness the lattice structures, and the improvement comes from the exploration of redundant information and multi-granularity representations.
no code implementations • 17 Mar 2021 • Juntao Li, Chang Liu, Chongyang Tao, Zhangming Chan, Dongyan Zhao, Min Zhang, Rui Yan
To fill the gap between these up-to-date methods and the real-world applications, we incorporate user-specific dialogue history into the response selection and propose a personalized hybrid matching network (PHMN).
no code implementations • 10 Mar 2021 • Mingfei Guo, Xiuying Chen, Juntao Li, Dongyan Zhao, Rui Yan
Automatically identifying fake news from the Internet is a challenging problem in deception detection tasks.
1 code implementation • 14 Dec 2020 • Xiuying Chen, Zhi Cui, Jiayi Zhang, Chen Wei, Jianwei Cui, Bin Wang, Dongyan Zhao, Rui Yan
Hence, in this paper, we propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question, where the question is focused on the omitted information in the dialog.
no code implementations • 14 Dec 2020 • Mingzhe Li, Xiuying Chen, Min Yang, Shen Gao, Dongyan Zhao, Rui Yan
In this paper, we propose a Disentanglement-based Attractive Headline Generator (DAHG) that generates headline which captures the attractive content following the attractive style.
no code implementations • COLING 2020 • Wenpeng Hu, Ran Le, Bing Liu, Jinwen Ma, Dongyan Zhao, Rui Yan
Understanding neural models is a major topic of interest in the deep learning community.
no code implementations • 14 Nov 2020 • Shen Gao, Xiuying Chen, Zhaochun Ren, Dongyan Zhao, Rui Yan
To generate more meaningful answers, in this paper, we propose a novel generative neural model, called the Meaningful Product Answer Generator (MPAG), which alleviates the safe answer problem by taking product reviews, product attributes, and a prototype answer into consideration.
no code implementations • 5 Nov 2020 • Shen Gao, Xiuying Chen, Li Liu, Dongyan Zhao, Rui Yan
Hence, in this paper, we propose to recommend an appropriate sticker to user based on multi-turn dialog context and sticker using history of user.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Xiaohan Yu, Quzhe Huang, Zheng Wang, Yansong Feng, Dongyan Zhao
Code comments are vital for software maintenance and comprehension, but many software projects suffer from the lack of meaningful and up-to-date comments in practice.
1 code implementation • EMNLP 2020 • Xueliang Zhao, Wei Wu, Can Xu, Chongyang Tao, Dongyan Zhao, Rui Yan
We study knowledge-grounded dialogue generation with pre-trained language models.
1 code implementation • EMNLP 2020 • Mingzhe Li, Xiuying Chen, Shen Gao, Zhangming Chan, Dongyan Zhao, Rui Yan
Hence, in this paper, we propose the task of Video-based Multimodal Summarization with Multimodal Output (VMSMO) to tackle such a problem.
no code implementations • 14 Sep 2020 • Ruijian Xu, Chongyang Tao, Daxin Jiang, Xueliang Zhao, Dongyan Zhao, Rui Yan
To address these issues, in this paper, we propose learning a context-response matching model with auxiliary self-supervised tasks designed for the dialogue data based on pre-trained language models.
Ranked #2 on
Conversational Response Selection
on E-commerce
no code implementations • 23 Jun 2020 • Zechang Li, Yuxuan Lai, Yansong Feng, Dongyan Zhao
In this paper, we propose a novel semantic parser for domain adaptation, where we have much fewer annotated data in the target domain compared to the source domain.
1 code implementation • 17 May 2020 • Juntao Li, Chang Liu, Jian Wang, Lidong Bing, Hongsong Li, Xiaozhong Liu, Dongyan Zhao, Rui Yan
We manually collect a new and high-quality paired dataset, where each pair contains an unordered product attribute set in the source language and an informative product description in the target language.
1 code implementation • ACL 2020 • Yuting Wu, Xiao Liu, Yansong Feng, Zheng Wang, Dongyan Zhao
This paper presents Neighborhood Matching Network (NMN), a novel entity alignment framework for tackling the structural heterogeneity challenge.
no code implementations • 10 May 2020 • Shen Gao, Xiuying Chen, Zhaochun Ren, Dongyan Zhao, Rui Yan
Text summarization is the research area aiming at creating a short and condensed version of the original document, which conveys the main idea of the document in a few words.
1 code implementation • 10 Mar 2020 • Shen Gao, Xiuying Chen, Chang Liu, Li Liu, Dongyan Zhao, Rui Yan
Stickers with vivid and engaging expressions are becoming increasingly popular in online messaging apps, and some works are dedicated to automatically select sticker response by matching text labels of stickers with previous utterances.
no code implementations • ICLR 2020 • Xueliang Zhao, Wei Wu, Chongyang Tao, Can Xu, Dongyan Zhao, Rui Yan
In such a low-resource setting, we devise a disentangled response decoder in order to isolate parameters that depend on knowledge-grounded dialogues from the entire generation model.
1 code implementation • 26 Nov 2019 • Yuan Ye, Yansong Feng, Bingfeng Luo, Yuxuan Lai, Dongyan Zhao
However, such models often make predictions for each entity pair individually, thus often fail to solve the inconsistency among different predictions, which can be characterized by discrete relation constraints.
1 code implementation • COLING 2020 • Wenpeng Hu, Mengyu Wang, Bing Liu, Feng Ji, Haiqing Chen, Dongyan Zhao, Jinwen Ma, Rui Yan
The key idea of the proposed approach is to use a Forward Transformation to transform dense representations to sparse representations.
1 code implementation • 7 Nov 2019 • Zhenxin Fu, Feng Ji, Wenpeng Hu, Wei Zhou, Dongyan Zhao, Haiqing Chen, Rui Yan
Information-seeking conversation system aims at satisfying the information needs of users through conversations.
no code implementations • IJCNLP 2019 • Zhangming Chan, Juntao Li, Xiaopeng Yang, Xiuying Chen, Wenpeng Hu, Dongyan Zhao, Rui Yan
In this work, we improve the WAE for response generation.
no code implementations • IJCNLP 2019 • Ran Le, Wenpeng Hu, Mingyue Shang, Zhenjun You, Lidong Bing, Dongyan Zhao, Rui Yan
Previous research on dialogue systems generally focuses on the conversation between two participants, yet multi-party conversations which involve more than two participants within one session bring up a more complicated but realistic scenario.
no code implementations • IJCNLP 2019 • Jia Li, Chongyang Tao, Wei Wu, Yansong Feng, Dongyan Zhao, Rui Yan
We study how to sample negative examples to automatically construct a training set for effective model learning in retrieval-based dialogue systems.
no code implementations • IJCNLP 2019 • Zhangming Chan, Xiuying Chen, Yongliang Wang, Juntao Li, Zhiqiang Zhang, Kun Gai, Dongyan Zhao, Rui Yan
Different from other text generation tasks, in product description generation, it is of vital importance to generate faithful descriptions that stick to the product attribute information.
no code implementations • IJCNLP 2019 • Jizhi Tang, Yansong Feng, Dongyan Zhao
News streams contain rich up-to-date information which can be used to update knowledge graphs (KGs).
no code implementations • 28 Oct 2019 • Xiuying Chen, Daorui Xiao, Shen Gao, Guojun Liu, Wei. Lin, Bo Zheng, Dongyan Zhao, Rui Yan
Sponsored search optimizes revenue and relevance, which is estimated by Revenue Per Mille (RPM).
no code implementations • 6 Oct 2019 • Chen Chen, Lisong Qiu, Zhenxin Fu, Dongyan Zhao, Junfei Liu, Rui Yan
Existing dialog systems are all monolingual, where features shared among different languages are rarely explored.
no code implementations • 25 Sep 2019 • Wenpeng Hu, Ran Le, Bing Liu, Feng Ji, Haiqing Chen, Dongyan Zhao, Jinwen Ma, Rui Yan
Positive-unlabeled (PU) learning learns a binary classifier using only positive and unlabeled examples without labeled negative examples.
no code implementations • IJCNLP 2019 • Mingyue Shang, Piji Li, Zhenxin Fu, Lidong Bing, Dongyan Zhao, Shuming Shi, Rui Yan
Text style transfer task requires the model to transfer a sentence of one style to another style while retaining its original content meaning, which is a challenging problem that has long suffered from the shortage of parallel data.
1 code implementation • IJCNLP 2019 • Yuting Wu, Xiao Liu, Yansong Feng, Zheng Wang, Dongyan Zhao
Entity alignment is a viable means for integrating heterogeneous knowledge among different knowledge graphs (KGs).
Ranked #18 on
Entity Alignment
on DBP15k zh-en
(using extra training data)
1 code implementation • IJCNLP 2019 • Shen Gao, Xiuying Chen, Piji Li, Zhangming Chan, Dongyan Zhao, Rui Yan
There are two main challenges in this task: (1) the model needs to incorporate learned patterns from the prototype, but (2) should avoid copying contents other than the patternized words---such as irrelevant facts---into the generated summaries.
1 code implementation • 2 Sep 2019 • Zechang Li, Yuxuan Lai, Yuxi Xie, Yansong Feng, Dongyan Zhao
The sketch is a high-level structure of the logical form exclusive of low-level details such as entities and predicates.
1 code implementation • 22 Aug 2019 • Yuting Wu, Xiao Liu, Yansong Feng, Zheng Wang, Rui Yan, Dongyan Zhao
Entity alignment is the task of linking entities with the same real-world identity from different knowledge graphs (KGs), which has been recently dominated by embedding-based methods.
Ranked #20 on
Entity Alignment
on DBP15k zh-en
(using extra training data)
1 code implementation • IJCAI 2019 2019 • Xiuying Chen, Zhangming Chan, Shen Gao, Meng-Hsuan Yu, Dongyan Zhao, Rui Yan
Timeline summarization targets at concisely summarizing the evolution trajectory along the timeline and existing timeline summarization approaches are all based on extractive methods. In this paper, we propose the task of abstractive timeline summarization, which tends to concisely paraphrase the information in the time-stamped events. Unlike traditional document summarization, timeline summarization needs to model the time series information of the input events and summarize important events in chronological order. To tackle this challenge, we propose a memory-based timeline summarization model (MTS). Concretely, we propose a time-event memory to establish a timeline, and use the time position of events on this timeline to guide generation process. Besides, in each decoding step, we incorporate event-level information into word-level attention to avoid confusion between events. Extensive experiments are conducted on a large-scale real-world dataset, and the results show that MTS achieves the state-of-the-art performance in terms of both automatic and human evaluations.
Ranked #1 on
Timeline Summarization
on MTS
1 code implementation • ACL 2019 • Chongyang Tao, Wei Wu, Can Xu, Wenpeng Hu, Dongyan Zhao, Rui Yan
Currently, researchers have paid great attention to retrieval-based dialogues in open-domain.
Ranked #10 on
Conversational Response Selection
on E-commerce
no code implementations • ACL 2019 • Lisong Qiu, Juntao Li, Wei Bi, Dongyan Zhao, Rui Yan
Due to its potential applications, open-domain dialogue generation has become popular and achieved remarkable progress in recent years, but sometimes suffers from generic responses.
no code implementations • 11 Jun 2019 • Xueliang Zhao, Chongyang Tao, Wei Wu, Can Xu, Dongyan Zhao, Rui Yan
We present a document-grounded matching network (DGMN) for response selection that can power a knowledge-aware retrieval-based chatbot system.
no code implementations • ACL 2019 • Jiazhan Feng, Chongyang Tao, Wei Wu, Yansong Feng, Dongyan Zhao, Rui Yan
Under the framework, we simultaneously learn two matching models with independent training sets.
1 code implementation • 31 May 2019 • Wenpeng Hu, Zhangming Chan, Bing Liu, Dongyan Zhao, Jinwen Ma, Rui Yan
Existing neural models for dialogue response generation assume that utterances are sequentially organized.
no code implementations • ICLR 2019 • Wenpeng Hu, Zhou Lin, Bing Liu, Chongyang Tao, Zhengwei Tao, Jinwen Ma, Dongyan Zhao, Rui Yan
Several continual learning methods have been proposed to address the problem.
no code implementations • ICLR 2019 • Wenpeng Hu, Zhengwei Tao, Zhanxing Zhu, Bing Liu, Zhou Lin, Jinwen Ma, Dongyan Zhao, Rui Yan
A large amount of parallel data is needed to train a strong neural machine translation (NMT) system.
1 code implementation • 25 Feb 2019 • Yuxuan Lai, Yansong Feng, Xiaohan Yu, Zheng Wang, Kun Xu, Dongyan Zhao
Short text matching often faces the challenges that there are great word mismatch and expression diversity between the two texts, which would be further aggravated in languages like Chinese where there is no natural space to segment words explicitly.
1 code implementation • 23 Jan 2019 • Shen Gao, Zhaochun Ren, Yihong Eric Zhao, Dongyan Zhao, Dawei Yin, Rui Yan
In this paper, we propose the task of product-aware answer generation, which tends to generate an accurate and complete answer from large-scale unlabeled e-commerce reviews and product attributes.
Ranked #2 on
Question Answering
on JD Product Question Answer
no code implementations • 13 Dec 2018 • Shen Gao, Xiuying Chen, Piji Li, Zhaochun Ren, Lidong Bing, Dongyan Zhao, Rui Yan
To tackle this problem, we propose the task of reader-aware abstractive summary generation, which utilizes the reader comments to help the model produce better summary about the main aspect.
Ranked #1 on
Reader-Aware Summarization
on RASG
no code implementations • 13 Dec 2018 • Mingyue Shang, Zhenxin Fu, Hongzhi Yin, Bo Tang, Dongyan Zhao, Rui Yan
In this paper, we incorporate the logic information with the help of the Natural Language Inference (NLI) task to the Story Cloze Test (SCT).
no code implementations • 19 Nov 2018 • Lili Yao, Ruijian Xu, Chao Li, Dongyan Zhao, Rui Yan
To build an open-domain multi-turn conversation system is one of the most interesting and challenging tasks in Artificial Intelligence.
2 code implementations • 14 Nov 2018 • Lili Yao, Nanyun Peng, Ralph Weischedel, Kevin Knight, Dongyan Zhao, Rui Yan
Automatic storytelling is challenging since it requires generating long, coherent natural language to describes a sensible sequence of events.
no code implementations • 9 Nov 2018 • Li-Wei Chen, Yansong Feng, Songfang Huang, Bingfeng Luo, Dongyan Zhao
Relation extraction is the task of identifying predefined relationship between entities, and plays an essential role in information extraction, knowledge base construction, question answering and so on.
1 code implementation • EMNLP 2018 • Feifan Fan, Yansong Feng, Dongyan Zhao
We propose a fine-grained attention mechanism, which can capture the word-level interaction between aspect and context.
no code implementations • EMNLP 2018 • Juntao Li, Yan Song, Haisong Zhang, Dongmin Chen, Shuming Shi, Dongyan Zhao, Rui Yan
It is a challenging task to automatically compose poems with not only fluent expressions but also aesthetic wording.
1 code implementation • EMNLP 2018 • Xiuying Chen, Shen Gao, Chongyang Tao, Yan Song, Dongyan Zhao, Rui Yan
In this paper, we introduce Iterative Text Summarization (ITS), an iteration-based model for supervised extractive text summarization, inspired by the observation that it is often necessary for a human to read an article multiple times in order to fully understand and summarize its contents.
Ranked #14 on
Extractive Text Summarization
on CNN / Daily Mail
no code implementations • 22 Aug 2018 • Chongyang Tao, Wei Wu, Can Xu, Yansong Feng, Dongyan Zhao, Rui Yan
In this paper, we study context-response matching with pre-trained contextualized representations for multi-turn response selection in retrieval-based chatbots.
no code implementations • ACL 2018 • Yanyan Jia, Yuan Ye, Yansong Feng, Yuxuan Lai, Rui Yan, Dongyan Zhao
Identifying long-span dependencies between discourse units is crucial to improve discourse parsing performance.
no code implementations • ACL 2018 • Bingfeng Luo, Yansong Feng, Zheng Wang, Songfang Huang, Rui Yan, Dongyan Zhao
The success of many natural language processing (NLP) tasks is bound by the number and quality of annotated data, but there is often a shortage of such training data.
no code implementations • 8 May 2018 • Xiaowei Tong, Zhenxin Fu, Mingyue Shang, Dongyan Zhao, Rui Yan
Automatic evaluating the performance of Open-domain dialogue system is a challenging problem.
no code implementations • ICLR 2018 • Ning Miao, Hengliang Wang, Ran Le, Chongyang Tao, Mingyue Shang, Rui Yan, Dongyan Zhao
Traditional recurrent neural network (RNN) or convolutional neural net- work (CNN) based sequence-to-sequence model can not handle tree structural data well.
no code implementations • ICLR 2018 • Wenpeng Hu, Bing Liu, Rui Yan, Dongyan Zhao, Jinwen Ma
In the paper, we propose a new question generation problem, which also requires the input of a target topic in addition to a piece of descriptive text.
no code implementations • ICLR 2018 • Yiping Song, Rui Yan, Cheng-Te Li, Jian-Yun Nie, Ming Zhang, Dongyan Zhao
Human-computer conversation systems have attracted much attention in Natural Language Processing.
no code implementations • 11 Dec 2017 • Ying Zeng, Yansong Feng, Rong Ma, Zheng Wang, Rui Yan, Chongde Shi, Dongyan Zhao
We show that this large volume of training data not only leads to a better event extractor, but also allows us to detect multiple typed events.
2 code implementations • 18 Nov 2017 • Zhenxin Fu, Xiaoye Tan, Nanyun Peng, Dongyan Zhao, Rui Yan
Results show that the proposed content preservation metric is highly correlate to human judgments, and the proposed models are able to generate sentences with higher style transfer strength and similar content preservation score comparing to auto-encoder.
Ranked #5 on
Unsupervised Text Style Transfer
on Yelp
no code implementations • IJCNLP 2017 • Yiping Song, Zhiliang Tian, Dongyan Zhao, Ming Zhang, Rui Yan
However, traditional seq2seq suffer from a severe weakness: during beam search decoding, they tend to rank universal replies at the top of the candidate list, resulting in the lack of diversity among candidate replies.
no code implementations • EMNLP 2017 • Lili Yao, Yaoyuan Zhang, Yansong Feng, Dongyan Zhao, Rui Yan
The study on human-computer conversation systems is a hot research topic nowadays.
no code implementations • EMNLP 2017 • Bingfeng Luo, Yansong Feng, Jianbo Xu, Xiang Zhang, Dongyan Zhao
The charge prediction task is to determine appropriate charges for a given case, which is helpful for legal assistant systems where the user input is fact description.
no code implementations • ACL 2017 • Zhiliang Tian, Rui Yan, Lili Mou, Yiping Song, Yansong Feng, Dongyan Zhao
Generative conversational systems are attracting increasing attention in natural language processing (NLP).
no code implementations • ACL 2017 • Bingfeng Luo, Yansong Feng, Zheng Wang, Zhanxing Zhu, Songfang Huang, Rui Yan, Dongyan Zhao
We show that the dynamic transition matrix can effectively characterize the noise in the training data built by distant supervision.
no code implementations • 7 Apr 2017 • Yaoyuan Zhang, Zhenxu Ye, Yansong Feng, Dongyan Zhao, Rui Yan
For word-level studies, words are simplified but also have potential grammar errors due to different usages of words before and after simplification.
1 code implementation • 11 Jan 2017 • Chongyang Tao, Lili Mou, Dongyan Zhao, Rui Yan
Open-domain human-computer conversation has been attracting increasing attention over the past few years.
no code implementations • COLING 2016 • Kun Xu, Yansong Feng, Songfang Huang, Dongyan Zhao
While these systems are able to provide more precise answers than information retrieval (IR) based QA systems, the natural incompleteness of KB inevitably limits the question scope that the system can answer.
2 code implementations • 23 Oct 2016 • Yiping Song, Rui Yan, Xiang Li, Dongyan Zhao, Ming Zhang
In this paper, we propose a novel ensemble of retrieval-based and generation-based dialog systems in the open domain.
1 code implementation • ACL 2016 • Kun Xu, Siva Reddy, Yansong Feng, Songfang Huang, Dongyan Zhao
Existing knowledge-based question answering systems often rely on small annotated training data.
no code implementations • EMNLP 2015 • Kun Xu, Yansong Feng, Songfang Huang, Dongyan Zhao
Syntactic features play an essential role in identifying relationship in a sentence.
Ranked #3 on
Relation Classification
on SemEval 2010 Task 8