1 code implementation • EMNLP (insights) 2021 • Qi Zhu, Yuxian Gu, Lingxiao Luo, Bing Li, Cheng Li, Wei Peng, Minlie Huang, Xiaoyan Zhu
Further pre-training language models on in-domain data (domain-adaptive pre-training, DAPT) or task-relevant data (task-adaptive pre-training, TAPT) before fine-tuning has been shown to improve downstream tasks’ performances.
1 code implementation • EMNLP 2021 • Hao Zhou, Minlie Huang, Yong liu, Wei Chen, Xiaoyan Zhu
Generating informative and appropriate responses is challenging but important for building human-like dialogue systems.
no code implementations • 4 Dec 2022 • Qi Zhu, Fei Mi, Zheng Zhang, Yasheng Wang, Yitong Li, Xin Jiang, Qun Liu, Xiaoyan Zhu, Minlie Huang
For the former, the grounding knowledge consists of keywords extracted from the response.
1 code implementation • 17 Oct 2022 • Yuxian Gu, Pei Ke, Xiaoyan Zhu, Minlie Huang
Recently, instruction tuning (IT), which fine-tunes a pre-trained language model on a massive collection of tasks described via human-craft instructions, has been shown effective in instruction learning for unseen tasks.
1 code implementation • 6 Jun 2022 • Pei Ke, Haozhe Ji, Zhenyu Yang, Yi Huang, Junlan Feng, Xiaoyan Zhu, Minlie Huang
Despite the success of text-to-text pre-trained models in various natural language generation (NLG) tasks, the generation performance is largely restricted by the number of labeled data in downstream tasks, particularly in data-to-text generation tasks.
1 code implementation • ACL 2022 • Pei Ke, Hao Zhou, Yankai Lin, Peng Li, Jie zhou, Xiaoyan Zhu, Minlie Huang
Existing reference-free metrics have obvious limitations for evaluating controlled text generation models.
1 code implementation • 17 Mar 2022 • Yuxian Gu, Jiaxin Wen, Hao Sun, Yi Song, Pei Ke, Chujie Zheng, Zheng Zhang, Jianzhu Yao, Xiaoyan Zhu, Jie Tang, Minlie Huang
Large-scale pre-training has shown remarkable performance in building open-domain dialogue systems.
1 code implementation • ACL 2022 • Qi Zhu, Bing Li, Fei Mi, Xiaoyan Zhu, Minlie Huang
A desirable dialog system should be able to continually learn new skills without forgetting old ones, and thereby adapt to new domains or tasks in its life cycle.
1 code implementation • ACL 2022 • Siyang Liu, Sahand Sabour, Yinhe Zheng, Pei Ke, Xiaoyan Zhu, Minlie Huang
We provide both empirical and theoretical evidence to show that our method effectively removes the biases existing in the original distinct score.
1 code implementation • Findings (ACL) 2022 • Hao Sun, Guangxuan Xu, Jiawen Deng, Jiale Cheng, Chujie Zheng, Hao Zhou, Nanyun Peng, Xiaoyan Zhu, Minlie Huang
We propose a taxonomy for dialogue safety specifically designed to capture unsafe behaviors in human-bot dialogue settings, with focuses on context-sensitive unsafety, which is under-explored in prior works.
2 code implementations • 3 Aug 2021 • Hao Zhou, Pei Ke, Zheng Zhang, Yuxian Gu, Yinhe Zheng, Chujie Zheng, Yida Wang, Chen Henry Wu, Hao Sun, Xiaocong Yang, Bosi Wen, Xiaoyan Zhu, Minlie Huang, Jie Tang
Although pre-trained language models have remarkably enhanced the generation ability of dialogue systems, open-domain Chinese dialogue systems are still limited by the dialogue data and the model size compared with English ones.
2 code implementations • 20 Jun 2021 • Zhengyan Zhang, Yuxian Gu, Xu Han, Shengqi Chen, Chaojun Xiao, Zhenbo Sun, Yuan YAO, Fanchao Qi, Jian Guan, Pei Ke, Yanzheng Cai, Guoyang Zeng, Zhixing Tan, Zhiyuan Liu, Minlie Huang, Wentao Han, Yang Liu, Xiaoyan Zhu, Maosong Sun
We present a suite of cost-effective techniques for the use of PLMs to deal with the efficiency issues of pre-training, fine-tuning, and inference.
1 code implementation • Findings (ACL) 2021 • Pei Ke, Haozhe Ji, Yu Ran, Xin Cui, LiWei Wang, Linfeng Song, Xiaoyan Zhu, Minlie Huang
Existing pre-trained models for knowledge-graph-to-text (KG-to-text) generation simply fine-tune text-to-text pre-trained models such as BART or T5 on KG-to-text datasets, which largely ignore the graph structure during encoding and lack elaborate pre-training tasks to explicitly model graph-text alignments.
Ranked #1 on
KG-to-Text Generation
on WebQuestions
1 code implementation • Findings (ACL) 2021 • Fei Huang, Zikai Chen, Chen Henry Wu, Qihan Guo, Xiaoyan Zhu, Minlie Huang
First, we observe that most words in the transferred sentence can be aligned with related words in the source sentence, so we explicitly model word alignments to suppress irrelevant words.
1 code implementation • ACL 2021 • Yilin Niu, Fei Huang, Jiaming Liang, Wenkai Chen, Xiaoyan Zhu, Minlie Huang
In this paper, we present a novel SEmantic-based Question Answering method (SEQA) for unsupervised commonsense question answering.
no code implementations • 15 Jan 2021 • Guangtao Wang, Qinbao Song, Xiaoyan Zhu
Recommending appropriate algorithms to a classification problem is one of the most challenging issues in the field of data mining.
no code implementations • 1 Jan 2021 • Fei Huang, Jian Guan, Pei Ke, Qihan Guo, Xiaoyan Zhu, Minlie Huang
Despite the great success of Generative Adversarial Networks (GANs) in generating high-quality images, GANs for text generation still face two major challenges: first, most text GANs are unstable in training mainly due to ineffective optimization of the generator, and they heavily rely on maximum likelihood pretraining; second, most text GANs adopt autoregressive generators without latent variables, which largely limits the ability to learn latent representations for natural language text.
3 code implementations • 1 Dec 2020 • Zhengyan Zhang, Xu Han, Hao Zhou, Pei Ke, Yuxian Gu, Deming Ye, Yujia Qin, Yusheng Su, Haozhe Ji, Jian Guan, Fanchao Qi, Xiaozhi Wang, Yanan Zheng, Guoyang Zeng, Huanqi Cao, Shengqi Chen, Daixuan Li, Zhenbo Sun, Zhiyuan Liu, Minlie Huang, Wentao Han, Jie Tang, Juanzi Li, Xiaoyan Zhu, Maosong Sun
However, applying GPT-3 to address Chinese NLP tasks is still challenging, as the training corpus of GPT-3 is primarily English, and the parameters are not publicly available.
1 code implementation • EMNLP 2020 • Haozhe Ji, Pei Ke, Shaohan Huang, Furu Wei, Xiaoyan Zhu, Minlie Huang
Despite the success of generative pre-trained language models on a series of text generation tasks, they still suffer in cases where reasoning over underlying commonsense knowledge is required during generation.
2 code implementations • 10 Aug 2020 • Yida Wang, Pei Ke, Yinhe Zheng, Kaili Huang, Yong Jiang, Xiaoyan Zhu, Minlie Huang
The cleaned dataset and the pre-training models will facilitate the research of short-text conversation modeling.
1 code implementation • 9 Jun 2020 • Mantong Zhou, Zhouxing Shi, Minlie Huang, Xiaoyan Zhu
During document retrieval, a candidate document is scored by considering its relationship to the question and other documents.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Zheng Zhang, Lizi Liao, Xiaoyan Zhu, Tat-Seng Chua, Zitao Liu, Yan Huang, Minlie Huang
Most existing approaches for goal-oriented dialogue policy learning used reinforcement learning, which focuses on the target agent policy and simply treat the opposite agent policy as part of the environment.
1 code implementation • ACL 2020 • Hao Zhou, Chujie Zheng, Kaili Huang, Minlie Huang, Xiaoyan Zhu
The research of knowledge-driven conversational systems is largely limited due to the lack of dialog data which consist of multi-turn conversations on multiple topics and with knowledge annotations.
no code implementations • 17 Mar 2020 • Zheng Zhang, Ryuichi Takanobu, Qi Zhu, Minlie Huang, Xiaoyan Zhu
Due to the significance and value in human-computer interaction and natural language processing, task-oriented dialog systems are attracting more and more attention in both academic and industrial communities.
2 code implementations • TACL 2020 • Qi Zhu, Kaili Huang, Zheng Zhang, Xiaoyan Zhu, Minlie Huang
To advance multi-domain (cross-domain) dialogue modeling as well as alleviate the shortage of Chinese task-oriented datasets, we propose CrossWOZ, the first large-scale Chinese Cross-Domain Wizard-of-Oz task-oriented dataset.
1 code implementation • ACL 2020 • Qi Zhu, Zheng Zhang, Yan Fang, Xiang Li, Ryuichi Takanobu, Jinchao Li, Baolin Peng, Jianfeng Gao, Xiaoyan Zhu, Minlie Huang
We present ConvLab-2, an open-source toolkit that enables researchers to build task-oriented dialogue systems with state-of-the-art models, perform an end-to-end evaluation, and diagnose the weakness of systems.
1 code implementation • 3 Feb 2020 • Fei Huang, Dazhen Wan, Zhihong Shao, Pei Ke, Jian Guan, Yilin Niu, Xiaoyan Zhu, Minlie Huang
In text generation evaluation, many practical issues, such as inconsistent experimental settings and metric implementations, are often ignored but lead to unfair evaluation and untenable conclusions.
1 code implementation • TACL 2020 • Jian Guan, Fei Huang, Zhihao Zhao, Xiaoyan Zhu, Minlie Huang
To further capture the causal and temporal dependencies between the sentences in a reasonable story, we employ multi-task learning which combines a discriminative objective to distinguish true and fake stories during fine-tuning.
no code implementations • 16 Nov 2019 • Mantong Zhou, Minlie Huang, Xiaoyan Zhu
In this paper, we address the over-confidence issue and the over-sensitivity issue existing in current RC models simultaneously with the help of external linguistic knowledge.
1 code implementation • EMNLP 2020 • Pei Ke, Haozhe Ji, Siyang Liu, Xiaoyan Zhu, Minlie Huang
To benefit the downstream tasks in sentiment analysis, we propose a novel language representation model called SentiLARE, which introduces word-level linguistic knowledge including part-of-speech tag and sentiment polarity (inferred from SentiWordNet) into pre-trained models.
no code implementations • 7 Sep 2019 • Jingwen Fu, Xiaoyan Zhu, Yingbin Li
Automatic defect recognition is one of the research hotspots in steel production, but most of the current methods mainly extract features manually and use machine learning classifiers to recognize defects, which cannot tackle the situation, where there are few data available to train and confine to a certain scene.
1 code implementation • IJCNLP 2019 • Pei Ke, Fei Huang, Minlie Huang, Xiaoyan Zhu
The generator is optimized with maximum likelihood estimation augmented by the discriminator's rewards instead of policy gradient.
1 code implementation • IJCNLP 2019 • Zhihong Shao, Minlie Huang, Jiangtao Wen, Wenfei Xu, Xiaoyan Zhu
Existing neural methods for data-to-text generation are still struggling to produce long and diverse texts: they are insufficient to model input data dynamically during generation, to capture inter-sentence coherence, or to generate diversified expressions.
no code implementations • 13 May 2019 • Minlie Huang, Xiaoyan Zhu, Jianfeng Gao
This paper reviews the recent works on neural approaches that are devoted to addressing three challenges in developing such systems: semantics, consistency, and interactiveness.
no code implementations • 27 Feb 2019 • Hao Zhou, Minlie Huang, Yishun Mao, Changlei Zhu, Peng Shu, Xiaoyan Zhu
Second, the inefficient ad impression issue: a large proportion of search queries, which are unpopular yet relevant to many ad keywords, have no ads presented on their search result pages.
no code implementations • 12 Feb 2019 • Dongliang Xu, Bailing Wang, XiaoJiang Du, Xiaoyan Zhu, zhitao Guan, Xiaoyan Yu, Jingyu Liu
However, the advantages of convolutional neural networks depend on the data used by the training classifier, particularly the size of the training set.
no code implementations • 3 Nov 2018 • Jun Feng, Minlie Huang, Yijie Zhang, Yang Yang, Xiaoyan Zhu
Experimental results show that our model is effective to extract relation mentions from noisy data.
no code implementations • 17 Sep 2018 • Jun Feng, Heng Li, Minlie Huang, Shichen Liu, Wenwu Ou, Zhirong Wang, Xiaoyan Zhu
The first one is lack of collaboration between scenarios meaning that each strategy maximizes its own objective but ignores the goals of other strategies, leading to a sub-optimal overall performance.
Multi-agent Reinforcement Learning
reinforcement-learning
+1
2 code implementations • 24 Aug 2018 • Jun Feng, Minlie Huang, Li Zhao, Yang Yang, Xiaoyan Zhu
In this paper, we propose a novel model for relation classification at the sentence level from noisy data.
no code implementations • COLING 2018 • Naitong Yu, Jie Zhang, Minlie Huang, Xiaoyan Zhu
Delete-based models have the strong ability to delete undesired words, while generate-based models are able to reorder or rephrase the words, which are more coherent to human sentence compression.
1 code implementation • ACL 2018 • Pei Ke, Jian Guan, Minlie Huang, Xiaoyan Zhu
Experiments show that our model outperforms state-of-the-art baselines, and it has the ability to generate responses with both controlled sentence function and informative content.
no code implementations • 1 May 2018 • Zheng Zhang, Minlie Huang, Zhongzhou Zhao, Feng Ji, Haiqing Chen, Xiaoyan Zhu
Dialogue management (DM) decides the next action of a dialogue system according to the current dialogue state, and thus plays a central role in task-oriented dialogue systems.
1 code implementation • The Web Conference (WWW) 2018 • Yequan Wang, Aixin Sun, Jialong Han, Ying Liu, Xiaoyan Zhu
Based on capsule representation, the probability module computes the capsule’s state probability.
Ranked #5 on
Sentiment Analysis
on MR
1 code implementation • COLING 2018 • Mantong Zhou, Minlie Huang, Xiaoyan Zhu
Multi-relation Question Answering is a challenging task, due to the requirement of elaborated analysis on questions and reasoning over multiple fact triples in knowledge base.
no code implementations • ACL 2017 • Qiao Qian, Minlie Huang, Jinhao Lei, Xiaoyan Zhu
This paper deals with sentence-level sentiment classification.
1 code implementation • 9 Jun 2017 • Qiao Qian, Minlie Huang, Haizhou Zhao, Jingfang Xu, Xiaoyan Zhu
Endowing a chatbot with personality or an identity is quite challenging but critical to deliver more realistic and natural conversations.
6 code implementations • 4 Apr 2017 • Hao Zhou, Minlie Huang, Tianyang Zhang, Xiaoyan Zhu, Bing Liu
Perception and expression of emotion are key factors to the success of dialogue systems or conversational agents.
no code implementations • COLING 2016 • Hao Zhou, Minlie Huang, Xiaoyan Zhu
Most tranditional QA systems based on templates or rules tend to generate rigid and stylised responses without the natural variation of human language.
no code implementations • COLING 2016 • Naitong Yu, Minlie Huang, Yuanyuan Shi, Xiaoyan Zhu
The main idea of our method is to leverage phrase properties to choose a subset of optimal phrases for generating the final summary.
1 code implementation • COLING 2016 • Jun Feng, Minlie Huang, Yang Yang, Xiaoyan Zhu
Knowledge embedding, which projects triples in a given knowledge base to d-dimensional vectors, has attracted considerable research efforts recently.
no code implementations • 12 Nov 2016 • Qiao Qian, Minlie Huang, Jinhao Lei, Xiaoyan Zhu
In this paper, we propose simple models trained with sentence-level annotation, but also attempt to generating linguistically coherent representations by employing regularizers that model the linguistic role of sentiment lexicons, negation words, and intensity words.
no code implementations • 27 Aug 2016 • Han Xiao, Minlie Huang, Xiaoyan Zhu
Since both aspects and categories are semantics-relevant, the collection of categories in each aspect is treated as the semantic representation of this triple.
no code implementations • 17 Apr 2016 • Han Xiao, Minlie Huang, Xiaoyan Zhu
To this end, this paper proposes a semantic representation method for knowledge graph \textbf{(KSR)}, which imposes a two-level hierarchical generative process that globally extracts many aspects and then locally assigns a specific category in each aspect for every triple.
no code implementations • 15 Dec 2015 • Han Xiao, Minlie Huang, Xiaoyan Zhu
Knowledge graph embedding aims at offering a numerical knowledge representation paradigm by transforming the entities and relations into continuous vector space.
no code implementations • 18 Sep 2015 • Han Xiao, Minlie Huang, Yu Hao, Xiaoyan Zhu
Knowledge representation is a major topic in AI, and many studies attempt to represent entities and relations of knowledge base in a continuous vector space.
no code implementations • 18 Sep 2015 • Han Xiao, Minlie Huang, Yu Hao, Xiaoyan Zhu
Recently, knowledge graph embedding, which projects symbolic entities and relations into continuous vector space, has become a new, hot topic in artificial intelligence.
no code implementations • 11 Jun 2015 • Han Xiao, Xiaoyan Zhu
Margin-Based Principle has been proposed for a long time, it has been proved that this principle could reduce the structural risk and improve the performance in both theoretical and practical aspects.
no code implementations • 11 Jun 2015 • Han Xiao, Xiaoyan Zhu
Entropy-Based Principle is the principle with which we could estimate the unknown distribution under some limited conditions.
no code implementations • 20 May 2015 • Jun Feng, Mantong Zhou, Yu Hao, Minlie Huang, Xiaoyan Zhu
TransF regards relation as translation between head entity vector and tail entity vector with flexible magnitude.