no code implementations • ICLR 2019 • Pengfei Liu, Xuanjing Huang
In this paper, we describe a general framework to systematically analyze current neural models for multi-task learning, in which we find that existing models expect to disentangle features into different spaces while features learned in practice are still entangled in shared space, leaving potential hazards for other training or unseen tasks.
no code implementations • ACL 2022 • Qin Liu, Rui Zheng, Bao Rong, Jingyi Liu, Zhihua Liu, Zhanzhan Cheng, Liang Qiao, Tao Gui, Qi Zhang, Xuanjing Huang
Adversarial robustness has attracted much attention recently, and the mainstream solution is adversarial training.
1 code implementation • ACL 2022 • Rui Zheng, Bao Rong, Yuhao Zhou, Di Liang, Sirui Wang, Wei Wu, Tao Gui, Qi Zhang, Xuanjing Huang
Recent works on Lottery Ticket Hypothesis have shown that pre-trained language models (PLMs) contain smaller matching subnetworks(winning tickets) which are capable of reaching accuracy comparable to the original models.
no code implementations • Findings (ACL) 2022 • Jianhan Xu, Cenyuan Zhang, Xiaoqing Zheng, Linyang Li, Cho-Jui Hsieh, Kai-Wei Chang, Xuanjing Huang
Most of the existing defense methods improve the adversarial robustness by making the models adapt to the training set augmented with some adversarial examples.
1 code implementation • EMNLP 2020 • Siyuan Wang, Zhongyu Wei, Zhihao Fan, Zengfeng Huang, Weijian Sun, Qi Zhang, Xuanjing Huang
Human evaluation also proves that our model is able to generate relevant and informative questions.
1 code implementation • ACL 2022 • Zichu Fei, Qi Zhang, Tao Gui, Di Liang, Sirui Wang, Wei Wu, Xuanjing Huang
CQG employs a simple method to generate the multi-hop questions that contain key entities in multi-hop reasoning chains, which ensure the complexity and quality of the questions.
no code implementations • EMNLP 2020 • Qinzhuo Wu, Qi Zhang, Jinlan Fu, Xuanjing Huang
With the advancements in natural language processing tasks, math word problem solving has received increasing attention.
1 code implementation • 19 Apr 2022 • Wei Chen, Zhiwei Li, Hongyi Fang, Qianyuan Yao, Cheng Zhong, Jianye Hao, Qi Zhang, Xuanjing Huang, Jiajie Peng, Zhongyu Wei
In recent years, interest has arisen in using machine learning to improve the efficiency of automatic medical consultation and enhance patient experience.
no code implementations • ACL 2022 • Xiao Wang, Shihan Dou, Limao Xiong, Yicheng Zou, Qi Zhang, Tao Gui, Liang Qiao, Zhanzhan Cheng, Xuanjing Huang
NER model has achieved promising performance on standard NER benchmarks.
1 code implementation • Findings (ACL) 2022 • Tianxiang Sun, Xiangyang Liu, Wei Zhu, Zhichao Geng, Lingling Wu, Yilong He, Yuan Ni, Guotong Xie, Xuanjing Huang, Xipeng Qiu
Previous works usually adopt heuristic metrics such as the entropy of internal outputs to measure instance difficulty, which suffers from generalization and threshold-tuning.
no code implementations • 16 Feb 2022 • Shihan Dou, Rui Zheng, Ting Wu, Songyang Gao, Qi Zhang, Yueming Wu, Xuanjing Huang
However, down-weighting these samples obstructs the model in learning from the non-biased parts of these samples.
1 code implementation • 29 Jan 2022 • Zejun Li, Zhihao Fan, Huaixiao Tou, Zhongyu Wei, Xuanjing Huang
We introduce concepts in both modalities to construct two-level semantic representations for language and vision.
1 code implementation • 10 Jan 2022 • Tianxiang Sun, Yunfan Shao, Hong Qian, Xuanjing Huang, Xipeng Qiu
In such a scenario, which we call Language-Model-as-a-Service (LMaaS), the gradients of PTMs are usually unavailable.
no code implementations • 14 Oct 2021 • Xin Zhou, Ruotian Ma, Tao Gui, Yiding Tan, Qi Zhang, Xuanjing Huang
Specifically, for each task, a label word set is first constructed by selecting a high-frequency word for each class respectively, and then, task-specific vectors are inserted into the inputs and optimized to manipulate the model predictions towards the corresponding label words.
1 code implementation • 13 Oct 2021 • Xiangyang Liu, Tianxiang Sun, Junliang He, Jiawen Wu, Lingling Wu, Xinyu Zhang, Hao Jiang, Zhao Cao, Xuanjing Huang, Xipeng Qiu
ELUE is dedicated to depict the Pareto Frontier for various language understanding tasks, such that it can tell whether and how much a method achieves Pareto improvement.
1 code implementation • 6 Oct 2021 • Linyang Li, Demin Song, Ruotian Ma, Xipeng Qiu, Xuanjing Huang
Pre-trained models are widely used in fine-tuning downstream tasks with linear classifiers optimized by the cross-entropy loss, which might face robustness and stability problems.
1 code implementation • 28 Sep 2021 • Ruotian Ma, Xin Zhou, Tao Gui, Yiding Tan, Linyang Li, Qi Zhang, Xuanjing Huang
Prompt-based methods have been successfully applied in sentence-level few-shot learning tasks, mostly owing to the sophisticated design of templates and label words.
1 code implementation • 26 Sep 2021 • Tianxiang Sun, Xiangyang Liu, Xipeng Qiu, Xuanjing Huang
In this paper, we review such phenomenon of paradigm shifts in recent years, highlighting several paradigms that have the potential to solve different NLP tasks.
no code implementations • 12 Sep 2021 • Zhihao Fan, Zhongyu Wei, Zejun Li, Siyuan Wang, Haijun Shan, Xuanjing Huang, Jianqing Fan
Existing research for image text retrieval mainly relies on sentence-level supervision to distinguish matched and mismatched sentences for a query image.
no code implementations • 10 Sep 2021 • Yitao Liu, Tianxiang Sun, Xipeng Qiu, Xuanjing Huang
This one-way interaction leads to the teacher's inability to perceive the characteristics of the student and its training progress.
no code implementations • 19 Aug 2021 • Tong Liu, Siyuan Wang, Jingchao Fu, Lei Chen, Zhongyu Wei, Yaqi Liu, Heng Ye, Liaosa Xu, Weiqiang Wan, Xuanjing Huang
Existing system dealing with online complaint provides a final decision without explanations.
no code implementations • ACL 2021 • Xinyi Mou, Zhongyu Wei, Lei Chen, Shangyi Ning, Yancheng He, Changjian Jiang, Xuanjing Huang
In addition, we propose a novel task, namely hashtag usage prediction to model the ideology of legislators on Twitter.
1 code implementation • ACL 2021 • Qinzhuo Wu, Qi Zhang, Zhongyu Wei, Xuanjing Huang
In recent years, math word problem solving has received considerable attention and achieved promising results, but previous methods rarely take numerical values into consideration.
no code implementations • ACL 2021 • Yi Zhou, Xiaoqing Zheng, Cho-Jui Hsieh, Kai-Wei Chang, Xuanjing Huang
Although deep neural networks have achieved prominent performance on many NLP tasks, they are vulnerable to adversarial examples.
1 code implementation • ACL 2021 • Ruotian Ma, Tao Gui, Linyang Li, Qi Zhang, Yaqian Zhou, Xuanjing Huang
In this work, we propose the use of negative training (NT), in which a model is trained using complementary labels regarding that ``the instance does not belong to these complementary labels".
no code implementations • 21 Jun 2021 • Zhihao Fan, Zhongyu Wei, Siyuan Wang, Ruize Wang, Zejun Li, Haijun Shan, Xuanjing Huang
Considering that theme concepts can be learned from both images and captions, we propose two settings for their representations learning based on TTN.
1 code implementation • ACL 2021 • Jinlan Fu, Xuanjing Huang, PengFei Liu
Recent years have seen the paradigm shift of Named Entity Recognition (NER) systems from sequence labeling to span prediction.
Ranked #4 on
Named Entity Recognition
on WNUT 2017
1 code implementation • ACL 2021 • Chong Li, Cenyuan Zhang, Xiaoqing Zheng, Xuanjing Huang
A sequence-to-sequence learning with neural networks has empirically proven to be an effective framework for Chinese Spelling Correction (CSC), which takes a sentence with some spelling errors as input and outputs the corrected one.
1 code implementation • ACL 2021 • Xiaonan Li, Yunfan Shao, Tianxiang Sun, Hang Yan, Xipeng Qiu, Xuanjing Huang
To alleviate this problem, we extend the recent successful early-exit mechanism to accelerate the inference of PTMs for sequence labeling tasks.
no code implementations • 28 May 2021 • Tianxiang Sun, Yunhua Zhou, Xiangyang Liu, Xinyu Zhang, Hao Jiang, Zhao Cao, Xuanjing Huang, Xipeng Qiu
In this paper, we show that a novel objective function for the training of the ensemble internal classifiers can be naturally induced from the perspective of ensemble learning and information theory.
1 code implementation • 8 May 2021 • Jiehang Zeng, Xiaoqing Zheng, Jianhan Xu, Linyang Li, Liping Yuan, Xuanjing Huang
Recently, few certified defense methods have been developed to provably guarantee the robustness of a text classifier to adversarial synonym substitutions.
no code implementations • 17 Apr 2021 • Lu Ji, Jing Li, Zhongyu Wei, Qi Zhang, Xuanjing Huang
Numerous online conversations are produced on a daily basis, resulting in a pressing need to conversation understanding.
no code implementations • NAACL 2021 • Jinlan Fu, Liangjing Feng, Qi Zhang, Xuanjing Huang, PengFei Liu
The development of neural networks and pretraining techniques has spawned many sentence-level tagging systems that achieved superior performance on typical benchmarks.
1 code implementation • 7 Apr 2021 • Chenxin An, Ming Zhong, Yiran Chen, Danqing Wang, Xipeng Qiu, Xuanjing Huang
Previous work for text summarization in scientific domain mainly focused on the content of the input document, but seldom considering its citation network.
no code implementations • NAACL 2021 • Zhihao Fan, Yeyun Gong, Dayiheng Liu, Zhongyu Wei, Siyuan Wang, Jian Jiao, Nan Duan, Ruofei Zhang, Xuanjing Huang
We therefore introduce a new layer named dynamic mask attention network (DMAN) with a learnable mask matrix which is able to model localness adaptively.
Ranked #9 on
Machine Translation
on IWSLT2014 German-English
no code implementations • 22 Mar 2021 • Liping Yuan, Jiangtao Feng, Xiaoqing Zheng, Xuanjing Huang
The key idea is that at each time step, the network takes as input a ``bundle'' of similar words predicted at the previous step instead of a single ground truth.
1 code implementation • ACL 2021 • Tao Gui, Xiao Wang, Qi Zhang, Qin Liu, Yicheng Zou, Xin Zhou, Rui Zheng, Chong Zhang, Qinzhuo Wu, Jiacheng Ye, Zexiong Pang, Yongxin Zhang, Zhengyan Li, Ruotian Ma, Zichu Fei, Ruijian Cai, Jun Zhao, Xingwu Hu, Zhiheng Yan, Yiding Tan, Yuan Hu, Qiyuan Bian, Zhihua Liu, Bolin Zhu, Shan Qin, Xiaoyu Xing, Jinlan Fu, Yue Zhang, Minlong Peng, Xiaoqing Zheng, Yaqian Zhou, Zhongyu Wei, Xipeng Qiu, Xuanjing Huang
To guarantee user acceptability, all the text transformations are linguistically based, and we provide a human evaluation for each one.
no code implementations • 21 Mar 2021 • Zejun Li, Zhongyu Wei, Zhihao Fan, Haijun Shan, Xuanjing Huang
In this paper, we focus on the problem of unsupervised image-sentence matching.
no code implementations • 29 Dec 2020 • Linyang Li, Yunfan Shao, Demin Song, Xipeng Qiu, Xuanjing Huang
The substitutions in the generated adversarial examples are not characters or words but \textit{'pieces'}, which are more natural to Chinese readers.
1 code implementation • EMNLP 2020 • Tao Gui, Jiacheng Ye, Qi Zhang, Zhengyan Li, Zichu Fei, Yeyun Gong, Xuanjing Huang
Conditional random fields (CRF) for label decoding has become ubiquitous in sequence labeling tasks.
1 code implementation • 14 Dec 2020 • Yicheng Zou, Jun Lin, Lujun Zhao, Yangyang Kang, Zhuoren Jiang, Changlong Sun, Qi Zhang, Xuanjing Huang, Xiaozhong Liu
Automatic chat summarization can help people quickly grasp important information from numerous chat messages.
1 code implementation • 14 Dec 2020 • Yicheng Zou, Lujun Zhao, Yangyang Kang, Jun Lin, Minlong Peng, Zhuoren Jiang, Changlong Sun, Qi Zhang, Xuanjing Huang, Xiaozhong Liu
In a customer service system, dialogue summarization can boost service efficiency by automatically creating summaries for long spoken dialogues in which customers and agents try to address issues about specific topics.
no code implementations • 12 Dec 2020 • Yichao Luo, Zhengyan Li, Bingning Wang, Xiaoyu Xing, Qi Zhang, Xuanjing Huang
Keyphrase Generation (KG) is the task of generating central topics from a given document or literary work, which captures the crucial information necessary to understand the content.
no code implementations • COLING 2020 • Lei Chen, Zhongyu Wei, Jing Li, Baohua Zhou, Qi Zhang, Xuanjing Huang
Previous work for rumor resolution concentrates on exploiting time-series characteristics or modeling topology structure separately.
no code implementations • COLING 2020 • Zhihao Fan, Yeyun Gong, Zhongyu Wei, Siyuan Wang, Yameng Huang, Jian Jiao, Xuanjing Huang, Nan Duan, Ruofei Zhang
Commonsense generation aims at generating plausible everyday scenario description based on a set of provided concepts.
no code implementations • 16 Nov 2020 • Jingjing Gong, Hang Yan, Yining Zheng, Xipeng Qiu, Xuanjing Huang
A lot of natural language processing problems need to encode the text sequence as a fix-length vector, which usually involves aggregation process of combining the representations of all the words, such as pooling or self-attention.
1 code implementation • EMNLP 2020 • Jinlan Fu, PengFei Liu, Qi Zhang, Xuanjing Huang
The performance of the Chinese Word Segmentation (CWS) systems has gradually reached a plateau with the rapid development of deep neural networks, especially the successful use of large pre-trained models.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Lu Liu, Yi Zhou, Jianhan Xu, Xiaoqing Zheng, Kai-Wei Chang, Xuanjing Huang
The words in each sentence of a source language corpus are rearranged to meet the word order in a target language under the guidance of a part-of-speech based language model (LM).
no code implementations • Findings of the Association for Computational Linguistics 2020 • Yanjian Zhang, Qin Chen, Yiteng Zhang, Zhongyu Wei, Yixu Gao, Jiajie Peng, Zengfeng Huang, Weijian Sun, Xuanjing Huang
Terms contained in Gene Ontology (GO) have been widely used in biology and bio-medicine.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Minlong Peng, Ruotian Ma, Qi Zhang, Lujun Zhao, Mengxi Wei, Changlong Sun, Xuanjing Huang
In this work, we explore the way to quickly adjust an existing named entity recognition (NER) system to make it capable of recognizing entity types not defined in the system.
2 code implementations • Findings of the Association for Computational Linguistics 2020 • Yiran Chen, PengFei Liu, Ming Zhong, Zi-Yi Dou, Danqing Wang, Xipeng Qiu, Xuanjing Huang
In this paper, we perform an in-depth analysis of characteristics of different datasets and investigate the performance of different summarization models under a cross-dataset setting, in which a summarizer trained on one corpus will be evaluated on a range of out-of-domain corpora.
1 code implementation • COLING 2020 • Tianxiang Sun, Yunfan Shao, Xipeng Qiu, Qipeng Guo, Yaru Hu, Xuanjing Huang, Zheng Zhang
With the emerging branch of incorporating factual knowledge into pre-trained language models such as BERT, most existing models consider shallow, static, and separately pre-trained entity embeddings, which limits the performance gains of these models.
1 code implementation • ACL 2021 • Zhichao Geng, Hang Yan, Xipeng Qiu, Xuanjing Huang
The joint-model is trained and evaluated on 13 corpora of four tasks, yielding near state-of-the-art (SOTA) performance in dependency parsing and NER, achieving SOTA performance in CWS and POS.
1 code implementation • EMNLP 2020 • Xiaoyu Xing, Zhijing Jin, Di Jin, Bingning Wang, Qi Zhang, Xuanjing Huang
Based on the SemEval 2014 dataset, we construct the Aspect Robustness Test Set (ARTS) as a comprehensive probe of the aspect robustness of ABSA models.
no code implementations • ACL 2020 • Xiaoqing Zheng, Jiehang Zeng, Yi Zhou, Cho-Jui Hsieh, Minhao Cheng, Xuanjing Huang
Despite achieving prominent performance on many important tasks, it has been reported that neural networks are vulnerable to adversarial examples.
1 code implementation • 20 Jun 2020 • Yi Zhou, Xiaoqing Zheng, Cho-Jui Hsieh, Kai-Wei Chang, Xuanjing Huang
Despite neural networks have achieved prominent performance on many natural language processing (NLP) tasks, they are vulnerable to adversarial examples.
no code implementations • EMNLP 2020 • Ruize Wang, Duyu Tang, Nan Duan, Wanjun Zhong, Zhongyu Wei, Xuanjing Huang, Daxin Jiang, Ming Zhou
We study the detection of propagandistic text fragments in news articles.
3 code implementations • 29 Apr 2020 • Kangenbei Liao, Qianlong Liu, Zhongyu Wei, Baolin Peng, Qin Chen, Weijian Sun, Xuanjing Huang
In this paper, we focus on automatic disease diagnosis with reinforcement learning (RL) methods in task-oriented dialogues setting.
1 code implementation • ACL 2020 • Danqing Wang, PengFei Liu, Yining Zheng, Xipeng Qiu, Xuanjing Huang
An intuitive way is to put them in the graph-based neural network, which has a more complex structure for capturing inter-sentence relationships.
1 code implementation • ACL 2020 • Xiaonan Li, Hang Yan, Xipeng Qiu, Xuanjing Huang
Recently, the character-word lattice structure has been proved to be effective for Chinese named entity recognition (NER) by incorporating the word information.
Ranked #2 on
Chinese Named Entity Recognition
on MSRA
2 code implementations • ACL 2020 • Ming Zhong, PengFei Liu, Yiran Chen, Danqing Wang, Xipeng Qiu, Xuanjing Huang
This paper creates a paradigm shift with regard to the way we build neural extractive summarization systems.
Ranked #1 on
Text Summarization
on BBC XSum
no code implementations • 13 Apr 2020 • Zhen Ke, Liang Shi, Erli Meng, Bin Wang, Xipeng Qiu, Xuanjing Huang
Besides, the pre-trained BERT language model has been also introduced into the MCCWS task in a multi-task learning framework.
3 code implementations • 18 Mar 2020 • Xipeng Qiu, Tianxiang Sun, Yige Xu, Yunfan Shao, Ning Dai, Xuanjing Huang
Recently, the emergence of pre-trained models (PTMs) has brought natural language processing (NLP) to a new era.
1 code implementation • 24 Feb 2020 • Yige Xu, Xipeng Qiu, Ligao Zhou, Xuanjing Huang
Fine-tuning pre-trained language models like BERT has become an effective way in NLP and yields state-of-the-art results on many downstream tasks.
1 code implementation • Findings (ACL) 2021 • Ruize Wang, Duyu Tang, Nan Duan, Zhongyu Wei, Xuanjing Huang, Jianshu ji, Guihong Cao, Daxin Jiang, Ming Zhou
We study the problem of injecting knowledge into large pre-trained models like BERT and RoBERTa.
Ranked #1 on
Entity Typing
on Open Entity
1 code implementation • 12 Jan 2020 • Jinlan Fu, PengFei Liu, Qi Zhang, Xuanjing Huang
While neural network-based models have achieved impressive performance on a large body of NLP tasks, the generalization behavior of different models remains poorly understood: Does this excellent performance imply a perfect generalization model, or are there still some limitations?
1 code implementation • 17 Dec 2019 • Yi Zhou, Xiaoqing Zheng, Xuanjing Huang
Inspired by a concept of content-addressable retrieval from cognitive science, we propose a novel fragment-based model augmented with a lexicon-based memory for Chinese NER, in which both the character-level and word-level features are combined to generate better feature representations for possible name candidates.
1 code implementation • 18 Nov 2019 • Tao Gui, Lizhi Qing, Qi Zhang, Jiacheng Ye, HangYan, Zichu Fei, Xuanjing Huang
In order to effectively reduce the impact of non-ideal auxiliary tasks on the main task, we further proposed a novel meta-learning-based multi-task learning approach, which trained the shared hidden layers on auxiliary tasks, while the meta-optimization objective was to minimize the loss on the main task, ensuring that the optimizing direction led to an improvement on the main task.
1 code implementation • 12 Nov 2019 • Tianxiang Sun, Yunfan Shao, Xiaonan Li, PengFei Liu, Hang Yan, Xipeng Qiu, Xuanjing Huang
Most existing deep multi-task learning models are based on parameter sharing, such as hard sharing, hierarchical sharing, and soft sharing.
no code implementations • COLING 2020 • Ruize Wang, Zhongyu Wei, Ying Cheng, Piji Li, Haijun Shan, Ji Zhang, Qi Zhang, Xuanjing Huang
Visual storytelling aims to generate a narrative paragraph from a sequence of images automatically.
1 code implementation • NAACL 2021 • Lu Ji, Zhongyu Wei, Jing Li, Qi Zhang, Xuanjing Huang
In this paper, we focus on extracting interactive argument pairs from two posts with opposite stances to a certain topic.
no code implementations • IJCNLP 2019 • Tao Gui, Yicheng Zou, Qi Zhang, Minlong Peng, Jinlan Fu, Zhongyu Wei, Xuanjing Huang
Recurrent neural networks (RNN) used for Chinese named entity recognition (NER) that sequentially track character and word information have achieved great success.
Ranked #10 on
Chinese Named Entity Recognition
on OntoNotes 4
no code implementations • IJCNLP 2019 • Di Liang, Fubao Zhang, Qi Zhang, Xuanjing Huang
However, in the process of reasoning, the role of the two sentences is obviously different, and the sentence pairs for NLI are asymmetrical corpora.
no code implementations • WS 2019 • Ming Zhong, Danqing Wang, PengFei Liu, Xipeng Qiu, Xuanjing Huang
In this paper, we take stock of the current state of summarization datasets and explore how different factors of datasets influence the generalization behaviour of neural extractive summarization models.
no code implementations • 25 Sep 2019 • Jinlan Fu, PengFei Liu, Xuanjing Huang
With the proliferation of models for natural language processing (NLP) tasks, it is even harder to understand the differences between models and their relative merits.
no code implementations • COLING 2020 • Minlong Peng, Qi Zhang, Xuanjing Huang
To address this problem, we propose a modification to DIRL, obtaining a novel weighted domain-invariant representation learning (WDIRL) framework.
no code implementations • 30 Aug 2019 • Danqing Wang, PengFei Liu, Ming Zhong, Jie Fu, Xipeng Qiu, Xuanjing Huang
Although domain shift has been well explored in many NLP applications, it still has received little attention in the domain of extractive text summarization.
3 code implementations • IJCNLP 2019 • Luyao Huang, Chi Sun, Xipeng Qiu, Xuanjing Huang
Word Sense Disambiguation (WSD) aims to find the exact sense of an ambiguous word in a particular context.
Ranked #3 on
Word Sense Disambiguation
on WiC-TSV
2 code implementations • ACL 2020 • Ruotian Ma, Minlong Peng, Qi Zhang, Xuanjing Huang
This method avoids designing a complicated sequence modeling architecture, and for any neural NER model, it requires only subtle adjustment of the character representation layer to introduce the lexicon information.
Ranked #6 on
Chinese Named Entity Recognition
on Resume NER
no code implementations • 25 Jul 2019 • Lin Zehui, PengFei Liu, Luyao Huang, Junkun Chen, Xipeng Qiu, Xuanjing Huang
Variants dropout methods have been designed for the fully-connected layer, convolutional layer and recurrent layer in neural networks, and shown to be effective to avoid overfitting.
2 code implementations • ACL 2019 • Ming Zhong, PengFei Liu, Danqing Wang, Xipeng Qiu, Xuanjing Huang
The recent years have seen remarkable success in the use of deep neural networks on text summarization.
Ranked #6 on
Extractive Text Summarization
on CNN / Daily Mail
no code implementations • ACL 2019 • Zhenqiao Song, Xiaoqing Zheng, Lu Liu, Mu Xu, Xuanjing Huang
It is desirable for dialog systems to have capability to express specific emotions during a conversation, which has a direct, quantifiable impact on improvement of their usability and user satisfaction.
1 code implementation • ACL 2019 • Zhihao Fan, Zhongyu Wei, Siyuan Wang, Xuanjing Huang
Existing research usually employs the architecture of CNN-RNN that views the generation as a sequential decision-making process and the entire dataset vocabulary is used as decoding space.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Xipeng Qiu, Hengzhi Pei, Hang Yan, Xuanjing Huang
Multi-criteria Chinese word segmentation (MCCWS) aims to exploit the relations among the multiple heterogeneous segmentation criteria and further improve the performance of each single criterion.
1 code implementation • ACL 2019 • Minlong Peng, Xiaoyu Xing, Qi Zhang, Jinlan Fu, Xuanjing Huang
In this work, we explore the way to perform named entity recognition (NER) using only unlabeled data and named entity dictionaries.
1 code implementation • 29 May 2019 • Minlong Peng, Qi Zhang, Xiaoyu Xing, Tao Gui, Jinlan Fu, Xuanjing Huang
However, representations of unseen or rare words trained on the end task are usually poor for appreciable performance.
15 code implementations • 14 May 2019 • Chi Sun, Xipeng Qiu, Yige Xu, Xuanjing Huang
Language model pre-training has proven to be useful in learning universal language representations.
Ranked #1 on
Text Classification
on Yahoo! Answers
4 code implementations • ACL 2019 • Ning Dai, Jianze Liang, Xipeng Qiu, Xuanjing Huang
Disentangling the content and style in the latent space is prevalent in unpaired text style transfer.
1 code implementation • TACL 2020 • Hang Yan, Xipeng Qiu, Xuanjing Huang
Our graph-based joint model achieves better performance than previous joint models and state-of-the-art results in both Chinese word segmentation and dependency parsing.
1 code implementation • NAACL 2019 • Chi Sun, Xipeng Qiu, Xuanjing Huang
Chinese is a logographic writing system, and the shape of Chinese characters contain rich syntactic and semantic information.
no code implementations • 21 Nov 2018 • Pengfei Liu, Shuaichen Chang, Xuanjing Huang, Jian Tang, Jackie Chi Kit Cheung
Recently, a large number of neural mechanisms and models have been proposed for sequence learning, of which self-attention, as exemplified by the Transformer model, and graph neural networks (GNNs) have attracted much attention.
1 code implementation • 9 Nov 2018 • Tao Gui, Qi Zhang, Lujun Zhao, Yaosong Lin, Minlong Peng, Jingjing Gong, Xuanjing Huang
In recent years, long short-term memory (LSTM) has been successfully used to model sequential data of variable length.
Ranked #22 on
Sentiment Analysis
on IMDb
no code implementations • WS 2018 • Yunfan Gu, Zhongyu Wei, Maoran Xu, Hao Fu, Yang Liu, Xuanjing Huang
In this paper, we propose to incorporate topic aspects information for online comments convincingness evaluation.
no code implementations • 23 Oct 2018 • Pengfei Liu, Xuanjing Huang
In this paper, we describe a general framework: Parameters Read-Write Networks (PRaWNs) to systematically analyze current neural models for multi-task learning, in which we find that existing models expect to disentangle features into different spaces while features learned in practice are still entangled in shared space, leaving potential hazards for other training or unseen tasks.
no code implementations • EMNLP 2018 • Tao Gui, Qi Zhang, Jingjing Gong, Minlong Peng, Di Liang, Keyu Ding, Xuanjing Huang
However, from a linguistic perspective, Twitter users not only tend to mimic the formal expressions of traditional media, like news, but they also appear to be developing linguistically informal styles.
Ranked #2 on
Part-Of-Speech Tagging
on Ritter
no code implementations • EMNLP 2018 • Jingjing Gong, Xipeng Qiu, Xinchi Chen, Dong Liang, Xuanjing Huang
Attention-based neural models have achieved great success in natural language inference (NLI).
no code implementations • EMNLP 2018 • Yucheng Wang, Zhongyu Wei, Yaqian Zhou, Xuanjing Huang
Automatic essay scoring (AES) is the task of assigning grades to essays without human interference.
no code implementations • 24 Sep 2018 • Shuyang Cao, Xipeng Qiu, Xuanjing Huang
Neural architecture for named entity recognition has achieved great success in the field of natural language processing.
no code implementations • 23 Aug 2018 • Junkun Chen, Kaiyu Chen, Xinchi Chen, Xipeng Qiu, Xuanjing Huang
Designing shared neural architecture plays an important role in multi-task learning.
no code implementations • 21 Aug 2018 • Chi Sun, Hang Yan, Xipeng Qiu, Xuanjing Huang
Therefore, with the aim of representing words in a highly efficient way, we propose to operate a Gaussian word embedding model with a loss function based on the Wasserstein distance.
no code implementations • COLING 2018 • Yicheng Zou, Tao Gui, Qi Zhang, Xuanjing Huang
Attention mechanisms have been leveraged for sentiment classification tasks because not all words have the same importance.
1 code implementation • COLING 2018 • Lu Ji, Zhongyu Wei, Xiangkun Hu, Yang Liu, Qi Zhang, Xuanjing Huang
In this paper, we investigate the issue of persuasiveness evaluation for argumentative comments.
no code implementations • COLING 2018 • Zhihao Fan, Zhongyu Wei, Siyuan Wang, Yang Liu, Xuanjing Huang
Visual Question Generation (VQG) aims to ask natural questions about an image automatically.
no code implementations • ACL 2018 • Zhongyu Wei, Qianlong Liu, Baolin Peng, Huaixiao Tou, Ting Chen, Xuanjing Huang, Kam-Fai Wong, Xiangying Dai
In this paper, we make a move to build a dialogue system for automatic diagnosis.
no code implementations • ACL 2018 • Minlong Peng, Qi Zhang, Yu-Gang Jiang, Xuanjing Huang
And we introduce a few target domain labeled data for learning domain-specific information.
2 code implementations • COLING 2018 • Jingjing Gong, Xipeng Qiu, Shaojing Wang, Xuanjing Huang
The dynamic routing policy is dynamically deciding that what and how much information need be transferred from each word to the final encoding of the text sequence.
Ranked #33 on
Sentiment Analysis
on IMDb
3 code implementations • 30 Apr 2018 • Zhan Shi, Xinchi Chen, Xipeng Qiu, Xuanjing Huang
Similar to the adversarial models, the reward and policy function in IRL are optimized alternately.
no code implementations • 25 Feb 2018 • Junkun Chen, Xipeng Qiu, Pengfei Liu, Xuanjing Huang
Specifically, we use a shared meta-network to capture the meta-knowledge of semantic composition and generate the parameters of the task-specific semantic composition models.
no code implementations • 25 Feb 2018 • Jinyue Su, Jiacheng Xu, Xipeng Qiu, Xuanjing Huang
Generating plausible and fluent sentence with desired properties has long been a challenge.
no code implementations • NeurIPS 2017 • Yuanbin Wu, Man Lan, Shiliang Sun, Qi Zhang, Xuanjing Huang
In this work, we try to understand the differences between exact and approximate inference algorithms in structured prediction.
no code implementations • EMNLP 2017 • Tao Gui, Qi Zhang, Haoran Huang, Minlong Peng, Xuanjing Huang
In this work, we study the problem of part-of-speech tagging for Tweets.
Ranked #3 on
Part-Of-Speech Tagging
on Ritter
no code implementations • EMNLP 2017 • Pengfei Liu, Kaiyu Qian, Xipeng Qiu, Xuanjing Huang
Idioms are peculiar linguistic constructions that impose great challenges for representing the semantics of language, especially in current prevailing end-to-end neural models, which assume that the semantics of a phrase or sentence can be literally composed from its constitutive words.
no code implementations • 2 Jul 2017 • Xinchi Chen, Zhan Shi, Xipeng Qiu, Xuanjing Huang
In this paper, we propose a new neural model to incorporate the word-level information for Chinese word segmentation.
1 code implementation • 9 Jun 2017 • Xipeng Qiu, Jingjing Gong, Xuanjing Huang
In this paper, we give an overview for the shared task at the CCF Conference on Natural Language Processing \& Chinese Computing (NLPCC 2017): Chinese News Headline Categorization.
no code implementations • 11 May 2017 • Pengfei Liu, Xipeng Qiu, Xuanjing Huang
Tree-structured neural networks have proven to be effective in learning semantic representations by exploiting syntactic information.
no code implementations • ACL 2017 • Xinchi Chen, Zhan Shi, Xipeng Qiu, Xuanjing Huang
Different linguistic perspectives causes many diverse segmentation criteria for Chinese word segmentation (CWS).
no code implementations • ACL 2017 • Pengfei Liu, Xipeng Qiu, Xuanjing Huang
Neural network models have shown their promising opportunities for multi-task learning, which focus on learning the shared layers to extract the common and task-invariant features.
1 code implementation • COLING 2016 • Yatian Shen, Xuanjing Huang
Nowadays, neural networks play an important role in the task of relation classification.
Ranked #15 on
Relation Extraction
on SemEval-2010 Task 8
no code implementations • COLING 2016 • Haoran Huang, Qi Zhang, Yeyun Gong, Xuanjing Huang
By incorporating the hierarchical attention mechanism, the relative improvement in the proposed method over the state-of-the-art method is around 67. 9{\%} in the F1-score.
no code implementations • 26 Nov 2016 • Jiacheng Xu, Kan Chen, Xipeng Qiu, Xuanjing Huang
In this paper, we propose a novel deep architecture to utilize both structural and textual information of entities.
no code implementations • 16 Nov 2016 • Xinchi Chen, Xipeng Qiu, Xuanjing Huang
Recently, neural network models for natural language processing tasks have been increasingly focused on for their ability of alleviating the burden of manual feature engineering.
no code implementations • 15 Nov 2016 • Jingjing Gong, Xinchi Chen, Xipeng Qiu, Xuanjing Huang
However, it is nontrivial for pair-wise models to incorporate the contextual sentence information.
no code implementations • 23 Sep 2016 • Pengfei Liu, Xipeng Qiu, Xuanjing Huang
Neural network based models have achieved impressive results on various specific tasks.
no code implementations • 20 Aug 2016 • Jifan Chen, Kan Chen, Xipeng Qiu, Qi Zhang, Xuanjing Huang, Zheng Zhang
To prove the effectiveness of our model, we evaluate it on four tasks, including word similarity, reverse dictionaries, Wiki link prediction, and document classification.
no code implementations • 23 Jul 2016 • Xinchi Chen, Xipeng Qiu, Xuanjing Huang
Sentence ordering is a general and critical task for natural language generation applications.
no code implementations • 22 Jul 2016 • PengFei Liu, Xipeng Qiu, Xuanjing Huang
Introducing attentional mechanism in neural network is a powerful concept, and has achieved impressive results in many natural language processing tasks.
no code implementations • EMNLP 2016 • Pengfei Liu, Xipeng Qiu, Xuanjing Huang
Recently, there is rising interest in modelling the interactions of two sentences with deep neural networks.
Ranked #70 on
Natural Language Inference
on SNLI
1 code implementation • 17 May 2016 • Pengfei Liu, Xipeng Qiu, Xuanjing Huang
Neural network based methods have obtained great progress on a variety of natural language processing tasks.
no code implementations • 22 Apr 2016 • Peng Qian, Xipeng Qiu, Xuanjing Huang
Recently, the long short-term memory neural network (LSTM) has attracted wide interest due to its success in many tasks.
no code implementations • 19 Nov 2015 • Xinchi Chen, Xipeng Qiu, Jingxiang Jiang, Xuanjing Huang
In this paper, we propose the Gaussian mixture skip-gram (GMSG) model to learn the Gaussian mixture embeddings for words based on skip-gram framework.
no code implementations • 28 May 2015 • Xipeng Qiu, Peng Qian, Liusong Yin, Shiyu Wu, Xuanjing Huang
In this paper, we give an overview for the shared task at the 4th CCF Conference on Natural Language Processing \& Chinese Computing (NLPCC 2015): Chinese word segmentation and part-of-speech (POS) tagging for micro-blog texts.
no code implementations • IJCNLP 2015 • Chenxi Zhu, Xipeng Qiu, Xinchi Chen, Xuanjing Huang
In this work, we address the problem to model all the nodes (words or phrases) in a dependency tree with the dense representations.