no code implementations • 17 May 2023 • Daowan Peng, Wei Wei, Xian-Ling Mao, Yuanyuan Fu, Dangyang Chen
Generalization beyond in-domain experience to out-of-distribution data is of paramount significance in the AI domain.
no code implementations • 15 May 2023 • Zewen Chi, Heyan Huang, Xian-Ling Mao
Recent studies have exhibited remarkable capabilities of pre-trained multilingual Transformers, especially cross-lingual transferability.
no code implementations • 4 May 2023 • Sen Zhao1, Wei Wei, Yifan Liu, Ziyang Wang, Wendi Li, Xian-Ling Mao, Shuai Zhu, Minghui Yang, Zujie Wen
Conversational recommendation systems (CRS) aim to timely and proactively acquire user dynamic preferred attributes through conversations for item recommendation.
1 code implementation • 3 May 2023 • Yuxiang Nie, Heyan Huang, Wei Wei, Xian-Ling Mao
To alleviate the problem, it might be possible to generate long-document QA pairs via unsupervised question answering (UQA) methods.
no code implementations • COLING 2022 • Ziming Huang, Zhuoxuan Jiang, Ke Wang, Juntao Li, Shanshan Feng, Xian-Ling Mao
Although most existing methods can fulfil this requirement, they can only model single-source dialog data and cannot effectively capture the underlying knowledge of relations among data and subtasks.
no code implementations • 19 Dec 2022 • Xiao Zhang, Heyan Huang, Zewen Chi, Xian-Ling Mao
Open-retrieval conversational machine reading comprehension (OCMRC) simulates real-life conversational interaction scenes.
1 code implementation • 5 Dec 2022 • Tian Lan, Yixuan Su, Shuhang Liu, Heyan Huang, Xian-Ling Mao
In this study, we formulate open-ended text generation from a new perspective, i. e., we view it as an exploration process within a directed graph.
no code implementations • 28 Nov 2022 • Shuo Liang, Wei Wei, Xian-Ling Mao, Yuanyuan Fu, Rui Fang, Dangyang Chen
Hence, we propose a novel approach, Span TAgging and Greedy infErence (STAGE), to extract sentiment triplets in span-level, where each span may consist of multiple words and play different roles simultaneously.
no code implementations • 17 Oct 2022 • Xiaofei Wen, Wei Wei, Xian-Ling Mao
To address the problem, in this paper we propose a novel approach, named Sequential Global Topic Attention (SGTA) to exploit topic transition over all conversations in a subtle way for better modeling post-to-response topic-transition and guiding the response generation to the current conversation.
no code implementations • 17 Oct 2022 • Ruihan Zhang, Wei Wei, Xian-Ling Mao, Rui Fang, Dangyang Chen
Conventional event detection models under supervised learning settings suffer from the inability of transfer to newly-emerged event types owing to lack of sufficient annotations.
1 code implementation • 11 Oct 2022 • Yuxiang Nie, Heyan Huang, Wei Wei, Xian-Ling Mao
The proposed model mainly focuses on the evidence selection phase of long document question answering.
1 code implementation • COLING 2022 • Xiao Zhang, Heyan Huang, Zewen Chi, Xian-Ling Mao
Conversational machine reading comprehension (CMRC) aims to assist computers to understand an natural language text and thereafter engage in a multi-turn conversation to answer questions related to the text.
1 code implementation • 23 Sep 2022 • Rong-Cheng Tu, Xian-Ling Mao, Kevin Qinghong Lin, Chengfei Cai, Weize Qin, Hongfa Wang, Wei Wei, Heyan Huang
Recently, to improve the unsupervised image retrieval performance, plenty of unsupervised hashing methods have been proposed by designing a semantic similarity matrix, which is based on the similarities between image features extracted by a pre-trained CNN model.
no code implementations • 27 Aug 2022 • Ziyang Wang, Huoyu Liu, Wei Wei, Yue Hu, Xian-Ling Mao, Shaojian He, Rui Fang, Dangyang Chen
Different from the previous contrastive learning-based methods for SR, MCLSR learns the representations of users and items through a cross-view contrastive learning paradigm from four specific views at two different levels (i. e., interest- and feature-level).
1 code implementation • COLING 2022 • Yuxiang Nie, Heyan Huang, Zewen Chi, Xian-Ling Mao
Previous works usually make use of heuristic rules as well as pre-trained models to construct data and train QA models.
1 code implementation • 22 Aug 2022 • Ding Zou, Wei Wei, Ziyang Wang, Xian-Ling Mao, Feida Zhu, Rui Fang, Dangyang Chen
Specifically, we first construct local and non-local graphs for user/item in KG, exploring more KG facts for KGR.
1 code implementation • 18 Jun 2022 • Ziyang Wang, Wei Wei, Chenwei Xu, Jun Xu, Xian-Ling Mao
Existing studies on person-job fit, however, mainly focus on calculating the similarity between the candidate resumes and the job postings on the basis of their contents, without taking the recruiters' experience (i. e., historical successful recruitment records) into consideration.
no code implementations • 11 May 2022 • Yu-Ming Shang, Heyan Huang, Xin Sun, Wei Wei, Xian-Ling Mao
Extracting relational triples from unstructured text is an essential task in natural language processing and knowledge graph construction.
2 code implementations • 20 Apr 2022 • Zewen Chi, Li Dong, Shaohan Huang, Damai Dai, Shuming Ma, Barun Patra, Saksham Singhal, Payal Bajaj, Xia Song, Xian-Ling Mao, Heyan Huang, Furu Wei
We also present a comprehensive analysis on the representation and routing behaviors of our models.
1 code implementation • 19 Apr 2022 • Ding Zou, Wei Wei, Xian-Ling Mao, Ziyang Wang, Minghui Qiu, Feida Zhu, Xin Cao
Different from traditional contrastive learning methods which generate two graph views by uniform data augmentation schemes such as corruption or dropping, we comprehensively consider three different graph views for KG-aware recommendation, including global-level structural view, local-level collaborative and semantic views.
1 code implementation • ACL 2022 • Heqi Zheng, Xiao Zhang, Zewen Chi, Heyan Huang, Tan Yan, Tian Lan, Wei Wei, Xian-Ling Mao
In this paper, we propose XPR, a cross-lingual phrase retriever that extracts phrase representations from unlabeled example sentences.
1 code implementation • Findings (ACL) 2022 • Shuo Liang, Wei Wei, Xian-Ling Mao, Fei Wang, Zhiyong He
Aspect-based sentiment analysis (ABSA) is a fine-grained sentiment analysis task that aims to align aspects and corresponding sentiments for aspect-specific sentiment polarity inference.
no code implementations • 6 Apr 2022 • Sheng-Fu Wang, Shu-Hang Liu, Tian-Yi Che, Yi-Fan Lu, Song-Xiao Yang, Heyan Huang, Xian-Ling Mao
Specifically, taking a paper as a basic and separate unit, existing PDF Readers cannot access extended information about the paper, such as corresponding videos, blogs and codes.
no code implementations • 10 Mar 2022 • Yu-Ming Shang, Heyan Huang, Xian-Ling Mao
Joint entity and relation extraction is an essential task in natural language processing and knowledge graph construction.
1 code implementation • 13 Oct 2021 • Tian Lan, Deng Cai, Yan Wang, Yixuan Su, Heyan Huang, Xian-Ling Mao
In this study, we present a solution to directly select proper responses from a large corpus or even a nonparallel corpus that only consists of unpaired sentences, using a dense retrieval model.
no code implementations • 23 Sep 2021 • Zewen Chi, Heyan Huang, Luyang Liu, Yu Bai, Xian-Ling Mao
The success of pretrained cross-lingual language models relies on two essential abilities, i. e., generalization ability for learning downstream tasks in a source language, and cross-lingual transferability for transferring the task knowledge to other languages.
1 code implementation • Findings (EMNLP) 2021 • Weiran Pan, Wei Wei, Xian-Ling Mao
Knowledge graph entity typing aims to infer entities' missing types in knowledge graphs which is an important but under-explored issue.
3 code implementations • ACL 2022 • Zewen Chi, Shaohan Huang, Li Dong, Shuming Ma, Bo Zheng, Saksham Singhal, Payal Bajaj, Xia Song, Xian-Ling Mao, Heyan Huang, Furu Wei
In this paper, we introduce ELECTRA-style tasks to cross-lingual language model pre-training.
Ranked #1 on
Zero-Shot Cross-Lingual Transfer
on XTREME
1 code implementation • ACL 2021 • Zewen Chi, Li Dong, Bo Zheng, Shaohan Huang, Xian-Ling Mao, Heyan Huang, Furu Wei
The cross-lingual language models are typically pretrained with masked language modeling on multilingual text or parallel sentences.
1 code implementation • 9 Jun 2021 • Ziyang Wang, Wei Wei, Gao Cong, Xiao-Li Li, Xian-Ling Mao, Minghui Qiu
In GCE-GNN, we propose a novel global-level item representation learning layer, which employs a session-aware attention mechanism to recursively incorporate the neighbors' embeddings of each node on the global graph.
1 code implementation • Findings (ACL) 2021 • Heng-Da Xu, Zhongli Li, Qingyu Zhou, Chao Li, Zizhen Wang, Yunbo Cao, Heyan Huang, Xian-Ling Mao
Chinese Spell Checking (CSC) aims to detect and correct erroneous characters for user-generated text in the Chinese language.
Ranked #2 on
Chinese Spell Checking
on SIGHAN 2015
1 code implementation • ACL 2021 • Puhai Yang, Heyan Huang, Xian-Ling Mao
Thus, in this paper, we will study and discuss how the context information of different granularity affects dialogue state tracking.
1 code implementation • 2 Jan 2021 • Houjin Yu, Xian-Ling Mao, Zewen Chi, Wei Wei, Heyan Huang
Recently, it has attracted much attention to build reliable named entity recognition (NER) systems using limited annotated data.
Ranked #3 on
Named Entity Recognition (NER)
on SciERC
(using extra training data)
Low Resource Named Entity Recognition
named-entity-recognition
+2
no code implementations • 21 Dec 2020 • Tian Lan, Xian-Ling Mao, Zhipeng Zhao, Wei Wei, Heyan Huang
Since the pre-trained language models are widely used, retrieval-based open-domain dialog systems, have attracted considerable attention from researchers recently.
1 code implementation • 17 Dec 2020 • Tian Lan, Xian-Ling Mao, Xiaoyan Gao, Wei Wei, Heyan Huang
Specifically, in our proposed DSHC model, a hashing optimizing module that consists of two autoencoder models is stacked on a trained dense representation model, and three loss functions are designed to optimize it.
no code implementations • 10 Dec 2020 • Ziyang Wang, Wei Wei, Xian-Ling Mao, Xiao-Li Li, Shanshan Feng
In RNMSR, we propose to learn the user preference from both instance-level and group-level, respectively: (i) instance-level, which employs GNNs on a similarity-based item-pairwise session graph to capture the users' preference in instance-level.
no code implementations • 20 Nov 2020 • Ziyang Wang, Wei Wei, Gao Cong, Xiao-Li Li, Xian-Ling Mao, Minghui Qiu, Shanshan Feng
Based on BGNN, we propose a novel approach, called Session-based Recommendation with Global Information (SRGI), which infers the user preferences via fully exploring global item-transitions over all sessions from two different perspectives: (i) Fusion-based Model (SRGI-FM), which recursively incorporates the neighbor embeddings of each node on global graph into the learning process of session level item representation; and (ii) Constrained-based Model (SRGI-CM), which treats the global-level item-transition information as a constraint to ensure the learned item embeddings are consistent with the global item-transition.
no code implementations • 16 Nov 2020 • Ziyang Wang, Wei Wei, Xian-Ling Mao, Guibing Guo, Pan Zhou, Shanshan Feng
Due to the huge commercial interests behind online reviews, a tremendousamount of spammers manufacture spam reviews for product reputation manipulation.
no code implementations • 6 Nov 2020 • Rong-Cheng Tu, Xian-Ling Mao, Rongxin Tu, Binbin Bian, Wei Wei, Heyan Huang
Finally, by minimizing the novel \textit{margin-dynamic-softmax loss}, the modality-specific hashing networks can be trained to generate hash codes which can simultaneously preserve the cross-modal similarity and abundant semantic information well.
no code implementations • 26 Oct 2020 • Jia-Nan Guo, Xian-Ling Mao, Shu-Yang Lin, Wei Wei, Heyan Huang
However, nearly all the existing network embedding based methods are hard to capture the actual category features of a node because of the linearly inseparable problem in low-dimensional space; meanwhile they cannot incorporate simultaneously network structure information and node label information into network embedding.
no code implementations • 7 Aug 2020 • Tian Lan, Xian-Ling Mao, Wei Wei, He-Yan Huang
Thus, in this paper, we will measure systematically nearly all representative hierarchical and non-hierarchical models over the same experimental settings to check which kind is better.
4 code implementations • NAACL 2021 • Zewen Chi, Li Dong, Furu Wei, Nan Yang, Saksham Singhal, Wenhui Wang, Xia Song, Xian-Ling Mao, He-Yan Huang, Ming Zhou
In this work, we present an information-theoretic framework that formulates cross-lingual language model pre-training as maximizing mutual information between multilingual-multi-granularity texts.
Ranked #16 on
Zero-Shot Cross-Lingual Transfer
on XTREME
no code implementations • 3 Jul 2020 • Heng-Da Xu, Xian-Ling Mao, Zewen Chi, Jing-Jing Zhu, Fanshu Sun, He-Yan Huang
Specifically, KW-Seq2Seq first uses a keywords decoder to predict some topic keywords, and then generates the final response under the guidance of them.
1 code implementation • EMNLP (NLP-COVID19) 2020 • Yong Hu, He-Yan Huang, Anfan Chen, Xian-Ling Mao
Therefore, in this paper, we release Weibo-COV, a first large-scale COVID-19 social media dataset from Weibo, covering more than 30 million tweets from 1 November 2019 to 30 April 2020.
Social and Information Networks
1 code implementation • 8 May 2020 • Puhai Yang, He-Yan Huang, Xian-Ling Mao
As a key component in a dialogue system, dialogue state tracking plays an important role.
no code implementations • 21 Apr 2020 • Yuming Shang, Heyan Huang, Xin Sun, Xian-Ling Mao
Then, we borrow the idea of Coulomb's Law from physics and introduce the concept of attractive force and repulsive force to this graph to learn correlation and mutual exclusion between relations.
1 code implementation • 6 Apr 2020 • Tian Lan, Xian-Ling Mao, Wei Wei, Xiaoyan Gao, He-Yan Huang
Through extensive experiments, the learning-based metrics are demonstrated that they are the most effective evaluation metrics for open-domain generative dialogue systems.
no code implementations • 20 Dec 2019 • Tian Lan, Xian-Ling Mao, He-Yan Huang, Wei Wei
Intuitively, a dialogue model that can control the timing of talking autonomously based on the conversation context can chat with humans more naturally.
no code implementations • 11 Nov 2019 • Zhuoxuan Jiang, Ziming Huang, Dong Sheng Li, Xian-Ling Mao
In this paper, we propose a novel joint end-to-end model by multi-task representation learning, which can capture the knowledge from heterogeneous information through automatically learning knowledgeable low-dimensional embeddings from data, named with DialogAct2Vec.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Zewen Chi, Li Dong, Furu Wei, Xian-Ling Mao, He-Yan Huang
Multilingual pretrained language models (such as multilingual BERT) have achieved impressive results for cross-lingual transfer.
no code implementations • 8 Nov 2019 • Tan Yan, He-Yan Huang, Xian-Ling Mao
We introduce a new scientific named entity recognizer called SEPT, which stands for Span Extractor with Pre-trained Transformers.
1 code implementation • 23 Sep 2019 • Zewen Chi, Li Dong, Furu Wei, Wenhui Wang, Xian-Ling Mao, He-Yan Huang
In this work we focus on transferring supervision signals of natural language generation (NLG) tasks between multiple languages.
no code implementations • 17 Sep 2019 • Tian Lan, Xian-Ling Mao, He-Yan Huang
As far as we know, the existing task-oriented dialogue systems obtain the dialogue policy through classification, which can assign either a dialogue act and its corresponding parameters or multiple dialogue acts without their corresponding parameters for a dialogue action.
no code implementations • 5 Sep 2019 • Wei Wei, Ling Cheng, Xian-Ling Mao, Guangyou Zhou, Feida Zhu
Recently, automatic image caption generation has been an important focus of the work on multimodal translation task.
1 code implementation • 25 Aug 2019 • Yong Hu, He-Yan Huang, Tian Lan, Xiaochi Wei, Yuxiang Nie, Jiarui Qi, Liner Yang, Xian-Ling Mao
Second language acquisition (SLA) modeling is to predict whether second language learners could correctly answer the questions according to what they have learned.
no code implementations • 24 Aug 2019 • Wei Wei, Zanbo Wang, Xian-Ling Mao, Guangyou Zhou, Pan Zhou, Sheng Jiang
Sequence labeling is a fundamental task in natural language processing and has been widely studied.
no code implementations • WS 2019 • Zhuoxuan Jiang, Xian-Ling Mao, Ziming Huang, Jie Ma, Shaochun Li
Learning an efficient manager of dialogue agent from data with little manual intervention is important, especially for goal-oriented dialogues.
1 code implementation • 13 Aug 2019 • Zewen Chi, He-Yan Huang, Heng-Da Xu, Houjin Yu, Wanxuan Yin, Xian-Ling Mao
It also attracts lots of attention to recognize the table structures in PDF files.
no code implementations • 12 Aug 2019 • Jia-Nan Guo, Xian-Ling Mao, Xiao-Jian Jiang, Ying-Xiang Sun, Wei Wei, He-Yan Huang
Network embedding is a promising way of network representation, facilitating many signed social network processing and analysis tasks such as link prediction and node classification.
no code implementations • 29 Jul 2019 • Rong-Cheng Tu, Xian-Ling Mao, Bing Ma, Yong Hu, Tan Yan, Wei Wei, He-Yan Huang
Specifically, by an iterative optimization algorithm, DCHUC jointly learns unified hash codes for image-text pairs in a database and a pair of hash functions for unseen query image-text pairs.
no code implementations • 24 Nov 2018 • Rong-Cheng Tu, Xian-Ling Mao, Bo-Si Feng, Bing-Bing Bian, Yu-shu Ying
Recently, similarity-preserving hashing methods have been extensively studied for large-scale image retrieval.
no code implementations • 7 Apr 2017 • Dan Wang, He-Yan Huang, Chi Lu, Bo-Si Feng, Liqiang Nie, Guihua Wen, Xian-Ling Mao
Specifically, we define a novel similarity formula for hierarchical labeled data by weighting each layer, and design a deep convolutional neural network to obtain a hash code for each data point.
no code implementations • 7 Apr 2017 • Yi-Kun Tang, Xian-Ling Mao, He-Yan Huang, Guihua Wen
Recently, topic modeling has been widely used to discover the abstract topics in text corpora.
no code implementations • COLING 2016 • Xian-Ling Mao, Yi-Jing Hao, Qiang Zhou, Wen-Qing Yuan, Liner Yang, He-Yan Huang
Recently, topic modeling has been widely applied in data mining due to its powerful ability.