no code implementations • EMNLP 2021 • Tianfu Zhang, Heyan Huang, Chong Feng, Longbing Cao
Multi-head self-attention recently attracts enormous interest owing to its specialized functions, significant parallelizable computation, and flexible extensibility.
no code implementations • 16 May 2023 • Bo wang, Heyan Huang, Xiaochi Wei, Ge Shi, Xiao Liu, Chong Feng, Tong Zhou, Shuaiqiang Wang, Dawei Yin
Event extraction aims to recognize pre-defined event triggers and arguments from texts, which suffer from the lack of high-quality annotations.
no code implementations • 15 May 2023 • Zewen Chi, Heyan Huang, Xian-Ling Mao
Recent studies have exhibited remarkable capabilities of pre-trained multilingual Transformers, especially cross-lingual transferability.
1 code implementation • 3 May 2023 • Yuxiang Nie, Heyan Huang, Wei Wei, Xian-Ling Mao
To alleviate the problem, it might be possible to generate long-document QA pairs via unsupervised question answering (UQA) methods.
no code implementations • 19 Dec 2022 • Xiao Zhang, Heyan Huang, Zewen Chi, Xian-Ling Mao
Open-retrieval conversational machine reading comprehension (OCMRC) simulates real-life conversational interaction scenes.
1 code implementation • 5 Dec 2022 • Tian Lan, Yixuan Su, Shuhang Liu, Heyan Huang, Xian-Ling Mao
In this study, we formulate open-ended text generation from a new perspective, i. e., we view it as an exploration process within a directed graph.
no code implementations • 8 Nov 2022 • Jinta Weng, Yue Hu, Zhihong Tian, Heyan Huang
The effectiveness of proposed ConsPrompt is demonstrated in five different few-shot learning tasks and shown the similarity-based sampling strategy is more effective than label-based in combining contrastive learning.
1 code implementation • 3 Nov 2022 • Peiyuan Gong, Xuebo Liu, Heyan Huang, Min Zhang
Pretraining-based (PT-based) automatic evaluation metrics (e. g., BERTScore and BARTScore) have been widely used in several sentence generation tasks (e. g., machine translation and text summarization) due to their better correlation with human judgments over traditional overlap-based methods.
1 code implementation • 11 Oct 2022 • Yuxiang Nie, Heyan Huang, Wei Wei, Xian-Ling Mao
The proposed model mainly focuses on the evidence selection phase of long document question answering.
1 code implementation • COLING 2022 • Xiao Zhang, Heyan Huang, Zewen Chi, Xian-Ling Mao
Conversational machine reading comprehension (CMRC) aims to assist computers to understand an natural language text and thereafter engage in a multi-turn conversation to answer questions related to the text.
1 code implementation • 23 Sep 2022 • Rong-Cheng Tu, Xian-Ling Mao, Kevin Qinghong Lin, Chengfei Cai, Weize Qin, Hongfa Wang, Wei Wei, Heyan Huang
Recently, to improve the unsupervised image retrieval performance, plenty of unsupervised hashing methods have been proposed by designing a semantic similarity matrix, which is based on the similarities between image features extracted by a pre-trained CNN model.
1 code implementation • COLING 2022 • Yuxiang Nie, Heyan Huang, Zewen Chi, Xian-Ling Mao
Previous works usually make use of heuristic rules as well as pre-trained models to construct data and train QA models.
no code implementations • 15 Jun 2022 • Mukhlis Amien, Chong Feng, Heyan Huang
Twitter contains an abundance of linguistic data from the real world.
no code implementations • ACL 2022 • Xiao Liu, Heyan Huang, Ge Shi, Bo wang
We consider event extraction in a generative manner with template-based conditional generation.
no code implementations • 11 May 2022 • Yu-Ming Shang, Heyan Huang, Xin Sun, Wei Wei, Xian-Ling Mao
Extracting relational triples from unstructured text is an essential task in natural language processing and knowledge graph construction.
no code implementations • 27 Apr 2022 • Yizhe Yang, Yang Gao, Jiawei Li, Heyan Huang
Besides, a Ground Graph Aware Transformer ($G^2AT$) is proposed to enhance knowledge grounded response generation.
2 code implementations • 20 Apr 2022 • Zewen Chi, Li Dong, Shaohan Huang, Damai Dai, Shuming Ma, Barun Patra, Saksham Singhal, Payal Bajaj, Xia Song, Xian-Ling Mao, Heyan Huang, Furu Wei
We also present a comprehensive analysis on the representation and routing behaviors of our models.
1 code implementation • ACL 2022 • Heqi Zheng, Xiao Zhang, Zewen Chi, Heyan Huang, Tan Yan, Tian Lan, Wei Wei, Xian-Ling Mao
In this paper, we propose XPR, a cross-lingual phrase retriever that extracts phrase representations from unlabeled example sentences.
no code implementations • COLING 2022 • Xiaochen Liu, Yang Gao, Yu Bai, Jiawei Li, Yinan Hu, Heyan Huang, Boxing Chen
Few-shot abstractive summarization has become a challenging task in natural language generation.
no code implementations • 6 Apr 2022 • Sheng-Fu Wang, Shu-Hang Liu, Tian-Yi Che, Yi-Fan Lu, Song-Xiao Yang, Heyan Huang, Xian-Ling Mao
Specifically, taking a paper as a basic and separate unit, existing PDF Readers cannot access extended information about the paper, such as corresponding videos, blogs and codes.
1 code implementation • 31 Mar 2022 • Mingjie Chen, Yanghao Zhou, Heyan Huang, Thomas Hain
It was shown recently that a combination of ASR and TTS models yield highly competitive performance on standard voice conversion tasks such as the Voice Conversion Challenge 2020 (VCC2020).
1 code implementation • CCL 2022 • Mucheng Ren, Heyan Huang, Yuxiang Zhou, Qianwen Cao, Yuan Bu, Yang Gao
Therefore, in this paper, we focus on the core task of the TCM diagnosis and treatment system -- syndrome differentiation (SD) -- and we introduce the first public large-scale dataset for SD, called TCM-SD.
no code implementations • 10 Mar 2022 • Yu-Ming Shang, Heyan Huang, Xian-Ling Mao
Joint entity and relation extraction is an essential task in natural language processing and knowledge graph construction.
1 code implementation • 15 Oct 2021 • Yu Bai, Heyan Huang, Kai Fan, Yang Gao, Yiming Zhu, Jiaao Zhan, Zewen Chi, Boxing Chen
Through introducing compression rate, the information ratio between the source and the target text, we regard the MT task as a special CLS task with a compression rate of 100%.
1 code implementation • 13 Oct 2021 • Tian Lan, Deng Cai, Yan Wang, Yixuan Su, Heyan Huang, Xian-Ling Mao
In this study, we present a solution to directly select proper responses from a large corpus or even a nonparallel corpus that only consists of unpaired sentences, using a dense retrieval model.
no code implementations • 23 Sep 2021 • Zewen Chi, Heyan Huang, Luyang Liu, Yu Bai, Xian-Ling Mao
The success of pretrained cross-lingual language models relies on two essential abilities, i. e., generalization ability for learning downstream tasks in a source language, and cross-lingual transferability for transferring the task knowledge to other languages.
3 code implementations • ACL 2022 • Zewen Chi, Shaohan Huang, Li Dong, Shuming Ma, Bo Zheng, Saksham Singhal, Payal Bajaj, Xia Song, Xian-Ling Mao, Heyan Huang, Furu Wei
In this paper, we introduce ELECTRA-style tasks to cross-lingual language model pre-training.
Ranked #1 on
Zero-Shot Cross-Lingual Transfer
on XTREME
1 code implementation • ACL 2021 • Zewen Chi, Li Dong, Bo Zheng, Shaohan Huang, Xian-Ling Mao, Heyan Huang, Furu Wei
The cross-lingual language models are typically pretrained with masked language modeling on multilingual text or parallel sentences.
no code implementations • Findings (ACL) 2021 • Mucheng Ren, Heyan Huang, Yang Gao
Qualitative relationships illustrate how changing one property (e. g., moving velocity) affects another (e. g., kinetic energy) and constitutes a considerable portion of textual knowledge.
1 code implementation • ACL 2021 • Yu Bai, Yang Gao, Heyan Huang
Employing one unified decoder to generate the sequential concatenation of monolingual and cross-lingual summaries, MCLAS makes the monolingual summarization task a prerequisite of the cross-lingual summarization (CLS) task.
Abstractive Text Summarization
Cross-Lingual Abstractive Summarization
+1
1 code implementation • Findings (ACL) 2021 • Heng-Da Xu, Zhongli Li, Qingyu Zhou, Chao Li, Zizhen Wang, Yunbo Cao, Heyan Huang, Xian-Ling Mao
Chinese Spell Checking (CSC) aims to detect and correct erroneous characters for user-generated text in the Chinese language.
Ranked #2 on
Chinese Spell Checking
on SIGHAN 2015
1 code implementation • ACL 2021 • Puhai Yang, Heyan Huang, Xian-Ling Mao
Thus, in this paper, we will study and discuss how the context information of different granularity affects dialogue state tracking.
1 code implementation • EMNLP 2021 • Zewen Chi, Li Dong, Shuming Ma, Shaohan Huang Xian-Ling Mao, Heyan Huang, Furu Wei
Multilingual T5 (mT5) pretrains a sequence-to-sequence model on massive monolingual texts, which has shown promising results on many cross-lingual tasks.
1 code implementation • 2 Jan 2021 • Houjin Yu, Xian-Ling Mao, Zewen Chi, Wei Wei, Heyan Huang
Recently, it has attracted much attention to build reliable named entity recognition (NER) systems using limited annotated data.
Ranked #3 on
Named Entity Recognition (NER)
on SciERC
(using extra training data)
Low Resource Named Entity Recognition
named-entity-recognition
+2
no code implementations • 1 Jan 2021 • Xiao Liu, Heyan Huang, Yue Zhang
News-driven stock prediction investigates the correlation between news events and stock price movements.
no code implementations • 21 Dec 2020 • Tian Lan, Xian-Ling Mao, Zhipeng Zhao, Wei Wei, Heyan Huang
Since the pre-trained language models are widely used, retrieval-based open-domain dialog systems, have attracted considerable attention from researchers recently.
1 code implementation • 17 Dec 2020 • Tian Lan, Xian-Ling Mao, Xiaoyan Gao, Wei Wei, Heyan Huang
Specifically, in our proposed DSHC model, a hashing optimizing module that consists of two autoencoder models is stacked on a trained dense representation model, and three loss functions are designed to optimize it.
no code implementations • 6 Nov 2020 • Rong-Cheng Tu, Xian-Ling Mao, Rongxin Tu, Binbin Bian, Wei Wei, Heyan Huang
Finally, by minimizing the novel \textit{margin-dynamic-softmax loss}, the modality-specific hashing networks can be trained to generate hash codes which can simultaneously preserve the cross-modal similarity and abundant semantic information well.
no code implementations • 26 Oct 2020 • Jia-Nan Guo, Xian-Ling Mao, Shu-Yang Lin, Wei Wei, Heyan Huang
However, nearly all the existing network embedding based methods are hard to capture the actual category features of a node because of the linearly inseparable problem in low-dimensional space; meanwhile they cannot incorporate simultaneously network structure information and node label information into network embedding.
no code implementations • 21 Oct 2020 • Puhai Yang, Heyan Huang, Xianling Mao
Scalability for handling unknown slot values is a important problem in dialogue state tracking (DST).
1 code implementation • EMNLP 2020 • Mucheng Ren, Xiubo Geng, Tao Qin, Heyan Huang, Daxin Jiang
We focus on the task of reasoning over paragraph effects in situation, which requires a model to understand the cause and effect described in a background paragraph, and apply the knowledge to a novel situation.
no code implementations • 21 Apr 2020 • Yuming Shang, Heyan Huang, Xin Sun, Xian-Ling Mao
Then, we borrow the idea of Coulomb's Law from physics and introduce the concept of attractive force and repulsive force to this graph to learn correlation and mutual exclusion between relations.