2 code implementations • EMNLP 2020 • Tiangang Zhu, Yue Wang, Haoran Li, Youzheng Wu, Xiaodong He, Bo-Wen Zhou
We annotate a multimodal product attribute value dataset that contains 87, 194 instances, and the experimental results on this dataset demonstrate that explicitly modeling the relationship between attributes and values facilitates our method to establish the correspondence between them, and selectively utilizing visual product information is necessary for the task.
no code implementations • ACL 2020 • Song Xu, Haoran Li, Peng Yuan, Youzheng Wu, Xiaodong He, Bo-Wen Zhou
Copy module has been widely equipped in the recent abstractive summarization models, which facilitates the decoder to extract words from the source into the summary.
no code implementations • 13 Apr 2020 • Tae Jin Park, Kyu J. Han, Jing Huang, Xiaodong He, Bo-Wen Zhou, Panayiotis Georgiou, Shrikanth Narayanan
This work presents a novel approach for speaker diarization to leverage lexical information provided by automatic speech recognition.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+4
no code implementations • 4 Apr 2020 • Ming Tu, Jing Huang, Xiaodong He, Bo-Wen Zhou
We validate the proposed GSN on two NLP tasks: interpretable multi-hop reading comprehension on HotpotQA and graph based fact verification on FEVER.
no code implementations • LREC 2020 • Meng Chen, Ruixue Liu, Lei Shen, Shaozu Yuan, Jingyan Zhou, Youzheng Wu, Xiaodong He, Bo-Wen Zhou
Human conversations are complicated and building a human-like dialogue agent is an extremely challenging task.
no code implementations • ACL 2020 • Yun Tang, Jing Huang, Guangtao Wang, Xiaodong He, Bo-Wen Zhou
Translational distance-based knowledge graph embedding has shown progressive improvements on the link prediction task, from TransE to the latest state-of-the-art RotatE.
Ranked #19 on
Link Prediction
on FB15k-237
1 code implementation • 1 Nov 2019 • Ming Tu, Kevin Huang, Guangtao Wang, Jing Huang, Xiaodong He, Bo-Wen Zhou
Interpretable multi-hop reading comprehension (RC) over multiple documents is a challenging problem because it demands reasoning over multiple information sources and explaining the answer prediction by providing supporting evidences.
no code implementations • CONLL 2019 • Kevin Huang, Yun Tang, Jing Huang, Xiaodong He, Bo-Wen Zhou
We test the relation module on the SQuAD 2. 0 dataset using both the BiDAF and BERT models as baseline readers.
no code implementations • 23 Oct 2019 • Kevin Huang, Yun Tang, Jing Huang, Xiaodong He, Bo-Wen Zhou
In this paper, we aim to improve a MRC model's ability to determine whether a question has an answer in a given context (e. g. the recently proposed SQuAD 2. 0 task).
1 code implementation • 29 Aug 2019 • Shuaichen Chang, PengFei Liu, Yun Tang, Jing Huang, Xiaodong He, Bo-Wen Zhou
Recent years have seen great success in the use of neural seq2seq models on the text-to-SQL task.
no code implementations • 12 Jun 2019 • Ming Tu, Jing Huang, Xiaodong He, Bo-Wen Zhou
In this paper, we propose a new end-to-end graph neural network (GNN) based algorithm for MIL: we treat each bag as a graph and use GNN to learn the bag embedding, in order to explore the useful structural information among instances in bags.
no code implementations • 28 May 2019 • Pengcheng Li, Jin-Feng Yi, Bo-Wen Zhou, Lijun Zhang
In this paper, we improve the robustness of DNNs by utilizing techniques of Distance Metric Learning.
no code implementations • ACL 2019 • Ming Tu, Guangtao Wang, Jing Huang, Yun Tang, Xiaodong He, Bo-Wen Zhou
We introduce a heterogeneous graph with different types of nodes and edges, which is named as Heterogeneous Document-Entity (HDE) graph.
no code implementations • 22 Apr 2019 • Lan-Zhe Guo, Yu-Feng Li, Ming Li, Jin-Feng Yi, Bo-Wen Zhou, Zhi-Hua Zhou
We guide the optimization of label quality through a small amount of validation data, and to ensure the safeness of performance while maximizing performance gain.
no code implementations • 21 Feb 2019 • Yun Tang, Guohong Ding, Jing Huang, Xiaodong He, Bo-Wen Zhou
This paper aims to improve the widely used deep speaker embedding x-vector model.
no code implementations • 26 Jan 2019 • Yu Cheng, Mo Yu, Xiaoxiao Guo, Bo-Wen Zhou
Our meta metric learning approach consists of task-specific learners, that exploit metric learning to handle flexible labels, and a meta learner, that discovers good parameters and gradient decent to specify the metrics in task-specific learners.
1 code implementation • 11 Nov 2018 • Chao Shang, Yun Tang, Jing Huang, Jinbo Bi, Xiaodong He, Bo-Wen Zhou
The recent graph convolutional network (GCN) provides another way of learning graph node embedding by successfully utilizing graph connectivity structure.
Ranked #30 on
Link Prediction
on FB15k-237
no code implementations • ICLR 2019 • Zaiyi Chen, Zhuoning Yuan, Jin-Feng Yi, Bo-Wen Zhou, Enhong Chen, Tianbao Yang
For example, there is still a lack of theories of convergence for SGD and its variants that use stagewise step size and return an averaged solution in practice.
3 code implementations • 16 Jun 2018 • Wenhan Xiong, Xiaoxiao Guo, Mo Yu, Shiyu Chang, Bo-Wen Zhou, William Yang Wang
We investigate the task of learning to follow natural language instructions by jointly reasoning with visual observations and language inputs.
2 code implementations • NAACL 2018 • Mo Yu, Xiaoxiao Guo, Jin-Feng Yi, Shiyu Chang, Saloni Potdar, Yu Cheng, Gerald Tesauro, Haoyu Wang, Bo-Wen Zhou
We study few-shot learning in natural language domains.
no code implementations • ACL 2017 • Mingbo Ma, Liang Huang, Bing Xiang, Bo-Wen Zhou
Question classification is an important task with wide applications.
no code implementations • 28 Sep 2017 • Mingbo Ma, Kai Zhao, Liang Huang, Bing Xiang, Bo-Wen Zhou
In order to utilize the potential benefits from their correlations, we propose a jointly trained model for learning the two tasks simultaneously via Long Short-Term Memory (LSTM) networks.
no code implementations • 19 Sep 2017 • Wei Zhang, Bo-Wen Zhou
Learning to remember long sequences remains a challenging task for recurrent neural networks.
1 code implementation • 31 Aug 2017 • Shuohang Wang, Mo Yu, Xiaoxiao Guo, Zhiguo Wang, Tim Klinger, Wei zhang, Shiyu Chang, Gerald Tesauro, Bo-Wen Zhou, Jing Jiang
Second, we propose a novel method that jointly trains the Ranker along with an answer-generation Reader model, based on reinforcement learning.
Ranked #4 on
Open-Domain Question Answering
on Quasar
no code implementations • 26 Aug 2017 • Mo Yu, Xiaoxiao Guo, Jin-Feng Yi, Shiyu Chang, Saloni Potdar, Gerald Tesauro, Haoyu Wang, Bo-Wen Zhou
We propose a new method to measure task similarities with cross-task transfer performance matrix for the deep learning scenario.
no code implementations • 1 Aug 2017 • Ramesh Nallapati, Igor Melnyk, Abhishek Kumar, Bo-Wen Zhou
We present a new topic model that generates documents by sampling a topic for one whole sentence at a time, and generating the words in the sentence using an RNN decoder that is conditioned on the topic of the sentence.
no code implementations • 7 Jul 2017 • Cicero Nogueira dos Santos, Kahini Wadhawan, Bo-Wen Zhou
We propose discriminative adversarial networks (DAN) for semi-supervised learning and loss function learning.
no code implementations • ACL 2017 • Mo Yu, Wenpeng Yin, Kazi Saidul Hasan, Cicero dos Santos, Bing Xiang, Bo-Wen Zhou
Relation detection is a core component for many NLP applications including Knowledge Base Question Answering (KBQA).
52 code implementations • 9 Mar 2017 • Zhouhan Lin, Minwei Feng, Cicero Nogueira dos santos, Mo Yu, Bing Xiang, Bo-Wen Zhou, Yoshua Bengio
This paper proposes a new model for extracting an interpretable sentence embedding by introducing self-attention.
1 code implementation • 15 Jan 2017 • Feifei Zhai, Saloni Potdar, Bing Xiang, Bo-Wen Zhou
Many natural language understanding (NLU) tasks, such as shallow parsing (i. e., text chunking) and semantic slot filling, require the assignment of representative labels to the meaningful chunks in a sentence.
no code implementations • 18 Nov 2016 • Wei Zhang, Minwei Feng, Yunhui Zheng, Yufei Ren, Yandong Wang, Ji Liu, Peng Liu, Bing Xiang, Li Zhang, Bo-Wen Zhou, Fei Wang
By evaluating the NLC workloads, we show that only the conservative hyper-parameter setup (e. g., small mini-batch size and small learning rate) can guarantee acceptable model accuracy for a wide range of customers.
no code implementations • 14 Nov 2016 • Ramesh Nallapati, Bo-Wen Zhou, Mingbo Ma
The Selector architecture, on the other hand, is free to pick one sentence at a time in any arbitrary order to piece together the summary.
7 code implementations • 14 Nov 2016 • Ramesh Nallapati, FeiFei Zhai, Bo-Wen Zhou
We present SummaRuNNer, a Recurrent Neural Network (RNN) based sequence model for extractive summarization of documents and show that it achieves performance better than or comparable to state-of-the-art.
Ranked #8 on
Text Summarization
on CNN / Daily Mail (Anonymized)
no code implementations • 31 Oct 2016 • Yang Yu, Wei zhang, Kazi Hasan, Mo Yu, Bing Xiang, Bo-Wen Zhou
This paper proposes dynamic chunk reader (DCR), an end-to-end neural reading comprehension (RC) model that is able to extract and rank a set of answer candidates from a given document to answer questions.
Ranked #49 on
Question Answering
on SQuAD1.1 dev
no code implementations • COLING 2016 • Wenpeng Yin, Mo Yu, Bing Xiang, Bo-Wen Zhou, Hinrich Schütze
In fact selection, we match the subject entity in a fact candidate with the entity mention in the question by a character-level convolutional neural network (char-CNN), and match the predicate in that fact with the question by a word-level CNN (word-CNN).
4 code implementations • 2 Jun 2016 • Iulian Vlad Serban, Tim Klinger, Gerald Tesauro, Kartik Talamadupula, Bo-Wen Zhou, Yoshua Bengio, Aaron Courville
We introduce the multiresolution recurrent neural network, which extends the sequence-to-sequence framework to model natural language generation as two parallel discrete stochastic processes: a sequence of high-level coarse tokens, and a sequence of natural language tokens.
Ranked #1 on
Dialogue Generation
on Ubuntu Dialogue (Activity)
no code implementations • ACL 2016 • Caglar Gulcehre, Sungjin Ahn, Ramesh Nallapati, Bo-Wen Zhou, Yoshua Bengio
At each time-step, the decision of which softmax layer to use choose adaptively made by an MLP which is conditioned on the context.~We motivate our work from a psychological evidence that humans naturally have a tendency to point towards objects in the context or the environment when the name of an object is not known.~We observe improvements on two tasks, neural machine translation on the Europarl English to French parallel corpora and text summarization on the Gigaword dataset using our proposed model.
4 code implementations • CONLL 2016 • Ramesh Nallapati, Bo-Wen Zhou, Cicero Nogueira dos santos, Caglar Gulcehre, Bing Xiang
In this work, we model abstractive text summarization using Attentional Encoder-Decoder Recurrent Neural Networks, and show that they achieve state-of-the-art performance on two different corpora.
Ranked #10 on
Text Summarization
on DUC 2004 Task 1
3 code implementations • 11 Feb 2016 • Cicero dos Santos, Ming Tan, Bing Xiang, Bo-Wen Zhou
In this work, we propose Attentive Pooling (AP), a two-way attention mechanism for discriminative model training.
Ranked #2 on
Question Answering
on SemEvalCQA
no code implementations • EMNLP 2016 • Gakuto Kurata, Bing Xiang, Bo-Wen Zhou, Mo Yu
Recurrent Neural Network (RNN) and one of its specific architectures, Long Short-Term Memory (LSTM), have been widely used for sequence labeling.
8 code implementations • TACL 2016 • Wenpeng Yin, Hinrich Schütze, Bing Xiang, Bo-Wen Zhou
(ii) We propose three attention schemes that integrate mutual influence between sentences into CNN; thus, the representation of each sentence takes into consideration its counterpart.
no code implementations • 19 Nov 2015 • James Cross, Bing Xiang, Bo-Wen Zhou
We propose two methods of learning vector representations of words and phrases that each combine sentence context with structural features extracted from dependency trees.
2 code implementations • 12 Nov 2015 • Ming Tan, Cicero dos Santos, Bing Xiang, Bo-Wen Zhou
One direction is to define a more composite representation for questions and answers by combining convolutional neural network with the basic framework.
no code implementations • 3 Nov 2015 • Minwei Feng, Bing Xiang, Bo-Wen Zhou
This paper is an empirical study of the distributed deep learning for question answering subtasks: answer selection and question classification.
no code implementations • 26 Oct 2015 • Yang Yu, Wei zhang, Chung-Wei Hang, Bing Xiang, Bo-Wen Zhou
In this paper we explore deep learning models with memory component or attention mechanism for question answering task.
no code implementations • 14 Oct 2015 • Wei Zhang, Yang Yu, Bo-Wen Zhou
Neural Turing Machines (NTM) contain memory component that simulates "working memory" in the brain to store and retrieve information to ease simple algorithms learning.
2 code implementations • 7 Aug 2015 • Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bo-Wen Zhou
We apply a general deep learning framework to address the non-factoid question answering task.
1 code implementation • IJCNLP 2015 • Mingbo Ma, Liang Huang, Bing Xiang, Bo-Wen Zhou
In sentence modeling and classification, convolutional neural network approaches have recently achieved state-of-the-art results, but all such efforts process word vectors sequentially and neglect long-distance dependencies.
no code implementations • 1 Jun 2015 • Chang Wang, Liangliang Cao, Bo-Wen Zhou
In this paper, we present a novel approach for medical synonym extraction.
2 code implementations • IJCNLP 2015 • Cicero Nogueira dos Santos, Bing Xiang, Bo-Wen Zhou
Relation classification is an important semantic processing task for which state-ofthe-art systems still rely on costly handcrafted features.
Ranked #27 on
Relation Extraction
on SemEval-2010 Task-8