Disconnected Recurrent Neural Networks for Text Categorization

ACL 2018  ·  Baoxin Wang ·

Recurrent neural network (RNN) has achieved remarkable performance in text categorization. RNN can model the entire sequence and capture long-term dependencies, but it does not do well in extracting key patterns. In contrast, convolutional neural network (CNN) is good at extracting local and position-invariant features. In this paper, we present a novel model named disconnected recurrent neural network (DRNN), which incorporates position-invariance into RNN. By limiting the distance of information flow in RNN, the hidden state at each time step is restricted to represent words near the current position. The proposed model makes great improvements over RNN and CNN models and achieves the best performance on several benchmark datasets for text categorization.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Text Classification AG News DRNN Error 5.53 # 5
Sentiment Analysis Amazon Review Full DRNN Accuracy 64.43 # 3
Sentiment Analysis Amazon Review Polarity DRNN Accuracy 96.49 # 4
Text Classification DBpedia DRNN Error 0.81 # 8
Text Classification Yahoo! Answers DRNN Accuracy 76.26 # 2
Sentiment Analysis Yelp Binary classification DRNN Error 2.73 # 9
Sentiment Analysis Yelp Fine-grained classification DRNN Error 30.85 # 7

Methods


No methods listed for this paper. Add relevant methods here