An interpretable LSTM neural network for autoregressive exogenous model

14 Apr 2018 Tian Guo Tao Lin Yao Lu

In this paper, we propose an interpretable LSTM recurrent neural network, i.e., multi-variable LSTM for time series with exogenous variables. Currently, widely used attention mechanism in recurrent neural networks mostly focuses on the temporal aspect of data and falls short of characterizing variable importance... (read more)

PDF Abstract
No code implementations yet. Submit your code now


  Add Datasets introduced or used in this paper

Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper

Graph Embeddings
Sigmoid Activation
Activation Functions
Tanh Activation
Activation Functions
Recurrent Neural Networks