Dynamic Evaluation of Neural Sequence Models

We present methodology for using dynamic evaluation to improve neural sequence models. Models are adapted to recent history via a gradient descent based mechanism, causing them to assign higher probabilities to re-occurring sequential patterns. Dynamic evaluation outperforms existing adaptation approaches in our comparisons. Dynamic evaluation improves the state-of-the-art word-level perplexities on the Penn Treebank and WikiText-2 datasets to 51.1 and 44.3 respectively, and the state-of-the-art character-level cross-entropies on the text8 and Hutter Prize datasets to 1.19 bits/char and 1.08 bits/char respectively.

PDF Abstract ICML 2018 PDF ICML 2018 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Language Modelling Hutter Prize mLSTM + dynamic eval Bit per Character (BPC) 1.08 # 10
Number of params 46M # 10
Language Modelling Penn Treebank (Word Level) AWD-LSTM + dynamic eval Validation perplexity 51.6 # 11
Test perplexity 51.1 # 14
Params 24M # 7
Language Modelling Text8 mLSTM + dynamic eval Bit per Character (BPC) 1.19 # 15
Number of params 45M # 8
Language Modelling WikiText-2 AWD-LSTM + dynamic eval Validation perplexity 46.4 # 10
Test perplexity 44.3 # 18
Number of params 33M # 23

Methods


No methods listed for this paper. Add relevant methods here