Deep Residual Output Layers for Neural Language Generation

14 May 2019  ·  Nikolaos Pappas, James Henderson ·

Many tasks, including language generation, benefit from learning the structure of the output space, particularly when the space of output labels is large and the data is sparse. State-of-the-art neural language models indirectly capture the output space structure in their classifier weights since they lack parameter sharing across output labels. Learning shared output label mappings helps, but existing methods have limited expressivity and are prone to overfitting. In this paper, we investigate the usefulness of more powerful shared mappings for output labels, and propose a deep residual output mapping with dropout between layers to better capture the structure of the output space and avoid overfitting. Evaluations on three language generation tasks show that our output label mapping can match or improve state-of-the-art recurrent and self-attention architectures, and suggest that the classifier does not necessarily need to be high-rank to better model natural language if it is better at capturing the structure of the output space.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Language Modelling Penn Treebank (Word Level) AWD-LSTM-DRILL + dynamic eval Validation perplexity 49.5 # 10
Test perplexity 49.4 # 12
Params 24M # 7
Language Modelling Penn Treebank (Word Level) AWD-LSTM-DRILL Validation perplexity 58.2 # 21
Test perplexity 55.7 # 25
Params 24M # 7
Language Modelling WikiText-2 AWD-LSTM-DRILL Validation perplexity 64.9 # 20
Test perplexity 61.9 # 28
Number of params 34M # 20
Language Modelling WikiText-2 AWD-LSTM-DRILL + dynamic eval Validation perplexity 43.9 # 9
Test perplexity 42.0 # 17
Number of params 34M # 20
Machine Translation WMT2014 English-German Transformer-DRILL Base BLEU score 28.1 # 49
Hardware Burden None # 1
Operations per network pass None # 1

Methods