no code implementations • 17 May 2017 • Xu Tian, Jun Zhang, Zejun Ma, Yi He, Juan Wei
The system which combined frame retaining with frame stacking could reduces the time consumption of both training and decoding.
no code implementations • 21 Mar 2017 • Xu Tian, Jun Zhang, Zejun Ma, Yi He, Juan Wei, Peihao Wu, Wenchang Situ, Shuai Li, Yang Zhang
It is a competitive framework that LSTM models of more than 7 layers are successfully trained on Shenma voice search data in Mandarin and they outperform the deep LSTM models trained by conventional approach.
no code implementations • 3 Mar 2017 • Xu Tian, Jun Zhang, Zejun Ma, Yi He, Juan Wei
As training data rapid growth, large-scale parallel training with multi-GPUs cluster is widely applied in the neural network model learning currently. We present a new approach that applies exponential moving average method in large-scale parallel training of neural network model.