Title: BAYESIAN RECURRENT NEURAL NETWORK LANGUAGE MODEL
Authors: Chien, Jen-Tzung
Ku, Yuan-Chu
電機學院
College of Electrical and Computer Engineering
Keywords: Recurrent neural network;language model;Bayesian learning;Hessian matrix
Issue Date: 2014
Abstract: This paper presents a Bayesian approach to construct the recurrent neural network language model (RNN-LM) for speech recognition. Our idea is to regularize the RNN-LM by compensating the uncertainty of the estimated model parameters which is represented by a Gaussian prior. The objective function in Bayesian RNN (BRNN) is formed as the regularized cross entropy error function. The regularized model is not only constructed by training the regularized parameters according to the maximum a posteriori criterion but also estimating the Gaussian hyperparameter by maximizing the marginal likelihood. A rapid approximation to Hessian matrix is developed by selecting a small set of salient outer-products and illustrated to be effective for BRNN-LM. BRNN-LM achieves sparser model than RNN-LM. Experiments on different corpora show promising improvement by applying BRNN-LM using different amount of training data.
URI: http://hdl.handle.net/11536/135882
ISBN: 978-1-4799-7129-9
Journal: 2014 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY SLT 2014
Begin Page: 206
End Page: 211
Appears in Collections:Conferences Paper