標題: | MARKOV RECURRENT NEURAL NETWORK LANGUAGE MODEL |
作者: | Chien, Jen-Tzung Kuo, Che-Yu 電機工程學系 Department of Electrical and Computer Engineering |
關鍵字: | language model;neural Markov process;recurrent neural network |
公開日期: | 1-一月-2019 |
摘要: | Recurrent neural network (RNN) has achieved a great success in language modeling where the temporal information based on deterministic state is continuously extracted and evolved through time. Such a simple deterministic transition function using input-to-hidden and hidden-to-hidden weights is usually insufficient to reflect the diversities and variations of latent variable structure behind the heterogeneous natural language. This paper presents a new stochastic Markov RNN (MRNN) to strengthen the learning capability in language model where the trajectory of word sequences is driven by a neural Markov process with Markov state transitions based on a K-state long short-term memory model. A latent state machine is constructed to characterize the complicated semantics in the structured lexical patterns. Gumbel-softmax is introduced to implement the stochastic backpropatation algorithm with discrete states. The parallel computation for rapid realization of MRNN is presented. The variational Bayesian learning procedure is implemented. Experiments demonstrate the merits of stochastic and diverse representation using MRNN language model where the overhead of parameters and computations is limited. |
URI: | http://hdl.handle.net/11536/154482 |
ISBN: | 978-1-7281-0306-8 |
期刊: | 2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019) |
起始頁: | 807 |
結束頁: | 813 |
顯示於類別: | 會議論文 |