Full metadata record
DC FieldValueLanguage
dc.contributor.authorKuo, Che-Yuen_US
dc.contributor.authorChien, Jen-Tzungen_US
dc.date.accessioned2019-04-02T06:04:19Z-
dc.date.available2019-04-02T06:04:19Z-
dc.date.issued2018-01-01en_US
dc.identifier.issn2161-0363en_US
dc.identifier.urihttp://hdl.handle.net/11536/150841-
dc.description.abstractDeep learning has achieved great success in many real-world applications. For speech and language processing, recurrent neural networks are learned to characterize sequential patterns and extract the temporal information based on dynamic states which are evolved through time and stored as an internal memory. Traditionally, simple transition function using input-to-hidden and hidden-to-hidden weights is insufficient. To strengthen the learning capability, it is crucial to explore the diversity of latent structure in sequential signals and learn the stochastic trajectory of signal transitions to improve sequential prediction. This paper proposes the stochastic modeling of transitions in deep sequential learning. Our idea is to enhance latent variable representation by discovering the Markov state transitions in sequential data based on a K-state long short-term memory (LSTM) model. Such a latent state machine is capable of learning the complicated latent semantics in highly structured and heterogeneous sequential data. Gumbel-softmax is introduced to implement stochastic learning procedure with discrete states. Experimental results on visual and text language modeling illustrate the merit of the proposed stochastic transitions in sequential prediction with limited amount of parameters.en_US
dc.language.isoen_USen_US
dc.subjectDeep learningen_US
dc.subjectrecurrent neural networken_US
dc.subjectstochastic transitionen_US
dc.subjectdiscrete latent structureen_US
dc.titleMARKOV RECURRENT NEURAL NETWORKSen_US
dc.typeProceedings Paperen_US
dc.identifier.journal2018 IEEE 28TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP)en_US
dc.contributor.department電機工程學系zh_TW
dc.contributor.departmentDepartment of Electrical and Computer Engineeringen_US
dc.identifier.wosnumberWOS:000450651000063en_US
dc.citation.woscount0en_US
Appears in Collections:Conferences Paper