Full metadata record
DC FieldValueLanguage
dc.contributor.authorLiao, YFen_US
dc.contributor.authorChen, SHen_US
dc.date.accessioned2014-12-08T15:44:06Z-
dc.date.available2014-12-08T15:44:06Z-
dc.date.issued2001-03-01en_US
dc.identifier.issn1063-6676en_US
dc.identifier.urihttp://dx.doi.org/10.1109/89.905999en_US
dc.identifier.urihttp://hdl.handle.net/11536/29782-
dc.description.abstractA new modular recurrent neural network (MRNN)-based method for continuous Mandarin speech recognition (CMSR) is proposed. The MRNN recognizer is composed of four main modules. The first is a sub-MRNN module whose function is to generate discriminant functions for all 412 base-syllables. It accomplishes the task by using four recurrent neural network (RNN) submodules. The second is an RNN module which is designed to detect syllable boundaries for providing timing cues in order to help solve the time-alignment problem. The third is also an RNN module whose function is to generate discriminant functions for 143 intersyllable diphone-like units to compensate the intersyllable coarticulation effect. The fourth is a dynamic programming (DP)-based recognition search module. Its function is to integrate the other three modules and solve the time-alignment problem for generating the recognized base-syllable sequence. A new multilevel pruning scheme designed to speed up the recognition process is also proposed. The whole MRNN can be trained by a sophisticated three-stage minimum classification error/generalized probabilistic descent (MCE/GPD) algorithm. Experimental results showed that the proposed method performed better than the maximum likelihood (ML)-trained hidden Markov model (HMM) method and is comparable to the MCE/GPD-trained HMM method. The multilevel pruning scheme was also found to be very efficient.en_US
dc.language.isoen_USen_US
dc.subjectMandarin speech recognitionen_US
dc.subjectMCE/GPD algorithmsen_US
dc.subjectmodular recurrent neural networksen_US
dc.titleA modular RNN-based method for continuous Mandarin speech recognitionen_US
dc.typeArticleen_US
dc.identifier.doi10.1109/89.905999en_US
dc.identifier.journalIEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSINGen_US
dc.citation.volume9en_US
dc.citation.issue3en_US
dc.citation.spage252en_US
dc.citation.epage263en_US
dc.contributor.department電信工程研究所zh_TW
dc.contributor.departmentInstitute of Communications Engineeringen_US
dc.identifier.wosnumberWOS:000167288600007-
dc.citation.woscount3-
Appears in Collections:Articles


Files in This Item:

  1. 000167288600007.pdf

If it is a zip file, please download the file and unzip it, then open index.html in a browser to view the full text content.