Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Lo, Yu-Wen | en_US |
dc.contributor.author | Shen, Yih-Liang | en_US |
dc.contributor.author | Liao, Yuan-Fu | en_US |
dc.contributor.author | Chi, Tai-Shih | en_US |
dc.date.accessioned | 2019-04-02T06:04:14Z | - |
dc.date.available | 2019-04-02T06:04:14Z | - |
dc.date.issued | 2018-01-01 | en_US |
dc.identifier.uri | http://hdl.handle.net/11536/150765 | - |
dc.description.abstract | Before the era of the neural network (NN), features extracted from auditory models have been applied to various speech applications and been demonstrated more robust against noise than conventional speech-processing features. What's the role of auditory models in the current NN era? Are they obsolete? To answer this question, we construct a NN with a generative auditory model embedded to process speech signals. The generative auditory model consists of two stages, the stage of spectrum estimation in the logarithmic-frequency axis by the cochlea and the stage of spectral-temporal analysis in the modulation domain by the auditory cortex. The NN is evaluated in a simple speaker identification task. Experiment results show that the auditory model embedded NN is still more robust against noise, especially in low SNR conditions, than the randomly-initialized NN in speaker identification. | en_US |
dc.language.iso | en_US | en_US |
dc.subject | generative auditory model | en_US |
dc.subject | convolutional neural network | en_US |
dc.subject | multi-resolution | en_US |
dc.subject | speaker identification | en_US |
dc.title | A GENERATIVE AUDITORY MODEL EMBEDDED NEURAL NETWORK FOR SPEECH PROCESSING | en_US |
dc.type | Proceedings Paper | en_US |
dc.identifier.journal | 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) | en_US |
dc.citation.spage | 5179 | en_US |
dc.citation.epage | 5183 | en_US |
dc.contributor.department | 電機工程學系 | zh_TW |
dc.contributor.department | Department of Electrical and Computer Engineering | en_US |
dc.identifier.wosnumber | WOS:000446384605070 | en_US |
dc.citation.woscount | 0 | en_US |
Appears in Collections: | Conferences Paper |