Full metadata record
DC FieldValueLanguage
dc.contributor.authorYeh, Lan-Yingen_US
dc.contributor.authorChi, Tai-Shihen_US
dc.date.accessioned2014-12-08T15:37:01Z-
dc.date.available2014-12-08T15:37:01Z-
dc.date.issued2010en_US
dc.identifier.isbn978-1-61782-123-3en_US
dc.identifier.urihttp://hdl.handle.net/11536/25432-
dc.description.abstractSpeech emotion recognition is mostly considered in clean speech. In this paper, joint spectro-temporal features (RS features) are extracted from an auditory model and are applied to detect the emotion status of noisy speech. The noisy speech is derived from the Berlin Emotional Speech database with added white and babble noises under various SNR levels. The clean train/noisy test scenario is investigated to simulate conditions with unknown noisy sources. The sequential forward floating selection (SFFS) method is adopted to demonstrate the redundancy of RS features and further dimensionality reduction is conducted. Compared to conventional MFCCs plus prosodic features, RS features show higher recognition rates especially in low SNR conditions.en_US
dc.language.isoen_USen_US
dc.subjectEmotion recognitionen_US
dc.subjectrobusten_US
dc.subjectspectro-temporal modulationsen_US
dc.titleSpectro-Temporal Modulations for Robust Speech Emotion Recognitionen_US
dc.typeArticleen_US
dc.identifier.journal11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 1-4en_US
dc.citation.spage789en_US
dc.citation.epage792en_US
dc.contributor.department電機工程學系zh_TW
dc.contributor.departmentDepartment of Electrical and Computer Engineeringen_US
dc.identifier.wosnumberWOS:000294382400194-
Appears in Collections:Conferences Paper