完整後設資料紀錄
DC 欄位語言
dc.contributor.authorHwang, Hsin-Teen_US
dc.contributor.authorTsao, Yuen_US
dc.contributor.authorWang, Hsin-Minen_US
dc.contributor.authorWang, Yih-Ruen_US
dc.contributor.authorChen, Sin-Horngen_US
dc.date.accessioned2014-12-08T15:30:03Z-
dc.date.available2014-12-08T15:30:03Z-
dc.date.issued2012en_US
dc.identifier.isbn978-1-4673-2507-3en_US
dc.identifier.urihttp://hdl.handle.net/11536/21518-
dc.description.abstractIn this paper, we propose a maximum mutual information (MMI) training criterion to refine the parameters of the joint density GMM (JDGMM) set to tackle the over-smoothing issue in voice conversion (VC). Conventionally, the maximum likelihood (ML) criterion is used to train a JDGMM set, which characterizes the joint property of the source and target feature vectors. The MMI training criterion, on the other hand, updates the parameters of the JDGMM set to increase its capability on modeling the dependency between the source and target feature vectors, and thus to make the converted sounds closer to the natural ones. The subjective listening test demonstrates that the quality and individuality of the converted speech by the proposed ML followed by MMI (ML+MMI) training method is better that by the ML training method.en_US
dc.language.isoen_USen_US
dc.subjectVoice conversionen_US
dc.subjectmutual informationen_US
dc.subjectGMMen_US
dc.titleEXPLORING MUTUAL INFORMATION FOR GMM-BASED SPECTRAL CONVERSIONen_US
dc.typeProceedings Paperen_US
dc.identifier.journal2012 8TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSINGen_US
dc.citation.spage50en_US
dc.citation.epage54en_US
dc.contributor.department電機資訊學士班zh_TW
dc.contributor.departmentUndergraduate Honors Program of Electrical Engineering and Computer Scienceen_US
dc.identifier.wosnumberWOS:000316984700018-
顯示於類別:會議論文