完整後設資料紀錄
DC 欄位語言
dc.contributor.authorYen, Frederick Z.en_US
dc.contributor.authorHuang, Mao-Changen_US
dc.contributor.authorChi, Tai-Shihen_US
dc.date.accessioned2017-04-21T06:49:29Z-
dc.date.available2017-04-21T06:49:29Z-
dc.date.issued2015en_US
dc.identifier.isbn978-1-5108-1790-6en_US
dc.identifier.urihttp://hdl.handle.net/11536/136225-
dc.description.abstractA two-stage singing voice separation algorithm using spectrotemporal modulation features is proposed in this paper. First, music clips are transformed into auditory spectrograms and the spectral-temporal modulation contents of all time-frequency (T-F) units of the auditory spectrograms are extracted using an auditory model. Then, T-F units are sequentially clustered using the expectation-maximization (EM) algorithm into percussive, harmonic and vocal units through the proposed two-stage algorithm. Lastly, the singing voice is synthesized from clustered vocal T-F units via time-frequency masking. The algorithm was evaluated using the MIR-1K dataset and demonstrated better separation results than our previously proposed one-stage algorithm.en_US
dc.language.isoen_USen_US
dc.subjectsinging voice separationen_US
dc.subjectspectro-temporal modulationen_US
dc.subjectauditory scene analysisen_US
dc.titleA Two-stage Singing Voice Separation Algorithm Using Spectro-temporal Modulation Featuresen_US
dc.typeProceedings Paperen_US
dc.identifier.journal16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5en_US
dc.citation.spage3321en_US
dc.citation.epage3324en_US
dc.contributor.department電機學院zh_TW
dc.contributor.departmentCollege of Electrical and Computer Engineeringen_US
dc.identifier.wosnumberWOS:000380581601230en_US
dc.citation.woscount0en_US
顯示於類別:會議論文