完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.author | Yen, Frederick Z. | en_US |
dc.contributor.author | Huang, Mao-Chang | en_US |
dc.contributor.author | Chi, Tai-Shih | en_US |
dc.date.accessioned | 2017-04-21T06:49:29Z | - |
dc.date.available | 2017-04-21T06:49:29Z | - |
dc.date.issued | 2015 | en_US |
dc.identifier.isbn | 978-1-5108-1790-6 | en_US |
dc.identifier.uri | http://hdl.handle.net/11536/136225 | - |
dc.description.abstract | A two-stage singing voice separation algorithm using spectrotemporal modulation features is proposed in this paper. First, music clips are transformed into auditory spectrograms and the spectral-temporal modulation contents of all time-frequency (T-F) units of the auditory spectrograms are extracted using an auditory model. Then, T-F units are sequentially clustered using the expectation-maximization (EM) algorithm into percussive, harmonic and vocal units through the proposed two-stage algorithm. Lastly, the singing voice is synthesized from clustered vocal T-F units via time-frequency masking. The algorithm was evaluated using the MIR-1K dataset and demonstrated better separation results than our previously proposed one-stage algorithm. | en_US |
dc.language.iso | en_US | en_US |
dc.subject | singing voice separation | en_US |
dc.subject | spectro-temporal modulation | en_US |
dc.subject | auditory scene analysis | en_US |
dc.title | A Two-stage Singing Voice Separation Algorithm Using Spectro-temporal Modulation Features | en_US |
dc.type | Proceedings Paper | en_US |
dc.identifier.journal | 16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5 | en_US |
dc.citation.spage | 3321 | en_US |
dc.citation.epage | 3324 | en_US |
dc.contributor.department | 電機學院 | zh_TW |
dc.contributor.department | College of Electrical and Computer Engineering | en_US |
dc.identifier.wosnumber | WOS:000380581601230 | en_US |
dc.citation.woscount | 0 | en_US |
顯示於類別: | 會議論文 |