Full metadata record
DC FieldValueLanguage
dc.contributor.authorGao, Pingen_US
dc.contributor.authorYou, Cheng-Youen_US
dc.contributor.authorChi, Tai-Shihen_US
dc.date.accessioned2020-10-05T02:01:29Z-
dc.date.available2020-10-05T02:01:29Z-
dc.date.issued2019-01-01en_US
dc.identifier.isbn978-1-7281-3248-8en_US
dc.identifier.issn2309-9402en_US
dc.identifier.urihttp://hdl.handle.net/11536/155269-
dc.description.abstractThe melody extraction can be considered as a sequence-to-sequence task or a classification task. Many recent models based on semantic segmentation have been proven very effective in melody extraction. In this paper, we built up a fully convolutional network (FCN) for melody extraction from polyphonic music. Inspired by the state-of-the-art architecture of the semantic segmentation, we constructed the encoder in a dense way and designed the decoder accordingly for audio processing. The combined frequency and periodicity (CFP) representation, which contains spectral and cepstral information, was adopted as the input feature of the proposed model. We conducted performance comparison between the proposed model and several methods on various datasets. Experimental results show the proposed model achieves state-of-the-art performance with less computation and fewer parameters.en_US
dc.language.isoen_USen_US
dc.titleA Multi-Scale Fully Convolutional Network for Singing Melody Extractionen_US
dc.typeProceedings Paperen_US
dc.identifier.journal2019 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC)en_US
dc.citation.spage1288en_US
dc.citation.epage1293en_US
dc.contributor.department電機工程學系zh_TW
dc.contributor.departmentDepartment of Electrical and Computer Engineeringen_US
dc.identifier.wosnumberWOS:000555696900216en_US
dc.citation.woscount0en_US
Appears in Collections:Conferences Paper