完整後設資料紀錄
DC 欄位語言
dc.contributor.authorLin, Wei-weien_US
dc.contributor.authorMak, Man-Waien_US
dc.contributor.authorChien, Jen-Tzungen_US
dc.date.accessioned2019-04-02T05:58:17Z-
dc.date.available2019-04-02T05:58:17Z-
dc.date.issued2018-12-01en_US
dc.identifier.issn2329-9290en_US
dc.identifier.urihttp://dx.doi.org/10.1109/TASLP.2018.2866707en_US
dc.identifier.urihttp://hdl.handle.net/11536/148095-
dc.description.abstractLike many machine learning tasks, the performance of speaker verification (SV) systems degrades when training and test data come from very different distributions. What's more, both training and test data themselves could be composed of heterogeneous subsets. These multisource mismatches are detrimental to SV performance. This paper proposes incorporating maximum mean discrepancy (MMD) into the loss function of autoencoders to reduce these mismatches. MMD is a non-parametric method for measuring the distance between two probability distributions. With a properly chosen kernel, MMD can match up to infinite moments of data distributions. We generalize MMD to measure the discrepancies of multiple distributions. We call the generalized MMD domainwise MMD. Using domainwise MMD as an objective function, we propose two autoencoders, namely nuisance-attribute autoencoder (NAE) and domain-invariant autoencoder (DAE), for multisource i-vector adaptation. NAE encodes the features that cause most of the multisource mismatch measured by domainwise MMD. DAE directly encodes the features that minimize the multisource mismatch. Using these MMD-based autoencoders as a preprocessing step for PLDA training, we achieve a relative improvement of 19.2% EER on the NIST 2016 SRE compared to PLDA without adaptation. We also found that MMD-based autoencoders are more robust to unseen domains. In the domain robustness experiments, MMD-based autoencoders show 6.8% and 5.2% improvements over IDVC on female and male Cantonese speakers, respectively.en_US
dc.language.isoen_USen_US
dc.subjectSpeaker verificationen_US
dc.subjectdomain adaptationen_US
dc.subjecti-vectorsen_US
dc.subjectmaximum mean discrepancyen_US
dc.titleMultisource I-Vectors Domain Adaptation Using Maximum Mean Discrepancy Based Autoencodersen_US
dc.typeArticleen_US
dc.identifier.doi10.1109/TASLP.2018.2866707en_US
dc.identifier.journalIEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSINGen_US
dc.citation.volume26en_US
dc.citation.spage2412en_US
dc.citation.epage2422en_US
dc.contributor.department電機工程學系zh_TW
dc.contributor.departmentDepartment of Electrical and Computer Engineeringen_US
dc.identifier.wosnumberWOS:000443761500003en_US
dc.citation.woscount1en_US
顯示於類別:期刊論文