完整後設資料紀錄
DC 欄位語言
dc.contributor.author林建廷zh_TW
dc.contributor.author王逸如zh_TW
dc.contributor.authorLin, Chien-Tingen_US
dc.contributor.authorWang, Yih-Ruen_US
dc.date.accessioned2018-01-24T07:37:34Z-
dc.date.available2018-01-24T07:37:34Z-
dc.date.issued2016en_US
dc.identifier.urihttp://etd.lib.nctu.edu.tw/cdrfb3/record/nctu/#GT070360313en_US
dc.identifier.urihttp://hdl.handle.net/11536/139182-
dc.description.abstract目前深層類神經網路已成為語音辨識領域中相當熱門的研究,本論文使用Kaldi speech recognition toolkit的環境為基礎實現共享隱藏層式深層類神經網路(Shared-hidden-layer language DNN, SHL-DNN)聲學模型訓練,因為類神經網路中的隱藏層可視為一連串複雜的特徵轉換,並且描述著各種語音之發音特徵,因此不同語言間可以共享。由於中文語料的不足,故使用較充足的英文語料,透過跨語言模型轉換技術(Cross-lingual model transfer),以改善中文語音之辨識率、降低各語者語音辨認率的變異程度。後續加入語言模型建立語音辨識系統,調整解碼過程中的參數並考量實際時間係數(Real Time Factor, RTF),找出最佳操作點,以得到即時性與辨識率兼顧的辨識系統。在訓練語料的部分,本實驗使用長達960個小時的Librispeech英文語料作為來源語言(Source language),目標語言(Target language)則為約24小時的中文語料TCC300,為了檢測系統強健度(Robustness),測試語料除了TCC300以外,另外加入1.9個小時的Sinica COSPRO02語料庫。zh_TW
dc.description.abstractDeep neural network (DNN) has been a popular research in automatic speech recognition. In this dissertation, we focus on implementing shared-hidden-layer language DNN acoustic model. The DNNs can be considered as a model that learns a complicated nonlinearity feature transformation and describe the pronouncing of different phonemes. Therefore, the hidden layers of DNNs could be shared for different languages. Unfortunately, there are only few small Taiwanese Mandarin speech corpora available. Consequently, in this paper, we evaluate cross-lingual model transfer approach to improve the performance of Mandarin-speech recognition and reduce the standard deviation of phone error rate with respect to different speakers. Moreover, we introduce a trigram language model to build a speech recognition system and tune the parameters using in decoding to receive the best operation point. Finally, we can build a real-time and high efficiency speech recognition system. In this experiment, a large 960 hours English corpus, Librispeech, was treated as the source language and a small 24 hours Mandarin corpus, TCC300, was treated as the target language. In the testing data, in addition to TCC300, we add an extra small 1.9 hours’ corpus, COSPRO02, to test the robustness of this system.en_US
dc.language.isozh_TWen_US
dc.subject深層類神經網路zh_TW
dc.subject中文語音辨識zh_TW
dc.subject跨語zh_TW
dc.subjectDeep neural networken_US
dc.subjectMandarin speech recognitionen_US
dc.subjectCross lingualen_US
dc.title使用跨語語料改進深層類神經網路中文語音辨認器之效能zh_TW
dc.titleUsing cross lingual data to improve the performance of DNN-based Mandarin-speech recognitionen_US
dc.typeThesisen_US
dc.contributor.department電信工程研究所zh_TW
顯示於類別:畢業論文