完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.author | Chien, Jen-Tzung | en_US |
dc.contributor.author | Chang, Ying-Lan | en_US |
dc.date.accessioned | 2018-08-21T05:56:37Z | - |
dc.date.available | 2018-08-21T05:56:37Z | - |
dc.date.issued | 2013-01-01 | en_US |
dc.identifier.issn | 2308-457X | en_US |
dc.identifier.uri | http://hdl.handle.net/11536/146415 | - |
dc.description.abstract | This paper presents a nonparametric interpretation for modem language model based on the hierarchical Pitman-Yor and Dirichlet (HPYD) process. We propose the HPYD language model (HPYD-LM) which flexibly conducts backoff smoothing and topic clustering through Bayesian nonparametric learning. The nonparametric priors of backoff n-grams and latent topics are tightly coupled in a compound process. A hybrid probability measure is drawn to build the smoothed topic-based LM. The model structure is automatically determined from training data. A new Chinese restaurant scenario is proposed to implement HPYD-LM via Gibbs sampling. This process reflects the power-law property and extracts the semantic topics from natural language. The superiority of HPYD-LM to the related LMs is demonstrated by the experiments on different corpora in terms of perplexity and word error rate. | en_US |
dc.language.iso | en_US | en_US |
dc.subject | language model | en_US |
dc.subject | backoff model | en_US |
dc.subject | topic model | en_US |
dc.subject | Bayesian learning | en_US |
dc.title | Hierarchical Pitman-Yor and Dirichlet Process for Language Model | en_US |
dc.type | Proceedings Paper | en_US |
dc.identifier.journal | 14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5 | en_US |
dc.citation.spage | 2211 | en_US |
dc.citation.epage | 2215 | en_US |
dc.contributor.department | 電機工程學系 | zh_TW |
dc.contributor.department | Department of Electrical and Computer Engineering | en_US |
dc.identifier.wosnumber | WOS:000395050000467 | en_US |
顯示於類別: | 會議論文 |