完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.author | Chang, Yen-Wei | en_US |
dc.contributor.author | Peng, Wen-Hsiao | en_US |
dc.date.accessioned | 2020-01-02T00:03:28Z | - |
dc.date.available | 2020-01-02T00:03:28Z | - |
dc.date.issued | 2019-01-01 | en_US |
dc.identifier.isbn | 978-1-5386-9552-4 | en_US |
dc.identifier.issn | 1945-7871 | en_US |
dc.identifier.uri | http://dx.doi.org/10.1109/ICME.2019.00096 | en_US |
dc.identifier.uri | http://hdl.handle.net/11536/153328 | - |
dc.description.abstract | This paper tackles the problem of learning a questioner in the goal-oriented visual dialog task. Several previous works adopt model-free reinforcement learning. Most pretrain the model from a finite set of human-generated data. We argue that using limited demonstrations to kick-start the questioner is insufficient due to the large policy search space. Inspired by a recently proposed information theoretic approach, we develop two analytic experts to serve as a source of high-quality demonstrations for imitation learning. We then take advantage of reinforcement learning to refine the model towards the goal-oriented objective. Experimental results on the GuessWhat?! dataset show that our method has the combined merits of imitation and reinforcement learning, achieving the state-of-the-art performance. | en_US |
dc.language.iso | en_US | en_US |
dc.title | LEARNING GOAL-ORIENTED VISUAL DIALOG AGENTS: IMITATING AND SURPASSING ANALYTIC EXPERTS | en_US |
dc.type | Proceedings Paper | en_US |
dc.identifier.doi | 10.1109/ICME.2019.00096 | en_US |
dc.identifier.journal | 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME) | en_US |
dc.citation.spage | 520 | en_US |
dc.citation.epage | 525 | en_US |
dc.contributor.department | 交大名義發表 | zh_TW |
dc.contributor.department | National Chiao Tung University | en_US |
dc.identifier.wosnumber | WOS:000501820600088 | en_US |
dc.citation.woscount | 0 | en_US |
顯示於類別: | 會議論文 |