完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.author | Lin, Chien-Yu | en_US |
dc.contributor.author | Lai, Bo-Cheng | en_US |
dc.date.accessioned | 2018-08-21T05:57:09Z | - |
dc.date.available | 2018-08-21T05:57:09Z | - |
dc.date.issued | 2018-01-01 | en_US |
dc.identifier.issn | 2153-6961 | en_US |
dc.identifier.uri | http://hdl.handle.net/11536/147112 | - |
dc.description.abstract | Sparsity is widely observed in convolutional neural networks by zeroing a large portion of both activations and weights without impairing the result. By keeping the data in a compressed-sparse format, the energy consumption could be considerably cut down due to less memory traffic. However, the wide SIMD-like MAC engine adopted in many CNN accelerators can not support the compressed input due to the data misalignment. In this work, a novel Dual Indexing Module (DIM) is proposed to efficiently handle the alignment issue where activations and weights are both kept in compressed-sparse format. The DIM is implemented in a representative SIMD-like CNN accelerator, and able to exploit both compressed-sparse activations and weights. The synthesis results with 40nm technology have shown that DIM can enhance up to 46% of energy consumption and 55.4% Energy-Delay-Product (EDP). | en_US |
dc.language.iso | en_US | en_US |
dc.title | Supporting Compressed-Sparse Activations and Weights on SIMD-like Accelerator for Sparse Convolutional Neural Networks | en_US |
dc.type | Proceedings Paper | en_US |
dc.identifier.journal | 2018 23RD ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC) | en_US |
dc.citation.spage | 105 | en_US |
dc.citation.epage | 110 | en_US |
dc.contributor.department | 電子工程學系及電子研究所 | zh_TW |
dc.contributor.department | Department of Electronics Engineering and Institute of Electronics | en_US |
dc.identifier.wosnumber | WOS:000426987100017 | en_US |
顯示於類別: | 會議論文 |