| 標題: | Enhancing Utilization of SIMD-Like Accelerator for Sparse Convolutional Neural Networks |
| 作者: | Lai, Bo-Cheng Pan, Jyun-Wei Lin, Chien-Yu 電子工程學系及電子研究所 Department of Electronics Engineering and Institute of Electronics |
| 關鍵字: | Load balance;machine learning;single-instruction-multiple-data (SIMD) architecture;sparse convolutional neural networks (CNNs) |
| 公開日期: | 1-五月-2019 |
| 摘要: | Although the existing single-instruction-multiple-data-like (SIMD) accelerators can handle the compressed format of sparse convolutional neural networks, the sparse and irregular distributions of nonzero elements cause low utilization of multipliers in a processing engine (PE) and imbalanced computation between PEs. This brief addresses the above issues by proposing a data screening and task mapping (DSTM) accelerator which integrates a series of techniques, including software refinement and hardware modules. An efficient indexing module is introduced to identify the effectual computation pairs and skip unnecessary computation in a fine-grained manner. The intra-PE load imbalance is alleviated with weight data rearrangement. An effective task sharing mechanism further balances the computation between PEs. When compared with the state-of-the-art SIMD-like accelerator, the proposed DSTM enhances the average PE utilization by 3.5x. The overall processing throughput is 59.7% higher than the previous design. |
| URI: | http://dx.doi.org/10.1109/TVLSI.2019.2897052 http://hdl.handle.net/11536/152414 |
| ISSN: | 1063-8210 |
| DOI: | 10.1109/TVLSI.2019.2897052 |
| 期刊: | IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS |
| Volume: | 27 |
| Issue: | 5 |
| 起始頁: | 1218 |
| 結束頁: | 1222 |
| 顯示於類別: | 期刊論文 |

