Lowering dynamic power of a stream-based CNN hardware accelerator

Custom hardware accelerators of Convolutional Neural Networks (CNN) provide a promising solution to meet real-time constraints for a wide range of applications on low-cost embedded devices. In this work, we aim to lower the dynamic power of a stream-based CNN hardware accelerator by reducing the com...

全面介紹

Saved in:
書目詳細資料
Main Authors: Piyasena, Duvindu, Wickramasinghe, Rukshan, Paul, Debdeep, Lam, Siew-Kei, Wu, Meiqing
其他作者: School of Computer Science and Engineering
格式: Conference or Workshop Item
語言:English
出版: 2021
主題:
在線閱讀:https://hdl.handle.net/10356/147509
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
實物特徵
總結:Custom hardware accelerators of Convolutional Neural Networks (CNN) provide a promising solution to meet real-time constraints for a wide range of applications on low-cost embedded devices. In this work, we aim to lower the dynamic power of a stream-based CNN hardware accelerator by reducing the computational redundancies in the CNN layers. In particular, we investigate the redundancies due to the downsampling effect of max pooling layers which are prevalent in state-of-the-art CNNs, and propose an approximation method to reduce the overall computations. The experimental results show that the proposed method leads to lower dynamic power without sacrificing accuracy.