z-logo
open-access-imgOpen Access
Implementation of energy‐efficient fast convolution algorithm for deep convolutional neural networks based on FPGA
Author(s) -
Li W.J.,
Ruan S.J.,
Yang D.S.
Publication year - 2020
Publication title -
electronics letters
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.375
H-Index - 146
ISSN - 1350-911X
DOI - 10.1049/el.2019.4188
Subject(s) - convolutional neural network , computation , computer science , convolution (computer science) , field programmable gate array , chip , algorithm , parallel computing , process (computing) , energy (signal processing) , state (computer science) , artificial neural network , computer engineering , computer hardware , artificial intelligence , mathematics , telecommunications , statistics , operating system
The state‐of‐the‐art convolutional neural networks (CNNs) have been widely applied to many deep neural networks models. As the model becomes more accurate, both the number of computation and the data accesses are significantly increased. The proposed design uses the row stationary with network‐on‐chip and the fast convolution algorithm in process elements to reduce the number of computation and data accesses simultaneously. The experimental evaluation which using the CNN layers of VGG‐16 with a batch size of three shows that the proposed design is more energy‐efficient than the state‐of‐the‐art work. The proposed design improves the total GOPs of the algorithm by 1.497 times and reduces the on‐chip memory and off‐chip memory accesses by 1.07 and 1.46 times than prior work, respectively.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here