z-logo
open-access-imgOpen Access
Deep compression of convolutional neural networks with low‐rank approximation
Author(s) -
Astrid Marcella,
Lee SeungIk
Publication year - 2018
Publication title -
etri journal
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.295
H-Index - 46
eISSN - 2233-7326
pISSN - 1225-6463
DOI - 10.4218/etrij.2018-0065
Subject(s) - computer science , inference , singular value decomposition , convolutional neural network , deep neural networks , artificial neural network , computer engineering , algorithm , artificial intelligence
The application of deep neural networks (DNNs) to connect the world with cyber physical systems (CPSs) has attracted much attention. However, DNNs require a large amount of memory and computational cost, which hinders their use in the relatively low‐end smart devices that are widely used in CPSs. In this paper, we aim to determine whether DNNs can be efficiently deployed and operated in low‐end smart devices. To do this, we develop a method to reduce the memory requirement of DNNs and increase the inference speed, while maintaining the performance (for example, accuracy) close to the original level. The parameters of DNNs are decomposed using a hybrid of canonical polyadic–singular value decomposition, approximated using a tensor power method, and fine‐tuned by performing iterative one‐shot hybrid fine‐tuning to recover from a decreased accuracy. In this study, we evaluate our method on frequently used networks. We also present results from extensive experiments on the effects of several fine‐tuning methods, the importance of iterative fine‐tuning, and decomposition techniques. We demonstrate the effectiveness of the proposed method by deploying compressed networks in smartphones.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here