TY - GEN
T1 - Summary of Convolutional Neural Network Compression Technology
AU - Zhang, Yabo
AU - DIng, Wenrui
AU - Liu, Chunlei
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/10
Y1 - 2019/10
N2 - Deep convolutional neural networks (DCNNs) obtain dramatically accuracy improvement in the area of computer vision for recent years. However, because of their large demand for memory, power, and computational ability, it is difficult for DCNNs to be employed in the light-weight devices such as mobile. Therefore, a natural idea is compressing the DCNNs, while reduce the demand for the memory and computational ability and not reduce the classification accuracy too much. In recent years, model compression gain a lot of improvement. And we summaries the compression methods in this paper. DCNNs mainly classify three types, feed-forward deep networks (FFDN), feed-back deep network (FBDN) and bi-directional deep networks (BDDN). The methods of compression and acceleration are roughly categorized into four schemes: parameter pruning and sharing, low-rank factorization, transfer /compact convolutional filters and knowledge distillation. In this paper, we introduce the performance, related applications, advantages and drawbacks of each CNNs and techniques for compacting and accelerating DCNNs model. In the last, we summarize the paper and propose the possible future work in model compression.
AB - Deep convolutional neural networks (DCNNs) obtain dramatically accuracy improvement in the area of computer vision for recent years. However, because of their large demand for memory, power, and computational ability, it is difficult for DCNNs to be employed in the light-weight devices such as mobile. Therefore, a natural idea is compressing the DCNNs, while reduce the demand for the memory and computational ability and not reduce the classification accuracy too much. In recent years, model compression gain a lot of improvement. And we summaries the compression methods in this paper. DCNNs mainly classify three types, feed-forward deep networks (FFDN), feed-back deep network (FBDN) and bi-directional deep networks (BDDN). The methods of compression and acceleration are roughly categorized into four schemes: parameter pruning and sharing, low-rank factorization, transfer /compact convolutional filters and knowledge distillation. In this paper, we introduce the performance, related applications, advantages and drawbacks of each CNNs and techniques for compacting and accelerating DCNNs model. In the last, we summarize the paper and propose the possible future work in model compression.
KW - convolutional neural network (CNN)
KW - deep learning
KW - model compression and acceleration
KW - object recognition
UR - https://www.scopus.com/pages/publications/85080858668
U2 - 10.1109/ICUS48101.2019.8995969
DO - 10.1109/ICUS48101.2019.8995969
M3 - 会议稿件
AN - SCOPUS:85080858668
T3 - Proceedings of the 2019 IEEE International Conference on Unmanned Systems, ICUS 2019
SP - 480
EP - 483
BT - Proceedings of the 2019 IEEE International Conference on Unmanned Systems, ICUS 2019
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2019 IEEE International Conference on Unmanned Systems, ICUS 2019
Y2 - 17 October 2019 through 19 October 2019
ER -