TY - GEN
T1 - Hierarchical Channel Clustering with Bayesian-Optimized Pruning for High-Fidelity CNN Compression
AU - Liu, Xiaoxin
AU - Pan, Xiong
AU - Tong, Junhua
AU - Cheng, Jingchun
AU - Song, Jiajie
AU - He, Luyue
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Parameter pruning in deep neural networks is a widely adopted technique for model compression. Conventional pruning methods primarily evaluate weight importance using predefined importance factors; however, they often result in significant accuracy degradation at high compression rates. An alternative strategy involves categorizing weights and selecting the most informative representatives within each category. Both approaches eliminate a substantial number of weights based on specific evaluation criteria, potentially overlooking their collective impact on model accuracy. To address this limitation, this study introduces a weight clustering and merging approach for compressing convolutional neural networks. A weight feature matrix is first constructed based on channel relationships, followed by hierarchical clustering using Ward's distance. The pruning ratio for each individual layer is adjusted through Bayesian optimization, followed by a reconstruction of weights guided by the geometric relationships among features. In comparison with recent leading approaches, our method attains superior performance in terms of accuracy. In certain parameter settings, it even outperforms the unpruned baseline model.
AB - Parameter pruning in deep neural networks is a widely adopted technique for model compression. Conventional pruning methods primarily evaluate weight importance using predefined importance factors; however, they often result in significant accuracy degradation at high compression rates. An alternative strategy involves categorizing weights and selecting the most informative representatives within each category. Both approaches eliminate a substantial number of weights based on specific evaluation criteria, potentially overlooking their collective impact on model accuracy. To address this limitation, this study introduces a weight clustering and merging approach for compressing convolutional neural networks. A weight feature matrix is first constructed based on channel relationships, followed by hierarchical clustering using Ward's distance. The pruning ratio for each individual layer is adjusted through Bayesian optimization, followed by a reconstruction of weights guided by the geometric relationships among features. In comparison with recent leading approaches, our method attains superior performance in terms of accuracy. In certain parameter settings, it even outperforms the unpruned baseline model.
KW - Bayesian Optimization
KW - Model Compression
KW - Neural Network Pruning
KW - Weight Clustering
UR - https://www.scopus.com/pages/publications/105019927524
U2 - 10.1109/ACIT65614.2025.11185684
DO - 10.1109/ACIT65614.2025.11185684
M3 - 会议稿件
AN - SCOPUS:105019927524
T3 - Proceedings - International Conference on Advanced Computer Information Technologies, ACIT
SP - 981
EP - 986
BT - 2025 15th International Conference on Advanced Computer Information Technologies, ACIT 2025 - Conference Proceedings
PB - Institute of Electrical and Electronics Engineers
T2 - 15th International Conference on Advanced Computer Information Technologies, ACIT 2025
Y2 - 17 September 2025 through 19 September 2025
ER -