학술논문

Neural Network Compression Using Higher-Order Statistics and Auxiliary Reconstruction Losses
Document Type
Conference
Source
2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) Computer Vision and Pattern Recognition Workshops (CVPRW),2020 IEEE/CVF Conference on. :3077-3086 Jun, 2020
Subject
Computing and Processing
Gaussian distribution
Training
Higher order statistics
Measurement
Neural networks
Machine learning
Computational complexity
Language
ISSN
2160-7516
Abstract
In this paper, the problem of pruning and compressing the weights of various layers of deep neural networks is investigated. The proposed method aims to remove redundant filters from the network to reduce computational complexity and storage requirements, while improving the performance of the original network. More specifically, a novel filter selection criterion is introduced based on the fact that filters whose weights follow a Gaussian distribution correspond to hidden units that do not capture important aspects of data. To this end, Higher Order Statistics (HOS) are used and filters with low cumulant values that do not deviate significantly from Gaussian distribution are identified and removed from the network. In addition, a novel pruning strategy is proposed aiming to decide on the pruning ratio of each layer using the Shapiro-Wilk normality test. The use of auxiliary MSE losses (intermediate and after the softmax layer) during the fine-tuning phase further improves the overall performance of the compressed network. Extensive experiments with different network architectures and comparison with state-of-the-art approaches on well-known public datasets, such as CIFAR-10, CIFAR-100 and ILSCVR-12, demonstrate the great potential of the proposed approach.