학술논문

PSM-nets: Compressing Neural Networks with Product of Sparse Matrices
Document Type
Conference
Source
2021 International Joint Conference on Neural Networks (IJCNN) Neural Networks (IJCNN), 2021 International Joint Conference on. :1-8 Jul, 2021
Subject
Bioengineering
Computing and Processing
Robotics and Control Systems
Signal Processing and Analysis
Deep learning
Tensors
Neural networks
Sparse matrices
Matrix decomposition
Optimization
Residual neural networks
neural networks
compression
sparsity
Language
ISSN
2161-4407
Abstract
Over-parameterization of neural networks is a well known issue that comes along with their great performance. Among the many approaches proposed to tackle this problem, low-rank tensor decompositions are largely investigated to compress deep neural networks. Such techniques rely on a low-rank assumption of the layer weight tensors that does not always hold in practice. Following this observation, this paper studies sparsity inducing techniques to build new sparse matrix product layers for high-rate neural networks compression. Specifically, we explore recent advances in sparse optimization to replace each layer's weight matrix, either convolutional or fully connected, by a product of sparse matrices. Our experiments validate that our approach provides a better compression-accuracy trade-off than most popular low-rank-based compression techniques.