2 code implementations • 15 Sep 2023 • Shaowu Chen, Weize Sun, Lei Huang, Xiaopeng Li, Qingyuan Wang, Deepu John
In Stage 1, POCKET utilizes dynamically varying penalties to efficiently achieve group sparsity within the classifier, removing features associated with zero weights and their corresponding kernels.
1 code implementation • 16 Feb 2023 • Shaowu Chen, Weize Sun, Lei Huang
Filter pruning has attracted increasing attention in recent years for its capacity in compressing and accelerating convolutional neural networks.
no code implementations • 22 Oct 2022 • Xuhua Li, Weize Sun, Lei Huang, Shaowu Chen
Filter pruning is a common method to achieve model compression and acceleration in deep neural networks (DNNs). Some research regarded filter pruning as a combinatorial optimization problem and thus used evolutionary algorithms (EA) to prune filters of DNNs.
1 code implementation • 9 Jul 2021 • Shaowu Chen, Jiahao Zhou, Weize Sun, Lei Huang
To overcome this problem, we propose to compress CNNs and alleviate performance degradation via joint matrix decomposition, which is different from existing works that compressed layers separately.
Efficient Neural Network Matrix Factorization / Decomposition +1
1 code implementation • IEEE Journal of Selected Topics in Signal Processing 2020 • Weize Sun, Shaowu Chen, Lei Huang, Hing Cheung So, Min Xie
Comparing to the state-of-the-art independent matrix and tensor decomposition based methods, our model can obtain a better network performance under the same compression ratio.