1 code implementation • ICCV 2023 • Man Yao, Jiakui Hu, Guangshe Zhao, Yaoyuan Wang, Ziyang Zhang, Bo Xu, Guoqi Li
In this work, we pose and focus on three key questions regarding the inherent redundancy in SNNs.
no code implementations • 20 May 2023 • Man Yao, Yuhong Chou, Guangshe Zhao, Xiawu Zheng, Yonghong Tian, Bo Xu, Guoqi Li
LTH opens up a new path for network pruning.
no code implementations • 28 Sep 2022 • Man Yao, Guangshe Zhao, Hengyu Zhang, Yifan Hu, Lei Deng, Yonghong Tian, Bo Xu, Guoqi Li
On ImageNet-1K, we achieve top-1 accuracy of 75. 92% and 77. 08% on single/4-step Res-SNN-104, which are state-of-the-art results in SNNs.
no code implementations • ICCV 2021 • Man Yao, Huanhuan Gao, Guangshe Zhao, Dingheng Wang, Yihan Lin, ZhaoXu Yang, Guoqi Li
However, when aggregating individual events into frames with a new higher temporal resolution, existing SNN models do not attach importance to that the serial frames have different signal-to-noise ratios since event streams are sparse and non-uniform.
Ranked #6 on Audio Classification on SHD
no code implementations • 21 Aug 2020 • Dingheng Wang, Bijiao Wu, Guangshe Zhao, Man Yao, Hengnu Chen, Lei Deng, Tianyi Yan, Guoqi Li
Recurrent neural networks (RNNs) are powerful in the tasks oriented to sequential data, such as natural language processing and video recognition.
no code implementations • 29 Jun 2020 • Bijiao Wu, Dingheng Wang, Guangshe Zhao, Lei Deng, Guoqi Li
We further theoretically and experimentally discover that the HT format has better performance on compressing weight matrices, while the TT format is more suited for compressing convolutional kernels.
no code implementations • 8 Dec 2019 • Dingheng Wang, Guangshe Zhao, Guoqi Li, Lei Deng, Yang Wu
However, due to the higher dimension of convolutional kernels, the space complexity of 3DCNNs is generally larger than that of traditional two dimensional convolutional neural networks (2DCNNs).
Ranked #1 on Quantization on Knowledge-based: