no code implementations • 25 Jan 2024 • Tong Niu, Haoyu Huang, Yu Du, Weihao Zhang, Luping Shi, Rong Zhao
Given the escalating intricacy and multifaceted nature of contemporary social systems, manually generating solutions to address pertinent social issues has become a formidable task.
1 code implementation • 11 Nov 2022 • Hao Zheng, Hui Lin, Rong Zhao, Luping Shi
In this paper, we propose a brain-inspired hybrid neural network (HNN) that introduces temporal binding theory originated from neuroscience into ANNs by integrating spike timing dynamics (via spiking neural networks, SNNs) with reconstructive attention (by ANNs).
no code implementations • 5 Jun 2020 • Yujie Wu, Rong Zhao, Jun Zhu, Feng Chen, Mingkun Xu, Guoqi Li, Sen Song, Lei Deng, Guanrui Wang, Hao Zheng, Jing Pei, Youhui Zhang, Mingguo Zhao, Luping Shi
We demonstrate the advantages of this model in multiple different tasks, including few-shot learning, continual learning, and fault-tolerance learning in neuromorphic vision sensors.
1 code implementation • 20 Dec 2019 • Faqiang Liu, Mingkun Xu, Guoqi Li, Jing Pei, Luping Shi, Rong Zhao
Generative adversarial networks have achieved remarkable performance on various tasks but suffer from training instability.
no code implementations • 25 Sep 2019 • Faqiang Liu, Mingkun Xu, Guoqi Li, Jing Pei, Luping Shi
Generative adversarial networks have achieved remarkable performance on various tasks but suffer from sensitivity to hyper-parameters, training instability, and mode collapse.
no code implementations • 15 Sep 2019 • Zheyu Yang, Yujie Wu, Guanrui Wang, Yukuan Yang, Guoqi Li, Lei Deng, Jun Zhu, Luping Shi
To the best of our knowledge, DashNet is the first framework that can integrate and process ANNs and SNNs in a hybrid paradigm, which provides a novel solution to achieve both effectiveness and efficiency for high-speed object tracking.
1 code implementation • NeurIPS 2019 • Shuang Wu, Guanrui Wang, Pei Tang, Feng Chen, Luping Shi
Compact convolutional neural networks gain efficiency mainly through depthwise convolutions, expanded channels and complex topologies, which contrarily aggravate the training process.
no code implementations • 16 Sep 2018 • Yujie Wu, Lei Deng, Guoqi Li, Jun Zhu, Luping Shi
Spiking neural networks (SNNs) that enables energy efficient implementation on emerging neuromorphic hardware are gaining more attention.
no code implementations • 17 Mar 2018 • Hongmin Li, Luping Shi
Object tracking based on retina-inspired and event-based dynamic vision sensor (DVS) is challenging for the noise events, rapid change of event-stream shape, chaos of complex background textures, and occlusion.
no code implementations • 27 Feb 2018 • Shuang Wu, Guoqi Li, Lei Deng, Liu Liu, Yuan Xie, Luping Shi
Batch Normalization (BN) has been proven to be quite effective at accelerating and improving the training of deep neural networks (DNNs).
3 code implementations • ICLR 2018 • Shuang Wu, Guoqi Li, Feng Chen, Luping Shi
Researches on deep neural networks with discrete parameters and their deployment in embedded systems have been active and promising topics.
no code implementations • 7 Feb 2018 • Hongmin Li, Guoqi Li, Hanchao Liu, Luping Shi
Firstly, the event number of each pixel of the HR DVS image is determined with a sparse signal representation based method to obtain the HR event-count map from that of the LR DVS recording.
1 code implementation • 8 Jun 2017 • Yujie Wu, Lei Deng, Guoqi Li, Jun Zhu, Luping Shi
By simultaneously considering the layer-by-layer spatial domain (SD) and the timing-dependent temporal domain (TD) in the training phase, as well as an approximated derivative for the spike activity, we propose a spatio-temporal backpropagation (STBP) training framework without using any complicated technology.