1 code implementation • 20 Mar 2025 • Sidi Yang, Binxiao Huang, Yulun Zhang, Dahai Yu, Yujiu Yang, Ngai Wong
While deep neural networks have revolutionized image denoising capabilities, their deployment on edge devices remains challenging due to substantial computational and memory requirements.
no code implementations • 20 Mar 2025 • Shiyong Liu, Xiao Tang, Zhihao LI, Yingfan He, Chongjie Ye, Jianzhuang Liu, Binxiao Huang, Shunbo Zhou, Xiaofei Wu
Cameras in such regions exhibit stronger correlations and a higher average contribution, facilitating high-quality scene reconstruction.
no code implementations • 18 Jan 2025 • Jiaqi Lin, Zhihao LI, Binxiao Huang, Xiao Tang, Jianzhuang Liu, Shiyong Liu, Xiaofei Wu, Fenglong Song, Wenming Yang
We validate our method on several appearance-variant scenes, and demonstrate that it achieves state-of-the-art rendering quality with minimal training time and memory usage, without compromising rendering speeds.
no code implementations • 9 May 2024 • Binxiao Huang, Jason Chun Lok, Chang Liu, Ngai Wong
To exploit the abundant information contained in the input data to output label mapping, our scheme utilizes the network trained from the clean dataset as a trigger generator to produce poisons that significantly raise the success rate of backdoor attacks versus conventional approaches.
1 code implementation • 28 Mar 2024 • Sidi Yang, Binxiao Huang, Mingdeng Cao, Yatai Ji, Hanzhong Guo, Ngai Wong, Yujiu Yang
Existing enhancement models often optimize for high performance while falling short of reducing hardware inference time and power consumption, especially on edge devices with constrained computing and storage resources.
no code implementations • 28 Dec 2023 • Jason Chun Lok Li, Chang Liu, Binxiao Huang, Ngai Wong
Existing approaches to Implicit Neural Representation (INR) can be interpreted as a global scene representation via a linear combination of Fourier bases of different frequencies.
1 code implementation • 11 Dec 2023 • Binxiao Huang, Jason Chun Lok Li, Jie Ran, Boyu Li, Jiajun Zhou, Dahai Yu, Ngai Wong
Conventional super-resolution (SR) schemes make heavy use of convolutional neural networks (CNNs), which involve intensive multiply-accumulate (MAC) operations, and require specialized hardware such as graphics processing units.
no code implementations • 14 Nov 2023 • Rui Lin, Jason Chun Lok Li, Jiajun Zhou, Binxiao Huang, Jie Ran, Ngai Wong
Most deep neural networks (DNNs) consist fundamentally of convolutional and/or fully connected layers, wherein the linear transform can be cast as the product between a filter matrix and a data matrix obtained by arranging feature tensors into columns.
no code implementations • 25 Jun 2023 • Binxiao Huang, Rui Lin, Chaofan Tao, Ngai Wong
Deep neural networks (DNNs) are incredibly vulnerable to crafted, imperceptible adversarial perturbations.
no code implementations • 24 Dec 2022 • Binxiao Huang, Chaofan Tao, Rui Lin, Ngai Wong
Deep neural networks are incredibly vulnerable to crafted, human-imperceptible adversarial perturbations.
1 code implementation • 20 Jul 2022 • Chang Liu, Xiaoyan Qian, Binxiao Huang, Xiaojuan Qi, Edmund Lam, Siew-Chong Tan, Ngai Wong
By enriching the sparse point clouds, our method achieves 4. 48\% and 4. 03\% better 3D AP on KITTI moderate and hard samples, respectively, versus the state-of-the-art autolabeler.
no code implementations • 16 Mar 2022 • Binxiao Huang, Chaofan Tao, Rui Lin, Ngai Wong
We are hopeful this work can shed light on the design of more robust neural networks.