no code implementations • ECCV 2020 • Xiangyu He, Zitao Mo, Ke Cheng, Weixiang Xu, Qinghao Hu, Peisong Wang, Qingshan Liu, Jian Cheng
The matrix composed of basis vectors is referred to as the proxy matrix, and auxiliary variables serve as the coefficients of this linear combination.
1 code implementation • 23 Sep 2024 • Zeyu Zhu, Peisong Wang, Qinghao Hu, Gang Li, Xiaoyao Liang, Jian Cheng
However, through an in-depth analysis, we observe that the efficiency of existing sampling-based training frameworks is still limited due to the key bottlenecks lying in all three phases of sampling-based training, i. e., subgraph sample, memory IO, and computation.
1 code implementation • 19 Aug 2024 • Yukang Chen, Fuzhao Xue, Dacheng Li, Qinghao Hu, Ligeng Zhu, Xiuyu Li, Yunhao Fang, Haotian Tang, Shang Yang, Zhijian Liu, Ethan He, Hongxu Yin, Pavlo Molchanov, Jan Kautz, Linxi Fan, Yuke Zhu, Yao Lu, Song Han
We introduce the long-context Multi-Modal Sequence Parallelism (MM-SP) system that efficiently parallelizes long video training and inference, enabling 2M context length training on 256 GPUs without any gradient checkpointing.
no code implementations • 19 Jul 2024 • Meng Zhang, Jie Sun, Qinghao Hu, Peng Sun, Zeke Wang, Yonggang Wen, Tianwei Zhang
While there emerge inspiring algorithm advancements, their practical adoption is still limited, particularly on real-world graphs involving up to millions of nodes.
1 code implementation • 12 Mar 2024 • Qinghao Hu, Zhisheng Ye, Zerui Wang, Guoteng Wang, Meng Zhang, Qiaoling Chen, Peng Sun, Dahua Lin, Xiaolin Wang, Yingwei Luo, Yonggang Wen, Tianwei Zhang
Large Language Models (LLMs) have presented impressive performance across several transformative tasks.
1 code implementation • 8 Dec 2023 • Xiaozhe Yao, Qinghao Hu, Ana Klimovic
Fine-tuning large language models (LLMs) greatly improves model quality for downstream tasks.
1 code implementation • 20 Sep 2023 • Xingting Yao, Qinghao Hu, Fei Zhou, Tielong Liu, Zitao Mo, Zeyu Zhu, Zhengyang Zhuge, Jian Cheng
In SpikingNeRF, each sampled point on the ray is matched to a particular time step and represented in a hybrid manner where the voxel grids are maintained as well.
no code implementations • 2 Mar 2023 • Meng Zhang, Qinghao Hu, Peng Sun, Yonggang Wen, Tianwei Zhang
Training Graph Neural Networks (GNNs) on large graphs is challenging due to the conflict between the high memory demand and limited GPU memory.
1 code implementation • 1 Feb 2023 • Zeyu Zhu, Fanrong Li, Zitao Mo, Qinghao Hu, Gang Li, Zejian Liu, Xiaoyao Liang, Jian Cheng
Through an in-depth analysis of the topology of GNNs, we observe that the topology of the graph leads to significant differences between nodes, and most of the nodes in a graph appear to have a small aggregation value.
1 code implementation • 25 Nov 2022 • Yiqun Chen, Qiang Chen, Qinghao Hu, Jian Cheng
In this paper, we revisit these two assignment methods and find that bringing one-to-many assignment back to end-to-end fully convolutional detectors helps with model convergence.
1 code implementation • 22 Oct 2022 • Xiangyu Chen, Qinghao Hu, Kaidong Li, Cuncong Zhong, Guanghui Wang
After carefully examining the self-attention modules, we discover that the number of trivial attention weights is far greater than the important ones and the accumulated trivial weights are dominating the attention in Vision Transformers due to their large quantity, which is not handled by the attention itself.
1 code implementation • 3 Aug 2022 • Qinghao Hu, Gang Li, Qiman Wu, Jian Cheng
In this paper, we propose the PArallel Low-precision Quantization (PalQuant) method that approximates high-precision computations via learning parallel low-precision representations from scratch.
no code implementations • 24 May 2022 • Wei Gao, Qinghao Hu, Zhisheng Ye, Peng Sun, Xiaolin Wang, Yingwei Luo, Tianwei Zhang, Yonggang Wen
Deep learning (DL) shows its prosperity in a wide variety of fields.
3 code implementations • CVPR 2022 • Qiang Chen, Qiman Wu, Jian Wang, Qinghao Hu, Tao Hu, Errui Ding, Jian Cheng, Jingdong Wang
We propose MixFormer to find a solution.
1 code implementation • 4 Apr 2022 • Weixiang Xu, Xiangyu He, Tianli Zhao, Qinghao Hu, Peisong Wang, Jian Cheng
The latest STTN shows that ResNet-18 with ternary weights and ternary activations achieves up to 68. 2% Top-1 accuracy on ImageNet.
no code implementations • ICCV Workshop 2021 • Xing Lan, Qinghao Hu, Jian Cheng
The statistical re- sults show the NME generated by quantization error is even larger than 1/3 of the SOTA item, which is a serious obsta- cle for making a new breakthrough in face alignment.
Ranked #8 on Face Alignment on COFW
1 code implementation • 3 Sep 2021 • Qinghao Hu, Peng Sun, Shengen Yan, Yonggang Wen, Tianwei Zhang
Modern GPU datacenters are critical for delivering Deep Learning (DL) models and services in both the research community and industry.
no code implementations • 1 Sep 2021 • Tianli Zhao, Qinghao Hu, Xiangyu He, Weixiang Xu, Jiaxing Wang, Cong Leng, Jian Cheng
Acceleration of deep neural networks to meet a specific latency constraint is essential for their deployment on mobile devices.
1 code implementation • 7 Apr 2021 • Xing Lan, Qinghao Hu, Qiang Chen, Jian Xue, Jian Cheng
In particular, our HIH reaches 4. 08 NME (Normalized Mean Error) on WFLW, and 3. 21 on COFW, which exceeds previous methods by a significant margin.
Ranked #4 on Face Alignment on WFW (Extra Data)
no code implementations • 21 Jan 2021 • Xiangyu He, Qinghao Hu, Peisong Wang, Jian Cheng
Convolutional neural networks are able to learn realistic image priors from numerous training samples in low-level image generation and restoration.
1 code implementation • ACM MM 2020 • Xing Lan, Qinghao Hu, Fangzhou Xiong, Cong Leng, Jian Cheng
Face alignment is an important task in the field of multi-media.
Ranked #9 on Face Alignment on COFW (using extra training data)
no code implementations • 24 Sep 2019 • Fanrong Li, Zitao Mo, Peisong Wang, Zejian Liu, Jiayun Zhang, Gang Li, Qinghao Hu, Xiangyu He, Cong Leng, Yang Zhang, Jian Cheng
As a case study, we evaluate our object detection system on a real-world surveillance video with input size of 512x512, and it turns out that the system can achieve an inference speed of 18 fps at the cost of 6. 9W (with display) with an mAP of 66. 4 verified on the PASCAL VOC 2012 dataset.
1 code implementation • 29 Jul 2019 • Xianyang Li, Feng Wang, Qinghao Hu, Cong Leng
With the development of convolutional neural network, significant progress has been made in computer vision tasks.
1 code implementation • 23 Jul 2019 • Xiangyu He, Ke Cheng, Qiang Chen, Qinghao Hu, Peisong Wang, Jian Cheng
Long-range dependencies modeling, widely used in capturing spatiotemporal correlation, has shown to be effective in CNN dominated computer vision tasks.
Ranked #212 on Object Detection on COCO test-dev (using extra training data)
no code implementations • ECCV 2018 • Qinghao Hu, Gang Li, Peisong Wang, Yifan Zhang, Jian Cheng
In this paper, we propose a novel semi-binary decomposition method which decomposes a matrix into two binary matrices and a diagonal matrix.
no code implementations • ECCV 2018 • Guan'an Wang, Qinghao Hu, Jian Cheng, Zeng-Guang Hou
Secondly, we design novel structure of the generative model and the discriminative model to learn the distribution of triplet-wise information in a semi-supervised way.
1 code implementation • CVPR 2018 • Peisong Wang, Qinghao Hu, Yifan Zhang, Chunjie Zhang, Yang Liu, Jian Cheng
In this paper, we propose a simple yet effective Two-Step Quantization (TSQ) framework, by decomposing the network quantization problem into two steps: code learning and transformation function learning based on the learned codes.
no code implementations • 8 Feb 2018 • Qinghao Hu, Peisong Wang, Jian Cheng
To achieve this goal, we propose a novel approach named BWNH to train Binary Weight Networks via Hashing.
no code implementations • 3 Feb 2018 • Jian Cheng, Peisong Wang, Gang Li, Qinghao Hu, Hanqing Lu
As for hardware implementation of deep neural networks, a batch of accelerators based on FPGA/ASIC have been proposed in recent years.
1 code implementation • CVPR 2016 • Jiaxiang Wu, Cong Leng, Yuhang Wang, Qinghao Hu, Jian Cheng
Recently, convolutional neural networks (CNN) have demonstrated impressive performance in various computer vision tasks.