no code implementations • 27 Dec 2024 • Yuanze Li, Chun-Mei Feng, Qilong Wang, Guanglei Yang, WangMeng Zuo
Human beings can leverage knowledge from relative tasks to improve learning on a primary task.
1 code implementation • 16 Dec 2024 • Tianyi Zhu, Dongwei Ren, Qilong Wang, Xiaohe Wu, WangMeng Zuo
Generative inbetweening aims to generate intermediate frame sequences by utilizing two key frames as input.
no code implementations • 6 Dec 2024 • Xiaojie Yin, Qilong Wang, Bing Cao, QinGhua Hu
Recently, many studies have been conducted to enhance the zero-shot generalization ability of vision-language models (e. g., CLIP) by addressing the semantic misalignment between image and text embeddings in downstream tasks.
no code implementations • 28 Nov 2024 • Yilong Wang, Zilin Gao, Qilong Wang, Zhaofeng Chen, Peihua Li, QinGhua Hu
To effectively and efficiently explore the potential of pre-trained models in transferring to target domain, our TAMT proposes a Hierarchical Temporal Tuning Network (HTTN), whose core involves local temporal-aware adapters (TAA) and a global temporal-aware moment tuning (GTMT).
no code implementations • 20 Nov 2024 • Bing Cao, Quanhao Lu, Jiekang Feng, Pengfei Zhu, QinGhua Hu, Qilong Wang
The dynamic imbalance of the fore-background is a major challenge in video object counting, which is usually caused by the sparsity of foreground objects.
1 code implementation • 3 Nov 2024 • Bing Cao, Xingxin Xu, Pengfei Zhu, Qilong Wang, QinGhua Hu
To address this issue, we propose a conditional controllable fusion (CCF) framework for general image fusion tasks without specific training.
no code implementations • 22 Aug 2024 • Shaobo Wang, Yantai Yang, Qilong Wang, Kaixin Li, Linfeng Zhang, Junchi Yan
Our findings suggest that prioritizing the synthesis of easier samples from the original dataset can enhance the quality of distilled datasets, especially in low IPC (image-per-class) settings.
1 code implementation • 21 Aug 2024 • Wei Shang, Dongwei Ren, Wanying Zhang, Qilong Wang, Pengfei Zhu, WangMeng Zuo
Subsequently, to effectively train the DRSC network, we propose a self-supervised learning strategy that ensures cycle consistency between input and reconstructed dual reversed RS images.
no code implementations • 13 Jun 2024 • Wenlong Yu, Dongyue Chen, Qilong Wang, QinGhua Hu
Likewise, we propose a Feature Structuralized Domain Generalization (FSDG) model, wherein features experience structuralization into common, specific, and confounding segments, harmoniously aligned with their relevant semantic concepts, to elevate performance in FGDG.
1 code implementation • CVPR 2024 • Yuwei Tang, Zhenyi Lin, Qilong Wang, Pengfei Zhu, QinGhua Hu
To this end, we disassemble three key components involved in computation of logit bias (i. e., logit features, logit predictor, and logit fusion) and empirically analyze the effect on performance of few-shot classification.
1 code implementation • 10 Dec 2023 • Yunheng Li, Zhongyu Li, ShangHua Gao, Qilong Wang, Qibin Hou, Ming-Ming Cheng
Effectively modeling discriminative spatio-temporal information is essential for segmenting activities in long action sequences.
1 code implementation • ICCV 2023 • Mingze Gao, Qilong Wang, Zhenyi Lin, Pengfei Zhu, QinGhua Hu, Jingbo Zhou
Distinguished from LP which builds a linear classification head based on the mean of final features (e. g., word tokens for ViT) or classification tokens, our MP performs a linear classifier on feature distribution, which provides the stronger representation ability by exploiting richer statistical information inherent in features.
1 code implementation • NeurIPS 2022 Conference 2023 • Qilong Wang, Mingze Gao, Zhaolin Zhang, Jiangtao Xie, Peihua Li, QinGhua Hu
Particularly, we for the first time show that \textit{effective post-normalization can make a good trade-off between representation decorrelation and information preservation for GCP, which are crucial to alleviate over-fitting and increase representation ability of deep GCP networks, respectively}.
no code implementations • CVPR 2023 • Zixuan Qin, Liu Yang, Qilong Wang, Yahong Han, QinGhua Hu
When there are large differences in data distribution among clients, it is crucial for federated learning to design a reliable client selection strategy and an interpretable client communication framework to better utilize group knowledge.
1 code implementation • CVPR 2022 • Jiangtao Xie, Fei Long, Jiaming Lv, Qilong Wang, Peihua Li
Few-shot classification is a challenging problem as only very few training examples are given for each new task.
1 code implementation • NeurIPS 2021 • Zilin Gao, Qilong Wang, Bingbing Zhang, QinGhua Hu, Peihua Li
Then, a temporal covariance pooling performs temporal pooling of the attentive covariance representations to characterize both intra-frame correlations and inter-frame cross-correlations of the calibrated features.
1 code implementation • ICCV 2021 • Bowen Dong, Zitong Huang, Yuelin Guo, Qilong Wang, Zhenxing Niu, WangMeng Zuo
In this paper, we defend the problem setting for improving localization performance by leveraging the bounding box regression knowledge from a well-annotated auxiliary dataset.
1 code implementation • CVPR 2021 • Longyin Wen, Dawei Du, Pengfei Zhu, QinGhua Hu, Qilong Wang, Liefeng Bo, Siwei Lyu
To promote the developments of object detection, tracking and counting algorithms in drone-captured videos, we construct a benchmark with a new drone-captured largescale dataset, named as DroneCrowd, formed by 112 video clips with 33, 600 HD frames in various scenarios.
1 code implementation • 22 Apr 2021 • Jiangtao Xie, Ruiren Zeng, Qilong Wang, Ziqi Zhou, Peihua Li
Therefore, we propose a new classification paradigm, where the second-order, cross-covariance pooling of visual tokens is combined with class token for final classification.
1 code implementation • CVPR 2020 • Qilong Wang, Li Zhang, Banggu Wu, Dongwei Ren, Peihua Li, WangMeng Zuo, QinGhua Hu
Recent works have demonstrated that global covariance pooling (GCP) has the ability to improve performance of deep convolutional neural networks (CNNs) on visual classification task.
1 code implementation • 4 Dec 2019 • Longyin Wen, Dawei Du, Pengfei Zhu, QinGhua Hu, Qilong Wang, Liefeng Bo, Siwei Lyu
This paper proposes a space-time multi-scale attention network (STANet) to solve density map estimation, localization and tracking in dense crowds of video clips captured by drones with arbitrary crowd density, perspective, and flight altitude.
12 code implementations • CVPR 2020 • Qilong Wang, Banggu Wu, Pengfei Zhu, Peihua Li, WangMeng Zuo, QinGhua Hu
By dissecting the channel attention module in SENet, we empirically show avoiding dimensionality reduction is important for learning channel attention, and appropriate cross-channel interaction can preserve performance while significantly decreasing model complexity.
Ranked #5 on Object Detection on PKU-DDD17-Car
1 code implementation • CVPR 2020 • Dongwei Ren, Kai Zhang, Qilong Wang, QinGhua Hu, WangMeng Zuo
To connect MAP and deep models, we in this paper present two generative networks for respectively modeling the deep priors of clean image and blur kernel, and propose an unconstrained neural optimization solution to blind deconvolution.
1 code implementation • 2 Jun 2019 • Hao Wang, Qilong Wang, Fan Yang, Weiqi Zhang, WangMeng Zuo
For guiding our IS to obtain better object performance, we explore issues of instance imbalance and class importance in datasets, which frequently occur and bring adverse effect on detection performance.
1 code implementation • CVPR 2019 • Qilong Wang, Peihua Li, Qinghua Hu, Pengfei Zhu, Wangmeng Zuo
To handle this issue, this paper proposes a novel deep global generalized Gaussian network (3G-Net), whose core is to estimate a global covariance of generalized Gaussian for modeling the last convolutional activations.
3 code implementations • 15 Apr 2019 • Qilong Wang, Jiangtao Xie, WangMeng Zuo, Lei Zhang, Peihua Li
The proposed methods are highly modular, readily plugged into existing deep CNNs.
Ranked #1 on Image Classification on iNaturalist (Top 3 Error metric)
1 code implementation • NeurIPS 2018 • Qilong Wang, Zilin Gao, Jiangtao Xie, WangMeng Zuo, Peihua Li
However, both GAP and existing HOP methods assume unimodal distributions, which cannot fully capture statistics of convolutional activations, limiting representation ability of deep CNNs, especially for samples with complex contents.
1 code implementation • CVPR 2019 • Zilin Gao, Jiangtao Xie, Qilong Wang, Peihua Li
Deep Convolutional Networks (ConvNets) are fundamental to, besides large-scale visual recognition, a lot of vision tasks.
2 code implementations • CVPR 2018 • Hao Wang, Qilong Wang, Mingqi Gao, Peihua Li, WangMeng Zuo
Our MLKP can be efficiently computed on a modified multi-scale feature map using a low-dimensional polynomial kernel approximation. Moreover, different from existing orderless global representations based on high-order statistics, our proposed MLKP is location retentive and sensitive so that it can be flexibly adopted to object detection.
4 code implementations • CVPR 2018 • Peihua Li, Jiangtao Xie, Qilong Wang, Zilin Gao
Towards addressing this problem, we propose an iterative matrix square root normalization method for fast end-to-end training of global covariance pooling networks.
Ranked #14 on Fine-Grained Image Classification on CUB-200-2011
Fine-Grained Image Classification Fine-Grained Image Recognition
no code implementations • CVPR 2017 • Qilong Wang, Peihua Li, Lei Zhang
Recently, plugging trainable structural layers into deep convolutional neural networks (CNNs) as image representations has made promising progress.
3 code implementations • CVPR 2017 • Hongliang Yan, Yukang Ding, Peihua Li, Qilong Wang, Yong Xu, WangMeng Zuo
Specifically, we introduce class-specific auxiliary weights into the original MMD for exploiting the class prior probability on source and target domains, whose challenge lies in the fact that the class label in target domain is unavailable.
1 code implementation • ICCV 2017 • Peihua Li, Jiangtao Xie, Qilong Wang, WangMeng Zuo
The main challenges involved are robust covariance estimation given a small sample of large-dimensional features and usage of the manifold structure of covariance matrices.
no code implementations • CVPR 2016 • Qilong Wang, Peihua Li, WangMeng Zuo, Lei Zhang
Infinite dimensional covariance descriptors can provide richer and more discriminative information than their low dimensional counterparts.
no code implementations • 9 Jul 2015 • Qilong Wang, Peihua Li, Lei Zhang, WangMeng Zuo
The bag-of-features (BoF) model for image classification has been thoroughly studied over the last decade.
no code implementations • CVPR 2015 • Peihua Li, Xiaoxiao Lu, Qilong Wang
The locality-constrained linear coding (LLC) is a very successful feature coding method in image classification.