no code implementations • ICML 2020 • Yonggang Zhang, Ya Li, Tongliang Liu, Xinmei Tian
To obtain sufficient knowledge for crafting adversarial examples, previous methods query the target model with inputs that are perturbed with different searching directions.
no code implementations • 20 Feb 2025 • Fangming Cui, Jan Fong, Rongfei Zeng, Xinmei Tian, Jun Yu
In this paper, we propose a novel method called Similarity Paradigm with Textual Regularization (SPTR) for prompt learning without forgetting.
no code implementations • 11 Feb 2025 • Yuzhu Chen, Yingjie Wang, Shi Fu, Li Shen, Yongcheng Jing, Xinmei Tian, DaCheng Tao
This paper studies the crucial impact of initialization on the convergence properties of Low-Rank Adaptation (LoRA).
no code implementations • 8 Dec 2024 • Jun Nie, Yonggang Zhang, Tongliang Liu, Yiu-ming Cheung, Bo Han, Xinmei Tian
In this work, we propose a novel approach for detecting AI-generated images by leveraging predictive uncertainty to mitigate misuse and associated risks.
no code implementations • 3 Sep 2024 • Wei Chen, Zhen Huang, Liang Xie, Binbin Lin, Houqiang Li, Le Lu, Xinmei Tian, Deng Cai, Yonggang Zhang, Wenxiao Wang, Xu Shen, Jieping Ye
Recent works propose to employ supervised fine-tuning (SFT) to mitigate the sycophancy issue, while it typically leads to the degeneration of LLMs' general capability.
no code implementations • 3 Sep 2024 • Wei zhang, Chaoqun Wan, Yonggang Zhang, Yiu-ming Cheung, Xinmei Tian, Xu Shen, Jieping Ye
In this work, we delve into uncovering a specific mechanism by which LLMs execute calculations.
no code implementations • 29 Jul 2024 • Fangming Cui, Xun Yang, Chao Wu, Liang Xiao, Xinmei Tian
Specifically, we propose a textual external layer and learnable visual embeddings for adapting VLMs to downstream tasks.
2 code implementations • 22 Feb 2024 • Yonggang Zhang, Zhiqin Yang, Xinmei Tian, Nannan Wang, Tongliang Liu, Bo Han
Federated semi-supervised learning (FSSL) has emerged as a powerful paradigm for collaboratively training machine learning models using distributed data with label deficiency.
no code implementations • 19 Feb 2024 • Shi Fu, Sen Zhang, Yingjie Wang, Xinmei Tian, DaCheng Tao
This paper tackles the emerging challenge of training generative models within a self-consuming loop, wherein successive generations of models are recursively trained on mixtures of real and synthetic data from previous generations.
no code implementations • 10 Feb 2024 • Zhenheng Tang, Yonggang Zhang, Shaohuai Shi, Xinmei Tian, Tongliang Liu, Bo Han, Xiaowen Chu
First, we analyze the generalization contribution of local training and conclude that this generalization contribution is bounded by the conditional Wasserstein distance between the data distribution of different clients.
no code implementations • CVPR 2024 • Zhiyuan Yu, Li Shen, Liang Ding, Xinmei Tian, Yixin Chen, DaCheng Tao
To address these challenges we introduce PreBackRazor a novel activation pruning scheme offering both computational and memory efficiency through a sparsified backpropagation strategy which judiciously avoids unnecessary activation pruning and storage and gradient computation.
no code implementations • CVPR 2024 • Wei zhang, Chaoqun Wan, Tongliang Liu, Xinmei Tian, Xu Shen, Jieping Ye
This limitation hinders the potential of language supervision emphasized in CLIP and restricts the learning of temporal features as the text encoder has demonstrated limited proficiency in motion understanding.
1 code implementation • 4 Dec 2023 • Kaiwen Yang, Tao Shen, Xinmei Tian, Xiubo Geng, Chongyang Tao, DaCheng Tao, Tianyi Zhou
QVix enables a wider exploration of visual scenes, improving the LVLMs' reasoning accuracy and depth in tasks such as visual question answering and visual entailment.
no code implementations • 13 Nov 2023 • Zeqiao Zhou, Yuxuan Du, Xu-Fei Yin, Shanshan Zhao, Xinmei Tian, DaCheng Tao
DQS incorporates two essential components: a Graph Neural Network (GNN) predictor and a trigonometric interpolation algorithm.
2 code implementations • NeurIPS 2023 • Zhiqin Yang, Yonggang Zhang, Yu Zheng, Xinmei Tian, Hao Peng, Tongliang Liu, Bo Han
Comprehensive experiments demonstrate the efficacy of FedFed in promoting model performance.
1 code implementation • 15 Sep 2023 • Zhihao Hu, Yiran Xu, Mengnan Du, Jindong Gu, Xinmei Tian, Fengxiang He
Our adaptive reweighing method prioritizes samples closer to the decision boundary and assigns a higher weight to improve the generalizability of fair classifiers.
1 code implementation • 16 Jun 2023 • Shuangtong Li, Tianyi Zhou, Xinmei Tian, DaCheng Tao
We propose "Structured Cooperative Learning (SCooL)", in which a cooperation graph across devices is generated by a graphical model prior to automatically coordinate mutual learning between devices.
1 code implementation • 27 Apr 2023 • Rui Dai, Yonggang Zhang, Zhen Fang, Bo Han, Xinmei Tian
We show that MODE can endow models with provable generalization performance on unknown target domains.
1 code implementation • 12 Apr 2023 • Chengchao Xu, Xinmei Tian
To mitigate the hard-fitting issue, we propose to perform a semantic-aware mixup (SAM) for domain generalization, where whether to perform mixup depends on the semantic and domain information.
no code implementations • 7 Apr 2023 • Li Shen, Yan Sun, Zhiyuan Yu, Liang Ding, Xinmei Tian, DaCheng Tao
The field of deep learning has witnessed significant progress, particularly in computer vision (CV), natural language processing (NLP), and speech.
1 code implementation • 2 Nov 2022 • Kaiwen Yang, Yanchao Sun, Jiahao Su, Fengxiang He, Xinmei Tian, Furong Huang, Tianyi Zhou, DaCheng Tao
In experiments, we show that our method consistently brings non-trivial improvements to the three aforementioned learning tasks from both efficiency and final performance, either or not combined with strong pre-defined augmentations, e. g., on medical images when domain knowledge is unavailable and the existing augmentation techniques perform poorly.
1 code implementation • 29 Sep 2022 • Chenghao Sun, Yonggang Zhang, Wan Chaoqun, Qizhou Wang, Ya Li, Tongliang Liu, Bo Han, Xinmei Tian
As it is hard to mitigate the approximation error with few available samples, we propose Error TransFormer (ETF) for lightweight attacks.
4 code implementations • 19 Jul 2022 • Yuning Lu, Liangjian Wen, Jianzhuang Liu, Yajing Liu, Xinmei Tian
Specifically, we maximize the mutual information (MI) of instances and their representations with a low-bias MI estimator to perform self-supervised pre-training.
cross-domain few-shot learning
Unsupervised Few-Shot Image Classification
+1
1 code implementation • 21 Jun 2022 • Fuchen Long, Ting Yao, Zhaofan Qiu, Xinmei Tian, Jiebo Luo, Tao Mei
The video-to-text/video-to-query projections over text prototypes/query vocabulary then start the text-to-query or query-to-text calibration to estimate the amendment to query or text.
1 code implementation • 11 Jun 2022 • Wei Li, Qiming Zhang, Jing Zhang, Zhen Huang, Xinmei Tian, DaCheng Tao
To address these issues, we establish a new high-quality dataset named RealRain-1k, consisting of $1, 120$ high-resolution paired clean and rainy images with low- and high-density rain streaks, respectively.
1 code implementation • 1 Jun 2022 • Rong Dai, Li Shen, Fengxiang He, Xinmei Tian, DaCheng Tao
In this work, we propose a novel personalized federated learning framework in a decentralized (peer-to-peer) communication protocol named Dis-PFL, which employs personalized sparse masks to customize sparse local models on the edge.
no code implementations • CVPR 2022 • Yuning Lu, Jianzhuang Liu, Yonggang Zhang, Yajing Liu, Xinmei Tian
We present prompt distribution learning for effectively adapting a pre-trained vision-language model to address downstream recognition tasks.
no code implementations • CVPR 2022 • Chaoqun Wan, Xu Shen, Yonggang Zhang, Zhiheng Yin, Xinmei Tian, Feng Gao, Jianqiang Huang, Xian-Sheng Hua
Taking meta features as reference, we propose compositional operations to eliminate irrelevant features of local convolutional features by an addressing process and then to reformulate the convolutional feature maps as a composition of related meta features.
Ranked #5 on
Single-Source Domain Generalization
on Digits-five
no code implementations • CVPR 2022 • Shuangtong Li, Tianyi Zhou, Xinmei Tian, DaCheng Tao
Decentralized learning (DL) can exploit the images distributed over devices on a network topology to train a global model but is not designed to train personalized models for different tasks or optimize the topology.
no code implementations • 14 Dec 2021 • Yang Chen, Yingwei Pan, Yu Wang, Ting Yao, Xinmei Tian, Tao Mei
From this point, we present a particular paradigm of self-supervised learning tailored for domain adaptation, i. e., Transferrable Contrastive Learning (TCL), which links the SSL and the desired cross-domain transferability congruently.
no code implementations • ICCV 2021 • Yang Chen, Yu Wang, Yingwei Pan, Ting Yao, Xinmei Tian, Tao Mei
Correspondingly, we also propose a novel "jury" mechanism, which is particularly effective in learning useful semantic feature commonalities among domains.
Ranked #45 on
Domain Generalization
on PACS
no code implementations • NeurIPS 2021 • Kaiwen Yang, Tianyi Zhou, Yonggang Zhang, Xinmei Tian, DaCheng Tao
In this paper, we propose ''class-disentanglement'' that trains a variational autoencoder $G(\cdot)$ to extract this class-dependent information as $x - G(x)$ via a trade-off between reconstructing $x$ by $G(x)$ and classifying $x$ by $D(x-G(x))$, where the former competes with the latter in decomposing $x$ so the latter retains only necessary information for classification in $x-G(x)$.
no code implementations • 29 Sep 2021 • Kaiwen Yang, Tianyi Zhou, Xinmei Tian, DaCheng Tao
We then adversarially perturb $G(x)$ in the VAE's bottleneck space and adds it back to the original $R(x)$ as an augmentation, which is therefore sufficiently challenging for contrastive learning and meanwhile preserves the sample identity intact.
1 code implementation • CVPR 2021 • Zhen Huang, Xu Shen, Jun Xing, Tongliang Liu, Xinmei Tian, Houqiang Li, Bing Deng, Jianqiang Huang, Xian-Sheng Hua
The inheritance part is learned with a similarity loss to transfer the existing learned knowledge from the teacher model to the student model, while the exploration part is encouraged to learn representations different from the inherited ones with a dis-similarity loss.
no code implementations • 29 Jun 2021 • Kaiwen Yang, Xinmei Tian
Domain adversarial learning is a promising domain generalization method that aims to remove domain information in the latent representation through adversarial training.
Domain Generalization
Generalizable Person Re-identification
1 code implementation • ICLR 2022 • Yonggang Zhang, Mingming Gong, Tongliang Liu, Gang Niu, Xinmei Tian, Bo Han, Bernhard Schölkopf, Kun Zhang
The adversarial vulnerability of deep neural networks has attracted significant attention in machine learning.
1 code implementation • ICCV 2021 • Zhen Huang, Dixiu Xue, Xu Shen, Xinmei Tian, Houqiang Li, Jianqiang Huang, Xian-Sheng Hua
Second, different body parts possess different scales, and even the same part in different frames can appear at different locations and scales.
Ranked #4 on
Gait Recognition
on OUMVLP
1 code implementation • 26 Nov 2020 • Zhen Huang, Xu Shen, Xinmei Tian, Houqiang Li, Jianqiang Huang, Xian-Sheng Hua
The topology of the adjacency graph is a key factor for modeling the correlations of the input skeletons.
1 code implementation • ECCV 2020 • Fuchen Long, Ting Yao, Zhaofan Qiu, Xinmei Tian, Jiebo Luo, Tao Mei
In this paper, we introduce a new design of transfer learning type to learn action localization for a large set of action categories, but only on action moments from the categories of interest and temporal annotations of untrimmed videos from a small set of action classes.
1 code implementation • 28 Nov 2019 • Xu Shen, Xinmei Tian, Anfeng He, Shaoyan Sun, DaCheng Tao
In this paper, we propose randomly transforming (rotation, scale, and translation) feature maps of CNNs during the training stage.
1 code implementation • 28 Nov 2019 • Xu Shen, Xinmei Tian, Tongliang Liu, Fang Xu, DaCheng Tao
On the one hand, continuous dropout is considerably closer to the activation characteristics of neurons in the human brain than traditional binary dropout.
no code implementations • 28 Nov 2019 • Xu Shen, Xinmei Tian, Shaoyan Sun, DaCheng Tao
Convolutional Neural Networks (CNNs) have demonstrated state-of-the-art performance on many visual recognition tasks.
1 code implementation • CVPR 2019 • Jiwei Yang, Xu Shen, Jun Xing, Xinmei Tian, Houqiang Li, Bing Deng, Jianqiang Huang, Xian-Sheng Hua
The proposed quantization function can be learned in a lossless and end-to-end manner and works for any weights and activations of neural networks in a simple and uniform way.
1 code implementation • CVPR 2019 • Fuchen Long, Ting Yao, Zhaofan Qiu, Xinmei Tian, Jiebo Luo, Tao Mei
Temporally localizing actions in a video is a fundamental challenge in video understanding.
no code implementations • 26 Aug 2019 • Yang Chen, Yingwei Pan, Ting Yao, Xinmei Tian, Tao Mei
Unsupervised image-to-image translation is the task of translating an image from one domain to another in the absence of any paired training examples and tends to be more applicable to practical applications.
no code implementations • CVPR 2019 • Zhaofan Qiu, Ting Yao, Chong-Wah Ngo, Xinmei Tian, Tao Mei
Diffusions effectively interact two aspects of information, i. e., localized and holistic, for more powerful way of representation learning.
Ranked #10 on
Action Recognition
on UCF101
1 code implementation • CVPR 2019 • Qi Cai, Yingwei Pan, Chong-Wah Ngo, Xinmei Tian, Ling-Yu Duan, Ting Yao
The whole architecture is then optimized with three consistency regularizations: 1) region-level consistency to align the region-level predictions between teacher and student, 2) inter-graph consistency for matching the graph structures between teacher and student, and 3) intra-graph consistency to enhance the similarity between regions of same class within the graph of student.
1 code implementation • CVPR 2019 • Chang Chen, Zhiwei Xiong, Xinmei Tian, Zheng-Jun Zha, Feng Wu
Existing methods for single image super-resolution (SR) are typically evaluated with synthetic degradation models such as bicubic or Gaussian downsampling.
no code implementations • 3 Apr 2019 • Ya Li, Xinmei Tian, Tongliang Liu, DaCheng Tao
The objective of our proposed method is to transform the features from different tasks into a common feature space in which the tasks are closely related and the shared parameters can be better optimized.
no code implementations • 5 Sep 2018 • Anfeng He, Chong Luo, Xinmei Tian, Wen-Jun Zeng
Recently, Siamese network based trackers have received tremendous interest for their fast tracking speed and high performance.
Ranked #10 on
Visual Object Tracking
on VOT2017/18
no code implementations • ECCV 2018 • Ya Li, Xinmei Tian, Mingming Gong, Yajing Liu, Tongliang Liu, Kun Zhang, DaCheng Tao
Under the assumption that the conditional distribution $P(Y|X)$ remains unchanged across domains, earlier approaches to domain generalization learned the invariant representation $T(X)$ by minimizing the discrepancy of the marginal distribution $P(T(X))$.
Ranked #76 on
Domain Generalization
on PACS
no code implementations • ECCV 2018 • Chang Chen, Zhiwei Xiong, Xinmei Tian, Feng Wu
Boosting is a classic algorithm which has been successfully applied to diverse computer vision tasks.
1 code implementation • 23 Jul 2018 • Ya Li, Mingming Gong, Xinmei Tian, Tongliang Liu, DaCheng Tao
With the conditional invariant representation, the invariance of the joint distribution $\mathbb{P}(h(X), Y)$ can be guaranteed if the class prior $\mathbb{P}(Y)$ does not change across training and test domains.
no code implementations • 16 May 2018 • Xikun Zhang, Chang Xu, Xinmei Tian, DaCheng Tao
Considering the complementarity between graph node convolution and graph edge convolution, we additionally construct two hybrid neural networks to combine graph node convolutional neural network and graph edge convolutional neural network using shared intermediate layers.
1 code implementation • CVPR 2018 • Anfeng He, Chong Luo, Xinmei Tian, Wen-Jun Zeng
SA-Siam is composed of a semantic branch and an appearance branch.
Ranked #1 on
Visual Object Tracking
on OTB-50