1 code implementation • 22 Jul 2024 • Yongcan Yu, Lijun Sheng, Ran He, Jian Liang
In particular, the memory bank is dynamically updated by selecting low-entropy and label-consistent samples in a class-balanced manner.
1 code implementation • 15 Mar 2024 • Yuting Xu, Jian Liang, Lijun Sheng, Xiao-Yu Zhang
The deepfake threats to society and cybersecurity have provoked significant public apprehension, driving intensified efforts within the realm of deepfake video detection.
1 code implementation • 6 Feb 2024 • Zhengbo Wang, Jian Liang, Lijun Sheng, Ran He, Zilei Wang, Tieniu Tan
Extensive results on 17 datasets validate that our method surpasses or achieves comparable results with state-of-the-art methods on few-shot classification, imbalanced learning, and out-of-distribution generalization.
1 code implementation • 28 Nov 2023 • Lijun Sheng, Zhengbo Wang, Jian Liang
Our solution adopts a two-stage source-free domain adaptation framework with a Swin Transformer backbone to achieve knowledge transfer from the USA (source) domain to Asia (target) domain.
no code implementations • 9 Oct 2023 • Yuhe Ding, Bo Jiang, Lijun Sheng, Aihua Zheng, Jian Liang
Transferability estimation aims to provide heuristics for quantifying how suitable a pre-trained model is for a specific downstream task, without fine-tuning them all.
1 code implementation • 24 Aug 2023 • Jian Liang, Lijun Sheng, Zhengbo Wang, Ran He, Tieniu Tan
The emergence of vision-language models, such as CLIP, has spurred a significant research effort towards their application for downstream supervised learning tasks.
1 code implementation • 6 Jul 2023 • Yongcan Yu, Lijun Sheng, Ran He, Jian Liang
To implement this benchmark, we have developed a unified framework in PyTorch, which allows for consistent evaluation and comparison of the TTA methods across the different datasets and network architectures.
1 code implementation • ICCV 2023 • Lijun Sheng, Jian Liang, Ran He, Zilei Wang, Tieniu Tan
To address this issue, we propose a model preprocessing framework, named AdaptGuard, to improve the security of model adaptation algorithms.
2 code implementations • 29 May 2022 • Yuhe Ding, Lijun Sheng, Jian Liang, Aihua Zheng, Ran He
First of all, to avoid additional parameters and explore the information in the source model, ProxyMix defines the weights of the classifier as the class prototypes and then constructs a class-balanced proxy source domain by the nearest neighbors of the prototypes to bridge the unseen source domain and the target domain.