no code implementations • 27 Aug 2024 • Lei Liu, Li Liu, Yawen Cui
Even in the era of large models, one of the well-known issues in continual learning (CL) is catastrophic forgetting, which is significantly challenging when the continual data stream exhibits a long-tailed distribution, termed as Long-Tailed Continual Learning (LTCL).
no code implementations • 22 Aug 2024 • Yuhao Wang, Chao Hao, Yawen Cui, Xinqi Su, Weicheng Xie, Tao Tan, Zitong Yu
This significantly enhances the report generation capability and clinical effectiveness of multi-modal large language models in the field of radiology reportgeneration.
1 code implementation • 20 Aug 2024 • Xinqi Su, Yawen Cui, Ajian Liu, Xun Lin, Yuhao Wang, Haochen Liang, Wenhui Li, Zitong Yu
In current web environment, fake news spreads rapidly across online social networks, posing serious threats to society.
2 code implementations • 31 Jul 2024 • Chunhui Zhang, Yawen Cui, Weilin Lin, Guanjie Huang, Yan Rong, Li Liu, Shiguang Shan
To address this gap, this work conducts a systematic review on SAM for videos in the era of foundation models.
1 code implementation • 15 Apr 2024 • Xinyu Xie, Yawen Cui, Chio-in Ieong, Tao Tan, Xiaozhi Zhang, Xubin Zheng, Zitong Yu
In this paper, we propose FusionMamba, a novel dynamic feature enhancement method for multimodal image fusion with Mamba.
no code implementations • ICCV 2023 • Jiacong Xu, Yi Zhang, Jiawei Peng, Wufei Ma, Artur Jesslen, Pengliang Ji, Qixin Hu, Jiehua Zhang, Qihao Liu, Jiahao Wang, Wei Ji, Chen Wang, Xiaoding Yuan, Prakhar Kaushik, Guofeng Zhang, Jie Liu, Yushan Xie, Yawen Cui, Alan Yuille, Adam Kortylewski
Animal3D consists of 3379 images collected from 40 mammal species, high-quality annotations of 26 keypoints, and importantly the pose and shape parameters of the SMAL model.
Ranked #1 on Animal Pose Estimation on Animal3D
no code implementations • 17 Aug 2023 • Shuangpeng Han, Rizhao Cai, Yawen Cui, Zitong Yu, Yongjian Hu, Alex Kot
To further improve generalization, we conduct hyperbolic contrastive learning for the bonafide only while relaxing the constraints on diverse spoofing attacks.
no code implementations • 26 Jul 2023 • Zitong Yu, Rizhao Cai, Yawen Cui, Ajian Liu, Changsheng chen
Recently, vision transformer based multimodal learning methods have been proposed to improve the robustness of face anti-spoofing (FAS) systems.
1 code implementation • 14 May 2023 • Chunhui Zhang, Li Liu, Yawen Cui, Guanjie Huang, Weilin Lin, Yiqian Yang, Yuehong Hu
As the first to comprehensively review the progress of segmenting anything task for vision and beyond based on the foundation model of SAM, this work focuses on its applications to various tasks and data types by discussing its historical development, recent progress, and profound impact on broad applications.
no code implementations • ICCV 2023 • Rizhao Cai, Yawen Cui, Zhi Li, Zitong Yu, Haoliang Li, Yongjian Hu, Alex Kot
To alleviate the forgetting of previous domains without using previous data, we propose the Proxy Prototype Contrastive Regularization (PPCR) to constrain the continual learning with previous domain knowledge from the proxy prototypes.
1 code implementation • 12 Feb 2023 • Yawen Cui, Zitong Yu, Rizhao Cai, Xun Wang, Alex C. Kot, Li Liu
The goal of Few-Shot Continual Learning (FSCL) is to incrementally learn novel tasks with limited labeled samples and preserve previous capabilities simultaneously, while current FSCL methods are all for the class-incremental purpose.
no code implementations • 11 Feb 2023 • Zitong Yu, Rizhao Cai, Yawen Cui, Xin Liu, Yongjian Hu, Alex Kot
In this paper, we investigate three key factors (i. e., inputs, pre-training, and finetuning) in ViT for multimodal FAS with RGB, Infrared (IR), and Depth.
no code implementations • 7 Feb 2023 • Zitong Yu, Yuming Shen, Jingang Shi, Hengshuang Zhao, Yawen Cui, Jiehua Zhang, Philip Torr, Guoying Zhao
As key modules in PhysFormer, the temporal difference transformers first enhance the quasi-periodic rPPG features with temporal difference guided global attention, and then refine the local spatio-temporal representation against interference.
1 code implementation • 24 Jan 2023 • Yawen Cui, Wanxia Deng, Haoyu Chen, Li Liu
Given a model well-trained with a large-scale base dataset, Few-Shot Class-Incremental Learning (FSCIL) aims at incrementally learning novel classes from a few labeled samples by avoiding overfitting, without catastrophically forgetting all encountered classes previously.
class-incremental learning Few-Shot Class-Incremental Learning +2
no code implementations • 20 Jul 2022 • Yawen Cui, Zitong Yu, Wei Peng, Li Liu
Few-Shot Class-Incremental Learning (FSCIL) aims at incrementally learning novel classes from a few labeled samples by avoiding the overfitting and catastrophic forgetting simultaneously.
class-incremental learning Few-Shot Class-Incremental Learning +3