Search Results for author: Fang Peng

Found 4 papers, 3 papers with code

HiVG: Hierarchical Multimodal Fine-grained Modulation for Visual Grounding

1 code implementation20 Apr 2024 Linhui Xiao, Xiaoshan Yang, Fang Peng, YaoWei Wang, Changsheng Xu

Specifically, HiVG consists of a multi-layer adaptive cross-modal bridge and a hierarchical multimodal low-rank adaptation (Hi LoRA) paradigm.

CLIP-VG: Self-paced Curriculum Adapting of CLIP for Visual Grounding

1 code implementation15 May 2023 Linhui Xiao, Xiaoshan Yang, Fang Peng, Ming Yan, YaoWei Wang, Changsheng Xu

In order to utilize vision and language pre-trained models to address the grounding problem, and reasonably take advantage of pseudo-labels, we propose CLIP-VG, a novel method that can conduct self-paced curriculum adapting of CLIP with pseudo-language labels.

Transfer Learning Visual Grounding

SgVA-CLIP: Semantic-guided Visual Adapting of Vision-Language Models for Few-shot Image Classification

no code implementations28 Nov 2022 Fang Peng, Xiaoshan Yang, Linhui Xiao, YaoWei Wang, Changsheng Xu

Although significant progress has been made in few-shot learning, most of existing few-shot image classification methods require supervised pre-training on a large amount of samples of base classes, which limits their generalization ability in real world application.

Few-Shot Image Classification Few-Shot Learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.