Search Results for author: Bingqi Ma

Found 5 papers, 1 papers with code

MoVA: Adapting Mixture of Vision Experts to Multimodal Context

1 code implementation19 Apr 2024 Zhuofan Zong, Bingqi Ma, Dazhong Shen, Guanglu Song, Hao Shao, Dongzhi Jiang, Hongsheng Li, Yu Liu

Although some large-scale pretrained vision encoders such as vision encoders in CLIP and DINOv2 have brought promising performance, we found that there is still no single vision encoder that can dominate various image content understanding, e. g., the CLIP vision encoder leads to outstanding results on general image understanding but poor performance on document or chart content.

Language Modelling Large Language Model

MedFLIP: Medical Vision-and-Language Self-supervised Fast Pre-Training with Masked Autoencoder

no code implementations7 Mar 2024 Lei LI, Tianfang Zhang, Xinglin Zhang, Jiaqi Liu, Bingqi Ma, Yan Luo, Tao Chen

Within the domain of medical analysis, extensive research has explored the potential of mutual learning between Masked Autoencoders(MAEs) and multimodal data.

Representation Learning Zero-Shot Learning

Towards Large-scale Masked Face Recognition

no code implementations25 Oct 2023 Manyuan Zhang, Bingqi Ma, Guanglu Song, Yunxiao Wang, Hongsheng Li, Yu Liu

During the COVID-19 coronavirus epidemic, almost everyone is wearing masks, which poses a huge challenge for deep learning-based face recognition algorithms.

Face Recognition

Rethinking Robust Representation Learning Under Fine-grained Noisy Faces

no code implementations8 Aug 2022 Bingqi Ma, Guanglu Song, Boxiao Liu, Yu Liu

To better understand this, we reformulate the noise type of each class in a more fine-grained manner as N-identities|K^C-clusters.

Face Recognition Representation Learning

Target-Relevant Knowledge Preservation for Multi-Source Domain Adaptive Object Detection

no code implementations CVPR 2022 Jiaxi Wu, Jiaxin Chen, Mengzhe He, Yiru Wang, Bo Li, Bingqi Ma, Weihao Gan, Wei Wu, Yali Wang, Di Huang

Specifically, TRKP adopts the teacher-student framework, where the multi-head teacher network is built to extract knowledge from labeled source domains and guide the student network to learn detectors in unlabeled target domain.

Disentanglement Domain Adaptation +2

Cannot find the paper you are looking for? You can Submit a new open access paper.