Search Results for author: Jinlong He

Found 2 papers, 2 papers with code

PeFoMed: Parameter Efficient Fine-tuning of Multimodal Large Language Models for Medical Imaging

1 code implementation5 Jan 2024 Gang Liu, Jinlong He, Pengfei Li, Genrong He, Zhaolin Chen, Shenjun Zhong

In this paper, we propose a parameter efficient framework for fine-tuning MLLMs, specifically validated on medical visual question answering (Med-VQA) and medical report generation (MRG) tasks, using public benchmark datasets.

 Ranked #1 on Medical Visual Question Answering on VQA-RAD (using extra training data)

Medical Report Generation Medical Visual Question Answering +4

Masked Vision and Language Pre-training with Unimodal and Multimodal Contrastive Losses for Medical Visual Question Answering

1 code implementation11 Jul 2023 Pengfei Li, Gang Liu, Jinlong He, Zixu Zhao, Shenjun Zhong

Medical visual question answering (VQA) is a challenging task that requires answering clinical questions of a given medical image, by taking consider of both visual and language information.

Medical Visual Question Answering

Cannot find the paper you are looking for? You can Submit a new open access paper.