Search Results for author: Shangyu Xing

Found 5 papers, 2 papers with code

GePBench: Evaluating Fundamental Geometric Perception for Multimodal Large Language Models

no code implementations30 Dec 2024 Shangyu Xing, Changhao Xiang, Yuteng Han, Yifan Yue, Zhen Wu, Xinyu Liu, Zhangtai Wu, Fei Zhao, Xinyu Dai

To address this limitation, we introduce GePBench, a novel benchmark designed to assess the geometric perception capabilities of MLLMs.

EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models

1 code implementation15 Feb 2024 Shangyu Xing, Fei Zhao, Zhen Wu, Tuo An, WeiHao Chen, Chunhui Li, Jianbing Zhang, Xinyu Dai

Multimodal large language models (MLLMs) have attracted increasing attention in the past few years, but they may still generate descriptions that include objects not present in the corresponding images, a phenomenon known as object hallucination.

Hallucination Object Hallucination

DRIN: Dynamic Relation Interactive Network for Multimodal Entity Linking

1 code implementation9 Oct 2023 Shangyu Xing, Fei Zhao, Zhen Wu, Chunhui Li, Jianbing Zhang, Xinyu Dai

Multimodal Entity Linking (MEL) is a task that aims to link ambiguous mentions within multimodal contexts to referential entities in a multimodal knowledge base.

Entity Linking Relation

Cannot find the paper you are looking for? You can Submit a new open access paper.