Search Results for author: Haozhan Shen

Found 2 papers, 2 papers with code

GroundVLP: Harnessing Zero-shot Visual Grounding from Vision-Language Pre-training and Open-Vocabulary Object Detection

1 code implementation22 Dec 2023 Haozhan Shen, Tiancheng Zhao, Mingwei Zhu, Jianwei Yin

Visual grounding, a crucial vision-language task involving the understanding of the visual context based on the query expression, necessitates the model to capture the interactions between objects, as well as various spatial and attribute information.

Attribute object-detection +2

VL-CheckList: Evaluating Pre-trained Vision-Language Models with Objects, Attributes and Relations

1 code implementation1 Jul 2022 Tiancheng Zhao, Tianqi Zhang, Mingwei Zhu, Haozhan Shen, Kyusong Lee, Xiaopeng Lu, Jianwei Yin

Inspired by the CheckList for testing natural language processing, we exploit VL-CheckList, a novel framework to understand the capabilities of VLP models.

Cannot find the paper you are looking for? You can Submit a new open access paper.