Search Results for author: Boshen Xu

Found 6 papers, 3 papers with code

EgoNCE++: Do Egocentric Video-Language Models Really Understand Hand-Object Interactions?

1 code implementation28 May 2024 Boshen Xu, Ziheng Wang, Yang Du, Zhinan Song, Sipeng Zheng, Qin Jin

Due to the occurrence of diverse EgoHOIs in the real world, we propose an open-vocabulary benchmark named EgoHOIBench to reveal the diminished performance of current egocentric video-language models (EgoVLM) on fined-grained concepts, indicating that these models still lack a full spectrum of egocentric understanding.

Action Recognition Attribute +2

SPAFormer: Sequential 3D Part Assembly with Transformers

1 code implementation9 Mar 2024 Boshen Xu, Sipeng Zheng, Qin Jin

We introduce SPAFormer, an innovative model designed to overcome the combinatorial explosion challenge in the 3D Part Assembly (3D-PA) task.

POV: Prompt-Oriented View-Agnostic Learning for Egocentric Hand-Object Interaction in the Multi-View World

1 code implementation9 Mar 2024 Boshen Xu, Sipeng Zheng, Qin Jin

We humans are good at translating third-person observations of hand-object interactions (HOI) into an egocentric view.

Open-Category Human-Object Interaction Pre-Training via Language Modeling Framework

no code implementations CVPR 2023 Sipeng Zheng, Boshen Xu, Qin Jin

Human-object interaction (HOI) has long been plagued by the conflict between limited supervised data and a vast number of possible interaction combinations in real life.

Human-Object Interaction Detection Language Modelling

Cannot find the paper you are looking for? You can Submit a new open access paper.