Search Results for author: JunBo Wang

Found 8 papers, 1 papers with code

RPMArt: Towards Robust Perception and Manipulation for Articulated Objects

no code implementations24 Mar 2024 JunBo Wang, Wenhai Liu, Qiaojun Yu, Yang You, Liu Liu, Weiming Wang, Cewu Lu

Our primary contribution is a Robust Articulation Network (RoArtNet) that is able to predict both joint parameters and affordable points robustly by local feature learning and point tuple voting.

ManiPose: A Comprehensive Benchmark for Pose-aware Object Manipulation in Robotics

no code implementations20 Mar 2024 Qiaojun Yu, Ce Hao, JunBo Wang, Wenhai Liu, Liu Liu, Yao Mu, Yang You, Hengxu Yan, Cewu Lu

Robotic manipulation in everyday scenarios, especially in unstructured environments, requires skills in pose-aware object manipulation (POM), which adapts robots' grasping and handling according to an object's 6D pose.

Motion Planning Pose Estimation

Spatial-Temporal Hypergraph Neural Network for Traffic Forecasting

no code implementations24 Oct 2023 Chengzhi Yao, Zhi Li, JunBo Wang

To tackle the above issues, we focus on the essence of traffic system and propose STHODE: Spatio-Temporal Hypergraph Neural Ordinary Differential Equation Network, which combines road network topology and traffic dynamics to capture high-order spatio-temporal dependencies in traffic data.

GAMMA: Generalizable Articulation Modeling and Manipulation for Articulated Objects

1 code implementation28 Sep 2023 Qiaojun Yu, JunBo Wang, Wenhai Liu, Ce Hao, Liu Liu, Lin Shao, Weiming Wang, Cewu Lu

Results show that GAMMA significantly outperforms SOTA articulation modeling and manipulation algorithms in unseen and cross-category articulated objects.

Manner Of Articulation Detection Robot Manipulation +1

M3: Multimodal Memory Modelling for Video Captioning

no code implementations CVPR 2018 Junbo Wang, Wei Wang, Yan Huang, Liang Wang, Tieniu Tan

Inspired by the facts that memory modelling poses potential advantages to long-term sequential problems [35] and working memory is the key factor of visual attention [33], we propose a Multimodal Memory Model (M3) to describe videos, which builds a visual and textual shared memory to model the long-term visual-textual dependency and further guide visual attention on described visual targets to solve visual-textual alignments.

Sentence Video Captioning

Multimodal Memory Modelling for Video Captioning

no code implementations17 Nov 2016 Junbo Wang, Wei Wang, Yan Huang, Liang Wang, Tieniu Tan

In this paper, we propose a Multimodal Memory Model (M3) to describe videos, which builds a visual and textual shared memory to model the long-term visual-textual dependency and further guide global visual attention on described targets.

Sentence Video Captioning

Cannot find the paper you are looking for? You can Submit a new open access paper.