no code implementations • 2 Jul 2023 • Hao-Shu Fang, Hongjie Fang, Zhenyu Tang, Jirong Liu, Chenxi Wang, JunBo Wang, Haoyi Zhu, Cewu Lu
A key challenge in robotic manipulation in open domains is how to acquire diverse and generalizable skills for robots.
no code implementations • CVPR 2023 • Jirong Liu, Ruo Zhang, Hao-Shu Fang, Minghao Gou, Hongjie Fang, Chenxi Wang, Sheng Xu, Hengxu Yan, Cewu Lu
Reactive grasping, which enables the robot to successfully grasp dynamic moving objects, is of great interest in robotics.
1 code implementation • 14 Nov 2022 • Yong-Lu Li, Hongwei Fan, Zuoyu Qiu, Yiming Dou, Liang Xu, Hao-Shu Fang, Peiyang Guo, Haisheng Su, Dongliang Wang, Wei Wu, Cewu Lu
In daily HOIs, humans often interact with a variety of objects, e. g., holding and touching dozens of household items in cleaning.
7 code implementations • 7 Nov 2022 • Hao-Shu Fang, Jiefeng Li, Hongyang Tang, Chao Xu, Haoyi Zhu, Yuliang Xiu, Yong-Lu Li, Cewu Lu
Accurate whole-body multi-person pose estimation and tracking is an important yet challenging topic in computer vision.
1 code implementation • 11 Oct 2022 • Haoyi Zhu, Hao-Shu Fang, Cewu Lu
In this paper, we focus on a rarely discussed but important setting: can we train one model that can represent multiple scenes, with 360$^\circ $ insufficient views and RGB-D images?
no code implementations • 23 Jun 2022 • Minghao Gou, Haolin Pan, Hao-Shu Fang, Ziyuan Liu, Cewu Lu, Ping Tan
In this paper, we propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
1 code implementation • 17 Feb 2022 • Hongjie Fang, Hao-Shu Fang, Sheng Xu, Cewu Lu
However, the majority of current grasping algorithms would fail in this case since they heavily rely on the depth image, while ordinary depth sensors usually fail to produce accurate depth information for transparent objects owing to the reflection and refraction of light.
Ranked #1 on Transparent Object Depth Estimation on TransCG
3 code implementations • 14 Feb 2022 • Yong-Lu Li, Xinpeng Liu, Xiaoqian Wu, Yizhuo Li, Zuoyu Qiu, Liang Xu, Yue Xu, Hao-Shu Fang, Cewu Lu
Human activity understanding is of widespread interest in artificial intelligence and spans diverse applications like health care and behavior analysis.
no code implementations • CVPR 2022 • Jianhua Sun, YuXuan Li, Liang Chai, Hao-Shu Fang, Yong-Lu Li, Cewu Lu
Human trajectory prediction task aims to analyze human future movements given their past status, which is a crucial step for many autonomous systems such as self-driving cars and social robots.
no code implementations • 23 Mar 2021 • Hanwen Cao, Hao-Shu Fang, Wenhai Liu, Cewu Lu
Meanwhile, we propose a method to predict numerous suction poses from an RGB-D image of a cluttered scene and demonstrate our superiority against several previous methods.
1 code implementation • ICCV 2021 • Jianhua Sun, YuXuan Li, Hao-Shu Fang, Cewu Lu
Multimodal prediction results are essential for trajectory prediction task as there is no single correct answer for the future.
1 code implementation • 3 Mar 2021 • Minghao Gou, Hao-Shu Fang, Zhanda Zhu, Sheng Xu, Chenxi Wang, Cewu Lu
In the first stage, an encoder-decoder like convolutional neural network Angle-View Net(AVN) is proposed to predict the SO(3) orientation of the gripper at every location of the image.
1 code implementation • ICCV 2021 • Chenxi Wang, Hao-Shu Fang, Minghao Gou, Hongjie Fang, Jin Gao, Cewu Lu
To quickly detect graspness in practice, we develop a neural network named graspness model to approximate the searching process.
Ranked #3 on Robotic Grasping on GraspNet-1Billion
1 code implementation • 2 Oct 2020 • Hao-Shu Fang, Yichen Xie, Dian Shao, Cewu Lu
On the other hand, existing one-stage methods mainly focus on the union regions of interactions, which introduce unnecessary visual information as disturbances to HOI detection.
Ranked #15 on Human-Object Interaction Detection on V-COCO
no code implementations • 2 Oct 2020 • Yichen Xie, Hao-Shu Fang, Dian Shao, Yong-Lu Li, Cewu Lu
Human-object interaction (HOI) detection requires a large amount of annotated data.
Ranked #68 on Domain Generalization on PACS
1 code implementation • CVPR 2020 • Hao-Shu Fang, Chenxi Wang, Minghao Gou, Cewu Lu
In this work, we contribute a large-scale grasp pose detection dataset with a unified evaluation system.
Ranked #6 on Robotic Grasping on GraspNet-1Billion
2 code implementations • CVPR 2020 • Yong-Lu Li, Liang Xu, Xinpeng Liu, Xijie Huang, Yue Xu, Shiyi Wang, Hao-Shu Fang, Ze Ma, Mingyang Chen, Cewu Lu
In light of this, we propose a new path: infer human part states first and then reason out the activities based on part-level semantics.
Ranked #3 on Human-Object Interaction Detection on HICO
no code implementations • 31 Dec 2019 • Hao-Shu Fang, Chenxi Wang, Minghao Gou, Cewu Lu
Object grasping is critical for many applications, which is also a challenging computer vision problem.
3 code implementations • ICCV 2019 • Hao-Shu Fang, Jianhua Sun, Runzhong Wang, Minghao Gou, Yong-Lu Li, Cewu Lu
With the guidance of such map, we boost the performance of R101-Mask R-CNN on instance segmentation from 35. 7 mAP to 37. 9 mAP without modifying the backbone or network structure.
Ranked #78 on Instance Segmentation on COCO test-dev
no code implementations • ICCV 2019 • Jinkun Cao, Hongyang Tang, Hao-Shu Fang, Xiaoyong Shen, Cewu Lu, Yu-Wing Tai
Therefore, the easily available human pose dataset, which is of a much larger scale than our labeled animal dataset, provides important prior knowledge to boost up the performance on animal pose estimation.
4 code implementations • 13 Apr 2019 • Yong-Lu Li, Liang Xu, Xinpeng Liu, Xijie Huang, Yue Xu, Mingyang Chen, Ze Ma, Shiyi Wang, Hao-Shu Fang, Cewu Lu
To address these and promote the activity understanding, we build a large-scale Human Activity Knowledge Engine (HAKE) based on the human body part states.
Ranked #2 on Human-Object Interaction Detection on HICO (using extra training data)
6 code implementations • 4 Dec 2018 • Zelin Zhao, Gao Peng, Haoyu Wang, Hao-Shu Fang, Chengkun Li, Cewu Lu
In this paper, we present an accurate yet effective solution for 6D pose estimation from an RGB image.
Ranked #17 on 6D Pose Estimation using RGB on LineMOD
3 code implementations • CVPR 2019 • Jiefeng Li, Can Wang, Hao Zhu, Yihuan Mao, Hao-Shu Fang, Cewu Lu
In this paper, we propose a novel and efficient method to tackle the problem of pose estimation in the crowd and a new dataset to better evaluate algorithms.
Ranked #6 on Multi-Person Pose Estimation on OCHuman
3 code implementations • CVPR 2019 • Yong-Lu Li, Siyuan Zhou, Xijie Huang, Liang Xu, Ze Ma, Hao-Shu Fang, Yan-Feng Wang, Cewu Lu
On account of the generalization of interactiveness, interactiveness network is a transferable knowledge learner and can be cooperated with any HOI detection models to achieve desirable results.
Ranked #29 on Human-Object Interaction Detection on V-COCO
1 code implementation • ECCV 2018 • Hao-Shu Fang, Jinkun Cao, Yu-Wing Tai, Cewu Lu
We propose a new pairwise body-part attention model which can learn to focus on crucial parts, and their correlations for HOI recognition.
Ranked #5 on Human-Object Interaction Detection on HICO
1 code implementation • CVPR 2018 • Hao-Shu Fang, Guansong Lu, Xiaolin Fang, Jianwen Xie, Yu-Wing Tai, Cewu Lu
In this paper, we present a novel method to generate synthetic human part segmentation data using easily-obtained human keypoint annotations.
Ranked #4 on Human Part Segmentation on PASCAL-Part (using extra training data)
no code implementations • 17 Oct 2017 • Hao-Shu Fang, Yuanlu Xu, Wenguan Wang, Xiaobai Liu, Song-Chun Zhu
In this paper, we propose a pose grammar to tackle the problem of 3D human pose estimation.
Ranked #1 on 3D Absolute Human Pose Estimation on Human3.6M (Average MPJPE (mm) metric)
14 code implementations • ICCV 2017 • Hao-Shu Fang, Shuqin Xie, Yu-Wing Tai, Cewu Lu
In this paper, we propose a novel regional multi-person pose estimation (RMPE) framework to facilitate pose estimation in the presence of inaccurate human bounding boxes.
Ranked #1 on Pose Estimation on UAV-Human