Search Results for author: Zhenjia Xu

Found 12 papers, 5 papers with code

DoughNet: A Visual Predictive Model for Topological Manipulation of Deformable Objects

no code implementations18 Apr 2024 Dominik Bauer, Zhenjia Xu, Shuran Song

Manipulation of elastoplastic objects like dough often involves topological changes such as splitting and merging.

Denoising

XSkill: Cross Embodiment Skill Discovery

1 code implementation19 Jul 2023 Mengda Xu, Zhenjia Xu, Cheng Chi, Manuela Veloso, Shuran Song

Human demonstration videos are a widely available data source for robot learning and an intuitive user interface for expressing desired behavior.

Imitation Learning Robot Manipulation

Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners

no code implementations4 Jul 2023 Allen Z. Ren, Anushri Dixit, Alexandra Bodrova, Sumeet Singh, Stephen Tu, Noah Brown, Peng Xu, Leila Takayama, Fei Xia, Jake Varley, Zhenjia Xu, Dorsa Sadigh, Andy Zeng, Anirudha Majumdar

Large language models (LLMs) exhibit a wide range of promising capabilities -- from step-by-step planning to commonsense reasoning -- that may provide utility for robots, but remain prone to confidently hallucinated predictions.

Conformal Prediction Language Modelling +1

Towards Generalist Robots: A Promising Paradigm via Generative Simulation

no code implementations17 May 2023 Zhou Xian, Theophile Gervet, Zhenjia Xu, Yi-Ling Qiao, Tsun-Hsuan Wang, Yian Wang

This document serves as a position paper that outlines the authors' vision for a potential pathway towards generalist robots.

Scene Generation

FluidLab: A Differentiable Environment for Benchmarking Complex Fluid Manipulation

1 code implementation4 Mar 2023 Zhou Xian, Bo Zhu, Zhenjia Xu, Hsiao-Yu Tung, Antonio Torralba, Katerina Fragkiadaki, Chuang Gan

We identify several challenges for fluid manipulation learning by evaluating a set of reinforcement learning and trajectory optimization methods on our platform.

Benchmarking

BusyBot: Learning to Interact, Reason, and Plan in a BusyBoard Environment

1 code implementation17 Jul 2022 Zeyi Liu, Zhenjia Xu, Shuran Song

We introduce BusyBoard, a toy-inspired robot learning environment that leverages a diverse set of articulated objects and inter-object functional relations to provide rich visual feedback for robot interactions.

Causal Discovery Robot Manipulation +2

UMPNet: Universal Manipulation Policy Network for Articulated Objects

no code implementations13 Sep 2021 Zhenjia Xu, Zhanpeng He, Shuran Song

We introduce the Universal Manipulation Policy Network (UMPNet) -- a single image-based policy network that infers closed-loop action sequences for manipulating arbitrary articulated objects.

Attribute

AdaGrasp: Learning an Adaptive Gripper-Aware Grasping Policy

1 code implementation28 Nov 2020 Zhenjia Xu, Beichun Qi, Shubham Agrawal, Shuran Song

We propose AdaGrasp, a method to learn a single grasping policy that generalizes to novel grippers.

Robotics

Learning 3D Dynamic Scene Representations for Robot Manipulation

2 code implementations3 Nov 2020 Zhenjia Xu, Zhanpeng He, Jiajun Wu, Shuran Song

3D scene representation for robot manipulation should capture three key object properties: permanency -- objects that become occluded over time continue to exist; amodal completeness -- objects have 3D occupancy, even if only partial observations are available; spatiotemporal continuity -- the movement of each object is continuous over space and time.

Model Predictive Control Robot Manipulation

Modeling Parts, Structure, and System Dynamics via Predictive Learning

no code implementations ICLR 2019 Zhenjia Xu, Zhijian Liu, Chen Sun, Kevin Murphy, William T. Freeman, Joshua B. Tenenbaum, Jiajun Wu

Humans easily recognize object parts and their hierarchical structure by watching how they move; they can then predict how each part moves in the future.

Object

Unsupervised Discovery of Parts, Structure, and Dynamics

no code implementations12 Mar 2019 Zhenjia Xu, Zhijian Liu, Chen Sun, Kevin Murphy, William T. Freeman, Joshua B. Tenenbaum, Jiajun Wu

Humans easily recognize object parts and their hierarchical structure by watching how they move; they can then predict how each part moves in the future.

Object

Cannot find the paper you are looking for? You can Submit a new open access paper.