Search Results for author: Zhenjia Xu

Found 17 papers, 5 papers with code

Sim-and-Real Co-Training: A Simple Recipe for Vision-Based Robotic Manipulation

no code implementations31 Mar 2025 Abhiram Maddukuri, Zhenyu Jiang, Lawrence Yunliang Chen, Soroush Nasiriany, Yuqi Xie, Yu Fang, Wenqi Huang, Zu Wang, Zhenjia Xu, Nikita Chernyadev, Scott Reed, Ken Goldberg, Ajay Mandlekar, Linxi Fan, Yuke Zhu

Large real-world robot datasets hold great potential to train generalist robot models, but scaling real-world human data collection is time-consuming and resource-intensive.

DexMimicGen: Automated Data Generation for Bimanual Dexterous Manipulation via Imitation Learning

no code implementations31 Oct 2024 Zhenyu Jiang, Yuqi Xie, Kevin Lin, Zhenjia Xu, Weikang Wan, Ajay Mandlekar, Linxi Fan, Yuke Zhu

To this end, we introduce DexMimicGen, a large-scale automated data generation system that synthesizes trajectories from a handful of human demonstrations for humanoid robots with dexterous hands.

Imitation Learning

One-Step Diffusion Policy: Fast Visuomotor Policies via Diffusion Distillation

no code implementations28 Oct 2024 Zhendong Wang, Zhaoshuo Li, Ajay Mandlekar, Zhenjia Xu, Jiaojiao Fan, Yashraj Narang, Linxi Fan, Yuke Zhu, Yogesh Balaji, Mingyuan Zhou, Ming-Yu Liu, Yu Zeng

Diffusion models, praised for their success in generative tasks, are increasingly being applied to robotics, demonstrating exceptional performance in behavior cloning.

Denoising

Flow as the Cross-Domain Manipulation Interface

no code implementations21 Jul 2024 Mengda Xu, Zhenjia Xu, Yinghao Xu, Cheng Chi, Gordon Wetzstein, Manuela Veloso, Shuran Song

By leveraging real-world human videos and simulated robot play data, we bypass the challenges of teleoperating physical robots in the real world, resulting in a scalable system for diverse tasks.

Object

DoughNet: A Visual Predictive Model for Topological Manipulation of Deformable Objects

no code implementations18 Apr 2024 Dominik Bauer, Zhenjia Xu, Shuran Song

Manipulation of elastoplastic objects like dough often involves topological changes such as splitting and merging.

Denoising

XSkill: Cross Embodiment Skill Discovery

1 code implementation19 Jul 2023 Mengda Xu, Zhenjia Xu, Cheng Chi, Manuela Veloso, Shuran Song

Human demonstration videos are a widely available data source for robot learning and an intuitive user interface for expressing desired behavior.

Imitation Learning Robot Manipulation

Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners

no code implementations4 Jul 2023 Allen Z. Ren, Anushri Dixit, Alexandra Bodrova, Sumeet Singh, Stephen Tu, Noah Brown, Peng Xu, Leila Takayama, Fei Xia, Jake Varley, Zhenjia Xu, Dorsa Sadigh, Andy Zeng, Anirudha Majumdar

Large language models (LLMs) exhibit a wide range of promising capabilities -- from step-by-step planning to commonsense reasoning -- that may provide utility for robots, but remain prone to confidently hallucinated predictions.

Conformal Prediction Language Modeling +2

Towards Generalist Robots: A Promising Paradigm via Generative Simulation

no code implementations17 May 2023 Zhou Xian, Theophile Gervet, Zhenjia Xu, Yi-Ling Qiao, Tsun-Hsuan Wang, Yian Wang

This document serves as a position paper that outlines the authors' vision for a potential pathway towards generalist robots.

Scene Generation

FluidLab: A Differentiable Environment for Benchmarking Complex Fluid Manipulation

1 code implementation4 Mar 2023 Zhou Xian, Bo Zhu, Zhenjia Xu, Hsiao-Yu Tung, Antonio Torralba, Katerina Fragkiadaki, Chuang Gan

We identify several challenges for fluid manipulation learning by evaluating a set of reinforcement learning and trajectory optimization methods on our platform.

Benchmarking

BusyBot: Learning to Interact, Reason, and Plan in a BusyBoard Environment

1 code implementation17 Jul 2022 Zeyi Liu, Zhenjia Xu, Shuran Song

We introduce BusyBoard, a toy-inspired robot learning environment that leverages a diverse set of articulated objects and inter-object functional relations to provide rich visual feedback for robot interactions.

Causal Discovery Robot Manipulation +2

UMPNet: Universal Manipulation Policy Network for Articulated Objects

no code implementations13 Sep 2021 Zhenjia Xu, Zhanpeng He, Shuran Song

We introduce the Universal Manipulation Policy Network (UMPNet) -- a single image-based policy network that infers closed-loop action sequences for manipulating arbitrary articulated objects.

Attribute

AdaGrasp: Learning an Adaptive Gripper-Aware Grasping Policy

1 code implementation28 Nov 2020 Zhenjia Xu, Beichun Qi, Shubham Agrawal, Shuran Song

We propose AdaGrasp, a method to learn a single grasping policy that generalizes to novel grippers.

Robotics

Learning 3D Dynamic Scene Representations for Robot Manipulation

2 code implementations3 Nov 2020 Zhenjia Xu, Zhanpeng He, Jiajun Wu, Shuran Song

3D scene representation for robot manipulation should capture three key object properties: permanency -- objects that become occluded over time continue to exist; amodal completeness -- objects have 3D occupancy, even if only partial observations are available; spatiotemporal continuity -- the movement of each object is continuous over space and time.

Model Predictive Control Robot Manipulation

Modeling Parts, Structure, and System Dynamics via Predictive Learning

no code implementations ICLR 2019 Zhenjia Xu, Zhijian Liu, Chen Sun, Kevin Murphy, William T. Freeman, Joshua B. Tenenbaum, Jiajun Wu

Humans easily recognize object parts and their hierarchical structure by watching how they move; they can then predict how each part moves in the future.

Object

Unsupervised Discovery of Parts, Structure, and Dynamics

no code implementations12 Mar 2019 Zhenjia Xu, Zhijian Liu, Chen Sun, Kevin Murphy, William T. Freeman, Joshua B. Tenenbaum, Jiajun Wu

Humans easily recognize object parts and their hierarchical structure by watching how they move; they can then predict how each part moves in the future.

Object

Cannot find the paper you are looking for? You can Submit a new open access paper.