Search Results for author: Kuan Fang

Found 27 papers, 4 papers with code

Temporal Representation Alignment: Successor Features Enable Emergent Compositionality in Robot Instruction Following

no code implementations8 Feb 2025 Vivek Myers, Bill Chunyuan Zheng, Anca Dragan, Kuan Fang, Sergey Levine

Effective task representations should facilitate compositionality, such that after learning a variety of basic tasks, an agent can perform compound tasks consisting of multiple steps simply by composing the representations of the constituent steps together.

Instruction Following

Planning-Guided Diffusion Policy Learning for Generalizable Contact-Rich Bimanual Manipulation

no code implementations3 Dec 2024 Xuanlin Li, Tong Zhao, Xinghao Zhu, Jiuguang Wang, Tao Pang, Kuan Fang

Contact-rich bimanual manipulation involves precise coordination of two arms to change object states through strategically selected contacts and motions.

Data Augmentation

KALIE: Fine-Tuning Vision-Language Models for Open-World Manipulation without Robot Data

no code implementations21 Sep 2024 Grace Tang, Swetha Rajkumar, Yifei Zhou, Homer Rich Walke, Sergey Levine, Kuan Fang

Building generalist robotic systems involves effectively endowing robots with the capabilities to handle novel objects in an open-world setting.

Policy Adaptation via Language Optimization: Decomposing Tasks for Few-Shot Imitation

no code implementations29 Aug 2024 Vivek Myers, Bill Chunyuan Zheng, Oier Mees, Sergey Levine, Kuan Fang

Learned language-conditioned robot policies often struggle to effectively adapt to new real-world tasks even when pre-trained across a diverse set of instructions.

Robot Manipulation

Affordance-Guided Reinforcement Learning via Visual Prompting

no code implementations14 Jul 2024 Olivia Y. Lee, Annie Xie, Kuan Fang, Karl Pertsch, Chelsea Finn

On real-world manipulation tasks specified by natural language descriptions, KAGI improves the sample efficiency of autonomous RL and enables successful task completion in 30K online fine-tuning steps.

reinforcement-learning Reinforcement Learning +3

MOKA: Open-World Robotic Manipulation through Mark-Based Visual Prompting

no code implementations5 Mar 2024 Fangchen Liu, Kuan Fang, Pieter Abbeel, Sergey Levine

Central to our approach is a compact point-based representation of affordance, which bridges the VLM's predictions on observed images and the robot's actions in the physical world.

In-Context Learning Object Rearrangement +3

Multi-Stage Cable Routing through Hierarchical Imitation Learning

no code implementations18 Jul 2023 Jianlan Luo, Charles Xu, Xinyang Geng, Gilbert Feng, Kuan Fang, Liam Tan, Stefan Schaal, Sergey Levine

In such settings, learning individual primitives for each stage that succeed with a high enough rate to perform a complete temporally extended task is impractical: if each stage must be completed successfully and has a non-negligible probability of failure, the likelihood of successful completion of the entire task becomes negligible.

Imitation Learning

Goal Representations for Instruction Following: A Semi-Supervised Language Interface to Control

no code implementations30 Jun 2023 Vivek Myers, Andre He, Kuan Fang, Homer Walke, Philippe Hansen-Estruch, Ching-An Cheng, Mihai Jalobeanu, Andrey Kolobov, Anca Dragan, Sergey Levine

Our method achieves robust performance in the real world by learning an embedding from the labeled data that aligns language not to the goal image, but rather to the desired change between the start and goal images that the instruction corresponds to.

Instruction Following

Stabilizing Contrastive RL: Techniques for Robotic Goal Reaching from Offline Data

1 code implementation6 Jun 2023 Chongyi Zheng, Benjamin Eysenbach, Homer Walke, Patrick Yin, Kuan Fang, Ruslan Salakhutdinov, Sergey Levine

Robotic systems that rely primarily on self-supervised learning have the potential to decrease the amount of human annotation and engineering effort required to learn control strategies.

Contrastive Learning Data Augmentation +2

Active Task Randomization: Learning Robust Skills via Unsupervised Generation of Diverse and Feasible Tasks

no code implementations11 Nov 2022 Kuan Fang, Toki Migimatsu, Ajay Mandlekar, Li Fei-Fei, Jeannette Bohg

ATR selects suitable tasks, which consist of an initial environment state and manipulation goal, for learning robust skills by balancing the diversity and feasibility of the tasks.

Diversity

Generalization with Lossy Affordances: Leveraging Broad Offline Data for Learning Visuomotor Tasks

no code implementations12 Oct 2022 Kuan Fang, Patrick Yin, Ashvin Nair, Homer Walke, Gengchen Yan, Sergey Levine

The utilization of broad datasets has proven to be crucial for generalization for a wide range of fields.

Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space

no code implementations17 May 2022 Kuan Fang, Patrick Yin, Ashvin Nair, Sergey Levine

Our experimental results show that PTP can generate feasible sequences of subgoals that enable the policy to efficiently solve the target tasks.

reinforcement-learning Reinforcement Learning (RL)

Discovering Generalizable Skills via Automated Generation of Diverse Tasks

no code implementations26 Jun 2021 Kuan Fang, Yuke Zhu, Silvio Savarese, Li Fei-Fei

To encourage generalizable skills to emerge, our method trains each skill to specialize in the paired task and maximizes the diversity of the generated tasks.

Diversity Hierarchical Reinforcement Learning +2

Synergies Between Affordance and Geometry: 6-DoF Grasp Detection via Implicit Representations

1 code implementation4 Apr 2021 Zhenyu Jiang, Yifeng Zhu, Maxwell Svetlik, Kuan Fang, Yuke Zhu

The experimental results in simulation and on the real robot have demonstrated that the use of implicit neural representations and joint learning of grasp affordance and 3D reconstruction have led to state-of-the-art grasping results.

3D Reconstruction Multi-Task Learning

Adaptive Procedural Task Generation for Hard-Exploration Problems

no code implementations ICLR 2021 Kuan Fang, Yuke Zhu, Silvio Savarese, Li Fei-Fei

To enable curriculum learning in the absence of a direct indicator of learning progress, we propose to train the task generator by balancing the agent's performance in the generated tasks and the similarity to the target tasks.

SERank: Optimize Sequencewise Learning to Rank Using Squeeze-and-Excitation Network

1 code implementation7 Jun 2020 RuiXing Wang, Kuan Fang, RiKang Zhou, Zhan Shen, LiWen Fan

Recently, there are a few methods have been proposed which focused on mining information across ranking candidates list for further improvements, such as learning multivariant scoring function or learning contextual embedding.

Learning-To-Rank Question Answering

Dynamics Learning with Cascaded Variational Inference for Multi-Step Manipulation

no code implementations29 Oct 2019 Kuan Fang, Yuke Zhu, Animesh Garg, Silvio Savarese, Li Fei-Fei

The fundamental challenge of planning for multi-step manipulation is to find effective and plausible action sequences that lead to the task goal.

Variational Inference

KETO: Learning Keypoint Representations for Tool Manipulation

no code implementations26 Oct 2019 Zengyi Qin, Kuan Fang, Yuke Zhu, Li Fei-Fei, Silvio Savarese

For this purpose, we present KETO, a framework of learning keypoint representations of tool-based manipulation.

Robotics

Learning Task-Oriented Grasping for Tool Manipulation from Simulated Self-Supervision

no code implementations25 Jun 2018 Kuan Fang, Yuke Zhu, Animesh Garg, Andrey Kurenkov, Viraj Mehta, Li Fei-Fei, Silvio Savarese

We perform both simulated and real-world experiments on two tool-based manipulation tasks: sweeping and hammering.

Recurrent Autoregressive Networks for Online Multi-Object Tracking

no code implementations7 Nov 2017 Kuan Fang, Yu Xiang, Xiaocheng Li, Silvio Savarese

The external memory explicitly stores previous inputs of each trajectory in a time window, while the internal memory learns to summarize long-term tracking history and associate detections by processing the external memory.

Multi-Object Tracking Object +1

DeLay: Robust Spatial Layout Estimation for Cluttered Indoor Scenes

no code implementations CVPR 2016 Saumitro Dasgupta, Kuan Fang, Kevin Chen, Silvio Savarese

We consider the problem of estimating the spatial layout of an indoor scene from a monocular RGB image, modeled as the projection of a 3D cuboid.

Cannot find the paper you are looking for? You can Submit a new open access paper.