Search Results for author: David Hsu

Found 48 papers, 15 papers with code

DESPOT: Online POMDP Planning with Regularization

1 code implementation NeurIPS 2013 Nan Ye, Adhiraj Somani, David Hsu, Wee Sun Lee

We show that the best policy obtained from a DESPOT is near-optimal, with a regret bound that depends on the representation size of the optimal policy.

Autonomous Driving

SUMMIT: A Simulator for Urban Driving in Massive Mixed Traffic

3 code implementations11 Nov 2019 Panpan Cai, Yiyuan Lee, Yuanfu Luo, David Hsu

Autonomous driving in an unregulated urban crowd is an outstanding challenge, especially, in the presence of many aggressive, high-speed traffic participants.

Robotics Multiagent Systems

Particle Filter Networks with Application to Visual Localization

2 code implementations23 May 2018 Peter Karkus, David Hsu, Wee Sun Lee

Particle filtering is a powerful approach to sequential state estimation and finds application in many domains, including robot localization, object tracking, etc.

Object Tracking Visual Localization

Particle Filter Recurrent Neural Networks

1 code implementation30 May 2019 Xiao Ma, Peter Karkus, David Hsu, Wee Sun Lee

Recurrent neural networks (RNNs) have been extraordinarily successful for prediction with sequential data.

General Classification Stock Price Prediction +2

QMDP-Net: Deep Learning for Planning under Partial Observability

2 code implementations NeurIPS 2017 Peter Karkus, David Hsu, Wee Sun Lee

It is a recurrent policy network, but it represents a policy for a parameterized set of tasks by connecting a model with a planning algorithm that solves the model, thus embedding the solution structure of planning in a network learning architecture.

HyP-DESPOT: A Hybrid Parallel Algorithm for Online Planning under Uncertainty

1 code implementation17 Feb 2018 Panpan Cai, Yuanfu Luo, David Hsu, Wee Sun Lee

Planning under uncertainty is critical for robust robot performance in uncertain, dynamic environments, but it incurs high computational cost.

Computational Efficiency

Discriminative Particle Filter Reinforcement Learning for Complex Partial Observations

1 code implementation ICLR 2020 Xiao Ma, Peter Karkus, David Hsu, Wee Sun Lee, Nan Ye

The particle filter maintains a belief using learned discriminative update, which is trained end-to-end for decision making.

Atari Games Decision Making +3

LEADER: Learning Attention over Driving Behaviors for Planning under Uncertainty

1 code implementation23 Sep 2022 Mohamad H. Danesh, Panpan Cai, David Hsu

To address this, we propose a new algorithm, LEarning Attention over Driving bEhavioRs (LEADER), that learns to attend to critical human behaviors during planning.

Autonomous Driving

Contrastive Variational Reinforcement Learning for Complex Observations

1 code implementation6 Aug 2020 Xiao Ma, Siwei Chen, David Hsu, Wee Sun Lee

This paper presents Contrastive Variational Reinforcement Learning (CVRL), a model-based method that tackles complex visual observations in DRL.

Atari Games Continuous Control +4

MAGIC: Learning Macro-Actions for Online POMDP Planning

1 code implementation7 Nov 2020 Yiyuan Lee, Panpan Cai, David Hsu

The partially observable Markov decision process (POMDP) is a principled general framework for robot decision making under uncertainty, but POMDP planning suffers from high computational complexity, when long-term planning is required.

Computational Efficiency Decision Making +1

Intention-Net: Integrating Planning and Deep Learning for Goal-Directed Autonomous Navigation

2 code implementations16 Oct 2017 Wei Gao, David Hsu, Wee Sun Lee, ShengMei Shen, Karthikk Subramanian

How can a delivery robot navigate reliably to a destination in a new office building, with minimal prior information?

Autonomous Navigation Navigate

GAMMA: A General Agent Motion Model for Autonomous Driving

1 code implementation4 Jun 2019 Yuanfu Luo, Panpan Cai, Yiyuan Lee, David Hsu

Further, the computational efficiency and the flexibility of GAMMA enable (i) simulation of mixed urban traffic at many locations worldwide and (ii) planning for autonomous driving in dense traffic with uncertain driver behaviors, both in real-time.

Autonomous Driving Collision Avoidance +2

Push-Net: Deep Planar Pushing for Objects with Unknown Physical Properties

1 code implementation Robotics: Science and Systems 2018 Jue Kun Li, David Hsu, Wee Sun Lee

This paper introduces Push-Net, a deep recurrent neural network model, which enables a robot to push objects of unknown physical properties for re-positioning and re-orientation, using only visual camera images as input.

DinerDash Gym: A Benchmark for Policy Learning in High-Dimensional Action Space

1 code implementation13 Jul 2020 Siwei Chen, Xiao Ma, David Hsu

It has been arduous to assess the progress of a policy learning algorithm in the domain of hierarchical task with high dimensional action space due to the lack of a commonly accepted benchmark.

Atari Games

Hindsight Trust Region Policy Optimization

1 code implementation29 Jul 2019 Hanbo Zhang, Site Bai, Xuguang Lan, David Hsu, Nanning Zheng

We propose \emph{Hindsight Trust Region Policy Optimization}(HTRPO), a new RL algorithm that extends the highly successful TRPO algorithm with \emph{hindsight} to tackle the challenge of sparse rewards.

Atari Games Policy Gradient Methods +1

Interactive Visual Grounding of Referring Expressions for Human-Robot Interaction

no code implementations11 Jun 2018 Mohit Shridhar, David Hsu

The first stage uses a neural network to generate visual descriptions of objects, compares them with the input language expression, and identifies a set of candidate objects.

Question Generation Question-Generation +1

Trust-Aware Decision Making for Human-Robot Collaboration: Model Learning and Planning

no code implementations12 Jan 2018 Min Chen, Stefanos Nikolaidis, Harold Soh, David Hsu, Siddhartha Srinivasa

The trust-POMDP model provides a principled approach for the robot to (i) infer the trust of a human teammate through interaction, (ii) reason about the effect of its own actions on human trust, and (iii) choose actions that maximize team performance over the long term.

Decision Making

Grounding Spatio-Semantic Referring Expressions for Human-Robot Interaction

no code implementations18 Jul 2017 Mohit Shridhar, David Hsu

A core issue for the system is semantic and spatial grounding, which is to infer objects and their spatial relationships from images and natural language expressions.

Object

Factored Contextual Policy Search with Bayesian Optimization

no code implementations6 Dec 2016 Peter Karkus, Andras Kupcsik, David Hsu, Wee Sun Lee

Scarce data is a major challenge to scaling robot learning to truly complex tasks, as we need to generalize locally learned policies over different "contexts".

Active Learning Bayesian Optimization +2

POMDP-lite for Robust Robot Planning under Uncertainty

no code implementations16 Feb 2016 Min Chen, Emilio Frazzoli, David Hsu, Wee Sun Lee

We show that a POMDP-lite is equivalent to a set of fully observable Markov decision processes indexed by a hidden parameter and is useful for modeling a variety of interesting robotic tasks.

Reinforcement Learning (RL)

Exploration in Interactive Personalized Music Recommendation: A Reinforcement Learning Approach

no code implementations6 Nov 2013 Xinxi Wang, Yi Wang, David Hsu, Ye Wang

Current music recommender systems typically act in a greedy fashion by recommending songs with the highest user ratings.

Bayesian Inference Music Recommendation +4

Integrating Algorithmic Planning and Deep Learning for Partially Observable Navigation

no code implementations17 Jul 2018 Peter Karkus, David Hsu, Wee Sun Lee

We propose to take a novel approach to robot system design where each building block of a larger system is represented as a differentiable program, i. e. a deep neural network.

Navigate Robot Navigation

Monte Carlo Value Iteration with Macro-Actions

no code implementations NeurIPS 2011 Zhan Lim, Lee Sun, David Hsu

The recently introduced Monte Carlo Value Iteration (MCVI) can tackle POMDPs with very large discrete state spaces or continuous state spaces, but its performance degrades when faced with long planning horizons.

Solving the Perspective-2-Point Problem for Flying-Camera Photo Composition

no code implementations CVPR 2018 Ziquan Lan, David Hsu, Gim Hee Lee

The user, instead of holding a camera in hand and manually searching for a viewpoint, will interact directly with image contents in the viewfinder through simple gestures, and the flying camera will achieve the desired viewpoint through the autonomous flying capability of the drone.

Monte Carlo Bayesian Reinforcement Learning

no code implementations27 Jun 2012 Yi Wang, Kok Sung Won, David Hsu, Wee Sun Lee

Bayesian reinforcement learning (BRL) encodes prior knowledge of the world in a model and represents uncertainty in model parameters by maintaining a probability distribution over them.

reinforcement-learning Reinforcement Learning (RL)

Differentiable Algorithm Networks for Composable Robot Learning

no code implementations28 May 2019 Peter Karkus, Xiao Ma, David Hsu, Leslie Pack Kaelbling, Wee Sun Lee, Tomas Lozano-Perez

This paper introduces the Differentiable Algorithm Network (DAN), a composable architecture for robot learning systems.

Navigate

LeTS-Drive: Driving in a Crowd by Learning from Tree Search

no code implementations29 May 2019 Panpan Cai, Yuanfu Luo, Aseem Saxena, David Hsu, Wee Sun Lee

LeTS-Drive leverages the robustness of planning and the runtime efficiency of learning to enhance the performance of both.

Autonomous Driving Imitation Learning

Robot Capability and Intention in Trust-based Decisions across Tasks

no code implementations3 Sep 2019 Yaqi Xie, Indu P Bodala, Desmond C. Ong, David Hsu, Harold Soh

In this paper, we present results from a human-subject study designed to explore two facets of human mental models of robots---inferred capability and intention---and their relationship to overall trust and eventual decisions.

PORCA: Modeling and Planning for Autonomous Driving among Many Pedestrians

no code implementations30 May 2018 Yuanfu Luo, Panpan Cai, Aniket Bera, David Hsu, Wee Sun Lee, Dinesh Manocha

Our planning system combines a POMDP algorithm with the pedestrian motion model and runs in near real time.

Robotics

Closing the Planning-Learning Loop with Application to Autonomous Driving

no code implementations11 Jan 2021 Panpan Cai, David Hsu

To achieve real-time performance for large-scale planning, this work introduces a new algorithm Learning from Tree Search for Driving (LeTS-Drive), which integrates planning and learning in a closed loop, and applies it to autonomous driving in crowded urban traffic in simulation.

Autonomous Driving Robotics

Learning Latent Graph Dynamics for Visual Manipulation of Deformable Objects

no code implementations25 Apr 2021 Xiao Ma, David Hsu, Wee Sun Lee

Manipulating deformable objects, such as ropes and clothing, is a long-standing challenge in robotics, because of their large degrees of freedom, complex non-linear dynamics, and self-occlusion in visual perception.

Contrastive Learning Deformable Object Manipulation +3

Differentiable SLAM-net: Learning Particle SLAM for Visual Navigation

no code implementations CVPR 2021 Peter Karkus, Shaojun Cai, David Hsu

We introduce the Differentiable SLAM Network (SLAM-net) along with a navigation architecture to enable planar robot navigation in previously unseen indoor environments.

Robot Navigation Simultaneous Localization and Mapping +1

Ab Initio Particle-based Object Manipulation

no code implementations19 Jul 2021 Siwei Chen, Xiao Ma, Yunfan Lu, David Hsu

Like the model-based analytic approaches to manipulation, the particle representation enables the robot to reason about the object's geometry and dynamics in order to choose suitable manipulation actions.

Object Robot Manipulation

INVIGORATE: Interactive Visual Grounding and Grasping in Clutter

no code implementations25 Aug 2021 Hanbo Zhang, Yunfan Lu, Cunjun Yu, David Hsu, Xuguang Lan, Nanning Zheng

This paper presents INVIGORATE, a robot system that interacts with human through natural language and grasps a specified object in clutter.

Blocking Object +5

Deep Visual Navigation under Partial Observability

no code implementations16 Sep 2021 Bo Ai, Wei Gao, Vinay, David Hsu

Importantly, we integrate the multiple neural network modules into a unified controller that achieves robust performance for visual navigation in complex, partially observable environments.

Imitation Learning Navigate +1

Context-Hierarchy Inverse Reinforcement Learning

no code implementations25 Feb 2022 Wei Gao, David Hsu, Wee Sun Lee

To solve these issues, we present Context Hierarchy IRL(CHIRL), a new IRL algorithm that exploits the context to scale up IRL and learn reward functions of complex behaviors.

Autonomous Driving reinforcement-learning +1

Receding Horizon Inverse Reinforcement Learning

no code implementations9 Jun 2022 Yiqing Xu, Wei Gao, David Hsu

Inverse reinforcement learning (IRL) seeks to infer a cost function that explains the underlying goals and preferences of expert demonstrations.

reinforcement-learning Reinforcement Learning (RL)

Partially Observable Markov Decision Processes in Robotics: A Survey

no code implementations21 Sep 2022 Mikko Lauri, David Hsu, Joni Pajarinen

Noisy sensing, imperfect control, and environment changes are defining characteristics of many real-world robot tasks.

Autonomous Driving

Differentiable Parsing and Visual Grounding of Natural Language Instructions for Object Placement

no code implementations1 Oct 2022 Zirui Zhao, Wee Sun Lee, David Hsu

Natural language generally describes objects and spatial relations with compositionality and ambiguity, two major obstacles to effective language grounding.

Object Relational Reasoning +1

The Planner Optimization Problem: Formulations and Frameworks

no code implementations12 Mar 2023 Yiyuan Lee, Katie Lee, Panpan Cai, David Hsu, Lydia E. Kavraki

Identifying internal parameters for planning is crucial to maximizing the performance of a planner.

On the Effective Horizon of Inverse Reinforcement Learning

no code implementations13 Jul 2023 Yiqing Xu, Finale Doshi-Velez, David Hsu

Inverse reinforcement learning (IRL) algorithms often rely on (forward) reinforcement learning or planning over a given time horizon to compute an approximately optimal policy for a hypothesized reward function and then match this policy with expert demonstrations.

Computational Efficiency reinforcement-learning

"Tidy Up the Table": Grounding Common-sense Objective for Tabletop Object Rearrangement

no code implementations21 Jul 2023 Yiqing Xu, David Hsu

Tidying up a messy table may appear simple for humans, but articulating clear criteria for tidiness is challenging due to the ambiguous nature of common sense reasoning.

Common Sense Reasoning

Invariance is Key to Generalization: Examining the Role of Representation in Sim-to-Real Transfer for Visual Navigation

no code implementations23 Oct 2023 Bo Ai, Zhanxin Wu, David Hsu

The data-driven approach to robot control has been gathering pace rapidly, yet generalization to unseen task domains remains a critical challenge.

Visual Navigation

LLM-State: Open World State Representation for Long-horizon Task Planning with Large Language Model

no code implementations29 Nov 2023 Siwei Chen, Anxing Xiao, David Hsu

We propose an open state representation that provides continuous expansion and updating of object attributes from the LLM's inherent capabilities for context understanding and historical action reasoning.

Decision Making Language Modelling +1

LLMs for Robotic Object Disambiguation

no code implementations7 Jan 2024 Connie Jiang, Yiqing Xu, David Hsu

The advantages of pre-trained large language models (LLMs) are apparent in a variety of language processing tasks.

Decision Making Navigate +2

On the Empirical Complexity of Reasoning and Planning in LLMs

no code implementations17 Apr 2024 Liwei Kang, Zirui Zhao, David Hsu, Wee Sun Lee

We found that if problems can be decomposed into a sequence of reasoning steps and learning to predict the next step has a low sample and computational complexity, explicitly outlining the reasoning chain with all necessary information for predicting the next step may improve performance.

Cannot find the paper you are looking for? You can Submit a new open access paper.