Search Results for author: Jaime F. Fisac

Found 21 papers, 8 papers with code

Who Plays First? Optimizing the Order of Play in Stackelberg Games with Many Robots

no code implementations14 Feb 2024 Haimin Hu, Gabriele Dragotto, Zixu Zhang, Kaiqu Liang, Bartolomeo Stellato, Jaime F. Fisac

To solve the problem, we introduce Branch and Play (B&P), an efficient and exact algorithm that provably converges to a socially optimal order of play and its Stackelberg equilibrium.

Trajectory Planning valid

Deception Game: Closing the Safety-Learning Loop in Interactive Robot Autonomy

no code implementations3 Sep 2023 Haimin Hu, Zixu Zhang, Kensuke Nakamura, Andrea Bajcsy, Jaime F. Fisac

An outstanding challenge for the widespread deployment of robotic systems like autonomous vehicles is ensuring safe interaction with humans without sacrificing performance.

Autonomous Vehicles Reinforcement Learning (RL)

Active Uncertainty Reduction for Safe and Efficient Interaction Planning: A Shielding-Aware Dual Control Approach

1 code implementation1 Feb 2023 Haimin Hu, David Isele, Sangjae Bae, Jaime F. Fisac

To ensure the safe operation of the interacting agents, we use a runtime safety filter (also referred to as a "shielding" scheme), which overrides the robot's dual control policy with a safety fallback strategy when a safety-critical event is imminent.

Autonomous Vehicles Model Predictive Control +1

Active Uncertainty Reduction for Human-Robot Interaction: An Implicit Dual Control Approach

2 code implementations15 Feb 2022 Haimin Hu, Jaime F. Fisac

The ability to accurately predict human behavior is central to the safety and efficiency of robot autonomy in interactive settings.

Model Predictive Control Motion Planning

Sim-to-Lab-to-Real: Safe Reinforcement Learning with Shielding and Generalization Guarantees

no code implementations20 Jan 2022 Kai-Chieh Hsu, Allen Z. Ren, Duy Phuong Nguyen, Anirudha Majumdar, Jaime F. Fisac

To improve safety, we apply a dual policy setup where a performance policy is trained using the cumulative task reward and a backup (safety) policy is trained by solving the Safety Bellman Equation based on Hamilton-Jacobi (HJ) reachability analysis.

reinforcement-learning Reinforcement Learning (RL) +1

Safety and Liveness Guarantees through Reach-Avoid Reinforcement Learning

1 code implementation23 Dec 2021 Kai-Chieh Hsu, Vicenç Rubies-Royo, Claire J. Tomlin, Jaime F. Fisac

Recent successes in reinforcement learning methods to approximately solve optimal control problems with performance objectives make their application to certification problems attractive; however, the Lagrange-type objective used in reinforcement learning is not suitable to encode temporal logic requirements.

Q-Learning reinforcement-learning +1

ProBF: Learning Probabilistic Safety Certificates with Barrier Functions

1 code implementation22 Dec 2021 Athindran Ramesh Kumar, Sulin Liu, Jaime F. Fisac, Ryan P. Adams, Peter J. Ramadge

In practice, we have inaccurate knowledge of the system dynamics, which can lead to unsafe behaviors due to unmodeled residual dynamics.

SHARP: Shielding-Aware Robust Planning for Safe and Efficient Human-Robot Interaction

1 code implementation2 Oct 2021 Haimin Hu, Kensuke Nakamura, Jaime F. Fisac

Leveraging recent work on Bayesian human motion prediction, the resulting robot policy proactively balances nominal performance with the risk of high-cost emergency maneuvers triggered by low-probability human behaviors.

Human motion prediction Motion Planning +1

Back to the Future: Efficient, Time-Consistent Solutions in Reach-Avoid Games

1 code implementation16 Sep 2021 Dennis R. Anthony, Duy P. Nguyen, David Fridovich-Keil, Jaime F. Fisac

We study the class of reach-avoid dynamic games in which multiple agents interact noncooperatively, and each wishes to satisfy a distinct target criterion while avoiding a failure criterion.

Motion Planning

Quantifying Hypothesis Space Misspecification in Learning from Human-Robot Demonstrations and Physical Corrections

no code implementations3 Feb 2020 Andreea Bobu, Andrea Bajcsy, Jaime F. Fisac, Sampada Deglurkar, Anca D. Dragan

Recent work focuses on how robots can use such input - like demonstrations or corrections - to learn intended objectives.

LESS is More: Rethinking Probabilistic Models of Human Behavior

no code implementations13 Jan 2020 Andreea Bobu, Dexter R. R. Scobee, Jaime F. Fisac, S. Shankar Sastry, Anca D. Dragan

A common model is the Boltzmann noisily-rational decision model, which assumes people approximately optimize a reward function and choose trajectories in proportion to their exponentiated reward.

Econometrics

Safely Probabilistically Complete Real-Time Planning and Exploration in Unknown Environments

no code implementations19 Nov 2018 David Fridovich-Keil, Jaime F. Fisac, Claire J. Tomlin

We present a new framework for motion planning that wraps around existing kinodynamic planners and guarantees recursive feasibility when operating in a priori unknown, static environments.

Robotics Systems and Control

Hierarchical Game-Theoretic Planning for Autonomous Vehicles

no code implementations13 Oct 2018 Jaime F. Fisac, Eli Bronstein, Elis Stefansson, Dorsa Sadigh, S. Shankar Sastry, Anca D. Dragan

This mutual dependence, best captured by dynamic game theory, creates a strong coupling between the vehicle's planning and its predictions of other drivers' behavior, and constitutes an open problem with direct implications on the safety and viability of autonomous driving technology.

Autonomous Driving Decision Making +1

Learning under Misspecified Objective Spaces

1 code implementation11 Oct 2018 Andreea Bobu, Andrea Bajcsy, Jaime F. Fisac, Anca D. Dragan

Learning robot objective functions from human input has become increasingly important, but state-of-the-art techniques assume that the human's desired objective lies within the robot's hypothesis space.

An Efficient, Generalized Bellman Update For Cooperative Inverse Reinforcement Learning

no code implementations ICML 2018 Dhruv Malik, Malayandi Palaniappan, Jaime F. Fisac, Dylan Hadfield-Menell, Stuart Russell, Anca D. Dragan

We apply this update to a variety of POMDP solvers and find that it enables us to scale CIRL to non-trivial problems, with larger reward parameter spaces, and larger action spaces for both robot and human.

reinforcement-learning Reinforcement Learning (RL)

Generating Plans that Predict Themselves

no code implementations14 Feb 2018 Jaime F. Fisac, Chang Liu, Jessica B. Hamrick, S. Shankar Sastry, J. Karl Hedrick, Thomas L. Griffiths, Anca D. Dragan

We introduce $t$-\ACty{}: a measure that quantifies the accuracy and confidence with which human observers can predict the remaining robot plan from the overall task goal and the observed initial $t$ actions in the plan.

Planning, Fast and Slow: A Framework for Adaptive Real-Time Safe Trajectory Planning

2 code implementations12 Oct 2017 David Fridovich-Keil, Sylvia L. Herbert, Jaime F. Fisac, Sampada Deglurkar, Claire J. Tomlin

Motion planning is an extremely well-studied problem in the robotics community, yet existing work largely falls into one of two categories: computationally efficient but with few if any safety guarantees, or able to give stronger guarantees but at high computational cost.

Systems and Control Computer Science and Game Theory

Pragmatic-Pedagogic Value Alignment

no code implementations20 Jul 2017 Jaime F. Fisac, Monica A. Gates, Jessica B. Hamrick, Chang Liu, Dylan Hadfield-Menell, Malayandi Palaniappan, Dhruv Malik, S. Shankar Sastry, Thomas L. Griffiths, Anca D. Dragan

In robotics, value alignment is key to the design of collaborative robots that can integrate into human workflows, successfully inferring and adapting to their users' objectives as they go.

Decision Making

FaSTrack: a Modular Framework for Fast and Guaranteed Safe Motion Planning

no code implementations21 Mar 2017 Sylvia L. Herbert, Mo Chen, SooJean Han, Somil Bansal, Jaime F. Fisac, Claire J. Tomlin

We propose a new algorithm FaSTrack: Fast and Safe Tracking for High Dimensional systems.

Robotics

Cannot find the paper you are looking for? You can Submit a new open access paper.