Search Results for author: David Fridovich-Keil

Found 24 papers, 7 papers with code

Decomposing Control Lyapunov Functions for Efficient Reinforcement Learning

1 code implementation18 Mar 2024 Antonio Lopez, David Fridovich-Keil

In this paper, we build from existing work that reshapes the reward function in RL by introducing a Control Lyapunov Function (CLF), which is demonstrated to reduce the sample complexity.

reinforcement-learning Reinforcement Learning (RL)

Auto-Encoding Bayesian Inverse Games

no code implementations14 Feb 2024 Xinjie Liu, Lasse Peters, Javier Alonso-Mora, Ufuk Topcu, David Fridovich-Keil

When multiple agents interact in a common environment, each agent's actions impact others' future decisions, and noncooperative dynamic games naturally capture this coupling.

Motion Planning

An Investigation of Time Reversal Symmetry in Reinforcement Learning

no code implementations28 Nov 2023 Brett Barkley, Amy Zhang, David Fridovich-Keil

We observe that utilizing the structure of time reversal in an MDP allows every environment transition experienced by an agent to be transformed into a feasible reverse-time transition, effectively doubling the number of experiences in the environment.

Data Augmentation Friction +2

Encouraging Inferable Behavior for Autonomy: Repeated Bimatrix Stackelberg Games with Observations

no code implementations30 Sep 2023 Mustafa O. Karabag, Sophia Smith, David Fridovich-Keil, Ufuk Topcu

As a converse result, we also provide a game where the required number of interactions is lower bounded by a function of the desired inferability loss.

Decision Making

Symbolic Regression on Sparse and Noisy Data with Gaussian Processes

no code implementations20 Sep 2023 Junette Hsin, Shubhankar Agarwal, Adam Thorpe, Luis Sentis, David Fridovich-Keil

To overcome this, we combine Gaussian process regression with a sparse identification of nonlinear dynamics (SINDy) method to denoise the data and identify nonlinear dynamical equations.

Gaussian Processes regression +1

Active Inverse Learning in Stackelberg Trajectory Games

no code implementations15 Aug 2023 Yue Yu, Jacob Levy, Negar Mehr, David Fridovich-Keil, Ufuk Topcu

We formulate an inverse learning problem in a Stackelberg game between a leader and a follower, where each player's action is the trajectory of a dynamical system.

Enabling Efficient, Reliable Real-World Reinforcement Learning with Approximate Physics-Based Models

1 code implementation16 Jul 2023 Tyler Westenbroek, Jacob Levy, David Fridovich-Keil

We focus on developing efficient and reliable policy optimization strategies for robot learning with real-world data.

Policy Gradient Methods

Scenario-Game ADMM: A Parallelized Scenario-Based Solver for Stochastic Noncooperative Games

no code implementations4 Apr 2023 Jingqi Li, Chih-Yuan Chiu, Lasse Peters, Fernando Palafox, Mustafa Karabag, Javier Alonso-Mora, Somayeh Sojoudi, Claire Tomlin, David Fridovich-Keil

To accommodate this, we decompose the approximated game into a set of smaller games with few constraints for each sampled scenario, and propose a decentralized, consensus-based ADMM algorithm to efficiently compute a generalized Nash equilibrium (GNE) of the approximated game.

Decision Making

GrAVITree: Graph-based Approximate Value Function In a Tree

no code implementations18 Jan 2023 Patrick H. Washington, David Fridovich-Keil, Mac Schwager

In this paper, we introduce GrAVITree, a tree- and sampling-based algorithm to compute a near-optimal value function and corresponding feedback policy for indefinite time-horizon, terminal state-constrained nonlinear optimal control problems.

Robust Forecasting for Robotic Control: A Game-Theoretic Approach

no code implementations22 Sep 2022 Shubhankar Agarwal, David Fridovich-Keil, Sandeep P. Chinchali

In order to model real-world factors affecting future forecasts, we introduce the notion of an adversary, which perturbs observed historical time series to increase a robot's ultimate control cost.

Self-Driving Cars Time Series +1

Back to the Future: Efficient, Time-Consistent Solutions in Reach-Avoid Games

1 code implementation16 Sep 2021 Dennis R. Anthony, Duy P. Nguyen, David Fridovich-Keil, Jaime F. Fisac

We study the class of reach-avoid dynamic games in which multiple agents interact noncooperatively, and each wishes to satisfy a distinct target criterion while avoiding a failure criterion.

Motion Planning

Approximate Solutions to a Class of Reachability Games

no code implementations1 Nov 2020 David Fridovich-Keil, Claire J. Tomlin

In this paper, we present a method for finding approximate Nash equilibria in a broad class of reachability games.

Collision Avoidance

Technical Report: Adaptive Control for Linearizable Systems Using On-Policy Reinforcement Learning

no code implementations6 Apr 2020 Tyler Westenbroek, Eric Mazumdar, David Fridovich-Keil, Valmik Prabhu, Claire J. Tomlin, S. Shankar Sastry

This paper proposes a framework for adaptively learning a feedback linearization-based tracking controller for an unknown system using discrete-time model-free policy-gradient parameter update rules.

reinforcement-learning Reinforcement Learning (RL)

Feedback Linearization for Unknown Systems via Reinforcement Learning

no code implementations29 Oct 2019 Tyler Westenbroek, David Fridovich-Keil, Eric Mazumdar, Shreyas Arora, Valmik Prabhu, S. Shankar Sastry, Claire J. Tomlin

We present a novel approach to control design for nonlinear systems which leverages model-free policy optimization techniques to learn a linearizing controller for a physical plant with unknown dynamics.

reinforcement-learning Reinforcement Learning (RL)

An Iterative Quadratic Method for General-Sum Differential Games with Feedback Linearizable Dynamics

1 code implementation1 Oct 2019 David Fridovich-Keil, Vicenc Rubies-Royo, Claire J. Tomlin

Iterative linear-quadratic (ILQ) methods are widely used in the nonlinear optimal control community.

Systems and Control Computer Science and Game Theory Multiagent Systems Robotics Systems and Control

Efficient Iterative Linear-Quadratic Approximations for Nonlinear Multi-Player General-Sum Differential Games

1 code implementation10 Sep 2019 David Fridovich-Keil, Ellis Ratner, Anca D. Dragan, Claire J. Tomlin

We benchmark our method in a three-player general-sum simulated example, in which it takes < 0. 75 s to identify a solution and < 50 ms to solve warm-started subproblems in a receding horizon.

Systems and Control Robotics Systems and Control

Safely Probabilistically Complete Real-Time Planning and Exploration in Unknown Environments

no code implementations19 Nov 2018 David Fridovich-Keil, Jaime F. Fisac, Claire J. Tomlin

We present a new framework for motion planning that wraps around existing kinodynamic planners and guarantees recursive feasibility when operating in a priori unknown, static environments.

Robotics Systems and Control

A Successive-Elimination Approach to Adaptive Robotic Sensing

no code implementations27 Sep 2018 Esther Rolf, David Fridovich-Keil, Max Simchowitz, Benjamin Recht, Claire Tomlin

We study an adaptive source seeking problem, in which a mobile robot must identify the strongest emitter(s) of a signal in an environment with background emissions.

Trajectory Planning

Towards Distributed Energy Services: Decentralizing Optimal Power Flow with Machine Learning

no code implementations14 Jun 2018 Roel Dobbe, Oscar Sondermeijer, David Fridovich-Keil, Daniel Arnold, Duncan Callaway, Claire Tomlin

We consider distribution systems with multiple controllable Distributed Energy Resources (DERs) and present a data-driven approach to learn control policies for each DER to reconstruct and mimic the solution to a centralized OPF problem from solely locally available information.

BIG-bench Machine Learning

Planning, Fast and Slow: A Framework for Adaptive Real-Time Safe Trajectory Planning

2 code implementations12 Oct 2017 David Fridovich-Keil, Sylvia L. Herbert, Jaime F. Fisac, Sampada Deglurkar, Claire J. Tomlin

Motion planning is an extremely well-studied problem in the robotics community, yet existing work largely falls into one of two categories: computationally efficient but with few if any safety guarantees, or able to give stronger guarantees but at high computational cost.

Systems and Control Computer Science and Game Theory

Fully Decentralized Policies for Multi-Agent Systems: An Information Theoretic Approach

no code implementations NeurIPS 2017 Roel Dobbe, David Fridovich-Keil, Claire Tomlin

Learning cooperative policies for multi-agent systems is often challenged by partial observability and a lack of coordination.

Cannot find the paper you are looking for? You can Submit a new open access paper.