1 code implementation • 18 Mar 2024 • Antonio Lopez, David Fridovich-Keil
In this paper, we build from existing work that reshapes the reward function in RL by introducing a Control Lyapunov Function (CLF), which is demonstrated to reduce the sample complexity.
no code implementations • 14 Feb 2024 • Xinjie Liu, Lasse Peters, Javier Alonso-Mora, Ufuk Topcu, David Fridovich-Keil
When multiple agents interact in a common environment, each agent's actions impact others' future decisions, and noncooperative dynamic games naturally capture this coupling.
no code implementations • 28 Nov 2023 • Brett Barkley, Amy Zhang, David Fridovich-Keil
We observe that utilizing the structure of time reversal in an MDP allows every environment transition experienced by an agent to be transformed into a feasible reverse-time transition, effectively doubling the number of experiences in the environment.
no code implementations • 30 Sep 2023 • Mustafa O. Karabag, Sophia Smith, David Fridovich-Keil, Ufuk Topcu
As a converse result, we also provide a game where the required number of interactions is lower bounded by a function of the desired inferability loss.
no code implementations • 20 Sep 2023 • Junette Hsin, Shubhankar Agarwal, Adam Thorpe, Luis Sentis, David Fridovich-Keil
To overcome this, we combine Gaussian process regression with a sparse identification of nonlinear dynamics (SINDy) method to denoise the data and identify nonlinear dynamical equations.
no code implementations • 14 Sep 2023 • Jiankai Sun, Shreyas Kousik, David Fridovich-Keil, Mac Schwager
Connected autonomous vehicles (CAVs) promise to enhance safety, efficiency, and sustainability in urban transportation.
no code implementations • 15 Aug 2023 • Yue Yu, Jacob Levy, Negar Mehr, David Fridovich-Keil, Ufuk Topcu
We formulate an inverse learning problem in a Stackelberg game between a leader and a follower, where each player's action is the trajectory of a dynamical system.
1 code implementation • 16 Jul 2023 • Tyler Westenbroek, Jacob Levy, David Fridovich-Keil
We focus on developing efficient and reliable policy optimization strategies for robot learning with real-world data.
no code implementations • 4 Apr 2023 • Jingqi Li, Chih-Yuan Chiu, Lasse Peters, Fernando Palafox, Mustafa Karabag, Javier Alonso-Mora, Somayeh Sojoudi, Claire Tomlin, David Fridovich-Keil
To accommodate this, we decompose the approximated game into a set of smaller games with few constraints for each sampled scenario, and propose a decentralized, consensus-based ADMM algorithm to efficiently compute a generalized Nash equilibrium (GNE) of the approximated game.
1 code implementation • 31 Mar 2023 • Shenghui Chen, Yue Yu, David Fridovich-Keil, Ufuk Topcu
Markov games model interactions among multiple players in a stochastic, dynamic environment.
no code implementations • 18 Jan 2023 • Patrick H. Washington, David Fridovich-Keil, Mac Schwager
In this paper, we introduce GrAVITree, a tree- and sampling-based algorithm to compute a near-optimal value function and corresponding feedback policy for indefinite time-horizon, terminal state-constrained nonlinear optimal control problems.
no code implementations • 22 Sep 2022 • Shubhankar Agarwal, David Fridovich-Keil, Sandeep P. Chinchali
In order to model real-world factors affecting future forecasts, we introduce the notion of an adversary, which perturbs observed historical time series to increase a robot's ultimate control cost.
1 code implementation • 16 Sep 2021 • Dennis R. Anthony, Duy P. Nguyen, David Fridovich-Keil, Jaime F. Fisac
We study the class of reach-avoid dynamic games in which multiple agents interact noncooperatively, and each wishes to satisfy a distinct target criterion while avoiding a failure criterion.
no code implementations • 1 Nov 2020 • David Fridovich-Keil, Claire J. Tomlin
In this paper, we present a method for finding approximate Nash equilibria in a broad class of reachability games.
no code implementations • 6 Apr 2020 • Tyler Westenbroek, Eric Mazumdar, David Fridovich-Keil, Valmik Prabhu, Claire J. Tomlin, S. Shankar Sastry
This paper proposes a framework for adaptively learning a feedback linearization-based tracking controller for an unknown system using discrete-time model-free policy-gradient parameter update rules.
no code implementations • 29 Oct 2019 • Tyler Westenbroek, David Fridovich-Keil, Eric Mazumdar, Shreyas Arora, Valmik Prabhu, S. Shankar Sastry, Claire J. Tomlin
We present a novel approach to control design for nonlinear systems which leverages model-free policy optimization techniques to learn a linearizing controller for a physical plant with unknown dynamics.
1 code implementation • 1 Oct 2019 • David Fridovich-Keil, Vicenc Rubies-Royo, Claire J. Tomlin
Iterative linear-quadratic (ILQ) methods are widely used in the nonlinear optimal control community.
Systems and Control Computer Science and Game Theory Multiagent Systems Robotics Systems and Control
1 code implementation • 10 Sep 2019 • David Fridovich-Keil, Ellis Ratner, Anca D. Dragan, Claire J. Tomlin
We benchmark our method in a three-player general-sum simulated example, in which it takes < 0. 75 s to identify a solution and < 50 ms to solve warm-started subproblems in a receding horizon.
Systems and Control Robotics Systems and Control
no code implementations • 19 Nov 2018 • David Fridovich-Keil, Jaime F. Fisac, Claire J. Tomlin
We present a new framework for motion planning that wraps around existing kinodynamic planners and guarantees recursive feasibility when operating in a priori unknown, static environments.
Robotics Systems and Control
no code implementations • 27 Sep 2018 • Esther Rolf, David Fridovich-Keil, Max Simchowitz, Benjamin Recht, Claire Tomlin
We study an adaptive source seeking problem, in which a mobile robot must identify the strongest emitter(s) of a signal in an environment with background emissions.
no code implementations • 14 Jun 2018 • Roel Dobbe, Oscar Sondermeijer, David Fridovich-Keil, Daniel Arnold, Duncan Callaway, Claire Tomlin
We consider distribution systems with multiple controllable Distributed Energy Resources (DERs) and present a data-driven approach to learn control policies for each DER to reconstruct and mimic the solution to a centralized OPF problem from solely locally available information.
no code implementations • 31 May 2018 • Jaime F. Fisac, Andrea Bajcsy, Sylvia L. Herbert, David Fridovich-Keil, Steven Wang, Claire J. Tomlin, Anca D. Dragan
In order to safely operate around humans, robots can employ predictive models of human motion.
2 code implementations • 12 Oct 2017 • David Fridovich-Keil, Sylvia L. Herbert, Jaime F. Fisac, Sampada Deglurkar, Claire J. Tomlin
Motion planning is an extremely well-studied problem in the robotics community, yet existing work largely falls into one of two categories: computationally efficient but with few if any safety guarantees, or able to give stronger guarantees but at high computational cost.
Systems and Control Computer Science and Game Theory
no code implementations • NeurIPS 2017 • Roel Dobbe, David Fridovich-Keil, Claire Tomlin
Learning cooperative policies for multi-agent systems is often challenged by partial observability and a lack of coordination.