no code implementations • 19 Mar 2024 • Xun Shen, Ye Wang, Kazumune Hashimoto, Yuhu Wu, Sebastien Gros
The existing methods of computing probabilistic reachable sets normally assume that the uncertainties are independent of the state.
no code implementations • 1 Jan 2024 • Sangjun Bae, Balazs Kulcsar, Sebastien Gros
In the environment, we propose a Q-learning-based PeDP to maximize fast-EVCS' revenue.
no code implementations • 20 Mar 2023 • Anilkumar Parsi, Marcell Bartos, Amber Srivastava, Sebastien Gros, Roy S. Smith
A novel perspective on the design of robust model predictive control (MPC) methods is presented, whereby closed-loop constraint satisfaction is ensured using recursive feasibility of the MPC optimization.
no code implementations • 24 Feb 2023 • Erlend Torje Berg Lundby, Adil Rasheed, Ivar Johan Halvorsen, Dirk Reinhardt, Sebastien Gros, Jan Tommy Gravdahl
This simulated dataset can be used in a static deep active learning acquisition scheme referred to as global explorations.
no code implementations • 4 Jan 2023 • Shambhuraj Sawant, Akhil S Anand, Dirk Reinhardt, Sebastien Gros
The state-of-the-art learning methods use RL to improve the performance of parameterized MPC schemes.
no code implementations • 9 Oct 2022 • Arash Bahari Kordabad, Mario Zanon, Sebastien Gros
This paper shows that the optimal policy and value functions of a Markov Decision Process (MDP), either discounted or not, can be captured by a finite-horizon undiscounted Optimal Control Problem (OCP), even if based on an inexact model.
no code implementations • 10 Jun 2022 • Hossein Nejatbakhsh Esfahani, Sebastien Gros
In this paper, we propose a learning-based Model Predictive Control (MPC) approach for the polytopic Linear Parameter-Varying (LPV) systems with inexact scheduling parameters (as exogenous signals with inexact bounds), where the Linear Time Invariant (LTI) models (vertices) captured by combinations of the scheduling parameters becomes wrong.
no code implementations • 18 May 2022 • Shambhuraj Sawant, Sebastien Gros
We propose simple tools to promote structures in the QP, pushing it to resemble a linear MPC scheme.
no code implementations • 26 Apr 2022 • Huang Zhang, Yang Su, Faisal Altaf, Torsten Wik, Sebastien Gros
For that reason, various data-driven methods have been proposed for point prediction of battery cycle life with minimum knowledge of the battery degradation mechanisms.
no code implementations • 31 Mar 2022 • Arash Bahari Kordabad, Sebastien Gros
This paper discusses the functional stability of closed-loop Markov Chains under optimal policies resulting from a discounted optimality criterion, forming Markov Decision Processes (MDPs).
no code implementations • 25 Mar 2022 • Arash Bahari Kordabad, Hossein Nejatbakhsh Esfahani, WenQi Cai, Sebastien Gros
We show that the approximate Hessian converges to the exact Hessian at the optimal policy, and allows for a superlinear convergence in the learning, provided that the policy parametrization is rich.
no code implementations • 19 Nov 2021 • Robert Hult, Mario Zanon, Sebastien Gros, Paolo Falcone
In this paper, we consider the optimal coordination of automated vehicles at intersections under fixed crossing orders.
no code implementations • 19 Nov 2021 • Hossein Nejatbakhsh Esfahani, Behdad Aminian, Esten Ingar Grøtli, Sebastien Gros
The aim of this paper is to propose a high performance control approach for trajectory tracking of Autonomous Underwater Vehicles (AUVs).
no code implementations • 7 Nov 2021 • Eivind Bøhn, Sebastien Gros, Signe Moe, Tor Arne Johansen
Its high computational complexity results in high power consumption from the control algorithm, which could account for a significant share of the energy resources in battery-powered embedded systems.
no code implementations • 16 Jun 2021 • WenQi Cai, Arash B. Kordabad, Hossein N. Esfahani, Anastasios M. Lekkas, Sebastien Gros
In this work, we propose a Model Predictive Control (MPC)-based Reinforcement Learning (RL) method for Autonomous Surface Vehicles (ASVs).
no code implementations • 24 May 2021 • Arash Bahari Kordabad, Sebastien Gros
In the Economic Nonlinear Model Predictive (ENMPC) context, closed-loop stability relates to the existence of a storage function satisfying a dissipation inequality.
no code implementations • 6 Apr 2021 • Hossein Nejatbakhsh Esfahani, Arash Bahari Kordabad, Sebastien Gros
We present a Reinforcement Learning-based Robust Nonlinear Model Predictive Control (RL-RNMPC) framework for controlling nonlinear systems in the presence of disturbances and uncertainties.
no code implementations • 6 Apr 2021 • Arash Bahari Kordabad, Hossein Nejatbakhsh Esfahani, Sebastien Gros
In this paper, we discuss the deterministic policy gradient using the Actor-Critic methods based on the linear compatible advantage function approximator, where the input spaces are continuous.
no code implementations • 6 Apr 2021 • Arash Bahari Kordabad, WenQi Cai, Sebastien Gros
In this paper, we are interested in optimal control problems with purely economic costs, which often yield optimal policies having a (nearly) bang-bang structure.
no code implementations • 22 Mar 2021 • Hossein Nejatbakhsh Esfahani, Arash Bahari Kordabad, Sebastien Gros
This paper proposes an observer-based framework for solving Partially Observable Markov Decision Processes (POMDPs) when an accurate model is not available.
no code implementations • 22 Feb 2021 • Eivind Bøhn, Sebastien Gros, Signe Moe, Tor Arne Johansen
Model predictive control (MPC) is a powerful trajectory optimization control technique capable of controlling complex nonlinear systems while respecting system constraints and ensuring safe operation.
1 code implementation • 26 Nov 2020 • Eivind Bøhn, Sebastien Gros, Signe Moe, Tor Arne Johansen
In control applications there is often a compromise that needs to be made with regards to the complexity and performance of the controller and the computational resources that are available.
no code implementations • 3 Apr 2020 • Sebastien Gros, Mario Zanon
Model Predictive Control has been recently proposed as policy approximation for Reinforcement Learning, offering a path towards safe and explainable Reinforcement Learning.
no code implementations • 2 Apr 2020 • Sebastien Gros, Mario Zanon, Alberto Bemporad
For all its successes, Reinforcement Learning (RL) still struggles to deliver formal guarantees on the closed-loop behavior of the learned policy.