Search Results for author: Santiago Paternain

Found 19 papers, 2 papers with code

Adaptive Primal-Dual Method for Safe Reinforcement Learning

no code implementations1 Feb 2024 Weiqin Chen, James Onyejizu, Long Vu, Lan Hoang, Dharmashankar Subramanian, Koushik Kar, Sandipan Mishra, Santiago Paternain

In this paper, we propose, analyze and evaluate adaptive primal-dual (APD) methods for SRL, where two adaptive LRs are adjusted to the Lagrangian multipliers so as to optimize the policy in each iteration.

reinforcement-learning Safe Reinforcement Learning

Learning Non-myopic Power Allocation in Constrained Scenarios

1 code implementation18 Jan 2024 Arindam Chowdhury, Santiago Paternain, Gunjan Verma, Ananthram Swami, Santiago Segarra

The problem of optimal power allocation -- for maximizing a given network utility metric -- under instantaneous constraints has recently gained significant popularity.

Decision Making

Blackout Mitigation via Physics-guided RL

no code implementations17 Jan 2024 Anmol Dwivedi, Santiago Paternain, Ali Tajer

This paper considers the sequential design of remedial control actions in response to system anomalies for the ultimate objective of preventing blackouts.

Reinforcement Learning (RL)

Constrained Learning with Non-Convex Losses

no code implementations8 Mar 2021 Luiz F. O. Chamon, Santiago Paternain, Miguel Calvo-Fullana, Alejandro Ribeiro

In this paper, we overcome this issue by learning in the empirical dual domain, where constrained statistical learning problems become unconstrained and deterministic.

Adversarial Robustness Fairness +1

Towards Safe Continuing Task Reinforcement Learning

no code implementations24 Feb 2021 Miguel Calvo-Fullana, Luiz F. O. Chamon, Santiago Paternain

However, to transfer from learning safety to learning safely, there are two hurdles that need to be overcome: (i) it has to be possible to learn the policy without having to re-initialize the system; and (ii) the rollouts of the system need to be in themselves safe.

reinforcement-learning Reinforcement Learning (RL) +1

State Augmented Constrained Reinforcement Learning: Overcoming the Limitations of Learning with Rewards

no code implementations23 Feb 2021 Miguel Calvo-Fullana, Santiago Paternain, Luiz F. O. Chamon, Alejandro Ribeiro

Thus, as we illustrate by an example, while previous methods can fail at finding optimal policies, running the dual dynamics while executing the augmented policy yields an algorithm that provably samples actions from the optimal policy.

reinforcement-learning Reinforcement Learning (RL)

Sufficiently Accurate Model Learning for Planning

no code implementations11 Feb 2021 Clark Zhang, Santiago Paternain, Alejandro Ribeiro

This paper introduces the constrained Sufficiently Accurate model learning approach, provides examples of such problems, and presents a theorem on how close some approximate solutions can be.

Trust but Verify: Assigning Prediction Credibility by Counterfactual Constrained Learning

no code implementations24 Nov 2020 Luiz F. O. Chamon, Santiago Paternain, Alejandro Ribeiro

Prediction credibility measures, in the form of confidence intervals or probability distributions, are fundamental in statistics and machine learning to characterize model robustness, detect out-of-distribution samples (outliers), and protect against adversarial attacks.

counterfactual

Policy Gradient for Continuing Tasks in Non-stationary Markov Decision Processes

no code implementations16 Oct 2020 Santiago Paternain, Juan Andres Bazerque, Alejandro Ribeiro

To that end we compute unbiased stochastic gradients of the value function which we use as ascent directions to update the policy.

Navigate

Counterfactual Programming for Optimal Control

no code implementations L4DC 2020 Luiz F.O. Chamon, Santiago Paternain, Alejandro Ribeiro

In recent years, considerable work has been done to tackle the issue of designing control laws based on observations to allow unknown dynamical systems to perform pre-specified tasks.

counterfactual

The empirical duality gap of constrained statistical learning

no code implementations12 Feb 2020 Luiz. F. O. Chamon, Santiago Paternain, Miguel Calvo-Fullana, Alejandro Ribeiro

This paper is concerned with the study of constrained statistical learning problems, the unconstrained version of which are at the core of virtually all of modern information processing.

Constrained Reinforcement Learning Has Zero Duality Gap

no code implementations NeurIPS 2019 Santiago Paternain, Luiz. F. O. Chamon, Miguel Calvo-Fullana, Alejandro Ribeiro

The later is generally addressed by formulating the conflicting requirements as a constrained RL problem and solved using Primal-Dual methods.

reinforcement-learning Reinforcement Learning (RL)

Sparse multiresolution representations with adaptive kernels

no code implementations7 May 2019 Maria Peifer, Luiz. F. O. Chamon, Santiago Paternain, Alejandro Ribeiro

To address the complexity issues, we then write the function estimation problem as a sparse functional program that explicitly minimizes the support of the representation leading to low complexity solutions.

Sufficiently Accurate Model Learning

no code implementations19 Feb 2019 Clark Zhang, Arbaaz Khan, Santiago Paternain, Alejandro Ribeiro

In this paper, we investigate a method to regularize model learning techniques to provide better error characteristics for traditional control and planning algorithms.

Decentralized Online Learning with Kernels

no code implementations11 Oct 2017 Alec Koppel, Santiago Paternain, Cedric Richard, Alejandro Ribeiro

That is, we establish that with constant step-size selections agents' functions converge to a neighborhood of the globally optimal one while satisfying the consensus constraints as the penalty parameter is increased.

General Classification Multi-class Classification +2

Cannot find the paper you are looking for? You can Submit a new open access paper.