Search Results for author: Javad Lavaei

Found 23 papers, 3 papers with code

Exact Recovery Guarantees for Parameterized Non-linear System Identification Problem under Adversarial Attacks

no code implementations30 Aug 2024 Haixiang Zhang, Baturalp Yalcin, Javad Lavaei, Eduardo D. Sontag

In this work, we study the system identification problem for parameterized non-linear systems using basis functions under adversarial attacks.

A CMDP-within-online framework for Meta-Safe Reinforcement Learning

no code implementations26 May 2024 Vanshaj Khattar, Yuhao Ding, Bilgehan Sel, Javad Lavaei, Ming Jin

Meta-reinforcement learning has widely been used as a learning-to-learn framework to solve unseen tasks with limited experience.

Meta-Learning Meta Reinforcement Learning +3

Absence of spurious solutions far from ground truth: A low-rank analysis with high-order losses

no code implementations10 Mar 2024 Ziye Ma, Ying Chen, Javad Lavaei, Somayeh Sojoudi

Matrix sensing problems exhibit pervasive non-convexity, plaguing optimization with a proliferation of suboptimal spurious solutions.

Exact Recovery for System Identification with More Corrupt Data than Clean Data

no code implementations17 May 2023 Baturalp Yalcin, Haixiang Zhang, Javad Lavaei, Murat Arcak

We highlight that the attack vectors are allowed to be correlated with each other in this work, whereas we make some assumptions about the times at which the attacks happen.

Scalable Multi-Agent Reinforcement Learning with General Utilities

no code implementations15 Feb 2023 Donghao Ying, Yuhao Ding, Alec Koppel, Javad Lavaei

The objective is to find a localized policy that maximizes the average of the team's local utility functions without the full observability of each agent in the team.

Multi-agent Reinforcement Learning reinforcement-learning +2

Non-stationary Risk-sensitive Reinforcement Learning: Near-optimal Dynamic Regret, Adaptive Detection, and Separation Design

no code implementations19 Nov 2022 Yuhao Ding, Ming Jin, Javad Lavaei

We study risk-sensitive reinforcement learning (RL) based on an entropic risk measure in episodic non-stationary Markov decision processes (MDPs).

Reinforcement Learning (RL)

Learning of Dynamical Systems under Adversarial Attacks -- Null Space Property Perspective

no code implementations4 Oct 2022 Han Feng, Baturalp Yalcin, Javad Lavaei

We study the identification of a linear time-invariant dynamical system affected by large-and-sparse disturbances modeling adversarial attacks or faults.

Policy-based Primal-Dual Methods for Concave CMDP with Variance Reduction

1 code implementation22 May 2022 Donghao Ying, Mengzi Amy Guo, Hyunin Lee, Yuhao Ding, Javad Lavaei, Zuo-Jun Max Shen

In the exact setting, we prove an $O(T^{-1/3})$ convergence rate for both the average optimality gap and constraint violation, which further improves to $O(T^{-1/2})$ under strong concavity of the objective in the occupancy measure.

Provably Efficient Primal-Dual Reinforcement Learning for CMDPs with Non-stationary Objectives and Constraints

no code implementations28 Jan 2022 Yuhao Ding, Javad Lavaei

We consider primal-dual-based reinforcement learning (RL) in episodic constrained Markov decision processes (CMDPs) with non-stationary objectives and constraints, which plays a central role in ensuring the safety of RL in time-varying environments.

Reinforcement Learning (RL) Safe Exploration

Stochastic $L^\natural$-convex Function Minimization

no code implementations NeurIPS 2021 Haixiang Zhang, Zeyu Zheng, Javad Lavaei

When applied to a stochastic submodular function, the computational complexity of the proposed algorithms is lower than that of the existing stochastic submodular minimization algorithms.

On the Global Optimum Convergence of Momentum-based Policy Gradient

no code implementations19 Oct 2021 Yuhao Ding, Junzi Zhang, Javad Lavaei

For the generic Fisher-non-degenerate policy parametrizations, our result is the first single-loop and finite-batch PG algorithm achieving $\tilde{O}(\epsilon^{-3})$ global optimality sample complexity.

Factorization Approach for Low-complexity Matrix Completion Problems: Exponential Number of Spurious Solutions and Failure of Gradient Methods

no code implementations19 Oct 2021 Baturalp Yalcin, Haixiang Zhang, Javad Lavaei, Somayeh Sojoudi

It is well-known that the Burer-Monteiro (B-M) factorization approach can efficiently solve low-rank matrix optimization problems under the RIP condition.

Matrix Completion

A Dual Approach to Constrained Markov Decision Processes with Entropy Regularization

no code implementations17 Oct 2021 Donghao Ying, Yuhao Ding, Javad Lavaei

We study entropy-regularized constrained Markov decision processes (CMDPs) under the soft-max parameterization, in which an agent aims to maximize the entropy-regularized value function while satisfying constraints on the expected total utility.

Sharp Restricted Isometry Property Bounds for Low-rank Matrix Recovery Problems with Corrupted Measurements

no code implementations18 May 2021 Ziye Ma, Yingjie Bi, Javad Lavaei, Somayeh Sojoudi

By analyzing the landscape of the non-convex problem, we first propose a global guarantee on the maximum distance between an arbitrary local minimizer and the ground truth under the assumption that the RIP constant is smaller than $1/2$.

Matrix Completion Retrieval

When Does MAML Objective Have Benign Landscape?

no code implementations31 May 2020 Igor Molybog, Javad Lavaei

The paper studies the complexity of the optimization problem behind the Model-Agnostic Meta-Learning (MAML) algorithm.

Decision Making Meta-Learning

Large-Scale Traffic Signal Offset Optimization

1 code implementation19 Nov 2019 Yi Ouyang, Richard Y. Zhang, Javad Lavaei, Pravin Varaiya

The offset optimization problem seeks to coordinate and synchronize the timing of traffic signals throughout a network in order to enhance traffic flow and reduce stops and delays.

Optimization and Control Systems and Control Systems and Control

Sharp Restricted Isometry Bounds for the Inexistence of Spurious Local Minima in Nonconvex Matrix Recovery

no code implementations7 Jan 2019 Richard Y. Zhang, Somayeh Sojoudi, Javad Lavaei

Using the technique, we prove that in the case of a rank-1 ground truth, an RIP constant of $\delta<1/2$ is both necessary and sufficient for exact recovery from any arbitrary initial point (such as a random point).

Stability-certified reinforcement learning: A control-theoretic perspective

no code implementations26 Oct 2018 Ming Jin, Javad Lavaei

We investigate the important problem of certifying stability of reinforcement learning policies when interconnected with nonlinear dynamical systems.

reinforcement-learning Reinforcement Learning +1

How Much Restricted Isometry is Needed In Nonconvex Matrix Recovery?

no code implementations NeurIPS 2018 Richard Y. Zhang, Cédric Josz, Somayeh Sojoudi, Javad Lavaei

When the linear measurements of an instance of low-rank matrix recovery satisfy a restricted isometry property (RIP)---i. e. they are approximately norm-preserving---the problem is known to contain no spurious local minima, so exact recovery is guaranteed.

Sparse Semidefinite Programs with Guaranteed Near-Linear Time Complexity via Dualized Clique Tree Conversion

1 code implementation10 Oct 2017 Richard Y. Zhang, Javad Lavaei

Clique tree conversion solves large-scale semidefinite programs by splitting an $n\times n$ matrix variable into up to $n$ smaller matrix variables, each representing a principal submatrix of up to $\omega\times\omega$.

Optimization and Control

Cannot find the paper you are looking for? You can Submit a new open access paper.