Search Results for author: Mahyar Fazlyab

Found 12 papers, 4 papers with code

Learning Region of Attraction for Nonlinear Systems

no code implementations2 Oct 2021 Shaoru Chen, Mahyar Fazlyab, Manfred Morari, George J. Pappas, Victor M. Preciado

Estimating the region of attraction (ROA) of general nonlinear autonomous systems remains a challenging problem and requires a case-by-case analysis.

DeepSplit: Scalable Verification of Deep Neural Networks via Operator Splitting

no code implementations16 Jun 2021 Shaoru Chen, Eric Wong, J. Zico Kolter, Mahyar Fazlyab

Analyzing the worst-case performance of deep neural networks against input perturbations amounts to solving a large-scale non-convex optimization problem, for which several past works have proposed convex relaxations as a promising alternative.

Image Classification reinforcement-learning

On Centralized and Distributed Mirror Descent: Convergence Analysis Using Quadratic Constraints

no code implementations29 May 2021 Youbang Sun, Mahyar Fazlyab, Shahin Shahrampour

Our numerical experiments on strongly convex problems indicate that our framework certifies superior convergence rates compared to the existing rates for distributed GD.

Performance Bounds for Neural Network Estimators: Applications in Fault Detection

no code implementations22 Mar 2021 Navid Hashemi, Mahyar Fazlyab, Justin Ruths

We exploit recent results in quantifying the robustness of neural networks to input variations to construct and tune a model-based anomaly detector, where the data-driven estimator model is provided by an autoregressive neural network.

Fault Detection

Learning Lyapunov Functions for Hybrid Systems

no code implementations22 Dec 2020 Shaoru Chen, Mahyar Fazlyab, Manfred Morari, George J. Pappas, Victor M. Preciado

By designing the learner and the verifier according to the analytic center cutting-plane method from convex optimization, we show that when the set of Lyapunov functions is full-dimensional in the parameter space, our method finds a Lyapunov function in a finite number of steps.

Optimization and Control

Certifying Incremental Quadratic Constraints for Neural Networks via Convex Optimization

no code implementations10 Dec 2020 Navid Hashemi, Justin Ruths, Mahyar Fazlyab

Abstracting neural networks with constraints they impose on their inputs and outputs can be very useful in the analysis of neural network classifiers and to derive optimization-based algorithms for certification of stability and robustness of feedback systems involving neural networks.

Enforcing robust control guarantees within neural network policies

1 code implementation ICLR 2021 Priya L. Donti, Melrose Roderick, Mahyar Fazlyab, J. Zico Kolter

When designing controllers for safety-critical systems, practitioners often face a challenging tradeoff between robustness and performance.

Robust Deep Learning as Optimal Control: Insights and Convergence Guarantees

no code implementations L4DC 2020 Jacob H. Seidman, Mahyar Fazlyab, Victor M. Preciado, George J. Pappas

By interpreting the min-max problem as an optimal control problem, it has recently been shown that one can exploit the compositional structure of neural networks in the optimization problem to improve the training time significantly.

Robust classification

Reach-SDP: Reachability Analysis of Closed-Loop Systems with Neural Network Controllers via Semidefinite Programming

no code implementations16 Apr 2020 Haimin Hu, Mahyar Fazlyab, Manfred Morari, George J. Pappas

There has been an increasing interest in using neural networks in closed-loop control systems to improve performance and reduce computational costs for on-line implementation.

Probabilistic Verification and Reachability Analysis of Neural Networks via Semidefinite Programming

1 code implementation9 Oct 2019 Mahyar Fazlyab, Manfred Morari, George J. Pappas

In this context, we discuss two relevant problems: (i) probabilistic safety verification, in which the goal is to find an upper bound on the probability of violating a safety specification; and (ii) confidence ellipsoid estimation, in which given a confidence ellipsoid for the input of the neural network, our goal is to compute a confidence ellipsoid for the output.

Efficient and Accurate Estimation of Lipschitz Constants for Deep Neural Networks

1 code implementation NeurIPS 2019 Mahyar Fazlyab, Alexander Robey, Hamed Hassani, Manfred Morari, George J. Pappas

The resulting SDP can be adapted to increase either the estimation accuracy (by capturing the interaction between activation functions of different layers) or scalability (by decomposition and parallel implementation).

Safety Verification and Robustness Analysis of Neural Networks via Quadratic Constraints and Semidefinite Programming

4 code implementations4 Mar 2019 Mahyar Fazlyab, Manfred Morari, George J. Pappas

Certifying the safety or robustness of neural networks against input uncertainties and adversarial attacks is an emerging challenge in the area of safe machine learning and control.

Cannot find the paper you are looking for? You can Submit a new open access paper.