1 code implementation • 4 May 2023 • Ziheng Cheng, Junzi Zhang, Akshay Agrawal, Stephen Boyd
Laplacian regularized stratified models (LRSM) are models that utilize the explicit or implicit network structure of the sub-problems as defined by the categorical features called strata (e. g., age, region, time, forecast horizon, etc.
1 code implementation • 3 Mar 2021 • Akshay Agrawal, Alnur Ali, Stephen Boyd
Our software scales to data sets with millions of items and tens of millions of distortion functions.
1 code implementation • 7 Jun 2020 • Akshay Agrawal, Shane Barratt, Stephen Boyd
A convex optimization model predicts an output from an input by solving a convex optimization problem.
2 code implementations • 27 Apr 2020 • Akshay Agrawal, Stephen Boyd
We use the adjoint of the derivative to implement differentiable log-log convex optimization layers in PyTorch and TensorFlow.
Optimization and Control
no code implementations • L4DC 2020 • Akshay Agrawal, Shane Barratt, Stephen Boyd, Bartolomeo Stellato
Common examples of such convex optimization control policies (COCPs) include the linear quadratic regulator (LQR), convex model predictive control (MPC), and convex control-Lyapunov or approximate dynamic programming (ADP) policies.
1 code implementation • NeurIPS 2019 • Akshay Agrawal, Brandon Amos, Shane Barratt, Stephen Boyd, Steven Diamond, Zico Kolter
In this paper, we propose an approach to differentiating through disciplined convex programs, a subclass of convex optimization problems used by domain-specific languages (DSLs) for convex optimization.
1 code implementation • 2 May 2019 • Akshay Agrawal, Stephen Boyd
We present a composition rule involving quasiconvex functions that generalizes the classical composition rule for convex functions.
Optimization and Control Mathematical Software
1 code implementation • 19 Apr 2019 • Akshay Agrawal, Shane Barratt, Stephen Boyd, Enzo Busseti, Walaa M. Moursi
These correspond to computing an approximate new solution, given a perturbation to the cone program coefficients (i. e., perturbation analysis), and to computing the gradient of a function of the solution with respect to the coefficients.
Optimization and Control
1 code implementation • 27 Feb 2019 • Akshay Agrawal, Akshay Naresh Modi, Alexandre Passos, Allen Lavoie, Ashish Agarwal, Asim Shankar, Igor Ganichev, Josh Levenberg, Mingsheng Hong, Rajat Monga, Shanqing Cai
TensorFlow Eager is a multi-stage, Python-embedded domain-specific language for hardware-accelerated machine learning, suitable for both interactive research and production.
1 code implementation • 13 Sep 2017 • Akshay Agrawal, Robin Verschueren, Steven Diamond, Stephen Boyd
We describe a modular rewriting system for translating optimization problems written in a domain-specific language to forms compatible with low-level solver interfaces.
Optimization and Control Mathematical Software