Search Results for author: Asuman Ozdaglar

Found 36 papers, 3 papers with code

Do LLM Agents Have Regret? A Case Study in Online Learning and Games

no code implementations25 Mar 2024 Chanwoo Park, Xiangyu Liu, Asuman Ozdaglar, Kaiqing Zhang

To better understand the limits of LLM agents in these interactive environments, we propose to study their interactions in benchmark decision-making settings in online learning and game theory, through the performance metric of \emph{regret}.

Decision Making

Matching of Users and Creators in Two-Sided Markets with Departures

no code implementations30 Dec 2023 Daniel Huttenlocher, Hannah Li, Liang Lyu, Asuman Ozdaglar, James Siderius

Most of the existing literature on platform recommendation algorithms largely focuses on user preferences and decisions, and does not simultaneously address creator incentives.

Two-Timescale Q-Learning with Function Approximation in Zero-Sum Stochastic Games

no code implementations8 Dec 2023 Zaiwei Chen, Kaiqing Zhang, Eric Mazumdar, Asuman Ozdaglar, Adam Wierman

Specifically, through a change of variable, we show that the update equation of the slow-timescale iterates resembles the classical smoothed best-response dynamics, where the regularized Nash gap serves as a valid Lyapunov function.

Q-Learning valid

EM for Mixture of Linear Regression with Clustered Data

no code implementations22 Aug 2023 Amirhossein Reisizadeh, Khashayar Gatmiry, Asuman Ozdaglar

In many settings however, heterogeneous data may be generated in clusters with shared structures, as is the case in several applications such as federated learning where a common latent variable governs the distribution of all the samples generated by a client.

Federated Learning regression

Multi-Player Zero-Sum Markov Games with Networked Separable Interactions

no code implementations NeurIPS 2023 Chanwoo Park, Kaiqing Zhang, Asuman Ozdaglar

We study a new class of Markov games, \emph(multi-player) zero-sum Markov Games} with \emph{Networked separable interactions} (zero-sum NMGs), to model the local interaction structure in non-cooperative multi-agent sequential decision-making.

Decision Making

Revisiting the Linear-Programming Framework for Offline RL with General Function Approximation

no code implementations28 Dec 2022 Asuman Ozdaglar, Sarath Pattathil, Jiawei Zhang, Kaiqing Zhang

Offline reinforcement learning (RL) aims to find an optimal policy for sequential decision-making using a pre-collected dataset, without further interaction with the environment.

Decision Making Offline RL +1

The Power of Regularization in Solving Extensive-Form Games

no code implementations19 Jun 2022 Mingyang Liu, Asuman Ozdaglar, Tiancheng Yu, Kaiqing Zhang

Second, we show that regularized counterfactual regret minimization (\texttt{Reg-CFR}), with a variant of optimistic mirror descent algorithm as regret-minimizer, can achieve $O(1/T^{1/4})$ best-iterate, and $O(1/T^{3/4})$ average-iterate convergence rate for finding NE in EFGs.

counterfactual

Optimal and Differentially Private Data Acquisition: Central and Local Mechanisms

no code implementations10 Jan 2022 Alireza Fallah, Ali Makhdoumi, Azarakhsh Malekian, Asuman Ozdaglar

We consider a platform's problem of collecting data from privacy sensitive users to estimate an underlying parameter of interest.

Independent Learning in Stochastic Games

no code implementations23 Nov 2021 Asuman Ozdaglar, Muhammed O. Sayin, Kaiqing Zhang

We focus on the development of simple and independent learning dynamics for stochastic games: each agent is myopic and chooses best-response type actions to other agents' strategy without any coordination with her opponent.

Autonomous Driving Reinforcement Learning (RL)

A Wasserstein Minimax Framework for Mixed Linear Regression

1 code implementation14 Jun 2021 Theo Diamandis, Yonina C. Eldar, Alireza Fallah, Farzan Farnia, Asuman Ozdaglar

We propose an optimal transport-based framework for MLR problems, Wasserstein Mixed Linear Regression (WMLR), which minimizes the Wasserstein distance between the learned and target mixture regression models.

Federated Learning regression

Decentralized Q-Learning in Zero-sum Markov Games

no code implementations NeurIPS 2021 Muhammed O. Sayin, Kaiqing Zhang, David S. Leslie, Tamer Basar, Asuman Ozdaglar

The key challenge in this decentralized setting is the non-stationarity of the environment from an agent's perspective, since both her own payoffs and the system evolution depend on the actions of other agents, and each agent adapts her policies simultaneously and independently.

Multi-agent Reinforcement Learning Q-Learning

Generalization of Model-Agnostic Meta-Learning Algorithms: Recurring and Unseen Tasks

no code implementations NeurIPS 2021 Alireza Fallah, Aryan Mokhtari, Asuman Ozdaglar

In this paper, we study the generalization properties of Model-Agnostic Meta-Learning (MAML) algorithms for supervised learning problems.

Generalization Bounds Meta-Learning

Personalized Federated Learning with Theoretical Guarantees: A Model-Agnostic Meta-Learning Approach

2 code implementations NeurIPS 2020 Alireza Fallah, Aryan Mokhtari, Asuman Ozdaglar

In this paper, we study a personalized variant of the federated learning in which our goal is to find an initial shared model that current or new users can easily adapt to their local dataset by performing one or a few steps of gradient descent with respect to their own data.

Meta-Learning Personalized Federated Learning

Train simultaneously, generalize better: Stability of gradient-based minimax learners

no code implementations23 Oct 2020 Farzan Farnia, Asuman Ozdaglar

In this paper, we show that the optimization algorithm also plays a key role in the generalization performance of the trained minimax model.

Multi-agent Bayesian Learning with Adaptive Strategies: Convergence and Stability

no code implementations18 Oct 2020 Manxi Wu, Saurabh Amin, Asuman Ozdaglar

Any fixed point belief consistently estimates the payoff distribution given the fixed point strategy profile.

Fictitious play in zero-sum stochastic games

no code implementations8 Oct 2020 Muhammed O. Sayin, Francesca Parise, Asuman Ozdaglar

We present a novel variant of fictitious play dynamics combining classical fictitious play with Q-learning for stochastic games and analyze its convergence properties in two-player zero-sum stochastic games.

Q-Learning

Bayesian Learning with Adaptive Load Allocation Strategies

no code implementations L4DC 2020 Manxi Wu, Saurabh Amin, Asuman Ozdaglar

We study a Bayesian learning dynamics induced by agents who repeatedly allocate loads on a set of resources based on their belief of an unknown parameter that affects the cost distributions of resources.

GANs May Have No Nash Equilibria

no code implementations ICML 2020 Farzan Farnia, Asuman Ozdaglar

We discuss several numerical experiments demonstrating the existence of proximal equilibrium solutions in GAN minimax problems.

Personalized Federated Learning: A Meta-Learning Approach

no code implementations19 Feb 2020 Alireza Fallah, Aryan Mokhtari, Asuman Ozdaglar

In this paper, we study a personalized variant of the federated learning in which our goal is to find an initial shared model that current or new users can easily adapt to their local dataset by performing one or a few steps of gradient descent with respect to their own data.

Meta-Learning Personalized Federated Learning

An Optimal Multistage Stochastic Gradient Method for Minimax Problems

no code implementations13 Feb 2020 Alireza Fallah, Asuman Ozdaglar, Sarath Pattathil

Next, we propose a multistage variant of stochastic GDA (M-GDA) that runs in multiple stages with a particular learning rate decay schedule and converges to the exact solution of the minimax problem.

On the Convergence Theory of Debiased Model-Agnostic Meta-Reinforcement Learning

1 code implementation NeurIPS 2021 Alireza Fallah, Kristian Georgiev, Aryan Mokhtari, Asuman Ozdaglar

We consider Model-Agnostic Meta-Learning (MAML) methods for Reinforcement Learning (RL) problems, where the goal is to find a policy using data from several tasks represented by Markov Decision Processes (MDPs) that can be updated by one step of stochastic policy gradient for the realized MDP.

Meta-Learning Meta Reinforcement Learning +3

A Decentralized Proximal Point-type Method for Saddle Point Problems

no code implementations31 Oct 2019 Weijie Liu, Aryan Mokhtari, Asuman Ozdaglar, Sarath Pattathil, Zebang Shen, Nenggan Zheng

In this paper, we focus on solving a class of constrained non-convex non-concave saddle point problems in a decentralized manner by a group of nodes in a network.

Vocal Bursts Type Prediction

Robust Distributed Accelerated Stochastic Gradient Methods for Multi-Agent Networks

no code implementations19 Oct 2019 Alireza Fallah, Mert Gurbuzbalaban, Asuman Ozdaglar, Umut Simsekli, Lingjiong Zhu

When gradients do not contain noise, we also prove that distributed accelerated methods can \emph{achieve acceleration}, requiring $\mathcal{O}(\kappa \log(1/\varepsilon))$ gradient evaluations and $\mathcal{O}(\kappa \log(1/\varepsilon))$ communications to converge to the same fixed point with the non-accelerated variant where $\kappa$ is the condition number and $\varepsilon$ is the target accuracy.

Stochastic Optimization

On the Convergence Theory of Gradient-Based Model-Agnostic Meta-Learning Algorithms

no code implementations27 Aug 2019 Alireza Fallah, Aryan Mokhtari, Asuman Ozdaglar

We study the convergence of a class of gradient-based Model-Agnostic Meta-Learning (MAML) methods and characterize their overall complexity as well as their best achievable accuracy in terms of gradient norm for nonconvex loss functions.

Meta-Learning

A Unified Analysis of Extra-gradient and Optimistic Gradient Methods for Saddle Point Problems: Proximal Point Approach

no code implementations24 Jan 2019 Aryan Mokhtari, Asuman Ozdaglar, Sarath Pattathil

In this paper we consider solving saddle point problems using two variants of Gradient Descent-Ascent algorithms, Extra-gradient (EG) and Optimistic Gradient Descent Ascent (OGDA) methods.

A Universally Optimal Multistage Accelerated Stochastic Gradient Method

no code implementations NeurIPS 2019 Necdet Serhat Aybat, Alireza Fallah, Mert Gurbuzbalaban, Asuman Ozdaglar

We study the problem of minimizing a strongly convex, smooth function when we have noisy estimates of its gradient.

Escaping Saddle Points in Constrained Optimization

no code implementations NeurIPS 2018 Aryan Mokhtari, Asuman Ozdaglar, Ali Jadbabaie

We propose a generic framework that yields convergence to a second-order stationary point of the problem, if the convex set $\mathcal{C}$ is simple for a quadratic objective function.

Convergence Rate of Block-Coordinate Maximization Burer-Monteiro Method for Solving Large SDPs

no code implementations12 Jul 2018 Murat A. Erdogdu, Asuman Ozdaglar, Pablo A. Parrilo, Nuri Denizcan Vanli

Furthermore, incorporating Lanczos method to the block-coordinate maximization, we propose an algorithm that is guaranteed to return a solution that provides $1-O(1/r)$ approximation to the original SDP without any assumptions, where $r$ is the rank of the factorization.

Community Detection

Robust Accelerated Gradient Methods for Smooth Strongly Convex Functions

no code implementations27 May 2018 Necdet Serhat Aybat, Alireza Fallah, Mert Gurbuzbalaban, Asuman Ozdaglar

We study the trade-offs between convergence rate and robustness to gradient errors in designing a first-order algorithm.

When Cyclic Coordinate Descent Outperforms Randomized Coordinate Descent

no code implementations NeurIPS 2017 Mert Gurbuzbalaban, Asuman Ozdaglar, Pablo A. Parrilo, Nuri Vanli

The coordinate descent (CD) method is a classical optimization algorithm that has seen a revival of interest because of its competitive performance in machine learning applications.

Computing the Stationary Distribution Locally

no code implementations NeurIPS 2013 Christina E. Lee, Asuman Ozdaglar, Devavrat Shah

In this paper, we provide a novel algorithm that answers whether a chosen state in a MC has stationary probability larger than some $\Delta \in (0, 1)$.

Distributed Subgradient Methods and Quantization Effects

no code implementations8 Mar 2008 Angelia Nedić, Alex Olshevsky, Asuman Ozdaglar, John N. Tsitsiklis

We consider a convex unconstrained optimization problem that arises in a network of agents whose goal is to cooperatively optimize the sum of the individual agent objective functions through local computations and communications.

Optimization and Control

Cannot find the paper you are looking for? You can Submit a new open access paper.