no code implementations • ICML 2020 • Amélie Héliou, Panayotis Mertikopoulos, Zhengyuan Zhou
Motivated by applications to online advertising and recommender systems, we consider a game-theoretic model with delayed rewards and asynchronous, payoff-based feedback.
no code implementations • 29 Dec 2024 • Kyriakos Lotidis, Angeliki Giannou, Panayotis Mertikopoulos, Nicholas Bambos
Motivated by the success of Nesterov's accelerated gradient algorithm for convex minimization problems, we examine whether it is possible to achieve similar performance gains in the context of online learning in games.
no code implementations • 28 Dec 2024 • Davide Legacci, Panayotis Mertikopoulos, Christos H. Papadimitriou, Georgios Piliouras, Bary S. R. Pradelski
By contrast, in harmonic games - the strategic counterpart of potential games, where players have conflicting interests - very little is known outside the narrow subclass of 2-player zero-sum games with a fully-mixed equilibrium.
no code implementations • 25 Jul 2024 • Panayotis Mertikopoulos, William H. Sandholm
We consider a model of learning and evolution in games whose action sets are endowed with a partition-based similarity structure intended to capture exogenous similarities between strategies.
no code implementations • 13 Jun 2024 • Waïss Azizian, Franck Iutzeler, Jérôme Malick, Panayotis Mertikopoulos
Using an approach based on the theory of large deviations and randomly perturbed dynamical systems, we show that the long-run distribution of SGD resembles the Boltzmann-Gibbs distribution of equilibrium thermodynamics with temperature equal to the method's step-size and energy levels determined by the problem's objective and the statistics of the noise.
no code implementations • 27 May 2024 • Iosif Lytras, Panayotis Mertikopoulos
Motivated by applications to deep learning which often fail standard Lipschitz smoothness requirements, we examine the problem of sampling from distributions that are not log-concave and are only weakly dissipative, with log-gradients allowed to grow superlinearly at infinity.
no code implementations • 12 May 2024 • Davide Legacci, Panayotis Mertikopoulos, Bary Pradelski
In view of the complexity of the dynamics of learning in games, we seek to decompose a game into simpler components where the dynamics' long-run behavior is well understood.
no code implementations • NeurIPS 2023 • Iosif Sakos, Emmanouil-Vasileios Vlatakis-Gkaragkounis, Panayotis Mertikopoulos, Georgios Piliouras
A wide array of modern machine learning applications - from adversarial models to multi-agent reinforcement learning - can be formulated as non-cooperative games whose Nash equilibria represent the system's desired operational states.
no code implementations • 17 Nov 2023 • Francisca Vasconcelos, Emmanouil-Vasileios Vlatakis-Gkaragkounis, Panayotis Mertikopoulos, Georgios Piliouras, Michael I. Jordan
In 2008, Jain and Watrous proposed the first classical algorithm for computing equilibria in quantum zero-sum games using the Matrix Multiplicative Weight Updates (MMWU) method to achieve a convergence rate of $\mathcal{O}(d/\epsilon^2)$ iterations to $\epsilon$-Nash equilibria in the $4^d$-dimensional spectraplex.
no code implementations • 5 Feb 2023 • Kyriakos Lotidis, Panayotis Mertikopoulos, Nicholas Bambos
In this paper, we introduce a class of learning dynamics for general quantum games, that we call "follow the quantum regularized leader" (FTQL), in reference to the classical "follow the regularized leader" (FTRL) template for learning in finite games.
no code implementations • 15 Nov 2022 • Waïss Azizian, Franck Iutzeler, Jérôme Malick, Panayotis Mertikopoulos
For generality, we focus on local solutions of constrained, non-monotone variational inequalities, and we show that the convergence rate of a given method depends sharply on its associated Legendre exponent, a notion that measures the growth rate of the underlying Bregman function (Euclidean, entropic, or other) near a solution.
no code implementations • 23 Oct 2022 • Tianyi Lin, Panayotis Mertikopoulos, Michael I. Jordan
Specifically, we show that the proposed methods generate iterates that remain within a bounded set and that the averaged iterates converge to an $\epsilon$-saddle point within $O(\epsilon^{-2/3})$ iterations in terms of a restricted gap function.
no code implementations • 17 Oct 2022 • Angeliki Giannou, Kyriakos Lotidis, Panayotis Mertikopoulos, Emmanouil-Vasileios Vlatakis-Gkaragkounis
Learning in stochastic games is a notoriously difficult problem because, in addition to each other's strategic decisions, the players must also contend with the fact that the game itself evolves over time, possibly in a very complicated manner.
1 code implementation • 15 Jul 2022 • Marina Costantini, Nikolaos Liakopoulos, Panayotis Mertikopoulos, Thrasyvoulos Spyropoulos
In decentralized optimization environments, each agent $i$ in a network of $n$ nodes has its own private function $f_i$, and nodes communicate with their neighbors to cooperatively minimize the aggregate objective $\sum_{i=1}^n f_i$.
no code implementations • 19 Jun 2022 • Matthieu Martin, Panayotis Mertikopoulos, Thibaud Rahier, Houssam Zenati
In many online decision processes, the optimizing agent is called to choose between large numbers of alternatives with many inherent similarities; in turn, these similarities imply closely correlated losses that may confound standard discrete choice models and bandit algorithms.
no code implementations • 14 Jun 2022 • Mohammad Reza Karimi, Ya-Ping Hsieh, Panayotis Mertikopoulos, Andreas Krause
We examine a wide class of stochastic approximation algorithms for solving (stochastic) nonlinear problems on Riemannian manifolds.
no code implementations • 13 Jun 2022 • Yu-Guan Hsieh, Kimon Antonakopoulos, Volkan Cevher, Panayotis Mertikopoulos
We examine the problem of regret minimization when the learner is involved in a continuous game with other optimizing agents: in this case, if all players follow a no-regret algorithm, it is possible to achieve significantly lower regret relative to fully adversarial environments.
no code implementations • 8 Jun 2022 • Panayotis Mertikopoulos, Ya-Ping Hsieh, Volkan Cevher
We develop a flexible stochastic approximation framework for analyzing the long-run behavior of learning in games (both continuous and finite).
no code implementations • NeurIPS 2021 • Angeliki Giannou, Emmanouil-Vasileios Vlatakis-Gkaragkounis, Panayotis Mertikopoulos
In this paper, we examine the convergence rate of a wide range of regularized methods for learning in games.
no code implementations • NeurIPS 2021 • Dong Quan Vu, Kimon Antonakopoulos, Panayotis Mertikopoulos
We examine an adaptive learning framework for nonatomic congestion games where the players' cost functions may be subject to exogenous fluctuations (e. g., due to disturbances in the network, variations in the traffic going through a link).
no code implementations • NeurIPS 2021 • Kimon Antonakopoulos, Thomas Pethick, Ali Kavis, Panayotis Mertikopoulos, Volkan Cevher
Our first result is that the algorithm achieves the optimal rates of convergence for cocoercive problems when the profile of the randomness is known to the optimizer: $\mathcal{O}(1/\sqrt{T})$ for absolute noise profiles, and $\mathcal{O}(1/T)$ for relative ones.
no code implementations • 13 Sep 2021 • Amélie Héliou, Matthieu Martin, Panayotis Mertikopoulos, Thibaud Rahier
We propose a hierarchical version of dual averaging for zeroth-order online non-convex optimization - i. e., learning processes where, at each stage, the optimizer is facing an unknown non-convex loss function and only receives the incurred loss as feedback.
no code implementations • NeurIPS 2021 • Kimon Antonakopoulos, Panayotis Mertikopoulos
We propose a new family of adaptive first-order methods for a class of convex minimization problems that may fail to be Lipschitz continuous or smooth in the standard sense.
no code implementations • 6 Jul 2021 • Zhengyuan Zhou, Panayotis Mertikopoulos, Nicholas Bambos, Peter W. Glynn, Yinyu Ye
One of the most widely used methods for solving large-scale stochastic optimization problems is distributed asynchronous stochastic gradient descent (DASGD), a family of algorithms that result from parallelizing stochastic gradient descent on distributed computing architectures (possibly) asychronously.
no code implementations • 5 Jul 2021 • Waïss Azizian, Franck Iutzeler, Jérôme Malick, Panayotis Mertikopoulos
In this paper, we analyze the local convergence rate of optimistic mirror descent methods in stochastic variational inequalities, a class of optimization problems with important applications to learning theory and machine learning.
no code implementations • 4 Jul 2021 • Saeed Hadikhanloo, Rida Laraki, Panayotis Mertikopoulos, Sylvain Sorin
We examine the long-run behavior of a wide range of dynamics for learning in nonatomic games, in both discrete and continuous time.
no code implementations • 27 May 2021 • Yu-Guan Hsieh, Franck Iutzeler, Jérôme Malick, Panayotis Mertikopoulos
In networks of autonomous agents (e. g., fleets of vehicles, scattered sensors), the problem of minimizing the sum of the agents' local functions has received a lot of interest.
no code implementations • NeurIPS 2021 • Kimon Antonakopoulos, Panayotis Mertikopoulos
We propose a new family of adaptive first-order methods for a class of convex minimization problems that may fail to be Lipschitz continuous or smooth in the standard sense.
no code implementations • 26 Apr 2021 • Yu-Guan Hsieh, Kimon Antonakopoulos, Panayotis Mertikopoulos
In game-theoretic learning, several agents are simultaneously following their individual interests, so the environment is non-stationary from each player's perspective.
no code implementations • 12 Jan 2021 • Angeliki Giannou, Emmanouil-Vasileios Vlatakis-Gkaragkounis, Panayotis Mertikopoulos
This equivalence extends existing continuous-time versions of the folk theorem of evolutionary game theory to a bona fide algorithmic learning setting, and it provides a clear refinement criterion for the prediction of the day-to-day behavior of no-regret learning in games
no code implementations • 21 Dec 2020 • Yu-Guan Hsieh, Franck Iutzeler, Jérôme Malick, Panayotis Mertikopoulos
In this paper, we provide a general framework for studying multi-agent online learning problems in the presence of delays and asynchronicities.
no code implementations • ICLR 2021 • Kimon Antonakopoulos, E. Veronica Belmega, Panayotis Mertikopoulos
We present a new family of min-max optimization algorithms that automatically exploit the geometry of the gradient data observed at earlier iterations to perform more informative extra-gradient steps in later ones.
no code implementations • NeurIPS 2020 • Lampros Flokas, Emmanouil-Vasileios Vlatakis-Gkaragkounis, Thanasis Lianeas, Panayotis Mertikopoulos, Georgios Piliouras
Understanding the behavior of no-regret dynamics in general $N$-player games is a fundamental question in online learning and game theory.
no code implementations • NeurIPS 2020 • Amélie Héliou, Matthieu Martin, Panayotis Mertikopoulos, Thibaud Rahier
We consider the problem of online learning with non-convex losses.
no code implementations • 13 Oct 2020 • Nadav Hallak, Panayotis Mertikopoulos, Volkan Cevher
In this setting, the minimization of external regret is beyond reach for first-order methods, so we focus on a local regret measure defined via a proximal-gradient mapping.
no code implementations • NeurIPS 2020 • Panayotis Mertikopoulos, Nadav Hallak, Ali Kavis, Volkan Cevher
This paper analyzes the trajectories of stochastic gradient descent (SGD) to help understand the algorithm's convergence properties in non-convex problems.
no code implementations • 16 Jun 2020 • Ya-Ping Hsieh, Panayotis Mertikopoulos, Volkan Cevher
Compared to ordinary function minimization problems, min-max optimization algorithms encounter far greater challenges because of the existence of periodic cycles and similar phenomena.
no code implementations • ICLR 2020 • Kimon Antonakopoulos, E. Veronica Belmega, Panayotis Mertikopoulos
Motivated by applications to machine learning and imaging science, we study a class of online and stochastic optimization problems with loss functions that are not Lipschitz continuous; in particular, the loss functions encountered by the optimizer could exhibit gradient singularities or be singular themselves.
no code implementations • NeurIPS 2020 • Yu-Guan Hsieh, Franck Iutzeler, Jérôme Malick, Panayotis Mertikopoulos
Owing to their stability and convergence speed, extragradient methods have become a staple for solving large-scale saddle-point problems in machine learning.
no code implementations • ICML 2020 • Ahmet Alacaoglu, Yura Malitsky, Panayotis Mertikopoulos, Volkan Cevher
In this paper, we focus on a theory-practice gap for Adam and its variants (AMSgrad, AdamNC, etc.).
no code implementations • ICML 2020 • Tianyi Lin, Zhengyuan Zhou, Panayotis Mertikopoulos, Michael. I. Jordan
In this paper, we consider multi-agent learning via online gradient descent in a class of games called $\lambda$-cocoercive games, a fairly broad class of games that admits many Nash equilibria and that properly includes unconstrained strongly monotone games.
no code implementations • NeurIPS 2019 • Kimon Antonakopoulos, Veronica Belmega, Panayotis Mertikopoulos
Lipschitz continuity is a central requirement for achieving the optimal O(1/T) rate of convergence in monotone, deterministic variational inequalities (a setting that includes convex minimization, convex-concave optimization, nonatomic games, and many other problems).
no code implementations • NeurIPS 2019 • Yu-Guan Hsieh, Franck Iutzeler, Jérôme Malick, Panayotis Mertikopoulos
Variational inequalities have recently attracted considerable interest in machine learning as a flexible paradigm for models that go beyond ordinary loss function minimization (such as generative adversarial networks and related deep learning systems).
no code implementations • ICLR 2019 • Panayotis Mertikopoulos, Bruno Lecouat, Houssam Zenati, Chuan-Sheng Foo, Vijay Chandrasekhar, Georgios Piliouras
Owing to their connection with generative adversarial networks (GANs), saddle-point problems have recently attracted considerable interest in machine learning and beyond.
no code implementations • 9 Feb 2019 • Radu Ioan Bot, Panayotis Mertikopoulos, Mathias Staudigl, Phan Tu Vuong
We develop a new stochastic algorithm with variance reduction for solving pseudo-monotone stochastic variational inequalities.
no code implementations • NeurIPS 2018 • Zhengyuan Zhou, Panayotis Mertikopoulos, Susan Athey, Nicholas Bambos, Peter W. Glynn, Yinyu Ye
We consider a game-theoretical multi-agent learning problem where the feedback information can be lost during the learning process and rewards are given by a broad class of games known as variationally stable games.
no code implementations • NeurIPS 2018 • Mario Bravo, David Leslie, Panayotis Mertikopoulos
This paper examines the long-run behavior of learning with bandit feedback in non-cooperative concave games.
no code implementations • 3 Oct 2018 • Mario Bravo, David S. Leslie, Panayotis Mertikopoulos
This paper examines the long-run behavior of learning with bandit feedback in non-cooperative concave games.
no code implementations • 25 Sep 2018 • Immanuel M. Bomze, Panayotis Mertikopoulos, Werner Schachinger, Mathias Staudigl
In the case of linearly constrained quadratic programs (not necessarily convex), we also show that the method's convergence rate is $\mathcal{O}(1/k^\rho)$ for some $\rho\in(0, 1]$ that depends only on the choice of kernel function (i. e., not on the problem's primitives).
no code implementations • 10 Sep 2018 • Benoit Duvocelle, Panayotis Mertikopoulos, Mathias Staudigl, Dries Vermeulen
We examine the long-run behavior of multi-agent online learning in games that evolve over time.
no code implementations • 7 Jul 2018 • Panayotis Mertikopoulos, Bruno Lecouat, Houssam Zenati, Chuan-Sheng Foo, Vijay Chandrasekhar, Georgios Piliouras
Owing to their connection with generative adversarial networks (GANs), saddle-point problems have recently attracted considerable interest in machine learning and beyond.
no code implementations • ICML 2018 • Zhengyuan Zhou, Panayotis Mertikopoulos, Nicholas Bambos, Peter Glynn, Yinyu Ye, Li-Jia Li, Li Fei-Fei
One of the most widely used optimization methods for large-scale machine learning problems is distributed asynchronous stochastic gradient descent (DASGD).
no code implementations • 12 Apr 2018 • E. Veronica Belmega, Panayotis Mertikopoulos, Romain Negrel, Luca Sanguinetti
Spurred by the enthusiasm surrounding the "Big Data" paradigm, the mathematical and algorithmic tools of online optimization have found widespread use in problems where the trade-off between data exploration and exploitation plays a predominant role.
no code implementations • NeurIPS 2017 • Zhengyuan Zhou, Panayotis Mertikopoulos, Nicholas Bambos, Stephen Boyd, Peter W. Glynn
In this paper, we examine a class of non-convex stochastic optimization problems which we call variationally coherent, and which properly includes pseudo-/quasiconvex and star-convex optimization problems.
no code implementations • NeurIPS 2017 • Amélie Heliou, Johanne Cohen, Panayotis Mertikopoulos
This paper examines the equilibrium convergence properties of no-regret learning with exponential weights in potential games.
no code implementations • NeurIPS 2017 • Zhengyuan Zhou, Panayotis Mertikopoulos, Nicholas Bambos, Peter W. Glynn, Claire Tomlin
We consider a model of game-theoretic learning based on online mirror descent (OMD) with asynchronous and delayed feedback information.
no code implementations • 8 Sep 2017 • Panayotis Mertikopoulos, Christos Papadimitriou, Georgios Piliouras
Regularized learning is a fundamental technique in online optimization, machine learning and many other fields of computer science.
no code implementations • 18 Jun 2017 • Zhengyuan Zhou, Panayotis Mertikopoulos, Nicholas Bambos, Stephen Boyd, Peter Glynn
In this paper, we examine the convergence of mirror descent in a class of stochastic optimization problems that are not necessarily convex (or even quasi-convex), and which we call variationally coherent.
no code implementations • 21 Nov 2016 • Panayotis Mertikopoulos, Mathias Staudigl
In the vanishing noise limit, we show that the dynamics converge to the solution set of the underlying problem (a. s.).
no code implementations • 25 Aug 2016 • Panayotis Mertikopoulos, Zhengyuan Zhou
This paper examines the convergence of no-regret learning in games with continuous action sets.
no code implementations • 29 Jul 2016 • Johanne Cohen, Amélie Héliou, Panayotis Mertikopoulos
Motivated by applications to data networks where fast convergence is essential, we analyze the problem of learning in generic N-person games that admit a Nash equilibrium in pure strategies.
no code implementations • 3 Jun 2016 • Panayotis Mertikopoulos, E. Veronica Belmega, Romain Negrel, Luca Sanguinetti
In this paper, we investigate a distributed learning scheme for a broad class of stochastic optimization problems and games that arise in signal processing and wireless communications.
no code implementations • 4 May 2016 • Panayotis Mertikopoulos, Aris L. Moustakas, Anna Tzanakaki
Motivated by the massive deployment of power-hungry data centers for service provisioning, we examine the problem of routing in optical networks with the aim of minimizing traffic-driven power consumption.
no code implementations • 20 Dec 2014 • Mario Bravo, Panayotis Mertikopoulos
Motivated by the scarcity of accurate payoff feedback in practical applications of game theory, we examine a class of learning dynamics where players adjust their choices based on past payoff observations that are subject to noise and random disturbances.
no code implementations • 1 Dec 2014 • Steven Perkins, Panayotis Mertikopoulos, David S. Leslie
To do so, we extend the theory of finite-dimensional two-timescale stochastic approximation to an infinite-dimensional, Banach space setting, and we prove that the continuous dynamics of the process converge to equilibrium in the case of potential games.
no code implementations • 23 Jul 2014 • Panayotis Mertikopoulos, William H. Sandholm
We investigate a class of reinforcement learning dynamics where players adjust their strategies based on their actions' cumulative payoffs over time - specifically, by playing mixed strategies that maximize their expected cumulative payoff minus a regularization term.
no code implementations • 27 Jan 2014 • Joon Kwon, Panayotis Mertikopoulos
We consider a family of learning strategies for online optimization problems that evolve in continuous time and we show that they lead to no regret.
no code implementations • 9 Mar 2013 • Pierre Coucheney, Bruno Gaujal, Panayotis Mertikopoulos
Starting from a heuristic learning scheme for N-person games, we derive a new class of continuous-time learning dynamics consisting of a replicator-like drift adjusted by a penalty term that renders the boundary of the game's strategy space repelling.