no code implementations • 17 Apr 2024 • Yaqun Yang, Jinlong Lei, Guanghui Wen, Yiguang Hong
This paper considers a distributed adaptive optimization problem, where all agents only have access to their local cost functions with a common unknown parameter, whereas they mean to collaboratively estimate the true parameter and find the optimal solution over a connected network.
no code implementations • 14 Oct 2023 • Jianguo Chen, Jinlong Lei, HongSheng Qi, Yiguang Hong
This work studies the parameter identification problem of a generalized non-cooperative game, where each player's cost function is influenced by an observable signal and some unknown parameters.
no code implementations • 31 May 2023 • Xinlei Yi, Xiuxian Li, Tao Yang, Lihua Xie, Yiguang Hong, Tianyou Chai, Karl H. Johansson
Moreover, if the loss functions are strongly convex, then the network regret bound is reduced to $\mathcal{O}(\log(T))$, and the network cumulative constraint violation bound is reduced to $\mathcal{O}(\sqrt{\log(T)T})$ and $\mathcal{O}(\log(T))$ without and with Slater's condition, respectively.
no code implementations • 19 Jan 2023 • Guanpu Chen, Gehui Xu, Fengxiang He, Yiguang Hong, Leszek Rutkowski, DaCheng Tao
This paper takes conjugate transformation to the formulation of non-convex multi-player games, and casts the complementary problem into a variational inequality (VI) problem with a continuous pseudo-gradient mapping.
no code implementations • 16 Jul 2022 • Xiuxian Li, Min Meng, Yiguang Hong, Jie Chen
Game theory has by now found numerous applications in various fields, including economics, industry, jurisprudence, and artificial intelligence, where each player only cares about its own interest in a noncooperative or cooperative manner, but without obvious malice to other players.
no code implementations • 14 May 2022 • Wenting Liu, Jinlong Lei, Peng Yi, Yiguang Hong
This paper considers no-regret learning for repeated continuous-kernel games with lossy bandit feedback.
no code implementations • 9 Apr 2022 • Ziliang Lyu, Xiangru Xu, Yiguang Hong
This paper develops a small-gain technique for the safety analysis and verification of interconnected systems with high-relative-degree safety constraints.
no code implementations • 8 Mar 2022 • Angela Fontan, Lingfei Wang, Yiguang Hong, Guodong Shi, Claudio Altafini
For the time-varying case, convergence to consensus can be guaranteed by the existence of a common Lyapunov function for all the signed Laplacians.
no code implementations • NeurIPS 2021 • Xi Wang, Zhipeng Tu, Yiguang Hong, Yingyi Wu, Guodong Shi
We consider online optimization over Riemannian manifolds, where a learner attempts to minimize a sequence of time-varying loss functions defined on Riemannian manifolds.
no code implementations • 4 Sep 2021 • Xiuxian Li, Kuo-Yi Lin, Li Li, Yiguang Hong, Jie Chen
For the first two cases, it can be shown that the scaled signGD converges at a linear rate.
no code implementations • 26 May 2021 • Xinran Li, Kuo-Yi Lin, Min Meng, Xiuxian Li, Li Li, Yiguang Hong, Jie Chen
Due to the growing awareness of driving safety and the development of sophisticated technologies, advanced driving assistance system (ADAS) has been equipped in more and more vehicles with higher accuracy and lower price.
no code implementations • NeurIPS 2020 • Jinlong Lei, Peng Yi, Yiguang Hong, Jie Chen, Guodong Shi
The regret bounds scaling with respect to $T$ match those obtained by state-of-the-art algorithms and fundamental limits in the corresponding centralized online optimization problems, e. g., $\mathcal{O}(\sqrt{T}) $ and $\mathcal{O}(\ln(T)) $ regrets are established for convex and strongly convex losses with full gradient feedback and two-points information, respectively.