no code implementations • 6 Mar 2025 • Aleksandrs Slivkins, Yunzong Xu, Shiliang Zuo
Our characterization identifies a partial identifiability property of the problem instance as the necessary and sufficient condition for the asymptotic success.
no code implementations • 31 May 2024 • Shiliang Zuo
I then study a setting where the marginal utility of each task is unknown so that the optimal contract must be learned or estimated with observational data.
no code implementations • 11 Mar 2024 • Shiliang Zuo
This work studies the repeated principal-agent problem from an online learning perspective.
no code implementations • 12 Dec 2023 • Rohan Deb, Yikun Ban, Shiliang Zuo, Jingrui He, Arindam Banerjee
Based on such a perturbed prediction, we show a ${\mathcal{O}}(\log T)$ regret for online regression with both squared loss and KL loss, and subsequently convert these respectively to $\tilde{\mathcal{O}}(\sqrt{KT})$ and $\tilde{\mathcal{O}}(\sqrt{KL^*} + K)$ regret for NeuCB, where $L^*$ is the loss of the best policy.
no code implementations • 26 Jul 2023 • Shiliang Zuo
In a total of $C$ rounds, the signal may be corrupted, though the value of $C$ is \emph{unknown} to the learner.
no code implementations • NeurIPS 2021 • Zhuolin Yang, Linyi Li, Xiaojun Xu, Shiliang Zuo, Qian Chen, Pan Zhou, Benjamin I. P. Rubinstein, Ce Zhang, Bo Li
To answer these questions, in this work we first theoretically analyze and outline sufficient conditions for adversarial transferability between models; then propose a practical algorithm to reduce the transferability between base models within an ensemble to improve its robustness.
1 code implementation • NeurIPS 2021 • Zhuolin Yang, Linyi Li, Xiaojun Xu, Shiliang Zuo, Qian Chen, Benjamin Rubinstein, Pan Zhou, Ce Zhang, Bo Li
To answer these questions, in this work we first theoretically analyze and outline sufficient conditions for adversarial transferability between models; then propose a practical algorithm to reduce the transferability between base models within an ensemble to improve its robustness.
no code implementations • 21 Aug 2020 • Shiliang Zuo
I design attack strategies against UCB and Thompson Sampling that only spend $\widehat{O}(\sqrt{\log T})$ cost.
no code implementations • 17 Aug 2020 • Shiliang Zuo
In non-convex settings, it has been shown that gradient descent is able to escape saddle points asymptotically and converge to local minimizers [Lee et.