no code implementations • 21 Dec 2023 • Yikun Ban, Yunzhe Qi, Jingrui He
Different from existing related tutorials, (1) we focus on the exploration perspective of contextual bandits to alleviate the ``Matthew Effect'' in the recommender systems, i. e., the rich get richer and the poor get poorer, concerning the popularity of items; (2) in addition to the conventional linear contextual bandits, we will also dedicated to neural contextual bandits which have emerged as an important branch in recent years, to investigate how neural networks benefit contextual bandits for personalized recommendation both empirically and theoretically; (3) we will cover the latest topic, collaborative neural contextual bandits, to incorporate both user heterogeneity and user correlations customized for recommender system; (4) we will provide and discuss the new emerging challenges and open questions for neural contextual bandits with applications in the personalized recommendation, especially for large neural models.
no code implementations • 21 Aug 2023 • Yunzhe Qi, Yikun Ban, Jingrui He
Contextual bandits algorithms aim to choose the optimal arm with the highest reward out of a set of candidates based on the contextual information.
no code implementations • 8 Jun 2022 • Yunzhe Qi, Yikun Ban, Jingrui He
Contextual bandits aim to identify among a set of arms the optimal one with the highest reward based on their contextual information.
no code implementations • 31 Jan 2022 • Yikun Ban, Yunzhe Qi, Tianxin Wei, Jingrui He
Contextual multi-armed bandits provide powerful tools to solve the exploitation-exploration dilemma in decision making, with direct applications in the personalized recommendation.