no code implementations • 18 Feb 2024 • Jianling Wang, Haokai Lu, James Caverlee, Ed Chi, Minmin Chen
The reasoning and generalization capabilities of LLMs can help us better understand user preferences and item characteristics, offering exciting prospects to enhance recommendation systems.
no code implementations • 24 Oct 2023 • Zhenzhen Liu, Chao Wan, Varsha Kishore, Jin Peng Zhou, Minmin Chen, Kilian Q. Weinberger
The results show that CoBa is effective and efficient in reducing hallucination, and offers great adaptability and flexibility.
no code implementations • 29 Jul 2023 • Xinyang Yi, Shao-Chuan Wang, Ruining He, Hariharan Chandrasekaran, Charles Wu, Lukasz Heldt, Lichan Hong, Minmin Chen, Ed H. Chi
In this paper, we introduce Online Matching: a scalable closed-loop bandit system learning from users' direct feedback on items in real time.
no code implementations • 2 Jun 2023 • Jianling Wang, Haokai Lu, Sai Zhang, Bart Locanthi, HaoTing Wang, Dylan Greaves, Benjamin Lipshitz, Sriraj Badam, Ed H. Chi, Cristos Goodrow, Su-Lin Wu, Lexi Baugher, Minmin Chen
The multi-funnel setup effectively balances between coverage and relevance.
no code implementations • 2 Jun 2023 • Pan Li, Yuyan Wang, Ed H. Chi, Minmin Chen
For the aspect-based recommendation component, the extracted aspects are concatenated with the usual user and item features used by the recommendation model.
no code implementations • 2 Jun 2023 • Pan Li, Yuyan Wang, Ed H. Chi, Minmin Chen
Hierarchical structure, on the other hand, exists in a user's novelty-seeking intent, which is manifested as a static and intrinsic user preference for seeking novelty along with a dynamic session-based propensity.
Hierarchical Reinforcement Learning Recommendation Systems +1
no code implementations • 24 May 2023 • Konstantina Christakopoulou, Alberto Lalama, Cj Adams, Iris Qu, Yifat Amir, Samer Chucri, Pierce Vollucci, Fabio Soldo, Dina Bseiso, Sarah Scodel, Lucas Dixon, Ed H. Chi, Minmin Chen
We argue, and demonstrate through extensive experiments, that LLMs as foundation models can reason through user activities, and describe their interests in nuanced and interesting ways, similar to how a human would.
no code implementations • 12 May 2023 • Yi Su, Xiangyu Wang, Elaine Ya Le, Liang Liu, Yuening Li, Haokai Lu, Benjamin Lipshitz, Sriraj Badam, Lukasz Heldt, Shuchao Bi, Ed Chi, Cristos Goodrow, Su-Lin Wu, Lexi Baugher, Minmin Chen
We conduct live experiments on one of the largest short-form video recommendation platforms that serves billions of users to validate the new experiment designs, quantify the long-term values of exploration, and to verify the effectiveness of the adopted neural linear bandit algorithm for exploration.
no code implementations • 17 Nov 2022 • Bo Chang, Alexandros Karatzoglou, Yuyan Wang, Can Xu, Ed H. Chi, Minmin Chen
We demonstrate the effectiveness of the latent user intent modeling via offline analyses as well as live experiments on a large-scale industrial recommendation platform.
no code implementations • 30 Sep 2022 • Konstantina Christakopoulou, Can Xu, Sai Zhang, Sriraj Badam, Trevor Potter, Daniel Li, Hao Wan, Xinyang Yi, Ya Le, Chris Berg, Eric Bencomo Dixon, Ed H. Chi, Minmin Chen
How might we design Reinforcement Learning (RL)-based recommenders that encourage aligning user trajectories with the underlying user satisfaction?
no code implementations • 2 Apr 2022 • Jianling Wang, Ya Le, Bo Chang, Yuyan Wang, Ed H. Chi, Minmin Chen
Users who come to recommendation platforms are heterogeneous in activity levels.
no code implementations • 26 Jan 2022 • Bo Chang, Can Xu, Matthieu Lê, Jingchen Feng, Ya Le, Sriraj Badam, Ed Chi, Minmin Chen
Recurrent recommender systems have been successful in capturing the temporal dynamics in users' activity trajectories.
no code implementations • 6 May 2021 • Ruohan Zhan, Konstantina Christakopoulou, Ya Le, Jayden Ooi, Martin Mladenov, Alex Beutel, Craig Boutilier, Ed H. Chi, Minmin Chen
We then build a REINFORCE recommender agent, coined EcoAgent, to optimize a joint objective of user utility and the counterfactual utility lift of the provider associated with the recommended content, which we show to be equivalent to maximizing overall user utility and the utilities of all providers on the platform under some mild assumptions.
no code implementations • ICLR 2021 • Yijie Guo, Shengyu Feng, Nicolas Le Roux, Ed Chi, Honglak Lee, Minmin Chen
Many real-world applications of reinforcement learning (RL) require the agent to learn from a fixed set of trajectories, without collecting new interactions.
no code implementations • NeurIPS 2019 • Minmin Chen, Ramki Gummadi, Chris Harris, Dale Schuurmans
We investigate batch policy optimization for cost-sensitive classification and contextual bandits---two related tasks that obviate exploration but require generalizing from observed rewards to action selections in unseen contexts.
no code implementations • 25 Sep 2019 • Dar Gilboa, Bo Chang, Minmin Chen, Greg Yang, Samuel S. Schoenholz, Ed H. Chi, Jeffrey Pennington
We demonstrate the efficacy of our initialization scheme on multiple sequence tasks, on which it enables successful training while a standard initialization either fails completely or is orders of magnitude slower.
no code implementations • 23 May 2019 • Francois Belletti, Minmin Chen, Ed H. Chi
Characterizing temporal dependence patterns is a critical step in understanding the statistical properties of sequential data.
1 code implementation • ICLR 2019 • Bo Chang, Minmin Chen, Eldad Haber, Ed H. Chi
In this paper, we draw connections between recurrent networks and ordinary differential equations.
no code implementations • 22 Feb 2019 • Jiaxi Tang, Francois Belletti, Sagar Jain, Minmin Chen, Alex Beutel, Can Xu, Ed H. Chi
Our approach employs a mixture of models, each with a different temporal range.
no code implementations • 25 Jan 2019 • Dar Gilboa, Bo Chang, Minmin Chen, Greg Yang, Samuel S. Schoenholz, Ed H. Chi, Jeffrey Pennington
We demonstrate the efficacy of our initialization scheme on multiple sequence tasks, on which it enables successful training while a standard initialization either fails completely or is orders of magnitude slower.
1 code implementation • 6 Dec 2018 • Minmin Chen, Alex Beutel, Paul Covington, Sagar Jain, Francois Belletti, Ed Chi
The contributions of the paper are: (1) scaling REINFORCE to a production recommender system with an action space on the orders of millions; (2) applying off-policy correction to address data biases in learning from logged feedback collected from multiple behavior policies; (3) proposing a novel top-K off-policy correction to account for our policy recommending multiple items at a time; (4) showcasing the value of exploration.
no code implementations • ICML 2018 • Minmin Chen, Jeffrey Pennington, Samuel S. Schoenholz
We develop a theory for signal propagation in recurrent networks after random initialization using a combination of mean field theory and random matrix theory.
no code implementations • 18 Nov 2017 • Minmin Chen
We introduce MinimalRNN, a new recurrent neural network architecture that achieves comparable performance as the popular gated RNNs with a simplified structure.
1 code implementation • 8 Jul 2017 • Minmin Chen
The simple model architecture introduced by Doc2VecC matches or out-performs the state-of-the-art in generating high-quality document representations for sentiment analysis, document classification as well as semantic relatedness tasks.
Ranked #5 on Semantic Similarity on SICK
no code implementations • 27 Feb 2014 • Laurens van der Maaten, Minmin Chen, Stephen Tyree, Kilian Weinberger
In this paper, we propose a third, alternative approach to combat overfitting: we extend the training set with infinitely many artificial training examples that are obtained by corrupting the original training data.
no code implementations • 9 Oct 2012 • Zhixiang Xu, Matt J. Kusner, Kilian Q. Weinberger, Minmin Chen
Recently, machine learning algorithms have successfully entered large-scale real-world industrial applications (e. g. search engines and email spam filters).
no code implementations • NeurIPS 2011 • Minmin Chen, Kilian Q. Weinberger, John Blitzer
Our algorithm is a variant of co-training, and we name it CODA (Co-training for domain adaptation).