no code implementations • 7 Mar 2024 • Savvas Petridis, Ben Wedin, Ann Yuan, James Wexler, Nithum Thain
We also show that we can improve overall performance by learning unique prompts for different semantic regions of the training data and using a mixture-of-experts (MoE) architecture to route inputs at inference time.
no code implementations • 24 Oct 2023 • Savvas Petridis, Ben Wedin, James Wexler, Aaron Donsbach, Mahima Pushkarna, Nitesh Goyal, Carrie J. Cai, Michael Terry
Inspired by these findings, we developed ConstitutionMaker, an interactive tool for converting user feedback into principles, to steer LLM-based chatbots.
no code implementations • 26 Jul 2023 • Scott Sanner, Krisztian Balog, Filip Radlinski, Ben Wedin, Lucas Dixon
Inspired by recent successes of prompting paradigms for large language models (LLMs), we study their use for making recommendations from both item-based and language-based preferences in comparison to state-of-the-art item-based collaborative filtering (CF) methods.
no code implementations • 19 May 2022 • Filip Radlinski, Krisztian Balog, Fernando Diaz, Lucas Dixon, Ben Wedin
Natural interaction with recommendation and personalized search systems has received tremendous attention in recent years.
no code implementations • 26 Jan 2022 • Eldon Schoop, Ben Wedin, Andrei Kapishnikov, Tolga Bolukbasi, Michael Terry
Developing a suitable Deep Neural Network (DNN) often requires significant iteration, where different model versions are evaluated and compared.
1 code implementation • CVPR 2021 • Andrei Kapishnikov, Subhashini Venugopalan, Besim Avci, Ben Wedin, Michael Terry, Tolga Bolukbasi
To minimize the effect of this source of noise, we propose adapting the attribution path itself -- conditioning the path not just on the image but also on the model being explained.