no code implementations • Dirk Weissenborn, Douwe Kiela, Jason Weston, Kyunghyun Cho
Word inputs tend to be represented as single continuous vectors in deep neural networks.
no code implementations • ICLR 2019 • Harm de Vries, Kurt Shuster, Dhruv Batra, Devi Parikh, Jason Weston, Douwe Kiela
We introduce `"Talk The Walk", the first large-scale dialogue dataset grounded in action and perception.
no code implementations • 12 Feb 2025 • Jihoon Tack, Jack Lanchantin, Jane Yu, Andrew Cohen, Ilia Kulikov, Janice Lan, Shibo Hao, Yuandong Tian, Jason Weston, Xian Li
We propose Continuous Concept Mixing (CoCoMix), a novel pretraining framework that combines discrete next token prediction with continuous concepts.
no code implementations • 30 Jan 2025 • Ping Yu, Weizhe Yuan, Olga Golovneva, Tianhao Wu, Sainbayar Sukhbaatar, Jason Weston, Jing Xu
Using Llama 3. 1-8B-Instruct, RIP improves AlpacaEval2 LC Win Rate by 9. 4%, Arena-Hard by 8. 7%, and WildBench by 9. 9%.
no code implementations • 30 Jan 2025 • Jack Lanchantin, Angelica Chen, Shehzaad Dhuliawala, Ping Yu, Jason Weston, Sainbayar Sukhbaatar, Ilia Kulikov
In this work we introduce Diverse Preference Optimization (DivPO), an optimization method which learns to generate much more diverse responses than standard pipelines, while maintaining the quality of the generations.
no code implementations • 30 Jan 2025 • Swarnadeep Saha, Xian Li, Marjan Ghazvininejad, Jason Weston, Tianlu Wang
LLM-as-a-Judge models generate chain-of-thought (CoT) sequences intended to capture the step-bystep reasoning process that underlies the final evaluation of a response.
no code implementations • 18 Jan 2025 • Yen-Ting Lin, Di Jin, Tengyu Xu, Tianhao Wu, Sainbayar Sukhbaatar, Chen Zhu, Yun He, Yun-Nung Chen, Jason Weston, Yuandong Tian, Arash Rahnama, Sinong Wang, Hao Ma, Han Fang
Large language models (LLMs) have recently demonstrated remarkable success in mathematical reasoning.
1 code implementation • 13 Dec 2024 • Artidoro Pagnoni, Ram Pasunuru, Pedro Rodriguez, John Nguyen, Benjamin Muller, Margaret Li, Chunting Zhou, Lili Yu, Jason Weston, Luke Zettlemoyer, Gargi Ghosh, Mike Lewis, Ari Holtzman, Srinivasan Iyer
We introduce the Byte Latent Transformer (BLT), a new byte-level LLM architecture that, for the first time, matches tokenization-based LLM performance at scale with significant improvements in inference efficiency and robustness.
1 code implementation • 9 Dec 2024 • Shibo Hao, Sainbayar Sukhbaatar, DiJia Su, Xian Li, Zhiting Hu, Jason Weston, Yuandong Tian
For example, most word tokens are primarily for textual coherence and not essential for reasoning, while some critical tokens require complex planning and pose huge challenges to LLMs.
no code implementations • 5 Dec 2024 • Michihiro Yasunaga, Leonid Shamis, Chunting Zhou, Andrew Cohen, Jason Weston, Luke Zettlemoyer, Marjan Ghazvininejad
Recent approaches to large language model (LLM) alignment typically require millions of human annotations or rely on external aligned models for synthetic data generation.
no code implementations • 14 Nov 2024 • Shehzaad Dhuliawala, Ilia Kulikov, Ping Yu, Asli Celikyilmaz, Jason Weston, Sainbayar Sukhbaatar, Jack Lanchantin
During language model decoding, it is known that using higher temperature sampling gives more creative responses, while lower temperatures are more factually accurate.
no code implementations • 6 Nov 2024 • Archiki Prasad, Weizhe Yuan, Richard Yuanzhe Pang, Jing Xu, Maryam Fazel-Zarandi, Mohit Bansal, Sainbayar Sukhbaatar, Jason Weston, Jane Yu
Self-alignment, whereby models learn to improve themselves without human annotation, is a rapidly growing research area.
no code implementations • 14 Oct 2024 • Tianhao Wu, Janice Lan, Weizhe Yuan, Jiantao Jiao, Jason Weston, Sainbayar Sukhbaatar
LLMs are typically trained to answer user questions or follow instructions similarly to how human experts respond.
no code implementations • 22 Sep 2024 • Yiming Zhang, Jianfeng Chi, Hailey Nguyen, Kartikeya Upasani, Daniel M. Bikel, Jason Weston, Eric Michael Smith
In the context of language model safety, when a partial unsafe generation is produced, language models by their nature tend to happily keep on generating similarly unsafe additional text.
no code implementations • 12 Sep 2024 • Alisia Lupidi, Carlos Gemmell, Nicola Cancedda, Jane Dwivedi-Yu, Jason Weston, Jakob Foerster, Roberta Raileanu, Maria Lomeli
Our method improves performance by 25. 51% for TQA on WikiSQL and 22. 57% for MHQA on HotPotQA compared to the fine-tuned baselines.
no code implementations • 8 Aug 2024 • Thao Nguyen, Jeffrey Li, Sewoong Oh, Ludwig Schmidt, Jason Weston, Luke Zettlemoyer, Xian Li
We propose a new method, instruction back-and-forth translation, to construct high-quality synthetic data grounded in world knowledge for aligning large language models (LLMs).
no code implementations • 5 Aug 2024 • Tianlu Wang, Ilia Kulikov, Olga Golovneva, Ping Yu, Weizhe Yuan, Jane Dwivedi-Yu, Richard Yuanzhe Pang, Maryam Fazel-Zarandi, Jason Weston, Xian Li
Model-based evaluation is at the heart of successful model development -- as a reward model for training, and as a replacement for human evaluation.
no code implementations • 28 Jul 2024 • Tianhao Wu, Weizhe Yuan, Olga Golovneva, Jing Xu, Yuandong Tian, Jiantao Jiao, Jason Weston, Sainbayar Sukhbaatar
Large Language Models (LLMs) are rapidly surpassing human knowledge in many domains.
no code implementations • 8 Jul 2024 • Ping Yu, Jing Xu, Jason Weston, Ilia Kulikov
Large language models (LLMs) can spend extra compute during inference to generate intermediate thoughts, which helps to produce better final responses.
no code implementations • 25 Jun 2024 • Weizhe Yuan, Ilia Kulikov, Ping Yu, Kyunghyun Cho, Sainbayar Sukhbaatar, Jason Weston, Jing Xu
Aligned instruction following models can better fulfill user requests than their unaligned counterparts.
no code implementations • 29 May 2024 • Olga Golovneva, Tianlu Wang, Jason Weston, Sainbayar Sukhbaatar
Incorporating position encoding (PE) makes it possible to address by position, such as attending to the i-th token.
no code implementations • 30 Apr 2024 • Richard Yuanzhe Pang, Weizhe Yuan, Kyunghyun Cho, He He, Sainbayar Sukhbaatar, Jason Weston
Iterative preference optimization methods have recently been shown to perform well for general instruction tuning tasks, but typically make little improvement on reasoning tasks (Yuan et al., 2024, Chen et al., 2024).
no code implementations • 20 Mar 2024 • Olga Golovneva, Zeyuan Allen-Zhu, Jason Weston, Sainbayar Sukhbaatar
Large language models (LLMs) have a surprising failure: when trained on "A has a feature B", they do not generalize to "B is a feature of A", which is termed the Reversal Curse.
1 code implementation • 12 Mar 2024 • Sainbayar Sukhbaatar, Olga Golovneva, Vasu Sharma, Hu Xu, Xi Victoria Lin, Baptiste Rozière, Jacob Kahn, Daniel Li, Wen-tau Yih, Jason Weston, Xian Li
We investigate efficient methods for training Large Language Models (LLMs) to possess capabilities in multiple specialized domains, such as coding, math reasoning and world knowledge.
Ranked #32 on
Multi-task Language Understanding
on MMLU
1 code implementation • 21 Feb 2024 • Dheeraj Mekala, Jason Weston, Jack Lanchantin, Roberta Raileanu, Maria Lomeli, Jingbo Shang, Jane Dwivedi-Yu
Teaching language models to use tools is an important milestone towards building general assistants, but remains an open problem.
3 code implementations • 18 Jan 2024 • Weizhe Yuan, Richard Yuanzhe Pang, Kyunghyun Cho, Xian Li, Sainbayar Sukhbaatar, Jing Xu, Jason Weston
We posit that to achieve superhuman agents, future models require superhuman feedback in order to provide an adequate training signal.
1 code implementation • 27 Dec 2023 • Jing Xu, Andrew Lee, Sainbayar Sukhbaatar, Jason Weston
Practitioners commonly align large language models using pairwise preferences, i. e., given labels of the type response A is preferred to response B for a given input.
1 code implementation • 20 Nov 2023 • Jason Weston, Sainbayar Sukhbaatar
Soft attention in Transformer-based Large Language Models (LLMs) is susceptible to incorporating irrelevant information from the context into its latent representations, which adversely affects next token generations.
no code implementations • 14 Nov 2023 • Kumar Shridhar, Koustuv Sinha, Andrew Cohen, Tianlu Wang, Ping Yu, Ram Pasunuru, Mrinmaya Sachan, Jason Weston, Asli Celikyilmaz
In recent years, Large Language Models (LLMs) have demonstrated remarkable generative abilities, but can they judge the quality of their own generations?
Ranked #55 on
Arithmetic Reasoning
on GSM8K
no code implementations • 23 Oct 2023 • Swarnadeep Saha, Omer Levy, Asli Celikyilmaz, Mohit Bansal, Jason Weston, Xian Li
Large Language Models (LLMs) are frequently used for multi-faceted language generation and evaluation tasks that involve satisfying intricate user constraints or taking into account multiple aspects and criteria.
no code implementations • 8 Oct 2023 • Howard Chen, Ramakanth Pasunuru, Jason Weston, Asli Celikyilmaz
Large language models (LLMs) have advanced in large strides due to the effectiveness of the self-attention mechanism that processes and compares all tokens at once.
1 code implementation • 20 Sep 2023 • Shehzaad Dhuliawala, Mojtaba Komeili, Jing Xu, Roberta Raileanu, Xian Li, Asli Celikyilmaz, Jason Weston
Generation of plausible yet incorrect factual information, termed hallucination, is an unsolved issue in large language models.
2 code implementations • 11 Aug 2023 • Xian Li, Ping Yu, Chunting Zhou, Timo Schick, Omer Levy, Luke Zettlemoyer, Jason Weston, Mike Lewis
We present a scalable method to build a high quality instruction following language model by automatically labelling human-written text with corresponding instructions.
no code implementations • 26 Jul 2023 • Richard Yuanzhe Pang, Stephen Roller, Kyunghyun Cho, He He, Jason Weston
We study improving social conversational agents by learning from natural dialogue between users and a deployed model, without extra annotations.
1 code implementation • 23 Jun 2023 • Weizhe Yuan, Kyunghyun Cho, Jason Weston
Natural language (NL) feedback offers rich insights into user experience.
no code implementations • 7 Jun 2023 • Jing Xu, Da Ju, Joshua Lane, Mojtaba Komeili, Eric Michael Smith, Megan Ung, Morteza Behrooz, William Ngan, Rashel Moritz, Sainbayar Sukhbaatar, Y-Lan Boureau, Jason Weston, Kurt Shuster
We present BlenderBot 3x, an update on the conversational model BlenderBot 3, which is now trained using organic conversation and feedback data from participating users of the system in order to improve both its skills and safety.
no code implementations • 7 Jun 2023 • Morteza Behrooz, William Ngan, Joshua Lane, Giuliano Morse, Benjamin Babcock, Kurt Shuster, Mojtaba Komeili, Moya Chen, Melanie Kambadur, Y-Lan Boureau, Jason Weston
Publicly deploying research chatbots is a nuanced topic involving necessary risk-benefit analyses.
no code implementations • 9 May 2023 • Imanol Schlag, Sainbayar Sukhbaatar, Asli Celikyilmaz, Wen-tau Yih, Jason Weston, Jürgen Schmidhuber, Xian Li
In recent years, large pre-trained language models (LLMs) have demonstrated the ability to follow instructions and perform novel tasks from a few examples.
no code implementations • 26 Apr 2023 • Jimmy Wei, Kurt Shuster, Arthur Szlam, Jason Weston, Jack Urbanek, Mojtaba Komeili
We compare models trained on our new dataset to existing pairwise-trained dialogue models, as well as large language models with few-shot prompting.
no code implementations • 14 Feb 2023 • Kushal Arora, Timothy J. O'Donnell, Doina Precup, Jason Weston, Jackie C. K. Cheung
State-of-the-art language generation models can degenerate when applied to open-ended generation problems such as text completion, story generation, or dialog modeling.
no code implementations • 13 Jan 2023 • Alexander Gurung, Mojtaba Komeili, Arthur Szlam, Jason Weston, Jack Urbanek
While language models have become more capable of producing compelling language, we find there are still gaps in maintaining consistency, especially when describing events in a dynamically changing world.
no code implementations • 10 Nov 2022 • Leonard Adolphs, Tianyu Gao, Jing Xu, Kurt Shuster, Sainbayar Sukhbaatar, Jason Weston
Standard language model training employs gold human documents or human-human interaction data, and treats all training data as positive examples.
no code implementations • 28 Oct 2022 • Weiyan Shi, Emily Dinan, Kurt Shuster, Jason Weston, Jing Xu
Deployed dialogue agents have the potential to integrate human feedback to continuously improve themselves.
no code implementations • 5 Aug 2022 • Jing Xu, Megan Ung, Mojtaba Komeili, Kushal Arora, Y-Lan Boureau, Jason Weston
We then study various algorithms for improving from such feedback, including standard supervised learning, rejection sampling, model-guiding and reward-based learning, in order to make recommendations on which type of feedback and algorithms work best.
no code implementations • 5 Aug 2022 • Da Ju, Jing Xu, Y-Lan Boureau, Jason Weston
The promise of interaction between intelligent conversational agents and humans is that models can learn from such feedback in order to improve.
5 code implementations • 5 Aug 2022 • Kurt Shuster, Jing Xu, Mojtaba Komeili, Da Ju, Eric Michael Smith, Stephen Roller, Megan Ung, Moya Chen, Kushal Arora, Joshua Lane, Morteza Behrooz, William Ngan, Spencer Poff, Naman Goyal, Arthur Szlam, Y-Lan Boureau, Melanie Kambadur, Jason Weston
We present BlenderBot 3, a 175B parameter dialogue model capable of open-domain conversation with access to the internet and a long-term memory, and having been trained on a large number of user defined tasks.
1 code implementation • 15 Jun 2022 • Kushal Arora, Kurt Shuster, Sainbayar Sukhbaatar, Jason Weston
Current language models achieve low perplexity but their resulting generations still suffer from toxic responses, repetitiveness and contradictions.
1 code implementation • 24 Mar 2022 • Kurt Shuster, Mojtaba Komeili, Leonard Adolphs, Stephen Roller, Arthur Szlam, Jason Weston
We show that, when using SeeKeR as a dialogue model, it outperforms the state-of-the-art model BlenderBot 2 (Chen et al., 2021) on open-domain knowledge-grounded conversations for the same number of parameters, in terms of consistency, knowledge and per-turn engagingness.
no code implementations • NLP4ConvAI (ACL) 2022 • Eric Michael Smith, Orion Hsu, Rebecca Qian, Stephen Roller, Y-Lan Boureau, Jason Weston
At the heart of improving conversational AI is the open problem of how to evaluate conversations.
no code implementations • Findings (NAACL) 2022 • Kurt Shuster, Jack Urbanek, Arthur Szlam, Jason Weston
State-of-the-art dialogue models still often stumble with regards to factual accuracy and self-contradiction.
no code implementations • 9 Nov 2021 • Leonard Adolphs, Kurt Shuster, Jack Urbanek, Arthur Szlam, Jason Weston
Large language models can produce fluent dialogue but often hallucinate factual inaccuracies.
1 code implementation • 18 Oct 2021 • Sam Shleifer, Jason Weston, Myle Ott
The extra operations incur negligible compute cost (+0. 4% parameter increase), but improve pretraining perplexity and downstream task performance for both causal and masked language models ranging from 125 Million to 2. 7 Billion parameters.
no code implementations • ACL 2022 • Jing Xu, Arthur Szlam, Jason Weston
Despite recent improvements in open-domain dialogue models, state of the art models are trained and evaluated on short conversations with little context.
no code implementations • ACL 2022 • Mojtaba Komeili, Kurt Shuster, Jason Weston
The largest store of continually updating knowledge on our planet can be accessed via internet search.
no code implementations • NeurIPS 2021 • Stephen Roller, Sainbayar Sukhbaatar, Arthur Szlam, Jason Weston
We investigate the training of sparse layers that use different parameters for different inputs based on hashing in large Transformer models.
1 code implementation • 8 Jun 2021 • Da Ju, Stephen Roller, Sainbayar Sukhbaatar, Jason Weston
Attention mechanisms have become a standard tool for sequence modeling tasks, in particular by stacking self-attention layers over the entire input sequence as in the Transformer architecture.
no code implementations • NAACL 2021 • Jing Xu, Da Ju, Margaret Li, Y-Lan Boureau, Jason Weston, Emily Dinan
Conversational agents trained on large unlabeled corpora of human interactions will learn patterns and mimic behaviors therein, which include offensive or otherwise toxic behavior.
1 code implementation • 13 May 2021 • Sainbayar Sukhbaatar, Da Ju, Spencer Poff, Stephen Roller, Arthur Szlam, Jason Weston, Angela Fan
We demonstrate that Expire-Span can help models identify and retain critical information and show it can achieve strong performance on reinforcement learning tasks specifically designed to challenge this functionality.
Ranked #4 on
Language Modelling
on enwik8
no code implementations • Findings (EMNLP) 2021 • Kurt Shuster, Spencer Poff, Moya Chen, Douwe Kiela, Jason Weston
Despite showing increasingly human-like conversational abilities, state-of-the-art dialogue models often suffer from factual incorrectness and hallucination of knowledge (Roller et al., 2020).
no code implementations • ACL 2021 • Yixin Nie, Mary Williamson, Mohit Bansal, Douwe Kiela, Jason Weston
To quantify how well natural language understanding models can capture consistency in a general conversation, we introduce the DialoguE COntradiction DEtection task (DECODE) and a new conversational dataset containing both human-human and human-bot contradictory dialogues.
no code implementations • 14 Oct 2020 • Jing Xu, Da Ju, Margaret Li, Y-Lan Boureau, Jason Weston, Emily Dinan
Models trained on large unlabeled corpora of human interactions will learn patterns and mimic behaviors therein, which include offensive or otherwise toxic behavior and unwanted biases.
no code implementations • EMNLP 2021 • Kurt Shuster, Eric Michael Smith, Da Ju, Jason Weston
Recent work in open-domain conversational agents has demonstrated that significant improvements in model engagingness and humanness metrics can be achieved via massive scaling in both pre-training data and model size (Adiwardana et al., 2020; Roller et al., 2020).
Ranked #1 on
Visual Dialog
on Wizard of Wikipedia
no code implementations • NAACL 2021 • Prithviraj Ammanabrolu, Jack Urbanek, Margaret Li, Arthur Szlam, Tim Rocktäschel, Jason Weston
We seek to create agents that both act and communicate with other agents in pursuit of a goal.
no code implementations • 18 Aug 2020 • Kurt Shuster, Jack Urbanek, Emily Dinan, Arthur Szlam, Jason Weston
As argued in de Vries et al. (2020), crowdsourced data has the issues of lack of naturalness and relevance to real-world use cases, while the static dataset paradigm does not allow for a model to learn from its experiences of using language (Silver et al., 2013).
no code implementations • ACL 2020 • Kurt Shuster, Samuel Humeau, Antoine Bordes, Jason Weston
To test such models, we collect a dataset of grounded human-human conversations, where speakers are asked to play roles given a provided emotional mood or style, as the use of such traits is also a key factor in engagingness (Guo et al., 2019).
no code implementations • 22 Jun 2020 • Stephen Roller, Y-Lan Boureau, Jason Weston, Antoine Bordes, Emily Dinan, Angela Fan, David Gunning, Da Ju, Margaret Li, Spencer Poff, Pratik Ringshia, Kurt Shuster, Eric Michael Smith, Arthur Szlam, Jack Urbanek, Mary Williamson
We present our view of what is necessary to build an engaging open-domain conversational agent: covering the qualities of such an agent, the pieces of the puzzle that have been built so far, and the gaping holes we have not filled yet.
no code implementations • EMNLP 2020 • Emily Dinan, Angela Fan, Ledell Wu, Jason Weston, Douwe Kiela, Adina Williams
We show our classifiers prove valuable for a variety of important applications, such as controlling for gender bias in generative models, detecting gender bias in arbitrary text, and shed light on offensive language in terms of genderedness.
8 code implementations • EACL 2021 • Stephen Roller, Emily Dinan, Naman Goyal, Da Ju, Mary Williamson, Yinhan Liu, Jing Xu, Myle Ott, Kurt Shuster, Eric M. Smith, Y-Lan Boureau, Jason Weston
Building open-domain chatbots is a challenging area for machine learning research.
2 code implementations • ACL 2020 • Eric Michael Smith, Mary Williamson, Kurt Shuster, Jason Weston, Y-Lan Boureau
Being engaging, knowledgeable, and empathetic are all desirable general qualities in a conversational agent.
no code implementations • 7 Feb 2020 • Shrimai Prabhumoye, Margaret Li, Jack Urbanek, Emily Dinan, Douwe Kiela, Jason Weston, Arthur Szlam
Dialogue research tends to distinguish between chit-chat and goal-oriented tasks.
2 code implementations • ICLR 2020 • Samuel Humeau, Kurt Shuster, Marie-Anne Lachaux, Jason Weston
The use of deep pre-trained transformers has led to remarkable progress in a number of applications (Devlin et al., 2018).
no code implementations • 28 Dec 2019 • Da Ju, Kurt Shuster, Y-Lan Boureau, Jason Weston
As single-task accuracy on individual language and image tasks has improved substantially in the last few years, the long-term goal of a generally skilled agent that can both see and talk becomes more feasible to explore.
no code implementations • 21 Nov 2019 • Xinyi Wang, Jason Weston, Michael Auli, Yacine Jernite
Neural sequence to sequence models are well established for applications which can be cast as mapping a single input sequence into a single output sequence.
Ranked #6 on
Open-Domain Question Answering
on ELI5
no code implementations • 20 Nov 2019 • Angela Fan, Jack Urbanek, Pratik Ringshia, Emily Dinan, Emma Qian, Siddharth Karamcheti, Shrimai Prabhumoye, Douwe Kiela, Tim Rocktaschel, Arthur Szlam, Jason Weston
We show that the game environments created with our approach are cohesive, diverse, and preferred by human evaluators compared to other machine learning based world construction algorithms.
no code implementations • EMNLP 2020 • Emily Dinan, Angela Fan, Adina Williams, Jack Urbanek, Douwe Kiela, Jason Weston
Models often easily learn biases present in the training data, and their predictions directly reflect this bias.
1 code implementation • ACL 2020 • Margaret Li, Stephen Roller, Ilia Kulikov, Sean Welleck, Y-Lan Boureau, Kyunghyun Cho, Jason Weston
Generative dialogue models currently suffer from a number of problems which standard maximum likelihood training does not address.
no code implementations • ACL 2020 • Kurt Shuster, Da Ju, Stephen Roller, Emily Dinan, Y-Lan Boureau, Jason Weston
We introduce dodecaDialogue: a set of 12 tasks that measures if a conversational agent can communicate engagingly with personality and empathy, ask questions, answer questions by utilizing knowledge resources, discuss topics and situations, and perceive and converse about images.
no code implementations • IJCNLP 2019 • Ethan Perez, Siddharth Karamcheti, Rob Fergus, Jason Weston, Douwe Kiela, Kyunghyun Cho
We propose a system that finds the strongest supporting evidence for a given answer to a question, using passage-based question-answering (QA) as a testbed.
2 code implementations • ACL 2020 • Yixin Nie, Adina Williams, Emily Dinan, Mohit Bansal, Jason Weston, Douwe Kiela
We introduce a new large-scale NLI benchmark dataset, collected via an iterative, adversarial human-and-model-in-the-loop procedure.
1 code implementation • 12 Sep 2019 • Ethan Perez, Siddharth Karamcheti, Rob Fergus, Jason Weston, Douwe Kiela, Kyunghyun Cho
We propose a system that finds the strongest supporting evidence for a given answer to a question, using passage-based question-answering (QA) as a testbed.
1 code implementation • IJCNLP 2019 • Dongyeop Kang, Anusha Balakrishnan, Pararth Shah, Paul Crook, Y-Lan Boureau, Jason Weston
These issues can be alleviated by treating recommendation as an interactive dialogue task instead, where an expert recommender can sequentially ask about someone's preferences, react to their requests, and recommend more appropriate items.
no code implementations • 6 Sep 2019 • Margaret Li, Jason Weston, Stephen Roller
While dialogue remains an important end-goal of natural language research, the difficulty of evaluation is an oft-quoted reason why it remains troublesome to make real progress towards its solution.
no code implementations • IJCNLP 2019 • Emily Dinan, Samuel Humeau, Bharath Chintagunta, Jason Weston
The detection of offensive language in the context of a dialogue has become an increasingly important application of natural language processing.
6 code implementations • ICLR 2020 • Sean Welleck, Ilia Kulikov, Stephen Roller, Emily Dinan, Kyunghyun Cho, Jason Weston
Neural text generation is a key tool in natural language applications, but it is well known there are major problems at its core.
3 code implementations • ACL 2019 • Angela Fan, Yacine Jernite, Ethan Perez, David Grangier, Jason Weston, Michael Auli
We introduce the first large-scale corpus for long-form question answering, a task requiring elaborate and in-depth answers to open-ended questions.
1 code implementation • 22 Jul 2019 • Arthur Szlam, Jonathan Gray, Kavya Srinet, Yacine Jernite, Armand Joulin, Gabriel Synnaeve, Douwe Kiela, Haonan Yu, Zhuoyuan Chen, Siddharth Goyal, Demi Guo, Danielle Rothermel, C. Lawrence Zitnick, Jason Weston
In this document we describe a rationale for a research program aimed at building an open "assistant" in the game Minecraft, in order to make progress on the problems of natural language understanding and learning from dialogue.
7 code implementations • 22 Apr 2019 • Samuel Humeau, Kurt Shuster, Marie-Anne Lachaux, Jason Weston
The use of deep pre-trained bidirectional transformers has led to remarkable progress in a number of applications (Devlin et al., 2018).
1 code implementation • IJCNLP 2019 • Jack Urbanek, Angela Fan, Siddharth Karamcheti, Saachi Jain, Samuel Humeau, Emily Dinan, Tim Rocktäschel, Douwe Kiela, Arthur Szlam, Jason Weston
We analyze the ingredients necessary for successful grounding in this setting, and how each of these factors relate to agents that can talk and act successfully.
2 code implementations • NAACL 2019 • Abigail See, Stephen Roller, Douwe Kiela, Jason Weston
A good conversation requires balance -- between simplicity and detail; staying on topic and changing it; asking questions and answering them.
2 code implementations • 31 Jan 2019 • Emily Dinan, Varvara Logacheva, Valentin Malykh, Alexander Miller, Kurt Shuster, Jack Urbanek, Douwe Kiela, Arthur Szlam, Iulian Serban, Ryan Lowe, Shrimai Prabhumoye, Alan W. black, Alexander Rudnicky, Jason Williams, Joelle Pineau, Mikhail Burtsev, Jason Weston
We describe the setting and results of the ConvAI2 NeurIPS competition that aims to further the state-of-the-art in open-domain chatbots.
1 code implementation • ACL 2019 • Braden Hancock, Antoine Bordes, Pierre-Emmanuel Mazaré, Jason Weston
As our agent engages in conversation, it also estimates user satisfaction in its responses.
2 code implementations • ICLR 2019 • Emily Dinan, Stephen Roller, Kurt Shuster, Angela Fan, Michael Auli, Jason Weston
In open-domain dialogue intelligent agents should exhibit the use of knowledge, however there are few convincing demonstrations of this to date.
1 code implementation • WS 2019 • Ilia Kulikov, Alexander H. Miller, Kyunghyun Cho, Jason Weston
We investigate the impact of search strategies in neural dialogue modeling.
3 code implementations • 2 Nov 2018 • Kurt Shuster, Samuel Humeau, Antoine Bordes, Jason Weston
To test such models, we collect a dataset of grounded human-human conversations, where speakers are asked to play roles given a provided emotional mood or style, as the use of such traits is also a key factor in engagingness (Guo et al., 2019).
Ranked #2 on
Text Retrieval
on Image-Chat
no code implementations • ACL 2019 • Sean Welleck, Jason Weston, Arthur Szlam, Kyunghyun Cho
Consistency is a long standing issue faced by dialogue models.
no code implementations • CVPR 2019 • Kurt Shuster, Samuel Humeau, Hexiang Hu, Antoine Bordes, Jason Weston
While such tasks are useful to verify that a machine understands the content of an image, they are not engaging to humans as captions.
1 code implementation • WS 2018 • Jasmijn Bastings, Marco Baroni, Jason Weston, Kyunghyun Cho, Douwe Kiela
Lake and Baroni (2018) recently introduced the SCAN data set, which consists of simple commands paired with action sequences and is intended to test the strong generalization abilities of recurrent sequence-to-sequence models.
1 code implementation • WS 2018 • Jason Weston, Emily Dinan, Alexander H. Miller
Sequence generation models for dialogue are known to have several problems: they tend to produce short, generic sentences that are uninformative and unengaging.
1 code implementation • 9 Jul 2018 • Harm de Vries, Kurt Shuster, Dhruv Batra, Devi Parikh, Jason Weston, Douwe Kiela
We introduce "Talk The Walk", the first large-scale dialogue dataset grounded in action and perception.
no code implementations • 19 Apr 2018 • Cinjon Resnick, Ilya Kulikov, Kyunghyun Cho, Jason Weston
Interest in emergent communication has recently surged in Machine Learning.
15 code implementations • ACL 2018 • Saizheng Zhang, Emily Dinan, Jack Urbanek, Arthur Szlam, Douwe Kiela, Jason Weston
Chit-chat models are known to have several problems: they lack specificity, do not display a consistent personality and are often not very captivating.
Ranked #5 on
Dialogue Generation
on Persona-Chat
(using extra training data)
no code implementations • ICLR 2018 • Zhilin Yang, Saizheng Zhang, Jack Urbanek, Will Feng, Alexander H. Miller, Arthur Szlam, Douwe Kiela, Jason Weston
Contrary to most natural language processing research, which makes use of static datasets, humans learn language interactively, grounded in an environment.
no code implementations • ICLR 2018 • Jason Lee, Kyunghyun Cho, Jason Weston, Douwe Kiela
While most machine translation systems to date are trained on large parallel corpora, humans learn language in a different way: by being grounded in an environment and interacting with other humans.
3 code implementations • 12 Sep 2017 • Ledell Wu, Adam Fisch, Sumit Chopra, Keith Adams, Antoine Bordes, Jason Weston
A framework for training and evaluating AI models on a variety of openly available dialogue datasets.
22 code implementations • EMNLP 2017 • Alexander H. Miller, Will Feng, Adam Fisch, Jiasen Lu, Dhruv Batra, Antoine Bordes, Devi Parikh, Jason Weston
We introduce ParlAI (pronounced "par-lay"), an open-source software platform for dialog research implemented in Python, available at http://parl. ai.
10 code implementations • ACL 2017 • Danqi Chen, Adam Fisch, Jason Weston, Antoine Bordes
This paper proposes to tackle open- domain question answering using Wikipedia as the unique knowledge source: the answer to any factoid question is a text span in a Wikipedia article.
Ranked #1 on
Open-Domain Question Answering
on SQuAD1.1
2 code implementations • 15 Dec 2016 • Jiwei Li, Alexander H. Miller, Sumit Chopra, Marc'Aurelio Ranzato, Jason Weston
A good dialogue agent should have the ability to interact with users by both responding to questions and by asking questions, and importantly to learn from both types of interaction.
5 code implementations • 12 Dec 2016 • Mikael Henaff, Jason Weston, Arthur Szlam, Antoine Bordes, Yann Lecun
The EntNet sets a new state-of-the-art on the bAbI tasks, and is the first method to solve all the tasks in the 10k training examples setting.
Ranked #5 on
Procedural Text Understanding
on ProPara
2 code implementations • 29 Nov 2016 • Jiwei Li, Alexander H. Miller, Sumit Chopra, Marc'Aurelio Ranzato, Jason Weston
An important aspect of developing conversational agents is to give a bot the ability to improve through communicating with humans and to learn from the mistakes that it makes.
2 code implementations • EMNLP 2016 • Alexander Miller, Adam Fisch, Jesse Dodge, Amir-Hossein Karimi, Antoine Bordes, Jason Weston
Directly reading documents and being able to answer questions from them is an unsolved challenge.
Ranked #12 on
Question Answering
on WikiQA
6 code implementations • 24 May 2016 • Antoine Bordes, Y-Lan Boureau, Jason Weston
We show similar result patterns on data extracted from an online concierge service.
no code implementations • NeurIPS 2016 • Jason Weston
A long-term goal of machine learning research is to build an intelligent dialog agent.
1 code implementation • 21 Nov 2015 • Jesse Dodge, Andreea Gane, Xiang Zhang, Antoine Bordes, Sumit Chopra, Alexander Miller, Arthur Szlam, Jason Weston
A long-term goal of machine learning is to build intelligent conversational agents.
3 code implementations • 7 Nov 2015 • Felix Hill, Antoine Bordes, Sumit Chopra, Jason Weston
We introduce a new test of how well language models capture meaning in children's books.
4 code implementations • EMNLP 2015 • Alexander M. Rush, Sumit Chopra, Jason Weston
Summarization based on text extraction is inherently limited, but generation-style abstractive methods have proven challenging to build.
Ranked #1 on
Extractive Text Summarization
on DUC 2004 Task 1
3 code implementations • 5 Jun 2015 • Antoine Bordes, Nicolas Usunier, Sumit Chopra, Jason Weston
Training large-scale question answering systems is complicated because training sources usually cover a small portion of the range of possible questions.
Ranked #1 on
Question Answering
on WebQuestions
(F1 metric)
44 code implementations • NeurIPS 2015 • Sainbayar Sukhbaatar, Arthur Szlam, Jason Weston, Rob Fergus
For the former our approach is competitive with Memory Networks, but with less supervision.
Ranked #6 on
Question Answering
on bAbi
20 code implementations • 19 Feb 2015 • Jason Weston, Antoine Bordes, Sumit Chopra, Alexander M. Rush, Bart van Merriënboer, Armand Joulin, Tomas Mikolov
One long-term goal of machine learning research is to produce methods that are applicable to reasoning and natural language, in particular building an intelligent dialogue agent.
5 code implementations • 15 Oct 2014 • Jason Weston, Sumit Chopra, Antoine Bordes
We describe a new class of learning models called memory networks.
1 code implementation • EMNLP 2014 • Antoine Bordes, Sumit Chopra, Jason Weston
Training our system using pairs of questions and structured representations of their answers, and pairs of question paraphrases, yields competitive results on a competitive benchmark of the literature.
Ranked #2 on
Question Answering
on WebQuestions
(F1 metric)
no code implementations • 16 Apr 2014 • Antoine Bordes, Jason Weston, Nicolas Usunier
Building computers able to answer questions on any subject is a long standing goal of artificial intelligence.
Ranked #1 on
Question Answering
on Reverb
8 code implementations • NeurIPS 2013 • Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, Oksana Yakhnenko
We consider the problem of embedding entities and relationships of multi-relational data in low-dimensional vector spaces.
Ranked #5 on
Link Prediction
on FB122
no code implementations • EMNLP 2013 • Jason Weston, Antoine Bordes, Oksana Yakhnenko, Nicolas Usunier
This paper proposes a novel approach for relation extraction from free text which is trained to jointly use information from the text and from existing knowledge.
no code implementations • 26 Apr 2013 • Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, Oksana Yakhnenko
We consider the problem of embedding entities and relations of knowledge bases in low-dimensional vector spaces.
no code implementations • 15 Jan 2013 • Xavier Glorot, Antoine Bordes, Jason Weston, Yoshua Bengio
Large-scale relational learning becomes crucial for handling the huge amounts of structured data generated daily in many application domains ranging from computational biology or information retrieval, to natural language processing.
no code implementations • 26 May 2011 • Jason Weston, Samy Bengio, Philippe Hamel
Music prediction tasks range from predicting tags given a song or clip of audio, predicting the name of the artist, or predicting related songs given a song, clip, artist name or tag.
1 code implementation • 2 Mar 2011 • Ronan Collobert, Jason Weston, Leon Bottou, Michael Karlen, Koray Kavukcuoglu, Pavel Kuksa
We propose a unified neural network architecture and learning algorithm that can be applied to various natural language processing tasks including: part-of-speech tagging, chunking, named entity recognition, and semantic role labeling.
no code implementations • NeurIPS 2010 • Samy Bengio, Jason Weston, David Grangier
Multi-class classification becomes challenging at test time when the number of classes is very large and testing against every possible class can become computationally infeasible.
no code implementations • NeurIPS 2009 • Bing Bai, Jason Weston, David Grangier, Ronan Collobert, Kunihiko Sadamasa, Yanjun Qi, Corinna Cortes, Mehryar Mohri
We present a class of nonlinear (polynomial) models that are discriminatively trained to directly map from the word content in a query-document or document-document pair to a ranking score.