1 code implementation • 21 Jun 2023 • Alessandro Sordoni, Xingdi Yuan, Marc-Alexandre Côté, Matheus Pereira, Adam Trischler, Ziang Xiao, Arian Hosseini, Friederike Niedtner, Nicolas Le Roux
We call the stacked architecture a \emph{Deep Language Network} (DLN).
no code implementations • 21 May 2023 • Cédric Colas, Laetitia Teodorescu, Pierre-Yves Oudeyer, Xingdi Yuan, Marc-Alexandre Côté
Without relying on any hand-coded goal representations, reward functions or curriculum, we show that LMA3 agents learn to master a large diversity of skills in a task-agnostic text-based environment.
no code implementations • 15 May 2023 • Arjun Subramonian, Xingdi Yuan, Hal Daumé III, Su Lin Blodgett
Progress in NLP is increasingly measured through benchmarks; hence, contextualizing progress requires understanding when and why practitioners may disagree about the validity of benchmarks.
no code implementations • 17 Apr 2023 • Ziang Xiao, Xingdi Yuan, Q. Vera Liao, Rania Abdelghani, Pierre-Yves Oudeyer
In this study, we explored the use of large language models (LLMs) in supporting deductive coding, a major category of qualitative analysis where researchers use pre-determined codebooks to label the data into a fixed set of codes.
no code implementations • 10 Feb 2023 • Laetitia Teodorescu, Xingdi Yuan, Marc-Alexandre Côté, Pierre-Yves Oudeyer
We show the importance of selectivity from the social peer's feedback; that experience replay needs to over-sample examples of rare goals; and that following self-generated goal sequences where the agent's competence is intermediate leads to significant improvements in final performance.
no code implementations • 25 Nov 2022 • Rania Abdelghani, Yen-Hsiang Wang, Xingdi Yuan, Tong Wang, Pauline Lucas, Hélène Sauzéon, Pierre-Yves Oudeyer
In this context, we propose to leverage advances in the natural language processing field (NLP) and investigate the efficiency of using a large language model (LLM) for automating the production of the pedagogical content of a curious question-asking (QA) training.
no code implementations • 22 Sep 2022 • Xingdi Yuan, Tong Wang, Yen-Hsiang Wang, Emery Fine, Rania Abdelghani, Pauline Lucas, Hélène Sauzéon, Pierre-Yves Oudeyer
Large Language Models (LLMs) have in recent years demonstrated impressive prowess in natural language generation.
1 code implementation • 20 Aug 2022 • Rui Meng, Tong Wang, Xingdi Yuan, Yingbo Zhou, Daqing He
Finally, we fine-tune the model with limited data with true labels to fully adapt it to the target domain.
no code implementations • 12 May 2022 • Iou-Jen Liu, Xingdi Yuan, Marc-Alexandre Côté, Pierre-Yves Oudeyer, Alexander G. Schwing
In order to study how agents can be taught to query external knowledge via language, we first introduce two new environments: the grid-world-based Q-BabyAI and the text-based Q-TextWorld.
no code implementations • 9 Mar 2022 • Nathaniel Weir, Xingdi Yuan, Marc-Alexandre Côté, Matthew Hausknecht, Romain Laroche, Ida Momennejad, Harm van Seijen, Benjamin Van Durme
Humans have the capability, aided by the expressive compositionality of their language, to learn quickly by demonstration.
1 code implementation • EMNLP 2021 • Xingdi Yuan
Interactive machine reading comprehension (iMRC) is machine comprehension tasks where knowledge sources are partially observable.
1 code implementation • ACL 2021 • Rui Meng, Khushboo Thaker, Lei Zhang, Yue Dong, Xingdi Yuan, Tong Wang, Daqing He
Faceted summarization provides briefings of a document from different perspectives.
Ranked #1 on
Unsupervised Extractive Summarization
on FacetSum
no code implementations • ICLR 2021 • Mohit Shridhar, Xingdi Yuan, Marc-Alexandre Cote, Yonatan Bisk, Adam Trischler, Matthew Hausknecht
ALFWorld enables the creation of a new BUTLER agent whose abstract knowledge, learned in TextWorld, corresponds directly to concrete, visually grounded actions.
1 code implementation • 8 Oct 2020 • Mohit Shridhar, Xingdi Yuan, Marc-Alexandre Côté, Yonatan Bisk, Adam Trischler, Matthew Hausknecht
ALFWorld enables the creation of a new BUTLER agent whose abstract knowledge, learned in TextWorld, corresponds directly to concrete, visually grounded actions.
1 code implementation • NAACL 2021 • Rui Meng, Xingdi Yuan, Tong Wang, Sanqiang Zhao, Adam Trischler, Daqing He
Recent years have seen a flourishing of neural keyphrase generation (KPG) works, including the release of several large-scale datasets and a host of new models to tackle them.
1 code implementation • NeurIPS 2020 • Shengding Hu, Zheng Xiong, Meng Qu, Xingdi Yuan, Marc-Alexandre Côté, Zhiyuan Liu, Jian Tang
Graph neural networks (GNNs) have been attracting increasing popularity due to their simplicity and effectiveness in a variety of fields.
1 code implementation • ICLR 2020 • Jie Fu, Xue Geng, Zhijian Duan, Bohan Zhuang, Xingdi Yuan, Adam Trischler, Jie Lin, Chris Pal, Hao Dong
To our knowledge, existing methods overlook the fact that although the student absorbs extra knowledge from the teacher, both models share the same input data -- and this data is the only medium by which the teacher's knowledge can be demonstrated.
1 code implementation • NeurIPS 2020 • Ashutosh Adhikari, Xingdi Yuan, Marc-Alexandre Côté, Mikuláš Zelinka, Marc-Antoine Rondeau, Romain Laroche, Pascal Poupart, Jian Tang, Adam Trischler, William L. Hamilton
Playing text-based games requires skills in processing natural language and sequential decision making.
1 code implementation • 21 Oct 2019 • Mikuláš Zelinka, Xingdi Yuan, Marc-Alexandre Côté, Romain Laroche, Adam Trischler
We are interested in learning how to update Knowledge Graphs (KG) from text.
2 code implementations • 11 Sep 2019 • Matthew Hausknecht, Prithviraj Ammanabrolu, Marc-Alexandre Côté, Xingdi Yuan
A hallmark of human intelligence is the ability to understand and communicate with language.
1 code implementation • 9 Sep 2019 • Rui Meng, Xingdi Yuan, Tong Wang, Peter Brusilovsky, Adam Trischler, Daqing He
Recently, concatenating multiple keyphrases as a target sequence has been proposed as a new learning paradigm for keyphrase generation.
1 code implementation • IJCNLP 2019 • Xingdi Yuan, Marc-Alexandre Cote, Jie Fu, Zhouhan Lin, Christopher Pal, Yoshua Bengio, Adam Trischler
In QAit, an agent must interact with a partially observable text-based environment to gather information required to answer questions.
1 code implementation • ACL 2020 • Xingdi Yuan, Jie Fu, Marc-Alexandre Cote, Yi Tay, Christopher Pal, Adam Trischler
Existing machine reading comprehension (MRC) models do not scale effectively to real-world applications like web-level information retrieval and question answering (QA).
no code implementations • ACL 2019 • Yi Tay, Shuohang Wang, Luu Anh Tuan, Jie Fu, Minh C. Phan, Xingdi Yuan, Jinfeng Rao, Siu Cheung Hui, Aston Zhang
This paper tackles the problem of reading comprehension over long narratives where documents easily span over thousands of tokens.
2 code implementations • 3 Dec 2018 • Ruo Yu Tao, Marc-Alexandre Côté, Xingdi Yuan, Layla El Asri
To solve a text-based game, an agent needs to formulate valid text commands for a given context and find the ones that lead to success.
no code implementations • ICLR 2019 • Rajarshi Das, Tsendsuren Munkhdalai, Xingdi Yuan, Adam Trischler, Andrew McCallum
We harness and extend a recently proposed machine reading comprehension (MRC) model to query for entity states, since these states are generally communicated in spans of text and MRC models perform well in extracting entity-centric spans.
Ranked #3 on
Procedural Text Understanding
on ProPara
1 code implementation • ACL 2020 • Xingdi Yuan, Tong Wang, Rui Meng, Khushboo Thaker, Peter Brusilovsky, Daqing He, Adam Trischler
With both previous and new evaluation metrics, our model outperforms strong baselines on all datasets.
no code implementations • WS 2018 • S Subramanian, eep, Tong Wang, Xingdi Yuan, Saizheng Zhang, Adam Trischler, Yoshua Bengio
We propose a two-stage neural model to tackle question generation from documents.
2 code implementations • 29 Jun 2018 • Xingdi Yuan, Marc-Alexandre Côté, Alessandro Sordoni, Romain Laroche, Remi Tachet des Combes, Matthew Hausknecht, Adam Trischler
We propose a recurrent RL agent with an episodic exploration mechanism that helps discovering good policies in text-based game environments.
1 code implementation • 29 Jun 2018 • Marc-Alexandre Côté, Ákos Kádár, Xingdi Yuan, Ben Kybartas, Tavian Barnes, Emery Fine, James Moore, Ruo Yu Tao, Matthew Hausknecht, Layla El Asri, Mahmoud Adada, Wendy Tay, Adam Trischler
We introduce TextWorld, a sandbox learning environment for the training and evaluation of RL agents on text-based games.
no code implementations • ICML 2018 • Tsendsuren Munkhdalai, Xingdi Yuan, Soroush Mehri, Adam Trischler
We describe a mechanism by which artificial neural networks can learn rapid adaptation - the ability to adapt on the fly, with little data, to new tasks - that we call conditionally shifted neurons.
no code implementations • 14 Jun 2017 • Sandeep Subramanian, Tong Wang, Xingdi Yuan, Saizheng Zhang, Yoshua Bengio, Adam Trischler
We propose a two-stage neural model to tackle question generation from documents.
no code implementations • 5 Jun 2017 • Tong Wang, Xingdi Yuan, Adam Trischler
We propose a generative machine comprehension model that learns jointly to ask and answer questions based on documents.
4 code implementations • WS 2017 • Xingdi Yuan, Tong Wang, Caglar Gulcehre, Alessandro Sordoni, Philip Bachman, Sandeep Subramanian, Saizheng Zhang, Adam Trischler
We propose a recurrent neural model that generates natural-language questions from documents, conditioned on answers.
1 code implementation • WS 2017 • Adam Trischler, Tong Wang, Xingdi Yuan, Justin Harris, Alessandro Sordoni, Philip Bachman, Kaheer Suleman
We present NewsQA, a challenging machine comprehension dataset of over 100, 000 human-generated question-answer pairs.
no code implementations • EMNLP 2016 • Adam Trischler, Zheng Ye, Xingdi Yuan, Kaheer Suleman
We present the EpiReader, a novel model for machine comprehension of text.
Ranked #7 on
Question Answering
on Children's Book Test
1 code implementation • ACL 2016 • Adam Trischler, Zheng Ye, Xingdi Yuan, Jing He, Phillip Bachman, Kaheer Suleman
The parallel hierarchy enables our model to compare the passage, question, and answer from a variety of trainable perspectives, as opposed to using a manually designed, rigid feature set.
Ranked #1 on
Question Answering
on MCTest-160