no code implementations • NAACL (DeeLIO) 2021 • Vidhisha Balachandran, Bhuwan Dhingra, Haitian Sun, Michael Collins, William Cohen
We create a subset of the NQ data, Factual Questions (FQ), where the questions have evidence in the KB in the form of paths that link question entities to answer entities but still must be answered using text, to facilitate further research into KB integration methods.
no code implementations • 3 Mar 2024 • Rohan Kumar, Youngmin Kim, Sunitha Ravi, Haitian Sun, Christos Faloutsos, Ruslan Salakhutdinov, Minji Yoon
Pretrained Large Language Models (LLMs) have gained significant attention for addressing open-domain Question Answering (QA).
2 code implementations • 8 Nov 2023 • Tal Schuster, Adam D. Lelkes, Haitian Sun, Jai Gupta, Jonathan Berant, William W. Cohen, Donald Metzler
Experimenting with several LLMs in various settings, we find this task to be surprisingly challenging, demonstrating the importance of QuoteSum for developing and studying such consolidation capabilities.
no code implementations • 16 Aug 2023 • Haitian Sun, William W. Cohen, Ruslan Salakhutdinov
Many open-domain questions are under-specified and thus have multiple possible answers, each of which is correct under a different interpretation of the question.
2 code implementations • 23 Feb 2023 • Yang Chen, Hexiang Hu, Yi Luan, Haitian Sun, Soravit Changpinyo, Alan Ritter, Ming-Wei Chang
In this study, we introduce InfoSeek, a visual question answering dataset tailored for information-seeking questions that cannot be answered with only common sense knowledge.
Ranked #3 on Visual Question Answering (VQA) on InfoSeek
no code implementations • 25 May 2022 • Haitian Sun, William W. Cohen, Ruslan Salakhutdinov
Even more challenging, we only provide evidences for a subset of the conditions, so some questions may not have deterministic answers.
2 code implementations • ACL 2022 • Haitian Sun, William W. Cohen, Ruslan Salakhutdinov
In addition to conditional answers, the dataset also features: (1) long context documents with information that is related in logically complex ways; (2) multi-hop questions that require compositional logical reasoning; (3) a combination of extractive questions, yes/no questions, questions with multiple answers, and not-answerable questions; (4) questions asked without knowing the answers.
no code implementations • NAACL 2021 • Pat Verga, Haitian Sun, Livio Baldini Soares, William Cohen
Past research has demonstrated that large neural language models (LMs) encode surprising amounts of factual information: however, augmenting or modifying this information requires modifying a corpus and retraining, which is computationally expensive.
no code implementations • 1 Jun 2021 • Haitian Sun, William W. Cohen, Ruslan Salakhutdinov
We propose a new model, DocHopper, that iteratively attends to different parts of long, hierarchically structured documents to answer complex questions.
Ranked #2 on Question Answering on ConditionalQA
no code implementations • 14 Feb 2021 • Haitian Sun, Pat Verga, Bhuwan Dhingra, Ruslan Salakhutdinov, William W. Cohen
We present the Open Predicate Query Language (OPQL); a method for constructing a virtual KB (VKB) trained entirely from text.
no code implementations • NAACL 2021 • Bill Yuchen Lin, Haitian Sun, Bhuwan Dhingra, Manzil Zaheer, Xiang Ren, William W. Cohen
As a step towards making commonsense reasoning research more realistic, we propose to study open-ended commonsense reasoning (OpenCSR) -- the task of answering a commonsense question without any pre-defined choices -- using as a resource only a corpus of commonsense facts written in natural language.
no code implementations • 2 Jul 2020 • Pat Verga, Haitian Sun, Livio Baldini Soares, William W. Cohen
Massive language models are the core of modern NLP modeling and have been shown to encode impressive amounts of commonsense and factual information.
1 code implementation • NeurIPS 2020 • Haitian Sun, Andrew O. Arnold, Tania Bedrax-Weiss, Fernando Pereira, William W. Cohen
We address this problem with a novel QE method that is more faithful to deductive reasoning, and show that this leads to better performance on complex queries to incomplete KBs.
1 code implementation • ICLR 2020 • William W. Cohen, Haitian Sun, R. Alex Hofer, Matthew Siegler
We describe a novel way of representing a symbolic knowledge base (KB) called a sparse-matrix reified KB.
no code implementations • 24 May 2019 • William W. Cohen, Haitian Sun, R. Alex Hofer, Matthew Siegler
We present efficient differentiable implementations of second-order multi-hop reasoning using a large symbolic knowledge base (KB).
no code implementations • IJCNLP 2019 • Haitian Sun, Tania Bedrax-Weiss, William W. Cohen
We focus on a setting in which a corpus is supplemented with a large but incomplete KB, and on questions that require non-trivial (e. g., ``multi-hop'') reasoning.
2 code implementations • EMNLP 2018 • Haitian Sun, Bhuwan Dhingra, Manzil Zaheer, Kathryn Mazaitis, Ruslan Salakhutdinov, William W. Cohen
In this paper we look at a more practical setting, namely QA over the combination of a KB and entity-linked text, which is appropriate when an incomplete KB is available with a large text corpus.
Graph Representation Learning Open-Domain Question Answering
no code implementations • NeurIPS 2018 • Haitian Sun, William W. Cohen, Lidong Bing
We propose a technique for declaratively specifying strategies for semi-supervised learning (SSL).