1 code implementation • 21 Dec 2022 • John Wieting, Jonathan H. Clark, William W. Cohen, Graham Neubig, Taylor Berg-Kirkpatrick
Contrastive learning has been successfully used for retrieval of semantically aligned sentences, but it often requires large batch sizes or careful engineering to work well.
1 code implementation • 15 Dec 2022 • Bernd Bohnet, Vinh Q. Tran, Pat Verga, Roee Aharoni, Daniel Andor, Livio Baldini Soares, Massimiliano Ciaramita, Jacob Eisenstein, Kuzman Ganchev, Jonathan Herzig, Kai Hui, Tom Kwiatkowski, Ji Ma, Jianmo Ni, Lierni Sestorain Saralegui, Tal Schuster, William W. Cohen, Michael Collins, Dipanjan Das, Donald Metzler, Slav Petrov, Kellie Webster
We take human annotations as a gold standard and show that a correlated automatic metric is suitable for development.
1 code implementation • 22 Nov 2022 • Wenhu Chen, Xueguang Ma, Xinyi Wang, William W. Cohen
By combining PoT with self-consistency decoding, we can achieve SoTA performance on all math problem datasets and near-SoTA performance on financial datasets.
1 code implementation • 22 Oct 2022 • Vidhisha Balachandran, Hannaneh Hajishirzi, William W. Cohen, Yulia Tsvetkov
Abstractive summarization models often generate inconsistent summaries containing factual errors or hallucinated content.
no code implementations • 6 Oct 2022 • Wenhu Chen, Hexiang Hu, Xi Chen, Pat Verga, William W. Cohen
While language Models store a massive amount of world knowledge implicitly in their parameters, even very large models often fail to encode information about rare entities and events, while incurring huge computational costs.
no code implementations • 29 Sep 2022 • Wenhu Chen, Hexiang Hu, Chitwan Saharia, William W. Cohen
To further evaluate the capabilities of the model, we introduce EntityDrawBench, a new benchmark that evaluates image generation for diverse entities, from frequent to rare, across multiple object categories including dogs, foods, landmarks, birds, and characters.
no code implementations • 25 Sep 2022 • Julian Martin Eisenschlos, Jeremy R. Cole, Fangyu Liu, William W. Cohen
We introduce a new in-context learning paradigm to measure Large Language Models' (LLMs) ability to learn novel words during inference.
no code implementations • 1 Jul 2022 • Wenhu Chen, William W. Cohen, Michiel de Jong, Nitish Gupta, Alessandro Presta, Pat Verga, John Wieting
In this position paper, we propose a new approach to generating a type of knowledge base (KB) from text, based on question generation and entity linking.
no code implementations • 25 May 2022 • Haitian Sun, William W. Cohen, Ruslan Salakhutdinov
Even more challenging, we only provide evidences for a subset of the conditions, so some questions may not have deterministic answers.
1 code implementation • 14 Feb 2022 • Yi Tay, Vinh Q. Tran, Mostafa Dehghani, Jianmo Ni, Dara Bahri, Harsh Mehta, Zhen Qin, Kai Hui, Zhe Zhao, Jai Gupta, Tal Schuster, William W. Cohen, Donald Metzler
In this paper, we demonstrate that information retrieval can be accomplished with a single Transformer, in which all information about the corpus is encoded in the parameters of the model.
1 code implementation • 17 Dec 2021 • Siddhant Arora, Danish Pruthi, Norman Sadeh, William W. Cohen, Zachary C. Lipton, Graham Neubig
Through our evaluation, we observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
2 code implementations • ACL 2022 • Haitian Sun, William W. Cohen, Ruslan Salakhutdinov
In addition to conditional answers, the dataset also features: (1) long context documents with information that is related in logically complex ways; (2) multi-hop questions that require compositional logical reasoning; (3) a combination of extractive questions, yes/no questions, questions with multiple answers, and not-answerable questions; (4) questions asked without knowing the answers.
1 code implementation • AKBC 2021 • Keshav Kolluru, Martin Rezk, Pat Verga, William W. Cohen, Partha Talukdar
This makes it challenging to link KG facts to sentences in languages other than the limited set of languages.
1 code implementation • EMNLP 2021 • Julian Martin Eisenschlos, Maharshi Gor, Thomas Müller, William W. Cohen
However, more than 20% of relational tables on the web have 20 or more rows (Cafarella et al., 2008), and these large tables present a challenge for current Transformer models, which are typically limited to 512 tokens.
Ranked #2 on
Question Answering
on HybridQA
no code implementations • 29 Jun 2021 • Bhuwan Dhingra, Jeremy R. Cole, Julian Martin Eisenschlos, Daniel Gillick, Jacob Eisenstein, William W. Cohen
We introduce a diagnostic dataset aimed at probing LMs for factual knowledge that changes over time and highlight problems with LMs at either end of the spectrum -- those trained on specific slices of temporal data, as well as those trained on a wide range of temporal data.
no code implementations • 1 Jun 2021 • Haitian Sun, William W. Cohen, Ruslan Salakhutdinov
We propose a new model, DocHopper, that iteratively attends to different parts of long, hierarchically structured documents to answer complex questions.
Ranked #2 on
Question Answering
on ConditionalQA
no code implementations • 5 Apr 2021 • Avishai Zagoury, Einat Minkov, Idan Szpektor, William W. Cohen
Here we study using such LMs to fill in entities in human-authored comparative questions, like ``Which country is older, India or ______?''
no code implementations • 14 Feb 2021 • Haitian Sun, Pat Verga, Bhuwan Dhingra, Ruslan Salakhutdinov, William W. Cohen
We present the Open Predicate Query Language (OPQL); a method for constructing a virtual KB (VKB) trained entirely from text.
1 code implementation • 1 Dec 2020 • Danish Pruthi, Rachit Bansal, Bhuwan Dhingra, Livio Baldini Soares, Michael Collins, Zachary C. Lipton, Graham Neubig, William W. Cohen
While many methods purport to explain predictions by highlighting salient features, what aims these explanations serve and how they ought to be evaluated often go unstated.
no code implementations • NAACL 2021 • Bill Yuchen Lin, Haitian Sun, Bhuwan Dhingra, Manzil Zaheer, Xiang Ren, William W. Cohen
As a step towards making commonsense reasoning research more realistic, we propose to study open-ended commonsense reasoning (OpenCSR) -- the task of answering a commonsense question without any pre-defined choices -- using as a resource only a corpus of commonsense facts written in natural language.
1 code implementation • ICLR 2021 • Wenhu Chen, Ming-Wei Chang, Eva Schlinger, William Wang, William W. Cohen
In open question answering (QA), the answer to a question is produced by retrieving and then analyzing documents that might contain answers to the question.
Ranked #1 on
Question Answering
on OTT-QA
no code implementations • 2 Jul 2020 • Pat Verga, Haitian Sun, Livio Baldini Soares, William W. Cohen
Massive language models are the core of modern NLP modeling and have been shown to encode impressive amounts of commonsense and factual information.
1 code implementation • NeurIPS 2020 • Haitian Sun, Andrew O. Arnold, Tania Bedrax-Weiss, Fernando Pereira, William W. Cohen
We address this problem with a novel QE method that is more faithful to deductive reasoning, and show that this leads to better performance on complex queries to incomplete KBs.
1 code implementation • ICLR 2020 • Bhuwan Dhingra, Manzil Zaheer, Vidhisha Balachandran, Graham Neubig, Ruslan Salakhutdinov, William W. Cohen
In particular, we describe a neural module, DrKIT, that traverses textual data like a KB, softly following paths of relations between mentions of entities in the corpus.
1 code implementation • ICLR 2020 • William W. Cohen, Haitian Sun, R. Alex Hofer, Matthew Siegler
We describe a novel way of representing a symbolic knowledge base (KB) called a sparse-matrix reified KB.
no code implementations • NeurIPS 2019 • Fan Yang, Liu Leqi, Yifan Wu, Zachary C. Lipton, Pradeep Ravikumar, William W. Cohen, Tom Mitchell
The ability to inferring latent psychological traits from human behavior is key to developing personalized human-interacting machine learning systems.
no code implementations • 8 Nov 2019 • Andrew O. Arnold, William W. Cohen
We focus on the problem of search in the multilingual setting.
3 code implementations • IJCNLP 2019 • Qiao Jin, Bhuwan Dhingra, Zhengping Liu, William W. Cohen, Xinghua Lu
We introduce PubMedQA, a novel biomedical question answering (QA) dataset collected from PubMed abstracts.
Ranked #5 on
Question Answering
on PubMedQA
1 code implementation • ACL 2019 • Bhuwan Dhingra, Manaal Faruqui, Ankur Parikh, Ming-Wei Chang, Dipanjan Das, William W. Cohen
Automatically constructed datasets for generating text from semi-structured data (tables), such as WikiBio, often contain reference texts that diverge from the information in the corresponding semi-structured data.
no code implementations • 24 May 2019 • William W. Cohen, Haitian Sun, R. Alex Hofer, Matthew Siegler
We present efficient differentiable implementations of second-order multi-hop reasoning using a large symbolic knowledge base (KB).
no code implementations • 15 May 2019 • William W. Cohen, Matthew Siegler, Alex Hofer
Large knowledge bases (KBs) are useful for many AI tasks, but are difficult to integrate into modern gradient-based learning systems.
no code implementations • ICLR 2019 • Zihang Dai*, Zhilin Yang*, Yiming Yang, William W. Cohen, Jaime Carbonell, Quoc V. Le, Ruslan Salakhutdinov
Moreover, Transformer-XL is up to 1, 800+ times faster than vanilla Transformer during evaluation.
no code implementations • IJCNLP 2019 • Haitian Sun, Tania Bedrax-Weiss, William W. Cohen
We focus on a setting in which a corpus is supplemented with a large but incomplete KB, and on questions that require non-trivial (e. g., ``multi-hop'') reasoning.
1 code implementation • WS 2019 • Qiao Jin, Bhuwan Dhingra, William W. Cohen, Xinghua Lu
For this we use the pre-trained LMs as fixed feature extractors and restrict the downstream task models to not have additional sequence modeling layers.
no code implementations • 15 Jan 2019 • Samira Abnar, Tania Bedrax-Weiss, Tom Kwiatkowski, William W. Cohen
Current state-of-the-art question answering models reason over an entire passage, not incrementally.
no code implementations • NeurIPS 2018 • Zhilin Yang, Jake Zhao, Bhuwan Dhingra, Kaiming He, William W. Cohen, Ruslan R. Salakhutdinov, Yann Lecun
We also show that the learned graphs are generic enough to be transferred to different embeddings on which the graphs have not been trained (including GloVe embeddings, ELMo embeddings, and task-specific RNN hidden units), or embedding-free units such as image pixels.
3 code implementations • EMNLP 2018 • Zhilin Yang, Peng Qi, Saizheng Zhang, Yoshua Bengio, William W. Cohen, Ruslan Salakhutdinov, Christopher D. Manning
Existing question answering (QA) datasets fail to train QA systems to perform complex reasoning and provide explanations for answers.
Ranked #36 on
Question Answering
on HotpotQA
2 code implementations • EMNLP 2018 • Haitian Sun, Bhuwan Dhingra, Manzil Zaheer, Kathryn Mazaitis, Ruslan Salakhutdinov, William W. Cohen
In this paper we look at a more practical setting, namely QA over the combination of a KB and entity-linked text, which is appropriate when an incomplete KB is available with a large text corpus.
Graph Representation Learning
Open-Domain Question Answering
1 code implementation • 14 Jun 2018 • Zhilin Yang, Jake Zhao, Bhuwan Dhingra, Kaiming He, William W. Cohen, Ruslan Salakhutdinov, Yann Lecun
We also show that the learned graphs are generic enough to be transferred to different embeddings on which the graphs have not been trained (including GloVe embeddings, ELMo embeddings, and task-specific RNN hidden unit), or embedding-free units such as image pixels.
no code implementations • NeurIPS 2018 • Haitian Sun, William W. Cohen, Lidong Bing
We propose a technique for declaratively specifying strategies for semi-supervised learning (SSL).
no code implementations • NAACL 2018 • Bhuwan Dhingra, Qiao Jin, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
Many problems in NLP require aggregating information from multiple mentions of the same entity which may be far apart in the text.
Ranked #6 on
Question Answering
on WikiHop
no code implementations • ICLR 2018 • Fan Yang, Jiazhong Nie, William W. Cohen, Ni Lao
Existing end-to-end deep QA models (Miller et al., 2016; Weston et al., 2014) need to read the entire text after observing the question, and therefore their complexity in responding a question is linear in the text size.
no code implementations • 17 Nov 2017 • Fan Yang, Jiazhong Nie, William W. Cohen, Ni Lao
Though deep neural networks have great success in natural language processing, they are limited at more knowledge intensive AI tasks, such as open-domain Question Answering (QA).
9 code implementations • ICLR 2018 • Zhilin Yang, Zihang Dai, Ruslan Salakhutdinov, William W. Cohen
We formulate language modeling as a matrix factorization problem, and show that the expressiveness of Softmax-based models (including the majority of neural language models) is limited by a Softmax bottleneck.
Ranked #10 on
Language Modelling
on Penn Treebank (Word Level)
no code implementations • 17 Jul 2017 • William W. Cohen, Fan Yang, Kathryn Rivard Mazaitis
We present an implementation of a probabilistic first-order logic called TensorLog, in which classes of logical queries are compiled into differentiable functions in a neural-network infrastructure such as Tensorflow or Theano.
1 code implementation • 12 Jul 2017 • Bhuwan Dhingra, Kathryn Mazaitis, William W. Cohen
ClueWeb09 serves as the background corpus for extracting these answers.
1 code implementation • NeurIPS 2017 • Zihang Dai, Zhilin Yang, Fan Yang, William W. Cohen, Ruslan Salakhutdinov
Semi-supervised learning methods based on generative adversarial networks (GANs) obtained strong empirical results, but it is not clear 1) how the discriminator benefits from joint training with a generator, and 2) why good semi-supervised classification performance and a good generator cannot be obtained at the same time.
4 code implementations • 18 Mar 2017 • Zhilin Yang, Ruslan Salakhutdinov, William W. Cohen
Recent papers have shown that neural networks obtain state-of-the-art performance on several different sequence tagging tasks.
Ranked #10 on
Part-Of-Speech Tagging
on Penn Treebank
no code implementations • 7 Mar 2017 • Bhuwan Dhingra, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
We introduce a model that encodes such graphs as explicit memory in recurrent neural networks, and use it to model coreference relations in text.
Ranked #1 on
Question Answering
on CNN / Daily Mail
no code implementations • 5 Mar 2017 • Lidong Bing, William W. Cohen, Bhuwan Dhingra
We propose a general approach to modeling semi-supervised learning (SSL) algorithms.
no code implementations • 2 Mar 2017 • Bhuwan Dhingra, Hanxiao Liu, Ruslan Salakhutdinov, William W. Cohen
The focus of past machine learning research for Reading Comprehension tasks has been primarily on the design of novel deep learning architectures.
1 code implementation • NeurIPS 2017 • Fan Yang, Zhilin Yang, William W. Cohen
We propose a framework, Neural Logic Programming, that combines the parameter and structure learning of first-order logical rules in an end-to-end differentiable model.
no code implementations • ACL 2017 • Zhilin Yang, Junjie Hu, Ruslan Salakhutdinov, William W. Cohen
In this framework, we train a generative model to generate questions based on the unlabeled text, and combine model-generated questions with human-generated questions for training question answering models.
1 code implementation • 6 Nov 2016 • Zhilin Yang, Bhuwan Dhingra, Ye Yuan, Junjie Hu, William W. Cohen, Ruslan Salakhutdinov
Previous work combines word-level and character-level representations using concatenation or scalar weighting, which is suboptimal for high-level tasks like reading comprehension.
Ranked #50 on
Question Answering
on SQuAD1.1 dev
no code implementations • 10 Jun 2016 • Lidong Bing, Bhuwan Dhingra, Kathryn Mazaitis, Jong Hyuk Park, William W. Cohen
We propose a framework to improve performance of distantly-supervised relation extraction, by jointly learning to solve two related tasks: concept-instance extraction and relation extraction.
4 code implementations • ACL 2017 • Bhuwan Dhingra, Hanxiao Liu, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
In this paper we study the problem of answering cloze-style questions over documents.
Ranked #1 on
Question Answering
on Children's Book Test
no code implementations • NeurIPS 2016 • Zhilin Yang, Ye Yuan, Yuexin Wu, Ruslan Salakhutdinov, William W. Cohen
We propose a novel extension of the encoder-decoder framework, called a review network.
1 code implementation • 20 May 2016 • William W. Cohen
Then, for each type of query to the factor graph, the message-passing steps required to perform belief propagation (BP) are "unrolled" into a function, which is differentiable.
1 code implementation • ACL 2016 • Bhuwan Dhingra, Zhong Zhou, Dylan Fitzpatrick, Michael Muehl, William W. Cohen
Text from social media provides a set of challenges that can cause traditional NLP approaches to fail.
21 code implementations • 29 Mar 2016 • Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
We present a semi-supervised learning framework based on graph embeddings.
Ranked #1 on
Node Classification
on USA Air-Traffic
no code implementations • 13 Feb 2016 • Abhinav Maurya, Kenton Murray, Yandong Liu, Chris Dyer, William W. Cohen, Daniel B. Neill
Many methods have been proposed for detecting emerging events in text streams using topic modeling.
no code implementations • 4 Jan 2016 • Lidong Bing, Mingyang Ling, Richard C. Wang, William W. Cohen
Distant labeling for information extraction (IE) suffers from noisy training data.
no code implementations • 1 May 2015 • Dana Movshovitz-Attias, William W. Cohen
To this end, we develop a similarity measure for Java classes using distributional information about how they are used in software, which we combine with corpus statistics on the distribution of contexts in which the classes appear in text.
no code implementations • 12 Apr 2014 • William Yang Wang, Kathryn Mazaitis, Ni Lao, Tom Mitchell, William W. Cohen
We show that the problem of constructing proofs for this logic is related to computation of personalized PageRank (PPR) on a linearized version of the proof space, and using on this connection, we develop a proveably-correct approximate grounding scheme, based on the PageRank-Nibble algorithm.
no code implementations • 1 Jul 2013 • Bhavana Dalvi, William W. Cohen, Jamie Callan
In multiclass semi-supervised learning (SSL), it is sometimes the case that the number of classes present in the data is not known, and hence no labeled examples are provided for some classes.
no code implementations • 1 Jul 2013 • Bhavana Dalvi, William W. Cohen, Jamie Callan
We describe a open-domain information extraction method for extracting concept-instance pairs from an HTML corpus.
no code implementations • 11 Jun 2013 • William W. Cohen, David P. Redlawsk, Douglas Pierce
We consider scenarios in which this effect arises in a model of rational decision making which includes the possibility of deceptive information.
no code implementations • 10 May 2013 • William Yang Wang, Kathryn Mazaitis, William W. Cohen
In many probabilistic first-order representation systems, inference is performed by "grounding"---i. e., mapping it to a propositional representation, and then performing propositional inference.
no code implementations • NeurIPS 2010 • Ni Lao, Jun Zhu, Liu Liu, Yandong Liu, William W. Cohen
Markov networks (MNs) can incorporate arbitrarily complex features in modeling relational data.