no code implementations • NAACL (DeeLIO) 2021 • Vidhisha Balachandran, Bhuwan Dhingra, Haitian Sun, Michael Collins, William Cohen
We create a subset of the NQ data, Factual Questions (FQ), where the questions have evidence in the KB in the form of paths that link question entities to answer entities but still must be answered using text, to facilitate further research into KB integration methods.
no code implementations • 23 Jun 2024 • Roy Xie, Junlin Wang, Ruomin Huang, Minxing Zhang, Rong Ge, Jian Pei, Neil Zhenqiang Gong, Bhuwan Dhingra
We propose ReCaLL (Relative Conditional Log-Likelihood), a novel membership inference attack (MIA) to detect LLMs' pretraining data by leveraging their conditional language modeling capabilities.
1 code implementation • 10 Jun 2024 • Junlin Wang, Tianyi Yang, Roy Xie, Bhuwan Dhingra
With the proliferation of LLM-integrated applications such as GPT-s, millions are deployed, offering valuable services through proprietary instruction prompts.
no code implementations • 6 Jun 2024 • Adam Fisch, Joshua Maynez, R. Alex Hofer, Bhuwan Dhingra, Amir Globerson, William W. Cohen
Prediction-powered inference (PPI) is a method that improves statistical estimates based on limited human-labeled data.
no code implementations • 21 May 2024 • Raghuveer Thirukovalluru, Yukun Huang, Bhuwan Dhingra
In this paper, we introduce Atomic Self-Consistency (ASC), a technique for improving the recall of relevant information in an LLM response.
1 code implementation • 17 May 2024 • Rickard Stureborg, Sanxing Chen, Ruoyu Xie, Aayushi Patel, Christopher Li, Chloe Qinyu Zhu, Tingnan Hu, Jun Yang, Bhuwan Dhingra
We define the task of tailoring vaccine interventions to a Common-Ground Opinion (CGO).
no code implementations • 9 May 2024 • R. Alex Hofer, Joshua Maynez, Bhuwan Dhingra, Adam Fisch, Amir Globerson, William W. Cohen
Prediction-powered inference (PPI) is a method that improves statistical estimates based on limited human-labeled data.
no code implementations • 15 Apr 2024 • Sanxing Chen, Sam Wiseman, Bhuwan Dhingra
The desire and ability to seek new information strategically are fundamental to human learning but often overlooked in current language agent evaluation.
no code implementations • 1 Apr 2024 • Deqing Fu, Ruohao Guo, Ghazal Khalighinejad, Ollie Liu, Bhuwan Dhingra, Dani Yogatama, Robin Jia, Willie Neiswanger
Current foundation models exhibit impressive capabilities when prompted either with text only or with both image and text inputs.
1 code implementation • 1 Mar 2024 • Ghazal Khalighinejad, Defne Circi, L. C. Brinson, Bhuwan Dhingra
This paper investigates the use of large language models (LLMs) for extracting sample lists of polymer nanocomposites (PNCs) from full-length materials science research papers.
1 code implementation • 27 Feb 2024 • Roy Xie, Chengxuan Huang, Junlin Wang, Bhuwan Dhingra
Large language models (LLMs) have significantly transformed the educational landscape.
no code implementations • 9 Feb 2024 • Yukun Huang, Yixin Liu, Raghuveer Thirukovalluru, Arman Cohan, Bhuwan Dhingra
Addressing this gap, we introduce a unified calibration framework, in which both the correctness of the LLMs' responses and their associated confidence levels are treated as distributions across a range of scores.
no code implementations • 1 Feb 2024 • Chloe Qinyu Zhu, Rickard Stureborg, Bhuwan Dhingra
Vaccine concerns are an ever-evolving target, and can shift quickly as seen during the COVID-19 pandemic.
no code implementations • 30 May 2023 • Anni Chen, Bhuwan Dhingra
Since the introduction of the SemEval 2020 Task 11 (Martino et al., 2020a), several approaches have been proposed in the literature for classifying propaganda based on the rhetorical techniques used to influence readers.
no code implementations • 24 May 2023 • Jeremy R. Cole, Michael J. Q. Zhang, Daniel Gillick, Julian Martin Eisenschlos, Bhuwan Dhingra, Jacob Eisenstein
We investigate question answering from this perspective, focusing on answering a subset of questions with a high degree of accuracy, from a set of questions in which many are inherently ambiguous.
no code implementations • 22 Mar 2023 • Jeremy R. Cole, Aditi Chaudhary, Bhuwan Dhingra, Partha Talukdar
First, we find that SSM alone improves the downstream performance on three temporal tasks by an avg.
1 code implementation • 9 Mar 2023 • Dev Seth, Rickard Stureborg, Danish Pruthi, Bhuwan Dhingra
In this work, we address this gap by learning models that predict the legibility of a perturbed string, and rank candidate perturbations based on their legibility.
no code implementations • 1 Mar 2023 • Jeremy R. Cole, Palak Jain, Julian Martin Eisenschlos, Michael J. Q. Zhang, Eunsol Choi, Bhuwan Dhingra
We propose representing factual changes between paired documents as question-answer pairs, where the answer to the same question differs between two versions.
1 code implementation • 14 Sep 2022 • Jacob Tyo, Bhuwan Dhingra, Zachary C. Lipton
Despite decades of research on authorship attribution (AA) and authorship verification (AV), inconsistent dataset splits/filtering and mismatched evaluation methods make it difficult to assess the state of the art.
1 code implementation • nlppower (ACL) 2022 • Phyllis Ang, Bhuwan Dhingra, Lisa Wu Wills
In this work, we perform a systematic study of this accuracy vs. efficiency trade-off on two widely used long-sequence models - Longformer-Encoder-Decoder (LED) and Big Bird - during fine-tuning and inference on four datasets from the SCROLLS benchmark.
no code implementations • 12 Apr 2022 • Ivan Stelmakh, Yi Luan, Bhuwan Dhingra, Ming-Wei Chang
In contrast to existing long-form QA tasks (such as ELI5), ASQA admits a clear notion of correctness: a user faced with a good summary should be able to answer different interpretations of the original ambiguous question.
no code implementations • 29 Jun 2021 • Bhuwan Dhingra, Jeremy R. Cole, Julian Martin Eisenschlos, Daniel Gillick, Jacob Eisenstein, William W. Cohen
We introduce a diagnostic dataset aimed at probing LMs for factual knowledge that changes over time and highlight problems with LMs at either end of the spectrum -- those trained on specific slices of temporal data, as well as those trained on a wide range of temporal data.
1 code implementation • NAACL 2021 • Julian Martin Eisenschlos, Bhuwan Dhingra, Jannis Bulian, Benjamin Börschinger, Jordan Boyd-Graber
We release FoolMeTwice (FM2 for short), a large dataset of challenging entailment pairs collected through a fun multi-player game.
no code implementations • 14 Feb 2021 • Haitian Sun, Pat Verga, Bhuwan Dhingra, Ruslan Salakhutdinov, William W. Cohen
We present the Open Predicate Query Language (OPQL); a method for constructing a virtual KB (VKB) trained entirely from text.
1 code implementation • 1 Dec 2020 • Danish Pruthi, Rachit Bansal, Bhuwan Dhingra, Livio Baldini Soares, Michael Collins, Zachary C. Lipton, Graham Neubig, William W. Cohen
While many methods purport to explain predictions by highlighting salient features, what aims these explanations serve and how they ought to be evaluated often go unstated.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Danish Pruthi, Bhuwan Dhingra, Graham Neubig, Zachary C. Lipton
For many prediction tasks, stakeholders desire not only predictions but also supporting evidence that a human can use to verify its correctness.
no code implementations • NAACL 2021 • Bill Yuchen Lin, Haitian Sun, Bhuwan Dhingra, Manzil Zaheer, Xiang Ren, William W. Cohen
As a step towards making commonsense reasoning research more realistic, we propose to study open-ended commonsense reasoning (OpenCSR) -- the task of answering a commonsense question without any pre-defined choices -- using as a resource only a corpus of commonsense facts written in natural language.
1 code implementation • EMNLP 2020 • Ankur P. Parikh, Xuezhi Wang, Sebastian Gehrmann, Manaal Faruqui, Bhuwan Dhingra, Diyi Yang, Dipanjan Das
We present ToTTo, an open-domain English table-to-text dataset with over 120, 000 training examples that proposes a controlled generation task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description.
Ranked #3 on Data-to-Text Generation on ToTTo
1 code implementation • ICLR 2020 • Bhuwan Dhingra, Manzil Zaheer, Vidhisha Balachandran, Graham Neubig, Ruslan Salakhutdinov, William W. Cohen
In particular, we describe a neural module, DrKIT, that traverses textual data like a KB, softly following paths of relations between mentions of entities in the corpus.
3 code implementations • ACL 2020 • Danish Pruthi, Mansi Gupta, Bhuwan Dhingra, Graham Neubig, Zachary C. Lipton
Attention mechanisms are ubiquitous components in neural architectures applied to natural language processing.
3 code implementations • IJCNLP 2019 • Qiao Jin, Bhuwan Dhingra, Zhengping Liu, William W. Cohen, Xinghua Lu
We introduce PubMedQA, a novel biomedical question answering (QA) dataset collected from PubMed abstracts.
Ranked #8 on Question Answering on PubMedQA
1 code implementation • ACL 2019 • Bhuwan Dhingra, Manaal Faruqui, Ankur Parikh, Ming-Wei Chang, Dipanjan Das, William W. Cohen
Automatically constructed datasets for generating text from semi-structured data (tables), such as WikiBio, often contain reference texts that diverge from the information in the corresponding semi-structured data.
3 code implementations • ACL 2019 • Danish Pruthi, Bhuwan Dhingra, Zachary C. Lipton
To combat adversarial spelling mistakes, we propose placing a word recognition model in front of the downstream classifier.
no code implementations • NAACL 2019 • Hao Peng, Ankur P. Parikh, Manaal Faruqui, Bhuwan Dhingra, Dipanjan Das
We propose a novel conditioned text generation model.
1 code implementation • WS 2019 • Qiao Jin, Bhuwan Dhingra, William W. Cohen, Xinghua Lu
For this we use the pre-trained LMs as fixed feature extractors and restrict the downstream task models to not have additional sequence modeling layers.
no code implementations • NeurIPS 2018 • Zhilin Yang, Jake Zhao, Bhuwan Dhingra, Kaiming He, William W. Cohen, Ruslan R. Salakhutdinov, Yann Lecun
We also show that the learned graphs are generic enough to be transferred to different embeddings on which the graphs have not been trained (including GloVe embeddings, ELMo embeddings, and task-specific RNN hidden units), or embedding-free units such as image pixels.
no code implementations • WS 2018 • Qiao Jin, Bhuwan Dhingra, William Cohen, Xinghua Lu
There are millions of articles in PubMed database.
2 code implementations • EMNLP 2018 • Haitian Sun, Bhuwan Dhingra, Manzil Zaheer, Kathryn Mazaitis, Ruslan Salakhutdinov, William W. Cohen
In this paper we look at a more practical setting, namely QA over the combination of a KB and entity-linked text, which is appropriate when an incomplete KB is available with a large text corpus.
Graph Representation Learning Open-Domain Question Answering
1 code implementation • 14 Jun 2018 • Zhilin Yang, Jake Zhao, Bhuwan Dhingra, Kaiming He, William W. Cohen, Ruslan Salakhutdinov, Yann Lecun
We also show that the learned graphs are generic enough to be transferred to different embeddings on which the graphs have not been trained (including GloVe embeddings, ELMo embeddings, and task-specific RNN hidden unit), or embedding-free units such as image pixels.
no code implementations • WS 2018 • Bhuwan Dhingra, Christopher J. Shallue, Mohammad Norouzi, Andrew M. Dai, George E. Dahl
Ideally, we could incorporate our prior knowledge of this hierarchical structure into unsupervised learning algorithms that work on text data.
no code implementations • NAACL 2018 • Bhuwan Dhingra, Qiao Jin, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
Many problems in NLP require aggregating information from multiple mentions of the same entity which may be far apart in the text.
Ranked #7 on Question Answering on WikiHop
no code implementations • NAACL 2018 • Bhuwan Dhingra, Danish Pruthi, Dheeraj Rajagopal
Recent success of deep learning models for the task of extractive Question Answering (QA) is hinged on the availability of large annotated corpora.
1 code implementation • 12 Jul 2017 • Bhuwan Dhingra, Kathryn Mazaitis, William W. Cohen
ClueWeb09 serves as the background corpus for extracting these answers.
no code implementations • 26 Mar 2017 • Yusuke Watanabe, Bhuwan Dhingra, Ruslan Salakhutdinov
Open domain Question Answering (QA) systems must interact with external knowledge sources, such as web pages, to find relevant information.
no code implementations • 7 Mar 2017 • Bhuwan Dhingra, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
We introduce a model that encodes such graphs as explicit memory in recurrent neural networks, and use it to model coreference relations in text.
Ranked #1 on Question Answering on CNN / Daily Mail
no code implementations • 5 Mar 2017 • Lidong Bing, William W. Cohen, Bhuwan Dhingra
We propose a general approach to modeling semi-supervised learning (SSL) algorithms.
no code implementations • 2 Mar 2017 • Bhuwan Dhingra, Hanxiao Liu, Ruslan Salakhutdinov, William W. Cohen
The focus of past machine learning research for Reading Comprehension tasks has been primarily on the design of novel deep learning architectures.
10 code implementations • 17 Dec 2016 • Xiujun Li, Zachary C. Lipton, Bhuwan Dhingra, Lihong Li, Jianfeng Gao, Yun-Nung Chen
Then, one can train reinforcement learning agents in an online fashion as they interact with the simulator.
1 code implementation • 6 Nov 2016 • Zhilin Yang, Bhuwan Dhingra, Ye Yuan, Junjie Hu, William W. Cohen, Ruslan Salakhutdinov
Previous work combines word-level and character-level representations using concatenation or scalar weighting, which is suboptimal for high-level tasks like reading comprehension.
Ranked #50 on Question Answering on SQuAD1.1 dev
1 code implementation • ACL 2017 • Bhuwan Dhingra, Lihong Li, Xiujun Li, Jianfeng Gao, Yun-Nung Chen, Faisal Ahmed, Li Deng
In this paper, we address this limitation by replacing symbolic queries with an induced "soft" posterior distribution over the KB that indicates which entities the user is interested in.
no code implementations • 10 Jun 2016 • Lidong Bing, Bhuwan Dhingra, Kathryn Mazaitis, Jong Hyuk Park, William W. Cohen
We propose a framework to improve performance of distantly-supervised relation extraction, by jointly learning to solve two related tasks: concept-instance extraction and relation extraction.
4 code implementations • ACL 2017 • Bhuwan Dhingra, Hanxiao Liu, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
In this paper we study the problem of answering cloze-style questions over documents.
Ranked #1 on Question Answering on Children's Book Test
1 code implementation • ACL 2016 • Bhuwan Dhingra, Zhong Zhou, Dylan Fitzpatrick, Michael Muehl, William W. Cohen
Text from social media provides a set of challenges that can cause traditional NLP approaches to fail.