no code implementations • EMNLP (NLP4ConvAI) 2021 • Pei Zhou, Behnam Hedayatnia, Karthik Gopalakrishnan, Seokhwan Kim, Jay Pujara, Xiang Ren, Yang Liu, Dilek Hakkani-Tur
We further investigate can such models identify when to generate implicit background knowledge and when it is not necessary.
no code implementations • EMNLP 2021 • Avijit Thawani, Jay Pujara, Filip Ilievski
This paper studies the effect of using six different number encoders on the task of masked word prediction (MWP), as a proxy for evaluating literacy.
no code implementations • 24 Feb 2025 • Dong-Ho Lee, Hyundong Cho, Jonathan May, Jay Pujara
Asking questions is a fundamental aspect of learning that facilitates deeper understanding.
1 code implementation • 18 Feb 2025 • Dong-Ho Lee, Adyasha Maharana, Jay Pujara, Xiang Ren, Francesco Barbieri
Building on these insights, we introduce two benchmark tasks: (1) persona simulation where a model continues a conversation on behalf of a specific user given prior dialogue context; and (2) memory probing where a model answers targeted questions requiring long-term memory of past interactions.
no code implementations • 2 Oct 2024 • Kian Ahrabian, Alon Benhaim, Barun Patra, Jay Pujara, Saksham Singhal, Xia Song
However, its main limitation is incompatibility with decoder-only transformers out of the box.
1 code implementation • 26 Sep 2024 • Yifan Jiang, Kriti Aggarwal, Tanmay Laud, Kashif Munir, Jay Pujara, Subhabrata Mukherjee
Our experiments reveal that all LLMs are vulnerable to RED QUEEN ATTACK, reaching 87. 62% attack success rate on GPT-4o and 75. 4% on Llama3-70B.
no code implementations • 21 Jul 2024 • Kian Ahrabian, Xihui Lin, Barun Patra, Vishrav Chaudhary, Alon Benhaim, Jay Pujara, Xia Song
With the growing utilization of large language models (LLMs) across domains, alignment towards human preferences has become one of the most critical aspects of training models.
1 code implementation • 21 Apr 2024 • Yifan Jiang, Jiarui Zhang, Kexuan Sun, Zhivar Sourati, Kian Ahrabian, Kaixin Ma, Filip Ilievski, Jay Pujara
Further analysis of perception questions reveals that MLLMs struggle to comprehend the visual features (near-random performance) and even count the panels in the puzzle ( <45%), hindering their ability for abstract reasoning.
no code implementations • 26 Mar 2024 • Saurav Joshi, Filip Ilievski, Jay Pujara
According to WWF, 1. 1 billion people lack access to water, and 2. 7 billion experience water scarcity at least one month a year.
1 code implementation • 18 Feb 2024 • Ju-Hyung Lee, Dong-Ho Lee, Joohan Lee, Jay Pujara
The burgeoning field of on-device AI communication, where devices exchange information directly through embedded foundation models, such as language models (LMs), requires robust, efficient, and generalizable communication frameworks.
2 code implementations • 6 Feb 2024 • Pei Zhou, Jay Pujara, Xiang Ren, Xinyun Chen, Heng-Tze Cheng, Quoc V. Le, Ed H. Chi, Denny Zhou, Swaroop Mishra, Huaixiu Steven Zheng
We introduce SELF-DISCOVER, a general framework for LLMs to self-discover the task-intrinsic reasoning structures to tackle complex reasoning problems that are challenging for typical prompting methods.
no code implementations • 22 Jan 2024 • Kian Ahrabian, Zhivar Sourati, Kexuan Sun, Jiarui Zhang, Yifan Jiang, Fred Morstatter, Jay Pujara
While large language models (LLMs) are still being adopted to new domains and utilized in novel applications, we are experiencing an influx of the new generation of foundation models, namely multi-modal large language models (MLLMs).
1 code implementation • 15 Dec 2023 • Pegah Jandaghi, XiangHai Sheng, Xinyi Bai, Jay Pujara, Hakim Sidahmed
Training Natural Language Processing (NLP) models on a diverse and comprehensive persona-based dataset can lead to conversational models that create a deeper connection with the user, and maintain their engagement.
1 code implementation • 31 Oct 2023 • Dong-Ho Lee, Jay Pujara, Mohit Sewak, Ryen W. White, Sujay Kumar Jauhar
In our experiments we demonstrate that instruction-following LLMs are highly cost-effective data creators, and that models trained with these data exhibit performance better than those trained with human-labeled data (by up to 17. 5%) on out-of-distribution evaluation, while maintaining comparable performance on in-distribution tasks.
Ranked #4 on
Visual Question Answering
on ViP-Bench
1 code implementation • 17 Oct 2023 • Avijit Thawani, Saurabh Ghanekar, Xiaoyuan Zhu, Jay Pujara
Language models typically tokenize text into subwords, using a deterministic, hand-engineered heuristic of combining characters into longer surface-level strings such as 'ing' or whole words.
no code implementations • 9 Oct 2023 • Avijit Thawani, Jay Pujara, Ashwin Kalyan
Despite recent successes in language models, their ability to represent numbers is insufficient.
1 code implementation • 4 Oct 2023 • Pei Zhou, Aman Madaan, Srividya Pranavi Potharaju, Aditya Gupta, Kevin R. McKee, Ari Holtzman, Jay Pujara, Xiang Ren, Swaroop Mishra, Aida Nematzadeh, Shyam Upadhyay, Manaal Faruqui
We propose a new evaluation paradigm for large language models (LLMs): Thinking for Doing (T4D), which requires models to connect inferences about others' mental states to actions in social scenarios.
no code implementations • NAACL (sdp) 2021 • Lee Kezar, Jay Pujara
Using a corpus of scholarly documents across 19 disciplines and state-of-the-art language modeling techniques, we learn a fixed set of domain-agnostic descriptors for document sections and "retrofit" the corpus to these descriptors (also referred to as "normalization").
no code implementations • 18 May 2023 • Jihyung Moon, Dong-Ho Lee, Hyundong Cho, Woojeong Jin, Chan Young Park, Minwoo Kim, Jonathan May, Jay Pujara, Sungjoon Park
Previous approaches to detecting toxic language and norm violations have been primarily concerned with conversations from online forums and social media, such as Reddit and Twitter.
1 code implementation • 17 May 2023 • Dong-Ho Lee, Kian Ahrabian, Woojeong Jin, Fred Morstatter, Jay Pujara
This shows that prior semantic knowledge is unnecessary; instead, LLMs can leverage the existing patterns in the context to achieve such performance.
no code implementations • 4 Feb 2023 • Kian Ahrabian, Xinwei Du, Richard Delwin Myloth, Arun Baalaaji Sankar Ananthan, Jay Pujara
In this paper, we present PubGraph, a new resource for studying scientific progress that takes the form of a large-scale knowledge graph (KG) with more than 385M entities, 13B main edges, and 1. 5B qualifier edges.
no code implementations • 20 Dec 2022 • Pei Zhou, Andrew Zhu, Jennifer Hu, Jay Pujara, Xiang Ren, Chris Callison-Burch, Yejin Choi, Prithviraj Ammanabrolu
We propose a novel task, G4C, to study teacher-student natural language interactions in a goal-driven and grounded environment.
no code implementations • 16 Nov 2022 • Pei Zhou, Hyundong Cho, Pegah Jandaghi, Dong-Ho Lee, Bill Yuchen Lin, Jay Pujara, Xiang Ren
Human communication relies on common ground (CG), the mutual knowledge and beliefs shared by participants, to produce coherent and interesting conversations.
no code implementations • 30 Oct 2022 • Dong-Ho Lee, Akshen Kadakia, Brihi Joshi, Aaron Chan, Ziyi Liu, Kiran Narahari, Takashi Shibuya, Ryosuke Mitani, Toshiyuki Sekiya, Jay Pujara, Xiang Ren
Explanation-based model debugging aims to resolve spurious biases by showing human users explanations of model behavior, asking users to give feedback on the behavior, then using the feedback to update the model.
no code implementations • 27 Oct 2022 • Ju-Hyung Lee, Dong-Ho Lee, Eunsoo Sheen, Thomas Choi, Jay Pujara
In this work, we propose a realistic semantic network called seq2seq-SC, designed to be compatible with 5G NR and capable of working with generalized text datasets using a pre-trained language model.
no code implementations • 2 Oct 2022 • Filip Ilievski, Jay Pujara, Kartik Shenoy
Analogical reasoning methods have been built over various resources, including commonsense knowledge bases, lexical resources, language models, or their combination.
1 code implementation • 14 Jun 2022 • Thiloshon Nagarajah, Filip Ilievski, Jay Pujara
Experiments with language models and neuro-symbolic AI reasoners on these tasks reveal that state-of-the-art methods can be applied to reason by analogy with a limited success, motivating the need for further research towards comprehensive and scalable analogical reasoning by AI.
1 code implementation • 12 May 2022 • Alon Albalak, Yi-Lin Tuan, Pegah Jandaghi, Connor Pryor, Luke Yoffe, Deepak Ramachandran, Lise Getoor, Jay Pujara, William Yang Wang
Task transfer, transferring knowledge contained in related tasks, holds the promise of reducing the quantity of labeled data required to fine-tune language models.
no code implementations • ACL 2022 • Woojeong Jin, Dong-Ho Lee, Chenguang Zhu, Jay Pujara, Xiang Ren
Pre-trained language models are still far from human performance in tasks that need understanding of properties (e. g. appearance, measurable quantity) and affordances of everyday objects in the real world since the text lacks such information due to reporting bias.
no code implementations • 19 Jan 2022 • Ehsan Qasemi, Lee Kezar, Jay Pujara, Pedro Szekely
Language models (LMs) show state of the art performance for common sense (CS) question answering, but whether this ability implies a human-level mastery of CS remains an open question.
1 code implementation • ACL 2022 • Dong-Ho Lee, Akshen Kadakia, Kangmin Tan, Mahak Agarwal, Xinyu Feng, Takashi Shibuya, Ryosuke Mitani, Toshiyuki Sekiya, Jay Pujara, Xiang Ren
We also find that good demonstration can save many labeled examples and consistency in demonstration contributes to better performance.
no code implementations • ACL 2022 • Pei Zhou, Karthik Gopalakrishnan, Behnam Hedayatnia, Seokhwan Kim, Jay Pujara, Xiang Ren, Yang Liu, Dilek Hakkani-Tur
Implicit knowledge, such as common sense, is key to fluid human conversations.
1 code implementation • SIGDIAL (ACL) 2021 • Pei Zhou, Karthik Gopalakrishnan, Behnam Hedayatnia, Seokhwan Kim, Jay Pujara, Xiang Ren, Yang Liu, Dilek Hakkani-Tur
Moreover, existing dialogue datasets do not explicitly focus on exhibiting commonsense as a facet.
no code implementations • 10 Sep 2021 • Dong-Ho Lee, Ravi Kiran Selvam, Sheikh Muhammad Sarwar, Bill Yuchen Lin, Fred Morstatter, Jay Pujara, Elizabeth Boschee, James Allan, Xiang Ren
Deep neural models for named entity recognition (NER) have shown impressive results in overcoming label scarcity and generalizing to unseen entities by leveraging distant supervision and auxiliary information such as explanations.
Low Resource Named Entity Recognition
named-entity-recognition
+2
1 code implementation • Findings (EMNLP) 2021 • Fei Wang, Kexuan Sun, Jay Pujara, Pedro Szekely, Muhao Chen
From one perspective, our system conducts masked salient token prediction to enhance the model for alignment and reasoning between the table and the statement.
Ranked #10 on
Table-based Fact Verification
on TabFact
no code implementations • AKBC Workshop CSKB 2021 • Filip Ilievski, Jay Pujara, Hanzhi Zhang
Our method aligns story types with commonsense axioms, and queries to a commonsense knowledge graph, enabling the generation of hundreds of thousands of stories.
1 code implementation • 4 May 2021 • Fei Wang, Kexuan Sun, Muhao Chen, Jay Pujara, Pedro Szekely
The task of natural language table retrieval (NLTR) seeks to retrieve semantically relevant tables based on natural language queries.
no code implementations • Findings (EMNLP) 2021 • Pei Zhou, Pegah Jandaghi, Bill Yuchen Lin, Justin Cho, Jay Pujara, Xiang Ren
Humans use commonsense reasoning (CSR) implicitly to produce natural and coherent responses in conversations.
no code implementations • NAACL 2021 • Avijit Thawani, Jay Pujara, Pedro A. Szekely, Filip Ilievski
NLP systems rarely give special consideration to numbers found in text.
no code implementations • EMNLP 2021 • Ninareh Mehrabi, Pei Zhou, Fred Morstatter, Jay Pujara, Xiang Ren, Aram Galstyan
In addition, we analyze two downstream models that use ConceptNet as a source for commonsense knowledge and find the existence of biases in those models as well.
no code implementations • EMNLP 2021 • Pei Zhou, Rahul Khanna, Seyeon Lee, Bill Yuchen Lin, Daniel Ho, Jay Pujara, Xiang Ren
Pre-trained language models (PTLMs) have achieved impressive performance on commonsense inference benchmarks, but their ability to employ commonsense to make robust inferences, which is crucial for effective communications with humans, is debated.
no code implementations • 16 Jan 2020 • Pegah Jandaghi, Jay Pujara
Our system finds patterns in time series data and ranks these patterns based on empirical observations of human behavior using utility estimation.
no code implementations • 16 Nov 2017 • Dhanya Sridhar, Jay Pujara, Lise Getoor
Knowledge bases (KB) constructed through information extraction from text play an important role in query answering and reasoning.
1 code implementation • EMNLP 2017 • Jay Pujara, Eriq Augustine, Lise Getoor
Knowledge graph (KG) embedding techniques use structured relationships between entities to learn low-dimensional representations of entities and relations.
no code implementations • 4 Jul 2016 • Jay Pujara, Lise Getoor
A common theme in this research has been the importance of incorporating relational features into the resolution process.
no code implementations • 2 Jul 2016 • Shobeir Fakhraei, Dhanya Sridhar, Jay Pujara, Lise Getoor
A neighborhood graph, which represents the instances as vertices and their relations as weighted edges, is the basis of many semi-supervised and relational models for node labeling and link prediction.