Search Results for author: Jay Pujara

Found 42 papers, 14 papers with code

Numeracy enhances the Literacy of Language Models

no code implementations EMNLP 2021 Avijit Thawani, Jay Pujara, Filip Ilievski

This paper studies the effect of using six different number encoders on the task of masked word prediction (MWP), as a proxy for evaluating literacy.

Sentence

MARVEL: Multidimensional Abstraction and Reasoning through Visual Evaluation and Learning

no code implementations21 Apr 2024 Yifan Jiang, Jiarui Zhang, Kexuan Sun, Zhivar Sourati, Kian Ahrabian, Kaixin Ma, Filip Ilievski, Jay Pujara

Further analysis of perception questions reveals that MLLMs struggle to comprehend the visual features (near-random performance) and even count the panels in the puzzle ( <45%), hindering their ability for abstract reasoning.

Visual Reasoning

Knowledge-Powered Recommendation for an Improved Diet Water Footprint

no code implementations26 Mar 2024 Saurav Joshi, Filip Ilievski, Jay Pujara

According to WWF, 1. 1 billion people lack access to water, and 2. 7 billion experience water scarcity at least one month a year.

graph construction Knowledge Graphs

Integrating Pre-Trained Language Model with Physical Layer Communications

1 code implementation18 Feb 2024 Ju-Hyung Lee, Dong-Ho Lee, Joohan Lee, Jay Pujara

The burgeoning field of on-device AI communication, where devices exchange information directly through embedded foundation models, such as language models (LMs), requires robust, efficient, and generalizable communication frameworks.

Language Modelling

Self-Discover: Large Language Models Self-Compose Reasoning Structures

2 code implementations6 Feb 2024 Pei Zhou, Jay Pujara, Xiang Ren, Xinyun Chen, Heng-Tze Cheng, Quoc V. Le, Ed H. Chi, Denny Zhou, Swaroop Mishra, Huaixiu Steven Zheng

We introduce SELF-DISCOVER, a general framework for LLMs to self-discover the task-intrinsic reasoning structures to tackle complex reasoning problems that are challenging for typical prompting methods.

Math

The Curious Case of Nonverbal Abstract Reasoning with Multi-Modal Large Language Models

1 code implementation22 Jan 2024 Kian Ahrabian, Zhivar Sourati, Kexuan Sun, Jiarui Zhang, Yifan Jiang, Fred Morstatter, Jay Pujara

While large language models (LLMs) are still being adopted to new domains and utilized in novel applications, we are experiencing an influx of the new generation of foundation models, namely multi-modal large language models (MLLMs).

Faithful Persona-based Conversational Dataset Generation with Large Language Models

1 code implementation15 Dec 2023 Pegah Jandaghi, XiangHai Sheng, Xinyi Bai, Jay Pujara, Hakim Sidahmed

Training Natural Language Processing (NLP) models on a diverse and comprehensive persona-based dataset can lead to conversational models that create a deeper connection with the user, and maintain their engagement.

Chatbot

Making Large Language Models Better Data Creators

1 code implementation31 Oct 2023 Dong-Ho Lee, Jay Pujara, Mohit Sewak, Ryen W. White, Sujay Kumar Jauhar

In our experiments we demonstrate that instruction-following LLMs are highly cost-effective data creators, and that models trained with these data exhibit performance better than those trained with human-labeled data (by up to 17. 5%) on out-of-distribution evaluation, while maintaining comparable performance on in-distribution tasks.

Instruction Following Prompt Engineering +1

Learn Your Tokens: Word-Pooled Tokenization for Language Modeling

1 code implementation17 Oct 2023 Avijit Thawani, Saurabh Ghanekar, Xiaoyuan Zhu, Jay Pujara

Language models typically tokenize text into subwords, using a deterministic, hand-engineered heuristic of combining characters into longer surface-level strings such as 'ing' or whole words.

Language Modelling

Estimating Numbers without Regression

no code implementations9 Oct 2023 Avijit Thawani, Jay Pujara, Ashwin Kalyan

Despite recent successes in language models, their ability to represent numbers is insufficient.

Language Modelling regression

How FaR Are Large Language Models From Agents with Theory-of-Mind?

no code implementations4 Oct 2023 Pei Zhou, Aman Madaan, Srividya Pranavi Potharaju, Aditya Gupta, Kevin R. McKee, Ari Holtzman, Jay Pujara, Xiang Ren, Swaroop Mishra, Aida Nematzadeh, Shyam Upadhyay, Manaal Faruqui

We propose a new evaluation paradigm for large language models (LLMs): Thinking for Doing (T4D), which requires models to connect inferences about others' mental states to actions in social scenarios.

In-Context Learning Question Answering

Finding Pragmatic Differences Between Disciplines

no code implementations NAACL (sdp) 2021 Lee Kezar, Jay Pujara

Using a corpus of scholarly documents across 19 disciplines and state-of-the-art language modeling techniques, we learn a fixed set of domain-agnostic descriptors for document sections and "retrofit" the corpus to these descriptors (also referred to as "normalization").

Document Summarization document understanding +2

Analyzing Norm Violations in Live-Stream Chat

no code implementations18 May 2023 Jihyung Moon, Dong-Ho Lee, Hyundong Cho, Woojeong Jin, Chan Young Park, Minwoo Kim, Jonathan May, Jay Pujara, Sungjoon Park

Previous approaches to detecting toxic language and norm violations have been primarily concerned with conversations from online forums and social media, such as Reddit and Twitter.

Temporal Knowledge Graph Forecasting Without Knowledge Using In-Context Learning

1 code implementation17 May 2023 Dong-Ho Lee, Kian Ahrabian, Woojeong Jin, Fred Morstatter, Jay Pujara

This shows that prior semantic knowledge is unnecessary; instead, LLMs can leverage the existing patterns in the context to achieve such performance.

In-Context Learning

PubGraph: A Large-Scale Scientific Knowledge Graph

no code implementations4 Feb 2023 Kian Ahrabian, Xinwei Du, Richard Delwin Myloth, Arun Baalaaji Sankar Ananthan, Jay Pujara

In this paper, we present PubGraph, a new resource for studying scientific progress that takes the form of a large-scale knowledge graph (KG) with more than 385M entities, 13B main edges, and 1. 5B qualifier edges.

Community Detection Knowledge Graph Completion +2

Reflect, Not Reflex: Inference-Based Common Ground Improves Dialogue Response Quality

no code implementations16 Nov 2022 Pei Zhou, Hyundong Cho, Pegah Jandaghi, Dong-Ho Lee, Bill Yuchen Lin, Jay Pujara, Xiang Ren

Human communication relies on common ground (CG), the mutual knowledge and beliefs shared by participants, to produce coherent and interesting conversations.

Response Generation

XMD: An End-to-End Framework for Interactive Explanation-Based Debugging of NLP Models

no code implementations30 Oct 2022 Dong-Ho Lee, Akshen Kadakia, Brihi Joshi, Aaron Chan, Ziyi Liu, Kiran Narahari, Takashi Shibuya, Ryosuke Mitani, Toshiyuki Sekiya, Jay Pujara, Xiang Ren

Explanation-based model debugging aims to resolve spurious biases by showing human users explanations of model behavior, asking users to give feedback on the behavior, then using the feedback to update the model.

text-classification Text Classification

Seq2Seq-SC: End-to-End Semantic Communication Systems with Pre-trained Language Model

no code implementations27 Oct 2022 Ju-Hyung Lee, Dong-Ho Lee, Eunsoo Sheen, Thomas Choi, Jay Pujara

In this work, we propose a realistic semantic network called seq2seq-SC, designed to be compatible with 5G NR and capable of working with generalized text datasets using a pre-trained language model.

Language Modelling Semantic Similarity +1

Does Wikidata Support Analogical Reasoning?

no code implementations2 Oct 2022 Filip Ilievski, Jay Pujara, Kartik Shenoy

Analogical reasoning methods have been built over various resources, including commonsense knowledge bases, lexical resources, language models, or their combination.

Understanding Narratives through Dimensions of Analogy

1 code implementation14 Jun 2022 Thiloshon Nagarajah, Filip Ilievski, Jay Pujara

Experiments with language models and neuro-symbolic AI reasoners on these tasks reveal that state-of-the-art methods can be applied to reason by analogy with a limited success, motivating the need for further research towards comprehensive and scalable analogical reasoning by AI.

FETA: A Benchmark for Few-Sample Task Transfer in Open-Domain Dialogue

1 code implementation12 May 2022 Alon Albalak, Yi-Lin Tuan, Pegah Jandaghi, Connor Pryor, Luke Yoffe, Deepak Ramachandran, Lise Getoor, Jay Pujara, William Yang Wang

Task transfer, transferring knowledge contained in related tasks, holds the promise of reducing the quantity of labeled data required to fine-tune language models.

Dialogue Understanding Domain Adaptation +1

Leveraging Visual Knowledge in Language Tasks: An Empirical Study on Intermediate Pre-training for Cross-modal Knowledge Transfer

no code implementations ACL 2022 Woojeong Jin, Dong-Ho Lee, Chenguang Zhu, Jay Pujara, Xiang Ren

Pre-trained language models are still far from human performance in tasks that need understanding of properties (e. g. appearance, measurable quantity) and affordances of everyday objects in the real world since the text lacks such information due to reporting bias.

Image Captioning Language Modelling +1

Evaluating Machine Common Sense via Cloze Testing

no code implementations19 Jan 2022 Ehsan Qasemi, Lee Kezar, Jay Pujara, Pedro Szekely

Language models (LMs) show state of the art performance for common sense (CS) question answering, but whether this ability implies a human-level mastery of CS remains an open question.

Common Sense Reasoning Open-Ended Question Answering +1

AutoTriggER: Label-Efficient and Robust Named Entity Recognition with Auxiliary Trigger Extraction

no code implementations10 Sep 2021 Dong-Ho Lee, Ravi Kiran Selvam, Sheikh Muhammad Sarwar, Bill Yuchen Lin, Fred Morstatter, Jay Pujara, Elizabeth Boschee, James Allan, Xiang Ren

Deep neural models for named entity recognition (NER) have shown impressive results in overcoming label scarcity and generalizing to unseen entities by leveraging distant supervision and auxiliary information such as explanations.

Low Resource Named Entity Recognition named-entity-recognition +2

Table-based Fact Verification with Salience-aware Learning

1 code implementation Findings (EMNLP) 2021 Fei Wang, Kexuan Sun, Jay Pujara, Pedro Szekely, Muhao Chen

From one perspective, our system conducts masked salient token prediction to enhance the model for alignment and reasoning between the table and the statement.

counterfactual Data Augmentation +2

Story Generation with Commonsense Knowledge Graphs and Axioms

no code implementations AKBC Workshop CSKB 2021 Filip Ilievski, Jay Pujara, Hanzhi Zhang

Our method aligns story types with commonsense axioms, and queries to a commonsense knowledge graph, enabling the generation of hundreds of thousands of stories.

Common Sense Reasoning Knowledge Graphs +1

Retrieving Complex Tables with Multi-Granular Graph Representation Learning

1 code implementation4 May 2021 Fei Wang, Kexuan Sun, Muhao Chen, Jay Pujara, Pedro Szekely

The task of natural language table retrieval (NLTR) seeks to retrieve semantically relevant tables based on natural language queries.

Graph Representation Learning Natural Language Queries +2

Lawyers are Dishonest? Quantifying Representational Harms in Commonsense Knowledge Resources

no code implementations EMNLP 2021 Ninareh Mehrabi, Pei Zhou, Fred Morstatter, Jay Pujara, Xiang Ren, Aram Galstyan

In addition, we analyze two downstream models that use ConceptNet as a source for commonsense knowledge and find the existence of biases in those models as well.

RICA: Evaluating Robust Inference Capabilities Based on Commonsense Axioms

no code implementations EMNLP 2021 Pei Zhou, Rahul Khanna, Seyeon Lee, Bill Yuchen Lin, Daniel Ho, Jay Pujara, Xiang Ren

Pre-trained language models (PTLMs) have achieved impressive performance on commonsense inference benchmarks, but their ability to employ commonsense to make robust inferences, which is crucial for effective communications with humans, is debated.

Human-like Time Series Summaries via Trend Utility Estimation

no code implementations16 Jan 2020 Pegah Jandaghi, Jay Pujara

Our system finds patterns in time series data and ranks these patterns based on empirical observations of human behavior using utility estimation.

Time Series Time Series Analysis

Using Noisy Extractions to Discover Causal Knowledge

no code implementations16 Nov 2017 Dhanya Sridhar, Jay Pujara, Lise Getoor

Knowledge bases (KB) constructed through information extraction from text play an important role in query answering and reasoning.

Causal Discovery

Sparsity and Noise: Where Knowledge Graph Embeddings Fall Short

1 code implementation EMNLP 2017 Jay Pujara, Eriq Augustine, Lise Getoor

Knowledge graph (KG) embedding techniques use structured relationships between entities to learn low-dimensional representations of entities and relations.

Knowledge Graph Embeddings Knowledge Graphs +1

Generic Statistical Relational Entity Resolution in Knowledge Graphs

no code implementations4 Jul 2016 Jay Pujara, Lise Getoor

A common theme in this research has been the importance of incorporating relational features into the resolution process.

Entity Resolution Knowledge Graphs

Adaptive Neighborhood Graph Construction for Inference in Multi-Relational Networks

no code implementations2 Jul 2016 Shobeir Fakhraei, Dhanya Sridhar, Jay Pujara, Lise Getoor

A neighborhood graph, which represents the instances as vertices and their relations as weighted edges, is the basis of many semi-supervised and relational models for node labeling and link prediction.

graph construction Link Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.