no code implementations • 2 Dec 2024 • Shwetha Somasundaram, Anirudh Phukan, Apoorv Saxena
To reduce the latency associated with autoretrogressive LLM inference, speculative decoding has emerged as a novel decoding paradigm, where future tokens are drafted and verified in parallel.
no code implementations • 28 Nov 2024 • Anirudh Phukan, Divyansh, Harshit Kumar Morj, Vaishnavi, Apoorv Saxena, Koustava Goswami
The rapid development of Large Multimodal Models (LMMs) has significantly advanced multimodal understanding by harnessing the language abilities of Large Language Models (LLMs) and integrating modality-specific encoders.
no code implementations • 25 Sep 2024 • Pritika Ramu, Koustava Goswami, Apoorv Saxena, Balaji Vasan Srinivasan
In this paper, we propose and investigate a novel approach to the factual decomposition of generated answers for attribution, employing template-based in-context learning.
no code implementations • 1 Jun 2024 • Sambaran Bandyopadhyay, Himanshu Maheshwari, Anandhavelu Natarajan, Apoorv Saxena
Generating presentation slides from a long document with multimodal elements such as text and images is an important task.
1 code implementation • 28 May 2024 • Anirudh Phukan, Shwetha Somasundaram, Apoorv Saxena, Koustava Goswami, Balaji Vasan Srinivasan
Attributing model generations to the input source document is essential to ensure trustworthiness and reliability.
no code implementations • 29 Jan 2024 • Shivanshu Shekhar, Tanishq Dubey, Koyel Mukherjee, Apoorv Saxena, Atharv Tyagi, Nishanth Kotla
In this work, we propose optimizing the usage costs of LLMs by estimating their output quality (without actually invoking the LLMs), and then solving an optimization routine for the LLM selection to either keep costs under a budget, or minimize the costs, in a quality and latency aware manner.
no code implementations • 3 Jan 2024 • Himanshu Maheshwari, Koustava Goswami, Apoorv Saxena, Balaji Vasan Srinivasan
Our architecture is based on two parts: a the first part contains an image captioning model that takes in an image that the brand wants to post online and gives a plain English caption; b the second part takes in the generated caption along with the target brand personality and outputs a catchy personality-aligned social media caption.
no code implementations • 22 Nov 2023 • Inderjeet Nair, Shwetha Somasundaram, Apoorv Saxena, Koustava Goswami
We address the task of evidence retrieval for long document question answering, which involves locating relevant paragraphs within a document to answer a question.
no code implementations • ICCV 2023 • Aishwarya Agarwal, Srikrishna Karanam, K J Joseph, Apoorv Saxena, Koustava Goswami, Balaji Vasan Srinivasan
First, our attention segregation loss reduces the cross-attention overlap between attention maps of different concepts in the text prompt, thereby reducing the confusion/conflict among various concepts and the eventual capture of all concepts in the generated output.
1 code implementation • 22 May 2023 • Adrian Kochsiek, Apoorv Saxena, Inderjeet Nair, Rainer Gemulla
We propose KGT5-context, a simple sequence-to-sequence model for link prediction (LP) in knowledge graphs (KG).
Ranked #2 on
Link Prediction
on Wikidata5M
no code implementations • 12 Oct 2022 • Aditya Sharma, Apoorv Saxena, Chitrank Gupta, Seyed Mehran Kazemi, Partha Talukdar, Soumen Chakrabarti
Recent years have witnessed much interest in temporal reasoning over knowledge graphs (KG) for complex question answering (QA), but there remains a substantial gap in human capabilities.
Ranked #3 on
Question Answering
on TimeQuestions
1 code implementation • ACL 2022 • Apoorv Saxena, Adrian Kochsiek, Rainer Gemulla
These methods have recently been applied to KG link prediction and question answering over incomplete KGs (KGQA).
Ranked #6 on
Link Prediction
on Wikidata5M
2 code implementations • ACL 2021 • Apoorv Saxena, Soumen Chakrabarti, Partha Talukdar
Temporal Knowledge Graphs (Temporal KGs) extend regular Knowledge Graphs by providing temporal scopes (start and end times) on each edge in the KG.
Ranked #4 on
Question Answering
on Complex-CronQuestions
2 code implementations • ACL 2020 • Apoorv Saxena, Aditay Tripathi, Partha Talukdar
In a separate line of research, KG embedding methods have been proposed to reduce KG sparsity by performing missing link prediction.
Ranked #6 on
Question Answering
on MultiTQ