Search Results for author: Siddhant Garg

Found 24 papers, 9 papers with code

Measuring Retrieval Complexity in Question Answering Systems

no code implementations5 Jun 2024 Matteo Gabburo, Nicolaas Paul Jedema, Siddhant Garg, Leonardo F. R. Ribeiro, Alessandro Moschitti

Further investigation reveals that RC scores strongly correlate with both QA performance and expert judgment across five of the six studied benchmarks, indicating that RC is an effective measure of question difficulty.

Question Answering Retrieval

SQUARE: Automatic Question Answering Evaluation using Multiple Positive and Negative References

no code implementations21 Sep 2023 Matteo Gabburo, Siddhant Garg, Rik Koncel Kedziorski, Alessandro Moschitti

Evaluation of QA systems is very challenging and expensive, with the most reliable approach being human annotations of correctness of answers for questions.

Answer Selection Sentence

Learning Answer Generation using Supervision from Automatic Question Answering Evaluators

no code implementations24 May 2023 Matteo Gabburo, Siddhant Garg, Rik Koncel-Kedziorski, Alessandro Moschitti

Recent studies show that sentence-level extractive QA, i. e., based on Answer Sentence Selection (AS2), is outperformed by Generation-based QA (GenQA) models, which generate answers using the top-k answer sentences ranked by AS2 models (a la retrieval-augmented generation style).

Answer Generation Question Answering +2

Context-Aware Transformer Pre-Training for Answer Sentence Selection

no code implementations24 May 2023 Luca Di Liello, Siddhant Garg, Alessandro Moschitti

Answer Sentence Selection (AS2) is a core component for building an accurate Question Answering pipeline.

Ranked #4 on Question Answering on TrecQA (using extra training data)

Question Answering Sentence

Structured Pruning for Multi-Task Deep Neural Networks

no code implementations13 Apr 2023 Siddhant Garg, Lijun Zhang, Hui Guan

Numerous structured pruning methods are already developed that can readily achieve speedups in single-task models, but the pruning of multi-task networks has not yet been extensively studied.

Model Compression

Knowledge Transfer from Answer Ranking to Answer Generation

no code implementations23 Oct 2022 Matteo Gabburo, Rik Koncel-Kedziorski, Siddhant Garg, Luca Soldaini, Alessandro Moschitti

In this paper, we propose to train a GenQA model by transferring knowledge from a trained AS2 model, to overcome the aforementioned issue.

Answer Generation Question Answering +2

Pre-training Transformer Models with Sentence-Level Objectives for Answer Sentence Selection

no code implementations20 May 2022 Luca Di Liello, Siddhant Garg, Luca Soldaini, Alessandro Moschitti

An important task for designing QA systems is answer sentence selection (AS2): selecting the sentence containing (or constituting) the answer to a question from a set of retrieved relevant documents.

Answer Selection Sentence

Paragraph-based Transformer Pre-training for Multi-Sentence Inference

1 code implementation NAACL 2022 Luca Di Liello, Siddhant Garg, Luca Soldaini, Alessandro Moschitti

Our evaluation on three AS2 and one fact verification datasets demonstrates the superiority of our pre-training technique over the traditional ones for transformers used as joint models for multi-candidate inference tasks, as well as when used as cross-encoders for sentence-pair formulations of these tasks.

Answer Selection Fact Verification +1

Self-Labeling Refinement for Robust Representation Learning with Bootstrap Your Own Latent

no code implementations9 Apr 2022 Siddhant Garg, Dhruval Jain

Using the proposed loss functions, we are able to surpass the performance of Vanilla BYOL (71. 04%) by training the BYOL framework using CCSL loss (76. 87%) on the STL10 dataset.

Representation Learning

A Simple Approach to Image Tilt Correction with Self-Attention MobileNet for Smartphones

no code implementations31 Oct 2021 Siddhant Garg, Debi Prasanna Mohanty, Siva Prasad Thota, Sukumar Moharana

With the combination of our novel approach and the architecture, we present state-of-the-art results on detecting the image tilt angle on mobile devices as compared to the MobileNetV3 model.

Multi-Label Learning

Attentive Walk-Aggregating Graph Neural Networks

1 code implementation6 Oct 2021 Mehmet F. Demirel, Shengchao Liu, Siddhant Garg, Zhenmei Shi, YIngyu Liang

Our experiments demonstrate the strong performance of AWARE in graph-level prediction tasks in the standard setting in the domains of molecular property prediction and social networks.

Molecular Property Prediction Property Prediction

Will this Question be Answered? Question Filtering via Answer Model Distillation for Efficient Question Answering

no code implementations EMNLP 2021 Siddhant Garg, Alessandro Moschitti

In this paper we propose a novel approach towards improving the efficiency of Question Answering (QA) systems by filtering out questions that will not be answered by them.

Question Answering

Functional Regularization for Representation Learning: A Unified Theoretical Perspective

1 code implementation NeurIPS 2020 Siddhant Garg, YIngyu Liang

Unsupervised and self-supervised learning approaches have become a crucial tool to learn representations for downstream prediction tasks.

Representation Learning Self-Supervised Learning +1

Can Adversarial Weight Perturbations Inject Neural Backdoors?

1 code implementation4 Aug 2020 Siddhant Garg, Adarsh Kumar, Vibhor Goel, YIngyu Liang

We introduce adversarial perturbations in the model weights using a composite loss on the predictions of the original model and the desired trigger through projected gradient descent.

Advances in Quantum Deep Learning: An Overview

no code implementations8 May 2020 Siddhant Garg, Goutham Ramakrishnan

The last few decades have seen significant breakthroughs in the fields of deep learning and quantum computing.

Deep Learning

Beyond Fine-tuning: Few-Sample Sentence Embedding Transfer

no code implementations Asian Chapter of the Association for Computational Linguistics 2020 Siddhant Garg, Rohit Kumar Sharma, YIngyu Liang

In this paper we show that concatenating the embeddings from the pre-trained model with those from a simple sentence embedding model trained only on the target data, can improve over the performance of FT for few-sample tasks.

Dimensionality Reduction General Classification +5

BAE: BERT-based Adversarial Examples for Text Classification

2 code implementations EMNLP 2020 Siddhant Garg, Goutham Ramakrishnan

Modern text classification models are susceptible to adversarial examples, perturbed versions of the original text indiscernible by humans which get misclassified by the model.

Adversarial Text General Classification +2

Cross-Shape Attention for Part Segmentation of 3D Point Clouds

1 code implementation20 Mar 2020 Marios Loizou, Siddhant Garg, Dmitry Petrov, Melinos Averkiou, Evangelos Kalogerakis

The mechanism assesses both the degree of interaction between points and also mediates feature propagation across shapes, improving the accuracy and consistency of the resulting point-wise feature representations for shape segmentation.

3D Semantic Segmentation Retrieval +1

Stochastic Bandits with Delayed Composite Anonymous Feedback

no code implementations2 Oct 2019 Siddhant Garg, Aditya Kumar Akash

The complexity of this problem stems from the anonymous feedback to the player and the stochastic generation of the reward.

Data Ordering Patterns for Neural Machine Translation: An Empirical Study

no code implementations23 Sep 2019 Siddhant Garg

Recent works show that ordering of the training data affects the model performance for Neural Machine Translation.

Machine Translation Translation

Surprisingly Easy Hard-Attention for Sequence to Sequence Learning

1 code implementation EMNLP 2018 Shiv Shankar, Siddhant Garg, Sunita Sarawagi

In this paper we show that a simple beam approximation of the joint distribution between attention and output is an easy, accurate, and efficient attention mechanism for sequence to sequence learning.

Hard Attention Image Captioning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.