Search Results for author: Krishna Srinivasan

Found 11 papers, 4 papers with code

Ambiguity-Aware In-Context Learning with Large Language Models

no code implementations14 Sep 2023 Lingyu Gao, Aditi Chaudhary, Krishna Srinivasan, Kazuma Hashimoto, Karthik Raman, Michael Bendersky

In-context learning (ICL) i. e. showing LLMs only a few task-specific demonstrations has led to downstream gains with no task-specific fine-tuning required.

In-Context Learning Semantic Similarity +3

Exploring the Viability of Synthetic Query Generation for Relevance Prediction

no code implementations19 May 2023 Aditi Chaudhary, Karthik Raman, Krishna Srinivasan, Kazuma Hashimoto, Mike Bendersky, Marc Najork

While our experiments demonstrate that these modifications help improve performance of QGen techniques, we also find that QGen approaches struggle to capture the full nuance of the relevance label space and as a result the generated queries are not faithful to the desired relevance label.

Information Retrieval Question Answering +2

QUILL: Query Intent with Large Language Models using Retrieval Augmentation and Multi-stage Distillation

no code implementations27 Oct 2022 Krishna Srinivasan, Karthik Raman, Anupam Samanta, Lingrui Liao, Luca Bertelli, Mike Bendersky

Thus, in this paper we make the following contributions: (1) We demonstrate that Retrieval Augmentation of queries provides LLMs with valuable additional context enabling improved understanding.

Feature Engineering Knowledge Distillation +1

Transforming Sequence Tagging Into A Seq2Seq Task

no code implementations16 Mar 2022 Karthik Raman, Iftekhar Naim, Jiecao Chen, Kazuma Hashimoto, Kiran Yalasangi, Krishna Srinivasan

Pretrained, large, generative language models (LMs) have had great success in a wide range of sequence tagging and structured prediction tasks.

Hallucination Structured Prediction +1

MURAL: Multimodal, Multitask Retrieval Across Languages

no code implementations10 Sep 2021 Aashi Jain, Mandy Guo, Krishna Srinivasan, Ting Chen, Sneha Kudugunta, Chao Jia, Yinfei Yang, Jason Baldridge

Both image-caption pairs and translation pairs provide the means to learn deep representations of and connections between languages.

Image-text matching Retrieval +5

DICT-MLM: Improved Multilingual Pre-Training using Bilingual Dictionaries

no code implementations23 Oct 2020 Aditi Chaudhary, Karthik Raman, Krishna Srinivasan, Jiecao Chen

In particular, by requiring the model to predict the language-specific token, the MLM objective disincentivizes learning a language-agnostic representation -- which is a key goal of multilingual pre-training.

Language Modelling Masked Language Modeling +1

Cannot find the paper you are looking for? You can Submit a new open access paper.