Search Results for author: Gergely Szilvasy

Found 5 papers, 2 papers with code

Evaluation data contamination in LLMs: how do we measure it and (when) does it matter?

no code implementations6 Nov 2024 Aaditya K. Singh, Muhammed Yusuf Kocyigit, Andrew Poulton, David Esiobu, Maria Lomeli, Gergely Szilvasy, Dieuwke Hupkes

We propose a novel analysis method called ConTAM, and show with a large scale survey of existing and novel n-gram based contamination metrics across 13 benchmarks and 7 models from 2 different families that ConTAM can be used to better understand evaluation data contamination and its effects.

Specificity

Vector search with small radiuses

no code implementations16 Mar 2024 Gergely Szilvasy, Pierre-Emmanuel Mazaré, Matthijs Douze

Although convenient to compute, this metric is distantly related to the end-to-end accuracy of a full system that integrates vector search.

Image Retrieval Retrieval

In-context Pretraining: Language Modeling Beyond Document Boundaries

1 code implementation16 Oct 2023 Weijia Shi, Sewon Min, Maria Lomeli, Chunting Zhou, Margaret Li, Gergely Szilvasy, Rich James, Xi Victoria Lin, Noah A. Smith, Luke Zettlemoyer, Scott Yih, Mike Lewis

Large language models (LMs) are currently trained to predict tokens given document prefixes, enabling them to directly perform long-form generation and prompting-style tasks which can be reduced to document completion.

In-Context Learning Language Modelling +1

RA-DIT: Retrieval-Augmented Dual Instruction Tuning

no code implementations2 Oct 2023 Xi Victoria Lin, Xilun Chen, Mingda Chen, Weijia Shi, Maria Lomeli, Rich James, Pedro Rodriguez, Jacob Kahn, Gergely Szilvasy, Mike Lewis, Luke Zettlemoyer, Scott Yih

Retrieval-augmented language models (RALMs) improve performance by accessing long-tail and up-to-date knowledge from external data stores, but are challenging to build.

Ranked #21 on Question Answering on TriviaQA (using extra training data)

Few-Shot Learning Open-Domain Question Answering +1

Cannot find the paper you are looking for? You can Submit a new open access paper.