Search Results for author: Daniel Campos

Found 35 papers, 10 papers with code

ColBERT-serve: Efficient Multi-Stage Memory-Mapped Scoring

no code implementations21 Apr 2025 Kaili Huang, Thejas Venkatesh, Uma Dingankar, Antonio Mallia, Daniel Campos, Jian Jiao, Christopher Potts, Matei Zaharia, Kwabena Boahen, Omar Khattab, Saarthak Sarup, Keshav Santhanam

We study serving retrieval models, specifically late interaction models like ColBERT, to many concurrent users at once and under a small budget, in which the index may not fit in memory.

Support Evaluation for the TREC 2024 RAG Track: Comparing Human versus LLM Judges

no code implementations21 Apr 2025 Nandan Thakur, Ronak Pradeep, Shivani Upadhyay, Daniel Campos, Nick Craswell, Jimmy Lin

Retrieval-augmented generation (RAG) enables large language models (LLMs) to generate answers with citations from source documents containing "ground truth", thereby reducing system hallucinations.

The Great Nugget Recall: Automating Fact Extraction and RAG Evaluation with Large Language Models

no code implementations21 Apr 2025 Ronak Pradeep, Nandan Thakur, Shivani Upadhyay, Daniel Campos, Nick Craswell, Jimmy Lin

In the context of the TREC 2024 RAG Track, we calibrate a fully automatic approach against strategies where nuggets are created manually or semi-manually by human assessors and then assigned manually to system answers.

CORD: Balancing COnsistency and Rank Distillation for Robust Retrieval-Augmented Generation

no code implementations19 Dec 2024 Youngwon Lee, Seung-won Hwang, Daniel Campos, Filip Graliński, Zhewei Yao, Yuxiong He

With the adoption of retrieval-augmented generation (RAG), large language models (LLMs) are expected to ground their generation to the retrieved contexts.

Position RAG +1

Inference Scaling for Bridging Retrieval and Augmented Generation

no code implementations14 Dec 2024 Youngwon Lee, Seung-won Hwang, Daniel Campos, Filip Graliński, Zhewei Yao, Yuxiong He

Retrieval-augmented generation (RAG) has emerged as a popular approach to steering the output of a large language model (LLM) by incorporating retrieved contexts as inputs.

Language Modeling Language Modelling +2

Arctic-Embed 2.0: Multilingual Retrieval Without Compromise

no code implementations3 Dec 2024 Puxuan Yu, Luke Merrick, Gaurav Nuti, Daniel Campos

This paper presents the training methodology of Arctic-Embed 2. 0, a set of open-source text embedding models built for accurate and efficient multilingual retrieval.

Representation Learning Retrieval

Initial Nugget Evaluation Results for the TREC 2024 RAG Track with the AutoNuggetizer Framework

1 code implementation14 Nov 2024 Ronak Pradeep, Nandan Thakur, Shivani Upadhyay, Daniel Campos, Nick Craswell, Jimmy Lin

Within the TREC setup, we are able to calibrate our fully automatic process against a manual process whereby nuggets are created by human assessors semi-manually and then assigned manually to system answers.

Question Answering RAG

A Large-Scale Study of Relevance Assessments with Large Language Models: An Initial Look

no code implementations13 Nov 2024 Shivani Upadhyay, Ronak Pradeep, Nandan Thakur, Daniel Campos, Nick Craswell, Ian Soboroff, Hoa Trang Dang, Jimmy Lin

This paper reports on the results of a large-scale evaluation (the TREC 2024 RAG Track) where four different relevance assessment approaches were deployed in situ: the "standard" fully manual process that NIST has implemented for decades and three different alternatives that take advantage of LLMs to different extents using the open-source UMBRELA tool.

Information Retrieval RAG

SuffixDecoding: A Model-Free Approach to Speeding Up Large Language Model Inference

no code implementations7 Nov 2024 Gabriele Oliaro, Zhihao Jia, Daniel Campos, Aurick Qiao

For open-ended chat and code generation tasks, SuffixDecoding achieves up to $1. 4\times$ higher output throughput than SpecInfer and up to $1. 1\times$ lower time-per-token (TPOT) latency.

Code Generation Language Modeling +4

Ragnarök: A Reusable RAG Framework and Baselines for TREC 2024 Retrieval-Augmented Generation Track

1 code implementation24 Jun 2024 Ronak Pradeep, Nandan Thakur, Sahel Sharifymoghaddam, Eric Zhang, Ryan Nguyen, Daniel Campos, Nick Craswell, Jimmy Lin

In our work, we lay out the steps we've made towards making this track a reality -- we describe the details of our reusable framework, Ragnar\"ok, explain the curation of the new MS MARCO V2. 1 collection choice, release the development topics for the track, and standardize the I/O definitions which assist the end user.

Benchmarking RAG +1

Synthetic Test Collections for Retrieval Evaluation

1 code implementation13 May 2024 Hossein A. Rahmani, Nick Craswell, Emine Yilmaz, Bhaskar Mitra, Daniel Campos

Previous studies demonstrate that LLMs have the potential to generate synthetic relevance judgments for use in the evaluation of IR systems.

Information Retrieval Retrieval

Arctic-Embed: Scalable, Efficient, and Accurate Text Embedding Models

no code implementations8 May 2024 Luke Merrick, Danmei Xu, Gaurav Nuti, Daniel Campos

This report describes the training dataset creation and recipe behind the family of \texttt{arctic-embed} text embedding models (a set of five models ranging from 22 to 334 million parameters with weights open-sourced under an Apache-2 license).

Retrieval

Overview of the TREC 2023 Product Product Search Track

no code implementations14 Nov 2023 Daniel Campos, Surya Kallumadi, Corby Rosset, Cheng Xiang Zhai, Alessandro Magnani

The focus this year was the creation of a reusable collection and evaluation of the impact of the use of metadata and multi-modal data on retrieval accuracy.

Retrieval

Noise-Robust Dense Retrieval via Contrastive Alignment Post Training

no code implementations6 Apr 2023 Daniel Campos, ChengXiang Zhai, Alessandro Magnani

The success of contextual word representations and advances in neural information retrieval have made dense vector-based retrieval a standard approach for passage and document ranking.

Data Augmentation Document Ranking +3

To Asymmetry and Beyond: Structured Pruning of Sequence to Sequence Models for Improved Inference Efficiency

no code implementations5 Apr 2023 Daniel Campos, ChengXiang Zhai

Sequence-to-sequence language models can be used to produce abstractive summaries which are coherent, relevant, and concise.

Decoder

Dense Sparse Retrieval: Using Sparse Language Models for Inference Efficient Dense Retrieval

no code implementations31 Mar 2023 Daniel Campos, ChengXiang Zhai

Vector-based retrieval systems have become a common staple for academic and industrial search applications because they provide a simple and scalable way of extending the search to leverage contextual representations for documents and queries.

Retrieval TriviaQA

Quick Dense Retrievers Consume KALE: Post Training Kullback Leibler Alignment of Embeddings for Asymmetrical dual encoders

no code implementations31 Mar 2023 Daniel Campos, Alessandro Magnani, ChengXiang Zhai

In this paper, we consider the problem of improving the inference latency of language model-based dense retrieval systems by introducing structural compression and model size asymmetry between the context and query encoders.

Knowledge Distillation Language Modeling +4

oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes

no code implementations30 Mar 2023 Daniel Campos, Alexandre Marques, Mark Kurtz, ChengXiang Zhai

In this paper, we introduce the range of oBERTa language models, an easy-to-use set of language models which allows Natural Language Processing (NLP) practitioners to obtain between 3. 8 and 24. 3 times faster models without expertise in model compression.

Knowledge Distillation Model Compression +3

Compressing Cross-Lingual Multi-Task Models at Qualtrics

no code implementations29 Nov 2022 Daniel Campos, Daniel Perry, Samir Joshi, Yashmeet Gambhir, Wei Du, Zhengzheng Xing, Aaron Colak

Experience management is an emerging business area where organizations focus on understanding the feedback of customers and employees in order to improve their end-to-end experiences.

Management Model Compression +3

Sparse*BERT: Sparse Models Generalize To New tasks and Domains

no code implementations25 May 2022 Daniel Campos, Alexandre Marques, Tuan Nguyen, Mark Kurtz, ChengXiang Zhai

Our experimentation shows that models that are pruned during pretraining using general domain masked language models can transfer to novel domains and tasks without extensive hyperparameter exploration or specialized approaches.

Quantization

IMG2SMI: Translating Molecular Structure Images to Simplified Molecular-input Line-entry System

no code implementations3 Sep 2021 Daniel Campos, Heng Ji

A large portion of chemistry literature focuses on new molecules and reactions between molecules.

Image Captioning

Curriculum learning for language modeling

1 code implementation4 Aug 2021 Daniel Campos

Language Models like ELMo and BERT have provided robust representations of natural language, which serve as the language understanding component for a diverse range of downstream tasks. Curriculum learning is a method that employs a structured training regime instead, which has been leveraged in computer vision and machine translation to improve model training speed and model performance.

Language Modeling Language Modelling +2

MS MARCO: Benchmarking Ranking Models in the Large-Data Regime

no code implementations9 May 2021 Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, Jimmy Lin

Evaluation efforts such as TREC, CLEF, NTCIR and FIRE, alongside public leaderboard such as MS MARCO, are intended to encourage research and track our progress, addressing big questions in our field.

Benchmarking

TREC Deep Learning Track: Reusable Test Collections in the Large Data Regime

no code implementations19 Apr 2021 Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, Ellen M. Voorhees, Ian Soboroff

The TREC Deep Learning (DL) Track studies ad hoc search in the large data regime, meaning that a large set of human-labeled training data is available.

Selection bias

Informational entropy thresholds as a physical mechanism to explain power-law time distributions in sequential decision-making

no code implementations17 Feb 2021 Javier Cristín, Vicenç Méndez, Daniel Campos

While frameworks based on physical grounds (like the Drift-Diffusion Model) have been exhaustively used in psychology and neuroscience to describe perceptual decision-making in humans, analogous approaches for more complex situations like sequential (tree-like) decision making are still absent.

Decision Making Physics and Society Disordered Systems and Neural Networks Adaptation and Self-Organizing Systems

Overview of the TREC 2020 deep learning track

1 code implementation15 Feb 2021 Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos

This is the second year of the TREC Deep Learning Track, with the goal of studying ad hoc ranking in the large training data regime.

Deep Learning Passage Retrieval +1

ORCAS: 18 Million Clicked Query-Document Pairs for Analyzing Search

no code implementations9 Jun 2020 Nick Craswell, Daniel Campos, Bhaskar Mitra, Emine Yilmaz, Bodo Billerbeck

Users of Web search engines reveal their information needs through queries and clicks, making click logs a useful asset for information retrieval.

Information Retrieval Retrieval

On the Reliability of Test Collections for Evaluating Systems of Different Types

no code implementations28 Apr 2020 Emine Yilmaz, Nick Craswell, Bhaskar Mitra, Daniel Campos

As deep learning based models are increasingly being used for information retrieval (IR), a major challenge is to ensure the availability of test collections for measuring their quality.

Deep Learning Fairness +3

XGLUE: A New Benchmark Dataset for Cross-lingual Pre-training, Understanding and Generation

2 code implementations3 Apr 2020 Yaobo Liang, Nan Duan, Yeyun Gong, Ning Wu, Fenfei Guo, Weizhen Qi, Ming Gong, Linjun Shou, Daxin Jiang, Guihong Cao, Xiaodong Fan, Ruofei Zhang, Rahul Agrawal, Edward Cui, Sining Wei, Taroon Bharti, Ying Qiao, Jiun-Hung Chen, Winnie Wu, Shuguang Liu, Fan Yang, Daniel Campos, Rangan Majumder, Ming Zhou

In this paper, we introduce XGLUE, a new benchmark dataset that can be used to train large-scale cross-lingual pre-trained models using multilingual and bilingual corpora and evaluate their performance across a diverse set of cross-lingual tasks.

Natural Language Understanding XLM-R

Overview of the TREC 2019 deep learning track

2 code implementations17 Mar 2020 Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, Ellen M. Voorhees

The Deep Learning Track is a new track for TREC 2019, with the goal of studying ad hoc ranking in a large data regime.

Deep Learning Passage Retrieval +2

Open Domain Web Keyphrase Extraction Beyond Language Modeling

2 code implementations IJCNLP 2019 Lee Xiong, Chuan Hu, Chenyan Xiong, Daniel Campos, Arnold Overwijk

This paper studies keyphrase extraction in real-world scenarios where documents are from diverse domains and have variant content quality.

Keyphrase Extraction Language Modeling +1

MS MARCO: A Human Generated MAchine Reading COmprehension Dataset

13 code implementations28 Nov 2016 Payal Bajaj, Daniel Campos, Nick Craswell, Li Deng, Jianfeng Gao, Xiaodong Liu, Rangan Majumder, Andrew McNamara, Bhaskar Mitra, Tri Nguyen, Mir Rosenberg, Xia Song, Alina Stoica, Saurabh Tiwary, Tong Wang

The size of the dataset and the fact that the questions are derived from real user search queries distinguishes MS MARCO from other well-known publicly available datasets for machine reading comprehension and question-answering.

Benchmarking Machine Reading Comprehension +1

Cannot find the paper you are looking for? You can Submit a new open access paper.