Search Results for author: Negar Arabzadeh

Found 19 papers, 9 papers with code

Generative Information Retrieval Evaluation

no code implementations11 Apr 2024 Marwah Alaofi, Negar Arabzadeh, Charles L. A. Clarke, Mark Sanderson

We resolve this apparent circularity in two ways: 1) by viewing LLM-based assessment as a form of "slow search", where a slower IR system is used for evaluation and training of a faster production IR system; and 2) by recognizing a continuing need to ground evaluation in human assessment, even if the characteristics of that human assessment must change.

Information Retrieval Retrieval

A Comparison of Methods for Evaluating Generative IR

2 code implementations5 Apr 2024 Negar Arabzadeh, Charles L. A. Clarke

Given that Gen-IR systems do not generate responses from a fixed set, we assume that methods for Gen-IR evaluation must largely depend on LLM-generated labels.

Information Retrieval Language Modelling +2

Query Performance Prediction using Relevance Judgments Generated by Large Language Models

1 code implementation1 Apr 2024 Chuan Meng, Negar Arabzadeh, Arian Askari, Mohammad Aliannejadi, Maarten de Rijke

This allows us to predict any IR evaluation measure using the generated relevance judgments as pseudo-labels; Also, this allows us to interpret predicted IR evaluation measures, and identify, track and rectify errors in generated relevance judgments to improve QPP quality.

Information Retrieval Language Modelling +2

Towards better Human-Agent Alignment: Assessing Task Utility in LLM-Powered Applications

no code implementations14 Feb 2024 Negar Arabzadeh, Julia Kiseleva, Qingyun Wu, Chi Wang, Ahmed Awadallah, Victor Dibia, Adam Fourney, Charles Clarke

The rapid development in the field of Large Language Models (LLMs) has led to a surge in applications that facilitate collaboration among multiple agents to assist humans in their daily tasks.

Math

Fréchet Distance for Offline Evaluation of Information Retrieval Systems with Sparse Labels

no code implementations31 Jan 2024 Negar Arabzadeh, Charles L. A. Clarke

The rapid advancement of natural language processing, information retrieval (IR), computer vision, and other technologies has presented significant challenges in evaluating the performance of these systems.

Information Retrieval Retrieval +1

Adapting Standard Retrieval Benchmarks to Evaluate Generated Answers

no code implementations9 Jan 2024 Negar Arabzadeh, Amin Bigdeli, Charles L. A. Clarke

In the second, we compare generated answers to the top results retrieved by a diverse set of retrieval models, ranging from traditional approaches to advanced methods, allowing us to measure improvements without human judgments.

Information Retrieval Retrieval

Retrieving Supporting Evidence for Generative Question Answering

no code implementations20 Sep 2023 Siqing Huo, Negar Arabzadeh, Charles L. A. Clarke

After presenting a question to an LLM and receiving a generated answer, we query the corpus with the combination of the question + generated answer.

Generative Question Answering Open-Domain Question Answering +1

Retrieving Supporting Evidence for LLMs Generated Answers

no code implementations23 Jun 2023 Siqing Huo, Negar Arabzadeh, Charles L. A. Clarke

After presenting a question to an LLM and receiving a generated answer, we query the corpus with the combination of the question + generated answer.

Open-Domain Question Answering Retrieval

Unsupervised Question Clarity Prediction Through Retrieved Item Coherency

no code implementations9 Aug 2022 Negar Arabzadeh, Mahsa Seifikar, Charles L. A. Clarke

While the research community has paid substantial attention to the problem of predicting query ambiguity in traditional search contexts, researchers have paid relatively little attention to predicting when this ambiguity is sufficient to warrant clarification in the context of conversational systems.

Conversational Question Answering Retrieval

Early Stage Sparse Retrieval with Entity Linking

no code implementations9 Aug 2022 Dahlia Shehata, Negar Arabzadeh, Charles L. A. Clarke

In this work, we propose boosting the performance of sparse retrievers by expanding both the queries and the documents with linked entities in two formats for the entity names: 1) explicit and 2) hashed.

Entity Linking Information Retrieval +1

PREME: Preference-based Meeting Exploration through an Interactive Questionnaire

1 code implementation5 May 2022 Negar Arabzadeh, Ali Ahmadvand, Julia Kiseleva, Yang Liu, Ahmed Hassan Awadallah, Ming Zhong, Milad Shokouhi

The recent increase in the volume of online meetings necessitates automated tools for managing and organizing the material, especially when an attendee has missed the discussion and needs assistance in quickly exploring it.

Making Large Language Models Interactive: A Pioneer Study on Supporting Complex Information-Seeking Tasks with Implicit Constraints

no code implementations2 May 2022 Ali Ahmadvand, Negar Arabzadeh, Julia Kiseleva, Patricio Figueroa Sanz, Xin Deng, Sujay Jauhar, Michael Gamon, Eugene Agichtein, Ned Friend, Aniruddha

Current interactive systems with natural language interfaces lack the ability to understand a complex information-seeking request which expresses several implicit constraints at once, and there is no prior information about user preferences e. g.,"find hiking trails around San Francisco which are accessible with toddlers and have beautiful scenery in summer", where output is a list of possible suggestions for users to start their exploration.

Hallucination Retrieval

Predicting Efficiency/Effectiveness Trade-offs for Dense vs. Sparse Retrieval Strategy Selection

no code implementations22 Sep 2021 Negar Arabzadeh, Xinyi Yan, Charles L. A. Clarke

These hybrid retrievers leverage low-cost, exact-matching based sparse retrievers along with dense retrievers to bridge the semantic gaps between query and documents.

Information Retrieval Retrieval

Shallow pooling for sparse labels

1 code implementation31 Aug 2021 Negar Arabzadeh, Alexandra Vtyurina, Xinyi Yan, Charles L. A. Clarke

To test this observation, we employed crowdsourced workers to make preference judgments between the top item returned by a modern neural ranking stack and a judged relevant item.

Passage Ranking

Cannot find the paper you are looking for? You can Submit a new open access paper.