no code implementations • 19 Dec 2024 • Xueguang Ma, Shengyao Zhuang, Bevan Koopman, Guido Zuccon, Wenhu Chen, Jimmy Lin
Generation with source attribution is important for enhancing the verifiability of retrieval-augmented generation (RAG) systems.
1 code implementation • 26 Nov 2024 • Shuai Wang, Shengyao Zhuang, Bevan Koopman, Guido Zuccon
In this reproducibility study, we implement and evaluate both versions of 2D Matryoshka Training on STS tasks and extend our analysis to retrieval tasks.
1 code implementation • 17 Oct 2024 • Shengyao Zhuang, Shuai Wang, Bevan Koopman, Guido Zuccon
Effective approaches that can scale embedding model depth (i. e. layers) and embedding size allow for the creation of models that are highly scalable across different computational resources and task requirements.
1 code implementation • 9 Oct 2024 • Shengyao Zhuang, Bevan Koopman, Guido Zuccon
In this paper, we take a new look at Vec2Text and investigate how much of a threat it poses to the different attacks of corpus poisoning, whereby an attacker injects adversarial passages into a retrieval corpus with the intention of misleading dense retrievers.
no code implementations • 5 Aug 2024 • Ekaterina Khramtsova, Mahsa Baktashmotlagh, Guido Zuccon, Xi Wang, Mathieu Salzmann
In this work, we propose a source-free approach centred on uncertainty-based estimation, using a generative model for calibration in the absence of source data.
no code implementations • 9 Jul 2024 • Ekaterina Khramtsova, Teerapong Leelanupab, Shengyao Zhuang, Mahsa Baktashmotlagh, Guido Zuccon
In this demo we present a web-based application for selecting an effective pre-trained dense retriever to use on a private collection.
1 code implementation • 30 Jun 2024 • Xinyu Mao, Shengyao Zhuang, Bevan Koopman, Guido Zuccon
In this paper, we propose an alternative approach that still relies on neural models, but leverages dense representations and relevance feedback to enhance screening prioritisation, without the need for costly model fine-tuning and inference.
1 code implementation • 20 Jun 2024 • Shuoqi Sun, Shengyao Zhuang, Shuai Wang, Guido Zuccon
We provide a systematic understanding of the impact of specific components and wordings used in prompts on the effectiveness of rankers based on zero-shot Large Language Models (LLMs).
1 code implementation • 13 May 2024 • Ferdinand Schlatt, Maik Fröbe, Harrisen Scells, Shengyao Zhuang, Bevan Koopman, Guido Zuccon, Benno Stein, Martin Potthast, Matthias Hagen
Cross-encoders distilled from large language models (LLMs) are often more effective re-rankers than cross-encoders fine-tuned on manually labeled data.
1 code implementation • 29 Apr 2024 • Shengyao Zhuang, Xueguang Ma, Bevan Koopman, Jimmy Lin, Guido Zuccon
Utilizing large language models (LLMs) for zero-shot document ranking is done in one of two ways: (1) prompt-based re-ranking methods, which require no further training but are only feasible for re-ranking a handful of candidate documents due to computational costs; and (2) unsupervised contrastive trained dense retrieval methods, which can retrieve relevant documents from the entire corpus but require a large amount of paired text data for contrastive training.
1 code implementation • 10 Apr 2024 • Ferdinand Schlatt, Maik Fröbe, Harrisen Scells, Shengyao Zhuang, Bevan Koopman, Guido Zuccon, Benno Stein, Martin Potthast, Matthias Hagen
Existing cross-encoder re-rankers can be categorized as pointwise, pairwise, or listwise models.
1 code implementation • 20 Feb 2024 • Shengyao Zhuang, Bevan Koopman, Xiaoran Chu, Guido Zuccon
The emergence of Vec2Text -- a method for text embedding inversion -- has raised serious privacy concerns for dense retrieval systems which use text embeddings, such as those offered by OpenAI and Cohere.
no code implementations • 19 Feb 2024 • Shuai Wang, Ekaterina Khramtsova, Shengyao Zhuang, Guido Zuccon
Federated search systems aggregate results from multiple search engines, selecting appropriate sources to enhance result quality and align with user intent.
no code implementations • 19 Feb 2024 • Shuai Wang, Shengyao Zhuang, Guido Zuccon
With this respect, we identify three avenues, each characterised by different trade-offs in terms of computational cost, effectiveness and robustness : (1) use LLMs to stem the vocabulary for a collection, i. e., the set of unique words that appear in the collection (vocabulary stemming), (2) use LLMs to stem each document separately (contextual stemming), and (3) use LLMs to extract from each document entities that should not be stemmed, then use vocabulary stemming to stem the rest of the terms (entity-based contextual stemming).
1 code implementation • 7 Feb 2024 • Ekaterina Khramtsova, Shengyao Zhuang, Mahsa Baktashmotlagh, Guido Zuccon
In this paper we present Large Language Model Assisted Retrieval Model Ranking (LARMOR), an effective unsupervised approach that leverages LLMs for selecting which dense retriever to use on a test corpus (target).
no code implementations • 31 Jan 2024 • Shuai Wang, Shengyao Zhuang, Bevan Koopman, Guido Zuccon
Our ReSLLM method exploits LLMs to drive the selection of resources in federated search in a zero-shot setting.
1 code implementation • 24 Jan 2024 • Shuyi Wang, Bing Liu, Guido Zuccon
In a FOLTR system, a ranker is learned by aggregating local updates to the global ranking model.
no code implementations • 24 Jan 2024 • Hang Li, Chuting Yu, Ahmed Mourad, Bevan Koopman, Guido Zuccon
This paper considers Pseudo-Relevance Feedback (PRF) methods for dense retrievers in a resource constrained environment such as that of cheap cloud instances or embedded systems (e. g., smartphones and smartwatches), where memory and CPU are limited and GPUs are not present.
1 code implementation • 16 Jan 2024 • Xinyu Mao, Bevan Koopman, Guido Zuccon
In this context, we show that there is no need for further pre-training if a domain-specific BERT backbone is used within the active learning pipeline.
no code implementations • 12 Jan 2024 • Shuai Wang, Harrisen Scells, Shengyao Zhuang, Martin Potthast, Bevan Koopman, Guido Zuccon
Systematic reviews are crucial for evidence-based medicine as they comprehensively analyse published research findings on specific questions.
no code implementations • 3 Jan 2024 • Shengyao Zhuang, Bevan Koopman, Guido Zuccon
We describe team ielab from CSIRO and The University of Queensland's approach to the 2023 TREC Clinical Trials Track.
no code implementations • 8 Nov 2023 • Lukas Gienapp, Harrisen Scells, Niklas Deckers, Janek Bevendorff, Shuai Wang, Johannes Kiesel, Shahbaz Syed, Maik Fröbe, Guido Zuccon, Benno Stein, Matthias Hagen, Martin Potthast
To lay a foundation for developing new evaluation methods for generative retrieval systems, we survey the relevant literature from the fields of information retrieval and natural language processing, identify search tasks and system architectures in generative retrieval, develop a new user model, and study its operationalization.
1 code implementation • 20 Oct 2023 • Shengyao Zhuang, Bing Liu, Bevan Koopman, Guido Zuccon
In the field of information retrieval, Query Likelihood Models (QLMs) rank documents based on the probability of generating the query given the content of a document.
1 code implementation • 14 Oct 2023 • Shengyao Zhuang, Honglei Zhuang, Bevan Koopman, Guido Zuccon
We propose a novel zero-shot document ranking approach based on Large Language Models (LLMs): the Setwise prompting approach.
no code implementations • 18 Sep 2023 • Ekaterina Khramtsova, Shengyao Zhuang, Mahsa Baktashmotlagh, Xi Wang, Guido Zuccon
We propose the new problem of choosing which dense retrieval model to use when searching on a new collection for which no labels are available, i. e. in a zero-shot setting.
no code implementations • 17 Sep 2023 • Guido Zuccon, Bevan Koopman, Razia Shaik
We find that ChatGPT provides correct or partially correct answers in about half of the cases (50. 6% of the times), but its suggested references only exist 14% of the times.
1 code implementation • 12 Sep 2023 • Sophia Althammer, Guido Zuccon, Sebastian Hofstätter, Suzan Verberne, Allan Hanbury
We further find that gains provided by AL strategies come at the expense of more assessments (thus higher annotation costs) and AL strategies underperform random selection when comparing effectiveness given a fixed annotation cost.
1 code implementation • 11 Sep 2023 • Shuai Wang, Harrisen Scells, Martin Potthast, Bevan Koopman, Guido Zuccon
Our best approach is not only viable based on the information available at the time of screening, but also has similar effectiveness to the final title.
no code implementations • 4 Jul 2023 • Shuyi Wang, Guido Zuccon
For this, FOLTR trains learning to rank models in an online manner -- i. e. by exploiting users' interactions with the search systems (queries, clicks), rather than labels -- and federatively -- i. e. by not aggregating interaction data in a central server for training purposes, but by training instances of a model on each user device on their own private data, and then sharing the model updates, not the data, across a set of users that have formed the federation.
no code implementations • 30 Jun 2023 • Wojciech Kusa, Guido Zuccon, Petr Knoth, Allan Hanbury
We find that accounting for the difference in review outcomes leads to a different assessment of the quality of a system than if traditional evaluation measures were used.
1 code implementation • 29 Jun 2023 • Joel Mackenzie, Shengyao Zhuang, Guido Zuccon
The SPLADE (SParse Lexical AnD Expansion) model is a highly effective approach to learned sparse retrieval, where documents are represented by term impact scores derived from large language models.
1 code implementation • 29 Jun 2023 • Guido Zuccon, Harrisen Scells, Shengyao Zhuang
As in other fields of artificial intelligence, the information retrieval community has grown interested in investigating the power consumption associated with neural models, particularly models of search.
1 code implementation • 6 May 2023 • Shengyao Zhuang, Linjun Shou, Guido Zuccon
Effective cross-lingual dense retrieval methods that rely on multilingual pre-trained language models (PLMs) need to be trained to encompass both the relevance matching task and the cross-language alignment task.
1 code implementation • 17 Apr 2023 • Shengyao Zhuang, Linjun Shou, Jian Pei, Ming Gong, Houxing Ren, Guido Zuccon, Daxin Jiang
To address this challenge, we propose ToRoDer (TypOs-aware bottlenecked pre-training for RObust DEnse Retrieval), a novel re-training strategy for DRs that increases their robustness to misspelled queries while preserving their effectiveness in downstream retrieval tasks.
no code implementations • 23 Feb 2023 • Guido Zuccon, Bevan Koopman
Aside from measuring the effectiveness of ChatGPT in this context, we show that the knowledge passed in the prompt can overturn the knowledge encoded in the model and this is, in our experiments, to the detriment of answer correctness.
no code implementations • 3 Feb 2023 • Shuai Wang, Harrisen Scells, Bevan Koopman, Guido Zuccon
The ability of ChatGPT to follow complex instructions and generate queries with high precision makes it a valuable tool for researchers conducting systematic reviews, particularly for rapid reviews where time is a constraint and often trading-off higher precision for lower recall is acceptable.
1 code implementation • 21 Dec 2022 • Bevan Koopman, Ahmed Mourad, Hang Li, Anton van der Vegt, Shengyao Zhuang, Simon Gibson, Yash Dang, David Lawrence, Guido Zuccon
On the basis of these needs we release an information retrieval test collection comprising real questions, a large collection of scientific documents split in passages, and ground truth relevance assessments indicating which passages are relevant to each question.
no code implementations • 18 Dec 2022 • Shuai Wang, Harrisen Scells, Bevan Koopman, Guido Zuccon
An empirical analysis compares how effective neural methods compare to traditional methods for this task.
1 code implementation • 18 Dec 2022 • Shuai Wang, Hang Li, Guido Zuccon
One challenge to creating an effective systematic review Boolean query is the selection of effective MeSH Terms to include in the query.
1 code implementation • 29 Nov 2022 • Bing Liu, Harrisen Scells, Wen Hua, Guido Zuccon, Genghong Zhao, Xia Zhang
Making compatible predictions thus should be one of the goals of training an EA model along with fitting the labelled data: this aspect however is neglected in current methods.
1 code implementation • 29 Nov 2022 • Bing Liu, Tiancheng Lan, Wen Hua, Guido Zuccon
Entity Alignment (EA), which aims to detect entity mappings (i. e. equivalent entity pairs) in different Knowledge Graphs (KGs), is critical for KG fusion.
1 code implementation • 19 Sep 2022 • Shuai Wang, Harrisen Scells, Bevan Koopman, Guido Zuccon
However, identifying the correct MeSH terms to include in a query is difficult: information experts are often unfamiliar with the MeSH database and unsure about the appropriateness of MeSH terms for a query.
1 code implementation • 22 Aug 2022 • Bing Liu, Wen Hua, Guido Zuccon, Genghong Zhao, Xia Zhang
To include in the EA subtasks a high proportion of the potential mappings originally present in the large EA task, we devise a counterpart discovery method that exploits the locality principle of the EA task and the power of trained EA models.
no code implementations • 9 Jul 2022 • Ekaterina Khramtsova, Guido Zuccon, Xi Wang, Mahsa Baktashmotlagh
This paper performs a detailed analysis of the effectiveness of topological properties for image classification in various training scenarios, defined by: the number of training samples, the complexity of the training data and the complexity of the backbone network.
1 code implementation • 21 Jun 2022 • Shengyao Zhuang, Houxing Ren, Linjun Shou, Jian Pei, Ming Gong, Guido Zuccon, Daxin Jiang
This problem is further exacerbated when using DSI for cross-lingual retrieval, where document text and query text are in different languages.
no code implementations • 12 May 2022 • Hang Li, Ahmed Mourad, Bevan Koopman, Guido Zuccon
Pseudo-Relevance Feedback (PRF) assumes that the top results retrieved by a first-stage ranker are relevant to the original query and uses them to improve the query representation for a second round of retrieval.
no code implementations • 30 Apr 2022 • Hang Li, Shuai Wang, Shengyao Zhuang, Ahmed Mourad, Xueguang Ma, Jimmy Lin, Guido Zuccon
In this paper we consider the problem of combining the relevance signals from sparse and dense retrievers in the context of Pseudo Relevance Feedback (PRF).
1 code implementation • 20 Apr 2022 • Shuyi Wang, Guido Zuccon
A well-known factor that affects the performance of federated learning systems, and that poses serious challenges to these approaches, is that there may be some type of bias in the way data is distributed across clients.
1 code implementation • 6 Apr 2022 • Shuai Wang, Harrisen Scells, Justin Clark, Bevan Koopman, Guido Zuccon
However, we show pseudo seed studies are not representative of real seed studies used by information specialists.
1 code implementation • 1 Apr 2022 • Shengyao Zhuang, Guido Zuccon
We then demonstrate that the root cause of this resides in the input tokenization strategy employed by BERT.
1 code implementation • 1 Apr 2022 • Shengyao Zhuang, Hang Li, Guido Zuccon
We then exploit such historic implicit interactions to improve the effectiveness of a DR. A key challenge that we study is the effect that biases in the click signal, such as position bias, have on the DRs.
1 code implementation • 25 Feb 2022 • Shengyao Zhuang, Guido Zuccon
A simple and efficient strategy to validate deep learning checkpoints is the addition of validation loops to execute during training.
no code implementations • 15 Feb 2022 • Daniel Locke, Guido Zuccon
Case law retrieval is the retrieval of judicial decisions relevant to a legal question.
1 code implementation • 5 Jan 2022 • Shengyao Zhuang, Zhihao Qiao, Guido Zuccon
Online learning to rank (OLTR) aims to learn a ranker directly from implicit feedback derived from users' interactions, such as clicks.
1 code implementation • 13 Dec 2021 • Hang Li, Shengyao Zhuang, Ahmed Mourad, Xueguang Ma, Jimmy Lin, Guido Zuccon
Finally, we contribute a study of the generalisability of the ANCE-PRF method when dense retrievers other than ANCE are used for the first round of retrieval and for encoding the PRF signal.
1 code implementation • 8 Dec 2021 • Shuai Wang, Harrisen Scells, Ahmed Mourad, Guido Zuccon
Our results also indicate that our reproduced screening prioritisation method, (1) is generalisable across datasets of similar and different topicality compared to the original implementation, (2) that when using multiple seed studies, the effectiveness of the method increases using our techniques to enable this, (3) and that the use of multiple seed studies produces more stable rankings compared to single seed studies.
1 code implementation • EMNLP 2021 • Bing Liu, Harrisen Scells, Guido Zuccon, Wen Hua, Genghong Zhao
Entity Alignment (EA) aims to match equivalent entities across different Knowledge Graphs (KGs) and is an essential step of KG fusion.
2 code implementations • EMNLP 2021 • Shengyao Zhuang, Guido Zuccon
Our experimental results on the MS MARCO passage ranking dataset show that, with our proposed typos-aware training, DR and BERT re-ranker can become robust to typos in queries, resulting in significantly improved effectiveness compared to models trained without appropriately accounting for typos.
1 code implementation • 25 Aug 2021 • Hang Li, Ahmed Mourad, Shengyao Zhuang, Bevan Koopman, Guido Zuccon
Text-based PRF results show that the use of PRF had a mixed effect on deep rerankers across different datasets.
1 code implementation • 19 Aug 2021 • Shengyao Zhuang, Guido Zuccon
BERT-based information retrieval models are expensive, in both time (query latency) and computational resources (energy, hardware cost), making many of these models impractical especially under resource constraints.
no code implementations • ALTA 2016 • Mahnoosh Kholghi, Lance De Vine, Laurianne Sitbon, Guido Zuccon, Anthony Nguyen
This study investigates the use of unsupervised word embeddings and sequence features for sample representation in an active learning framework built to extract clinical concepts from clinical free text.
no code implementations • LREC 2016 • Lorraine Goeuriot, Liadh Kelly, Guido Zuccon, Joao Palotti
In this paper we present the datasets created by CLEF eHealth Lab from 2013-2015 for evaluation of search solutions to support common people finding health information online.