8 code implementations • 28 Jul 2020 • Craig Macdonald, Nicola Tonellotto
The advent of deep machine learning platforms such as Tensorflow and Pytorch, developed in expressive high-level languages such as Python, have allowed more expressive representations of deep neural network architectures.
1 code implementation • 6 Jul 2022 • Aleksandr Petrov, Craig Macdonald
Hence, we propose a novel Recency-based Sampling of Sequences training objective that addresses both limitations.
1 code implementation • 15 Jul 2022 • Aleksandr Petrov, Craig Macdonald
We also propose our own implementation of BERT4Rec based on the Hugging Face Transformers library, which we demonstrate replicates the originally reported results on 3 out 4 datasets, while requiring up to 95% less training time to converge.
3 code implementations • 21 Jun 2021 • Xiao Wang, Craig Macdonald, Nicola Tonellotto, Iadh Ounis
In particular, based on the pseudo-relevant set of documents identified using a first-pass dense retrieval, we extract representative feedback embeddings (using KMeans clustering) -- while ensuring that these embeddings discriminate among passages (based on IDF) -- which are then added to the query representation.
Ranked #1 on TREC 2019 Passage Ranking on MSMARCO
1 code implementation • 13 Aug 2021 • Craig Macdonald, Nicola Tonellotto, Iadh Ounis
The advent of contextualised language models has brought gains in search effectiveness, not just when applied for re-ranking the output of classical weighting models such as BM25, but also when used directly for passage indexing and retrieval, a technique which is called dense retrieval.
1 code implementation • 23 Aug 2021 • Nicola Tonellotto, Craig Macdonald
Recent advances in dense retrieval techniques have offered the promise of being able not just to re-rank documents using contextualised language models such as BERT, but also to use such models to identify documents from the collection in the first place.
1 code implementation • 25 Aug 2021 • Craig Macdonald, Nicola Tonellotto
In this work, we investigate the use of ANN scores for ranking the candidate documents, in order to decrease the number of candidate documents being fully scored.
1 code implementation • 9 Jan 2023 • Mitko Gospodinov, Sean MacAvaney, Craig Macdonald
Doc2Query -- the process of expanding the content of a document before indexing using a sequence-to-sequence model -- has emerged as a prominent technique for improving the first-stage retrieval effectiveness of search engines.
2 code implementations • 14 Aug 2023 • Aleksandr Petrov, Craig Macdonald
A large catalogue size is one of the central challenges in training recommendation models: a large number of items makes them memory and computationally inefficient to compute scores for all items during training, forcing these models to deploy negative sampling.
2 code implementations • 18 Mar 2020 • Jimmy Lin, Joel Mackenzie, Chris Kamphuis, Craig Macdonald, Antonio Mallia, Michał Siedlaczek, Andrew Trotman, Arjen de Vries
There exists a natural tension between encouraging a diverse ecosystem of open-source search engines and supporting fair, replicable comparisons across those systems.
1 code implementation • 18 Aug 2022 • Sean MacAvaney, Nicola Tonellotto, Craig Macdonald
Search systems often employ a re-ranking pipeline, wherein documents (or passages) from an initial pool of candidates are assigned new ranking scores.
1 code implementation • 17 Sep 2019 • Zaiqiao Meng, Richard McCreadie, Craig Macdonald, Iadh Ounis
We train our VBCAR model based on the Bayesian Skip-gram framework coupled with the amortized variational inference so that it can learn more expressive latent representations that integrate both the non-linearity and Bayesian behaviour.
1 code implementation • 29 Mar 2024 • Aleksandr V. Petrov, Sean MacAvaney, Craig Macdonald
However, Cross-Encoders based on large transformer models (such as BERT or T5) are computationally expensive and allow for scoring only a small number of documents within a reasonably small latency window.
1 code implementation • 18 Apr 2021 • Amir H. Jadidinejad, Craig Macdonald, Iadh Ounis
Using the relative comparison of many recommendation models as in the typical offline evaluation of recommender systems, and based on the Kendall rank correlation coefficient, we show that our proposed evaluation methodology exhibits statistically significant improvements of 14% and 40% on the examined open loop datasets (Yahoo!
no code implementations • 22 Jun 2016 • Xiao Yang, Craig Macdonald, Iadh Ounis
In this paper, using a Twitter election classification task that aims to detect election-related tweets, we investigate the impact of the background dataset used to train the embedding models, the context window size and the dimensionality of word embeddings on the classification performance.
no code implementations • ACL 2014 • Miles Osborne, Sean Moran, Richard McCreadie, Alex Von Lunen, er, Martin Sykora, Elizabeth Cano, Neil Ireson, Craig Macdonald, Iadh Ounis, Yulan He, Tom Jackson, Fabio Ciravegna, Ann O{'}Brien
no code implementations • 5 Jul 2019 • Graham McDonald, Craig Macdonald, Iadh Ounis
However, many government documents contain sensitive information, such as personal or confidential information.
no code implementations • 15 Jul 2020 • Xiao Wang, Craig Macdonald, Iadh Ounis
Query reformulations have long been a key mechanism to alleviate the vocabulary-mismatch problem in information retrieval, for example by expanding the queries with related query terms or by generating paraphrases of the queries.
no code implementations • 26 Jul 2020 • Zaiqiao Meng, Richard McCreadie, Craig Macdonald, Iadh Ounis
In this paper, we both show that there is no standard splitting strategy and that the selection of splitting strategy can have a strong impact on the ranking of recommender systems.
no code implementations • 14 Aug 2020 • Xi Wang, Iadh Ounis, Craig Macdonald
However, a classification model that learns to classify binary instances with incomplete positive labels while assuming all unlabelled data to be negative examples will often generate a biased classifier.
no code implementations • 5 Feb 2021 • Xi Wang, Iadh Ounis, Craig Macdonald
Furthermore, inspired by the users' information adoption framework, we integrate two loss functions and a negative sampling strategy into our proposed RPRM model, to ensure that the properties of reviews are correlated with the users' preferences.
no code implementations • 8 Jul 2021 • Zaiqiao Meng, Siwei Liu, Craig Macdonald, Iadh Ounis
For the GCN-P model, two single-relational graphs are constructed from all the users' and items' side information respectively, to pre-train entity representations by using the Graph Convolutional Networks.
no code implementations • 9 Aug 2021 • Sean MacAvaney, Craig Macdonald, Roderick Murray-Smith, Iadh Ounis
Existing approaches often rely on massive query logs and interaction data to generate a variety of possible query intents, which then can be used to re-rank documents.
no code implementations • EMNLP (scai) 2020 • Sarawoot Kongyoung, Craig Macdonald, Iadh Ounis
Furthermore, we propose a novel hybrid dynamic method combining Abridged Linear for the main task with a Loss-Balanced Task Weighting (LBTW) for the auxiliary tasks, so as to automatically fine-tune task weighting during learning, ensuring that each of the task’s weights is adjusted by the relative importance of the different tasks.
no code implementations • 26 Nov 2021 • Sean MacAvaney, Craig Macdonald, Iadh Ounis
We present ir-measures, a new tool that makes it convenient to calculate a diverse set of evaluation measures used in information retrieval.
no code implementations • 21 Jan 2022 • Sean MacAvaney, Craig Macdonald, Iadh Ounis
Given that web documents are prone to change over time, we study the differences present between a version of the corpus containing documents as they appeared in 2017 (which has been used by several recent works) and a new version we construct that includes documents close to as they appeared at the time the query log was produced (2006).
no code implementations • 19 Nov 2022 • Ting Su, Craig Macdonald, Iadh Ounis
Our results show that the neural language models significantly outperform traditional TF. IDF and LSTM methods.
no code implementations • 19 Nov 2022 • Ting Su, Craig Macdonald, Iadh Ounis
We conclude that the Twitter users' friendship and followers network information can significantly outperform language-based approaches, as well as the existing state-of-the-art fake news detection models that use a more sophisticated network structure, in classifying fake news on Twitter.
no code implementations • 19 Jun 2023 • Aleksandr V. Petrov, Craig Macdonald
This paper presents the GPTRec sequential recommendation model, which is based on the GPT-2 architecture.
no code implementations • 1 Aug 2023 • Xiao Wang, Sean MacAvaney, Craig Macdonald, Iadh Ounis
GenQR directly reformulates the user's input query, while GenPRF provides additional context for the query by making use of pseudo-relevance feedback information.
no code implementations • 21 Aug 2023 • Zixuan Yi, Iadh Ounis, Craig Macdonald
In our research, we introduce the Personalised Graph Prompt-based Recommendation (PGPRec) framework.
no code implementations • 17 Oct 2023 • Maria Vlachou, Craig Macdonald
Our approach introduces a new setting for obtaining richer information on query differences in dense QPP that can explain potential unstable performance of existing predictors and outlines the unique characteristics of different query types on dense retrieval models.
no code implementations • 31 Oct 2023 • Zixuan Yi, Zijun Long, Iadh Ounis, Craig Macdonald, Richard McCreadie
In recent years, the rapid growth of online multimedia services, such as e-commerce platforms, has necessitated the development of personalised recommendation approaches that can encode diverse content about each item.
no code implementations • 27 Nov 2023 • Siwei Liu, Xi Wang, Craig Macdonald, Iadh Ounis
We propose a novel recommendation model, the Social-aware Gaussian Pre-trained model (SGP), which encodes the user social relations and interaction data at the pre-training stage in a Graph Neural Network (GNN).
no code implementations • 11 Dec 2023 • Aleksandr V. Petrov, Craig Macdonald
Sequential Recommendation is a popular recommendation task that uses the order of user-item interaction to model evolving users' interests and sequential patterns in their behaviour.
no code implementations • 11 Jan 2024 • Maria Vlachou, Craig Macdonald
Still, evaluating systems in offline settings with simulators suffers from problems, such as focusing entirely on a single target item (not addressing the exploratory nature of a recommender system), and exhibiting extreme patience (consistent feedback over a large number of turns).
1 code implementation • 7 Mar 2024 • Aleksandr Petrov, Craig Macdonald
Adaptations of Transformer models, such as BERT4Rec and SASRec, achieve state-of-the-art performance in the sequential recommendation task according to accuracy-based metrics, such as NDCG.