1 code implementation • 24 May 2023 • Błażej Leporowski, Arian Bakhtiarnia, Nicole Bonnici, Adrian Muscat, Luca Zanella, Yiming Wang, Alexandros Iosifidis
We introduce the first audio-visual dataset for traffic anomaly detection taken from real-world scenes, called MAVAD, with a diverse range of weather and illumination conditions.
no code implementations • 7 Feb 2023 • Brandon Birmingham, Adrian Muscat
To analyse the use and choice of keywords in context of this approach, this study analysed the generation of image captions based on (a) keywords extracted from gold standard captions and (b) from automatically detected keywords.
no code implementations • PVLAM (LREC) 2022 • Marc Tanti, Shaun Abdilla, Adrian Muscat, Claudia Borg, Reuben A. Farrugia, Albert Gatt
To encourage the development of more human-focused descriptions, we developed a new data set of facial descriptions based on the CelebA image data set.
1 code implementation • 14 May 2021 • Marc Tanti, Camille Berruyer, Paul Tafforeau, Adrian Muscat, Reuben Farrugia, Kenneth Scerri, Gianluca Valentino, V. Armando Solé, Johann A. Briffa
Propagation Phase Contrast Synchrotron Microtomography (PPC-SR${\mu}$CT) is the gold standard for non-invasive and non-destructive access to internal structures of archaeological remains.
no code implementations • 12 Jan 2021 • Stefan Cassar, Adrian Muscat, Dylan Seychell
The problem is casted as a classification one and geometrical features based on object bounding boxes, object labels and scene attributes are computed and used as inputs to pattern recognition models to predict relative depth.
no code implementations • 5 Jan 2021 • Daniel Cauchi, Adrian Muscat
When dealing with multi-class classification problems, it is common practice to build a model consisting of a series of binary classifiers using a learning paradigm which dictates how the classifiers are built and combined to discriminate between the individual classes.
1 code implementation • 28 Aug 2019 • Michael P. J. Camilleri, Adrian Muscat, Victor Buttigieg, Maria Attard
In this paper we describe $Vja\dot{g}\dot{g}$, a battery-aware journey detection algorithm that executes on the mobile device.
no code implementations • 26 Mar 2019 • Noel Mizzi, Adrian Muscat
Visual Relationship Detection is defined as, given an image composed of a subject and an object, the correct relation is predicted.
1 code implementation • WS 2018 • Anja Belz, Adrian Muscat, Pierre Anguill, Mouhamadou Sow, Ga{\'e}tan Vincent, Yassine Zinessabah
The dataset incorporates (i) the labelled object bounding boxes from VOC2008, (ii) geometrical, language and depth features for each object, and (iii) for each pair of objects in both orders, (a) the single best preposition and (b) the set of possible prepositions in the given language that describe the spatial relationship between the two objects.
no code implementations • WS 2018 • Br Birmingham, on, Adrian Muscat, Anja Belz
Detection of spatial relations between objects in images is currently a popular subject in image description research.
1 code implementation • 12 Oct 2018 • Marc Tanti, Albert Gatt, Adrian Muscat
Image caption generation systems are typically evaluated against reference outputs.
1 code implementation • LREC 2018 • Albert Gatt, Marc Tanti, Adrian Muscat, Patrizia Paggio, Reuben A. Farrugia, Claudia Borg, Kenneth P. Camilleri, Mike Rosner, Lonneke van der Plas
To gain a better understanding of the variation we find in face description and the possible issues that this may raise, we also conducted an annotation study on a subset of the corpus.
no code implementations • WS 2017 • Br Birmingham, on, Adrian Muscat
On the other hand, images with two image objects were better described with template-generated sentences composed of object labels and prepositions.
no code implementations • 15 Jan 2016 • Raffaella Bernardi, Ruket Cakici, Desmond Elliott, Aykut Erdem, Erkut Erdem, Nazli Ikizler-Cinbis, Frank Keller, Adrian Muscat, Barbara Plank
Automatic description generation from natural images is a challenging problem that has recently received a large amount of interest from the computer vision and natural language processing communities.