1 code implementation • 24 Feb 2021 • Robert-George Colt, Csongor-Huba Várady, Riccardo Volpi, Luigi Malagò
We focus on automatic feature extraction for raw audio heartbeat sounds, aimed at anomaly detection applications in healthcare.
no code implementations • 20 Sep 2020 • Goffredo Chirco, Luigi Malagò, Giovanni Pistone
In a non-parametric formalism, we consider the full set of positive probability functions on a finite sample space, and we provide a specific expression for the tangent and cotangent spaces over the statistical manifold, in terms of a Hilbert bundle structure that we call the Statistical Bundle.
1 code implementation • NeurIPS Workshop DL-IG 2020 • Csongor Várady, Riccardo Volpi, Luigi Malagò, Nihat Ay
These models are commonly trained using a two-step optimization algorithm called Wake-Sleep (WS) and more recently by improved versions, such as Reweighted Wake-Sleep (RWS) and Bidirectional Helmholtz Machines (BiHM).
no code implementations • 24 Jul 2020 • Alexandra-Ioana Albu, Alina Enescu, Luigi Malagò
In presence of an additional dataset of unlabelled data containing also anomalies, the task can be framed as a semi-supervised task with negative and unlabelled sample points.
no code implementations • 4 May 2020 • Héctor J. Hortúa, Riccardo Volpi, Luigi Malagò
Upcoming experiments such as Hydrogen Epoch of Reionization Array (HERA) and Square Kilometre Array (SKA) are intended to measure the 21cm signal over a wide range of redshifts, representing an incredible opportunity in advancing our understanding about the nature of cosmic Reionization.
no code implementations • 4 Dec 2019 • Riccardo Volpi, Luigi Malagò
Learning an embedding for a large collection of items is a popular approach to overcome the computational limitations associated to one-hot encodings.
2 code implementations • 19 Nov 2019 • Hector J. Hortua, Riccardo Volpi, Dimitri Marinelli, Luigi Malagò
In the second part of the paper, we present a guide to the training and calibration of a successful multi-channel BNN for the CMB temperature and polarization map.
no code implementations • ICLR Workshop DeepGenStruct 2019 • Septimia Sârbu, Luigi Malagò
In training, they exploit the power of variational inference, by optimizing a lower bound on the model evidence.
no code implementations • 5 Jul 2018 • Septimia Sârbu, Riccardo Volpi, Alexandra Peşte, Luigi Malagò
In this paper we propose two novel bounds for the log-likelihood based on Kullback-Leibler and the R\'{e}nyi divergences, which can be used for variational inference and in particular for the training of Variational AutoEncoders.