Search Results for author: Michele Scarpiniti

Found 9 papers, 0 papers with code

AFAFed -- Protocol analysis

no code implementations29 Jun 2022 Enzo Baccarelli, Michele Scarpiniti, Alireza Momenzadeh, Sima Sarv Ahrabi

The convergence properties of AFAFed under (possibly) non-convex loss functions is guaranteed by a set of new analytical bounds, which formally unveil the impact on the resulting AFAFed convergence rate of a number of Federated Learning (FL) parameters, like, first and second moments of the per-coworker number of consecutive model updates, data skewness, communication packet-loss probability, and maximum/minimum values of the (adaptively tuned) mixing coefficient used for model aggregation.

Fairness Federated Learning

Gomoku: analysis of the game and of the player Wine

no code implementations1 Nov 2021 Lorenzo Piazzo, Michele Scarpiniti, Enzo Baccarelli

Gomoku, also known as five in a row, is a classical board game, ideally suited for quickly testing novel Artificial Intelligence (AI) techniques.

A New Class of Efficient Adaptive Filters for Online Nonlinear Modeling

no code implementations19 Apr 2021 Danilo Comminiello, Alireza Nezamdoust, Simone Scardapane, Michele Scarpiniti, Amir Hussain, Aurelio Uncini

In order to make this class of functional link adaptive filters (FLAFs) efficient, we propose low-complexity expansions and frequency-domain adaptation of the parameters.

Acoustic echo cancellation Domain Adaptation

Combined Sparse Regularization for Nonlinear Adaptive Filters

no code implementations24 Jul 2020 Danilo Comminiello, Michele Scarpiniti, Simone Scardapane, Luis A. Azpicueta-Ruiz, Aurelio Uncini

Nonlinear adaptive filters often show some sparse behavior due to the fact that not all the coefficients are equally useful for the modeling of any nonlinearity.

Why should we add early exits to neural networks?

no code implementations27 Apr 2020 Simone Scardapane, Michele Scarpiniti, Enzo Baccarelli, Aurelio Uncini

Deep neural networks are generally designed as a stack of differentiable layers, in which a prediction is obtained only after running the full stack.

A Multimodal Deep Network for the Reconstruction of T2W MR Images

no code implementations8 Aug 2019 Antonio Falvo, Danilo Comminiello, Simone Scardapane, Michele Scarpiniti, Aurelio Uncini

In this paper, we present a deep learning method that is able to reconstruct subsampled MR images obtained by reducing the k-space data, while maintaining a high image quality that can be used to observe brain lesions.

Efficient data augmentation using graph imputation neural networks

no code implementations20 Jun 2019 Indro Spinelli, Simone Scardapane, Michele Scarpiniti, Aurelio Uncini

Recently, data augmentation in the semi-supervised regime, where unlabeled data vastly outnumbers labeled data, has received a considerable attention.

Data Augmentation Imputation

Effective Blind Source Separation Based on the Adam Algorithm

no code implementations25 May 2016 Michele Scarpiniti, Simone Scardapane, Danilo Comminiello, Raffaele Parisi, Aurelio Uncini

In this paper, we derive a modified InfoMax algorithm for the solution of Blind Signal Separation (BSS) problems by using advanced stochastic methods.

blind source separation Stochastic Optimization

Cannot find the paper you are looking for? You can Submit a new open access paper.