1 code implementation • 12 Oct 2023 • Giuseppe Floris, Raffaele Mura, Luca Scionis, Giorgio Piras, Maura Pintor, Ambra Demontis, Battista Biggio
Evaluating the adversarial robustness of machine learning models using gradient-based attacks is challenging.
no code implementations • 12 Oct 2023 • Giorgio Piras, Maura Pintor, Ambra Demontis, Battista Biggio
Neural network pruning has shown to be an effective technique for reducing the network size, trading desirable properties like generalization and robustness to adversarial attacks for higher sparsity.
no code implementations • 19 Sep 2023 • Emanuele Ledda, Daniele Angioni, Giorgio Piras, Giorgio Fumera, Battista Biggio, Fabio Roli
Machine-learning models can be fooled by adversarial examples, i. e., carefully-crafted input perturbations that force models to output wrong predictions.
no code implementations • 10 Aug 2022 • Giorgio Piras, Maura Pintor, Luca Demetrio, Battista Biggio
One of the most common causes of lack of continuity of online systems stems from a widely popular Cyber Attack known as Distributed Denial of Service (DDoS), in which a network of infected devices (botnet) gets exploited to flood the computational capacity of services through the commands of an attacker.