no code implementations • 25 Jan 2021 • Arnaud Van Looveren, Janis Klaise, Giovanni Vacanti, Oliver Cobb
Counterfactual instances offer human-interpretable insight into the local behaviour of machine learning models.
1 code implementation • 13 Jul 2020 • Janis Klaise, Arnaud Van Looveren, Clive Cox, Giovanni Vacanti, Alexandru Coca
The machine learning lifecycle extends beyond the deployment stage.
1 code implementation • 21 Feb 2020 • Giovanni Vacanti, Arnaud Van Looveren
We present a novel adversarial detection and correction method for machine learning classifiers. The detector consists of an autoencoder trained with a custom loss function based on the Kullback-Leibler divergence between the classifier predictions on the original and reconstructed instances. The method is unsupervised, easy to train and does not require any knowledge about the underlying attack.