Search Results for author: Nils Lukas

Found 10 papers, 6 papers with code

Universal Backdoor Attacks

1 code implementation30 Nov 2023 Benjamin Schneider, Nils Lukas, Florian Kerschbaum

We demonstrate the effectiveness and robustness of our universal backdoor attacks by controlling models with up to 6, 000 classes while poisoning only 0. 15% of the training dataset.

Data Poisoning

Leveraging Optimization for Adaptive Attacks on Image Watermarks

no code implementations29 Sep 2023 Nils Lukas, Abdulrahman Diaa, Lucas Fenaux, Florian Kerschbaum

A core security property of watermarking is robustness, which states that an attacker can only evade detection by substantially degrading image quality.

Pick your Poison: Undetectability versus Robustness in Data Poisoning Attacks

no code implementations7 May 2023 Nils Lukas, Florian Kerschbaum

Our research points to intrinsic flaws in current attack evaluation methods and raises the bar for all data poisoning attackers who must delicately balance this trade-off to remain robust and undetectable.

Data Poisoning Image Classification

PTW: Pivotal Tuning Watermarking for Pre-Trained Image Generators

2 code implementations14 Apr 2023 Nils Lukas, Florian Kerschbaum

We propose an adaptive attack that can successfully remove any watermarking with access to only 200 non-watermarked images.

DeepFake Detection Face Swapping

Analyzing Leakage of Personally Identifiable Information in Language Models

1 code implementation1 Feb 2023 Nils Lukas, Ahmed Salem, Robert Sim, Shruti Tople, Lukas Wutschitz, Santiago Zanella-Béguelin

Understanding the risk of LMs leaking Personally Identifiable Information (PII) has received less attention, which can be attributed to the false assumption that dataset curation techniques such as scrubbing are sufficient to prevent PII leakage.

Sentence

Feature Grinding: Efficient Backdoor Sanitation in Deep Neural Networks

no code implementations29 Sep 2021 Nils Lukas, Charles Zhang, Florian Kerschbaum

Feature Grinding requires at most six percent of the model's training time on CIFAR-10 and at most two percent on ImageNet for sanitizing the surveyed backdoors.

Backdoor Attack

SoK: How Robust is Image Classification Deep Neural Network Watermarking? (Extended Version)

1 code implementation11 Aug 2021 Nils Lukas, Edward Jiang, Xinda Li, Florian Kerschbaum

Watermarking should be robust against watermark removal attacks that derive a surrogate model that evades provenance verification.

Image Classification

Deep Neural Network Fingerprinting by Conferrable Adversarial Examples

1 code implementation ICLR 2021 Nils Lukas, Yuxuan Zhang, Florian Kerschbaum

We propose a fingerprinting method for deep neural network classifiers that extracts a set of inputs from the source model so that only surrogates agree with the source model on the classification of such inputs.

Model extraction Transfer Learning

On the Robustness of the Backdoor-based Watermarking in Deep Neural Networks

no code implementations18 Jun 2019 Masoumeh Shafieinejad, Jiaqi Wang, Nils Lukas, Xinda Li, Florian Kerschbaum

We focus on backdoor-based watermarking and propose two -- a black-box and a white-box -- attacks that remove the watermark.

Cannot find the paper you are looking for? You can Submit a new open access paper.