Search Results for author: Shruti Tople

Found 27 papers, 13 papers with code

Closed-Form Bounds for DP-SGD against Record-level Inference

no code implementations22 Feb 2024 Giovanni Cherubin, Boris Köpf, Andrew Paverd, Shruti Tople, Lukas Wutschitz, Santiago Zanella-Béguelin

This paper presents a new approach to evaluate the privacy of machine learning models against specific record-level threats, such as membership and attribute inference, without the indirection through DP.

Attribute

Rethinking Privacy in Machine Learning Pipelines from an Information Flow Control Perspective

no code implementations27 Nov 2023 Lukas Wutschitz, Boris Köpf, Andrew Paverd, Saravan Rajmohan, Ahmed Salem, Shruti Tople, Santiago Zanella-Béguelin, Menglin Xia, Victor Rühle

In this paper, we take an information flow control perspective to describe machine learning systems, which allows us to leverage metadata such as access control policies and define clear-cut privacy and confidentiality guarantees with interpretable information flows.

Retrieval

SoK: Memorization in General-Purpose Large Language Models

no code implementations24 Oct 2023 Valentin Hartmann, Anshuman Suri, Vincent Bindschaedler, David Evans, Shruti Tople, Robert West

A major part of this success is due to their huge training datasets and the unprecedented number of model parameters, which allow them to memorize large amounts of information contained in the training data.

Memorization Question Answering

Why Train More? Effective and Efficient Membership Inference via Memorization

no code implementations12 Oct 2023 Jihye Choi, Shruti Tople, Varun Chandrasekaran, Somesh Jha

Many practical black-box MIAs require query access to the data distribution (the same distribution where the private data is drawn) to train shadow models.

Memorization

Investigating the Effect of Misalignment on Membership Privacy in the White-box Setting

1 code implementation8 Jun 2023 Ana-Maria Cretu, Daniel Jones, Yves-Alexandre de Montjoye, Shruti Tople

We here present the first systematic analysis of the causes of misalignment in shadow models and show the use of a different weight initialisation to be the main cause.

On the Efficacy of Differentially Private Few-shot Image Classification

1 code implementation2 Feb 2023 Marlon Tobaben, Aliaksandra Shysheya, John Bronskill, Andrew Paverd, Shruti Tople, Santiago Zanella-Beguelin, Richard E Turner, Antti Honkela

There has been significant recent progress in training differentially private (DP) models which achieve accuracy that approaches the best non-private models.

Federated Learning Few-Shot Image Classification

Analyzing Leakage of Personally Identifiable Information in Language Models

1 code implementation1 Feb 2023 Nils Lukas, Ahmed Salem, Robert Sim, Shruti Tople, Lukas Wutschitz, Santiago Zanella-Béguelin

Understanding the risk of LMs leaking Personally Identifiable Information (PII) has received less attention, which can be attributed to the false assumption that dataset curation techniques such as scrubbing are sufficient to prevent PII leakage.

Sentence

Invariant Aggregator for Defending against Federated Backdoor Attacks

no code implementations4 Oct 2022 Xiaoyang Wang, Dimitrios Dimitriadis, Sanmi Koyejo, Shruti Tople

Federated learning enables training high-utility models across several clients without directly sharing their private data.

Federated Learning Model Optimization

Membership Inference Attacks and Generalization: A Causal Perspective

1 code implementation18 Sep 2022 Teodora Baluta, Shiqi Shen, S. Hitarth, Shruti Tople, Prateek Saxena

Our causal models also show a new connection between generalization and MI attacks via their shared causal factors.

Distribution inference risks: Identifying and mitigating sources of leakage

2 code implementations18 Sep 2022 Valentin Hartmann, Léo Meynent, Maxime Peyrard, Dimitrios Dimitriadis, Shruti Tople, Robert West

We identify three sources of leakage: (1) memorizing specific information about the $\mathbb{E}[Y|X]$ (expected label given the feature values) of interest to the adversary, (2) wrong inductive bias of the model, and (3) finiteness of the training data.

Inductive Bias

Bayesian Estimation of Differential Privacy

1 code implementation10 Jun 2022 Santiago Zanella-Béguelin, Lukas Wutschitz, Shruti Tople, Ahmed Salem, Victor Rühle, Andrew Paverd, Mohammad Naseri, Boris Köpf, Daniel Jones

Our Bayesian method exploits the hypothesis testing interpretation of differential privacy to obtain a posterior for $\varepsilon$ (not just a confidence interval) from the joint posterior of the false positive and false negative rates of membership inference attacks.

The Connection between Out-of-Distribution Generalization and Privacy of ML Models

1 code implementation7 Oct 2021 Divyat Mahajan, Shruti Tople, Amit Sharma

Through extensive evaluation on a synthetic dataset and image datasets like MNIST, Fashion-MNIST, and Chest X-rays, we show that a lower OOD generalization gap does not imply better robustness to MI attacks.

Domain Generalization Out-of-Distribution Generalization

Grey-box Extraction of Natural Language Models

no code implementations1 Jan 2021 Santiago Zanella-Beguelin, Shruti Tople, Andrew Paverd, Boris Köpf

This is true even for queries that are entirely in-distribution, making extraction attacks indistinguishable from legitimate use; (ii) with fine-tuned base layers, the effectiveness of algebraic attacks decreases with the learning rate, showing that fine-tuning is not only beneficial for accuracy but also indispensable for model confidentiality.

Model extraction

MACE: A Flexible Framework for Membership Privacy Estimation in Generative Models

no code implementations11 Sep 2020 Yixi Xu, Sumit Mukherjee, Xiyang Liu, Shruti Tople, Rahul Dodhia, Juan Lavista Ferres

In this work, we propose the first formal framework for membership privacy estimation in generative models.

SOTERIA: In Search of Efficient Neural Networks for Private Inference

1 code implementation25 Jul 2020 Anshul Aggarwal, Trevor E. Carlson, Reza Shokri, Shruti Tople

In this setting, our objective is to protect the confidentiality of both the users' input queries as well as the model parameters at the server, with modest computation and communication overhead.

Neural Architecture Search

Domain Generalization using Causal Matching

1 code implementation arXiv 2020 Divyat Mahajan, Shruti Tople, Amit Sharma

In the domain generalization literature, a common objective is to learn representations independent of the domain after conditioning on the class label.

Data Augmentation Domain Generalization +1

Leakage of Dataset Properties in Multi-Party Machine Learning

1 code implementation12 Jun 2020 Wanrong Zhang, Shruti Tople, Olga Ohrimenko

Using multiple machine learning models, we show that leakage occurs even if the sensitive attribute is not included in the training data and has a low correlation with other attributes or the target variable.

Attribute BIG-bench Machine Learning

FALCON: Honest-Majority Maliciously Secure Framework for Private Deep Learning

1 code implementation5 Apr 2020 Sameer Wagh, Shruti Tople, Fabrice Benhamouda, Eyal Kushilevitz, Prateek Mittal, Tal Rabin

For private training, we are about 6x faster than SecureNN, 4. 4x faster than ABY3 and about 2-60x more communication efficient.

To Transfer or Not to Transfer: Misclassification Attacks Against Transfer Learned Text Classifiers

no code implementations8 Jan 2020 Bijeeta Pal, Shruti Tople

Thus, our results motivate the need for designing training techniques that are robust to unintended feature learning, specifically for transfer learned models.

Binary Classification Fake News Detection +2

Analyzing Information Leakage of Updates to Natural Language Models

no code implementations17 Dec 2019 Santiago Zanella-Béguelin, Lukas Wutschitz, Shruti Tople, Victor Rühle, Andrew Paverd, Olga Ohrimenko, Boris Köpf, Marc Brockschmidt

To continuously improve quality and reflect changes in data, machine learning applications have to regularly retrain and update their core models.

Language Modelling

An Empirical Study on the Intrinsic Privacy of SGD

1 code implementation5 Dec 2019 Stephanie L. Hyland, Shruti Tople

Introducing noise in the training of machine learning systems is a powerful way to protect individual privacy via differential privacy guarantees, but comes at a cost to utility.

Inference Attack Membership Inference Attack +1

Collaborative Machine Learning Markets with Data-Replication-Robust Payments

no code implementations8 Nov 2019 Olga Ohrimenko, Shruti Tople, Sebastian Tschiatschek

We study the problem of collaborative machine learning markets where multiple parties can achieve improved performance on their machine learning tasks by combining their training data.

BIG-bench Machine Learning

Alleviating Privacy Attacks via Causal Learning

1 code implementation ICML 2020 Shruti Tople, Amit Sharma, Aditya Nori

Such privacy risks are exacerbated when a model's predictions are used on an unseen data distribution.

Analyzing Privacy Loss in Updates of Natural Language Models

no code implementations25 Sep 2019 Shruti Tople, Marc Brockschmidt, Boris Köpf, Olga Ohrimenko, Santiago Zanella-Béguelin

To continuously improve quality and reflect changes in data, machine learning-based services have to regularly re-train and update their core models.

Privado: Practical and Secure DNN Inference with Enclaves

no code implementations1 Oct 2018 Karan Grover, Shruti Tople, Shweta Shinde, Ranjita Bhagwan, Ramachandran Ramjee

In this paper, we ask a timely question: "Can third-party cloud services use Intel SGX enclaves to provide practical, yet secure DNN Inference-as-a-service?"

Cannot find the paper you are looking for? You can Submit a new open access paper.