Search Results for author: Liwei Song

Found 9 papers, 5 papers with code

Towards Personalized Evaluation of Large Language Models with An Anonymous Crowd-Sourcing Platform

no code implementations13 Mar 2024 Mingyue Cheng, Hao Zhang, Jiqian Yang, Qi Liu, Li Li, Xin Huang, Liwei Song, Zhi Li, Zhenya Huang, Enhong Chen

Through this gateway, users have the opportunity to submit their questions, testing the models on a personalized and potentially broader range of capabilities.

Language Modelling Large Language Model

Mitigating Membership Inference Attacks by Self-Distillation Through a Novel Ensemble Architecture

no code implementations15 Oct 2021 Xinyu Tang, Saeed Mahloujifar, Liwei Song, Virat Shejwalkar, Milad Nasr, Amir Houmansadr, Prateek Mittal

The goal of this work is to train ML models that have high membership privacy while largely preserving their utility; we therefore aim for an empirical membership privacy guarantee as opposed to the provable privacy guarantees provided by techniques like differential privacy, as such techniques are shown to deteriorate model utility.

Privacy Preserving

A Critical Evaluation of Open-World Machine Learning

no code implementations8 Jul 2020 Liwei Song, Vikash Sehwag, Arjun Nitin Bhagoji, Prateek Mittal

With our evaluation across 6 OOD detectors, we find that the choice of in-distribution data, model architecture and OOD data have a strong impact on OOD detection performance, inducing false positive rates in excess of $70\%$.

BIG-bench Machine Learning Out of Distribution (OOD) Detection

Universal Adversarial Attacks with Natural Triggers for Text Classification

1 code implementation NAACL 2021 Liwei Song, Xinwei Yu, Hsuan-Tung Peng, Karthik Narasimhan

Recent work has demonstrated the vulnerability of modern text classifiers to universal adversarial attacks, which are input-agnostic sequences of words added to text processed by classifiers.

General Classification text-classification +1

Systematic Evaluation of Privacy Risks of Machine Learning Models

1 code implementation24 Mar 2020 Liwei Song, Prateek Mittal

Machine learning models are prone to memorizing sensitive data, making them vulnerable to membership inference attacks in which an adversary aims to guess if an input sample was used to train the model.

BIG-bench Machine Learning Inference Attack

Towards Probabilistic Verification of Machine Unlearning

1 code implementation9 Mar 2020 David Marco Sommer, Liwei Song, Sameer Wagh, Prateek Mittal

In this work, we take the first step in proposing a formal framework to study the design of such verification mechanisms for data deletion requests -- also known as machine unlearning -- in the context of systems that provide machine learning as a service (MLaaS).

backdoor defense Machine Unlearning +1

Privacy Risks of Securing Machine Learning Models against Adversarial Examples

1 code implementation24 May 2019 Liwei Song, Reza Shokri, Prateek Mittal

To perform the membership inference attacks, we leverage the existing inference methods that exploit model predictions.

Adversarial Defense BIG-bench Machine Learning +1

Better the Devil you Know: An Analysis of Evasion Attacks using Out-of-Distribution Adversarial Examples

no code implementations5 May 2019 Vikash Sehwag, Arjun Nitin Bhagoji, Liwei Song, Chawin Sitawarin, Daniel Cullina, Mung Chiang, Prateek Mittal

A large body of recent work has investigated the phenomenon of evasion attacks using adversarial examples for deep learning systems, where the addition of norm-bounded perturbations to the test inputs leads to incorrect output classification.

Autonomous Driving General Classification

Inaudible Voice Commands

1 code implementation24 Aug 2017 Liwei Song, Prateek Mittal

Voice assistants like Siri enable us to control IoT devices conveniently with voice commands, however, they also provide new attack opportunities for adversaries.

Cryptography and Security

Cannot find the paper you are looking for? You can Submit a new open access paper.