Search Results for author: Ashish Hooda

Found 8 papers, 2 papers with code

PRP: Propagating Universal Perturbations to Attack Large Language Model Guard-Rails

no code implementations24 Feb 2024 Neal Mangaokar, Ashish Hooda, Jihye Choi, Shreyas Chandrashekaran, Kassem Fawaz, Somesh Jha, Atul Prakash

More recent LLMs often incorporate an additional layer of defense, a Guard Model, which is a second LLM that is designed to check and moderate the output response of the primary LLM.

Language Modelling Large Language Model

Theoretically Principled Trade-off for Stateful Defenses against Query-Based Black-Box Attacks

no code implementations30 Jul 2023 Ashish Hooda, Neal Mangaokar, Ryan Feng, Kassem Fawaz, Somesh Jha, Atul Prakash

This work aims to address this gap by offering a theoretical characterization of the trade-off between detection and false positive rates for stateful defenses.

Stateful Defenses for Machine Learning Models Are Not Yet Secure Against Black-box Attacks

1 code implementation11 Mar 2023 Ryan Feng, Ashish Hooda, Neal Mangaokar, Kassem Fawaz, Somesh Jha, Atul Prakash

Such stateful defenses aim to defend against black-box attacks by tracking the query history and detecting and rejecting queries that are "similar" and thus preventing black-box attacks from finding useful gradients and making progress towards finding adversarial attacks within a reasonable query budget.

SkillFence: A Systems Approach to Practically Mitigating Voice-Based Confusion Attacks

no code implementations16 Dec 2022 Ashish Hooda, Matthew Wallace, Kushal Jhunjhunwalla, Earlence Fernandes, Kassem Fawaz

Our key insight is that we can interpret a user's intentions by analyzing their activity on counterpart systems of the web and smartphones.

Re-purposing Perceptual Hashing based Client Side Scanning for Physical Surveillance

no code implementations8 Dec 2022 Ashish Hooda, Andrey Labunets, Tadayoshi Kohno, Earlence Fernandes

Content scanning systems employ perceptual hashing algorithms to scan user content for illegal material, such as child pornography or terrorist recruitment flyers.

D4: Detection of Adversarial Diffusion Deepfakes Using Disjoint Ensembles

no code implementations11 Feb 2022 Ashish Hooda, Neal Mangaokar, Ryan Feng, Kassem Fawaz, Somesh Jha, Atul Prakash

D4 uses an ensemble of models over disjoint subsets of the frequency spectrum to significantly improve adversarial robustness.

Adversarial Robustness DeepFake Detection +1

Invisible Perturbations: Physical Adversarial Examples Exploiting the Rolling Shutter Effect

2 code implementations CVPR 2021 Athena Sayles, Ashish Hooda, Mohit Gupta, Rahul Chatterjee, Earlence Fernandes

By contrast, we contribute a procedure to generate, for the first time, physical adversarial examples that are invisible to human eyes.

Object

Cannot find the paper you are looking for? You can Submit a new open access paper.