no code implementations • 3 Mar 2025 • Kyle Domico, Jean-Charles Noirot Ferrand, Ryan Sheatsley, Eric Pauley, Josiah Hanna, Patrick McDaniel
Reinforcement learning (RL) offers powerful techniques for solving complex sequential decision-making tasks from experience.
no code implementations • 27 Jan 2025 • Jean-Charles Noirot Ferrand, Yohan Beugin, Eric Pauley, Ryan Sheatsley, Patrick McDaniel
We observe that alignment embeds a safety classifier in the target model that is responsible for deciding between refusal and compliance.
1 code implementation • 13 Dec 2024 • Blaine Hoak, Ryan Sheatsley, Patrick McDaniel
Bias significantly undermines both the accuracy and trustworthiness of machine learning models.
2 code implementations • 3 Oct 2024 • Xiaogeng Liu, Peiran Li, Edward Suh, Yevgeniy Vorobeychik, Zhuoqing Mao, Somesh Jha, Patrick McDaniel, Huan Sun, Bo Li, Chaowei Xiao
In this paper, we propose AutoDAN-Turbo, a black-box jailbreak method that can automatically discover as many jailbreak strategies as possible from scratch, without any human intervention or predefined scopes (e. g., specified candidate strategies), and use them for red-teaming.
no code implementations • 16 Sep 2024 • Blaine Hoak, Patrick McDaniel
Through this, we create the Prompted Textures Dataset (PTD), a dataset of 362, 880 texture images that span 56 textures.
1 code implementation • 14 Mar 2024 • Blaine Hoak, Patrick McDaniel
In this work, we investigate \textit{texture learning}: the identification of textures learned by object classification models, and the extent to which they rely on these textures.
no code implementations • 28 Feb 2024 • Fangzhou Wu, Ning Zhang, Somesh Jha, Patrick McDaniel, Chaowei Xiao
Large Language Model (LLM) systems are inherently compositional, with individual LLM serving as the core foundation with additional layers of objects such as plugins, sandbox, and so on.
1 code implementation • 22 Feb 2024 • Jiongxiao Wang, Jiazhao Li, Yiquan Li, Xiangyu Qi, Junjie Hu, Yixuan Li, Patrick McDaniel, Muhao Chen, Bo Li, Chaowei Xiao
In particular, service providers will construct prefixed safety examples with a secret prompt, acting as a "backdoor trigger".
no code implementations • 17 Oct 2023 • Kunyang Li, Kyle Domico, Jean-Charles Noirot Ferrand, Patrick McDaniel
The transferability of these adversarial examples is measured by evaluating each set on other models to determine which models offer more adversarial strength, and consequently, more robustness against these attacks.
no code implementations • 9 Sep 2022 • Ryan Sheatsley, Blaine Hoak, Eric Pauley, Patrick McDaniel
From our evaluation we find that attack performance to be highly contextual: the domain, model robustness, and threat model can have a profound influence on attack efficacy.
no code implementations • 1 May 2022 • Valentin Vie, Ryan Sheatsley, Sophia Beyda, Sushrut Shringarputale, Kevin Chan, Trent Jaeger, Patrick McDaniel
We evaluate the performance of the algorithms against two dominant planning algorithms used in commercial applications (D* Lite and Fast Downward) and show both are vulnerable to extremely limited adversarial action.
no code implementations • 4 Apr 2022 • Kyle Domico, Ryan Sheatsley, Yohan Beugin, Quinn Burke, Patrick McDaniel
They result from high sun activity, which are induced from cool areas on the Sun known as sunspots.
no code implementations • 21 Feb 2022 • Ahmed Abdou, Ryan Sheatsley, Yohan Beugin, Tyler Shipp, Patrick McDaniel
To harden these systems the ever-growing field of Adversarial Machine Learning has proposed new attack and defense mechanisms.
no code implementations • 21 Feb 2022 • Ryan Sheatsley, Matthew Durbin, Azaree Lintereur, Patrick McDaniel
With four and eight detector arrays, we collect counts of gamma-rays as features for a suite of machine learning models to localize radioactive material.
no code implementations • 18 May 2021 • Ryan Sheatsley, Blaine Hoak, Eric Pauley, Yohan Beugin, Michael J. Weisman, Patrick McDaniel
Machine learning is vulnerable to adversarial examples-inputs designed to cause models to perform poorly.
no code implementations • 2 Nov 2020 • Ryan Sheatsley, Nicolas Papernot, Michael Weisman, Gunjan Verma, Patrick McDaniel
To assess how these algorithms perform, we evaluate them in constrained (e. g., network intrusion detection) and unconstrained (e. g., image recognition) domains.
no code implementations • 17 Feb 2020 • Michael Norris, Berkay Celik, Patrick McDaniel, Gang Tan, Prasanna Venkatesh, Shulin Zhao, Anand Sivasubramaniam
IoT devices are decentralized and deployed in un-stable environments, which causes them to be prone to various kinds of faults, such as device failure and network disruption.
Software Engineering Performance
no code implementations • 24 Nov 2019 • Leonardo Babun, Z. Berkay Celik, Patrick McDaniel, A. Selcuk Uluagac
We designed and built IoTWatcH based on an IoT privacy survey that considers the privacy needs of IoT users.
no code implementations • 22 Nov 2019 • Amit Kumar Sikder, Leonardo Babun, Z. Berkay Celik, Abbas Acar, Hidayet Aksu, Patrick McDaniel, Engin Kirda, A. Selcuk Uluagac
Users can specify their desired access control settings using the interaction module which are translated into access control policies in the backend server.
Cryptography and Security
no code implementations • 30 Aug 2019 • Dan Boneh, Andrew J. Grotto, Patrick McDaniel, Nicolas Papernot
Popular culture has contemplated societies of thinking machines for generations, envisioning futures from utopian to dystopian.
1 code implementation • 22 Oct 2018 • Dang Tu Nguyen, Chengyu Song, Zhiyun Qian, Srikanth V. Krishnamurthy, Edward J. M. Colbert, Patrick McDaniel
In this paper, we design IoTSan, a novel practical system that uses model checking as a building block to reveal "interaction-level" flaws by identifying events that can lead the system to unsafe states.
Cryptography and Security
1 code implementation • 18 Sep 2018 • Z. Berkay Celik, Earlence Fernandes, Eric Pauley, Gang Tan, Patrick McDaniel
Based on a study of five IoT programming platforms, we identify the key insights resulting from works in both the program analysis and security communities and relate the efficacy of program-analysis techniques to security and privacy issues.
Cryptography and Security Programming Languages
4 code implementations • 13 Mar 2018 • Nicolas Papernot, Patrick McDaniel
However, deep learning is often criticized for its lack of robustness in adversarial settings (e. g., vulnerability to adversarial inputs) and general inability to rationalize its predictions.
1 code implementation • 22 Feb 2018 • Z. Berkay Celik, Leonardo Babun, Amit K. Sikder, Hidayet Aksu, Gang Tan, Patrick McDaniel, A. Selcuk Uluagac
Through this effort, we introduce a rigorously grounded framework for evaluating the use of sensitive information in IoT apps---and therein provide developers, markets, and consumers a means of identifying potential threats to security and privacy.
Cryptography and Security Programming Languages
13 code implementations • ICLR 2018 • Florian Tramèr, Alexey Kurakin, Nicolas Papernot, Ian Goodfellow, Dan Boneh, Patrick McDaniel
We show that this form of adversarial training converges to a degenerate global minimum, wherein small curvature artifacts near the data points obfuscate a linear approximation of the loss.
1 code implementation • 15 May 2017 • Nicolas Papernot, Patrick McDaniel
Machine learning is vulnerable to adversarial examples: inputs carefully modified to force misclassification.
2 code implementations • 11 Apr 2017 • Florian Tramèr, Nicolas Papernot, Ian Goodfellow, Dan Boneh, Patrick McDaniel
Adversarial examples are maliciously perturbed inputs designed to mislead machine learning (ML) models at test-time.
no code implementations • 21 Feb 2017 • Kathrin Grosse, Praveen Manoharan, Nicolas Papernot, Michael Backes, Patrick McDaniel
Specifically, we augment our ML model with an additional output, in which the model is trained to classify all adversarial inputs.
no code implementations • 26 Nov 2016 • Z. Berkay Celik, David Lopez-Paz, Patrick McDaniel
In this paper, we present privacy distillation, a mechanism which allows patients to control the type and amount of information they wish to disclose to the healthcare providers for use in statistical models.
no code implementations • 11 Nov 2016 • Nicolas Papernot, Patrick McDaniel, Arunesh Sinha, Michael Wellman
Advances in machine learning (ML) in recent years have enabled a dizzying array of applications such as data analytics, autonomous systems, and security diagnostics.
13 code implementations • 3 Oct 2016 • Nicolas Papernot, Fartash Faghri, Nicholas Carlini, Ian Goodfellow, Reuben Feinman, Alexey Kurakin, Cihang Xie, Yash Sharma, Tom Brown, Aurko Roy, Alexander Matyasko, Vahid Behzadan, Karen Hambardzumyan, Zhishuai Zhang, Yi-Lin Juang, Zhi Li, Ryan Sheatsley, Abhibhav Garg, Jonathan Uesato, Willi Gierke, Yinpeng Dong, David Berthelot, Paul Hendricks, Jonas Rauber, Rujun Long, Patrick McDaniel
An adversarial example library for constructing attacks, building defenses, and benchmarking both
no code implementations • 14 Jun 2016 • Kathrin Grosse, Nicolas Papernot, Praveen Manoharan, Michael Backes, Patrick McDaniel
Deep neural networks, like many other machine learning models, have recently been shown to lack robustness against adversarially crafted inputs.
no code implementations • 24 May 2016 • Nicolas Papernot, Patrick McDaniel, Ian Goodfellow
We demonstrate our attacks on two commercial machine learning classification systems from Amazon (96. 19% misclassification rate) and Google (88. 94%) using only 800 queries of the victim model, thereby showing that existing machine learning approaches are in general vulnerable to systematic black-box attacks regardless of their structure.
1 code implementation • 28 Apr 2016 • Nicolas Papernot, Patrick McDaniel, Ananthram Swami, Richard Harang
Machine learning models are frequently used to solve complex security problems, as well as to make decisions in sensitive situations like guiding autonomous vehicles or predicting financial market behaviors.
no code implementations • 31 Mar 2016 • Z. Berkay Celik, Patrick McDaniel, Rauf Izmailov, Nicolas Papernot, Ryan Sheatsley, Raquel Alvarez, Ananthram Swami
In this paper, we consider an alternate learning approach that trains models using "privileged" information--features available at training time but not at runtime--to improve the accuracy and resilience of detection systems.
18 code implementations • 8 Feb 2016 • Nicolas Papernot, Patrick McDaniel, Ian Goodfellow, Somesh Jha, Z. Berkay Celik, Ananthram Swami
Our attack strategy consists in training a local model to substitute for the target DNN, using inputs synthetically generated by an adversary and labeled by the target DNN.
11 code implementations • 24 Nov 2015 • Nicolas Papernot, Patrick McDaniel, Somesh Jha, Matt Fredrikson, Z. Berkay Celik, Ananthram Swami
In this work, we formalize the space of adversaries against deep neural networks (DNNs) and introduce a novel class of algorithms to craft adversarial samples based on a precise understanding of the mapping between inputs and outputs of DNNs.
2 code implementations • 14 Nov 2015 • Nicolas Papernot, Patrick McDaniel, Xi Wu, Somesh Jha, Ananthram Swami
In this work, we introduce a defensive mechanism called defensive distillation to reduce the effectiveness of adversarial samples on DNNs.