no code implementations • 2 Oct 2019 • Lakshya Jain, Wilson Wu, Steven Chen, Uyeong Jang, Varun Chandrasekaran, Sanjit Seshia, Somesh Jha
In this paper we explore semantic adversarial examples (SAEs) where an attacker creates perturbations in the semantic space representing the environment that produces input for the ML model.
no code implementations • ICLR 2020 • Uyeong Jang, Susmit Jha, Somesh Jha
These defenses rely on the assumption that data lie in a manifold of a lower dimension than the input space.
no code implementations • ICLR 2018 • Xi Wu, Uyeong Jang, Lingjiao Chen, Somesh Jha
Interestingly, we find that a recent objective by Madry et al. encourages training a model that satisfies well our formal version of the goodness property, but has a weak control of points that are wrong but with low confidence.
no code implementations • ICML 2018 • Xi Wu, Uyeong Jang, Jiefeng Chen, Lingjiao Chen, Somesh Jha
In this paper we study leveraging confidence information induced by adversarial training to reinforce adversarial robustness of a given adversarially trained model.