Provable Defenses against Spatially Transformed Adversarial Inputs: Impossibility and Possibility Results

ICLR 2019 Xinyang ZhangYifan HuangChanh NguyenShouling JiTing Wang

One intriguing property of neural networks is their inherent vulnerability to adversarial inputs, which are maliciously crafted samples to trigger target networks to misbehave. The state-of-the-art attacks generate adversarial inputs using either pixel perturbation or spatial transformation... (read more)

PDF Abstract

Code


No code implementations yet. Submit your code now

Tasks


Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.