no code implementations • 28 May 2019 • Pu Zhao, Siyue Wang, Cheng Gongye, Yanzhi Wang, Yunsi Fei, Xue Lin
Despite the great achievements of deep neural networks (DNNs), the vulnerability of state-of-the-art DNNs raises security concerns of DNNs in many application domains requiring high reliability. We propose the fault sneaking attack on DNNs, where the adversary aims to misclassify certain input images into any target labels by modifying the DNN parameters.
no code implementations • 7 Sep 2021 • Amel Nestor Docena, Thomas Wahl, Trevor Pearce, Yunsi Fei
We demonstrate the impact of the use of solvers in terms of functionality and search efficiency, using a case study for the detection of Trojan attacks on Neural Networks.
no code implementations • 27 Mar 2023 • Ruyi Ding, Cheng Gongye, Siyue Wang, Aidong Ding, Yunsi Fei
Inspired by the fact that electromagnetic (EM) emanations of a model inference are dependent on both operations and data and may contain footprints of different input classes, we propose a framework, EMShepherd, to capture EM traces of model execution, perform processing on traces and exploit them for adversarial detection.
no code implementations • ICCV 2023 • Ruyi Ding, Shijin Duan, Xiaolin Xu, Yunsi Fei
Graph neural networks (GNNs) have brought superb performance to various applications utilizing graph structural data, such as social analysis and fraud detection.