no code implementations • 6 Apr 2021 • Yiming Li, Tongqing Zhai, Yong Jiang, Zhifeng Li, Shu-Tao Xia
We demonstrate that this attack paradigm is vulnerable when the trigger in testing images is not consistent with the one used for training.
1 code implementation • 22 Oct 2020 • Tongqing Zhai, Yiming Li, Ziqi Zhang, Baoyuan Wu, Yong Jiang, Shu-Tao Xia
We also demonstrate that existing backdoor attacks cannot be directly adopted in attacking speaker verification.
no code implementations • 9 Apr 2020 • Yiming Li, Tongqing Zhai, Baoyuan Wu, Yong Jiang, Zhifeng Li, Shu-Tao Xia
Backdoor attack intends to inject hidden backdoor into the deep neural networks (DNNs), such that the prediction of the infected model will be maliciously changed if the hidden backdoor is activated by the attacker-defined trigger, while it performs well on benign samples.