1 code implementation • 28 Feb 2025 • Jiaqi Bai, Hongcheng Guo, Zhongyuan Peng, Jian Yang, Zhoujun Li, Mohan Li, Zhihong Tian
Furthermore, we propose an entropy-based noise-controlling strategy to enable the injected noise to be adaptively constrained regarding the smoothness of the similarity distribution.
1 code implementation • 16 Jan 2025 • Yixiao Xu, Binxing Fang, Rui Wang, Yinghai Zhou, YuAn Liu, Mohan Li, Zhihong Tian
Guided by the model, we further introduce: (1) a similarity-based training-free watermarking method for plug-and-play and flexible watermarking, and (2) a distribution-based multi-step watermark information transmission strategy for robust watermarking.
no code implementations • CVPR 2025 • Keke Tang, Chao Hou, Weilong Peng, Xiang Fang, Zhize Wu, Yongwei Nie, Wenping Wang, Zhihong Tian
We attribute this issue to the inherent over-complexity of DNNs and investigate two key aspects: capacity and nonlinearity.
no code implementations • 26 Dec 2024 • Keke Tang, Weiyao Ke, Weilong Peng, Xiaofei Wang, Ziyong Du, Zhize Wu, Peican Zhu, Zhihong Tian
In this paper, we attribute the inadequate imperceptibility of adversarial attacks on point clouds to deviations from the underlying surface.
no code implementations • CVPR 2024 • Keke Tang, Chao Hou, Weilong Peng, Runnan Chen, Peican Zhu, Wenping Wang, Zhihong Tian
Deep neural networks (DNNs) often display overconfidence when encountering out-of-distribution (OOD) samples posing significant challenges in real-world applications.
no code implementations • 6 Oct 2021 • Lei Zhang, Shuaimin Jiang, Xiajiong Shen, Brij B. Gupta, Zhihong Tian
To address this imbalance, an intrusion detection system called pretraining Wasserstein generative adversarial network intrusion detection system (PWG-IDS) is proposed in this paper.
no code implementations • 29 Sep 2021 • Xin Zhang, Yanhua Li, Ziming Zhang, Christopher Brinton, Zhenming Liu, Zhi-Li Zhang, Hui Lu, Zhihong Tian
State-of-the-art imitation learning (IL) approaches, e. g, GAIL, apply adversarial training to minimize the discrepancy between expert and learner behaviors, which is prone to unstable training and mode collapse.
no code implementations • 13 Sep 2021 • Bin Zhu, Zhaoquan Gu, Le Wang, Zhihong Tian
Recent work shows that deep neural networks are vulnerable to adversarial examples.
no code implementations • 8 Sep 2021 • Xugong Qin, Yu Zhou, Youhui Guo, Dayan Wu, Zhihong Tian, Ning Jiang, Hongbin Wang, Weiping Wang
We propose to use an MLP decoder instead of the "deconv-conv" decoder in the mask head, which alleviates the issue and promotes robustness significantly.
no code implementations • ICCV 2021 • Keke Tang, Dingruibo Miao, Weilong Peng, Jianpeng Wu, Yawen Shi, Zhaoquan Gu, Zhihong Tian, Wenping Wang
Overconfident predictions on out-of-distribution (OOD) samples is a thorny issue for deep neural networks.
Generative Adversarial Network
Out of Distribution (OOD) Detection
1 code implementation • 2021 IEEE International Conference on Data Mining (ICDM) 2021 • Jing Wen, Bi-Yi Chen, Chang-Dong Wang, Zhihong Tian
However, recommender systems suffer from interaction data sparsity and data noise problems in reality.
no code implementations • 27 Nov 2019 • Keke Tang, Peng Song, Yuexin Ma, Zhaoquan Gu, Yu Su, Zhihong Tian, Wenping Wang
High-level (e. g., semantic) features encoded in the latter layers of convolutional neural networks are extensively exploited for image classification, leaving low-level (e. g., color) features in the early layers underexplored.