no code implementations • 14 Oct 2024 • Zhanpeng Zhou, Mingze Wang, Yuchen Mao, Bingrui Li, Junchi Yan
Specifically, we find that SAM efficiently selects flatter minima late in training.
no code implementations • 7 Oct 2024 • Bingrui Li, Wei Huang, Andi Han, Zhanpeng Zhou, Taiji Suzuki, Jun Zhu, Jianfei Chen
We also show that Adam behaves similarly to SignGD in terms of both optimization and generalization in this setting.
1 code implementation • 6 Feb 2024 • Zhanpeng Zhou, Zijun Chen, Yilan Chen, Bo Zhang, Junchi Yan
The pretraining-finetuning paradigm has become the prevailing trend in modern deep learning.
no code implementations • 10 Oct 2023 • Yiting Chen, Zhanpeng Zhou, Junchi Yan
In this paper, we expand the concept of equivalent feature and provide the definition of what we call functionally equivalent features.
no code implementations • 17 Oct 2022 • Ling Tang, Wen Shen, Zhanpeng Zhou, Yuefeng Chen, Quanshi Zhang
In this paper, we prove the representation defects of a cascaded convolutional decoder network, considering the capacity of representing different frequency components of an input sample.
no code implementations • 30 May 2022 • Zhanpeng Zhou, Wen Shen, Huixin Chen, Ling Tang, Quanshi Zhang
In this paper, we prove the effects of the BN operation on the back-propagation of the first and second derivatives of the loss.
1 code implementation • NeurIPS 2021 • Jie Ren, Die Zhang, Yisen Wang, Lu Chen, Zhanpeng Zhou, Yiting Chen, Xu Cheng, Xin Wang, Meng Zhou, Jie Shi, Quanshi Zhang
This paper provides a unified view to explain different adversarial attacks and defense methods, i. e. the view of multi-order interactions between input variables of DNNs.
1 code implementation • 5 Nov 2021 • Jie Ren, Die Zhang, Yisen Wang, Lu Chen, Zhanpeng Zhou, Yiting Chen, Xu Cheng, Xin Wang, Meng Zhou, Jie Shi, Quanshi Zhang
This paper provides a unified view to explain different adversarial attacks and defense methods, \emph{i. e.} the view of multi-order interactions between input variables of DNNs.
no code implementations • 29 Sep 2021 • Jie Ren, Zhanpeng Zhou, Qirui Chen, Quanshi Zhang
In the computation of Shapley values, people usually set an input variable to its baseline value to represent the absence of this variable.
1 code implementation • 22 May 2021 • Jie Ren, Zhanpeng Zhou, Qirui Chen, Quanshi Zhang
Masking some input variables of a deep neural network (DNN) and computing output changes on the masked input sample represent a typical way to compute attributions of input variables in the sample.
1 code implementation • 12 Mar 2021 • Jie Ren, Die Zhang, Yisen Wang, Lu Chen, Zhanpeng Zhou, Yiting Chen, Xu Cheng, Xin Wang, Meng Zhou, Jie Shi, Quanshi Zhang
This paper provides a unified view to explain different adversarial attacks and defense methods, i. e. the view of multi-order interactions between input variables of DNNs.