1 code implementation • ICCV 2023 • Liang Zhang, Nathaniel Xu, Pengfei Yang, Gaojie Jin, Cheng-Chao Huang, Lijun Zhang
Firstly, the previous definitions of robustness in trajectory prediction are ambiguous.
no code implementations • 10 Feb 2023 • Pengfei Yang, Zhiming Chi, Zongxin Liu, Mengyu Zhao, Cheng-Chao Huang, Shaowei Cai, Lijun Zhang
Moreover, based on the framework, we propose the multi-objective DNN repair problem and give an algorithm based on our incremental SMT solving algorithm.
no code implementations • 23 Nov 2022 • Renjue Li, Tianhang Qin, Pengfei Yang, Cheng-Chao Huang, Youcheng Sun, Lijun Zhang
The safety properties proved in the resulting surrogate model apply to the original ADS with a probabilistic guarantee.
1 code implementation • IEEE Transactions on Parallel and Distributed Systems 2022 • Wenkai Lv, Quan Wang, Pengfei Yang
Then, we propose a multi-objective microservice deployment problem (MMDP) in edge computing.
no code implementations • 23 Jan 2022 • Gaojie Jin, Xinping Yi, Pengfei Yang, Lijun Zhang, Sven Schewe, Xiaowei Huang
While dropout is known to be a successful regularization technique, insights into the mechanisms that lead to this success are still lacking.
no code implementations • 5 Jun 2021 • Renjue Li, Hanwei Zhang, Pengfei Yang, Cheng-Chao Huang, Aimin Zhou, Bai Xue, Lijun Zhang
In this paper, we propose a framework of filter-based ensemble of deep neuralnetworks (DNNs) to defend against adversarial attacks.
1 code implementation • 25 Jan 2021 • Renjue Li, Pengfei Yang, Cheng-Chao Huang, Youcheng Sun, Bai Xue, Lijun Zhang
It is shown that DeepPAC outperforms the state-of-the-art statistical method PROVERO, and it achieves more practical robustness analysis than the formal verification tool ERAN.
1 code implementation • 15 Oct 2020 • Pengfei Yang, Renjue Li, Jianlin Li, Cheng-Chao Huang, Jingyi Wang, Jun Sun, Bai Xue, Lijun Zhang
The core idea is to make use of the obtained constraints of the abstraction to infer new bounds for the neurons.
no code implementations • 27 Aug 2019 • Shengyu Zhu, Biao Chen, Zhitang Chen, Pengfei Yang
With Sanov's theorem, we derive a sufficient condition for one-sample tests to achieve the optimal error exponent in the universal setting, i. e., for any distribution defining the alternative hypothesis.
no code implementations • 26 Feb 2019 • Jianlin Li, Pengfei Yang, Jiangchao Liu, Liqian Chen, Xiaowei Huang, Lijun Zhang
Several verification approaches have been developed to automatically prove or disprove safety properties of DNNs.
no code implementations • 21 Feb 2018 • Shengyu Zhu, Biao Chen, Pengfei Yang, Zhitang Chen
We show that two classes of Maximum Mean Discrepancy (MMD) based tests attain this optimality on $\mathbb R^d$, while the quadratic-time Kernel Stein Discrepancy (KSD) based tests achieve the maximum exponential decay rate under a relaxed level constraint.