1 code implementation • 25 Jul 2024 • Zihui Wu, Haichang Gao, Jianping He, Ping Wang
Large language models (LLMs) have demonstrated remarkable capabilities, but their power comes with significant security considerations.
no code implementations • 29 May 2024 • Zihui Wu, Yu Sun, Yifan Chen, Bingliang Zhang, Yisong Yue, Katherine L. Bouman
Diffusion models (DMs) have recently shown outstanding capability in modeling complex image distributions, making them expressive image priors for solving Bayesian inverse problems.
1 code implementation • 16 Oct 2023 • Yu Sun, Zihui Wu, Yifan Chen, Berthy T. Feng, Katherine L. Bouman
PMC is able to incorporate expressive score-based generative priors for high-quality image reconstruction while also performing uncertainty quantification via posterior sampling.
no code implementations • NeurIPS 2023 • Xinyi Wu, Amir Ajorlou, Zihui Wu, Ali Jadbabaie
Oversmoothing in Graph Neural Networks (GNNs) refers to the phenomenon where increasing network depth leads to homogeneous node representations.
no code implementations • 24 May 2023 • Zihui Wu, Haichang Gao, Bingqian Zhou, Ping Wang
To tackle this problem, we propose a simple but effective strategy called Adversarial Function Matching (AdvFunMatch), which aims to match distributions for all data points within the $\ell_p$-norm ball of the training data, in accordance with consistent teaching.
no code implementations • 25 Apr 2023 • Zihui Wu, Tianwei Yin, Yu Sun, Robert Frost, Andre van der Kouwe, Adrian V. Dalca, Katherine L. Bouman
Traditional CS-MRI methods often separately address measurement subsampling, image reconstruction, and task prediction, resulting in a suboptimal end-to-end performance.
no code implementations • 26 Aug 2022 • Zihui Wu, Haichang Gao, Bingqian Zhou, Xiaoyan Guo, Shudong Zhang
In addition, we discuss the function of entropy in TRADES, and we find that models with high entropy can be better robustness learners.
no code implementations • 24 May 2022 • Shudong Zhang, Haichang Gao, Tianwei Zhang, Yunyi Zhou, Zihui Wu
Adversarial training (AT) has proven to be one of the most effective ways to defend Deep Neural Networks (DNNs) against adversarial attacks.
no code implementations • 29 Sep 2021 • Zihui Wu, Haichang Gao, Shudong Zhang, Yipeng Gao
Then, we explored the effect of another classic smoothing regularizer, namely, the maximum entropy (ME), and we have found ME can also help reduce both inter-class similarity and intra-class variance.
1 code implementation • 13 May 2021 • Tianwei Yin, Zihui Wu, He Sun, Adrian V. Dalca, Yisong Yue, Katherine L. Bouman
In this paper, we leverage the sequential nature of MRI measurements, and propose a fully differentiable framework that jointly learns a sequential sampling policy simultaneously with a reconstruction strategy.
no code implementations • 5 Jun 2020 • Yu Sun, Zihui Wu, Xiaojian Xu, Brendt Wohlberg, Ulugbek S. Kamilov
Plug-and-play priors (PnP) is a broadly applicable methodology for solving inverse problems by exploiting statistical priors specified as denoisers.
no code implementations • 4 Sep 2019 • Zihui Wu, Yu Sun, Jiaming Liu, Ulugbek S. Kamilov
Regularization by denoising (RED) is a powerful framework for solving imaging inverse problems.