Search Results for author: Zhiqi Bu

Found 29 papers, 13 papers with code

Pre-training Differentially Private Models with Limited Public Data

no code implementations28 Feb 2024 Zhiqi Bu, Xinwei Zhang, Mingyi Hong, Sheng Zha, George Karypis

The superior performance of large foundation models relies on the use of massive amounts of high-quality data, which often contain sensitive, private and copyrighted material that requires formal protection.

Differentially Private SGD Without Clipping Bias: An Error-Feedback Approach

no code implementations24 Nov 2023 Xinwei Zhang, Zhiqi Bu, Zhiwei Steven Wu, Mingyi Hong

In our work, we propose a new error-feedback (EF) DP algorithm as an alternative to DPSGD-GC, which not only offers a diminishing utility bound without inducing a constant clipping bias, but more importantly, it allows for an arbitrary choice of clipping threshold that is independent of the problem.

Zero redundancy distributed learning with differential privacy

no code implementations20 Nov 2023 Zhiqi Bu, Justin Chiu, Ruixuan Liu, Sheng Zha, George Karypis

Deep learning using large models have achieved great success in a wide range of domains.

Privacy Preserving

On the accuracy and efficiency of group-wise clipping in differentially private optimization

no code implementations30 Oct 2023 Zhiqi Bu, Ruixuan Liu, Yu-Xiang Wang, Sheng Zha, George Karypis

Recent advances have substantially improved the accuracy, memory cost, and training speed of differentially private (DP) deep learning, especially on large vision and language models with millions to billions of parameters.

Coupling public and private gradient provably helps optimization

no code implementations2 Oct 2023 Ruixuan Liu, Zhiqi Bu, Yu-Xiang Wang, Sheng Zha, George Karypis

The success of large neural networks is crucially determined by the availability of data.

MISNN: Multiple Imputation via Semi-parametric Neural Networks

no code implementations2 May 2023 Zhiqi Bu, Zongyu Dai, Yiliang Zhang, Qi Long

Multiple imputation (MI) has been widely applied to missing value problems in biomedical, social and econometric research, in order to avoid improper inference in the downstream data analysis.

feature selection Imputation +1

Multiple Imputation with Neural Network Gaussian Process for High-dimensional Incomplete Data

1 code implementation23 Nov 2022 Zongyu Dai, Zhiqi Bu, Qi Long

Single imputation methods such as matrix completion methods do not adequately account for imputation uncertainty and hence would yield improper statistical inference.

Imputation Matrix Completion

Differentially Private Optimizers Can Learn Adversarially Robust Models

no code implementations16 Nov 2022 Yuan Zhang, Zhiqi Bu

Machine learning models have shone in a variety of domains and attracted increasing attention from both the security and the privacy communities.

Adversarial Robustness

Accelerating Adversarial Perturbation by 50% with Semi-backward Propagation

no code implementations9 Nov 2022 Zhiqi Bu

Adversarial perturbation plays a significant role in the field of adversarial robustness, which solves a maximization problem over the input data.

Adversarial Robustness

Differentially Private Bias-Term only Fine-tuning of Foundation Models

1 code implementation30 Sep 2022 Zhiqi Bu, Yu-Xiang Wang, Sheng Zha, George Karypis

We study the problem of differentially private (DP) fine-tuning of large pre-trained models -- a recent privacy-preserving approach suitable for solving downstream tasks with sensitive data.

Privacy Preserving

Differentially Private Optimization on Large Model at Small Cost

2 code implementations30 Sep 2022 Zhiqi Bu, Yu-Xiang Wang, Sheng Zha, George Karypis

Our implementation achieves state-of-the-art (SOTA) accuracy with very small extra cost: on GPT2 and at almost the same memory cost (<1% overhead), BK has 1. 03X the time complexity of the standard training (0. 83X training speed in practice), and 0. 61X the time complexity of the most efficient DP implementation (1. 36X training speed in practice).

Privacy Preserving

CEDAR: Communication Efficient Distributed Analysis for Regressions

no code implementations1 Jul 2022 Changgee Chang, Zhiqi Bu, Qi Long

We provide theoretical investigation for the asymptotic properties of the proposed method for statistical inference as well as differential privacy, and evaluate its performance in simulations and real data analyses in comparison with several recently developed methods.

Scalable and Efficient Training of Large Convolutional Neural Networks with Differential Privacy

1 code implementation21 May 2022 Zhiqi Bu, Jialin Mao, Shiyun Xu

Large convolutional neural networks (CNN) can be difficult to train in the differentially private (DP) regime, since the optimization algorithms require a computationally expensive operation, known as the per-sample gradient clipping.

Sparse Neural Additive Model: Interpretable Deep Learning with Feature Selection via Group Sparsity

no code implementations25 Feb 2022 Shiyun Xu, Zhiqi Bu, Pratik Chaudhari, Ian J. Barnett

In order to empower NAM with feature selection and improve the generalization, we propose the sparse neural additive models (SNAM) that employ the group sparsity regularization (e. g. Group LASSO), where each feature is learned by a sub-network whose trainable parameters are clustered as a group.

Additive models feature selection +1

Multiple Imputation via Generative Adversarial Network for High-dimensional Blockwise Missing Value Problems

no code implementations21 Dec 2021 Zongyu Dai, Zhiqi Bu, Qi Long

Missing data are present in most real world problems and need careful handling to preserve the prediction accuracy and statistical consistency in the downstream analysis.

Generative Adversarial Network Imputation

Practical Adversarial Training with Differential Privacy for Deep Learning

no code implementations29 Sep 2021 Zhiqi Bu, Ping Li, Weijie Zhao

In this work, we propose the practical adversarial training with differential privacy (DP-Adv), to combine the backbones from both communities and deliver robust and private models with high accuracy.

Differentially Private Bayesian Neural Networks on Accuracy, Privacy and Reliability

no code implementations18 Jul 2021 Qiyiwen Zhang, Zhiqi Bu, Kan Chen, Qi Long

Interestingly, we show a new equivalence between DP-SGD and DP-SGLD, implying that some non-Bayesian DP training naturally allows for uncertainty quantification.

Uncertainty Quantification

Privacy Amplification via Iteration for Shuffled and Online PNSGD

no code implementations20 Jun 2021 Matteo Sordello, Zhiqi Bu, Jinshuo Dong

We then analyze the online setting and provide a faster decaying scheme for the magnitude of the injected noise that also guarantees the convergence of privacy loss.

Accuracy, Interpretability, and Differential Privacy via Explainable Boosting

1 code implementation17 Jun 2021 Harsha Nori, Rich Caruana, Zhiqi Bu, Judy Hanwen Shen, Janardhan Kulkarni

We show that adding differential privacy to Explainable Boosting Machines (EBMs), a recent method for training interpretable ML models, yields state-of-the-art accuracy while protecting privacy.

regression

On the Convergence and Calibration of Deep Learning with Differential Privacy

1 code implementation15 Jun 2021 Zhiqi Bu, Hua Wang, Zongyu Dai, Qi Long

Differentially private (DP) training preserves the data privacy usually at the cost of slower convergence (and thus lower accuracy), as well as more severe mis-calibration than its non-private counterpart.

Characterizing the SLOPE Trade-off: A Variational Perspective and the Donoho-Tanner Limit

1 code implementation27 May 2021 Zhiqi Bu, Jason Klusowski, Cynthia Rush, Weijie J. Su

Sorted l1 regularization has been incorporated into many methods for solving high-dimensional statistical estimation problems, including the SLOPE estimator in linear regression.

Variable Selection

Efficient Designs of SLOPE Penalty Sequences in Finite Dimension

1 code implementation14 Feb 2021 Yiliang Zhang, Zhiqi Bu

In this paper, we propose two efficient algorithms to design the possibly high-dimensional SLOPE penalty, in order to minimize the mean squared error.

Fast and Memory Efficient Differentially Private-SGD via JL Projections

no code implementations NeurIPS 2021 Zhiqi Bu, Sivakanth Gopi, Janardhan Kulkarni, Yin Tat Lee, Judy Hanwen Shen, Uthaipon Tantipongpipat

Unlike previous attempts to make DP-SGD faster which work only on a subset of network architectures or use compiler techniques, we propose an algorithmic solution which works for any network in a black-box manner which is the main contribution of this paper.

FAST DIFFERENTIALLY PRIVATE-SGD VIA JL PROJECTIONS

no code implementations1 Jan 2021 Zhiqi Bu, Sivakanth Gopi, Janardhan Kulkarni, Yin Tat Lee, Uthaipon Tantipongpipat

Differentially Private-SGD (DP-SGD) of Abadi et al. (2016) and its variations are the only known algorithms for private training of large scale neural networks.

DebiNet: Debiasing Linear Models with Nonlinear Overparameterized Neural Networks

1 code implementation1 Nov 2020 Shiyun Xu, Zhiqi Bu

Recent years have witnessed strong empirical performance of over-parameterized neural networks on various tasks and many advances in the theory, e. g. the universal approximation and provable convergence to global minimum.

feature selection valid

A Dynamical View on Optimization Algorithms of Overparameterized Neural Networks

1 code implementation25 Oct 2020 Zhiqi Bu, Shiyun Xu, Kan Chen

When equipped with efficient optimization algorithms, the over-parameterized neural networks have demonstrated high level of performance even though the loss function is non-convex and non-smooth.

The Complete Lasso Tradeoff Diagram

2 code implementations NeurIPS 2020 Hua Wang, Yachong Yang, Zhiqi Bu, Weijie J. Su

A fundamental problem in the high-dimensional regression is to understand the tradeoff between type I and type II errors or, equivalently, false discovery rate (FDR) and power in variable selection.

Statistics Theory Information Theory Information Theory Statistics Theory

Deep Learning with Gaussian Differential Privacy

3 code implementations26 Nov 2019 Zhiqi Bu, Jinshuo Dong, Qi Long, Weijie J. Su

Leveraging the appealing properties of $f$-differential privacy in handling composition and subsampling, this paper derives analytically tractable expressions for the privacy guarantees of both stochastic gradient descent and Adam used in training deep neural networks, without the need of developing sophisticated techniques as [3] did.

General Classification Image Classification +2

Algorithmic Analysis and Statistical Estimation of SLOPE via Approximate Message Passing

1 code implementation NeurIPS 2019 Zhiqi Bu, Jason Klusowski, Cynthia Rush, Weijie Su

SLOPE is a relatively new convex optimization procedure for high-dimensional linear regression via the sorted l1 penalty: the larger the rank of the fitted coefficient, the larger the penalty.

Cannot find the paper you are looking for? You can Submit a new open access paper.