no code implementations • 1 Jun 2025 • Xiang Li, Jiayi Xin, Qi Long, Weijie J. Su
Accurate evaluation of large language models (LLMs) is crucial for understanding their capabilities and guiding their development.
no code implementations • 27 May 2025 • Zhekun Shi, Kaizhao Liu, Qi Long, Weijie J. Su, Jiancong Xiao
However, using raw preference as the payoff in the game highly limits the potential of the game-theoretic LLM alignment framework.
no code implementations • 4 May 2025 • Jiancong Xiao, BoJian Hou, Zhanliang Wang, Ruochen Jin, Qi Long, Weijie J. Su, Li Shen
However, as models are further fine-tuned for better performance, they enter the non-calibratable regime.
1 code implementation • 14 Mar 2025 • Kaizhao Liu, Qi Long, Zhekun Shi, Weijie J. Su, Jiancong Xiao
As a blessing, we prove that this condition holds with high probability under the probabilistic preference model, thereby highlighting the statistical possibility of preserving minority preferences without explicit regularization in aligning LLMs.
no code implementations • 25 Feb 2025 • Wenlong Ji, Weizhe Yuan, Emily Getzen, Kyunghyun Cho, Michael I. Jordan, Song Mei, Jason E Weston, Weijie J. Su, Jing Xu, Linjun Zhang
Large Language Models (LLMs) have emerged as transformative tools in artificial intelligence (AI), exhibiting remarkable capabilities across diverse tasks such as text generation, reasoning, and decision-making.
1 code implementation • 21 Nov 2024 • Xiang Li, Feng Ruan, Huiyuan Wang, Qi Long, Weijie J. Su
We prove that the Tr-GoF test achieves optimality in robust detection of the Gumbel-max watermark in a certain asymptotic regime of substantial text modifications and vanishing watermark signals.
1 code implementation • 17 Nov 2024 • Yangxinyu Xie, Xiang Li, Tanwi Mallick, Weijie J. Su, Ruixun Zhang
Watermarking language models is essential for distinguishing between human and machine-generated text and thus maintaining the integrity and trustworthiness of digital communication.
no code implementations • 22 Oct 2024 • Mingzhi Wang, Chengdong Ma, Qizhi Chen, Linjian Meng, Yang Han, Jiancong Xiao, Zhaowei Zhang, Jing Huo, Weijie J. Su, Yaodong Yang
Self-play methods have demonstrated remarkable success in enhancing model capabilities across various domains.
2 code implementations • 11 Oct 2024 • Buxin Su, Weijie J. Su, Chendi Wang
The U. S. Decennial Census serves as the foundation for many high-profile policy decision-making processes, including federal funding allocation and redistricting.
no code implementations • 14 Sep 2024 • Weijie J. Su
We review techniques that render $f$-differential privacy a unified framework for analyzing privacy bounds in data analysis and machine learning.
1 code implementation • 24 Aug 2024 • Hangfeng He, Weijie J. Su
Large language models (LLMs) have been widely employed across various application domains, yet their black-box nature poses significant challenges to understanding how these models process input data internally to make predictions.
1 code implementation • 16 Jun 2024 • Bowen Jiang, Yangxinyu Xie, Zhuoqun Hao, Xiaomeng Wang, Tanwi Mallick, Weijie J. Su, Camillo J. Taylor, Dan Roth
This study introduces a hypothesis-testing framework to assess whether large language models (LLMs) possess genuine reasoning abilities or primarily depend on token bias.
no code implementations • 8 Jun 2024 • Jiancong Xiao, Ruoyu Sun, Qi Long, Weijie J. Su
We aim to construct a new cover that possesses two properties: 1) compatibility with adversarial examples, and 2) precision comparable to covers used in standard settings.
1 code implementation • 5 Jun 2024 • Hiroaki Chiba-Okabe, Weijie J. Su
As a practical implementation, we introduce PREGen (Prompt Rewriting-Enhanced Genericization), which combines our genericization method with an existing mitigation technique.
1 code implementation • 1 Jun 2024 • Bowen Jiang, Yangxinyu Xie, Xiaomeng Wang, Yuan Yuan, Zhuoqun Hao, Xinyi Bai, Weijie J. Su, Camillo J. Taylor, Tanwi Mallick
This work discusses how to build more rational language and multimodal agents and what criteria define rationality in intelligent systems.
1 code implementation • 26 May 2024 • Jiancong Xiao, Ziniu Li, Xingyu Xie, Emily Getzen, Cong Fang, Qi Long, Weijie J. Su
To mitigate this algorithmic bias, we introduce preference matching (PM) RLHF, a novel approach that provably aligns LLMs with the preference distribution of the reward model under the Bradley--Terry--Luce/Plackett--Luce model.
no code implementations • 14 May 2024 • Chendi Wang, Yuqing Zhu, Weijie J. Su, Yu-Xiang Wang
A recent study by De et al. (2022) has reported that large-scale representation learning through pre-training on a public dataset significantly enhances differentially private (DP) learning in downstream tasks, despite the high dimensionality of the feature space.
no code implementations • 22 Apr 2024 • Jiachen T. Wang, Zhun Deng, Hiroaki Chiba-Okabe, Boaz Barak, Weijie J. Su
Generative artificial intelligence (AI) systems are trained on large data corpora to generate new pieces of text, images, videos, and other media.
1 code implementation • 1 Apr 2024 • Xiang Li, Feng Ruan, Huiyuan Wang, Qi Long, Weijie J. Su
In particular, we derive optimal detection rules for these watermarks under our framework.
no code implementations • 8 Mar 2024 • Huiying Zhong, Zhun Deng, Weijie J. Su, Zhiwei Steven Wu, Linjun Zhang
Our work \textit{initiates} the theoretical study of multi-party RLHF that explicitly models the diverse preferences of multiple individuals.
no code implementations • 6 Jul 2023 • Xinming Tu, James Zou, Weijie J. Su, Linjun Zhang
LLMs can also play a significant role in the classroom as interactive teaching and learning tools, contributing to personalized education.
no code implementations • 9 Jun 2023 • Hua Wang, Sheng Gao, Huanyu Zhang, Weijie J. Su, Milan Shen
In our paper, we introduce DP-HyPO, a pioneering framework for ``adaptive'' private hyperparameter optimization, aiming to bridge the gap between private and non-private hyperparameter optimization.
1 code implementation • 28 May 2023 • Ziang Song, Tianle Cai, Jason D. Lee, Weijie J. Su
This insight allows us to derive closed-form expressions for the reward distribution associated with a set of utility functions in an asymptotic regime.
no code implementations • 27 May 2023 • Lei Wu, Weijie J. Su
By contrast, for gradient descent (GD), the stability imposes a similar constraint but only on the largest eigenvalue of Hessian.
no code implementations • 21 Apr 2023 • Yuling Yan, Weijie J. Su, Jianqing Fan
In 2023, the International Conference on Machine Learning (ICML) required authors with multiple submissions to rank their submissions based on perceived quality.
2 code implementations • 31 Oct 2022 • Hangfeng He, Weijie J. Su
While deep learning has enabled significant advances in many areas of science, its black-box nature hinders architecture design for future artificial intelligence applications and interpretation for high-stakes decision makings.
1 code implementation • 29 Sep 2022 • Yizhou Liu, Weijie J. Su, Tongyang Li
Classical algorithms are often not effective for solving nonconvex optimization problems where local minima are separated by high barriers.
1 code implementation • 9 Jun 2022 • Hua Wang, Sheng Gao, Huanyu Zhang, Milan Shen, Weijie J. Su
Many modern machine learning algorithms are composed of simple private algorithms; thus, an increasingly important problem is to efficiently compute the overall privacy loss under composition.
no code implementations • 6 Jun 2022 • Zhun Deng, Jiayao Zhang, Linjun Zhang, Ting Ye, Yates Coley, Weijie J. Su, James Zou
Specifically, FIFA encourages both classification and fairness generalization and can be flexibly combined with many existing fair learning methods with logits-based losses.
1 code implementation • 31 Jan 2022 • Jiayao Zhang, Hongming Zhang, Weijie J. Su, Dan Roth
Commonsense causality reasoning (CCR) aims at identifying plausible causes and effects in natural language descriptions that are deemed reasonable by an average person.
no code implementations • 17 Dec 2021 • Weijie J. Su
To advance deep learning methodologies in the next decade, a theoretical framework for reasoning about modern neural networks is needed.
no code implementations • 27 Oct 2021 • Weijie J. Su
To address this withholding of information, in this paper, I introduce the Isotonic Mechanism, a simple and efficient approach to improving imprecise raw scores by leveraging certain information that the owner is incentivized to provide.
1 code implementation • NeurIPS 2021 • Jiayao Zhang, Hua Wang, Weijie J. Su
Our main finding uncovers a sharp phase transition phenomenon regarding the {intra-class impact: if the SDEs are locally elastic in the sense that the impact is more significant on samples from the same class as the input, the features of the training data become linearly separable, meaning vanishing training loss; otherwise, the features are not separable, regardless of how long the training time is.
no code implementations • ICLR 2022 • Wenlong Ji, Yiping Lu, Yiliang Zhang, Zhun Deng, Weijie J. Su
We prove that gradient flow on this model converges to critical points of a minimum-norm separation problem exhibiting neural collapse in its global minimizer.
1 code implementation • ICLR 2022 • Shuxiao Chen, Koby Crammer, Hangfeng He, Dan Roth, Weijie J. Su
In this paper, we introduce Target-Aware Weighted Training (TAWT), a weighted training algorithm for cross-task learning based on minimizing a representation-based task distance between the source and target tasks.
1 code implementation • 27 May 2021 • Zhiqi Bu, Jason Klusowski, Cynthia Rush, Weijie J. Su
Sorted l1 regularization has been incorporated into many methods for solving high-dimensional statistical estimation problems, including the SLOPE estimator in linear regression.
no code implementations • 18 May 2021 • Gang Qiao, Weijie J. Su, Li Zhang
Being able to efficiently and accurately select the top-$k$ elements with differential privacy is an integral component of various private data analysis tasks.
no code implementations • 5 Apr 2021 • Jinshuo Dong, Aaron Roth, Weijie J. Su
In this rejoinder, we aim to address two broad issues that cover most comments made in the discussion.
no code implementations • NeurIPS 2021 • Jinshuo Dong, Weijie J. Su, Linjun Zhang
The central question, therefore, is to understand which noise distribution optimizes the privacy-accuracy trade-off, especially when the dimension of the answer vector is high.
no code implementations • 2 Mar 2021 • Shuxiao Chen, Qinqing Zheng, Qi Long, Weijie J. Su
A widely recognized difficulty in federated learning arises from the statistical heterogeneity among clients: local datasets often come from different but not entirely unrelated distributions, and personalization is, therefore, necessary to achieve optimal results from each individual's perspective.
1 code implementation • 22 Feb 2021 • Qinqing Zheng, Shuxiao Chen, Qi Long, Weijie J. Su
Federated learning (FL) is a training paradigm where the clients collaboratively learn models by repeatedly sharing information without compromising much on the privacy of their local sensitive data.
1 code implementation • 29 Jan 2021 • Cong Fang, Hangfeng He, Qi Long, Weijie J. Su
More importantly, when moving to the imbalanced case, our analysis of the Layer-Peeled Model reveals a hitherto unknown phenomenon that we term \textit{Minority Collapse}, which fundamentally limits the performance of deep learning models on the minority classes.
no code implementations • 27 Oct 2020 • Zhun Deng, Hangfeng He, Weijie J. Su
Given that, we propose \emph{locally elastic stability} as a weaker and distribution-dependent stability notion, which still yields exponential generalization bounds.
no code implementations • 22 Oct 2020 • Fan Yang, Hongyang R. Zhang, Sen Wu, Christopher Ré, Weijie J. Su
For example, we can identify a phase transition in the high-dimensional linear regression setting from positive transfer to negative transfer under a model shift between the source and target tasks.
1 code implementation • NeurIPS 2020 • Shuxiao Chen, Hangfeng He, Weijie J. Su
As a popular approach to modeling the dynamics of training overparametrized neural networks (NNs), the neural tangent kernels (NTK) are known to fall behind real-world NNs in generalization ability.
no code implementations • ICML 2020 • Zhun Deng, Hangfeng He, Jiaoyang Huang, Weijie J. Su
An acknowledged weakness of neural networks is their vulnerability to adversarial perturbations to the inputs.
2 code implementations • NeurIPS 2020 • Hua Wang, Yachong Yang, Zhiqi Bu, Weijie J. Su
A fundamental problem in the high-dimensional regression is to understand the tradeoff between type I and type II errors or, equivalently, false discovery rate (FDR) and power in variable selection.
Statistics Theory Information Theory Information Theory Statistics Theory
no code implementations • 15 Apr 2020 • Bin Shi, Weijie J. Su, Michael. I. Jordan
In this paper, we present a general theoretical analysis of the effect of the learning rate in stochastic gradient descent (SGD).
1 code implementation • ICML 2020 • Qinqing Zheng, Jinshuo Dong, Qi Long, Weijie J. Su
To address this question, we introduce a family of analytical and sharp privacy bounds under composition using the Edgeworth expansion in the framework of the recently proposed f-differential privacy.
3 code implementations • 26 Nov 2019 • Zhiqi Bu, Jinshuo Dong, Qi Long, Weijie J. Su
Leveraging the appealing properties of $f$-differential privacy in handling composition and subsampling, this paper derives analytically tractable expressions for the privacy guarantees of both stochastic gradient descent and Adam used in training deep neural networks, without the need of developing sophisticated techniques as [3] did.
1 code implementation • ICLR 2020 • Hangfeng He, Weijie J. Su
This phenomenon is shown to persist for neural networks with nonlinear activation functions through extensive simulations on real-life and synthetic datasets, whereas this is not observed in linear classifiers.
3 code implementations • 7 May 2019 • Jinshuo Dong, Aaron Roth, Weijie J. Su
More precisely, the privacy guarantees of \emph{any} hypothesis testing based definition of privacy (including original DP) converges to GDP in the limit under composition.
no code implementations • NeurIPS 2019 • Bin Shi, Simon S. Du, Weijie J. Su, Michael. I. Jordan
We study first-order optimization methods obtained by discretizing ordinary differential equations (ODEs) corresponding to Nesterov's accelerated gradient methods (NAGs) and Polyak's heavy-ball method.
no code implementations • 21 Oct 2018 • Bin Shi, Simon S. Du, Michael. I. Jordan, Weijie J. Su
We also show that these ODEs are more accurate surrogates for the underlying algorithms; in particular, they not only distinguish between NAG-SC and Polyak's heavy-ball method, but they allow the identification of a term that we refer to as "gradient correction" that is present in NAG-SC but not in the heavy-ball method and is responsible for the qualitative difference in convergence of the two methods.
no code implementations • 11 Jul 2018 • Cynthia Dwork, Weijie J. Su, Li Zhang
Differential privacy provides a rigorous framework for privacy-preserving data analysis.
1 code implementation • 1 Jul 2018 • Edgar Dobriban, Weijie J. Su
In this paper, we propose methods that are robust to large and unequal noise in different observational units (i. e., heteroskedasticity) for statistical inference in linear regression.
Statistics Theory Methodology Statistics Theory
no code implementations • 13 Feb 2018 • Weijie J. Su, Yuancheng Zhu
Stochastic gradient descent (SGD) is an immensely popular approach for online learning in settings where data arrives in a stream or data sizes are very large.
no code implementations • 10 Aug 2017 • Weijie J. Su
In a regime of certain sparsity levels, however, three examples of sequential procedures--forward stepwise, the lasso, and least angle regression--are shown to include the first spurious variable unexpectedly early.
1 code implementation • 11 Oct 2016 • Jingshu Wang, Lin Gui, Weijie J. Su, Chiara Sabatti, Art B. Owen
Replicability is a fundamental quality of scientific discoveries: we are interested in those signals that are detectable in different laboratories, study populations, across time etc.
Methodology