no code implementations • 11 Mar 2024 • Weixin Liang, Zachary Izzo, Yaohui Zhang, Haley Lepp, Hancheng Cao, Xuandong Zhao, Lingjiao Chen, Haotian Ye, Sheng Liu, Zhi Huang, Daniel A. McFarland, James Y. Zou
We present an approach for estimating the fraction of text in a large corpus which is likely to be substantially modified or produced by a large language model (LLM).
1 code implementation • 29 Apr 2023 • Zachary Izzo, Ruishan Liu, James Zou
To do this, simple parametric models are frequently used (e. g. coefficients of linear regression) but usually fitted on the whole dataset.
no code implementations • 12 Nov 2022 • Zachary Izzo, Jinsung Yoon, Sercan O. Arik, James Zou
However, DP's strong theoretical guarantees often come at the cost of a large drop in its utility for machine learning, and DP guarantees themselves can be difficult to interpret.
no code implementations • 14 Oct 2022 • Yuhua Zhu, Zachary Izzo, Lexing Ying
The optimal policy for the limiting HJB equation can be explicitly obtained for several common bandit problems, and we give numerical methods to solve the HJB equation when an explicit solution is not available.
no code implementations • 19 Sep 2022 • Yiping Lu, Wenlong Ji, Zachary Izzo, Lexing Ying
In this paper, we propose importance tempering to improve the decision boundary and achieve consistently better results for overparameterized models.
no code implementations • 13 Dec 2021 • Zachary Izzo, James Zou, Lexing Ying
A recent line of work has focused on training machine learning (ML) models in the performative setting, i. e. when the data distribution reacts to the deployed model.
no code implementations • NeurIPS 2021 • Zachary Izzo, Sandeep Silwal, Samson Zhou
In order to cope with this "curse of dimensionality," we study dimensionality reduction techniques for the Wasserstein barycenter problem.
1 code implementation • 15 Feb 2021 • Zachary Izzo, Lexing Ying, James Zou
Performative distribution shift captures the setting where the choice of which ML model is deployed changes the data distribution.
no code implementations • 24 Feb 2020 • Zachary Izzo, Mary Anne Smart, Kamalika Chaudhuri, James Zou
Deleting data from a trained machine learning (ML) model is a critical task in many applications.