no code implementations • 16 Apr 2024 • Eric Yeats, Cameron Darwin, Eduardo Ortega, Frank Liu, Hai Li
We leverage diffusion models to study the robustness-performance tradeoff of robust classifiers.
no code implementations • 3 Apr 2024 • Jingyang Zhang, Jingwei Sun, Eric Yeats, Yang Ouyang, Martin Kuo, Jianyi Zhang, Hao Yang, Hai Li
The problem of pre-training data detection for large language models (LLMs) has received growing attention due to its implications in critical issues like copyright violation and test data contamination.
no code implementations • 11 Dec 2023 • Eric Yeats, Cameron Darwin, Frank Liu, Hai Li
Quantification of the number of variables needed to locally explain complex data is often the first step to better understanding it.
1 code implementation • 8 Feb 2023 • Eric Yeats, Frank Liu, Hai Li
Disentangled learning representations have promising utility in many applications, but they currently suffer from serious reliability issues.
1 code implementation • 21 Sep 2022 • Eric Yeats, Frank Liu, David Womble, Hai Li
We present a self-supervised method to disentangle factors of variation in high-dimensional data that does not rely on prior knowledge of the underlying variation profile (e. g., no assumptions on the number or distribution of the individual latent variables to be extracted).