Search Results for author: Marvin Li

Found 4 papers, 1 papers with code

Bias Begets Bias: The Impact of Biased Embeddings on Diffusion Models

no code implementations15 Sep 2024 Sahil Kuchlous, Marvin Li, Jeffrey G. Wang

By specifically adapting the perspective of embedding spaces, we establish new fairness conditions for diffusion model development and evaluation.

Diversity Fairness

Critical windows: non-asymptotic theory for feature emergence in diffusion models

no code implementations3 Mar 2024 Marvin Li, Sitan Chen

Additionally, preliminary experiments on Stable Diffusion suggest critical windows may serve as a useful tool for diagnosing fairness and privacy violations in real-world diffusion models.

Fairness Image Generation

Pandora's White-Box: Precise Training Data Detection and Extraction in Large Language Models

1 code implementation26 Feb 2024 Jeffrey G. Wang, Jason Wang, Marvin Li, Seth Neel

In fine-tuning, we find that a simple attack based on the ratio of the loss between the base and fine-tuned models is able to achieve near-perfect MIA performance; we then leverage our MIA to extract a large fraction of the fine-tuning dataset from fine-tuned Pythia and Llama models.

Language Modelling

MoPe: Model Perturbation-based Privacy Attacks on Language Models

no code implementations22 Oct 2023 Marvin Li, Jason Wang, Jeffrey Wang, Seth Neel

In this paper, we present Model Perturbations (MoPe), a new method to identify with high confidence if a given text is in the training data of a pre-trained language model, given white-box access to the models parameters.

Language Modelling Memorization

Cannot find the paper you are looking for? You can Submit a new open access paper.