Search Results for author: Holden Lee

Found 24 papers, 3 papers with code

How Flawed is ECE? An Analysis via Logit Smoothing

no code implementations15 Feb 2024 Muthu Chidambaram, Holden Lee, Colin McSwiggen, Semon Rezchikov

Informally, a model is calibrated if its predictions are correct with a probability that matches the confidence of the prediction.

Image Classification

Principled Gradient-based Markov Chain Monte Carlo for Text Generation

no code implementations29 Dec 2023 Li Du, Afra Amini, Lucas Torroba Hennigen, Xinyan Velocity Yu, Jason Eisner, Holden Lee, Ryan Cotterell

Recent papers have demonstrated the possibility of energy-based text generation by adapting gradient-based sampling algorithms, a paradigm of MCMC algorithms that promises fast convergence.

Language Modelling Text Generation

Improved Bound for Mixing Time of Parallel Tempering

no code implementations3 Apr 2023 Holden Lee, Zeyu Shen

In this paper, we present a new lower bound for parallel tempering on the spectral gap that has a polynomial dependence on all parameters except $\log L$, where $(L + 1)$ is the number of levels.

Fisher information lower bounds for sampling

no code implementations5 Oct 2022 Sinho Chewi, Patrik Gerber, Holden Lee, Chen Lu

We prove two lower bounds for the complexity of non-log-concave sampling within the framework of Balasubramanian et al. (2022), who introduced the use of Fisher information (FI) bounds as a notion of approximate first-order stationarity in sampling.

Pitfalls of Gaussians as a noise distribution in NCE

no code implementations1 Oct 2022 Holden Lee, Chirag Pabbaraju, Anish Sevekari, Andrej Risteski

Noise Contrastive Estimation (NCE) is a popular approach for learning probability density functions parameterized up to a constant of proportionality.

Convergence of score-based generative modeling for general data distributions

no code implementations26 Sep 2022 Holden Lee, Jianfeng Lu, Yixin Tan

Score-based generative modeling (SGM) has grown to be a hugely successful method for learning to generate samples from complex data distributions such as that of images and audio.

Denoising

Convergence for score-based generative modeling with polynomial complexity

no code implementations13 Jun 2022 Holden Lee, Jianfeng Lu, Yixin Tan

Using our guarantee, we give a theoretical analysis of score-based generative modeling, which transforms white-noise input into samples from a learned data distribution given score estimates at different noise scales.

Sampling Approximately Low-Rank Ising Models: MCMC meets Variational Methods

no code implementations17 Feb 2022 Frederic Koehler, Holden Lee, Andrej Risteski

We consider Ising models on the hypercube with a general interaction matrix $J$, and give a polynomial time sampling algorithm when all but $O(1)$ eigenvalues of $J$ lie in an interval of length one, a situation which occurs in many models of interest.

Variational Inference

Universal Approximation Using Well-Conditioned Normalizing Flows

no code implementations NeurIPS 2021 Holden Lee, Chirag Pabbaraju, Anish Prasad Sevekari, Andrej Risteski

As ill-conditioned Jacobians are an obstacle for likelihood-based training, the fundamental question remains: which distributions can be approximated using well-conditioned affine coupling flows?

Universal Approximation for Log-concave Distributions using Well-conditioned Normalizing Flows

no code implementations ICML Workshop INNF 2021 Holden Lee, Chirag Pabbaraju, Anish Sevekari, Andrej Risteski

As ill-conditioned Jacobians are an obstacle for likelihood-based training, the fundamental question remains: which distributions can be approximated using well-conditioned affine coupling flows?

Improved rates for prediction and identification for partially observed linear dynamical systems

no code implementations NeurIPS 2021 Holden Lee

Identification of a linear time-invariant dynamical system from partial observations is a fundamental problem in control theory.

Improved rates for prediction and identification of partially observed linear dynamical systems

1 code implementation19 Nov 2020 Holden Lee

Identification of a linear time-invariant dynamical system from partial observations is a fundamental problem in control theory.

Efficient sampling from the Bingham distribution

no code implementations30 Sep 2020 Rong Ge, Holden Lee, Jianfeng Lu, Andrej Risteski

We give a algorithm for exact sampling from the Bingham distribution $p(x)\propto \exp(x^\top A x)$ on the sphere $\mathcal S^{d-1}$ with expected runtime of $\operatorname{poly}(d, \lambda_{\max}(A)-\lambda_{\min}(A))$.

No-Regret Prediction in Marginally Stable Systems

no code implementations6 Feb 2020 Udaya Ghai, Holden Lee, Karan Singh, Cyril Zhang, Yi Zhang

This requires a refined regret analysis, including a structural lemma showing the current state of the system to be a small linear combination of past states, even if the state grows polynomially.

LEMMA

Estimating Normalizing Constants for Log-Concave Distributions: Algorithms and Lower Bounds

no code implementations8 Nov 2019 Rong Ge, Holden Lee, Jianfeng Lu

Estimating the normalizing constant of an unnormalized probability distribution has important applications in computer science, statistical physics, machine learning, and statistics.

Robust guarantees for learning an autoregressive filter

no code implementations23 May 2019 Holden Lee, Cyril Zhang

The optimal predictor for a linear dynamical system (with hidden state and Gaussian noise) takes the form of an autoregressive linear filter, namely the Kalman filter.

Time Series Time Series Prediction

Online Sampling from Log-Concave Distributions

1 code implementation NeurIPS 2019 Holden Lee, Oren Mangoubi, Nisheeth K. Vishnoi

Given a sequence of convex functions $f_0, f_1, \ldots, f_T$, we study the problem of sampling from the Gibbs distribution $\pi_t \propto e^{-\sum_{k=0}^tf_k}$ for each epoch $t$ in an online manner.

regression

Simulated Tempering Langevin Monte Carlo II: An Improved Proof using Soft Markov Chain Decomposition

no code implementations29 Nov 2018 Rong Ge, Holden Lee, Andrej Risteski

Previous approaches rely on decomposing the state space as a partition of sets, while our approach can be thought of as decomposing the stationary measure as a mixture of distributions (a "soft partition").

Spectral Filtering for General Linear Dynamical Systems

no code implementations NeurIPS 2018 Elad Hazan, Holden Lee, Karan Singh, Cyril Zhang, Yi Zhang

We give a polynomial-time algorithm for learning latent-state linear dynamical systems without system identification, and without assumptions on the spectral radius of the system's transition matrix.

Towards Provable Control for Unknown Linear Dynamical Systems

no code implementations ICLR 2018 Sanjeev Arora, Elad Hazan, Holden Lee, Karan Singh, Cyril Zhang, Yi Zhang

We study the control of symmetric linear dynamical systems with unknown dynamics and a hidden state.

On the ability of neural nets to express distributions

no code implementations22 Feb 2017 Holden Lee, Rong Ge, Tengyu Ma, Andrej Risteski, Sanjeev Arora

We take a first cut at explaining the expressivity of multilayer nets by giving a sufficient criterion for a function to be approximable by a neural network with $n$ hidden layers.

Cannot find the paper you are looking for? You can Submit a new open access paper.