Search Results for author: Bryon Aragam

Found 37 papers, 17 papers with code

On the Origins of Linear Representations in Large Language Models

no code implementations6 Mar 2024 Yibo Jiang, Goutham Rajendran, Pradeep Ravikumar, Bryon Aragam, Victor Veitch

To that end, we introduce a simple latent variable model to abstract and formalize the concept dynamics of the next token prediction.

Language Modelling Large Language Model

Optimal estimation of Gaussian (poly)trees

1 code implementation9 Feb 2024 Yuhao Wang, Ming Gao, Wai Ming Tai, Bryon Aragam, Arnab Bhattacharyya

We develop optimal algorithms for learning undirected Gaussian trees and directed Gaussian polytrees from data.

Inconsistency of cross-validation for structure learning in Gaussian graphical models

no code implementations28 Dec 2023 Zhao Lyu, Wai Ming Tai, Mladen Kolar, Bryon Aragam

In this paper, we highlight the inherent limitations of cross-validation when employed to discern the structure of a Gaussian graphical model.

Model Selection

iSCAN: Identifying Causal Mechanism Shifts among Nonlinear Additive Noise Models

1 code implementation NeurIPS 2023 Tianyu Chen, Kevin Bello, Bryon Aragam, Pradeep Ravikumar

Structural causal models (SCMs) are widely used in various disciplines to represent causal relationships among variables in complex systems.

Learning Linear Causal Representations from Interventions under General Nonlinear Mixing

no code implementations NeurIPS 2023 Simon Buchholz, Goutham Rajendran, Elan Rosenfeld, Bryon Aragam, Bernhard Schölkopf, Pradeep Ravikumar

We study the problem of learning causal representations from unknown, latent interventions in a general setting, where the latent distribution is Gaussian but the mixing function is completely general.

counterfactual

Neuro-Causal Factor Analysis

no code implementations31 May 2023 Alex Markham, MingYu Liu, Bryon Aragam, Liam Solus

Factor analysis (FA) is a statistical tool for studying how observed variables with some mutual dependences can be expressed as functions of mutually independent unobserved factors, and it is widely applied throughout the psychological, biological, and physical sciences.

Causal Discovery

Optimizing NOTEARS Objectives via Topological Swaps

1 code implementation26 May 2023 Chang Deng, Kevin Bello, Bryon Aragam, Pradeep Ravikumar

In this work, we delve into the optimization challenges associated with this class of non-convex programs.

Learning Mixtures of Gaussians with Censored Data

no code implementations6 May 2023 Wai Ming Tai, Bryon Aragam

We study the problem of learning mixtures of Gaussians with censored data.

DAGMA: Learning DAGs via M-matrices and a Log-Determinant Acyclicity Characterization

3 code implementations16 Sep 2022 Kevin Bello, Bryon Aragam, Pradeep Ravikumar

From the optimization side, we drop the typically used augmented Lagrangian scheme and propose DAGMA ($\textit{DAGs via M-matrices for Acyclicity}$), a method that resembles the central path for barrier methods.

Causal Discovery

Identifiability of deep generative models without auxiliary information

no code implementations20 Jun 2022 Bohdan Kivva, Goutham Rajendran, Pradeep Ravikumar, Bryon Aragam

We prove identifiability of a broad class of deep latent variable models that (a) have universal approximation capabilities and (b) are the decoders of variational autoencoders that are commonly used in practice.

A non-graphical representation of conditional independence via the neighbourhood lattice

no code implementations12 Jun 2022 Arash A. Amini, Bryon Aragam, Qing Zhou

We introduce and study the neighbourhood lattice decomposition of a distribution, which is a compact, non-graphical representation of conditional independence that is valid in the absence of a faithful graphical representation.

valid

Tight Bounds on the Hardness of Learning Simple Nonparametric Mixtures

no code implementations28 Mar 2022 Bryon Aragam, Wai Ming Tai

Combining these bounds, we conclude that the optimal sample complexity of this problem properly lies in between polynomial and exponential, which is not common in learning theory.

Density Estimation Learning Theory

Optimal estimation of Gaussian DAG models

1 code implementation25 Jan 2022 Ming Gao, Wai Ming Tai, Bryon Aragam

In other words, at least for Gaussian models with equal error variances, learning a directed graphical model is statistically no more difficult than learning an undirected graphical model.

Tradeoffs of Linear Mixed Models in Genome-wide Association Studies

no code implementations5 Nov 2021 Haohan Wang, Bryon Aragam, Eric Xing

Motivated by empirical arguments that are well-known from the genome-wide association studies (GWAS) literature, we study the statistical properties of linear mixed models (LMMs) applied to GWAS.

NOTMAD: Estimating Bayesian Networks with Sample-Specific Structures and Parameters

1 code implementation1 Nov 2021 Ben Lengerich, Caleb Ellington, Bryon Aragam, Eric P. Xing, Manolis Kellis

We encode the acyclicity constraint as a smooth regularization loss which is back-propagated to the mixing function; in this way, NOTMAD shares information between context-specific acyclic graphs, enabling the estimation of Bayesian network structures and parameters at even single-sample resolution.

Efficient Bayesian network structure learning via local Markov boundary search

1 code implementation NeurIPS 2021 Ming Gao, Bryon Aragam

Perhaps surprisingly, we show that for certain graph ensembles, a simple forward greedy search algorithm (i. e. without a backward pruning phase) suffices to learn the Markov boundary of each node.

Structure learning in polynomial time: Greedy algorithms, Bregman information, and exponential families

no code implementations NeurIPS 2021 Goutham Rajendran, Bohdan Kivva, Ming Gao, Bryon Aragam

Greedy algorithms have long been a workhorse for learning graphical models, and more broadly for learning statistical models with sparse structure.

Uniform Consistency in Nonparametric Mixture Models

no code implementations31 Aug 2021 Bryon Aragam, Ruiyi Yang

We study uniform consistency in nonparametric mixture models as well as closely related mixture of regression (also known as mixed regression) models, where the regression functions are allowed to be nonparametric and the error distributions are assumed to be convolutions of a Gaussian density.

regression

Learning latent causal graphs via mixture oracles

1 code implementation NeurIPS 2021 Bohdan Kivva, Goutham Rajendran, Pradeep Ravikumar, Bryon Aragam

We study the problem of reconstructing a causal graphical model from data in the presence of latent variables.

Fundamental Limits and Tradeoffs in Invariant Representation Learning

no code implementations NeurIPS 2023 Han Zhao, Chen Dan, Bryon Aragam, Tommi S. Jaakkola, Geoffrey J. Gordon, Pradeep Ravikumar

A wide range of machine learning applications such as privacy-preserving learning, algorithmic fairness, and domain adaptation/generalization among others, involve learning invariant representations of the data that aim to achieve two competing goals: (a) maximize information or accuracy with respect to a target response, and (b) maximize invariance or independence with respect to a set of protected features (e. g., for fairness, privacy, etc).

Domain Adaptation Fairness +4

A polynomial-time algorithm for learning nonparametric causal graphs

1 code implementation NeurIPS 2020 Ming Gao, Yi Ding, Bryon Aragam

We establish finite-sample guarantees for a polynomial-time algorithm for learning a nonlinear, nonparametric directed acyclic graphical (DAG) model from data.

DYNOTEARS: Structure Learning from Time-Series Data

4 code implementations2 Feb 2020 Roxana Pamfil, Nisara Sriwattanaworachai, Shaan Desai, Philip Pilgerstorfer, Paul Beaumont, Konstantinos Georgatzis, Bryon Aragam

Compared to state-of-the-art methods for learning dynamic Bayesian networks, our method is both scalable and accurate on real data.

Time Series Time Series Analysis

Automated Dependence Plots

2 code implementations2 Dec 2019 David I. Inouye, Liu Leqi, Joon Sik Kim, Bryon Aragam, Pradeep Ravikumar

To address these drawbacks, we formalize a method for automating the selection of interesting PDPs and extend PDPs beyond showing single features to show the model response along arbitrary directions, for example in raw feature space or a latent space arising from some generative model.

Model Selection Selection bias

Globally optimal score-based learning of directed acyclic graphs in high-dimensions

no code implementations NeurIPS 2019 Bryon Aragam, Arash Amini, Qing Zhou

We prove that $\Omega(s\log p)$ samples suffice to learn a sparse Gaussian directed acyclic graph (DAG) from data, where $s$ is the maximum Markov blanket size.

valid

Learning Sample-Specific Models with Low-Rank Personalized Regression

1 code implementation NeurIPS 2019 Benjamin Lengerich, Bryon Aragam, Eric P. Xing

Modern applications of machine learning (ML) deal with increasingly heterogeneous datasets comprised of data collected from overlapping latent subpopulations.

regression

Learning Sparse Nonparametric DAGs

2 code implementations29 Sep 2019 Xun Zheng, Chen Dan, Bryon Aragam, Pradeep Ravikumar, Eric P. Xing

We develop a framework for learning sparse nonparametric directed acyclic graphs (DAGs) from data.

Causal Discovery

On perfectness in Gaussian graphical models

no code implementations3 Sep 2019 Arash A. Amini, Bryon Aragam, Qing Zhou

Knowing when a graphical model is perfect to a distribution is essential in order to relate separation in the graph to conditional independence in the distribution, and this is particularly important when performing inference from data.

The Sample Complexity of Semi-Supervised Learning with Nonparametric Mixture Models

no code implementations NeurIPS 2018 Chen Dan, Liu Leqi, Bryon Aragam, Pradeep K. Ravikumar, Eric P. Xing

We study the sample complexity of semi-supervised learning (SSL) and introduce new assumptions based on the mismatch between a mixture model learned from unlabeled data and the true mixture model induced by the (unknown) class conditional distributions.

Binary Classification Classification +2

Fault Tolerance in Iterative-Convergent Machine Learning

no code implementations17 Oct 2018 Aurick Qiao, Bryon Aragam, Bingjing Zhang, Eric P. Xing

In this paper, we develop a general framework to quantify the effects of calculation errors on iterative-convergent algorithms and use this framework to design new strategies for checkpoint-based fault tolerance.

BIG-bench Machine Learning

Sample Complexity of Nonparametric Semi-Supervised Learning

no code implementations NeurIPS 2018 Chen Dan, Liu Leqi, Bryon Aragam, Pradeep Ravikumar, Eric P. Xing

We study the sample complexity of semi-supervised learning (SSL) and introduce new assumptions based on the mismatch between a mixture model learned from unlabeled data and the true mixture model induced by the (unknown) class conditional distributions.

Binary Classification Classification +2

Identifiability of Nonparametric Mixture Models and Bayes Optimal Clustering

no code implementations12 Feb 2018 Bryon Aragam, Chen Dan, Eric P. Xing, Pradeep Ravikumar

Motivated by problems in data clustering, we establish general conditions under which families of nonparametric mixture models are identifiable, by introducing a novel framework involving clustering overfitted \emph{parametric} (i. e. misspecified) mixture models.

Clustering Nonparametric Clustering

The neighborhood lattice for encoding partial correlations in a Hilbert space

1 code implementation3 Nov 2017 Arash A. Amini, Bryon Aragam, Qing Zhou

We study the computational complexity of computing these structures and show that under a sparsity assumption, they can be computed in polynomial time, even in the absence of the assumption of perfectness to a graph.

Dimensionality Reduction regression

Learning Large-Scale Bayesian Networks with the sparsebn Package

2 code implementations11 Mar 2017 Bryon Aragam, Jiaying Gu, Qing Zhou

To meet this challenge, we have developed a new R package called sparsebn for learning the structure of large, sparse graphical models with a focus on Bayesian networks.

Learning Directed Acyclic Graphs with Penalized Neighbourhood Regression

1 code implementation29 Nov 2015 Bryon Aragam, Arash A. Amini, Qing Zhou

We study a family of regularized score-based estimators for learning the structure of a directed acyclic graph (DAG) for a multivariate normal distribution from high-dimensional data with $p\gg n$.

regression

Concave Penalized Estimation of Sparse Gaussian Bayesian Networks

no code implementations4 Jan 2014 Bryon Aragam, Qing Zhou

We develop a penalized likelihood estimation framework to estimate the structure of Gaussian Bayesian networks from observational data.

Cannot find the paper you are looking for? You can Submit a new open access paper.