1 code implementation • 22 May 2024 • Hugh Dance, Benjamin Bloem-Reddy
We identify a local symmetry property satisfied by a large class of causal models under such interventions.
1 code implementation • 29 Aug 2023 • Gian Carlo Diluvi, Benjamin Bloem-Reddy, Trevor Campbell
First, we develop a measure-preserving and discrete (MAD) invertible map that leaves the discrete target invariant, and then create a mixed variational flow (MAD Mix) based on that map.
1 code implementation • 28 Jul 2023 • Kenny Chiu, Benjamin Bloem-Reddy
Finally, we apply them to testing for symmetry in geomagnetic satellite data and in two problems from high-energy particle physics.
no code implementations • 2 Jun 2022 • Quanhan Xi, Benjamin Bloem-Reddy
Most modern probabilistic generative models, such as the variational autoencoder (VAE), have certain indeterminacies that are unresolvable even with an infinite amount of data.
no code implementations • NeurIPS Workshop ICBINB 2021 • Benjamin Bloem-Reddy
Drawing on empirical research and theories of beauty from psychology, and on philosophical investigations into the role of beauty in scientific research, I argue that aesthetic considerations are valuable to machine research in two modes.
1 code implementation • NeurIPS 2021 • Yann Dubois, Benjamin Bloem-Reddy, Karen Ullrich, Chris J. Maddison
Most data is automatically collected and only ever "seen" by algorithms.
Ranked #1 on Image Compression on Oxford-IIIT Pet Dataset (using extra training data)
no code implementations • 8 Oct 2020 • Saeid Naderiparizi, Kenny Chiu, Benjamin Bloem-Reddy, Frank Wood
We aim this work to be a counterpoint to a recent trend in the literature that stresses achieving good samples when the amount of conditioning data is large.
no code implementations • 1 May 2020 • Clare Lyle, Mark van der Wilk, Marta Kwiatkowska, Yarin Gal, Benjamin Bloem-Reddy
Many real world data analysis problems exhibit invariant structure, and models that take advantage of this structure have shown impressive empirical performance, particularly in deep learning.
no code implementations • 18 Jan 2019 • Benjamin Bloem-Reddy, Yee Whye Teh
Treating neural network inputs and outputs as random variables, we characterize the structure of neural networks that can be used to model data that are invariant or equivariant under the action of a compact group.
no code implementations • 13 Jul 2018 • Martin Tegner, Benjamin Bloem-Reddy, Stephen Roberts
We consider the problem of inferring a latent function in a probabilistic model of data.
1 code implementation • 9 Jul 2018 • Benjamin Bloem-Reddy, Adam Foster, Emile Mathieu, Yee Whye Teh
Empirical evidence suggests that heavy-tailed degree distributions occurring in many real networks are well-approximated by power laws with exponents $\eta$ that may take values either less than and greater than two.
1 code implementation • 19 Dec 2016 • Benjamin Bloem-Reddy, Peter Orbanz
We introduce a class of generative network models that insert edges by connecting the starting and terminal vertices of a random walk on the network graph.