1 code implementation • 15 Jun 2021 • Vikram Voleti, Chris Finlay, Adam Oberman, Christopher Pal
In this work we introduce a Multi-Resolution variant of such models (MRCNF), by characterizing the conditional distribution over the additional information required to generate a fine image that is consistent with the coarse image.
Ranked #6 on Image Generation on ImageNet 64x64 (Bits per dim metric)
no code implementations • ICML Workshop INNF 2021 • Vikram Voleti, Chris Finlay, Adam M Oberman, Christopher Pal
Recent work has shown that Continuous Normalizing Flows (CNFs) can serve as generative models of images with exact likelihood calculation and invertible generation/density estimation.
1 code implementation • 5 Oct 2020 • Ryan Campbell, Chris Finlay, Adam M Oberman
Machine learning models are vulnerable to adversarial attacks.
1 code implementation • 10 Jul 2020 • Chris Finlay, Bruce A. Bassett
Multiple studies have suggested the spread of COVID-19 is affected by factors such as climate, BCG vaccinations, pollution and blood type.
1 code implementation • 10 Jun 2020 • Ryan Campbell, Chris Finlay, Adam M. Oberman
We present a deterministic method to compute the Gaussian average of neural networks used in regression and classification.
no code implementations • 10 Jun 2020 • Chris Finlay, Augusto Gerolin, Adam M. Oberman, Aram-Alexandre Pooladian
We approach the problem of learning continuous normalizing flows from a dual perspective motivated by entropy-regularized optimal transport, in which continuous normalizing flows are cast as gradients of scalar potential functions.
1 code implementation • 18 May 2020 • Alireza Vafaei Sadr, Bruce A. Bassett, Nadeem Oozeer, Yabebal Fantaye, Chris Finlay
Our results strongly suggest that deep learning on simulations, boosted by transfer learning on real data, will likely play a key role in the future of RFI flagging of radio astronomy data.
Instrumentation and Methods for Astrophysics
2 code implementations • ICML 2020 • Chris Finlay, Jörn-Henrik Jacobsen, Levon Nurbekyan, Adam M. Oberman
Training neural ODEs on large datasets has not been tractable due to the necessity of allowing the adaptive numerical ODE solver to refine its step size to very small values.
Ranked #1 on Density Estimation on CelebA-HQ 256x256
1 code implementation • 4 Oct 2019 • Aram-Alexandre Pooladian, Chris Finlay, Adam M. Oberman
Successfully training deep neural networks often requires either batch normalization, appropriate weight initialization, both of which come with their own challenges.
no code implementations • 25 Sep 2019 • Chris Finlay, Adam M. Oberman
It is well-known that the softmax values of the network are not estimates of the probabilities of class labels.
2 code implementations • 5 Aug 2019 • Aram-Alexandre Pooladian, Chris Finlay, Tim Hoheisel, Adam Oberman
This includes, but is not limited to, $\ell_1, \ell_2$, and $\ell_\infty$ perturbations; the $\ell_0$ counting "norm" (i. e. true sparseness); and the total variation seminorm, which is a (non-$\ell_p$) convolutional dissimilarity measuring local pixel changes.
1 code implementation • 27 May 2019 • Chris Finlay, Adam M. Oberman
In this work we revisit gradient regularization for adversarial robustness with some new ingredients.
1 code implementation • ICCV 2019 • Chris Finlay, Aram-Alexandre Pooladian, Adam M. Oberman
Adversarial attacks formally correspond to an optimization problem: find a minimum norm image perturbation, constrained to cause misclassification.
1 code implementation • 21 Mar 2019 • Adam M. Oberman, Chris Finlay, Alexander Iannantuono, Tiago Salvador
While the accuracy of modern deep learning models has significantly improved in recent years, the ability of these models to generate uncertainty estimates has not progressed to the same degree.
1 code implementation • ICLR 2019 • Chris Finlay, Adam Oberman, Bilal Abbasi
We augment adversarial training (AT) with worst case adversarial training (WCAT) which improves adversarial robustness by 11% over the current state-of-the-art result in the $\ell_2$ norm on CIFAR-10.
no code implementations • 28 Aug 2018 • Chris Finlay, Jeff Calder, Bilal Abbasi, Adam Oberman
In this work we study input gradient regularization of deep neural networks, and demonstrate that such regularization leads to generalization proofs and improved adversarial robustness.