Search Results for author: Vardan Papyan

Found 17 papers, 6 papers with code

Pushing Boundaries: Mixup's Influence on Neural Collapse

no code implementations9 Feb 2024 Quinn Fisher, Haoming Meng, Vardan Papyan

These findings are unexpected, as mixed-up features are not simple convex combinations of feature class means (as one might get, for example, by training mixup with the mean squared error loss).

Data Augmentation

Residual Alignment: Uncovering the Mechanisms of Residual Networks

no code implementations NeurIPS 2023 Jianing Li, Vardan Papyan

Our measurements reveal a process called Residual Alignment (RA) characterized by four properties: (RA1) intermediate representations of a given input are equispaced on a line, embedded in high dimensional space, as observed by Gai and Zhang [2021]; (RA2) top left and right singular vectors of Residual Jacobians align with each other and across different depths; (RA3) Residual Jacobians are at most rank C for fully-connected ResNets, where C is the number of classes; and (RA4) top singular values of Residual Jacobians scale inversely with depth.

Out of the Ordinary: Spectrally Adapting Regression for Covariate Shift

no code implementations29 Dec 2023 Benjamin Eyre, Elliot Creager, David Madras, Vardan Papyan, Richard Zemel

Designing deep neural network classifiers that perform robustly on distributions differing from the available training data is an active area of machine learning research.

Out-of-Distribution Generalization regression

LLM Censorship: A Machine Learning Challenge or a Computer Security Problem?

no code implementations20 Jul 2023 David Glukhov, Ilia Shumailov, Yarin Gal, Nicolas Papernot, Vardan Papyan

Specifically, we demonstrate that semantic censorship can be perceived as an undecidable problem, highlighting the inherent challenges in censorship that arise due to LLMs' programmatic and instruction-following capabilities.

Computer Security Instruction Following

Neural Collapse Under MSE Loss: Proximity to and Dynamics on the Central Path

1 code implementation ICLR 2022 X. Y. Han, Vardan Papyan, David L. Donoho

The analytically-tractable MSE loss offers more mathematical opportunities than the hard-to-analyze CE loss, inspiring us to leverage MSE loss towards the theoretical investigation of NC.

Traces of Class/Cross-Class Structure Pervade Deep Learning Spectra

no code implementations27 Aug 2020 Vardan Papyan

Numerous researchers recently applied empirical spectral analysis to the study of modern deep learning classifiers.

Prevalence of Neural Collapse during the terminal phase of deep learning training

1 code implementation18 Aug 2020 Vardan Papyan, X. Y. Han, David L. Donoho

Modern practice for training classification deepnets involves a Terminal Phase of Training (TPT), which begins at the epoch where training error first vanishes; During TPT, the training error stays effectively zero while training loss is pushed towards zero.

Inductive Bias

Degrees of Freedom Analysis of Unrolled Neural Networks

no code implementations10 Jun 2019 Morteza Mardani, Qingyun Sun, Vardan Papyan, Shreyas Vasanawala, John Pauly, David Donoho

Leveraging the Stein's Unbiased Risk Estimator (SURE), this paper analyzes the generalization risk with its bias and variance components for recurrent unrolled networks.

Image Restoration

Measuring the Spectrum of Deepnet Hessians

no code implementations17 May 2019 Vardan Papyan

We apply state-of-the-art tools in modern high-dimensional numerical linear algebra to approximate efficiently the spectrum of the Hessian of modern deepnets, with tens of millions of parameters, trained on real data.

Measurements of Three-Level Hierarchical Structure in the Outliers in the Spectrum of Deepnet Hessians

1 code implementation24 Jan 2019 Vardan Papyan

We expose a structure in the derivative of the logits with respect to the parameters of the model, which is used to explain the existence of outliers in the spectrum of the Hessian.

The Full Spectrum of Deepnet Hessians at Scale: Dynamics with SGD Training and Sample Size

2 code implementations16 Nov 2018 Vardan Papyan

We decompose the Hessian into different components and study the dynamics with training and sample size of each term individually.

Neural Proximal Gradient Descent for Compressive Imaging

1 code implementation NeurIPS 2018 Morteza Mardani, Qingyun Sun, Shreyas Vasawanala, Vardan Papyan, Hatef Monajemi, John Pauly, David Donoho

Recovering high-resolution images from limited sensory data typically leads to a serious ill-posed inverse problem, demanding inversion algorithms that effectively capture the prior information.

Recurrent Generative Adversarial Networks for Proximal Learning and Automated Compressive Image Recovery

no code implementations27 Nov 2017 Morteza Mardani, Hatef Monajemi, Vardan Papyan, Shreyas Vasanawala, David Donoho, John Pauly

Building effective priors is however challenged by the low train and test overhead dictated by real-time tasks; and the need for retrieving visually "plausible" and physically "feasible" images with minimal hallucination.

Denoising Hallucination +1

Multi-Layer Convolutional Sparse Modeling: Pursuit and Dictionary Learning

no code implementations29 Aug 2017 Jeremias Sulam, Vardan Papyan, Yaniv Romano, Michael Elad

We show that the training of the filters is essential to allow for non-trivial signals in the model, and we derive an online algorithm to learn the dictionaries from real data, effectively resulting in cascaded sparse convolutional layers.

Dictionary Learning

Convolutional Dictionary Learning via Local Processing

1 code implementation ICCV 2017 Vardan Papyan, Yaniv Romano, Jeremias Sulam, Michael Elad

Convolutional Sparse Coding (CSC) is an increasingly popular model in the signal and image processing communities, tackling some of the limitations of traditional patch-based sparse representations.

Dictionary Learning Image Inpainting +1

Multimodal Latent Variable Analysis

no code implementations25 Nov 2016 Vardan Papyan, Ronen Talmon

The first step in our analysis is to find the common source of variability present in all sensor measurements.

Convolutional Neural Networks Analyzed via Convolutional Sparse Coding

no code implementations27 Jul 2016 Vardan Papyan, Yaniv Romano, Michael Elad

This is shown to be tightly connected to CNN, so much so that the forward pass of the CNN is in fact the thresholding pursuit serving the ML-CSC model.

Attribute

Cannot find the paper you are looking for? You can Submit a new open access paper.