Search Results for author: Ravid Shwartz-Ziv

Found 17 papers, 8 papers with code

Simplifying Neural Network Training Under Class Imbalance

1 code implementation NeurIPS 2023 Ravid Shwartz-Ziv, Micah Goldblum, Yucen Lily Li, C. Bayan Bruss, Andrew Gordon Wilson

Real-world datasets are often highly class-imbalanced, which can adversely impact the performance of deep learning models.

Data Augmentation

Sudden Drops in the Loss: Syntax Acquisition, Phase Transitions, and Simplicity Bias in MLMs

no code implementations13 Sep 2023 Angelica Chen, Ravid Shwartz-Ziv, Kyunghyun Cho, Matthew L. Leavitt, Naomi Saphra

Most interpretability research in NLP focuses on understanding the behavior and features of a fully trained model.

Variance-Covariance Regularization Improves Representation Learning

no code implementations23 Jun 2023 Jiachen Zhu, Katrina Evtimova, Yubei Chen, Ravid Shwartz-Ziv, Yann Lecun

In summary, VCReg offers a universally applicable regularization framework that significantly advances transfer learning and highlights the connection between gradient starvation, neural collapse, and feature transferability.

Long-tail Learning Representation Learning +2

Reverse Engineering Self-Supervised Learning

1 code implementation NeurIPS 2023 Ido Ben-Shaul, Ravid Shwartz-Ziv, Tomer Galanti, Shai Dekel, Yann Lecun

Self-supervised learning (SSL) is a powerful tool in machine learning, but understanding the learned representations and their underlying mechanisms remains a challenge.

Clustering Representation Learning +1

To Compress or Not to Compress- Self-Supervised Learning and Information Theory: A Review

no code implementations19 Apr 2023 Ravid Shwartz-Ziv, Yann Lecun

Information theory, and notably the information bottleneck principle, has been pivotal in shaping deep neural networks.

Self-Supervised Learning

An Information-Theoretic Perspective on Variance-Invariance-Covariance Regularization

no code implementations1 Mar 2023 Ravid Shwartz-Ziv, Randall Balestriero, Kenji Kawaguchi, Tim G. J. Rudner, Yann Lecun

In this paper, we provide an information-theoretic perspective on Variance-Invariance-Covariance Regularization (VICReg) for self-supervised learning.

Self-Supervised Learning Transfer Learning

What Do We Maximize in Self-Supervised Learning?

no code implementations20 Jul 2022 Ravid Shwartz-Ziv, Randall Balestriero, Yann Lecun

In this paper, we examine self-supervised learning methods, particularly VICReg, to provide an information-theoretical understanding of their construction.

Self-Supervised Learning Transfer Learning

Pre-Train Your Loss: Easy Bayesian Transfer Learning with Informative Priors

1 code implementation20 May 2022 Ravid Shwartz-Ziv, Micah Goldblum, Hossein Souri, Sanyam Kapoor, Chen Zhu, Yann Lecun, Andrew Gordon Wilson

Deep learning is increasingly moving towards a transfer learning paradigm whereby large foundation models are fine-tuned on downstream tasks, starting from an initialization learned on the source task.

Transfer Learning

Information Flow in Deep Neural Networks

no code implementations10 Feb 2022 Ravid Shwartz-Ziv

Then, we propose using the Information Bottleneck (IB) theory to explain deep learning systems.

Spatial-Temporal Convolutional Network for Spread Prediction of COVID-19

no code implementations27 Dec 2020 Ravid Shwartz-Ziv, Itamar Ben Ari, Amitai Armon

In this work we present a spatial-temporal convolutional neural network for predicting future COVID-19 related symptoms severity among a population, per region, given its past reported symptoms.

The Dual Information Bottleneck

1 code implementation8 Jun 2020 Zoe Piran, Ravid Shwartz-Ziv, Naftali Tishby

The Information Bottleneck (IB) framework is a general characterization of optimal representations obtained using a principled approach for balancing accuracy and complexity.

Information Plane

Information in Infinite Ensembles of Infinitely-Wide Neural Networks

1 code implementation pproximateinference AABI Symposium 2019 Ravid Shwartz-Ziv, Alexander A. Alemi

In this preliminary work, we study the generalization properties of infinite ensembles of infinitely-wide neural networks.

REPRESENTATION COMPRESSION AND GENERALIZATION IN DEEP NEURAL NETWORKS

no code implementations ICLR 2019 Ravid Shwartz-Ziv, Amichai Painsky, Naftali Tishby

Specifically, we show that the training of the network is characterized by a rapid increase in the mutual information (MI) between the layers and the target label, followed by a longer decrease in the MI between the layers and the input variable.

Information Plane

Opening the Black Box of Deep Neural Networks via Information

13 code implementations2 Mar 2017 Ravid Shwartz-Ziv, Naftali Tishby

Previous work proposed to analyze DNNs in the \textit{Information Plane}; i. e., the plane of the Mutual Information values that each layer preserves on the input and output variables.

Information Plane

Cannot find the paper you are looking for? You can Submit a new open access paper.