Search Results for author: Eric Gan

Found 5 papers, 1 papers with code

Investigating the Benefits of Projection Head for Representation Learning

no code implementations18 Mar 2024 Yihao Xue, Eric Gan, Jiayi Ni, Siddharth Joshi, Baharan Mirzasoleiman

An effective technique for obtaining high-quality representations is adding a projection head on top of the encoder during training, then discarding it and using the pre-projection representations.

Contrastive Learning Data Augmentation +1

Inference and Interference: The Role of Clipping, Pruning and Loss Landscapes in Differentially Private Stochastic Gradient Descent

no code implementations12 Nov 2023 Lauren Watson, Eric Gan, Mohan Dantam, Baharan Mirzasoleiman, Rik Sarkar

Differentially private stochastic gradient descent (DP-SGD) is known to have poorer training and test performance on large neural networks, compared to ordinary stochastic gradient descent (SGD).

Dimensionality Reduction

Identifying Spurious Biases Early in Training through the Lens of Simplicity Bias

no code implementations30 May 2023 Yu Yang, Eric Gan, Gintare Karolina Dziugaite, Baharan Mirzasoleiman

In this work, we provide the first theoretical analysis of the effect of simplicity bias on learning spurious correlations.

Inductive Bias

Which Features are Learnt by Contrastive Learning? On the Role of Simplicity Bias in Class Collapse and Feature Suppression

no code implementations25 May 2023 Yihao Xue, Siddharth Joshi, Eric Gan, Pin-Yu Chen, Baharan Mirzasoleiman

However, supervised CL is prone to collapsing representations of subclasses within a class by not capturing all their features, and unsupervised CL may suppress harder class-relevant features by focusing on learning easy class-irrelevant features; both significantly compromise representation quality.

Contrastive Learning Representation Learning

Seeing is Believing: Brain-Inspired Modular Training for Mechanistic Interpretability

1 code implementation4 May 2023 Ziming Liu, Eric Gan, Max Tegmark

We introduce Brain-Inspired Modular Training (BIMT), a method for making neural networks more modular and interpretable.

Cannot find the paper you are looking for? You can Submit a new open access paper.