no code implementations • ICCV 2023 • Yonatan Dukler, Benjamin Bowman, Alessandro Achille, Aditya Golatkar, Ashwin Swaminathan, Stefano Soatto
We present Synergy Aware Forgetting Ensemble (SAFE), a method to adapt large models on a diverse collection of data while minimizing the expected cost to remove the influence of training samples from the trained model.
no code implementations • CVPR 2023 • Rajshekhar Das, Yonatan Dukler, Avinash Ravichandran, Ashwin Swaminathan
Prompt learning is an efficient approach to adapt transformers by inserting learnable set of parameters into the input and intermediate representations of a pre-trained model.
no code implementations • ICLR 2022 • Yonatan Dukler, Alessandro Achille, Giovanni Paolini, Avinash Ravichandran, Marzia Polito, Stefano Soatto
A learning task is a function from a training set to the validation error, which can be represented by a trained deep neural network (DNN).
no code implementations • ICML 2020 • Yonatan Dukler, Quanquan Gu, Guido Montúfar
The success of deep neural networks is in part due to the use of normalization layers.
no code implementations • 25 Sep 2019 • Yonatan Dukler, Quanquan Gu, Guido Montufar
We present a proof of convergence for ReLU networks trained with weight normalization.
no code implementations • 15 Sep 2019 • Alex Tong Lin, Yonatan Dukler, Wuchen Li, Guido Montufar
We propose regularization strategies for learning discriminative models that are robust to in-class variations of the input data.