no code implementations • 30 May 2023 • Yu Yang, Eric Gan, Gintare Karolina Dziugaite, Baharan Mirzasoleiman
We further show that if spurious features have a small enough noise-to-signal ratio, the network's output on the majority of examples in a class will be almost exclusively determined by the spurious features and will be nearly invariant to the core feature.
no code implementations • 25 May 2023 • Yihao Xue, Siddharth Joshi, Eric Gan, Pin-Yu Chen, Baharan Mirzasoleiman
However, supervised CL is prone to collapsing representations of subclasses within a class by not capturing all their features, and unsupervised CL may suppress harder class-relevant features by focusing on learning easy class-irrelevant features; both significantly compromise representation quality.
no code implementations • 23 May 2023 • Yihao Xue, Ali Payani, Yu Yang, Baharan Mirzasoleiman
Machine learning models pre-trained on large datasets have achieved remarkable convergence and robustness properties.
no code implementations • 8 Apr 2023 • Yu Yang, Besmira Nushi, Hamid Palangi, Baharan Mirzasoleiman
Spurious correlations that degrade model generalization or lead the model to be right for the wrong reasons are one of the main robustness concerns for real-world deployments.
no code implementations • 24 Mar 2023 • Shayan Fazeli, Lionel Levine, Mehrab Beikzadeh, Baharan Mirzasoleiman, Bita Zadeh, Tara Peris, Majid Sarrafzadeh
Recent advances in remote health monitoring systems have significantly benefited patients and played a crucial role in improving their quality of life.
no code implementations • 20 Mar 2023 • Evan Becker, Jingdong Gao, Ted Zadouri, Baharan Mirzasoleiman
This implies that for a particular run of the algorithms, the solution may be much worse than the provided guarantee in expectation.
no code implementations • 13 Mar 2023 • Wenhan Yang, Baharan Mirzasoleiman
Contrastive vision-language representation learning has achieved state-of-the-art performance for zero-shot classification, by learning from millions of image-caption pairs crawled from the internet.
no code implementations • 11 Mar 2023 • Wenhan Yang, Baharan Mirzasoleiman
Effectively, the high-pass filter captures the dissimilarity between nodes in a neighborhood and the low-pass filter captures the similarity between neighboring nodes. Contrasting the two filtered views allows HLCL to learn rich node representations for graphs, under heterophily and homophily. Empirically, HLCL outperforms state-of-the-art graph CL methods on benchmark heterophily datasets and large-scale real-world datasets by up to 10%.
2 code implementations • 18 Feb 2023 • Siddharth Joshi, Baharan Mirzasoleiman
Interestingly, we also discover the subsets that contribute the most to contrastive learning are those that contribute the least to supervised learning.
no code implementations • 31 Jan 2023 • Omead Pooladzandi, Pasha Khosravi, Erik Nijkamp, Baharan Mirzasoleiman
Generative models have the ability to synthesize data points drawn from the data distribution, however, not all generated samples are high quality.
2 code implementations • 18 Oct 2022 • Yu Yang, Tian Yu Liu, Baharan Mirzasoleiman
Data poisoning causes misclassification of test time target examples by injecting maliciously crafted samples in the training data.
1 code implementation • 15 Oct 2022 • Tian Yu Liu, Baharan Mirzasoleiman
To address this, we propose a rigorous technique to select subsets of data points that when augmented, closely capture the training dynamics of full data augmentation.
no code implementations • 17 Aug 2022 • Yihao Xue, Kyle Whitecross, Baharan Mirzasoleiman
We show that under a sufficiently large noise-to-sample size ratio, generalization error eventually increases with model size.
1 code implementation • 14 Aug 2022 • Tian Yu Liu, Yu Yang, Baharan Mirzasoleiman
We make the key observation that attacks introduce local sharp regions of high training loss, which when minimized, results in learning the adversarial perturbations and makes the attack successful.
no code implementations • 28 Jul 2022 • Omead Pooladzandi, David Davini, Baharan Mirzasoleiman
We propose AdaCore, a method that leverages the geometry of the data to extract subsets of the training examples for efficient machine learning.
no code implementations • 29 Jan 2022 • Yihao Xue, Kyle Whitecross, Baharan Mirzasoleiman
Self-supervised Contrastive Learning (CL) has been recently shown to be very effective in preventing deep networks from overfitting noisy labels.
1 code implementation • 6 May 2021 • Ahmad Khajehnejad, Moein Khajehnejad, Mahmoudreza Babaei, Krishna P. Gummadi, Adrian Weller, Baharan Mirzasoleiman
The potential for machine learning systems to amplify social inequities and unfairness is receiving increasing popular and academic attention.
no code implementations • NeurIPS 2020 • Baharan Mirzasoleiman, Kaidi Cao, Jure Leskovec
Modern neural networks have the capacity to overfit noisy labels frequently found in real-world datasets.
no code implementations • 15 Nov 2020 • Baharan Mirzasoleiman, Kaidi Cao, Jure Leskovec
Modern neural networks have the capacity to overfit noisy labels frequently found in real-world datasets.
Ranked #36 on
Image Classification
on mini WebVision 1.0
no code implementations • 25 Sep 2019 • Baharan Mirzasoleiman, Jeff Bilmes, Jure Leskovec
But because at each epoch the gradients are computed only on the subset S, we obtain a speedup that is inversely proportional to the size of S. Our subset selection algorithm is fully general and can be applied to most IG methods.
1 code implementation • ICLR 2020 • Cody Coleman, Christopher Yeh, Stephen Mussmann, Baharan Mirzasoleiman, Peter Bailis, Percy Liang, Jure Leskovec, Matei Zaharia
By removing hidden layers from the target model, using smaller architectures, and training for fewer epochs, we create proxies that are an order of magnitude faster to train.
3 code implementations • ICML 2020 • Baharan Mirzasoleiman, Jeff Bilmes, Jure Leskovec
Here we develop CRAIG, a method to select a weighted subset (or coreset) of training data that closely estimates the full gradient by maximizing a submodular function.
no code implementations • 3 Jun 2019 • Saeed Vahidian, Baharan Mirzasoleiman, Alexander Cloninger
In a number of situations, collecting a function value for every data point may be prohibitively expensive, and random sampling ignores any structure in the underlying data.
no code implementations • 16 May 2019 • Junaid Ali, Mahmoudreza Babaei, Abhijnan Chakraborty, Baharan Mirzasoleiman, Krishna P. Gummadi, Adish Singla
As we show in this paper, the time-criticality of the information could further exacerbate the disparity of influence across groups.
Social and Information Networks Computers and Society
no code implementations • ICLR 2019 • Cody Coleman, Stephen Mussmann, Baharan Mirzasoleiman, Peter Bailis, Percy Liang, Jure Leskovec, Matei Zaharia
In our approach, we first train a small proxy model quickly, which we then use to estimate the utility of individual training data points, and then select the most informative ones for training the large target model.
no code implementations • NeurIPS 2018 • Elahe Ghalebi, Baharan Mirzasoleiman, Radu Grosu, Jure Leskovec
We propose a novel framework for providing a non-parametric dynamic network model--based on a mixture of coupled hierarchical Dirichlet processes-- based on data capturing cascade node infection times.
no code implementations • ICML 2017 • Baharan Mirzasoleiman, Amin Karbasi, Andreas Krause
How can we summarize a dynamic data stream when elements selected for the summary can be deleted at any time?
1 code implementation • 12 Jun 2017 • Baharan Mirzasoleiman, Stefanie Jegelka, Andreas Krause
The need for real time analysis of rapidly producing data streams (e. g., video and image streams) motivated the design of streaming algorithms that can efficiently extract and summarize useful information from massive data "on the fly".
Data Structures and Algorithms Information Retrieval
no code implementations • NeurIPS 2016 • Baharan Mirzasoleiman, Morteza Zadimoghaddam, Amin Karbasi
The goal is to provide a succinct summary of massive dataset, ideally as small as possible, from which customized summaries can be built for each user, i. e. it can contain elements from the public data (for diversity) and users' private data (for personalization).
no code implementations • 17 Jun 2016 • Andrew An Bian, Baharan Mirzasoleiman, Joachim M. Buhmann, Andreas Krause
Submodular continuous functions are a category of (generally) non-convex/non-concave functions with a wide spectrum of applications.
no code implementations • NeurIPS 2015 • Baharan Mirzasoleiman, Amin Karbasi, Ashwinkumar Badanidiyuru, Andreas Krause
In this paper, we formalize this challenge as a submodular cover problem.
no code implementations • 3 Nov 2014 • Baharan Mirzasoleiman, Amin Karbasi, Rik Sarkar, Andreas Krause
Such problems can often be reduced to maximizing a submodular set function subject to various constraints.
no code implementations • 28 Sep 2014 • Baharan Mirzasoleiman, Ashwinkumar Badanidiyuru, Amin Karbasi, Jan Vondrak, Andreas Krause
Is it possible to maximize a monotone submodular function faster than the widely used lazy greedy algorithm (also known as accelerated greedy), both in theory and practice?
no code implementations • NeurIPS 2013 • Baharan Mirzasoleiman, Amin Karbasi, Rik Sarkar, Andreas Krause
Such problems can often be reduced to maximizing a submodular set function subject to cardinality constraints.