Search Results for author: Amin Karbasi

Found 72 papers, 5 papers with code

Streaming Submodular Maximization under a k-Set System Constraint

no code implementations ICML 2020 Ran Haba, Ehsan Kazemi, Moran Feldman, Amin Karbasi

Moreover, we propose the first streaming algorithms for monotone submodular maximization subject to $k$-extendible and $k$-system constraints.

Data Summarization

Data-driven mapping between functional connectomes using optimal transport

no code implementations2 Jul 2021 Javid Dadashkarimi, Amin Karbasi, Dustin Scheinost

Being able to map connectomes and derived results between different atlases without additional pre-processing is a crucial step in improving interpretation and generalization between studies that use different atlases.

Time Series

An Exponential Improvement on the Memorization Capacity of Deep Threshold Networks

no code implementations NeurIPS 2021 Shashank Rajput, Kartik Sreenivasan, Dimitris Papailiopoulos, Amin Karbasi

Recently, Vershynin (2020) settled a long standing question by Baum (1988), proving that \emph{deep threshold} networks can memorize $n$ points in $d$ dimensions using $\widetilde{\mathcal{O}}(e^{1/\delta^2}+\sqrt{n})$ neurons and $\widetilde{\mathcal{O}}(e^{1/\delta^2}(d+\sqrt{n})+n)$ weights, where $\delta$ is the minimum distance between the points.

Submodular + Concave

no code implementations NeurIPS 2021 Siddharth Mitra, Moran Feldman, Amin Karbasi

It has been well established that first order optimization methods can converge to the maximal objective value of concave functions and provide constant factor approximation guarantees for (non-convex/non-concave) continuous submodular functions.

Parallelizing Thompson Sampling

no code implementations NeurIPS 2021 Amin Karbasi, Vahab Mirrokni, Mohammad Shadravan

How can we make use of information parallelism in online decision making problems while efficiently balancing the exploration-exploitation trade-off?

Decision Making

Learning and Certification under Instance-targeted Poisoning

no code implementations18 May 2021 Ji Gao, Amin Karbasi, Mohammad Mahmoody

In this paper, we study PAC learnability and certification of predictions under instance-targeted poisoning attacks, where the adversary who knows the test instance may change a fraction of the training set with the goal of fooling the learner at the test instance.

The Power of Subsampling in Submodular Maximization

no code implementations6 Apr 2021 Christopher Harshaw, Ehsan Kazemi, Moran Feldman, Amin Karbasi

We propose subsampling as a unified algorithmic technique for submodular maximization in centralized and online settings.

Video Summarization

Federated Functional Gradient Boosting

no code implementations11 Mar 2021 Zebang Shen, Hamed Hassani, Satyen Kale, Amin Karbasi

First, in the semi-heterogeneous setting, when the marginal distributions of the feature vectors on client machines are identical, we develop the federated functional gradient boosting (FFGB) method that provably converges to the global minimum.

Federated Learning

Batched Neural Bandits

no code implementations25 Feb 2021 Quanquan Gu, Amin Karbasi, Khashayar Khosravi, Vahab Mirrokni, Dongruo Zhou

In many sequential decision-making problems, the individuals are split into several batches and the decision-maker is only allowed to change her policy at the end of batches.

Decision Making

Online MAP Inference of Determinantal Point Processes

no code implementations NeurIPS 2020 Aditya Bhaskara, Amin Karbasi, Silvio Lattanzi, Morteza Zadimoghaddam

In this paper, we provide an efficient approximation algorithm for finding the most likelihood configuration (MAP) of size $k$ for Determinantal Point Processes (DPP) in the online setting where the data points arrive in an arbitrary order and the algorithm cannot discard the selected elements from its local memory.

Point Processes

How Do You Want Your Greedy: Simultaneous or Repeated?

1 code implementation29 Sep 2020 Moran Feldman, Christopher Harshaw, Amin Karbasi

We also present SubmodularGreedy. jl, a Julia package which implements these algorithms and may be downloaded at https://github. com/crharshaw/SubmodularGreedy. jl .

Multiple Descent: Design Your Own Generalization Curve

no code implementations NeurIPS 2021 Lin Chen, Yifei Min, Mikhail Belkin, Amin Karbasi

This paper explores the generalization loss of linear regression in variably parameterized families of models, both under-parameterized and over-parameterized.

Safe Learning under Uncertain Objectives and Constraints

no code implementations23 Jun 2020 Mohammad Fereydounian, Zebang Shen, Aryan Mokhtari, Amin Karbasi, Hamed Hassani

More precisely, by assuming that Reliable-FW has access to a (stochastic) gradient oracle of the objective function and a noisy feasibility oracle of the safety polytope, it finds an $\epsilon$-approximate first-order stationary point with the optimal ${\mathcal{O}}({1}/{\epsilon^2})$ gradient oracle complexity (resp.

Continuous Submodular Maximization: Beyond DR-Submodularity

no code implementations NeurIPS 2020 Moran Feldman, Amin Karbasi

We first prove that a simple variant of the vanilla coordinate ascent, called Coordinate-Ascent+, achieves a $(\frac{e-1}{2e-1}-\varepsilon)$-approximation guarantee while performing $O(n/\varepsilon)$ iterations, where the computational complexity of each iteration is roughly $O(n/\sqrt{\varepsilon}+n\log n)$ (here, $n$ denotes the dimension of the optimization problem).

Meta Learning in the Continuous Time Limit

no code implementations19 Jun 2020 Ruitu Xu, Lin Chen, Amin Karbasi

In this paper, we establish the ordinary differential equation (ODE) that underlies the training dynamics of Model-Agnostic Meta-Learning (MAML).


Submodularity in Action: From Machine Learning to Signal Processing Applications

no code implementations17 Jun 2020 Ehsan Tohidi, Rouhollah Amiri, Mario Coutino, David Gesbert, Geert Leus, Amin Karbasi

We introduce a variety of submodular-friendly applications, and elucidate the relation of submodularity to convexity and concavity which enables efficient optimization.

The Curious Case of Adversarially Robust Models: More Data Can Help, Double Descend, or Hurt Generalization

no code implementations25 Feb 2020 Yifei Min, Lin Chen, Amin Karbasi

In the medium adversary regime, with more training data, the generalization loss exhibits a double descent curve, which implies the existence of an intermediate stage where more training data hurts the generalization.

Classification General Classification

More Data Can Expand the Generalization Gap Between Adversarially Robust and Standard Models

no code implementations ICML 2020 Lin Chen, Yifei Min, Mingrui Zhang, Amin Karbasi

Despite remarkable success in practice, modern machine learning models have been found to be susceptible to adversarial attacks that make human-imperceptible perturbations to the data, but result in serious and potentially dangerous prediction errors.

Submodular Maximization Through Barrier Functions

no code implementations NeurIPS 2020 Ashwinkumar Badanidiyuru, Amin Karbasi, Ehsan Kazemi, Jan Vondrak

In this paper, we introduce a novel technique for constrained submodular maximization, inspired by barrier functions in continuous optimization.

Regularized Submodular Maximization at Scale

no code implementations10 Feb 2020 Ehsan Kazemi, Shervin Minaee, Moran Feldman, Amin Karbasi

In this paper, we propose scalable methods for maximizing a regularized submodular function $f = g - \ell$ expressed as the difference between a monotone submodular function $g$ and a modular function $\ell$.

Data Summarization Point Processes +1

Streaming Submodular Maximization under a $k$-Set System Constraint

1 code implementation9 Feb 2020 Ran Haba, Ehsan Kazemi, Moran Feldman, Amin Karbasi

In this paper, we propose a novel framework that converts streaming algorithms for monotone submodular maximization into streaming algorithms for non-monotone submodular maximization.

Data Summarization

Stochastic Continuous Greedy ++: When Upper and Lower Bounds Match

no code implementations NeurIPS 2019 Amin Karbasi, Hamed Hassani, Aryan Mokhtari, Zebang Shen

Concretely, for a monotone and continuous DR-submodular function, \SCGPP achieves a tight $[(1-1/e)\OPT -\epsilon]$ solution while using $O(1/\epsilon^2)$ stochastic gradients and $O(1/\epsilon)$ calls to the linear optimization oracle.

Adaptivity in Adaptive Submodularity

no code implementations9 Nov 2019 Hossein Esfandiari, Amin Karbasi, Vahab Mirrokni

We propose an efficient semi adaptive policy that with $O(\log n \times\log k)$ adaptive rounds of observations can achieve an almost tight $1-1/e-\epsilon$ approximation guarantee with respect to an optimal policy that carries out $k$ actions in a fully sequential manner.

Active Learning Decision Making +1

Minimax Regret of Switching-Constrained Online Convex Optimization: No Phase Transition

no code implementations NeurIPS 2020 Lin Chen, Qian Yu, Hannah Lawrence, Amin Karbasi

To establish the dimension-independent upper bound, we next show that a mini-batching algorithm provides an $ O(\frac{T}{\sqrt{K}}) $ upper bound, and therefore conclude that the minimax regret of switching-constrained OCO is $ \Theta(\frac{T}{\sqrt{K}}) $ for any $K$.

Regret Bounds for Batched Bandits

no code implementations11 Oct 2019 Hossein Esfandiari, Amin Karbasi, Abbas Mehrabian, Vahab Mirrokni

We present simple and efficient algorithms for the batched stochastic multi-armed bandit and batched stochastic linear bandit problems.

Multi-Armed Bandits

One Sample Stochastic Frank-Wolfe

no code implementations10 Oct 2019 Mingrui Zhang, Zebang Shen, Aryan Mokhtari, Hamed Hassani, Amin Karbasi

One of the beauties of the projected gradient descent method lies in its rather simple mechanism and yet stable behavior with inexact, stochastic gradients, which has led to its wide-spread use in many machine learning applications.

Submodular Streaming in All its Glory: Tight Approximation, Minimum Memory and Low Adaptive Complexity

no code implementations2 May 2019 Ehsan Kazemi, Marko Mitrovic, Morteza Zadimoghaddam, Silvio Lattanzi, Amin Karbasi

We show how one can achieve the tight $(1/2)$-approximation guarantee with $O(k)$ shared memory while minimizing not only the required rounds of computations but also the total number of communicated bits.

Data Summarization

Submodular Maximization Beyond Non-negativity: Guarantees, Fast Algorithms, and Applications

1 code implementation19 Apr 2019 Christopher Harshaw, Moran Feldman, Justin Ward, Amin Karbasi

It is generally believed that submodular functions -- and the more general class of $\gamma$-weakly submodular functions -- may only be optimized under the non-negativity assumption $f(S) \geq 0$.

Experimental Design

Stochastic Conditional Gradient++

no code implementations19 Feb 2019 Hamed Hassani, Amin Karbasi, Aryan Mokhtari, Zebang Shen

It is known that this rate is optimal in terms of stochastic gradient evaluations.

Stochastic Optimization

Adaptive Sequence Submodularity

1 code implementation NeurIPS 2019 Marko Mitrovic, Ehsan Kazemi, Moran Feldman, Andreas Krause, Amin Karbasi

In many machine learning applications, one needs to interactively select a sequence of items (e. g., recommending movies based on a user's feedback) or make sequential decisions in a certain order (e. g., guiding an agent through a series of states).

Decision Making Link Prediction +1

Black Box Submodular Maximization: Discrete and Continuous Settings

no code implementations28 Jan 2019 Lin Chen, Mingrui Zhang, Hamed Hassani, Amin Karbasi

In this paper, we consider the problem of black box continuous submodular maximization where we only have access to the function values and no information about the derivatives is provided.

Unconstrained Submodular Maximization with Constant Adaptive Complexity

no code implementations15 Nov 2018 Lin Chen, Moran Feldman, Amin Karbasi

In this paper, we consider the unconstrained submodular maximization problem.

Scalable Deletion-Robust Submodular Maximization: Data Summarization with Privacy and Fairness Constraints

no code implementations ICML 2018 Ehsan Kazemi, Morteza Zadimoghaddam, Amin Karbasi

Can we efficiently extract useful information from a large user-generated dataset while protecting the privacy of the users and/or ensuring fairness in representation?

Data Summarization Fairness +1

Data Summarization at Scale: A Two-Stage Submodular Approach

no code implementations ICML 2018 Marko Mitrovic, Ehsan Kazemi, Morteza Zadimoghaddam, Amin Karbasi

The sheer scale of modern datasets has resulted in a dire need for summarization techniques that identify representative elements in a dataset.

Data Summarization

Projection-Free Bandit Convex Optimization

no code implementations18 May 2018 Lin Chen, Mingrui Zhang, Amin Karbasi

In this paper, we propose the first computationally efficient projection-free algorithm for bandit convex optimization (BCO).

Matrix Completion

Stochastic Conditional Gradient Methods: From Convex Minimization to Submodular Maximization

no code implementations24 Apr 2018 Aryan Mokhtari, Hamed Hassani, Amin Karbasi

Further, for a monotone and continuous DR-submodular function and subject to a general convex body constraint, we prove that our proposed method achieves a $((1-1/e)OPT-\eps)$ guarantee with $O(1/\eps^3)$ stochastic gradient computations.

Stochastic Optimization

Projection-Free Online Optimization with Stochastic Gradient: From Convexity to Submodularity

no code implementations ICML 2018 Lin Chen, Christopher Harshaw, Hamed Hassani, Amin Karbasi

We also propose One-Shot Frank-Wolfe, a simpler algorithm which requires only a single stochastic gradient estimate in each round and achieves an $O(T^{2/3})$ stochastic regret bound for convex and continuous submodular optimization.

Comparison Based Learning from Weak Oracles

no code implementations20 Feb 2018 Ehsan Kazemi, Lin Chen, Sanjoy Dasgupta, Amin Karbasi

More specifically, we aim at devising efficient algorithms to locate a target object in a database equipped with a dissimilarity metric via invocation of the weak comparison oracle.

Do Less, Get More: Streaming Submodular Maximization with Subsampling

no code implementations NeurIPS 2018 Moran Feldman, Amin Karbasi, Ehsan Kazemi

In this paper, we develop the first one-pass streaming algorithm for submodular maximization that does not evaluate the entire stream even once.

Video Summarization

Online Continuous Submodular Maximization

no code implementations16 Feb 2018 Lin Chen, Hamed Hassani, Amin Karbasi

For such settings, we then propose an online stochastic gradient ascent algorithm that also achieves a regret bound of $O(\sqrt{T})$ regret, albeit against a weaker $1/2$-approximation to the best feasible solution in hindsight.

Interactive Submodular Bandit

no code implementations NeurIPS 2017 Lin Chen, Andreas Krause, Amin Karbasi

We then receive a noisy feedback about the utility of the action (e. g., ratings) which we model as a submodular function over the context-action space.

Data Summarization News Recommendation

Deletion-Robust Submodular Maximization at Scale

no code implementations20 Nov 2017 Ehsan Kazemi, Morteza Zadimoghaddam, Amin Karbasi

Can we efficiently extract useful information from a large user-generated dataset while protecting the privacy of the users and/or ensuring fairness in representation.

Fairness Feature Selection

Conditional Gradient Method for Stochastic Submodular Maximization: Closing the Gap

no code implementations5 Nov 2017 Aryan Mokhtari, Hamed Hassani, Amin Karbasi

More precisely, for a monotone and continuous DR-submodular function and subject to a \textit{general} convex body constraint, we prove that \alg achieves a $[(1-1/e)\text{OPT} -\eps]$ guarantee (in expectation) with $\mathcal{O}{(1/\eps^3)}$ stochastic gradient computations.

Gradient Methods for Submodular Maximization

no code implementations NeurIPS 2017 Hamed Hassani, Mahdi Soltanolkotabi, Amin Karbasi

Despite the apparent lack of convexity in such functions, we prove that stochastic projected gradient methods can provide strong approximation guarantees for maximizing continuous submodular functions with convex constraints.

Active Learning

Probabilistic Submodular Maximization in Sub-Linear Time

no code implementations ICML 2017 Serban Stan, Morteza Zadimoghaddam, Andreas Krause, Amin Karbasi

As a remedy, we introduce the problem of sublinear time probabilistic submodular maximization: Given training examples of functions (e. g., via user feature vectors), we seek to reduce the ground set so that optimizing new functions drawn from the same distribution will provide almost as much value when restricted to the reduced ground set as when using the full set.

Recommendation Systems

Weakly Submodular Maximization Beyond Cardinality Constraints: Does Randomization Help Greedy?

no code implementations ICML 2018 Lin Chen, Moran Feldman, Amin Karbasi

In this paper, we prove that a randomized version of the greedy algorithm (previously used by Buchbinder et al. (2014) for a different problem) achieves an approximation ratio of $(1 + 1/\gamma)^{-2}$ for the maximization of a weakly submodular function subject to a general matroid constraint, where $\gamma$ is a parameter measuring the distance of the function from submodularity.

Greed is Good: Near-Optimal Submodular Maximization via Greedy Optimization

no code implementations5 Apr 2017 Moran Feldman, Christopher Harshaw, Amin Karbasi

Sample Greedy achieves $(k + 3)$-approximation with only $O(nr/k)$ function evaluations.

Streaming Weak Submodularity: Interpreting Neural Networks on the Fly

1 code implementation NeurIPS 2017 Ethan R. Elenberg, Alexandros G. Dimakis, Moran Feldman, Amin Karbasi

In many machine learning applications, it is important to explain the predictions of a black-box classifier.

Fast Distributed Submodular Cover: Public-Private Data Summarization

no code implementations NeurIPS 2016 Baharan Mirzasoleiman, Morteza Zadimoghaddam, Amin Karbasi

The goal is to provide a succinct summary of massive dataset, ideally as small as possible, from which customized summaries can be built for each user, i. e. it can contain elements from the public data (for diversity) and users' private data (for personalization).

Data Summarization Recommendation Systems

Tradeoffs for Space, Time, Data and Risk in Unsupervised Learning

no code implementations2 May 2016 Mario Lucic, Mesrob I. Ohannessian, Amin Karbasi, Andreas Krause

Using k-means clustering as a prototypical unsupervised learning problem, we show how we can strategically summarize the data (control space) in order to trade off risk and time when data is generated by a probabilistic model.

Submodular Variational Inference for Network Reconstruction

no code implementations29 Mar 2016 Lin Chen, Forrest W. Crawford, Amin Karbasi

In real-world and online social networks, individuals receive and transmit information in real time.

Variational Inference

Near-Optimal Active Learning of Halfspaces via Query Synthesis in the Noisy Setting

no code implementations11 Mar 2016 Lin Chen, Hamed Hassani, Amin Karbasi

This problem has recently gained a lot of interest in automated science and adversarial reverse engineering for which only heuristic algorithms are known.

Active Learning

Seeing the Unseen Network: Inferring Hidden Social Ties from Respondent-Driven Sampling

no code implementations13 Nov 2015 Lin Chen, Forrest W. Crawford, Amin Karbasi

Learning about the social structure of hidden and hard-to-reach populations --- such as drug users and sex workers --- is a major goal of epidemiological and public health research on risk behaviors and disease prevention.

Stochastic Optimization Time Series

Fast Mixing for Discrete Point Processes

no code implementations6 Jun 2015 Patrick Rebeschini, Amin Karbasi

We show that if the set function (not necessarily submodular) displays a natural notion of decay of correlation, then, for $\beta$ small enough, it is possible to design fast mixing Markov chain Monte Carlo methods that yield error bounds on marginal approximations that do not depend on the size of the set $V$.

Point Processes

Distributed Submodular Maximization

no code implementations3 Nov 2014 Baharan Mirzasoleiman, Amin Karbasi, Rik Sarkar, Andreas Krause

Such problems can often be reduced to maximizing a submodular set function subject to various constraints.

Lazier Than Lazy Greedy

no code implementations28 Sep 2014 Baharan Mirzasoleiman, Ashwinkumar Badanidiyuru, Amin Karbasi, Jan Vondrak, Andreas Krause

Is it possible to maximize a monotone submodular function faster than the widely used lazy greedy algorithm (also known as accelerated greedy), both in theory and practice?

Data Summarization

Convolutional Neural Associative Memories: Massive Capacity with Noise Tolerance

no code implementations24 Jul 2014 Amin Karbasi, Amir Hesam Salavati, Amin Shokrollahi

The resulting network has a retrieval capacity that is exponential in the size of the network.

Noise Facilitation in Associative Memories of Exponential Capacity

no code implementations13 Mar 2014 Amin Karbasi, Amir Hesam Salavati, Amin Shokrollahi, Lav R. Varshney

More surprisingly, we show that internal noise actually improves the performance of the recall phase while the pattern retrieval capacity remains intact, i. e., the number of stored patterns does not reduce with noise (up to a threshold).


Near-Optimally Teaching the Crowd to Classify

no code implementations10 Feb 2014 Adish Singla, Ilija Bogunovic, Gábor Bartók, Amin Karbasi, Andreas Krause

How should we present training examples to learners to teach them classification rules?

Noise-Enhanced Associative Memories

no code implementations NeurIPS 2013 Amin Karbasi, Amir Hesam Salavati, Amin Shokrollahi, Lav R. Varshney

More surprisingly, we show that internal noise actually improves the performance of the recall phase.


Neural Networks Built from Unreliable Components

no code implementations26 Jan 2013 Amin Karbasi, Amir Hesam Salavati, Amin Shokrollahi, Lav Varshney

Recent advances in associative memory design through strutured pattern sets and graph-based inference algorithms have allowed the reliable learning and retrieval of an exponential number of patterns.

Coupled Neural Associative Memories

no code implementations8 Jan 2013 Amin Karbasi, Amir Hesam Salavati, Amin Shokrollahi

We propose a novel architecture to design a neural associative memory that is capable of learning a large number of patterns and recalling them later in presence of noise.

From Small-World Networks to Comparison-Based Search

no code implementations15 Jul 2011 Amin Karbasi, Stratis Ioannidis, Laurent Massoulie

In short, a user searching for a target object navigates through a database in the following manner: the user is asked to select the object most similar to her target from a small list of objects.

Cannot find the paper you are looking for? You can Submit a new open access paper.