no code implementations • 23 Oct 2024 • Rana Shahout, Cong Liang, Shiji Xin, Qianru Lao, Yong Cui, Minlan Yu, Michael Mitzenmacher
In this paper, we propose LAMPS, a novel LLM inference framework for augmented LLMs.
no code implementations • 1 Oct 2024 • Rana Shahout, Eran Malach, Chunwei Liu, Weifan Jiang, Minlan Yu, Michael Mitzenmacher
Using these predictions, we propose a prediction-based SRPT variant with limited preemption designed to account for memory overhead in LLM systems.
no code implementations • 17 Sep 2024 • Rana Shahout, Ibrahim Sabek, Michael Mitzenmacher
We show how to utilize machine learning approaches to improve sliding window algorithms for approximate frequency estimation problems, under the ``algorithms with predictions'' framework.
no code implementations • 1 Jul 2024 • Wenchen Han, Shay Vargaftik, Michael Mitzenmacher, Brad Karp, Ran Ben Basat
Gradient aggregation has long been identified as a major bottleneck in today's large-scale distributed machine learning training systems.
no code implementations • 24 Jun 2024 • Rana Shahout, Michael Mitzenmacher
In this paper, we present the first learned competing-counter-based algorithm, called LSS, for identifying heavy hitters, top k, and flow frequency estimation that utilizes the well-known Space Saving algorithm.
no code implementations • 5 Feb 2024 • Rana Shahout, Michael Mitzenmacher
SkipPredict prioritizes predicted short jobs over long jobs, and for the latter, SkipPredict applies a second round of more detailed "expensive predictions" to approximate Shortest Remaining Processing Time for these jobs.
no code implementations • 5 Feb 2024 • Ran Ben-Basat, Yaniv Ben-Itzhak, Michael Mitzenmacher, Shay Vargaftik
Quantization is a fundamental optimization for many machine-learning use cases, including compressing gradients, model weights and activations, and datasets.
1 code implementation • 16 Feb 2023 • Minghao Li, Ran Ben Basat, Shay Vargaftik, ChonLam Lao, Kevin Xu, Michael Mitzenmacher, Minlan Yu
To address this bottleneck and accelerate training, a widely-deployed approach is compression.
2 code implementations • 30 Jun 2022 • Eric R. Knorr, Baptiste Lemaire, Andrew Lim, Siqiang Luo, Huanchen Zhang, Stratos Idreos, Michael Mitzenmacher
We introduce Proteus, a novel self-designing approximate range filter, which configures itself based on sampled data in order to optimize its false positive rate (FPR) for a given space requirement.
no code implementations • 26 May 2022 • Ran Ben Basat, Shay Vargaftik, Amit Portnoy, Gil Einziger, Yaniv Ben-Itzhak, Michael Mitzenmacher
Distributed Mean Estimation (DME), in which $n$ clients communicate vectors to a parameter server that estimates their average, is a fundamental building block in communication-efficient federated learning.
no code implementations • 5 Mar 2022 • Maximilian Lam, Michael Mitzenmacher, Vijay Janapa Reddi, Gu-Yeon Wei, David Brooks
This enables an online phase where securely computing the result of a nonlinear function requires just a single round of communication, with communication cost equal to twice the number of bits of the input to the nonlinear function.
1 code implementation • 29 Sep 2021 • Max Lam, Michael Mitzenmacher, Vijay Janapa Reddi, Gu-Yeon Wei, David Brooks
Multiparty computation approaches to private neural network inference require significant communication between server and client, incur tremendous runtime penalties, and cost massive storage overheads.
1 code implementation • 19 Aug 2021 • Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, Michael Mitzenmacher
Distributed Mean Estimation (DME) is a central building block in federated learning, where clients send local gradients to a parameter server for averaging and updating the model.
1 code implementation • 10 Jun 2021 • Maximilian Lam, Gu-Yeon Wei, David Brooks, Vijay Janapa Reddi, Michael Mitzenmacher
We show that aggregated model updates in federated learning may be insecure.
1 code implementation • NeurIPS 2021 • Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, Michael Mitzenmacher
We consider the problem where $n$ clients transmit $d$-dimensional real-valued vectors using $d(1+o(1))$ bits each, in a manner that allows the receiver to approximately reconstruct their mean.
no code implementations • ICLR 2021 • Kapil Vaidya, Eric Knorr, Michael Mitzenmacher, Tim Kraska
Bloom filters are space-efficient probabilistic data structures that are used to test whether an element is a member of a set, and may return false positives.
no code implementations • 5 Oct 2020 • Ran Ben-Basat, Michael Mitzenmacher, Shay Vargaftik
We consider both the biased and unbiased estimation problems and aim to minimize the cost.
no code implementations • 27 Jun 2020 • Michael Mitzenmacher
Motivated by recent work on scheduling with predicted job sizes, we consider the performance of scheduling algorithms with minimal advice, namely a single bit.
no code implementations • 5 Jun 2020 • Kapil Vaidya, Eric Knorr, Tim Kraska, Michael Mitzenmacher
Bloom filters are space-efficient probabilistic data structures that are used to test whether an element is a member of a set, and may return false positives.
no code implementations • 21 Sep 2019 • Kasper Green Larsen, Michael Mitzenmacher, Charalampos E. Tsourakakis
The goal is to recover $n$ discrete variables $g_i \in \{0, \ldots, k-1\}$ (up to some global offset) given noisy observations of a set of their pairwise differences $\{(g_i - g_j) \bmod k\}$; specifically, with probability $\frac{1}{k}+\delta$ for some $\delta > 0$ one obtains the correct answer, and with the remaining probability one obtains a uniformly random incorrect answer.
no code implementations • 23 May 2019 • Michael Mitzenmacher, Matteo Dell'Amico
The supermarket model refers to a system with a large number of queues, where new customers choose d queues at random and join the one with the fewest customers.
no code implementations • 30 Jan 2019 • Hossein Esfandiari, Mohammadtaghi Hajiaghayi, Brendan Lucier, Michael Mitzenmacher
We consider online variations of the Pandora's box problem (Weitzman.
no code implementations • 3 Jan 2019 • Michael Mitzenmacher
Recent work has suggested enhancing Bloom filters by using a pre-filter, based on applying machine learning to determine a function that models the data set the Bloom filter is meant to represent.
no code implementations • NeurIPS 2018 • Michael Mitzenmacher
Recent work has suggested enhancing Bloom filters by using a pre-filter, based on applying machine learning to determine a function that models the data set the Bloom filter is meant to represent.
no code implementations • NeurIPS 2018 • Diana Cai, Michael Mitzenmacher, Ryan P. Adams
The count-min sketch is a time- and memory-efficient randomized data structure that provides a point estimate of the number of times an item has appeared in a data stream.
no code implementations • 5 Mar 2018 • Michael Mitzenmacher
We provide a simple method for improving the performance of the recently introduced learned Bloom filters, by showing that they perform better when the learned function is sandwiched between two Bloom filters.
Data Structures and Algorithms
no code implementations • 3 Feb 2018 • Michael Mitzenmacher
Recent work has suggested enhancing Bloom filters by using a pre-filter, based on applying machine learning to model the data set the Bloom filter is meant to represent.
Data Structures and Algorithms
1 code implementation • 19 Sep 2017 • Charalampos E. Tsourakakis, Michael Mitzenmacher, Kasper Green Larsen, Jarosław Błasiok, Ben Lawson, Preetum Nakkiran, Vasileios Nakos
The {\em edge sign prediction problem} aims to predict whether an interaction between a pair of nodes will be positive or negative.
no code implementations • 3 Sep 2017 • Max Hopkins, Michael Mitzenmacher, Sebastian Wagner-Carena
JPEG is one of the most widely used image formats, but in some ways remains surprisingly unoptimized, perhaps because some natural optimizations would go outside the standard that defines JPEG.
no code implementations • NeurIPS 2016 • Ping Li, Michael Mitzenmacher, Martin Slawski
Random projections constitute a simple, yet effective technique for dimensionality reduction with applications in learning and search problems.
1 code implementation • 30 May 2016 • David Eppstein, Michael T. Goodrich, Jenny Lam, Nil Mamano, Michael Mitzenmacher, Manuel Torres
We introduce models and algorithmic foundations for graph watermarking.
Multimedia Data Structures and Algorithms
no code implementations • 21 Feb 2016 • Ping Li, Michael Mitzenmacher, Anshumali Shrivastava
In this paper, we focus on a simple 2-bit coding scheme.
no code implementations • 21 Aug 2014 • Yakir A. Reshef, David N. Reshef, Pardis C. Sabeti, Michael Mitzenmacher
Introducing MIC_* also enables us to reason about the properties of MIC more abstractly: for instance, we show that MIC_* is continuous and that there is a sense in which it is a canonical "smoothing" of mutual information.
no code implementations • 31 Mar 2014 • Ping Li, Michael Mitzenmacher, Anshumali Shrivastava
This technical note compares two coding (quantization) schemes for random projections in the context of sub-linear time approximate near neighbor search.
no code implementations • 9 Aug 2013 • Ping Li, Michael Mitzenmacher, Anshumali Shrivastava
The method of random projections has become very popular for large-scale applications in statistical learning, information retrieval, bio-informatics and other applications.
no code implementations • 27 Jan 2013 • David Reshef, Yakir Reshef, Michael Mitzenmacher, Pardis Sabeti
A measure of dependence is said to be equitable if it gives similar scores to equally noisy relationships of different types.
3 code implementations • 12 Jan 2011 • Michael T. Goodrich, Michael Mitzenmacher
We present a version of the Bloom filter data structure that supports not only the insertion, deletion, and lookup of key-value pairs, but also allows a complete listing of its contents with high probability, as long the number of key-value pairs is below a designed threshold.
Data Structures and Algorithms Databases