Search Results for author: Michael Mitzenmacher

Found 37 papers, 9 papers with code

Don't Stop Me Now: Embedding Based Scheduling for LLMs

no code implementations1 Oct 2024 Rana Shahout, Eran Malach, Chunwei Liu, Weifan Jiang, Minlan Yu, Michael Mitzenmacher

Using these predictions, we propose a prediction-based SRPT variant with limited preemption designed to account for memory overhead in LLM systems.

Blocking Large Language Model +1

Learning-Augmented Frequency Estimation in Sliding Windows

no code implementations17 Sep 2024 Rana Shahout, Ibrahim Sabek, Michael Mitzenmacher

We show how to utilize machine learning approaches to improve sliding window algorithms for approximate frequency estimation problems, under the ``algorithms with predictions'' framework.

Beyond Throughput and Compression Ratios: Towards High End-to-end Utility of Gradient Compression

no code implementations1 Jul 2024 Wenchen Han, Shay Vargaftik, Michael Mitzenmacher, Brad Karp, Ran Ben Basat

Gradient aggregation has long been identified as a major bottleneck in today's large-scale distributed machine learning training systems.

Quantization

Learning-Based Heavy Hitters and Flow Frequency Estimation in Streams

no code implementations24 Jun 2024 Rana Shahout, Michael Mitzenmacher

In this paper, we present the first learned competing-counter-based algorithm, called LSS, for identifying heavy hitters, top k, and flow frequency estimation that utilizes the well-known Space Saving algorithm.

SkipPredict: When to Invest in Predictions for Scheduling

no code implementations5 Feb 2024 Rana Shahout, Michael Mitzenmacher

SkipPredict prioritizes predicted short jobs over long jobs, and for the latter, SkipPredict applies a second round of more detailed "expensive predictions" to approximate Shortest Remaining Processing Time for these jobs.

Scheduling

Optimal and Near-Optimal Adaptive Vector Quantization

no code implementations5 Feb 2024 Ran Ben-Basat, Yaniv Ben-Itzhak, Michael Mitzenmacher, Shay Vargaftik

Quantization is a fundamental optimization for many machine-learning use cases, including compressing gradients, model weights and activations, and datasets.

Quantization

Proteus: A Self-Designing Range Filter

2 code implementations30 Jun 2022 Eric R. Knorr, Baptiste Lemaire, Andrew Lim, Siqiang Luo, Huanchen Zhang, Stratos Idreos, Michael Mitzenmacher

We introduce Proteus, a novel self-designing approximate range filter, which configures itself based on sampled data in order to optimize its false positive rate (FPR) for a given space requirement.

QUIC-FL: Quick Unbiased Compression for Federated Learning

no code implementations26 May 2022 Ran Ben Basat, Shay Vargaftik, Amit Portnoy, Gil Einziger, Yaniv Ben-Itzhak, Michael Mitzenmacher

Distributed Mean Estimation (DME), in which $n$ clients communicate vectors to a parameter server that estimates their average, is a fundamental building block in communication-efficient federated learning.

Federated Learning Quantization

Tabula: Efficiently Computing Nonlinear Activation Functions for Secure Neural Network Inference

no code implementations5 Mar 2022 Maximilian Lam, Michael Mitzenmacher, Vijay Janapa Reddi, Gu-Yeon Wei, David Brooks

This enables an online phase where securely computing the result of a nonlinear function requires just a single round of communication, with communication cost equal to twice the number of bits of the input to the nonlinear function.

Quantization

Tabula: Efficiently Computing Nonlinear Activation Functions for Private Neural Network Inference

1 code implementation29 Sep 2021 Max Lam, Michael Mitzenmacher, Vijay Janapa Reddi, Gu-Yeon Wei, David Brooks

Multiparty computation approaches to private neural network inference require significant communication between server and client, incur tremendous runtime penalties, and cost massive storage overheads.

EDEN: Communication-Efficient and Robust Distributed Mean Estimation for Federated Learning

1 code implementation19 Aug 2021 Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, Michael Mitzenmacher

Distributed Mean Estimation (DME) is a central building block in federated learning, where clients send local gradients to a parameter server for averaging and updating the model.

Federated Learning

DRIVE: One-bit Distributed Mean Estimation

1 code implementation NeurIPS 2021 Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, Michael Mitzenmacher

We consider the problem where $n$ clients transmit $d$-dimensional real-valued vectors using $d(1+o(1))$ bits each, in a manner that allows the receiver to approximately reconstruct their mean.

Federated Learning

Partitioned Learned Bloom Filters

no code implementations ICLR 2021 Kapil Vaidya, Eric Knorr, Michael Mitzenmacher, Tim Kraska

Bloom filters are space-efficient probabilistic data structures that are used to test whether an element is a member of a set, and may return false positives.

How to send a real number using a single bit (and some shared randomness)

no code implementations5 Oct 2020 Ran Ben-Basat, Michael Mitzenmacher, Shay Vargaftik

We consider both the biased and unbiased estimation problems and aim to minimize the cost.

Queues with Small Advice

no code implementations27 Jun 2020 Michael Mitzenmacher

Motivated by recent work on scheduling with predicted job sizes, we consider the performance of scheduling algorithms with minimal advice, namely a single bit.

Scheduling

Partitioned Learned Bloom Filter

no code implementations5 Jun 2020 Kapil Vaidya, Eric Knorr, Tim Kraska, Michael Mitzenmacher

Bloom filters are space-efficient probabilistic data structures that are used to test whether an element is a member of a set, and may return false positives.

Optimal Learning of Joint Alignments with a Faulty Oracle

no code implementations21 Sep 2019 Kasper Green Larsen, Michael Mitzenmacher, Charalampos E. Tsourakakis

The goal is to recover $n$ discrete variables $g_i \in \{0, \ldots, k-1\}$ (up to some global offset) given noisy observations of a set of their pairwise differences $\{(g_i - g_j) \bmod k\}$; specifically, with probability $\frac{1}{k}+\delta$ for some $\delta > 0$ one obtains the correct answer, and with the remaining probability one obtains a uniformly random incorrect answer.

The Supermarket Model with Known and Predicted Service Times

no code implementations23 May 2019 Michael Mitzenmacher, Matteo Dell'Amico

The supermarket model refers to a system with a large number of queues, where new customers choose d queues at random and join the one with the fewest customers.

A Model for Learned Bloom Filters, and Optimizing by Sandwiching

no code implementations3 Jan 2019 Michael Mitzenmacher

Recent work has suggested enhancing Bloom filters by using a pre-filter, based on applying machine learning to determine a function that models the data set the Bloom filter is meant to represent.

BIG-bench Machine Learning

A Model for Learned Bloom Filters and Optimizing by Sandwiching

no code implementations NeurIPS 2018 Michael Mitzenmacher

Recent work has suggested enhancing Bloom filters by using a pre-filter, based on applying machine learning to determine a function that models the data set the Bloom filter is meant to represent.

BIG-bench Machine Learning

A Bayesian Nonparametric View on Count-Min Sketch

no code implementations NeurIPS 2018 Diana Cai, Michael Mitzenmacher, Ryan P. Adams

The count-min sketch is a time- and memory-efficient randomized data structure that provides a point estimate of the number of times an item has appeared in a data stream.

Optimizing Learned Bloom Filters by Sandwiching

no code implementations5 Mar 2018 Michael Mitzenmacher

We provide a simple method for improving the performance of the recently introduced learned Bloom filters, by showing that they perform better when the learned function is sandwiched between two Bloom filters.

Data Structures and Algorithms

A Model for Learned Bloom Filters and Related Structures

no code implementations3 Feb 2018 Michael Mitzenmacher

Recent work has suggested enhancing Bloom filters by using a pre-filter, based on applying machine learning to model the data set the Bloom filter is meant to represent.

Data Structures and Algorithms

Predicting Positive and Negative Links with Noisy Queries: Theory & Practice

1 code implementation19 Sep 2017 Charalampos E. Tsourakakis, Michael Mitzenmacher, Kasper Green Larsen, Jarosław Błasiok, Ben Lawson, Preetum Nakkiran, Vasileios Nakos

The {\em edge sign prediction problem} aims to predict whether an interaction between a pair of nodes will be positive or negative.

Clustering

Simulated Annealing for JPEG Quantization

no code implementations3 Sep 2017 Max Hopkins, Michael Mitzenmacher, Sebastian Wagner-Carena

JPEG is one of the most widely used image formats, but in some ways remains surprisingly unoptimized, perhaps because some natural optimizations would go outside the standard that defines JPEG.

Quantization

Quantized Random Projections and Non-Linear Estimation of Cosine Similarity

no code implementations NeurIPS 2016 Ping Li, Michael Mitzenmacher, Martin Slawski

Random projections constitute a simple, yet effective technique for dimensionality reduction with applications in learning and search problems.

Dimensionality Reduction LEMMA +1

Models and Algorithms for Graph Watermarking

1 code implementation30 May 2016 David Eppstein, Michael T. Goodrich, Jenny Lam, Nil Mamano, Michael Mitzenmacher, Manuel Torres

We introduce models and algorithmic foundations for graph watermarking.

Multimedia Data Structures and Algorithms

Theoretical Foundations of Equitability and the Maximal Information Coefficient

no code implementations21 Aug 2014 Yakir A. Reshef, David N. Reshef, Pardis C. Sabeti, Michael Mitzenmacher

Introducing MIC_* also enables us to reason about the properties of MIC more abstractly: for instance, we show that MIC_* is continuous and that there is a sense in which it is a canonical "smoothing" of mutual information.

Coding for Random Projections and Approximate Near Neighbor Search

no code implementations31 Mar 2014 Ping Li, Michael Mitzenmacher, Anshumali Shrivastava

This technical note compares two coding (quantization) schemes for random projections in the context of sub-linear time approximate near neighbor search.

Quantization

Coding for Random Projections

no code implementations9 Aug 2013 Ping Li, Michael Mitzenmacher, Anshumali Shrivastava

The method of random projections has become very popular for large-scale applications in statistical learning, information retrieval, bio-informatics and other applications.

Information Retrieval Quantization +1

Equitability Analysis of the Maximal Information Coefficient, with Comparisons

no code implementations27 Jan 2013 David Reshef, Yakir Reshef, Michael Mitzenmacher, Pardis Sabeti

A measure of dependence is said to be equitable if it gives similar scores to equally noisy relationships of different types.

Mutual Information Estimation

Invertible Bloom Lookup Tables

3 code implementations12 Jan 2011 Michael T. Goodrich, Michael Mitzenmacher

We present a version of the Bloom filter data structure that supports not only the insertion, deletion, and lookup of key-value pairs, but also allows a complete listing of its contents with high probability, as long the number of key-value pairs is below a designed threshold.

Data Structures and Algorithms Databases

Cannot find the paper you are looking for? You can Submit a new open access paper.