no code implementations • 6 Jul 2024 • Nina Narodytska, Shay Vargaftik
Large Language Models (LLMs) have made significant progress in assisting users to query databases in natural language.
no code implementations • 1 Jul 2024 • Wenchen Han, Shay Vargaftik, Michael Mitzenmacher, Brad Karp, Ran Ben Basat
Gradient aggregation has long been identified as a major bottleneck in today's large-scale distributed machine learning training systems.
no code implementations • 5 Feb 2024 • Ran Ben-Basat, Yaniv Ben-Itzhak, Michael Mitzenmacher, Shay Vargaftik
Quantization is a fundamental optimization for many machine-learning use cases, including compressing gradients, model weights and activations, and datasets.
1 code implementation • 16 Feb 2023 • Minghao Li, Ran Ben Basat, Shay Vargaftik, ChonLam Lao, Kevin Xu, Michael Mitzenmacher, Minlan Yu
To address this bottleneck and accelerate training, a widely-deployed approach is compression.
no code implementations • 1 Feb 2023 • Ron Dorfman, Shay Vargaftik, Yaniv Ben-Itzhak, Kfir Y. Levy
Many compression techniques have been proposed to reduce the communication overhead of Federated Learning training procedures.
no code implementations • 13 Oct 2022 • Yaniv Ben-Itzhak, Helen Möllering, Benny Pinkas, Thomas Schneider, Ajith Suresh, Oleksandr Tkachenko, Shay Vargaftik, Christian Weinert, Hossein Yalame, Avishay Yanai
In this paper, we unite both research directions by introducing ScionFL, the first secure aggregation framework for FL that operates efficiently on quantized inputs and simultaneously provides robustness against malicious clients.
no code implementations • 26 May 2022 • Ran Ben Basat, Shay Vargaftik, Amit Portnoy, Gil Einziger, Yaniv Ben-Itzhak, Michael Mitzenmacher
Distributed Mean Estimation (DME), in which $n$ clients communicate vectors to a parameter server that estimates their average, is a fundamental building block in communication-efficient federated learning.
1 code implementation • 18 May 2022 • Changgang Zheng, Mingyuan Zang, Xinpeng Hong, Riyad Bensoussane, Shay Vargaftik, Yaniv Ben-Itzhak, Noa Zilberman
To date, no general solution has been provided for mapping machine learning algorithms to programmable network devices.
no code implementations • 17 May 2022 • Changgang Zheng, Zhaoqi Xiong, Thanh T Bui, Siim Kaupmees, Riyad Bensoussane, Antoine Bernabeu, Shay Vargaftik, Yaniv Ben-Itzhak, Noa Zilberman
In this paper, we introduce IIsy, implementing machine learning classification models in a hybrid fashion using off-the-shelf network devices.
1 code implementation • 19 Aug 2021 • Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, Michael Mitzenmacher
Distributed Mean Estimation (DME) is a central building block in federated learning, where clients send local gradients to a parameter server for averaging and updating the model.
1 code implementation • NeurIPS 2021 • Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, Michael Mitzenmacher
We consider the problem where $n$ clients transmit $d$-dimensional real-valued vectors using $d(1+o(1))$ bits each, in a manner that allows the receiver to approximately reconstruct their mean.
no code implementations • 5 Oct 2020 • Ran Ben-Basat, Michael Mitzenmacher, Shay Vargaftik
We consider both the biased and unbiased estimation problems and aim to minimize the cost.
no code implementations • 26 Sep 2019 • Shay Vargaftik, Isaac Keslassy, Ariel Orda, Yaniv Ben-Itzhak
Decision-tree-based ensemble classification methods (DTEMs) are a prevalent tool for supervised anomaly detection.
2 code implementations • 23 Dec 2018 • Gal Mendelson, Shay Vargaftik, Katherine Barabash, Dean Lorenz, Isaac Keslassy, Ariel Orda
Consistent hashing (CH) is a central building block in many networking applications, from datacenter load-balancing to distributed storage.
Data Structures and Algorithms