no code implementations • 1 May 2024 • Sadegh Farhadkhani, Rachid Guerraoui, Nirupam Gupta, Rafael Pinot
It has been argued that the seemingly weaker threat model where only workers' local datasets get poisoned is more reasonable.
no code implementations • 20 Feb 2024 • Youssef Allouah, Sadegh Farhadkhani, Rachid Guerraoui, Nirupam Gupta, Rafael Pinot, Geovani Rizk, Sasha Voitovych
The natural approach to robustify FL against adversarial clients is to replace the simple averaging operation at the server in the standard $\mathsf{FedAvg}$ algorithm by a \emph{robust averaging rule}.
no code implementations • 9 Feb 2023 • Youssef Allouah, Rachid Guerraoui, Nirupam Gupta, Rafael Pinot, John Stephan
The latter amortizes the dependence on the dimension in the error (caused by adversarial workers and DP), while being agnostic to the statistical properties of the data.
no code implementations • 3 Feb 2023 • Youssef Allouah, Sadegh Farhadkhani, Rachid Guerraoui, Nirupam Gupta, Rafael Pinot, John Stephan
Byzantine machine learning (ML) aims to ensure the resilience of distributed learning algorithms to misbehaving (or Byzantine) machines.
no code implementations • 16 Nov 2022 • Shuo Liu, Nirupam Gupta, Nitin H. Vaidya
In particular, we introduce the notion of $(f, r; \epsilon)$-resilience to characterize how well the true solution is approximated in the presence of up to $f$ Byzantine faulty agents, and up to $r$ slow agents (or stragglers) -- smaller $\epsilon$ represents a better approximation.
no code implementations • 30 Sep 2022 • El-Mahdi El-Mhamdi, Sadegh Farhadkhani, Rachid Guerraoui, Nirupam Gupta, Lê-Nguyên Hoang, Rafael Pinot, Sébastien Rouault, John Stephan
Large AI Models (LAIMs), of which large language models are the most prominent recent example, showcase some impressive performance.
1 code implementation • 22 Sep 2022 • Sadegh Farhadkhani, Rachid Guerraoui, Nirupam Gupta, Lê Nguyên Hoang, Rafael Pinot, John Stephan
We present MoNNA, a new algorithm that (a) is provably robust under standard assumptions and (b) has a gradient computation overhead that is linear in the fraction of faulty machines, which is conjectured to be tight.
no code implementations • 24 May 2022 • Sadegh Farhadkhani, Rachid Guerraoui, Nirupam Gupta, Rafael Pinot, John Stephan
We present \emph{RESAM (RESilient Averaging of Momentums)}, a unified framework that makes it simple to establish optimal Byzantine resilience, relying only on standard machine learning assumptions.
no code implementations • 21 Oct 2021 • Shuo Liu, Nirupam Gupta, Nitin Vaidya
We demonstrate, both theoretically and empirically, the merits of our proposed redundancy model in improving the robustness of DGD against asynchronous and Byzantine agents, and their extensions to distributed stochastic gradient descent (D-SGD) for robust distributed machine learning with asynchronous and Byzantine agents.
no code implementations • 8 Oct 2021 • Rachid Guerraoui, Nirupam Gupta, Rafael Pinot, Sebastien Rouault, John Stephan
Privacy and Byzantine resilience (BR) are two crucial requirements of modern-day distributed machine learning.
no code implementations • 26 Aug 2021 • Nirupam Gupta, Thinh T. Doan, Nitin Vaidya
However, we do not know of any such techniques for the federated local SGD algorithm - a more commonly used method for federated machine learning.
no code implementations • 19 Aug 2021 • Kushal Chakrabarti, Nirupam Gupta, Nikhil Chopra
The system comprises multiple agents in this problem, each with a set of local data points and an associated local cost function.
1 code implementation • 16 Feb 2021 • Rachid Guerraoui, Nirupam Gupta, Rafaël Pinot, Sébastien Rouault, John Stephan
This paper addresses the problem of combining Byzantine resilience with privacy in machine learning (ML).
no code implementations • 28 Jan 2021 • Nirupam Gupta, Nitin H. Vaidya
We consider the problem of Byzantine fault-tolerance in the peer-to-peer (P2P) distributed gradient-descent method -- a prominent algorithm for distributed optimization in a P2P system.
Distributed Optimization Distributed, Parallel, and Cluster Computing
no code implementations • 26 Jan 2021 • Kushal Chakrabarti, Nirupam Gupta, Nikhil Chopra
This paper considers the problem of multi-agent distributed linear regression in the presence of system noises.
no code implementations • 15 Nov 2020 • Kushal Chakrabarti, Nirupam Gupta, Nikhil Chopra
The recently proposed Iteratively Pre-conditioned Gradient-descent (IPG) method has been shown to converge faster than other existing distributed algorithms that solve this problem.
no code implementations • 11 Aug 2020 • Nirupam Gupta, Shuo Liu, Nitin H. Vaidya
We show that the CGE gradient-filter guarantees fault-tolerance against a bounded fraction of Byzantine agents under standard stochastic assumptions, and is computationally simpler compared to many existing gradient-filters such as multi-KRUM, geometric median-of-means, and the spectral filters.
no code implementations • 6 Aug 2020 • Kushal Chakrabarti, Nirupam Gupta, Nikhil Chopra
In this problem, the system comprises multiple agents, each having a set of local data points, that are connected to a server.
no code implementations • 13 Mar 2020 • Kushal Chakrabarti, Nirupam Gupta, Nikhil Chopra
In this problem, there are multiple agents in the system, and each agent only knows its local cost function.
no code implementations • 19 Dec 2019 • Nirupam Gupta, Nitin H. Vaidya
The coding schemes use the concept of reactive redundancy for isolating Byzantine workers that eventually send faulty information.
no code implementations • 20 Mar 2019 • Nirupam Gupta, Nitin H. Vaidya
This paper considers the problem of Byzantine fault tolerance in distributed linear regression in a multi-agent system.