no code implementations • 25 May 2023 • Yangsibo Huang, Haotian Jiang, Daogao Liu, Mohammad Mahdian, Jieming Mao, Vahab Mirrokni
In this paper, we study the setting in which data owners train machine learning models collaboratively under a privacy notion called joint differential privacy [Kearns et al., 2018].
1 code implementation • 31 Jan 2023 • Jacob Imola, Alessandro Epasto, Mohammad Mahdian, Vincent Cohen-Addad, Vahab Mirrokni
Then, we exhibit a polynomial-time approximation algorithm with $O(|V|^{2. 5}/ \epsilon)$-additive error, and an exponential-time algorithm that meets the lower bound.
no code implementations • 19 Jul 2022 • Mohammad Mahdian, Jieming Mao, Kangning Wang
In our model, the task is to pick the highest one out of $n$ values.
no code implementations • NeurIPS 2020 • Alessandro Epasto, Mohammad Mahdian, Jieming Mao, Vahab Mirrokni, Lijie Ren
But at the same time, more noise might need to be added to the algorithm in order to keep the algorithm differentially private and this might hurt the algorithm’s performance.
no code implementations • NeurIPS 2020 • Alessandro Epasto, Mohammad Mahdian, Vahab Mirrokni, Emmanouil Zampetakis
A soft-max function has two main efficiency measures: (1) approximation - which corresponds to how well it approximates the maximum function, (2) smoothness - which shows how sensitive it is to changes of its input.
no code implementations • 22 Oct 2020 • Alessandro Epasto, Mohammad Mahdian, Vahab Mirrokni, Manolis Zampetakis
A soft-max function has two main efficiency measures: (1) approximation - which corresponds to how well it approximates the maximum function, (2) smoothness - which shows how sensitive it is to changes of its input.
no code implementations • NeurIPS 2020 • Sara Ahmadian, Alessandro Epasto, Marina Knittel, Ravi Kumar, Mohammad Mahdian, Benjamin Moseley, Philip Pham, Sergei Vassilvitskii, Yuyan Wang
As machine learning has become more prevalent, researchers have begun to recognize the necessity of ensuring machine learning systems are fair.
1 code implementation • 6 Feb 2020 • Sara Ahmadian, Alessandro Epasto, Ravi Kumar, Mohammad Mahdian
We define a fairlet decomposition with cost similar to the $k$-median cost and this allows us to obtain approximation algorithms for a wide range of fairness constraints.
no code implementations • 29 May 2019 • Sara Ahmadian, Alessandro Epasto, Ravi Kumar, Mohammad Mahdian
In this paper we consider clustering problems in which each point is endowed with a color.
no code implementations • NeurIPS 2019 • Santiago Balseiro, Negin Golrezaei, Mohammad Mahdian, Vahab Mirrokni, Jon Schneider
We consider the variant of this problem where in addition to receiving the reward $r_{i, t}(c)$, the learner also learns the values of $r_{i, t}(c')$ for some other contexts $c'$ in set $\mathcal{O}_i(c)$; i. e., the rewards that would have been achieved by performing that action under different contexts $c'\in \mathcal{O}_i(c)$.
no code implementations • NeurIPS 2016 • Aris Anagnostopoulos, Jakub Łącki, Silvio Lattanzi, Stefano Leonardi, Mohammad Mahdian
In many of these applications, the input graph evolves over time in a continual and decentralized manner, and, to maintain a good clustering, the clustering algorithm needs to repeatedly probe the graph.