no code implementations • 18 Jul 2023 • Sumegha Garg, Christopher Jung, Omer Reingold, Aaron Roth
We develop a new online multicalibration algorithm that is well defined for infinite benchmark classes $F$, and is oracle efficient (i. e. for any class $F$, the algorithm has the form of an efficient reduction to a no-regret learning algorithm for $F$).
no code implementations • 5 Jul 2021 • Sumegha Garg, Pravesh K. Kothari, Pengda Liu, Ran Raz
We show that any learning algorithm for the learning problem corresponding to $M$, with error, requires either a memory of size at least $\Omega\left(\frac{k \cdot \ell}{\varepsilon} \right)$, or at least $2^{\Omega(r)}$ samples.
no code implementations • 17 May 2020 • Mark Braverman, Sumegha Garg
Showing that if the objective is to maximize the efficiency of the classification process (defined as the accuracy of the outcome minus the sunk cost of the qualified players manipulating their features to gain a better outcome), then using randomized classifiers (that is, ones where the probability of a given feature vector to be accepted by the classifier is strictly between 0 and 1) is necessary.
no code implementations • 22 Apr 2019 • Sumegha Garg, Michael P. Kim, Omer Reingold
As algorithmic prediction systems have become widespread, fears that these systems may inadvertently discriminate against members of underrepresented populations have grown.
no code implementations • 8 Aug 2017 • Sumegha Garg, Ran Raz, Avishay Tal
We show that any learning algorithm for the learning problem corresponding to $M$ requires either a memory of size at least $\Omega\left(k \cdot \ell \right)$, or at least $2^{\Omega(r)}$ samples.