no code implementations • 8 Sep 2023 • Lydia T. Liu, Solon Barocas, Jon Kleinberg, Karen Levy
Through a simple model encompassing actions, latent states, and measurements, we demonstrate that pure outcome prediction rarely results in the most effective policy for taking actions, even when combined with other measurements.
no code implementations • 8 Sep 2022 • Lydia T. Liu, Serena Wang, Tolani Britton, Rediet Abebe
We find that a cross-disciplinary gap exists and is particularly salient in two parts of the ML life cycle: the formulation of an ML problem from education goals and the translation of predictions to interventions.
no code implementations • 16 Sep 2021 • Lydia T. Liu, Nikhil Garg, Christian Borgs
Strategic classification studies the design of a classifier robust to the manipulation of input by strategic individuals.
no code implementations • 14 Dec 2020 • Lydia T. Liu, Feng Ruan, Horia Mania, Michael I. Jordan
We study two-sided matching markets in which one side of the market (the players) does not have a priori knowledge about its preferences for the other side (the arms) and is required to learn its preferences from experience.
1 code implementation • ICML 2020 • Esther Rolf, Max Simchowitz, Sarah Dean, Lydia T. Liu, Daniel Björkegren, Moritz Hardt, Joshua Blumenstock
Our theoretical results characterize the optimal strategies in this class, bound the Pareto errors due to inaccuracies in the scores, and show an equivalence between optimal strategies and a rich class of fairness-constrained profit-maximizing policies.
no code implementations • 12 Jun 2019 • Lydia T. Liu, Horia Mania, Michael. I. Jordan
Stable matching, a classical model for two-sided markets, has long been studied with little consideration for how each side's preferences are learned.
1 code implementation • 20 Dec 2018 • Zhizhen Zhao, Lydia T. Liu, Amit Singer
The second is steerable PCA, a fast and accurate procedure for including all planar rotations for PCA.
no code implementations • 29 Aug 2018 • Lydia T. Liu, Max Simchowitz, Moritz Hardt
We show that under reasonable conditions, the deviation from satisfying group calibration is upper bounded by the excess risk of the learned score relative to the Bayes optimal score function.
no code implementations • NeurIPS 2018 • Chi Jin, Lydia T. Liu, Rong Ge, Michael. I. Jordan
Our objective is to find the $\epsilon$-approximate local minima of the underlying function $F$ while avoiding the shallow local minima---arising because of the tolerance $\nu$---which exist only in $f$.
3 code implementations • ICML 2018 • Lydia T. Liu, Sarah Dean, Esther Rolf, Max Simchowitz, Moritz Hardt
Fairness in machine learning has predominantly been studied in static classification settings without concern for how decisions change the underlying population over time.
1 code implementation • 17 Nov 2016 • Lydia T. Liu, Edgar Dobriban, Amit Singer
We develop $e$PCA (exponential family PCA), a new methodology for PCA on exponential family distributions.
Methodology