no code implementations • 19 Apr 2024 • Jessica Dai, Eve Fleisig
We analyze the problem settings of social choice and RLHF, identify key differences between them, and discuss how these differences may affect the RLHF interpretation of well-known technical results in social choice.
no code implementations • 10 Jan 2024 • Jessica Dai, Bailey Flanigan, Nika Haghtalab, Meena Jagadeesan, Chara Podimata
A common explanation for negative user impacts of content recommender systems is misalignment between the platform's objective and user welfare.
no code implementations • 15 May 2022 • Jessica Dai, Sohini Upadhyay, Ulrich Aivodji, Stephen H. Bach, Himabindu Lakkaraju
We then leverage these properties to propose a novel evaluation framework which can quantitatively measure disparities in the quality of explanations output by state-of-the-art methods.
no code implementations • 14 Mar 2022 • Kweku Kwegyir-Aggrey, A. Feder Cooper, Jessica Dai, John Dickerson, Keegan Hines, Suresh Venkatasubramanian
We study the problem of post-processing a supervised machine-learned regressor to maximize fair binary classification at all decision thresholds.
no code implementations • 24 Jun 2021 • Jessica Dai, Sohini Upadhyay, Stephen H. Bach, Himabindu Lakkaraju
In situations where explanations of black-box models may be useful, the fairness of the black-box is also often a relevant concern.
no code implementations • 7 Nov 2020 • Jessica Dai, Sina Fazelpour, Zachary C. Lipton
If k% of employers were to voluntarily adopt a fairness-promoting intervention, should we expect k% progress (in aggregate) towards the benefits of universal adoption, or will the dynamics of partial compliance wash out the hoped-for benefits?