Search Results for author: Ben Packer

Found 10 papers, 2 papers with code

FRAPPÉ: A Group Fairness Framework for Post-Processing Everything

1 code implementation5 Dec 2023 Alexandru Ţifrea, Preethi Lahoti, Ben Packer, Yoni Halpern, Ahmad Beirami, Flavien Prost

Despite achieving promising fairness-error trade-offs, in-processing mitigation techniques for group fairness cannot be employed in numerous practical applications with limited computation resources or no access to the training pipeline of the prediction model.

Fairness

Causally motivated Shortcut Removal Using Auxiliary Labels

1 code implementation13 May 2021 Maggie Makar, Ben Packer, Dan Moldovan, Davis Blalock, Yoni Halpern, Alexander D'Amour

Shortcut learning, in which models make use of easy-to-represent but unstable associations, is a major failure mode for robust machine learning.

Causal Inference Disentanglement +1

CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation

no code implementations EMNLP 2020 Tianlu Wang, Xuezhi Wang, Yao Qin, Ben Packer, Kang Li, Jilin Chen, Alex Beutel, Ed Chi

Experiments on real-world NLP datasets demonstrate that our method can generate more diverse and fluent adversarial texts, compared to many existing adversarial text generation approaches.

Adversarial Text Attribute +3

Can We Improve Model Robustness through Secondary Attribute Counterfactuals?

no code implementations EMNLP 2021 Ananth Balashankar, Xuezhi Wang, Ben Packer, Nithum Thain, Ed Chi, Alex Beutel

By implementing RDI in the context of toxicity detection, we find that accounting for secondary attributes can significantly improve robustness, with improvements in sliced accuracy on the original dataset up to 7% compared to existing robustness methods.

Attribute coreference-resolution +3

Flexible text generation for counterfactual fairness probing

no code implementations NAACL (WOAH) 2022 Zee Fryer, Vera Axelrod, Ben Packer, Alex Beutel, Jilin Chen, Kellie Webster

A common approach for testing fairness issues in text-based classifiers is through the use of counterfactuals: does the classifier output change if a sensitive attribute in the input is changed?

Attribute counterfactual +2

Simpson's Paradox in Recommender Fairness: Reconciling differences between per-user and aggregated evaluations

no code implementations14 Oct 2022 Flavien Prost, Ben Packer, Jilin Chen, Li Wei, Pierre Kremp, Nicholas Blumm, Susan Wang, Tulsee Doshi, Tonia Osadebe, Lukasz Heldt, Ed H. Chi, Alex Beutel

We reconcile these notions and show that the tension is due to differences in distributions of users where items are relevant, and break down the important factors of the user's recommendations.

Fairness Recommendation Systems

Striving for data-model efficiency: Identifying data externalities on group performance

no code implementations11 Nov 2022 Esther Rolf, Ben Packer, Alex Beutel, Fernando Diaz

Building trustworthy, effective, and responsible machine learning systems hinges on understanding how differences in training data and modeling decisions interact to impact predictive performance.

Improving Classifier Robustness through Active Generation of Pairwise Counterfactuals

no code implementations22 May 2023 Ananth Balashankar, Xuezhi Wang, Yao Qin, Ben Packer, Nithum Thain, Jilin Chen, Ed H. Chi, Alex Beutel

We demonstrate that with a small amount of human-annotated counterfactual data (10%), we can generate a counterfactual augmentation dataset with learned labels, that provides an 18-20% improvement in robustness and a 14-21% reduction in errors on 6 out-of-domain datasets, comparable to that of a fully human-annotated counterfactual dataset for both sentiment classification and question paraphrase tasks.

counterfactual Data Augmentation +2

Towards A Scalable Solution for Improving Multi-Group Fairness in Compositional Classification

no code implementations11 Jul 2023 James Atwood, Tina Tian, Ben Packer, Meghana Deodhar, Jilin Chen, Alex Beutel, Flavien Prost, Ahmad Beirami

Despite the rich literature on machine learning fairness, relatively little attention has been paid to remediating complex systems, where the final prediction is the combination of multiple classifiers and where multiple groups are present.

Fairness

Improving Diversity of Demographic Representation in Large Language Models via Collective-Critiques and Self-Voting

no code implementations25 Oct 2023 Preethi Lahoti, Nicholas Blumm, Xiao Ma, Raghavendra Kotikalapudi, Sahitya Potluri, Qijun Tan, Hansa Srinivasan, Ben Packer, Ahmad Beirami, Alex Beutel, Jilin Chen

A crucial challenge for generative large language models (LLMs) is diversity: when a user's prompt is under-specified, models may follow implicit assumptions while generating a response, which may result in homogenization of the responses, as well as certain demographic groups being under-represented or even erased from the generated responses.

Cannot find the paper you are looking for? You can Submit a new open access paper.