no code implementations • ACL 2022 • Marco Tulio Ribeiro, Scott Lundberg
Current approaches to testing and debugging NLP models rely on highly variable human creativity and extensive labor, or only work for a very restrictive class of bugs.
no code implementations • 28 May 2023 • Zexue He, Marco Tulio Ribeiro, Fereshte Khani
Even when aggregate accuracy is high, state-of-the-art NLP models often fail systematically on specific subgroups of data, resulting in unfair outcomes and eroding user trust.
no code implementations • 20 May 2023 • Fereshte Khani, Marco Tulio Ribeiro
Our main insight is learning a \emph{local} model for each concept, and a \emph{global} model to integrate the original data with all concepts.
no code implementations • 19 Apr 2023 • Charvi Rastogi, Marco Tulio Ribeiro, Nicholas King, Harsha Nori, Saleema Amershi
Through the design process we highlight the importance of sensemaking and human-AI communication to leverage complementary strengths of humans and generative models in collaborative auditing.
2 code implementations • 22 Mar 2023 • Sébastien Bubeck, Varun Chandrasekaran, Ronen Eldan, Johannes Gehrke, Eric Horvitz, Ece Kamar, Peter Lee, Yin Tat Lee, Yuanzhi Li, Scott Lundberg, Harsha Nori, Hamid Palangi, Marco Tulio Ribeiro, Yi Zhang
We contend that (this early version of) GPT-4 is part of a new cohort of LLMs (along with ChatGPT and Google's PaLM for example) that exhibit more general intelligence than previous AI models.
Ranked #33 on Arithmetic Reasoning on GSM8K
2 code implementations • 16 Mar 2023 • Bhargavi Paranjape, Scott Lundberg, Sameer Singh, Hannaneh Hajishirzi, Luke Zettlemoyer, Marco Tulio Ribeiro
We introduce Automatic Reasoning and Tool-use (ART), a framework that uses frozen LLMs to automatically generate intermediate reasoning steps as a program.
1 code implementation • 14 Feb 2023 • Tongshuang Wu, Hua Shen, Daniel S. Weld, Jeffrey Heer, Marco Tulio Ribeiro
ScatterShot iteratively slices unlabeled data into task-specific patterns, samples informative inputs from underexplored or not-yet-saturated slices in an active learning manner, and helps users label more efficiently with the help of an LLM and the current example set.
6 code implementations • 8 Dec 2022 • Gabriel Ilharco, Marco Tulio Ribeiro, Mitchell Wortsman, Suchin Gururangan, Ludwig Schmidt, Hannaneh Hajishirzi, Ali Farhadi
Changing how pre-trained models behave -- e. g., improving their performance on a downstream task or mitigating biases learned during pre-training -- is a common practice when developing machine learning systems.
1 code implementation • ICCV 2023 • Irena Gao, Gabriel Ilharco, Scott Lundberg, Marco Tulio Ribeiro
Vision models often fail systematically on groups of data that share common semantic characteristics (e. g., rare objects or unusual scenes), but identifying these failure modes is a challenge.
1 code implementation • 7 Nov 2022 • Shikhar Murty, Christopher D. Manning, Scott Lundberg, Marco Tulio Ribeiro
Current approaches for fixing systematic problems in NLP models (e. g. regex patches, finetuning on more data) are either brittle, or labor-intensive and liable to shortcuts.
1 code implementation • NAACL 2022 • Yilun Zhou, Marco Tulio Ribeiro, Julie Shah
Interpretability methods are developed to understand the working mechanisms of black-box models, which is crucial to their responsible deployment.
no code implementations • 3 Jun 2021 • Gregory Plumb, Marco Tulio Ribeiro, Ameet Talwalkar
Image classifiers often use spurious patterns, such as "relying on the presence of a person to detect a tennis racket, which do not generalize.
1 code implementation • 27 Apr 2021 • Yilun Zhou, Serena Booth, Marco Tulio Ribeiro, Julie Shah
Feature attribution methods are exceedingly popular in interpretable machine learning.
1 code implementation • ACL 2021 • Tongshuang Wu, Marco Tulio Ribeiro, Jeffrey Heer, Daniel S. Weld
While counterfactual examples are useful for analysis and training of NLP models, current generation methods either rely on manual labor to create very few counterfactuals, or only instantiate limited types of perturbations such as paraphrases or word substitutions.
no code implementations • 26 Jun 2020 • Gagan Bansal, Tongshuang Wu, Joyce Zhou, Raymond Fok, Besmira Nushi, Ece Kamar, Marco Tulio Ribeiro, Daniel S. Weld
However, prior studies observed improvements from explanations only when the AI, alone, outperformed both the human and the best team.
4 code implementations • ACL 2020 • Marco Tulio Ribeiro, Tongshuang Wu, Carlos Guestrin, Sameer Singh
Although measuring held-out accuracy has been the primary approach to evaluate generalization, it often overestimates the performance of NLP models, while alternative approaches for evaluating models either focus on individual tasks or on specific behaviors.
1 code implementation • ACL 2019 • Marco Tulio Ribeiro, Carlos Guestrin, Sameer Singh
Although current evaluation of question-answering systems treats predictions in isolation, we need to consider the relationship between predictions to measure true understanding.
1 code implementation • ACL 2019 • Tongshuang Wu, Marco Tulio Ribeiro, Jeffrey Heer, Daniel Weld
Though error analysis is crucial to understanding and improving NLP models, the common practice of manual, subjective categorization of a small sample of errors can yield biased and incomplete conclusions.
1 code implementation • ACL 2018 • Marco Tulio Ribeiro, Sameer Singh, Carlos Guestrin
Complex machine learning models for NLP are often brittle, making different predictions for input instances that are extremely similar semantically.
no code implementations • 22 Nov 2016 • Sameer Singh, Marco Tulio Ribeiro, Carlos Guestrin
Recent work in model-agnostic explanations of black-box machine learning has demonstrated that interpretability of complex models does not have to come at the cost of accuracy or model flexibility.
no code implementations • 17 Nov 2016 • Marco Tulio Ribeiro, Sameer Singh, Carlos Guestrin
At the core of interpretable machine learning is the question of whether humans are able to make accurate predictions about a model's behavior.
no code implementations • 16 Jun 2016 • Marco Tulio Ribeiro, Sameer Singh, Carlos Guestrin
Understanding why machine learning models behave the way they do empowers both system designers and end-users in many ways: in model selection, feature engineering, in order to trust and act upon the predictions, and in more intuitive user interfaces.
27 code implementations • 16 Feb 2016 • Marco Tulio Ribeiro, Sameer Singh, Carlos Guestrin
Despite widespread adoption, machine learning models remain mostly black boxes.