no code implementations • 11 Apr 2024 • Ian Tenney, Ryan Mullins, Bin Du, Shree Pandya, Minsuk Kahng, Lucas Dixon
We present Sequence Salience, a visual tool for interactive prompt debugging with input salience methods.
2 code implementations • 1 Apr 2024 • Seongmin Lee, Zijie J. Wang, Aishwarya Chakravarthy, Alec Helbling, Shengyun Peng, Mansi Phute, Duen Horng Chau, Minsuk Kahng
Our library offers a new way to quickly attribute an LLM's text generation to training data points to inspect model behaviors, enhance its trustworthiness, and compare model-generated text with user-provided text.
1 code implementation • 21 Feb 2024 • Emily Reif, Crystal Qian, James Wexler, Minsuk Kahng
Making sense of unstructured text datasets is perennially difficult, yet increasingly relevant with Large Language Models.
no code implementations • 21 Feb 2024 • Crystal Qian, Emily Reif, Minsuk Kahng
We find that although data quality is a top priority, there is little consensus around what data quality is and how to evaluate it.
no code implementations • 16 Feb 2024 • Minsuk Kahng, Ian Tenney, Mahima Pushkarna, Michael Xieyang Liu, James Wexler, Emily Reif, Krystal Kallarackal, Minsuk Chang, Michael Terry, Lucas Dixon
Automatic side-by-side evaluation has emerged as a promising approach to evaluating the quality of responses from large language models (LLMs).
1 code implementation • 14 Feb 2024 • Jessica Quaye, Alicia Parrish, Oana Inel, Charvi Rastogi, Hannah Rose Kirk, Minsuk Kahng, Erin Van Liemt, Max Bartolo, Jess Tsang, Justin White, Nathan Clement, Rafael Mosquera, Juan Ciro, Vijay Janapa Reddi, Lora Aroyo
By focusing on ``implicitly adversarial'' prompts (those that trigger T2I models to generate unsafe images for non-obvious reasons), we isolate a set of difficult safety issues that human creativity is well-suited to uncover.
1 code implementation • ICCV 2023 • Eric Slyman, Minsuk Kahng, Stefan Lee
Recent work in vision-and-language demonstrates that large-scale pretraining can learn generalizable models that are efficiently transferable to downstream tasks.
1 code implementation • 19 May 2023 • Emily Reif, Minsuk Kahng, Savvas Petridis
We present LinguisticLens, a novel inter-active visualization tool for making sense of and analyzing syntactic diversity of LLM-generated datasets.
no code implementations • 4 Jun 2022 • Kin-Ho Lam, Delyar Tabatabai, Jed Irvine, Donald Bertucci, Anita Ruangrotsakun, Minsuk Kahng, Alan Fern
Reinforcement learning (RL) agents are commonly evaluated via their expected value over a distribution of test scenarios.
1 code implementation • 14 May 2022 • Donald Bertucci, Md Montaser Hamid, Yashwanthi Anand, Anita Ruangrotsakun, Delyar Tabatabai, Melissa Perez, Minsuk Kahng
In this paper, we present DendroMap, a novel approach to interactively exploring large-scale image datasets for machine learning (ML).
no code implementations • 28 Sep 2021 • Kin-Ho Lam, Zhengxian Lin, Jed Irvine, Jonathan Dodge, Zeyad T Shureih, Roli Khanna, Minsuk Kahng, Alan Fern
We describe a user interface and case study, where a small group of AI experts and developers attempt to identify reasoning flaws due to inaccurate agent learning.
no code implementations • 13 Sep 2021 • Li Fuxin, Zhongang Qi, Saeed Khorram, Vivswan Shitole, Prasad Tadepalli, Minsuk Kahng, Alan Fern
This paper summarizes our endeavors in the past few years in terms of explaining image classifiers, with the aim of including negative results and insights we have gained.
no code implementations • 18 Aug 2021 • Matthew L. Olson, Thuy-Vy Nguyen, Gaurav Dixit, Neale Ratzlaff, Weng-Keen Wong, Minsuk Kahng
Identifying covariate shift is crucial for making machine learning systems robust in the real world and for detecting training data biases that are not reflected in test data.
1 code implementation • NeurIPS 2021 • Vivswan Shitole, Li Fuxin, Minsuk Kahng, Prasad Tadepalli, Alan Fern
Attention maps are a popular way of explaining the decisions of convolutional networks for image classification.
5 code implementations • 30 Apr 2020 • Zijie J. Wang, Robert Turko, Omar Shaikh, Haekyu Park, Nilaksh Das, Fred Hohman, Minsuk Kahng, Duen Horng Chau
Deep learning's great success motivates many practitioners and students to learn about this exciting technology.
no code implementations • 7 Jan 2020 • Zijie J. Wang, Robert Turko, Omar Shaikh, Haekyu Park, Nilaksh Das, Fred Hohman, Minsuk Kahng, Duen Horng Chau
The success of deep learning solving previously-thought hard problems has inspired many non-experts to learn and understand this exciting technology.
1 code implementation • 10 Apr 2019 • Ángel Alexander Cabrera, Will Epperson, Fred Hohman, Minsuk Kahng, Jamie Morgenstern, Duen Horng Chau
We present FairVis, a mixed-initiative visual analytics system that integrates a novel subgroup discovery technique for users to audit the fairness of machine learning models.
1 code implementation • 5 Sep 2018 • Minsuk Kahng, Nikhil Thorat, Duen Horng Chau, Fernanda Viégas, Martin Wattenberg
Recent success in deep learning has generated immense interest among practitioners and students, inspiring many to learn about this new technology.
no code implementations • 21 Jan 2018 • Fred Hohman, Minsuk Kahng, Robert Pienta, Duen Horng Chau
We present a survey of the role of visual analytics in deep learning research, which highlights its short yet impactful history and thoroughly summarizes the state-of-the-art using a human-centered interrogative framework, focusing on the Five W's and How (Why, Who, What, How, When, and Where).
no code implementations • 6 Apr 2017 • Minsuk Kahng, Pierre Y. Andrews, Aditya Kalro, Duen Horng Chau
While deep learning models have achieved state-of-the-art accuracies for many prediction tasks, understanding these models remains a challenge.