Search Results for author: Alex Mei

Found 7 papers, 3 papers with code

ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models

1 code implementation14 Oct 2023 Alex Mei, Sharon Levy, William Yang Wang

As large language models are integrated into society, robustness toward a suite of prompts is increasingly important to maintain reliability in a high-variance environment. Robustness evaluations must comprehensively encapsulate the various settings in which a user may invoke an intelligent system.

Let's Think Frame by Frame with VIP: A Video Infilling and Prediction Dataset for Evaluating Video Chain-of-Thought

1 code implementation23 May 2023 Vaishnavi Himakunthala, Andy Ouyang, Daniel Rose, Ryan He, Alex Mei, Yujie Lu, Chinmay Sonar, Michael Saxon, William Yang Wang

Despite exciting recent results showing vision-language systems' capacity to reason about images using natural language, their capacity for video reasoning remains under-explored.

Descriptive Video Prediction

Users are the North Star for AI Transparency

no code implementations9 Mar 2023 Alex Mei, Michael Saxon, Shiyu Chang, Zachary C. Lipton, William Yang Wang

We conduct a broad literature survey, identifying many clusters of similar conceptions of transparency, tying each back to our north star with analysis of how it furthers or hinders our ideal AI transparency goals.

Foveate, Attribute, and Rationalize: Towards Physically Safe and Trustworthy AI

1 code implementation19 Dec 2022 Alex Mei, Sharon Levy, William Yang Wang

Users' physical safety is an increasing concern as the market for intelligent systems continues to grow, where unconstrained systems may recommend users dangerous actions that can lead to serious injury.

Attribute

Mitigating Covertly Unsafe Text within Natural Language Systems

no code implementations17 Oct 2022 Alex Mei, Anisha Kabir, Sharon Levy, Melanie Subbiah, Emily Allaway, John Judge, Desmond Patton, Bruce Bimber, Kathleen McKeown, William Yang Wang

An increasingly prevalent problem for intelligent technologies is text safety, as uncontrolled systems may generate recommendations to their users that lead to injury or life-threatening consequences.

Cannot find the paper you are looking for? You can Submit a new open access paper.