Search Results for author: Anhong Guo

Found 8 papers, 2 papers with code

ProgramAlly: Creating Custom Visual Access Programs via Multi-Modal End-User Programming

no code implementations20 Aug 2024 Jaylin Herskovitz, Andi Xu, Rahaf Alharbi, Anhong Guo

Existing visual assistive technologies are built for simple and common use cases, and have few avenues for blind people to customize their functionalities.

VRCopilot: Authoring 3D Layouts with Generative AI Models in VR

no code implementations18 Aug 2024 Lei Zhang, Jin Pan, Jacob Gettig, Steve Oney, Anhong Guo

Through a series of user studies, we evaluated the potential and challenges in manual, scaffolded, and automatic creation in immersive authoring.

EditScribe: Non-Visual Image Editing with Natural Language Verification Loops

no code implementations13 Aug 2024 Ruei-Che Chang, Yuxuan Liu, Lotus Zhang, Anhong Guo

To address this, we developed EditScribe, a prototype system that makes image editing accessible using natural language verification loops powered by large multimodal models.

WorldScribe: Towards Context-Aware Live Visual Descriptions

no code implementations13 Aug 2024 Ruei-Che Chang, Yuxuan Liu, Anhong Guo

In this work, we develop WorldScribe, a system that generates automated live real-world visual descriptions that are customizable and adaptive to users' contexts: (i) WorldScribe's descriptions are tailored to users' intents and prioritized based on semantic relevance.

Designing Disaggregated Evaluations of AI Systems: Choices, Considerations, and Tradeoffs

no code implementations10 Mar 2021 Solon Barocas, Anhong Guo, Ece Kamar, Jacquelyn Krones, Meredith Ringel Morris, Jennifer Wortman Vaughan, Duncan Wadsworth, Hanna Wallach

Disaggregated evaluations of AI systems, in which system performance is assessed and reported separately for different groups of people, are conceptually simple.

StateLens: A Reverse Engineering Solution for Making Existing Dynamic Touchscreens Accessible

1 code implementation20 Aug 2019 Anhong Guo, Junhan Kong, Michael Rivera, Frank F. Xu, Jeffrey P. Bigham

Second, using the state diagrams, StateLens automatically generates conversational agents to guide blind users through specifying the tasks that the interface can perform, allowing the StateLens iOS application to provide interactive guidance and feedback so that blind users can access the interface.

VizWiz Grand Challenge: Answering Visual Questions from Blind People

1 code implementation CVPR 2018 Danna Gurari, Qing Li, Abigale J. Stangl, Anhong Guo, Chi Lin, Kristen Grauman, Jiebo Luo, Jeffrey P. Bigham

The study of algorithms to automatically answer visual questions currently is motivated by visual question answering (VQA) datasets constructed in artificial VQA settings.

Question Answering Visual Question Answering

Cannot find the paper you are looking for? You can Submit a new open access paper.