Search Results for author: Liangtai Sun

Found 4 papers, 1 papers with code

MULTI: Multimodal Understanding Leaderboard with Text and Images

no code implementations5 Feb 2024 Zichen Zhu, Yang Xu, Lu Chen, Jingkai Yang, Yichuan Ma, Yiming Sun, Hailin Wen, Jiaqi Liu, Jinyu Cai, Yingzi Ma, Situo Zhang, Zihan Zhao, Liangtai Sun, Kai Yu

Rapid progress in multimodal large language models (MLLMs) highlights the need to introduce challenging yet realistic benchmarks to the academic community, while existing benchmarks primarily focus on understanding simple natural images and short context.

In-Context Learning

META-GUI: Towards Multi-modal Conversational Agents on Mobile GUI

no code implementations23 May 2022 Liangtai Sun, Xingyu Chen, Lu Chen, Tianle Dai, Zichen Zhu, Kai Yu

However, this API-based architecture greatly limits the information-searching capability of intelligent assistants and may even lead to task failure if TOD-specific APIs are not available or the task is too complicated to be executed by the provided APIs.


Cannot find the paper you are looking for? You can Submit a new open access paper.