Search Results for author: Yanda Li

Found 12 papers, 4 papers with code

MTPChat: A Multimodal Time-Aware Persona Dataset for Conversational Agents

no code implementations9 Feb 2025 Wanqi Yang, Yanda Li, Meng Fang, Ling Chen

Understanding temporal dynamics is critical for conversational agents, enabling effective content analysis and informed decision-making.

Decision Making

Who Can Withstand Chat-Audio Attacks? An Evaluation Benchmark for Large Language Models

1 code implementation22 Nov 2024 Wanqi Yang, Yanda Li, Meng Fang, Yunchao Wei, Tianyi Zhou, Ling Chen

We evaluate six state-of-the-art LLMs with voice interaction capabilities, including Gemini-1. 5-Pro, GPT-4o, and others, using three distinct evaluation methods on the CAA benchmark.

Foundations and Recent Trends in Multimodal Mobile Agents: A Survey

1 code implementation4 Nov 2024 Biao Wu, Yanda Li, Meng Fang, Zirui Song, Zhiwei Zhang, Yunchao Wei, Ling Chen

This survey provides a comprehensive review of mobile agent technologies, focusing on recent advancements that enhance real-time adaptability and multimodal interaction.

multimodal interaction Survey

Enhancing Temporal Sensitivity and Reasoning for Time-Sensitive Question Answering

no code implementations25 Sep 2024 Wanqi Yang, Yanda Li, Meng Fang, Ling Chen

Time-Sensitive Question Answering (TSQA) demands the effective utilization of specific temporal contexts, encompassing multiple time-evolving facts, to address time-sensitive questions.

Question Answering

AppAgent v2: Advanced Agent for Flexible Mobile Interactions

no code implementations5 Aug 2024 Yanda Li, Chi Zhang, Wanqi Yang, Bin Fu, Pei Cheng, Xin Chen, Ling Chen, Yunchao Wei

In the deployment phase, RAG technology enables efficient retrieval and update from this knowledge base, thereby empowering the agent to perform tasks effectively and accurately.

RAG

Continual Learning for Temporal-Sensitive Question Answering

no code implementations17 Jul 2024 Wanqi Yang, Yunqiu Xu, Yanda Li, Kunze Wang, Binbin Huang, Ling Chen

In this study, we explore an emerging research area of Continual Learning for Temporal Sensitive Question Answering (CLTSQA).

Continual Learning Contrastive Learning +1

Adaptive Reinforcement Learning Planning: Harnessing Large Language Models for Complex Information Extraction

no code implementations17 Jun 2024 Zepeng Ding, Ruiyang Ke, Wenhao Huang, Guochao Jiang, Yanda Li, Deqing Yang, Jiaqing Liang

Existing research on large language models (LLMs) shows that they can solve information extraction tasks through multi-step planning.

Missing Elements

Tokenization Matters! Degrading Large Language Models through Challenging Their Tokenization

no code implementations27 May 2024 Dixuan Wang, Yanda Li, Junyuan Jiang, Zepeng Ding, Guochao Jiang, Jiaqing Liang, Deqing Yang

Our empirical results reveal that our ADT is highly effective on challenging the tokenization of leading LLMs, including GPT-4o, Llama-3, Qwen2. 5-max and so on, thus degrading these LLMs' capabilities.

MMAC-Copilot: Multi-modal Agent Collaboration Operating Copilot

no code implementations28 Apr 2024 Zirui Song, Yaohang Li, Meng Fang, Yanda Li, Zhenhao Chen, Zecheng Shi, Yuan Huang, Xiuying Chen, Ling Chen

To address this, we propose the Multi-Modal Agent Collaboration framework (MMAC-Copilot), a framework utilizes the collective expertise of diverse agents to enhance interaction ability with application.

Hallucination Language Modeling +2

Reason from Fallacy: Enhancing Large Language Models' Logical Reasoning through Logical Fallacy Understanding

no code implementations4 Apr 2024 Yanda Li, Dixuan Wang, Jiaqing Liang, Guochao Jiang, Qianyu He, Yanghua Xiao, Deqing Yang

Large Language Models (LLMs) have demonstrated good performance in many reasoning tasks, but they still struggle with some complicated reasoning tasks including logical reasoning.

Logical Fallacies Logical Reasoning

Disentangled Pre-training for Image Matting

1 code implementation3 Apr 2023 Yanda Li, Zilong Huang, Gang Yu, Ling Chen, Yunchao Wei, Jianbo Jiao

The pre-training task is designed in a similar manner as image matting, where random trimap and alpha matte are generated to achieve an image disentanglement objective.

Disentanglement Image Matting

Cannot find the paper you are looking for? You can Submit a new open access paper.