Search Results for author: Qinlin Zhao

Found 3 papers, 1 papers with code

DyVal 2: Dynamic Evaluation of Large Language Models by Meta Probing Agents

no code implementations21 Feb 2024 Kaijie Zhu, Jindong Wang, Qinlin Zhao, Ruochen Xu, Xing Xie

Our multifaceted analysis demonstrated the strong correlation between the basic abilities and an implicit Matthew effect on model size, i. e., larger models possess stronger correlations of the abilities.

Data Augmentation

PromptBench: A Unified Library for Evaluation of Large Language Models

1 code implementation13 Dec 2023 Kaijie Zhu, Qinlin Zhao, Hao Chen, Jindong Wang, Xing Xie

The evaluation of large language models (LLMs) is crucial to assess their performance and mitigate potential security risks.

Prompt Engineering

CompeteAI: Understanding the Competition Behaviors in Large Language Model-based Agents

no code implementations26 Oct 2023 Qinlin Zhao, Jindong Wang, Yixuan Zhang, Yiqiao Jin, Kaijie Zhu, Hao Chen, Xing Xie

Large language models (LLMs) have been widely used as agents to complete different tasks, such as personal assistance or event planning.

Language Modelling Large Language Model

Cannot find the paper you are looking for? You can Submit a new open access paper.