Search Results for author: Zongjie Li

Found 6 papers, 2 papers with code

An Empirical Study on Large Language Models in Accuracy and Robustness under Chinese Industrial Scenarios

no code implementations27 Jan 2024 Zongjie Li, Wenying Qiu, Pingchuan Ma, Yichen Li, You Li, Sijia He, Baozheng Jiang, Shuai Wang, Weixi Gu

In this paper, we present a comprehensive empirical study on the accuracy and robustness of LLMs in the context of the Chinese industrial production area.

VRPTEST: Evaluating Visual Referring Prompting in Large Multimodal Models

no code implementations7 Dec 2023 Zongjie Li, Chaozheng Wang, Chaowei Liu, Pingchuan Ma, Daoyuan Wu, Shuai Wang, Cuiyun Gao

With recent advancements in Large Multimodal Models (LMMs) across various domains, a novel prompting method called visual referring prompting has emerged, showing significant potential in enhancing human-computer interaction within multimodal systems.

InstructTA: Instruction-Tuned Targeted Attack for Large Vision-Language Models

no code implementations4 Dec 2023 Xunguang Wang, Zhenlan Ji, Pingchuan Ma, Zongjie Li, Shuai Wang

Initially, we utilize a public text-to-image generative model to "reverse" the target response into a target image, and employ GPT-4 to infer a reasonable instruction $\boldsymbol{p}^\prime$ from the target response.

Adversarial Attack Language Modelling +2

Benchmarking and Explaining Large Language Model-based Code Generation: A Causality-Centric Approach

1 code implementation10 Oct 2023 Zhenlan Ji, Pingchuan Ma, Zongjie Li, Shuai Wang

We illustrate the insights that our framework can provide by studying over 3 popular LLMs with over 12 prompt adjustment strategies.

Benchmarking Code Generation +2

Split and Merge: Aligning Position Biases in Large Language Model based Evaluators

no code implementations29 Sep 2023 Zongjie Li, Chaozheng Wang, Pingchuan Ma, Daoyuan Wu, Shuai Wang, Cuiyun Gao, Yang Liu

Specifically, PORTIA splits the answers into multiple segments, aligns similar content across candidate answers, and then merges them back into a single prompt for evaluation by LLMs.

Language Modelling Large Language Model +1

"Oops, Did I Just Say That?" Testing and Repairing Unethical Suggestions of Large Language Models with Suggest-Critique-Reflect Process

1 code implementation4 May 2023 Pingchuan Ma, Zongjie Li, Ao Sun, Shuai Wang

Moreover, we propose a novel on-the-fly (OTF) repairing scheme that repairs unethical suggestions made by LLMs in real-time.

Moral Scenarios

Cannot find the paper you are looking for? You can Submit a new open access paper.