no code implementations • 29 Nov 2024 • Shengming Zhao, Yuheng Huang, Jiayang Song, Zhijie Wang, Chengcheng Wan, Lei Ma
Retrieval-Augmented Generation (RAG) is a pivotal technique for enhancing the capability of large language models (LLMs) and has demonstrated promising efficacy across a diverse spectrum of tasks.
no code implementations • 7 Oct 2024 • Zhijie Wang, Zhehua Zhou, Jiayang Song, Yuheng Huang, Zhan Shu, Lei Ma
This highlights the need for a reliable testing and evaluation platform.
no code implementations • 20 Aug 2024 • Xuan Xie, Jiayang Song, Yuheng Huang, Da Song, Fuyuan Zhang, Felix Juefei-Xu, Lei Ma
Large Language Models (LLMs) are widely used in many different domains, but because of their limited interpretability, there are questions about how trustworthy they are in various perspectives, e. g., truthfulness and toxicity.
no code implementations • 7 Aug 2024 • Renzhi Wang, Zhehua Zhou, Jiayang Song, Xuan Xie, Xiaofei Xie, Lei Ma
If an unsafe action is detected, MORTAR then initiates a repair process to correct it.
no code implementations • 7 Aug 2024 • Yuheng Huang, Jiayang Song, Qiang Hu, Felix Juefei-Xu, Lei Ma
Given that LLMs' diverse task-handling abilities stem from large volumes of training data, a comprehensive evaluation also necessitates abundant, well-annotated, and representative test data to assess LLM performance across various downstream tasks.
no code implementations • 10 Jul 2024 • Jiayang Song, Yuheng Huang, Zhehua Zhou, Lei Ma
As safety remains a crucial concern throughout the development lifecycle of Large Language Models (LLMs), researchers and industrial practitioners have increasingly focused on safeguarding and aligning LLM behaviors with human preferences and ethical standards.
no code implementations • 6 Jun 2024 • Zhehua Zhou, Xuan Xie, Jiayang Song, Zhan Shu, Lei Ma
To address this issue, we introduce in this work a novel Generalizable Safety enhancer (GenSafe) that is able to overcome the challenge of data insufficiency and enhance the performance of SRL approaches.
no code implementations • 12 Apr 2024 • Xuan Xie, Jiayang Song, Zhehua Zhou, Yuheng Huang, Da Song, Lei Ma
To bridge this gap, we conduct in this work a comprehensive evaluation of the effectiveness of existing online safety analysis methods on LLMs.
no code implementations • 22 Oct 2023 • Da Song, Xuan Xie, Jiayang Song, Derui Zhu, Yuheng Huang, Felix Juefei-Xu, Lei Ma
the trustworthiness perspective, is bound to and enriches the abstract model with semantics, which enables more detailed analysis applications for diverse purposes.
1 code implementation • 13 Sep 2023 • Jiayang Song, Zhehua Zhou, Jiawei Liu, Chunrong Fang, Zhan Shu, Lei Ma
Then, the performance of the reward function is assessed, and the results are presented back to the LLM for guiding its self-refinement process.
2 code implementations • 26 Aug 2023 • Zhehua Zhou, Jiayang Song, Kunpeng Yao, Zhan Shu, Lei Ma
Motivated by the substantial achievements observed in Large Language Models (LLMs) in the field of natural language processing, recent research has commenced investigations into the application of LLMs for complex, long-horizon sequential task planning challenges in robotics.
no code implementations • 16 Jul 2023 • Yuheng Huang, Jiayang Song, Zhijie Wang, Shengming Zhao, Huaming Chen, Felix Juefei-Xu, Lei Ma
In particular, we experiment with twelve uncertainty estimation methods and four LLMs on four prominent natural language processing (NLP) tasks to investigate to what extent uncertainty estimation techniques could help characterize the prediction risks of LLMs.