no code implementations • 30 Mar 2024 • Shu Yang, Jiayuan Su, Han Jiang, Mengdi Li, Keyuan Cheng, Muhammad Asif Ali, Lijie Hu, Di Wang
With the rise of large language models (LLMs), ensuring they embody the principles of being helpful, honest, and harmless (3H), known as Human Alignment, becomes crucial.
no code implementations • 2 Mar 2024 • Jiayuan Su, Jing Luo, Hongwei Wang, Lu Cheng
This study aims to address the pervasive challenge of quantifying uncertainty in large language models (LLMs) without logit-access.