Search Results for author: Hanyu Duan

Found 4 papers, 1 papers with code

Do LLMs Know about Hallucination? An Empirical Investigation of LLM's Hidden States

no code implementations15 Feb 2024 Hanyu Duan, Yi Yang, Kar Yan Tam

More specifically, we check whether and how an LLM reacts differently in its hidden states when it answers a question right versus when it hallucinates.

Hallucination

Bias A-head? Analyzing Bias in Transformer-Based Language Model Attention Heads

no code implementations17 Nov 2023 Yi Yang, Hanyu Duan, Ahmed Abbasi, John P. Lalor, Kar Yan Tam

Although a burgeoning literature has emerged on stereotypical bias mitigation in PLMs, such as work on debiasing gender and racial stereotyping, how such biases manifest and behave internally within PLMs remains largely unknown.

Fairness Language Modelling

Exploring the Relationship between In-Context Learning and Instruction Tuning

no code implementations17 Nov 2023 Hanyu Duan, Yixuan Tang, Yi Yang, Ahmed Abbasi, Kar Yan Tam

In this work, we explore the relationship between ICL and IT by examining how the hidden states of LLMs change in these two paradigms.

In-Context Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.