Search Results for author: Weihong Zhong

Found 14 papers, 6 papers with code

Improving Contextual Faithfulness of Large Language Models via Retrieval Heads-Induced Optimization

no code implementations23 Jan 2025 Lei Huang, Xiaocheng Feng, Weitao Ma, Yuchun Fan, Xiachong Feng, Yangfan Ye, Weihong Zhong, Yuxuan Gu, Baoxin Wang, Dayong Wu, Guoping Hu, Bing Qin

In this work, we identify a salient correlation between LFQA faithfulness and retrieval heads, a set of attention heads responsible for retrieving contextual information.

Long Form Question Answering Retrieval

Cross-Lingual Text-Rich Visual Comprehension: An Information Theory Perspective

no code implementations23 Dec 2024 Xinmiao Yu, Xiaocheng Feng, Yun Li, Minghui Liao, Ya-Qi Yu, Xiachong Feng, Weihong Zhong, Ruihan Chen, Mengkang Hu, Jihao Wu, Dandan Tu, Duyu Tang, Bing Qin

To mitigate this issue, we propose MVCL-MI (Maximization of Vision-Language Cross-Lingual Mutual Information), where a visual-text cross-lingual alignment is built by maximizing mutual information between the model's outputs and visual information.

Question Answering Visual Question Answering

Length Controlled Generation for Black-box LLMs

no code implementations19 Dec 2024 Yuxuan Gu, Wenjie Wang, Xiaocheng Feng, Weihong Zhong, Kun Zhu, Lei Huang, Tat-Seng Chua, Bing Qin

Large language models (LLMs) have demonstrated impressive instruction following capabilities, while still struggling to accurately manage the length of the generated text, which is a fundamental requirement in many real-world applications.

Abstractive Text Summarization Instruction Following

Discrete Modeling via Boundary Conditional Diffusion Processes

no code implementations29 Oct 2024 Yuxuan Gu, Xiaocheng Feng, Lei Huang, Yingsheng Wu, Zekun Zhou, Weihong Zhong, Kun Zhu, Bing Qin

Experimental results indicate that our approach achieves strong performance in both language modeling and discrete image generation tasks.

Image Generation Language Modeling +1

Advancing Large Language Model Attribution through Self-Improving

no code implementations17 Oct 2024 Lei Huang, Xiaocheng Feng, Weitao Ma, Liang Zhao, Yuchun Fan, Weihong Zhong, Dongliang Xu, Qing Yang, Hongtao Liu, Bing Qin

Teaching large language models (LLMs) to generate text with citations to evidence sources can mitigate hallucinations and enhance verifiability in information-seeking systems.

Language Modeling Language Modelling +2

Extending Context Window of Large Language Models from a Distributional Perspective

1 code implementation2 Oct 2024 Yingsheng Wu, Yuxuan Gu, Xiaocheng Feng, Weihong Zhong, Dongliang Xu, Qing Yang, Hongtao Liu, Bing Qin

However, existing scaling methods often rely on empirical approaches and lack a profound understanding of the internal distribution within RoPE, resulting in suboptimal performance in extending the context window length.

16k 8k

Learning Fine-Grained Grounded Citations for Attributed Large Language Models

1 code implementation8 Aug 2024 Lei Huang, Xiaocheng Feng, Weitao Ma, Yuxuan Gu, Weihong Zhong, Xiachong Feng, Weijiang Yu, Weihua Peng, Duyu Tang, Dandan Tu, Bing Qin

Despite the impressive performance on information-seeking tasks, large language models (LLMs) still struggle with hallucinations.

In-Context Learning

Investigating and Mitigating the Multimodal Hallucination Snowballing in Large Vision-Language Models

1 code implementation30 Jun 2024 Weihong Zhong, Xiaocheng Feng, Liang Zhao, Qiming Li, Lei Huang, Yuxuan Gu, Weitao Ma, Yuan Xu, Bing Qin

To mitigate this, we further propose a training-free method called Residual Visual Decoding, where we revise the output distribution of LVLMs with the one derived from the residual visual input, providing models with direct access to the visual information.

Hallucination multimodal interaction

Unveiling Entity-Level Unlearning for Large Language Models: A Comprehensive Analysis

no code implementations22 Jun 2024 Weitao Ma, Xiaocheng Feng, Weihong Zhong, Lei Huang, Yangfan Ye, Xiachong Feng, Bing Qin

Large language model unlearning has garnered increasing attention due to its potential to address security and privacy concerns, leading to extensive research in the field.

Language Modeling Language Modelling +1

STOA-VLP: Spatial-Temporal Modeling of Object and Action for Video-Language Pre-training

no code implementations20 Feb 2023 Weihong Zhong, Mao Zheng, Duyu Tang, Xuan Luo, Heng Gong, Xiaocheng Feng, Bing Qin

Although large-scale video-language pre-training models, which usually build a global alignment between the video and the text, have achieved remarkable progress on various downstream tasks, the idea of adopting fine-grained information during the pre-training stage is not well explored.

Language Modelling Object +5

Controllable Text Generation via Probability Density Estimation in the Latent Space

1 code implementation16 Dec 2022 Yuxuan Gu, Xiaocheng Feng, Sicheng Ma, Lingyuan Zhang, Heng Gong, Weihong Zhong, Bing Qin

Previous work on controllable text generation has explored the idea of control from the latent space, such as optimizing a representation with attribute-related classifiers or sampling a representation from relevant discrete samples.

Attribute Density Estimation +2

Cannot find the paper you are looking for? You can Submit a new open access paper.