no code implementations • 14 Apr 2025 • Hao Yin, Guangzong Si, Zilei Wang
Contrastive decoding strategies are widely used to reduce hallucinations in multimodal large language models (MLLMs).
1 code implementation • 17 Mar 2025 • Hao Yin, Guangzong Si, Zilei Wang
However, these methods present two main limitations: (1) bluntly suppressing language priors can compromise coherence and accuracy of generated content, and (2) processing contrastive inputs adds computational load, significantly slowing inference speed.
1 code implementation • 17 Mar 2025 • Hao Yin, Guangzong Si, Zilei Wang
Multimodal large language models (MLLMs) improve performance on vision-language tasks by integrating visual features from pre-trained vision encoders into large language models (LLMs).