Search Results

A Comprehensive Survey of Hallucination Mitigation Techniques in Large Language Models

1 code implementation2 Jan 2024

As Large Language Models (LLMs) continue to advance in their ability to write human-like text, a key challenge remains around their tendency to hallucinate generating content that appears factual but is ungrounded.

Financial Analysis Hallucination +2

SynCode: LLM Generation with Grammar Augmentation

1 code implementation3 Mar 2024

We present SynCode, a novel framework for efficient and general syntactical decoding with LLMs, to address this challenge.

Code Generation valid

BatGPT: A Bidirectional Autoregessive Talker from Generative Pre-trained Transformer

1 code implementation1 Jul 2023

BatGPT is a large-scale language model designed and trained jointly by Wuhan University and Shanghai Jiao Tong University.

Language Modeling Language Modelling +2

Granite Guardian

1 code implementation10 Dec 2024

We introduce the Granite Guardian models, a suite of safeguards designed to provide risk detection for prompts and responses, enabling safe and responsible use in combination with any large language model (LLM).

Hallucination Language Modeling +2

ML-Bench: Evaluating Large Language Models and Agents for Machine Learning Tasks on Repository-Level Code

1 code implementation16 Nov 2023

Despite Large Language Models (LLMs) like GPT-4 achieving impressive results in function-level code generation, they struggle with repository-scale code understanding (e. g., coming up with the right arguments for calling routines), requiring a deeper comprehension of complex file interactions.

Code Generation Navigate +1

Cognitive Mirage: A Review of Hallucinations in Large Language Models

1 code implementation13 Sep 2023

Our contribution are threefold: (1) We provide a detailed and complete taxonomy for hallucinations appearing in text generation tasks; (2) We provide theoretical analyses of hallucinations in LLMs and provide existing detection and improvement methods; (3) We propose several research directions that can be developed in the future.

Hallucination Text Generation

Agent AI: Surveying the Horizons of Multimodal Interaction

1 code implementation7 Jan 2024

To accelerate research on agent-based multimodal intelligence, we define "Agent AI" as a class of interactive systems that can perceive visual stimuli, language inputs, and other environmentally-grounded data, and can produce meaningful embodied actions.

A Survey on Hallucination in Large Vision-Language Models

1 code implementation1 Feb 2024

In this comprehensive survey, we dissect LVLM-related hallucinations in an attempt to establish an overview and facilitate future mitigation.

Hallucination Survey

PhD: A ChatGPT-Prompted Visual hallucination Evaluation Dataset

1 code implementation17 Mar 2024

This paper contributes a ChatGPT-Prompted visual hallucination evaluation Dataset (PhD) for objective VHE at a large scale.

Attribute Common Sense Reasoning +4

Aligning Modalities in Vision Large Language Models via Preference Fine-tuning

1 code implementation18 Feb 2024

This procedure is not perfect and can cause the model to hallucinate - provide answers that do not accurately reflect the image, even when the core LLM is highly factual and the vision backbone has sufficiently complete representations.

Hallucination Instruction Following +1