Question Answering
2874 papers with code • 143 benchmarks • 360 datasets
Question Answering is the task of answering questions (typically reading comprehension questions), but abstaining when presented with a question that cannot be answered based on the provided context.
Question answering can be segmented into domain-specific tasks like community question answering and knowledge-base question answering. Popular benchmark datasets for evaluation question answering systems include SQuAD, HotPotQA, bAbI, TriviaQA, WikiQA, and many others. Models for question answering are typically evaluated on metrics like EM and F1. Some recent top performing models are T5 and XLNet.
( Image credit: SQuAD )
Libraries
Use these libraries to find Question Answering models and implementationsDatasets
Subtasks
- Open-Ended Question Answering
- Open-Domain Question Answering
- Conversational Question Answering
- Answer Selection
- Answer Selection
- Knowledge Base Question Answering
- Community Question Answering
- Zero-Shot Video Question Answer
- Multiple Choice Question Answering (MCQA)
- Long Form Question Answering
- Science Question Answering
- Generative Question Answering
- Cross-Lingual Question Answering
- Mathematical Question Answering
- Temporal/Casual QA
- Logical Reasoning Question Answering
- Multilingual Machine Comprehension in English Hindi
- True or False Question Answering
- Question Quality Assessment
Latest papers
EuSQuAD: Automatically Translated and Aligned SQuAD2.0 for Basque
The widespread availability of Question Answering (QA) datasets in English has greatly facilitated the advancement of the Natural Language Processing (NLP) field.
Consistency Training by Synthetic Question Generation for Conversational Question Answering
In our novel model-agnostic approach, referred to as CoTaH (Consistency-Trained augmented History), we augment the historical information with synthetic questions and subsequently employ consistency training to train a model that utilizes both real and augmented historical data to implicitly make the reasoning robust to irrelevant history.
ViTextVQA: A Large-Scale Visual Question Answering Dataset for Evaluating Vietnamese Text Comprehension in Images
Visual Question Answering (VQA) is a complicated task that requires the capability of simultaneously processing natural language and images.
Spiral of Silences: How is Large Language Model Killing Information Retrieval? -- A Case Study on Open Domain Question Answering
The practice of Retrieval-Augmented Generation (RAG), which integrates Large Language Models (LLMs) with retrieval systems, has become increasingly prevalent.
TextCoT: Zoom In for Enhanced Multimodal Text-Rich Image Understanding
The image overview stage provides a comprehensive understanding of the global scene information, and the coarse localization stage approximates the image area containing the answer based on the question asked.
Bridging Vision and Language Spaces with Assignment Prediction
This paper introduces VLAP, a novel approach that bridges pretrained vision models and large language models (LLMs) to make frozen LLMs understand the visual world.
Constructing Benchmarks and Interventions for Combating Hallucinations in LLMs
In this work, we first introduce an approach for constructing datasets based on the model knowledge for detection and intervention methods in closed-book and open-book question-answering settings.
CuriousLLM: Elevating Multi-Document QA with Reasoning-Infused Knowledge Graph Prompting
In the field of Question Answering (QA), unifying large language models (LLMs) with external databases has shown great success.
Synthetic Dataset Creation and Fine-Tuning of Transformer Models for Question Answering in Serbian
In this paper, we focus on generating a synthetic question answering (QA) dataset using an adapted Translate-Align-Retrieve method.
Enhancing Visual Question Answering through Question-Driven Image Captions as Prompts
This study explores the impact of incorporating image captioning as an intermediary process within the VQA pipeline.