Reading Comprehension

568 papers with code • 7 benchmarks • 95 datasets

Most current question answering datasets frame the task as reading comprehension where the question is about a paragraph or document and the answer often is a span in the document.

Some specific tasks of reading comprehension include multi-modal machine reading comprehension and textual machine reading comprehension, among others. In the literature, machine reading comprehension can be divide into four categories: cloze style, multiple choice, span prediction, and free-form answer. Read more about each category here.

Benchmark datasets used for testing a model's reading comprehension abilities include MovieQA, ReCoRD, and RACE, among others.

The Machine Reading group at UCL also provides an overview of reading comprehension tasks.

Figure source: A Survey on Machine Reading Comprehension: Tasks, Evaluation Metrics and Benchmark Datasets

Libraries

Use these libraries to find Reading Comprehension models and implementations
4 papers
1,102
2 papers
8,533
See all 6 libraries.

AC-EVAL: Evaluating Ancient Chinese Language Understanding in Large Language Models

yuting-wei/ac-eval 11 Mar 2024

Given the importance of ancient Chinese in capturing the essence of rich historical and cultural heritage, the rapid advancements in Large Language Models (LLMs) necessitate benchmarks that can effectively evaluate their understanding of ancient contexts.

3
11 Mar 2024

Video Relationship Detection Using Mixture of Experts

shibshib/Moe-VRD IEEE Access 2023

Secondly, classifiers trained by a single, monolithic neural network often lack stability and generalization.

2
06 Mar 2024

Scaling Rectified Flow Transformers for High-Resolution Image Synthesis

hxixixh/adaflow 5 Mar 2024

Rectified flow is a recent generative model formulation that connects data and noise in a straight line.

3
05 Mar 2024

PoTeC: A German Naturalistic Eye-tracking-while-reading Corpus

dili-lab/potec 1 Mar 2024

The Potsdam Textbook Corpus (PoTeC) is a naturalistic eye-tracking-while-reading corpus containing data from 75 participants reading 12 scientific texts.

8
01 Mar 2024

Causal Orthogonalization: Multicollinearity, Economic Interpretability, and the Gram-Schmidt Process

crossrm/gsls 27 Feb 2024

This paper considers the problem of interpreting orthogonalization model coefficients.

1
27 Feb 2024

VlogQA: Task, Dataset, and Baseline Models for Vietnamese Spoken-Based Machine Reading Comprehension

sonlam1102/vlogqa 5 Feb 2024

This paper presents the development process of a Vietnamese spoken language corpus for machine reading comprehension (MRC) tasks and provides insights into the challenges and opportunities associated with using real-world data for machine reading comprehension tasks.

0
05 Feb 2024

An Information-Theoretic Approach to Analyze NLP Classification Tasks

wangluran/nlp-element-influence 1 Feb 2024

This work provides an information-theoretic framework to analyse the influence of inputs for text classification tasks.

1
01 Feb 2024

Large Language Models are Superpositions of All Characters: Attaining Arbitrary Role-play via Self-Alignment

ofa-sys/ditto 23 Jan 2024

Nevertheless, we posit that LLMs inherently harbor role-play capabilities, owing to the extensive knowledge of characters and potential dialogues ingrained in their vast training corpora.

105
23 Jan 2024

Knowledge Fusion of Large Language Models

fanqiwan/fusellm 19 Jan 2024

In this paper, we introduce the notion of knowledge fusion for LLMs, aimed at combining the capabilities of existing LLMs and transferring them into a single LLM.

326
19 Jan 2024

Improving Domain Adaptation through Extended-Text Reading Comprehension

microsoft/lmops 14 Jan 2024

To enhance the domain-specific capabilities of large language models, continued pre-training on a domain-specific corpus is a prevalent method.

3,174
14 Jan 2024