Text Summarization

368 papers with code • 33 benchmarks • 87 datasets

Text Summarization is a natural language processing (NLP) task that involves condensing a lengthy text document into a shorter, more compact version while still retaining the most important information and meaning. The goal is to produce a summary that accurately represents the content of the original text in a concise form.

There are different approaches to text summarization, including extractive methods that identify and extract important sentences or phrases from the text, and abstractive methods that generate new text based on the content of the original text.

Libraries

Use these libraries to find Text Summarization models and implementations

Most implemented papers

Leveraging BERT for Extractive Text Summarization on Lectures

dmmiller612/lecture-summarizer 7 Jun 2019

This paper reports on the project called Lecture Summarization Service, a python based RESTful service that utilizes the BERT model for text embeddings and KMeans clustering to identify sentences closes to the centroid for summary selection.

Incorporating Copying Mechanism in Sequence-to-Sequence Learning

allenai/allennlp-models ACL 2016

CopyNet can nicely integrate the regular way of word generation in the decoder with the new copying mechanism which can choose sub-sequences in the input sequence and put them at proper places in the output sequence.

SummaRuNNer: A Recurrent Neural Network based Sequence Model for Extractive Summarization of Documents

kedz/nnsum 14 Nov 2016

We present SummaRuNNer, a Recurrent Neural Network (RNN) based sequence model for extractive summarization of documents and show that it achieves performance better than or comparable to state-of-the-art.

MASS: Masked Sequence to Sequence Pre-training for Language Generation

microsoft/MASS 7 May 2019

Pre-training and fine-tuning, e. g., BERT, have achieved great success in language understanding by transferring knowledge from rich-resource pre-training task to the low/zero-resource downstream tasks.

NLG Evaluation Metrics Beyond Correlation Analysis: An Empirical Metric Preference Checklist

inimah/metric-preference-checklist 15 May 2023

Our proposed framework provides access: (i) for verifying whether automatic metrics are faithful to human preference, regardless of their correlation level to human; and (ii) for inspecting the strengths and limitations of NLG systems via pairwise evaluation.

Leveraging Pre-trained Checkpoints for Sequence Generation Tasks

huggingface/transformers TACL 2020

Unsupervised pre-training of large neural models has recently revolutionized Natural Language Processing.

Neural Abstractive Text Summarization with Sequence-to-Sequence Models

tshi04/NATS 5 Dec 2018

As part of this survey, we also develop an open source library, namely, Neural Abstractive Text Summarizer (NATS) toolkit, for the abstractive text summarization.

SummEval: Re-evaluating Summarization Evaluation

Yale-LILY/SummEval 24 Jul 2020

The scarcity of comprehensive up-to-date studies on evaluation metrics for text summarization and the lack of consensus regarding evaluation protocols continue to inhibit progress.

A Neural Attention Model for Abstractive Sentence Summarization

tensorflow/models EMNLP 2015

Summarization based on text extraction is inherently limited, but generation-style abstractive methods have proven challenging to build.

Pretraining-Based Natural Language Generation for Text Summarization

nayeon7lee/bert-summarization CONLL 2019

For the decoder, there are two stages in our model, in the first stage, we use a Transformer-based decoder to generate a draft output sequence.