Text Generation

526 papers with code • 13 benchmarks • 66 datasets

Text generation is the task of generating text with the goal of appearing indistinguishable to human-written text.

( Image credit: Adversarial Ranking for Language Generation )

Greatest papers with code

Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks

huggingface/transformers NeurIPS 2020

Large pre-trained language models have been shown to store factual knowledge in their parameters, and achieve state-of-the-art results when fine-tuned on downstream NLP tasks.

Question Answering Text Generation

Plug and Play Language Models: A Simple Approach to Controlled Text Generation

huggingface/transformers ICLR 2020

Large transformer-based language models (LMs) trained on huge text corpora have shown unparalleled generation capabilities.

Language Modelling Text Generation

BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension

huggingface/transformers ACL 2020

We evaluate a number of noising approaches, finding the best performance by both randomly shuffling the order of the original sentences and using a novel in-filling scheme, where spans of text are replaced with a single mask token.

Abstractive Text Summarization Denoising +4

HuggingFace's Transformers: State-of-the-art Natural Language Processing

huggingface/transformers 9 Oct 2019

Transformer architectures have facilitated building higher-capacity models and pretraining has made it possible to effectively utilize this capacity for a wide variety of tasks.

Text Generation Transfer Learning

Language Models are Unsupervised Multitask Learners

huggingface/transformers Preprint 2019

Natural language processing tasks, such as question answering, machine translation, reading comprehension, and summarization, are typically approached with supervised learning on taskspecific datasets.

 Ranked #1 on Language Modelling on enwik8 (using extra training data)

Common Sense Reasoning Data-to-Text Generation +6

Stepwise Extractive Summarization and Planning with Structured Transformers

google-research/google-research EMNLP 2020

We propose encoder-centric stepwise models for extractive summarization using structured transformers -- HiBERT and Extended Transformers.

Extractive Summarization Table-to-Text Generation

Towards Automatic Evaluation of Dialog Systems: A Model-Free Off-Policy Evaluation Approach

google-research/google-research 20 Feb 2021

An ideal environment for evaluating dialog systems, also known as the Turing test, needs to involve human interaction, which is usually not affordable for large-scale experiments.

Model-based Reinforcement Learning Text Generation

fairseq: A Fast, Extensible Toolkit for Sequence Modeling

pytorch/fairseq NAACL 2019

fairseq is an open-source sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling, and other text generation tasks.

Language Modelling Text Generation

Pre-trained Language Model Representations for Language Generation

pytorch/fairseq NAACL 2019

Pre-trained language model representations have been successful in a wide range of language understanding tasks.

Abstractive Text Summarization Language Modelling +2

Mixture Models for Diverse Machine Translation: Tricks of the Trade

pytorch/fairseq 20 Feb 2019

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Machine Translation Text Generation