HTLM: Hyper-Text Pre-Training and Prompting of Language Models

We introduce HTLM, a hyper-text language model trained on a large-scale web crawl. Modeling hyper-text has a number of advantages: (1) it is easily gathered at scale, (2) it provides rich document-level and end-task-adjacent supervision (e.g. class and id attributes often encode document category information), and (3) it allows for new structured prompting that follows the established semantics of HTML (e.g. to do zero-shot summarization by infilling title tags for a webpage that contains the input text). We show that pretraining with a BART-style denoising loss directly on simplified HTML provides highly effective transfer for a wide range of end tasks and supervision levels. HTLM matches or exceeds the performance of comparably sized text-only LMs for zero-shot prompting and fine-tuning for classification benchmarks, while also setting new state-of-the-art performance levels for zero-shot summarization. We also find that hyper-text prompts provide more value to HTLM, in terms of data efficiency, than plain text prompts do for existing LMs, and that HTLM is highly effective at auto-prompting itself, by simply generating the most likely hyper-text formatting for any available training data. We will release all code and models to support future HTLM research.

PDF Abstract ICLR 2022 PDF ICLR 2022 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Table-to-Text Generation DART HTLM (fine-tuning) BLEU 47.2 # 1
METEOR 0.39 # 1
TER 0.44 # 1
Mover 0.51 # 1
BERT 0.94 # 1
BLEURT 0.4 # 1
Table-to-Text Generation DART GPT-2-Large (fine-tuning) BLEU 47.0 # 2
METEOR 0.39 # 1
TER 0.46 # 2
Mover 0.51 # 1
BERT 0.94 # 1
BLEURT 0.4 # 1
Table-to-Text Generation E2E HTLM (fine-tuning) BLEU 70.3 # 1
NIST 8.90 # 1
METEOR 46.3 # 1
ROUGE-L 70.8 # 1
CIDEr 2.47 # 1
Table-to-Text Generation E2E GPT-2-Large (fine-tuning) BLEU 68.5 # 2
NIST 8.78 # 2
METEOR 46.0 # 2
ROUGE-L 69.9 # 2
CIDEr 2.45 # 2
Data-to-Text Generation WebNLG HTML (fine-tuning) BLEU 65.4 # 4
Table-to-Text Generation WebNLG (All) GPT-2-Large (fine-tuning) BLEU 55.5 # 2
METEOR 0.42 # 1
TER 0.42 # 1
Table-to-Text Generation WebNLG (All) HTLM (fine-tuning) BLEU 55.6 # 1
METEOR 0.42 # 1
TER 0.4 # 2
Data-to-Text Generation WebNLG Full HTLM (prefix 0.1%) BLEU 56.3 # 6
Table-to-Text Generation WebNLG (Seen) HTLM (fine-tuning) BLEU 65.4 # 1
METEOR 0.46 # 1
TER 0.33 # 1
Table-to-Text Generation WebNLG (Seen) GPT-2-Large (fine-tuning) BLEU 65.3 # 2
METEOR 0.46 # 1
TER 0.33 # 1
Table-to-Text Generation WebNLG (Unseen) HTLM (fine-tuning) BLEU 48.4 # 1
METEOR 0.39 # 1
TER 0.51 # 2
Table-to-Text Generation WebNLG (Unseen) GPT-2-Large (fine-tuning) BLEU 43.1 # 2
METEOR 0.38 # 2
TER 0.53 # 1

Methods


No methods listed for this paper. Add relevant methods here