Multi-Document Summarization withDeterminantal Point Process Attention

The ability to convey relevant and diverse information is critical in multi-documentsummarization and yet remains elusive for neural seq-to-seq models whose outputs are of-ten redundant and fail to correctly cover important details. In this work, we propose anattention mechanism which encourages greater focus onrelevanceanddiversity. Attentionweights are computed based on (proportional) probabilities given by Determinantal PointProcesses (DPPs) defined on the set of content units to be summarized. DPPs have beensuccessfully used in extractive summarisation, here we use them to select relevant anddiverse content for neural abstractive summarisation. We integrate DPP-based attentionwith various seq-to-seq architectures ranging from CNNs to LSTMs, and Transformers.Experimental evaluation shows that our attention mechanism consistently improves sum-marization and delivers performance comparable with the state-of-the-art on the MultiNewsdataset.

PDF

Datasets


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Multi-Document Summarization Multi-News CTF+DPP ROUGE-2 15.94 # 3
ROUGE-1 45.84 # 3
ROUGE-SU4 19.19 # 2

Methods


No methods listed for this paper. Add relevant methods here