no code implementations • NAACL (DeeLIO) 2021 • Hyeju Jang, Seojin Bang, Wen Xiao, Giuseppe Carenini, Raymond Ng, Young ji Lee
Text classification has wide-ranging applications in various domains.
1 code implementation • EMNLP (ACL) 2021 • Raymond Li, Wen Xiao, Lanjun Wang, Hyeju Jang, Giuseppe Carenini
Transformers are the dominant architecture in NLP, but their training and fine-tuning is still very challenging.
no code implementations • 12 Jan 2022 • Qingyong Hu, Bo Yang, Sheikh Khalid, Wen Xiao, Niki Trigoni, Andrew Markham
Each point in the dataset has been labelled with fine-grained semantic annotations, resulting in a dataset that is three times the size of the previous existing largest photogrammetric point cloud dataset.
no code implementations • 10 Dec 2021 • Raymond Li, Wen Xiao, Lanjun Wang, Giuseppe Carenini
The transformer multi-head self-attention mechanism has been thoroughly investigated recently.
1 code implementation • ACL 2022 • Wen Xiao, Iz Beltagy, Giuseppe Carenini, Arman Cohan
We introduce PRIMERA, a pre-trained model for multi-document representation with a focus on summarization that reduces the need for dataset-specific architectures and large amounts of fine-tuning labeled data.
Ranked #1 on
Multi-Document Summarization
on Multi-News
1 code implementation • 31 Aug 2021 • Raymond Li, Wen Xiao, Lanjun Wang, Hyeju Jang, Giuseppe Carenini
Transformers are the dominant architecture in NLP, but their training and fine-tuning is still very challenging.
no code implementations • ACL 2021 • Patrick Huber, Wen Xiao, Giuseppe Carenini
Aiming for a better integration of data-driven and linguistically-inspired approaches, we explore whether RST Nuclearity, assigning a binary assessment of importance between text segments, can be replaced by automatically generated, real-valued scores, in what we call a Weighted-RST framework.
no code implementations • ACL 2021 • Linzi Xing, Wen Xiao, Giuseppe Carenini
In news articles the lead bias is a common phenomenon that usually dominates the learning signals for neural extractive summarizers, severely limiting their performance on data with different or even no bias.
1 code implementation • NAACL 2021 • Wen Xiao, Patrick Huber, Giuseppe Carenini
Previous work indicates that discourse information benefits summarization.
no code implementations • EMNLP (CODI) 2020 • Wen Xiao, Patrick Huber, Giuseppe Carenini
The multi-head self-attention of popular transformer models is widely used within Natural Language Processing (NLP), including for the task of extractive summarization.
1 code implementation • Asian Chapter of the Association for Computational Linguistics 2020 • Wen Xiao, Giuseppe Carenini
Our analysis of large summarization datasets indicates that redundancy is a very serious problem when summarizing long documents.
Ranked #9 on
Text Summarization
on Pubmed
2 code implementations • CVPR 2021 • Qingyong Hu, Bo Yang, Sheikh Khalid, Wen Xiao, Niki Trigoni, Andrew Markham
An essential prerequisite for unleashing the potential of supervised deep learning algorithms in the area of 3D scene understanding is the availability of large-scale and richly annotated datasets.
1 code implementation • IJCNLP 2019 • Wen Xiao, Giuseppe Carenini
In this paper, we propose a novel neural single document extractive summarization model for long documents, incorporating both the global context of the whole document and the local context within the current topic.
Ranked #13 on
Text Summarization
on arXiv