PubLayNet: largest dataset ever for document layout analysis

16 Aug 2019  ·  Xu Zhong, Jianbin Tang, Antonio Jimeno Yepes ·

Recognizing the layout of unstructured digital documents is an important step when parsing the documents into structured machine-readable format for downstream applications. Deep neural networks that are developed for computer vision have been proven to be an effective method to analyze layout of document images. However, document layout datasets that are currently publicly available are several magnitudes smaller than established computing vision datasets. Models have to be trained by transfer learning from a base model that is pre-trained on a traditional computer vision dataset. In this paper, we develop the PubLayNet dataset for document layout analysis by automatically matching the XML representations and the content of over 1 million PDF articles that are publicly available on PubMed Central. The size of the dataset is comparable to established computer vision datasets, containing over 360 thousand document images, where typical document layout elements are annotated. The experiments demonstrate that deep neural networks trained on PubLayNet accurately recognize the layout of scientific articles. The pre-trained models are also a more effective base mode for transfer learning on a different document domain. We release the dataset (https://github.com/ibm-aur-nlp/PubLayNet) to support development and evaluation of more advanced models for document layout analysis.

PDF Abstract

Datasets


Introduced in the Paper:

PubLayNet
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Document Layout Analysis PubLayNet val Mask RCNN Text 0.916 # 11
Title 0.840 # 11
List 0.886 # 11
Table 0.960 # 12
Figure 0.949 # 11
Overall 0.910 # 11
Document Layout Analysis PubLayNet val Faster RCNN Text 0.910 # 12
Title 0.826 # 12
List 0.883 # 12
Table 0.954 # 13
Figure 0.937 # 12
Overall 0.902 # 12

Methods


No methods listed for this paper. Add relevant methods here