TextSnake: A Flexible Representation for Detecting Text of Arbitrary Shapes

Driven by deep neural networks and large scale datasets, scene text detection methods have progressed substantially over the past years, continuously refreshing the performance records on various standard benchmarks. However, limited by the representations (axis-aligned rectangles, rotated rectangles or quadrangles) adopted to describe text, existing methods may fall short when dealing with much more free-form text instances, such as curved text, which are actually very common in real-world scenarios. To tackle this problem, we propose a more flexible representation for scene text, termed as TextSnake, which is able to effectively represent text instances in horizontal, oriented and curved forms. In TextSnake, a text instance is described as a sequence of ordered, overlapping disks centered at symmetric axes, each of which is associated with potentially variable radius and orientation. Such geometry attributes are estimated via a Fully Convolutional Network (FCN) model. In experiments, the text detector based on TextSnake achieves state-of-the-art or comparable performance on Total-Text and SCUT-CTW1500, the two newly published benchmarks with special emphasis on curved text in natural images, as well as the widely-used datasets ICDAR 2015 and MSRA-TD500. Specifically, TextSnake outperforms the baseline on Total-Text by more than 40% in F-measure.

PDF Abstract ECCV 2018 PDF ECCV 2018 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Scene Text Detection ICDAR 2015 PixelLink+VGG16 2s F-Measure 83.7 # 29
Precision 85.5 # 32
Recall 82 # 23
Scene Text Detection MSRA-TD500 TextSnake Recall 73.9 # 14
Precision 83.2 # 16
F-Measure 78.3 # 15
Curved Text Detection SCUT-CTW1500 TextSnake [[Long et al.(2018)Long, Ruan, Zhang, He, Wu, and Yao]] F-Measure 75.6% # 2
Scene Text Detection SCUT-CTW1500 TextSnake F-Measure 75.6 # 16
Precision 67.9 # 16
Recall 85.3 # 4
Scene Text Detection Total-Text TextSnake F-Measure 78.4% # 24
Precision 82.7 # 20
Recall 74.5 # 21

Methods


No methods listed for this paper. Add relevant methods here