Arbitrary-Shaped Text Detection With Adaptive Text Region Representation

Text detection/localization, as an important task in computer vision, has witnessed substantial advancements in methodology and performance with convolutional neural networks. However, the vast majority of popular methods use rectangles or quadrangles to describe text regions. These representations...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2020, Vol.8, p.102106-102118
Hauptverfasser: Jiang, Xiufeng, Xu, Shugong, Zhang, Shunqing, Cao, Shan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Text detection/localization, as an important task in computer vision, has witnessed substantial advancements in methodology and performance with convolutional neural networks. However, the vast majority of popular methods use rectangles or quadrangles to describe text regions. These representations have inherent drawbacks, especially relating to dense adjacent text and loose regional text boundaries, which usually cause difficulty detecting arbitrarily shaped text. In this paper, we propose a novel text region representation method, with a robust pipeline, which can precisely detect dense adjacent text instances with arbitrary shapes. We consider a text instance to be composed of an adaptive central text region mask and a corresponding expanding ratio between the central text region and the full text region. More specifically, our pipeline generates adaptive central text regions and corresponding expanding ratios with a proposed training strategy, followed by a new proposed post-processing algorithm which expands central text regions to the complete text instance with the corresponding expanding ratios. We demonstrated that our new text region representation is effective, and that the pipeline can precisely detect closely adjacent text instances of arbitrary shapes. Experimental results on common datasets demonstrate superior performance of our work.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2020.2999069