A Fast Scene Text Detector Using Knowledge Distillation

Incidental scene text detection is a challenging problem because of arbitrary orientation, low resolution, perspective distortion, and variant aspect ratios of text in natural images. In this paper, we present an end-to-end trainable deep model, which can effectively and efficiently locate multi-ori...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2019, Vol.7, p.22588-22598
Hauptverfasser: Yang, Peng, Zhang, Fanlong, Yang, Guowei
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Incidental scene text detection is a challenging problem because of arbitrary orientation, low resolution, perspective distortion, and variant aspect ratios of text in natural images. In this paper, we present an end-to-end trainable deep model, which can effectively and efficiently locate multi-oriented scene text. Our detector includes a student network and a teacher network, and they inherit complex VGGNet and lightweight PVANet architecture, respectively. While deploying for text detection, the teacher network is used to guide the training process of a student via knowledge distilling so as to maintain the tradeoff between accuracy and efficiency. We have evaluated the proposed detector on three popular benchmarks, and it achieves F-measures of 83.7%, 57.27%, and 90% on ICDAR2015 Incidental Scene Text, COCO-Text, and ICDAR2013, respectively, which outperforms the most state-of-the-art methods.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2019.2895330