TY - GEN
T1 - A New Semi-automatic Annotation Model via Semantic Boundary Estimation for Scene Text Detection
AU - Zhuang, Zhenzhou
AU - Liu, Zonghao
AU - Lam, Kin Man
AU - Huang, Shuangping
AU - Dai, Gang
N1 - Funding Information:
Acknowledgements. This research is supported in part by Guangdong Basic and Applied Basic Research Foundation (No. 2021A1515012282) and the Alibaba Innovative Research (AIR) Program.
Publisher Copyright:
© 2021, Springer Nature Switzerland AG.
PY - 2021/9
Y1 - 2021/9
N2 - Manually annotating a data set for scene text detection is extremely time-consuming. In this paper, we propose a new semi-automatic annotation model to produce tight polygonal annotations for text instances in scene images, based on the input of manually annotated text center lines. Our approach first generates multiple candidate boundaries, which share the same input center line. Then, by training a fastidious content recognizer, optimal boundary selection is performed. The bounded text region, which achieves the smallest recognition loss, is selected as the tightest of the text. As this optimal boundary estimation is guided by semantic recognition, our method is called Semantic Boundary Estimation. Experiment results show that only half clicks compared to manually annotated polygon, are input to annotate center line, and precise polygon text region annotation is automatically produced. A high recall of more than 95% at IoU > 0.5 and 80% at IoU > 0.7 is achieved, demonstrating the high agreement with the original ground truth. In addition, using the generated annotations on benchmarks, such as Total-Text, CTW1500 and ICDAR2015, to train state-of-the-art detectors can achieve similar performance to those trained with manual annotations. This further verifies the good annotation performance. A annotation toolkit based on the proposed model is available at CenterlineAnnotation.
AB - Manually annotating a data set for scene text detection is extremely time-consuming. In this paper, we propose a new semi-automatic annotation model to produce tight polygonal annotations for text instances in scene images, based on the input of manually annotated text center lines. Our approach first generates multiple candidate boundaries, which share the same input center line. Then, by training a fastidious content recognizer, optimal boundary selection is performed. The bounded text region, which achieves the smallest recognition loss, is selected as the tightest of the text. As this optimal boundary estimation is guided by semantic recognition, our method is called Semantic Boundary Estimation. Experiment results show that only half clicks compared to manually annotated polygon, are input to annotate center line, and precise polygon text region annotation is automatically produced. A high recall of more than 95% at IoU > 0.5 and 80% at IoU > 0.7 is achieved, demonstrating the high agreement with the original ground truth. In addition, using the generated annotations on benchmarks, such as Total-Text, CTW1500 and ICDAR2015, to train state-of-the-art detectors can achieve similar performance to those trained with manual annotations. This further verifies the good annotation performance. A annotation toolkit based on the proposed model is available at CenterlineAnnotation.
KW - Scene text detection
KW - Semantic boundary estimation
KW - Semi-automatic annotation algorithm
UR - http://www.scopus.com/inward/record.url?scp=85115328506&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-86334-0_17
DO - 10.1007/978-3-030-86334-0_17
M3 - Conference article published in proceeding or book
AN - SCOPUS:85115328506
SN - 9783030863333
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 257
EP - 273
BT - Document Analysis and Recognition - ICDAR 2021 - 16th International Conference, Proceedings
A2 - Lladós, Josep
A2 - Lopresti, Daniel
A2 - Uchida, Seiichi
PB - Springer Science and Business Media Deutschland GmbH
T2 - 16th International Conference on Document Analysis and Recognition, ICDAR 2021
Y2 - 5 September 2021 through 10 September 2021
ER -