
Semantic Text Segmentation from Synthetic Images of Full-Text Documents
Author(s) -
Lukáš Bureš,
Ivan Gruber,
Petr Neduchal,
Miroslav Hlaváč,
Marek Hrúz
Publication year - 2019
Publication title -
trudy spiiran
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.242
H-Index - 9
eISSN - 2078-9599
pISSN - 2078-9181
DOI - 10.15622/sp.2019.18.6.1381-1406
Subject(s) - computer science , artificial intelligence , optical character recognition , segmentation , test set , pattern recognition (psychology) , convolutional neural network , set (abstract data type) , character (mathematics) , image (mathematics) , computer vision , programming language , geometry , mathematics
An algorithm (divided into multiple modules) for generating images of full-text documents is presented. These images can be used to train, test, and evaluate models for Optical Character Recognition (OCR).
The algorithm is modular, individual parts can be changed and tweaked to generate desired images. A method for obtaining background images of paper from already digitized documents is described. For this, a novel approach based on Variational AutoEncoder (VAE) to train a generative model was used.
These backgrounds enable the generation of similar background images as the training ones on the fly.The module for printing the text uses large text corpora, a font, and suitable positional and brightness character noise to obtain believable results (for natural-looking aged documents).
A few types of layouts of the page are supported. The system generates a detailed, structured annotation of the synthesized image. Tesseract OCR to compare the real-world images to generated images is used. The recognition rate is very similar, indicating the proper appearance of the synthetic images. Moreover, the errors which were made by the OCR system in both cases are very similar. From the generated images, fully-convolutional encoder-decoder neural network architecture for semantic segmentation of individual characters was trained. With this architecture, the recognition accuracy of 99.28% on a test set of synthetic documents is reached.