학술논문

Distilling GRU with Data Augmentation for Unconstrained Handwritten Text Recognition
Document Type
Conference
Source
2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR) ICFHR Frontiers in Handwriting Recognition (ICFHR), 2018 16th International Conference on. :56-61 Aug, 2018
Subject
Computing and Processing
Text recognition
Trajectory
Feature extraction
Training
Logic gates
Handwriting recognition
Acceleration
unconstrained
text recognition
data augmentation
rnn
Language
Abstract
Handwritten texts with various styles, such as horizontal, overlapping, vertical, and multi-lines texts, are commonly observed in the community. However, most existing handwriting recognition methods only concentrate on one specific kind of text style. In this paper, we focus on the problem of new unconstrained handwritten text recognition and propose distilling gated recurrent unit (GRU) with a new data augmentation technology to model the complex sequential dynamic of unconstrained handwriting text of various styles. The proposed data augmentation method can synthesize realistic handwritten text datasets including horizontal, vertical, overlap, right-down, screw-rotation, and multi-line situation, which render our framework robust for general purposes. The recommended distilling GRU can not only accelerate the training speed through the distilling stage but also maintain the original recognition accuracy. Experiments on our synthesized handwritten test sets show that the proposed multi-layer GRU performs well on the unconstrained handwriting text recognition problem. On the ICDAR2013 handwritten text recognition benchmark dataset, the proposed framework demonstrates comparable performance with state-of-the-art techniques.