TY - GEN
T1 - How Does a CNN Manage Different Printing Types?
AU - Ide, Shota
AU - Uchida, Seiichi
N1 - Funding Information:
ACKNOWLEDGMENT This research was partially supported by MEXT-Japan (Grant No. 26240024 and 17H06100) and Kakihara Foundation.
Publisher Copyright:
© 2017 IEEE.
PY - 2017/7/2
Y1 - 2017/7/2
N2 - In past OCR research, different OCR engines are used for different printing types, i.e., machine-printed characters, handwritten characters, and decorated fonts. A recent research, however, reveals that convolutional neural networks (CNN) can realize a universal OCR, which can deal with any printing types without pre-classification into individual types. In this paper, we analyze how CNN for universal OCR manage the different printing types. More specifically, we try to find where a handwritten character of a class and a machine-printed character of the same class are 'fused' in CNN. For analysis, we use two different approaches. The first approach is statistical analysis for detecting the CNN units which are sensitive (or insensitive) to type difference. The second approach is network-based visualization of pattern distribution in each layer. Both analyses suggest the same trend that types are not fully fused in convolutional layers but the distributions of the same class from different types become closer in upper layers.
AB - In past OCR research, different OCR engines are used for different printing types, i.e., machine-printed characters, handwritten characters, and decorated fonts. A recent research, however, reveals that convolutional neural networks (CNN) can realize a universal OCR, which can deal with any printing types without pre-classification into individual types. In this paper, we analyze how CNN for universal OCR manage the different printing types. More specifically, we try to find where a handwritten character of a class and a machine-printed character of the same class are 'fused' in CNN. For analysis, we use two different approaches. The first approach is statistical analysis for detecting the CNN units which are sensitive (or insensitive) to type difference. The second approach is network-based visualization of pattern distribution in each layer. Both analyses suggest the same trend that types are not fully fused in convolutional layers but the distributions of the same class from different types become closer in upper layers.
UR - http://www.scopus.com/inward/record.url?scp=85045222486&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85045222486&partnerID=8YFLogxK
U2 - 10.1109/ICDAR.2017.167
DO - 10.1109/ICDAR.2017.167
M3 - Conference contribution
AN - SCOPUS:85045222486
T3 - Proceedings of the International Conference on Document Analysis and Recognition, ICDAR
SP - 1004
EP - 1009
BT - Proceedings - 14th IAPR International Conference on Document Analysis and Recognition, ICDAR 2017
PB - IEEE Computer Society
T2 - 14th IAPR International Conference on Document Analysis and Recognition, ICDAR 2017
Y2 - 9 November 2017 through 15 November 2017
ER -