Digit Recognition Using Composite Features With Decision Tree Strategy.

Authors

  • Chung Hsing Chen National Kaohsiung University of Science and Technology image/svg+xml
  • Ko Wei Huang National Kaohsiung University of Science and Technology image/svg+xml

DOI:

https://doi.org/10.9781/ijimai.2022.12.001

Keywords:

Decision Tree, E13B Fonts, Feature Extraction, Image Classification, Multilayer Perceptron
Supporting Agencies
Plustek Inc. provided the samples for this study at no cost. We thank Bob Lin, General Manager at Plustek Inc., for his constant support, and ADView Technology for providing an Nvidia GPU for E13B model training, making this study possible. In addition, the identification framework of this study has obtained the Republic of China Patent No. M617631. The Plustek iKnow application software and SDK have been imported, which can be used by other software developers for authorization. This work was supported in part by the Ministry of Science and Technology, Taiwan, R.O.C., under the grant ID: MOST 110-2222-E-992-006-.

Abstract

At present, check transactions are one of the most common forms of money transfer in the market. The information for check exchange is printed using magnetic ink character recognition (MICR), widely used in the banking industry, primarily for processing check transactions. However, the magnetic ink card reader is specialized and expensive, resulting in general accounting departments or bookkeepers using manual data registration instead. An organization that deals with parts or corporate services might have to process 300 to 400 checks each day, which would require a considerable amount of labor to perform the registration process. The cost of a single-sided scanner is only 1/10 of the MICR; hence, using image recognition technology is an economical solution. In this study, we aim to use multiple features for character recognition of E13B, comprising ten numbers and four symbols. For the numeric part, we used statistical features such as image density features, geometric features, and simple decision trees for classification. The symbols of E13B are composed of three distinct rectangles, classified according to their size and relative position. Using the same sample set, MLP, LetNet-5, Alexnet, and hybrid CNN-SVM were used to train the numerical part of the artificial intelligence network as the experimental control group to verify the accuracy and speed of the proposed method. The results of this study were used to verify the performance and usability of the proposed method. Our proposed method obtained all test samples correctly, with a recognition rate close to 100%. A prediction time of less than one millisecond per character, with an average value of 0.03 ms, was achieved, over 50 times faster than state-of-the-art methods. The accuracy rate is also better than all comparative state-of-the-art methods. The proposed method was also applied to an embedded device to ensure the CPU would be used for verification instead of a high-end GPU.

Downloads

Download data is not yet available.

References

L. Deng, “The mnist database of handwritten digit images for machine learning research [best of the web],” IEEE signal processing magazine, vol. 29, no. 6, pp. 141–142, 2012.

X. Corporation, “Generic micr fundamentals guide,” Xerox Corporation, 2012.

A. Choudhary, S. Ahlawat, R. Rishi, “A binarization feature extraction approach to ocr: Mlp vs. rbf,” in International Conference on Distributed Computing and Internet Technology, 2014, pp. 341–346, Springer.

I. B. Cruz, A. Díaz Sardiñas, R. Bello Pérez, Y. Sardiñas Oliva, “Learning optimization in a mlp neural network applied to ocr,” in Mexican International Conference on Artificial Intelligence, 2002, pp. 292–300, Springer.

A. Choudhary, R. Rishi, S. Ahlawat, “Off-line handwritten character recognition using features extracted from binarization technique,” Aasri Procedia, vol. 4, pp. 306–312, 2013.

A. F. Agarap, “An architecture combining convolutional neural network and support vector machine for image classification,” arXiv preprint arXiv:1712.03541, 2017.

Y. LeCun, L. Bottou, Y. Bengio, P. Haffner, “Gradient- based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 1998.

Z. Zhong, L. Jin, Z. Xie, “High performance offline handwritten chinese character recognition using googlenet and directional feature maps,” in 2015 13th International Conference on Document Analysis and Recognition, 2015, pp. 846–850, IEEE.

C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, A. Rabinovich, “Going deeper with convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 1–9.

K. Simonyan, A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.

A. Krizhevsky, I. Sutskever, G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Advances in neural information processing systems, vol. 25, 2012.

N. Sharma, B. Kumar, V. Singh, “Recognition of off- line hand printed english characters, numerals and special symbols,” in 2014 5th International Conference- Confluence The Next Generation Information Technology Summit, 2014, pp. 640–645, IEEE.

I. O. for Standardization, “Information processing — magnetic ink character recognition — part 1: Print specifications for e13b,” International Organization for Standardization, 2018.

Y. Yang, X. Lijia, C. Chen, “English character recognition based on feature combination,” Procedia Engineering, vol. 24, pp. 159–164, 2011.

M. Rani, Y. K. Meena, “An efficient feature extraction method for handwritten character recognition,” in International Conference on Swarm, Evolutionary, and Memetic Computing, 2011, pp. 302–309, Springer.

S. B. Moussa, A. Zahour, A. Benabdelhafid, A. M. Alimi, “New features using fractal multi-dimensions for generalized arabic font recognition,” Pattern Recognition Letters, vol. 31, no. 5, pp. 361–371, 2010.

H. Bay, T. Tuytelaars, L. V. Gool, “Surf: Speeded up robust features,” in European conference on computer vision, 2006, pp. 404–417, Springer.

L. Wang, S. Bi, X. Lu, Y. Gu, C. Zhai, “Deformation measurement of highspeed rotating drone blades based on digital image correlation combined with ring projection transform and orientation codes,” Measurement, vol. 148, p. 106899, 2019.

K. K. Shreyas, S. Rajeev, K. Panetta, S. S. Agaian, “Fingerprint authentication using geometric features,” in 2017 IEEE International Symposium on Technologies for Homeland Security, 2017, pp. 1–7, IEEE.

T. Kobayashi, “Bfo meets hog: feature extraction based on histograms of oriented pdf gradients for image classification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2013, pp. 747–754.

J. R. Quinlan, “Induction of decision trees,” Machine learning, vol. 1, no. 1, pp. 81–106, 1986.

J.-M. Park, C. G. Looney, H.-C. Chen, “Fast connected component labeling algorithm using a divide and conquer technique,” Computers and Their Applications, vol. 4, no. 20, p. 0, 2000.

F. Kimura, M. Shridhar, “Handwritten numerical recognition based on multiple algorithms,” Pattern recognition, vol. 24, no. 10, pp. 969–983, 1991.

P. Singh, S. Budhiraja, “Feature extraction and classification techniques in ocr systems for handwritten gurmukhi script–a survey,” International Journal of Engineering Research and Applications, vol. 1, no. 4, pp. 1736–1739, 2011.

R. Verma, D. J. Ali, “A-survey of feature extraction and classification techniques in ocr systems,” International Journal of Computer Applications & Information Technology, vol. 1, no. 3, pp. 1–3, 2012.

D. Varshni, K. Thakral, L. Agarwal, R. Nijhawan, A. Mittal, “Pneumonia detection using cnn based feature extraction,” in 2019 IEEE international conference on electrical, computer and communication technologies, 2019, pp. 1–7, IEEE.

A. Yang, X. Yang, W. Wu, H. Liu, Y. Zhuansun, “Research on feature extraction of tumor image based on convolutional neural network,” IEEE access, vol. 7, pp. 24204–24213, 2019.

G. S. Lehal, “Optical character recognition of gurmukhi script using multiple classifiers,” in Proceedings of the international workshop on multilingual OCR, 2009, pp. 1–9.

T. Kobayashi, A. Hidaka, T. Kurita, “Selection of histograms of oriented gradients features for pedestrian detection,” in International conference on neural information processing, 2007, pp. 598–607, Springer.

S. Singh, A. Aggarwal, R. Dhir, “Use of gabor filters for recognition of handwritten gurmukhi character,” International Journal of Advanced Research in Computer Science and Software Engineering, vol. 2, no. 5, 2012.

A. Shawon, M. J.-U. Rahman, F. Mahmud, M. A. Zaman, “Bangla handwritten digit recognition using deep cnn for large and unbiased dataset,” in 2018 International Conference on Bangla Speech and Language Processing, 2018, pp. 1–6, IEEE.

V. Rajinikanth, S. Kadry, R. González-Crespo, E. Verdú, “A study on RGB image multi-thresholding using kapur/tsallis entropy and moth-flame algorithm,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 5, no. 2, 163-171, 2021.

S. Acharya, A. K. Pant, P. K. Gyawali, “Deep learning based large scale handwritten devanagari character recognition,” in 2015 9th International conference on software, knowledge, information management and applications, 2015, pp. 1–6, IEEE.

I. Ramadhan, P. Sukarno, M. A. Nugroho, “Comparative analysis of k-nearest neighbor and decision tree in detecting distributed denial of service,” in 2020 8th International Conference on Information and Communication Technology, 2020, pp. 1–4, IEEE.

T. A. Assegie, P. S. Nair, “Handwritten digits recognition with decision tree classification: a machine learning approach,” International Journal of Electrical and Computer Engineering, vol. 9, no. 5, pp. 4446–4451, 2019.

S. Ahlawat, A. Choudhary, “Hybrid cnn-svm classifier for handwritten digit recognition,” Procedia Computer Science, vol. 167, pp. 2554–2560, 2020.

A. A. Barbhuiya, R. K. Karsh, R. Jain, “Cnn based feature extraction and classification for sign language,” Multimedia Tools and Applications, vol. 80, no. 2, pp. 3051–3069, 2021.

V. Dogra, S. Verma, N. Jhanjhi, U. Ghosh, D.-N. Le, et al., “A comparative analysis of machine learning models for banking news extraction by multiclass classification with imbalanced datasets of financial news: Challenges and solutions.,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 7, no. 3, 2022.

M. Khari, A. K. Garg, R. G. Crespo, E. Verdú, “Gesture recognition of RGB and RGB-D static images using convolutional neural networks.,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 5, no. 7, pp. 22–27, 2019.

J. D. Rodriguez, A. Perez, J. A. Lozano, “Sensitivity analysis of k-fold cross validation in prediction error estimation,” IEEE transactions on pattern analysis and machine intelligence, vol. 32, no. 3, pp. 569–575, 2009.

Downloads

Published

2023-06-01
Metrics
Views/Downloads
  • Abstract
    146
  • PDF
    114

How to Cite

Hsing Chen, C. and Wei Huang, K. (2023). Digit Recognition Using Composite Features With Decision Tree Strategy. International Journal of Interactive Multimedia and Artificial Intelligence, 8(2), 98–107. https://doi.org/10.9781/ijimai.2022.12.001