JACIII Vol.28 No.1 pp. 103-110
doi: 10.20965/jaciii.2024.p0103


Burnt-in Text Recognition from Medical Imaging Modalities: Existing Machine Learning Practices

Efosa Osagie, Wei Ji, and Na Helian ORCID Icon

Department of Computer Science, University of Hertfordshire
College Lane Campus, Hatfield, Hertfordshire AL 9, United Kingdom

Corresponding author

February 2, 2023
August 17, 2023
January 20, 2024
medical image character recognition, OCR challenges, burned-in text, medical imaging, medical image processing

In recent times, medical imaging has become a significant component of clinical diagnosis and examinations to detect and evaluate various medical conditions. The interpretation of these medical examinations and the patient’s demographics are usually textual data, which is burned in on the pixel content of medical imaging modalities (MIM). Example of these MIM includes ultrasound and X-ray imaging. As artificial intelligence advances for medical applications, there is a high demand for the accessibility of these burned-in textual data for various needs. This article aims to review the significance of burned-in textual data recognition in MIM and recent research regarding the machine learning approach, challenges, and open issues for further investigation on this application. The review describes the significant problems in this study area as low resolution and background interference of textual data. Finally, the review suggests applying more advanced deep learning ensemble algorithms as possible solutions.

Medical image character recognition using deep learning

Medical image character recognition using deep learning

Cite this article as:
E. Osagie, W. Ji, and N. Helian, “Burnt-in Text Recognition from Medical Imaging Modalities: Existing Machine Learning Practices,” J. Adv. Comput. Intell. Intell. Inform., Vol.28 No.1, pp. 103-110, 2024.
Data files:
  1. [1] K. Baskar, “A survey on feature selection techniques in medical image processing,” Computer Science, 2018. [Accessed November 6, 2022]
  2. [2] J. Z. Wang, “Security filtering of medical images using OCR,” Proc. of the 3rd All-Russian Scientific Conf., pp. 118-122, 2001.
  3. [3] T. Davenport and R. Kalakota, “The potential for artificial intelligence in healthcare,” Future Healthcare J., Vol.6, No.2, pp. 94-98, 2019.
  4. [4] C. B. Collin et al., “Computational models for clinical applications in personalized medicine—guidelines and recommendations for data integration and model validation,” J. of Personalized Medicine, Vol.12, No.2, Article No.166, 2022.
  5. [5] V. Ehrenstein et al., “Obtaining data from electronic health records,” R. E. Gliklich, M. B. Leavy, and N. A. Dreyer (Eds.), “Tools and Technologies for Registry Interoperability, Registries for Evaluating Patient Outcomes: A User’s Guide, 3rd Edition, Addendum 2,” Agency for Healthcare Research and Quality, 2019. [Accessed November 12, 2022]
  6. [6] A. Paulsen, S. Overgaard, and J. M. Lauritsen, “Quality of data entry using single entry, double entry and automated forms processing—an example based on a study of patient-reported outcomes,” PLOS ONE, Vol.7, No.4, Article No.e35087, 2012.
  7. [7] R. J. McDonald et al., “The effects of changes in utilization and technological advancements of cross-sectional imaging on radiologist workload,” Academic Radiology, Vol.22, No.9, pp. 1191-1198, 2015.
  8. [8] B. Bergeron, “Clinical data capture: OMR and OCR and your flatbed scanner,” MedGenMed, Vol.7, No.2, Article No.66, 2005.
  9. [9] R. Rabiei et al., “Prediction of breast cancer using machine learning approaches,” J. of Biomedical Physics and Engineering, Vol.12, No.3, pp. 297-308, 2022.
  10. [10] Y. Amethiya et al., “Comparative analysis of breast cancer detection using machine learning and biosensors,” Intelligent Medicine, Vol.2, No.2, pp. 69-81, 2022.
  11. [11] W. Newhauser et al., “Anonymization of DICOM electronic medical records for radiation therapy,” Computers in Biology and Medicine, Vol.53, pp. 134-140, 2014.
  12. [12] E. Menasalvas and C. Gonzalo-Martin, “Challenges of medical text and image processing: Machine learning approaches,” A. Holzinger (Ed.), “Machine Learning for Health Informatics: State-of-the-Art and Future Challenges,” pp. 221-242, Springer, 2016.
  13. [13] H. Kawano et al., “Structure extraction from decorated characters by graph spectral decomposition and component selection criterion,” J. Adv. Comput. Intell. Intell. Inform., Vol.14, No.2, pp. 179-184, 2010.
  14. [14] H. Miyao et al., “Printed Japanese character recognition using multiple commercial OCRs,” J. Adv. Comput. Intell. Intell. Inform., Vol.8, No.2, pp. 200-207, 2004.
  15. [15] J. Park et al., “Multi-lingual optical character recognition system using the reinforcement learning of character segmenter,” IEEE Access, Vol.8, pp. 174437-174448, 2020.
  16. [16] P. Vcelak et al., “Identification and classification of DICOM files with burned-in text content,” Int. J. of Medical Informatics, Vol.126, pp. 128-137, 2019.
  17. [17] K. Mohsenzadegan, V. Tavakkoli, and K. Kyamakya, “Deep neural network concept for a blind enhancement of document-images in the presence of multiple distortions,” Applied Sciences, Vol.12, No.19, Article No.9601, 2022.
  18. [18] G. K. Tsui and T. Chan, “Automatic selective removal of embedded patient information from image content of DICOM files,” American J. of Roentgenology, Vol.198, No.4, pp. 769-772, 2012.
  19. [19] E. Monteiro, C. Costa, and J. L. Oliveira, “A machine learning methodology for medical imaging anonymization,” 2015 37th Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 1381-1384, 2015.
  20. [20] Y. Ma and Y. Wang, “Text detection in medical images using local feature extraction and supervised learning,” 2015 12th Int. Conf. on Fuzzy Systems and Knowledge Discovery (FSKD), pp. 953-958, 2015.
  21. [21] C. Reul et al., “Expectation-driven text extraction from medical ultrasound images,” A. Hoerbst et al. (Eds.), “Exploring Complexity in Health: An Interdisciplinary Systems Approach,” pp. 712-716, IOS Press, 2016.
  22. [22] E. Monteiro, C. Costa, and J. L. Oliveira, “A de-identification pipeline for ultrasound medical images in DICOM format,” J. of Medical Systems, Vol.41, No.5, Article No.89, 2017.
  23. [23] J. M. Silva et al., “Controlled searching in reversibly de-identified medical imaging archives,” J. of Biomedical Informatics, Vol.77, pp. 81-90, 2018.
  24. [24] X. Xu, W. Wang, and Q. Liu, “Medical image character recognition based on multi-scale neural convolutional network,” 2021 Int. Conf. on Security, Pattern Analysis, and Cybernetics (SPAC), pp. 408-412, 2021.
  25. [25] M. Antunes, R. Machado, and A. Silva, “Anonymization of burned-in annotations in ultrasound imaging,” Eletrónica e Telecomunicações, Vol.5, No.3, pp. 360-364, 2011.
  26. [26] J. P. Segal and R. Hansen, “Medical images, social media and consent,” Nature Reviews Gastroenterology & Hepatology, Vol.18, No.8, pp. 517-518, 2021.
  27. [27] A. Badano et al., “Consistency and standardization of color in medical imaging: A consensus report,” J. of Digital Imaging, Vol.28, No.1, pp. 41-52, 2015.
  28. [28] M. Kociołek, M. Strzelecki, and R. Obuchowicz, “Does image normalization and intensity resolution impact texture classification?,” Computerized Medical Imaging and Graphics, Vol.81, Article No.101716, 2020.
  29. [29] L. Maier-Hein et al., “Why rankings of biomedical image analysis competitions should be interpreted with care,” Nature Communications, Vol.9, No.1, Article No.5217, 2018.
  30. [30] M. Aljabri et al., “Towards a better understanding of annotation tools for medical imaging: A survey,” Multimedia Tools and Applications, Vol.81, No.18, pp. 25877-25911, 2022.
  31. [31] Y. Li, B. Sixou, and F. Peyrin, “A review of the deep learning methods for medical images super resolution problems,” IRBM, Vol.42, No.2, pp. 120-133, 2021.
  32. [32] H. Michalak and K. Okarma, “Improvement of image binarization methods using image preprocessing with local entropy filtering for alphanumerical character recognition purposes,” Entropy, Vol.21, No.6, Article No.562, 2019.
  33. [33] W. Bieniecki, S. Grabowski, and W. Rozenberg, “Image preprocessing for improving OCR accuracy,” 2007 Int. Conf. on Perspective Technologies and Methods in MEMS Design, pp. 75-80, 2007.
  34. [34] S. Nomura et al., “Morphological preprocessing method to thresholding degraded word images,” Pattern Recognition Letters, Vol.30, No.8, pp. 729-744, 2009.
  35. [35] J. M. D. Delgado and L. Oyedele, “Deep learning with small datasets: Using autoencoders to address limited datasets in construction management,” Applied Soft Computing, Vol.112, Article No.107836, 2021.
  36. [36] M. Li, R. Poovendran, and S. Narayanan, “Protecting patient privacy against unauthorized release of medical images in a group communication environment,” Computerized Medical Imaging and Graphics, Vol.29, No.5, pp. 367-383, 2005.
  37. [37] X. Qin, F. M. Bui, and H. H. Nguyen, “Learning from an imbalanced and limited dataset and an application to medical imaging,” 2019 IEEE Pacific Rim Conf. on Communications, Computers and Signal Processing (PACRIM), 2019.
  38. [38] D. Pal et al., “MSHSCNN: Multi-scale hybrid-Siamese network to differentiate visually similar character classes,” 2021 9th European Workshop on Visual Information Processing (EUVIP), 2021.
  39. [39] P. Inkeaw et al., “Recognition of similar characters using gradient features of discriminative regions,” Expert Systems with Applications, Vol.134, pp. 120-137, 2019.
  40. [40] A. Safaei, “Text-based multi-dimensional medical images retrieval according to the features-usage correlation,” Medical & Biological Engineering & Computing, Vol.59, No.10, pp. 1993-2017, 2021.
  41. [41] D. Cireşan, U. Meier, and J. Schmidhuber, “Multi-column deep neural networks for image classification,” arXiv: 1202.2745, 2012.
  42. [42] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Communications of the ACM, Vol.60, No.6, pp. 84-90, 2017.
  43. [43] L. Lam and S. Y. Suen, “Application of majority voting to pattern recognition: An analysis of its behavior and performance,” IEEE Trans. on Systems, Man, and Cybernetics – Part A: Systems and Humans, Vol.27, No.5, pp. 553-568, 1997.
  44. [44] J. Shlens, “A tutorial on principal component analysis,” arXiv: 1404.1100, 2014.
  45. [45] Z. M. Kovács-V, “A novel architecture for high quality hand-printed character recognition,” Pattern Recognition, Vol.28, No.11, pp. 1685-1692, 1995.

*This site is desgined based on HTML5 and CSS3 for modern browsers, e.g. Chrome, Firefox, Safari, Edge, Opera.

Last updated on Feb. 19, 2024