single-jc.php

JACIII Vol.28 No.4 pp. 835-844
doi: 10.20965/jaciii.2024.p0835
(2024)

Research Paper:

Gradually Vanishing Bridge Based on Multi-Kernel Maximum Mean Discrepancy for Breast Ultrasound Image Classification

Bo Xu* ORCID Icon, Cuier Tan* ORCID Icon, Ying Wu**,***,† ORCID Icon, and Faming Li*

*School of Information, Guangdong University of Finance and Economics
No.21 Luntou Road, Haizhu District, Guangzhou, Guangdong 510320, China

**Department of Ultrasound, The First Affiliated Hospital of Jinan University
No.613 West Huangpu Avenue, Tianhe District, Guangzhou, Guangdong 510320, China

***Department of Ultrasound, Chaoshan Hospital, The First Affiliated Hospital of Jinan University
Chaozhou, Guangdong 515799, China

Corresponding author

Received:
November 12, 2023
Accepted:
February 27, 2024
Published:
July 20, 2024
Keywords:
breast ultrasound image, deep transfer learning, unsupervised domain adaptation, adversarial domain adaptive
Abstract

This study seeks to enhance the classification performance of breast ultrasound images, addressing the challenges of difficult and costly collection of breast ultrasound datasets as well as the discrepancies in feature distribution of the collected datasets. Performance is enhanced by using a mix of generative adversarial networks (GAN) and domain adaptive networks. First, an adaptive layer is first added to the basic model of the gradually vanishing bridge (GVB), to better match the feature distributions of the source and target domains of the dataset. The multi-kernel maximum mean discrepancy (MK-MMD), which is the most efficient of existing adaptive approaches, is implemented in the fully connected layer of the original model’s feature extraction network. Finally, through the process of fine-tuning, the model that has the highest level of overall performance is determined. In experiments, the proposed method surpassed the conventional unsupervised domain adaptation (DDC) and adversarial domain adaptation (MK_DAAN, GVB) in performance, achieving 85.11% accuracy, 97.48% recall, and 0.92 F1-score.

Cite this article as:
B. Xu, C. Tan, Y. Wu, and F. Li, “Gradually Vanishing Bridge Based on Multi-Kernel Maximum Mean Discrepancy for Breast Ultrasound Image Classification,” J. Adv. Comput. Intell. Intell. Inform., Vol.28 No.4, pp. 835-844, 2024.
Data files:
References
  1. [1] H. Sung et al., “Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries,” CA: A Cancer J. for Clinicians, Vol.71, No.3, pp. 209-249, 2021. https://doi.org/10.3322/caac.21660
  2. [2] S. Lei et al., “Breast cancer incidence and mortality in women in China: temporal trends and projections to 2030,” Cancer Biology & Medicine, Vol.18, No.3, pp. 900-909, 2021. https://doi.org/10.20892/j.issn.2095-3941.2020.0523
  3. [3] K. D. Marcomini, E. F. C. Fleury, H. Schiabel, and R. M. Nishikawa, “Proposal of semi-automatic classification of breast lesions for strain sonoelastography using a dedicated CAD system,” Proc. of 13th Int. Workshop on Breast Imaging (IWDM 2016), pp. 454-460, 2016. https://doi.org/10.1007/978-3-319-41546-8_57
  4. [4] Y. Yu, “Advances in application of artificial intelligence in medical image analysis,” Chinese J. of Medical Imaging Technology, pp. 1808-1812, 2019.
  5. [5] I. Goodfellow et al., “Generative adversarial networks,” Communications of the ACM, Vol.63, No.11, pp. 139-144, 2020. https://doi.org/10.1145/3422622
  6. [6] S. Cui, S. Wang, J. Zhuo, C. Su, Q. Huang, and Q. Tian, “Gradually vanishing bridge for adversarial domain adaptation,” Proc. of 2020 IEEE/CVF Conf. on Computer Vision and Pattern Recognition, pp. 12455-12464, 2020. https://doi.org/10.48550/arXiv.2003.13183
  7. [7] S. Han et al., “A deep learning framework for supporting the classification of breast lesions in ultrasound images,” Physics in Medicine & Biology, Vol.62, No.19, pp. 7714-7728, 2017. https://doi.org/10.1088/1361-6560/aa82ec
  8. [8] C. Szegedy et al., “Going deeper with convolutions,” Proc. of 2015 IEEE Conf. on Computer Vision and Pattern Recognition, 2015. https://doi.org/10.1109/CVPR.2015.7298594
  9. [9] A. Rakhlin, A. Shvets, V. Iglovikov, and A. A. Kalinin, “Deep convolutional neural networks for breast cancer histology image analysis,” 15th Int. Conf. on Image Analysis and Recognition (ICIAR 2018), pp. 737-744, 2018. https://doi.org/10.1007/978-3-319-93000-8_83
  10. [10] T. Xiao, L. Liu, K. Li, W. Qin, S. Yu, and Z. Li, “Comparison of transferred deep neural networks in ultrasonic breast masses discrimination,” BioMed Research Int., Vol.2018, Article No.4605191, 2018. https://doi.org/10.1155/2018/4605191
  11. [11] K. He et al., “Deep residual learning for image recognition,” Proc. of 2016 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 770-778, 2016. https://doi.org/10.1109/CVPR.2016.90
  12. [12] C. Szegedy et al., “Rethinking the inception architecture for computer vision,” Proc. of 2016 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 2818-2826, 2016. https://doi.org/10.1109/CVPR.2016.308
  13. [13] F. Chollet, “Xception: Deep learning with depthwise separable convolutions,” Proc. of 2017 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 1800-1807, 2017. https://doi.ieeecomputersociety.org/10.1109/CVPR.2017.195
  14. [14] S.-W. Chen, Y.-J. Liu, D. Liu, C. Su, D. Zhao, L.-X. Lin, and P.-H. Zhang, “AlexNet model and adaptive contrast enhancement based ultrasound imaging classification,” Computer Science, Vol.46, Issue 6A, pp. 146-152, 2019.
  15. [15] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet Classification With Deep Convolutional Neural Networks,” Advances in Neural Information Processing Systems, Vol.25, 2012.
  16. [16] H. Zhang, L. Han, K. Chen, Y. Peng, and J. Lin, “Diagnostic efficiency of the breast ultrasound computer-aided prediction model based on convolutional neural network in breast cancer,” J. of Digital Imaging, Vol.33, pp. 1218-1223, 2020. https://doi.org/10.1007/s10278-020-00357-7
  17. [17] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint, arXiv:1409.1556, 2014. https://doi.org/10.48550/arXiv.1409.1556
  18. [18] J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “How transferable are features in deep neural networks?,” Advances in Neural Information Processing Systems, Vol.27, 2014. https://doi.org/10.48550/arXiv.1411.1792
  19. [19] E. Tzeng, J. Hoffman, N. Zhang, K. Saenko, and T. Darrell, “Deep domain confusion: Maximizing for domain invariance,” arXiv preprint, arXiv:1412.3474, 2014. https://doi.org/10.48550/arXiv.1412.3474
  20. [20] M. Long, Y. Cao, J. Wang, and M. Jordan, “Learning transferable features with deep adaptation networks,” Proc. of Int. Conf. on Machine Learning Research (PMLR), pp. 97-105, 2015.
  21. [21] A. Gretton et al., “Optimal kernel choice for large-scale two-sample tests,” Advances in Neural Information Processing Systems, Vol.25, 2012.
  22. [22] F. Zhuang, X. Cheng, P. Luo, S. J. Pan, and Q. He, “Supervised representation learning: Transfer learning with deep autoencoders,” 24th Int. Conf. on Artificial Intelligence, pp. 4119-4125, 2015.
  23. [23] M. Long, J. Wang, Y. Cao, J. Sun, and P. S. Yu, “Deep learning of transferable representation for scalable domain adaptation,” IEEE Trans. on Knowledge and Data Engineering, Vol.28, No.8, pp. 2027-2040, 2016. https://doi.org/10.1109/TKDE.2016.2554549
  24. [24] B. Sun and K. Saenko, “Deep CORAL: Correlation Alignment for Deep Domain Adaptation,” arXiv preprint, arXiv:1607.01719, 2016. https://doi.org/10.48550/arXiv.1607.01719
  25. [25] P. Wei, Y. Ke, and C. K. Goh, “Deep nonlinear feature coding for unsupervised domain adaptation,” Proc. of the 25th Int. Joint Conf. on Artificial Intelligence (IJCAI 2016), pp. 2189-2195, 2016.
  26. [26] J. Zhuo, S. Wang, W. Zhang, and Q. Huang, “Deep unsupervised convolutional domain adaptation,” Proc. of the 25th ACM Int. Conf. on Multimedia, pp. 261-269, 2017. https://doi.org/10.1145/3123266.3123292
  27. [27] Z. Luo, Y. Zou, J. Hoffman, and L. Fei-Fei, “Label efficient learning of transferable representations across domains and tasks,” Advances in Neural Information Processing Systems, Vol.30, 2017. https://doi.org/10.48550/arXiv.1712.00123
  28. [28] Y. Ganin et al., “Domain-adversarial training of neural networks,” J. of Machine Learning Research, Vol.17, No.1, pp. 2030-2096, 2016.
  29. [29] K. Bousmalis, G. Trigeorgis, N. Silberman, D. Krishnan, and D. Erhan, “Domain separation networks,” Advances in Neural Information Processing Systems, Vol.29, 2016. https://doi.org/10.48550/arXiv.1608.06019
  30. [30] E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation,” Proc. of 2017 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 2962-2971, 2017. https://doi.org/10.1109/CVPR.2017.316
  31. [31] C. Yu, J. Wang, Y. Chen, and M. Huang, “Transfer learning with dynamic adversarial adaptation network,” 2019 IEEE Int. Conf. on Data Mining (ICDM), pp. 778-786, 2019. https://doi.org/10.1109/ICDM.2019.00088
  32. [32] B. Xu, K. Wu, Y. Wu, J. He, and C. Chen, “Dynamic adversarial domain adaptation based on multikernel maximum mean discrepancy for breast ultrasound image classification,” Expert Systems with Applications, Vol.207, Article No.117978, 2022. https://doi.org/10.1016/j.eswa.2022.117978
  33. [33] K. M. Borgwardt et al., “Integrating structured biological data by kernel maximum mean discrepancy,” Bioinformatics, Vol.22, Issue 14, pp. e49-e57, 2006. https://doi.org/10.1093/bioinformatics/btl242
  34. [34] M. H. Yap et al., “Automated breast ultrasound lesions detection using convolutional neural networks,” IEEE J. of Biomedical and Health Informatics, Vol.22, No.4, pp. 1218-1226, 2018.

*This site is desgined based on HTML5 and CSS3 for modern browsers, e.g. Chrome, Firefox, Safari, Edge, Opera.

Last updated on Sep. 09, 2024