Research Paper:
Gradually Vanishing Bridge Based on Multi-Kernel Maximum Mean Discrepancy for Breast Ultrasound Image Classification
Bo Xu* , Cuier Tan* , Ying Wu**,***, , and Faming Li*
*School of Information, Guangdong University of Finance and Economics
No.21 Luntou Road, Haizhu District, Guangzhou, Guangdong 510320, China
**Department of Ultrasound, The First Affiliated Hospital of Jinan University
No.613 West Huangpu Avenue, Tianhe District, Guangzhou, Guangdong 510320, China
***Department of Ultrasound, Chaoshan Hospital, The First Affiliated Hospital of Jinan University
Chaozhou, Guangdong 515799, China
Corresponding author
This study seeks to enhance the classification performance of breast ultrasound images, addressing the challenges of difficult and costly collection of breast ultrasound datasets as well as the discrepancies in feature distribution of the collected datasets. Performance is enhanced by using a mix of generative adversarial networks (GAN) and domain adaptive networks. First, an adaptive layer is first added to the basic model of the gradually vanishing bridge (GVB), to better match the feature distributions of the source and target domains of the dataset. The multi-kernel maximum mean discrepancy (MK-MMD), which is the most efficient of existing adaptive approaches, is implemented in the fully connected layer of the original model’s feature extraction network. Finally, through the process of fine-tuning, the model that has the highest level of overall performance is determined. In experiments, the proposed method surpassed the conventional unsupervised domain adaptation (DDC) and adversarial domain adaptation (MK_DAAN, GVB) in performance, achieving 85.11% accuracy, 97.48% recall, and 0.92 F1-score.
- [1] H. Sung et al., “Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries,” CA: A Cancer J. for Clinicians, Vol.71, No.3, pp. 209-249, 2021. https://doi.org/10.3322/caac.21660
- [2] S. Lei et al., “Breast cancer incidence and mortality in women in China: temporal trends and projections to 2030,” Cancer Biology & Medicine, Vol.18, No.3, pp. 900-909, 2021. https://doi.org/10.20892/j.issn.2095-3941.2020.0523
- [3] K. D. Marcomini, E. F. C. Fleury, H. Schiabel, and R. M. Nishikawa, “Proposal of semi-automatic classification of breast lesions for strain sonoelastography using a dedicated CAD system,” Proc. of 13th Int. Workshop on Breast Imaging (IWDM 2016), pp. 454-460, 2016. https://doi.org/10.1007/978-3-319-41546-8_57
- [4] Y. Yu, “Advances in application of artificial intelligence in medical image analysis,” Chinese J. of Medical Imaging Technology, pp. 1808-1812, 2019.
- [5] I. Goodfellow et al., “Generative adversarial networks,” Communications of the ACM, Vol.63, No.11, pp. 139-144, 2020. https://doi.org/10.1145/3422622
- [6] S. Cui, S. Wang, J. Zhuo, C. Su, Q. Huang, and Q. Tian, “Gradually vanishing bridge for adversarial domain adaptation,” Proc. of 2020 IEEE/CVF Conf. on Computer Vision and Pattern Recognition, pp. 12455-12464, 2020. https://doi.org/10.48550/arXiv.2003.13183
- [7] S. Han et al., “A deep learning framework for supporting the classification of breast lesions in ultrasound images,” Physics in Medicine & Biology, Vol.62, No.19, pp. 7714-7728, 2017. https://doi.org/10.1088/1361-6560/aa82ec
- [8] C. Szegedy et al., “Going deeper with convolutions,” Proc. of 2015 IEEE Conf. on Computer Vision and Pattern Recognition, 2015. https://doi.org/10.1109/CVPR.2015.7298594
- [9] A. Rakhlin, A. Shvets, V. Iglovikov, and A. A. Kalinin, “Deep convolutional neural networks for breast cancer histology image analysis,” 15th Int. Conf. on Image Analysis and Recognition (ICIAR 2018), pp. 737-744, 2018. https://doi.org/10.1007/978-3-319-93000-8_83
- [10] T. Xiao, L. Liu, K. Li, W. Qin, S. Yu, and Z. Li, “Comparison of transferred deep neural networks in ultrasonic breast masses discrimination,” BioMed Research Int., Vol.2018, Article No.4605191, 2018. https://doi.org/10.1155/2018/4605191
- [11] K. He et al., “Deep residual learning for image recognition,” Proc. of 2016 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 770-778, 2016. https://doi.org/10.1109/CVPR.2016.90
- [12] C. Szegedy et al., “Rethinking the inception architecture for computer vision,” Proc. of 2016 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 2818-2826, 2016. https://doi.org/10.1109/CVPR.2016.308
- [13] F. Chollet, “Xception: Deep learning with depthwise separable convolutions,” Proc. of 2017 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 1800-1807, 2017. https://doi.ieeecomputersociety.org/10.1109/CVPR.2017.195
- [14] S.-W. Chen, Y.-J. Liu, D. Liu, C. Su, D. Zhao, L.-X. Lin, and P.-H. Zhang, “AlexNet model and adaptive contrast enhancement based ultrasound imaging classification,” Computer Science, Vol.46, Issue 6A, pp. 146-152, 2019.
- [15] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet Classification With Deep Convolutional Neural Networks,” Advances in Neural Information Processing Systems, Vol.25, 2012.
- [16] H. Zhang, L. Han, K. Chen, Y. Peng, and J. Lin, “Diagnostic efficiency of the breast ultrasound computer-aided prediction model based on convolutional neural network in breast cancer,” J. of Digital Imaging, Vol.33, pp. 1218-1223, 2020. https://doi.org/10.1007/s10278-020-00357-7
- [17] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint, arXiv:1409.1556, 2014. https://doi.org/10.48550/arXiv.1409.1556
- [18] J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “How transferable are features in deep neural networks?,” Advances in Neural Information Processing Systems, Vol.27, 2014. https://doi.org/10.48550/arXiv.1411.1792
- [19] E. Tzeng, J. Hoffman, N. Zhang, K. Saenko, and T. Darrell, “Deep domain confusion: Maximizing for domain invariance,” arXiv preprint, arXiv:1412.3474, 2014. https://doi.org/10.48550/arXiv.1412.3474
- [20] M. Long, Y. Cao, J. Wang, and M. Jordan, “Learning transferable features with deep adaptation networks,” Proc. of Int. Conf. on Machine Learning Research (PMLR), pp. 97-105, 2015.
- [21] A. Gretton et al., “Optimal kernel choice for large-scale two-sample tests,” Advances in Neural Information Processing Systems, Vol.25, 2012.
- [22] F. Zhuang, X. Cheng, P. Luo, S. J. Pan, and Q. He, “Supervised representation learning: Transfer learning with deep autoencoders,” 24th Int. Conf. on Artificial Intelligence, pp. 4119-4125, 2015.
- [23] M. Long, J. Wang, Y. Cao, J. Sun, and P. S. Yu, “Deep learning of transferable representation for scalable domain adaptation,” IEEE Trans. on Knowledge and Data Engineering, Vol.28, No.8, pp. 2027-2040, 2016. https://doi.org/10.1109/TKDE.2016.2554549
- [24] B. Sun and K. Saenko, “Deep CORAL: Correlation Alignment for Deep Domain Adaptation,” arXiv preprint, arXiv:1607.01719, 2016. https://doi.org/10.48550/arXiv.1607.01719
- [25] P. Wei, Y. Ke, and C. K. Goh, “Deep nonlinear feature coding for unsupervised domain adaptation,” Proc. of the 25th Int. Joint Conf. on Artificial Intelligence (IJCAI 2016), pp. 2189-2195, 2016.
- [26] J. Zhuo, S. Wang, W. Zhang, and Q. Huang, “Deep unsupervised convolutional domain adaptation,” Proc. of the 25th ACM Int. Conf. on Multimedia, pp. 261-269, 2017. https://doi.org/10.1145/3123266.3123292
- [27] Z. Luo, Y. Zou, J. Hoffman, and L. Fei-Fei, “Label efficient learning of transferable representations across domains and tasks,” Advances in Neural Information Processing Systems, Vol.30, 2017. https://doi.org/10.48550/arXiv.1712.00123
- [28] Y. Ganin et al., “Domain-adversarial training of neural networks,” J. of Machine Learning Research, Vol.17, No.1, pp. 2030-2096, 2016.
- [29] K. Bousmalis, G. Trigeorgis, N. Silberman, D. Krishnan, and D. Erhan, “Domain separation networks,” Advances in Neural Information Processing Systems, Vol.29, 2016. https://doi.org/10.48550/arXiv.1608.06019
- [30] E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation,” Proc. of 2017 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 2962-2971, 2017. https://doi.org/10.1109/CVPR.2017.316
- [31] C. Yu, J. Wang, Y. Chen, and M. Huang, “Transfer learning with dynamic adversarial adaptation network,” 2019 IEEE Int. Conf. on Data Mining (ICDM), pp. 778-786, 2019. https://doi.org/10.1109/ICDM.2019.00088
- [32] B. Xu, K. Wu, Y. Wu, J. He, and C. Chen, “Dynamic adversarial domain adaptation based on multikernel maximum mean discrepancy for breast ultrasound image classification,” Expert Systems with Applications, Vol.207, Article No.117978, 2022. https://doi.org/10.1016/j.eswa.2022.117978
- [33] K. M. Borgwardt et al., “Integrating structured biological data by kernel maximum mean discrepancy,” Bioinformatics, Vol.22, Issue 14, pp. e49-e57, 2006. https://doi.org/10.1093/bioinformatics/btl242
- [34] M. H. Yap et al., “Automated breast ultrasound lesions detection using convolutional neural networks,” IEEE J. of Biomedical and Health Informatics, Vol.22, No.4, pp. 1218-1226, 2018.
This article is published under a Creative Commons Attribution-NoDerivatives 4.0 Internationa License.