Abstract
Breast tumor segmentation plays a critical role in the early diagnosis and treatment of breast diseases. However, existing segmentation methods yet have limitations in dealing with the complex structures and low contrast of breast ultrasound images. To address this issue, we propose a novel dual-attention patching network (DPNet) for breast tumor segmentation. DPNet adopts a dual-attention structure, making efficient use of spatial information, and employs a specific fusion strategy to maximize the integration of global context and local details. To further enhance the integration of global context and local details, we introduce the spatial feature aggregation block (SFA) and the space-channel position hybrid block (SCPH). The SFA incorporates global and local attention pathways, suppressing irrelevant information and emphasizing valuable features, effectively integrating long-range spatial dependencies with fine-grained features of interest. Moreover, SCPH enhances the receptive field and improves the understanding of image structures and details by fusing spatial and channel positional information. Experimental results on the BUSI dataset demonstrate the effectiveness of our DPNet, achieving remarkable performance with Dice score of 81.10%, IoU of 73.03%, and accuracy of 96.29%. This significantly enhances the segmentation performance of breast tumor and outperforms other state-of-the-art methods in terms of segmentation accuracy and robustness.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
No datasets were generated or analysed during the current study.
References
Tanaka, H., Chiu, S.-W., Watanabe, T., Kaoku, S., Yamaguchi, T.: Computer-aided diagnosis system for breast ultrasound images using deep learning. Phys. Med. Biol. 64(23), 235013 (2019)
Wei, Q., Yan, Y.-J., Wu, G.-G., Ye, X.-R., Jiang, F., Liu, J., Wang, G., Wang, Y., Song, J., Pan, Z.-P.: The diagnostic performance of ultrasound computer-aided diagnosis system for distinguishing breast masses: a prospective multicenter study. Eur. Radiol. 32(6), 4046–4055 (2022)
Pedraza, L., Vargas, C., Narváez, F., Durán, O., Muñoz, E., Romero, E.: An open access thyroid ultrasound image database. In: 10th International Symposium on Medical Information Processing and Analysis, vol. 9287, pp. 188–193. SPIE (2015)
Xu, C., Qi, Y., Wang, Y., Lou, M., Pi, J., Ma, Y.: Arf-net: an adaptive receptive field network for breast mass segmentation in whole mammograms and ultrasound images. Biomed. Signal Process. Control 71, 103178 (2022)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pp. 234–241. Springer (2015)
Azad, R., Aghdam, E.K., Rauland, A., Jia, Y., Avval, A.H., Bozorgpour, A., Karimijafarbigloo, S., Cohen, J.P., Adeli, E., Merhof, D.: Medical image segmentation review: the success of U-Net. arXiv preprint arXiv:2211.14830 (2022)
Wang, P., Chen, P., Yuan, Y., Liu, D., Huang, Z., Hou, X., Cottrell, G.: Understanding convolution for semantic segmentation. (2018). arxiv:abs/1702.08502
Pan, X., Ge, C., Lu, R., Song, S., Chen, G., Huang, Z., Huang, G.: On the integration of self-attention and convolution. (2022). arxiv:abs/2111.14556
Lou, M., Zhou, H.-Y., Yang, S., Yu, Y.: TransXNet: learning both global and local dynamics with a dual dynamic token mixer for visual recognition. (2023). arxiv:abs/2310.19380
Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: Unet++: a nested u-net architecture for medical image segmentation. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 20, 2018, Proceedings 4, pp. 3–11. Springer (2018)
Oktay, O., Schlemper, J., Le Folgoc, L., Lee, M., Heinrich, M., Misawa, K., Mori, K., McDonagh, S., Hammerla, N.Y., Kainz, B., : Attention U-Net: learning where to look for the pancreas. In: Medical Imaging with Deep Learning (2022)
Zhang, Z., Liu, Q., Wang, Y.: Road extraction by deep residual U-Net. IEEE Geosci. Remote Sens. Lett. 15(5), 749–753 (2018)
Valanarasu, J.M.J., Patel, V.M.: Unext: MLP-based rapid medical image segmentation network. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 23–33. Springer (2022)
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)
Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)
Guo, M.-H., Xu, T.-X., Liu, J.-J., Liu, Z.-N., Jiang, P.-T., Mu, T.-J., Zhang, S.-H., Martin, R.R., Cheng, M.-M., Hu, S.-M.: Attention mechanisms in computer vision: a survey. Comput. Vis. Media 8(3), 331–368 (2022)
Li, K., Wang, Y., Zhang, J., Gao, P., Song, G., Liu, Y., Li, H., Qiao, Y.: Uniformer: unifying convolution and self-attention for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. (2023)
Cao, H., Wang, Y., Chen, J., Jiang, D., Zhang, X., Tian, Q., Wang, M.: Swin-Unet: Unet-like pure transformer for medical image segmentation. In: European Conference on Computer Vision, pp. 205–218. Springer (2022)
Huang, X., Deng, Z., Li, D., Yuan, X., Fu, Y.: Missformer: an effective transformer for 2D medical image segmentation. IEEE Trans. Med. Imaging (2022)
Wang, H., Xie, S., Lin, L., Iwamoto, Y., Han, X.-H., Chen, Y.-W., Tong, R.: Mixed transformer U-Net for medical image segmentation. In: ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2390–2394. IEEE (2022)
Zhang, Y., Liu, H., Hu, Q.: Transfuse: fusing transformers and CNNs for medical image segmentation. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part I 24, pp. 14–24. Springer (2021)
Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., Lu, L., Yuille, A.L., Zhou, Y.: TransUnet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:2102.04306 (2021)
Wang, H., Cao, P., Wang, J., Zaiane, O.R.: UCTransNet: rethinking the skip connections in U-Net from a channel-wise perspective with transformer. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, pp. 2441–2449 (2022)
Lou, M., Meng, J., Qi, Y., Li, X., Ma, Y.: MCRNet: multi-level context refinement network for semantic segmentation in breast ultrasound imaging. Neurocomputing 470, 154–169 (2022)
Lou, M., Qi, Y., Meng, J., Xu, C., Wang, Y., Pi, J., Ma, Y.: DCANet: dual contextual affinity network for mass segmentation in whole mammograms. Med. Phys. 48(8), 4291–4303 (2021)
Oza, P., Sharma, P., Patel, S., Kumar, P.: Deep convolutional neural networks for computer-aided breast cancer diagnostic: a survey. Neural Comput. Appl. 34(3), 1815–1836 (2022)
Farshad, A., Yeganeh, Y., Gehlbach, P., Navab, N.: Y-Net: a spatiospectral dual-encoder network for medical image segmentation. In: International conference on medical image computing and computer-assisted intervention, pp. 582–592. Springer (2022)
Zhang, X., Lin, C.: Multi-decoding network with attention learning for edge detection. Neural Process. Lett. 1–18 (2022)
Luo, X., Cao, S., Wang, Z., Chen, Y.: LCDA-Net: efficient image dehazing with contrast-regularized and dilated attention. Neural Process. Lett. 1–22 (2023)
Tang, F., Huang, Q., Wang, J., Hou, X., Su, J., Liu, J.: DUAT: dual-aggregation transformer network for medical image segmentation. arXiv preprint arXiv:2212.11677 (2022)
Trockman, A., Kolter, J.Z.: Patches are all you need? arXiv preprint arXiv:2201.09792 (2022)
Tang, F., Wang, L., Ning, C., Xian, M., Ding, J.: CMU-Net: a strong ConvMixer-based medical ultrasound image segmentation network. arXiv preprint arXiv:2210.13012 (2022)
Al-Dhabyani, W., Gomaa, M., Khaled, H., Fahmy, A.: Dataset of breast ultrasound images. Data Brief 28, 104863 (2020)
Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al.: PyTorch: an imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 32 (2019)
Funding
The work is supported by the Key Science and Technology Program of Henan Province, China (242102210051).
Author information
Authors and Affiliations
Contributions
The authors participate in the writing of each chapter and discuss changes.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships.
Ethical approval
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Liu, S., Liu, D. & Lin, Y. DPNet: a dual-attention patching network for breast tumor segmentation in an ultrasound image. Multimedia Systems 30, 355 (2024). https://doi.org/10.1007/s00530-024-01562-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00530-024-01562-y