Selective Deeply Supervised Multi-Scale Attention Network for Brain Tumor Segmentation
<p>Illustrations of brain tumor regions in an MRI slice from the BraTS 2020 database. From <b>left</b> to <b>right</b>: FLAIR, T1, T1ce, and T2 slices.</p> "> Figure 2
<p>The illustration of the preprocessing stage, which includes scan refinement and image enhancement using cropping and histogram equalization, respectively.</p> "> Figure 3
<p>Selective deeply supervised multi-scale attention network (SDS-MSA-Net) takes 2D and 3D inputs to segment three types of brain tumor regions. SDS-MSA-Net produces four outputs, which enable it to be trained with the selective deep supervision technique.</p> "> Figure 4
<p>Illustration of the architecture of Res block, Conv block, bridge block, DeConv block, and auxiliary block. (<b>a</b>) Res blocks and (<b>c</b>) bridge blocks are used in the 3D encoding unit to extract and to downscale the dimensions of the meaningful features, respectively; (<b>b</b>) Conv blocks are employed in the 2D encoding unit; (<b>d</b>) DeConv block is used in the decoder block to upscale the refined features; finally, (<b>e</b>) the auxiliary block employed in the SDS block to immediately upscale the features from intermediate layers of the decoder block to produce the segmentation mask of the selected brain tumor region(s).</p> "> Figure 5
<p>The schematic of the attention unit (AU) that uses additive attention is illustrated. AG is being utilized in the decoder block in the proposed SDS-MSA-Net (<a href="#sensors-23-02346-f003" class="html-fig">Figure 3</a>). The input features (<span class="html-italic">x</span>) are scaled with attention coefficients (<math display="inline"><semantics> <mi>α</mi> </semantics></math>) computed in AU. Spatial regions are selected by analyzing both the activations and contextual information provided by the gating signal (<span class="html-italic">g</span>), which is collected from a coarser scale. AUs are employed in the proposed MSA-Net at the decoder block to refine the coarse features coming from the encoder block.</p> "> Figure 6
<p>Learning curves for different training schemes.</p> "> Figure 7
<p>Results of SDS-MSA-Net compared with three downgraded variants (attention UNet, MS-CNN, and MSA-CNN). Note: Red, blue and green colors indicate the whole, core and enhanced tumor regions, respectively.</p> ">
Abstract
:1. Introduction
- This study proposes a novel selective deeply supervised multi-scale attention network (SDS-MSA-Net) framework that combines global and local features to improve the performance of brain tumor segmentation.
- The proposed model incorporates selective deep supervision as a novel training approach to improve the performance of the model for the task at hand. By adding auxiliary outputs at various levels of the network, we aim to achieve improved performance, faster convergence, and better generalization of the model.
- The presented methodology underwent a comprehensive evaluation for the task of brain tumor segmentation on the BraTS2020 dataset [13]. Our framework demonstrates substantial progress in the segmentation of both the enhanced and core brain tumor regions, as evidenced by the improvement in the Dice score, which serves as a metric for the efficacy of our proposed framework.
2. Related Work
3. Materials and Methods
3.1. Data and Preprocessing
3.2. Selective Deeply Supervised Multi-Scale Convolutional Neural Network
3.2.1. Encoder Block
3.2.2. Decoder Block
3.2.3. Selective Deep Supervision Block
3.2.4. Implementation Details and Training Strategy
3.3. Performance Measures
- Dice Similarity Coefficient: The evaluation of the proposed framework’s performance utilizes the Dice similarity coefficient (DSC) [33]. The DSC measures the degree of overlap between the ground truth mask and the predicted mask, with values ranging from 0 to 1. A value of 1 represents complete overlap and a value of 0 represents no overlap. The DSC is defined as follows:
- Sensitivity: To measure the pixel classification performance proposed framework, the used sensitivity (SEN) can be defined as follows:
- Specificity: To measure the correctness of the segmentation area produced by the proposed framework, the used Specificity can be defined as follows:
- Hausdorff Distance: The Hausdorff Distance (HD) is a widely used metric in the assessment of medical segmentation [34]. The Hausdorff distance is an important measure in brain tumor segmentation because it provides a quantitative way to evaluate the similarity between two sets of points, such as the ground truth segmentation and the predicted segmentation. It calculates the differences between two sets of points, with the directed Hausdorff distance between two sets ( and ) defined as the maximum distance between each point and its nearest neighbor .
4. Results and Discussion
4.1. Benchmarking Results
4.2. Impact of Selective Deep Supervision on Training
4.3. Qualitative Analysis
4.4. Overall Performance Analysis
5. Limitations
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Society, N.B.T. Brain Tumor Facts and Statistics. 2021. Available online: https://braintumor.org/brain-tumor-information/brain-tumor-facts/ (accessed on 14 February 2023).
- Agravat, R.R.; Raval, M.S. 3D semantic segmentation of brain tumor for overall survival prediction. In Proceedings of the International MICCAI Brainlesion Workshop, Lima, Peru, 4 October 2020; pp. 215–227. [Google Scholar]
- Pereira, S.; Pinto, A.; Alves, V.; Silva, C.A. Brain tumor segmentation using convolutional neural networks in MRI images. IEEE Trans. Med. Imaging 2016, 35, 1240–1251. [Google Scholar] [CrossRef] [PubMed]
- Saouli, R.; Akil, M.; Kachouri, R. Fully automatic brain tumor segmentation using end-to-end incremental deep neural networks in MRI images. Comput. Methods Programs Biomed. 2018, 166, 39–49. [Google Scholar]
- Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.M.; Larochelle, H. Brain tumor segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Dong, H.; Yang, G.; Liu, F.; Mo, Y.; Guo, Y. Automatic brain tumor detection and segmentation using U-Net based fully convolutional networks. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis, Edinburgh, UK, 11–13 July 2017; pp. 506–517. [Google Scholar]
- Li, Z.; Sun, Y.; Zhang, L.; Tang, J. CTNet: Context-based tandem network for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9904–9917. [Google Scholar] [CrossRef]
- Sun, Y.; Li, Z. SSA: Semantic Structure Aware Inference for Weakly Pixel-Wise Dense Predictions without Cost. arXiv 2021, arXiv:2111.03392. [Google Scholar]
- Usman, M.; Lee, B.D.; Byon, S.S.; Kim, S.H.; Lee, B.i.; Shin, Y.G. Volumetric lung nodule segmentation using adaptive roi with multi-view residual learning. Sci. Rep. 2020, 10, 12839. [Google Scholar] [CrossRef]
- Nguyen, D.H.; Nguyen, D.M.; Truong, M.T.; Nguyen, T.; Tran, K.T.; Triet, N.A.; Bao, P.T.; Nguyen, B.T. ASMCNN: An Efficient Brain Extraction Using Active Shape Model and Convolutional Neural Networks. Inf. Sci. 2022, 591, 25–48. [Google Scholar] [CrossRef]
- Mahbub, M.K.; Biswas, M.; Gaur, L.; Alenezi, F.; Santosh, K. Deep features to detect pulmonary abnormalities in chest X-rays due to infectious diseaseX: COVID-19, pneumonia, and tuberculosis. Inf. Sci. 2022, 592, 389–401. [Google Scholar] [CrossRef]
- Latif, S.; Usman, M.; Manzoor, S.; Iqbal, W.; Qadir, J.; Tyson, G.; Castro, I.; Razi, A.; Boulos, M.N.K.; Weller, A.; et al. Leveraging data science to combat COVID-19: A comprehensive review. IEEE Trans. Artif. Intell. 2020, 1, 85–103. [Google Scholar] [CrossRef]
- Ghaffari, M.; Sowmya, A.; Oliver, R. Automated brain tumour segmentation using cascaded 3d densely-connected u-net. In Proceedings of the International MICCAI Brainlesion Workshop, Lima, Peru, 4 October 2020; pp. 481–491. [Google Scholar]
- Ullah, Z.; Usman, M.; Gwak, J. MTSS-AAE: Multi-task semi-supervised adversarial autoencoding for COVID-19 detection based on chest X-ray images. Expert Syst. Appl. 2023, 216, 119475. [Google Scholar] [CrossRef]
- Tampu, I.E.; Haj-Hosseini, N.; Eklund, A. Does Anatomical Contextual Information Improve 3D U-Net-Based Brain Tumor Segmentation? Diagnostics 2021, 11, 1159. [Google Scholar] [CrossRef] [PubMed]
- Ahmad, P.; Qamar, S.; Shen, L.; Saeed, A. Context aware 3D UNet for brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Lima, Peru, 4 October 2020; pp. 207–218. [Google Scholar]
- Wang, Y.; Zhang, Y.; Hou, F.; Liu, Y.; Tian, J.; Zhong, C.; Zhang, Y.; He, Z. Modality-pairing learning for brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Lima, Peru, 4 October 2020; pp. 230–240. [Google Scholar]
- Mora Ballestar, L.; Vilaplana, V. Brain Tumor Segmentation using 3D-CNNs with Uncertainty Estimation. arXiv 2020, arXiv:2009.12188. [Google Scholar]
- Huang, G.; Chen, D.; Li, T.; Wu, F.; Van Der Maaten, L.; Weinberger, K.Q. Multi-scale dense networks for resource efficient image classification. arXiv 2017, arXiv:1703.09844. [Google Scholar]
- Jiang, Z.; Ding, C.; Liu, M.; Tao, D. Two-Stage Cascaded U-Net: 1st Place Solution to BraTS Challenge 2019 Segmentation Task. In Proceedings of the International MICCAI Brainlesion Workshop, Shenzhen, China, 17 October 2019; pp. 231–241. [Google Scholar]
- Myronenko, A. 3D MRI brain tumor segmentation using autoencoder regularization. In Proceedings of the International MICCAI Brainlesion Workshop, Granada, Spain, 16 September 2018; pp. 311–320. [Google Scholar]
- Isensee, F.; Kickingereder, P.; Wick, W.; Bendszus, M.; Maier-Hein, K.H. No new-net. In Proceedings of the International MICCAI Brainlesion Workshop, Granada, Spain, 16 September 2018; pp. 234–244. [Google Scholar]
- Zhao, X.; Wu, Y.; Song, G.; Li, Z.; Zhang, Y.; Fan, Y. A deep learning model integrating FCNNs and CRFs for brain tumor segmentation. Med. Image Anal. 2018, 43, 98–111. [Google Scholar] [CrossRef]
- Zhang, T.; Li, A.; Wang, M.; Wu, X.; Qiu, B. Multiple attention fully convolutional network for automated ventricle segmentation in cardiac magnetic resonance imaging. J. Med. Imaging Health Inform. 2019, 9, 1037–1045. [Google Scholar] [CrossRef]
- Liu, X.; Zhang, L.; Li, T.; Wang, D.; Wang, Z. Dual attention guided multi-scale CNN for fine-grained image classification. Inf. Sci. 2021, 573, 37–45. [Google Scholar] [CrossRef]
- Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
- Ullah, Z.; Usman, M.; Jeon, M.; Gwak, J. Cascade multiscale residual attention cnns with adaptive roi for automatic brain tumor segmentation. Inf. Sci. 2022, 608, 1541–1556. [Google Scholar] [CrossRef]
- Gross, S.; Wilber, M. Training and Investigating Residual Nets. 2016. Available online: http://torch.ch/blog/2016/02/04/resnets.html (accessed on 19 February 2023).
- Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
- Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. J. Mach. Learn. Res. 2010, 9, 249–256. [Google Scholar]
- Bradley, D.M. Learning in Modular Systems; Carnegie Mellon University ProQuest Dissertations Publishing: Pittsburgh, PA, USA, 2010. [Google Scholar]
- Dou, Q.; Yu, L.; Chen, H.; Jin, Y.; Yang, X.; Qin, J.; Heng, P.A. 3D deeply supervised network for automated segmentation of volumetric medical images. Med. Image Anal. 2017, 41, 40–54. [Google Scholar] [CrossRef]
- Wang, S.; Zhou, M.; Liu, Z.; Liu, Z.; Gu, D.; Zang, Y.; Dong, D.; Gevaert, O.; Tian, J. Central focused convolutional neural networks: Developing a data-driven model for lung nodule segmentation. Med. Image Anal. 2017, 40, 172–183. [Google Scholar] [CrossRef] [PubMed]
- Wang, G.; Li, W.; Ourselin, S.; Vercauteren, T. Automatic brain tumor segmentation using cascaded anisotropic convolutional neural networks. In Proceedings of the International MICCAI Brainlesion Workshop, Quebec, QC, Canada, 14 September 2017; pp. 178–190. [Google Scholar]
- Nguyen, H.T.; Le, T.T.; Nguyen, T.V.; Nguyen, N.T. Enhancing MRI brain tumor segmentation with an additional classification network. arXiv 2020, arXiv:2009.12111. [Google Scholar]
Datasets | Total No. of Scans | Total No. of Training Scans | Total No. of Validation Scans | Total No. of Testing Scans |
---|---|---|---|---|
BraTS2020 | 369 | 240 | 18 | 111 |
BraTS2019 | 335 | 218 | 17 | 100 |
BraTS2017 | 285 | 185 | 14 | 86 |
Method | Dice | Sensitivity | Specificity | Hausdorff | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
WT | TC | ET | WT | TC | ET | WT | TC | ET | WT | TC | ET | |
Attention-UNet [29] | 86.85 | 71.92 | 72.85 | 84.22 | 68.16 | 76.09 | 83.94 | 79.11 | 80.67 | 7.89 | 10.42 | 9.22 |
3D Multi-Scale-Net | 87.83 | 80.26 | 75.63 | 88.81 | 74.18 | 78.35 | 85.22 | 78.25 | 79.45 | 6.28 | 8.43 | 7.14 |
MSA-Net | 88.65 | 84.73 | 78.95 | 89.34 | 87.12 | 81.57 | 88.42 | 81.57 | 84.32 | 5.62 | 7.91 | 5.96 |
DS-MSA-Net | 89.19 | 84.96 | 78.81 | 89.87 | 88.59 | 81.94 | 88.02 | 80.93 | 84.10 | 5.11 | 6.91 | 6.04 |
SDS-MSA-Net | 90.24 | 86.93 | 80.64 | 93.16 | 90.78 | 92.41 | 92.65 | 94.12 | 93.06 | 4.27 | 6.32 | 5.87 |
Method | Dice Score | Hausdorff | ||||
---|---|---|---|---|---|---|
WT | TC | ET | WT | TC | ET | |
Mina et al. [13] | 90.0 ± – | 82.0± – | 78.0 ± – | 5.14 ± – | 6.64 ± – | 7.71 ± – |
Rupal and Mehul [2] | 87.0 ± 9.3 | 72.0 ± 28.4 | 73.0 ±30.7 | 9.47 ± 15.2 | 14.53 ± 38.06 | 34.19 ±109.14 |
Laura and Veronica [18] | 84.0 ± – | 75.0 ± – | 62.0 ± – | 20.4 ± – | 12.17 ± – | 47.7 ± – |
Hieu et al. [35] | 89.99 ± – | 84.22 ± – | 78.43 ± – | 5.68 ± – | 9.56 ± – | 24.02 ± – |
Wang et al. [17] | 90.0 ± 7.8 | 85.0 ± 13.0 | 78.0 ± 27.6 | 4.39 ± 7.62 | 8.34 ± 10.17 | 32.25 ± 105.54 |
Parvez et al. [16] | 89.12 ± – | 84.74 ± – | 79.12 ± – | – | – | – |
Lulian et al. [15] | 89.85 ± – | 82.36 ± – | 70.07 ± – | 13.67 ± – | 15.62 ± – | 99.54 ± – |
SDS-MSA-Net | 90.24 ± 7.21 | 86.23 ± 10.58 | 80.64 ± 18.46 | 4.27 ± 9.56 | 6.32 ± 9.88 | 5.87 ± 22.13 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Rehman, A.; Usman, M.; Shahid, A.; Latif, S.; Qadir, J. Selective Deeply Supervised Multi-Scale Attention Network for Brain Tumor Segmentation. Sensors 2023, 23, 2346. https://doi.org/10.3390/s23042346
Rehman A, Usman M, Shahid A, Latif S, Qadir J. Selective Deeply Supervised Multi-Scale Attention Network for Brain Tumor Segmentation. Sensors. 2023; 23(4):2346. https://doi.org/10.3390/s23042346
Chicago/Turabian StyleRehman, Azka, Muhammad Usman, Abdullah Shahid, Siddique Latif, and Junaid Qadir. 2023. "Selective Deeply Supervised Multi-Scale Attention Network for Brain Tumor Segmentation" Sensors 23, no. 4: 2346. https://doi.org/10.3390/s23042346