Bioengineering 10 00957 v2
Bioengineering 10 00957 v2
Article
RGSB-UNet: Hybrid Deep Learning Framework for Tumour
Segmentation in Digital Pathology Images
Tengfei Zhao 1 , Chong Fu 1,2,3, * , Ming Tie 4 , Chiu-Wing Sham 5 and Hongfeng Ma 6
1 School of Computer Science and Engineering, Northeastern University, Shenyang 110819, China
2 Engineering Research Center of Security Technology of Complex Network System, Ministry of Education,
Shenyang 110819, China
3 Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University,
Shenyang 110819, China
4 Science and Technology on Space Physics Laboratory, Beijing 100076, China
5 School of Computer Science, The University of Auckland, Auckland 1142, New Zealand
6 Dopamine Group Ltd., Auckland 1542, New Zealand
* Correspondence: fuchong@mail.neu.edu.cn
Abstract: Colorectal cancer (CRC) is a prevalent gastrointestinal tumour with high incidence and
mortality rates. Early screening for CRC can improve cure rates and reduce mortality. Recently, deep
convolution neural network (CNN)-based pathological image diagnosis has been intensively studied
to meet the challenge of time-consuming and labour-intense manual analysis of high-resolution
whole slide images (WSIs). Despite the achievements made, deep CNN-based methods still suffer
from some limitations, and the fundamental problem is that they cannot capture global features. To
address this issue, we propose a hybrid deep learning framework (RGSB-UNet) for automatic tumour
segmentation in WSIs. The framework adopts a UNet architecture that consists of the newly-designed
residual ghost block with switchable normalization (RGS) and the bottleneck transformer (BoT) for
downsampling to extract refined features, and the transposed convolution and 1 × 1 convolution with
ReLU for upsampling to restore the feature map resolution to that of the original image. The proposed
framework combines the advantages of the spatial-local correlation of CNNs and the long-distance
feature dependencies of BoT, ensuring its capacity of extracting more refined features and robustness
Citation: Zhao, T.; Fu, C.; Tie, M.;
to varying batch sizes. Additionally, we consider a class-wise dice loss (CDL) function to train the
Sham, C.-W.; Ma, H. RGSB-UNet:
segmentation network. The proposed network achieves state-of-the-art segmentation performance
Hybrid Deep Learning Framework
under small batch sizes. Experimental results on DigestPath2019 and GlaS datasets demonstrate that
for Tumour Segmentation in Digital
Pathology Images. Bioengineering
our proposed model produces superior evaluation scores and state-of-the-art segmentation results.
2023, 10, 957. https://doi.org/
10.3390/bioengineering10080957 Keywords: hybrid deep learning framework; tumour segmentation; whole slide image; Residual-
Ghost-SN; bottleneck transformer
Academic Editors: Yan Pei
and Jijiang Yang
stability under small batch sizes. The proposed encoder–decoder architecture utilizes a
newly designed encoder that includes residual ghost blocks with switchable normalization
(RGS) and a bottleneck transformer block (BoT) for downsampling, while the decoder em-
ploys transpose convolution for upsampling [23–27]. By leveraging the benefits of CNNs
and the transformer, the proposed encoder uses RGS and BoT as downsampling operations
to extract more refined features from input images. The operation extracts local informa-
tion, and the multi-head self-attention (MHSA) in the BoT models global dependency [27].
Experimental results demonstrate that the proposed model can accurately segment the
tumour and produce a more refined boundary, leading to improved segmentation accuracy
under small batch sizes. The primary contributions of our study are outlined below:
• We propose a deep hybrid network that combines a transformer and CNN for auto-
matic tumour region segmentation in pathology images of the colon.
• A newly-designed feature extraction block RGS is presented. The block can adaptively
determine the optimal combination of normalizers for each layer, making our model
robust to varying batch sizes.
• Our novel hybrid backbone encoder, which includes RGS and BoT blocks, can extract
more refined features.
• Experimental results demonstrate that the proposed RGSB-UNet achieves higher
evaluation scores and produces finer segmentation results than state-of-the-art seg-
mentation methods under small batch sizes.
The remainder of this paper is structured as follows. In Section 2, we present the
proposed network architecture. Section 3 describes the datasets and evaluation criteria used
in our experiments, while Section 4 presents our experimental results. Finally, in Section 5,
we summarize the study results and suggest potential avenues for future research.
2. Proposed Method
2.1. Network Architecture
Our proposed deep learning framework for colon pathology WSI analysis is illustrated
in Figure 1. As shown in Figure 2, to extract relevant features from original images, we start
with 512 × 512 × 3 image patches using dense cropping methods. The encoder includes a
novel downsampling operation that combines RGS and BoT blocks as the feature extraction
backbone. The details of the design of the encoder and decoder, GBS, RGS, and BoT will be
discussed below.
Figure 1. An overview of RGSB-UNet. The TRCCR denotes transposed convolution, ReLU, concate-
nate, convolution, and ReLU.
Bioengineering 2023, 10, 957 4 of 16
ReLU
BoT
3×3 conv s=1 p=1
2×2 AP s=2 p=0
concatenate
ReLU
3×3 Tconv s=2 p=1
Figure 2. Schematic diagram of RGSB-UNet. RGS denotes the proposed residual ghost block with
switchable normalization, and BoT denotes the bottleneck transformer. MP and AP denote the max
and average pooling, respectively. Tconv denotes the transposed convolution used for upsampling.
SN
ReLU
SN
ReLU
Concatenate
Figure 3. Schematic diagram of Ghost block with switchable normalization. The dash box denotes
the cheap operation that uses a 3 × 3 group convolution in the ghost block.
GBS GBS
SN SN
SN SN
GSB GBS
SN SN
Add Add
(a) (b)
Figure 4. Schematic diagram of the proposed bottleneck. (a) RGS Bottleneck. (b) Bottleneck trans-
former. GBS and SN denote the ghost block with switchable normalization and switchable normaliza-
tion, respectively. MHSA denotes multi-head self-attention.
z
H×W×d
Self-Attention Layer H*W×H*W H*W×d Multi-headed Self-attention
softmax
b4
b3 Concatenate
2
b
H* W×H*W H*W×H*W b1
qrT qkT Self-Attention Layer
content-content
content-position Self-Attention Layer head=4
H×W×d Self-Attention Layer Linear
Self-Attention Layer
q k v
a1
H×1×d a1
1×W×d
H×W×d H×W×d H×W×d a1
a1
WQ:1×1 WK:1×1 WV:1×1 b1
Rh Rw
H×W×d
(a) x (b)
Figure 5. Schematic diagram of (a) self-attention [26] and (b) multi-head self-attention.
where yi is the binary label of pixel i, ŷi is the predicted probability, and N is the total
number of pixels in a patch. e is a small number to avoid the denominator becoming 0.
The presence of a lesion area determines the patch label (y p ). The CDL function can
alleviate pixel-level class imbalance, resulting in an all-zero mask when training nega-
tive samples.
2|YA ∩ YP |
DSC = , (2)
|YA | + |YP |
|YA ∩ YP |
JI = , (3)
|YA | + |YP | − |YA ∩ YP |
and
|YP | − |YA |
RVD = , (4)
|YA |
where YA is the set of lesion pixels in the annotation, and YP is the corresponding set of
lesion pixels in the segmentation result.
Bioengineering 2023, 10, 957 8 of 16
We use pixel accuracy (PA) and area under the curve (AUC) to measure the classi-
fication performance of our proposed model. AUC is defined as the area of the receiver
operating characteristic (ROC) curve, determined by the true positive rate (TPR) and false
positive rate (FPR). TPR, FPR, and Precision are defined as follows:
TP
TPR = , (5)
TP + FN
FP
FPR = , (6)
FP + TN
and
TP
Precision = , (7)
TP + FP
where TP, FP, TN, and FN are true positives, false positives, true negatives, and false
negatives, respectively.
AUC and PA are defined as
Z 1
AUC = TPR( FPR−1 ( x ))dx = P( X1 > X0 ) (8)
x =0
and
TP + TN
PA = , (9)
TP + TN + FP + FN
where X0 and X1 are the scores for the negative and positive instances, respectively.
The simulations were run on a station equipped with an NVIDIA GeForce RTX 3090
GPU and Intel(R) Xeon(R) CPU E5-2680v4×2. We augmented the training data during
training. Table 1 lists the detailed hyperparameters of the proposed framework. We
embarked on an iterative journey of manual tuning, wherein we systematically explored
and fine-tuned various hyperparameters within our framework. By meticulously adjusting
parameters such as learning rates, batch sizes, and model architecture, we meticulously
tracked the impact of each modification on the overall performance metrics. This exhaustive
process allowed us to discover the optimal combination of hyperparameters, leading to a
highly refined and efficient version of our framework that exhibits superior accuracy and
generalization on diverse datasets.
Hyperparameters Value
Crop Method Dense Crop
Crop Stride 512
Crop Patch Size 512 × 512 × 3
Batch Size 2
MHSA Head 4
Optimizer SGD
Learning Rate 1.0 × e−2
Weight Deacy 1.0 × e−4
Momentum 0.9
Epoch Number 500
Bioengineering 2023, 10, 957 10 of 16
4. Experimental Results
Table 2 shows the results of the ablation study, which demonstrate the performance
gains when integrating different blocks into UNet, including residual block (RSB), residual
ghost block (RGB), RGS, and BoT. Especially, our proposed RGSB-UNet achieves the
highest DSC score of 0.8336. We further analyze the performance of different batch sizes
and MHSA head numbers based on RGSB-UNet. As is shown in Table 3, the proposed
network maintains high performance even with small batch sizes. We tried different small
batch sizes in our experiments. We prove that batch size is no longer a strict limitation for
the proposed network. In addition, the head numbers of MHSA impact the performance
of the proposed network. We have tried different numbers of heads for the MHSA in
the proposed network to search for the best results, and our network achieved optimal
performance when the heads are four. When integrating RGS and BoT together to the UNet,
the segmentation model produces the best performance, which indicates that these blocks
can improve the performance of pathology image segmentation.
Table 2. Performance gains by integrating different blocks into UNet on the DigestPath2019 dataset.
RSB and RGB denote the residual block and residual ghost block with batch normalization, respectively.
Table 3. Effect of batch size and MHSA head on model performance. The best results are marked
in bold.
Batch Size 1 2
MHSA Head 1 2 4 1 2 4
DSC 0.8126 0.8241 0.8220 0.8331 0.8263 0.8294 0.8250 0.8336
Table 4 compares the performance of the proposed and other popular models in terms
of six metrics on the DigestPath2019 dataset; the numbers in bold indicate the best results for
each metric. As can be seen from this table, under a small batch of two, our proposed model
achieves the highest DSC, PA, JI, and Precision; it also achieves the second best RVD and
AUC. Furthermore, although DeepLab with Xception backbone outperforms other models
in terms of RVD, and the CAC-UNet (first place) achieves the highest AUC, our model
performs significantly better in the other three metrics. In Figure 7, we illustrate the results
of tumour segmentation on the sample images and compare them with that of [20,21,37–46].
As shown in this figure, the mask predicted by the proposed network is extremely close
to the ground truth. Compared with other leading networks, our proposed network can
successfully segment tumour regions with nearly overlapping margins, indicated in the
red boxes. Overall, our proposed model can capture more refined features and achieve
state-of-the-art accuracy in tumour segmentation tasks.
Bioengineering 2023, 10, 957 11 of 16
Figure 7. Segmentation results of different networks on the DigestPath2019 dataset. In the superim-
posed images, the areas marked in green represent the ground truth.
Bioengineering 2023, 10, 957 12 of 16
Table 4. Comparative results for tumour segmentation on the DigestPath2019 dataset. The best
results are marked in bold.
Table 5. Comparative results for gland segmentation on the GlaS dataset. The best results are marked
in bold.
Figure 8. Segmentation results of different networks on the GlaS dataset. In the superimposed images,
the areas marked in green represent the ground truth.
Bioengineering 2023, 10, 957 14 of 16
5. Conclusions
In this paper, we propose a hybrid deep learning framework for segmenting tumours
in WSIs. Our model employs an encoder–decoder architecture, with a newly designed RGS
block and a BoT block in the decoder part. These blocks are implemented to capture more
refined features and improve network stability, particularly when working with small batch
sizes. To evaluate the performance of our approach, we conducted extensive experiments
on the DigestPath2019 and GlaS datasets, and the results indicate that our model achieved
state-of-the-art segmentation accuracy.
Our proposed framework is generic and can be easily applied to other histopathology
image analysis tasks. In addition, the decoder architecture proposed in this study is flexible
and can be incorporated into other deep CNNs for histopathology image analysis. However,
we are yet to conduct experiments using natural images; therefore the superiority of our
approach in this context cannot be guaranteed. We consider this an open problem and plan
to conduct further research to provide a theoretical analysis with complete proof.
Author Contributions: Conceptualization, T.Z. and C.F.; methodology, T.Z. and C.F.; validation,
T.Z.; formal analysis, T.Z., M.T. and C.-W.S.; investigation, T.Z. and C.F.; resources, T.Z.; data curation,
T.Z., H.M. and C.F.; writing—original draft preparation, T.Z.; writing—review and editing, T.Z., C.F.
and C.-W.S.; visualization, T.Z.; supervision, T.Z.; project administration, T.Z.; funding acquisition,
C.F. All authors have read and agreed to the published version of the manuscript.
Funding: This research is supported by the National Natural Science Foundation of China (No.
62032013), and the Fundamental Research Funds for the Central Universities (No. N2224001-7).
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The data that support the findings of this study are available from the
corresponding author upon reasonable request.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Siegel, R.L.; Miller, K.D.; Goding Sauer, A.; Fedewa, S.A.; Butterly, L.F.; Anderson, J.C.; Cercek, A.; Smith, R.A.; Jemal, A.
Colorectal cancer statistics, 2020. CA Cancer J. Clin. 2020, 70, 145–164. [CrossRef] [PubMed]
2. Xia, C.; Dong, X.; Li, H.; Cao, M.; Sun, D.; He, S.; Yang, F.; Yan, X.; Zhang, S.; Li, N.; et al. Cancer statistics in China and United
States, 2022: Profiles, trends, and determinants. Chin. Med. J. 2022, 135, 584–590. [CrossRef]
3. Vega, P.; Valentín, F.; Cubiella, J. Colorectal cancer diagnosis: Pitfalls and opportunities. World J. Gastrointest. Oncol. 2015, 7, 422.
[CrossRef] [PubMed]
4. Song, W.; Fu, C.; Zheng, Y.; Cao, L.; Tie, M. A practical medical image cryptosystem with parallel acceleration. J. Ambient. Intell.
Humaniz. Comput. 2022, 14, 9853–9867. [CrossRef]
5. Wang, S.; Yang, D.M.; Rong, R.; Zhan, X.; Xiao, G. Pathology image analysis using segmentation deep learning algorithms. Am. J.
Pathol. 2019, 189, 1686–1698. [CrossRef]
6. Kumar, N.; Gupta, R.; Gupta, S. Whole slide imaging (WSI) in pathology: Current perspectives and future directions. J. Digit.
Imaging 2020, 33, 1034–1040. [CrossRef]
7. Wright, A.M.; Smith, D.; Dhurandhar, B.; Fairley, T.; Scheiber-Pacht, M.; Chakraborty, S.; Gorman, B.K.; Mody, D.; Coffey, D.M.
Digital slide imaging in cervicovaginal cytology: A pilot study. Arch. Pathol. Lab. Med. 2013, 137, 618–624. [CrossRef]
8. Zhang, W.; Fu, C.; Chang, X.; Zhao, T.; Li, X.; Sham, C.W. A more compact object detector head network with feature enhancement
and relational reasoning. Neurocomputing 2022, 499, 23–34. [CrossRef]
9. Xu, J.; Luo, X.; Wang, G.; Gilmore, H.; Madabhushi, A. A deep convolutional neural network for segmenting and classifying
epithelial and stromal regions in histopathological images. Neurocomputing 2016, 191, 214–223. [CrossRef]
10. Liu, Y.; Gadepalli, K.; Norouzi, M.; Dahl, G.E.; Kohlberger, T.; Boyko, A.; Venugopalan, S.; Timofeev, A.; Nelson, P.Q.; Corrado,
G.S.; et al. Detecting Cancer Metastases on Gigapixel Pathology Images. arXiv 2017, arXiv:1703.02442.
11. Wang, S.; Chen, A.; Yang, L.; Cai, L.; Xie, Y.; Fujimoto, J.; Gazdar, A.; Xiao, G. Comprehensive analysis of lung cancer pathology
images to discover tumor shape and boundary features that predict survival outcome. Sci. Rep. 2018, 8, 10393. [CrossRef]
[PubMed]
12. Johnson, J.W. Adapting mask-rcnn for automatic nucleus segmentation. arXiv 2018, arXiv:1805.00500.
13. Fan, K.; Wen, S.; Deng, Z. Deep learning for detecting breast cancer metastases on WSI. In Innovation in Medicine and Healthcare
Systems, and Multimedia; Springer: Berlin/Heidelberg, Germany, 2019; pp. 137–145.
Bioengineering 2023, 10, 957 15 of 16
14. Cho, S.; Jang, H.; Tan, J.W.; Jeong, W.K. DeepScribble: Interactive Pathology Image Segmentation Using Deep Neural Networks
with Scribbles. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), Nice, France, 13–16
April 2021; pp. 761–765.
15. Zhai, Z.; Wang, C.; Sun, Z.; Cheng, S.; Wang, K. Deep Neural Network Guided by Attention Mechanism for Segmentation of
Liver Pathology Image. In Proceedings of the 2021 Chinese Intelligent Systems Conference, Fuzhou, China, 16–17 October 2021;
Springer: Berlin/Heidelberg, Germany, 2022; pp. 425–435.
16. Deng, X.; Liu, E.; Li, S.; Duan, Y.; Xu, M. Interpretable Multi-Modal Image Registration Network Based on Disentangled
Convolutional Sparse Coding. IEEE Trans. Image Process. 2023, 32, 1078–1091. [CrossRef] [PubMed]
17. Jin, K.; Yan, Y.; Wang, S.; Yang, C.; Chen, M.; Liu, X.; Terasaki, H.; Yeo, T.H.; Singh, N.G.; Wang, Y.; et al. iERM: An Interpretable
Deep Learning System to Classify Epiretinal Membrane for Different Optical Coherence Tomography Devices: A Multi-Center
Analysis. J. Clin. Med. 2023, 12, 400. [CrossRef] [PubMed]
18. Xiong, S.; Li, B.; Zhu, S. DCGNN: A single-stage 3D object detection network based on density clustering and graph neural
network. Complex Intell. Syst. 2022, 9, 3399–3408. [CrossRef]
19. Qaiser, T.; Tsang, Y.W.; Taniyama, D.; Sakamoto, N.; Nakane, K.; Epstein, D.; Rajpoot, N. Fast and accurate tumor segmentation of
histology images using persistent homology and deep convolutional features. Med. Image Anal. 2019, 55, 1–14. [CrossRef]
20. Zhu, C.; Mei, K.; Peng, T.; Luo, Y.; Liu, J.; Wang, Y.; Jin, M. Multi-level colonoscopy malignant tissue detection with adversarial
CAC-UNet. Neurocomputing 2021, 438, 165–183. [CrossRef]
21. Feng, R.; Liu, X.; Chen, J.; Chen, D.Z.; Gao, H.; Wu, J. A deep learning approach for colonoscopy pathology WSI analysis:
Accurate segmentation and classification. IEEE J. Biomed. Health Inform. 2020, 25, 3700–3708. [CrossRef]
22. Song, W.; Fu, C.; Zheng, Y.; Cao, L.; Tie, M.; Sham, C.W. Protection of image ROI using chaos-based encryption and DCNN-based
object detection. Neural Comput. Appl. 2022, 34, 5743–5756. [CrossRef]
23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
24. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C. GhostNet: More Features From Cheap Operations. In Proceedings of the 2020
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020.
25. Luo, P.; Ren, J.; Peng, Z.; Zhang, R.; Li, J. Differentiable Learning-to-Normalize via Switchable Normalization. In Proceedings of
the 7th International Conference on Learning Representations, ICLR, New Orleans, LA, USA, 6–9 May 2019.
26. Srinivas, A.; Lin, T.Y.; Parmar, N.; Shlens, J.; Abbeel, P.; Vaswani, A. Bottleneck transformers for visual recognition. In Proceedings
of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16514–16524.
27. Cordonnier, J.B.; Loukas, A.; Jaggi, M. On the Relationship between Self-Attention and Convolutional Layers. In Proceedings of
the 8th International Conference on Learning Representations, ICLR, Addis Ababa, Ethiopia, 26–30 April 2020.
28. Dumoulin, V.; Visin, F. A guide to convolution arithmetic for deep learning. Stat 2018, 1050, 11. [CrossRef]
29. Luo, P.; Zhang, R.; Ren, J.; Peng, Z.; Li, J. Switchable Normalization for Learning-to-Normalize Deep Representation. IEEE Trans.
Pattern Anal. Mach. Intell. 2021, 43, 712–728. [CrossRef] [PubMed]
30. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. PMLR 2015,
37, 448–456.
31. Huang, X.; Belongie, S. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the IEEE
International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1510–1519.
32. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer normalization. arXiv 2016, arXiv:1607.06450.
33. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In
Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571.
34. Qadri, S.F.; Lin, H.; Shen, L.; Ahmad, M.; Qadri, S.; Khan, S.; Khan, M.; Zareen, S.S.; Akbar, M.A.; Bin Heyat, M.B.; et al. CT-Based
Automatic Spine Segmentation Using Patch-Based Deep Learning. Int. J. Intell. Syst. 2023, 2023, 2345835. [CrossRef]
35. Da, Q.; Huang, X.; Li, Z.; Zuo, Y.; Zhang, C.; Liu, J.; Chen, W.; Li, J.; Xu, D.; Hu, Z.; et al. DigestPath: A benchmark dataset
with challenge review for the pathological detection and segmentation of digestive-system. Med. Image Anal. 2022, 80, 102485.
[CrossRef]
36. Sirinukunwattana, K.; Pluim, J.P.; Chen, H.; Qi, X.; Heng, P.A.; Guo, Y.B.; Wang, L.Y.; Matuszewski, B.J.; Bruni, E.; Sanchez, U.;
et al. Gland segmentation in colon histology images: The glas challenge contest. Med. Image Anal. 2017, 35, 489–502. [CrossRef]
37. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the
International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October
2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241.
38. Sri, S.V.; Kavya, S. Lung Segmentation Using Deep Learning. Asian J. Appl. Sci. Technol. AJAST 2021, 5, 10–19. [CrossRef]
39. Pravitasari, A.A.; Iriawan, N.; Almuhayar, M.; Azmi, T.; Irhamah, I.; Fithriasari, K.; Purnami, S.W.; Ferriastuti, W. UNet-VGG16
with transfer learning for MRI-based brain tumor segmentation. TELKOMNIKA Telecommun. Comput. Electron. Control. 2020,
18, 1310–1318. [CrossRef]
40. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. ResUNet-a: A deep learning framework for semantic segmentation of remotely
sensed data. ISPRS J. Photogramm. Remote Sens. 2020, 162, 94–114. [CrossRef]
41. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.W.; Heng, P.A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor
segmentation from CT volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [CrossRef]
Bioengineering 2023, 10, 957 16 of 16
42. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image
Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision
Support, Granada, Spain, 20 September 2018; Stoyanov, D., Taylor, Z., Carneiro, G., Syeda-Mahmood, T., Martel, A., Maier-Hein,
L., Tavares, J.M.R., Bradley, A., Papa, J.P., Belagiannis, V., et al., Eds.; Springer International Publishing: Cham, Switzerland, 2018;
pp. 3–11.
43. Huang, H.; Lin, L.; Tong, R.; Hu, H.; Zhang, Q.; Iwamoto, Y.; Han, X.; Chen, Y.W.; Wu, J. UNet 3+: A Full-Scale Connected UNet
for Medical Image Segmentation. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech
and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [CrossRef]
44. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional
nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [CrossRef] [PubMed]
45. Chen, H.; Qi, X.; Yu, L.; Heng, P.A. DCAN: Deep contour-aware networks for accurate gland segmentation. In Proceedings of the
IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2487–2496.
46. Wang, X.; Yao, L.; Wang, X.; Paik, H.Y.; Wang, S. Global Convolutional Neural Processes. In Proceedings of the 2021 IEEE
International Conference on Data Mining (ICDM), Auckland, New Zealand, 7–10 December 2021; pp. 699–708.
47. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.
IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.