[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (534)

Search Parameters:
Keywords = Mask RCNN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
32 pages, 11565 KiB  
Article
Advanced Segmentation of Gastrointestinal (GI) Cancer Disease Using a Novel U-MaskNet Model
by Aditya Pal, Hari Mohan Rai, Mohamed Ben Haj Frej and Abdul Razaque
Life 2024, 14(11), 1488; https://doi.org/10.3390/life14111488 - 15 Nov 2024
Viewed by 339
Abstract
The purpose of this research is to contribute to the development of approaches for the classification and segmentation of various gastrointestinal (GI) cancer diseases, such as dyed lifted polyps, dyed resection margins, esophagitis, normal cecum, normal pylorus, normal Z line, polyps, and ulcerative [...] Read more.
The purpose of this research is to contribute to the development of approaches for the classification and segmentation of various gastrointestinal (GI) cancer diseases, such as dyed lifted polyps, dyed resection margins, esophagitis, normal cecum, normal pylorus, normal Z line, polyps, and ulcerative colitis. This research is relevant and essential because of the current challenges related to the absence of efficient diagnostic tools for early diagnostics of GI cancers, which are fundamental for improving the diagnosis of these common diseases. To address the above challenges, we propose a new hybrid segmentation model, U-MaskNet, which is a combination of U-Net and Mask R-CNN models. Here, U-Net is utilized for pixel-wise classification and Mask R-CNN for instance segmentation, together forming a solution for classifying and segmenting GI cancer. The Kvasir dataset, which includes 8000 endoscopic images of various GI cancers, is utilized to validate the proposed methodology. The experimental results clearly demonstrated that the novel proposed model provided superior segmentation compared to other well-known models, such as DeepLabv3+, FCN, and DeepMask, as well as improved classification performance compared to state-of-the-art (SOTA) models, including LeNet-5, AlexNet, VGG-16, ResNet-50, and the Inception Network. The quantitative analysis revealed that our proposed model outperformed the other models, achieving a precision of 98.85%, recall of 98.49%, and F1 score of 98.68%. Additionally, the novel model achieved a Dice coefficient of 94.35% and IoU of 89.31%. Consequently, the developed model increased the accuracy and reliability in detecting and segmenting GI cancer, and it was proven that the proposed model can potentially be used for improving the diagnostic process and, consequently, patient care in the clinical environment. This work highlights the benefits of integrating the U-Net and Mask R-CNN models, opening the way for further research in medical image segmentation. Full article
(This article belongs to the Section Medical Research)
Show Figures

Figure 1

Figure 1
<p>Proposed architecture of U-MaskNet used in our research for GI image segmentation.</p>
Full article ">Figure 2
<p>Dataset overview: sample GI images.</p>
Full article ">Figure 3
<p>Detailed architecture of the VGG19 convolutional neural network.</p>
Full article ">Figure 4
<p>Proposed U-MaskNet architecture used in our methodology.</p>
Full article ">Figure 5
<p>Training curves of the DeepLabv3+ model: (<b>a</b>) Dice plot, (<b>b</b>) IoU plot, (<b>c</b>) loss plot, (<b>d</b>) precision plot, and (<b>e</b>) recall plot.</p>
Full article ">Figure 6
<p>Training curves of the FCN model: (<b>a</b>) Dice plot, (<b>b</b>) IoU plot, (<b>c</b>) loss plot, (<b>d</b>) precision plot, and (<b>e</b>) recall plot.</p>
Full article ">Figure 7
<p>Training curves of the DeepMask model: (<b>a</b>) Dice plot, (<b>b</b>) IoU plot, (<b>c</b>) loss plot, (<b>d</b>) precision plot, and (<b>e</b>) recall plot.</p>
Full article ">Figure 8
<p>Training curves of the U-MaskNet model: (<b>a</b>) Dice plot, (<b>b</b>) IoU plot, (<b>c</b>) loss plot, (<b>d</b>) precision plot, and (<b>e</b>) recall plot.</p>
Full article ">Figure 9
<p>Qualitative classification results of GI diseases.</p>
Full article ">Figure 10
<p>Segmentation performance of GI cancer using U-MaskNet.</p>
Full article ">Figure 11
<p>(<b>a</b>) Training set confusion matrix, (<b>b</b>) validation set confusion matrix, and (<b>c</b>) test set confusion matrix.</p>
Full article ">Figure 11 Cont.
<p>(<b>a</b>) Training set confusion matrix, (<b>b</b>) validation set confusion matrix, and (<b>c</b>) test set confusion matrix.</p>
Full article ">Figure 12
<p>Performance comparison of segmentation models.</p>
Full article ">Figure 13
<p>Visualized performance of the proposed U-MaskNet model compared to other state-of-the-art models.</p>
Full article ">
28 pages, 45529 KiB  
Article
High-Quality Damaged Building Instance Segmentation Based on Improved Mask Transfiner Using Post-Earthquake UAS Imagery: A Case Study of the Luding Ms 6.8 Earthquake in China
by Kangsan Yu, Shumin Wang, Yitong Wang and Ziying Gu
Remote Sens. 2024, 16(22), 4222; https://doi.org/10.3390/rs16224222 - 13 Nov 2024
Viewed by 466
Abstract
Unmanned aerial systems (UASs) are increasingly playing a crucial role in earthquake emergency response and disaster assessment due to their ease of operation, mobility, and low cost. However, post-earthquake scenes are complex, with many forms of damaged buildings. UAS imagery has a high [...] Read more.
Unmanned aerial systems (UASs) are increasingly playing a crucial role in earthquake emergency response and disaster assessment due to their ease of operation, mobility, and low cost. However, post-earthquake scenes are complex, with many forms of damaged buildings. UAS imagery has a high spatial resolution, but the resolution is inconsistent between different flight missions. These factors make it challenging for existing methods to accurately identify individual damaged buildings in UAS images from different scenes, resulting in coarse segmentation masks that are insufficient for practical application needs. To address these issues, this paper proposed DB-Transfiner, a building damage instance segmentation method for post-earthquake UAS imagery based on the Mask Transfiner network. This method primarily employed deformable convolution in the backbone network to enhance adaptability to collapsed buildings of arbitrary shapes. Additionally, it used an enhanced bidirectional feature pyramid network (BiFPN) to integrate multi-scale features, improving the representation of targets of various sizes. Furthermore, a lightweight Transformer encoder has been used to process edge pixels, enhancing the efficiency of global feature extraction and the refinement of target edges. We conducted experiments on post-disaster UAS images collected from the 2022 Luding earthquake with a surface wave magnitude (Ms) of 6.8 in the Sichuan Province of China. The results demonstrated that the average precisions (AP) of DB-Transfiner, APbox and APseg, are 56.42% and 54.85%, respectively, outperforming all other comparative methods. Our model improved the original model by 5.00% and 4.07% in APbox and APseg, respectively. Importantly, the APseg of our model was significantly higher than the state-of-the-art instance segmentation model Mask R-CNN, with an increase of 9.07%. In addition, we conducted applicability testing, and the model achieved an average correctness rate of 84.28% for identifying images from different scenes of the same earthquake. We also applied the model to the Yangbi earthquake scene and found that the model maintained good performance, demonstrating a certain level of generalization capability. This method has high accuracy in identifying and assessing damaged buildings after earthquakes and can provide critical data support for disaster loss assessment. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>The study area and UAS orthophotos after the earthquake in Luding County, Sichuan Province. (<b>A</b>) study area; (<b>B</b>) UAS orthophotos: (<b>a</b>,<b>c</b>) Moxi town; (<b>b</b>,<b>d</b>,<b>g</b>) Detuo town; (<b>e</b>) Fawang village; (<b>f</b>) Wandong village.</p>
Full article ">Figure 2
<p>The samples of damaged buildings and labels: (<b>a</b>) Field investigation photos; (<b>b</b>) UAS images, the red fan-shaped marker representing the viewing angle of the observation location; (<b>c</b>) Labeled bounding boxes; (<b>d</b>) Labeled instance masks, the color of the polygon masks represents different instance objects.</p>
Full article ">Figure 3
<p>The network architecture of Mask Transfiner.</p>
Full article ">Figure 4
<p>The improved network architecture for DB-Transfiner. Deformable convolution is employed in the backbone. The FPN is replaced by enhanced BiFPN to fuse the multi-scale features, and, in this study, a lightweight sequence encoder is adopted for efficiency.</p>
Full article ">Figure 5
<p>Deformable convolution feature extraction module. Arrows indicate the type of convolution used at each stage. The first two stages use standard convolution, and the last three stages use deformable convolution. (<b>a</b>) Standard convolution; (<b>b</b>) Deformable convolution.</p>
Full article ">Figure 6
<p>Replacing FPN with enhanced BiFPN to improve feature fusion network.</p>
Full article ">Figure 7
<p>Lightweight sequence encoder to improve the efficiency of the network, using a Transformer structure with an eight-headed self-attention mechanism instead of three Transformer structures with four-headed self-attention mechanisms.</p>
Full article ">Figure 8
<p>Loss curve during DB-Transfiner training.</p>
Full article ">Figure 9
<p>Comparison of the performance of all models based on the metrics <span class="html-italic">AP</span> (%).</p>
Full article ">Figure 10
<p>Visualization of the prediction results of different network models. The colored bounding boxes and polygons represent the detection and segmentation results, respectively. (<b>a</b>) Annotated images; (<b>b</b>) Mask R-CNN; (<b>c</b>) Mask Transfiner; (<b>d</b>) DB-Transfiner.</p>
Full article ">Figure 11
<p>Visualization of instance mask results of different network models. The colored polygons represent the recognized instance objects. ① and ② represent two typical damaged buildings with the same level of destruction. (<b>a</b>) Original images; (<b>b</b>) Annotated results; (<b>c</b>) Mask R-CNN; (<b>d</b>) Mask Transfiner; (<b>e</b>) DB-Transfiner.</p>
Full article ">Figure 12
<p>Visualization of heatmaps: (<b>a</b>) The original images; (<b>b</b>) The heatmaps of Conv2_x layer of the DCNM; (<b>c</b>) The heatmaps of Conv5_x layer of the DCNM; (<b>d</b>) The heatmaps of <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>N</mi> </mrow> <mrow> <mn>5</mn> </mrow> </msub> </mrow> </semantics></math> layer of the MEFM; (<b>e</b>) The final results. The colored borders represent the model’s predicted different instance objects.</p>
Full article ">Figure 13
<p>The visualization of feature maps before and after the LTGM. The colored borders represent the different instance objects.</p>
Full article ">Figure 14
<p>Results of damaged building classification in Fawang village (<a href="#remotesensing-16-04222-f001" class="html-fig">Figure 1</a>B(e)). Red indicates correct detections, green indicates incorrect detections, and yellow indicates missed.</p>
Full article ">Figure 15
<p>Results of damaged building classification in Wandong village and Detuo town (<a href="#remotesensing-16-04222-f001" class="html-fig">Figure 1</a>B(f,g)). Red indicates correct detections, green indicates incorrect detections, and yellow indicates missed.</p>
Full article ">Figure 16
<p>Example of UAV imagery from the Yangbi earthquake in Yunnan, China: (<b>a</b>) Huaian village; (<b>b</b>) Yangbi town.</p>
Full article ">Figure 17
<p>UAS imagery samples of damaged buildings from the Yangbi earthquake. (<b>a</b>) The red irregular polygons denote the damaged buildings. (<b>b</b>) The bounding boxes and polygon masks are the visualized results of our model. The colors represent different instance objects.</p>
Full article ">Figure 18
<p>Examples of densely built-up areas. The red boxes indicate buildings with blurred contour information caused by shadows and occlusions.</p>
Full article ">
16 pages, 6553 KiB  
Article
Cucumber Leaf Segmentation Based on Bilayer Convolutional Network
by Tingting Qian, Yangxin Liu, Shenglian Lu, Linyi Li, Xiuguo Zheng, Qingqing Ju, Yiyang Li, Chun Xie and Guo Li
Agronomy 2024, 14(11), 2664; https://doi.org/10.3390/agronomy14112664 - 12 Nov 2024
Viewed by 415
Abstract
When monitoring crop growth using top-down images of the plant canopies, leaves in agricultural fields appear very dense and significantly overlap each other. Moreover, the image can be affected by external conditions such as background environment and light intensity, impacting the effectiveness of [...] Read more.
When monitoring crop growth using top-down images of the plant canopies, leaves in agricultural fields appear very dense and significantly overlap each other. Moreover, the image can be affected by external conditions such as background environment and light intensity, impacting the effectiveness of image segmentation. To address the challenge of segmenting dense and overlapping plant leaves under natural lighting conditions, this study employed a Bilayer Convolutional Network (BCNet) method for accurate leaf segmentation across various lighting environments. The major contributions of this study are as follows: (1) Utilized Fully Convolutional Object Detection (FCOS) for plant leaf detection, incorporating ResNet-50 with the Convolutional Block Attention Module (CBAM) and Feature Pyramid Network (FPN) to enhance Region of Interest (RoI) feature extraction from canopy top-view images. (2) Extracted the sub-region of the RoI based on the position of the detection box, using this region as input for the BCNet, ensuring precise segmentation. (3) Utilized instance segmentation of canopy top-view images using BCNet, improving segmentation accuracy. (4) Applied the Varifocal Loss Function to improve the classification loss function in FCOS, leading to better performance metrics. The experimental results on cucumber canopy top-view images captured in glass greenhouse and plastic greenhouse environments show that our method is highly effective. For cucumber leaves at different growth stages and under various lighting conditions, the Precision, Recall and Average Precision (AP) metrics for object recognition are 97%, 94% and 96.57%, respectively. For instance segmentation, the Precision, Recall and Average Precision (AP) metrics are 87%, 83% and 84.71%, respectively. Our algorithm outperforms commonly used deep learning algorithms such as Faster R-CNN, Mask R-CNN, YOLOv4 and PANet, showcasing its superior capability in complex agricultural settings. The results of this study demonstrate the potential of our method for accurate recognition and segmentation of highly overlapping leaves in diverse agricultural environments, significantly contributing to the application of deep learning algorithms in smart agriculture. Full article
(This article belongs to the Special Issue AI, Sensors and Robotics for Smart Agriculture—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Bilayer decomposition diagram.</p>
Full article ">Figure 2
<p>Flow chart of image segmentation based on improved BCNet.</p>
Full article ">Figure 3
<p>Schematic diagram of image annotation method. <span class="html-fig-inline" id="agronomy-14-02664-i001"><img alt="Agronomy 14 02664 i001" src="/agronomy/agronomy-14-02664/article_deploy/html/images/agronomy-14-02664-i001.png"/></span> means labeled, <span class="html-fig-inline" id="agronomy-14-02664-i002"><img alt="Agronomy 14 02664 i002" src="/agronomy/agronomy-14-02664/article_deploy/html/images/agronomy-14-02664-i002.png"/></span> means unlabeled.</p>
Full article ">Figure 4
<p>Schematic diagram of image expansion scheme.</p>
Full article ">Figure 5
<p>Segmentation effect of cucumber plant images in glass greenhouses. (<b>a</b>) Early Growth Stage, Sunny. (<b>b</b>) Early Growth Stage, Cloudy. (<b>c</b>) Metaphase Growth Stage, Sunny. (<b>d</b>) Metaphase Growth Stage. Cloudy. (<b>e</b>) Terminal Growth Stage, Sunny. (<b>f</b>) Terminal Growth Stage, Cloudy.</p>
Full article ">Figure 6
<p>Effect of cucumber plant image segmentation in plastic greenhouses. (<b>a</b>) Early Growth Stage, Sunny. (<b>b</b>) Early Growth Stage, Cloudy. (<b>c</b>) Terminal Growth Stage, Sunny. (<b>d</b>) Terminal Growth Stage, Cloudy.</p>
Full article ">Figure 7
<p>Target recognition and instance segmentation P-R curve of six models: (<b>a</b>) Object detection P-R curve; (<b>b</b>) Example split P-R curve.</p>
Full article ">Figure 8
<p>Effect of cucumber plant image segmentation in glass greenhouse: (<b>a</b>) improved BCNet; (<b>b</b>) BCNet; (<b>c</b>) PANet; (<b>d</b>) Mask R-CNN; (<b>e</b>) YOLOv4; (<b>f</b>) Faster R-CNN.</p>
Full article ">
17 pages, 6601 KiB  
Article
Deep Learning-Driven Virtual Furniture Replacement Using GANs and Spatial Transformer Networks
by Resmy Vijaykumar, Muneer Ahmad, Maizatul Akmar Ismail, Iftikhar Ahmad and Neelum Noreen
Mathematics 2024, 12(22), 3513; https://doi.org/10.3390/math12223513 - 11 Nov 2024
Viewed by 424
Abstract
This study proposes a Generative Adversarial Network (GAN)-based method for virtual furniture replacement within indoor scenes. The proposed method addresses the challenge of accurately positioning new furniture in an indoor space by combining image reconstruction with geometric matching through combining spatial transformer networks [...] Read more.
This study proposes a Generative Adversarial Network (GAN)-based method for virtual furniture replacement within indoor scenes. The proposed method addresses the challenge of accurately positioning new furniture in an indoor space by combining image reconstruction with geometric matching through combining spatial transformer networks and GANs. The system leverages deep learning architectures like Mask R-CNN for executing image segmentation and generating masks, and it employs DeepLabv3+, EdgeConnect algorithms, and ST-GAN networks for carrying out virtual furniture replacement. With the proposed system, furniture shoppers can obtain a virtual shopping experience, providing an easier way to understand the aesthetic effects of furniture rearrangement without putting in effort to physically move furniture. The proposed system has practical applications in the furnishing industry and interior design practices, providing a cost-effective and efficient alternative to physical furniture replacement. The results indicate that the proposed method achieves accurate positioning of new furniture in indoor scenes with minimal distortion or displacement. The proposed system is limited to 2D front-view images of furniture and indoor scenes. Future work would involve synthesizing 3D scenes and expanding the system to replace furniture images photographed from different angles. This would enhance the efficiency and practicality of the proposed system for virtual furniture replacement in indoor scenes. Full article
(This article belongs to the Special Issue Advances and Applications of Artificial Intelligence Technologies)
Show Figures

Figure 1

Figure 1
<p>Virtual Furniture Replacement Process Flow.</p>
Full article ">Figure 2
<p>Convolutional Neural Network [<a href="#B2-mathematics-12-03513" class="html-bibr">2</a>].</p>
Full article ">Figure 3
<p>DeepLab v3+ Architecture [<a href="#B10-mathematics-12-03513" class="html-bibr">10</a>].</p>
Full article ">Figure 4
<p>An edge map is created [<a href="#B4-mathematics-12-03513" class="html-bibr">4</a>].</p>
Full article ">Figure 5
<p>Structure of GAN for Image Replacement.</p>
Full article ">Figure 6
<p>STN in ST-GAN [<a href="#B3-mathematics-12-03513" class="html-bibr">3</a>] (p. 3).</p>
Full article ">Figure 7
<p>Discriminator D in ST-GAN [<a href="#B3-mathematics-12-03513" class="html-bibr">3</a>] (p. 4).</p>
Full article ">Figure 8
<p>Data Flow Diagram.</p>
Full article ">Figure 9
<p>Creation of Masks for Furniture Objects.</p>
Full article ">Figure 10
<p>Furniture Object Removal and Image Replacement.</p>
Full article ">Figure 11
<p>Image Obtained Without Use of GAN.</p>
Full article ">Figure 12
<p>Virtual Furniture Replacement Results.</p>
Full article ">Figure 13
<p>Inpainted Images.</p>
Full article ">Figure 14
<p>Partially Inpainted Images.</p>
Full article ">Figure 15
<p>Advantages of using GANs for Image Replacement.</p>
Full article ">
20 pages, 8552 KiB  
Article
A Heatmap-Supplemented R-CNN Trained Using an Inflated IoU for Small Object Detection
by Justin Butler and Henry Leung
Remote Sens. 2024, 16(21), 4065; https://doi.org/10.3390/rs16214065 - 31 Oct 2024
Viewed by 660
Abstract
Object detection architectures struggle to detect small objects across applications including remote sensing and autonomous vehicles. Specifically, for unmanned aerial vehicles, poor detection of small objects directly limits this technology’s applicability. Objects both appear smaller than they are in large-scale images captured in [...] Read more.
Object detection architectures struggle to detect small objects across applications including remote sensing and autonomous vehicles. Specifically, for unmanned aerial vehicles, poor detection of small objects directly limits this technology’s applicability. Objects both appear smaller than they are in large-scale images captured in aerial imagery and are represented by reduced information in high-altitude imagery. This paper presents a new architecture, CR-CNN, which predicts independent regions of interest from two unique prediction branches within the first stage of the network: a conventional R-CNN convolutional backbone and an hourglass backbone. Utilizing two independent sources within the first stage, our approach leads to an increase in successful predictions of regions that contain smaller objects. Anchor-based methods such as R-CNNs also utilize less than half the number of small objects compared to larger ones during training due to the poor intersection over union (IoU) scores between the generated anchors and the groundtruth—further reducing their performance on small objects. Therefore, we also propose artificially inflating the IoU of smaller objects during training using a simple, size-based Gaussian multiplier—leading to an increase in the quantity of small objects seen per training cycle based on an increase in the number of anchor–object pairs during training. This architecture and training strategy led to improved detection overall on two challenging aerial-based datasets heavily composed of small objects while predicting fewer false positives compared to Mask R-CNN. These results suggest that while new and unique architectures will continue to play a part in advancing the field of object detection, the training methodologies and strategies used will also play a valuable role. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Example Aerial-Cars image. The additional annotated objects (denoted by red squares) are primarily located towards the horizon and in the upper-left and -right corners.</p>
Full article ">Figure 2
<p>Histogram of object sizes in the synthetic and aerial-based data explored in this work.</p>
Full article ">Figure 3
<p>Complete CR-CNN architecture. The portions within the red hashed borders are standard implementations of Mask R-CNN utilizing ResNeXt-101 and PANet. The secondary RoI prediction branch proposed in this work is contained within the blue hashed border with the RPN utilizing the architecture of CR-CNN-off or CR-CNN-abs, as discussed in Materials and Methods.</p>
Full article ">Figure 4
<p>A cropped image from the Aerial-Cars detection dataset demonstrating how a portion of the anchors (green) is overlaid onto each object’s groundtruth label (red) in order to calculate the IoU. Artificially inflating the IoU of the object–anchor pairs for small objects increases the number of anchors used during training—effectively increasing the object size of the label annotation.</p>
Full article ">Figure 5
<p>Various IoU inflation functions tested for improved detection of small objects. The Gaussian function resulted in the largest increase in small object performance with minimal decreases in larger object accuracy.</p>
Full article ">Figure 6
<p>CR-CNN predictions on Aerial-Cars Figure MOS157 [<a href="#B54-remotesensing-16-04065" class="html-bibr">54</a>] with a prediction threshold of 0.25. Predictions are represented by each blue box.</p>
Full article ">Figure 7
<p>CR-CNN predictions on Aerial-Cars Figure DJI-0762-00001 [<a href="#B54-remotesensing-16-04065" class="html-bibr">54</a>] with a prediction threshold of 0.25. This image represents one of the poorer performances of CR-CNN on the data, with missed predictions in regions with changing light and contrast. Predictions are represented by each blue box.</p>
Full article ">Figure 8
<p>Average number of matched anchors.</p>
Full article ">
19 pages, 8993 KiB  
Article
Segmentation-Based Detection for Luffa Seedling Grading Using the Seg-FL Model
by Sheng Jiang, Fangnan Xie, Jiangbo Ao, Yechen Wei, Jingye Lu, Shilei Lyu and Zhen Li
Agronomy 2024, 14(11), 2557; https://doi.org/10.3390/agronomy14112557 - 31 Oct 2024
Viewed by 316
Abstract
This study addresses the issue of inaccurate and error-prone grading judgments in luffa plug seedlings. A new Seg-FL seedling segmentation model is proposed as an extension of the YOLOv5s-Seg model. The small leaves of early-stage luffa seedlings are liable to be mistaken for [...] Read more.
This study addresses the issue of inaccurate and error-prone grading judgments in luffa plug seedlings. A new Seg-FL seedling segmentation model is proposed as an extension of the YOLOv5s-Seg model. The small leaves of early-stage luffa seedlings are liable to be mistaken for impurities in the plug trays. To address this issue, cross-scale connections and weighted feature fusion are introduced in order to integrate feature information from different levels, thereby improving the recognition and segmentation accuracy of seedlings or details by refining the PANet structure. To address the ambiguity of seedling edge information during segmentation, an efficient channel attention module is incorporated to enhance the network’s focus on seedling edge information and suppress irrelevant features, thus sharpening the model’s focus on luffa seedlings. By optimizing the CIoU loss function, the calculation of overlapping areas, center point distances, and aspect ratios between predicted and ground truth boxes is preserved, thereby accelerating the convergence process and reducing the computational resource requirements on edge devices. The experimental results demonstrate that the proposed model attains a mean average precision of 97.03% on a self-compiled luffa plug seedling dataset, representing a 6.23 percentage point improvement over the original YOLOv5s-Seg. Furthermore, compared to the YOLACT++, FCN, and Mask R-CNN segmentation models, the improved model displays increases in [email protected] of 12.93%, 13.73%, and 10.53%, respectively, and improvements in precision of 15.73%, 16.93%, and 13.33%, respectively. This research not only validates the viability of the enhanced model for luffa seedling grading but also provides tangible technical support for the automation of grading in agricultural production. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Data collection points; (<b>b</b>) greenhouse; (<b>c</b>) typical hole tray seedling.</p>
Full article ">Figure 1 Cont.
<p>(<b>a</b>) Data collection points; (<b>b</b>) greenhouse; (<b>c</b>) typical hole tray seedling.</p>
Full article ">Figure 2
<p>Image acquisition system.</p>
Full article ">Figure 3
<p>(<b>a</b>) No seedings; (<b>b</b>) weak seedings; (<b>c</b>) strong seedings.</p>
Full article ">Figure 4
<p>(<b>a</b>) Flip horizontal; (<b>b</b>) gray processing; (<b>c</b>) random cropping.</p>
Full article ">Figure 5
<p>YOLOv5s seg network structure.</p>
Full article ">Figure 6
<p>(<b>a</b>) The architectures of PANet; (<b>b</b>) the architectures of BiFPN.</p>
Full article ">Figure 7
<p>ECA mechanism structure diagram.</p>
Full article ">Figure 8
<p>Schematic diagram of angle cost calculation.</p>
Full article ">Figure 9
<p>Seg FL algorithm structure diagram.</p>
Full article ">Figure 10
<p>Network training loss curve.</p>
Full article ">Figure 11
<p>P-R curves of two models. (<b>a</b>) No seedings; (<b>b</b>) weak seedings; (<b>c</b>) strong seedings.</p>
Full article ">Figure 12
<p>The segmentation result images before model improvement. (<b>a</b>) No seedings; (<b>b</b>) weak seedings; (<b>c</b>) strong seedings.</p>
Full article ">Figure 13
<p>The segmentation result images following model improvement. (<b>a</b>) No seedings; (<b>b</b>) weak seedings; (<b>c</b>) strong seedings.</p>
Full article ">
20 pages, 4073 KiB  
Article
Individual Tree Crown Detection and Classification of Live and Dead Trees Using a Mask Region-Based Convolutional Neural Network (Mask R-CNN)
by Shilong Yao, Zhenbang Hao, Christopher J. Post, Elena A. Mikhailova and Lili Lin
Forests 2024, 15(11), 1900; https://doi.org/10.3390/f15111900 - 28 Oct 2024
Viewed by 641
Abstract
Mapping the distribution of living and dead trees in forests, particularly in ecologically fragile areas where forests serve as crucial ecological environments, is essential for assessing forest health, carbon storage capacity, and biodiversity. Convolutional neural networks, including Mask R-CNN, can assist in rapid [...] Read more.
Mapping the distribution of living and dead trees in forests, particularly in ecologically fragile areas where forests serve as crucial ecological environments, is essential for assessing forest health, carbon storage capacity, and biodiversity. Convolutional neural networks, including Mask R-CNN, can assist in rapid and accurate forest monitoring. In this study, Mask R-CNN was employed to detect the crowns of Casuarina equisetifolia and to distinguish between live and dead trees in the Pingtan Comprehensive Pilot Zone, Fujian, China. High-resolution images of five plots were obtained using a multispectral Unmanned Aerial Vehicle. Six band combinations and derivatives, RGB, RGB-digital surface model (DSM), Multispectral, Multispectral-DSM, Vegetation Index, and Vegetation-Index-DSM, were used for tree crown detection and classification of live and dead trees. Five-fold cross-validation was employed to divide the manually annotated dataset of 21,800 live trees and 7157 dead trees into training and validation sets, which were used for training and validating the Mask R-CNN models. The results demonstrate that the RGB band combination achieved the most effective detection performance for live trees (average F1 score = 74.75%, IoU = 70.85%). The RGB–DSM combination exhibited the highest accuracy for dead trees (average F1 score = 71.16%, IoU = 68.28%). The detection performance for dead trees was lower than for live trees, which may be due to the similar spectral features across the images and the similarity of dead trees to the background, resulting in false identification. For the simultaneous detection of living and dead trees, the RGB combination produced the most promising results (average F1 score = 74.18%, IoU = 69.8%). It demonstrates that the Mask R-CNN model can achieve promising results for the detection of live and dead trees. Our study could provide forest managers with detailed information on the forest condition, which has the potential to improve forest management. Full article
Show Figures

Figure 1

Figure 1
<p>Experimental site: (<b>a</b>) Location of the study site; (<b>b</b>) Location and distribution of 5 plots (different colored points represent different sample plots); (<b>c</b>) Example of an UAV image in the study area, illustrated using RGB bands; (<b>d</b>) Example of manually delineated live and dead <span class="html-italic">C. equisetifolia</span> tree crowns (live trees are marked in green, dead trees in red).</p>
Full article ">Figure 2
<p>Research flowchart for this study. The light red color represents the products derived from UAV imagery. The light blue color represents the application of the Mask R-CNN model. The light green color represents a five-fold cross-validation.</p>
Full article ">Figure 3
<p>The Mask R-CNN architecture and workflow for individual tree crown detection.</p>
Full article ">Figure 4
<p>Average accuracy of individual tree crown detection and delineation for live trees using different input images. The different colors of the scatter plots represent different sample plots. (<b>a</b>) Recall for different sample plots; (<b>b</b>) Precision for different sample plots; (<b>c</b>) F1 Score of different sample plots. (<b>d</b>) IoU for different sample plots.</p>
Full article ">Figure 5
<p>Average accuracy of individual tree crown detection and delineation for dead trees using different input images. The different colors of the scatter plots represent different sample plots. (<b>a</b>) Recall for different sample plots. (<b>b</b>) Precision for different sample plots. (<b>c</b>) F1 score of different sample plots. (<b>d</b>) IoU for different sample plots.</p>
Full article ">Figure 6
<p>Average accuracy of individual tree crown detection and delineation for live and dead trees using different input images. The different colors of the scatter plots represent different sample plots. (<b>a</b>) Recall for different sample plots; (<b>b</b>) Precision for different sample plots; (<b>c</b>) F1 score of different sample plots; (<b>d</b>) IoU for different sample plots.</p>
Full article ">Figure 7
<p>The manually annotated living and dead tree crowns may have erroneous features due to overlapping crown conditions. (<b>a</b>,<b>b</b>) Dead tree crowns may contain features of live tree crowns. (<b>c</b>,<b>d</b>) Live tree crowns may contain features of dead tree crowns.</p>
Full article ">Figure 8
<p>Detection results of the Mask R-CNN model using different band combinations in sample plot 4 and sample plot 2. (<b>a</b>) Example of the UAV image from sample plot 4, shown in RGB bands. (<b>b</b>) Example of the UAV image from sample plot 2, shown in RGB bands. (<b>c</b>–<b>h</b>) For RGB, RGB–DSM, M, M–DSM, VI, and VI–DSM combinations. (<b>i</b>–<b>n</b>) For RGB, RGB–DSM, M, M–DSM, VI, and VI–DSM combinations.</p>
Full article ">
19 pages, 14722 KiB  
Article
Log Volume Measurement and Counting Based on Improved Cascade Mask R-CNN and Deep SORT
by Chunjiang Yu, Yongke Sun, Yong Cao, Lei Liu and Xiaotao Zhou
Forests 2024, 15(11), 1884; https://doi.org/10.3390/f15111884 - 26 Oct 2024
Viewed by 520
Abstract
Logs require multiple verifications to ensure accurate volume and quantity measurements. Log end detection is a crucial step in measuring log volume and counting logs. Currently, this task primarily relies on the Mask R-CNN instance segmentation model. However, the Feature Pyramid Network (FPN) [...] Read more.
Logs require multiple verifications to ensure accurate volume and quantity measurements. Log end detection is a crucial step in measuring log volume and counting logs. Currently, this task primarily relies on the Mask R-CNN instance segmentation model. However, the Feature Pyramid Network (FPN) in Mask R-CNN may compromise accuracy due to feature redundancy during multi-scale fusion, particularly with small objects. Moreover, counting logs in a single image is challenging due to their large size and stacking. To address the above issues, we propose an improved log segmentation model based on Cascade Mask R-CNN. This method uses ResNet for multi-scale feature extraction and integrates a hierarchical Convolutional Block Attention Module (CBAM) to refine feature weights and enhance object emphasis. Then, a Region Proposal Network (RPN) is employed to generate log segmentation proposals. Finally, combined with Deep SORT, the model tracks log ends in video streams and counts the number of logs in the stack. Experiments demonstrate the effectiveness of our method, achieving an average precision (AP) of 82.3, APs of 75.3 for small, APm of 70.9 for medium, and APl of 86.2 for large objects. These results represent improvements of 1.8%, 3.7%, 2.6%, and 1.4% over Mask R-CNN, respectively. The detection rate reached 98.6%, with a counting accuracy of 95%. Compared to manually measured volumes, our method shows a low error rate of 4.07%. Full article
(This article belongs to the Section Wood Science and Forest Products)
Show Figures

Figure 1

Figure 1
<p>Original images and augmentation images.</p>
Full article ">Figure 2
<p>The workflow.</p>
Full article ">Figure 3
<p>The overall network structure.</p>
Full article ">Figure 4
<p>Improved Backbone, ResNet50-CBAM.</p>
Full article ">Figure 5
<p>Convolutional Block Attention module. It contains a channel attention module and a spatial attention module.</p>
Full article ">Figure 6
<p>Log end images detected using different backbone networks. These images were taken in the field at customs ports. The algorithm uses Cascade Mask R-CNN.</p>
Full article ">Figure 7
<p>Heatmaps generated by different backbone networks. It compared attention of the models to objects after fusing attention mechanisms in different backbone networks.</p>
Full article ">Figure 8
<p>Deep SORT tracking results. (<b>a</b>–<b>f</b>) represent the tracking state of the log end at different frames.</p>
Full article ">Figure 9
<p>Tracking count results of Deep SORT with different output confidence levels for object detection models.</p>
Full article ">Figure 10
<p>Log stacks in outdoor scenes for log tracking counting and volume measurement experiments.</p>
Full article ">Figure 11
<p>Short and long diameter search results, where the two straight lines represent the short and long diameters, respectively.</p>
Full article ">
19 pages, 6528 KiB  
Article
Estimation of Tree Diameter at Breast Height from Aerial Photographs Using a Mask R-CNN and Bayesian Regression
by Kyeongnam Kwon, Seong-kyun Im, Sung Yong Kim, Ye-eun Lee and Chun Geun Kwon
Forests 2024, 15(11), 1881; https://doi.org/10.3390/f15111881 - 25 Oct 2024
Viewed by 734
Abstract
A probabilistic estimation model for forest biomass using unmanned aerial vehicle (UAV) photography was developed. We utilized a machine-learning-based object detection algorithm, a mask region-based convolutional neural network (Mask R-CNN), to detect trees in aerial photographs. Subsequently, Bayesian regression was used to calibrate [...] Read more.
A probabilistic estimation model for forest biomass using unmanned aerial vehicle (UAV) photography was developed. We utilized a machine-learning-based object detection algorithm, a mask region-based convolutional neural network (Mask R-CNN), to detect trees in aerial photographs. Subsequently, Bayesian regression was used to calibrate the model based on an allometric model using the estimated crown diameter (CD) obtained from aerial photographs and analyzed the diameter at breast height (DBH) data acquired through terrestrial laser scanning. The F1 score of the Mask R-CNN for individual tree detection was 0.927. Moreover, CD estimation using the Mask R-CNN was acceptable (rRMSE = 10.17%). Accordingly, the probabilistic DBH estimation model was successfully calibrated using Bayesian regression. A predictive distribution accurately predicted the validation data, with 98.6% and 56.7% of the data being within the 95% and 50% prediction intervals, respectively. Furthermore, the estimated uncertainty of the probabilistic model was more practical and reliable compared to traditional ordinary least squares (OLS). Our model can be applied to estimate forest biomass at the individual tree level. Particularly, the probabilistic approach of this study provides a benefit for risk assessments. Additionally, since the workflow is not interfered by the tree canopy, it can effectively estimate forest biomass in dense canopy conditions. Full article
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Schematics of the overall procedure for establishing a probabilistic estimation model and estimation process.</p>
Full article ">Figure 2
<p>(<b>a</b>) Location of the study site. (<b>b</b>) Aerial photographs with regions of interest.</p>
Full article ">Figure 3
<p>Mask region-based convolutional neural network (Mask R-CNN) structure.</p>
Full article ">Figure 4
<p>Schematics of crown diameter (CD) calculation methods from a segmentation mask.</p>
Full article ">Figure 5
<p>Training result of tree segmentation (<b>a</b>) without and (<b>b</b>) with large-scale jittering.</p>
Full article ">Figure 6
<p>Examples of tree detection results using a Mask region-based convolutional neural network (Mask R-CNN).</p>
Full article ">Figure 7
<p>Plots of measured and estimated crown diameter (CD) of various methods with identity lines.</p>
Full article ">Figure 8
<p>Plots of metrics from terrestrial laser scanning (TLS) measured data and allometric models. CD: crown diameter, DBH: diameter at breast height, TH: tree height.</p>
Full article ">Figure 9
<p>Scatter plots of (<b>a</b>) estimated and (<b>b</b>) measured crown diameter (CD) versus diameter at breast height (DBH) with a posterior mean from Bayesian regression and ordinary least square fitted to measured data.</p>
Full article ">Figure 10
<p>Scatter plots of estimated crown diameter (CD) using (<b>a</b>) calibration data and (<b>b</b>) validation data versus diameter at breast height (DBH) with posterior predictive distribution from Bayesian regression.</p>
Full article ">Figure 11
<p>Estimated uncertainty of the ordinary least-square-fitted function compared to the result of the probabilistic estimation model with a scatter plot of diameter at breast height (DBH) versus estimated crown diameter (CD).</p>
Full article ">Figure 12
<p>A scatter plot of estimated crown diameter (CD) versus terrestrial laser scanning (TLS) measured diameter at breast height (DBH) with probabilistic estimation model based on (red) terrestrial laser scanning (TLS) measured DBH and (blue) field survey DBH.</p>
Full article ">Figure 13
<p>A plot of allometric correlations between tree height (TH) and diameter at breast height (DBH) with measured data using terrestrial laser scanning (TLS) [<a href="#B65-forests-15-01881" class="html-bibr">65</a>,<a href="#B66-forests-15-01881" class="html-bibr">66</a>].</p>
Full article ">Figure 14
<p>Plots of probabilistic estimation for tree height (TH) using TH allometric correlations based on diameter at breast height (DBH) with measured crown diameter (CD) and TH from terrestrial laser scanning (TLS) [<a href="#B65-forests-15-01881" class="html-bibr">65</a>,<a href="#B66-forests-15-01881" class="html-bibr">66</a>].</p>
Full article ">
25 pages, 24844 KiB  
Article
Individual Tree Crown Delineation Using Airborne LiDAR Data and Aerial Imagery in the Taiga–Tundra Ecotone
by Yuanyuan Lin, Hui Li, Linhai Jing, Haifeng Ding and Shufang Tian
Remote Sens. 2024, 16(21), 3920; https://doi.org/10.3390/rs16213920 - 22 Oct 2024
Viewed by 785
Abstract
The circumpolar Taiga–Tundra Ecotone significantly influences the feedback mechanism of global climate change. Achieving large-scale individual tree crown (ITC) extraction in the transition zone is crucial for estimating vegetation biomass in the transition zone and studying plants’ response to climate change. This study [...] Read more.
The circumpolar Taiga–Tundra Ecotone significantly influences the feedback mechanism of global climate change. Achieving large-scale individual tree crown (ITC) extraction in the transition zone is crucial for estimating vegetation biomass in the transition zone and studying plants’ response to climate change. This study employed aerial images and airborne LiDAR data covering several typical transitional zone regions in northern Finland to explore the ITC delineation method based on deep learning. First, this study developed an improved multi-scale ITC delineation method to enable the semi-automatic assembly of the ITC sample collection. This approach led to the creation of an individual tree dataset containing over 20,000 trees in the transitional zone. Then, this study explored the ITC delineation method using the Mask R-CNN model. The accuracies of the Mask R-CNN model were compared with two traditional ITC delineation methods: the improved multi-scale ITC delineation method and the local maxima clustering method based on point cloud distribution. For trees with a height greater than 1.3 m, the Mask R-CNN model achieved an overall recall rate (Ar) of 96.60%. Compared to the two conventional ITC delineation methods, the Ar of Mask R-CNN showed an increase of 1.99 and 5.52 points in percentage, respectively, indicating that the Mask R-CNN model can significantly improve the accuracy of ITC delineation. These results highlight the potential of Mask R-CNN in extracting low trees with relatively small crowns in transitional zones using high-resolution aerial imagery and low-density airborne point cloud data for the first time. Full article
(This article belongs to the Special Issue Remote Sensing of Savannas and Woodlands II)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>The location of the study area and the distribution of images used for training and validation. (<b>a</b>) The location of the study area; (<b>b</b>–<b>f</b>) the false color aerial orthophotos of the five validation areas of plot 1, plot 2, plot 3, plot 4, and plot 5, respectively (shown in a combination of near-infrared, red, and green bands).</p>
Full article ">Figure 2
<p>The CHM of plot 2 (<b>a</b>) and the corresponding false color aerial orthophoto showing individual trees in 3D (<b>b</b>).</p>
Full article ">Figure 3
<p>The CHM optimization results and ITC delineation results using different filtering methods. (<b>a</b>) The original CHM; (<b>b</b>) ITC delineation results (the polygons in blue) using the original CHM; (<b>c</b>) the CHM obtained by low-pass filtering; and (<b>d</b>) ITC delineation results (the polygons in blue) using the low-pass filtered CHM.</p>
Full article ">Figure 4
<p>An example of manually delineated ITCs. (<b>a</b>) A sub-area of the study region, in which the yellow rectangular frame represents the location of plot 3; (<b>b</b>) the manually delineated tree crowns of plot 3 (overlapping the aerial imagery); and (<b>c</b>) the manually delineated tree crowns of plot 3 (overlapping the CHM).</p>
Full article ">Figure 5
<p>The flowchart of the improved MSAS method.</p>
Full article ">Figure 6
<p>An example of scale analysis and the mean value of different images derived from a series of morphologically opened images of the CHM. (<b>a</b>) A sub-area of the CHM; (<b>b</b>) the corresponding false color aerial imagery of (<b>a</b>); and (<b>c</b>) the mean values of the different images derived from a series of opened images obtained from the CHM using a disk SE with a gradually increasing diameter.</p>
Full article ">Figure 7
<p>One of the nine training images and a sub-image of the training image. (<b>a</b>) One of the nine training images; (<b>b</b>) a sub-image of (<b>a</b>) and training patterns; (<b>c</b>) CHM of the sub-image; and (<b>d</b>) CHM of the sub-image and training patterns.</p>
Full article ">Figure 8
<p>The workflow for ITC delineation using Mask R-CNN. The FCN generates labels and bounding boxes for each object, and the Fully Convolutional Nets produce masks including the pixels of each object.</p>
Full article ">Figure 9
<p>The workflow for ITC delineation using Li’s method; “<span class="html-italic">i</span>” represents the iteration count.</p>
Full article ">Figure 10
<p>The diagram of classifying target tree crowns into five categories according to their spatial relationship with reference tree crowns: (<b>a</b>) nearly matched; (<b>b</b>) matched; (<b>c</b>) merged; (<b>d</b>) split; and (<b>e</b>) missing.</p>
Full article ">Figure 11
<p>The reference tree crowns (<b>a</b>) and the ITC delineation results of plot 1 using Mask R-CNN (<b>b</b>), the improved MSAS method (<b>c</b>), and Li’s method (<b>d</b>). The tree crowns are shown in polygons with different colors. The reference tree crowns (<b>e</b>) and the ITC delineation results of plot 2 using Mask R-CNN (<b>f</b>), the improved MSAS method (<b>g</b>), and Li’s method (<b>h</b>); (<b>i</b>–<b>l</b>) are the local detail maps of (<b>e</b>–<b>h</b>), respectively.</p>
Full article ">Figure 12
<p>The reference tree crowns (<b>a</b>) and the ITC delineation results of plot 3 using Mask R-CNN (<b>b</b>), the improved MSAS method (<b>c</b>), and Li’s method (<b>d</b>). The tree crowns are shown in polygons with different colors. (<b>e</b>–<b>h</b>) are the local detail maps of (<b>a</b>–<b>d</b>), respectively.</p>
Full article ">Figure 13
<p>The reference tree crowns (<b>a</b>) and the ITC delineation results of plot 4 using Mask R-CNN (<b>b</b>), the improved MSAS method (<b>c</b>), and Li’s method (<b>d</b>) The tree crowns are shown in polygons with different colors. (<b>e1</b>–<b>h2</b>) are the local detail maps of (<b>a</b>–<b>d</b>), respectively.</p>
Full article ">Figure 14
<p>The reference tree crowns (<b>a</b>) and the ITC delineation results of plot 5 using Mask R-CNN (<b>b</b>), the improved MSAS method (<b>c</b>), and Li’s method (<b>d</b>) The tree crowns are shown in polygons with different colors. (<b>e1</b>–<b>h2</b>) are the local detail maps of (<b>a</b>–<b>d</b>), respectively.</p>
Full article ">
21 pages, 15071 KiB  
Article
MaskAppendix: Backbone-Enriched Mask R-CNN Based on Grad-CAM for Automatic Appendix Segmentation
by Emre Dandıl, Betül Tiryaki Baştuğ, Mehmet Süleyman Yıldırım, Kadir Çorbacı and Gürkan Güneri
Diagnostics 2024, 14(21), 2346; https://doi.org/10.3390/diagnostics14212346 - 22 Oct 2024
Viewed by 506
Abstract
Background: A leading cause of emergency abdominal surgery, appendicitis is a common condition affecting millions of people worldwide. Automatic and accurate segmentation of the appendix from medical imaging is a challenging task, due to its small size, variability in shape, and proximity to [...] Read more.
Background: A leading cause of emergency abdominal surgery, appendicitis is a common condition affecting millions of people worldwide. Automatic and accurate segmentation of the appendix from medical imaging is a challenging task, due to its small size, variability in shape, and proximity to other anatomical structures. Methods: In this study, we propose a backbone-enriched Mask R-CNN architecture (MaskAppendix) on the Detectron platform, enhanced with Gradient-weighted Class Activation Mapping (Grad-CAM), for precise appendix segmentation on computed tomography (CT) scans. In the proposed MaskAppendix deep learning model, ResNet101 network is used as the backbone. By integrating Grad-CAM into the MaskAppendix network, our model improves feature localization, allowing it to better capture subtle variations in appendix morphology. Results: We conduct extensive experiments on a dataset of abdominal CT scans, demonstrating that our method achieves state-of-the-art performance in appendix segmentation, outperforming traditional segmentation techniques in terms of both accuracy and robustness. In the automatic segmentation of the appendix region in CT slices, a DSC score of 87.17% was achieved with the proposed approach, and the results obtained have the potential to improve clinical diagnostic accuracy. Conclusions: This framework provides an effective tool for aiding clinicians in the diagnosis of appendicitis and other related conditions, reducing the potential for diagnostic errors and enhancing clinical workflow efficiency. Full article
(This article belongs to the Topic AI in Medical Imaging and Image Processing)
Show Figures

Figure 1

Figure 1
<p>Overall architecture of the proposed MaskAppendix model for automatic segmentation of the appendix.</p>
Full article ">Figure 2
<p>Original CT in dataset slices and the appendix region masks in these CT slices extracted by expert physicians.</p>
Full article ">Figure 3
<p>The proposed Mask R-CNN framework for automatic segmentation of the appendix on CT images.</p>
Full article ">Figure 4
<p>(<b>a</b>) Internal structure of the ResNet backbone network, (<b>b</b>) depth generation with residual block in ResNet50 and ResNet101.</p>
Full article ">Figure 5
<p>(<b>a</b>) Progression of total loss values during training using ResNet50 and ResNet101 backbone in Mask R-CNN network, (<b>b</b>) DSC progression using Mask R-CNN with ResNet50 and Mask R-CNN with ResNet101 on the test set.</p>
Full article ">Figure 6
<p>The successful automatic segmentation of the appendix in CT slices using the proposed MaskAppendix model.</p>
Full article ">Figure 7
<p>Boxplot evaluation of the results of ablation experiments carried out to verify the effect of ResNet50 and ResNet101 modules proposed as backbone in the Mask R-CNN architecture on segmentation performance. Boxplot for (<b>a</b>) DSC, (<b>b</b>) JSI, (<b>c</b>) VOE, (<b>d</b>) ASD, and (<b>e</b>) HD95.</p>
Full article ">Figure 8
<p>Appendix regions segmented with lower performance using the proposed MaskAppendix model in some CT slices.</p>
Full article ">Figure 9
<p>Evaluation of visual results of heatmap feature outputs with Grad-CAM with MaskAppendix segmentation result.</p>
Full article ">
19 pages, 13917 KiB  
Article
TCSNet: A New Individual Tree Crown Segmentation Network from Unmanned Aerial Vehicle Images
by Yue Chi, Chenxi Wang, Zhulin Chen and Sheng Xu
Forests 2024, 15(10), 1814; https://doi.org/10.3390/f15101814 - 17 Oct 2024
Viewed by 640
Abstract
As the main area for photosynthesis in trees, the canopy absorbs a large amount of carbon dioxide and plays an irreplaceable role in regulating the carbon cycle in the atmosphere and mitigating climate change. Therefore, monitoring the growth of the canopy is crucial. [...] Read more.
As the main area for photosynthesis in trees, the canopy absorbs a large amount of carbon dioxide and plays an irreplaceable role in regulating the carbon cycle in the atmosphere and mitigating climate change. Therefore, monitoring the growth of the canopy is crucial. However, traditional field investigation methods are often limited by time-consuming and labor-intensive methods, as well as limitations in coverage, which may result in incomplete and inaccurate assessments. In response to the challenges encountered in the application of tree crown segmentation algorithms, such as adhesion between individual tree crowns and insufficient generalization ability of the algorithm, this study proposes an improved algorithm based on Mask R-CNN (Mask Region-based Convolutional Neural Network), which identifies irregular edges of tree crowns in RGB images obtained from drones. Firstly, it optimizes the backbone network by improving it to ResNeXt and embedding the SENet (Squeeze-and-Excitation Networks) module to enhance the model’s feature extraction capability. Secondly, the BiFPN-CBAM module is introduced to enable the model to learn and utilize features more effectively. Finally, it optimizes the mask loss function to the Boundary-Dice loss function to further improve the tree crown segmentation effect. In this study, TCSNet also incorporated the concept of panoptic segmentation, achieving the coherent and consistent segmentation of tree crowns throughout the entire scene through fine tree crown boundary recognition and integration. TCSNet was tested on three datasets with different geographical environments and forest types, namely artificial forests, natural forests, and urban forests, with artificial forests performing the best. Compared with the original algorithm, on the artificial forest dataset, the precision increased by 6.6%, the recall rate increased by 1.8%, and the F1-score increased by 4.2%, highlighting its potential and robustness in tree detection and segmentation. Full article
(This article belongs to the Special Issue Panoptic Segmentation of Tree Scenes from Mobile LiDAR Data)
Show Figures

Figure 1

Figure 1
<p>The proposed TCSNet structure.</p>
Full article ">Figure 2
<p>SE-ResNeXt structure.</p>
Full article ">Figure 3
<p>BiFPN and BiFPN-CBAM.</p>
Full article ">Figure 4
<p>Urban forest scene. The research area is located near Xuanwu Lake in Nanjing, Jiangsu Province, China. The forest type is mixed forest, and common tree species include camphor (<span class="html-italic">Camphora officinarum</span> Nees ex Wall), ginkgo (<span class="html-italic">Ginkgo biloba</span> L.), pine (<span class="html-italic">Pinus</span> L.), and willow (<span class="html-italic">Salix</span>).</p>
Full article ">Figure 5
<p>Artificial forest scene. The research area of the artificial forest dataset is located in Jiangsu Huanghai Haibin National Forest Park in Yancheng. Here, there are vast artificial ecological forests with extremely high forest coverage. Common tree species in the park include metasequoia (<span class="html-italic">Metasequoia glyptostroboides</span> Hu et Cheng) and poplar (<span class="html-italic">Populus</span> spp.).</p>
Full article ">Figure 6
<p>UAV image collection equipment: M350RTK.</p>
Full article ">Figure 7
<p>Data augmentation of poplar (<span class="html-italic">Populus</span> spp.) trees in artificial forest datasets. (<b>a</b>) Original image; (<b>b</b>) the image is obtained by rotation; (<b>c</b>) the image is obtained by changing the contrast; (<b>d</b>) the image is obtained by changing the saturation.</p>
Full article ">Figure 7 Cont.
<p>Data augmentation of poplar (<span class="html-italic">Populus</span> spp.) trees in artificial forest datasets. (<b>a</b>) Original image; (<b>b</b>) the image is obtained by rotation; (<b>c</b>) the image is obtained by changing the contrast; (<b>d</b>) the image is obtained by changing the saturation.</p>
Full article ">Figure 8
<p>Total loss. The total loss considers the training effect of multiple loss function comprehensive indicators.</p>
Full article ">Figure 9
<p>Classification loss. The classification loss focuses on evaluating the loss function of model prediction accuracy in classification tasks.</p>
Full article ">Figure 10
<p>Loss for bounding box regression. The loss for bounding box regression is used to measure the prediction error of bounding box regression in tree crown detection.</p>
Full article ">Figure 11
<p>Segmentation performance of each dataset. (<b>a</b>,<b>b</b>) and (<b>c</b>,<b>d</b>) respectively demonstrate the segmentation performance of TCSNet on two tree species. (<b>e</b>–<b>h</b>) demonstrate the segmentation performance of TCSNet in urban parks and green spaces. (<b>i</b>–<b>l</b>) demonstrated the segmentation performance of TCSNet in tropical rainforests.</p>
Full article ">Figure 12
<p>Differences in different datasets. Blue squares represent tree crowns that are difficult to segment. (<b>a</b>,<b>b</b>) Both belong to artificial forests, but (<b>a</b>) has a similar canopy size and a more orderly arrangement, so the segmentation effect is better. (<b>b</b>) More affected by grass, and the crown is irregular, the effect is average. (<b>c</b>) The canopy size is similar, some ordered and some disorderly, but it will still be affected by the shadow, and the segmentation effect is general. (<b>d</b>) The trees are natural forests with small gaps and inconsistent canopy sizes, so segmentation is the most difficult.</p>
Full article ">Figure 13
<p>Comparisons with other algorithms [<a href="#B29-forests-15-01814" class="html-bibr">29</a>,<a href="#B30-forests-15-01814" class="html-bibr">30</a>,<a href="#B31-forests-15-01814" class="html-bibr">31</a>]. We achieved most tree crown instances from input data.</p>
Full article ">
16 pages, 3825 KiB  
Article
Evolutionary Grid Optimization and Deep Learning for Improved In Vitro Cellular Spheroid Localization
by Jonas Schurr, Hannah Janout, Andreas Haghofer, Marian Fürsatz, Josef Scharinger, Stephan Winkler and Sylvia Nürnberger
Appl. Sci. 2024, 14(20), 9476; https://doi.org/10.3390/app14209476 - 17 Oct 2024
Viewed by 527
Abstract
The recently developed high-throughput system for cell spheroid generation (SpheroWell) is a promising technology for cost- and time-efficient in vitro analysis of, for example, chondrogenic differentiation. It is a compartmental growth surface where spheroids develop from a cell monolayer by self-assembling and aggregation. [...] Read more.
The recently developed high-throughput system for cell spheroid generation (SpheroWell) is a promising technology for cost- and time-efficient in vitro analysis of, for example, chondrogenic differentiation. It is a compartmental growth surface where spheroids develop from a cell monolayer by self-assembling and aggregation. In order to automatize the analysis of spheroids, we aimed to develop imaging software and improve the localization of cell compartments and fully formed spheroids. Our workflow provides automated detection and localization of spheroids in different formation stages within Petri dishes based on images created with a low-budget camera imaging setup. This automated detection enables a fast and inexpensive analysis workflow by processing a stack of images within a short period of time, which is essential for the extraction of early readout parameters. Our workflow combines image processing algorithms and deep learning-based image localization/segmentation methods like Mask R-CNN and Unet++. These methods are refined by an evolution strategy for automated grid detection, which is able to improve the overall segmentation and classification quality. Besides the already pre-trained neural networks and predefined image processing parameters, our evolution-based post-processing provides the required adaptability for our workflow to deliver a consistent and reproducible quality. This is especially important due to the use of a low-budget imaging setup with various light conditions. The to-be-detected objects of the three different stages show improved results using our evolutionary post-processing for monolayer and starting aggregation with Dice coefficients of 0.7301 and 0.8562, respectively, compared with the raw scores of 0.2879 and 0.8187. The Dice coefficient of the fully formed spheroids in both cases is 0.8829. With our algorithm, we provide automated analyses of cell spheroid by self-assembling in SpheroWell dishes, even if the images are created using a low-budget camera setup. Full article
(This article belongs to the Special Issue Evolutionary Algorithms and Their Real-World Applications)
Show Figures

Figure 1

Figure 1
<p>Example of a SpheroWell Petri dish image (<b>left</b>) with the individual states of forming spheroids as well as the corresponding object labels on the (<b>right</b>). Each labeled object is drawn by a polygon, and its intensity value represents a unique identification number. This number is used for additional data files, including the corresponding class label for each identification number.</p>
Full article ">Figure 2
<p>We used our dataset of 20 images of Petri dishes, along with class and object labels, to create and validate individual workflow modules. For the creation of our segmentation/instance segmentation models, we split the initial images into three datasets (training, validation, and testing). Using the training and validation split, we created our Mask R-CNN and our Unet++. Using our validation dataset, we selected the best model for both architectures based on the validation loss. The final testing of our model was split into two parts (binary segmentation and object detection), both using the test dataset. Solely using the Petri dish images without class labels, our evolutionary grid optimization created a grid for each of our images. Based on these grids, we created and tested our parameters for the grid-based filtering. The final testing of our workflow was also based on the separate test dataset.</p>
Full article ">Figure 3
<p>Overview of steps for the evolution strategy. 1: Extracted Petri dish with Hough Circles (insufficient extraction of grid cells), 2: Extracted Hough Lines 3: grid of a solution candidate, 4: Compute fitness of solution candidates with overlap based on Dice coefficient, 5: Final results.</p>
Full article ">Figure 4
<p>Simplified representation of the whole classification workflow, starting with the image of a Petri dish carrying the individual objects. Based on this image, our workflow generates a grid using the evolution strategy-based optimization process and provides the position of each grid cell, as well as a classification mask of all individual objects determined by our neural network. Both of these results are fed into the grid filtering mechanism, which processes the classification results based on the grid cells and outputs the final classification result for each object.</p>
Full article ">Figure 5
<p>Comparison of first epoch (<b>A1</b>,<b>B1</b>) and last epoch (<b>A2</b>,<b>B2</b>) of an evolution strategy run. Images A1 and A2 show a grid represented by a solution candidate, and images (<b>B1</b>,<b>B2</b>) show the final solution on the RGB image with the extracted Petri dish. The result clearly shows the improvement resulting in an almost perfectly aligned grid (fitness: 0.0844). Additionally, <b>A</b> and <b>B</b> show the robustness of the evolution strategy to noise and wrong lines of the preprocessing.</p>
Full article ">Figure 6
<p>Comparison of validated parameters and corresponding values used for the line extraction process by fitness value (lowest is best). Showing the influence and robustness of best and most important parameters and their corresponding values. Large differences in the minimum gap parameter can be observed.</p>
Full article ">Figure 7
<p>Comparison of validated parameters and corresponding values used for the ES by fitness value (lowest is best). Showing the influence and robustness of best and most important parameters and their corresponding values. Large differences in population size can be observed.</p>
Full article ">Figure 8
<p>Comparison of validated parameter values used for the ES by Dice coefficient (highest is best) based on a ground truth image for detected lines and corresponding. Showing the influence and robustness of best and most important parameters and their corresponding values. Large differences in population size can be observed.</p>
Full article ">Figure 9
<p>Comparison of the raw classification result and the filtered/corrected version in comparison with the ground truth. (blue: monolayer, green: forming state, yellow: final spheroid).</p>
Full article ">
18 pages, 8011 KiB  
Article
Intelligent Vision System with Pruning and Web Interface for Real-Time Defect Detection on African Plum Surfaces
by Arnaud Nguembang Fadja, Sain Rigobert Che and Marcellin Atemkemg
Information 2024, 15(10), 635; https://doi.org/10.3390/info15100635 - 14 Oct 2024
Viewed by 698
Abstract
Agriculture stands as the cornerstone of Africa’s economy, supporting over 60% of the continent’s labor force. Despite its significance, the quality assessment of agricultural products remains a challenging task, particularly at a large scale, consuming valuable time and resources. The African plum is [...] Read more.
Agriculture stands as the cornerstone of Africa’s economy, supporting over 60% of the continent’s labor force. Despite its significance, the quality assessment of agricultural products remains a challenging task, particularly at a large scale, consuming valuable time and resources. The African plum is an agricultural fruit that is widely consumed across West and Central Africa but remains underrepresented in AI research. In this paper, we collected a dataset of 2892 African plum samples from fields in Cameroon representing the first dataset of its kind for training AI models. The dataset contains images of plums annotated with quality grades. We then trained and evaluated various state-of-the-art object detection and image classification models, including YOLOv5, YOLOv8, YOLOv9, Fast R-CNN, Mask R-CNN, VGG-16, DenseNet-121, MobileNet, and ResNet, on this African plum dataset. Our experimentation resulted in mean average precision scores ranging from 88.2% to 89.9% and accuracies between 86% and 91% for the object detection models and the classification models, respectively. We then performed model pruning to reduce model sizes while preserving performance, achieving up to 93.6% mean average precision and 99.09% accuracy after pruning YOLOv5, YOLOv8 and ResNet by 10–30%. We deployed the high-performing YOLOv8 system in a web application, offering an accessible AI-based quality assessment tool tailored for African plums. To the best of our knowledge, this represents the first such solution for assessing this underrepresented fruit, empowering farmers with efficient tools. Our approach integrates agriculture and AI to fill a key gap. Full article
Show Figures

Figure 1

Figure 1
<p>Sample images showcasing plum fruits on the fruit tree [<a href="#B27-information-15-00635" class="html-bibr">27</a>].</p>
Full article ">Figure 2
<p>The YOLOv5 is structured into three primary segments: the backbone, neck, and output [<a href="#B41-information-15-00635" class="html-bibr">41</a>].</p>
Full article ">Figure 3
<p>Overview of the key steps in our implementation. These structured steps ensure efficient implementation of the project.</p>
Full article ">Figure 4
<p>Sample images showcasing the labeling of good and defective plums with and without the background category. (<b>a</b>) Labeling of a good plum with the background class. (<b>b</b>) Labeling of a good plum. (<b>c</b>) Labeling of a defective plum with the background class. (<b>d</b>) Labeling of a defective plum.</p>
Full article ">Figure 5
<p>YOLOv5 training performance. This figure shows the training curves for the YOLOv5 object detection model. The top plot displays the loss function during the training process, which includes components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 6
<p>YOLOv8 training and evaluation. This figure presents the performance metrics for the YOLOv8 object detection model during the training and evaluation phases. The top plot shows the training loss, which is composed of components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 7
<p>YOLOv9 training and evaluation. This figure presents the performance metrics for the YOLOv9 object detection model during the training and evaluation phases. The top plot shows the training loss, which is composed of components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 8
<p>Fast R-CNN training and evaluation metrics. This figure shows the training and validation metrics for the Fast R-CNN object detection model. The blue line represents the overall training loss, which includes components for bounding box regression, object classification, and region proposal classification. The orange and green lines show the validation metrics for the classification loss and the regression loss, respectively. These metrics indicate the model’s performance in generating accurate region proposals and classifying/localizing detected objects.</p>
Full article ">Figure 9
<p>Mask R-CNN training and evaluation metrics. This figure presents the training and validation performance metrics for the Mask R-CNN instance segmentation model. The blue line represents the overall training loss, which includes components for bounding box regression, object classification, and region proposal classification. The orange and green lines show the validation metrics for the classification loss and the regression loss, respectively. These metrics indicate the model’s performance in generating accurate region proposals and classifying/localizing detected objects.</p>
Full article ">Figure 10
<p>Training and validation metrics for the VGG-16 model. The top curves represent training (green) and validation (red) accuracy, while the bottom curves depict training (green) and validation (red) loss. The model demonstrates rapid generalization from a strong initial point, as indicated by the swift convergence of accuracy and loss metrics.</p>
Full article ">Figure 11
<p>Training and validation metrics for the DenseNet-121 model. The top curves represent training (green) and validation (red) accuracy, while the bottom curves depict training (blue) and validation (yellow) loss. The model demonstrates rapid generalization from a strong initial point, as indicated by the swift convergence of accuracy and loss metrics.</p>
Full article ">Figure 12
<p>Model predictions with background class: YOLOv5, YOLOv8, and YOLOv9. (<b>a</b>) YOLOv5 good fruit prediction. (<b>b</b>) YOLOv8 good fruit prediction. (<b>c</b>) YOLOv9 good fruit prediction. (<b>d</b>) YOLOv5 bad fruit prediction. (<b>e</b>) YOLOv8 bad fruit prediction. (<b>f</b>) YOLOv9 bad fruit prediction.</p>
Full article ">
21 pages, 26972 KiB  
Article
Defective Pennywort Leaf Detection Using Machine Vision and Mask R-CNN Model
by Milon Chowdhury, Md Nasim Reza, Hongbin Jin, Sumaiya Islam, Geung-Joo Lee and Sun-Ok Chung
Agronomy 2024, 14(10), 2313; https://doi.org/10.3390/agronomy14102313 - 9 Oct 2024
Viewed by 542
Abstract
Demand and market value for pennywort largely depend on the quality of the leaves, which can be affected by various ambient environment or fertigation variables during cultivation. Although early detection of defects in pennywort leaves would enable growers to take quick action, conventional [...] Read more.
Demand and market value for pennywort largely depend on the quality of the leaves, which can be affected by various ambient environment or fertigation variables during cultivation. Although early detection of defects in pennywort leaves would enable growers to take quick action, conventional manual detection is laborious and time consuming as well as subjective. Therefore, the objective of this study was to develop an automatic leaf defect detection algorithm for pennywort plants grown under controlled environment conditions, using machine vision and deep learning techniques. Leaf images were captured from pennywort plants grown in an ebb-and-flow hydroponic system under fluorescent light conditions in a controlled plant factory environment. Physically or biologically damaged leaves (e.g., curled, creased, discolored, misshapen, or brown spotted) were classified as defective leaves. Images were annotated using an online tool, and Mask R-CNN models were implemented with the integrated attention mechanisms, convolutional block attention module (CBAM) and coordinate attention (CA) and compared for improved image feature extraction. Transfer learning was employed to train the model with a smaller dataset, effectively reducing processing time. The improved models demonstrated significant advancements in accuracy and precision, with the CA-augmented model achieving the highest metrics, including a mean average precision (mAP) of 0.931 and an accuracy of 0.937. These enhancements enabled more precise localization and classification of leaf defects, outperforming the baseline Mask R-CNN model in complex visual recognition tasks. The final model was robust, effectively distinguishing defective leaves in challenging scenarios, making it highly suitable for applications in precision agriculture. Future research can build on this modeling framework, exploring additional variables to identify specific leaf abnormalities at earlier growth stages, which is crucial for production quality assurance. Full article
(This article belongs to the Special Issue Advanced Machine Learning in Agriculture)
Show Figures

Figure 1

Figure 1
<p>Experimental site and image acquisition: (<b>a</b>) cultivation shelf for pennywort seedling adaption with hydroponic system and ambient environment, (<b>b</b>) pennywort seedlings grown under fluorescent light, and (<b>c</b>) sample images of pennywort leaves grown in an ebb-and-flow type hydroponic system: malnourished leaves (<b>top</b>), healthy leaves (<b>bottom</b>).</p>
Full article ">Figure 2
<p>Pennywort leaf annotation: (<b>a</b>) original image of affected pennywort plants taken during the experiment, and (<b>b</b>) manually masked healthy and unhealthy leaves.</p>
Full article ">Figure 3
<p>Image augmentation: (<b>a</b>) original image, (<b>b</b>) horizontal flip, (<b>c</b>) vertical flip, (<b>d</b>) shift, (<b>e</b>) zoom, and (<b>f</b>) rotation.</p>
Full article ">Figure 4
<p>The Mask R-CNN architecture with RPN and FPN was used in this study for detecting defective pennywort leaves.</p>
Full article ">Figure 5
<p>(<b>a</b>) The backbone feature extraction network (modified from [<a href="#B31-agronomy-14-02313" class="html-bibr">31</a>]), (<b>b</b>) anchor generation principle (modified from [<a href="#B29-agronomy-14-02313" class="html-bibr">29</a>]), and (<b>c</b>) ROI Align output achieved through grid points of bilinear interpolation (modified from [<a href="#B30-agronomy-14-02313" class="html-bibr">30</a>]), used in this study for detecting defective pennywort leaves.</p>
Full article ">Figure 6
<p>Illustration of feature extraction through the implemented algorithm for defective pennywort leaves.</p>
Full article ">Figure 7
<p>Illustration of CBAM model structure used in this study for detecting defective pennywort leaves: (<b>a</b>) convolutional block attention module, (<b>b</b>) channel attention module, and (<b>c</b>) spatial attention module.</p>
Full article ">Figure 8
<p>Structure of the coordinate attention (CA) mechanism used in this study for detecting defective pennywort leaves.</p>
Full article ">Figure 9
<p>Schematic diagrams for integrating ResNet-101 with attention mechanism modules: (<b>a</b>) ResNet-101+CBAM, and (<b>b</b>) ResNet-101+CA.</p>
Full article ">Figure 10
<p>Loss and accuracy variation of the Mask-RCNN and improved Mask-RCNN models: (<b>a</b>) loss variation for Mask-RCNN_ResNet-101, Mask-RCNN_ResNet-101+CBAM, and Mask-RCNN_ResNet-101+CA, and (<b>b</b>) accuracy variation for Mask-RCNN_ResNet-101, Mask-RCNN_ResNet-101+CBAM, and Mask-RCNN_ResNet-101+CA.</p>
Full article ">Figure 11
<p>Heatmap generated from the images and using the pre-trained models: (<b>a</b>) original image, (<b>b</b>) heatmap of Mask-RCNN_ResNet-101 model, (<b>c</b>) heatmap of Mask-RCNN_ResNet-101+CBAM, and (<b>d</b>) heatmap of Mask-RCNN_ResNet-101+CA model.</p>
Full article ">Figure 12
<p>Output results of the defective pennywort leaf detection in the test images using: (<b>a</b>) an annotated image, (<b>b</b>) the Mask R-CNN model, (<b>c</b>) the improved Mask-RCNN model with CBAM, and (<b>d</b>) the improved Mask-RCNN model with CA.</p>
Full article ">Figure 13
<p>Detection inaccuracies in test images: (<b>a</b>) annotated image and (<b>b</b>) false negative detection from the Mask RCNN model and the improved Mask RCNN models.</p>
Full article ">Figure 14
<p>Visualization of defective leaf segmentation results; (<b>a</b>) original annotated image, (<b>b</b>) ground truth, (<b>c</b>) segmentation result of Mask-RCNN model, (<b>d</b>) segmentation result of improved Mask-RCNN model with CBAM, and (<b>e</b>) segmentation result of improved Mask-RCNN model with CA.</p>
Full article ">Figure 15
<p>Precision- Recall (P-R) curve to evaluate the proposed models performance used in this study.</p>
Full article ">
Back to TopTop