[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,460)

Search Parameters:
Keywords = YOLOv5s

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
14 pages, 4478 KiB  
Article
A New Kiwi Fruit Detection Algorithm Based on an Improved Lightweight Network
by Yi Yang, Lijun Su, Aying Zong, Wanghai Tao, Xiaoping Xu, Yixin Chai and Weiyi Mu
Agriculture 2024, 14(10), 1823; https://doi.org/10.3390/agriculture14101823 (registering DOI) - 16 Oct 2024
Abstract
To address the challenges associated with kiwi fruit detection methods, such as low average accuracy, inaccurate recognition of fruits, and long recognition time, this study proposes a novel kiwi fruit recognition method based on an improved lightweight network S-YOLOv4-tiny detection algorithm. Firstly, the [...] Read more.
To address the challenges associated with kiwi fruit detection methods, such as low average accuracy, inaccurate recognition of fruits, and long recognition time, this study proposes a novel kiwi fruit recognition method based on an improved lightweight network S-YOLOv4-tiny detection algorithm. Firstly, the YOLOv4-tiny algorithm utilizes the CSPdarknet53-tiny network as a backbone feature extraction network, replacing the CSPdarknet53 network in the YOLOv4 algorithm to enhance the speed of kiwi fruit recognition. Additionally, a squeeze-and-excitation network has been incorporated into the S-YOLOv4-tiny detection algorithm to improve accurate image extraction of kiwi fruit characteristics. Finally, enhancing dataset pictures using mosaic methods has improved precision in the characteristic recognition of kiwi fruits. The experimental results demonstrate that the recognition and positioning of kiwi fruits have yielded improved outcomes. The mean average precision (mAP) stands at 89.75%, with a detection precision of 93.96% and a single-picture detection time of 8.50 ms. Compared to the YOLOv4-tiny detection algorithm network, the network in this study exhibits a 7.07% increase in mean average precision and a 1.16% acceleration in detection time. Furthermore, an enhancement method based on the Squeeze-and-Excitation Network (SENet) is proposed, as opposed to the convolutional block attention module (CBAM) and efficient channel attention (ECA). This approach effectively addresses issues related to slow training speed and low recognition accuracy of kiwi fruit, offering valuable technical insights for efficient mechanical picking methods. Full article
Show Figures

Figure 1

Figure 1
<p>Kiwi fruit images with the different degrees of occlusion.</p>
Full article ">Figure 2
<p>YOLOv4-tiny Network structure.</p>
Full article ">Figure 3
<p>The structure of the squeeze-and-excitation network.</p>
Full article ">Figure 4
<p>P-R curves for the four detection methods.</p>
Full article ">Figure 5
<p>The mean average accuracy changes with the number of iterations when the IOU value is 0.5.</p>
Full article ">Figure 6
<p>The detection of the different fruit occlusions.</p>
Full article ">Figure 7
<p>Image detection results of kiwi fruit by different mainstream networks.</p>
Full article ">
19 pages, 5406 KiB  
Article
An Automatic Movement Monitoring Method for Group-Housed Pigs
by Ziyuan Liang, Aijun Xu, Junhua Ye, Suyin Zhou, Xiaoxing Weng and Sian Bao
Animals 2024, 14(20), 2985; https://doi.org/10.3390/ani14202985 (registering DOI) - 16 Oct 2024
Abstract
Continuous movement monitoring helps quickly identify pig abnormalities, enabling immediate action to enhance pig welfare. However, continuous and precise monitoring of daily pig movement on farms remains challenging. We present an approach to automatically and precisely monitor the movement of group-housed pigs. The [...] Read more.
Continuous movement monitoring helps quickly identify pig abnormalities, enabling immediate action to enhance pig welfare. However, continuous and precise monitoring of daily pig movement on farms remains challenging. We present an approach to automatically and precisely monitor the movement of group-housed pigs. The instance segmentation model YOLOv8m-seg was applied to detect the presence of pigs. We then applied a spatial moment algorithm to quantitatively summarize each detected pig’s contour as a corresponding center point. The agglomerative clustering (AC) algorithm was subsequently used to gather the pig center points of a single frame into one point representing the group-housed pigs’ position, and the movement volume was obtained by calculating the displacements of the clustered group-housed pigs’ center points of consecutive frames. We employed the method to monitor the movement of group-housed pigs from April to July 2023; more than 1500 h of top-down pig videos were recorded by a surveillance camera. The F1 scores of the trained YOLOv8m-seg model during training were greater than 90% across most confidence levels, and the model achieved an mAP50-95 of 0.96. The AC algorithm performs with an average extraction time of less than 1 millisecond; this method can run efficiently on commodity hardware. Full article
(This article belongs to the Section Pigs)
Show Figures

Figure 1

Figure 1
<p>Experimental conditions: (<b>a</b>) draft of the pigpen; (<b>b</b>) installation position of the dual sensor surveillance camera; (<b>c</b>) top-down camera view of the pigsty floor.</p>
Full article ">Figure 2
<p>Designed workflow of the pig movement monitoring method.</p>
Full article ">Figure 3
<p>Model structure of YOLOv8-seg: the segmentation and detection tasks begin with the (<b>a</b>) original image and output an (<b>b</b>) image with a bounding box and segmentation contour.</p>
Full article ">Figure 4
<p>Distinguishing the center point of a predicted pig contour. The images in the columns are described as follows: (1) prediction image, (2) mean coordinate, (3) least squares, (4) signed area, and (5) spatial moment. The different pig behavior patterns depicted in each row are as follows: (<b>a</b>) lying, (<b>b</b>) sitting, and (<b>c</b>) standing.</p>
Full article ">Figure 5
<p>Running times of different algorithms based on the test video (1166 frames): (<b>a</b>) Time spent on each frame. (<b>b</b>) Total time spent in progress (average of 30 repetitions).</p>
Full article ">Figure 6
<p>Complete distribution information of group pig positions. The information was obtained from 13 May to 8 July 2023, and every pig position was drawn at a given point.</p>
Full article ">Figure 7
<p>Changes in the position information over time: (<b>a</b>) Pigs positioned during two periods from 13 May to 9 June 2023 and 10 June to 8 July 2023. (<b>b</b>) The statistical variation in the number of pigs appearing in different regions during the two periods.</p>
Full article ">Figure 8
<p>Daily summed movement distances of group-housed pigs from 13 May 2023 to 8 July 2023.</p>
Full article ">Figure 9
<p>Movement characteristics of pigs in terms of days with the longest, shortest, and median movement distances; every subfigure starts at 0 a.m. and ends at 12 p.m.</p>
Full article ">Figure 10
<p>Various distribution locations of group-housed pigs: (<b>a</b>) pigs lying close to the corner; (<b>b</b>) pigs congregating near the door; (<b>c</b>) herd of pigs eating.</p>
Full article ">
17 pages, 3107 KiB  
Article
CL-YOLOv8: Crack Detection Algorithm for Fair-Faced Walls Based on Deep Learning
by Qinjun Li, Guoyu Zhang and Ping Yang
Appl. Sci. 2024, 14(20), 9421; https://doi.org/10.3390/app14209421 (registering DOI) - 16 Oct 2024
Viewed by 216
Abstract
Cracks pose a critical challenge in the preservation of historical buildings worldwide, particularly in fair-faced walls, where timely and accurate detection is essential to prevent further degradation. Traditional image processing methods have proven inadequate for effectively detecting building cracks. Despite global advancements in [...] Read more.
Cracks pose a critical challenge in the preservation of historical buildings worldwide, particularly in fair-faced walls, where timely and accurate detection is essential to prevent further degradation. Traditional image processing methods have proven inadequate for effectively detecting building cracks. Despite global advancements in deep learning, crack detection under diverse environmental and lighting conditions remains a significant technical hurdle, as highlighted by recent international studies. To address this challenge, we propose an enhanced crack detection algorithm, CL-YOLOv8 (ConvNeXt V2-LSKA-YOLOv8). By integrating the well-established ConvNeXt V2 model as the backbone network into YOLOv8, the algorithm benefits from advanced feature extraction techniques, leading to a superior detection accuracy. This choice leverages ConvNeXt V2’s recognized strengths, providing a robust foundation for improving the overall model performance. Additionally, by introducing the LSKA (Large Separable Kernel Attention) mechanism into the SPPF structure, the feature receptive field is enlarged and feature correlations are strengthened, further enhancing crack detection accuracy in diverse environments. This study also contributes to the field by significantly expanding the dataset for fair-faced wall crack detection, increasing its size sevenfold through data augmentation and the inclusion of additional data. Our experimental results demonstrate that CL-YOLOv8 outperforms mainstream algorithms such as Faster R-CNN, YOLOv5s, YOLOv7-tiny, SSD, and various YOLOv8n/s/m/l/x models. CL-YOLOv8 achieves an accuracy of 85.3%, a recall rate of 83.2%, and a mean average precision (mAP) of 83.7%. Compared to the YOLOv8n base model, CL-YOLOv8 shows improvements of 0.9%, 2.3%, and 3.9% in accuracy, recall rate, and mAP, respectively. These results underscore the effectiveness and superiority of CL-YOLOv8 in crack detection, positioning it as a valuable tool in the global effort to preserve architectural heritage. Full article
Show Figures

Figure 1

Figure 1
<p>The system diagram of YOLOv8n.</p>
Full article ">Figure 2
<p>System diagram comparing LKA and LSKA modules.</p>
Full article ">Figure 3
<p>LSKA module.</p>
Full article ">Figure 4
<p>Diagram of the LSKA module structure.</p>
Full article ">Figure 5
<p>Comparison diagram of ConvNeXt V1 and V2 modules.</p>
Full article ">Figure 6
<p>Improved system architecture diagram.</p>
Full article ">Figure 7
<p>Comparison of P–R figures before and after improvement.</p>
Full article ">Figure 8
<p>Comparison of F1-confidence curves before and after improvement.</p>
Full article ">Figure 9
<p>The detection effect of different models. YOLOv5s (<b>a</b>), YOLOv7-tiny (<b>b</b>), YOLOv8n (<b>c</b>), and CL-YOLOv8 (<b>d</b>).</p>
Full article ">
20 pages, 7477 KiB  
Article
A Ship’s Maritime Critical Target Identification Method Based on Lightweight and Triple Attention Mechanisms
by Pu Wang, Shenhua Yang, Guoquan Chen, Weijun Wang, Zeyang Huang and Yuanliang Jiang
J. Mar. Sci. Eng. 2024, 12(10), 1839; https://doi.org/10.3390/jmse12101839 - 14 Oct 2024
Viewed by 333
Abstract
The ability to classify and recognize maritime targets based on visual images plays an important role in advancing ship intelligence and digitalization. The current target recognition algorithms for common maritime targets, such as buoys, reefs, other ships, and bridges of different colors, face [...] Read more.
The ability to classify and recognize maritime targets based on visual images plays an important role in advancing ship intelligence and digitalization. The current target recognition algorithms for common maritime targets, such as buoys, reefs, other ships, and bridges of different colors, face challenges such as incomplete classification, low recognition accuracy, and a large number of model parameters. To address these issues, this paper proposes a novel maritime target recognition method called DTI-YOL (DualConv Triple Attention InnerEIOU-You Only Look Once). This method is based on a triple attention mechanism designed to enhance the model’s ability to classify and recognize buoys of different colors in the channel while also making the feature extraction network more lightweight. First, the lightweight double convolution kernel feature extraction layer is constructed using group convolution technology to replace the Conv structure of YOLOv9 (You Only Look Once Version 9), effectively reducing the number of parameters in the original model. Second, an improved three-branch structure is designed to capture cross-dimensional interactions of input image features. This structure forms a triple attention mechanism that accounts for the mutual dependencies between input channels and spatial positions, allowing for the calculation of attention weights for targets such as bridges, buoys, and other ships. Finally, InnerEIoU is used to replace CIoU to improve the loss function, thereby optimizing loss regression for targets with large scale differences. To verify the effectiveness of these algorithmic improvements, the DTI-YOLO algorithm was tested on a self-made dataset of 2300 ship navigation images. The experimental results show that the average accuracy of this method in identifying seven types of targets—including buoys, bridges, islands and reefs, container ships, bulk carriers, passenger ships, and other ships—reached 92.1%, with a 12% reduction in the number of parameters. This enhancement improves the model’s ability to recognize and distinguish different targets and buoy colors. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>DTI-YOLO network structure diagram.</p>
Full article ">Figure 2
<p>Internal structure of 1 × 1 and 3 × 3 double convolution kernel.</p>
Full article ">Figure 3
<p>Double convolution structure.</p>
Full article ">Figure 4
<p>Structure of group convolution technique.</p>
Full article ">Figure 5
<p>Internal structure of the triple attention mechanism.</p>
Full article ">Figure 6
<p>Schematic diagram of EIoU losses.</p>
Full article ">Figure 7
<p>Schematic diagram of inner loss structure.</p>
Full article ">Figure 8
<p>(<b>a</b>) Example plot of the dataset; (<b>b</b>) labeling plot.</p>
Full article ">Figure 9
<p>Sample categories and number of samples in the Harborships dataset.</p>
Full article ">Figure 10
<p>Comparison of ablation experiment effect and heat map: (<b>a</b>) YOLOv9 recognition effect; (<b>b</b>) YOLOv9+ DualConv recognition effect; (<b>c</b>) YOLOv9 + Attention recognition effect.</p>
Full article ">Figure 11
<p>Comparison of mAP curves.</p>
Full article ">Figure 12
<p>Comparison of model precision and recall.</p>
Full article ">Figure 13
<p>Comparison of YOLOv9 and DTI-YOLO algorithm target identification and heat map results.</p>
Full article ">Figure 14
<p>Comparison of P-R curves of YOLOv9 (<b>a</b>) and DTI-YOLO (<b>b</b>) algorithms.</p>
Full article ">Figure 15
<p>Comparison of target recognition results: (<b>a</b>) Original figure; (<b>b</b>) YOLOv5 recognition result; (<b>c</b>) YOLOv7 recognition result; (<b>d</b>) YOLOv8 recognition result; (<b>e</b>) YOLOv9 recognition result; (<b>f</b>) SSD recognition result; (<b>g</b>) Faster-RCNN recognition result; (<b>h</b>) DTI-YOLO algorithm recognition result.</p>
Full article ">Figure 15 Cont.
<p>Comparison of target recognition results: (<b>a</b>) Original figure; (<b>b</b>) YOLOv5 recognition result; (<b>c</b>) YOLOv7 recognition result; (<b>d</b>) YOLOv8 recognition result; (<b>e</b>) YOLOv9 recognition result; (<b>f</b>) SSD recognition result; (<b>g</b>) Faster-RCNN recognition result; (<b>h</b>) DTI-YOLO algorithm recognition result.</p>
Full article ">
18 pages, 8011 KiB  
Article
Intelligent Vision System with Pruning and Web Interface for Real-Time Defect Detection on African Plum Surfaces
by Arnaud Nguembang Fadja, Sain Rigobert Che and Marcellin Atemkemg
Information 2024, 15(10), 635; https://doi.org/10.3390/info15100635 (registering DOI) - 14 Oct 2024
Viewed by 417
Abstract
Agriculture stands as the cornerstone of Africa’s economy, supporting over 60% of the continent’s labor force. Despite its significance, the quality assessment of agricultural products remains a challenging task, particularly at a large scale, consuming valuable time and resources. The African plum is [...] Read more.
Agriculture stands as the cornerstone of Africa’s economy, supporting over 60% of the continent’s labor force. Despite its significance, the quality assessment of agricultural products remains a challenging task, particularly at a large scale, consuming valuable time and resources. The African plum is an agricultural fruit that is widely consumed across West and Central Africa but remains underrepresented in AI research. In this paper, we collected a dataset of 2892 African plum samples from fields in Cameroon representing the first dataset of its kind for training AI models. The dataset contains images of plums annotated with quality grades. We then trained and evaluated various state-of-the-art object detection and image classification models, including YOLOv5, YOLOv8, YOLOv9, Fast R-CNN, Mask R-CNN, VGG-16, DenseNet-121, MobileNet, and ResNet, on this African plum dataset. Our experimentation resulted in mean average precision scores ranging from 88.2% to 89.9% and accuracies between 86% and 91% for the object detection models and the classification models, respectively. We then performed model pruning to reduce model sizes while preserving performance, achieving up to 93.6% mean average precision and 99.09% accuracy after pruning YOLOv5, YOLOv8 and ResNet by 10–30%. We deployed the high-performing YOLOv8 system in a web application, offering an accessible AI-based quality assessment tool tailored for African plums. To the best of our knowledge, this represents the first such solution for assessing this underrepresented fruit, empowering farmers with efficient tools. Our approach integrates agriculture and AI to fill a key gap. Full article
Show Figures

Figure 1

Figure 1
<p>Sample images showcasing plum fruits on the fruit tree [<a href="#B27-information-15-00635" class="html-bibr">27</a>].</p>
Full article ">Figure 2
<p>The YOLOv5 is structured into three primary segments: the backbone, neck, and output [<a href="#B41-information-15-00635" class="html-bibr">41</a>].</p>
Full article ">Figure 3
<p>Overview of the key steps in our implementation. These structured steps ensure efficient implementation of the project.</p>
Full article ">Figure 4
<p>Sample images showcasing the labeling of good and defective plums with and without the background category. (<b>a</b>) Labeling of a good plum with the background class. (<b>b</b>) Labeling of a good plum. (<b>c</b>) Labeling of a defective plum with the background class. (<b>d</b>) Labeling of a defective plum.</p>
Full article ">Figure 5
<p>YOLOv5 training performance. This figure shows the training curves for the YOLOv5 object detection model. The top plot displays the loss function during the training process, which includes components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 6
<p>YOLOv8 training and evaluation. This figure presents the performance metrics for the YOLOv8 object detection model during the training and evaluation phases. The top plot shows the training loss, which is composed of components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 7
<p>YOLOv9 training and evaluation. This figure presents the performance metrics for the YOLOv9 object detection model during the training and evaluation phases. The top plot shows the training loss, which is composed of components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 8
<p>Fast R-CNN training and evaluation metrics. This figure shows the training and validation metrics for the Fast R-CNN object detection model. The blue line represents the overall training loss, which includes components for bounding box regression, object classification, and region proposal classification. The orange and green lines show the validation metrics for the classification loss and the regression loss, respectively. These metrics indicate the model’s performance in generating accurate region proposals and classifying/localizing detected objects.</p>
Full article ">Figure 9
<p>Mask R-CNN training and evaluation metrics. This figure presents the training and validation performance metrics for the Mask R-CNN instance segmentation model. The blue line represents the overall training loss, which includes components for bounding box regression, object classification, and region proposal classification. The orange and green lines show the validation metrics for the classification loss and the regression loss, respectively. These metrics indicate the model’s performance in generating accurate region proposals and classifying/localizing detected objects.</p>
Full article ">Figure 10
<p>Training and validation metrics for the VGG-16 model. The top curves represent training (green) and validation (red) accuracy, while the bottom curves depict training (green) and validation (red) loss. The model demonstrates rapid generalization from a strong initial point, as indicated by the swift convergence of accuracy and loss metrics.</p>
Full article ">Figure 11
<p>Training and validation metrics for the DenseNet-121 model. The top curves represent training (green) and validation (red) accuracy, while the bottom curves depict training (blue) and validation (yellow) loss. The model demonstrates rapid generalization from a strong initial point, as indicated by the swift convergence of accuracy and loss metrics.</p>
Full article ">Figure 12
<p>Model predictions with background class: YOLOv5, YOLOv8, and YOLOv9. (<b>a</b>) YOLOv5 good fruit prediction. (<b>b</b>) YOLOv8 good fruit prediction. (<b>c</b>) YOLOv9 good fruit prediction. (<b>d</b>) YOLOv5 bad fruit prediction. (<b>e</b>) YOLOv8 bad fruit prediction. (<b>f</b>) YOLOv9 bad fruit prediction.</p>
Full article ">
21 pages, 4007 KiB  
Article
Lightweight Detection of Broccoli Heads in Complex Field Environments Based on LBDC-YOLO
by Zhiyu Zuo, Sheng Gao, Haitao Peng, Yue Xue, Lvhua Han, Guoxin Ma and Hanping Mao
Agronomy 2024, 14(10), 2359; https://doi.org/10.3390/agronomy14102359 - 13 Oct 2024
Viewed by 391
Abstract
Robotically selective broccoli harvesting requires precise lightweight detection models to efficiently detect broccoli heads. Therefore, this study introduces a lightweight and high-precision detection model named LBDC-YOLO (Lightweight Broccoli Detection in Complex Environment—You Look Only Once), based on the improved YOLOv8 (You Look Only [...] Read more.
Robotically selective broccoli harvesting requires precise lightweight detection models to efficiently detect broccoli heads. Therefore, this study introduces a lightweight and high-precision detection model named LBDC-YOLO (Lightweight Broccoli Detection in Complex Environment—You Look Only Once), based on the improved YOLOv8 (You Look Only Once, Version 8). The model incorporates the Slim-neck design paradigm based on GSConv to reduce computational complexity. Furthermore, Triplet Attention is integrated into the backbone network to capture cross-dimensional interactions between spatial and channel dimensions, enhancing the model’s feature extraction capability under multiple interfering factors. The original neck network structure is replaced with a BiFPN (Bidirectional Feature Pyramid Network), optimizing the cross-layer connection structure, and employing weighted fusion methods for better integration of multi-scale features. The model undergoes training and testing on a dataset constructed in real field conditions, featuring broccoli images under various influencing factors. Experimental results demonstrate that LBDC-YOLO achieves an average detection accuracy of 94.44% for broccoli. Compared to the original YOLOv8n, LBDC-YOLO achieves a 32.1% reduction in computational complexity, a 47.8% decrease in parameters, a 44.4% reduction in model size, and a 0.47 percentage point accuracy improvement. When compared to models such as YOLOv5n, YOLOv5s, and YOLOv7-tiny, LBDC-YOLO exhibits higher detection accuracy and lower computational complexity, presenting clear advantages for broccoli detection tasks in complex field environments. The results of this study provide an accurate and lightweight method for the detection of broccoli heads in complex field environments. This work aims to inspire further research in precision agriculture and to advance knowledge in model-assisted agricultural practices. Full article
Show Figures

Figure 1

Figure 1
<p>Different conditions of broccoli heads. (<b>a</b>) Broccoli heads in direct sunlight; (<b>b</b>) broccoli heads in soft and even light; (<b>c</b>) occluded broccoli heads; (<b>d</b>) broccoli heads in partial shadows; (<b>e</b>) broccoli heads in complete shadow; (<b>f</b>) wet broccoli heads.</p>
Full article ">Figure 2
<p>Network structure of LBDC-YOLO. Red rectangles in the output image indicate broccoli heads detected by the model. The different colored boxes in the figure represent modules with different functions.</p>
Full article ">Figure 3
<p>Schematic diagram of GSConv principle.</p>
Full article ">Figure 4
<p>Structure of Triplet.</p>
Full article ">Figure 5
<p>Structure of BiFPN. (<b>a</b>) Simplified structure of BiFPN; (<b>b</b>) BiFPN structure in LBDC-YOLO.</p>
Full article ">Figure 6
<p><span class="html-italic">AP</span><sub>0.5–0.95</sub> curve of LBDC-YOLO.</p>
Full article ">Figure 7
<p>Visualization results of the model. (<b>a</b>) Original image; (<b>b</b>) original image with annotations; (<b>c</b>–<b>e</b>) are visualization heatmaps of shallow, intermediate, and deep feature maps of the YOLOv8n model, respectively; (<b>f</b>–<b>h</b>) are visualization heatmaps of shallow, intermediate, and deep feature maps of the YOLOv8n model with Slim-neck, respectively; (<b>i</b>–<b>k</b>) are visualization heatmaps of shallow, intermediate, and deep feature maps of the YOLOv8n model with Slim-neck and Triplet, respectively; (<b>l</b>–<b>n</b>) are visualization heatmaps of shallow, intermediate, and deep feature maps of the LBDC-YOLO model, respectively.</p>
Full article ">Figure 8
<p>LBDC-YOLO and YOLOv8n model detection results. (<b>a</b>,<b>c</b>,<b>e</b>,<b>g</b>,<b>i</b>,<b>k</b>) show the detection results using the LBDC-YOLO model; (<b>b</b>,<b>d</b>,<b>f</b>,<b>h</b>,<b>j</b>,<b>l</b>) show the detection results using the YOLOv8n model. The environmental effects on the broccoli head in each image are listed in the first column. The red squares in the figures are model detection results, and the red arrows are used to indicate the position of the local zoom in the original figure.</p>
Full article ">
25 pages, 9054 KiB  
Article
Object Detection Algorithm for Citrus Fruits Based on Improved YOLOv5 Model
by Yao Yu, Yucheng Liu, Yuanjiang Li, Changsu Xu and Yunwu Li
Agriculture 2024, 14(10), 1798; https://doi.org/10.3390/agriculture14101798 - 12 Oct 2024
Viewed by 482
Abstract
To address the challenges of missed and false detections in citrus fruit detection caused by environmental factors such as leaf occlusion, fruit overlap, and variations in natural light in hilly and mountainous orchards, this paper proposes a citrus detection model based on an [...] Read more.
To address the challenges of missed and false detections in citrus fruit detection caused by environmental factors such as leaf occlusion, fruit overlap, and variations in natural light in hilly and mountainous orchards, this paper proposes a citrus detection model based on an improved YOLOv5 algorithm. By introducing receptive field convolutions with full 3D weights (RFCF), the model overcomes the issue of parameter sharing in convolution operations, enhancing detection accuracy. A focused linear attention (FLA) module is incorporated to improve the expressive power of the self-attention mechanism while maintaining computational efficiency. Additionally, anchor boxes were re-clustered based on the shape characteristics of target objects, and the boundary box loss function was improved to Foal-EIoU, boosting the model’s localization ability. Experiments conducted on a citrus fruit dataset labeled using LabelImg, collected from hilly and mountainous areas, showed a detection precision of 95.83% and a mean average precision (mAP) of 79.68%. This research not only significantly improves detection performance in complex environments but also provides crucial data support for precision tasks such as orchard localization and intelligent picking, demonstrating strong potential for practical applications in smart agriculture. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>The specific steps for data augmentation.</p>
Full article ">Figure 2
<p>Improved YOLOv5 network structure.</p>
Full article ">Figure 3
<p>The process of full three-dimensional weight computation based on the energy function.</p>
Full article ">Figure 4
<p>The specific process of self-attention module.</p>
Full article ">Figure 5
<p>K-means++ algorithm flow.</p>
Full article ">Figure 6
<p>The results after re-clustering (in the figure (x) represents the cluster center).</p>
Full article ">Figure 7
<p>The detection results of the aforementioned improvement strategy are visualized using heatmaps to demonstrate the localization.</p>
Full article ">Figure 8
<p>Comparison of the detection effects of fruits under dense distribution (the yellow arrows indicate the cases of false negatives and positives).</p>
Full article ">Figure 9
<p>Comparison of the detection effects under obstruction by branches and leaves.</p>
Full article ">Figure 10
<p>Loss function comparison.</p>
Full article ">Figure 11
<p>Comparison of the experimental results of citrus fruits under different spatial distributions on sunny days.</p>
Full article ">Figure 12
<p>Comparison of the experimental results of citrus fruits under different spatial distributions on rainy days.</p>
Full article ">Figure 13
<p>Comparison of detection effect under different spatial distributions.</p>
Full article ">
24 pages, 10818 KiB  
Article
ADL-YOLOv8: A Field Crop Weed Detection Model Based on Improved YOLOv8
by Zhiyu Jia, Ming Zhang, Chang Yuan, Qinghua Liu, Hongrui Liu, Xiulin Qiu, Weiguo Zhao and Jinlong Shi
Agronomy 2024, 14(10), 2355; https://doi.org/10.3390/agronomy14102355 - 12 Oct 2024
Viewed by 445
Abstract
This study presents an improved weed detection model, ADL-YOLOv8, designed to enhance detection accuracy for small targets while achieving model lightweighting. It addresses the challenge of attaining both high accuracy and low memory usage in current intelligent weeding equipment. By overcoming this issue, [...] Read more.
This study presents an improved weed detection model, ADL-YOLOv8, designed to enhance detection accuracy for small targets while achieving model lightweighting. It addresses the challenge of attaining both high accuracy and low memory usage in current intelligent weeding equipment. By overcoming this issue, the research not only reduces the hardware costs of automated impurity removal equipment but also enhances software recognition accuracy, contributing to reduced pesticide use and the promotion of sustainable agriculture. The ADL-YOLOv8 model incorporates a lighter AKConv network for better processing of specific features, an ultra-lightweight DySample upsampling module to improve accuracy and efficiency, and the LSKA-Attention mechanism for enhanced detection, particularly of small targets. On the same dataset, ADL-YOLOv8 demonstrated a 2.2% increase in precision, a 2.45% rise in recall, a 3.07% boost in [email protected], and a 1.9% enhancement in [email protected]. The model’s size was cut by 15.77%, and its computational complexity was reduced by 10.98%. These findings indicate that ADL-YOLOv8 not only exceeds the original YOLOv8n model but also surpasses the newer YOLOv9t and YOLOv10n in overall performance. The improved algorithm model makes the hardware cost required for embedded terminals lower. Full article
(This article belongs to the Special Issue Robotics and Automation in Farming)
Show Figures

Figure 1

Figure 1
<p>The type of destination in the dataset: (<b>a</b>) bluegrass; (<b>b</b>) chenopodium album; (<b>c</b>) cirsium setosum; (<b>d</b>) corn; (<b>e</b>) sedge; (<b>f</b>) portulaca oleracea.</p>
Full article ">Figure 1 Cont.
<p>The type of destination in the dataset: (<b>a</b>) bluegrass; (<b>b</b>) chenopodium album; (<b>c</b>) cirsium setosum; (<b>d</b>) corn; (<b>e</b>) sedge; (<b>f</b>) portulaca oleracea.</p>
Full article ">Figure 2
<p>Original and after data augmentation image: (<b>a</b>) original image; (<b>b</b>) charge light; (<b>c</b>) Gaussian noise; (<b>d</b>) crop image; (<b>e</b>) flipping; (<b>f</b>) flipping + crop + Gaussian noise.</p>
Full article ">Figure 3
<p>Structure of the source YOLOv8.</p>
Full article ">Figure 4
<p>Improved structure of YOLOv8. In YOLO-Head, the asterisk (*) represents arithmetic multiplication to obtain the number of channels of the convolution kernel.</p>
Full article ">Figure 5
<p>Initial sampling coordinates generated by the algorithm for arbitrary convolution kernel sizes. It provides initial sampling shapes for irregular convolution kernel sizes.</p>
Full article ">Figure 6
<p>The structure of AKConv.</p>
Full article ">Figure 7
<p>YOLOv8 PAN-FPN.</p>
Full article ">Figure 8
<p>Sampling-based dynamic upsampling.</p>
Full article ">Figure 9
<p>(<b>a</b>) In the case closest to initialization, all 4 offsets share the same initial position, ignoring positional relationships. (<b>b</b>) In bilinear initialization, the initial positions are separated to achieve uniform distribution. However, without offset modulation, the offset ranges typically overlap. (<b>c</b>) The offset ranges are constrained to reduce overlapping.</p>
Full article ">Figure 10
<p>Static sampling set generator.</p>
Full article ">Figure 11
<p>Dynamic sampling set generator.</p>
Full article ">Figure 12
<p>(<b>a</b>–<b>d</b>) Comparison of different designs for large kernel attention modules.</p>
Full article ">Figure 13
<p>(<b>a</b>) The improved YOLOv8 model. (<b>b</b>) The original YOLOv8 model.</p>
Full article ">Figure 14
<p>Detection of a large number of small targets.</p>
Full article ">Figure 15
<p>Multi-object detection.</p>
Full article ">Figure 15 Cont.
<p>Multi-object detection.</p>
Full article ">Figure 16
<p>Detection of occluded targets.</p>
Full article ">
22 pages, 6449 KiB  
Article
Development of a Smart Material Resource Planning System in the Context of Warehouse 4.0
by Oleksandr Sokolov, Angelina Iakovets, Vladyslav Andrusyshyn and Justyna Trojanowska
Eng 2024, 5(4), 2588-2609; https://doi.org/10.3390/eng5040136 (registering DOI) - 12 Oct 2024
Viewed by 397
Abstract
This study explores enhancing decision-making processes in inventory management and production operations by integrating a developed system. The proposed solution improves the decision-making process, managing the material supply of the product and inventory management in general. Based on the researched issues, the shortcomings [...] Read more.
This study explores enhancing decision-making processes in inventory management and production operations by integrating a developed system. The proposed solution improves the decision-making process, managing the material supply of the product and inventory management in general. Based on the researched issues, the shortcomings of modern enterprise resource planning systems (ERPs) were considered in the context of Warehouse 4.0. Based on the problematic areas of material accounting in manufacturing enterprises, a typical workplace was taken as a basis, which creates a gray area for warehouse systems and does not provide the opportunity of quality-managing the company’s inventory. The main tool for collecting and processing data from the workplace was the neural network. A mobile application was proposed for processing and converting the collected data for the decision-maker on material management. The YOLOv8 convolutional neural network was used to identify materials and production parts. A laboratory experiment was conducted using 3D-printed models of commercially available products at the SmartTechLab laboratory of the Technical University of Košice to evaluate the system’s effectiveness. The data from the network evaluation was obtained with the help of the ONNX format of the network for further use in conjunction with the C++ OpenCV library. The results were normalized and illustrated by diagrams. The designed system works on the principle of client–server communication; it can be easily integrated into the enterprise resource planning system. The proposed system has potential for further development, such as the expansion of the product database, facilitating efficient interaction with production systems in accordance with the circular economy, Warehouse 4.0, and lean manufacturing principles. Full article
(This article belongs to the Special Issue Feature Papers in Eng 2024)
Show Figures

Figure 1

Figure 1
<p>Block diagram of the proposed system.</p>
Full article ">Figure 2
<p>Workstation configuration.</p>
Full article ">Figure 3
<p>Mendix web interface for monitoring current state of operation in Warehouse 4.0.</p>
Full article ">Figure 4
<p>Microflow for decision-making process in Mendix database.</p>
Full article ">Figure 5
<p>Structure of YOLOv8 [<a href="#B36-eng-05-00136" class="html-bibr">36</a>].</p>
Full article ">Figure 6
<p>Used photo for training neural network.</p>
Full article ">Figure 7
<p>Some 3D-printed turbine components.</p>
Full article ">Figure 8
<p>(<b>a</b>) YOLOv8 confusion matrix; (<b>b</b>) YOLOv8 confusion matrix normalized.</p>
Full article ">Figure 9
<p>F1–confidence curve.</p>
Full article ">Figure 10
<p>Label distribution histogram.</p>
Full article ">Figure 11
<p>Scatter plots.</p>
Full article ">Figure 12
<p>(<b>a</b>) Precision–confidence curve; (<b>b</b>) completeness curve.</p>
Full article ">Figure 13
<p>Precision–Recall curve.</p>
Full article ">Figure 14
<p>Loss plots.</p>
Full article ">
18 pages, 3999 KiB  
Article
SS-YOLOv8: A Lightweight Algorithm for Surface Litter Detection
by Zhipeng Fan, Zheng Qin, Wei Liu, Ming Chen and Zeguo Qiu
Appl. Sci. 2024, 14(20), 9283; https://doi.org/10.3390/app14209283 (registering DOI) - 12 Oct 2024
Viewed by 383
Abstract
With the advancement of science and technology, pollution in rivers and water surfaces has increased, impacting both ecology and public health. Timely identification of surface waste is crucial for effective cleanup. Traditional edge detection devices struggle with limited memory and resources, making the [...] Read more.
With the advancement of science and technology, pollution in rivers and water surfaces has increased, impacting both ecology and public health. Timely identification of surface waste is crucial for effective cleanup. Traditional edge detection devices struggle with limited memory and resources, making the YOLOv8 algorithm inefficient. This paper introduces a lightweight network model for detecting water surface litter. We enhance the CSP Bottleneck with a two-convolutions (C2f) module to improve image recognition tasks. By implementing the powerful intersection over union 2 (PIoU2), we enhance model accuracy over the original CIoU. Our novel Shared Convolutional Detection Head (SCDH) minimizes parameters, while the scale layer optimizes feature scaling. Using a slimming pruning method, we further reduce the model’s size and computational needs. Our model achieves a mean average precision (mAP) of 79.9% on the surface litter dataset, with a compact size of 2.3 MB and a processing rate of 128 frames per second, meeting real-time detection requirements. This work significantly contributes to efficient environmental monitoring and offers a scalable solution for deploying advanced detection models on resource-constrained devices. Full article
(This article belongs to the Topic Computer Vision and Image Processing, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>YOLOv8 architecture diagram.</p>
Full article ">Figure 2
<p>Schematic diagram of horizontal gradient extraction.</p>
Full article ">Figure 3
<p>Schematic diagram of vertical gradient extraction.</p>
Full article ">Figure 4
<p>WS-C2f modular structure.</p>
Full article ">Figure 5
<p>Head structure.</p>
Full article ">Figure 6
<p>SCDH-Detect structure.</p>
Full article ">Figure 7
<p>PIoU loss function diagram.</p>
Full article ">Figure 8
<p>Schematic diagram of pruning.</p>
Full article ">Figure 9
<p>Selected samples of the dataset.</p>
Full article ">Figure 10
<p>(<b>a</b>) Experiment A; (<b>b</b>) Experiment B; (<b>c</b>) Experiment C.</p>
Full article ">Figure 11
<p>Comparison of the number of channels before and after pruning.</p>
Full article ">Figure 12
<p>(<b>a</b>) SS-YOLOv8n heat map. (<b>b</b>) YOLOv8n heat map. (<b>c</b>) YOLOv8-ghost heat map. (<b>d</b>) YOLOv8-mobilentv4 heat map.</p>
Full article ">
24 pages, 11251 KiB  
Article
A Novel YOLOv10-DECA Model for Real-Time Detection of Concrete Cracks
by Chaokai Zhang, Ningbo Peng, Jiaheng Yan, Lixu Wang, Yinjia Chen, Zhancheng Zhou and Ye Zhu
Buildings 2024, 14(10), 3230; https://doi.org/10.3390/buildings14103230 - 11 Oct 2024
Viewed by 375
Abstract
The You Only Look Once (YOLO) series algorithms have been widely adopted in concrete crack detection, with attention mechanisms frequently being incorporated to enhance recognition accuracy and efficiency. However, existing research is confronted by two primary challenges: the suboptimal performance of attention mechanism [...] Read more.
The You Only Look Once (YOLO) series algorithms have been widely adopted in concrete crack detection, with attention mechanisms frequently being incorporated to enhance recognition accuracy and efficiency. However, existing research is confronted by two primary challenges: the suboptimal performance of attention mechanism modules and the lack of explanation regarding how these mechanisms influence the model’s decision-making process to improve accuracy. To address these issues, a novel Dynamic Efficient Channel Attention (DECA) module is proposed in this study, which is designed to enhance the performance of the YOLOv10 model in concrete crack detection, and the effectiveness of this module is visually demonstrated through the application of interpretable analysis algorithms. In this paper, a concrete dataset with a complex background is used. Experimental results indicate that the DECA module significantly improves the model’s accuracy in crack localization and the detection of discontinuous cracks, outperforming the existing Efficient Channel Attention (ECA). When compared to the similarly sized YOLOv10n model, the proposed YOLOv10-DECA model demonstrates improvements of 4.40%, 3.06%, 4.48%, and 5.56% in precision, recall, mAP50, and mAP50-95 metrics, respectively. Moreover, even when compared with the larger YOLOv10s model, these performance indicators are increased by 2.00%, 0.04%, 2.27%, and 1.12%, respectively. In terms of speed evaluation, owing to the lightweight design of the DECA module, the YOLOv10-DECA model achieves an inference speed of 78 frames per second, which is 2.5 times faster than YOLOv10s, thereby fully meeting the requirements for real-time detection. These results demonstrate that an optimized balance between accuracy and speed in concrete crack detection tasks has been achieved by the YOLOv10-DECA model. Consequently, this study provides valuable insights for future research and applications in this field. Full article
Show Figures

Figure 1

Figure 1
<p>Framework of research methodology.</p>
Full article ">Figure 2
<p>YOLOv10-DECA.</p>
Full article ">Figure 3
<p>Diagram of Efficient Channel Attention (ECA) module [<a href="#B48-buildings-14-03230" class="html-bibr">48</a>]. Reprinted/adapted with permission from Ref. [<a href="#B48-buildings-14-03230" class="html-bibr">48</a>]. Copyright 2020, IEEE.</p>
Full article ">Figure 4
<p>The Dynamic Efficient Channel Attention (DECA) module proposed by us.</p>
Full article ">Figure 5
<p>Sample images after applying image enhancement. The label “0” indicates the class index of the detected target. In this study, only one type of object is detected, which is cracks.</p>
Full article ">Figure 6
<p>Distribution of cracks: (<b>a</b>) scatter plot of relative crack size distribution. The darker the blue color in the graph, the greater the number of cracks distributed at that location; (<b>b</b>) scatter plot of relative crack position distribution in the image. The darker the blue color in the graph, the greater the number of cracks at that size scale.</p>
Full article ">Figure 7
<p>Loss value reduction curves and improvement curves for precision, recall, mAP50, and mAP50-95 during the training process.</p>
Full article ">Figure 7 Cont.
<p>Loss value reduction curves and improvement curves for precision, recall, mAP50, and mAP50-95 during the training process.</p>
Full article ">Figure 8
<p>Histogram of YOLOv10n, YOLOv10s, YOLOv10-ECA, and YOLOv10-DECA accuracy evaluation results.</p>
Full article ">Figure 9
<p>PR curves for YOLOv10n, YOLOv10s, YOLOv10-ECA, and YOLOv10-DECA.</p>
Full article ">Figure 10
<p>YOLOv10-DECA crack identification results.</p>
Full article ">Figure 11
<p>Comparison of crack identification results in complex backgrounds.</p>
Full article ">Figure 12
<p>Class activation heatmaps by Score-CAM. The highlighted areas in the figure are the regions of interest for the model.</p>
Full article ">
20 pages, 6554 KiB  
Article
An Efficient UAV Image Object Detection Algorithm Based on Global Attention and Multi-Scale Feature Fusion
by Rui Qian and Yong Ding
Electronics 2024, 13(20), 3989; https://doi.org/10.3390/electronics13203989 - 10 Oct 2024
Viewed by 804
Abstract
Object detection technology holds significant promise in unmanned aerial vehicle (UAV) applications. However, traditional methods face challenges in detecting denser, smaller, and more complex targets within UAV aerial images. To address issues such as target occlusion and dense small objects, this paper proposes [...] Read more.
Object detection technology holds significant promise in unmanned aerial vehicle (UAV) applications. However, traditional methods face challenges in detecting denser, smaller, and more complex targets within UAV aerial images. To address issues such as target occlusion and dense small objects, this paper proposes a multi-scale object detection algorithm based on YOLOv5s. A novel feature extraction module, DCNCSPELAN4, which combines CSPNet and ELAN, is introduced to enhance the receptive field of feature extraction while maintaining network efficiency. Additionally, a lightweight Vision Transformer module, the CloFormer Block, is integrated to provide the network with a global receptive field. Moreover, the algorithm incorporates a three-scale feature fusion (TFE) module and a scale sequence feature fusion (SSFF) module in the neck network to effectively leverage multi-scale spatial information across different feature maps. To address dense small objects, an additional small object detection head was added to the detection layer. The original large object detection head was removed to reduce computational load. The proposed algorithm has been evaluated through ablation experiments and compared with other state-of-the-art methods on the VisDrone2019 and AU-AIR datasets. The results demonstrate that our algorithm outperforms other baseline methods in terms of both accuracy and speed. Compared to the YOLOv5s baseline model, the enhanced algorithm achieves improvements of 12.4% and 8.4% in AP50 and AP metrics, respectively, with only a marginal parameter increase of 0.3 M. These experiments validate the effectiveness of our algorithm for object detection in drone imagery. Full article
Show Figures

Figure 1

Figure 1
<p>Sample images taken from UAVs.</p>
Full article ">Figure 2
<p>Improved YOLOv5 network structure.</p>
Full article ">Figure 3
<p>The structure of (<b>a</b>) CSPNet, (<b>b</b>) ELAN, and (<b>c</b>) CSPELAN. CSPELAN extends the convolution module in ELAN to arbitrary computable modules modeled after CSPNet.</p>
Full article ">Figure 4
<p>The overall structure of DCNCSPELAN4. For the DCN structure. The gray grid simulates the distribution of targets in aerial photography, while the solid circle and hollow circle represent the receptive fields of DCN and regular convolutions in UAV images, respectively.</p>
Full article ">Figure 5
<p>The structure of CloFormer Block, consisting of a global branch and a local branch.</p>
Full article ">Figure 6
<p>Structure of the three-scale feature fusion operation (<b>a</b>) SSFF and (<b>b</b>) THE.</p>
Full article ">Figure 7
<p>Experimental results for all categories of Visdrone2019-test.</p>
Full article ">Figure 8
<p>Confusion matrix of (<b>a</b>) the original YOLOv5s model, (<b>b</b>) TPH-YOLOv5, and (<b>c</b>) the improved YOLOv5s model.</p>
Full article ">Figure 9
<p>Performance of the original YOLOv5, TPH-YOLOv5, and the proposed algorithm on the AU-AIR and Visdrone2019-test. (<b>a</b>) Original YOLOv5; (<b>b</b>) TPH-YOLOv5; (<b>c</b>) the proposed algorithm.</p>
Full article ">Figure 10
<p>The detection effect of our improved YOLOv5s in the urban traffic supervision scenario. (<b>a</b>) Daytime urban traffic scenario; (<b>b</b>) nighttime urban traffic scenario.</p>
Full article ">Figure 11
<p>The detection effect of our improved YOLOv5s in the urban streets supervision scenario. (<b>a</b>) Daytime urban streets scenario; (<b>b</b>) nighttime urban streets scenario.</p>
Full article ">
23 pages, 4476 KiB  
Article
YOLOv5s-Based Image Identification of Stripe Rust and Leaf Rust on Wheat at Different Growth Stages
by Qian Jiang, Hongli Wang, Zhenyu Sun, Shiqin Cao and Haiguang Wang
Plants 2024, 13(20), 2835; https://doi.org/10.3390/plants13202835 - 10 Oct 2024
Viewed by 461
Abstract
Stripe rust caused by Puccinia striiformis f. sp. tritici and leaf rust caused by Puccinia triticina, are two devastating diseases on wheat, which seriously affect the production safety of wheat. Timely detection and identification of the two diseases are essential for taking effective disease [...] Read more.
Stripe rust caused by Puccinia striiformis f. sp. tritici and leaf rust caused by Puccinia triticina, are two devastating diseases on wheat, which seriously affect the production safety of wheat. Timely detection and identification of the two diseases are essential for taking effective disease management measures to reduce wheat yield losses. To realize the accurate identification of wheat stripe rust and wheat leaf rust during the different growth stages, in this study, the image-based identification of wheat stripe rust and wheat leaf rust during different growth stages was investigated based on deep learning using image processing technology. Based on the YOLOv5s model, we built identification models of wheat stripe rust and wheat leaf rust during the seedling stage, stem elongation stage, booting stage, inflorescence emergence stage, anthesis stage, milk development stage, and all the growth stages. The models were tested on the different testing sets in the different individual growth stages and in all the growth stages. The results showed that the models performed differently in disease image identification. The model based on the disease images acquired during an individual growth stage was not suitable for the identification of the disease images acquired during the other individual growth stages, except for the model based on the disease images acquired during the milk development stage, which had acceptable identification performance on the testing sets in the anthesis stage and the milk development stage. In addition, the results demonstrated that wheat growth stages had a great influence on the image identification of the two diseases. The model built based on the disease images acquired in all the growth stages produced acceptable identification results. Mean F1 Score values between 64.06% and 79.98% and mean average precision (mAP) values between 66.55% and 82.80% were achieved on each testing set composed of the disease images acquired during an individual growth stage and on the testing set composed of the disease images acquired during all the growth stages. This study provides a basis for the image-based identification of wheat stripe rust and wheat leaf rust during the different growth stages, and it provides a reference for the accurate identification of other plant diseases. Full article
(This article belongs to the Special Issue Plant Pathology and Epidemiology for Grain, Pulses, and Cereal Crops)
Show Figures

Figure 1

Figure 1
<p>The workflow for building the YOLOv5s-based image identification models of wheat stripe rust and wheat leaf rust in the different growth stages.</p>
Full article ">Figure 2
<p>Images of wheat stripe rust and wheat leaf rust with complex backgrounds and single backgrounds acquired in an indoor environment and in the field. (<b>a</b>,<b>b</b>): Images of wheat stripe rust and wheat leaf rust, respectively, with complex backgrounds acquired in an indoor environment; (<b>c</b>,<b>d</b>): Images of wheat stripe rust and wheat leaf rust, respectively, with complex backgrounds acquired in the field; (<b>e</b>,<b>f</b>): Images of wheat stripe rust and wheat leaf rust, respectively, with single backgrounds acquired in an indoor environment; (<b>g</b>,<b>h</b>): Images of wheat stripe rust and wheat leaf rust, respectively, with single backgrounds acquired in the field.</p>
Full article ">
24 pages, 12126 KiB  
Article
Efficient Optimized YOLOv8 Model with Extended Vision
by Qi Zhou, Zhou Wang, Yiwen Zhong, Fenglin Zhong and Lijin Wang
Sensors 2024, 24(20), 6506; https://doi.org/10.3390/s24206506 - 10 Oct 2024
Viewed by 622
Abstract
In the field of object detection, enhancing algorithm performance in complex scenarios represents a fundamental technological challenge. To address this issue, this paper presents an efficient optimized YOLOv8 model with extended vision (YOLO-EV), which optimizes the performance of the YOLOv8 model through a [...] Read more.
In the field of object detection, enhancing algorithm performance in complex scenarios represents a fundamental technological challenge. To address this issue, this paper presents an efficient optimized YOLOv8 model with extended vision (YOLO-EV), which optimizes the performance of the YOLOv8 model through a series of innovative improvement measures and strategies. First, we propose a multi-branch group-enhanced fusion attention (MGEFA) module and integrate it into YOLO-EV, which significantly boosts the model’s feature extraction capabilities. Second, we enhance the existing spatial pyramid pooling fast (SPPF) layer by integrating large scale kernel attention (LSKA), improving the model’s efficiency in processing spatial information. Additionally, we replace the traditional IOU loss function with the Wise-IOU loss function, thereby enhancing localization accuracy across various target sizes. We also introduce a P6 layer to augment the model’s detection capabilities for multi-scale targets. Through network structure optimization, we achieve higher computational efficiency, ensuring that YOLO-EV consumes fewer computational resources than YOLOv8s. In the validation section, preliminary tests on the VOC12 dataset demonstrate YOLO-EV’s effectiveness in standard object detection tasks. Moreover, YOLO-EV has been applied to the CottonWeedDet12 and CropWeed datasets, which are characterized by complex scenes, diverse weed morphologies, significant occlusions, and numerous small targets. Experimental results indicate that YOLO-EV exhibits superior detection accuracy in these complex agricultural environments compared to the original YOLOv8s and other state-of-the-art models, effectively identifying and locating various types of weeds, thus demonstrating its significant practical application potential. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

Figure 1
<p>The images from the VOC2012 dataset.</p>
Full article ">Figure 2
<p>Visualization of data augmentation during training.</p>
Full article ">Figure 3
<p>YOLOv8 model architecture.</p>
Full article ">Figure 4
<p>YOLO-EV model architecture. Compared to YOLOv8, YOLO-EV incorporates the innovative MGEFAC2f module, SPPF_LSKA, and an additional P6 layer, resulting in an extra detection head. Moreover, it replaces the traditional IoU metric with Wise-IoU. The specific implementation details will be elaborated in the following sections.</p>
Full article ">Figure 5
<p>MGEFA network architecture.</p>
Full article ">Figure 6
<p>MGEFAC2f module structure.</p>
Full article ">Figure 7
<p>LSKA module and SPPF_LSKA module.</p>
Full article ">Figure 8
<p>Comparison of ablation study results.</p>
Full article ">Figure 9
<p>Comparison chart of experimental results between the improved model YOLO-EV and YOLOv8s.</p>
Full article ">Figure 10
<p>Comparison of experimental results for the CWD12 dataset.</p>
Full article ">Figure 11
<p>Display of experimental results for the CWD12 dataset.</p>
Full article ">Figure 12
<p>Comparison of experimental results for the CropWeed dataset.</p>
Full article ">Figure 13
<p>Display of experimental results for the CropWeed dataset.</p>
Full article ">
18 pages, 19441 KiB  
Article
YOLOv5s-ECCW: A Lightweight Detection Model for Sugarcane Smut in Natural Environments
by Min Yu, Fengbing Li, Xiupeng Song, Xia Zhou, Xiaoqiu Zhang, Zeping Wang, Jingchao Lei, Qiting Huang, Guanghu Zhu, Weihua Huang, Hairong Huang, Xiaohang Chen, Yunhai Yang, Dongmei Huang, Qiufang Li, Hui Fang and Meixin Yan
Agronomy 2024, 14(10), 2327; https://doi.org/10.3390/agronomy14102327 - 10 Oct 2024
Viewed by 342
Abstract
Sugarcane smut, a serious disease caused by the fungus Sporosorium scitamineum, can result in 30% to 100% cane loss. The most affordable and efficient measure of preventing and handling sugarcane smut disease is to select disease-resistant varieties. A comprehensive evaluation of disease [...] Read more.
Sugarcane smut, a serious disease caused by the fungus Sporosorium scitamineum, can result in 30% to 100% cane loss. The most affordable and efficient measure of preventing and handling sugarcane smut disease is to select disease-resistant varieties. A comprehensive evaluation of disease resistance based on the incidence of smut disease is essential during the selection process, necessitating the rapid and accurate identification of sugarcane smut. Traditional identification methods, which rely on visual observation of symptoms, are time-consuming, costly, and inefficient. To address these limitations, we present the lightweight sugarcane smut detection model (YOLOv5s-ECCW), which incorporates several innovative features. Specifically, the EfficientNetV2 is incorporated into the YOLOv5 network to achieve model compression while maintaining high detection accuracy. The convolutional block attention mechanism (CBAM) is added to the backbone network to improve its feature extraction capability and suppress irrelevant information. The C3STR module is used to replace the C3 module, enhancing the ability to capture global large targets. The WIoU loss function is used in place of the CIoU one to improve the bounding box regression’s accuracy. The experimental results demonstrate that the YOLOv5s-ECCW model achieves a mean average precision (mAP) of 97.8% with only 4.9 G FLOPs and 3.25 M parameters. Compared with the original YOLOv5, our improvements include a 0.2% increase in mAP, a 54% reduction in parameters, and a 70.3% decrease in computational requirements. The proposed model outperforms YOLOv4, SSD, YOLOv5, and YOLOv8 in terms of accuracy, efficiency, and model size. The YOLOv5s-ECCW model meets the urgent need for the accurate real-time identification of sugarcane smut, supporting better disease management and selection of resistant varieties. Full article
(This article belongs to the Special Issue In-Field Detection and Monitoring Technology in Precision Agriculture)
Show Figures

Figure 1

Figure 1
<p>Focus operation diagram. A value was obtained at each pixel interval in an image to concentrate width and height information in the channel space. The input channels were expanded by 4, resulting in a spliced image with 12 channels instead of the original RGB 3-channel model.</p>
Full article ">Figure 2
<p>Structure of the CBAM module. The channel attention module and spatial attention module sequentially refine feature maps. ⨁ denotes the addition of two features. <span class="html-fig-inline" id="agronomy-14-02327-i001"><img alt="Agronomy 14 02327 i001" src="/agronomy/agronomy-14-02327/article_deploy/html/images/agronomy-14-02327-i001.png"/></span> denotes the Sigmoid activation function. ⨂ denotes the multiplication of the input feature maps by the corresponding attention module.</p>
Full article ">Figure 3
<p>Structure of the C3 and C3STR module. (<b>a</b>) Original C3 module. (<b>b</b>) C3STR module. The C3STR module utilizes the Swin transformer (STR) to reduce the model’s parameters. The STR module is a pairwise combination of two different Swin transformer blocks (W-MSA, SW-MSA). The dashed box shows the detailed structure of the STR module.</p>
Full article ">Figure 4
<p>Overall structure of YOLOv5s-ECCW network. Backbone: EfficientnetV2, the convolutional block attention mechanism (CBAM module), and SPP. Neck: FPN + PAN with C3STR. Head: three detection heads detect small, medium, and large objects, respectively. Firstly, 640 × 640 RGB images are given as the input, then the image features are extracted and fused through backbone and neck. Finally, three detection heads with three different sizes are the output.</p>
Full article ">Figure 5
<p>Comparison of mAPs of different backbone networks. The horizontal axis represents training epochs. The vertical axis represents the mean accuracy precision (mAP) at an IoU threshold of 0.5. The lines represent the continuous change in the mAP value as the number of training rounds increases.</p>
Full article ">Figure 6
<p>Detection performance of different improved models. This bar chart contains a variety of evaluation indicators to describe the detection performance of different improved models. The symbol ‘-’ in the figure indicates that the opposite number of the parameter is taken.</p>
Full article ">Figure 7
<p>Confusion matrix. The horizontal axis represents the ground truth classes and the vertical axis represents the predicted classes. Each cell element represents the proportion of the number of the predicted class to the total number of the true class. The diagonal elements represent correctly classified outcomes. All other off-diagonal elements along a column are wrong predictions.</p>
Full article ">Figure 8
<p>Confidence and accuracy metrics of the YOLOv5s-ECCW. (<b>A</b>) The F1 curve. The model had an F1 score of 0.95 in the training set. (<b>B</b>) The P-R curve. The mAP@0.5 value of the training set was 0.964. (<b>C</b>) The P curve. The model’s accuracy consistently exceeded 80% when the confidence level reached 0.1 or higher. (<b>D</b>) The R curve. The recall remained high for confidence levels below 0.8 but gradually decreased beyond that threshold (0.8).</p>
Full article ">Figure 9
<p>Validation set prediction results. This is an example of the prediction results on the validation set during model training. It contains 16 images randomly combined from the validation set. Each image contains the target prediction category with its confidence level.</p>
Full article ">Figure 10
<p>The trend of each loss function in the training process. This model has four loss functions. The box_loss denotes a regression error. The obj_loss represents a confidence error. The cls_loss represents a target category loss function. Total_loss represents the sum of the first three losses.</p>
Full article ">Figure 11
<p>Recognition results. A visualization of the results of detecting images outside the dataset using four object detection algorithms (YOLOv4,YOLOv5,YOLOv8,YOLOv5s-ECCW). Each predicted bounding box shows the predicted label of the detected smut and the confidence of the predicted result. To make the prediction frames in images clearer, some of the detection images are cropped in size.</p>
Full article ">
Back to TopTop