[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

Search Results (131)

Search Parameters:
Keywords = BiFormer

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 14422 KiB  
Article
YOLO-SegNet: A Method for Individual Street Tree Segmentation Based on the Improved YOLOv8 and the SegFormer Network
by Tingting Yang, Suyin Zhou, Aijun Xu, Junhua Ye and Jianxin Yin
Agriculture 2024, 14(9), 1620; https://doi.org/10.3390/agriculture14091620 (registering DOI) - 15 Sep 2024
Abstract
In urban forest management, individual street tree segmentation is a fundamental method to obtain tree phenotypes, which is especially critical. Most existing tree image segmentation models have been evaluated on smaller datasets and lack experimental verification on larger, publicly available datasets. Therefore, this [...] Read more.
In urban forest management, individual street tree segmentation is a fundamental method to obtain tree phenotypes, which is especially critical. Most existing tree image segmentation models have been evaluated on smaller datasets and lack experimental verification on larger, publicly available datasets. Therefore, this paper, based on a large, publicly available urban street tree dataset, proposes YOLO-SegNet for individual street tree segmentation. In the first stage of the street tree object detection task, the BiFormer attention mechanism was introduced into the YOLOv8 network to increase the contextual information extraction and improve the ability of the network to detect multiscale and multishaped targets. In the second-stage street tree segmentation task, the SegFormer network was proposed to obtain street tree edge information more efficiently. The experimental results indicate that our proposed YOLO-SegNet method, which combines YOLOv8+BiFormer and SegFormer, achieved a 92.0% mean intersection over union (mIoU), 95.9% mean pixel accuracy (mPA), and 97.4% accuracy on a large, publicly available urban street tree dataset. Compared with those of the fully convolutional neural network (FCN), lite-reduced atrous spatial pyramid pooling (LR-ASPP), pyramid scene parsing network (PSPNet), UNet, DeepLabv3+, and HRNet, the mIoUs of our YOLO-SegNet increased by 10.5, 9.7, 5.0, 6.8, 4.5, and 2.7 percentage points, respectively. The proposed method can effectively support smart agroforestry development. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>(<b>A</b>) is the number distribution of street tree images; (<b>B</b>) is the street tree image annotation.</p>
Full article ">Figure 2
<p>Examples of street tree object detection and instance segmentation annotated images for different tree species.</p>
Full article ">Figure 3
<p>YOLO-SegNet model. The CBS is the basic module, including the Conv2d layer, BatchNorm2d layer, and Sigmoid Linear Unit (SiLU) layer. The function of the CBS module is to introduce a cross-stage partial connection to improve the feature expression ability and information transfer efficiency. The role of the Spatial Pyramid Pooling Fast (SPPF) module is to fuse larger-scale global information to improve the performance of object detection. The bottleneck block can reduce the computational complexity and the number of parameters.</p>
Full article ">Figure 4
<p>(<b>A</b>) The overall architecture of BiFormer; (<b>B</b>) details of a BiFormer block.</p>
Full article ">Figure 5
<p>(<b>a</b>) Vanilla attention. (<b>b</b>–<b>d</b>) Local window [<a href="#B40-agriculture-14-01620" class="html-bibr">40</a>,<a href="#B42-agriculture-14-01620" class="html-bibr">42</a>], axial stripe [<a href="#B39-agriculture-14-01620" class="html-bibr">39</a>], and dilated window [<a href="#B41-agriculture-14-01620" class="html-bibr">41</a>,<a href="#B42-agriculture-14-01620" class="html-bibr">42</a>]. (<b>e</b>) Deformable attention [<a href="#B43-agriculture-14-01620" class="html-bibr">43</a>]. (<b>f</b>) Bilevel routing attention, BRA [<a href="#B6-agriculture-14-01620" class="html-bibr">6</a>].</p>
Full article ">Figure 6
<p>Gathering key–value pairs in the top <span class="html-italic">k</span> related windows.</p>
Full article ">Figure 7
<p>(<b>A</b>,<b>B</b>) are the loss function curves of the object detection network on the train and validation sets, respectively; (<b>C</b>,<b>D</b>) are the loss function curves of tree classification on the train and validation sets, respectively; (<b>E</b>–<b>H</b>) are the change curves of the four segmentation indicator values on the validation set, respectively.</p>
Full article ">Figure 8
<p>(<b>A</b>) Thermal map examples of YOLOv8 series models and YOLOv8m+BiFormer in the training process; (<b>B</b>) example results of the different object detection models on the test set.</p>
Full article ">Figure 9
<p>(<b>A</b>) The training loss function curves of the segmentation models without the object detection module. (<b>B</b>) The training loss function curves of the segmentation models with the object detection module.</p>
Full article ">Figure 10
<p>Performance of different segmentation models on the validation and test sets: (<b>A<sub>1</sub></b>,<b>A<sub>2</sub></b>) the segmentation results on the validation set; (<b>B<sub>1</sub></b>,<b>B<sub>2</sub></b>) the segmentation results on the test set.</p>
Full article ">Figure 11
<p>Results of the different segmentation models on the test set.</p>
Full article ">
15 pages, 7669 KiB  
Article
Advanced Multi-Label Fire Scene Image Classification via BiFormer, Domain-Adversarial Network and GCN
by Yu Bai, Dan Wang, Qingliang Li, Taihui Liu and Yuheng Ji
Fire 2024, 7(9), 322; https://doi.org/10.3390/fire7090322 (registering DOI) - 15 Sep 2024
Abstract
Detecting wildfires presents significant challenges due to the presence of various potential targets in fire imagery, such as smoke, vehicles, and people. To address these challenges, we propose a novel multi-label classification model based on BiFormer’s feature extraction method, which constructs sparse region-indexing [...] Read more.
Detecting wildfires presents significant challenges due to the presence of various potential targets in fire imagery, such as smoke, vehicles, and people. To address these challenges, we propose a novel multi-label classification model based on BiFormer’s feature extraction method, which constructs sparse region-indexing relations and performs feature extraction only in key regions, thereby facilitating more effective capture of flame characteristics. Additionally, we introduce a feature screening method based on a domain-adversarial neural network (DANN) to minimize misclassification by accurately determining feature domains. Furthermore, a feature discrimination method utilizing a Graph Convolutional Network (GCN) is proposed, enabling the model to capture label correlations more effectively and improve performance by constructing a label correlation matrix. This model enhances cross-domain generalization capability and improves recognition performance in fire scenarios. In the experimental phase, we developed a comprehensive dataset by integrating multiple fire-related public datasets, and conducted detailed comparison and ablation experiments. Results from the tenfold cross-validation demonstrate that the proposed model significantly improves recognition of multi-labeled images in fire scenarios. Compared with the baseline model, the mAP increased by 4.426%, CP by 4.14% and CF1 by 7.04%. Full article
Show Figures

Figure 1

Figure 1
<p>Rescaled samples of fire images from CFDB.</p>
Full article ">Figure 2
<p>Rescaled samples of fire images from KT.</p>
Full article ">Figure 3
<p>Rescaled samples of fire images from VOC2012.</p>
Full article ">Figure 4
<p>Model framework diagram.</p>
Full article ">Figure 5
<p>An example of the conditional probability relationship between two labels is provided. Typically, when the image contains “flame”, there is a high likelihood that “smoke” is also present. However, if “smoke” is observed, “flame” may not necessarily be present.</p>
Full article ">Figure 6
<p>BiFormer Block operational flow (<span class="html-fig-inline" id="fire-07-00322-i001"><img alt="Fire 07 00322 i001" src="/fire/fire-07-00322/article_deploy/html/images/fire-07-00322-i001.png"/></span> Represents a residual connection).</p>
Full article ">Figure 7
<p>Domain classification and label classification network architecture.</p>
Full article ">Figure 8
<p>Visualization of results (where red represents a higher level of concern).</p>
Full article ">Figure 9
<p>Visualization of multi-label classification results (where green means the prediction is correct and red means the prediction is incorrect).</p>
Full article ">Figure 10
<p>Accuracy comparisons with different values of τ.</p>
Full article ">Figure 11
<p>Example of predicting sun.</p>
Full article ">Figure 12
<p>Example of predicting clouds.</p>
Full article ">Figure 13
<p>Example of predicting fire and smog.</p>
Full article ">
19 pages, 8739 KiB  
Article
Evaluation of AV Deadheading Strategies
by Sruthi Mantri, David Bergman and Nicholas Lownes
Future Transp. 2024, 4(3), 1059-1077; https://doi.org/10.3390/futuretransp4030051 - 12 Sep 2024
Viewed by 190
Abstract
The transition of the vehicle fleet to incorporate AV will be a long and complex process. AVs will gradually form a larger and larger share of the fleet mix, offering opportunities and challenges for improved efficiency and safety. At any given point during [...] Read more.
The transition of the vehicle fleet to incorporate AV will be a long and complex process. AVs will gradually form a larger and larger share of the fleet mix, offering opportunities and challenges for improved efficiency and safety. At any given point during this transition a portion of the AV fleet will be consuming roadway capacity while deadheading, which means operating without passengers. Should these unoccupied vehicles simply utilize the shortest paths to their next destination, they will contribute to congestion for the rest of the roadway users without providing any benefit to human passengers. There is an opportunity to develop routing strategies for deadheading AVs that mitigate or eliminate their contribution to congestion while still serving the mobility needs of AV owners or passengers. Some of the AV fleet will be privately owned, while some will be part of a shared AV fleet. In the former, some AVs will be owned by households that are lower-income and benefit from the ability to have fewer vehicles to serve the mobility needs of the household. In these cases, it is especially important that deadheading AVs can meet household mobility needs while also limiting the contribution to roadway congestion. The aim of this study is to develop and evaluate routing strategies for deadheading autonomous vehicles (AVs) that balance the reduction of roadway congestion and the mobility needs of households. By proposing and testing a bi-objective program, this study seeks to identify effective methodologies for routing unoccupied AVs in a manner that mitigates their negative impact on traffic while still fulfilling essential transportation requirements of the household. Three strategies are proposed to deploy AV deadheading methodology to route deadheading vehicles on longer paths, reducing congestion for occupied vehicles, while still meeting the trip-making needs of households. Case studies on two transportation networks are presented alongside their practical implications and computational requirements. Full article
Show Figures

Figure 1

Figure 1
<p>Example network.</p>
Full article ">Figure 2
<p>Flow of the Strategy.</p>
Full article ">Figure 3
<p>Sioux Falls Network, Sioux Falls, North Dakota, USA (with permission from Taylor and Francis).</p>
Full article ">Figure 4
<p>Distribution of the delay of the deadheading vehicles at <span class="html-italic">e</span> = 0.1.</p>
Full article ">Figure 5
<p>Delay (<math display="inline"><semantics> <mrow> <msup> <mrow> <mi>μ</mi> </mrow> <mrow> <mi>r</mi> <mo>,</mo> <mi>s</mi> </mrow> </msup> <mo>)</mo> </mrow> </semantics></math> of the OVs and UVs—Strategy 1.</p>
Full article ">Figure 6
<p>Delay (<math display="inline"><semantics> <mrow> <msup> <mrow> <mi>μ</mi> </mrow> <mrow> <mi>r</mi> <mo>,</mo> <mi>s</mi> </mrow> </msup> <mo>)</mo> </mrow> </semantics></math> of the OVs and UVs—Strategy 2.</p>
Full article ">Figure 7
<p>Delay (<math display="inline"><semantics> <mrow> <msup> <mrow> <mi>μ</mi> </mrow> <mrow> <mi>r</mi> <mo>,</mo> <mi>s</mi> </mrow> </msup> <mo>)</mo> </mrow> </semantics></math> of the OVs and UVs—Strategy 3.</p>
Full article ">Figure 8
<p>Total System Travel Time Savings.</p>
Full article ">Figure 9
<p>Algorithm Convergence for Strategy 1—Sioux Falls Network.</p>
Full article ">Figure 9 Cont.
<p>Algorithm Convergence for Strategy 1—Sioux Falls Network.</p>
Full article ">Figure 10
<p>Delay (<math display="inline"><semantics> <mrow> <msup> <mrow> <mi>μ</mi> </mrow> <mrow> <mi>r</mi> <mo>,</mo> <mi>s</mi> </mrow> </msup> <mo>)</mo> </mrow> </semantics></math> of the OVs and UVs—Strategy 1.</p>
Full article ">Figure 11
<p>Delay (<math display="inline"><semantics> <mrow> <msup> <mrow> <mi>μ</mi> </mrow> <mrow> <mi>r</mi> <mo>,</mo> <mi>s</mi> </mrow> </msup> <mo>)</mo> </mrow> </semantics></math> of the OVs and—Strategy 2.</p>
Full article ">Figure 12
<p>Delay (<math display="inline"><semantics> <mrow> <msup> <mrow> <mi>μ</mi> </mrow> <mrow> <mi>r</mi> <mo>,</mo> <mi>s</mi> </mrow> </msup> <mo>)</mo> </mrow> </semantics></math> of the OVs and—Strategy 3.</p>
Full article ">Figure 13
<p>Total System Travel Time Plots.</p>
Full article ">Figure 14
<p>Sensitivity Analysis of TSTT.</p>
Full article ">
24 pages, 49819 KiB  
Article
Personnel Monitoring in Shipboard Surveillance Using Improved Multi-Object Detection and Tracking Algorithm
by Yiming Li, Bin Zhang, Yichen Liu, Huibing Wang and Shibo Zhang
Sensors 2024, 24(17), 5756; https://doi.org/10.3390/s24175756 - 4 Sep 2024
Viewed by 360
Abstract
Detecting and tracking personnel onboard is an important measure to prevent ships from being invaded by outsiders and ensure ship security. Ships are characterized by more cabins, numerous equipment, and dense personnel, so there are problems such as unpredictable personnel trajectories, frequent occlusions, [...] Read more.
Detecting and tracking personnel onboard is an important measure to prevent ships from being invaded by outsiders and ensure ship security. Ships are characterized by more cabins, numerous equipment, and dense personnel, so there are problems such as unpredictable personnel trajectories, frequent occlusions, and many small targets, which lead to the poor performance of existing multi-target-tracking algorithms on shipboard surveillance videos. This study conducts research in the context of onboard surveillance and proposes a multi-object detection and tracking algorithm for anti-intrusion on ships. First, this study designs the BR-YOLO network to provide high-quality object-detection results for the tracking algorithm. The shallow layers of its backbone network use the BiFormer module to capture dependencies between distant objects and reduce information loss. Second, the improved C2f module is used in the deep layer of BR-YOLO to introduce the RepGhost structure to achieve model lightweighting through reparameterization. Then, the Part OSNet network is proposed, which uses different pooling branches to focus on multi-scale features, including part-level features, thereby obtaining strong Re-ID feature representations and providing richer appearance information for personnel tracking. Finally, by integrating the appearance information for association matching, the tracking trajectory is generated in Tracking-By-Detection mode and validated on the self-constructed shipboard surveillance dataset. The experimental results show that the algorithm in this paper is effective in shipboard surveillance. Compared with the present mainstream algorithms, the MOTA, HOTP, and IDF1 are enhanced by about 10 percentage points, the MOTP is enhanced by about 7 percentage points, and IDs are also significantly reduced, which is of great practical significance for the prevention of intrusion by ship personnel. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>Structure of the tracking algorithm.</p>
Full article ">Figure 2
<p>Improved YOLOv8 network architecture.</p>
Full article ">Figure 3
<p>BRA structure.</p>
Full article ">Figure 4
<p>Network structure diagram of BiFormer and C2f.</p>
Full article ">Figure 5
<p>Comparison diagram of Bottleneck structure.</p>
Full article ">Figure 6
<p>Comparison diagram of C2f structure.</p>
Full article ">Figure 7
<p>Part OSNet backbone network schematic.</p>
Full article ">Figure 8
<p>OSNet foundation building blocks schematic.</p>
Full article ">Figure 9
<p>Operation flow of Part OSNet.</p>
Full article ">Figure 10
<p>Example of autonomous datasets.</p>
Full article ">Figure 11
<p>Results of Grad-CAM heat map visualization.</p>
Full article ">Figure 12
<p>Performance comparison of object-detection algorithm.</p>
Full article ">Figure 13
<p>Object Detection comparison experiments (green box in the figure indicates a missed target, white circle circled for redundant background).</p>
Full article ">Figure 14
<p>Performance comparison of tracking algorithm on the Bohai Sea Ro-Ro Ship Dataset.</p>
Full article ">Figure 15
<p>Performance comparison of tracking algorithm on MOT17.</p>
Full article ">Figure 16
<p>Results of multi-object-tracking algorithms (the green dashed line indicates a missed target, the circle indicates an ID error or skip, the blue dashed line indicates an incorrectly tracked target, and the yellow box indicates a misdetected target).</p>
Full article ">
24 pages, 5578 KiB  
Article
Study on Nighttime Pedestrian Trajectory-Tracking from the Perspective of Driving Blind Spots
by Wei Zhao, Congcong Ren and Ao Tan
Electronics 2024, 13(17), 3460; https://doi.org/10.3390/electronics13173460 - 31 Aug 2024
Viewed by 381
Abstract
With the acceleration of urbanization and the growing demand for traffic safety, developing intelligent systems capable of accurately recognizing and tracking pedestrian trajectories at night or under low-light conditions has become a research focus in the field of transportation. This study aims to [...] Read more.
With the acceleration of urbanization and the growing demand for traffic safety, developing intelligent systems capable of accurately recognizing and tracking pedestrian trajectories at night or under low-light conditions has become a research focus in the field of transportation. This study aims to improve the accuracy and real-time performance of nighttime pedestrian-detection and -tracking. A method that integrates the multi-object detection algorithm YOLOP with the multi-object tracking algorithm DeepSORT is proposed. The improved YOLOP algorithm incorporates the C2f-faster structure in the Backbone and Neck sections, enhancing feature extraction capabilities. Additionally, a BiFormer attention mechanism is introduced to focus on the recognition of small-area features, the CARAFE module is added to improve shallow feature fusion, and the DyHead dynamic target-detection head is employed for comprehensive fusion. In terms of tracking, the ShuffleNetV2 lightweight module is integrated to reduce model parameters and network complexity. Experimental results demonstrate that the proposed FBCD-YOLOP model improves lane detection accuracy by 5.1%, increases the IoU metric by 0.8%, and enhances detection speed by 25 FPS compared to the baseline model. The accuracy of nighttime pedestrian-detection reached 89.6%, representing improvements of 1.3%, 0.9%, and 3.8% over the single-task YOLO v5, multi-task TDL-YOLO, and the original YOLOP models, respectively. These enhancements significantly improve the model’s detection performance in complex nighttime environments. The enhanced DeepSORT algorithm achieved an MOTA of 86.3% and an MOTP of 84.9%, with ID switch occurrences reduced to 5. Compared to the ByteTrack and StrongSORT algorithms, MOTA improved by 2.9% and 0.4%, respectively. Additionally, network parameters were reduced by 63.6%, significantly enhancing the real-time performance of nighttime pedestrian-detection and -tracking, making it highly suitable for deployment on intelligent edge computing surveillance platforms. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Nighttime driver’s blind spot pedestrian-tracking technology route.</p>
Full article ">Figure 2
<p>Algorithm implementation flowchart.</p>
Full article ">Figure 3
<p>C2faster structural diagram. (<b>a</b>) FasterNet block, (<b>b</b>) C2f-faster.</p>
Full article ">Figure 4
<p>BiFormer attention mechanism structure diagram.</p>
Full article ">Figure 5
<p>CARAFE upsampling structure diagram.</p>
Full article ">Figure 6
<p>Dynamic detection head DyHead structure diagram.</p>
Full article ">Figure 7
<p>Improved YOLOP network structure diagram.</p>
Full article ">Figure 8
<p>ShuffleNetV2 structure diagram.</p>
Full article ">Figure 9
<p>DIoU schematic diagram.</p>
Full article ">Figure 10
<p>Improved DeepSORT structure flowchart.</p>
Full article ">Figure 11
<p>The lane line-detection results at night are presented. In Scene 1, the road at night is unobstructed and the lane lines are clear. In Scene 2, the road at night has obstructions. In Scene 3, the lane lines on the road at night are unclear.</p>
Full article ">Figure 12
<p>FBCD-YOLOP Training Process Results Diagram.</p>
Full article ">Figure 13
<p>The results of the loss during the training and validation process of the tracking algorithm.</p>
Full article ">Figure 14
<p>Nighttime pedestrian-tracking results. (<b>a</b>) A frame from the first video sequence showing the initial detection and tracking of pedestrians by the proposed algorithm. (<b>b</b>) The corresponding frame from the first video sequence where the IDS-tracking process is shown; the proposed algorithm accurately tracks pedestrian ID3 through the crowd, while the YOLOP-DeepSort algorithm exhibits ID switches (highlighted by orange circles). (<b>c</b>) A frame from the second video sequence showing the proposed algorithm’s detection of pedestrians with no ID changes or false detections. (<b>d</b>) The corresponding frame from the second video sequence where the YOLOP-DeepSort algorithm mistakenly identifies a tree trunk and a wall crack as pedestrians (highlighted by red circles), demonstrating the superiority of the proposed algorithm in avoiding false detections.</p>
Full article ">
23 pages, 25505 KiB  
Article
A New Method for Non-Destructive Identification and Tracking of Multi-Object Behaviors in Beef Cattle Based on Deep Learning
by Guangbo Li, Jiayong Sun, Manyu Guan, Shuai Sun, Guolong Shi and Changjie Zhu
Animals 2024, 14(17), 2464; https://doi.org/10.3390/ani14172464 - 24 Aug 2024
Viewed by 561
Abstract
The method proposed in this paper provides theoretical and practical support for the intelligent recognition and management of beef cattle. Accurate identification and tracking of beef cattle behaviors are essential components of beef cattle production management. Traditional beef cattle identification and tracking methods [...] Read more.
The method proposed in this paper provides theoretical and practical support for the intelligent recognition and management of beef cattle. Accurate identification and tracking of beef cattle behaviors are essential components of beef cattle production management. Traditional beef cattle identification and tracking methods are time-consuming and labor-intensive, which hinders precise cattle farming. This paper utilizes deep learning algorithms to achieve the identification and tracking of multi-object behaviors in beef cattle, as follows: (1) The beef cattle behavior detection module is based on the YOLOv8n algorithm. Initially, a dynamic snake convolution module is introduced to enhance the ability to extract key features of beef cattle behaviors and expand the model’s receptive field. Subsequently, the BiFormer attention mechanism is incorporated to integrate high-level and low-level feature information, dynamically and sparsely learning the behavioral features of beef cattle. The improved YOLOv8n_BiF_DSC algorithm achieves an identification accuracy of 93.6% for nine behaviors, including standing, lying, mounting, fighting, licking, eating, drinking, working, and searching, with average 50 and 50:95 precisions of 96.5% and 71.5%, showing an improvement of 5.3%, 5.2%, and 7.1% over the original YOLOv8n. (2) The beef cattle multi-object tracking module is based on the Deep SORT algorithm. Initially, the detector is replaced with YOLOv8n_BiF_DSC to enhance detection accuracy. Subsequently, the re-identification network model is switched to ResNet18 to enhance the tracking algorithm’s capability to gather appearance information. Finally, the trajectory generation and matching process of the Deep SORT algorithm is optimized with secondary IOU matching to reduce ID mismatching errors during tracking. Experimentation with five different complexity levels of test video sequences shows improvements in IDF1, IDS, MOTA, and MOTP, among other metrics, with IDS reduced by 65.8% and MOTA increased by 2%. These enhancements address issues of tracking omission and misidentification in sparse and long-range dense environments, thereby facilitating better tracking of group-raised beef cattle and laying a foundation for intelligent detection and tracking in beef cattle farming. Full article
(This article belongs to the Section Cattle)
Show Figures

Figure 1

Figure 1
<p>Data collection and dataset construction process.</p>
Full article ">Figure 2
<p>Data enhancement example.</p>
Full article ">Figure 3
<p>DarkLabel labeling interface. (“星期四” is Thursday).</p>
Full article ">Figure 4
<p>Sample appearance re-recognition dataset.</p>
Full article ">Figure 5
<p>Framework for nondestructive identification and tracking of beef cattle behavior.</p>
Full article ">Figure 6
<p>Deep SORT target tracking process.</p>
Full article ">Figure 7
<p>YOLOv8n_BiF_DSC algorithm flow.</p>
Full article ">Figure 8
<p>DSConv graphical representation. (<b>a</b>) DSConv coordinate calculation. (<b>b</b>) DSConv sensory field.</p>
Full article ">Figure 9
<p>BiFormer flowchart. (<b>a</b>) Overall structure. (<b>b</b>) Detailed structure of BiFormer module.</p>
Full article ">Figure 10
<p>ResNet18 structure diagram.</p>
Full article ">Figure 11
<p>Trajectory generation and matching process.</p>
Full article ">Figure 12
<p>Loss curve.</p>
Full article ">Figure 13
<p>Beef cattle behavioral detection chart.</p>
Full article ">Figure 14
<p>Visualization map.</p>
Full article ">Figure 15
<p>Convergence of loss value and top-1 error.</p>
Full article ">Figure 16
<p>Accuracy curve graphs ((<b>A</b>) accuracy curve of the original algorithm; (<b>B</b>) accuracy curve of ResNet18).</p>
Full article ">Figure 17
<p>Track results before and after improvements. (“星期四” is Thursday).</p>
Full article ">Figure 18
<p>Sparse cattle herd tracking results. (“星期四” is Thursday).</p>
Full article ">Figure 19
<p>Remote dense cattle herd tracking results. (“星期四” is Thursday).</p>
Full article ">
14 pages, 8316 KiB  
Article
Maize Anthesis-Silking Interval Estimation via Image Detection under Field Rail-Based Phenotyping Platform
by Lvhan Zhuang, Chuanyu Wang, Haoyuan Hao, Wei Song and Xinyu Guo
Agronomy 2024, 14(8), 1723; https://doi.org/10.3390/agronomy14081723 - 5 Aug 2024
Viewed by 510
Abstract
The Anthesis-Silking Interval (ASI) is a crucial indicator of the synchrony of reproductive development in maize, reflecting its sensitivity to adverse environmental conditions such as heat stress and drought. This paper presents an automated method for detecting the maize ASI index using a [...] Read more.
The Anthesis-Silking Interval (ASI) is a crucial indicator of the synchrony of reproductive development in maize, reflecting its sensitivity to adverse environmental conditions such as heat stress and drought. This paper presents an automated method for detecting the maize ASI index using a field high-throughput phenotyping platform. Initially, high temporal-resolution visible-light image sequences of maize plants from the tasseling to silking stage are collected using a field rail-based phenotyping platform. Then, the training results of different sizes of YOLOv8 models on this dataset are compared to select the most suitable base model for the task of detecting maize tassels and ear silks. The chosen model is enhanced by incorporating the SENetv2 and the dual-layer routing attention mechanism BiFormer, named SEBi-YOLOv8. The SEBi-YOLOv8 model, with these combined modules, shows improvements of 2.3% and 8.2% in mAP over the original model, reaching 0.989 and 0.886, respectively. Finally, SEBi-YOLOv8 is used for the dynamic detection of maize tassels and ear silks in maize populations. The experimental results demonstrate the method’s high detection accuracy, with a correlation coefficient (R2) of 0.987 and an RMSE of 0.316. Based on these detection results, the ASI indices of different inbred lines are calculated and compared. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Location of the experimental test field and the operating trajectory and original visible light images collected by the platform from the corn tasseling stage to the silking stage.</p>
Full article ">Figure 2
<p>Partial images of the dataset created.</p>
Full article ">Figure 3
<p>Flowchart of data processing for detection of flowering and silking in maize.</p>
Full article ">Figure 4
<p>BRA module.</p>
Full article ">Figure 5
<p>Comparison of ResNeXt, SENet, and SENetV2 modules.</p>
Full article ">Figure 6
<p>Structure of the SEBi-YOLOv8 model.</p>
Full article ">Figure 7
<p>SEBi-YOLOv8 model detection result images, where (<b>a</b>–<b>d</b>) are four different planting areas, with the original images on the left and the detection results on the right.</p>
Full article ">
19 pages, 10494 KiB  
Article
RT-DETR-Tomato: Tomato Target Detection Algorithm Based on Improved RT-DETR for Agricultural Safety Production
by Zhimin Zhao, Shuo Chen, Yuheng Ge, Penghao Yang, Yunkun Wang and Yunsheng Song
Appl. Sci. 2024, 14(14), 6287; https://doi.org/10.3390/app14146287 - 19 Jul 2024
Viewed by 978
Abstract
The detection of tomatoes is of vital importance for enhancing production efficiency, with image recognition-based tomato detection methods being the primary approach. However, these methods face challenges such as the difficulty in extracting small targets, low detection accuracy, and slow processing speeds. Therefore, [...] Read more.
The detection of tomatoes is of vital importance for enhancing production efficiency, with image recognition-based tomato detection methods being the primary approach. However, these methods face challenges such as the difficulty in extracting small targets, low detection accuracy, and slow processing speeds. Therefore, this paper proposes an improved RT-DETR-Tomato model for efficient tomato detection under complex environmental conditions. The model mainly consists of a Swin Transformer block, a BiFormer module, path merging, multi-scale convolutional layers, and fully connected layers. In this proposed model, Swin Transformer is chosen as the new backbone network to replace ResNet50 because of its superior ability to capture broader global dependency relationships and contextual information. Meanwhile, a lightweight BiFormer block is adopted in Swin Transformer to reduce computational complexity through content-aware flexible computation allocation. Experimental results show that the average accuracy of the final RT-DETR-Tomato model is greatly improved compared to the original model, and the model training time is greatly reduced, demonstrating better environmental adaptability. In the future, the RT-DETR-Tomato model can be integrated with intelligent patrol and picking robots, enabling precise identification of crops and ensuring the safety of crops and the smooth progress of agricultural production. Full article
Show Figures

Figure 1

Figure 1
<p>Architecture of the RT-DETR-Tomato-BS Model.</p>
Full article ">Figure 2
<p>Architecture of ResNet50.</p>
Full article ">Figure 3
<p>Architecture of the Swin Transformer.</p>
Full article ">Figure 4
<p>Structure of the Swin Transformer block.</p>
Full article ">Figure 5
<p>Patch merging layer, W-MSA, and SW-MSA modules.</p>
Full article ">Figure 6
<p>Framework of the BiFormer block.</p>
Full article ">Figure 7
<p>Schematic representation of the specific implementation module of BRA.</p>
Full article ">Figure 8
<p>Tomato image samples under different environments in the natural scene dataset: (<b>a</b>) Single target without occlusion, (<b>b</b>) Multiple targets with occlusion, (<b>c</b>) tomato cluster, (<b>d</b>) enhanced lighting, (<b>e</b>) diminished lighting, and (<b>f</b>) multiple targets with or without occlusion.</p>
Full article ">Figure 9
<p>Training performance at each stage of the model.</p>
Full article ">Figure 10
<p>Comparison of tomato image detection between RT-DETR and RT-DETR-Tomato-BS models based on the natural scene dataset.</p>
Full article ">Figure 11
<p>P–R curves of different methods for ablation study.</p>
Full article ">Figure 12
<p>Comparison of the precision curves for each model.</p>
Full article ">
14 pages, 4039 KiB  
Article
The Adaptive Alternation of Intestinal Microbiota and Regulation of Host Genes Jointly Promote Pigs to Digest Appropriate High-Fiber Diets
by Yunchao Zhang, Hui Li, Bengao Li, Jiayi He, Chen Peng, Yanshe Xie, Guiqing Huang, Pengju Zhao and Zhengguang Wang
Animals 2024, 14(14), 2076; https://doi.org/10.3390/ani14142076 - 16 Jul 2024
Viewed by 618
Abstract
Although studies have revealed the significant impact of dietary fiber on growth performance and nutrient digestibility, the specific characteristics of the intestinal microbiota and gene regulation in pigs capable of digesting high-fiber diets remained unclear. To investigate the traits associated with roughage tolerance [...] Read more.
Although studies have revealed the significant impact of dietary fiber on growth performance and nutrient digestibility, the specific characteristics of the intestinal microbiota and gene regulation in pigs capable of digesting high-fiber diets remained unclear. To investigate the traits associated with roughage tolerance in the Chinese indigenous pig breed, we conducted comparative analysis of growth performance, apparent fiber digestibility, intestinal microbiota, SCFA concentrations and intestinal transcriptome in Tunchang pigs, feeding them diets with different wheat bran levels. The results indicated that the growth performance of Tunchang pigs was not significantly impacted, and the apparent total tract digestibility of crude fiber was significantly improved with increasing dietary fiber content. High-fiber diets altered the diversity of intestinal microbiota, and increased the relative abundance of Prevotella, CF231, as well as the concentrations of isobutyrate, valerate and isovalerate. The LDA analysis identified potential microbial biomarkers that could be associated with roughage tolerance, such as Prevotella stercorea, and Eubacterium biforme. In addition, appropriate high-fiber diets containing 4.34% crude fiber upregulated the mRNA expressions of PYY, AQP8, and SLC5A8, while downregulating the mRNA expressions of CKM and CNN1.This indicated that appropriate high-fiber diets may inhibit intestine motility and increase the absorption of water and SCFAs. Full article
Show Figures

Figure 1

Figure 1
<p>The comparison of microbial diversity of Tunchang pigs among groups. Alpha diversity of cecal microbiota (<b>a</b>) and colonic microbiota (<b>b</b>). Anoism test of cecal microbiota (<b>c</b>) and colonic microbiota (<b>d</b>) based on Bray–Curtis distance.</p>
Full article ">Figure 2
<p>The intestinal microbial composition in the cecum and colon of Tunchang pigs evaluated at the phylum (<b>a</b>) and genus (<b>b</b>) level. The LDA analysis in the cecum (<b>c</b>) and colon (<b>d</b>).</p>
Full article ">Figure 3
<p>The predicted functions of cecal microbiota of Tunchang pigs based on PICRUSt analysis (<b>a</b>). Significantly changed microbial functions of KEGG (<b>b</b>) and MetaCyc (<b>c</b>). <sup>#</sup> <span class="html-italic">p</span> &lt; 0.05 compared with group A. Significantly changed SCFAs (<b>d</b>). * <span class="html-italic">p</span> &lt; 0.05; ns, not significant.</p>
Full article ">Figure 4
<p>The top 20 differentially expressed genes between group A and B (<b>a</b>) in the cecum of Tunchang pigs. KEGG enrichment analysis of differentially expressed genes (<b>b</b>). Relative expression of several genes as determined by quantitative real-time PCR (<b>c</b>). AQP8, aquaporin 8; SLC5A8, solute carrier family 5 member 8; PYY, peptide YY; CKM, creatine kinase, M-type; CNN1, calponin 1. <sup>#</sup> <span class="html-italic">p</span> &lt; 0.05, <sup>##</sup> <span class="html-italic">p</span> &lt; 0.01.</p>
Full article ">
17 pages, 4756 KiB  
Article
CFE-YOLOv8s: Improved YOLOv8s for Steel Surface Defect Detection
by Shuxin Yang, Yang Xie, Jianqing Wu, Weidong Huang, Hongsheng Yan, Jingyong Wang, Bi Wang, Xiangchun Yu, Qiang Wu and Fei Xie
Electronics 2024, 13(14), 2771; https://doi.org/10.3390/electronics13142771 - 15 Jul 2024
Viewed by 755
Abstract
Due to the low detection accuracy in steel surface defect detection and the constraints of limited hardware resources, we propose an improved model for steel surface defect detection, named CBiF-FC-EFC-YOLOv8s (CFE-YOLOv8s), including CBS-BiFormer (CBiF) modules, Faster-C2f (FC) modules, and EMA-Faster-C2f (EFC) modules. Firstly, [...] Read more.
Due to the low detection accuracy in steel surface defect detection and the constraints of limited hardware resources, we propose an improved model for steel surface defect detection, named CBiF-FC-EFC-YOLOv8s (CFE-YOLOv8s), including CBS-BiFormer (CBiF) modules, Faster-C2f (FC) modules, and EMA-Faster-C2f (EFC) modules. Firstly, because of the potential information loss that convolutional neural networks (CNN) may encounter when dealing with miniature targets, the CBiF combines CNN with Transformer to optimize local and global features. Secondly, to address the increased computational complexity caused by the extensive use of convolutional layers, the FC uses the FasterNet block to reduce redundant computations and memory access. Lastly, the EMA is incorporated into the FC to design the EFC module and enhance feature fusion capability while ensuring the light weight of the model. CFE-YOLOv8s achieves [email protected] values of 77.8% and 69.5% on the NEU-DET and GC10-DET datasets, respectively, representing enhancements of 3.1% and 2.8% over YOLOv8s, with reductions of 22% and 18% in model parameters and FLOPS. The CFE-YOLOv8s demonstrates superior overall performance and balance compared to other advanced models. Full article
(This article belongs to the Special Issue Machine Learning and Deep Learning Based Pattern Recognition)
Show Figures

Figure 1

Figure 1
<p>Six types of steel defects: (<b>a</b>) Inclusion. (<b>b</b>) Crazing. (<b>c</b>) Scratches. (<b>d</b>) Rolled-in scale. (<b>e</b>) Pitted surface. (<b>f</b>) Patches.</p>
Full article ">Figure 2
<p>The structure of the C2f module.</p>
Full article ">Figure 3
<p>The overall framework of CFE-YOLOv8s.</p>
Full article ">Figure 4
<p>The structure of the CBiF module.</p>
Full article ">Figure 5
<p>The concatenation of the CBiF module.</p>
Full article ">Figure 6
<p>The structure of the FasterNet block.</p>
Full article ">Figure 7
<p>The structure of the FC module.</p>
Full article ">Figure 8
<p>The structure of the EMA-Faster block in EFC.</p>
Full article ">Figure 9
<p>The experimental results of the CBiF module.</p>
Full article ">Figure 10
<p>Defect detection results on NEU-DET.</p>
Full article ">Figure 11
<p>FLOPS with different models.</p>
Full article ">Figure 12
<p>Defect detection results on GC10-DET.</p>
Full article ">Figure 13
<p>Heatmaps of YOLOv8s and CFE-YOLOv8s.</p>
Full article ">Figure 14
<p>Detection results of YOLOv8s and CFE-YOLOv8s: (<b>a</b>) detection results of YOLOv8s; (<b>b</b>) detection results of CEF-YOLOv8s.</p>
Full article ">
17 pages, 2713 KiB  
Article
Gasoline Engine Misfire Fault Diagnosis Method Based on Improved YOLOv8
by Zhichen Li, Zhao Qin, Weiping Luo and Xiujun Ling
Electronics 2024, 13(14), 2688; https://doi.org/10.3390/electronics13142688 - 9 Jul 2024
Viewed by 662
Abstract
In order to realize the online diagnosis and prediction of gasoline engine fire faults, this paper proposes an improved misfire fault detection algorithm model based on YOLOv8 for sound signals of gasoline engines. The improvement involves substituting a C2f module in the YOLOv8 [...] Read more.
In order to realize the online diagnosis and prediction of gasoline engine fire faults, this paper proposes an improved misfire fault detection algorithm model based on YOLOv8 for sound signals of gasoline engines. The improvement involves substituting a C2f module in the YOLOv8 backbone network by a BiFormer attention module and another C2f module substituted by a CBAM module that combines channel and spatial attention mechanisms which enhance the neural network’s capacity to extract the complex features. The normal and misfire sound signals of a gasoline engine are processed by wavelet transformation and converted to time–frequency images for the training, verification, and testing of convolutional neural network. The experimental results show that the precision of the improved YOLOv8 algorithm model is 99.71% for gasoline engine fire fault tests, which is 2 percentage points higher than for the YOLOv8 network model. The diagnosis time of each sound is less than 100 ms, making it suitable for developing IoT devices for gasoline engine misfire fault diagnosis and driverless vehicles. Full article
Show Figures

Figure 1

Figure 1
<p>Engine sound signal acquisition.</p>
Full article ">Figure 2
<p>Engine sound signal. (<b>a</b>) normal; (<b>b</b>) one-cylinder misfire; (<b>c</b>) two-cylinder misfire.</p>
Full article ">Figure 3
<p>Wavelet transformation time–frequency image. (<b>a</b>) normal; (<b>b</b>) one-cylinder misfire; (<b>c</b>) two-cylinder misfire.</p>
Full article ">Figure 4
<p>Structure of BiFormer.</p>
Full article ">Figure 5
<p>The overview of CBAM (Note: Pictures are from the Ref. [<a href="#B34-electronics-13-02688" class="html-bibr">34</a>]). (<b>A</b>) the structure of channel attention; (<b>B</b>) the structure of spatial attention; (<b>C</b>) the structure of CBAM attention.</p>
Full article ">Figure 5 Cont.
<p>The overview of CBAM (Note: Pictures are from the Ref. [<a href="#B34-electronics-13-02688" class="html-bibr">34</a>]). (<b>A</b>) the structure of channel attention; (<b>B</b>) the structure of spatial attention; (<b>C</b>) the structure of CBAM attention.</p>
Full article ">Figure 6
<p>Structural comparison of YOLOv8 and YOLOv8-CBBF. (<b>A</b>) Structure of YOLOv8; (<b>B</b>) Structure of YOLOv8-CBBF.</p>
Full article ">Figure 7
<p>YOLOv8-CBBF training process. (<b>A</b>) Train Loss; (<b>B</b>) Validation Loss; (<b>C</b>) Train accuracy.</p>
Full article ">
22 pages, 7148 KiB  
Article
An Improved YOLOv8n Used for Fish Detection in Natural Water Environments
by Zehao Zhang, Yi Qu, Tan Wang, Yuan Rao, Dan Jiang, Shaowen Li and Yating Wang
Animals 2024, 14(14), 2022; https://doi.org/10.3390/ani14142022 - 9 Jul 2024
Viewed by 695
Abstract
To improve detection efficiency and reduce cost consumption in fishery surveys, target detection methods based on computer vision have become a new method for fishery resource surveys. However, the specialty and complexity of underwater photography result in low detection accuracy, limiting its use [...] Read more.
To improve detection efficiency and reduce cost consumption in fishery surveys, target detection methods based on computer vision have become a new method for fishery resource surveys. However, the specialty and complexity of underwater photography result in low detection accuracy, limiting its use in fishery resource surveys. To solve these problems, this study proposed an accurate method named BSSFISH-YOLOv8 for fish detection in natural underwater environments. First, replacing the original convolutional module with the SPD-Conv module allows the model to lose less fine-grained information. Next, the backbone network is supplemented with a dynamic sparse attention technique, BiFormer, which enhances the model’s attention to crucial information in the input features while also optimizing detection efficiency. Finally, adding a 160 × 160 small target detection layer (STDL) improves sensitivity for smaller targets. The model scored 88.3% and 58.3% in the two indicators of mAP@50 and mAP@50:95, respectively, which is 2.0% and 3.3% higher than the YOLOv8n model. The results of this research can be applied to fishery resource surveys, reducing measurement costs, improving detection efficiency, and bringing environmental and economic benefits. Full article
(This article belongs to the Section Aquatic Animals)
Show Figures

Figure 1

Figure 1
<p>YOLOv8 network structure.</p>
Full article ">Figure 2
<p>SPD-Conv module structure.</p>
Full article ">Figure 3
<p>BiFormer attention structure.</p>
Full article ">Figure 4
<p>Target size distribution of the dataset (Darker colours represent a greater number of distributions).</p>
Full article ">Figure 5
<p>BSSFISH-YOLOv8 network structure.</p>
Full article ">Figure 6
<p>Examples of enhanced images: (<b>a</b>) vertical flip; (<b>b</b>) horizontal flip; (<b>c</b>) brightness adjustment; (<b>d</b>) Gaussian blur; (<b>e</b>) affine transformation translation; (<b>f</b>) affine transformation scaling; (<b>g</b>) channel addition; (<b>h</b>) rotate; (<b>i</b>) Gaussian noise.</p>
Full article ">Figure 7
<p>Typical images in the dataset: (<b>a</b>) blur; (<b>b</b>) occlusion; (<b>c</b>) small targets.</p>
Full article ">Figure 8
<p>Comparison of heat maps: (<b>a</b>) YOLOv8n; (<b>b</b>) BSSFISH-YOLOv8.</p>
Full article ">Figure 9
<p>Feature maps of different scales: (<b>a</b>) 160 × 160; (<b>b</b>) 80 × 80; (<b>c</b>) 40 × 40; (<b>d</b>) 20 × 20.</p>
Full article ">Figure 10
<p>Comparison of improvement effects. From top to bottom: original images; YOLOv8n; BSSFISH-YOLOv8. (<b>a</b>) blur; (<b>b</b>) occlusion; (<b>c</b>) small targets.</p>
Full article ">Figure 11
<p>mAP@50 curve.</p>
Full article ">Figure 12
<p>Confusion matrix for identification of different fish species. (<b>a</b>) quantitative information (<b>b</b>) ratio information.</p>
Full article ">Figure 13
<p>Demonstration of cases with higher and lower fish detection accuracy: (<b>a</b>) Blue Catfish; (<b>b</b>) Yellowfin Bream; (<b>c</b>) Eastern Striped Grunter.</p>
Full article ">
22 pages, 13347 KiB  
Article
Research on Automated Fiber Placement Surface Defect Detection Based on Improved YOLOv7
by Liwei Wen, Shihao Li, Zhentao Dong, Haiqing Shen and Entao Xu
Appl. Sci. 2024, 14(13), 5657; https://doi.org/10.3390/app14135657 - 28 Jun 2024
Viewed by 492
Abstract
Due to the black and glossy appearance of the carbon fiber prepreg bundle surface, the accurate identification of surface defects in automated fiber placement (AFP) presents a high level of difficulty. Currently, the enhanced YOLOv7 algorithm demonstrates certain performance advantages in this detection [...] Read more.
Due to the black and glossy appearance of the carbon fiber prepreg bundle surface, the accurate identification of surface defects in automated fiber placement (AFP) presents a high level of difficulty. Currently, the enhanced YOLOv7 algorithm demonstrates certain performance advantages in this detection task, yet issues with missed detections, false alarms, and low confidence levels persist. Therefore, this study proposes an improved YOLOv7 algorithm to further enhance the performance and generalization of surface defect detection in AFP. Firstly, to enhance the model’s feature extraction capability, the BiFormer attention mechanism is introduced to make the model pay more attention to small target defects, thereby improving feature discriminability. Next, the AFPN structure is used to replace the PAFPN at the neck layer to strengthen feature fusion, preserve semantic information to a greater extent, and finely integrate multi-scale features. Finally, WIoU is adopted to replace CIoU as the bounding box regression loss function, making it more sensitive to small targets, enabling more accurate prediction of object bounding boxes, and enhancing the model’s detection accuracy and generalization capability. Through a series of ablation experiments, the improved YOLOv7 shows a 10.5% increase in mAP and a 14 FPS increase in frame rate, with a maximum detection speed of 35 m/min during the AFP process, meeting the requirements of online detection and thus being able to be applied to surface defect detection in AFP operations. Full article
Show Figures

Figure 1

Figure 1
<p>Automatic fiber placement machine (column type).</p>
Full article ">Figure 2
<p>The structure of the AFPN [<a href="#B41-applsci-14-05657" class="html-bibr">41</a>] (the black arrows indicate convolution operations, while the aqua arrows indicate adaptive spatial fusion).</p>
Full article ">Figure 3
<p>Adaptive spatial feature fusion [<a href="#B41-applsci-14-05657" class="html-bibr">41</a>].</p>
Full article ">Figure 4
<p>The 8-tow automated fiber placement machine developed by Nanjing University of Aeronautics and Astronautics (NUAA).</p>
Full article ">Figure 5
<p>Defect detection system’s detection process.</p>
Full article ">Figure 6
<p>Surface defect data collection in AFP process.</p>
Full article ">Figure 7
<p>Comparison of Location loss.</p>
Full article ">Figure 8
<p>Comparison of Objectness loss and Classification loss trained based on five models mentioned above: (<b>a</b>) obj_loss; (<b>b</b>) cls_loss.</p>
Full article ">Figure 9
<p>Comparison chart of mAP for five algorithms.</p>
Full article ">Figure 10
<p>Comparison of detection results between original YOLOv7 (<b>a1</b>–<b>f1</b>) and Improved YOLOv7 (<b>a2</b>–<b>f2</b>) (overlap confidence in <b>e2</b> is 0.91, overlap confidence in <b>f1</b> is 0.89, and overlap confidence in <b>f2</b> is 0.94).</p>
Full article ">Figure 10 Cont.
<p>Comparison of detection results between original YOLOv7 (<b>a1</b>–<b>f1</b>) and Improved YOLOv7 (<b>a2</b>–<b>f2</b>) (overlap confidence in <b>e2</b> is 0.91, overlap confidence in <b>f1</b> is 0.89, and overlap confidence in <b>f2</b> is 0.94).</p>
Full article ">Figure 10 Cont.
<p>Comparison of detection results between original YOLOv7 (<b>a1</b>–<b>f1</b>) and Improved YOLOv7 (<b>a2</b>–<b>f2</b>) (overlap confidence in <b>e2</b> is 0.91, overlap confidence in <b>f1</b> is 0.89, and overlap confidence in <b>f2</b> is 0.94).</p>
Full article ">
16 pages, 4397 KiB  
Article
BPN-YOLO: A Novel Method for Wood Defect Detection Based on YOLOv7
by Rijun Wang, Yesheng Chen, Fulong Liang, Bo Wang, Xiangwei Mou and Guanghao Zhang
Forests 2024, 15(7), 1096; https://doi.org/10.3390/f15071096 - 25 Jun 2024
Viewed by 949
Abstract
The detection of wood defect is a crucial step in wood processing and manufacturing, determining the quality and reliability of wood products. To achieve accurate wood defect detection, a novel method named BPN-YOLO is proposed. The ordinary convolution in the ELAN module of [...] Read more.
The detection of wood defect is a crucial step in wood processing and manufacturing, determining the quality and reliability of wood products. To achieve accurate wood defect detection, a novel method named BPN-YOLO is proposed. The ordinary convolution in the ELAN module of the YOLOv7 backbone network is replaced with Pconv partial convolution, resulting in the P-ELAN module. Wood defect detection performance is improved by this modification while unnecessary redundant computations and memory accesses are reduced. Additionally, the Biformer attention mechanism is introduced to achieve more flexible computation allocation and content awareness. The IOU loss function is replaced with the NWD loss function, addressing the sensitivity of the IOU loss function to small defect location fluctuations. The BPN-YOLO model has been rigorously evaluated using an optimized wood defect dataset, and ablation and comparison experiments have been performed. The experimental results show that the mean average precision (mAP) of BPN-YOLO is improved by 7.4% relative to the original algorithm, which can better meet the need to accurately detecting surface defects on wood. Full article
(This article belongs to the Special Issue Wood Quality and Wood Processing)
Show Figures

Figure 1

Figure 1
<p>Wood defects [<a href="#B28-forests-15-01096" class="html-bibr">28</a>]: (<b>A</b>) Live_Knot, (<b>B</b>) Dead_Knot, (<b>C</b>) Quartzity, (<b>D</b>) Knot_with_crack, (<b>E</b>) Knot_missing, (<b>F</b>) Crack, (<b>G</b>) Overgrown, (<b>H</b>) Resin (Resin pocket), (<b>I</b>) Marrow (Pith), (<b>J</b>) Blue_stain.</p>
Full article ">Figure 2
<p>YOLOv7 network structure.</p>
Full article ">Figure 3
<p>BPN-YOLO network structure.</p>
Full article ">Figure 4
<p>P-ELAN Module.</p>
Full article ">Figure 5
<p>Structure of Biformer attention mechanism [<a href="#B37-forests-15-01096" class="html-bibr">37</a>].</p>
Full article ">Figure 6
<p>Precision-recall (P–R) curves: (<b>a</b>) YOLOv5; (<b>b</b>) YOLOv7; (<b>c</b>) YOLOv8; (<b>d</b>) YOLOv9; (<b>e</b>) RT-DETR; (<b>f</b>) BPN-YOLO.</p>
Full article ">Figure 7
<p>Comparison of detection results.</p>
Full article ">Figure 8
<p>Gradient-weighted class activation map (Grad-CAM).</p>
Full article ">
16 pages, 4417 KiB  
Article
UO-YOLO: Ureteral Orifice Detection Network Based on YOLO and Biformer Attention Mechanism
by Li Liang and Wang Yuanjun
Appl. Sci. 2024, 14(12), 5124; https://doi.org/10.3390/app14125124 - 12 Jun 2024
Viewed by 845
Abstract
Background and Purpose: In urological surgery, accurate localization of the ureteral orifice is crucial for procedures such as ureteral stent insertion, assessment of ureteral orifice lesions, and prostate tumor resection. Consequently, we have developed and validated a computer-assisted ureteral orifice detection system that [...] Read more.
Background and Purpose: In urological surgery, accurate localization of the ureteral orifice is crucial for procedures such as ureteral stent insertion, assessment of ureteral orifice lesions, and prostate tumor resection. Consequently, we have developed and validated a computer-assisted ureteral orifice detection system that combines the YOLO deep convolutional neural network and the attention mechanism. Data: The cases were partitioned into a training set and a validation set at a 4:1 ratio, with 84 cases comprising 820 images in the training set and 20 cases containing 223 images in the validation set. Method: We improved the YOLO network structure to accomplish the detection task. Based on the one-stage strategy, we replaced the backbone of YOLOv5 with a structure composed of ConvNeXt blocks. Additionally, we introduced GRN (Global Response Normalization) modules and SE blocks into the blocks to enhance deep feature diversity. In the feature enhancement section, we incorporated the BiFormer attention structure, which provides long-distance context dependencies without adding excessive computational costs. Finally, we improved the prediction box loss function to WIoU (Wise-IoU), enhancing the accuracy of the prediction boxes. Results: Testing on 223 cystoscopy images demonstrated a precision of 0.928 and recall of 0.756 for our proposed ureteral orifice detection network. With an overlap threshold of 0.5, the mAP of our proposed image detection system reached 0.896. The entire model achieved a single-frame detection speed of 5.7 ms on the platform, with a frame rate of 175FPS. Conclusion: We have enhanced a deep learning framework based on the one-stage YOLO strategy, suitable for real-time detection of the ureteral orifice in endoscopic scenarios. The system simultaneously maintains high accuracy and good real-time performance. This method holds substantial potential as an excellent learning and feedback system for trainees and new urologists in clinical settings. Full article
Show Figures

Figure 1

Figure 1
<p>Morphology of the ureteral orifice that is challenging to identify in ureteroscopy, where (<b>a</b>–<b>d</b>) represent images of actual scenes after cropping, and (<b>e</b>–<b>h</b>) are their corresponding images with annotated ureteral orifices.</p>
Full article ">Figure 2
<p>Dataset processing.</p>
Full article ">Figure 3
<p>Network structure.</p>
Full article ">Figure 4
<p>Comparison block of modules from different network architectures.</p>
Full article ">Figure 5
<p>BiFormer structure.</p>
Full article ">Figure 6
<p>The positional relationship between the predicted rectangle P and the ground truth rectangle G.</p>
Full article ">Figure 7
<p>Confusion matrix.</p>
Full article ">Figure 8
<p>The detection results of the ureteral orifice using the YOLO series algorithm models. We performed experiments on 20 cases and randomly selected the results of 8 cases to display in <a href="#applsci-14-05124-f008" class="html-fig">Figure 8</a>. The unshown data results are consistent with those displayed. The first column shows the ground truth, the second column shows the detection results of our proposed algorithm, the third column shows the results of YOLOv5n, the fourth column shows the results of YOLOv7m, and the fifth column shows the results of YOLOv8m. The images in different rows of each group come from different independent case videos to comprehensively demonstrate the detection effects.</p>
Full article ">
Back to TopTop