[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (447)

Search Parameters:
Keywords = accurate vehicle localization

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 4666 KiB  
Article
Quantifying Qiyi Glacier Surface Dirtiness Using UAV and Sentinel-2 Imagery
by Jiangtao Chen, Ninglian Wang, Yuwei Wu, Anan Chen, Chenlie Shi, Mingjie Zhao and Longjiang Xie
Remote Sens. 2024, 16(17), 3351; https://doi.org/10.3390/rs16173351 - 9 Sep 2024
Viewed by 210
Abstract
The glacier surface is composed not only of ice or snow but also of a heterogeneous mixture of various materials. The presence of light-absorbing impurities darkens the glacier surface, reducing local reflectance and thereby accelerating the glacier melting process. However, our understanding of [...] Read more.
The glacier surface is composed not only of ice or snow but also of a heterogeneous mixture of various materials. The presence of light-absorbing impurities darkens the glacier surface, reducing local reflectance and thereby accelerating the glacier melting process. However, our understanding of the spatial distribution of these impurities remains limited, and there is a lack of studies on quantifying the dirty degree of glacier surfaces. During the Sentinel satellite overpass on 21 August 2023, we used an ASD FieldSpec3 spectrometer to measure the reflectance spectra of glacier surfaces with varying degrees of dirtiness on the Qiyi glacier, Qinghai–Tibet Plateau. Using Multiple Endmember Spectral Mixture Analysis (MESMA), the Sentinel imagery was decomposed to generate fraction images of five primary ice surface materials as follows: coarse-grained snow, slightly dirty ice, moderately dirty ice, extremely dirty ice, and debris. Using unmanned aerial vehicle (UAV) imagery with a 0.05 m resolution, the primary ice surface was delineated and utilized as reference data to validate the fraction images. The findings revealed a strong correlation between the fraction images and the reference data (R2 ≥ 0.66, RMSE ≤ 0.21). Based on pixel-based classification from the UAV imagery, approximately 80% of the glacier surface is covered by slightly dirty ice (19.2%), moderately dirty ice (33.3%), extremely dirty ice (26.3%), and debris (1.2%), which significantly contributes to its darkening. Our study demonstrates the effectiveness of using Sentinel imagery in conjunction with MESMA to map the degree of glacier surface dirtiness accurately. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Location of the Qiyi glacier (red star). (<b>b</b>) A true-color RGB image (10 m resolution) of the glacier, with the blue curve outlining its boundary. Red circles represent spectral sampling points, yellow triangles indicate UAV ground control points, and pink rectangles delineate the validation areas. (<b>c</b>,<b>d</b>) are images of the glacier terminus taken on 31 July 2013, and 15 August 2023, respectively.</p>
Full article ">Figure 2
<p>(<b>a</b>) Spectral measurements were collected with a fiber optic probe ~1 m above the ice surface. (<b>b</b>) The actual measured spectral curves are depicted with solid black lines, while colored circles represent the reflectance values at the central wavelengths of Sentinel-2B bands (B2-B8A bands correspond to red to pink hues on the graph).</p>
Full article ">Figure 3
<p>Spectral curves after SRF conversion, where solid lines represent mean values and shaded areas denote standard deviations (<b>a</b>). Photographs of the following categories of ice are shown: (<b>b</b>) coarse-grained snow; (<b>c</b>) slightly dirty ice; (<b>d</b>) moderately dirty ice; (<b>e</b>) extremely dirty ice; and (<b>f</b>) supraglacial rivers. The spectrometer’s field of view is a ~50 cm diameter circle; a pen is placed for scale, aiming to provide readers with a sense of proportion for better comprehension.</p>
Full article ">Figure 4
<p>A comparison of measured reflectance and satellite products, where red pentagrams denote the sampling positions of the spectrometer. (<b>a</b>,<b>b</b>) represent relatively clean glacier surfaces, while (<b>c</b>,<b>d</b>) depict relatively dirty glacier surfaces. L2A denotes products produced by the ESA, FLAASH (10 m) signifies atmospheric correction through FLAASH, and L2A (Sen2cor) indicates correction via the Sen2cor plugin. SRF refers to spectral response function conversion, the green line represents the measured spectra, and L1C denotes ESA L1C products.</p>
Full article ">Figure 5
<p>(<b>a</b>) The UAV image and (<b>b</b>) the SVM-classified image.</p>
Full article ">Figure 6
<p>The final spectral endmembers for the following different glacier surface types: (<b>a</b>) coarse-grained snow; (<b>b</b>) slightly dirty ice; (<b>c</b>) moderately dirty ice; (<b>d</b>) extremely dirty ice; and (<b>e</b>) debris.</p>
Full article ">Figure 7
<p>Fraction images for the following five distinct ice surface types are presented: (<b>a</b>) coarse-grained snow; (<b>b</b>) slightly dirty ice; (<b>c</b>) moderately dirty ice; (<b>d</b>) extremely dirty ice; and (<b>e</b>) debris.</p>
Full article ">Figure 8
<p>A regression model was constructed to examine the relationship between MESMA fraction images and reference fraction (UAV images). The solid line illustrates the degree of fitting, while the shaded area represents the 95% confidence interval. The determination coefficient (R<sup>2</sup>) and root mean square error (RMSE) are presented, <span class="html-italic">n</span> = 330.</p>
Full article ">
19 pages, 5224 KiB  
Article
A Hybrid Algorithm of LSTM and Factor Graph for Improving Combined GNSS/INS Positioning Accuracy during GNSS Interruptions
by Fuchao Liu, Hailin Zhao and Wenjue Chen
Sensors 2024, 24(17), 5605; https://doi.org/10.3390/s24175605 - 29 Aug 2024
Viewed by 411
Abstract
In urban road environments, global navigation satellite system (GNSS) signals may be interrupted due to occlusion by buildings and obstacles, resulting in reduced accuracy and discontinuity of combined GNSS/inertial navigation system (INS) positioning. Improving the accuracy and robustness of combined GNSS/INS positioning systems [...] Read more.
In urban road environments, global navigation satellite system (GNSS) signals may be interrupted due to occlusion by buildings and obstacles, resulting in reduced accuracy and discontinuity of combined GNSS/inertial navigation system (INS) positioning. Improving the accuracy and robustness of combined GNSS/INS positioning systems for land vehicles in the presence of GNSS interruptions is a challenging task. The main objective of this paper is to develop a method for predicting GNSS information during GNSS outages based on a long short-term memory (LSTM) neural network to assist in factor graph-based combined GNSS/INS localization, which can provide a reliable combined localization solution during GNSS signal outages. In an environment with good GNSS signals, a factor graph fusion algorithm is used for data fusion of the combined positioning system, and an LSTM neural network prediction model is trained, and model parameters are determined using the INS velocity, inertial measurement unit (IMU) output, and GNSS position incremental data. In an environment with interrupted GNSS signals, the LSTM model is used to predict the GNSS positional increments and generate the pseudo-GNSS information and the solved results of INS for combined localization. In order to verify the performance and effectiveness of the proposed method, we conducted real-world road test experiments on land vehicles installed with GNSS receivers and inertial sensors. The experimental results show that, compared with the traditional combined GNSS/INS factor graph localization method, the proposed method can provide more accurate and robust localization results even in environments with frequent GNSS signal loss. Full article
(This article belongs to the Section Navigation and Positioning)
Show Figures

Figure 1

Figure 1
<p>Pre-integration schematic diagram.</p>
Full article ">Figure 2
<p>GNSS/INS Combined Positioning Factor Graph Model.</p>
Full article ">Figure 3
<p>LSTM-Assisted Model Structure Diagram.</p>
Full article ">Figure 4
<p>Basic LSTM Structure.</p>
Full article ">Figure 5
<p>Vehicle with navigation system equipment.</p>
Full article ">Figure 6
<p>Road Test Trajectory.</p>
Full article ">Figure 7
<p>Section 1 Trajectory Comparison.</p>
Full article ">Figure 8
<p>Section 1 East Error.</p>
Full article ">Figure 9
<p>Section 1 North Error.</p>
Full article ">Figure 10
<p>Section 2 Trajectory Comparison.</p>
Full article ">Figure 11
<p>Section 2 East Error.</p>
Full article ">Figure 12
<p>Section 2 North Error.</p>
Full article ">Figure 13
<p>Section 3 Trajectory Comparison.</p>
Full article ">Figure 14
<p>Section 3 Eastward Error.</p>
Full article ">Figure 15
<p>Section 3 Northward Error.</p>
Full article ">
21 pages, 4390 KiB  
Article
Mapping Shrub Biomass at 10 m Resolution by Integrating Field Measurements, Unmanned Aerial Vehicles, and Multi-Source Satellite Observations
by Wenchao Liu, Jie Wang, Yang Hu, Taiyong Ma, Munkhdulam Otgonbayar, Chunbo Li, You Li and Jilin Yang
Remote Sens. 2024, 16(16), 3095; https://doi.org/10.3390/rs16163095 - 22 Aug 2024
Viewed by 659
Abstract
Accurately estimating shrub biomass in arid and semi-arid regions is critical for understanding ecosystem productivity and carbon stocks at both local and global scales. Due to the short and sparse features of shrubs, capturing the shrub biomass accurately by satellite observations is challenging. [...] Read more.
Accurately estimating shrub biomass in arid and semi-arid regions is critical for understanding ecosystem productivity and carbon stocks at both local and global scales. Due to the short and sparse features of shrubs, capturing the shrub biomass accurately by satellite observations is challenging. Previous studies mostly used ground samples and satellite observations to estimate shrub biomass by establishing a direct connection, which was often hindered by the limited number of ground samples and spatial scale mismatch between samples and observations. Unmanned aerial vehicles (UAVs) provide opportunities to obtain more samples that are in line with the aspects of satellite observations (i.e., scale) for regional-scale shrub biomass estimations accurately with low costs. However, few studies have been conducted based on the air-space-ground-scale connection assisted by UAVs. Here we developed a framework for estimating 10 m shrub biomass at a regional scale by integrating ground measurements, UAV, Landsat, and Sentinel-1/2 observations. First, the spatial distribution map of shrublands and non-shrublands was generated in 2023 in the Helan Mountains of Ningxia province, China. This map had an F1 score of 0.92. Subsequently, the UAV-based shrub biomass map was estimated using an empirical model between the biomass and the crown area of shrubs, which was aggregated at a 10 m × 10 m grid to match the spatial resolution of Sentinel-1/2 images. Then, a regional-scale estimation model of shrub biomass was developed with a random forest regression (RFR) approach driven by ground biomass measurements, UAV-based biomass, and the optimal satellite metrics. Finally, the developed model was used to produce the biomass map of shrublands over the study area in 2023. The uncertainty of the resultant biomass map was characterized by the pixel-level standard deviation (SD) using the leave-one-out cross-validation (LOOCV) method. The results suggested that the integration of multi-scale observations from the ground, UAVs, and satellites provided a promising approach to obtaining the regional shrub biomass accurately. Our developed model, which integrates satellite spectral bands and vegetation indices (R2 = 0.62), outperformed models driven solely by spectral bands (R2 = 0.33) or vegetation indices (R2 = 0.55). In addition, our estimated biomass has an average uncertainty of less than 4%, with the lowest values (<2%) occurring in regions with high shrub coverage (>30%) and biomass production (>300 g/m2). This study provides a methodology to accurately monitor the shrub biomass from satellite images assisted by near-ground UAV observations as well as ground measurements. Full article
(This article belongs to the Special Issue Crops and Vegetation Monitoring with Remote/Proximal Sensing II)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>,<b>b</b>) Locations of the Helan Mountain in China and Ningxia province, and (<b>c</b>) the distribution of ground truth samples from field measurements, UAV, and visual interpretation.</p>
Full article ">Figure 2
<p>The workflow for estimating biomass of shrubland.</p>
Full article ">Figure 3
<p>(<b>a</b>) The original unmanned aerial vehicle (UAV) image. (<b>b</b>) The classified map of shrublands. (<b>c</b>) The fishnet constructed based on the UAV imagery. (<b>d</b>–<b>g</b>) The zoomed-in views of four sample points in (<b>b</b>).</p>
Full article ">Figure 4
<p>(<b>a</b>) The shrublands and other land over types of Helan Mountain, China, in 2023. (<b>b</b>–<b>i</b>) The zoom-in views of four example regions in the resultant map and the Google Earth images.</p>
Full article ">Figure 5
<p>The comparison of accuracy among the three models. The x-axis represents three models driven by the basic bands (SB), the vegetation indices (VI), and the combination of the basic bands and vegetation indices (SBVI). Their performance is evaluated using R<sup>2</sup> and RMSE.</p>
Full article ">Figure 6
<p>(<b>a</b>) The distribution of R<sup>2</sup> and EOPC within different ranges of shrub coverage. (<b>b</b>) The distribution of R<sup>2</sup> and EOUB within different ranges of shrub biomass. (<b>c</b>) The sensitivity of the biomass model to each variable examined by R<sup>2</sup> and RMSE. These analyses were conducted based on the ground samples. EOPC denotes the error of one percent coverage of shrub, calculated by RMSE/mean shrub coverage. EOUB denotes the error of one unit biomass, calculated by RMSE/mean shrub biomass.</p>
Full article ">Figure 7
<p>(<b>a</b>) The estimated distribution map of shrub biomass in the Helan Mountains. (<b>b</b>) The corresponding map of standard deviation (SD). (<b>c</b>) The distribution of EOPC within different ranges of shrub coverage. (<b>d</b>) The distribution of EOUB within different ranges of shrub biomass. These analyses were conducted based on the estimated distribution map of shrub biomass and the corresponding map of standard deviation. EOPC denotes the error of one percent coverage of shrub, calculated by SD/mean shrub coverage. EOUB denotes the error of one unit biomass, calculated by SD/mean shrub biomass.</p>
Full article ">Figure 8
<p>(<b>a</b>) The distribution of shrub biomass under precipitation gradients. (<b>b</b>) The distribution of shrub biomass under temperature gradients. (<b>c</b>) The distribution of shrub biomass within different ranges of aridity index. (<b>d</b>) The distribution of shrub biomass along elevation gradients.</p>
Full article ">
16 pages, 11179 KiB  
Article
Detection Method of Stator Coating Quality of Flat Wire Motor Based on Improved YOLOv8s
by Hongping Wang, Gong Chen, Xin Rong, Yiwen Zhang, Linsen Song and Xiao Shang
Sensors 2024, 24(16), 5392; https://doi.org/10.3390/s24165392 - 21 Aug 2024
Viewed by 351
Abstract
The stator of a flat wire motor is the core component of new energy vehicles. However, detecting quality defects in the coating process in real-time is a challenge. Moreover, the number of defects is large, and the pixels of a single defect are [...] Read more.
The stator of a flat wire motor is the core component of new energy vehicles. However, detecting quality defects in the coating process in real-time is a challenge. Moreover, the number of defects is large, and the pixels of a single defect are very few, which make it difficult to distinguish the defect features and make accurate detection more difficult. To solve this problem, this article proposes the YOLOv8s-DFJA network. The network is based on YOLOv8s, which uses DSFI-HEAD to replace the original detection head, realizing task alignment. It enhances joint features between the classification task and localization task and improves the ability of network detection. The LEFG module replaces the C2f module in the backbone of the YOLOv8s network that reduces the redundant parameters brought by the traditional BottleNeck structure. It also enhances the feature extraction and gradient flow ability to achieve the lightweight of the network. For this research, we produced our own dataset of stator coating quality regarding flat wire motors. Data augmentation technology (Gaussian noise, adjusting brightness, etc.) enriches the dataset, to a certain extent, which improves the robustness and generalization ability of YOLOv8s-DFJA. The experimental results show that in the performance of YOLOv8s-DFJA compared with YOLOv8s, the [email protected] index increased by 6.4%, the precision index increased by 1.1%, the recall index increased by 8.1%, the FPS index increased by 9.8FPS/s, and the parameters decreased by 3 Mb. Therefore, YOLOv8s-DFJA can be better realize the fast and accurate detection of the stator coating quality of flat wire motors. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>Image samples of stator coating area defects. The red ring is a bare defect, the blue ring is an adhesion defect, and the green ring is an impurity defect.</p>
Full article ">Figure 2
<p>YOLOv8s-DFJA network structure diagram.</p>
Full article ">Figure 3
<p>DSFI-HEAD module structure diagram. GNConv has the function of GroupNorm [<a href="#B31-sensors-24-05392" class="html-bibr">31</a>] module, which can improve the performance of detection head positioning and classification.</p>
Full article ">Figure 4
<p>LEFG module structure diagram.</p>
Full article ">Figure 5
<p>LEFG module dimension variation diagram.</p>
Full article ">Figure 6
<p>Presentation of datasets: (<b>a</b>) original stator image; (<b>b</b>) pretzel noise; (<b>c</b>) Gaussian blurring.</p>
Full article ">Figure 7
<p>Distribution of dataset information.</p>
Full article ">Figure 8
<p>LEFG module dimension variation diagram. PR curves and mAP for each defect category: (<b>a</b>) YOLOv8s; (<b>b</b>) YOLOv8s-DSFI-HEAD; (<b>c</b>) YOLOv8s-LFEG; (<b>d</b>) YOLOv8s-DFJA.</p>
Full article ">Figure 9
<p>Attention heatmap of the YOLOv8 model and YOLOv8-DFJA model.</p>
Full article ">Figure 10
<p>Detection effect diagram of each network. Missing detection targets are marked with thick red frame; false detection targets are marked with thick green frame; im, impurity.</p>
Full article ">Figure 11
<p>Detection effect of local enlargement.</p>
Full article ">
22 pages, 7164 KiB  
Article
LettuceNet: A Novel Deep Learning Approach for Efficient Lettuce Localization and Counting
by Aowei Ruan, Mengyuan Xu, Songtao Ban, Shiwei Wei, Minglu Tian, Haoxuan Yang, Annan Hu, Dong Hu and Linyi Li
Agriculture 2024, 14(8), 1412; https://doi.org/10.3390/agriculture14081412 - 20 Aug 2024
Viewed by 381
Abstract
Traditional lettuce counting relies heavily on manual labor, which is laborious and time-consuming. In this study, a simple and efficient method for localization and counting lettuce is proposed, based only on lettuce field images acquired by an unmanned aerial vehicle (UAV) equipped with [...] Read more.
Traditional lettuce counting relies heavily on manual labor, which is laborious and time-consuming. In this study, a simple and efficient method for localization and counting lettuce is proposed, based only on lettuce field images acquired by an unmanned aerial vehicle (UAV) equipped with an RGB camera. In this method, a new lettuce counting model based on the weak supervised deep learning (DL) approach is developed, called LettuceNet. The LettuceNet network adopts a more lightweight design that relies only on point-level labeled images to train and accurately predict the number and location information of high-density lettuce (i.e., clusters of lettuce with small planting spacing, high leaf overlap, and unclear boundaries between adjacent plants). The proposed LettuceNet is thoroughly assessed in terms of localization and counting accuracy, model efficiency, and generalizability using the Shanghai Academy of Agricultural Sciences-Lettuce (SAAS-L) and the Global Wheat Head Detection (GWHD) datasets. The results demonstrate that LettuceNet achieves superior counting accuracy, localization, and efficiency when employing the enhanced MobileNetV2 as the backbone network. Specifically, the counting accuracy metrics, including mean absolute error (MAE), root mean square error (RMSE), normalized root mean square error (nRMSE), and coefficient of determination (R2), reach 2.4486, 4.0247, 0.0276, and 0.9933, respectively, and the F-Score for localization accuracy is an impressive 0.9791. Moreover, the LettuceNet is compared with other existing widely used plant counting methods including Multi-Column Convolutional Neural Network (MCNN), Dilated Convolutional Neural Networks (CSRNets), Scale Aggregation Network (SANet), TasselNet Version 2 (TasselNetV2), and Focal Inverse Distance Transform Maps (FIDTM). The results indicate that our proposed LettuceNet performs the best among all evaluated merits, with 13.27% higher R2 and 72.83% lower nRMSE compared to the second most accurate SANet in terms of counting accuracy. In summary, the proposed LettuceNet has demonstrated great performance in the tasks of localization and counting of high-density lettuce, showing great potential for field application. Full article
Show Figures

Figure 1

Figure 1
<p>The location of the study area and distribution of the lettuce field.</p>
Full article ">Figure 2
<p>Examples of lettuce labeling for (<b>a</b>) a single variety of healthy lettuce individuals; and (<b>b</b>) tightly packed groups of lettuce (the green dots are point-level labels).</p>
Full article ">Figure 3
<p>The architecture of the LettuceNet. (The term layer m (1/n) in the figure indicates that after convolutional layer m, the size of the feature map is reduced by a factor of 1/n relative to the input image size. If 1/n is not specified, the feature map size remains the same as the previous layer. For example, the feature map size for Layer 6 is 1/16 of the input image size, which is identical to the feature map size for Layer 5 (i.e., 1/16). The feature map size for Layer 1 is the same as the input image size (i.e., 1/1), and so on; The number of feature maps refers to the number of output feature maps after convolutional processing. For the original input, the model has 3 feature maps, corresponding to the R, G, and B color channels; The term Rate refers to the dilation rate of the Atrous convolution; The notation k × k Conv indicates that the convolution kernel has dimensions of k × k; The Upsample by i refers to increasing the feature map size by a factor of i).</p>
Full article ">Figure 4
<p>Structure and internal details of the improved MobileNetV2. (The expansion_factor and bottleneck_num will be used to parameterize the convolutional layer respectively; Layer 1 performs only one deep convolution; from layer 2 to layer 8, each layer has the same internal structure of inverted residuals, but with different expansion and bottleneck coefficients; Layers 3–5 output feature maps to the DM, and layer 8 outputs feature maps to the MFFM (see the architecture of LettuceNet in <a href="#agriculture-14-01412-f003" class="html-fig">Figure 3</a>)).</p>
Full article ">Figure 5
<p>Comparison of LettuceNet operation efficiency using ResNet50, VGG16, MobileNetV2 and improved MobileNetV2 as the backbone network, respectively.</p>
Full article ">Figure 6
<p>The localization effects of the LettuceNet model using improved MobileNetV2 as a backbone network for lettuce counts on the five test images from the SAAS-L dataset for (<b>a</b>) clear borders, clear texture features, and tight arrangement; (<b>b</b>,<b>c</b>) unclear borders, fuzzy texture features, and tight arrangement; (<b>d</b>,<b>e</b>) relatively clear border and texture features, and compact irregular arrangement. (The red area in the second column represents the probability that the pixel point is a lettuce class, with darker colors representing a higher probability of being lettuce and vice versa. The blue area in the third column is the blobs consisting of neighboring pixel points with a probability greater than 0.5).</p>
Full article ">Figure 7
<p>Comparison of the overall performance for LettuceNet localization using different backbone networks for lettuce images with (<b>a</b>) clear borders, clear texture features, and tight arrangement; (<b>b</b>,<b>c</b>) unclear borders, fuzzy texture features, and tight arrangement; (<b>d</b>,<b>e</b>) relatively clear border and texture features, and overly tight arrangement. Red boxes indicate that the lettuce is undetected; green boxes indicate that two or more lettuces are detected as one.</p>
Full article ">Figure 8
<p>Comparison of local visualizations of LettuceNet localization using different backbone networks in randomly selected small area lettuce images with (<b>a</b>) clear borders, clear texture features, and tight arrangement; (<b>b</b>,<b>c</b>) unclear borders, fuzzy texture features, and tight arrangement; (<b>d</b>,<b>e</b>) relatively clear border and texture features, and overly tight arrangement. (Red boxes indicate that the lettuce is undetected; green boxes indicate that two or more lettuces are detected as one).</p>
Full article ">Figure 9
<p>Comparison of the proposed LettuceNet with MCNN, CSRNets, SANet, TasselNetV2, and FIDTM on the operational efficiency of lettuce counting tasks.</p>
Full article ">Figure 10
<p>LettuceNet visualization of counting results from the GWHD dataset of wheat heads with (<b>a</b>–<b>c</b>) obvious features and large differences from the background; (<b>d</b>,<b>e</b>) similar features and mixed with the background under strong light. (The first column shows the original RGB test images, the second column shows the heat map, and the third column shows the localization and counting map; The red area in the second column represents the probability that the pixel point is a wheat head class, with darker colors representing a higher probability of being wheat head and vice versa. The blue area in the third column is the blobs consisting of neighboring pixel points with a probability greater than 0.5.).</p>
Full article ">Figure 11
<p>Coefficients of determination of the LettuceNet on 24 original images (resolution 5472 × 4648). (The orange dots represent 24 counting experiments, and the green lines are 1:1 lines across the origin. The closer the orange dot is to the green line, the closer the predicted count value is to the actual manual count value).</p>
Full article ">Figure 12
<p>Localization results with LettuceNet model for a stitched lettuce image most affected by boundary effects (The blue areas represent individual lettuces and are blobs of adjacent pixels, each of which has a probability greater than 0.5 of belonging to the lettuces category; The red boxes indicate lettuces that were not detected due to boundary effects; The green boxes indicate lettuces that were repeatedly detected due to boundary effects). (<b>A</b>–<b>C</b>) areas where lettuce was not detected due to boundary effects; (<b>D</b>) areas where lettuce is repeatedly detected due to boundary effects.</p>
Full article ">
14 pages, 2331 KiB  
Article
Enhancing Weather Scene Identification Using Vision Transformer
by Christine Dewi, Muhammad Asad Arshed, Henoch Juli Christanto, Hafiz Abdul Rehman, Amgad Muneer and Shahzad Mumtaz
World Electr. Veh. J. 2024, 15(8), 373; https://doi.org/10.3390/wevj15080373 - 16 Aug 2024
Viewed by 636
Abstract
The accuracy of weather scene recognition is critical in a world where weather affects every aspect of our everyday lives, particularly in areas like intelligent transportation networks, autonomous vehicles, and outdoor vision systems. The importance of weather in many aspects of our life [...] Read more.
The accuracy of weather scene recognition is critical in a world where weather affects every aspect of our everyday lives, particularly in areas like intelligent transportation networks, autonomous vehicles, and outdoor vision systems. The importance of weather in many aspects of our life highlights the vital necessity for accurate information. Precise weather detection is especially crucial for industries like intelligent transportation, outside vision systems, and driverless cars. The outdated, unreliable, and time-consuming manual identification techniques are no longer adequate. Unmatched accuracy is required for local weather scene forecasting in real time. This work utilizes the capabilities of computer vision to address these important issues. Specifically, we employ the advanced Vision Transformer model to distinguish between 11 different weather scenarios. The development of this model results in a remarkable performance, achieving an accuracy rate of 93.54%, surpassing industry standards such as MobileNetV2 and VGG19. These findings advance computer vision techniques into new domains and pave the way for reliable weather scene recognition systems, promising extensive real-world applications across various industries. Full article
Show Figures

Figure 1

Figure 1
<p>Weather classification dataset: (<b>a</b>) dew, (<b>b</b>) fog–smog, (<b>c</b>) frost, (<b>d</b>) glaze, (<b>e</b>) hail, (<b>f</b>) lightning, (<b>g</b>) rain, (<b>h</b>) rainbow, (<b>i</b>) rime, (<b>j</b>) sandstorm (<b>k</b>) snow.</p>
Full article ">Figure 1 Cont.
<p>Weather classification dataset: (<b>a</b>) dew, (<b>b</b>) fog–smog, (<b>c</b>) frost, (<b>d</b>) glaze, (<b>e</b>) hail, (<b>f</b>) lightning, (<b>g</b>) rain, (<b>h</b>) rainbow, (<b>i</b>) rime, (<b>j</b>) sandstorm (<b>k</b>) snow.</p>
Full article ">Figure 2
<p>ViT abstract level architecture diagram [<a href="#B25-wevj-15-00373" class="html-bibr">25</a>].</p>
Full article ">Figure 3
<p>Confusion matrix of ViT model.</p>
Full article ">Figure 4
<p>Comparison of ViT, VGG-16 and MobileNetV2 models.</p>
Full article ">
20 pages, 31178 KiB  
Article
TC–Radar: Transformer–CNN Hybrid Network for Millimeter-Wave Radar Object Detection
by Fengde Jia, Chenyang Li, Siyi Bi, Junhui Qian, Leizhe Wei and Guohao Sun
Remote Sens. 2024, 16(16), 2881; https://doi.org/10.3390/rs16162881 - 7 Aug 2024
Viewed by 1077
Abstract
In smart transportation, assisted driving relies on data integration from various sensors, notably LiDAR and cameras. However, their optical performance can degrade under adverse weather conditions, potentially compromising vehicle safety. Millimeter-wave radar, which can overcome these issues more economically, has been re-evaluated. Despite [...] Read more.
In smart transportation, assisted driving relies on data integration from various sensors, notably LiDAR and cameras. However, their optical performance can degrade under adverse weather conditions, potentially compromising vehicle safety. Millimeter-wave radar, which can overcome these issues more economically, has been re-evaluated. Despite this, developing an accurate detection model is challenging due to significant noise interference and limited semantic information. To address these practical challenges, this paper presents the TC–Radar model, a novel approach that synergistically integrates the strengths of transformer and the convolutional neural network (CNN) to optimize the sensing potential of millimeter-wave radar in smart transportation systems. The rationale for this integration lies in the complementary nature of CNNs, which are adept at capturing local spatial features, and transformers, which excel at modeling long-range dependencies and global context within data. This hybrid approach allows for a more robust and accurate representation of radar signals, leading to enhanced detection performance. A key innovation of our approach is the introduction of the Cross-Attention (CA) module, which facilitates efficient and dynamic information exchange between the encoder and decoder stages of the network. This CA mechanism ensures that critical features are accurately captured and transferred, thereby significantly improving the overall network performance. In addition, the model contains the dense information fusion block (DIFB) to further enrich the feature representation by integrating different high-frequency local features. This integration process ensures thorough incorporation of key data points. Extensive tests conducted on the CRUW and CARRADA datasets validate the strengths of this method, with the model achieving an average precision (AP) of 83.99% and a mean intersection over union (mIoU) of 45.2%, demonstrating robust radar sensing capabilities. Full article
(This article belongs to the Special Issue Technical Developments in Radar—Processing and Application)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Overview of the radar signal processing chain. PC data need to go through multiple preprocessing processes, and the information is relatively sparse.</p>
Full article ">Figure 2
<p>Overall structure of the TC–Radar model. The top is the encode branch, and the bottom is the decode branch. The DIFB module and the Cross-Attention module act in these two parts, respectively, making full use of the multi-scale information.</p>
Full article ">Figure 3
<p>Overview of the DIFB framework, extracting local high-frequency information through parallel atrous convolutions with different expansion rates.</p>
Full article ">Figure 4
<p>Cross-attention module framework. (<b>a</b>) describes the data operation process, (<b>b</b>) represents the fusion method, and (<b>c</b>) shows the flow of cross information.</p>
Full article ">Figure 5
<p>Radar echo visual display, target information is often interfered by clutter.</p>
Full article ">Figure 6
<p>Comparison of model complexity. The results from the CRUW dataset (sections (<b>a</b>,<b>b</b>)) and the CARRADA dataset (sections (<b>c</b>,<b>d</b>)) demonstrate that TC–Radar achieves an optimal balance between detection accuracy and model complexity.</p>
Full article ">Figure 7
<p>Visual detection effects of different algorithms on the CRUW dataset. Green, red, and blue represent bicycle, pedestrian, and car targets, respectively. The darker the color, the higher the confidence level of the target.</p>
Full article ">Figure 8
<p>Comparing the visual segmentation effect of the baseline model on the CARRADA dataset, green, red, and blue represent bicycle, pedestrian, and car targets, respectively. It should be noted that TC–Radar only learns RA perspective data, while the baseline model requires joint learning from multiple perspectives.</p>
Full article ">Figure 9
<p>Visualization results of the contribution of each module. The first four rows and the last four rows of results are based on CARRADA and CRUW data, respectively.</p>
Full article ">Figure 10
<p>To validate the potential of radar as a low-cost alternative to optical sensors in low-light conditions, we conduct a comprehensive evaluation, termed the “super test”, using a nighttime dataset.</p>
Full article ">
26 pages, 14305 KiB  
Article
Evaluation of Traffic Sign Occlusion Rate Based on a 3D Point Cloud Space
by Ziqi Wang, Xiaofei Wang, Jun Li and Jiangbei Yao
Remote Sens. 2024, 16(16), 2872; https://doi.org/10.3390/rs16162872 - 6 Aug 2024
Viewed by 780
Abstract
The effectiveness of road signs is hindered by obstructions, such as vegetation, mutual obstruction of signs, or the road alignment itself. The traditional evaluation of road sign recognition effectiveness is conducted through in-vehicle field surveys. However, this method has several drawbacks, including discontinuous [...] Read more.
The effectiveness of road signs is hindered by obstructions, such as vegetation, mutual obstruction of signs, or the road alignment itself. The traditional evaluation of road sign recognition effectiveness is conducted through in-vehicle field surveys. However, this method has several drawbacks, including discontinuous identification, unclear positioning, incomplete coverage, and being time-consuming. Consequently, it is unable to effectively assess the recognition status of road signs at any arbitrary point within the road space. Therefore, this study employed laser scanning to construct a point-surface model, which was based on a point cloud algorithm and SLAM (Simultaneous Localization and Mapping), integrated LiDAR and inertial navigation system data, and optimized the point model after processing steps such as denoising, resampling, and three-dimensional model construction. Furthermore, a method for assessing the highway sign occlusion rate based on the picking algorithm was proposed. The algorithm was applied to an actual road environment, and the occlusion by other items was simulated. The results demonstrated the effectiveness of the method. This new method provides support for the fast and accurate calculation of road sign occlusion rates, which is of great importance for ensuring the safe traveling of vehicles. Full article
Show Figures

Figure 1

Figure 1
<p>Technical scheme.</p>
Full article ">Figure 2
<p>Installation of LiDAR and camera.</p>
Full article ">Figure 3
<p>Intensity information and ring information of LiDAR.</p>
Full article ">Figure 4
<p>Point cloud denoising effect.</p>
Full article ">Figure 5
<p>Real photo and point cloud shape of signs.</p>
Full article ">Figure 6
<p>Sign and plane point cloud recognized by the RANSAC algorithm.</p>
Full article ">Figure 7
<p>Flowchart of the RANSAC algorithm.</p>
Full article ">Figure 8
<p>Direction of surface normal vectors.</p>
Full article ">Figure 9
<p>Mesh reconstruction process.</p>
Full article ">Figure 10
<p>Scanning consistency display.</p>
Full article ">Figure 11
<p>Road conditions as shown in Google maps.</p>
Full article ">Figure 12
<p>Rosbag information.</p>
Full article ">Figure 13
<p>Point cloud map after the slope calculation.</p>
Full article ">Figure 14
<p>Point cloud model before and after slope segmentation.</p>
Full article ">Figure 15
<p>Point cloud model before and after the intensity segmentation of the non-ground point cloud.</p>
Full article ">Figure 16
<p>Three planar point clouds separated by RANSAC.</p>
Full article ">Figure 17
<p>Test points corresponding to the three signs.</p>
Full article ">Figure 17 Cont.
<p>Test points corresponding to the three signs.</p>
Full article ">Figure 18
<p>Additional test points corresponding to sign 1.</p>
Full article ">Figure 19
<p>Heat map of the occlusion rate for sign 1.</p>
Full article ">Figure 20
<p>Location of sign 1 and the tree point cloud.</p>
Full article ">Figure 21
<p>Field situation of sign 1.</p>
Full article ">Figure 22
<p>Field situations of sign 2 and sign 3.</p>
Full article ">Figure 23
<p>Sign 2, virtual sign 3, and added test points.</p>
Full article ">Figure 24
<p>Heat map of the occlusion rate for sign 2.</p>
Full article ">
21 pages, 12275 KiB  
Article
Segmentation Point Simultaneous Localization and Mapping: A Stereo Vision Simultaneous Localization and Mapping Method for Unmanned Surface Vehicles in Nearshore Environments
by Xiujing Gao, Xinzhi Lin, Fanchao Lin and Hongwu Huang
Electronics 2024, 13(16), 3106; https://doi.org/10.3390/electronics13163106 - 6 Aug 2024
Viewed by 744
Abstract
Unmanned surface vehicles (USVs) in nearshore areas are prone to environmental occlusion and electromagnetic interference, which can lead to the failure of traditional satellite-positioning methods. This paper utilizes a visual simultaneous localization and mapping (vSLAM) method to achieve USV positioning in nearshore environments. [...] Read more.
Unmanned surface vehicles (USVs) in nearshore areas are prone to environmental occlusion and electromagnetic interference, which can lead to the failure of traditional satellite-positioning methods. This paper utilizes a visual simultaneous localization and mapping (vSLAM) method to achieve USV positioning in nearshore environments. To address the issues of uneven feature distribution, erroneous depth information, and frequent viewpoint jitter in the visual data of USVs operating in nearshore environments, we propose a stereo vision SLAM system tailored for nearshore conditions: SP-SLAM (Segmentation Point-SLAM). This method is based on ORB-SLAM2 and incorporates a distance segmentation module, which filters feature points from different regions and adaptively adjusts the impact of outliers on iterative optimization, reducing the influence of erroneous depth information on motion scale estimation in open environments. Additionally, our method uses the Sum of Absolute Differences (SAD) for matching image blocks and quadric interpolation to obtain more accurate depth information, constructing a complete map. The experimental results on the USVInland dataset show that SP-SLAM solves the scaling constraint failure problem in nearshore environments and significantly improves the robustness of the stereo SLAM system in such environments. Full article
(This article belongs to the Section Electrical and Autonomous Vehicles)
Show Figures

Figure 1

Figure 1
<p>Nearshore environments image, (<b>a</b>) is the river channel, (<b>b</b>) is the coastal area, and (<b>c</b>) is the lake.</p>
Full article ">Figure 2
<p>Structure of the SP-SLAM system.</p>
Full article ">Figure 3
<p>Rotation invariance and scale invariance of ORB features.</p>
Full article ">Figure 4
<p>Morphological dilatation, (<b>a</b>) is the original image, (<b>b</b>) is the structural element, and (<b>c</b>) is the expanded image; the gray areas represent the pixel distribution before dilation, while the black areas denote the pixels filled during the dilation process.</p>
Full article ">Figure 5
<p>Distance segmentation, (<b>a</b>) is the original image, (<b>b</b>) is the image after distance segmentation.</p>
Full article ">Figure 6
<p>Quadric interpolation diagram.</p>
Full article ">Figure 7
<p>Three-dimensional point updating.</p>
Full article ">Figure 8
<p>USVInland Acquisition Platform: (<b>a</b>) is the physical picture of the unmanned ship, (<b>b</b>) is the front view of the unmanned ship, (<b>c</b>) is the top view of the unmanned ship.</p>
Full article ">Figure 9
<p>The distance segmentation effect processed by different structural parts is as follows: the first act is the original image, the second act is the binary image segmented by the iterative threshold, and the third to fifth lines are segmented by different structural elements.</p>
Full article ">Figure 10
<p>Nearshore feature extraction effect, (<b>a</b>–<b>d</b>) illustrate the extraction results in different scenarios, where the green points represent foreground feature points and the red points represent background feature points.</p>
Full article ">Figure 11
<p>USVInland real environment presentation.</p>
Full article ">Figure 12
<p>Comparison of motion trajectories.</p>
Full article ">Figure 13
<p>KITTI datasets 00, 01 sequence trajectory comparison diagram.</p>
Full article ">Figure 14
<p>N03-5 mapping effect display, (<b>a</b>) shows the real scene image and mapping results of the riverbanks, and (<b>b</b>) displays the satellite top view of the river and the overall mapping outcomes. The red line in the figure indicates the approximate path of the unmanned surface vessel.</p>
Full article ">Figure 15
<p>N02-4 mapping effect display, (<b>a</b>) shows the real scene images of both ends of the river, both of which are bridge scenes; (<b>b</b>) displays the satellite top view of the river and the overall mapping results. The red line in the figure indicates the approximate path of the USVs.</p>
Full article ">
18 pages, 31105 KiB  
Article
Global Path Planning of Unmanned Surface Vehicle in Complex Sea Areas Based on Improved Streamline Method
by Haoran Liu, Qihe Shan, Yuchi Cao and Qi Xu
J. Mar. Sci. Eng. 2024, 12(8), 1324; https://doi.org/10.3390/jmse12081324 - 5 Aug 2024
Viewed by 584
Abstract
In this paper, an innovative method is proposed to improve the global path planning of Unmanned Surface Vehicles (USV) in complex sea areas, combining fluid mechanic calculations with an improved A* algorithm. This method not only generates smooth paths but also ensures feasible [...] Read more.
In this paper, an innovative method is proposed to improve the global path planning of Unmanned Surface Vehicles (USV) in complex sea areas, combining fluid mechanic calculations with an improved A* algorithm. This method not only generates smooth paths but also ensures feasible global solutions, significantly enhancing the efficiency and safety of path planning. Firstly, in response to the water depths limitation, this study set up safe water depths, providing strong guarantees for the safe navigation of USVs in complex waters. Secondly, based on the hydrological and geographical characteristics of the study sea area, an accurate ocean environment model was constructed using Ansys Fluent software and computational fluid dynamics (CFD) technology, thus providing USVs with a feasible path solution on a global scale. Then, the local sea area with complex obstacles was converted into a grid map to facilitate detailed planning. Meanwhile, the improved A* algorithm was utilized for meticulous route optimization. Furthermore, by combining the results of local and global planning, the approach generated a comprehensive route that accounts for the complexities of the maritime environment while avoiding local optima. Finally, simulation results demonstrated that the algorithm proposed in this study shows faster pathfinding speed, shorter route distances, and higher route safety compared to other algorithms. Moreover, it remains stable and effective in real-world scenarios. Full article
(This article belongs to the Special Issue Unmanned Marine Vehicles: Perception, Planning, Control and Swarm)
Show Figures

Figure 1

Figure 1
<p>Global path planning design process.</p>
Full article ">Figure 2
<p>Principle diagram of streamline method.</p>
Full article ">Figure 3
<p>Search pattern of A* algorithm. (<b>a</b>) Four-directional search pattern; (<b>b</b>) Eight-directional search pattern.</p>
Full article ">Figure 4
<p>Diagram of safe distance.</p>
Full article ">Figure 5
<p>Inflection point diagram. (<b>a</b>) Eliminate collinear nodes. (<b>b</b>) Eliminate redundant inflection points.</p>
Full article ">Figure 6
<p>GeoNetworking-based danger zone mapping for USVs.</p>
Full article ">Figure 7
<p>Target area.</p>
Full article ">Figure 8
<p>Diagram of research ship.</p>
Full article ">Figure 9
<p>Schematic diagram of minimum safe water depths.</p>
Full article ">Figure 10
<p>Safe water depths processing diagram.</p>
Full article ">Figure 11
<p>Grid environment model.</p>
Full article ">Figure 12
<p>Grid environment model.</p>
Full article ">Figure 13
<p>Flow field grid division diagrams.</p>
Full article ">Figure 14
<p>Simulation calculation model diagram.</p>
Full article ">Figure 15
<p>Simulation results diagram. (<b>a</b>) Velocity cloud diagram; (<b>b</b>) Velocity vector diagram; (<b>c</b>) Streamline diagram.</p>
Full article ">Figure 16
<p>Global optimal route.</p>
Full article ">Figure 17
<p>Local segmentation path planning.</p>
Full article ">Figure 18
<p>Three types of safety distance diagrams.</p>
Full article ">Figure 19
<p>Comparison of the algorithms. (<b>a</b>) the traditional A* algorithm; (<b>b</b>) the RRT algorithm; (<b>c</b>) the improved A* algorithm.</p>
Full article ">Figure 20
<p>Path planning in dangerous areas under GeoNetworking scenarios.</p>
Full article ">Figure 21
<p>Environmental map.</p>
Full article ">Figure 22
<p>The result of the simulation process of the USV.</p>
Full article ">
17 pages, 15832 KiB  
Article
Development of Indicator for Piled Pier Health Evaluation in Vietnam Using Impact Vibration Test Approach
by Thi Bach Duong Nguyen, Jungwon Huh, Thanh Thai Vu, Minh Long Tran and Van Ha Mac
Buildings 2024, 14(8), 2366; https://doi.org/10.3390/buildings14082366 - 1 Aug 2024
Viewed by 702
Abstract
Vietnam’s seaport system currently includes 298 ports with 588 wharves (a total length of approximately 92,275 m), which is vital in developing Vietnam’s marine economy. The piled pier, a type of wharf structure, is widely used and accounts for up to 90%, while [...] Read more.
Vietnam’s seaport system currently includes 298 ports with 588 wharves (a total length of approximately 92,275 m), which is vital in developing Vietnam’s marine economy. The piled pier, a type of wharf structure, is widely used and accounts for up to 90%, while the remaining 10% is made up of other types of wharf structures, such as gravity and sheet pile quay walls. Most wharves have been operating for over 10 years and some for even more than 50 years. Noticeably, wharves are highly vulnerable and degrade rapidly due to many factors, especially heavy load impacts and severe environmental conditions. Additionally, wharves have a higher risk of deterioration than other inland infrastructure, such as buildings and bridges. Consequently, determining a wharf’s health is an important task in maintaining normal working conditions, extending its lifecycle, and avoiding other severe damage that could lead to dangers to the safety of vehicles, facilities, and humans. Moreover, regulated quality inspections usually include only simple inspections, e.g., displacement, settlement, geometric height, and tilt; the visual inspection and determination of dimensions by simple length-measuring equipment; concrete strength testing by ultrasonic and rebound hammers; and the experimental identification of the chloride ion concentration, chloride diffusion coefficient, corrosion activity of rebar in concrete, and steel thickness. These testing methods often give local results depending on the number of test samples. Therefore, advanced diagnostic techniques for assessing the technical condition of piled piers need to be studied. The impact vibration test (IVT) is a powerful non-destructive evaluation method that indicates the overall health of structures, e.g., underground and foundation structures, according to official standards. Hence, the IVT is expected to help engineers detect the potential deterioration of overall structures. It is fundamental that, if a structure is degraded, its natural frequency will be affected. A structure’s health index and technical condition are determined based on this change. However, the IVT does not seem to be widely applied to piled piers, with no published standard; hence, controversial issues related to accuracy and reliability still remain. This motivates the present study to recommend an adjusted factor (equal to 1.16) for the health index (classified in official standards for other structures) through numerical and experimental approaches before officially applying the IVT method to piled piers. The current work focuses on the health index using the design natural frequency, which is more practical in common cases where previous historical data and the standard natural frequency are unavailable. This study also examines a huge number of influencing factors and situations through theoretical analysis, experience, and field experiments to propose an adjusted indicator. The results are achieved with several assumptions of damages, such as the degradation of materials and local damages to structural components. With the proposed adjusted indicator, the overall health of piled piers can be assessed quickly and accurately by IVT inspections in cases of incidents, accidents due to collisions, cargo falls during loading and unloading, or subsidence and erosion due to natural disasters, storms, and floods. Full article
Show Figures

Figure 1

Figure 1
<p>Arrangement of filed IVT experiments for several ports.</p>
Full article ">Figure 2
<p>Procedure for assessing the health status of piled piers in the IVT method.</p>
Full article ">Figure 3
<p>Providing horizontal impact forces by: (<b>a</b>) the weight; (<b>b</b>) a mobile crane; (<b>c</b>) a tugboat; (<b>d</b>) a vessel; and (<b>e</b>) a movable gantry crane.</p>
Full article ">Figure 4
<p>Filed investigation of the wharves using the IVT method.</p>
Full article ">Figure 5
<p>Examples of measurement data for Gemarlink wharf.</p>
Full article ">Figure 6
<p>The horizontal natural frequencies for: (<b>a</b>) Tan Vu wharf; (<b>b</b>) Lach Huyen Port wharf; (<b>c</b>) Hiep Phuoc wharf; and (<b>d</b>) Gemarlink wharf.</p>
Full article ">Figure 6 Cont.
<p>The horizontal natural frequencies for: (<b>a</b>) Tan Vu wharf; (<b>b</b>) Lach Huyen Port wharf; (<b>c</b>) Hiep Phuoc wharf; and (<b>d</b>) Gemarlink wharf.</p>
Full article ">Figure 7
<p>Numerical simulation of a wharf.</p>
Full article ">Figure 8
<p>An example of IVT models with assumptions: (<b>a</b>) without erosion and (<b>b</b>) with erosion.</p>
Full article ">
27 pages, 122098 KiB  
Article
Multiple Object Tracking in Drone Aerial Videos by a Holistic Transformer and Multiple Feature Trajectory Matching Pattern
by Yubin Yuan, Yiquan Wu, Langyue Zhao, Yaxuan Pang and Yuqi Liu
Drones 2024, 8(8), 349; https://doi.org/10.3390/drones8080349 - 28 Jul 2024
Viewed by 410
Abstract
Drone aerial videos have immense potential in surveillance, rescue, agriculture, and urban planning. However, accurately tracking multiple objects in drone aerial videos faces challenges like occlusion, scale variations, and rapid motion. Current joint detection and tracking methods often compromise accuracy. We propose a [...] Read more.
Drone aerial videos have immense potential in surveillance, rescue, agriculture, and urban planning. However, accurately tracking multiple objects in drone aerial videos faces challenges like occlusion, scale variations, and rapid motion. Current joint detection and tracking methods often compromise accuracy. We propose a drone multiple object tracking algorithm based on a holistic transformer and multiple feature trajectory matching pattern to overcome these challenges. The holistic transformer captures local and global interaction information, providing precise detection and appearance features for tracking. The tracker includes three components: preprocessing, trajectory prediction, and matching. Preprocessing categorizes detection boxes based on scores, with each category adopting specific matching rules. Trajectory prediction employs the visual Gaussian mixture probability hypothesis density method to integrate visual detection results to forecast object motion accurately. The multiple feature pattern introduces Gaussian, Appearance, and Optimal subpattern assignment distances for different detection box types (GAO trajectory matching pattern) in the data association process, enhancing tracking robustness. We perform comparative validations on the vision-meets-drone (VisDrone) and the unmanned aerial vehicle benchmarks; the object detection and tracking (UAVDT) datasets affirm the algorithm’s effectiveness: it obtained 38.8% and 61.7% MOTA, respectively. Its potential for seamless integration into practical engineering applications offers enhanced situational awareness and operational efficiency in drone-based missions. Full article
Show Figures

Figure 1

Figure 1
<p>GAO-Tracker framework.</p>
Full article ">Figure 2
<p>Holistic trans-detector.</p>
Full article ">Figure 3
<p>Holistic transformer.</p>
Full article ">Figure 4
<p>Holistic self-attention. We initially partition the feature map into 4 × 4 grids. While the central 4 × 4 grid serves as the query window, we extract tokens at three granularity levels of 1 × 1, 2 × 2, and 4 × 4, respectively, from surrounding regions to serve as its keys and values. This results in tokens with dimensions of 8 × 8, 6 × 6, and 5 × 5. Ultimately, these tokens from the three levels are concatenated to compute the keys and values for the 4 × 4 = 16 tokens (queries) within the window.</p>
Full article ">Figure 5
<p>GAO trajectory matching module.</p>
Full article ">Figure 6
<p>Appear-IOU trajectory matching.</p>
Full article ">Figure 7
<p>Gau-IOU trajectory matching.</p>
Full article ">Figure 8
<p>OSPA-IOU trajectory matching.</p>
Full article ">Figure 9
<p>Comparison of detection results on the VisDrone dataset.</p>
Full article ">Figure 10
<p>Comparison of detection results on the UAVDT dataset.</p>
Full article ">Figure 11
<p>Tracking results of GAO-Tracker on the VisDrone dataset.</p>
Full article ">Figure 12
<p>Tracking results of GAO-Tracker on the UAVDT dataset.</p>
Full article ">
18 pages, 15267 KiB  
Article
Estimation of Forest Stand Volume in Coniferous Plantation from Individual Tree Segmentation Aspect Using UAV-LiDAR
by Xinshao Zhou, Kaisen Ma, Hua Sun, Chaokui Li and Yonghong Wang
Remote Sens. 2024, 16(15), 2736; https://doi.org/10.3390/rs16152736 - 26 Jul 2024
Viewed by 501
Abstract
The main problems of forest parameter extraction and forest stand volume estimation using unmanned aerial vehicle light detection and ranging (UAV-LiDAR) technology are the lack of precision in individual tree segmentation and the inability to directly obtain the diameter at breast height (DBH) [...] Read more.
The main problems of forest parameter extraction and forest stand volume estimation using unmanned aerial vehicle light detection and ranging (UAV-LiDAR) technology are the lack of precision in individual tree segmentation and the inability to directly obtain the diameter at breast height (DBH) parameter. To address such limitations, the study proposed an improved individual tree segmentation method combined with a DBH prediction model to obtain the tree height (H) and DBH for calculating the volume of trees, thus realizing the accurate estimation of forest stand volume from individual tree segmentation aspect. The method involves the following key steps: (1) The local maximum method with variable window combined with the Gaussian mixture model were used to detect the treetop position using the canopy height model for removing pits. (2) The measured tree DBH and H parameters of the sample trees were used to construct an optimal DBH-H prediction model. (3) The duality standing tree volume model was used to calculate the forest stand volume at the individual tree scale. The results showed that: (1) Individual tree segmentation based on the improved Gaussian mixture model with optimal accuracy, detection rate r, accuracy rate p, and composite score F were 89.10%, 95.21%, and 0.921, respectively. The coefficient of determination R2 of the accuracy of the extracted tree height parameter was 0.88, and the root mean square error RMSE was 0.84 m. (2) The Weibull model had the optimal model fit for DBH-H with predicted DBH parameter accuracy, the R2 and RMSE were 0.84 and 2.28 cm, respectively. (3) Using the correctly detected trees from the individual tree segmentation results combined with the duality standing tree volume model estimated the forest stand volume with an accuracy AE of 90.86%. In conclusion, using UAV-LiDAR technology, based on the individual tree segmentation method and the DBH-H model, it is possible to realize the estimation of forest stand volume at the individual tree scale, which helps to improve the estimation accuracy. Full article
(This article belongs to the Section Forest Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>The methodological framework of this study.</p>
Full article ">Figure 2
<p>Location of the study area. (<b>a</b>) and (<b>b</b>) illustrate the locations of the study area in Guangxi, China, while (<b>c</b>) presents the forest phase diagram associated with this area.</p>
Full article ">Figure 3
<p>Distribution of tree height and diameter at breast height.</p>
Full article ">Figure 4
<p>Logical schematic of the improved Gaussian mixture model individual tree segmentation method. (<b>a</b>,<b>b</b>) illustrate the normalized point cloud and the results of individual tree segmentation. (<b>c</b>–<b>e</b>) represent the three key processes of the improved Gaussian mixture model algorithm.</p>
Full article ">Figure 5
<p>Matching results of individual tree detections with measured positions: (<b>a</b>) WS, (<b>b</b>) QP, (<b>c</b>) GMM, and (<b>d</b>) IGMM.</p>
Full article ">Figure 6
<p>Plots of the <span class="html-italic">H<sub>L</sub></span> extracted from the individual tree point cloud against the measured <span class="html-italic">H</span>. In the figure, a circle denotes the position of a tree within the coordinate system, derived from the extracted tree height obtained using UAV-LiDAR point cloud data and the measured tree height. The red dashed line illustrates the linear fitting relationship between the two height measurements, while the blue line serves as the reference line indicating equality.</p>
Full article ">Figure 7
<p>Plots of the estimated <span class="html-italic">DBH</span> against the measured <span class="html-italic">DBH</span>. In the figure, a circle denotes the position of a tree within the coordinate system, derived from the extracted tree DBH predicted using Weibull model and the measured tree DBH. The red dashed line illustrates the linear fitting relationship between the two DBH measurements, while the blue line serves as the reference line indicating equality.</p>
Full article ">
18 pages, 11915 KiB  
Article
Detection of Individual Corn Crop and Canopy Delineation from Unmanned Aerial Vehicle Imagery
by Freda Dorbu and Leila Hashemi-Beni
Remote Sens. 2024, 16(14), 2679; https://doi.org/10.3390/rs16142679 - 22 Jul 2024
Viewed by 548
Abstract
Precise monitoring of individual crop growth and health status is crucial for precision agriculture practices. However, traditional inspection methods are time-consuming, labor-intensive, prone to human error, and may not provide the comprehensive coverage required for the detailed analysis of crop variability across an [...] Read more.
Precise monitoring of individual crop growth and health status is crucial for precision agriculture practices. However, traditional inspection methods are time-consuming, labor-intensive, prone to human error, and may not provide the comprehensive coverage required for the detailed analysis of crop variability across an entire field. This research addresses the need for efficient and high-resolution crop monitoring by leveraging Unmanned Aerial Vehicle (UAV) imagery and advanced computational techniques. The primary goal was to develop a methodology for the precise identification, extraction, and monitoring of individual corn crops throughout their growth cycle. This involved integrating UAV-derived data with image processing, computational geometry, and machine learning techniques. Bi-weekly UAV imagery was captured at altitudes of 40 m and 70 m from 30 April to 11 August, covering the entire growth cycle of the corn crop from planting to harvest. A time-series Canopy Height Model (CHM) was generated by analyzing the differences between the Digital Terrain Model (DTM) and the Digital Surface Model (DSM) derived from the UAV data. To ensure the accuracy of the elevation data, the DSM was validated against Ground Control Points (GCPs), adhering to standard practices in remote sensing data verification. Local spatial analysis and image processing techniques were employed to determine the local maximum height of each crop. Subsequently, a Voronoi data model was developed to delineate individual crop canopies, successfully identifying 13,000 out of 13,050 corn crops in the study area. To enhance accuracy in canopy size delineation, vegetation indices were incorporated into the Voronoi model segmentation, refining the initial canopy area estimates by eliminating interference from soil and shadows. The proposed methodology enables the precise estimation and monitoring of crop canopy size, height, biomass reduction, lodging, and stunted growth over time by incorporating advanced image processing techniques and integrating metrics for quantitative assessment of fields. Additionally, machine learning models were employed to determine relationships between the canopy sizes, crop height, and normalized difference vegetation index, with Polynomial Regression recording an R-squared of 11% compared to other models. This work contributes to the scientific community by demonstrating the potential of integrating UAV technology, computational geometry, and machine learning for accurate and efficient crop monitoring at the individual plant level. Full article
(This article belongs to the Special Issue Aerial Remote Sensing System for Agriculture)
Show Figures

Figure 1

Figure 1
<p>Automated framework for crop mapping and canopy characterization.</p>
Full article ">Figure 2
<p>A Voronoi diagram delineating points acquired from [<a href="#B31-remotesensing-16-02679" class="html-bibr">31</a>].</p>
Full article ">Figure 3
<p>A map of the estimated NDVI, GCC and ExG indices.</p>
Full article ">Figure 4
<p>Orthophoto of study area.</p>
Full article ">Figure 5
<p>(<b>a</b>) Randomly selected areas of the study site. (<b>b</b>) Spectral profile of randomly selected areas of the study area.</p>
Full article ">Figure 6
<p>(<b>a</b>) Initially generated CHM. (<b>b</b>) Filtered CHM. (<b>c</b>) Local peak height. The unit of measurement for the CHM, filtered CHM, and generated local peak height maps is feet.</p>
Full article ">Figure 7
<p>Local maximal results showing the peak height of each crop.</p>
Full article ">Figure 8
<p>A Voronoi map showing the canopy sizes and the individual crop peak heights with noise.</p>
Full article ">Figure 9
<p>Refined crop canopy sizes.</p>
Full article ">Figure 10
<p>The refinement process of each crop canopy size from (<b>a</b>) ExG to (<b>b</b>) segmented ExG. (<b>c</b>) Reclassified segmented ExG. (<b>d</b>) Conversion of segmented ExG raster to polygon. (<b>e</b>) Soil feature selection. (<b>f</b>) Overlay of modified Voronoi on the RGB image.</p>
Full article ">Figure 11
<p>Influence of canopy refinement.</p>
Full article ">Figure 12
<p>Predictive ability of SVM with RBF kernel, Polynomial Regression with Degree 2 models, and Gradient Boosting Machine. The red line represents perfect predictions. The closer the points are to this line, the more accurate the model’s predictions.</p>
Full article ">
24 pages, 6948 KiB  
Article
Expediting the Convergence of Global Localization of UAVs through Forward-Facing Camera Observation
by Zhenyu Li, Xiangyuan Jiang, Sile Ma, Xiaojing Ma, Zhenyi Lv, Hongliang Ding, Haiyan Ji and Zheng Sun
Drones 2024, 8(7), 335; https://doi.org/10.3390/drones8070335 - 19 Jul 2024
Viewed by 752
Abstract
In scenarios where the global navigation satellite system is unavailable, unmanned aerial vehicles (UAVs) can employ visual algorithms to process aerial images. These images are integrated with satellite maps and digital elevation models (DEMs) to achieve global localization. To address the localization challenge [...] Read more.
In scenarios where the global navigation satellite system is unavailable, unmanned aerial vehicles (UAVs) can employ visual algorithms to process aerial images. These images are integrated with satellite maps and digital elevation models (DEMs) to achieve global localization. To address the localization challenge in unfamiliar areas devoid of prior data, an iterative computation-based localization framework is commonly used. This framework iteratively refines its calculations using multiple observations from a downward-facing camera to determine an accurate global location. To improve the rate of convergence for localization, we introduced an innovative observation model. We derived a terrain descriptor from the images captured by a forward-facing camera and integrated it as supplementary observation into a point-mass filter (PMF) framework to enhance the confidence of the observation likelihood distribution. Furthermore, within this framework, the methods for the truncation of the convolution kernel and that of the probability distribution were developed, thereby enhancing the computational efficiency and convergence rate, respectively. The performance of the algorithm was evaluated using real UAV flight sequences, a satellite map, and a DEM in an area measuring 7.7 km × 8 km. The results demonstrate that this method significantly accelerates the localization convergence during both takeoff and ascent phases as well as during cruise flight. Additionally, it increases localization accuracy and robustness in complex environments, such as areas with uneven terrain and ambiguous scenes. The method is applicable to the localization of UAVs in large-scale unknown scenarios, thereby enhancing the flight safety and mission execution capabilities of UAVs. Full article
(This article belongs to the Special Issue Drone-Based Information Fusion to Improve Autonomous Navigation)
Show Figures

Figure 1

Figure 1
<p>Block diagram of the proposal. The first part (<a href="#sec3dot2-drones-08-00335" class="html-sec">Section 3.2</a>) describes the method of constructing a 2.5D descriptor based on two-view SFM; the second part (<a href="#sec3dot3-drones-08-00335" class="html-sec">Section 3.3</a>) outlines the matching procedure and the approach for obtaining the likelihood distribution; and the third part (<a href="#sec3dot4-drones-08-00335" class="html-sec">Section 3.4</a>) introduces the technique of fusion observations and accelerated iteration calculation based on PMF.</p>
Full article ">Figure 2
<p>Schematic of two-view SFM. (<b>a</b>) Two-view SFM process during the cruise flight stage, where the UAV maintains a constant altitude; (<b>b</b>) two-view SFM process during the ascend stage, where the UAV maintains a constant horizontal location.</p>
Full article ">Figure 3
<p>(<b>a</b>) A 2.5D DEM map generated from DEM; (<b>b</b>) schematic of mapping, matching, and similarity computation. First, 3D points are mapped to the 2.5D descriptor, and the map patch corresponding to grid <math display="inline"><semantics> <mi>k</mi> </semantics></math> in the 2.5D DEM map is matched to the descriptor, then the similarity <math display="inline"><semantics> <mrow> <msub> <mrow> <mo stretchy="false">(</mo> <msub> <mi>S</mi> <mi>p</mi> </msub> <mo stretchy="false">)</mo> </mrow> <mi>k</mi> </msub> </mrow> </semantics></math> is calculated and the likelihood probability distribution <math display="inline"><semantics> <mrow> <mi>p</mi> <mo stretchy="false">(</mo> <msub> <mi>z</mi> <mi>t</mi> </msub> <mo stretchy="false">(</mo> <mi>k</mi> <mo stretchy="false">)</mo> <mo>|</mo> <msub> <mi>χ</mi> <mi>t</mi> </msub> <mo stretchy="false">(</mo> <mi>k</mi> <mo stretchy="false">)</mo> <mo stretchy="false">)</mo> </mrow> </semantics></math> is obtained.</p>
Full article ">Figure 4
<p>Modeling of horizontal location noise and elevation noise for 3D points. The orange dashed line indicates the boundary of the influence of angular noise on the position of a 3D point. (<b>a</b>) Modeling of horizontal location noise for 3D points: noise introduced by the yaw angle of the camera (<b>left</b>) and odometry (<b>right</b>). (<b>b</b>) Modeling of elevation noise for 3D points: noise introduced by the pitch angle of the camera (<b>left</b>) and barometric measurements (<b>right</b>). The gray area represents the distribution of the noise.</p>
Full article ">Figure 5
<p>Illustrations of truncation methods: (<b>a</b>) convolution kernel truncation; (<b>b</b>) sliding window-based probability truncation.</p>
Full article ">Figure 6
<p>(<b>a</b>) Data acquisition platform; (<b>b</b>) algorithm-processing platform.</p>
Full article ">Figure 7
<p>(<b>a</b>) Screenshot of waypoint planning on ground station system; (<b>b</b>) sample of aerial images captured by a downward-facing camera; (<b>c</b>) sample of aerial images captured by a forward-facing camera.</p>
Full article ">Figure 8
<p>Results of image feature matching and point-cloud reconstruction. (<b>a</b>) Image feature matching of two keyframes during takeoff and ascend stage; (<b>b</b>) image feature matching of two keyframes during cruise flight stage; (<b>c</b>) 3D reconstruction results based on feature matching in (<b>b</b>). This perspective represents a top–down view.</p>
Full article ">Figure 9
<p>Visualization of the location trajectory, with 0 denoting the start-point and 74 representing the end-point in each graph. The actual flight path and the results of each experiment are represented by green and red curves, respectively. Figures (<b>a</b>–<b>f</b>) depict the localization trajectories for SAMPLE-00 to SAMPLE-05.</p>
Full article ">Figure 10
<p>Statistical graphs of localization error.</p>
Full article ">Figure 11
<p>Statistical graphs of localization standard deviation.</p>
Full article ">
Back to TopTop