[go: up one dir, main page]

Next Article in Journal
A Method for the Pattern Recognition of Acoustic Emission Signals Using Blind Source Separation and a CNN for Online Corrosion Monitoring in Pipelines with Interference from Flow-Induced Noise
Previous Article in Journal
A Spatio-Temporal Capsule Neural Network with Self-Correlation Routing for EEG Decoding of Semantic Concepts of Imagination and Perception Tasks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Instance Segmentation Method for Complex Elements of Farm UAV Aerial Survey Images

by
Feixiang Lv
1,
Taihong Zhang
1,2,3,*,
Yunjie Zhao
1,2,3,*,
Zhixin Yao
1,2,3 and
Xinyu Cao
1
1
School of Computer and Information Engineering, Xinjiang Agricultural University, Urumqi 830052, China
2
Ministry of Education Engineering, Research Center for Intelligent Agriculture, Urumqi 830052, China
3
Xinjiang Agricultural Informatization Engineering Technology Research Center, Urumqi 830052, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(18), 5990; https://doi.org/10.3390/s24185990
Submission received: 15 August 2024 / Revised: 9 September 2024 / Accepted: 14 September 2024 / Published: 15 September 2024
(This article belongs to the Section Intelligent Sensors)
Figure 1
<p>Farm scene mask image.</p> ">
Figure 2
<p>Data processing flowchart.</p> ">
Figure 3
<p>SparseInst network architecture. The SparseInst network architecture comprises three main components: the backbone, the encoder, and the IAM-based decoder. The backbone extracts multi-scale image features from the input image, specifically {stage2, stage3, stage4}. The encoder uses a pyramid pooling module (PPM) [<a href="#B30-sensors-24-05990" class="html-bibr">30</a>] to expand the receptive field and integrate the multi-scale features. The notation ‘4×’ or ‘2×’ indicates upsampling by a factor of 4 or 2, respectively. The IAM-based decoder is divided into two branches: the instance branch and the mask branch. The instance branch utilizes the ‘IAM’ module to predict instance activation maps (shown in the right column), which are used to extract instance features for recognition and mask generation. The mask branch provides mask features M, which are combined with the predicted kernels to produce segmentation masks.</p> ">
Figure 4
<p>Improved SparseInst neck network PPM refers to the pyramid pooling module, MSA refers to the multi-scale attention module, 2× and 4× denote upsampling by a factor of 2 and 4, respectively, 3 × 3 denotes a convolution operation with a kernel size of 3, + denotes element-wise summation, and CA refers to the coordinate attention module.</p> ">
Figure 5
<p>Channel attention mechanism. GAP stands for global average pooling, relu is the rectified linear unit activation function, σ represents the Sigmoid activation function, C denotes the number of channels, and × denotes element-wise multiplication.</p> ">
Figure 6
<p>Dense connection diagram padding refers to the dilation rate of the convolution kernel, and C denotes feature concatenation.</p> ">
Figure 7
<p>Multi-scale attention module (MSA). GAP stands for global average pooling, relu is the rectified linear unit activation function, <span class="html-italic">σ</span> represents the activation function, padding refers to the dilation rate coefficient, and c denotes concatenation. + is element-by-element addition. × is a matrix product.</p> ">
Figure 8
<p>PADPN network architecture.</p> ">
Figure 9
<p>Coordinate attention blocks X Y (avg pool) denote global pooling along the h and w directions, BatchNorm refers to batch normalization, non-linear represents the non-linear activation function, split denotes splitting along the channel dimension, and Sigmoid represents the activation function.</p> ">
Figure 10
<p>Visualization results.</p> ">
Figure 11
<p>High-resolution image visualization results.</p> ">
Figure 12
<p>HRSID visualization results.</p> ">
Review Reports Versions Notes

Abstract

:
Farm aerial survey layers can assist in unmanned farm operations, such as planning paths and early warnings. To address the inefficiencies and high costs associated with traditional layer construction, this study proposes a high-precision instance segmentation algorithm based on SparseInst. Considering the structural characteristics of farm elements, this study introduces a multi-scale attention module (MSA) that leverages the properties of atrous convolution to expand the sensory field. It enhances spatial and channel feature weights, effectively improving segmentation accuracy for large-scale and complex targets in the farm through three parallel dense connections. A bottom-up aggregation path is added to the feature pyramid fusion network, enhancing the model’s ability to perceive complex targets such as mechanized trails in farms. Coordinate attention blocks (CAs) are incorporated into the neck to capture richer contextual semantic information, enhancing farm aerial imagery scene recognition accuracy. To assess the proposed method, we compare it against existing mainstream object segmentation models, including the Mask R-CNN, Cascade–Mask, SOLOv2, and Condinst algorithms. The experimental results show that the improved model proposed in this study can be adapted to segment various complex targets in farms. The accuracy of the improved SparseInst model greatly exceeds that of Mask R-CNN and Cascade–Mask and is 10.8 and 12.8 percentage points better than the average accuracy of SOLOv2 and Condinst, respectively, with the smallest number of model parameters. The results show that the model can be used for real-time segmentation of targets under complex farm conditions.

1. Introduction

The construction of UAV aerial maps for farms has significant applications in crop monitoring, smart farm path planning [1], and farm machinery operations [2]. Particularly, it provides critical monitoring and decision-making support for unmanned farm machinery operations on large farms. When farm machines are manually operated, professionals are needed to drive large machinery, incurring labor costs. While driverless farm machines can avoid these issues, building a farm aerial survey layer is necessary to ensure safe autonomous operation. Deep learning technology enables precise recognition of farm images [3], allowing farm machinery to anticipate target locations within the farm layer, thereby avoiding potential risks. This capability enhances the accuracy of unmanned farm machinery decisions, improving both efficiency and safety. Consequently, the study of UAV farm image segmentation technology holds considerable practical significance [4].
Creating high-precision farm maps necessitates extensive detailed annotations. Traditional methods of layer construction rely on manual annotation, which is inefficient and prone to inaccuracies in boundary segmentation, resulting in significant labor and time expenditure. With the ongoing advancement of instance segmentation algorithms, both the accuracy and speed of segmentation have significantly improved. Instance segmentation methods [5,6], which differentiate targets and backgrounds at the pixel level, have thus become the standard for processing high-precision images. Compared to semantic segmentation, instance segmentation [7] offers the advantage of recognizing individual targets without confining objects to the same category. Unmanned farm machinery requires precise identification of all objects during operation, making instance segmentation highly suitable for constructing farm UAV aerial survey layers [8]. Unlike cityscapes and roads with regular geometries, farm scenes present unique challenges due to their natural features [9,10]. Farms typically consist of extensive arable land, mechanized paths, forest belts, and heathland, which feature minimal human intervention and exhibit complex characteristics, significant variations in scale among similar elements, and intricate edge details. These factors complicate pixel-level instance segmentation in farm imagery.
The rest of the paper is organized as follows. Section 2 presents research results on instance segmentation in the field of UAVs. Section 3 describes the process of collecting and constructing the farm aerial survey dataset. Section 4 details the SparseInst network and the improvement methods, including the multi-scale attention mechanism, aggregation path, and CA attention module, tailored to the characteristics of farm target structures. Section 5 analyzes the experimental results and includes the design and execution of comparison, ablation, and validation experiments on the HRSID. The conclusions are presented in Section 6.

2. Related Works

In recent years, extensive research has been conducted on the segmentation of remote sensing images. In 2017, Kaiming He et al. [11] introduced the Mask-RCNN instance segmentation network, which achieves pixel-level accuracy in image segmentation. Mask-RCNN operates as a two-stage network: it first detects candidate regions for object detection and subsequently segments instance masks based on these regions. Following Mask-RCNN, several networks such as Faster R-CNN [12], YOLACT [13], and PANet [14] have been developed, which build upon either candidate regions or anchor-based approaches for segmentation. Another category is single-stage segmentation networks that do not rely on anchors. These networks generate instance masks through aggregation and mask encoding rather than depending on anchor-based methods. Examples include CondInst [15], SOLO [16], SparseInst [17], and similar architectures. In the realm of remote sensing image segmentation, researchers have made substantial progress. For instance, Xiaocheng Zhou et al. [18] and colleagues achieved low-cost monitoring of young fir forests using UAV remote sensing images, with the measurement accuracy for individual saplings reaching 0.9968. Braga et al. [19] used the Mask R-CNN model to extract tree crowns from high-resolution WorldView-2 satellite imagery, achieving an F1 score of 0.86 for the extracted tree crowns. Jian et al. [20] designed a specialized CutPaste self-supervised method based on the Mask R-CNN network architecture to generate dense and reliable candidate anchors with the help of the region proposal network (RPN) network, enabling the segmentation and recognition of ships in remote sensing images. Yuanyuan Wang et al. [21] applied the fully convolutional network (FCN) semantic segmentation model to GaoFen-3 remote sensing images, completing the extraction and segmentation of water bodies, vegetation, and buildings. Beibei Xu et al. [10] proposed a new method based on the DeepLabv3plus network, combining the visible light color index with an instance segmentation method based on an encoder–decoder architecture. This approach achieves accurate detection and segmentation of densely distributed weeds and soybean crops, with a weed segmentation accuracy of 0.905. Yunpeng Wu et al. [22] developed a YOLARC framework for ranking potential safety hazards (PSHs) by utilizing UAVs to perform track inspections on high-speed railroads. YOLARC employs multiple convolutions with enlarged receptive fields to create a new backbone that can accurately predict regions with PSHs. Ran Zou et al. [23] evaluated the damage level of post-disaster buildings by integrating the mixed local channel attention (MLCA) mechanism into the YOLOv5-Seg backbone. The improved YOLOv5-Seg was three times more efficient than the Mask-R-CNN and YOLO V9-Seg models. Zhihao Guan et al. [24] developed an instance segmentation method (MaskSU R-CNN) based on the channel domain attention mechanism by incorporating the dual semantic attention (DSA) module into the ResNet backbone network. Using the fire image dataset FLAME, MaskSU R-CNN achieved 91.85% accuracy in early forest fire detection and segmentation. Huan Liu et al. [25] proposed a novel UAV remote sensing segmentation model called Clusterformer, which is an encoder–decoder structure, with the encoder consisting of a spatial channel feed-forward network (SC-FFN). The model is used to segment pine wilt disease (PWD), and its segmentation results on two PWD datasets are significantly better than those of existing state-of-the-art segmentation models.
Most studies in this area focus on target detection for segmentation, with a majority of segmented objects being relatively small. However, farm UAV images differ significantly from traditional datasets; for instance, a single piece of cultivated land in a farm image can encompass tens of thousands of pixels, and the narrow, complex structure of mechanized roads can cover the entire image. Using instance segmentation based on anchor frames often leads to issues such as high repetitiveness of anchors, excessively large anchor areas, excessive background features within anchors, and difficulties in target mask extraction. Therefore, this study adopts the SparseInst instance segmentation algorithm, which is based on pixel aggregation. This method utilizes instance activation maps to weigh and highlight relevant features, overcoming the limitations associated with anchor-based approaches. It proves to be advantageous for segmenting large-scale targets and complex features in farm images. The main contributions of this study are as follows.
In this study, the multi-scale attention module (MSA) is designed to address the challenges of target scale variations and irregular shapes in farm images. This module employs three atrous convolutions along with dense connectivity to extract multi-scale feature information, enabling it to capture a larger receptive field.
In this study, a bottom-up aggregation path is introduced to enhance the fusion of semantic information from low-level feature maps. This addition improves the model’s detection performance by more effectively integrating details from earlier layers with higher-level features.
Incorporating the CA attention mechanism into the pyramid fusion network significantly improves the convolutional neural network’s capability to capture long-range dependencies and extract detailed features.

3. Instance Dataset

The farm image data used in this study were collected from Huaxing Farm in Changji City, Xinjiang, China, Uygur Autonomous Region, using a DJI Mavic 3 UAV equipped with a visible light camera at a flight altitude of 100 m. The images were captured with a resolution of 5280 × 3956 pixels and a 24-bit JPG color depth. Instance segmentation labeling of the farm images was performed using the Labelme tool [26], resulting in a total of 1155 finely labeled images. The labeling mask is illustrated in Figure 1.
In order to maintain the target resolution of the images, scaling was not employed. However, due to the high computational demands associated with the original images, which have a large resolution [27], this experiment involves cropping the mask-labeled maps generated after labeling, along with the original images. A total of 10,395 images with a resolution of 1760 × 1319 were generated for each of the three image types—original, semantic labeled, and instance labeled—at a 1:9 ratio. The cropped mask maps were then restructured into a COCO [28] format dataset [29], as illustrated in Figure 2.
The farm dataset was split into training and test sets with a 4:1 ratio. The dataset was annotated with a total of eight categories, with the majority of farm elements being plotlands. After cropping at a 1:9 ratio, plotlands emerged as the most numerous category followed by forest belts and heathland. Outletpiles were distributed within the plots but were present at the smallest scale and in limited quantities, making them challenging to segment. The distribution of the farm categories is detailed in Table 1.

4. Methodology

4.1. SparseInst Instance Segmentation

Many instance segmentation networks combine target detection by adding a mask segmentation head beneath the generated anchor box to further segment instances. However, this approach is limited by the location of the anchor box, making it difficult to segment large targets. SparseInst [17] introduces a novel method using instance activation maps to extract feature weights, followed by a mask extraction branch to complete the segmentation. This avoids the limitations of traditional anchor-based mask extraction, making it well suited for instance segmentation of farm images. The method is innovative, achieves high accuracy, and uses a lightweight network architecture. The network structure is illustrated in Figure 3.
The SparseInst network constructs the instance activation mapping (IAM) segmentation decoder after the feature pyramid fusion network, consisting of an instance branch and a mask branch. First, the instance feature branch generates instance-aware weighted weights for the fused semantic features, which are then passed to the mask branch for precise instance segmentation. This design enhances the segmentation process by dynamically adapting to the features of each instance, improving accuracy, particularly for large and complex targets. This process involves four 3 × 3 convolutional layers, followed by a sigmoid activation function to produce feature probability values ranging from 0 to 1. The instance activation map is then derived by summing these values along the last dimension. This step can be expressed as follows. Given the input image feature X R D × ( H × W ) , the instance activation map A is calculated as A = F I A M ( X ) R N × ( H × W ) , where A represents the sparse N-instance activation map and F i a m ( · ) denotes the sigmoid nonlinear activation function. Subsequently, a multiplication operation is performed with the reshaped features A, aggregating target information based on the activation map. Finally, three linear layers are applied for classification, target scoring, and generating the mask kernel. The mask branch then further refines the features using four 3 × 3 convolutions. The mask kernel obtained from the instance branch is matrix multiplied by the output from the mask branch to generate the predicted mask information. This process enhances the instance-aware features through instance-activated mapping, as shown in Equation (1).
z = A ¯ X T R N × D
where z = { z i } N is the feature representation of N potential objects in the image and A is normalized for each instance map. The sparse instance-aware features { z i } N are directly used for subsequent recognition and instance-level segmentation.
Due to farm elements’ scale variations and irregular shapes, this study introduces the MSA. This module employs three atrous convolutions with dense connections to perform multi-scale feature extraction. Using atrous convolutions allows for a larger receptive field and more efficient feature extraction. Additionally, a bottom-up aggregation path is incorporated into the neck network, facilitating comprehensive information transfer across different levels. The CA mechanism is designed to capture long-range dependencies and preserve detailed positional information. By importing this mechanism into the feature fusion network, important features can be captured more effectively. The improved network structure is illustrated in Figure 4.

4.2. Multi-Scale Attention Mechanism

To address the complex target features and large-scale variations in farm scenes, which traditional attention mechanisms struggle to handle, this study introduces an MSA mechanism. This mechanism is tailored to the characteristics of farm targets and uses parallel atrous convolutions for multi-scale semantic extraction, enhancing both channel and spatial semantic information. Although the channel attention (SENet) [31] can effectively find out the importance level between channels, the full connectivity layer used introduces new parameters that tend to cause overfitting. In this study, two 1 × 1 convolutions of the pyramid split attention (PSA) module in the EPSANet [32] network are used to replace the fully connected layer. The structure is shown in Figure 5, and this approach captures the dependencies between channels while effectively avoiding the complex process of dimensionality reduction and dimensionality upgrading and reduces the number of parameters, which enhances the transfer of information between channels.
Spatial features can be enhanced using the simple attention mechanism (Simam) [33] parameterless module, which deals with each position of the feature map through a pixel-by-pixel adaptive mechanism, which means that each pixel’s position can be independently adjusted with its weight. In contrast, the SENet mechanism involves a more complex structure and greater computational demands because it processes global information. Compared to the Simam module, feature enhancement can be more precisely controlled, resulting in improved expressiveness of the features. The Simam module’s simple structure and absence of parameters reduce the risk of overfitting. The Simam mechanism emulates the spatial inhibition behavior of neurons. It identifies a measure of linear separability between the target neuron and other neurons, with the energy function defined as shown in Equation (2).
e t ( w t , b t , y , x i ) = y t t ^ 2 + 1 M 1 i = 1 M 1   y o x ^ i 2
where t and x i are the target neuron and other neurons in a single channel of the input feature, X R C × H × W is the index of the spatial dimension, and M = H × W is the number of neurons on that channel. w t and b t are the weights and biases of the transformation.
Despite the enhancement of semantic features through spatial and channel-wise attention, significant scale differences among targets in farm scenes persist. To address this, this study implements three parallel atrous convolutions for multi-scale feature extraction. A key advantage of atrous convolution is its ability to expand the receptive field [34] without increasing the number of parameters or computational cost. The receptive field is defined by the formula given in Equation (3).
F i = ( F i 1 1 ) × s t r i d e + k
where F i denotes the receptive field of the i layer, F i 1 denotes the receptive field of the i 1 layer, stride denotes the table step, and k denotes the convolutional kernel size. This is experiment number 3. Features at various scales are extracted by employing three different dilation rates of 1, 2, and 4. This multi-scale feature extraction capability allows the model to more effectively extract targets with significant scale variations. The structure is further optimized through dense connectivity with parallel connections [35]. As illustrated in Figure 6, the output of one atrous convolution [36] is used as the input for the subsequent atrous convolution. Additionally, the outputs of both the previous and current layers are concatenated along the channel dimension using the concat operation, which enhances the model’s expressive power.
The structure of the MSA network designed for this experiment is illustrated in Figure 7. The fully connected layer is replaced by two 1 × 1 convolutions to obtain channel semantic information weights, which are then matrix multiplied with the input features to produce the enhanced channel semantic information weighted map. Another branch applies three parallel atrous convolutions for large-scale feature extraction. However, this parallel design does not optimally utilize multi-scale semantic information and shares limitations similar to the feature pyramid network (FPN), often resulting in an incomplete representation of small-scale semantic information in the feature maps, leading to potential loss. To address this, a dense connection structure is introduced between the three atrous convolutions, as depicted in Figure 6. This design fuses small-scale and complex structural semantic information through channel concatenation, ensuring the retention of small-scale semantic information while extracting large-scale targets. Following a 1 × 1 dimensionality reduction, the output is summed with the channel information weighted map and the features enhanced by the Simam mechanism to generate the multi-scale semantic feature map. The results demonstrate that this multi-scale attention module (MSA) significantly enhances the extraction of large-scale target features within the farm environment.

4.3. Feature Fusion Network

In the SparseInst neck network, features from the last three stages of ResNet are utilized with a top-down path for fusing features at different scales. Despite this multi-scale fusion, the combination of detailed and high-level information may still be insufficient for capturing underlying semantics. This is because low-level semantic information often contains richer details than high-level semantic information. Specifically, features of small targets may become increasingly abstracted and lost as semantic information is extracted through successive convolutional layers. In this study, a bottom-up aggregation path is incorporated into the path aggregation network (PAFPN) structure. This enhancement allows for a more effective fusion of semantic information from low-level feature maps, thereby improving the model’s detection performance. The revised network architecture is illustrated in Figure 8.

4.4. CA Attention Mechanism

Considering the complexity of farm imagery, the CA attention mechanism [37] was added to the pyramid fusion network. This addition improves feature representation by introducing spatial coordinate information into the channel attention. The structure of the enhanced network is illustrated in Figure 9.
The coordinate attention (CA) mechanism encodes channel relationships and long-term dependencies in two steps: embedding coordinate information and generating coordinate attention. For the input, X encodes each channel with two spatial ranges of the pooled kernel ( H , 1 ) and ( 1 , W ) along the horizontal coordinates x c ( h , i ) and vertical coordinate x c ( j , w ) . This can be expressed by Equations (4) and (5).
z c h ( h ) = 1 W 0 j < W x c ( h , i )
z c w ( w ) = 1 H 0 j < H x c ( j , w )
In this context, z c h ( h ) and z c w ( w ) are the feature encodings of the coordinate direction. After aggregating the features along these two spatial directions and performing dimensionality reduction, the fused directional feature information is obtained. This information is then extended by applying the sigmoid activation function to derive the attention weights. The final output can be expressed by Equation (6).
y c ( i , j ) = x c ( i , j ) × g c h ( i ) × g c w ( j )
where gh and gw are spatial direction weights. The coordinate attention (CA) mechanism incorporates coordinate information, significantly enhancing the ability of convolutional neural networks to capture long-range dependencies and extract detailed features.
In this study, we enhance the performance of the SparseInst network for segmenting farm elements by improving the multi-scale feature fusion capabilities of the neck network. This is achieved by adding a bottom-up aggregation path to form a new feature fusion network. Furthermore, we design a multi-scale attention module tailored to the characteristics of farm elements and integrate a coordinate attention mechanism to meet the segmentation requirements for farm elements.

4.5. Loss Function

The SparseInst network produces a fixed-size set of predictions. However, because the number of predictions may differ from the number of actual objects, accurately matching each prediction with a real object can be challenging. Therefore, the label assignment problem is formulated as a two-way matching issue using a matching score based on the Dice coefficient, denoted as C ( i , k ) , which is used to measure the degree of matching between the i th prediction and the k th real object as in Equation (7).
C ( i , k ) = p i , c k 1 a D I C E ( m i , t k ) a
a is a hyperparameter that balances the effects of the categorization score and the Dice coefficient. P i , c k is the probability that the i th prediction belongs to the k th true object category. D I C E ( m i , t k ) is the Dice coefficient of the i th prediction mask and the k th true mask, which is computed by Equation (8) as follows.
D I C E ( m , t ) = 2 x , y m x y t x y x , y m x y 2 + x , y t x y 2
where m x y and t x y represent the pixel values of the i th prediction mask and the k th true mask at position ( x , y ) , respectively.
The training loss consists of the following components. Classification Loss employs Focal Loss [38] to deal with the class classification problem, especially to deal with class imbalance. Mask Loss evaluates the accuracy of the predicted mask in matching the true mask. Targeting Loss measures the probability of the target’s presence, using IoU-aware binary cross-entropy loss. The final training loss is defined as shown in Equation (9).
L = λ c L c l s + L m a s k + λ s L s
where λ c and λ s are weighting factors.

5. Experimental Results and Analysis

5.1. Evaluation Indicators

Average precision (AP) [39] is selected as the evaluation metric for the instance segmentation model in the experiments. Precision reflects the model’s accuracy, while recall evaluates the model’s ability to recognize instances. Average precision, which represents the area under the precision–recall curve, is used to comprehensively assess the model’s performance on the test set and is calculated as shown in Equation (10).
A P = 0 1 P r e c i s i o n ( R e c a l l ) d ( R e c a l l )
Precision and recall are calculated using Equations (11) and (12), respectively. The mean average precision (mAP) is the average of the AP values for all categories, as shown in Equation (13).
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
mAP = 1 N i = 1 n A P i
TP, FP, and FN represent true positives, false positives, and false negatives under different IoU thresholds, respectively. IoU measures the degree of overlap between the predicted and true masks, defined as the ratio of the intersection to the union of the predicted and true mask regions. This experiment uses the average precision to evaluate the model’s performance across different IoU thresholds, ranging from 0.5 to 0.95. Frames Per Second (FPS) and cross-validation metrics are also used to evaluate model performance.
  • AP50: average precision when the IoU threshold is 0.5.
  • AP75: average precision when the IoU threshold is 0.75.
  • APs: average precision for small objects (area < 322).
  • APm: average precision for medium objects (322 ≤ area < 962).
  • APl: average precision for large objects (area ≥ 962).

5.2. Experimental Environment

The operating system used in this experiment is Linux 3.10 with an NVIDIA A100 GPU with 16 GB memory. The deep learning framework employed is PyTorch 1.10.0 and the parallel architecture used is CUDA 11.2. Due to the significant category difference between this dataset and the COCO dataset, the initial weights from the COCO pre-trained model were not utilized. The images for pre-training were resized to 1056 × 790 pixels. The initial learning rate was set to 0.00005, with a momentum [40] factor of 0.9 and a weight decay [41] of 0.05. The total number of iterations was 200,000, and the batch size was 2. Data enhancements such as scaling, horizontal flipping, and vertical flipping of images are added during the training process.

5.3. Ablation Study

5.3.1. Comparison with Other Baselines

To verify the effectiveness of the improved model, comparisons were made with other segmentation algorithms, including the classical networks Mask R-CNN, Cascade–Mask [42], and the widely used SOLOv2 [43] and Condinst networks, based on mean average precision (mAP). In addition to these comparison experiments, cross-validation was performed to further assess the model’s generalization and robustness, ensuring that it did not overfit specific data. The network complexity (GFLOPs) and model inference speed (FPS) were also measured, with the results presented in Table 2.
The experimental results revealed that the segmentation accuracy of Mask R-CNN and Cascade–Mask networks is poor due to the large scale of the farm instance targets. Both networks utilize the RPN region proposal network for target detection, generating nine candidate regions with aspect ratios of 1:1, 1:2, and 2:1, and then segmenting the instance masks based on these regions. However, the complex structure and scale of farm targets make it difficult to generate accurate candidate regions, resulting in imprecise segmentation. In contrast, SOLOv2, CondInst, and SparseInst are single-stage anchor-free algorithms, which avoid the limitations of RPN-based detection. The results indicate that the improved SparseInst increases the average precision by 10.8 and 12.8 percentage points compared to these two algorithms. The cross-validation results further confirm that the model demonstrates strong generalization and robustness, performing well on five randomly divided datasets and outperforming other single-stage anchor-free networks. It also has the fastest inference speed and the lowest network complexity, making SparseInst the optimal choice for instance segmentation of farm images.

5.3.2. Ablation Study Results and Analysis

To validate the effectiveness of each improvement strategy, this study used SparseInst as the benchmark network. Ablation experiments were conducted to assess each enhancement, and the results are presented in Table 3.
The ablation experiment, which focused on a farm aerial survey scene, revealed that incorporating the MSA module into the feature fusion network enhances the average precision by 1.32 percentage points compared to the baseline network, and the AP75/% increases by 2.1 percentage points. In terms of average precision across different pixel areas, both the baseline network and the network with the MSA module show limited performance in distinguishing small targets (less than 32 × 32 pixels) and medium targets (between 96 × 96 pixels). However, the APl/% for large targets improved by 1.927 percentage points. These results demonstrate that the proposed MSA module, which utilizes atrous convolution, significantly benefits the segmentation of large-scale targets. The addition of an aggregation path to the feature fusion structure enhances the integration of complex multi-target features based on the MSA module, particularly improving the handling of narrow and variable targets, such as mechanized roads and highways. The incorporation of the CA module significantly enhances the accuracy of the overall network structure. The average accuracy improves by 2.103 percentage points compared to the baseline model. Additionally, the average precision at IoU thresholds of 50% (AP50) and 75% (AP75) increased by 0.708 and 3.167 percentage points, respectively. For large targets exceeding 96 × 96 pixels, the improvement over the baseline network is 2.69 percentage points. Overall, the enhanced SparseInst-based network structure demonstrates a significant improvement in average accuracy. However, both the baseline and improved networks continue to face challenges in detecting small-scale targets, such as outletpile.
Table 4 shows that this is the average AP/% accuracy of the improved network for the farm categories, with eight categories. Most of these categories consist of plotlands and adjacent forest belts, which represent the majority of the categories and exhibit the largest scales. The dataset also includes arable roads (paths between plots) and irrigation watercourses. Due to the UAV aerial imagery, these structures often span entire images, presenting narrow and irregular shapes with colors similar to the background, which makes segmentation challenging and tends to result in inaccuracies.
The results show that the highest accuracy of plotland and river reaches is more than 90%. The MSA module significantly improves the overall model accuracy, particularly enhancing the extraction capabilities for targets with complex and irregular structures, such as forest belts, arable roads, wild grasslands, watercourses, and rivers. This improvement demonstrates that densely connected atrous convolutions effectively capture spatial information across multiple scales. Notably, the largest gains are observed in the segmentation of watercourses and rivers, with accuracy improvements of 2.506 and 3.581 percentage points, respectively. The addition of an aggregation path to the feature fusion structure significantly enhances the segmentation performance for roads and mechanized roads, with improvements of 3.197 and 7.385 percentage points, respectively, compared to the baseline network. This modification also enhances information transfer and fusion, particularly for arable roads, which exhibit large variations in structural scales and complexity. Incorporating the CA module further enhances the network’s feature extraction capability. However, both the improved and baseline networks struggle with the accurate segmentation of very small-scale targets, such as outletpiles. This challenge arises because outlet piles, distributed within plots, have small shapes and colors that blend into the surrounding land, making them difficult to distinguish and segment effectively.

5.3.3. Visual Analysis of Experimental Outcomes

As shown in Figure 10, the segmentation results of the improved strategy are compared with those of the other two models. Due to the inapplicability of Mask R-CNN and Cascade–Mask to farm image segmentation, their results are low due to accuracy. The six images include categories such as plotlands, forest belts, highways, watercourses, rivers, arable roads, and highways. The results demonstrate that both SparseInst and the improved SparseInst achieve high accuracy in segmenting plotlands, highways, and rivers. Notably, the original SparseInst missed the detection and segmentation of forest belts, whereas the improved SparseInst achieved more comprehensive segmentation of forest belts, wild grasslands, and arable roads.
In order to verify the network’s performance on real-world images, the final trained model was used to recognize high-resolution farm images, as shown in Figure 11. Since the image resolution is 5280 × 3956, the available computational resources could not process such large-resolution images, so the network was trained using uniformly cropped images with a resolution of 1760 × 1319. As a result, the model demonstrated excellent segmentation performance on the cropped images but incomplete segmentation for targets in the uncropped high-resolution images. However, the improved SparseInst still outperforms the original SparseInst network for high-resolution farm image segmentation. This issue arises because the feature maps generated from low-resolution images during training are specific to certain scales and may not directly align when applied to high-resolution images, thus affecting the model’s segmentation accuracy.
In fact, in performing the construction of the farm layer, a complete map is stitched together from numerous small maps. The operation of this study for cropping the training image does not affect the construction of subsequent layers. However, to address this deficiency, plans could include image scaling, where the original image is cropped as close to its original resolution as possible, and image scaling is used to keep the image undistorted. Additionally, implementing multi-scale training using images of different resolutions during the training process can improve the model’s ability to generalize to images of varying scales.

5.3.4. HRSID Validation

For the segmentation of high-resolution farm remote sensing images, this study adopts the SparseInst single-stage anchorless instance segmentation algorithm. To further validate the effectiveness of the proposed approach, a comparative validation was conducted using the High-Resolution SAR Images Dataset (HRSID) with the same single-stage anchorless instance segmentation algorithm. The HRSID is a high-resolution synthetic aperture radar (SAR) remote sensing dataset for vessel detection, consisting of 5604 high-resolution SAR images. Since the SAR images are grayscale, with vessels and backgrounds distinguishable in black and white, it is easy to evaluate the accuracy of the network’s segmentation masks. Given the similarity between the HRSID and the UAV farm dataset in this study, both falling within the same category of remote sensing, the HRSID is suitable for validating the effectiveness of the improved method presented in this research, and the results are presented in Table 5.
The results indicate that the improved SparseInst network achieves the highest average accuracy and has the lowest complexity compared to the SOLOv2 and CondInst networks. The improved network increases the complexity compared to the baseline network but effectively enhances model accuracy. The segmentation results for vessels are illustrated in Figure 10. Validation experiments with the HRSID confirm that the algorithm performs effectively in single-stage anchorless instance segmentation and the model’s ability to generalize. The segmentation results are shown in Figure 12.

6. Conclusions

This paper focuses on UAV aerial survey images of farms, constructing a dedicated farm aerial survey image dataset. It proposes an instance segmentation algorithm with an improved SparseInst network, specifically designed to address the challenges of complex target structures and significant scale variations in farm images. We design the MSA module, introduce the CA module, and enhance the neck network. Based on ablation experiments and comparisons with other classical networks, the following conclusions are drawn.
(1) The experimental results on the self-constructed farm imagery dataset demonstrate that the improved SparseInst enhances the average accuracy by 2.103, 0.708, and 3.167 percentage points, respectively, compared to the baseline model. Notably, the proposed MSA has significantly improved the detection and segmentation of watercourses and river targets in complex scenes. The arable road is one of the most complex elements due to its color similarity to the surrounding background. However, the improved feature fusion network effectively enhances the perception of such challenging semantic information, enabling better recognition of arable roads.
(2) Comparison experiments with the Mask R-CNN, Cascade–Mask, SOLOv2, and Condinst algorithms demonstrated that the improved SparseInst outperforms SOLOv2 and Condinst by 10.8 and 12.8 percentage points, respectively. The experiments on the farm aerial survey dataset conclude that instance segmentation algorithms based on anchor-free methods are better suited for handling large-scale targets compared to those built on object detection frameworks.
The enhanced accuracy of the model presented in this paper meets the requirements for constructing farm layers, thereby improving manual construction efficiency and reducing labor costs. Additionally, it provides technical support for instance segmentation in UAV imagery.
Despite significant advancements in UAV farm image segmentation, current datasets still face limitations, particularly in the number of small target instances and the ability to segment targets with complex structures. Future research should focus on increasing the instances of small-scale and complex-structured targets and expanding the dataset to include a wider range of farm categories. Due to the model’s poor inference performance on high-resolution images, the next step is to create a multi-scale farm image dataset to improve the model’s segmentation performance on such images. Additionally, optimizing model parameters is crucial for enhancing both accuracy and generalization capability.

Author Contributions

Conceptualization, F.L.; methodology, F.L. and X.C.; software, X.C.; formal analysis, F.L. and T.Z.; investigation, F.L. and T.Z.; funding acquisition, T.Z.; supervision, Y.Z. and Z.Y.; writing—original draft preparation, F.L.; writing—review and editing, Y.Z. and Z.Y.; validation, F.L. and X.C.; data curation, F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Key R&D Program of China (2022ZD0115805) and the Provincial Key S&T Program of Xinjiang (2022A02011).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The data supporting the findings of this study are available upon request from the readers.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Urvina, R.P.; Guevara, C.L.; Vásconez, J.P.; Prado, A.J. An Integrated Route and Path Planning Strategy for Skid–Steer Mobile Robots in Assisted Harvesting Tasks with Terrain Traversability Constraints. Agriculture 2024, 14, 1206. [Google Scholar] [CrossRef]
  2. Wei, S.; Lu, Y. Adoption mode of agricultural machinery and food productivity: Evidence from China. Front. Sustain. Food Syst. 2024, 7, 1257918. [Google Scholar] [CrossRef]
  3. Hinton, G.E.; Salakhutdinov, R.R. Reducing the Dimensionality of Data with Neural Networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
  4. Du, X.; Huang, D.; Dai, L.; Du, X. Recognition of Plastic Film in Terrain-Fragmented Areas Based on Drone Visible Light Images. Agriculture 2024, 14, 736. [Google Scholar] [CrossRef]
  5. Zhang, Z.; Zhu, L. A review on unmanned aerial vehicle remote sensing: Platforms, sensors, data processing methods, and applications. Drones 2023, 7, 398. [Google Scholar] [CrossRef]
  6. Su, H.; Wei, S.; Liu, S.; Liang, J.; Wang, C.; Shi, J.; Zhang, X. HQ-ISNet: High-quality instance segmentation for remote sensing imagery. Remote Sens. 2020, 12, 989. [Google Scholar] [CrossRef]
  7. Yin, C.; Tang, J.; Yuan, T.; Xu, Z.; Wang, Y. Bridging the gap between semantic segmentation and instance segmentation. IEEE Trans. Multimed. 2021, 24, 4183–4196. [Google Scholar] [CrossRef]
  8. Bouguettaya, A.; Zarzour, H.; Kechida, A.; Taberkit, A.M. Deep learning techniques to classify agricultural crops through UAV imagery: A review. Neural Comput. Appl. 2022, 34, 9511–9536. [Google Scholar] [CrossRef] [PubMed]
  9. Sang, Y.; Han, J. Improved DeepLabV3+ algorithm for scene segmentation. Electro-Opt. Control 2022, 29, 47–52. [Google Scholar]
  10. Xu, B.; Fan, J.; Chao, J.; Arsenijevic, N.; Werle, R.; Zhang, Z. Instance segmentation method for weed detection using UAV imagery in soybean fields. Comput. Electron. Agric. 2023, 211, 107994. [Google Scholar] [CrossRef]
  11. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  12. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar]
  13. Bolya, D.; Zhou, C.; Xiao, F.; Lee, Y.J. Yolact: Real-time instance segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 20–26 October 2019. [Google Scholar]
  14. Wang, K.; Liew, J.H.; Zou, Y.; Zhou, D.; Feng, J. Panet: Few-shot image semantic segmentation with prototype alignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 20–26 October 2019. [Google Scholar]
  15. Tian, Z.; Zhang, B.; Chen, H.; Shen, C. Instance and panoptic segmentation using conditional convolutions. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 669–680. [Google Scholar] [CrossRef] [PubMed]
  16. Wang, X.; Kong, T.; Shen, C.; Jiang, Y.; Li, L. Solo: Segmenting objects by locations. In Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XVIII 16; Springer International Publishing: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  17. Cheng, T.; Wang, X.; Chen, S.; Zhang, W.; Zhang, Q. Sparse instance activation for real-time instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  18. Zhou, X.; Wang, H.; Chen, C.; Nagy, G.; Jancso, T.; Huang, H. Detection of Growth Change of Young Forest Based on UAV RGB Images at Single-Tree Level. Forests 2023, 14, 141. [Google Scholar] [CrossRef]
  19. Braga, J.R.G.; Peripato, V.; Dalagnol, R.; Ferreira, M.P.; Tarabalka, Y.; Aragäo, L.E.O.C.; Velho, H.F.d.C.; Shiguemori, E.H.; Wagner, F.H. Tree Crown Delineation Algorithm Based on a Convolutional Neural Network. Remote Sens. 2020, 12, 1288. [Google Scholar] [CrossRef]
  20. Jian, L.; Pu, Z.; Zhu, L.; Yao, T.; Liang, X. SS R-CNN: Self-Supervised Learning Improving MaskR-CNN for Ship Detection in Remote Sensing Images. Remote Sens. 2022, 14, 4383. [Google Scholar] [CrossRef]
  21. Wang, Y.; Chao, W.; Hong, Z. Integrating HA-α with fully convolutional networks for fully PolSAR classification. In Proceedings of the 2017 International Workshop on Remote Sensing with Intelligent Processing (RSIP), Shanghai, China, 18–21 May 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar]
  22. Wu, Y.; Meng, F.; Qin, Y.; Qian, Y.; Xu, F.; Jia, L. UAV imagery based potential safety hazard evaluation for high-speed railroad using Real-time instance segmentation. Adv. Eng. Inform. 2023, 55, 101819. [Google Scholar] [CrossRef]
  23. Zou, R.; Liu, J.; Pan, H.; Tang, D.; Zhou, R. An Improved Instance Segmentation Method for Fast Assessment of Damaged Buildings Based on Post-Earthquake UAV Images. Sensors 2024, 24, 4371. [Google Scholar] [CrossRef]
  24. Guan, Z.; Miao, X.; Mu, Y.; Sun, Q.; Ye, Q.; Gao, D. Forest fire segmentation from Aerial Imagery data Using an improved instance segmentation model. Remote Sens. 2022, 14, 3159. [Google Scholar] [CrossRef]
  25. Liu, H.; Li, W.; Jia, W.; Sun, H.; Zhang, M.; Song, L.; Gui, Y. Clusterformer for pine tree disease identification based on UAV remote sensing image segmentation. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5609215. [Google Scholar] [CrossRef]
  26. Torralba, A.; Russell, B.C.; Yuen, J. Labelme: Online image annotation and applications. Proc. IEEE 2010, 98, 1467–1484. [Google Scholar] [CrossRef]
  27. Talebi, H.; Milanfar, P. Learning to resize images for computer vision tasks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  28. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13; Springer International Publishing: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  29. Waqas Zamir, S.; Arora, A.; Gupta, A.; Khan, S.; Sun, G.; Shahbaz Khan, F.; Zhu, F.; Shao, L.; Xia, G.-S.; Bai, X. isaid: A large-scale dataset for instance segmentation in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  30. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  31. Cheng, D.; Meng, G.; Cheng, G.; Pan, C. SeNet: Structured edge network for sea–land segmentation. IEEE Geosci. Remote Sens. Lett. 2016, 14, 247–251. [Google Scholar] [CrossRef]
  32. Zhang, H.; Zu, K.; Lu, J.; Zou, Y.; Meng, D. EPSANet: An efficient pyramid squeeze attention block on convolutional neural network. In Proceedings of the Asian Conference on Computer Vision, Macao, China, 4–8 December 2022. [Google Scholar]
  33. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; PMLR: Seattle, WA, USA, 2021. [Google Scholar]
  34. Luo, W.; Li, Y.; Urtasun, R.; Zemel, R. Understanding the effective receptive field in deep convolutional neural networks. arXiv 2016, arXiv:1701.04128. [Google Scholar]
  35. Huang, G.; Liu, S.; Van der Maaten, L.; Weinberger, K.Q. Condensenet: An efficient densenet using learned group convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  36. Liu, J.; Li, C.; Liang, F.; Lin, C.; Sun, M.; Yan, J.; Ouyang, W.; Xu, D. Inception convolution with efficient dilation search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  37. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  38. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  39. Padilla, R.; Netto, S.L.; Da Silva, E.A. A survey on performance metrics for object-detection algorithms. In Proceedings of the 2020 International Conference on Systems, Signals and Image Processing (IWSSIP), Niteroi, Brazil, 1–3 July 2020; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  40. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  41. Ishii, M.; Sato, A. Layer-wise weight decay for deep neural networks. In Image and Video Technology: 8th Pacific-Rim Symposium, PSIVT 2017, Wuhan, China, 20–24 November 2017; Revised Selected Papers 8; Springer International Publishing: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  42. Wang, G.; Xiong, Z.; Liu, D.; Luo, C. Cascade mask generation framework for fast small object detection. In Proceedings of the 2018 IEEE International Conference on Multimedia and Expo (ICME), San Diego, CA, USA, 23–27 July 2018; IEEE: Piscataway, NJ, USA, 2018. [Google Scholar]
  43. Wang, X.; Zhang, R.; Kong, T.; Li, L.; Shen, C. Solov2: Dynamic and fast instance segmentation. Adv. Neural Inf. Process. Syst. 2020, 33, 17721–17732. [Google Scholar]
Figure 1. Farm scene mask image.
Figure 1. Farm scene mask image.
Sensors 24 05990 g001
Figure 2. Data processing flowchart.
Figure 2. Data processing flowchart.
Sensors 24 05990 g002
Figure 3. SparseInst network architecture. The SparseInst network architecture comprises three main components: the backbone, the encoder, and the IAM-based decoder. The backbone extracts multi-scale image features from the input image, specifically {stage2, stage3, stage4}. The encoder uses a pyramid pooling module (PPM) [30] to expand the receptive field and integrate the multi-scale features. The notation ‘4×’ or ‘2×’ indicates upsampling by a factor of 4 or 2, respectively. The IAM-based decoder is divided into two branches: the instance branch and the mask branch. The instance branch utilizes the ‘IAM’ module to predict instance activation maps (shown in the right column), which are used to extract instance features for recognition and mask generation. The mask branch provides mask features M, which are combined with the predicted kernels to produce segmentation masks.
Figure 3. SparseInst network architecture. The SparseInst network architecture comprises three main components: the backbone, the encoder, and the IAM-based decoder. The backbone extracts multi-scale image features from the input image, specifically {stage2, stage3, stage4}. The encoder uses a pyramid pooling module (PPM) [30] to expand the receptive field and integrate the multi-scale features. The notation ‘4×’ or ‘2×’ indicates upsampling by a factor of 4 or 2, respectively. The IAM-based decoder is divided into two branches: the instance branch and the mask branch. The instance branch utilizes the ‘IAM’ module to predict instance activation maps (shown in the right column), which are used to extract instance features for recognition and mask generation. The mask branch provides mask features M, which are combined with the predicted kernels to produce segmentation masks.
Sensors 24 05990 g003
Figure 4. Improved SparseInst neck network PPM refers to the pyramid pooling module, MSA refers to the multi-scale attention module, 2× and 4× denote upsampling by a factor of 2 and 4, respectively, 3 × 3 denotes a convolution operation with a kernel size of 3, + denotes element-wise summation, and CA refers to the coordinate attention module.
Figure 4. Improved SparseInst neck network PPM refers to the pyramid pooling module, MSA refers to the multi-scale attention module, 2× and 4× denote upsampling by a factor of 2 and 4, respectively, 3 × 3 denotes a convolution operation with a kernel size of 3, + denotes element-wise summation, and CA refers to the coordinate attention module.
Sensors 24 05990 g004
Figure 5. Channel attention mechanism. GAP stands for global average pooling, relu is the rectified linear unit activation function, σ represents the Sigmoid activation function, C denotes the number of channels, and × denotes element-wise multiplication.
Figure 5. Channel attention mechanism. GAP stands for global average pooling, relu is the rectified linear unit activation function, σ represents the Sigmoid activation function, C denotes the number of channels, and × denotes element-wise multiplication.
Sensors 24 05990 g005
Figure 6. Dense connection diagram padding refers to the dilation rate of the convolution kernel, and C denotes feature concatenation.
Figure 6. Dense connection diagram padding refers to the dilation rate of the convolution kernel, and C denotes feature concatenation.
Sensors 24 05990 g006
Figure 7. Multi-scale attention module (MSA). GAP stands for global average pooling, relu is the rectified linear unit activation function, σ represents the activation function, padding refers to the dilation rate coefficient, and c denotes concatenation. + is element-by-element addition. × is a matrix product.
Figure 7. Multi-scale attention module (MSA). GAP stands for global average pooling, relu is the rectified linear unit activation function, σ represents the activation function, padding refers to the dilation rate coefficient, and c denotes concatenation. + is element-by-element addition. × is a matrix product.
Sensors 24 05990 g007
Figure 8. PADPN network architecture.
Figure 8. PADPN network architecture.
Sensors 24 05990 g008
Figure 9. Coordinate attention blocks X Y (avg pool) denote global pooling along the h and w directions, BatchNorm refers to batch normalization, non-linear represents the non-linear activation function, split denotes splitting along the channel dimension, and Sigmoid represents the activation function.
Figure 9. Coordinate attention blocks X Y (avg pool) denote global pooling along the h and w directions, BatchNorm refers to batch normalization, non-linear represents the non-linear activation function, split denotes splitting along the channel dimension, and Sigmoid represents the activation function.
Sensors 24 05990 g009
Figure 10. Visualization results.
Figure 10. Visualization results.
Sensors 24 05990 g010
Figure 11. High-resolution image visualization results.
Figure 11. High-resolution image visualization results.
Sensors 24 05990 g011
Figure 12. HRSID visualization results.
Figure 12. HRSID visualization results.
Sensors 24 05990 g012
Table 1. The distribution of the number of elements in the farm imagery dataset.
Table 1. The distribution of the number of elements in the farm imagery dataset.
PlotlandForest_BeltOutletpileArable RoadWild_GraslandRiverHighwayWatercoursesTotal
9465285216892244549349197896824,180
Table 2. Comparison with other algorithms.
Table 2. Comparison with other algorithms.
ModelmAP/%AP50/%AP75/%APs/%APm/%APl/%GFLOPsCV (mAP)FPS
Mask R-CNN31.119.69.40012.522629.023.1
Cascade–Mask39.054.039.40.32.045.3174933.212.7
SOLOv251.066.752.50.79.859.522443.827.5
Condinst49.062.850.72.75.357.217824.321.6
Improving SparseInst61.880.962.63.26.070.715757.338.8
Note: mAP/% represents mean accuracy, AP50/% is average precision at an IoU threshold of 50%, AP75/% is average precision at an IoU threshold of 75%. APs/% is average precision for small objects, APm/% is for medium objects, APl/% is average precision for large objects. CV refers to the mean accuracy obtained from cross-validation.
Table 3. Ablation study results for six types of mean average precision (mAP).
Table 3. Ablation study results for six types of mean average precision (mAP).
MSAPAFPNCVmAP/%AP50/%AP75/%Aps/%APm/%APl/%
×××59.75180.22759.4653.5854.34068.057
××61.07179.99161.5653.0288.10569.984
×61.05079.44362.4213.3404.73569.882
61.85480.93562.6323.2646.00870.747
Note: ‘×’ indicates that the corresponding strategy is not used, ‘√’ indicates that the corresponding strategy is adopted.
Table 4. Mean average precision of farm categories under different improvement strategies (AP/%).
Table 4. Mean average precision of farm categories under different improvement strategies (AP/%).
CategoryOriginal SparseInst+MSA+MSA + PAFPN+MSA + PAFPN + CA
Plotland93.46494.20594.38294.587
Forest_Belt62.70565.74364.83166.861
Outletpile16.54114.08915.69015.741
Highway82.03882.09785.23582.872
Arable Road41.04545.70948.43047.844
Wild_Grassland30.69734.11932.91236.338
Watercourse52.53255.03850.10553.692
River93.98797.56896.81696.900
Table 5. Comparative validation of the HRSIDs.
Table 5. Comparative validation of the HRSIDs.
ModelmAP/%AP50/%AP75/%APs/%APm/%APl/%GFLOPsFPS
SOLOv227.257.022.226.534.52.517819.8
Condinst31.965.828.931.538.96.721011.2
SparseInst31.455.835.131.137.37.611826.0
Improving SparseInst34.561.038.734.240.46.313922.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lv, F.; Zhang, T.; Zhao, Y.; Yao, Z.; Cao, X. An Improved Instance Segmentation Method for Complex Elements of Farm UAV Aerial Survey Images. Sensors 2024, 24, 5990. https://doi.org/10.3390/s24185990

AMA Style

Lv F, Zhang T, Zhao Y, Yao Z, Cao X. An Improved Instance Segmentation Method for Complex Elements of Farm UAV Aerial Survey Images. Sensors. 2024; 24(18):5990. https://doi.org/10.3390/s24185990

Chicago/Turabian Style

Lv, Feixiang, Taihong Zhang, Yunjie Zhao, Zhixin Yao, and Xinyu Cao. 2024. "An Improved Instance Segmentation Method for Complex Elements of Farm UAV Aerial Survey Images" Sensors 24, no. 18: 5990. https://doi.org/10.3390/s24185990

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop