[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,334)

Search Parameters:
Keywords = F-CNN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 8420 KiB  
Article
CRAUnet++: A New Convolutional Neural Network for Land Surface Water Extraction from Sentinel-2 Imagery by Combining RWI with Improved Unet++
by Nan Li, Xiaohua Xu, Shifeng Huang, Yayong Sun, Jianwei Ma, He Zhu and Mengcheng Hu
Remote Sens. 2024, 16(18), 3391; https://doi.org/10.3390/rs16183391 - 12 Sep 2024
Viewed by 221
Abstract
Accurately mapping the surface water bodies through remote sensing technology is of great significance for water resources management, flood monitoring, and drought monitoring. At present, many scholars at home and abroad carry out research on deep learning image recognition algorithms based on convolutional [...] Read more.
Accurately mapping the surface water bodies through remote sensing technology is of great significance for water resources management, flood monitoring, and drought monitoring. At present, many scholars at home and abroad carry out research on deep learning image recognition algorithms based on convolutional neural networks, and a variety of variant-based convolutional neural networks are proposed to be applied to extract water bodies from remote sensing images. However, due to the low depth of convolutional layers employed and underutilization of water spectral feature information, most of the water body extraction methods based on convolutional neural networks (CNNs) for remote sensing images are limited in accuracy. In this study, we propose a novel surface water automatic extraction method based on the convolutional neural network (CRAUnet++) for Sentinel-2 images. The proposed method includes three parts: (1) substituting the feature extractor of the original Unet++ with ResNet34 to enhance the network’s complexity by increasing its depth; (2) Embedding the Spatial and Channel ‘Squeeze and Excitation’ (SCSE) module into the up-sampling stage of the network to suppress background features and amplify water body features; (3) adding the vegetation red edge-based water index (RWI) into the input data to maximize the utilization of water body spectral information of Sentinel-2 images without increasing the data processing time. To verify the performance and accuracy of the proposed algorithm, the ablation experiment under four different strategies and comparison experiment with different algorithms of RWI, FCN, SegNet, Unet, and DeepLab v3+ were conducted on Sentinel-2 images of the Poyang Lake. The experimental result shows that the precision, recall, F1, and IoU of CRAUnet++ are 95.99%, 96.41%, 96.19%, and 92.67%, respectively. CRAUnet++ has a good performance in extracting various types of water bodies and suppressing noises because it introduces SCSE attention mechanisms and combines surface water spectral features from RWI, exceeding that of the other five algorithms. The result demonstrates that CRAUnet++ has high validity and reliability in extracting surface water bodies based on Sentinel-2 images. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>The geographic location of the study area and the selected 4 Sentinel-2 Multi-Spectral Instrument (MSI) images.</p>
Full article ">Figure 2
<p>Flowchart of Sentinel-2 MSI images preprocessing. The first row shows the process of generating training sample data used as input to the model, and the second row shows the process of generating water body labels used to calculate model losses.</p>
Full article ">Figure 3
<p>Examples of data augmentation. (<b>a</b>) is the original image and corresponding label; (<b>b</b>) is flipping the original image up and down; (<b>c</b>) is flipping the original image left and right; (<b>d</b>) is rotating the original image at 180°.</p>
Full article ">Figure 4
<p>Reflectance of water, soil, and vegetation at different wavelengths [<a href="#B52-remotesensing-16-03391" class="html-bibr">52</a>].</p>
Full article ">Figure 5
<p>The structure of CRAUnet++. RGB+RWI indicates a combination of red, green, blue, and vegetation red edge based water index (RWI) images.</p>
Full article ">Figure 6
<p>The structure of ResNet34 feature extractor using in CRAUnet++. It consists of two Basic Blocks, which are represented by different colors in the figure. The first convolutional layer of the green BasicBlock has a stride of 1 and its input and output feature maps are of the same size. The first convolutional layer of the yellow BasicBlock has a stride of 2 and can downsample the input feature map twice.</p>
Full article ">Figure 7
<p>The structure of Spatial and Channel ‘Squeeze and Excitation’ (SCSE) module. The first row of this figure weights the input data from the spatial dimension, and the second row of this figure weights the input data from the channel dimension.</p>
Full article ">Figure 8
<p>Sentinel-2 images under different band combinations (<b>a</b>) true color composite image, (<b>b</b>) false color composite image, and (<b>c</b>) Vegetation red edge based water index (RWI) image.</p>
Full article ">Figure 9
<p>The trend of accuracy and loss values in the training and testing sets (the upper graph shows the trend of accuracy value change, the lower graph shows the trend of loss value change, the red line represents the testing set, and the blue line represents the training set).</p>
Full article ">Figure 10
<p>Visualization of water extraction results for ablation studies: (<b>a</b>) images; (<b>b</b>) labels; (<b>c</b>) Baseline; (<b>d</b>) Baseline + RWI; (<b>e</b>) Baseline + RWI + ResNet34; (<b>f</b>) Baseline + RWI + ResNet34 + SCSE. Black denotes the background, and white denotes water bodies.</p>
Full article ">Figure 11
<p>Visualization results of CRAUnet++, RWI, and CNN-based semantic segmentation networks on Sentinel-2 dataset: (<b>a</b>) images; (<b>b</b>) labels; (<b>c</b>) RWI; (<b>d</b>) FCN; (<b>e</b>) SegNet; (<b>f</b>) Unet; (<b>g</b>) DeepLab v3+; (<b>h</b>) CRAUnet++. Black denotes the background, and white denotes water bodies.</p>
Full article ">
15 pages, 6817 KiB  
Article
A Fully Connected Neural Network (FCNN) Model to Simulate Karst Spring Flowrates in the Umbria Region (Central Italy)
by Francesco Maria De Filippi, Matteo Ginesi and Giuseppe Sappa
Water 2024, 16(18), 2580; https://doi.org/10.3390/w16182580 - 12 Sep 2024
Viewed by 354
Abstract
In the last decades, climate change has led to increasingly frequent drought events within the Mediterranean area, creating an urgent need of a more sustainable management of groundwater resources exploited for drinking and agricultural purposes. One of the most challenging issues is to [...] Read more.
In the last decades, climate change has led to increasingly frequent drought events within the Mediterranean area, creating an urgent need of a more sustainable management of groundwater resources exploited for drinking and agricultural purposes. One of the most challenging issues is to provide reliable simulations and forecasts of karst spring discharges, whose reduced information, as well as the hydrological processes involving their feeding aquifers, is often a big issue for water service managers and researchers. In order to plan a sustainable water resource exploitation that could face future shortages, the groundwater availability should be assessed by continuously monitoring spring discharge during the hydrological year, using collected data to better understand the past behaviour and, possibly, forecast the future one in case of severe droughts. The aim of this paper is to understand the factors that govern different spring discharge patterns according to rainfall inputs and to present a model, based on artificial neural network (ANN) data training and cross-correlation analyses, to evaluate the discharge of some karst spring in the Umbria region (Central Italy). The model used is a fully connected neural network (FCNN) and has been used both for filling gaps in the spring discharge time series and for simulating the response of six springs to rainfall seasonal patterns from a 20-year continuous daily record, collected and provided by the Regional Environmental Protection Agency (ARPA) of the Umbria region. Full article
(This article belongs to the Special Issue Recent Advances in Karstic Hydrogeology, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Simplified hydrogeological map (IMP: impermeable, SP: semi-permeable, HP: highly permeable).</p>
Full article ">Figure 2
<p>Conceptual structure of FCNN for filling gaps within the discharge time series.</p>
Full article ">Figure 3
<p>Karst spring discharge time series: (<b>a</b>) raw data with gaps and (<b>b</b>) post-processed data after filling gaps.</p>
Full article ">Figure 4
<p>Conceptual structure of FCNN for simulating karst spring discharge behavior.</p>
Full article ">Figure 5
<p>Final plots of loss function and simulated vs. measured discharge values for the entire dataset of the six selected springs (Phyton 3.12.3). Red dotted lines in the scatter plot defines the 90% confidence interval.</p>
Full article ">Figure 6
<p>Comparison between measured (orange) and simulated (blue) spring flowrates (in L/s) in the time series 2000–2024: (<b>a</b>) Rasiglia; (<b>b</b>) Nocera; (<b>c</b>) San Giovenale.</p>
Full article ">Figure 7
<p>Comparison between measured (orange) and simulated (blue) spring flowrates (in L/s) in the time series 2000–2024: (<b>a</b>) Lupa; (<b>b</b>) Bagnara; (<b>c</b>) Acquabianca.</p>
Full article ">Figure 7 Cont.
<p>Comparison between measured (orange) and simulated (blue) spring flowrates (in L/s) in the time series 2000–2024: (<b>a</b>) Lupa; (<b>b</b>) Bagnara; (<b>c</b>) Acquabianca.</p>
Full article ">
26 pages, 1895 KiB  
Article
Enhanced Ischemic Stroke Lesion Segmentation in MRI Using Attention U-Net with Generalized Dice Focal Loss
by Beatriz P. Garcia-Salgado, Jose A. Almaraz-Damian, Oscar Cervantes-Chavarria, Volodymyr Ponomaryov, Rogelio Reyes-Reyes, Clara Cruz-Ramos and Sergiy Sadovnychiy
Appl. Sci. 2024, 14(18), 8183; https://doi.org/10.3390/app14188183 - 11 Sep 2024
Viewed by 302
Abstract
Ischemic stroke lesion segmentation in MRI images represents significant challenges, particularly due to class imbalance between foreground and background pixels. Several approaches have been developed to achieve higher F1-Scores in stroke lesion segmentation under this challenge. These strategies include convolutional neural networks (CNN) [...] Read more.
Ischemic stroke lesion segmentation in MRI images represents significant challenges, particularly due to class imbalance between foreground and background pixels. Several approaches have been developed to achieve higher F1-Scores in stroke lesion segmentation under this challenge. These strategies include convolutional neural networks (CNN) and models that represent a large number of parameters, which can only be trained on specialized computational architectures that are explicitly oriented to data processing. This paper proposes a lightweight model based on the U-Net architecture that handles an attention module and the Generalized Dice Focal loss function to enhance the segmentation accuracy in the class imbalance environment, characteristic of stroke lesions in MRI images. This study also analyzes the segmentation performance according to the pixel size of stroke lesions, giving insights into the loss function behavior using the public ISLES 2015 and ISLES 2022 MRI datasets. The proposed model can effectively segment small stroke lesions with F1-Scores over 0.7, particularly in FLAIR, DWI, and T2 sequences. Furthermore, the model shows reasonable convergence with their 7.9 million parameters at 200 epochs, making it suitable for practical implementation on mid and high-end general-purpose graphic processing units. Full article
(This article belongs to the Special Issue Advances in Computer Vision and Semantic Segmentation, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Scheme of the proposed model.</p>
Full article ">Figure 2
<p>Distribution of segmentation masks’ sizes (in pixels) with annotation of the first quartile (Q1), median (Q2), and third quartile (Q3): (<b>a</b>) ISLES 2015, (<b>b</b>) ISLES 2022.</p>
Full article ">Figure 3
<p>F1-Scores resulting from changing key hyperparameters <math display="inline"><semantics> <mrow> <msub> <mi>λ</mi> <mrow> <mi>F</mi> <mi>L</mi> </mrow> </msub> <mo>=</mo> <mi>FL</mi> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msub> <mi>λ</mi> <mrow> <mi>G</mi> <mi>D</mi> <mi>L</mi> </mrow> </msub> <mo>=</mo> <mi>GDL</mi> </mrow> </semantics></math>. The combination leading to the best results is highlighted in orange. (<b>a</b>) Experiments performed on FLAIR sequences of ISLES 2015. (<b>b</b>) Experiments performed on DWI modality of ISLES 2022.</p>
Full article ">Figure 4
<p>Learning curves comparison: (<b>a</b>) Proposed model, (<b>b</b>) SGD, (<b>c</b>) W/O A.M., (<b>d</b>) CBAM, (<b>e</b>) Dice Loss, (<b>f</b>) Focal Loss.</p>
Full article ">Figure 5
<p>Visual comparison of the model’s versions results: Ground truth masks are displayed in the first column (<b>a</b>,<b>g</b>,<b>m</b>,<b>s</b>). Results of Proposed model are given in the second column (<b>b</b>,<b>h</b>,<b>n</b>,<b>t</b>), of Dice Loss model in the third column (<b>c</b>,<b>i</b>,<b>o</b>,<b>u</b>), of Focal Loss model in the fourth column (<b>d</b>,<b>j</b>,<b>p</b>,<b>v</b>), of W/O A.M. model in the fifth column (<b>e</b>,<b>k</b>,<b>q</b>,<b>w</b>), and of CBAM model in the sixth column (<b>f</b>,<b>l</b>,<b>r</b>,<b>x</b>).</p>
Full article ">Figure 6
<p>Violin plot of the proposed model’s results on FLAIR images using the axial view, where the dot localizes the median, and the white line represents the mean: (<b>a</b>) IoU scores by mask’s size category, (<b>b</b>) F1-Scores by mask’s size category.</p>
Full article ">Figure 7
<p>Performance of the proposed model in segmenting small lesions on different MRI modalities using the ISLES 2015 dataset (dot and white line represent the median and the mean values): (<b>a</b>) IoU scores by MRI modality, (<b>b</b>) F1-Scores by MRI modality.</p>
Full article ">Figure 8
<p>Overall performance of the proposed model on different MRI modalities using the ISLES 2015 dataset (dot and white line represent the median and the mean values): (<b>a</b>) IoU scores in the coronal plane, (<b>b</b>) IoU scores in the sagittal plane.</p>
Full article ">Figure 9
<p>Examples of segmented FLAIR images in the coronal plane by the proposed method (second row) and their corresponding ground truth mask (first row) for mask categories Small (<b>a</b>,<b>e</b>), Medium Down (<b>b</b>,<b>f</b>), Medium Up (<b>c</b>,<b>g</b>), and Large (<b>d</b>,<b>h</b>).</p>
Full article ">Figure 10
<p>Examples of segmented FLAIR images in the sagittal plane by the proposed method (second row) and their corresponding ground truth mask (first row) for mask categories Small (<b>a</b>,<b>e</b>), Medium Down (<b>b</b>,<b>f</b>), Medium Up (<b>c</b>,<b>g</b>), and Large (<b>d</b>,<b>h</b>).</p>
Full article ">Figure 11
<p>Violin plot of the proposed model’s results on DWI and ADC images using the axial view, where the dot localizes the median, and the white line represents the mean: (<b>a</b>) F1-Scores by mask size category using DWI and configuration A (FL = 0.7, GDL = 0.3), (<b>b</b>) F1-Scores by mask size category using DWI and configuration B (FL = 0.9, GDL = 0.1), (<b>c</b>) F1-Scores by mask size category using ADC and configuration A (FL = 0.7, GDL = 0.3), (<b>d</b>) F1-Scores by mask size category using ADC and configuration B (FL = 0.9, GDL = 0.1).</p>
Full article ">Figure 12
<p>Violin plot of the non-segmented images’ mask size in pixels. Mean value is marked as a white line.</p>
Full article ">Figure 13
<p>Examples of ground truth mask of DWI images in the axial plane (first row) and the segmentation results by the proposed method using <math display="inline"><semantics> <mrow> <msub> <mi>λ</mi> <mrow> <mi>G</mi> <mi>D</mi> <mi>L</mi> </mrow> </msub> <mo>=</mo> <mn>0.3</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msub> <mi>λ</mi> <mrow> <mi>F</mi> <mi>L</mi> </mrow> </msub> <mo>=</mo> <mn>0.7</mn> </mrow> </semantics></math> (second row) and <math display="inline"><semantics> <mrow> <msub> <mi>λ</mi> <mrow> <mi>G</mi> <mi>D</mi> <mi>L</mi> </mrow> </msub> <mo>=</mo> <mn>0.1</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msub> <mi>λ</mi> <mrow> <mi>F</mi> <mi>L</mi> </mrow> </msub> <mo>=</mo> <mn>0.9</mn> </mrow> </semantics></math> (third row) for mask categories Small (<b>a</b>,<b>e</b>,<b>i</b>), Medium Down (<b>b</b>,<b>f</b>,<b>j</b>), Medium Up (<b>c</b>,<b>g</b>,<b>k</b>), and Large (<b>d</b>,<b>h</b>,<b>l</b>).</p>
Full article ">
23 pages, 30954 KiB  
Article
A Deep CNN-Based Salinity and Freshwater Fish Identification and Classification Using Deep Learning and Machine Learning
by Wahidur Rahman, Mohammad Motiur Rahman, Md Ariful Islam Mozumder, Rashadul Islam Sumon, Samia Allaoua Chelloug, Rana Othman Alnashwan and Mohammed Saleh Ali Muthanna
Sustainability 2024, 16(18), 7933; https://doi.org/10.3390/su16187933 - 11 Sep 2024
Viewed by 394
Abstract
Concerning the oversight and safeguarding of aquatic environments, it is necessary to ascertain the quantity of fish, their size, and their distribution. Many deep learning (DL), artificial intelligence (AI), and machine learning (ML) techniques have been developed to oversee and safeguard the fish [...] Read more.
Concerning the oversight and safeguarding of aquatic environments, it is necessary to ascertain the quantity of fish, their size, and their distribution. Many deep learning (DL), artificial intelligence (AI), and machine learning (ML) techniques have been developed to oversee and safeguard the fish species. Still, all the previous work had some limitations, such as a limited dataset, only binary class categorization, only employing one technique (ML/DL), etc. Therefore, in the proposed work, the authors develop an architecture that will eliminate all the limitations. Both DL and ML techniques were used in the suggested framework to identify and categorize multiple classes of the salinity and freshwater fish species. Two different datasets of fish images with thirteen fish species were employed in the current research. Seven CNN architectures were implemented to find out the important features of the fish images. Then, seven ML classifiers were utilized in the suggested work to identify the binary class (freshwater and salinity) of fish species. Following that, the multiclass classification of thirteen fish species was evaluated through the ML algorithms, where the present model diagnosed the freshwater or salinity fish in the specific fish species. To achieve the primary goals of the proposed study, several assessments of the experimental data are provided. The results of the investigation indicated that DenseNet121, EfficientNetB0, ResNet50, VGG16, and VGG19 architectures of the CNN with SVC ML technique achieved 100% accuracy, F1-score, precision, and recall for binary classification (freshwater/salinity) of fish images. Additionally, the ResNet50 architecture of the CNN with SVC ML technique achieved 98.06% and 100% accuracy for multiclass classification (freshwater and salinity fish species) of fish images. However, the proposed pipeline can be very effective in sustainable fish management in fish identification and classification. Full article
(This article belongs to the Special Issue Sustainable Engineering Applications of Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Overall system illustration.</p>
Full article ">Figure 2
<p>Learning curve of binary classification.</p>
Full article ">Figure 3
<p>Confusion matrix of binary classification.</p>
Full article ">Figure 4
<p>Bar chart of the accuracy, recall, precision, and F1-score of binary classification.</p>
Full article ">Figure 4 Cont.
<p>Bar chart of the accuracy, recall, precision, and F1-score of binary classification.</p>
Full article ">Figure 5
<p>The learning curve of multiclass classification (freshwater fish).</p>
Full article ">Figure 6
<p>Confusion matrix of multiclass classification (freshwater fish).</p>
Full article ">Figure 6 Cont.
<p>Confusion matrix of multiclass classification (freshwater fish).</p>
Full article ">Figure 7
<p>ROC of multiclass classification (freshwater fish).</p>
Full article ">Figure 8
<p>Accuracy, recall, precision, and F1-score of multiclass freshwater fish classification.</p>
Full article ">Figure 9
<p>Learning curve of multiclass classification (salinity fish).</p>
Full article ">Figure 10
<p>Confusion matrix of multiclass classification (salinity fish).</p>
Full article ">Figure 11
<p>ROC of multiclass classification (salinity fish).</p>
Full article ">Figure 12
<p>Accuracy, recall, precision, and F1-score of multiclass salinity fish classification.</p>
Full article ">
21 pages, 20841 KiB  
Article
Snow Detection in Gaofen-1 Multi-Spectral Images Based on Swin-Transformer and U-Shaped Dual-Branch Encoder Structure Network with Geographic Information
by Yue Wu, Chunxiang Shi, Runping Shen, Xiang Gu, Ruian Tie, Lingling Ge and Shuai Sun
Remote Sens. 2024, 16(17), 3327; https://doi.org/10.3390/rs16173327 - 8 Sep 2024
Viewed by 456
Abstract
Snow detection is imperative in remote sensing for various applications, including climate change monitoring, water resources management, and disaster warning. Recognizing the limitations of current deep learning algorithms in cloud and snow boundary segmentation, as well as issues like detail snow information loss [...] Read more.
Snow detection is imperative in remote sensing for various applications, including climate change monitoring, water resources management, and disaster warning. Recognizing the limitations of current deep learning algorithms in cloud and snow boundary segmentation, as well as issues like detail snow information loss and mountainous snow omission, this paper presents a novel snow detection network based on Swin-Transformer and U-shaped dual-branch encoder structure with geographic information (SD-GeoSTUNet), aiming to address the above issues. Initially, the SD-GeoSTUNet incorporates the CNN branch and Swin-Transformer branch to extract features in parallel and the Feature Aggregation Module (FAM) is designed to facilitate the detail feature aggregation via two branches. Simultaneously, an Edge-enhanced Convolution (EeConv) is introduced to promote snow boundary contour extraction in the CNN branch. In particular, auxiliary geographic information, including altitude, longitude, latitude, slope, and aspect, is encoded in the Swin-Transformer branch to enhance snow detection in mountainous regions. Experiments conducted on Levir_CS, a large-scale cloud and snow dataset originating from Gaofen-1, demonstrate that SD-GeoSTUNet achieves optimal performance with the values of 78.08%, 85.07%, and 92.89% for IoU_s, F1_s, and MPA, respectively, leading to superior cloud and snow boundary segmentation and thin cloud and snow detection. Further, ablation experiments reveal that integrating slope and aspect information effectively alleviates the omission of snow detection in mountainous areas and significantly exhibits the best vision under complex terrain. The proposed model can be used for remote sensing data with geographic information to achieve more accurate snow extraction, which is conducive to promoting the research of hydrology and agriculture with different geospatial characteristics. Full article
(This article belongs to the Section Environmental Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Overview of the proposed SD-GeoSTUNet. Hereinafter referred to as CNN branch and Swin-T branch, respectively. (<b>a</b>) Overview of Decoder Layer. (<b>b</b>) Overview of Feature Aggregation Module (FAM). (<b>c</b>) Overview of Residual Layer.</p>
Full article ">Figure 2
<p>(<b>a</b>) Overview of Edge-enhanced Convolution (EeConv), including a vanilla convolution, a central difference convolution (CDC), an angular difference convolution (ADC), a horizontal difference convolution (HDC), and a vertical difference convolution (VDC). (<b>b</b>) the principle of vertical difference convolution (VDC).</p>
Full article ">Figure 3
<p>The global distribution of the images in the experiment. Base map from Cartopy.</p>
Full article ">Figure 4
<p>Segmentation results of cloud and snow under different module combinations. (<b>a</b>) RGB true color image, (<b>b</b>) Label, (<b>c</b>) Concatenation, (<b>d</b>) FAM, (<b>e</b>) FAM + EeConv. In (<b>b</b>–<b>e</b>), black, blue, and white pixels represent background, cloud, and snow, respectively. The scene of the first row is at the center of 109.6°E, 48.8°N, and the date is 17 October 2017. The scene of the second row is at the center of 86.6°E, 28.5°N, and the date is 30 November 2016.</p>
Full article ">Figure 5
<p>Detection results of snow in mountain regions under different geographic information feature combination ablation experiments, and enlarged views marked with blue, green, and red boxes. (<b>a</b>) RGB true color image. (<b>b</b>) Label. (<b>c</b>) Experiment 1 detection results. (<b>d</b>) Experiment 2 detection results. (<b>e</b>) Experiment 3 detection results. In (<b>b</b>–<b>e</b>), black and white pixels represent the background and snow, respectively. The scene is at the center of 128.9°E, 44.7°N, and the date is 27 January 2018.</p>
Full article ">Figure 6
<p>Except for the background, the detection results of clouds and snow coexistence and the enlarged views (marked with the red box in the figure). (<b>a</b>) RGB true color image. (<b>b</b>) Label. (<b>c</b>) PSPNet. (<b>d</b>) Segformer. (<b>e</b>) U-Net. (<b>f</b>) CDNetV2. (<b>g</b>) GeoInfoNet. (<b>h</b>) SD-GeoSTUNet. In (<b>b</b>–<b>h</b>), black, blue, and white pixels represent the background, cloud, and snow, respectively. The scene of the first row is at the center of 104.2°E, 31.3°N, and the date is 29 March 2018. The scene of the third row is at the center of 104.6°E, 33.0°N, and the date is 29 March 2018.</p>
Full article ">Figure 7
<p>Except for the background, the detection results of pure snow and the enlarged views (marked with the red box in the figure). (<b>a</b>) RGB true color image. (<b>b</b>) Label. (<b>c</b>) PSPNet. (<b>d</b>) Segformer. (<b>e</b>) U-Net. (<b>f</b>) CDNetV2. (<b>g</b>) GeoInfoNet. (<b>h</b>) SD-GeoSTUNet. In (<b>b</b>–<b>h</b>), black, blue, and white pixels represent the background, cloud, and snow, respectively. The scene of the first row is at the center of 128.6°E, 51.3°N, and the date is 4 January 2016. The scene of the third row is at the center of 133.7°E, 50.9°N, and the date is 4 February 2018.</p>
Full article ">Figure 8
<p>Except for the background, the detection results of pure cloud and the enlarged views (marked with the red box in the figure). (<b>a</b>) RGB true color image. (<b>b</b>) Label. (<b>c</b>) PSPNet. (<b>d</b>) Segformer. (<b>e</b>) U-Net. (<b>f</b>) CDNetV2. (<b>g</b>) GeoInfoNet. (<b>h</b>) SD-GeoSTUNet. In (<b>b</b>–<b>h</b>), black, blue, and white pixels represent the background, cloud, and snow, respectively. This scene is at the center of 4.1°W, 54.2°N, and the date is 13 July 2016.</p>
Full article ">
15 pages, 4574 KiB  
Article
Student Behavior Recognition in Classroom Based on Deep Learning
by Qingzheng Jia and Jialiang He
Appl. Sci. 2024, 14(17), 7981; https://doi.org/10.3390/app14177981 - 6 Sep 2024
Viewed by 427
Abstract
With the widespread application of information technology in education, the real-time detection of student behavior in the classroom has become a key issue in improving teaching quality. This paper proposes a Student Behavior Detection (SBD) model that combines YOLOv5, the Contextual Attention (CA) [...] Read more.
With the widespread application of information technology in education, the real-time detection of student behavior in the classroom has become a key issue in improving teaching quality. This paper proposes a Student Behavior Detection (SBD) model that combines YOLOv5, the Contextual Attention (CA) mechanism and OpenPose, aiming to achieve efficient and accurate behavior recognition in complex classroom environments. By integrating YOLOv5 with the CA attention mechanism to enhance feature extraction capabilities, the model’s recognition performance in complex backgrounds, such as those with occlusion, is significantly improved. In addition, the feature map generated by the improved YOLOv5 is used to replace VGG-19 in OpenPose, which effectively improves the accuracy of student posture recognition. The experimental results demonstrate that the proposed model achieves a maximum mAP of 82.1% in complex classroom environments, surpassing Faster R-CNN by 5.2 percentage points and YOLOv5 by 4.6 percentage points. Additionally, the F1 score and R value of this model exhibit clear advantages over the other two traditional methods. This model offers an effective solution for intelligent classroom behavior analysis and the optimization of educational management. Full article
(This article belongs to the Special Issue Intelligent Techniques, Platforms and Applications of E-learning)
Show Figures

Figure 1

Figure 1
<p>Detection model diagram.</p>
Full article ">Figure 2
<p>Framework of Student Behavior Detection.</p>
Full article ">Figure 3
<p>YoLov5 network structure.</p>
Full article ">Figure 4
<p>CA attention mechanism.</p>
Full article ">Figure 5
<p>YOLOv5-A network structure.</p>
Full article ">Figure 6
<p>OpenPose network structure.</p>
Full article ">Figure 7
<p>Detection flow chart.</p>
Full article ">Figure 8
<p>Precision curve of YOLOv5 model.</p>
Full article ">Figure 9
<p>Precision curve of SBD model.</p>
Full article ">Figure 10
<p>PR curve of SBD model.</p>
Full article ">Figure 11
<p>Actual detection results of SBD.</p>
Full article ">
25 pages, 10917 KiB  
Article
Promoting Sustainable Development of Coal Mines: CNN Model Optimization for Identification of Microseismic Signals Induced by Hydraulic Fracturing in Coal Seams
by Nan Li, Yunpeng Zhang, Xiaosong Zhou, Lihong Sun, Xiaokai Huang, Jincheng Qiu, Yan Li and Xiaoran Wang
Sustainability 2024, 16(17), 7592; https://doi.org/10.3390/su16177592 - 2 Sep 2024
Viewed by 521
Abstract
Borehole hydraulic fracturing in coal seams can prevent dynamic coal mine disasters and promote the sustainability of the mining industry, and microseismic signal recognition is a prerequisite and foundation for microseismic monitoring technology that evaluates the effectiveness of hydraulic fracturing. This study constructed [...] Read more.
Borehole hydraulic fracturing in coal seams can prevent dynamic coal mine disasters and promote the sustainability of the mining industry, and microseismic signal recognition is a prerequisite and foundation for microseismic monitoring technology that evaluates the effectiveness of hydraulic fracturing. This study constructed ultra-lightweight CNN models specifically designed to identify microseismic waveforms induced by borehole hydraulic fracturing in coal seams, namely Ul-Inception28, Ul-ResNet12, Ul-MobileNet17, and Ul-TripleConv8. The three best-performing models were selected to create both a probability averaging ensemble CNN model and a voting ensemble CNN model. Additionally, an automatic threshold adjustment strategy for CNN identification was introduced. The relationships between feature map entropy, training data volume, and model performance were also analyzed. The results indicated that our in-house models surpassed the performance of the InceptionV3, ResNet50, and MobileNetV3 models from the TensorFlow Keras library. Notably, the voting ensemble CNN model achieved an improvement of at least 0.0452 in the F1 score compared to individual models. The automatic threshold adjustment strategy enhanced the identification threshold’s precision to 26 decimal places. However, a continuous zero-entropy value in the feature maps of various channels was found to detract from the model’s generalization performance. Moreover, the expanded training dataset, derived from thousands of waveforms, proved more compatible with CNN models comprising hundreds of thousands of parameters. The findings of this research significantly contribute to the prevention of dynamic coal mine disasters, potentially reducing casualties, economic losses, and promoting the sustainable progress of the coal mining industry. Full article
(This article belongs to the Section Hazards and Sustainability)
Show Figures

Figure 1

Figure 1
<p>Schematic diagram of different convolution methods.</p>
Full article ">Figure 2
<p>Structure of the Ul-Inception28 model.</p>
Full article ">Figure 3
<p>Structure of the Ul-ResNet-12 model.</p>
Full article ">Figure 4
<p>Structure of the Ul-MobileNet-17 model.</p>
Full article ">Figure 5
<p>Structure of the Ul-TripleConv8 model.</p>
Full article ">Figure 6
<p>Flowchart of automatic adjustment strategy for identification threshold.</p>
Full article ">Figure 7
<p>Flowchart of microseismic waveform identification by probability-averaged CNN model.</p>
Full article ">Figure 8
<p>Flowchart of microseismic waveform identification by voting ensemble CNN Model.</p>
Full article ">Figure 9
<p>Training and testing accuracy and loss functions of different CNN models.</p>
Full article ">Figure 10
<p>Ul-Mobilenet automatic threshold adjustment identification process.</p>
Full article ">Figure 11
<p>Examples of identified event results. (<b>a</b>) Microseismic event 1; (<b>b</b>) microseismic event 2.</p>
Full article ">Figure 12
<p>Time-domain images of three microseismic waveforms. (<b>a</b>) Background noise; (<b>b</b>) weak microseismic waveform; (<b>c</b>) high signal-to-noise ratio microseismic waveform.</p>
Full article ">Figure 13
<p>Feature maps of the first channel in the first convolutional layer of Ul-Inception28 for different images. (<b>a</b>) Background noise; (<b>b</b>) high signal-to-noise ratio microseismic waveform.</p>
Full article ">Figure 14
<p>Feature maps of the first channel in different convolutional layers of Ul-Inception28. (<b>a</b>) The 1st convolutional layer; (<b>b</b>) the 10th convolutional layer; (<b>c</b>) the 19th convolutional layer; (<b>d</b>) the 28th convolutional layer.</p>
Full article ">Figure 15
<p>Feature maps of the first channel in different convolutional layers of Ul-ResNet12. (<b>a</b>) The 1st convolutional layer; (<b>b</b>) the 5th convolutional layer; (<b>c</b>) the 9th convolutional layer; (<b>d</b>) the 12th convolutional layer.</p>
Full article ">Figure 16
<p>Feature maps of the first channel in different convolutional layers of Ul-MobileNet17. (<b>a</b>) The 1st convolutional layer; (<b>b</b>) the 6th convolutional layer; (<b>c</b>) the 11th convolutional layer; (<b>d</b>) the 17th convolutional layer.</p>
Full article ">Figure 16 Cont.
<p>Feature maps of the first channel in different convolutional layers of Ul-MobileNet17. (<b>a</b>) The 1st convolutional layer; (<b>b</b>) the 6th convolutional layer; (<b>c</b>) the 11th convolutional layer; (<b>d</b>) the 17th convolutional layer.</p>
Full article ">Figure 17
<p>Feature maps of the first channel in different convolutional layers of Ul-TripleConv8. (<b>a</b>) The first convolutional layer; (<b>b</b>) the third convolutional layer; (<b>c</b>) the fifth convolutional layer; (<b>d</b>) the eighth convolutional layer.</p>
Full article ">Figure 18
<p>Feature maps of different channels in the last convolutional layer of UI-mobilenet8. (<b>a</b>) The 64th channel; (<b>b</b>) the 128th channel; (<b>c</b>) the 192nd channel; (<b>d</b>) the 256th channel.</p>
Full article ">Figure 19
<p>Feature maps of different channels in the last convolutional layer of MobilenetV3. (<b>a</b>) The 64th channel; (<b>b</b>) the 128th channel; (<b>c</b>) the 192nd channel; (<b>d</b>) the 256th channel.</p>
Full article ">Figure 20
<p>Entropy values of feature maps from all channels of all convolutional layers across different models. (<b>a</b>) InceptionV3; (<b>b</b>) ResNet50; (<b>c</b>) MobileNetV3; (<b>d</b>) UI-TripleConv8; (<b>e</b>) UI-ResNet12; (<b>f</b>) UI-MobileNet17; (<b>g</b>) UI-Inception18.</p>
Full article ">Figure 20 Cont.
<p>Entropy values of feature maps from all channels of all convolutional layers across different models. (<b>a</b>) InceptionV3; (<b>b</b>) ResNet50; (<b>c</b>) MobileNetV3; (<b>d</b>) UI-TripleConv8; (<b>e</b>) UI-ResNet12; (<b>f</b>) UI-MobileNet17; (<b>g</b>) UI-Inception18.</p>
Full article ">Figure 20 Cont.
<p>Entropy values of feature maps from all channels of all convolutional layers across different models. (<b>a</b>) InceptionV3; (<b>b</b>) ResNet50; (<b>c</b>) MobileNetV3; (<b>d</b>) UI-TripleConv8; (<b>e</b>) UI-ResNet12; (<b>f</b>) UI-MobileNet17; (<b>g</b>) UI-Inception18.</p>
Full article ">Figure 21
<p>F1 scores of microseismic waveform recognition for different models trained on various datasets.</p>
Full article ">
25 pages, 12480 KiB  
Article
EFS-Former: An Efficient Network for Fruit Tree Leaf Disease Segmentation and Severity Assessment
by Donghui Jiang, Miao Sun, Shulong Li, Zhicheng Yang and Liying Cao
Agronomy 2024, 14(9), 1992; https://doi.org/10.3390/agronomy14091992 - 2 Sep 2024
Viewed by 466
Abstract
Fruit is a major source of vitamins, minerals, and dietary fiber in people’s daily lives. Leaf diseases caused by climate change and other factors have significantly reduced fruit production. Deep learning methods for segmenting leaf diseases can effectively mitigate this issue. However, challenges [...] Read more.
Fruit is a major source of vitamins, minerals, and dietary fiber in people’s daily lives. Leaf diseases caused by climate change and other factors have significantly reduced fruit production. Deep learning methods for segmenting leaf diseases can effectively mitigate this issue. However, challenges such as leaf folding, jaggedness, and light shading make edge feature extraction difficult, affecting segmentation accuracy. To address these problems, this paper proposes a method based on EFS-Former. The expanded local detail (ELD) module extends the model’s receptive field by expanding the convolution, better handling fine spots and effectively reducing information loss. H-attention reduces computational redundancy by superimposing multi-layer convolutions, significantly improving feature filtering. The parallel fusion architecture effectively utilizes the different feature extraction intervals of the convolutional neural network (CNN) and Transformer encoders, achieving comprehensive feature extraction and effectively fusing detailed and semantic information in the channel and spatial dimensions within the feature fusion module (FFM). Experiments show that, compared to DeepLabV3+, this method achieves 10.78%, 9.51%, 0.72%, and 8.00% higher scores for mean intersection over union (mIoU), mean pixel accuracy (mPA), accuracy (Acc), and F_score, respectively, while having 1.78 M fewer total parameters and 0.32 G lower floating point operations per second (FLOPS). Additionally, it effectively calculates the ratio of leaf area occupied by spots. This method is also effective in calculating the disease period by analyzing the ratio of leaf area occupied by diseased spots. The method’s overall performance is evaluated using mIoU, mPA, Acc, and F_score metrics, achieving 88.60%, 93.49%, 98.60%, and 95.90%, respectively. In summary, this study offers an efficient and accurate method for fruit tree leaf spot segmentation, providing a solid foundation for the precise analysis of fruit tree leaves and spots, and supporting smart agriculture for precision pesticide spraying. Full article
(This article belongs to the Special Issue The Applications of Deep Learning in Smart Agriculture)
Show Figures

Figure 1

Figure 1
<p>Original and annotated images: (<b>a</b>) apple spotted leaf drop disease, (<b>b</b>) early stage of grape brown spot disease, (<b>c</b>) late stage of grape brown spot disease, (<b>d</b>) early stage of grape black rot disease, (<b>e</b>) late stage of grape black rot disease, and (<b>f</b>) pomegranate cercospora spot.</p>
Full article ">Figure 2
<p>Original image and enhanced image: (<b>a</b>) original image, (<b>b</b>) reduced brightness and flipped, (<b>c</b>) randomly flipped, (<b>d</b>) random zoom, (<b>e</b>) white mask block added, (<b>f</b>) panning and increasing brightness.</p>
Full article ">Figure 3
<p>The overall architecture of EFS-Former, including the main structure of parallel fusion, CNN encoder, feature fusion module, and improved Transformer encoder.</p>
Full article ">Figure 4
<p>ELD module architecture diagram.</p>
Full article ">Figure 5
<p>Seg-Block architecture diagram.</p>
Full article ">Figure 6
<p>Overall architecture of FFM.</p>
Full article ">Figure 7
<p>Images of results generated in different models for four disease conditions of three fruits: (<b>a</b>) apple spotted leaf drop disease, (<b>b</b>) grape brown spot disease, (<b>c</b>) grape black rot disease, and (<b>d</b>) pomegranate cercospora spot.</p>
Full article ">Figure 8
<p>Visualization of segmentation results for different methods. (<b>a</b>) Apple spotted leaf drop disease, (<b>b</b>,<b>c</b>) late and early stages of grape black rot disease, (<b>d</b>,<b>e</b>) early and late stages of grape brown spot disease, and (<b>f</b>) pomegranate cercospora spot.</p>
Full article ">Figure 9
<p>Effective attention results of different models for indoor and outdoor leaves. (<b>a</b>) Grape brown spot disease, (<b>b</b>) apple spotted leaf drop disease, and (<b>c</b>) pomegranate cercospora spot.</p>
Full article ">Figure 10
<p>Results of different models for effective attention to leaf spots indoors and outdoors. (<b>a</b>) Grape brown spot disease, (<b>b</b>) apple spotted leaf drop disease, and (<b>c</b>) pomegranate cercospora spot.</p>
Full article ">
19 pages, 2702 KiB  
Article
Modeling and Forecasting Ionospheric foF2 Variation Based on CNN-BiLSTM-TPA during Low- and High-Solar Activity Years
by Baoyi Xu, Wenqiang Huang, Peng Ren, Yi Li and Zheng Xiang
Remote Sens. 2024, 16(17), 3249; https://doi.org/10.3390/rs16173249 - 2 Sep 2024
Viewed by 396
Abstract
The transmission of high-frequency signals over long distances depends on the ionosphere’s reflective properties, with the selection of operating frequencies being closely tied to variations in the ionosphere. The accurate prediction of ionospheric critical frequency foF2 and other parameters in low latitudes is [...] Read more.
The transmission of high-frequency signals over long distances depends on the ionosphere’s reflective properties, with the selection of operating frequencies being closely tied to variations in the ionosphere. The accurate prediction of ionospheric critical frequency foF2 and other parameters in low latitudes is of great significance for understanding ionospheric changes in high-frequency communications. Currently, deep learning algorithms demonstrate significant advantages in capturing characteristics of the ionosphere. In this paper, a state-of-the-art hybrid neural network is utilized in conjunction with a temporal pattern attention mechanism for predicting variations in the foF2 parameter during high- and low-solar activity years. Convolutional neural networks (CNNs) and bidirectional long short-term memory (BiLSTM), which is capable of extracting spatiotemporal features of ionospheric variations, are incorporated into a hybrid neural network. The foF2 data used for training and testing come from three observatories in Brisbane (27°53′S, 152°92′E), Darwin (12°45′S, 130°95′E) and Townsville (19°63′S, 146°85′E) in 2000, 2008, 2009 and 2014 (the peak or trough years of solar activity in solar cycles 23 and 24), using the advanced Australian Digital Ionospheric Sounder. The results show that the proposed model accurately captures the changes in ionospheric foF2 characteristics and outperforms International Reference Ionosphere 2020 (IRI-2020) and BiLSTM ionospheric prediction models. Full article
(This article belongs to the Special Issue Ionosphere Monitoring with Remote Sensing (3rd Edition))
Show Figures

Figure 1

Figure 1
<p>In 2009 and 2014, time series of hourly data: (<b>a</b>,<b>b</b>) geomagnetic index, Dst; (<b>c</b>,<b>d</b>) geomagnetic index, Ap; (<b>e</b>,<b>f</b>) solar activity index, F10.7; (<b>g</b>,<b>h</b>) geomagnetic index, IMF BZ; (<b>i</b>,<b>j</b>) solar activity index, SSN.</p>
Full article ">Figure 1 Cont.
<p>In 2009 and 2014, time series of hourly data: (<b>a</b>,<b>b</b>) geomagnetic index, Dst; (<b>c</b>,<b>d</b>) geomagnetic index, Ap; (<b>e</b>,<b>f</b>) solar activity index, F10.7; (<b>g</b>,<b>h</b>) geomagnetic index, IMF BZ; (<b>i</b>,<b>j</b>) solar activity index, SSN.</p>
Full article ">Figure 2
<p>Temporal pattern attention mechanism.</p>
Full article ">Figure 3
<p>Structure diagram of hybrid neural network.</p>
Full article ">Figure 4
<p>(<b>a</b>–<b>l</b>) Comparison of these ionospheric prediction models’ performance on test samples in 2009.</p>
Full article ">Figure 5
<p>(<b>a</b>–<b>l</b>) Comparison of these ionospheric prediction models’ performance on test samples in 2014.</p>
Full article ">Figure 6
<p>For 2008 and 2009, the model’s cumulative distributions of sample errors at Brisbane: (<b>a</b>,<b>d</b>) IRI-2020; (<b>b</b>,<b>e</b>) BiLSTM-foF2; (<b>c</b>,<b>f</b>) proposed model.</p>
Full article ">Figure 7
<p>For 2008 and 2009, the model’s cumulative distributions of sample errors at Darwin: (<b>a</b>,<b>d</b>) IRI-2020; (<b>b</b>,<b>e</b>) BiLSTM-foF2; (<b>c</b>,<b>f</b>) proposed model.</p>
Full article ">Figure 8
<p>For 2008 and 2009, the model’s cumulative distributions of sample errors at Townsvile: (<b>a</b>,<b>d</b>) IRI-2020; (<b>b</b>,<b>e</b>) BiLSTM-foF2; (<b>c</b>,<b>f</b>) proposed model.</p>
Full article ">Figure 9
<p>For 2000 and 2014, histograms and the curve of the normal distribution for prediction errors at Brisbane: (<b>a</b>,<b>d</b>) IRI-2020; (<b>b</b>,<b>e</b>) BiLSTM-foF2; (<b>c</b>,<b>f</b>) proposed model.</p>
Full article ">Figure 10
<p>For 2000 and 2014, histograms and the curve of the normal distribution for prediction errors at Darwin: (<b>a</b>,<b>d</b>) IRI-2020; (<b>b</b>,<b>e</b>) BiLSTM-foF2; (<b>c</b>,<b>f</b>) proposed model.</p>
Full article ">Figure 11
<p>For 2000 and 2014, histograms and the curve of the normal distribution for prediction errors at Townsvile: (<b>a</b>,<b>d</b>) IRI-2020; (<b>b</b>,<b>e</b>) BiLSTM-foF2; (<b>c</b>,<b>f</b>) proposed model.</p>
Full article ">
37 pages, 76788 KiB  
Article
Machine Learning-Based Remote Sensing Inversion of Non-Photosynthetic/Photosynthetic Vegetation Coverage in Desertified Areas and Its Response to Drought Analysis
by Zichen Guo, Shulin Liu, Kun Feng, Wenping Kang and Xiang Chen
Remote Sens. 2024, 16(17), 3226; https://doi.org/10.3390/rs16173226 - 31 Aug 2024
Viewed by 516
Abstract
Determining the responses of non-photosynthetic vegetation (NPV) and photosynthetic vegetation (PV) communities to climate change is crucial in illustrating the sensitivity and sustainability of these ecosystems. In this study, we evaluated the accuracy of inverting NPV and PV using Landsat imagery with random [...] Read more.
Determining the responses of non-photosynthetic vegetation (NPV) and photosynthetic vegetation (PV) communities to climate change is crucial in illustrating the sensitivity and sustainability of these ecosystems. In this study, we evaluated the accuracy of inverting NPV and PV using Landsat imagery with random forest (RF), backpropagation neural network (BPNN), and fully connected neural network (FCNN) models. Additionally, we inverted MODIS NPV and PV time-series data using spectral unmixing. Based on this, we analyzed the responses of NPV and PV to precipitation and drought across different ecological regions. The main conclusions are as follows: (1) In NPV remote sensing inversion, the softmax activation function demonstrates greater advantages over the ReLU activation function. Specifically, the use of the softmax function results in an approximate increase of 0.35 in the R2 value. (2) Compared with a five-layer FCNN with 128 neurons and a three-layer BPNN with 12 neurons, a random forest model with over 50 trees and 5 leaf nodes provides better inversion results for NPV and PV (R2_RF-NPV = 0.843, R2_RF-PV = 0.861). (3) Long-term drought or heavy rainfall events can affect the utilization of precipitation by NPV and PV. There is a high correlation between extreme precipitation events following prolonged drought and an increase in PV coverage. (4) Under long-term drought conditions, the vegetation in the study area responded to precipitation during the last winter and growing season. This study provides an illustration of the response of semi-arid ecosystems to drought and wetting events, thereby offering a data basis for the effect evaluation of afforestation projects. Full article
(This article belongs to the Special Issue Machine Learning for Spatiotemporal Remote Sensing Data (2nd Edition))
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Desertification types in the study area; (<b>b</b>) annual mean monthly precipitation in the study area; (<b>c</b>) location of the study area in a semi-arid region of China.</p>
Full article ">Figure 2
<p>Technical workflow diagram of this study.</p>
Full article ">Figure 3
<p>Box plot of non-photosynthetic vegetation coverage of different desertification types and degrees.</p>
Full article ">Figure 4
<p>Box plot of photosynthetic vegetation coverage of different desertification types and degrees.</p>
Full article ">Figure 5
<p>(<b>a</b>) Response of non-photosynthetic vegetation to annual precipitation. (<b>b</b>) Response of photosynthetic vegetation to annual precipitation. (<b>c</b>) Response of non-photosynthetic vegetation to annual mean temperature. (<b>d</b>) Response of photosynthetic vegetation to annual mean temperature.</p>
Full article ">Figure 6
<p>(<b>a</b>) Time-Lagged Response of NPV and PV to Monthly Precipitation on mobile dune desertification during Dry Years. (<b>b</b>) Time-Lagged Response of NPV and PV to Monthly Precipitation on coppice dune desertification during Dry Years. (<b>c</b>) Time-Lagged Response of NPV and PV to Monthly Precipitation on Gobi desertification during Dry Years. (<b>d</b>) Time-Lagged Response of NPV and PV to Monthly Precipitation on mobile dune desertification during Wet Years. (<b>e</b>) Time-Lagged Response of NPV and PV to Monthly Precipitation on coppice dune desertification Wet Years. (<b>f</b>) Time-Lagged Response of NPV and PV to Monthly Precipitation on Gobi desertification Wet Years. Figure note: MBD represents mobile dune desertification; CD represents coppice dune desertification; GD represents Gobi desertification; MD represents mild desertification; MOD represents moderate desertification; and SD represents severe desertification. K is the response degree of NPV and PV to precipitation in the desertification type and degree region.</p>
Full article ">Figure 7
<p>(<b>a</b>) Time-delay correlation (R<sup>2</sup>) of NPV and PV to Monthly Precipitation on mobile dune desertification during Dry Years. (<b>b</b>) Time-delay correlation (R<sup>2</sup>) of NPV and PV to Monthly Precipitation on coppice dune desertification during Dry Years. (<b>c</b>) Time-delay correlation (R<sup>2</sup>) of NPV and PV to Monthly Precipitation on Gobi desertification during Dry Years. (<b>d</b>) Time-delay correlation (R<sup>2</sup>) of NPV and PV to Monthly Precipitation on mobile dune desertification during Wet Years. (<b>e</b>) Time-delay correlation (R<sup>2</sup>) of NPV and PV to Monthly Precipitation on coppice dune desertification Wet Years. (<b>f</b>) Time-delay correlation (R<sup>2</sup>) of NPV and PV to Monthly Precipitation on Gobi desertifi-cation Wet Years. Figure note: MBD represents mobile dune desertification; CD represents coppice dune desertification; GD represents Gobi desertification; MD represents mild desertification; MOD represents moderate desertification; and SD represents severe desertification. R<sup>2</sup> is the correlation.</p>
Full article ">Figure 8
<p>(<b>a</b>) Response of NPV and PV to SPEI on mobile dune desertification in the Year 2000. (<b>b</b>) Response of NPV and PV to SPEI on coppice dune desertification in the Year 2000. (<b>c</b>) Response of NPV and PV to SPEI on Gobi desertification in the Year 2000. (<b>d</b>) Response of NPV and PV to SPEI on mobile dune desertification in the Year 2005. (<b>e</b>) Response of NPV and PV to SPEI on coppice dune desertification in the Year 2005. (<b>f</b>) Response of NPV and PV to SPEI on Gobi desertification in the Year 2005. (<b>g</b>) Response of NPV and PV to SPEI on mobile dune desertification in the Year 2010. (<b>h</b>) Response of NPV and PV to SPEI on coppice dune desertification in the Year 2010. (<b>i</b>) Response of NPV and PV to SPEI on Gobi desertification in the Year 2010. (<b>j</b>) Response of NPV and PV to SPEI on mobile dune desertification in the Year 2015. (<b>k</b>) Response of NPV and PV to SPEI on coppice dune desertification in the Year 2015. (<b>l</b>) Response of NPV and PV to SPEI on Gobi desertification in the Year 2015. Figure note: MBD represents mobile dune desertification; CD represents coppice dune desertification; GD represents Gobi desertification; MD represents mild desertification; MOD represents moderate desertification; and SD represents severe desertification. K is the response degree of NPV and PV to precipitation in the desertification type and degree region.</p>
Full article ">Figure 9
<p>(<b>a</b>) Correlation of NPV and PV to SPEI on mobile dune desertification in the Year 2000. (<b>b</b>) Correlation of NPV and PV to SPEI on coppice dune desertification in the Year 2000. (<b>c</b>) Correlation of NPV and PV to SPEI on Gobi desertification in the Year 2000. (<b>d</b>) Correlation of NPV and PV to SPEI on mobile dune desertification in the Year 2005. (<b>e</b>) Correlation of NPV and PV to SPEI on coppice dune desertification in the Year 2005. (<b>f</b>) Correlation of NPV and PV to SPEI on Gobi desertification in the Year 2005. (<b>g</b>) Correlation of NPV and PV to SPEI on mobile dune desertification in the Year 2010. (<b>h</b>) Correlation of NPV and PV to SPEI on coppice dune desertification in the Year 2010. (<b>i</b>) Correlation of NPV and PV to SPEI on Gobi desertification in the Year 2010. (<b>j</b>) Correlation of NPV and PV to SPEI on mobile dune desertification in the Year 2015. (<b>k</b>) Response of NPV and PV to SPEI on coppice dune desertification in the Year 2015. (<b>l</b>) Response of NPV and PV to SPEI on Gobi desertification in the Year 2015. Figure note: MBD represents mobile dune desertification; CD represents coppice dune desertification; GD represents Gobi desertification; MD represents mild desertification; MOD represents moderate desertification; and SD represents severe desertification. R<sup>2</sup> is the correlation.</p>
Full article ">Figure 10
<p>(<b>a</b>) Local NPV cover of MODIS in 2019; (<b>b</b>) SWIR67 Index in 2019; (<b>c</b>) Local NPV of MODIS in 2018.</p>
Full article ">Figure 11
<p>Spectra of photosynthetic and non-photosynthetic components of major vegetation types and major soil types in the study area.</p>
Full article ">Figure A1
<p>(<b>a</b>) Process of parameter selection for RF models. (<b>b</b>) Error distribution of random forest models. (<b>c</b>) Process of parameter selection for BPNN. (<b>d</b>) Process of parameter selection for FCNN.</p>
Full article ">Figure A2
<p>(<b>a</b>) The regression relationship between Landsat 8 B<sub>2</sub> and Landsat 5 B<sub>1</sub>. (<b>b</b>) The regression relationship between Landsat 8 B<sub>3</sub> and Landsat 5 B<sub>2</sub>. (<b>c</b>) The regression relationship between Landsat 8 B<sub>4</sub> and Landsat 5 B<sub>3</sub>. (<b>d</b>) The regression relationship between Landsat 8 B<sub>5</sub> and Landsat 5 B<sub>4</sub>. (<b>e</b>) The regression relationship between Landsat 8 B<sub>6</sub> and Landsat 5 B<sub>5</sub>. (<b>f</b>) The regression relationship between Landsat 8 B<sub>7</sub> and Landsat 5 B<sub>6</sub>.</p>
Full article ">Figure A3
<p>(<b>a</b>) Non-photosynthetic vegetation coverage at the end of the growing season in 2000. (<b>b</b>) Photosynthetic vegetation coverage at the end of the growing season in 2000.</p>
Full article ">Figure A4
<p>(<b>a</b>) Non-photosynthetic vegetation coverage at the end of the growing season in 2005. (<b>b</b>) Photosynthetic vegetation coverage at the end of the growing season in 2005.</p>
Full article ">Figure A5
<p>(<b>a</b>) Non-photosynthetic vegetation coverage at the end of the growing season in 2010. (<b>b</b>) Photosynthetic vegetation coverage at the end of the growing season in 2010.</p>
Full article ">Figure A6
<p>(<b>a</b>) Non-photosynthetic vegetation coverage at the end of the growing season in 2015. (<b>b</b>) Photosynthetic vegetation coverage at the end of the growing season in 2015.</p>
Full article ">Figure A7
<p>(<b>a</b>) Response degree of non-photosynthetic vegetation at the end of growing season to total precipitation in 1–9 months. (<b>b</b>) Response degree of non-photosynthetic vegetation at the end of growing season to total precipitation in 10–21 months. (<b>c</b>) Response degree of non-photosynthetic vegetation at the end of growing season to total precipitation in 22–33 months. (<b>d</b>) Response degree of non-photosynthetic vegetation at the end of growing season to total precipitation in 34–45 months. (<b>e</b>) Response degree of non-photosynthetic vegetation at the end of growing season to total precipitation in 46–57 months.</p>
Full article ">Figure A8
<p>(<b>a</b>) Response degree of photosynthetic vegetation at the end of growing season to total precipitation in 0–9 months. (<b>b</b>) Response degree of photosynthetic vegetation at the end of growing season to total precipitation in 9–21 months. (<b>c</b>) Response degree of photosynthetic vegetation at the end of growing season to total precipitation in 21–33 months. (<b>d</b>) Response degree of photosynthetic vegetation at the end of growing season to total precipitation in 33–45 months. (<b>e</b>) Response degree of photosynthetic vegetation at the end of growing season to total precipitation in 45–57 months.</p>
Full article ">Figure A9
<p>(<b>a</b>) Response of non-photosynthetic vegetation to 0–9 month mean temperature at the end of growing season. (<b>b</b>) Response of non-photosynthetic vegetation to 9–21 month mean temperature at the end of growing season. (<b>c</b>) Response of non-photosynthetic vegetation to 21–33 month mean temperature at the end of growing season. (<b>d</b>) Response of non-photosynthetic vegetation to 33–45 month mean temperature at the end of growing season. (<b>e</b>) Response of non-photosynthetic vegetation to 45–57 month mean temperature at the end of growing season.</p>
Full article ">Figure A10
<p>(<b>a</b>) Response of photosynthetic vegetation to 0–9 month mean temperature at the end of growing season. (<b>b</b>) Response of photosynthetic vegetation to 9–21 month mean temperature at the end of growing season. (<b>c</b>) Response of photosynthetic vegetation to 21–33 month mean temperature at the end of growing season. (<b>d</b>) Response of photosynthetic vegetation to 33–45 month mean temperature at the end of growing season. (<b>e</b>) Response of photosynthetic vegetation to 45–57 month mean temperature at the end of growing season.</p>
Full article ">Figure A11
<p>Time-delay responses of non-photosynthetic and photosynthetic vegetation coverage to monthly precipitation in different desertification types and degrees. Figure note: MBD represents mobile dune desertification; CD represents coppice dune desertification; GD represents Gobi desertification; MD represents mild desertification; MOD represents moderate desertification; and SD represents severe desertification. K is the response degree of NPV and PV to precipitation in the desertification type and degree region.</p>
Full article ">Figure A12
<p>Time-delay correlation (R<sup>2</sup>) of non-photosynthetic and photosynthetic vegetation cover with monthly precipitation in different desertification types and degrees. Figure note: MBD represents mobile dune desertification; CD represents coppice dune desertification; GD represents Gobi desertification; MD represents mild desertification; MOD represents moderate desertification; and SD represents severe desertification. R<sup>2</sup> is the correlation.</p>
Full article ">Figure A13
<p>(<b>a</b>) Spatial distribution map of 1-month SPEI series in September 2000. (<b>b</b>) Spatial distribution map of 3-month SPEI series in September 2000. (<b>c</b>) Spatial distribution map of 9-month SPEI series in September 2000. (<b>d</b>) Spatial distribution map of 12-month SPEI series in September 2000.</p>
Full article ">Figure A14
<p>(<b>a</b>) Spatial distribution map of 1-month SPEI series in September 2005. (<b>b</b>) Spatial distribution map of 3-month SPEI series in September 2005. (<b>c</b>) Spatial distribution map of 9-month SPEI series in September 2005. (<b>d</b>) Spatial distribution map of 12-month SPEI series in September 2005.</p>
Full article ">Figure A15
<p>(<b>a</b>) Spatial distribution map of 1-month SPEI series in September 2010. (<b>b</b>) Spatial distribution map of 3-month SPEI series in September 2010. (<b>c</b>) Spatial distribution map of 9-month SPEI series in September 2010. (<b>d</b>) Spatial distribution map of 12-month SPEI series in September 2010.</p>
Full article ">Figure A16
<p>(<b>a</b>) Spatial distribution map of 1-month SPEI series in September 2015. (<b>b</b>) Spatial distribution map of 3-month SPEI series in September 2015. (<b>c</b>) Spatial distribution map of 9-month SPEI series in September 2015. (<b>d</b>) Spatial distribution map of 12-month SPEI series in September 2015.</p>
Full article ">
32 pages, 10548 KiB  
Article
GAN-SkipNet: A Solution for Data Imbalance in Cardiac Arrhythmia Detection Using Electrocardiogram Signals from a Benchmark Dataset
by Hari Mohan Rai, Joon Yoo and Serhii Dashkevych
Mathematics 2024, 12(17), 2693; https://doi.org/10.3390/math12172693 - 29 Aug 2024
Cited by 1 | Viewed by 313
Abstract
Electrocardiography (ECG) plays a pivotal role in monitoring cardiac health, yet the manual analysis of ECG signals is challenging due to the complex task of identifying and categorizing various waveforms and morphologies within the data. Additionally, ECG datasets often suffer from a significant [...] Read more.
Electrocardiography (ECG) plays a pivotal role in monitoring cardiac health, yet the manual analysis of ECG signals is challenging due to the complex task of identifying and categorizing various waveforms and morphologies within the data. Additionally, ECG datasets often suffer from a significant class imbalance issue, which can lead to inaccuracies in detecting minority class samples. To address these challenges and enhance the effectiveness and efficiency of cardiac arrhythmia detection from imbalanced ECG datasets, this study proposes a novel approach. This research leverages the MIT-BIH arrhythmia dataset, encompassing a total of 109,446 ECG beats distributed across five classes following the Association for the Advancement of Medical Instrumentation (AAMI) standard. Given the dataset’s inherent class imbalance, a 1D generative adversarial network (GAN) model is introduced, incorporating the Bi-LSTM model to synthetically generate the two minority signal classes, which represent a mere 0.73% fusion (F) and 2.54% supraventricular (S) of the data. The generated signals are rigorously evaluated for similarity to real ECG data using three key metrics: mean squared error (MSE), structural similarity index (SSIM), and Pearson correlation coefficient (r). In addition to addressing data imbalance, the work presents three deep learning models tailored for ECG classification: SkipCNN (a convolutional neural network with skip connections), SkipCNN+LSTM, and SkipCNN+LSTM+Attention mechanisms. To further enhance efficiency and accuracy, the test dataset is rigorously assessed using an ensemble model, which consistently outperforms the individual models. The performance evaluation employs standard metrics such as precision, recall, and F1-score, along with their average, macro average, and weighted average counterparts. Notably, the SkipCNN+LSTM model emerges as the most promising, achieving remarkable precision, recall, and F1-scores of 99.3%, which were further elevated to an impressive 99.60% through ensemble techniques. Consequently, with this innovative combination of data balancing techniques, the GAN-SkipNet model not only resolves the challenges posed by imbalanced data but also provides a robust and reliable solution for cardiac arrhythmia detection. This model stands poised for clinical applications, offering the potential to be deployed in hospitals for real-time cardiac arrhythmia detection, thereby benefiting patients and healthcare practitioners alike. Full article
Show Figures

Figure 1

Figure 1
<p>Sample ECG beats in the AAMI classes. Zero padding was applied to standardize all segment lengths to 187.</p>
Full article ">Figure 2
<p>(<b>a</b>) Schema of Bi-LSTM architecture (<b>left</b>) and (<b>b</b>) schema of attention model architecture (<b>right</b>).</p>
Full article ">Figure 3
<p>Block diagram of the proposed methodology for the classification of cardiac arrhythmia.</p>
Full article ">Figure 4
<p>Proposed GAN architecture for minority ECG data augmentation.</p>
Full article ">Figure 5
<p>Layer-wise architectures of the proposed ECG classification deep models.</p>
Full article ">Figure 6
<p>Generation of ECG beats of the S class by the GAN model. Each graph depicts overlayed synthetic ECG signals with uniform fixed length (size 187) in common with the pre-processed input ECG signals; the graphs are shown from left to right and top to bottom in the temporal sequence of the generated signals at intervals of 900 training epochs. Compared with the beginning of training (<b>top left graph</b>), the synthetic signals demonstrate more convergence toward the end of training (<b>bottom right graph</b>).</p>
Full article ">Figure 7
<p>Real (<b>left</b>) and GAN-generated synthetic (<b>right</b>) ECG beats of the S class. The synthetic ECG beat appears visually realistic.</p>
Full article ">Figure 8
<p>Visual depiction of progressive discriminator training error (which quantifies differences between the real and synthetic ECG signals) and generator training error measured over 3000 epochs during ECG beat generation of the S class using the GAN model. The discriminator training error is reduced substantially with training, whereas the generator training error remains largely flat.</p>
Full article ">Figure 9
<p>Generation of ECG beats of the F class by the GAN model. Each graph depicts overlayed synthetic ECG signals with uniform fixed length (size 187) in common with the pre-processed input ECG signals; the graphs are shown from left to right and top to bottom in the temporal sequence of the generated signals at intervals of 900 training epochs. Compared with the beginning of training (<b>top left graph</b>), the synthetic signals demonstrate more convergence toward the end of training (<b>bottom right graph</b>).</p>
Full article ">Figure 10
<p>Real (<b>left</b>) and GAN-generated synthetic (<b>right</b>) ECG beats of the F class. The synthetic ECG beat appears visually realistic.</p>
Full article ">Figure 11
<p>Visual depiction of progressive discriminator training error (which quantifies differences between the real and synthetic ECG signals) and generator training error measured over 3000 epochs during ECG beat generation of the F class using the GAN model. The discriminator training error is reduced substantially with training, whereas the generator training error increases initially but then decreases and remains flat.</p>
Full article ">Figure 12
<p>Randomly selected real (<b>left</b>) and synthetic (<b>right</b>) ECG signals of the S class, with corresponding calculated similarity matching scores MSE, SSIM, and r.</p>
Full article ">Figure 13
<p>Randomly selected real (<b>left</b>) and synthetic (<b>right</b>) ECG signals of the F class, with corresponding calculated similarity matching scores MSE, SSIM, and r.</p>
Full article ">Figure 14
<p>Loss function curves (<b>left</b>) and performance metrics curves (<b>right</b>) during 100 epochs of training with the SkipCNN model.</p>
Full article ">Figure 15
<p>SkipCNN classification performance for individual ECG beat classes and across all arrhythmia classes in the test dataset.</p>
Full article ">Figure 16
<p>Loss function curves (<b>left</b>) and performance metrics curves (<b>right</b>) during 100 epochs of training with the SkipCNN+LSTM model.</p>
Full article ">Figure 17
<p>SkipCNN-LSTM classification performance for individual ECG beat classes and across all arrhythmia classes in the test dataset.</p>
Full article ">Figure 18
<p>Loss function curves (<b>left</b>) and performance metric curves (<b>right</b>) during 100 epochs of training with the SkipCNN+LSTM+Attention model.</p>
Full article ">Figure 19
<p>Arrhythmia detection outcomes in terms of performance metrics using the proposed SkipCNN+LSTM+Attention model.</p>
Full article ">Figure 20
<p>Confusion matrix of arrhythmia detection using the ensemble model.</p>
Full article ">
20 pages, 11706 KiB  
Article
Precision Medicine for Apical Lesions and Peri-Endo Combined Lesions Based on Transfer Learning Using Periapical Radiographs
by Pei-Yi Wu, Yi-Cheng Mao, Yuan-Jin Lin, Xin-Hua Li, Li-Tzu Ku, Kuo-Chen Li, Chiung-An Chen, Tsung-Yi Chen, Shih-Lun Chen, Wei-Chen Tu and Patricia Angela R. Abu
Bioengineering 2024, 11(9), 877; https://doi.org/10.3390/bioengineering11090877 - 29 Aug 2024
Viewed by 415
Abstract
An apical lesion is caused by bacteria invading the tooth apex through caries. Periodontal disease is caused by plaque accumulation. Peri-endo combined lesions include both diseases and significantly affect dental prognosis. The lack of clear symptoms in the early stages of onset makes [...] Read more.
An apical lesion is caused by bacteria invading the tooth apex through caries. Periodontal disease is caused by plaque accumulation. Peri-endo combined lesions include both diseases and significantly affect dental prognosis. The lack of clear symptoms in the early stages of onset makes diagnosis challenging, and delayed treatment can lead to the spread of symptoms. Early infection detection is crucial for preventing complications. PAs used as the database were provided by Chang Gung Memorial Medical Center, Taoyuan, Taiwan, with permission from the Institutional Review Board (IRB): 02002030B0. The tooth apex image enhancement method is a new technology in PA detection. This image enhancement method is used with convolutional neural networks (CNN) to classify apical lesions, peri-endo combined lesions, and asymptomatic cases, and to compare with You Only Look Once-v8-Oriented Bounding Box (YOLOv8-OBB) disease detection results. The contributions lie in the utilization of database augmentation and adaptive histogram equalization on individual tooth images, achieving the highest comprehensive validation accuracy of 95.23% with the ConvNextv2 model. Furthermore, the CNN outperformed YOLOv8 in identifying apical lesions, achieving an F1-Score of 92.45%. For the classification of peri-endo combined lesions, CNN attained the highest F1-Score of 96.49%, whereas YOLOv8 scored 88.49%. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Research flowchart.</p>
Full article ">Figure 2
<p>Manual annotation using Roboflow’s polygon tool.</p>
Full article ">Figure 3
<p>The first tooth rotated to a horizontal 0-degree image.</p>
Full article ">Figure 4
<p>Image expansion. (<b>a</b>) Original cropped image. (<b>b</b>) Cropped image expanded by 20 pixels horizontally and 40 pixels vertically.</p>
Full article ">Figure 5
<p>The Gaussian high-pass filter result. (<b>a</b>) The original image. (<b>b</b>) The result of the Gaussian high-pass filter. (<b>c</b>) The result of (<b>a</b>) minus (<b>b</b>).</p>
Full article ">Figure 6
<p>The adaptive histogram equalization. (<b>a</b>) Original image and histogram. (<b>b</b>) Enhanced image and histogram after adaptive histogram equalization.</p>
Full article ">Figure 7
<p>The flat-field correction result. (<b>a</b>) Original image. (<b>b</b>) Flat-field correction image.</p>
Full article ">Figure 8
<p>Linear transform model.</p>
Full article ">Figure 9
<p>The result of linear transformation. (<b>a</b>) Original image. (<b>b</b>) Linear transformation image.</p>
Full article ">Figure 10
<p>The result of negative film effect. (<b>a</b>) Original image. (<b>b</b>) Negative film effect image.</p>
Full article ">Figure 11
<p>The disease prediction results with YOLOv8 OBB.</p>
Full article ">Figure 12
<p>Single-tooth prediction results.</p>
Full article ">Figure 13
<p>Validation accuracy during the training process of the Places365-GoogLeNet model.</p>
Full article ">Figure 14
<p>Validation loss function during the training process of the Places365-GoogLeNet model.</p>
Full article ">Figure 15
<p>After adaptive histogram equalization, segmentation into multiple single-tooth images is conducted (numbered from left to right).</p>
Full article ">
14 pages, 7195 KiB  
Article
RHYTHMI: A Deep Learning-Based Mobile ECG Device for Heart Disease Prediction
by Alaa Eleyan, Ebrahim AlBoghbaish, Abdulwahab AlShatti, Ahmad AlSultan and Darbi AlDarbi
Appl. Syst. Innov. 2024, 7(5), 77; https://doi.org/10.3390/asi7050077 - 29 Aug 2024
Viewed by 448
Abstract
Heart disease, a global killer with many variations like arrhythmia and heart failure, remains a major health concern. Traditional risk factors include age, cholesterol, diabetes, and blood pressure. Fortunately, artificial intelligence (AI) offers a promising solution. We have harnessed the power of AI, [...] Read more.
Heart disease, a global killer with many variations like arrhythmia and heart failure, remains a major health concern. Traditional risk factors include age, cholesterol, diabetes, and blood pressure. Fortunately, artificial intelligence (AI) offers a promising solution. We have harnessed the power of AI, specifically deep learning and convolutional neural networks (CNNs), to develop Rhythmi, an innovative mobile ECG diagnosis device for heart disease detection. Rhythmi leverages extensive medical data from databases like MIT-BIH and BIDMC. These data empower the training and testing of the developed deep learning model to analyze ECG signals with accuracy, precision, sensitivity, specificity, and F1-score in identifying arrhythmias and other heart conditions, with performances reaching 98.52%, 98.55%, 98.52%, 99.26%, and 98.52%, respectively. Moreover, we tested Rhythmi in real time using a mobile device with a single-lead ECG sensor. This user-friendly prototype captures the ECG signal, transmits it to Rhythmi’s dedicated website, and provides instant diagnosis and feedback on the patient’s heart health. The developed mobile ECG diagnosis device addresses the main problems of traditional ECG diagnostic devices such as accessibility, cost, mobility, complexity, and data integration. However, we believe that despite the promising results, our system will still need intensive clinical validation in the future. Full article
Show Figures

Figure 1

Figure 1
<p>Number of available recordings per class (blue) vs. the number of used recordings per class (orange).</p>
Full article ">Figure 2
<p>Original signal/recording (65,535 features) vs. one segmented signal (500 features).</p>
Full article ">Figure 3
<p>Comparison between a noisy and a filtered signal.</p>
Full article ">Figure 4
<p>Database splitting ratio for the training and testing sets.</p>
Full article ">Figure 5
<p>Block diagram of the stages of the proposed deep learning model.</p>
Full article ">Figure 6
<p>The proposed deep learning model’s structure.</p>
Full article ">Figure 7
<p>ECG acquisition and electrode placements. The positive electrode (red) is placed on the left hand while both the negative (black) and the reference (white) electrodes are placed on the right hand.</p>
Full article ">Figure 8
<p>RHYTHMI’s application website (<a href="https://www.rhythmi.org/" target="_blank">https://www.rhythmi.org/</a>).</p>
Full article ">Figure 9
<p>A scene from a real-time testing trial of our developed mobile ECG device at one of the exhibitions we participated in.</p>
Full article ">Figure 10
<p>Rhythmi real-time ECG test results for normal (<b>left</b>) and abnormal (<b>right</b>) cases.</p>
Full article ">Figure 11
<p>Number of individuals tested at different exhibitions and competitions and their diagnosis.</p>
Full article ">Figure 12
<p>The proposed model’s accuracy and loss values tracked over multiple epochs for the training (blue) and validation (orange) datasets.</p>
Full article ">Figure 13
<p>The confusion matrix for the results of the proposed approach using the MIT-BIH and BIDMC databases.</p>
Full article ">Figure 14
<p>ROC curves for the training dataset (<b>top</b>) and the test dataset (<b>bottom</b>) for the three classes of the MIT-BIH and BIDMC databases.</p>
Full article ">
22 pages, 29298 KiB  
Article
Landslide Recognition Based on Machine Learning Considering Terrain Feature Fusion
by Jincan Wang, Zhiheng Wang, Liyao Peng and Chenzhihao Qian
ISPRS Int. J. Geo-Inf. 2024, 13(9), 306; https://doi.org/10.3390/ijgi13090306 - 28 Aug 2024
Viewed by 459
Abstract
Landslides are one of the major disasters that exist worldwide, posing a serious threat to human life and property safety. Rapid and accurate detection and mapping of landslides are crucial for risk assessment and humanitarian assistance in affected areas. To achieve this goal, [...] Read more.
Landslides are one of the major disasters that exist worldwide, posing a serious threat to human life and property safety. Rapid and accurate detection and mapping of landslides are crucial for risk assessment and humanitarian assistance in affected areas. To achieve this goal, this study proposes a landslide recognition method based on machine learning (ML) and terrain feature fusion. Taking the Dawan River Basin in Detuo Township and Tianwan Yi Ethnic Township as the research area, firstly, landslide-related data were compiled, including a landslide inventory based on field surveys, satellite images, historical data, high-resolution remote sensing images, and terrain data. Then, different training datasets for landslide recognition are constructed, including full feature datasets that fusion terrain features and remote sensing features and datasets that only contain remote sensing features. At the same time, different ratios of landslide to non-landslide (or positive/negative, P/N) samples are set in the training data. Subsequently, five ML algorithms, including Extreme Gradient Boost (XGBoost), Adaptive Boost (AdaBoost), Light Gradient Boost (LightGBM), Random Forest (RF), and Convolutional Neural Network (CNN), were used to train each training dataset, and landslide recognition was performed on the validation area. Finally, accuracy (A), precision (P), recall (R), F1 score (F1), and intersection over union (IOU) were selected to evaluate the landslide recognition ability of different models. The research results indicate that selecting ML models suitable for the study area and the ratio of the P/N samples can improve the A, R, F1, and IOU of landslide identification results, resulting in more accurate and reasonable landslide identification results; Fusion terrain features can make the model recognize landslides more comprehensively and align better with the actual conditions. The best-performing model in the study is LightGBM. When the input data includes all features and the P/N sample ratio is optimal, the A, P, R, F1, and IOU of landslide recognition results for this model are 97.47%, 85.40%, 76.95%, 80.95%, and 71.28%, respectively. Compared to the landslide recognition results using only remote sensing features, this model shows improvements of 4.51%, 35.66%, 5.41%, 22.27%, and 29.16% in A, P, R, F1, and IOU, respectively. This study serves as a valuable reference for the precise and comprehensive identification of landslide areas. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Mogangling landslide; and (<b>b</b>) Lantianwan landslide.</p>
Full article ">Figure 2
<p>The study area’s location.</p>
Full article ">Figure 3
<p>Basic data for landslide identification includes: (<b>a</b>) optical image on 10 September 2022; (<b>b</b>) DEM; (<b>c</b>) landslide area in Dadu River Basin, Tianwan Yi Ethnic Township; and (<b>d</b>) landslide area in Dadu River Basin, Detuo Township.</p>
Full article ">Figure 4
<p>XGBoost conceptual model.</p>
Full article ">Figure 5
<p>AdaBoost conceptual model.</p>
Full article ">Figure 6
<p>LightGBM conceptual model.</p>
Full article ">Figure 7
<p>Flow chart of the Random Forest model.</p>
Full article ">Figure 8
<p>CNN structure diagram.</p>
Full article ">Figure 9
<p>Landslide identification results based on XGBoost.</p>
Full article ">Figure 10
<p>Landslide identification results based on AdaBoost.</p>
Full article ">Figure 11
<p>Landslide identification results based on LightGBM.</p>
Full article ">Figure 12
<p>Landslide identification results based on RF.</p>
Full article ">Figure 13
<p>Landslide identification results based on CNN.</p>
Full article ">Figure 14
<p>Performance comparison of different models: (<b>a</b>) accuracy on the test set; (<b>b</b>) precision on the test set; (<b>c</b>) recall on the test set; (<b>d</b>) F1 score on the test set; and (<b>e</b>) IOU on the test set.</p>
Full article ">Figure 15
<p>Landslide identification effect with different ratios of P/N samples.</p>
Full article ">Figure 16
<p>Landslide recognition effect of different training data.</p>
Full article ">
16 pages, 2588 KiB  
Article
Development of a Machine Learning Model for the Classification of Enterobius vermicularis Egg
by Natthanai Chaibutr, Pongphan Pongpanitanont, Sakhone Laymanivong, Tongjit Thanchomnang and Penchom Janwan
J. Imaging 2024, 10(9), 212; https://doi.org/10.3390/jimaging10090212 - 28 Aug 2024
Viewed by 436
Abstract
Enterobius vermicularis (pinworm) infections are a significant global health issue, affecting children predominantly in environments like schools and daycares. Traditional diagnosis using the scotch tape technique involves examining E. vermicularis eggs under a microscope. This method is time-consuming and depends heavily on the [...] Read more.
Enterobius vermicularis (pinworm) infections are a significant global health issue, affecting children predominantly in environments like schools and daycares. Traditional diagnosis using the scotch tape technique involves examining E. vermicularis eggs under a microscope. This method is time-consuming and depends heavily on the examiner’s expertise. To improve this, convolutional neural networks (CNNs) have been used to automate the detection of pinworm eggs from microscopic images. In our study, we enhanced E. vermicularis egg detection using a CNN benchmarked against leading models. We digitized and augmented 40,000 images of E. vermicularis eggs (class 1) and artifacts (class 0) for comprehensive training, using an 80:20 training–validation and a five-fold cross-validation. The proposed CNN model showed limited initial performance but achieved 90.0% accuracy, precision, recall, and F1-score after data augmentation. It also demonstrated improved stability with an ROC-AUC metric increase from 0.77 to 0.97. Despite its smaller file size, our CNN model performed comparably to larger models. Notably, the Xception model achieved 99.0% accuracy, precision, recall, and F1-score. These findings highlight the effectiveness of data augmentation and advanced CNN architectures in improving diagnostic accuracy and efficiency for E. vermicularis infections. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>The workflow of an object detection system incorporating data augmentation techniques is illustrated. (<b>A</b>) This comprehensive process begins with data acquisition, followed by preprocessing and image augmentation. The augmented images are used for model training, after which the model undergoes validation and testing. Once trained, the model is applied to new test images to detect objects, with performance evaluated using the Intersection-over-Union (IoU) metric. (<b>B</b>) Various augmentation techniques are applied to the original dataset, creating an enhanced and diversified dataset.</p>
Full article ">Figure 2
<p>A series of microscopic images illustrating the impact of various image augmentation techniques on two distinct classes (class 0 and class 1). The first column features the original, unaltered images. Moving to the right, each subsequent column reveals the transformed images after applying different augmentation methods: Gaussian blur, mean filtering, Gaussian noise, and kernel sharpening. These techniques introduce a range of visual variations and distortions, enriching the dataset. By incorporating these augmentations, the goal is to bolster the robustness and enhance the generalization capabilities of the machine learning model trained on this enriched dataset. The upper row of images represents class 0, and the lower row represents class 1.</p>
Full article ">Figure 3
<p>Architectural design of proposed convolutional neural network (CNN).</p>
Full article ">Figure 4
<p>Outcomes of training and validating machine learning model. Results of five-fold cross-validation training loss. (<b>A</b>) Non-augmented image dataset. (<b>B</b>) Augmented image dataset. Relationship between prediction accuracy and number of folds used in cross-validation for image datasets. (<b>C</b>) Non-augmented image dataset. (<b>D</b>) Augmented image dataset.</p>
Full article ">Figure 5
<p>Receiver operating characteristic (ROC) curves for binary classification model trained on two different datasets. (<b>A</b>) Non-augmented image dataset. (<b>B</b>) Augmented image dataset. The orange line indicates the correct positives and misclassified negatives, highlighting the model’s class distinction ability. The blue dashed line indicates an AUC value of 0.5, suggesting no discriminative power.</p>
Full article ">Figure 6
<p>A detailed comparison between the object detection results of a highly trained machine learning model, Xception, and the annotations made by expert medical staff on the microscopic images. (<b>A</b>) The objects detected by the Xception model, highlighted by green bounding boxes. (<b>B</b>) The annotations made by a parasitology expert, indicated by red bounding boxes. (<b>C</b>) A combined view, displaying both the expert annotations (red bounding boxes) and the model’s predictions (green bounding boxes).</p>
Full article ">
Back to TopTop