[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (302)

Search Parameters:
Keywords = ConvLSTM

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 8824 KiB  
Article
Research on Short-Term Forecasting Model of Global Atmospheric Temperature and Wind in the near Space Based on Deep Learning
by Xingxin Sun, Chen Zhou, Jian Feng, Huiyun Yang, Yuqiang Zhang, Zhou Chen, Tong Xu, Zhongxin Deng, Zhengyu Zhao, Yi Liu and Ting Lan
Atmosphere 2024, 15(9), 1069; https://doi.org/10.3390/atmos15091069 - 4 Sep 2024
Viewed by 231
Abstract
Developing short-term forecasting models for global atmospheric temperature and wind in near space is crucial for understanding atmospheric dynamics and supporting human activities in this region. While numerical models have been extensively developed, deep learning techniques have recently shown promise in improving atmospheric [...] Read more.
Developing short-term forecasting models for global atmospheric temperature and wind in near space is crucial for understanding atmospheric dynamics and supporting human activities in this region. While numerical models have been extensively developed, deep learning techniques have recently shown promise in improving atmospheric forecasting accuracy. In this study, convolutional long short-term memory (ConvLSTM) and convolutional gated recurrent unit (ConvGRU) neural networks were applied to build for short-term global-scale forecasting model of atmospheric temperature and wind in near space based on the MERRA-2 reanalysis dataset from 2010–2022. The model results showed that the ConvGRU model outperforms the ConvLSTM model in the short-term forecast results. The ConvGRU model achieved a root mean square error in the first three hours of approximately 1.8 K for temperature predictions, and errors of 4.2 m/s and 3.8 m/s for eastward and northward wind predictions on all 72 isobaric surfaces. Specifically, at a higher altitude (on the 1.65 Pa isobaric surface, approximately 70 km above sea level), the ConvGRU model achieved a RMSE of about 2.85 K for temperature predictions, and 5.67 m/s and 5.17 m/s for eastward and northward wind. This finding is significantly meaningful for short-term temperature and wind forecasts in near space and for exploring the physical mechanisms related to temperature and wind variations in this region. Full article
(This article belongs to the Section Atmospheric Techniques, Instruments, and Modeling)
Show Figures

Figure 1

Figure 1
<p>The results from the MERRA2 reanalysis dataset at a height of 1.65 Pa on 1 June 2022. (<b>a</b>): temperature at 9:00 on 1 June 2022; (<b>b</b>): temperature at 9:00 on 1 December 2022; (<b>c</b>): eastward wind at 9:00 on 1 December 2022; (<b>d</b>): northward wind at 9:00 on 1 December 2022.</p>
Full article ">Figure 2
<p>The MSELoss curves during the training of ConvGRU and ConvLSTM. (<b>a</b>): The MSELoss curves during the training of ConvGRU with normalized data; (<b>b</b>): The MSELoss curves during the training of ConvGRU with unnormalized data; (<b>c</b>): The MSELoss curves during the training of ConvLSTM with normalized data; (<b>d</b>): The MSELoss curves during the training of ConvLSTM with unnormalized data.</p>
Full article ">Figure 3
<p>The structure of the ConvLSTM unit.</p>
Full article ">Figure 4
<p>The workflow diagram of the ConvLSTM model.</p>
Full article ">Figure 5
<p>The structure of the ConvGRU unit.</p>
Full article ">Figure 6
<p>The workflow diagram of the ConvGRU model.</p>
Full article ">Figure 7
<p>The predictive accuracy of eight-time segments per day. ((<b>a</b>–<b>c</b>): The MAE, R and RMSE Curve of temperature on 72 isobaric surfaces on global scales at eight time periods on 3 January 2022. (<b>d</b>–<b>f</b>): The MAE, R and RMSE Curve of northward wind on 72 isobaric surfaces on global scales at eight time periods on 3 January 2022. (<b>g</b>–<b>i</b>): The MAE, R and RMSE Curve of eastward wind on 72 isobaric surfaces on global scales at eight time periods on 3 January 2022).</p>
Full article ">Figure 8
<p>Predicted Temperature by ConvLSTM. (<b>a</b>): the actual temperature data on 3 June 2022; (<b>b</b>): the temperature predicted by GonvLSTM on 3 June 2022; (<b>c</b>): the differences between actual temperature data and predicted data on 3 June 2022; (<b>d</b>): the average and root mean square errors between the actual temperatures and predicted data on 3 June 2022; (<b>e</b>): the correlation coefficient, average error, and root mean square error between the actual data and predicted temperature data on 3 June 2022.</p>
Full article ">Figure 9
<p>Predicted Temperature by ConvGRU. (<b>a</b>): the actual temperature data on 3 June 2022; (<b>b</b>): the temperature predicted by GonvGRU on 3 June 2022; (<b>c</b>): the differences between actual temperature data and predicted data on 3 June 2022; (<b>d</b>): the average and root mean square errors between the actual temperatures and predicted data on 3 June 2022; (<b>e</b>): the correlation coefficient, average error, and root mean square error between the actual data and predicted temperature data on 3 June 2022.</p>
Full article ">Figure 10
<p>Predicted Eastward Wind by ConvLSTM. (<b>a</b>): the actual Eastward Wind data on 3 June 2022; (<b>b</b>): the Eastward Wind predicted by GonvLSTM on 3 June 2022; (<b>c</b>): the differences between actual Eastward Wind data and predicted data on 3 June 2022; (<b>d</b>): the average and root mean square errors between the actual Eastward Wind and predicted data on 3 June 2022; (<b>e</b>): the correlation coefficient, average error, and root mean square error between the actual data and predicted Eastward Wind data on 3 June 2022.</p>
Full article ">Figure 11
<p>Predicted Eastward by ConvGRU. (<b>a</b>): the actual Eastward Wind data on 3 June 2022; (<b>b</b>): the Eastward Wind predicted by GonvGRU on 3 June 2022; (<b>c</b>): the differences between actual Eastward Wind data and predicted data on 3 June 2022; (<b>d</b>): the average and root mean square errors between the actual Eastward Wind and predicted data on 3 June 2022; (<b>e</b>): the correlation coefficient, average error, and root mean square error between the actual data and predicted Eastward Wind data on 3 June 2022.</p>
Full article ">Figure 12
<p>Predicted Northward Wind by ConvLSTM. (<b>a</b>): the actual Northward Wind data on 3 June 2022; (<b>b</b>): the Northward Wind predicted by GonvLSTM on 3 June 2022; (<b>c</b>): the differences between actual Northward Wind data and predicted data on 3 June 2022; (<b>d</b>): the average and root mean square errors between the actual Northward Wind and predicted data on 3 June 2022; (<b>e</b>): the correlation coefficient, average error, and root mean square error between the actual data and predicted Northward Wind data on 3 June 2022.</p>
Full article ">Figure 13
<p>Predicted Northward wind by ConvGRU. (<b>a</b>): the actual Northward Wind data on 3 June 2022; (<b>b</b>): the Northward Wind predicted by GonvGRU on 3 June 2022; (<b>c</b>): the differences between actual Northward Wind data and predicted data on 3 June 2022; (<b>d</b>): the average and root mean square errors between the actual Northward Wind and predicted data on 3 June 2022; (<b>e</b>): the correlation coefficient, average error, and root mean square error between the actual data and predicted Northward Wind data on 3 June 2022.</p>
Full article ">Figure 14
<p>Model performance vs. HWM model. (<b>a</b>): the actual Eastward Wind data on 3 June 2022; (<b>b</b>): the predicted Eastward Wind data by HWM model; (<b>c</b>): the predicted Eastward Wind data by ConvGRU model; (<b>d</b>): the differences between actual Eastward Wind data and data predicted by HWM model; (<b>e</b>): the differences between actual Eastward Wind data and data predicted by ConvGRU model; (<b>f</b>): the differences between actual Eastward Wind data and data predicted by ConvLSTM model.</p>
Full article ">Figure 15
<p>Model performance at different 72 isobaric surfaces. (<b>a</b>): the accuracy of temperature; (<b>b</b>): the accuracy of Eastward Wind; (<b>c</b>) the accuracy of Northward Wind; (<b>d</b>): the correlation coefficient of temperature, Eastward Wind and Northward Wind.</p>
Full article ">Figure 16
<p>Model performance at different 96 latitudes surfaces. (<b>a</b>): the accuracy of temperature; (<b>b</b>): the accuracy of Eastward Wind; (<b>c</b>) the accuracy of Northward Wind; (<b>d</b>): the correlation coefficient of temperature, Eastward Wind and Northward Wind.</p>
Full article ">
20 pages, 22277 KiB  
Article
Attention-Based Spatiotemporal-Aware Network for Fine-Grained Visual Recognition
by Yili Ren, Ruidong Lu, Guan Yuan, Dashuai Hao and Hongjue Li
Appl. Sci. 2024, 14(17), 7755; https://doi.org/10.3390/app14177755 - 2 Sep 2024
Viewed by 430
Abstract
On public benchmarks, current macro facial expression recognition technologies have achieved significant success. However, in real-life scenarios, individuals may attempt to conceal their true emotions. Conventional expression recognition often overlooks subtle facial changes, necessitating more fine-grained micro-expression recognition techniques. Different with prevalent facial [...] Read more.
On public benchmarks, current macro facial expression recognition technologies have achieved significant success. However, in real-life scenarios, individuals may attempt to conceal their true emotions. Conventional expression recognition often overlooks subtle facial changes, necessitating more fine-grained micro-expression recognition techniques. Different with prevalent facial expressions, weak intensity and short duration are the two main obstacles for perceiving and interpreting a micro-expression correctly. Meanwhile, correlations between pixels of visual data in spatial and channel dimensions are ignored in most existing methods. In this paper, we propose a novel network structure, the Attention-based Spatiotemporal-aware network (ASTNet), for micro-expression recognition. In ASTNet, we combine ResNet and ConvLSTM as a holistic framework (ResNet-ConvLSTM) to extract the spatial and temporal features simultaneously. Moreover, we innovatively integrate two level attention mechanisms, channel-level attention and spatial-level attention, into the ResNet-ConvLSTM. Channel-level attention is used to discriminate the importance of different channels because the contributions for the overall presentation of micro-expression vary between channels. Spatial-level attention is leveraged to dynamically estimate weights for different regions due to the diversity of regions’ reflections to micro-expression. Extensive experiments conducted on two benchmark datasets demonstrate that ASTNet achieves performance improvements of 4.25–16.02% and 0.79–12.93% over several state-of-the-art methods. Full article
(This article belongs to the Topic Computer Vision and Image Processing, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Research framework.</p>
Full article ">Figure 2
<p>Framework of our network.</p>
Full article ">Figure 3
<p>Comparison of before and after micro-expression magnification.</p>
Full article ">Figure 4
<p>Comparison of before and after micro-expression magnification.</p>
Full article ">Figure 5
<p>Optical flow visualization.</p>
Full article ">Figure 6
<p>Residual block structure of ResNet.</p>
Full article ">Figure 7
<p>CBAM attention module.</p>
Full article ">Figure 8
<p>Channel attention module.</p>
Full article ">Figure 9
<p>Spatial attention module.</p>
Full article ">Figure 10
<p>The structure of ConvLSTM.</p>
Full article ">Figure 11
<p>Influence of different batch sizes on our network.</p>
Full article ">Figure 12
<p>Influence of different dropouts on our network.</p>
Full article ">Figure 13
<p>Confusion matrix.</p>
Full article ">Figure 14
<p>The multimodal teaching behavior intelligent analysis system for smart classrooms.</p>
Full article ">
17 pages, 3826 KiB  
Article
Prediction of Remaining Useful Life of Battery Using Partial Discharge Data
by Qaiser Hussain, Sunguk Yun, Jaekyun Jeong, Mangyu Lee and Jungeun Kim
Electronics 2024, 13(17), 3475; https://doi.org/10.3390/electronics13173475 - 1 Sep 2024
Viewed by 351
Abstract
Lithium-ion batteries are cornerstones of renewable technologies, which is why they are used in many applications, specifically in electric vehicles and portable electronics. The accurate estimation of the remaining useful life (RUL) of a battery is pertinent for durability, efficient operation, and stability. [...] Read more.
Lithium-ion batteries are cornerstones of renewable technologies, which is why they are used in many applications, specifically in electric vehicles and portable electronics. The accurate estimation of the remaining useful life (RUL) of a battery is pertinent for durability, efficient operation, and stability. In this study, we have proposed an approach to predict the RUL of a battery using partial discharge data from the battery cycles. Unlike other studies that use complete cycle data and face reproducibility issues, our research utilizes only partial data, making it both practical and reproducible. To analyze this partial data, we applied various deep learning methods and compared multiple models, among which ConvLSTM showed the best performance, with an RMSE of 0.0824. By comparing the performance of ConvLSTM at various ratios and ranges, we have confirmed that using partial data can achieve a performance equal to or better than that obtained when using complete cycle data. Full article
(This article belongs to the Special Issue Power Electronics and Renewable Energy System)
Show Figures

Figure 1

Figure 1
<p>The discharge capacity profile vs. voltage over the 400 cycles.</p>
Full article ">Figure 2
<p>Framework of our proposed study.</p>
Full article ">Figure 3
<p>Overall performance evaluation of ConvLSTM in 5%, 10%, 20%, and 100% ratio.</p>
Full article ">Figure 4
<p>Statistical values of discharge capacity across a 5% ratio.</p>
Full article ">Figure 5
<p>Comparison of performance across 5% ratio in marked ranges.</p>
Full article ">Figure 6
<p>Comparison of performance across 1% ratio in marked ranges.</p>
Full article ">Figure 7
<p>Statistical values of the discharge capacity across a 1% ratio.</p>
Full article ">Figure 8
<p>Performances of ConvLSTM in the 1%, 5%, 10%, and 20% ranges excluding the second input.</p>
Full article ">
27 pages, 16628 KiB  
Article
Predicting Ground Cover with Deep Learning Models—An Application of Spatio-Temporal Prediction Methods to Satellite-Derived Ground Cover Maps in the Great Barrier Reef Catchments
by Yongjing Mao, Ryan D. R. Turner, Joseph M. McMahon, Diego F. Correa, Debbie A. Chamberlain and Michael St. J. Warne
Remote Sens. 2024, 16(17), 3193; https://doi.org/10.3390/rs16173193 - 29 Aug 2024
Viewed by 562
Abstract
Livestock grazing is a major land use in the Great Barrier Reef Catchment Area (GBRCA). Heightened grazing density coupled with inadequate land management leads to accelerated soil erosion and increased sediment loads being transported downstream. Ultimately, these increased sediment loads impact the water [...] Read more.
Livestock grazing is a major land use in the Great Barrier Reef Catchment Area (GBRCA). Heightened grazing density coupled with inadequate land management leads to accelerated soil erosion and increased sediment loads being transported downstream. Ultimately, these increased sediment loads impact the water quality of the Great Barrier Reef (GBR) lagoon. Ground cover mapping has been adopted to monitor and assess the land condition in the GBRCA. However, accurate prediction of ground cover remains a vital knowledge gap to inform proactive approaches for improving land conditions. Herein, we explored two deep learning-based spatio-temporal prediction models, including convolutional LSTM (ConvLSTM) and Predictive Recurrent Neural Network (PredRNN), to predict future ground cover. The two models were evaluated on different spatial scales, ranging from a small site (i.e., <5 km2) to the entire GBRCA, with different quantities of training data. Following comparisons against 25% withheld testing data, we found the following: (1) both ConvLSTM and PredRNN accurately predicted the next-season ground cover for not only a single site but also the entire GBRCA. They achieved this with a Mean Absolute Error (MAE) under 5% and a Structural Similarity Index Measure (SSIM) exceeding 0.65; (2) PredRNN superseded ConvLSTM by providing more accurate next-season predictions with better training efficiency; (3) The accuracy of PredRNN varies seasonally and spatially, with lower accuracy observed for low ground cover, which is underestimated. The models assessed in this study can serve as an early-alert tool to produce high-accuracy and high-resolution ground cover prediction one season earlier than observation for the entire GBRCA, which enables local authorities and grazing property owners to take preventive measures to improve land conditions. This study also offers a new perspective on the future utilization of predictive spatio-temporal models, particularly over large spatial scales and across varying environmental sites. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Spatial distribution of training (green), validation (blue), and testing (red) sample sites. The extent of grazing area on the map corresponds to the grazing native vegetation class from the Queensland Land Use Mapping [<a href="#B57-remotesensing-16-03193" class="html-bibr">57</a>].</p>
Full article ">Figure 2
<p>An overview of the workflow.</p>
Full article ">Figure 3
<p>Overview of the model architecture.</p>
Full article ">Figure 4
<p>Time series of ground cover prediction (<b>a</b>) and scoring metrics (<b>b</b>,<b>c</b>) for PredRNN. In plot (<b>a</b>), the solid and dash lines are the predicted and target ground cover at prediction step 1, respectively. The grey bars are the cloud cover rate, and the black vertical line indicates the date with smallest ground cover. The solid line in the plot (<b>b</b>,<b>c</b>) shows the spatial average of MAE and SSIM at prediction step 1, respectively, and the bars are the corresponding decay rate from prediction step 1 to step 8 (season 9~16 in target sequence in <a href="#remotesensing-16-03193-f003" class="html-fig">Figure 3</a>).</p>
Full article ">Figure 5
<p>Comparison of predicted and target ground cover maps for different prediction steps of the data sequence highlighted in <a href="#remotesensing-16-03193-f004" class="html-fig">Figure 4</a>. The first row shows the context images used to generate predictions. The second and third rows show the target images and their predictions. The fourth row shows the absolute difference between the targets and predictions, while the last row shows the histogram comparison of observations (targets) and predictions.</p>
Full article ">Figure 6
<p>Statistics of scoring metrics of PredRNN model for training, validation, and testing sites. Plot (<b>a</b>), (<b>b</b>), (<b>c</b>), and (<b>d</b>) are the results for MAE, SSIM, MAE decay, and SSIM decay, respectively. The MAE and SSIM results are the spatial–temporal mean for the first prediction step. Numbers above/below the bars are the median values of the metrics among all 100 sites.</p>
Full article ">Figure 7
<p>Comparison of different models and training scales for the arid testing site (site 78) in <a href="#remotesensing-16-03193-f001" class="html-fig">Figure 1</a>. Plots in the first column (<b>a1</b>–<b>e1</b>) shows the temporal variation of accuracy. Ground cover, MAE, and SSIM are based on the prediction step 1. ConvLSTM_site and PredRNN_site (dash lines) are model-trained sites, specifically. ConvLSTM_GBRCA and PredRNN_GBRCA (solid lines) are models trained on the GBRCA scale. The brown vertical line indicates the date with the least ground cover. The boxplot in the second column (<b>a2</b>–<b>e2</b>) summarizes the statistics of the time series in the first column. The blue triangle is the mean. Boxes extend from 25th to 75th percentiles, and whiskers from 5th to 95th percentiles. Circles represent outliers.</p>
Full article ">Figure 8
<p>Comparison of different models and training strategies for target images in a single data sequence indicated in <a href="#remotesensing-16-03193-f007" class="html-fig">Figure 7</a>. First row shows results from observation. Second to fifth rows are prediction results from different trained models. The last row shows comparison of histograms.</p>
Full article ">Figure 9
<p>Spatial distribution of MAE and SSIM. The two metrics are aggregated as the spatio-temporal mean for the first prediction step. The smoke-white and light-blue backgrounds represent land and sea, respectively. Polygons with different colors are the grazing areas in different catchments. The varying sizes of scatter points indicate the magnitude of scoring metrics.</p>
Full article ">Figure 10
<p>Correlation of scoring metrics to different environmental factors. GC_mean and GC_STD are the spatial–temporal mean and standard deviation of ground cover for a site. The solid line in each pair plot is the linear fit for the scatter points. <span class="html-italic">ρ</span> is the Pearson correlation coefficient. The background color gradient indicates the magnitude of <span class="html-italic">ρ</span>.</p>
Full article ">Figure 11
<p>Ground cover for grazing areas in the GBRCA. (<b>a</b>,<b>d</b>) are the ground cover maps for June 2022 to August 2022 (the last season in the context sequence) from observation and prediction, respectively; (<b>b</b>,<b>e</b>) are the results for September 2022 to December 2022 (the first season in the target sequence); and (<b>c</b>,<b>f</b>) are the maps for the change in ground cover in the above two seasons. The inset plots in (<b>c</b>,<b>f</b>) compare the histograms of pixels in the two seasons.</p>
Full article ">Figure 12
<p>Accuracy of GBRCA ground cover prediction from September 2019 to September 2023. The solid and dashed lines in the first plot are the spatial average of observation and PredRNN prediction. The filled and unfilled bars represent the increment (negative value for decrement) of observation and prediction, respectively. In the second plot, the solid line with error bars represents the MAE and the standard deviation of the absolute error, while the dashed line represents the SSIM.</p>
Full article ">Figure 13
<p>Comparison of PredRNN model trained on sites in GBRCA and Burdekin. The first column (<b>a</b>) is the observation for the season from September to December 2022; the second column (<b>b</b>,<b>c</b>) shows predictions for the same season by different trained models. The last column (<b>d</b>,<b>e</b>) shows the errors (i.e., prediction–observation) of the different predictions. Inset plots in the last column compare the histograms for observation and model prediction.</p>
Full article ">
23 pages, 7568 KiB  
Article
1D-CLANet: A Novel Network for NLoS Classification in UWB Indoor Positioning System
by Qiu Wang, Mingsong Chen, Jiajie Liu, Yongcheng Lin, Kai Li, Xin Yan and Chizhou Zhang
Appl. Sci. 2024, 14(17), 7609; https://doi.org/10.3390/app14177609 - 28 Aug 2024
Viewed by 419
Abstract
Ultra-Wideband (UWB) technology is crucial for indoor localization systems due to its high accuracy and robustness in multipath environments. However, Non-Line-of-Sight (NLoS) conditions can cause UWB signal distortion, significantly reducing positioning accuracy. Thus, distinguishing between NLoS and LoS scenarios and mitigating positioning errors [...] Read more.
Ultra-Wideband (UWB) technology is crucial for indoor localization systems due to its high accuracy and robustness in multipath environments. However, Non-Line-of-Sight (NLoS) conditions can cause UWB signal distortion, significantly reducing positioning accuracy. Thus, distinguishing between NLoS and LoS scenarios and mitigating positioning errors is crucial for enhancing UWB system performance. This research proposes a novel 1D-ConvLSTM-Attention network (1D-CLANet) for extracting UWB temporal channel impulse response (CIR) features and identifying NLoS scenarios. The model combines the convolutional neural network (CNN) and Long Short-Term memory (LSTM) architectures to extract temporal CIR features and introduces the Squeeze-and-Excitation (SE) attention mechanism to enhance critical features. Integrating SE attention with LSTM outputs boosts the model’s ability to differentiate between various NLoS categories. Experimental results show that the proposed 1D-CLANet with SE attention achieves superior performance in differentiating multiple NLoS scenarios with limited computational resources, attaining an accuracy of 95.58%. It outperforms other attention mechanisms and the version of 1D-CLANet without attention. Compared to advanced methods, the SE-enhanced 1D-CLANet significantly improves the ability to distinguish between LoS and similar NLoS scenarios, such as human obstructions, enhancing overall recognition accuracy in complex environments. Full article
(This article belongs to the Section Electrical, Electronics and Communications Engineering)
Show Figures

Figure 1

Figure 1
<p>Example of NLoS and LoS propagation in a UWB IPS.</p>
Full article ">Figure 2
<p>Example of a trilateration-based 3-anchor positioning model: (<b>a</b>) positioning under LoS conditions, (<b>b</b>) positioning under NLoS conditions.</p>
Full article ">Figure 3
<p>CIR curve from typical (<b>a</b>) LoS, (<b>b</b>) other NLoS scenarios.</p>
Full article ">Figure 4
<p>The network structure diagram of 1D-CLANet.</p>
Full article ">Figure 5
<p>The structure diagram of 1D-CNN.</p>
Full article ">Figure 6
<p>The architecture of a LSTM cell.</p>
Full article ">Figure 7
<p>The architecture of SE attention block.</p>
Full article ">Figure 8
<p>Instruments and experimental environment. The anchor point (“<math display="inline"><semantics> <mo>Δ</mo> </semantics></math>”) and tag (“<math display="inline"><semantics> <mo>□</mo> </semantics></math>”) are positioned as shown. LoS ranging positions are shown in blue and NLoS ranging positions are shown in red. (<b>a</b>) Stairway passage, (<b>b</b>) office corridor.</p>
Full article ">Figure 9
<p>Explanation of network structures: (<b>a</b>) CNSM, (<b>b</b>) ResNet without ECA.</p>
Full article ">Figure 10
<p>ROC curve for different methods in NLoS binary classification.</p>
Full article ">Figure 11
<p>Performance comparison of different methods for NLoS multi-classification.</p>
Full article ">Figure 12
<p>Confusion matrix outcomes for multiclassification. The Sce.1, Sce.2, Sce.3, Sce.4, and Sce.5 correspond to LoS, human, glass, door, and wall, respectively. (<b>a</b>) LSTM. (<b>b</b>) SVM. (<b>c</b>) HQCNN. (<b>d</b>) MLP. (<b>e</b>) CNSM. (<b>f</b>) 1D-CLANet.</p>
Full article ">
12 pages, 6271 KiB  
Article
Prediction of PM2.5 Concentration on the Basis of Multitemporal Spatial Scale Fusion
by Sihan Li, Yu Sun and Pengying Wang
Appl. Sci. 2024, 14(16), 7152; https://doi.org/10.3390/app14167152 - 14 Aug 2024
Viewed by 567
Abstract
While machine learning methods have been successful in predicting air pollution, current deep learning models usually focus only on the time-based connection of air quality monitoring stations or the complex link between PM2.5 levels and explanatory factors. Due to the lack of effective [...] Read more.
While machine learning methods have been successful in predicting air pollution, current deep learning models usually focus only on the time-based connection of air quality monitoring stations or the complex link between PM2.5 levels and explanatory factors. Due to the lack of effective integration of spatial correlation, the prediction model shows poor performance in PM2.5 prediction tasks. Predicting air pollution levels accurately over a long period is difficult because of the changing levels of correlation between past pollution levels and the future. In order to address these challenges, the study introduces a Convolutional Long Short-Term Memory (ConvLSTM) network-based neural network model with multiple feature extraction for forecasting PM2.5 levels in air quality prediction. The technique is composed of three components. The model-building process of this article is as follows: Firstly, we create a complex network layout with multiple branches to capture various temporal features at different levels. Secondly, a convolutional module was introduced to enable the model to focus on identifying neighborhood units, extracting feature scales with high spatial correlation, and helping to improve the learning ability of ConvLSTM. Next, the module for spatiotemporal fusion prediction is utilized to make predictions of PM2.5 over time and space, generating fused prediction outcomes that combine characteristics from various scales. Comparative experiments were conducted. Experimental findings indicate that the proposed approach outperforms ConvLSTM in forecasting PM2.5 concentration for the following day, three days, and seven days, resulting in a lower root mean square error (RMSE). This approach excels in modeling spatiotemporal features and is well-suited for predicting PM2.5 levels in specific regions. Full article
Show Figures

Figure 1

Figure 1
<p>The study region along with the 30 stations for monitoring air quality.</p>
Full article ">Figure 2
<p>Layout of 30 air quality monitoring stations in Jilin Province.</p>
Full article ">Figure 3
<p>Multi-branch network architecture.</p>
Full article ">Figure 4
<p>ConvLSTM structure diagram.</p>
Full article ">Figure 5
<p>Comparison of Model MAE Errors.</p>
Full article ">Figure 6
<p>Forecast outcomes from the three techniques over a 1-day prediction interval with MAE Errors. Comparison of prediction results of different models within 1-day prediction interval.</p>
Full article ">Figure 7
<p>Forecast outcomes from the three techniques over a 3-day prediction interval with MAE errors. Comparison of prediction results of different models within 3-day prediction interval.</p>
Full article ">Figure 8
<p>Forecast outcomes from the three techniques over a 7-day prediction interval with MAE Errors. Comparison of prediction results of different models within 7-day prediction interval.</p>
Full article ">
26 pages, 9128 KiB  
Article
AI-Based Visual Early Warning System
by Zeena Al-Tekreeti, Jeronimo Moreno-Cuesta, Maria Isabel Madrigal Garcia and Marcos A. Rodrigues
Informatics 2024, 11(3), 59; https://doi.org/10.3390/informatics11030059 - 12 Aug 2024
Viewed by 1052
Abstract
Facial expressions are a universally recognised means of conveying internal emotional states across diverse human cultural and ethnic groups. Recent advances in understanding people’s emotions expressed through verbal and non-verbal communication are particularly noteworthy in the clinical context for the assessment of patients’ [...] Read more.
Facial expressions are a universally recognised means of conveying internal emotional states across diverse human cultural and ethnic groups. Recent advances in understanding people’s emotions expressed through verbal and non-verbal communication are particularly noteworthy in the clinical context for the assessment of patients’ health and well-being. Facial expression recognition (FER) plays an important and vital role in health care, providing communication with a patient’s feelings and allowing the assessment and monitoring of mental and physical health conditions. This paper shows that automatic machine learning methods can predict health deterioration accurately and robustly, independent of human subjective assessment. The prior work of this paper is to discover the early signs of deteriorating health that align with the principles of preventive reactions, improving health outcomes and human survival, and promoting overall health and well-being. Therefore, methods are developed to create a facial database mimicking the underlying muscular structure of the face, whose Action Unit motions can then be transferred to human face images, thus displaying animated expressions of interest. Then, building and developing an automatic system based on convolution neural networks (CNN) and long short-term memory (LSTM) to recognise patterns of facial expressions with a focus on patients at risk of deterioration in hospital wards. This research presents state-of-the-art results on generating and modelling synthetic database and automated deterioration prediction through FEs with 99.89% accuracy. The main contributions to knowledge from this paper can be summarized as (1) the generation of visual datasets mimicking real-life samples of facial expressions indicating health deterioration, (2) improvement of the understanding and communication with patients at risk of deterioration through facial expression analysis, and (3) development of a state-of-the-art model to recognize such facial expressions using a ConvLSTM model. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Facial expression areas that reveal if the patient is under deterioration or not. (<b>a</b>) The left avatar expresses a neutral expression, which is bounded by the blue rectangles. (<b>b</b>) The right avatar reveals deterioration status in the final stage, which is bounded by the red rectangles.</p>
Full article ">Figure 2
<p>Five classes along with the combination of Action Units.</p>
Full article ">Figure 3
<p>Frames of video sample after utilizing FOMM to transfer facial expressions from avatars to real facial images.</p>
Full article ">Figure 4
<p>Samples of five classes of facial frames representing five classes.</p>
Full article ">Figure 5
<p>Facial frames samples for each class after pre-processing using face mesh as a face detection technique.</p>
Full article ">Figure 6
<p>Number and ratio of samples in each class for the whole dataset. (<b>a</b>) The total number of samples is represented by column chart. (<b>b</b>) The ratio of samples in each class.</p>
Full article ">Figure 7
<p>Number of samples in training and test dataset. (<b>a</b>) The number of samples in the training dataset. (<b>b</b>) The number of samples in the test dataset.</p>
Full article ">Figure 8
<p>Number of training samples before and after oversampling method. (<b>a</b>) Number of samples of training dataset before oversampling. (<b>b</b>) Number of samples of training dataset after oversampling.</p>
Full article ">Figure 9
<p>Structure of ConvLSTM [<a href="#B40-informatics-11-00059" class="html-bibr">40</a>].</p>
Full article ">Figure 10
<p>The proposed model architecture.</p>
Full article ">Figure 11
<p>(<b>a</b>) Evaluation metrics of model performance. Accuracy of the proposed model. (<b>b</b>) Precision of the proposed model. (<b>c</b>) Recall of the proposed model.</p>
Full article ">Figure 12
<p>Loss, Mean Square Error and Mean Absolute Error. (<b>a</b>) Loss of the predicted model. (<b>b</b>) Mean Square Error of the predicted model. (<b>c</b>) Mean Absolute Error.</p>
Full article ">Figure 13
<p>Confusion matrix.</p>
Full article ">Figure 14
<p>Evaluation of the model.</p>
Full article ">Figure 15
<p>Classification report.</p>
Full article ">Figure 16
<p>Evaluating model by Receiver Operating Characteristics Curve (ROC) Precision-Recall Curve. (<b>a</b>) ROC. (<b>b</b>) Precision-Recall Curve.</p>
Full article ">Figure 17
<p>The percentage of accuracy of model prediction for unseen data for different classes. (<b>a</b>) Accuracy of prediction of unseen data predicted as Class FD1. (<b>b</b>) Accuracy of prediction unseen data predicted as Class FD2-L. (<b>c</b>) Accuracy of prediction of unseen data predicted as Class FD2-R. (<b>d</b>) Accuracy of prediction of unseen data predicted as Class FD3-L. (<b>e</b>) Accuracy of prediction of unseen data predicted as Class FD3-R.</p>
Full article ">Figure 17 Cont.
<p>The percentage of accuracy of model prediction for unseen data for different classes. (<b>a</b>) Accuracy of prediction of unseen data predicted as Class FD1. (<b>b</b>) Accuracy of prediction unseen data predicted as Class FD2-L. (<b>c</b>) Accuracy of prediction of unseen data predicted as Class FD2-R. (<b>d</b>) Accuracy of prediction of unseen data predicted as Class FD3-L. (<b>e</b>) Accuracy of prediction of unseen data predicted as Class FD3-R.</p>
Full article ">Figure 17 Cont.
<p>The percentage of accuracy of model prediction for unseen data for different classes. (<b>a</b>) Accuracy of prediction of unseen data predicted as Class FD1. (<b>b</b>) Accuracy of prediction unseen data predicted as Class FD2-L. (<b>c</b>) Accuracy of prediction of unseen data predicted as Class FD2-R. (<b>d</b>) Accuracy of prediction of unseen data predicted as Class FD3-L. (<b>e</b>) Accuracy of prediction of unseen data predicted as Class FD3-R.</p>
Full article ">
22 pages, 15279 KiB  
Article
Reconstruction of OFDM Signals Using a Dual Discriminator CGAN with BiLSTM and Transformer
by Yuhai Li, Youchen Fan, Shunhu Hou, Yufei Niu, You Fu and Hanzhe Li
Sensors 2024, 24(14), 4562; https://doi.org/10.3390/s24144562 - 14 Jul 2024
Viewed by 681
Abstract
Communication signal reconstruction technology represents a critical area of research within communication countermeasures and signal processing. Considering traditional OFDM signal reconstruction methods’ intricacy and suboptimal reconstruction performance, a dual discriminator CGAN model incorporating LSTM and Transformer is proposed. When reconstructing OFDM signals using [...] Read more.
Communication signal reconstruction technology represents a critical area of research within communication countermeasures and signal processing. Considering traditional OFDM signal reconstruction methods’ intricacy and suboptimal reconstruction performance, a dual discriminator CGAN model incorporating LSTM and Transformer is proposed. When reconstructing OFDM signals using the traditional CNN network, it becomes challenging to extract intricate temporal information. Therefore, the BiLSTM network is incorporated into the first discriminator to capture timing details of the IQ (In-phase and Quadrature-phase) sequence and constellation map information of the AP (Amplitude and Phase) sequence. Subsequently, following the addition of fixed position coding, these data are fed into the core network constructed based on the Transformer Encoder for further learning. Simultaneously, to capture the correlation between the two IQ signals, the VIT (Vision in Transformer) concept is incorporated into the second discriminator. The IQ sequence is treated as a single-channel two-dimensional image and segmented into pixel blocks containing IQ sequence through Conv2d. Fixed position coding is added and sent to the Transformer core network for learning. The generator network transforms input noise data into a dimensional space aligned with the IQ signal and embedding vector dimensions. It appends identical position encoding information to the IQ sequence before sending it to the Transformer network. The experimental results demonstrate that, under commonly utilized OFDM modulation formats such as BPSK, QPSK, and 16QAM, the time series waveform, constellation diagram, and spectral diagram exhibit high-quality reconstruction. Our algorithm achieves improved signal quality while managing complexity compared to other reconstruction methods. Full article
(This article belongs to the Special Issue Computer Vision Recognition and Communication Sensing System)
Show Figures

Figure 1

Figure 1
<p>CGAN model architecture.</p>
Full article ">Figure 2
<p>Flowchart of the OFDM baseband communication system.</p>
Full article ">Figure 3
<p>Signal reconstruction model.</p>
Full article ">Figure 4
<p>Improved CGAN model architecture.</p>
Full article ">Figure 5
<p>Discriminator network model.</p>
Full article ">Figure 6
<p>Generator network model.</p>
Full article ">Figure 7
<p>Reconstruction results of the time-domain waveform map and constellation diagram when SNR was 20 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Original signal time-domain waveform; (<b>b</b>) Reconstruction of signal time-domain waveform; (<b>c</b>) Original signal constellation diagram; (<b>d</b>) Reconstruction of signal constellation diagram.</p>
Full article ">Figure 8
<p>Reconstruction results of time-domain waveform map and constellation diagram when SNR was 15 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Original signal time-domain waveform; (<b>b</b>) Reconstruction of signal time-domain waveform; (<b>c</b>) Original signal constellation diagram; (<b>d</b>) Reconstruction of signal constellation diagram.</p>
Full article ">Figure 9
<p>Reconstruction results of time-domain waveform map and constellation diagram when SNR was 10 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Original signal time-domain waveform; (<b>b</b>) Reconstruction of signal time-domain waveform; (<b>c</b>) Original signal constellation diagram; (<b>d</b>) Reconstruction of signal constellation diagram.</p>
Full article ">Figure 10
<p>Visual comparison of time-domain waveform of OFDM symbol sequence when the SNR was 20 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Real part; (<b>b</b>) Imaginary part.</p>
Full article ">Figure 11
<p>Visual comparison of OFDM signal spectrogram when the SNR was 20 dB. (Left to right modulation style is BPSK, QPSK, 16QAM).</p>
Full article ">Figure 12
<p>Visual comparison of time-domain waveform of OFDM symbol sequence when the SNR was 15 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Real part; (<b>b</b>) Imaginary part.</p>
Full article ">Figure 13
<p>Visual comparison of OFDM signal spectrogram when the SNR was 15 dB. (Left to right modulation style is BPSK, QPSK, 16QAM).</p>
Full article ">Figure 14
<p>Visual comparison of time-domain waveform of OFDM symbol sequence when the SNR was 10 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Real part; (<b>b</b>) Imaginary part.</p>
Full article ">Figure 15
<p>Visual comparison of OFDM signal spectrogram when the SNR was 10 dB. (Left to right modulation style is BPSK, QPSK, 16QAM).</p>
Full article ">Figure 16
<p>Probability density distribution when the SNR was 20 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Real part; (<b>b</b>) Imaginary part.</p>
Full article ">Figure 17
<p>Probability density distribution when the SNR was 15 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Real part; (<b>b</b>) Imaginary part.</p>
Full article ">Figure 18
<p>Probability density distribution when the SNR was 10 dB. (Left to right modulation style is BPSK, QPSK, 16QAM). (<b>a</b>) Real part; (<b>b</b>) Imaginary part.</p>
Full article ">
18 pages, 5522 KiB  
Article
Application of Fast MEEMD–ConvLSTM in Sea Surface Temperature Predictions
by R. W. W. M. U. P. Wanigasekara, Zhenqiu Zhang, Weiqiang Wang, Yao Luo and Gang Pan
Remote Sens. 2024, 16(13), 2468; https://doi.org/10.3390/rs16132468 - 5 Jul 2024
Viewed by 452
Abstract
Sea Surface Temperature (SST) is of great importance to study several major phenomena due to ocean interactions with other earth systems. Previous studies on SST based on statistical inference methods were less accurate for longer prediction lengths. A considerable number of studies in [...] Read more.
Sea Surface Temperature (SST) is of great importance to study several major phenomena due to ocean interactions with other earth systems. Previous studies on SST based on statistical inference methods were less accurate for longer prediction lengths. A considerable number of studies in recent years involve machine learning for SST modeling. These models were able to mitigate this problem to some length by modeling SST patterns and trends. Sequence analysis by decomposition is used for SST forecasting in several studies. Ensemble Empirical Mode Decomposition (EEMD) has been proven in previous studies as a useful method for this. The application of EEMD in spatiotemporal modeling has been introduced as Multidimensional EEMD (MEEMD). The aim of this study is to employ fast MEEMD methods to decompose the SST spatiotemporal dataset and apply a Convolutional Long Short-Term Memory (ConvLSTM)-based model to model and forecast SST. The results show that the fast MEEMD method is capable of enhancing spatiotemporal SST modeling compared to the Linear Inverse Model (LIM) and ConvLSTM model without decomposition. The model was further validated by making predictions from April to May 2023 and comparing them to original SST values. There was a high consistency between predicted and real SST values. Full article
(This article belongs to the Special Issue Artificial Intelligence and Big Data for Oceanography)
Show Figures

Figure 1

Figure 1
<p>Study area.</p>
Full article ">Figure 2
<p>Normalized SST data processing workflow.</p>
Full article ">Figure 3
<p>SST Variation at a single random instance.</p>
Full article ">Figure 4
<p>Experimenting the optimal train–test split ratio: the figure represents training (blue) and validation (red) errors of two experiments in train–test ratio split modeling. Generally, the validation error should be lower than the training error in modeling. This allows a model to avoid overfitting to the training data and underfitting to the testing data. This way the model will make more reliable predictions.</p>
Full article ">Figure 5
<p>Fast MEEMD decomposed IMFs and residual functions visualized for the training (<b>left</b>) and testing (<b>right</b>) sets, respectively. The figure represents the SST dataset after preprocessing steps. The sum of all the IMFs and residual functions should, theoretically, create the exact same normalized SST time series as the original normalized SST time series it was generated from. However, in application there exists a small error called the reconstruction error [<a href="#B40-remotesensing-16-02468" class="html-bibr">40</a>]. This error is mostly negligible.</p>
Full article ">Figure 6
<p>Fast MEEMD–ConvLSTM model prediction accuracy metrics for different prediction lengths are illustrated in this figure. The model prediction performance metrics MAE ranges from 0.24–0.94 and RMSE ranges from 0.31–1.2 for 1–4-weeks-ahead prediction lengths.</p>
Full article ">Figure 7
<p>MAE of SST predictions before 2016 shows that the prediction performance near the Andaman and Nicobar Islands (near 10°N, 90°E) is slightly higher compared to the prediction performance that includes MHWs near this location. However, the removal of data from 2016–2023 seems to have an adverse effect on the model performance.</p>
Full article ">Figure 8
<p>RMSE error metric for SST predictions before 2016 indicates a lower error near the Andaman and Nicobar Islands (near 10°N, 90°E) compared to SST prediction capabilities that include the 2016 MHW near the area. As before, the removal of data for 2016–2023 period seems to have an adverse effect on the model capabilities, resulting in higher errors in some other areas.</p>
Full article ">Figure 9
<p>SST Prediction MAE (<b>left</b>) and RMSE (<b>right</b>) mean value comparison between different models is indicated in the figure. The fast MEEMD-based model has lower error compared to the ConvLSTM model without fast MEEMD decomposition. However, the EOF and Lasso regression-employed LIM model performed better for 1–3-weeks-ahead predictions. For 4-weeks-lead prediction the fast MEEMD–ConvLSTM model performed better than both of the other models.</p>
Full article ">Figure 10
<p>SST prediction vs true for the time period from 23th April 2023 to 21th May 2023 period. For all predictions the true and predicted values have slight variations that increase over prediction length.</p>
Full article ">
18 pages, 2578 KiB  
Article
Prediction on Demand for Regional Online Car-Hailing Travel Based on Self-Attention Memory and ConvLSTM
by Jianqi Li, Wenbao Zeng, Weiqi Liu and Rongjun Cheng
Sustainability 2024, 16(13), 5725; https://doi.org/10.3390/su16135725 - 4 Jul 2024
Viewed by 610
Abstract
High precision in forecasting travel demand for online car-hailing is crucial for traffic management to schedule vehicles, hence reducing energy consumption and achieving sustainable development. Netflix demand forecasting relies on the capture of spatiotemporal correlations. To extract the spatiotemporal information more fully, this [...] Read more.
High precision in forecasting travel demand for online car-hailing is crucial for traffic management to schedule vehicles, hence reducing energy consumption and achieving sustainable development. Netflix demand forecasting relies on the capture of spatiotemporal correlations. To extract the spatiotemporal information more fully, this study designs and develops a novel spatiotemporal prediction model with multidimensional inputs (MSACL) by embedding a self-attention memory (SAM) module into a convolutional long short-term memory neural network (ConvLSTM). The SAM module can extract features with long-range spatiotemporal dependencies. The experimental data are derived from the Chengdu City online car-hailing trajectory data set and the external factors data set. Comparative experiments demonstrate that the proposed model has higher accuracy. The proposed model outperforms the Sa-ConvLSTM model and has the highest prediction accuracy, shows a reduction in the mean absolute error (MAE) by 1.72, a reduction in the mean squared error (MSE) by 0.43, and an increase in the R-squared (R2) by 4%. In addition, ablation experiments illustrate the effectiveness of each component, where the external factor inputs have the least impact on the model accuracy, but the removal of the SAM module results in the most significant decrease in model accuracy. Full article
(This article belongs to the Special Issue Sustainable Transportation and Data Science Application)
Show Figures

Figure 1

Figure 1
<p>Unit structure of ConvLSTM.</p>
Full article ">Figure 2
<p>Unit structure of Sa-ConvLSTM.</p>
Full article ">Figure 3
<p>The structure of the SAM module.</p>
Full article ">Figure 4
<p>The structure of the MSACL.</p>
Full article ">Figure 5
<p>Comparison of model performance.</p>
Full article ">Figure 6
<p>Model prediction results: (<b>a</b>) ConvLSTM; (<b>b</b>) Sa-ConvLSTM; (<b>c</b>) MSACL; (<b>d</b>) true value.</p>
Full article ">
27 pages, 5261 KiB  
Article
Convformer: A Model for Reconstructing Ocean Subsurface Temperature and Salinity Fields Based on Multi-Source Remote Sensing Observations
by Tao Song, Guangxu Xu, Kunlin Yang, Xin Li and Shiqiu Peng
Remote Sens. 2024, 16(13), 2422; https://doi.org/10.3390/rs16132422 - 1 Jul 2024
Cited by 1 | Viewed by 887
Abstract
Observational data on ocean subsurface temperature and salinity are patently insufficient because in situ observations are complex and costly, while satellite remote-sensed measurements are abundant but mainly focus on sea surface data. To make up for the ocean interior data shortage and entirely [...] Read more.
Observational data on ocean subsurface temperature and salinity are patently insufficient because in situ observations are complex and costly, while satellite remote-sensed measurements are abundant but mainly focus on sea surface data. To make up for the ocean interior data shortage and entirely use the abundant satellite data, we developed a data-driven deep learning model named Convformer to reconstruct ocean subsurface temperature and salinity fields from satellite-observed sea surface data. Convformer is designed by deeply optimizing Vision Transformer and ConvLSTM, consisting of alternating residual connections between multiple temporal and spatial attention blocks. The input variables consist of sea surface temperature (SST), sea surface salinity (SSS), sea surface height (SSH), and sea surface wind (SSW). Our results demonstrate that Convformer exhibits superior performance in estimating the temperature-salinity structure of the tropical Pacific Ocean. The all-depth average root mean square error (RMSE) of the reconstructed subsurface temperature (ST)/subsurface salinity (SS) is 0.353 °C/0.0695 PSU, with correlation coefficients (R²) of 0.98663/0.99971. In the critical thermocline, although the root mean square errors of ST and SS reach 0.85 °C and 0.121 PSU, respectively, they remain smaller compared to other models. Furthermore, we assessed Convformer’s performance from various perspectives. Notably, we also delved into the potential of Convformer to extract physical and dynamic information from a model mechanism perspective. Our study offers a practical approach to reconstructing the subsurface temperature and salinity fields from satellite-observed sea surface data. Full article
(This article belongs to the Special Issue Artificial Intelligence and Big Data for Oceanography)
Show Figures

Figure 1

Figure 1
<p>Schematic diagram and partitioning of the study area. The study area is located in the central Pacific region (160°E–120°W, 30°S–30°N); the Argo buoys No. 5905969, 6902701, and 6902909 are used for vertical profiles verification.</p>
Full article ">Figure 2
<p>The architecture of Convformer and flowchart of Pacific ST/SS reconstruction using remote sensing data.</p>
Full article ">Figure 3
<p>Vertical RMSE profiles for each model from 50 m down to 1500 m of ST: (<b>a</b>) models comparison; (<b>b</b>–<b>e</b>) each model respectively.</p>
Full article ">Figure 4
<p>Bar graph for each model from 50 m down to 1900 m of ST. The legend colors correspond to various models, while the symbols indicate the correlation coefficients’ values for each model at a given depth.</p>
Full article ">Figure 5
<p>Vertical RMSE profiles for each model from 50 m down to 1500 m of SS: (<b>a</b>) models comparison; (<b>b</b>–<b>e</b>) each model respectively.</p>
Full article ">Figure 6
<p>Bar graph for each model from 50 m down to 1900 m of SS. The legend colors correspond to various models, while the symbols indicate the correlation coefficients’ values for each model at a given depth.</p>
Full article ">Figure 7
<p>Vertical RMSE profiles for Convformer with different remote sensing inputs schemes from 50 m down to 1500 m: (<b>a</b>) RMSE of ST; (<b>b</b>) RMSE of SS; (<b>c</b>) heat map of correlation coefficients.</p>
Full article ">Figure 8
<p>Vertical RMSE profiles for Convformer with different components in ablation study from 50 m down to 1500 m of ST: (<b>a</b>) models comparison; (<b>b</b>–<b>f</b>) each model respectively.</p>
Full article ">Figure 9
<p>RMSE of the ST estimation at each depth level and in each month of 2018.</p>
Full article ">Figure 10
<p>RMSE of the SS estimation at each depth level and in each month of 2018.</p>
Full article ">Figure 11
<p>The vertical profiles of the ocean subsurface temperature and salinity fields in March 2018 at the longitude of 188.5° for (<b>a</b>) Argo ST, (<b>b</b>) Argo SS, (<b>c</b>) Convformer-reconstructed ST, and (<b>d</b>) Convformer-reconstructed SS.</p>
Full article ">Figure 12
<p>Distribution of temperature field estimated by the (<b>i</b>) Convformer compared to the (<b>ii</b>) Argo ST and their (<b>iii</b>) difference (dS = STConvfomer − STArgo) at the depth of (<b>a</b>) 50 m, (<b>b</b>) 100 m, (<b>c</b>) 300 m, (<b>d</b>) 600 m, and (<b>e</b>) 1000 m.</p>
Full article ">Figure 13
<p>Distribution of salinity field estimated by the (<b>i</b>) Convformer compared to the (<b>ii</b>) Argo ST and their (<b>iii</b>) difference (dS = SSConvfomer − SSArgo) at the depth of (<b>a</b>) 50 m, (<b>b</b>) 100 m, (<b>c</b>) 300 m, (<b>d</b>) 600 m, and (<b>e</b>) 1000 m.</p>
Full article ">
18 pages, 10425 KiB  
Article
Simulation of Full Wavefield Data with Deep Learning Approach for Delamination Identification
by Saeed Ullah, Pawel Kudela, Abdalraheem A. Ijjeh, Eleni Chatzi and Wieslaw Ostachowicz
Appl. Sci. 2024, 14(13), 5438; https://doi.org/10.3390/app14135438 - 23 Jun 2024
Viewed by 536
Abstract
In this work, a novel approach of guided wave-based damage identification in composite laminates is proposed. The novelty of this research lies in the implementation of ConvLSTM-based autoencoders for the generation of full wavefield data of propagating guided waves in composite structures. The [...] Read more.
In this work, a novel approach of guided wave-based damage identification in composite laminates is proposed. The novelty of this research lies in the implementation of ConvLSTM-based autoencoders for the generation of full wavefield data of propagating guided waves in composite structures. The developed surrogate deep learning model takes as input full wavefield frames of propagating waves in a healthy plate, along with a binary image representing delamination, and predicts the frames of propagating waves in a plate, which contains single delamination. The evaluation of the surrogate model is ultrafast (less than 1 s). Therefore, unlike traditional forward solvers, the surrogate model can be employed efficiently in the inverse framework of damage identification. In this work, particle swarm optimisation is applied as a suitable tool to this end. The proposed method was tested on a synthetic dataset, thus showing that it is capable of estimating the delamination location and size with good accuracy. The test involved full wavefield data in the objective function of the inverse method, but it should be underlined as well that partial data with measurements can be implemented. This is extremely important for practical applications in structural health monitoring where only signals at a finite number of locations are available. Full article
Show Figures

Figure 1

Figure 1
<p>Flowchart of the proposed inverse method for damage identification.</p>
Full article ">Figure 2
<p>Numerical and experimental setup showing the types of possible delaminations (embedded and opened at the edge), piezoelectric transducer (PZT), and scanning laser Doppler vibrometer (SLDV).</p>
Full article ">Figure 3
<p>An exemplary mesh containing piezoelectric transducer (red) and random delamination (green) used for Lamb wave propagation modelling.</p>
Full article ">Figure 4
<p>The flowchart of the proposed DL model.</p>
Full article ">Figure 5
<p>The architecture of the proposed ConvLSTM autoencoder model.</p>
Full article ">Figure 6
<p>The training and validation PSNR metric as a function of the epoch.</p>
Full article ">Figure 7
<p>First scenario: comparison of predicted frames with the label frames at 10th, 20th, and 30th frames after the interaction with delamination.</p>
Full article ">Figure 8
<p>Second scenario: comparison of predicted frames with the label frames at 10th, 20th, and 30th frames after the interaction with delamination.</p>
Full article ">Figure 9
<p>Third scenario: comparison of predicted frames with the label frames at 10th, 20th, and 30th frames after the interaction with delamination.</p>
Full article ">Figure 10
<p>Delamination identification results: green—ground truth, red—prediction, yellow—intersection.</p>
Full article ">
22 pages, 5466 KiB  
Article
A Hybrid Convolutional–Long Short-Term Memory–Attention Framework for Short-Term Photovoltaic Power Forecasting, Incorporating Data from Neighboring Stations
by Feng Hu, Linghua Zhang and Jiaqi Wang
Appl. Sci. 2024, 14(12), 5189; https://doi.org/10.3390/app14125189 - 14 Jun 2024
Viewed by 480
Abstract
To enhance the safety of grid operations, this paper proposes a high-precision short-term photovoltaic (PV) power forecasting method that integrates information from surrounding PV stations and deep learning prediction models. The proposed method utilizes numerical weather prediction (NWP) data of the target PV [...] Read more.
To enhance the safety of grid operations, this paper proposes a high-precision short-term photovoltaic (PV) power forecasting method that integrates information from surrounding PV stations and deep learning prediction models. The proposed method utilizes numerical weather prediction (NWP) data of the target PV station and highly correlated features from nearby stations as inputs. This study first analyzes the correlation between irradiance and power sequences and calculates a comprehensive similarity index based on distance factors. Stations with high-similarity indices are selected as data sources. Subsequently, Bayesian optimization is employed to determine the optimal data fusion ratio. The selected data are then used to model power predictions through the convolutional long short-term memory with attention (Conv-LSTM-ATT) deep neural network. Experimental results show that the proposed model significantly outperforms three classical models in terms of forecasting accuracy. The data fusion strategy determined by Bayesian optimization reduces the root mean square error (RMSE) of the test set by 20.04%, 28.24%, and 30.94% under sunny, cloudy, and rainy conditions, respectively. Full article
Show Figures

Figure 1

Figure 1
<p>The spatiotemporal data-driven framework for photovoltaic (PV) power forecasting.</p>
Full article ">Figure 2
<p>Short-term photovoltaic (PV) power forecasting model based on deep learning.</p>
Full article ">Figure 3
<p>Structure diagram of CNN.</p>
Full article ">Figure 4
<p>The LSTM neuron structure.</p>
Full article ">Figure 5
<p>The attention-based Conv-LSTM module.</p>
Full article ">Figure 6
<p>Attention and the Conv-LSTM network.</p>
Full article ">Figure 7
<p>PV power generation data in three dimensions.</p>
Full article ">Figure 8
<p>Distribution diagram of weather types based on K-Means++ clustering.</p>
Full article ">Figure 9
<p>Forecast result under sunny weather.</p>
Full article ">Figure 10
<p>Forecast result under cloudy weather.</p>
Full article ">Figure 11
<p>Forecast result under rainy weather.</p>
Full article ">Figure 12
<p>Forecast result under sunny weather.</p>
Full article ">Figure 13
<p>Forecast result under cloudy weather.</p>
Full article ">Figure 14
<p>Forecast result under rainy weather.</p>
Full article ">
15 pages, 4923 KiB  
Article
Research on Grain Futures Price Prediction Based on a Bi-DSConvLSTM-Attention Model
by Bensheng Yun, Jiannan Lai, Yingfeng Ma and Yanan Zheng
Systems 2024, 12(6), 204; https://doi.org/10.3390/systems12060204 - 11 Jun 2024
Viewed by 792
Abstract
Grain is a commodity related to the livelihood of the nation’s people, and the volatility of its futures price affects risk management, investment decisions, and policy making. Therefore, it is very necessary to establish an accurate and efficient futures price prediction model. Aiming at [...] Read more.
Grain is a commodity related to the livelihood of the nation’s people, and the volatility of its futures price affects risk management, investment decisions, and policy making. Therefore, it is very necessary to establish an accurate and efficient futures price prediction model. Aiming at improving the accuracy and efficiency of the prediction model, so as to support reasonable decision making, this paper proposes a Bi-DSConvLSTM-Attention model for grain futures price prediction, which is based on the combination of a bidirectional long short-term memory neural network (BiLSTM), a depthwise separable convolutional long short-term memory neural network (DSConvLSTM), and an attention mechanism. Firstly, the mutual information is used to evaluate, sort, and select the features for dimension reduction. Secondly, the lightweight depthwise separable convolution (DSConv) is introduced to replace the standard convolution (SConv) in ConvLSTM without sacrificing its performance. Then, the self-attention mechanism is adopted to improve the accuracy. Finally, taking the wheat futures price prediction as an example, the model is trained and its performance is evaluated. Under the Bi-DSConvLSTM-Attention model, the experimental results of selecting the most relevant 1, 2, 3, 4, 5, 6, and 7 features as the inputs showed that the optimal number of features to be selected was 4. When the four best features were selected as the inputs, the RMSE, MAE, MAPE, and R2 of the prediction result of the Bi-DSConvLSTM-Attention model were 5.61, 3.63, 0.55, and 0.9984, respectively, which is a great improvement compared with the existing price-prediction models. Other experimental results demonstrated that the model also possesses a certain degree of generalization and is capable of obtaining positive returns. Full article
Show Figures

Figure 1

Figure 1
<p>Process of price prediction model.</p>
Full article ">Figure 2
<p>Comparison of standard convolution and depthwise separable convolution.</p>
Full article ">Figure 3
<p>Bi-DSConvLSTM-Attention model architecture.</p>
Full article ">Figure 4
<p>LSTM model structure.</p>
Full article ">Figure 5
<p>Ranking of mutual information values between some features and wheat futures price.</p>
Full article ">Figure 6
<p>Wheat futures price prediction.</p>
Full article ">Figure 7
<p>Soybean futures price prediction.</p>
Full article ">
33 pages, 2156 KiB  
Article
Identification of Optimal Data Augmentation Techniques for Multimodal Time-Series Sensory Data: A Framework
by Nazish Ashfaq, Muhammad Hassan Khan and Muhammad Adeel Nisar
Information 2024, 15(6), 343; https://doi.org/10.3390/info15060343 - 11 Jun 2024
Viewed by 1029
Abstract
Recently, the research community has shown significant interest in the continuous temporal data obtained from motion sensors in wearable devices. These data are useful for classifying and analysing different human activities in many application areas such as healthcare, sports and surveillance. The literature [...] Read more.
Recently, the research community has shown significant interest in the continuous temporal data obtained from motion sensors in wearable devices. These data are useful for classifying and analysing different human activities in many application areas such as healthcare, sports and surveillance. The literature has presented a multitude of deep learning models that aim to derive a suitable feature representation from temporal sensory input. However, the presence of a substantial quantity of annotated training data is crucial to adequately train the deep networks. Nevertheless, the data originating from the wearable devices are vast but ineffective due to a lack of labels which hinders our ability to train the models with optimal efficiency. This phenomenon leads to the model experiencing overfitting. The contribution of the proposed research is twofold: firstly, it involves a systematic evaluation of fifteen different augmentation strategies to solve the inadequacy problem of labeled data which plays a critical role in the classification tasks. Secondly, it introduces an automatic feature-learning technique proposing a Multi-Branch Hybrid Conv-LSTM network to classify human activities of daily living using multimodal data of different wearable smart devices. The objective of this study is to introduce an ensemble deep model that effectively captures intricate patterns and interdependencies within temporal data. The term “ensemble model” pertains to fusion of distinct deep models, with the objective of leveraging their own strengths and capabilities to develop a solution that is more robust and efficient. A comprehensive assessment of ensemble models is conducted using data-augmentation techniques on two prominent benchmark datasets: CogAge and UniMiB-SHAR. The proposed network employs a range of data-augmentation methods to improve the accuracy of atomic and composite activities. This results in a 5% increase in accuracy for composite activities and a 30% increase for atomic activities. Full article
(This article belongs to the Special Issue Human Activity Recognition and Biomedical Signal Processing)
Show Figures

Figure 1

Figure 1
<p>An illustration exhibiting the process of detecting human actions utilising raw sensory data using handcrafted feature-based encoding approaches.</p>
Full article ">Figure 2
<p>An illustration exhibiting the process of detecting human actions utilising raw sensory data using codebook feature-based encoding approaches.</p>
Full article ">Figure 3
<p>An illustration exhibiting the process of detecting human actions utilising raw sensory data using an automatic feature-learning approach.</p>
Full article ">Figure 4
<p>Flow of activity classification performed with our proposed approach.</p>
Full article ">Figure 5
<p>A visual representation of actual and transformed data after applying jittering on a CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 6
<p>A visual representation of actual data and transformed data after applying scaling on a CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 7
<p>A visual representation of actual data and transformed data after applying time warping on CogAge atomic (bending) activity. The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 8
<p>A visual representation of actual data and transformed data after applying linear interpolation on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 9
<p>A visual representation of actual data and transformed data after applying exponential moving median smoothing on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 10
<p>A visual representation of actual data and transformed data after applying channel permutation on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 11
<p>A visual representation of actual data and transformed data after applying rolling window averaging on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 12
<p>A visual representation of actual data and transformed data after applying sub averaging on the CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 13
<p>A visual representation of actual data and transformed data after applying sub cutmix on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 14
<p>A visual representation of actual data and transformed data after applying AugMix on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 15
<p>A visual representation of actual data and transformed data after applying mixup on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 16
<p>A visual representation of actual data and transformed data after applying cutmix on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 17
<p>A visual representation of actual data and transformed data after applying hide and seek on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 18
<p>A visual representation of actual data and transformed data after applying tsaug on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 19
<p>A visual representation of actual data and transformed data after applying sequential transformation on CogAge atomic dataset (bending activity). The <span class="html-italic">X</span>-axis represents time in milliseconds, while the <span class="html-italic">y</span>-axis represents data sequences for bending activity.</p>
Full article ">Figure 20
<p>A visual representation showcasing CNN architecture comprising of convolution, pooling and dense layer.</p>
Full article ">Figure 21
<p>A visual representation showcasing the internal arrangement of the LSTM architecture.</p>
Full article ">Figure 22
<p>A visual representation showcasing the architecture of the proposed ensemble model.</p>
Full article ">Figure 23
<p>A visual representation of the distribution of samples across various atomic activities in the CogAge dataset. The <span class="html-italic">X</span>-axis represents example counts for each activity, while the <span class="html-italic">Y</span>-axis represents activities.</p>
Full article ">Figure 24
<p>A visual representation of the distribution of samples across various composite activities in the CogAge dataset. The <span class="html-italic">X</span>-axis represents example counts for each activity, while the <span class="html-italic">Y</span>-axis represents activities.</p>
Full article ">Figure 25
<p>A visual representation of the distribution of samples across various composite activities in the UniMiB-SHAR dataset. The <span class="html-italic">X</span>-axis represents example counts for each activity, while the <span class="html-italic">Y</span>-axis represents activities.</p>
Full article ">Figure 26
<p>Training and validation accuracy graph of actual data for CogAge atomic activities.</p>
Full article ">Figure 27
<p>Training and validation accuracy graph of augmented data for CogAge atomic activities.</p>
Full article ">
Back to TopTop