[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (6,369)

Search Parameters:
Keywords = LSTM

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
31 pages, 6207 KiB  
Article
A Distributed VMD-BiLSTM Model for Taxi Demand Forecasting with GPS Sensor Data
by Hasan A. H. Naji, Qingji Xue and Tianfeng Li
Sensors 2024, 24(20), 6683; https://doi.org/10.3390/s24206683 (registering DOI) - 17 Oct 2024
Abstract
With the ubiquitous deployment of mobile and sensor technologies in modes of transportation, taxis have become a significant component of public transportation. However, vacant taxis represent an important waste of transportation resources. Forecasting taxi demand within a short time achieves a supply–demand balance [...] Read more.
With the ubiquitous deployment of mobile and sensor technologies in modes of transportation, taxis have become a significant component of public transportation. However, vacant taxis represent an important waste of transportation resources. Forecasting taxi demand within a short time achieves a supply–demand balance and reduces oil emissions. Although earlier studies have forwarded highly developed machine learning- and deep learning-based models to forecast taxicab demands, these models often face significant computational expenses and cannot effectively utilize large-scale trajectory sensor data. To address these challenges, in this paper, we propose a hybrid deep learning-based model for taxi demand prediction. In particular, the Variational Mode Decomposition (VMD) algorithm is integrated along with a Bidirectional Long Short-Term Memory (BiLSTM) model to perform the prediction process. The VMD algorithm is applied to decompose time series-aware traffic features into multiple sub-modes of different frequencies. After that, the BiLSTM method is utilized to predict time series data fed with the relevant demand features. To overcome the limitation of high computational expenses, the designed model is performed on the Spark distributed platform. The performance of the proposed model is tested using a real-world dataset, and it surpasses existing state-of-the-art predictive models in terms of accuracy, efficiency, and distributed performance. These findings provide insights for enhancing the efficiency of passenger search and increasing the profit of taxicabs. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

Figure 1
<p>The structure of the distributed VMD-BiLSTM prediction model.</p>
Full article ">Figure 2
<p>Road map network of Wuhan City.</p>
Full article ">Figure 3
<p>Typical trajectory of taxi trips.</p>
Full article ">Figure 4
<p>Study area of Wuchang district.</p>
Full article ">Figure 5
<p>The distribution of taxi demands on the weekdays and weekends.</p>
Full article ">Figure 6
<p>Taxi demand distribution in the target area during holidays.</p>
Full article ">Figure 7
<p>The distribution of taxi demands in the target area over 24 h.</p>
Full article ">Figure 8
<p>Schematic diagram of the VMD-BiLSTM model.</p>
Full article ">Figure 9
<p>Flowchart of VMD algorithm.</p>
Full article ">Figure 10
<p>The transformation of the taxi demands time series into a two-dimensional array.</p>
Full article ">Figure 11
<p>Architecture of bidirectional LSTM network.</p>
Full article ">Figure 12
<p>Distributed implementation of VMD-BiLSTM model on Spark.</p>
Full article ">Figure 13
<p>Results of our prediction model on Wuhan’s dataset. (<b>a</b>) 1 day; (<b>b</b>) 1 week; (<b>c</b>) 2 weeks; (<b>d</b>) 1 month; (<b>e</b>) 2 months; and (<b>f</b>) whole dataset.</p>
Full article ">Figure 13 Cont.
<p>Results of our prediction model on Wuhan’s dataset. (<b>a</b>) 1 day; (<b>b</b>) 1 week; (<b>c</b>) 2 weeks; (<b>d</b>) 1 month; (<b>e</b>) 2 months; and (<b>f</b>) whole dataset.</p>
Full article ">Figure 14
<p>VMD renderings.</p>
Full article ">Figure 14 Cont.
<p>VMD renderings.</p>
Full article ">Figure 15
<p>Wavelet threshold denoising method for VMD renderings (IMF1, IMF2, and IMF3).</p>
Full article ">Figure 15 Cont.
<p>Wavelet threshold denoising method for VMD renderings (IMF1, IMF2, and IMF3).</p>
Full article ">Figure 16
<p>Box-plot of MOEs for Wuhan dataset.</p>
Full article ">Figure 17
<p>Comparison of loss function of distributed VMD-BiLSM.</p>
Full article ">Figure 18
<p>Running time (seconds) of VMD-BiLSTM based on Spark platform.</p>
Full article ">Figure 19
<p>Scaleup comparative analysis of distributed VMD-BiLSTM for different computing nodes.</p>
Full article ">Figure 20
<p>Speedup comparative analysis of the proposed model for different computing nodes.</p>
Full article ">
17 pages, 4313 KiB  
Article
D3AT-LSTM: An Efficient Model for Spatiotemporal Temperature Prediction Based on Attention Mechanisms
by Ting Tian, Huijing Wu, Xianhua Liu and Qiao Hu
Electronics 2024, 13(20), 4089; https://doi.org/10.3390/electronics13204089 (registering DOI) - 17 Oct 2024
Abstract
Accurate temperature prediction is essential for economic production and human society’s daily life. However, most current methods only focus on time-series temperature modeling and prediction, ignoring the complex interplay of meteorological variables in the spatial domain. In this paper, a novel temperature prediction [...] Read more.
Accurate temperature prediction is essential for economic production and human society’s daily life. However, most current methods only focus on time-series temperature modeling and prediction, ignoring the complex interplay of meteorological variables in the spatial domain. In this paper, a novel temperature prediction model (D3AT-LSTM) is proposed by combining the three-dimensional convolutional neural network (3DCNN) and the attention-based gated cyclic network. Firstly, the historical meteorological series of eight surrounding pixels are combined to construct a multi-dimensional feature tensor that integrates variables from the temporal domain as the input data. Convolutional units are used to model and analyze the spatiotemporal patterns of the local sequence in CNN modules by combining them with parallel attention mechanisms. The fully connected layer finally makes the final temperature prediction. This method is subsequently compared with both classical and state-of-art prediction models such as ARIMA (AR), long short-term memory network (LSTM), and Transformer using three indices: the root mean square error (RMSE), the mean absolute error (MAE), and the coefficient of determination (R2). The results indicate that the D3AT-LSTM model can achieve good prediction accuracy compared to AR, LSTMs, and Transformer. Full article
Show Figures

Figure 1

Figure 1
<p>D3AT-LSTM network.</p>
Full article ">Figure 2
<p>3DCNN blocks.</p>
Full article ">Figure 3
<p>Convolution modules combined with parallel attention branch.</p>
Full article ">Figure 4
<p>Correlation coefficient between meteorological characteristics and air temperature.</p>
Full article ">Figure 5
<p>Supervoxel reconstruction process of meteorological characteristics.</p>
Full article ">Figure 6
<p>Search area of Station1–5.</p>
Full article ">Figure 7
<p>The accuracy of comparison experiment per hour.</p>
Full article ">Figure 8
<p>Thermal distribution of temperature predicted by each model.</p>
Full article ">Figure 9
<p>Accuracy curve of predicted accuracy at each time step of ablation experiment.</p>
Full article ">
17 pages, 4709 KiB  
Article
Top-Oil Temperature Prediction of Power Transformer Based on Long Short-Term Memory Neural Network with Self-Attention Mechanism Optimized by Improved Whale Optimization Algorithm
by Dexu Zou, He Xu, Hao Quan, Jianhua Yin, Qingjun Peng, Shan Wang, Weiju Dai and Zhihu Hong
Symmetry 2024, 16(10), 1382; https://doi.org/10.3390/sym16101382 (registering DOI) - 17 Oct 2024
Abstract
The operational stability of the power transformer is essential for maintaining the symmetry, balance, and security of power systems. Once the power transformer fails, it will lead to heightened instability within grid operations. Accurate prediction of oil temperature is crucial for efficient transformer [...] Read more.
The operational stability of the power transformer is essential for maintaining the symmetry, balance, and security of power systems. Once the power transformer fails, it will lead to heightened instability within grid operations. Accurate prediction of oil temperature is crucial for efficient transformer operation. To address challenges such as the difficulty in selecting model hyperparameters and incomplete consideration of temporal information in transformer oil temperature prediction, a novel model is constructed based on the improved whale optimization algorithm (IWOA) and long short-term memory (LSTM) neural network with self-attention (SA) mechanism. To incorporate holistic and local information, the SA is integrated with the LSTM model. Furthermore, the IWOA is employed in the optimization of the hyper-parameters for the LSTM-SA model. The standard IWOA is improved by incorporating adaptive parameters, thresholds, and a Latin hypercube sampling initialization strategy. The proposed method was applied and tested using real operational data from two transformers within a practical power grid. The results of the single-step prediction experiments demonstrate that the proposed method significantly improves the accuracy of oil temperature prediction for power transformers, with enhancements ranging from 1.06% to 18.85% compared to benchmark models. Additionally, the proposed model performs effectively across various prediction steps, consistently outperforming benchmark models. Full article
(This article belongs to the Special Issue Symmetry/Asymmetry Studies in Modern Power Systems)
Show Figures

Figure 1

Figure 1
<p>The basic construction of an oil-immersed transformer.</p>
Full article ">Figure 2
<p>Flow chart of IWOA-LSTM-SA.</p>
Full article ">Figure 3
<p>LSTM structure diagram.</p>
Full article ">Figure 4
<p>LSTM-SA structure.</p>
Full article ">Figure 5
<p>Flow chart of the IWOA.</p>
Full article ">Figure 6
<p>Average convergence curves for each algorithm.</p>
Full article ">Figure 6 Cont.
<p>Average convergence curves for each algorithm.</p>
Full article ">Figure 7
<p>The prediction results of IWOA-LSTM-SA.</p>
Full article ">Figure 8
<p>Training and testing errors over iterations.</p>
Full article ">Figure 9
<p>Performance comparison across models.</p>
Full article ">Figure 10
<p>Model residuals.</p>
Full article ">Figure 11
<p>Multi-step prediction performance comparison across models (one week).</p>
Full article ">
21 pages, 1156 KiB  
Article
EDSCVD: Enhanced Dual-Channel Smart Contract Vulnerability Detection Method
by Huaiguang Wu, Yibo Peng, Yaqiong He and Siqi Lu
Symmetry 2024, 16(10), 1381; https://doi.org/10.3390/sym16101381 - 17 Oct 2024
Viewed by 80
Abstract
Ensuring the absence of vulnerabilities or flaws in smart contracts before their deployment is crucial for the smooth progress of subsequent work. Existing detection methods heavily rely on expert rules, resulting in low robustness and accuracy. Therefore, we propose EDSCVD, an enhanced deep [...] Read more.
Ensuring the absence of vulnerabilities or flaws in smart contracts before their deployment is crucial for the smooth progress of subsequent work. Existing detection methods heavily rely on expert rules, resulting in low robustness and accuracy. Therefore, we propose EDSCVD, an enhanced deep learning vulnerability detection model based on dual-channel networks. Firstly, the contract fragments are preprocessed by BERT into the required word embeddings. Next, we utilized adversarial training FGM to the word embeddings to generate perturbations, thereby producing symmetric adversarial samples and enhancing the robustness of the model. Then, the dual-channel model combining BiLSTM and CNN is utilized for feature training to obtain more comprehensive and symmetric information on temporal and local contract features.Finally, the combined output features are passed through a classifier to classify and detect contract vulnerabilities. Experimental results show that our EDSCVD exhibits excellent detection performance in the detection of classical reentrancy vulnerabilities, timestamp dependencies, and integer overflow vulnerabilities. Full article
Show Figures

Figure 1

Figure 1
<p>Reentrancy source code. (Source: Own elaboration).</p>
Full article ">Figure 2
<p>Timestamp Dependency source code. (Source: Own elaboration).</p>
Full article ">Figure 3
<p>Integer Overflow source code. (∗ denotes the multiplication operator. <math display="inline"><semantics> <mrow> <mo>&amp;</mo> <mo>&amp;</mo> </mrow> </semantics></math> denotes a logical symbol used to combine two Boolean expressions). (Source: Own elaboration).</p>
Full article ">Figure 4
<p>The overall architecture of EDSCVD. (Source: Own elaboration).</p>
Full article ">Figure 5
<p>Contract fragment representation. (Source: Own elaboration).</p>
Full article ">Figure 6
<p>The structure of BERT. (Source: Own elaboration based on literature [<a href="#B49-symmetry-16-01381" class="html-bibr">49</a>]).</p>
Full article ">Figure 7
<p>Adversarial Training Methods FGM. (Source: Own elaboration).</p>
Full article ">Figure 8
<p>Dual-Channel Network Architecture. (Source: Own elaboration based on literature [<a href="#B49-symmetry-16-01381" class="html-bibr">49</a>]).</p>
Full article ">Figure 9
<p>Structure of a single LSTM module. (Source: Own elaboration based on literature [<a href="#B22-symmetry-16-01381" class="html-bibr">22</a>]).</p>
Full article ">Figure 10
<p>Multi-Head Attention Mechanisms. (Source: Own elaboration).</p>
Full article ">Figure 11
<p>Epochs and Evaluation Metrics in model training. (Source: Own elaboration).</p>
Full article ">
17 pages, 4788 KiB  
Article
Benchmark Dataset for Offshore Platform Motion Prediction and Its Applications
by Wenyin Pan, Xiaoxian Guo and Xin Li
J. Mar. Sci. Eng. 2024, 12(10), 1852; https://doi.org/10.3390/jmse12101852 - 17 Oct 2024
Viewed by 96
Abstract
The accurate prediction of offshore platform and ship motion is crucial for motion compensation devices and for helping the crew make informed decisions. Traditional time series and physical models are being replaced by machine learning models due to their simplicity and lower training [...] Read more.
The accurate prediction of offshore platform and ship motion is crucial for motion compensation devices and for helping the crew make informed decisions. Traditional time series and physical models are being replaced by machine learning models due to their simplicity and lower training cost. However, insufficient data has hindered model training, making evaluating and comparing different models difficult. This paper introduces a comprehensive motion dataset containing data of more than 400 pieces from tens of offshore platform tests conducted at the State Key Laboratory of Ocean Engineering, Shanghai Jiao Tong University. The dataset is divided into subsets tailored for four application scenarios, including specific types of offshore platforms, wave conditions, noise addition data, and transfer learning. A Convolutional Attention-based LSTM model that combines convolution and self-attention mechanisms is proposed to validate the dataset and improve the accuracy of motion prediction. The proposed model is compared with classical models using our introduced dataset, achieving 5–10% improvement and confirming the dataset’s high reliability and applicability, as well as the effectiveness of the Conv-Att-LSTM model. This development sets a new standard for motion prediction and furthers the application of machine learning in ocean engineering. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>An overview of motion prediction research history [<a href="#B4-jmse-12-01852" class="html-bibr">4</a>,<a href="#B5-jmse-12-01852" class="html-bibr">5</a>,<a href="#B6-jmse-12-01852" class="html-bibr">6</a>,<a href="#B7-jmse-12-01852" class="html-bibr">7</a>,<a href="#B8-jmse-12-01852" class="html-bibr">8</a>,<a href="#B9-jmse-12-01852" class="html-bibr">9</a>,<a href="#B10-jmse-12-01852" class="html-bibr">10</a>,<a href="#B11-jmse-12-01852" class="html-bibr">11</a>,<a href="#B12-jmse-12-01852" class="html-bibr">12</a>,<a href="#B13-jmse-12-01852" class="html-bibr">13</a>,<a href="#B14-jmse-12-01852" class="html-bibr">14</a>,<a href="#B15-jmse-12-01852" class="html-bibr">15</a>,<a href="#B17-jmse-12-01852" class="html-bibr">17</a>,<a href="#B18-jmse-12-01852" class="html-bibr">18</a>,<a href="#B19-jmse-12-01852" class="html-bibr">19</a>,<a href="#B20-jmse-12-01852" class="html-bibr">20</a>,<a href="#B21-jmse-12-01852" class="html-bibr">21</a>,<a href="#B22-jmse-12-01852" class="html-bibr">22</a>,<a href="#B23-jmse-12-01852" class="html-bibr">23</a>,<a href="#B24-jmse-12-01852" class="html-bibr">24</a>,<a href="#B25-jmse-12-01852" class="html-bibr">25</a>,<a href="#B32-jmse-12-01852" class="html-bibr">32</a>,<a href="#B33-jmse-12-01852" class="html-bibr">33</a>,<a href="#B34-jmse-12-01852" class="html-bibr">34</a>,<a href="#B35-jmse-12-01852" class="html-bibr">35</a>,<a href="#B36-jmse-12-01852" class="html-bibr">36</a>].</p>
Full article ">Figure 2
<p>Data description in dataset (<b>A</b>) The dataset includes a wide range of offshore platforms. (<b>B</b>) The dataset includes 1-year to 1000-year wave conditions. Tp stands for average peak period and Hs stands for wave height. (<b>C</b>) Cross-correlation between different channels in the dataset. (<b>D</b>) The time and spectrum domain of wave and motion data.</p>
Full article ">Figure 3
<p>Data processing progress.</p>
Full article ">Figure 4
<p>Illustration of Conv-Att-LSTM model.</p>
Full article ">Figure 5
<p>Box plots showing training results on (<b>A</b>) platform-based and (<b>B</b>) wave condition-based subsets. The left and right figures show accuracy and loss variation, respectively.</p>
Full article ">Figure 6
<p>Training on different noise levels. (<b>A</b>) shows the data under different added noise levels; (<b>B</b>) shows the box plots of training results; (<b>C</b>) shows the motion prediction under different noise situations.</p>
Full article ">Figure 7
<p>Training results of (<b>A</b>) domain adaptation and (<b>B</b>) domain generalization.</p>
Full article ">
23 pages, 5439 KiB  
Article
AMTCN: An Attention-Based Multivariate Temporal Convolutional Network for Electricity Consumption Prediction
by Wei Zhang, Jiaxuan Liu, Wendi Deng, Siyu Tang, Fan Yang, Ying Han, Min Liu and Renzhuo Wan
Electronics 2024, 13(20), 4080; https://doi.org/10.3390/electronics13204080 - 17 Oct 2024
Viewed by 134
Abstract
Accurate prediction of electricity consumption is crucial for energy management and allocation. This study introduces a novel approach, named Attention-based Multivariate Temporal Convolutional Network (AMTCN), for electricity consumption forecasting by integrating attention mechanisms with multivariate temporal convolutional networks. The method involves feature extraction [...] Read more.
Accurate prediction of electricity consumption is crucial for energy management and allocation. This study introduces a novel approach, named Attention-based Multivariate Temporal Convolutional Network (AMTCN), for electricity consumption forecasting by integrating attention mechanisms with multivariate temporal convolutional networks. The method involves feature extraction from diverse time series of different feature variables using dilated convolutional networks. Subsequently, attention mechanisms are employed to capture the correlation and contextually important information among various features, thereby enhancing the model’s predictive accuracy. The AMTCN method exhibits universality, making it applicable to various prediction tasks in different scenarios. Experimental evaluations are conducted on four distinct datasets, encompassing electricity consumption and weather temperature aspects. Comparative experiments with LSTM, ConvLSTM, GRU, and TCN—widely-used deep learning methods—demonstrate that our AMTCN model achieves significant improvements of 57% in MSE, 37% in MAE, 35% in RRSE, and 12% in CORR metrics, respectively. This research contributes a promising approach to accurate electricity consumption prediction, leveraging the synergy of attention mechanisms and multivariate temporal convolutional networks, with broad applicability in diverse forecasting scenarios. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>A dilated causal convolution with dilation factors d = 1, 2, 4 and filter size k = 3.</p>
Full article ">Figure 2
<p>TCN residual block. A 1 × 1 convolution is added when residual input and output have different dimensions.</p>
Full article ">Figure 3
<p>Overall architecture of the AMTCN model.</p>
Full article ">Figure 4
<p>Visualization of dilated convolution with different dilation factors.</p>
Full article ">Figure 5
<p>An overview of two residual blocks with asymmetric structure: Residual Block 1 with three layers of dilated convolution (<b>left</b>) and Residual Block 2 with four layers dilated (<b>right</b>).</p>
Full article ">Figure 6
<p>Multi-head attention consists of several attention layers running in parallel.</p>
Full article ">Figure 7
<p>The Pearson correlation coefficient between different power generation methods and consumption.</p>
Full article ">Figure 8
<p>Distribution of data used in the Electricity-R dataset for the consumption feature.</p>
Full article ">Figure 9
<p>The MSE of the predicted values of each network on the four datasets with a prediction window of 24.</p>
Full article ">Figure 10
<p>Predicted and actual values of the AMTCN model for seven consecutive days on four datasets.</p>
Full article ">Figure 11
<p>Prediction results for each network on the Electricity-R dataset for 7 consecutive days.</p>
Full article ">Figure 12
<p>MSE of the predicted values of each network on the two power consumption datasets for a prediction window of 12.</p>
Full article ">Figure 13
<p>Prediction results for each network on the Electricity-R dataset for 7 consecutive days with a prediction window of 12.</p>
Full article ">Figure 14
<p>MSE of the predicted values of each network on the two power consumption datasets for a prediction window of 6.</p>
Full article ">Figure 15
<p>Prediction results for each network on the Electricity-R dataset for 7 consecutive days with a prediction window of 6.</p>
Full article ">Figure 16
<p>Seven consecutive days of AMTCN model predictions for Electricity-R under three prediction windows.</p>
Full article ">Figure 17
<p>Overall structure of the MTCN ablation model.</p>
Full article ">Figure 18
<p>MSE of predicted values of MTCN and AMTCN on each dataset with a prediction window of 24.</p>
Full article ">
24 pages, 4102 KiB  
Article
Plastic Constitutive Training Method for Steel Based on a Recurrent Neural Network
by Tianwei Wang, Yongping Yu, Haisong Luo and Zhigang Wang
Buildings 2024, 14(10), 3279; https://doi.org/10.3390/buildings14103279 - 16 Oct 2024
Viewed by 290
Abstract
The deep learning steel plastic constitutive model training method was studied based on the recurrent neural network (RNN) model to improve the allocative efficiency of the deep learning steel plastic constitutive model and promote its application in practical engineering. Two linear hardening constitutive [...] Read more.
The deep learning steel plastic constitutive model training method was studied based on the recurrent neural network (RNN) model to improve the allocative efficiency of the deep learning steel plastic constitutive model and promote its application in practical engineering. Two linear hardening constitutive datasets of steel were constructed using the Gaussian stochastic process. The RNN, long short-term memory (LSTM), and gated recurrent unit (GRU) were used as models for training. The effects of the data pre-processing method, neural network structure, and training method on the model training were analyzed. The prediction ability of the model for different scale series and the corresponding data demand were evaluated. The results show that LSTM and the GRU are more suitable for stress–strain prediction. The marginal effect of the stacked neural network depth and number gradually decreases, and the hysteresis curve can be accurately predicted by a two-layer RNN. The optimal structure of the two models is A50-100 and B150-150. The prediction accuracy of the models increased with the decrease in batch size and the increase in training batch, and the training time also increased significantly. The decay learning rate method could balance the prediction accuracy and training time, and the optimal initial learning rate, batch size, and training batch were 0.001, 60, and 100, respectively. The deep learning plastic constitutive model based on the optimal parameters can accurately predict the hysteresis curve of steel, and the prediction abilities of the GRU are 6.13, 6.7, and 3.3 times those of LSTM in short, medium, and long sequences, respectively. Full article
(This article belongs to the Special Issue Intelligent Design, Green Construction, and Innovation)
Show Figures

Figure 1

Figure 1
<p>Traditional constitutive model construction and deep learning constitutive model construction flow.</p>
Full article ">Figure 2
<p>Linear hardening constitutive model: (<b>a</b>) linear isotropic hardening constitutive; (<b>b</b>) linear kinematic hardening constitutive.</p>
Full article ">Figure 3
<p>Original RNN structure.</p>
Full article ">Figure 4
<p>LSTM network structure.</p>
Full article ">Figure 5
<p>GRU network structure.</p>
Full article ">Figure 6
<p>Comparison of data pre-processing methods.</p>
Full article ">Figure 7
<p>The effect of the number of neurons on the model: (<b>a</b>) the effect on the model performance; (<b>b</b>) the effect on the training time.</p>
Full article ">Figure 8
<p>Influence of hidden layers on model performance: (<b>a</b>) influence on model performance; (<b>b</b>) influence on training time.</p>
Full article ">Figure 9
<p>Influence of neural network topology on model performance and training time: (<b>a</b>) influence on model performance; (<b>b</b>) influence on training time.</p>
Full article ">Figure 10
<p>Effects of training frequency and training batches on the model: (<b>a</b>) LSTM; (<b>b</b>) GRU.</p>
Full article ">Figure 11
<p>Influence of learning rate on the model: (<b>a</b>) LSTM; (<b>b</b>) GRU.</p>
Full article ">Figure 12
<p>Influence of learning rate on the number of iterations.</p>
Full article ">Figure 13
<p>Model prediction effects of different dataset sizes: (<b>a</b>–<b>g</b>) prediction curve of LSTM when the dataset size is 500–10,000; (<b>h</b>–<b>n</b>) prediction curve of GRU when the dataset size is 500–10,000.</p>
Full article ">Figure 13 Cont.
<p>Model prediction effects of different dataset sizes: (<b>a</b>–<b>g</b>) prediction curve of LSTM when the dataset size is 500–10,000; (<b>h</b>–<b>n</b>) prediction curve of GRU when the dataset size is 500–10,000.</p>
Full article ">Figure 14
<p>Prediction capabilities of LSTM and GRU: (<b>a</b>) LSTM; (<b>b</b>) GRU.</p>
Full article ">Figure 15
<p>Prediction effect of the model: (<b>a</b>–<b>c</b>) linear isotropic constitutive hardening, and (<b>d</b>–<b>f</b>) linear kinematic constitutive hardening.</p>
Full article ">
22 pages, 1654 KiB  
Article
A New Scene Sensing Model Based on Multi-Source Data from Smartphones
by Zhenke Ding, Zhongliang Deng, Enwen Hu, Bingxun Liu, Zhichao Zhang and Mingyang Ma
Sensors 2024, 24(20), 6669; https://doi.org/10.3390/s24206669 - 16 Oct 2024
Viewed by 178
Abstract
Smartphones with integrated sensors play an important role in people’s lives, and in advanced multi-sensor fusion navigation systems, the use of individual sensor information is crucial. Because of the different environments, the weights of the sensors will be different, which will also affect [...] Read more.
Smartphones with integrated sensors play an important role in people’s lives, and in advanced multi-sensor fusion navigation systems, the use of individual sensor information is crucial. Because of the different environments, the weights of the sensors will be different, which will also affect the method and results of multi-source fusion positioning. Based on the multi-source data from smartphone sensors, this study explores five types of information—Global Navigation Satellite System (GNSS), Inertial Measurement Units (IMUs), cellular networks, optical sensors, and Wi-Fi sensors—characterizing the temporal, spatial, and mathematical statistical features of the data, and it constructs a multi-scale, multi-window, and context-connected scene sensing model to accurately detect the environmental scene in indoor, semi-indoor, outdoor, and semi-outdoor spaces, thus providing a good basis for multi-sensor positioning in a multi-sensor navigation system. Detecting environmental scenes provides an environmental positioning basis for multi-sensor fusion localization. This model is divided into four main parts: multi-sensor-based data mining, a multi-scale convolutional neural network (CNN), a bidirectional long short-term memory (BiLSTM) network combined with contextual information, and a meta-heuristic optimization algorithm. Full article
(This article belongs to the Special Issue Smart Sensor Systems for Positioning and Navigation)
Show Figures

Figure 1

Figure 1
<p>Four scene classifications: (<b>a</b>) outdoor, (<b>b</b>) semi-outdoor, (<b>c</b>) semi-indoor, and (<b>d</b>) indoor.</p>
Full article ">Figure 2
<p>Satellite zenith view: (<b>a</b>) west indoor neighboring window, (<b>b</b>) south indoor neighbouring window, (<b>c</b>) indoor, and (<b>d</b>) open outdoor neighboring window.</p>
Full article ">Figure 3
<p>DOP change graph: (<b>a</b>) outdoor DOP change graph, and (<b>b</b>) indoor DOP change graph.</p>
Full article ">Figure 4
<p>Visible satellite map: (<b>a</b>) variation in the number of visible satellites. (<b>b</b>) Variation in the rate of change of visible satellites in different windows.</p>
Full article ">Figure 5
<p>Satellite signal quality map: (<b>a</b>) CNR variation and (<b>b</b>) DCNR variation.</p>
Full article ">Figure 6
<p>State of motion versus acceleration.</p>
Full article ">Figure 7
<p>Wi-Fi channel spectrum scan: (<b>a</b>) indoor, (<b>b</b>) outdoor.</p>
Full article ">Figure 8
<p>Visible AP distribution of Wi-Fi: (<b>a</b>) number distribution, (<b>b</b>) signal strength distribution.</p>
Full article ">Figure 9
<p>Variation of light sensors and cellular network sensors: (<b>a</b>) variation of indoor and outdoor light intensity over 24 h, (<b>b</b>) variation of the number of base stations receiving signals.</p>
Full article ">Figure 10
<p>An algorithmic model for the classification of complex indoor and outdoor scenes based on spatio-temporal features.</p>
Full article ">Figure 11
<p>Pearson correlation feature map.</p>
Full article ">Figure 12
<p>Schematic of a two-scale convolutional neural network.</p>
Full article ">Figure 13
<p>BiLSTM network structure diagram.</p>
Full article ">Figure 14
<p>Structure of the ablation experiment.</p>
Full article ">Figure 15
<p>Confusion matrix: (<b>a</b>) confusion matrix before WOA optimization. (<b>b</b>) confusion matrix after WOA optimisation.</p>
Full article ">Figure 16
<p>Comparison of the accuracy of different models.</p>
Full article ">Figure 17
<p>Comparison of accuracy in different scenarios.</p>
Full article ">
22 pages, 4866 KiB  
Article
TCEDN: A Lightweight Time-Context Enhanced Depression Detection Network
by Keshan Yan, Shengfa Miao, Xin Jin, Yongkang Mu, Hongfeng Zheng, Yuling Tian, Puming Wang, Qian Yu and Da Hu
Life 2024, 14(10), 1313; https://doi.org/10.3390/life14101313 - 16 Oct 2024
Viewed by 181
Abstract
The automatic video recognition of depression is becoming increasingly important in clinical applications. However, traditional depression recognition models still face challenges in practical applications, such as high computational costs, the poor application effectiveness of facial movement features, and spatial feature degradation due to [...] Read more.
The automatic video recognition of depression is becoming increasingly important in clinical applications. However, traditional depression recognition models still face challenges in practical applications, such as high computational costs, the poor application effectiveness of facial movement features, and spatial feature degradation due to model stitching. To overcome these challenges, this work proposes a lightweight Time-Context Enhanced Depression Detection Network (TCEDN). We first use attention-weighted blocks to aggregate and enhance video frame-level features, easing the model’s computational workload. Next, by integrating the temporal and spatial changes of video raw features and facial movement features in a self-learning weight manner, we enhance the precision of depression detection. Finally, a fusion network of 3-Dimensional Convolutional Neural Network (3D-CNN) and Convolutional Long Short-Term Memory Network (ConvLSTM) is constructed to minimize spatial feature loss by avoiding feature flattening and to achieve depression score prediction. Tests on the AVEC2013 and AVEC2014 datasets reveal that our approach yields results on par with state-of-the-art techniques for detecting depression using video analysis. Additionally, our method has significantly lower computational complexity than mainstream methods. Full article
(This article belongs to the Section Biochemistry, Biophysics and Computational Biology)
Show Figures

Figure 1

Figure 1
<p>The proposed TCEDN framework, as shown in the figure, consists of three key components: (<b>a</b>) the Attention-Weighted Aggregation Module (AWAM) for video frames, (<b>b</b>) the Self-Learning Time-Difference Weighting Module (STDWM), (<b>c</b>) the Concatenated Network based on ConvLSTM and 3D-CNN. Video data are input into the model in the form of a series of frames, first passing through the Attention-Weighted Aggregation Module to highlight important information in the video. The aggregation module aggregates the weighted frames to generate a comprehensive feature representation. The aggregated features then enter the STDWM, which can intelligently fuse and enhance time-difference features, capturing subtle changes in facial expressions in the video. The fused features are further input into the ConvLSTM-based 3D-CNN network, processed, and finally mapped to a specific score using a traditional 2D-CNN regression network.</p>
Full article ">Figure 2
<p>Attention-Weighted Aggregation Module. Feature aggregation module based on attention mechanism weighting, taking three adjacent frames as an example.</p>
Full article ">Figure 3
<p>Self-Learning Time-Difference Weighting Module. The proposed STDWM, as shown in the figure, consists of two components: the Time-Difference Module and the Self-Learning Weight Fusion Module (SWFM). The raw video data are first processed by the Time-Difference Module to obtain facial motion features, which are then combined with the primitive facial features and weighted through the SWFM to obtain enhanced video features.</p>
Full article ">Figure 4
<p>ConvLSTM is a hybrid model that combines CNN and LSTM. In the figure, the arrows indicate the direction of data flow, ∗ denotes the convolution operation, <math display="inline"><semantics> <mi>σ</mi> </semantics></math> represents the sigmoid activation function. By integrating the spatial feature extraction capability of CNN and the temporal modeling capability of LSTM, it can effectively handle spatiotemporal data and achieve good performance in tasks such as video prediction and time-series forecasting.</p>
Full article ">Figure 5
<p>The process of training models on AVEC2014 and AVEC2013 datasets. We recorded the MAE and RMSE metrics for both the training and validation sets to evaluate the model’s performance. As training progressed, we observed a gradual decrease in the MAE and RMSE metrics, eventually leveling off.</p>
Full article ">Figure 6
<p>The error distribution of videos with the minimum absolute error to videos with the maximum absolute error on two public datasets.</p>
Full article ">Figure 7
<p>Absolute error distribution of videos from lowest to highest depression scores on two public datasets.</p>
Full article ">Figure 8
<p>A visualization of the eye slice features of TCEDN on the AVEC2013 dataset: (<b>A</b>) Feature visualization for a BDI-II score of 8. (<b>B</b>) Feature visualization for a BDI-II score of 15. (<b>C</b>) Feature visualization for a BDI-II score of 22. (<b>D</b>) Feature visualization for a BDI-II score of 30.</p>
Full article ">Figure 9
<p>A visualization of the facial slice features of TCEDN on the AVEC2013 dataset: (<b>A</b>) Feature visualization for a BDI-II score of 1. (<b>B</b>) Feature visualization for a BDI-II score of 10. (<b>C</b>) Feature visualization for a BDI-II score of 18. (<b>D</b>) Feature visualization for a BDI-II score of 41.</p>
Full article ">Figure 10
<p>A visualization of the ocular slice features of TCEDN on the AVEC2014 dataset: (<b>A</b>) Feature visualization for a BDI-II score of 0. (<b>B</b>) Feature visualization for a BDI-II score of 5. (<b>C</b>) Feature visualization for a BDI-II score of 16. (<b>D</b>) Feature visualization for a BDI-II score of 25.</p>
Full article ">Figure 11
<p>A visualization of the facial slice features of TCEDN on the AVEC2014 dataset: (<b>A</b>) Feature visualization for a BDI-II score of 3. (<b>B</b>) Feature visualization for a BDI-II score of 11. (<b>C</b>) Feature visualization for a BDI-II score of 22. (<b>D</b>) Feature visualization for a BDI-II score of 37.</p>
Full article ">
24 pages, 10877 KiB  
Article
Modeling Temperature-Dependent Photoluminescence Dynamics of Colloidal CdS Quantum Dots Using Long Short-Term Memory (LSTM) Networks
by Ivan Malashin, Daniil Daibagya, Vadim Tynchenko, Vladimir Nelyub, Aleksei Borodulin, Andrei Gantimurov, Alexandr Selyukov, Sergey Ambrozevich, Mikhail Smirnov and Oleg Ovchinnikov
Materials 2024, 17(20), 5056; https://doi.org/10.3390/ma17205056 - 16 Oct 2024
Viewed by 309
Abstract
This study addresses the challenge of modeling temperature-dependent photoluminescence (PL) in CdS colloidal quantum dots (QD), where PL properties fluctuate with temperature, complicating traditional modeling approaches. The objective is to develop a predictive model capable of accurately capturing these variations using Long Short-Term [...] Read more.
This study addresses the challenge of modeling temperature-dependent photoluminescence (PL) in CdS colloidal quantum dots (QD), where PL properties fluctuate with temperature, complicating traditional modeling approaches. The objective is to develop a predictive model capable of accurately capturing these variations using Long Short-Term Memory (LSTM) networks, which are well suited for managing temporal dependencies in time-series data. The methodology involved training the LSTM model on experimental time-series data of PL intensity and temperature. Through numerical simulation, the model’s performance was assessed. Results demonstrated that the LSTM-based model effectively predicted PL trends under different temperature conditions. This approach could be applied in optoelectronics and quantum dot-based sensors for enhanced forecasting capabilities. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) TEM image of the synthesized semiconductor colloidal CdS QD. (<b>b</b>) Absorption (dashed curve) and luminescence (solid curve) spectra of CdS QD.</p>
Full article ">Figure 2
<p>Workflow of the experimental and analytical approach used in this study.</p>
Full article ">Figure 3
<p>Examples of images of spectra with approximations at different temperatures. (<b>a</b>) 85.6 K, (<b>b</b>) 182.6 K, (<b>c</b>) 300.2 K.</p>
Full article ">Figure 4
<p>The PL spectra of CdS quantum dots (QDs) were measured as the sample was heated from 84 K to 306 K, with curves recorded at intervals of 7 K.</p>
Full article ">Figure 5
<p>Temperature-dependence of trap state luminescence peak energy (<b>a</b>), FWHM (<b>b</b>), and integrated intensity (<b>c</b>) for CdS QDs. The approximation of the absorption spectrum by four Gaussian functions (G1–G4); the red dashed line shows the fitting result (<b>d</b>).</p>
Full article ">Figure 6
<p>LSTM prediction results.</p>
Full article ">Figure 7
<p>Overview of the LSTM model for predicting temperature-dependent PL of CdS QDs.</p>
Full article ">
21 pages, 2639 KiB  
Article
A Recurrent Deep Network for Gait Phase Identification from EMG Signals During Exoskeleton-Assisted Walking
by Bruna Maria Vittoria Guerra, Micaela Schmid, Stefania Sozzi, Serena Pizzocaro, Alessandro Marco De Nunzio and Stefano Ramat
Sensors 2024, 24(20), 6666; https://doi.org/10.3390/s24206666 - 16 Oct 2024
Viewed by 280
Abstract
Lower limb exoskeletons represent a relevant tool for rehabilitating gait in patients with lower limb movement disorders. Partial assistance exoskeletons adaptively provide the joint torque needed, on top of that produced by the patient, for a correct and stable gait, helping the patient [...] Read more.
Lower limb exoskeletons represent a relevant tool for rehabilitating gait in patients with lower limb movement disorders. Partial assistance exoskeletons adaptively provide the joint torque needed, on top of that produced by the patient, for a correct and stable gait, helping the patient to recover an autonomous gait. Thus, the device needs to identify the different phases of the gait cycle to produce precisely timed commands that drive its joint motors appropriately. In this study, EMG signals have been used for gait phase detection considering that EMG activations lead limb kinematics by at least 120 ms. We propose a deep learning model based on bidirectional LSTM to identify stance and swing gait phases from EMG data. We built a dataset of EMG signals recorded at 1500 Hz from four muscles from the dominant leg in a population of 26 healthy subjects walking overground (WO) and walking on a treadmill (WT) using a lower limb exoskeleton. The data were labeled with the corresponding stance or swing gait phase based on limb kinematics provided by inertial motion sensors. The model was studied in three different scenarios, and we explored its generalization abilities and evaluated its applicability to the online processing of EMG data. The training was always conducted on 500-sample sequences from WO recordings of 23 subjects. Testing always involved WO and WT sequences from the remaining three subjects. First, the model was trained and tested on 500 Hz EMG data, obtaining an overall accuracy on the WO and WT test datasets of 92.43% and 91.16%, respectively. The simulation of online operation required 127 ms to preprocess and classify one sequence. Second, the trained model was evaluated against a test set built on 1500 Hz EMG data. The accuracies were lower, yet the processing times were 11 ms faster. Third, we partially retrained the model on a subset of the 1500 Hz training dataset, achieving 87.17% and 89.64% accuracy on the 1500 Hz WO and WT test sets, respectively. Overall, the proposed deep learning model appears to be a valuable candidate for entering the control pipeline of a lower limb rehabilitation exoskeleton in terms of both the achieved accuracy and processing times. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

Figure 1
<p>(<b>A</b>): WO between the safety bars. The experimenter follows the subject, driving the exoskeleton through the control handles. (<b>B</b>): platform for patient rotation used to invert the walking direction while the subject maintains standing position.</p>
Full article ">Figure 2
<p>(<b>A</b>): ankle dorsiflexion setup, MVIC assessment for TA; (<b>B</b>): knee extension setup, MVIC assessment for VL; (<b>C</b>): ankle plantarflexion setup, MVIC assessment for SOL; (<b>D</b>): knee flexion setup, MVIC assessment for BF.</p>
Full article ">Figure 3
<p>HS and TO identification. (<b>A</b>) shows the hip angle (HA) calculated between the sacrum and thigh (black trace) and the HA after being high pass-filtered (HA hp, blue trace). Positive values signify hip flexion, while negative values indicate hip extension. The positive part of the filtered HA signal was set to zero, and then the signal obtained was rectified (blue trace, (<b>B</b>)). On this trace, the local maxima within each time window defined between each maximum and the following minimum of the HA signal (pale-colored rectangles in (<b>A</b>,<b>B</b>)) were identified (red dots in (<b>B</b>)). The instants corresponding to local maxima thus identified were the instants in which the HS event occurred. In (<b>C</b>), the black trace corresponded to the knee angle (KA). Positive values signify knee flexion, while negative values indicate knee extension. In a time window between each minimum and successive maximum of the KA signal (pale-colored rectangles in (<b>C</b>)), 60% of the KA maximum was searched (green dots in (<b>C</b>)). TO occurred at 60% of the KA maximum. Only a 10 s portion of the signals is presented in the figure for clarity.</p>
Full article ">Figure 4
<p>(<b>A</b>,<b>B</b>) illustrates labelled HA and KA data referring to WO condition. The light blue line labelled “Standing” corresponds to the quiet standing phase identified by the hidden Markov model. The data belonging to the stance phase are displayed in red, whereas those of the swing phase are in green. The dashed black line corresponds to the data excluded from the muscle labelling procedure, then discarded.</p>
Full article ">Figure 5
<p>10 s example of the processed and labelled EMG signals of the four muscles (SOL, TA, BF, VL) during the stance (in red) and swing phases (in green) in the WO scenario.</p>
Full article ">Figure 6
<p>ROC curves referring to WO (blue line) and WT test datasets (dashed red line), both under-sampled at 500 Hz.</p>
Full article ">Figure 7
<p>Comparison of ROC curves depicting the performance of the BiLSTM model on the 1500 Hz WO and WT test datasets. The performance of the original BiLSTM model is represented by the blue line (WO) and dashed red line (WT), while the retrained BiLSTM model is illustrated by the green line (WO) and dashed violet line (WT).</p>
Full article ">Figure 8
<p>Examples of the best and the worst input signals to the BiLSTM model from the WO test set are shown in S1 and S2 (<b>A</b>) and examples from the WT test set are shown in S3 and S2 (<b>B</b>). Signals represent normalized muscular activation of the soleus (SOL), tibialis (TA), biceps femoris (BF), and vastus lateralis (VL) during four gait cycles. Red traces indicate time intervals labeled as belonging to the stance phase while green traces indicate those labeled as swing phases. The difference in signal quality is evident, especially in the SOL, TA, and VL signals of S2 in both scenarios (WO and WT), where identifying the steps and distinguishing the stance and swing phases is challenging even visually.</p>
Full article ">
19 pages, 11753 KiB  
Article
Landslide Deformation Analysis and Prediction with a VMD-SA-LSTM Combined Model
by Chengzhi Wen, Hongling Tian, Xiaoyan Zeng, Xin Xia, Xiaobo Hu and Bo Pang
Water 2024, 16(20), 2945; https://doi.org/10.3390/w16202945 - 16 Oct 2024
Viewed by 308
Abstract
The evolution of landslides is influenced by the complex interplay of internal geological factors and external triggering factors, resulting in nonlinear dynamic changes. Although deep learning methods have demonstrated advantages in predicting multivariate landslide displacement, their performance is often constrained by the challenges [...] Read more.
The evolution of landslides is influenced by the complex interplay of internal geological factors and external triggering factors, resulting in nonlinear dynamic changes. Although deep learning methods have demonstrated advantages in predicting multivariate landslide displacement, their performance is often constrained by the challenges of extracting intricate features from extended time-series data. To address this challenge, we propose a novel displacement prediction model that integrates Variational Mode Decomposition (VMD), Self-Attention (SA), and Long Short-Term Memory (LSTM) networks. The model first employs VMD to decompose cumulative landslide displacement into trend, periodic, and stochastic components, followed by an assessment of the correlation between these components and the triggering factors using grey relational analysis. Subsequently, the self-attention mechanism is incorporated into the LSTM model to enhance its ability to capture complex dependencies. Finally, each displacement component is fed into the SA-LSTM model for separate predictions, which are then reconstructed to obtain the cumulative displacement prediction. Using the Zhonghai Village tunnel entrance (ZVTE) landslide as a case study, we validated the model with displacement data from GPS point 105 and made predictions for GPS point 104 to evaluate the model’s generalization capability. The results indicated that the RMSE and MAPE for SA-LSTM, LSTM, and TCN-LSTM at GPS point 105 were 0.3251 and 1.6785, 0.6248 and 2.9130, and 1.1777 and 5.5131, respectively. These findings demonstrate that SA-LSTM outperformed the other models in terms of complex feature extraction and accuracy. Furthermore, the RMSE and MAPE at GPS point 104 were 0.4232 and 1.0387, further corroborating the model’s strong extrapolation capability and its effectiveness in landslide monitoring. Full article
(This article belongs to the Special Issue Water, Geohazards, and Artificial Intelligence, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Layout of monitoring points at the ZVTE landslide.</p>
Full article ">Figure 2
<p>The monitoring curves of the cumulative displacement, rainfall, and soil moisture content of the ZVTE landslide.</p>
Full article ">Figure 3
<p>The principle of the self-attention mechanism.</p>
Full article ">Figure 4
<p>Gating mechanism of the Long Short-Term Memory Network.</p>
Full article ">Figure 5
<p>Structure of the SA-LSTM model.</p>
Full article ">Figure 6
<p>Deformation prediction process: (<b>a</b>) data preprocessing; (<b>b</b>) model prediction and validation.</p>
Full article ">Figure 7
<p>GPS105 cumulative displacement decomposition: (<b>a</b>) Trend component displacement. (<b>b</b>) Periodic component displacement. (<b>c</b>) Random component displacement.</p>
Full article ">Figure 8
<p>Trend component displacement prediction: (<b>a</b>) displacement prediction; (<b>b</b>) accuracy metrics comparison.</p>
Full article ">Figure 9
<p>Periodic component displacement prediction: (<b>a</b>) displacement prediction; (<b>b</b>) accuracy metrics comparison.</p>
Full article ">Figure 10
<p>Random component displacement prediction: (<b>a</b>) displacement prediction; (<b>b</b>) accuracy metrics.</p>
Full article ">Figure 11
<p>Cumulative displacement prediction: (<b>a</b>) displacement prediction; (<b>b</b>) accuracy metrics comparison.</p>
Full article ">Figure 12
<p>Cumulative displacement prediction at monitoring point GPS104.</p>
Full article ">
19 pages, 9417 KiB  
Hypothesis
Prediction of Traffic Volume Based on Deep Learning Model for AADT Correction
by Dae Cheol Han
Appl. Sci. 2024, 14(20), 9436; https://doi.org/10.3390/app14209436 (registering DOI) - 16 Oct 2024
Viewed by 271
Abstract
Accurate traffic volume data are crucial for effective traffic management, infrastructure development, and demand forecasting. This study addresses the challenges associated with traffic volume data collection, including, notably, equipment malfunctions that often result in missing data and inadequate anomaly detection. We have developed [...] Read more.
Accurate traffic volume data are crucial for effective traffic management, infrastructure development, and demand forecasting. This study addresses the challenges associated with traffic volume data collection, including, notably, equipment malfunctions that often result in missing data and inadequate anomaly detection. We have developed a deep-learning-based model to improve the reliability of predictions for annual average daily traffic volume. Utilizing a decade of traffic survey data (2010–2020) from the Korea Institute of Civil Engineering and Building Technology, we constructed a univariate time series prediction model across three consecutive sections. This model incorporates both raw and adjusted traffic volume data from 2017 to 2019, employing long short-term memory (LSTM) techniques to manage data discontinuities. A power function was integrated to simulate various error correction scenarios, thus enhancing the model’s resilience to prediction inaccuracies. The performance of the model was evaluated using certain metrics, such as the mean absolute error, the root mean squared error, and the coefficient of determination, thus validating the effectiveness of the deep learning approach in refining traffic volume estimations. Full article
(This article belongs to the Special Issue Intelligent Transportation System Technologies and Applications)
Show Figures

Figure 1

Figure 1
<p>Basic recursive structure of recurrent neural network.</p>
Full article ">Figure 2
<p>Development of the architecture for deep-learning-based models.</p>
Full article ">Figure 3
<p>Design procedure for verifying time series prediction performance.</p>
Full article ">Figure 4
<p>Prophet model’s prediction results (point 10001).</p>
Full article ">Figure 5
<p>Prophet model’s prediction results (point 10004).</p>
Full article ">Figure 6
<p>LSTM model’s prediction results (24 h).</p>
Full article ">Figure 7
<p>Comparison of performances between LSTM and power-function-based LSTM.</p>
Full article ">Figure 8
<p>Traffic volume prediction results of the power-function-based LSTM (24 h).</p>
Full article ">Figure 9
<p>LSTM model’s results (point 10001, 10% data missing).</p>
Full article ">Figure 10
<p>LSTM model’s prediction results (point 10001, 50% data missing).</p>
Full article ">
15 pages, 4238 KiB  
Article
An Improved PINN Algorithm for Shallow Water Equations Driven by Deep Learning
by Yanling Li, Qianxing Sun, Junfang Wei and Chunyan Huang
Symmetry 2024, 16(10), 1376; https://doi.org/10.3390/sym16101376 (registering DOI) - 16 Oct 2024
Viewed by 243
Abstract
Solving shallow water equations is crucial in science and engineering for understanding and predicting natural phenomena. To address the limitations of Physics-Informed Neural Network (PINN) in solving shallow water equations, we propose an improved PINN algorithm integrated with a deep learning framework. This [...] Read more.
Solving shallow water equations is crucial in science and engineering for understanding and predicting natural phenomena. To address the limitations of Physics-Informed Neural Network (PINN) in solving shallow water equations, we propose an improved PINN algorithm integrated with a deep learning framework. This algorithm introduces a regularization term as a penalty in the loss function, based on the PINN and Long Short-Term Memory (LSTM) models, and incorporates an attention mechanism to solve the original equation across the entire domain. Simulation experiments were conducted on one-dimensional and two-dimensional shallow water equations. The results indicate that, compared to the classical PINN algorithm, the improved algorithm shows significant advantages in handling discontinuities, such as sparse waves, in one-dimensional problems. It accurately captures sparse waves and avoids smoothing effects. In two-dimensional problems, the improved algorithm demonstrates good symmetry and effectively reduces non-physical oscillations. It also shows significant advantages in capturing details and handling complex phenomena, offering higher reliability and accuracy. The improved PINNs algorithm, which combines neural networks with physical mechanisms, can provide robust solutions and effectively avoid some of the shortcomings of classical PINNs methods. It also possesses high resolution and strong generalization capabilities, enabling accurate predictions at any given moment. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

Figure 1
<p>Diagrammatic sketch of PINN.</p>
Full article ">Figure 2
<p>LSTM network architecture diagram.</p>
Full article ">Figure 3
<p>Depicts a schematic diagram of the improved PINN structure.</p>
Full article ">Figure 4
<p>Comparative analysis of <math display="inline"><semantics> <mrow> <msub> <mi>L</mi> <mn>1</mn> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>L</mi> <mn>2</mn> </msub> </mrow> </semantics></math> regularization loss in one-dimensional and two-dimensional shallow water equations.</p>
Full article ">Figure 5
<p>Comparison of results on one dimension: (<b>a</b>) results of the classical PINN algorithm; (<b>b</b>) results of the improved PINN algorithm.</p>
Full article ">Figure 6
<p>One-dimensional loss function variation curve.</p>
Full article ">Figure 7
<p>Comparison of algorithm results at t = 0.03 and t = 0.21.</p>
Full article ">Figure 8
<p>Comparison of algorithm results at t = 0.30 and t = 0.37.</p>
Full article ">Figure 9
<p>Two-dimensional loss function variation curve.</p>
Full article ">
24 pages, 7433 KiB  
Article
Efficient Fault Warning Model Using Improved Red Deer Algorithm and Attention-Enhanced Bidirectional Long Short-Term Memory Network
by Yutian Wang and Mingli Wu
Processes 2024, 12(10), 2253; https://doi.org/10.3390/pr12102253 (registering DOI) - 15 Oct 2024
Viewed by 280
Abstract
The rapid advancement of industrial processes makes ensuring the stability of industrial equipment a critical factor in improving production efficiency and safeguarding operational safety. Fault warning systems, as a key technological means to enhance equipment stability, are increasingly gaining attention across industries. However, [...] Read more.
The rapid advancement of industrial processes makes ensuring the stability of industrial equipment a critical factor in improving production efficiency and safeguarding operational safety. Fault warning systems, as a key technological means to enhance equipment stability, are increasingly gaining attention across industries. However, as equipment structures and functions become increasingly complex, traditional fault warning methods face challenges such as limited prediction accuracy and difficulties in meeting real-time requirements. To address these challenges, this paper proposes an innovative hybrid fault warning method. The proposed approach integrates a multi-strategy improved red deer optimization algorithm (MIRDA), attention mechanism, and bidirectional long short-term memory network (BiLSTM). Firstly, the red deer optimization algorithm (RDA) is enhanced through improvements in population initialization strategy, adaptive optimal guidance strategy, chaos regulation factor, and double-sided mirror reflection theory, thereby enhancing its optimization performance. Subsequently, the MIRDA is employed to optimize the hyperparameters of the BiLSTM model incorporating an attention mechanism. A predictive model is then constructed based on the optimized Attention-BiLSTM, which, combined with a sliding window approach, provides robust support for fault threshold identification. The proposed algorithm’s efficacy is demonstrated through its application to real-world gas-fired power plant equipment fault cases. Comparative analyses with other advanced algorithms reveal its superior robustness and accuracy in efficiently issuing fault warnings. This research not only provides a more reliable safeguard for the stable operation of industrial equipment but also pioneers a new avenue for the application of metaheuristic algorithms. Full article
Show Figures

Figure 1

Figure 1
<p>Attention-BiLSTM network structure.</p>
Full article ">Figure 2
<p>Flowchart of MIRDA.</p>
Full article ">Figure 3
<p>Images of the feedwater pump.</p>
Full article ">Figure 4
<p>Predictive results on the testing set.</p>
Full article ">Figure 4 Cont.
<p>Predictive results on the testing set.</p>
Full article ">Figure 5
<p>The average of residuals obtained after processing with the sliding window method.</p>
Full article ">Figure 6
<p>Analysis results of the average residuals for the fault warning example.</p>
Full article ">Figure 7
<p>The 95% confidence intervals and standard deviation of algorithm performance (<a href="#sec5dot1-processes-12-02253" class="html-sec">Section 5.1</a>): RMSE (<b>a</b>), MAPE (<b>b</b>), standard deviation (<b>c</b>).</p>
Full article ">Figure 7 Cont.
<p>The 95% confidence intervals and standard deviation of algorithm performance (<a href="#sec5dot1-processes-12-02253" class="html-sec">Section 5.1</a>): RMSE (<b>a</b>), MAPE (<b>b</b>), standard deviation (<b>c</b>).</p>
Full article ">Figure 8
<p>The 95% confidence intervals and standard deviation of algorithm performance (<a href="#sec5dot2-processes-12-02253" class="html-sec">Section 5.2</a>): RMSE (<b>a</b>), MAPE (<b>b</b>), standard deviation (<b>c</b>).</p>
Full article ">Figure 8 Cont.
<p>The 95% confidence intervals and standard deviation of algorithm performance (<a href="#sec5dot2-processes-12-02253" class="html-sec">Section 5.2</a>): RMSE (<b>a</b>), MAPE (<b>b</b>), standard deviation (<b>c</b>).</p>
Full article ">Figure 9
<p>Graphical presentation of statistical results (<a href="#sec5dot1-processes-12-02253" class="html-sec">Section 5.1</a>): RMSE (<b>a</b>), MAPE (<b>b</b>).</p>
Full article ">Figure 10
<p>Graphical presentation of statistical results (<a href="#sec5dot2-processes-12-02253" class="html-sec">Section 5.2</a>): RMSE (<b>a</b>), MAPE (<b>b</b>).</p>
Full article ">
Back to TopTop