[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,292)

Search Parameters:
Keywords = GRU

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
31 pages, 6880 KiB  
Article
Multi-Dimensional Global Temporal Predictive Model for Multi-State Prediction of Marine Diesel Engines
by Liyong Ma, Siqi Chen, Shuli Jia, Yong Zhang and Hai Du
J. Mar. Sci. Eng. 2024, 12(8), 1370; https://doi.org/10.3390/jmse12081370 - 11 Aug 2024
Viewed by 351
Abstract
The reliability and stability of marine diesel engines are pivotal to the safety and economy of maritime operations. Accurate and efficient prediction of the states of these engines is essential for performance evaluation and operational continuity. This paper introduces a novel hybrid deep [...] Read more.
The reliability and stability of marine diesel engines are pivotal to the safety and economy of maritime operations. Accurate and efficient prediction of the states of these engines is essential for performance evaluation and operational continuity. This paper introduces a novel hybrid deep learning model, the multi-dimensional global temporal predictive (MDGTP) model, designed for synchronous multi-state prediction of marine diesel engines. The model incorporates parallel multi-head attention mechanisms, an enhanced long short-term memory (LSTM) with interleaved residual connections, and gated recurrent units (GRUs). Additionally, we propose a dynamic arithmetic tuna optimization algorithm, which synergizes tuna swarm optimization (TSO), and the arithmetic optimization algorithm (AOA) for hyperparameter optimization, thereby enhancing prediction accuracy. Comparative experiments using actual marine diesel engine data demonstrate that our model outperforms the LSTM, GRU, LSTM–GRU, support vector regression (SVR), random forest (RF), Gaussian process regression (GPR), and back propagation (BP) models, achieving the lowest root mean squared error (RMSE) and mean absolute error (MAE), as well as the highest Pearson correlation coefficient across three sampling periods. Ablation studies confirm the significance of each component in improving prediction accuracy. Our findings validate the efficacy of the proposed MDGTP model for predicting the multi-dimensional operating states of marine diesel engines. Full article
Show Figures

Figure 1

Figure 1
<p>The structure of multi-dimensional global temporal predictive model.</p>
Full article ">Figure 2
<p>The global extractor of the multi-dimensional global temporal predictive model.</p>
Full article ">Figure 3
<p>Temporal extractor module and unfolded GRU layer structure.</p>
Full article ">Figure 4
<p>Unfolded IRLSTM layer structure.</p>
Full article ">Figure 5
<p>The output layer structure.</p>
Full article ">Figure 6
<p>Optimization process using DATO.</p>
Full article ">Figure 7
<p>Comparison of predicted trends for eight operating states at a 30 s sampling period. (<b>a</b>) Diesel engine cooling water temperature. (<b>b</b>) Cooling water inlet temperature. (<b>c</b>) Cooling water outlet temperature. (<b>d</b>) Diesel engine lubricating oil temperature. (<b>e</b>) Lubricating oil outlet temperature. (<b>f</b>) Diesel engine lubricating oil pressure. (<b>g</b>) Diesel engine fuel pressure. (<b>h</b>) Diesel engine cooling water pressure.</p>
Full article ">Figure 8
<p>Comparison of predicted trends for eight operating states at a 60 s sampling period. (<b>a</b>) Diesel engine cooling water temperature. (<b>b</b>) Cooling water inlet temperature. (<b>c</b>) Cooling water outlet temperature. (<b>d</b>) Diesel engine lubricating oil temperature. (<b>e</b>) Lubricating oil outlet temperature. (<b>f</b>) Diesel engine lubricating oil pressure. (<b>g</b>) Diesel engine fuel pressure. (<b>h</b>) Diesel engine cooling water pressure.</p>
Full article ">Figure 9
<p>Comparison of predicted trends for eight operating states at a 90 s sampling period. (<b>a</b>) Diesel engine cooling water temperature. (<b>b</b>) Cooling water inlet temperature. (<b>c</b>) Cooling water outlet temperature. (<b>d</b>) Diesel engine lubricating oil temperature. (<b>e</b>) Lubricating oil outlet temperature. (<b>f</b>) Diesel engine lubricating oil pressure. (<b>g</b>) Diesel engine fuel pressure. (<b>h</b>) Diesel engine cooling water pressure.</p>
Full article ">Figure 10
<p>Performance metrics of ablation study for five models at 30 s sampling period. (<b>a</b>) RMSE. (<b>b</b>) MAE. (<b>c</b>) Pearson correlation coefficient <span class="html-italic">r</span>.</p>
Full article ">Figure 11
<p>Performance metrics of ablation study for five models at 60 s sampling period. (<b>a</b>) RMSE. (<b>b</b>) MAE. (<b>c</b>) Pearson correlation coefficient <span class="html-italic">r</span>.</p>
Full article ">Figure 12
<p>Performance metrics of ablation study for five models at 90 s sampling period. (<b>a</b>) RMSE. (<b>b</b>) MAE. (<b>c</b>) Pearson correlation coefficient <span class="html-italic">r</span>.</p>
Full article ">
21 pages, 2246 KiB  
Article
A Novel Rational Medicine Use System Based on Domain Knowledge Graph
by Chaoping Qin, Zhanxiang Wang, Jingran Zhao, Luyi Liu, Feng Xiao and Yi Han
Electronics 2024, 13(16), 3156; https://doi.org/10.3390/electronics13163156 - 9 Aug 2024
Viewed by 291
Abstract
Medication errors, which could often be detected in advance, are a significant cause of patient deaths each year, highlighting the critical importance of medication safety. The rapid advancement of data analysis technologies has made intelligent medication assistance applications possible, and these applications rely [...] Read more.
Medication errors, which could often be detected in advance, are a significant cause of patient deaths each year, highlighting the critical importance of medication safety. The rapid advancement of data analysis technologies has made intelligent medication assistance applications possible, and these applications rely heavily on medical knowledge graphs. However, current knowledge graph construction techniques are predominantly focused on general domains, leaving a gap in specialized fields, particularly in the medical domain for medication assistance. The specialized nature of medical knowledge and the distinct distribution of vocabulary between general and biomedical texts pose challenges. Applying general natural language processing techniques directly to the medical domain often results in lower accuracy due to the inadequate utilization of contextual semantics and entity information. To address these issues and enhance knowledge graph production, this paper proposes an optimized model for named entity recognition and relationship extraction in the Chinese medical domain. Key innovations include utilizing Medical Bidirectional Encoder Representations from Transformers (MCBERT) for character-level embeddings pre-trained on Chinese biomedical corpora, employing Bi-directional Gated Recurrent Unit (BiGRU) networks for extracting enriched contextual features, integrating a Conditional Random Field (CRF) layer for optimal label sequence output, using the Piecewise Convolutional Neural Network (PCNN) to capture comprehensive semantic information and fusing it with entity features for better classification accuracy, and implementing a microservices architecture for the medication assistance review system. These enhancements significantly improve the accuracy of entity relationship classification in Chinese medical texts. The model achieved good performance in recognizing most entity types, with an accuracy of 88.3%, a recall rate of 85.8%, and an F1 score of 87.0%. In the relationship extraction stage, the accuracy reached 85.7%, the recall rate 82.5%, and the F1 score 84.0%. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

Figure 1
<p>Overall design of Rational Medicine Use System.</p>
Full article ">Figure 2
<p>Candidate entries.</p>
Full article ">Figure 3
<p>Knowledge graph ontology structure.</p>
Full article ">Figure 4
<p>Knowledge graph medicine structure.</p>
Full article ">Figure 5
<p>Schematic of MCB-CRF model structure.</p>
Full article ">Figure 6
<p>Relationship extraction model structure diagram.</p>
Full article ">Figure 7
<p>Character and position vector representation.</p>
Full article ">Figure 8
<p>Schematic diagram of PCNN model.</p>
Full article ">Figure 9
<p>Comparison of experimental results.</p>
Full article ">Figure 10
<p>Prescribing information page.</p>
Full article ">
17 pages, 2681 KiB  
Article
A Parallel Prediction Model for Photovoltaic Power Using Multi-Level Attention and Similar Day Clustering
by Jinming Gao, Xianlong Su, Changsu Kim, Kerang Cao and Hoekyung Jung
Energies 2024, 17(16), 3958; https://doi.org/10.3390/en17163958 - 9 Aug 2024
Viewed by 263
Abstract
Photovoltaic (PV) power generation is significantly impacted by environmental factors that exhibit substantial uncertainty and volatility, posing a critical challenge for accurate PV power prediction in power system management. To address this, a parallel model is proposed for PV short-term prediction utilizing a [...] Read more.
Photovoltaic (PV) power generation is significantly impacted by environmental factors that exhibit substantial uncertainty and volatility, posing a critical challenge for accurate PV power prediction in power system management. To address this, a parallel model is proposed for PV short-term prediction utilizing a multi-level attention mechanism. Firstly, gray relation analysis (GRA) and an improved ISODATA algorithm are used to select a dataset of similar days with comparable meteorological characteristics to the forecast day. A transformer encoder layer with multi-head attention is then used to extract long-term dependency features. Concurrently, BiGRU, optimized with a Global Attention network, is used to capture global temporal features. Feature fusion is performed using Cross Attention, calculating attention weights to emphasize significant features and enhancing feature integration. Finally, high-precision predictions are achieved through a fully connected layer. Utilizing historical PV power generation data to predict power output under various weather conditions, the proposed model demonstrates superior performance across all three climate types compared to other models, achieving more reliable predictions. Full article
(This article belongs to the Section F5: Artificial Intelligence and Smart Energy)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Transformer–multi-head attention.</p>
Full article ">Figure 2
<p>GRU and BiGRU structure diagrams. (<b>a</b>) GRU structure diagram. (<b>b</b>) BiGRU structure diagram.</p>
Full article ">Figure 3
<p>Global Attention structure.</p>
Full article ">Figure 4
<p>Cross Attention structure.</p>
Full article ">Figure 5
<p>Multi-level attention parallel architecture.</p>
Full article ">Figure 6
<p>Clustering indicator.</p>
Full article ">Figure 7
<p>Results of different weather model predictions. (<b>a</b>) Sunny day model prediction results. (<b>b</b>) Cloudy day model prediction results. (<b>c</b>) Rainy day model prediction results.</p>
Full article ">
16 pages, 2022 KiB  
Article
A Novel Accident Duration Prediction Method Based on a Conditional Table Generative Adversarial Network and Transformer
by Yongdong Wang, Haonan Zhai, Xianghong Cao and Xin Geng
Sustainability 2024, 16(16), 6821; https://doi.org/10.3390/su16166821 - 9 Aug 2024
Viewed by 315
Abstract
The accurate duration prediction of road traffic accident is crucial for ensuring the safe and efficiency of transportation within social road networks. Such predictive capabilities provide significant support for informed decision-making by transportation administrators while also offering new technological support for the sustainable [...] Read more.
The accurate duration prediction of road traffic accident is crucial for ensuring the safe and efficiency of transportation within social road networks. Such predictive capabilities provide significant support for informed decision-making by transportation administrators while also offering new technological support for the sustainable development of modern road networks. This study introduced a novel predictive model for road traffic accident duration, integrating a Conditional Table Generative Adversarial Network (CTGAN) with a transformer architecture. We initially utilized CTGAN to augment and refine the historical accident dataset. Subsequently, we implemented a wavelet denoising technique to cleanse the expanded dataset. The core of our model lies in the application of the transformer mechanism, which was trained to forecast the accident duration with high precision. To prove the effectiveness of our proposed model, a series of comparative experiments were designed and executed. The experimental results show that the prediction error of CTGAN-Tr for accident duration in the accident area could reach below 0.8. Compared with other models, the MAE of CTGAN-Tr was reduced by 0.31 compared with GRU, and the correlation coefficient was increased by 0.2 compared with TCN. At the same time, the model can show excellent performance in the other two accident areas. The results of these experiments not only substantiate the performance of our model but also demonstrate its robustness and generalizability when applied to traffic accident data from other regions. Full article
Show Figures

Figure 1

Figure 1
<p>The framework of the proposed method.</p>
Full article ">Figure 2
<p>CTGAN structural framework diagram.</p>
Full article ">Figure 3
<p>Transformer structural framework diagram.</p>
Full article ">Figure 4
<p>Comparison of evaluation indicators.</p>
Full article ">Figure 5
<p>Comparison of different evaluation indicators in the ablation experiments.</p>
Full article ">Figure 6
<p>The comparison of different methods for evaluation indicators.</p>
Full article ">Figure 7
<p>The comparison of various models under different regions (CA LA: Los Angeles, California; FL MIA: Miami, Florida; TX HOU: Houston, Texas).</p>
Full article ">
15 pages, 2698 KiB  
Article
Salinity Prediction Based on Improved LSTM Model in the Qiantang Estuary, China
by Rong Zheng, Zhilin Sun, Jiange Jiao, Qianqian Ma and Liqin Zhao
J. Mar. Sci. Eng. 2024, 12(8), 1339; https://doi.org/10.3390/jmse12081339 - 7 Aug 2024
Viewed by 444
Abstract
Accurate prediction of estuarine salinity can effectively mitigate the adverse effects of saltwater intrusion and help ensure the safety of water resources in estuarine regions. Presently, diverse data-driven models, mainly neural network models, have been employed to predict tidal estuarine salinity and obtained [...] Read more.
Accurate prediction of estuarine salinity can effectively mitigate the adverse effects of saltwater intrusion and help ensure the safety of water resources in estuarine regions. Presently, diverse data-driven models, mainly neural network models, have been employed to predict tidal estuarine salinity and obtained considerable achievements. Due to the nonlinear and nonstationary features of estuarine salinity sequences, this paper proposed a multi-factor salinity prediction model using an enhanced Long Short-Term Memory (LSTM) network. To improve prediction accuracy, input variables of the model were determined through Grey Relational Analysis (GRA) combined with estuarine dynamic analysis, and hyperparameters for the LSTM model were optimized using a multi-strategy Improved Sparrow Search Algorithm (ISSA). The proposed ISSA-LSTM model was applied to predict salinity at the Cangqian and Qibao stations in the Qiantang Estuary of China, based on measured data from 2011–2012. The model performance is evaluated by mean absolute error (MAE), mean absolute percentage error (MAPE), root mean square error (RMSE), and Nash-Sutcliffe efficiency (NSE). The results show that compared to other models including Back Propagation neural network (BP), Gate Recurrent Unit (GRU), and LSTM model, the new model has smaller errors and higher prediction accuracy, with NSE improved by 8–32% and other metrics (MAP, MAPE, RMSE) improved by 15–67%. Meanwhile, compared with LSTM optimized with the original SSA (SSA-LSTM), MAE, MAPE, and RMSE values of the new model decreased by 13–16%, 15–16%, and 11–13%, and NSE value increased by 5–6%, indicating that the ISSA has a better hyperparameter optimization ability than the original SSA. Thus, the model provides a practical solution for the rapid and precise prediction of estuarine salinity. Full article
Show Figures

Figure 1

Figure 1
<p>A map of study area. (<b>a</b>,<b>b</b>) Monitoring stations along the Qiantang Estuary. The discharge data is provided by Fuchunjiang hydrological station (FCJ), the water level data is provided by Ganpu station (GP), the salinity data is provided by CQ and QB station, and the wind speed data is provided by Hangzhou station (HZ).</p>
Full article ">Figure 2
<p>Overall framework and flowchart of the ISSA-LSTM model. Part 1 is data preprocessing and feature selection; Part 2 is hyperparameters optimization by ISSA; Part 3 is the LSTM model.</p>
Full article ">Figure 3
<p>Flowchart of the SSA and ISSA. (<b>a</b>) SSA. (<b>b</b>) ISSA.</p>
Full article ">Figure 4
<p>Prediction results of different models. (<b>a</b>) CQ station. (<b>b</b>) QB station. The gray color block represents observed values of daily maximum salinity. The green, brown, purple, blue, and red lines represent the predicated results of BP, GRU, LSTM, SSA-LSTM, and ISSA-LSTM models. The light orange region is zoomed in and shown in the small window in the subgraph (7/15–8/15, 10/15–11/15).</p>
Full article ">Figure 5
<p>Scatterplot of observed values and predicted values of different models. (<b>a</b>) CQ station. (<b>b</b>) QB station. The green, brown, purple, blue, and red lines represent the results of BP, GRU, LSTM, SSA-LSTM, and ISSA-LSTM models.</p>
Full article ">Figure 6
<p>Comparison of prediction results under different discharge conditions. (<b>a</b>) CQ station. (<b>b</b>) QB station. The black solid line represents the salinity prediction result for the original discharge and the yellow and blue dash lines represent the salinity prediction results for discharge decreased or increased by 50%.</p>
Full article ">
20 pages, 3083 KiB  
Article
Efficient Detection of Irrelevant User Reviews Using Machine Learning
by Cheolgi Kim and Hyeon Gyu Kim
Appl. Sci. 2024, 14(16), 6900; https://doi.org/10.3390/app14166900 - 7 Aug 2024
Viewed by 352
Abstract
User reviews such as SNS feeds and blog writings have been widely used to extract opinions, complains, and requirements about a given place or product from users’ perspective. However, during the process of collecting them, a lot of reviews that are irrelevant to [...] Read more.
User reviews such as SNS feeds and blog writings have been widely used to extract opinions, complains, and requirements about a given place or product from users’ perspective. However, during the process of collecting them, a lot of reviews that are irrelevant to a given search keyword can be included in the results. Such irrelevant reviews may lead to distorted results in data analysis. In this paper, we discuss a method to detect irrelevant user reviews efficiently by combining various oversampling and machine learning algorithms. About 35,000 user reviews collected from 25 restaurants and 33 tourist attractions in Ulsan Metropolitan City, South Korea, were used for learning, where the ratio of irrelevant reviews in the two kinds of data sets was 53.7% and 71.6%, respectively. To deal with skewness in the collected reviews, oversampling algorithms such as SMOTE, Borderline-SMOTE, and ADASYN were used. To build a model for the detection of irrelevant reviews, RNN, LSTM, GRU, and BERT were adopted and compared, as they are known to provide high accuracy in text processing. The performance of the detection models was examined through experiments, and the results showed that the BERT model presented the best performance, with an F1 score of 0.965. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Combination of data imbalance processing algorithms and machine learning algorithms that can be used to implement a model for the detection of irrelevant reviews.</p>
Full article ">Figure 2
<p>Comparison of the models’ precision in terms of data imbalance processing algorithms obtained from (<b>a</b>) 25 review data sets for the restaurants and (<b>b</b>) 33 review data sets for the tourist attractions.</p>
Full article ">Figure 3
<p>Comparison of TP and FP values of the models obtained from 25 review data sets for the restaurants where the ratio of irrelevant reviews is about 53.7%.</p>
Full article ">Figure 4
<p>Comparison of TP and FP values of the models obtained from 33 review data sets for the tourist attractions where the ratio of irrelevant reviews is about 71.6%.</p>
Full article ">Figure 5
<p>Comparison of the models’ recall in terms of data imbalance processing algorithms obtained from (<b>a</b>) 25 review data sets for the restaurants and (<b>b</b>) 33 review data sets for the tourist attractions.</p>
Full article ">Figure 6
<p>Comparison of TP and FN of the models obtained from 25 review data sets for the restaurants where the ratio of irrelevant reviews is about 53.7%.</p>
Full article ">Figure 7
<p>Comparison of the TP and FN values of the models obtained from 33 review data sets for the tourist attractions where the ratio of irrelevant reviews is about 71.6%.</p>
Full article ">Figure 8
<p>Comparison of the models’ F1 scores in terms of data imbalance processing algorithms obtained from (<b>a</b>) 25 review data sets for the restaurants and (<b>b</b>) 33 review data sets for the tourist attractions.</p>
Full article ">Figure 9
<p>Comparison of the models’ balanced accuracy in terms of data imbalance processing algorithms, obtained from (<b>a</b>) 25 review data sets for the restaurants and (<b>b</b>) 33 review data sets for the tourist attractions.</p>
Full article ">
21 pages, 4580 KiB  
Article
Driving Attention State Detection Based on GRU-EEGNet
by Xiaoli Wu, Changcheng Shi and Lirong Yan
Sensors 2024, 24(16), 5086; https://doi.org/10.3390/s24165086 - 7 Aug 2024
Viewed by 300
Abstract
The present study utilizes the significant differences in θ, α, and β band power spectra observed in electroencephalograms (EEGs) during distracted versus focused driving. Three subtasks, visual distraction, auditory distraction, and cognitive distraction, were designed to appear randomly during driving simulations. The θ, [...] Read more.
The present study utilizes the significant differences in θ, α, and β band power spectra observed in electroencephalograms (EEGs) during distracted versus focused driving. Three subtasks, visual distraction, auditory distraction, and cognitive distraction, were designed to appear randomly during driving simulations. The θ, α, and β band power spectra of the EEG signals of the four driving attention states were extracted, and SVM, EEGNet, and GRU-EEGNet models were employed for the detection of the driving attention states, respectively. Online experiments were conducted. The extraction of the θ, α, and β band power spectrum features of the EEG signals was found to be a more effective method than the extraction of the power spectrum features of the whole EEG signals for the detection of driving attention states. The driving attention state detection accuracy of the proposed GRU-EEGNet model is improved by 6.3% and 12.8% over the EEGNet model and PSD_SVM method, respectively. The EEG decoding method combining EEG features and an improved deep learning algorithm, which effectively improves the driving attention state detection accuracy, was manually and preliminarily selected based on the results of existing studies. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

Figure 1
<p>Simulated driving experimental platform. (<b>a</b>) Simulated driving display with current lap and vehicle speed parameters displayed in the upper left corner and visual distraction subtasks displayed in the upper right corner. (<b>b</b>) Experimental road. (<b>c</b>) The experimental scene with a subject wearing an EEG cap on their head.</p>
Full article ">Figure 2
<p>Visual and auditory distraction paradigm. (<b>a</b>) Visual distraction paradigm. (<b>b</b>) Auditory distraction paradigm.</p>
Full article ">Figure 3
<p>Distracted versus focused driving power spectra by channel. (<b>A</b>) Visual distracted versus focused driving power spectra; (<b>B</b>) auditory distracted versus focused driving power spectra; (<b>C</b>) cognitive distracted versus focused driving power spectra.</p>
Full article ">Figure 4
<p>Power spectra of each channel in the θ, α, and β bands of the driving attention state. (<b>A</b>) visual distraction vs. focused driving; (<b>B</b>) auditory distraction vs. focused driving; (<b>C</b>) cognitive distraction vs. focused driving.</p>
Full article ">Figure 5
<p>Schematic diagram of EEGNet network structure.</p>
Full article ">Figure 6
<p>Schematic diagram of the internal structure of the GRU. (<b>A</b>) schematic diagram of a unit GRU structure; (<b>B</b>) schematic diagram of three GRU units connected in stages.</p>
Full article ">Figure 7
<p>Schematic of EEG decoding for GRU-EEGNet model.</p>
Full article ">Figure 8
<p>GRU network structure and parameters.</p>
Full article ">Figure 9
<p>Classification performance of six commonly used EEG classification methods (f: focused; v: visual; a: auditory; c: cognitive; fv: focused vs. visually distracted driving classification; fa: focused vs. auditory distracted driving classification; fc: focused vs. cognitive distracted driving classification; fva: focused, visually distracted vs. auditorily distracted driving triple classification; fvc: focused, visually distracted vs. cognitive distracted driving triple classification; fca: focused, cognitive distracted vs. auditorily distracted driving triple classification; fvac: focused, visually distracted, auditorily distracted vs. cognitively distracted driving quadruple classification).</p>
Full article ">Figure 10
<p>Average classification accuracy of full-band and three-band feature extraction methods.</p>
Full article ">Figure 11
<p>EEGNet and GRU-EEGNet classification performance. (<b>A</b>) Average classification accuracy of PSD_SVM, EEGNet, and GRU-EEGNet; (<b>B</b>) performance of EEGNet and GRU-EEGNet (The horizontal line in the box represents the median, and the red circle represents the outlier). (E(2): EEGNet 2 classification; G-E(3): GRU-EEGNet 3 classification).</p>
Full article ">Figure 12
<p>Schematic structure of driving attention state detection system.</p>
Full article ">Figure 13
<p>Driving attention state detection system interface.</p>
Full article ">
18 pages, 5562 KiB  
Article
A Stock Market Decision-Making Framework Based on CMR-DQN
by Xun Chen, Qin Wang, Chao Hu and Chengqi Wang
Appl. Sci. 2024, 14(16), 6881; https://doi.org/10.3390/app14166881 - 6 Aug 2024
Viewed by 673
Abstract
In the dynamic and uncertain stock market, precise forecasting and decision-making are crucial for profitability. Traditional deep neural networks (DNN) often struggle with capturing long-term dependencies and multi-scale features in complex financial time series data. To address these challenges, we introduce CMR-DQN, an [...] Read more.
In the dynamic and uncertain stock market, precise forecasting and decision-making are crucial for profitability. Traditional deep neural networks (DNN) often struggle with capturing long-term dependencies and multi-scale features in complex financial time series data. To address these challenges, we introduce CMR-DQN, an innovative framework that integrates discrete wavelet transform (DWT) for multi-scale data analysis, temporal convolutional network (TCN) for extracting deep temporal features, and a GRU–LSTM–Attention mechanism to enhance the model’s focus and memory. Additionally, CMR-DQN employs the Rainbow DQN reinforcement learning strategy to learn optimal trading strategies in a simulated environment. CMR-DQN significantly improved the total return rate on six selected stocks, with increases ranging from 20.37% to 55.32%. It also demonstrated substantial improvements over the baseline model in terms of Sharpe ratio and maximum drawdown, indicating increased excess returns per unit of total risk and reduced investment risk. These results underscore the efficiency and effectiveness of CMR-DQN in handling multi-scale time series data and optimizing stock market decisions. Full article
Show Figures

Figure 1

Figure 1
<p>The architecture of CMR-DQN framework.</p>
Full article ">Figure 2
<p>The structural diagram of DWT-TCN.</p>
Full article ">Figure 3
<p>Working diagram of Rainbow DQN.</p>
Full article ">Figure 4
<p>Dueling architecture network.</p>
Full article ">Figure 5
<p>The accumulation of rewards and the variation trend of the loss function during the training process of the CMR-DQN model on six datasets.</p>
Full article ">Figure 6
<p>Results of Different Models on Six Datasets.</p>
Full article ">
16 pages, 1818 KiB  
Article
FFA-BiGRU: Attention-Based Spatial-Temporal Feature Extraction Model for Music Emotion Classification
by Yuping Su, Jie Chen, Ruiting Chai, Xiaojun Wu and Yumei Zhang
Appl. Sci. 2024, 14(16), 6866; https://doi.org/10.3390/app14166866 - 6 Aug 2024
Viewed by 328
Abstract
Music emotion recognition is becoming an important research direction due to its great significance for music information retrieval, music recommendation, and so on. In the task of music emotion recognition, the key to achieving accurate emotion recognition lies in how to extract the [...] Read more.
Music emotion recognition is becoming an important research direction due to its great significance for music information retrieval, music recommendation, and so on. In the task of music emotion recognition, the key to achieving accurate emotion recognition lies in how to extract the affect-salient features fully. In this paper, we propose an end-to-end spatial-temporal feature extraction method named FFA-BiGRU for music emotion classification. Taking the log Mel-spectrogram of music audio as the input, this method employs an attention-based convolutional residual module named FFA, which serves as a spatial feature learning module to obtain multi-scale spatial features. In the FFA module, three group architecture blocks extract multi-level spatial features, each of which consists of a stack of multiple channel-spatial attention-based residual blocks. Then, the output features from FFA are fed into the bidirectional gated recurrent units (BiGRU) module to capture the temporal features of music further. In order to make full use of the extracted spatial and temporal features, the output feature maps of FFA and those of the BiGRU are concatenated in the channel dimension. Finally, the concatenated features are passed through fully connected layers to predict the emotion classification results. The experimental results of the EMOPIA dataset show that the proposed model achieves better classification accuracy than the existing baselines. Meanwhile, the ablation experiments also demonstrate the effectiveness of each part of the proposed method. Full article
Show Figures

Figure 1

Figure 1
<p>An overview of the proposed FFA-BiGRU model. The model consists of a spatial feature learning module (FFA), a temporal feature learning module (BiGRU), and an emotion prediction module.</p>
Full article ">Figure 2
<p>Channel-spatial attention mechanism. The input feature maps are first scaled by channel attention weights and then by spatial attention weights. The filter size of Conv layers in both the CA and SA blocks is 1 × 1.</p>
Full article ">Figure 3
<p>The architecture of the GRU unit. It has two doors: reset door <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>r</mi> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> and update door <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>z</mi> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 4
<p>The FC layer for emotion prediction.</p>
Full article ">Figure 5
<p>Confusion matrix of the proposed model.</p>
Full article ">
20 pages, 11655 KiB  
Article
Daily Runoff Prediction Based on FA-LSTM Model
by Qihui Chai, Shuting Zhang, Qingqing Tian, Chaoqiang Yang and Lei Guo
Water 2024, 16(16), 2216; https://doi.org/10.3390/w16162216 - 6 Aug 2024
Viewed by 587
Abstract
Accurate and reliable short-term runoff prediction plays a pivotal role in water resource management, agriculture, and flood control, enabling decision-makers to implement timely and effective measures to enhance water use efficiency and minimize losses. To further enhance the accuracy of runoff prediction, this [...] Read more.
Accurate and reliable short-term runoff prediction plays a pivotal role in water resource management, agriculture, and flood control, enabling decision-makers to implement timely and effective measures to enhance water use efficiency and minimize losses. To further enhance the accuracy of runoff prediction, this study proposes a FA-LSTM model that integrates the Firefly algorithm (FA) with the long short-term memory neural network (LSTM). The research focuses on historical daily runoff data from the Dahuangjiangkou and Wuzhou Hydrology Stations in the Xijiang River Basin. The FA-LSTM model is compared with RNN, LSTM, GRU, SVM, and RF models. The FA-LSTM model was used to carry out the generalization experiment in Qianjiang, Wuxuan, and Guigang hydrology stations. Additionally, the study analyzes the performance of the FA-LSTM model across different forecasting horizons (1–5 days). Four quantitative evaluation metrics—mean absolute error (MAE), root mean square error (RMSE), coefficient of determination (R2), and Kling–Gupta efficiency coefficient (KGE)—are utilized in the evaluation process. The results indicate that: (1) Compared to RNN, LSTM, GRU, SVM, and RF models, the FA-LSTM model exhibits the best prediction performance, with daily runoff prediction determination coefficients (R2) reaching as high as 0.966 and 0.971 at the Dahuangjiangkou and Wuzhou Stations, respectively, and the KGE is as high as 0.965 and 0.960, respectively. (2) FA-LSTM model was used to conduct generalization tests at Qianjiang, Wuxuan and Guigang hydrology stations, and its R2 and KGE are 0.96 or above, indicating that the model has good adaptability in different hydrology stations and strong robustness. (3) As the prediction period extends, the R2 and KGE of the FA-LSTM model show a decreasing trend, but the whole model still showed feasible forecasting ability. The FA-LSTM model introduced in this study presents an effective new approach for daily runoff prediction. Full article
(This article belongs to the Section Hydrology)
Show Figures

Figure 1

Figure 1
<p>Location of selected hydrological control station.</p>
Full article ">Figure 2
<p>RNN model basic structure diagram.</p>
Full article ">Figure 3
<p>LSTM model basic structure diagram.</p>
Full article ">Figure 4
<p>GRU model basic structure diagram.</p>
Full article ">Figure 5
<p>SVM model basic structure diagram.</p>
Full article ">Figure 6
<p>RF model schematic.</p>
Full article ">Figure 7
<p>FA flow chart.</p>
Full article ">Figure 8
<p>FA optimizes LSTM model flow chart.</p>
Full article ">Figure 9
<p><span class="html-italic">R</span><sup>2</sup> index trends of four hyperparameters of LSTM model.</p>
Full article ">Figure 10
<p>Comparison of predicted and observed runoff values.</p>
Full article ">Figure 11
<p>Scatter plots of predicted and observed values for (<b>a</b>) DHJK station and (<b>b</b>) WZ station.</p>
Full article ">Figure 12
<p>(<b>a</b>) Comparison of predicted and observed values at QJ station. (<b>b</b>) Comparison of predicted and observed values at WX station. (<b>c</b>) Comparison of predicted and observed values at GG station.</p>
Full article ">Figure 12 Cont.
<p>(<b>a</b>) Comparison of predicted and observed values at QJ station. (<b>b</b>) Comparison of predicted and observed values at WX station. (<b>c</b>) Comparison of predicted and observed values at GG station.</p>
Full article ">Figure 12 Cont.
<p>(<b>a</b>) Comparison of predicted and observed values at QJ station. (<b>b</b>) Comparison of predicted and observed values at WX station. (<b>c</b>) Comparison of predicted and observed values at GG station.</p>
Full article ">Figure 13
<p>Prediction results of FA-LSTM model for (<b>a</b>) DHJK station and (<b>b</b>) WZ station under different forecast periods.</p>
Full article ">Figure 14
<p>Taylor diagram for (<b>a</b>) DHJK Station and (<b>b</b>) WZ station.</p>
Full article ">Figure 15
<p>Violin plots of predicted and measured values for (<b>a</b>) DHJK station and (<b>b</b>) WZ station.</p>
Full article ">
46 pages, 8707 KiB  
Article
Design and Enhancement of a Fog-Enabled Air Quality Monitoring and Prediction System: An Optimized Lightweight Deep Learning Model for a Smart Fog Environmental Gateway
by Divya Bharathi Pazhanivel, Anantha Narayanan Velu and Bagavathi Sivakumar Palaniappan
Sensors 2024, 24(15), 5069; https://doi.org/10.3390/s24155069 - 5 Aug 2024
Viewed by 558
Abstract
Effective air quality monitoring and forecasting are essential for safeguarding public health, protecting the environment, and promoting sustainable development in smart cities. Conventional systems are cloud-based, incur high costs, lack accurate Deep Learning (DL)models for multi-step forecasting, and fail to optimize DL models [...] Read more.
Effective air quality monitoring and forecasting are essential for safeguarding public health, protecting the environment, and promoting sustainable development in smart cities. Conventional systems are cloud-based, incur high costs, lack accurate Deep Learning (DL)models for multi-step forecasting, and fail to optimize DL models for fog nodes. To address these challenges, this paper proposes a Fog-enabled Air Quality Monitoring and Prediction (FAQMP) system by integrating the Internet of Things (IoT), Fog Computing (FC), Low-Power Wide-Area Networks (LPWANs), and Deep Learning (DL) for improved accuracy and efficiency in monitoring and forecasting air quality levels. The three-layered FAQMP system includes a low-cost Air Quality Monitoring (AQM) node transmitting data via LoRa to the Fog Computing layer and then the cloud layer for complex processing. The Smart Fog Environmental Gateway (SFEG) in the FC layer introduces efficient Fog Intelligence by employing an optimized lightweight DL-based Sequence-to-Sequence (Seq2Seq) Gated Recurrent Unit (GRU) attention model, enabling real-time processing, accurate forecasting, and timely warnings of dangerous AQI levels while optimizing fog resource usage. Initially, the Seq2Seq GRU Attention model, validated for multi-step forecasting, outperformed the state-of-the-art DL methods with an average RMSE of 5.5576, MAE of 3.4975, MAPE of 19.1991%, R2 of 0.6926, and Theil’s U1 of 0.1325. This model is then made lightweight and optimized using post-training quantization (PTQ), specifically dynamic range quantization, which reduced the model size to less than a quarter of the original, improved execution time by 81.53% while maintaining forecast accuracy. This optimization enables efficient deployment on resource-constrained fog nodes like SFEG by balancing performance and computational efficiency, thereby enhancing the effectiveness of the FAQMP system through efficient Fog Intelligence. The FAQMP system, supported by the EnviroWeb application, provides real-time AQI updates, forecasts, and alerts, aiding the government in proactively addressing pollution concerns, maintaining air quality standards, and fostering a healthier and more sustainable environment. Full article
(This article belongs to the Special Issue Artificial Intelligence Methods for Smart Cities—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Post-training optimization methods provided by TensorFlow.</p>
Full article ">Figure 2
<p>A three-layered Fog Computing-based architecture of the proposed system.</p>
Full article ">Figure 3
<p>Hardware of the proposed FAQMP system. (<b>a</b>) Air Quality Monitoring (AQM) Sensor Node. (<b>b</b>) Smart Fog Environmental Gateway (SFEG).</p>
Full article ">Figure 4
<p>Architecture and data flow of the proposed Fog-enabled Air Quality Monitoring and Prediction (FAQMP) System.</p>
Full article ">Figure 5
<p>DL model deployment pipeline after model quantization.</p>
Full article ">Figure 6
<p>Real-time alerts triggered by anomalous AQI Levels via email.</p>
Full article ">Figure 7
<p>Graphical User Interface of the EnviroWeb application displaying the live pollutants, Air Quality Index (AQI) level, and recommendations for citizens in real time.</p>
Full article ">Figure 8
<p>City-wide implementation of the proposed FAQMP system.</p>
Full article ">Figure 9
<p>GRU architecture.</p>
Full article ">Figure 10
<p>Architecture of the Sequence-to-Sequence GRU Attention mechanism.</p>
Full article ">Figure 11
<p>Steps involved in multivariate multi-step air quality forecasting.</p>
Full article ">Figure 12
<p>Error metrics of DL models to forecast PM<sub>2.5</sub> over twelve time steps (t1–t12). (<b>a</b>) RMSE comparison; (<b>b</b>) MAE comparison; (<b>c</b>) MAPE comparison.</p>
Full article ">Figure 13
<p>Performance metrics of DL models to forecast PM<sub>2.5</sub> over twelve time steps (t1–t12). (<b>a</b>) R<sup>2</sup> comparison; (<b>b</b>) Theil’s U1 comparison.</p>
Full article ">Figure 14
<p>Error metrics of DL models to forecast PM<sub>10</sub> over twelve time steps (t1–t12). (<b>a</b>) RMSE comparison; (<b>b</b>) MAE comparison; (<b>c</b>) MAPE comparison.</p>
Full article ">Figure 15
<p>Performance metrics of DL models to forecast PM<sub>10</sub> over twelve time steps (t1–t12). (<b>a)</b> R<sup>2</sup> comparison; (<b>b</b>) Theil’s U1 comparison.</p>
Full article ">Figure 16
<p>Performance metrics (RMSE, MAE, MAPE, R<sup>2</sup>, and U1) of the compared models across all pollutants (PM2.5, PM10, NO2, SO2, CO, and O3) over 12 time steps (t1–t12): (<b>a</b>) Average RMSE; (<b>b</b>) Average MAE; (<b>c</b>) Average MAPE; (<b>d</b>) Average R<sup>2</sup>; (<b>e</b>) Average Theil’s U1; and Model 1—GRU, Model 2—LSTM-GRU, Model 3—Seq2Seq GRU, Model 4—GRU Autoencoder, Model 5—GRU-LSTM Autoencoder, Model 6—GRU Attention, Model 7—LSTM-GRU Attention, Model 8—Seq2Seq LSTM Attention, Model 9—Seq2Seq Bi-LSTM Attention, and Our model—Seq2Seq GRU Attention.</p>
Full article ">Figure 17
<p>TensorFlow Lite models—file size comparison.</p>
Full article ">
19 pages, 4827 KiB  
Article
Heart Murmur Quality Detection Using Deep Neural Networks with Attention Mechanism
by Tingwei Wu, Zhaohan Huang, Shilong Li, Qijun Zhao and Fan Pan
Appl. Sci. 2024, 14(15), 6825; https://doi.org/10.3390/app14156825 - 5 Aug 2024
Viewed by 451
Abstract
Heart murmurs play a critical role in assessing the condition of the heart. Murmur quality reflects the subjective human perception of heart murmurs and is an important characteristic strongly linked to cardiovascular diseases (CVDs). This study aims to use deep neural networks to [...] Read more.
Heart murmurs play a critical role in assessing the condition of the heart. Murmur quality reflects the subjective human perception of heart murmurs and is an important characteristic strongly linked to cardiovascular diseases (CVDs). This study aims to use deep neural networks to classify the patients’ murmur quality (i.e., harsh and blowing) from phonocardiogram (PCG) signals. The phonocardiogram recordings with murmurs used for this task are from the CirCor DigiScope Phonocardiogram dataset, which provides the murmur quality labels. The recordings were segmented, and a dataset of 1266 segments with average lengths of 4.1 s from 164 patients’ recordings was obtained. Each patient usually has multiple segments. A deep neural network model based on convolutional neural networks (CNNs) with channel attention and gated recurrent unit (GRU) networks was first used to extract features from the log-Mel spectrograms of segments. Then, the features of different segments from one patient were weighted by the proposed “Feature Attention” module based on the attention mechanism. The “Feature Attention” module contains a layer of global pooling and two fully connected layers. Through it, the different features can learn their weight, which can help the deep learning model distinguish the importance of different features of one patient. Finally, the detection results were produced. The cross-entropy loss function was used to train the model, and five-fold cross-validation was employed to evaluate the performance of the proposed methods. The accuracy of detecting the quality of patients’ murmurs is 73.6%. The F1-scores (precision and recall) for the murmurs of harsh and blowing are 76.8% (73.0%, 83.0%) and 67.8% (76.0%, 63.3%), respectively. The proposed methods have been thoroughly evaluated and have the potential to assist physicians with the diagnosis of cardiovascular diseases as well as explore the relationship between murmur quality and cardiovascular diseases in depth. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of the methodology to detect murmur quality using deep neural network model. The dataset used is the analysis dataset defined in <a href="#sec3-applsci-14-06825" class="html-sec">Section 3</a>.</p>
Full article ">Figure 2
<p>The waveforms and log-Mel spectrograms of two typical heart sound segments (two with systolic murmurs described as “Harsh” and “Blowing”). n.u. refers to normalized units.</p>
Full article ">Figure 3
<p>Structure of the proposed neural network model. N corresponds to the batch size. K, S, and P correspond to the kernel size, stride, and padding, respectively, in the CNN-Block. C, H, and W correspond to the channel, height, and width of the input feature maps, respectively, in the CNN-Block and SE-Block. H and L in Bi-GRU correspond to the feature size and sequence length of the input feature sequences, respectively.</p>
Full article ">Figure 4
<p>(<b>a</b>) Structure of the Feature Weighting method. (<b>b</b>) Structure of the Feature Attention module. N corresponds to the number of segments for one patient. The two figures do not contain the batch dimension.</p>
Full article ">Figure 5
<p>Methods for adding characteristic information.</p>
Full article ">Figure 6
<p>The accuracy of different types of PCGs, including different types of “Timing”, “Shape”, “Grading”, and “Pitch”. Specific values for each type of PCG can be found in <a href="#applsci-14-06825-t002" class="html-table">Table 2</a>.</p>
Full article ">Figure 7
<p>The waveforms and log-Mel spectrograms of different types of PCGs: (<b>a</b>) the correctly categorized PCG with the timing of “Holosystolic”, the grading of “III”, and the pitch of “High”; (<b>b</b>) the misclassified PCG with the timing of “Early-systolic”; (<b>c</b>) the misclassified PCG with the grading of “II”; (<b>d</b>) the misclassified PCG with the pitch of “Low”. For succinctness, the PCGs only show 0–1.5 s (about 3 cardiac cycles), not all the segments.</p>
Full article ">
22 pages, 6538 KiB  
Article
A Method for Predicting Tool Remaining Useful Life: Utilizing BiLSTM Optimized by an Enhanced NGO Algorithm
by Jianwei Wu, Jiaqi Wang and Huanguo Chen
Mathematics 2024, 12(15), 2404; https://doi.org/10.3390/math12152404 - 2 Aug 2024
Viewed by 305
Abstract
Predicting remaining useful life (RUL) is crucial for tool condition monitoring (TCM) systems. Inaccurate predictions can lead to premature tool replacements or excessive usage, resulting in resource wastage and potential equipment failures. This study introduces a novel tool RUL prediction method that integrates [...] Read more.
Predicting remaining useful life (RUL) is crucial for tool condition monitoring (TCM) systems. Inaccurate predictions can lead to premature tool replacements or excessive usage, resulting in resource wastage and potential equipment failures. This study introduces a novel tool RUL prediction method that integrates the enhanced northern goshawk optimization (MSANGO) algorithm with a bidirectional long short-term memory (BiLSTM) network. Initially, key statistical features are extracted from collected signal data using multivariate variational mode decomposition. This is followed by effective feature reduction, facilitated by the uniform information coefficient and Mann–Kendall trend tests. The RUL predictions are subsequently refined through a BiLSTM network, with the MSANGO algorithm optimizing the network parameters. Comparative evaluations with BiLSTM, BiGRU, and NGO-BiLSTM models, as well as tests on real-world datasets, demonstrate this method’s superior accuracy and generalizability in RUL prediction, enhancing the efficacy of tool management systems. Full article
Show Figures

Figure 1

Figure 1
<p>Flowchart of the MSANGO algorithm.</p>
Full article ">Figure 2
<p>Flowchart of the MSANGO algorithm. (<b>a</b>) F1 (<b>b</b>) F2 (<b>c</b>) F3 (<b>d</b>) F4 (<b>e</b>) F5 (<b>f</b>) F6.</p>
Full article ">Figure 3
<p>Cell structure diagram of LSTM.</p>
Full article ">Figure 4
<p>Schematic diagram of BiLSTM.</p>
Full article ">Figure 5
<p>Schematic diagram of the experimental setup.</p>
Full article ">Figure 6
<p>The frequency at which various features appear across the experimental groups, alongside the average UIC values and the mean |<span class="html-italic">Z<sub>MK</sub></span>| values associated with each feature: (<b>a</b>) D1; (<b>b</b>) D2; and (<b>c</b>) D3.</p>
Full article ">Figure 6 Cont.
<p>The frequency at which various features appear across the experimental groups, alongside the average UIC values and the mean |<span class="html-italic">Z<sub>MK</sub></span>| values associated with each feature: (<b>a</b>) D1; (<b>b</b>) D2; and (<b>c</b>) D3.</p>
Full article ">Figure 7
<p>Schematic of the BiLSTM network structure.</p>
Full article ">Figure 8
<p>Convergence trajectories of the NGO and MSANGO algorithms. (<b>a</b>) D1 (<b>b</b>) D2 (<b>c</b>) D3.</p>
Full article ">Figure 9
<p>The predictive results of the four models on experimental group D1. (<b>a</b>) BiLSTM (<b>b</b>) BiGRU (<b>c</b>) NGO-BiLSTM (<b>d</b>) MSANGO-BiLSTM.</p>
Full article ">Figure 10
<p>The predictive results of the four models on experimental group D2. (<b>a</b>) BiLSTM (<b>b</b>) BiGRU (<b>c</b>) NGO-BiLSTM (<b>d</b>) MSANGO-BiLSTM.</p>
Full article ">Figure 11
<p>The predictive results of the four models on experimental group D3. (<b>a</b>) BiLSTM (<b>b</b>) BiGRU (<b>c</b>) NGO-BiLSTM (<b>d</b>) MSANGO-BiLSTM.</p>
Full article ">Figure 11 Cont.
<p>The predictive results of the four models on experimental group D3. (<b>a</b>) BiLSTM (<b>b</b>) BiGRU (<b>c</b>) NGO-BiLSTM (<b>d</b>) MSANGO-BiLSTM.</p>
Full article ">Figure 12
<p>Comparative of evaluation metrics among each model. (<b>a</b>) D1 (<b>b</b>) D2 (<b>c</b>) D3.</p>
Full article ">
17 pages, 10211 KiB  
Article
Digital Self-Interference Cancellation for Full-Duplex Systems Based on CNN and GRU
by Jun Liu and Tian Ding
Electronics 2024, 13(15), 3041; https://doi.org/10.3390/electronics13153041 - 1 Aug 2024
Viewed by 295
Abstract
Self-interference (SI) represents a bottleneck in the performance of full-duplex (FD) communication systems, necessitating robust offsetting techniques to unlock the potential of FD systems. Currently, deep learning has been leveraged within the communication domain to address specific challenges and enhance efficiency. Inspired by [...] Read more.
Self-interference (SI) represents a bottleneck in the performance of full-duplex (FD) communication systems, necessitating robust offsetting techniques to unlock the potential of FD systems. Currently, deep learning has been leveraged within the communication domain to address specific challenges and enhance efficiency. Inspired by this, this paper reviews the self-interference cancellation (SIC) process in the digital domain focusing on SIC capability. The paper introduces a model architecture that integrates CNN and gated recurrent unit (GRU), while also incorporating residual networks and self-attention mechanisms to enhance the identification and elimination of SI. This model is named CGRSA-Net. Firstly, CNN is employed to capture local signal features in the time–frequency domain. Subsequently, a ResNet module is introduced to mitigate the gradient vanishing problem. Concurrently, GRU is utilized to dynamically capture and retain both long- and short-term dependencies during the communication process. Lastly, by integrating the self-attention mechanism, attention weights are flexibly assigned when processing sequence data, thereby focusing on the most important parts of the input sequence. Experimental results demonstrate that the proposed CGRSA-Net model achieves a minimum of 28% improvement in nonlinear SIC capability compared to polynomial and existing neural network-based eliminator. Additionally, through ablation experiments, we demonstrate that the various modules utilized in this paper effectively learn signal features and further enhance SIC performance. Full article
Show Figures

Figure 1

Figure 1
<p>Overall flow chart of the article.</p>
Full article ">Figure 2
<p>Full-duplex system model.</p>
Full article ">Figure 3
<p>CNN model structure.</p>
Full article ">Figure 4
<p>Residual block.</p>
Full article ">Figure 5
<p>GRU structure.</p>
Full article ">Figure 6
<p>Self-attention structure.</p>
Full article ">Figure 7
<p>The proposed SIC model.</p>
Full article ">Figure 8
<p>PSD of the SI after applying cancellation schemes.</p>
Full article ">Figure 9
<p>Loss value curves.</p>
Full article ">Figure 10
<p>PSD of the SI after applying various cancellation schemes.</p>
Full article ">
17 pages, 4924 KiB  
Article
Integrating Machine Learning with Intelligent Control Systems for Flow Rate Forecasting in Oil Well Operations
by Bibars Amangeldy, Nurdaulet Tasmurzayev, Shona Shinassylov, Aksultan Mukhanbet and Yedil Nurakhov
Automation 2024, 5(3), 343-359; https://doi.org/10.3390/automation5030021 - 1 Aug 2024
Viewed by 437
Abstract
This study addresses the integration of machine learning (ML) with supervisory control and data acquisition (SCADA) systems to enhance predictive maintenance and operational efficiency in oil well monitoring. We investigated the applicability of advanced ML models, including Long Short-Term Memory (LSTM), Bidirectional LSTM [...] Read more.
This study addresses the integration of machine learning (ML) with supervisory control and data acquisition (SCADA) systems to enhance predictive maintenance and operational efficiency in oil well monitoring. We investigated the applicability of advanced ML models, including Long Short-Term Memory (LSTM), Bidirectional LSTM (BiLSTM), and Momentum LSTM (MLSTM), on a dataset of 21,644 operational records. These models were trained to predict a critical operational parameter, FlowRate, which is essential for operational integrity and efficiency. Our results demonstrate substantial improvements in predictive accuracy: the LSTM model achieved an R2 score of 0.9720, the BiLSTM model reached 0.9725, and the MLSTM model topped at 0.9726, all with exceptionally low Mean Absolute Errors (MAEs) around 0.0090 for LSTM and 0.0089 for BiLSTM and MLSTM. These high R2 values indicate that our models can explain over 97% of the variance in the dataset, reflecting significant predictive accuracy. Such performance underscores the potential of integrating ML with SCADA systems for real-time applications in the oil and gas industry. This study quantifies ML’s integration benefits and sets the stage for further advancements in autonomous well-monitoring systems. Full article
Show Figures

Figure 1

Figure 1
<p>Workflow of the system.</p>
Full article ">Figure 2
<p>General view of dataset.</p>
Full article ">Figure 3
<p>Correlation matrix of variables.</p>
Full article ">Figure 4
<p>Unrolled BILSTM.</p>
Full article ">Figure 5
<p>Transformer architecture.</p>
Full article ">Figure 6
<p>Overview of the control box.</p>
Full article ">Figure 7
<p>Overview of the intelligent data collection device.</p>
Full article ">Figure 8
<p>Supervisory control and data acquisition system.</p>
Full article ">Figure 9
<p>Comparison of test and predicted results of LSTM.</p>
Full article ">Figure 10
<p>Comparison of test and forecast results, as well as forecasting future results for a month.</p>
Full article ">Figure 11
<p>Comparative graph of results of all algorithms.</p>
Full article ">Figure 12
<p>Analysis of the effectiveness of integrating machine learning with intelligent control systems for cost forecasting in the oil industry according to the Ishikawa diagram.</p>
Full article ">
Back to TopTop