[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,178)

Search Parameters:
Keywords = BiLSTM

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 5772 KiB  
Article
Optimizing Football Formation Analysis via LSTM-Based Event Detection
by Benjamin Orr, Ephraim Pan and Dah-Jye Lee
Electronics 2024, 13(20), 4105; https://doi.org/10.3390/electronics13204105 (registering DOI) - 18 Oct 2024
Viewed by 131
Abstract
The process of manually annotating sports footage is a demanding one. In American football alone, coaches spend thousands of hours reviewing and analyzing videos each season. We aim to automate this process by developing a system that generates comprehensive statistical reports from full-length [...] Read more.
The process of manually annotating sports footage is a demanding one. In American football alone, coaches spend thousands of hours reviewing and analyzing videos each season. We aim to automate this process by developing a system that generates comprehensive statistical reports from full-length football game videos. Having previously demonstrated the proof of concept for our system, here, we present optimizations to our preprocessing techniques along with an inventive method for multi-person event detection in sports videos. Employing a long short-term memory (LSTM)-based architecture to detect the snap in American football, we achieve an outstanding LSI (Levenshtein similarity index) of 0.9445, suggesting a normalized difference of less than 0.06 between predictions and ground truth labels. We also illustrate the utility of snap detection as a means of identifying the offensive players’ assuming of formation. Our results exhibit not only the success of our unique approach and underlying optimizations but also the potential for continued robustness as we pursue the development of our remaining system components. Full article
(This article belongs to the Special Issue Deep Learning for Computer Vision Application)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>An abstract depiction of the work described in this paper. Part of a larger system, our pipeline takes a football game video clip as input and outputs the point in the clip where the offensive players are in formation.</p>
Full article ">Figure 2
<p>An example of the stark differences between our previous dataset [<a href="#B14-electronics-13-04105" class="html-bibr">14</a>] (<b>left</b>) and one of our new datasets (<b>right</b>). The previous dataset is a collection of images captured from the Madden 2020 video game. The new datasets were sourced from real-world game footage. The real-world data contain variations in weather, lighting, perspective, and image quality.</p>
Full article ">Figure 3
<p>An overview of our system pipeline. The input is a video clip of a football game. The locations of players and numbers are recognized using YOLOv8x models, and field lines are detected using traditional computer vision techniques. The lines and number locations are used to transform the player locations into bird’s-eye view and project them onto a virtual football field. The localized player locations are then input to our event detection model (LSTM), which detects when the players are in formation.</p>
Full article ">Figure 4
<p>A snapshot of the manual labeling process for our player recognition dataset. To ensure accurate player locations, we hand-labeled each image with the utmost attention to detail, focusing on tightly positioning bounding boxes around each player. This included players who were not 100% visible to the camera—a common occurrence in our dataset and other similar datasets. We completed our manual labeling using the makesense.ai object labeling tool [<a href="#B19-electronics-13-04105" class="html-bibr">19</a>].</p>
Full article ">Figure 5
<p>Examples of labels in our event detection dataset. The dataset consists of preprocessed sequences of player locations extracted from video clips. By visualizing five frames from a single sequence, we can clearly understand the labeling methodology. Player locations were used to assign a label to each frame, dividing the sequence into three distinct segments: <span class="html-italic">pre-snap</span> for frames before formation, <span class="html-italic">snap</span> for frames during formation and the snap, and <span class="html-italic">post-snap</span> for frames after formation breaks.</p>
Full article ">Figure 6
<p>A diagram of a single LSTM cell, based on [<a href="#B21-electronics-13-04105" class="html-bibr">21</a>]. Taking in a current input <math display="inline"><semantics> <msub> <mi>x</mi> <mi>t</mi> </msub> </semantics></math> and previous cell and hidden states <math display="inline"><semantics> <msub> <mi>c</mi> <mrow> <mi>t</mi> <mo>−</mo> <mn>1</mn> </mrow> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>h</mi> <mrow> <mi>t</mi> <mo>−</mo> <mn>1</mn> </mrow> </msub> </semantics></math>, the cell makes use of four layers of weights (comprising three gates) to learn what to forget, retain, and output at each time step. New cell and hidden states <math display="inline"><semantics> <msub> <mi>c</mi> <mi>t</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>h</mi> <mi>t</mi> </msub> </semantics></math> are then passed to the next cell or used in the next time step, as appropriate.</p>
Full article ">Figure 7
<p>The block diagram for our event detection network. Consisting of four bidirectional LSTM layers, each sequence <span class="html-italic">x</span> is serially processed in both directions simultaneously by two sets of four LSTM cells (see <a href="#electronics-13-04105-f006" class="html-fig">Figure 6</a>): by the forward set as <math display="inline"><semantics> <msub> <mi>x</mi> <mn>0</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>x</mi> <mn>1</mn> </msub> </semantics></math>…<math display="inline"><semantics> <msub> <mi>x</mi> <mrow> <mi>T</mi> <mo>−</mo> <mn>1</mn> </mrow> </msub> </semantics></math> and the backward set as <math display="inline"><semantics> <msub> <mi>x</mi> <mrow> <mi>T</mi> <mo>−</mo> <mn>1</mn> </mrow> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>x</mi> <mrow> <mi>T</mi> <mo>−</mo> <mn>2</mn> </mrow> </msub> </semantics></math>…<math display="inline"><semantics> <msub> <mi>x</mi> <mn>0</mn> </msub> </semantics></math>. Each of the eight cells has a unique set of weights and states. The cells are connected via dropout layers of probability 0.3 and hidden and cell states of size 128. For each time step <span class="html-italic">t</span>, the final hidden states of both sets of LSTM cells are concatenated to form the final output <math display="inline"><semantics> <msub> <mi>h</mi> <mi>t</mi> </msub> </semantics></math>, now of size 256. This is input into a single linear layer, which outputs the probabilities for each of our three classes.</p>
Full article ">Figure 8
<p>The F1–confidence curve for our player recognition model. This plot shows F1 scores for the whole range of confidence values, where F1 is the harmonic mean of precision and recall, and confidence is a measure of the model’s certainty about its predictions. We achieved an F1 score of 96% at confidence values of up to 0.9, demonstrating notable improvements over the 90.3% accuracy of this stage in our previous work [<a href="#B14-electronics-13-04105" class="html-bibr">14</a>].</p>
Full article ">Figure 9
<p>The F1–confidence curve for our number recognition model (see <a href="#electronics-13-04105-f008" class="html-fig">Figure 8</a> for an explanation of F1 and confidence). Each thin line represents the F1 score for an individual output class, while the thick dark blue line shows the average F1 score across all classes. Examining this curve, we see that the model consistently performed with an F1 score of 97%. Reaching this result on a dataset that is larger and more representative of the real world shows great improvement over the 96% accuracy of our previous work [<a href="#B15-electronics-13-04105" class="html-bibr">15</a>].</p>
Full article ">Figure 10
<p>An image of players alongside their locations transformed to bird’s-eye view and projected to a virtual field. Notice the highly accurate positioning of each player’s processed location. These quality results are the product of our enhanced approaches to both player and number recognition.</p>
Full article ">Figure 11
<p>A frame that corresponds to the exact middle of the predicted <span class="html-italic">snap</span> portion of a sequence in the test set. As intended, the offense team is shown to be in formation and ready to imminently snap the ball.</p>
Full article ">
31 pages, 6207 KiB  
Article
A Distributed VMD-BiLSTM Model for Taxi Demand Forecasting with GPS Sensor Data
by Hasan A. H. Naji, Qingji Xue and Tianfeng Li
Sensors 2024, 24(20), 6683; https://doi.org/10.3390/s24206683 - 17 Oct 2024
Viewed by 202
Abstract
With the ubiquitous deployment of mobile and sensor technologies in modes of transportation, taxis have become a significant component of public transportation. However, vacant taxis represent an important waste of transportation resources. Forecasting taxi demand within a short time achieves a supply–demand balance [...] Read more.
With the ubiquitous deployment of mobile and sensor technologies in modes of transportation, taxis have become a significant component of public transportation. However, vacant taxis represent an important waste of transportation resources. Forecasting taxi demand within a short time achieves a supply–demand balance and reduces oil emissions. Although earlier studies have forwarded highly developed machine learning- and deep learning-based models to forecast taxicab demands, these models often face significant computational expenses and cannot effectively utilize large-scale trajectory sensor data. To address these challenges, in this paper, we propose a hybrid deep learning-based model for taxi demand prediction. In particular, the Variational Mode Decomposition (VMD) algorithm is integrated along with a Bidirectional Long Short-Term Memory (BiLSTM) model to perform the prediction process. The VMD algorithm is applied to decompose time series-aware traffic features into multiple sub-modes of different frequencies. After that, the BiLSTM method is utilized to predict time series data fed with the relevant demand features. To overcome the limitation of high computational expenses, the designed model is performed on the Spark distributed platform. The performance of the proposed model is tested using a real-world dataset, and it surpasses existing state-of-the-art predictive models in terms of accuracy, efficiency, and distributed performance. These findings provide insights for enhancing the efficiency of passenger search and increasing the profit of taxicabs. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

Figure 1
<p>The structure of the distributed VMD-BiLSTM prediction model.</p>
Full article ">Figure 2
<p>Road map network of Wuhan City.</p>
Full article ">Figure 3
<p>Typical trajectory of taxi trips.</p>
Full article ">Figure 4
<p>Study area of Wuchang district.</p>
Full article ">Figure 5
<p>The distribution of taxi demands on the weekdays and weekends.</p>
Full article ">Figure 6
<p>Taxi demand distribution in the target area during holidays.</p>
Full article ">Figure 7
<p>The distribution of taxi demands in the target area over 24 h.</p>
Full article ">Figure 8
<p>Schematic diagram of the VMD-BiLSTM model.</p>
Full article ">Figure 9
<p>Flowchart of VMD algorithm.</p>
Full article ">Figure 10
<p>The transformation of the taxi demands time series into a two-dimensional array.</p>
Full article ">Figure 11
<p>Architecture of bidirectional LSTM network.</p>
Full article ">Figure 12
<p>Distributed implementation of VMD-BiLSTM model on Spark.</p>
Full article ">Figure 13
<p>Results of our prediction model on Wuhan’s dataset. (<b>a</b>) 1 day; (<b>b</b>) 1 week; (<b>c</b>) 2 weeks; (<b>d</b>) 1 month; (<b>e</b>) 2 months; and (<b>f</b>) whole dataset.</p>
Full article ">Figure 13 Cont.
<p>Results of our prediction model on Wuhan’s dataset. (<b>a</b>) 1 day; (<b>b</b>) 1 week; (<b>c</b>) 2 weeks; (<b>d</b>) 1 month; (<b>e</b>) 2 months; and (<b>f</b>) whole dataset.</p>
Full article ">Figure 14
<p>VMD renderings.</p>
Full article ">Figure 14 Cont.
<p>VMD renderings.</p>
Full article ">Figure 15
<p>Wavelet threshold denoising method for VMD renderings (IMF1, IMF2, and IMF3).</p>
Full article ">Figure 15 Cont.
<p>Wavelet threshold denoising method for VMD renderings (IMF1, IMF2, and IMF3).</p>
Full article ">Figure 16
<p>Box-plot of MOEs for Wuhan dataset.</p>
Full article ">Figure 17
<p>Comparison of loss function of distributed VMD-BiLSM.</p>
Full article ">Figure 18
<p>Running time (seconds) of VMD-BiLSTM based on Spark platform.</p>
Full article ">Figure 19
<p>Scaleup comparative analysis of distributed VMD-BiLSTM for different computing nodes.</p>
Full article ">Figure 20
<p>Speedup comparative analysis of the proposed model for different computing nodes.</p>
Full article ">
21 pages, 1156 KiB  
Article
EDSCVD: Enhanced Dual-Channel Smart Contract Vulnerability Detection Method
by Huaiguang Wu, Yibo Peng, Yaqiong He and Siqi Lu
Symmetry 2024, 16(10), 1381; https://doi.org/10.3390/sym16101381 - 17 Oct 2024
Viewed by 219
Abstract
Ensuring the absence of vulnerabilities or flaws in smart contracts before their deployment is crucial for the smooth progress of subsequent work. Existing detection methods heavily rely on expert rules, resulting in low robustness and accuracy. Therefore, we propose EDSCVD, an enhanced deep [...] Read more.
Ensuring the absence of vulnerabilities or flaws in smart contracts before their deployment is crucial for the smooth progress of subsequent work. Existing detection methods heavily rely on expert rules, resulting in low robustness and accuracy. Therefore, we propose EDSCVD, an enhanced deep learning vulnerability detection model based on dual-channel networks. Firstly, the contract fragments are preprocessed by BERT into the required word embeddings. Next, we utilized adversarial training FGM to the word embeddings to generate perturbations, thereby producing symmetric adversarial samples and enhancing the robustness of the model. Then, the dual-channel model combining BiLSTM and CNN is utilized for feature training to obtain more comprehensive and symmetric information on temporal and local contract features.Finally, the combined output features are passed through a classifier to classify and detect contract vulnerabilities. Experimental results show that our EDSCVD exhibits excellent detection performance in the detection of classical reentrancy vulnerabilities, timestamp dependencies, and integer overflow vulnerabilities. Full article
Show Figures

Figure 1

Figure 1
<p>Reentrancy source code. (Source: Own elaboration).</p>
Full article ">Figure 2
<p>Timestamp Dependency source code. (Source: Own elaboration).</p>
Full article ">Figure 3
<p>Integer Overflow source code. (∗ denotes the multiplication operator. <math display="inline"><semantics> <mrow> <mo>&amp;</mo> <mo>&amp;</mo> </mrow> </semantics></math> denotes a logical symbol used to combine two Boolean expressions). (Source: Own elaboration).</p>
Full article ">Figure 4
<p>The overall architecture of EDSCVD. (Source: Own elaboration).</p>
Full article ">Figure 5
<p>Contract fragment representation. (Source: Own elaboration).</p>
Full article ">Figure 6
<p>The structure of BERT. (Source: Own elaboration based on literature [<a href="#B49-symmetry-16-01381" class="html-bibr">49</a>]).</p>
Full article ">Figure 7
<p>Adversarial Training Methods FGM. (Source: Own elaboration).</p>
Full article ">Figure 8
<p>Dual-Channel Network Architecture. (Source: Own elaboration based on literature [<a href="#B49-symmetry-16-01381" class="html-bibr">49</a>]).</p>
Full article ">Figure 9
<p>Structure of a single LSTM module. (Source: Own elaboration based on literature [<a href="#B22-symmetry-16-01381" class="html-bibr">22</a>]).</p>
Full article ">Figure 10
<p>Multi-Head Attention Mechanisms. (Source: Own elaboration).</p>
Full article ">Figure 11
<p>Epochs and Evaluation Metrics in model training. (Source: Own elaboration).</p>
Full article ">
22 pages, 1654 KiB  
Article
A New Scene Sensing Model Based on Multi-Source Data from Smartphones
by Zhenke Ding, Zhongliang Deng, Enwen Hu, Bingxun Liu, Zhichao Zhang and Mingyang Ma
Sensors 2024, 24(20), 6669; https://doi.org/10.3390/s24206669 - 16 Oct 2024
Viewed by 216
Abstract
Smartphones with integrated sensors play an important role in people’s lives, and in advanced multi-sensor fusion navigation systems, the use of individual sensor information is crucial. Because of the different environments, the weights of the sensors will be different, which will also affect [...] Read more.
Smartphones with integrated sensors play an important role in people’s lives, and in advanced multi-sensor fusion navigation systems, the use of individual sensor information is crucial. Because of the different environments, the weights of the sensors will be different, which will also affect the method and results of multi-source fusion positioning. Based on the multi-source data from smartphone sensors, this study explores five types of information—Global Navigation Satellite System (GNSS), Inertial Measurement Units (IMUs), cellular networks, optical sensors, and Wi-Fi sensors—characterizing the temporal, spatial, and mathematical statistical features of the data, and it constructs a multi-scale, multi-window, and context-connected scene sensing model to accurately detect the environmental scene in indoor, semi-indoor, outdoor, and semi-outdoor spaces, thus providing a good basis for multi-sensor positioning in a multi-sensor navigation system. Detecting environmental scenes provides an environmental positioning basis for multi-sensor fusion localization. This model is divided into four main parts: multi-sensor-based data mining, a multi-scale convolutional neural network (CNN), a bidirectional long short-term memory (BiLSTM) network combined with contextual information, and a meta-heuristic optimization algorithm. Full article
(This article belongs to the Special Issue Smart Sensor Systems for Positioning and Navigation)
Show Figures

Figure 1

Figure 1
<p>Four scene classifications: (<b>a</b>) outdoor, (<b>b</b>) semi-outdoor, (<b>c</b>) semi-indoor, and (<b>d</b>) indoor.</p>
Full article ">Figure 2
<p>Satellite zenith view: (<b>a</b>) west indoor neighboring window, (<b>b</b>) south indoor neighbouring window, (<b>c</b>) indoor, and (<b>d</b>) open outdoor neighboring window.</p>
Full article ">Figure 3
<p>DOP change graph: (<b>a</b>) outdoor DOP change graph, and (<b>b</b>) indoor DOP change graph.</p>
Full article ">Figure 4
<p>Visible satellite map: (<b>a</b>) variation in the number of visible satellites. (<b>b</b>) Variation in the rate of change of visible satellites in different windows.</p>
Full article ">Figure 5
<p>Satellite signal quality map: (<b>a</b>) CNR variation and (<b>b</b>) DCNR variation.</p>
Full article ">Figure 6
<p>State of motion versus acceleration.</p>
Full article ">Figure 7
<p>Wi-Fi channel spectrum scan: (<b>a</b>) indoor, (<b>b</b>) outdoor.</p>
Full article ">Figure 8
<p>Visible AP distribution of Wi-Fi: (<b>a</b>) number distribution, (<b>b</b>) signal strength distribution.</p>
Full article ">Figure 9
<p>Variation of light sensors and cellular network sensors: (<b>a</b>) variation of indoor and outdoor light intensity over 24 h, (<b>b</b>) variation of the number of base stations receiving signals.</p>
Full article ">Figure 10
<p>An algorithmic model for the classification of complex indoor and outdoor scenes based on spatio-temporal features.</p>
Full article ">Figure 11
<p>Pearson correlation feature map.</p>
Full article ">Figure 12
<p>Schematic of a two-scale convolutional neural network.</p>
Full article ">Figure 13
<p>BiLSTM network structure diagram.</p>
Full article ">Figure 14
<p>Structure of the ablation experiment.</p>
Full article ">Figure 15
<p>Confusion matrix: (<b>a</b>) confusion matrix before WOA optimization. (<b>b</b>) confusion matrix after WOA optimisation.</p>
Full article ">Figure 16
<p>Comparison of the accuracy of different models.</p>
Full article ">Figure 17
<p>Comparison of accuracy in different scenarios.</p>
Full article ">
24 pages, 7433 KiB  
Article
Efficient Fault Warning Model Using Improved Red Deer Algorithm and Attention-Enhanced Bidirectional Long Short-Term Memory Network
by Yutian Wang and Mingli Wu
Processes 2024, 12(10), 2253; https://doi.org/10.3390/pr12102253 (registering DOI) - 15 Oct 2024
Viewed by 303
Abstract
The rapid advancement of industrial processes makes ensuring the stability of industrial equipment a critical factor in improving production efficiency and safeguarding operational safety. Fault warning systems, as a key technological means to enhance equipment stability, are increasingly gaining attention across industries. However, [...] Read more.
The rapid advancement of industrial processes makes ensuring the stability of industrial equipment a critical factor in improving production efficiency and safeguarding operational safety. Fault warning systems, as a key technological means to enhance equipment stability, are increasingly gaining attention across industries. However, as equipment structures and functions become increasingly complex, traditional fault warning methods face challenges such as limited prediction accuracy and difficulties in meeting real-time requirements. To address these challenges, this paper proposes an innovative hybrid fault warning method. The proposed approach integrates a multi-strategy improved red deer optimization algorithm (MIRDA), attention mechanism, and bidirectional long short-term memory network (BiLSTM). Firstly, the red deer optimization algorithm (RDA) is enhanced through improvements in population initialization strategy, adaptive optimal guidance strategy, chaos regulation factor, and double-sided mirror reflection theory, thereby enhancing its optimization performance. Subsequently, the MIRDA is employed to optimize the hyperparameters of the BiLSTM model incorporating an attention mechanism. A predictive model is then constructed based on the optimized Attention-BiLSTM, which, combined with a sliding window approach, provides robust support for fault threshold identification. The proposed algorithm’s efficacy is demonstrated through its application to real-world gas-fired power plant equipment fault cases. Comparative analyses with other advanced algorithms reveal its superior robustness and accuracy in efficiently issuing fault warnings. This research not only provides a more reliable safeguard for the stable operation of industrial equipment but also pioneers a new avenue for the application of metaheuristic algorithms. Full article
Show Figures

Figure 1

Figure 1
<p>Attention-BiLSTM network structure.</p>
Full article ">Figure 2
<p>Flowchart of MIRDA.</p>
Full article ">Figure 3
<p>Images of the feedwater pump.</p>
Full article ">Figure 4
<p>Predictive results on the testing set.</p>
Full article ">Figure 4 Cont.
<p>Predictive results on the testing set.</p>
Full article ">Figure 5
<p>The average of residuals obtained after processing with the sliding window method.</p>
Full article ">Figure 6
<p>Analysis results of the average residuals for the fault warning example.</p>
Full article ">Figure 7
<p>The 95% confidence intervals and standard deviation of algorithm performance (<a href="#sec5dot1-processes-12-02253" class="html-sec">Section 5.1</a>): RMSE (<b>a</b>), MAPE (<b>b</b>), standard deviation (<b>c</b>).</p>
Full article ">Figure 7 Cont.
<p>The 95% confidence intervals and standard deviation of algorithm performance (<a href="#sec5dot1-processes-12-02253" class="html-sec">Section 5.1</a>): RMSE (<b>a</b>), MAPE (<b>b</b>), standard deviation (<b>c</b>).</p>
Full article ">Figure 8
<p>The 95% confidence intervals and standard deviation of algorithm performance (<a href="#sec5dot2-processes-12-02253" class="html-sec">Section 5.2</a>): RMSE (<b>a</b>), MAPE (<b>b</b>), standard deviation (<b>c</b>).</p>
Full article ">Figure 8 Cont.
<p>The 95% confidence intervals and standard deviation of algorithm performance (<a href="#sec5dot2-processes-12-02253" class="html-sec">Section 5.2</a>): RMSE (<b>a</b>), MAPE (<b>b</b>), standard deviation (<b>c</b>).</p>
Full article ">Figure 9
<p>Graphical presentation of statistical results (<a href="#sec5dot1-processes-12-02253" class="html-sec">Section 5.1</a>): RMSE (<b>a</b>), MAPE (<b>b</b>).</p>
Full article ">Figure 10
<p>Graphical presentation of statistical results (<a href="#sec5dot2-processes-12-02253" class="html-sec">Section 5.2</a>): RMSE (<b>a</b>), MAPE (<b>b</b>).</p>
Full article ">
14 pages, 346 KiB  
Article
An ELECTRA-Based Model for Power Safety Named Entity Recognition
by Peng Liu, Zhenfu Sun and Biao Zhou
Appl. Sci. 2024, 14(20), 9410; https://doi.org/10.3390/app14209410 - 15 Oct 2024
Viewed by 279
Abstract
Power safety named entity recognition (NER) is essential for determining the cause of faults, assessing potential risks, and planning maintenance schedules, contributing to the comprehension and analysis of power safety documentation content and structure. Such analysis is crucial for the development of a [...] Read more.
Power safety named entity recognition (NER) is essential for determining the cause of faults, assessing potential risks, and planning maintenance schedules, contributing to the comprehension and analysis of power safety documentation content and structure. Such analysis is crucial for the development of a knowledge graph within the power safety domain and the augmentation of the associated dataset. This paper introduces a power safety NER model using efficiently learning an encoder that classifies token replacements accurately (ELECTRA) model. This model employs root mean square layer normalization (RMSNorm) and the switched gated linear unit (SwiGLU) activation function, which substitutes the conventional layer normalization (LayerNorm) and the Gaussian error linear units (GeLU). This model also integrates bidirectional long short-term memory (BiLSTM) with conditional random fields (CRF) to bolster performance in NER tasks. Experimental results show that the improved ELECTRA model achieved an F1 value of 93% on the constructed power safety NER dataset. It outperforms the BERT-BiLSTM-CRF model, achieving a 3.3% performance improvement. Full article
Show Figures

Figure 1

Figure 1
<p>RS-ELECTRA-BiLSTM-CRF model structure diagram.</p>
Full article ">Figure 2
<p>ELECTRA pre-training process.</p>
Full article ">Figure 3
<p>Long and short memory network structure diagram.</p>
Full article ">Figure 4
<p>Performance comparison for each entity type of different models.</p>
Full article ">
19 pages, 1774 KiB  
Article
A Novel Approach to Predict the Asian Exchange Stock Market Index Using Artificial Intelligence
by Rohit Salgotra, Harmanjeet Singh, Gurpreet Kaur, Supreet Singh, Pratap Singh and Szymon Lukasik
Algorithms 2024, 17(10), 457; https://doi.org/10.3390/a17100457 (registering DOI) - 15 Oct 2024
Viewed by 301
Abstract
This study uses real-world illustrations to explore the application of deep learning approaches to predict economic information. In this, we investigate the effect of deep learning model architecture and time-series data properties on prediction accuracy. We aim to evaluate the predictive power of [...] Read more.
This study uses real-world illustrations to explore the application of deep learning approaches to predict economic information. In this, we investigate the effect of deep learning model architecture and time-series data properties on prediction accuracy. We aim to evaluate the predictive power of several neural network models using a financial time-series dataset. These models include Convolutional RNNs, Convolutional LSTMs, Convolutional GRUs, Convolutional Bi-directional RNNs, Convolutional Bi-directional LSTMs, and Convolutional Bi-directional GRUs. Our main objective is to utilize deep learning techniques for simultaneous predictions on multivariable time-series datasets. We utilize the daily fluctuations of six Asian stock market indices from 1 April 2020 to 31 March 2024. This study’s overarching goal is to evaluate deep learning models constructed using training data gathered during the early stages of the COVID-19 pandemic when the economy was hit hard. We find that the limitations prove that no single deep learning algorithm can reliably forecast financial data for every state. In addition, predictions obtained from solitary deep learning models are more precise when dealing with consistent time-series data. Nevertheless, the hybrid model performs better when analyzing time-series data with significant chaos. Full article
(This article belongs to the Special Issue Nature-Inspired Algorithms in Machine Learning (2nd Edition))
Show Figures

Figure 1

Figure 1
<p>Architecture of Conv. RNN, Conv. LSTM, Conv. GRU, Conv. BRNN, Conv. BLSTM, and Conv. BGRU.</p>
Full article ">Figure 2
<p>Proposed system architecture.</p>
Full article ">Figure 3
<p>The BSE index from the first of April 2020 to the last day of March 2024.</p>
Full article ">Figure 4
<p>The HSI index from the first of April 2020 to the last day of March 2024.</p>
Full article ">Figure 5
<p>The Taiwan index from the first of April 2020 to the last day of March 2024.</p>
Full article ">Figure 6
<p>The JKSE index from the first of April 2020 to the last day of March 2024.</p>
Full article ">Figure 7
<p>The JPXGY index from the first of April 2020 to the last day of March 2024.</p>
Full article ">Figure 8
<p>The KS11 index from the first of April 2020 to the last day of March 2024.</p>
Full article ">Figure 9
<p>Training and value loss of all six Asian indices with BGRU model training. (<b>a</b>) Training and value loss for BSE index with BGRU Model; (<b>b</b>) training and value loss for HSI index with BGRU Model; (<b>c</b>) training and value loss for TWII index with BGRU Model; (<b>d</b>) training and value loss for JKSE index with BGRU Model; (<b>e</b>) training and value loss for JPXGY index with BGRU Model; (<b>f</b>) training and value loss for KS11 index with BGRU Model.</p>
Full article ">Figure 10
<p>Actual and prediction values of all six Asian indices with BGRU model. (<b>a</b>) Actual and prediction values for BSE index with BGRU model; (<b>b</b>) actual and prediction values for HSI index with BGRU model; (<b>c</b>) actual and prediction values for TWII index with BGRU model; (<b>d</b>) actual and prediction values for JKSE index with BGRU model; (<b>e</b>) actual and prediction values for JPXGY index with BGRU model; (<b>f</b>) actual and prediction values for KS11 index with BGRU model.</p>
Full article ">Figure 11
<p>Training and Value Loss for BSE index with BGRU, Conv. LSTM, Conv. GRU, Conv. RNN, Conv. BLSTM, and Conv. BGRU models. (<b>a</b>) Training and value loss for BSE index with Conv. RNN; (<b>b</b>) training and value loss for BSE index with Conv. LSTM; (<b>c</b>) training and value loss for BSE index with Conv. GRU; (<b>d</b>) training and value loss for BSE index with Conv. BRNN; (<b>e</b>) training and value loss for BSE index with Conv. BLSTM; (<b>f</b>) training and value loss for BSE index with Conv. BGRU.</p>
Full article ">Figure 12
<p>Actual and prediction values for BSE index with Conv. RNN, Conv. LSTM, Conv. BRNN, Conv. GRU, Conv. BLSTM and Conv. BGRU models. (<b>a</b>) Actual and prediction values for BSE index with Conv. RNN model; (<b>b</b>) actual and prediction values for BSE index with Conv. LSTM model; (<b>c</b>) actual and prediction values for BSE index with Conv. BRNN model; (<b>d</b>) actual and prediction values for BSE index with Conv. RNN; (<b>e</b>) actual and prediction values for BSE index with Conv. BLSTM; (<b>f</b>) actual and prediction values for BSE index with Conv. BGRU.</p>
Full article ">
24 pages, 2131 KiB  
Article
Improving Text Classification in Agricultural Expert Systems with a Bidirectional Encoder Recurrent Convolutional Neural Network
by Xiaojuan Guo, Jianping Wang, Guohong Gao, Li Li, Junming Zhou and Yancui Li
Electronics 2024, 13(20), 4054; https://doi.org/10.3390/electronics13204054 (registering DOI) - 15 Oct 2024
Viewed by 372
Abstract
With the rapid development of internet and AI technologies, Agricultural Expert Systems (AESs) have become crucial for delivering technical support and decision-making in agricultural management. However, traditional natural language processing methods often struggle with specialized terminology and context, and they lack the adaptability [...] Read more.
With the rapid development of internet and AI technologies, Agricultural Expert Systems (AESs) have become crucial for delivering technical support and decision-making in agricultural management. However, traditional natural language processing methods often struggle with specialized terminology and context, and they lack the adaptability to handle complex text classifications. The diversity and evolving nature of agricultural texts make deep semantic understanding and integration of contextual knowledge especially challenging. To tackle these challenges, this paper introduces a Bidirectional Encoder Recurrent Convolutional Neural Network (AES-BERCNN) tailored for short-text classification in agricultural expert systems. We designed an Agricultural Text Encoder (ATE) with a six-layer transformer architecture to capture both preceding and following word information. A recursive convolutional neural network based on Gated Recurrent Units (GRUs) was also developed to merge contextual information and learn complex semantic features, which are then combined with the ATE output and refined through max-pooling to form the final feature representation. The AES-BERCNN model was tested on a self-constructed agricultural dataset, achieving an accuracy of 99.63% in text classification. Its generalization ability was further verified on the Tsinghua News dataset. Compared to other models such as TextCNN, DPCNN, BiLSTM, and BERT-based models, the AES-BERCNN shows clear advantages in agricultural text classification. This work provides precise and timely technical support for intelligent agricultural expert systems. Full article
Show Figures

Figure 1

Figure 1
<p>Preprocessing and construction dataset.</p>
Full article ">Figure 2
<p>Structure of AES-BERCNN.</p>
Full article ">Figure 3
<p>Structure of ATE.</p>
Full article ">Figure 4
<p>Structure of the agricultural text classifier.</p>
Full article ">Figure 5
<p>The accuracy, loss, and time of comparative models on agricultural question training set: (<b>a</b>) accuracy comparison, (<b>b</b>) loss comparison, and (<b>c</b>) time comparison.</p>
Full article ">Figure 6
<p>Comparative model experimental results of precision, recall, and F1 on agricultural question test dataset: (<b>a</b>) precision comparison, (<b>b</b>) recall comparison, and (<b>c</b>) F1 comparison.</p>
Full article ">Figure 7
<p>Confusion matrix of classification effect of each model: (<b>a</b>) BiLSTM, (<b>b</b>) BiGRU, (<b>c</b>) TextCNN, (<b>d</b>) DPCNN, (<b>e</b>) BERT-TextCNN, (<b>f</b>) BERT-DPCNN, (<b>g</b>) BERT-BiLSTM, (<b>h</b>) BERT-BiGRU, (<b>i</b>) ATE-DPCNN, (<b>j</b>) ATE-TextCNN, (<b>k</b>) ATE-BiLSTM, (<b>l</b>) ATE-BiGRU, and (<b>m</b>) ATE-BERCNN.</p>
Full article ">Figure 7 Cont.
<p>Confusion matrix of classification effect of each model: (<b>a</b>) BiLSTM, (<b>b</b>) BiGRU, (<b>c</b>) TextCNN, (<b>d</b>) DPCNN, (<b>e</b>) BERT-TextCNN, (<b>f</b>) BERT-DPCNN, (<b>g</b>) BERT-BiLSTM, (<b>h</b>) BERT-BiGRU, (<b>i</b>) ATE-DPCNN, (<b>j</b>) ATE-TextCNN, (<b>k</b>) ATE-BiLSTM, (<b>l</b>) ATE-BiGRU, and (<b>m</b>) ATE-BERCNN.</p>
Full article ">Figure 8
<p>The accuracy and loss of comparative models on Tsinghua training set: (<b>a</b>) accuracy comparison and (<b>b</b>) loss comparison.</p>
Full article ">
21 pages, 5946 KiB  
Article
A Method Based on Deep Learning for Severe Convective Weather Forecast: CNN-BiLSTM-AM (Version 1.0)
by Jianbin Zhang, Meng Yin, Pu Wang and Zhiqiu Gao
Atmosphere 2024, 15(10), 1229; https://doi.org/10.3390/atmos15101229 (registering DOI) - 15 Oct 2024
Viewed by 373
Abstract
In this study, we propose a model called CNN-BiLSTM-AM that utilizes deep learning techniques to forecast severe convective weather based on ERA5 hourly data and observations. The model integrates a CNN with a Bidirectional Long Short-Term Memory (BiLSTM) system and an Attention Mechanism [...] Read more.
In this study, we propose a model called CNN-BiLSTM-AM that utilizes deep learning techniques to forecast severe convective weather based on ERA5 hourly data and observations. The model integrates a CNN with a Bidirectional Long Short-Term Memory (BiLSTM) system and an Attention Mechanism (AM). The CNN is tasked with extracting features from the input data, while the BiLSTM effectively captures temporal dependencies. The AM enhances the results by considering the impact of past feature states on severe weather phenomena. Additionally, we assess the performance of our model in comparison to traditional network architectures, including ConvLSTM, Predrnn++, CNN, FC-LSTM, and LSTM. Our results indicate that the CNN-BiLSTM-AM model exhibits superior accuracy in precipitation forecasting. Especially with the extension of the forecast time, the model performs well across multiple evaluation metrics. Furthermore, an interpretability analysis of the convective weather mechanisms utilizing machine learning highlights the critical role of total precipitable water (PWAT) in short-term heavy precipitation forecasts. It also emphasizes the significant impact of regional variables on convective weather patterns and the role of convective available potential energy (CAPE) in fostering conditions conducive to convection. These findings not only confirm the effectiveness of deep learning in the automatic identification of severe weather features but also validate the suitability of the sample dataset employed. Given its remarkable performance and robustness, we advocate for the adoption of this model to enhance the forecast of severe convective weather across various business applications. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence in Atmospheric Sciences)
Show Figures

Figure 1

Figure 1
<p>The flowchart of the proposed model.</p>
Full article ">Figure 2
<p>The research area chosen for the study (the white box represents Henan); shaded areas represent topographical features distribution (unit: m).</p>
Full article ">Figure 3
<p>Scattered density plots of various DL models on the testing dataset (<b>a</b>): FC-LSTM, (<b>b</b>): ConvLSTM, (<b>c</b>): Predrnn++, (<b>d</b>): CNN, (<b>e</b>): CNN-BiLSTM-AM, and (<b>f</b>): LSTM).</p>
Full article ">Figure 4
<p>Taylor plots from six models.</p>
Full article ">Figure 5
<p>RMSE distribution of FC-LSTM (<b>a</b>), ConvLSTM (<b>b</b>), Predrnn++ (<b>c</b>), CNN (<b>d</b>), CNN-BiLSTM-AM (<b>e</b>), and LSTM (<b>f</b>) in Henan on 22 July 2022.</p>
Full article ">Figure 6
<p>TS score for convective weather event on 22 July 2022 (the red line is the CNN-BiLSTM-AM model score, and the green line is the WRF score).</p>
Full article ">Figure 7
<p>Assessment of precipitation histograms predicted by six models across various months: (<b>a</b>) denotes RMSE, (<b>b</b>) denotes FAR, (<b>c</b>) denotes MAR, (<b>d</b>) denotes POD, (<b>e</b>) denotes TS, and (<b>f</b>) denotes ACCURACY.</p>
Full article ">Figure 8
<p>Feature importance (<b>a</b>) and correlation coefficients (<b>b</b>) of each forecast factor.</p>
Full article ">
16 pages, 3230 KiB  
Article
Research on Predictive Analysis Method of Building Energy Consumption Based on TCN-BiGru-Attention
by Sijia Fu, Rui Zhu and Feiyang Yu
Appl. Sci. 2024, 14(20), 9373; https://doi.org/10.3390/app14209373 - 14 Oct 2024
Viewed by 324
Abstract
Building energy consumption prediction has always played a significant role in assessing building energy efficiency, building commissioning, and detecting and diagnosing building system faults. With the progress of society and economic development, building energy consumption is growing rapidly. Therefore, accurate and effective building [...] Read more.
Building energy consumption prediction has always played a significant role in assessing building energy efficiency, building commissioning, and detecting and diagnosing building system faults. With the progress of society and economic development, building energy consumption is growing rapidly. Therefore, accurate and effective building energy consumption prediction is the basis of energy conservation. Although there are currently a large number of energy consumption research methods, each method has different applicability and advantages and disadvantages. This study proposes a Time Convolution Network model based on an attention mechanism, which combines the ability of the Time Convolution Network model to capture ultra-long time series information with the ability of the BiGRU model to integrate contextual information, improve model parallelism, and reduce the risk of overfitting. In order to tune the hyperparameters in the structure of this prediction model, such as the learning rate, the size of the convolutional kernel, and the number of recurrent units, this study chooses to use the Golden Jackal Optimization Algorithm for optimization. The study shows that this optimized model has better accuracy than models such as LSTM, SVM, and CNN. Full article
(This article belongs to the Special Issue Air Quality Prediction Based on Machine Learning Algorithms II)
Show Figures

Figure 1

Figure 1
<p>CO<sub>2</sub> emissions by fuel from 1950 to 2022. Data source: Ritchie, H., Rosado, P., &amp; Roser, M. (2020). CO<sub>2</sub> emissions by fuel. Our World in Data. Retrieved from <a href="https://ourworldindata.org/emissions-by-fuel" target="_blank">https://ourworldindata.org/emissions-by-fuel</a> (accessed on 9 Feburary 2022).</p>
Full article ">Figure 2
<p>The framework of TCN.</p>
Full article ">Figure 3
<p>The framework of GRU.</p>
Full article ">Figure 4
<p>The framework of BiGRU.</p>
Full article ">Figure 5
<p>Proposed Attention-BiGRU-Attention framework.</p>
Full article ">Figure 6
<p>Flow chart of Golden Jackal Optimization Algorithm.</p>
Full article ">Figure 7
<p>Evolutionary iteration curves of GJO and others with different test sets (CEC2022-F9).</p>
Full article ">Figure 8
<p>Evolutionary iteration curves of GJO and others with different test sets (CEC2022-F10).</p>
Full article ">Figure 9
<p>Evolutionary iteration curves of GJO and others with different test sets (CEC2022-F12).</p>
Full article ">Figure 10
<p>Comparison of prediction and actual energy consumption.</p>
Full article ">Figure 11
<p>Radar chart.</p>
Full article ">Figure 12
<p>Comparison of the MAE, MAPE, RMSE data before and after the optimization.</p>
Full article ">
24 pages, 13862 KiB  
Article
Depth Video-Based Secondary Action Recognition in Vehicles via Convolutional Neural Network and Bidirectional Long Short-Term Memory with Spatial Enhanced Attention Mechanism
by Weirong Shao, Mondher Bouazizi and Ohtuski Tomoaki
Sensors 2024, 24(20), 6604; https://doi.org/10.3390/s24206604 - 13 Oct 2024
Viewed by 544
Abstract
Secondary actions in vehicles are activities that drivers engage in while driving that are not directly related to the primary task of operating the vehicle. Secondary Action Recognition (SAR) in drivers is vital for enhancing road safety and minimizing accidents related to distracted [...] Read more.
Secondary actions in vehicles are activities that drivers engage in while driving that are not directly related to the primary task of operating the vehicle. Secondary Action Recognition (SAR) in drivers is vital for enhancing road safety and minimizing accidents related to distracted driving. It also plays an important part in modern car driving systems such as Advanced Driving Assistance Systems (ADASs), as it helps identify distractions and predict the driver’s intent. Traditional methods of action recognition in vehicles mostly rely on RGB videos, which can be significantly impacted by external conditions such as low light levels. In this research, we introduce a novel method for SAR. Our approach utilizes depth-video data obtained from a depth sensor located in a vehicle. Our methodology leverages the Convolutional Neural Network (CNN), which is enhanced by the Spatial Enhanced Attention Mechanism (SEAM) and combined with Bidirectional Long Short-Term Memory (Bi-LSTM) networks. This method significantly enhances action recognition ability in depth videos by improving both the spatial and temporal aspects. We conduct experiments using K-fold cross validation, and the experimental results show that on the public benchmark dataset Drive&Act, our proposed method shows significant improvement in SAR compared to the state-of-the-art methods, reaching an accuracy of about 84% in SAR in depth videos. Full article
(This article belongs to the Special Issue Multi-Sensor Systems for Object Tracking—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>The structure of CNN-SEAM+Bi-LSTM model.</p>
Full article ">Figure 2
<p>The overview of our experiment.</p>
Full article ">Figure 3
<p>Examples of data augmentation.</p>
Full article ">Figure 4
<p>Data distribution.</p>
Full article ">Figure 5
<p>Depth video before SR.</p>
Full article ">Figure 6
<p>Depth video after SR.</p>
Full article ">Figure 7
<p>The detail of SR: The first row is the local zoom image before SR and the second row is the local zoom image after SR.</p>
Full article ">Figure 8
<p>The structure of the SEAM.</p>
Full article ">Figure 9
<p>The structure of CNN-SEAM.</p>
Full article ">Figure 10
<p>The working process of ToF.</p>
Full article ">Figure 11
<p>Examples of video data of Drive&amp;Act: The first line consists of depth videos, the second line consists of RGB videos, and the third line consists of IR videos.</p>
Full article ">Figure 12
<p>Model performance across 5 folds.</p>
Full article ">Figure 13
<p>Cumulative confusion matrix for the proposed model of SAR.</p>
Full article ">Figure 14
<p>The predictions and prediction probabilities of the proposed model for SAR for some sample clips. The green font indicates the correct prediction of our proposed method.</p>
Full article ">Figure 15
<p>The left image is an RGB video clip in light condition and the right image is an RGB video clip in dark condition.</p>
Full article ">Figure 16
<p>Image (<b>A</b>) is a frame from the original videos, and Image (<b>B</b>) is the frame from the video after adding ‘camera shake’ and ‘motion blur’ effects.</p>
Full article ">
29 pages, 6269 KiB  
Article
Malware Detection Based on API Call Sequence Analysis: A Gated Recurrent Unit–Generative Adversarial Network Model Approach
by Nsikak Owoh, John Adejoh, Salaheddin Hosseinzadeh, Moses Ashawa, Jude Osamor and Ayyaz Qureshi
Future Internet 2024, 16(10), 369; https://doi.org/10.3390/fi16100369 - 13 Oct 2024
Viewed by 569
Abstract
Malware remains a major threat to computer systems, with a vast number of new samples being identified and documented regularly. Windows systems are particularly vulnerable to malicious programs like viruses, worms, and trojans. Dynamic analysis, which involves observing malware behavior during execution in [...] Read more.
Malware remains a major threat to computer systems, with a vast number of new samples being identified and documented regularly. Windows systems are particularly vulnerable to malicious programs like viruses, worms, and trojans. Dynamic analysis, which involves observing malware behavior during execution in a controlled environment, has emerged as a powerful technique for detection. This approach often focuses on analyzing Application Programming Interface (API) calls, which represent the interactions between the malware and the operating system. Recent advances in deep learning have shown promise in improving malware detection accuracy using API call sequence data. However, the potential of Generative Adversarial Networks (GANs) for this purpose remains largely unexplored. This paper proposes a novel hybrid deep learning model combining Gated Recurrent Units (GRUs) and GANs to enhance malware detection based on API call sequences from Windows portable executable files. We evaluate our GRU–GAN model against other approaches like Bidirectional Long Short-Term Memory (BiLSTM) and Bidirectional Gated Recurrent Unit (BiGRU) on multiple datasets. Results demonstrated the superior performance of our hybrid model, achieving 98.9% accuracy on the most challenging dataset. It outperformed existing models in resource utilization, with faster training and testing times and low memory usage. Full article
(This article belongs to the Special Issue Privacy and Security in Computing Continuum and Data-Driven Workflows)
Show Figures

Figure 1

Figure 1
<p>Flowchart of the proposed method.</p>
Full article ">Figure 2
<p>The architecture of the proposed GRU–GAN model.</p>
Full article ">Figure 3
<p>Accuracy and validation loss of the BiLSTM and BiGRU models on dataset 1.</p>
Full article ">Figure 4
<p>Discriminator accuracy and validation loss of the GRU–GAN model on dataset 1.</p>
Full article ">Figure 5
<p>Accuracy and validation loss of the BiLSTM and BiGRU models on dataset 2.</p>
Full article ">Figure 6
<p>Discriminator accuracy and validation loss of the GRU–GAN model on dataset 2.</p>
Full article ">Figure 7
<p>Confusion matrix of the BiLSTM, BiGRU, and GRU–GAN models on datasets 1, 2 and 3.</p>
Full article ">Figure 8
<p>Evaluation metrics results of the BiLSTM, BiGRU, and GRU–GAN models on the three datasets.</p>
Full article ">Figure 9
<p>ROC curve results of the BiLSTM, BiGRU, and GRU–GAN models on the three datasets.</p>
Full article ">Figure 10
<p>Prediction performance of the BiLSTM, BiGRU, and GRU–GAN models on dataset 1.</p>
Full article ">Figure 11
<p>Prediction performance of the BiLSTM, BiGRU, and GRU–GAN models on dataset 2.</p>
Full article ">Figure 12
<p>Prediction performance of the BiLSTM, BiGRU, and GRU–GAN models on dataset 3.</p>
Full article ">Figure 13
<p>Computation resources used by the BiLSTM, BiGRU, and GRU–GAN models.</p>
Full article ">
23 pages, 7794 KiB  
Article
Efficient Method for Photovoltaic Power Generation Forecasting Based on State Space Modeling and BiTCN
by Guowei Dai, Shuai Luo, Hu Chen and Yulong Ji
Sensors 2024, 24(20), 6590; https://doi.org/10.3390/s24206590 - 13 Oct 2024
Viewed by 568
Abstract
As global carbon reduction initiatives progress and the new energy sector rapidly develops, photovoltaic (PV) power generation is playing an increasingly significant role in renewable energy. Accurate PV output forecasting, influenced by meteorological factors, is essential for efficient energy management. This paper presents [...] Read more.
As global carbon reduction initiatives progress and the new energy sector rapidly develops, photovoltaic (PV) power generation is playing an increasingly significant role in renewable energy. Accurate PV output forecasting, influenced by meteorological factors, is essential for efficient energy management. This paper presents an optimal hybrid forecasting strategy, integrating bidirectional temporal convolutional networks (BiTCN), dynamic convolution (DC), bidirectional long short-term memory networks (BiLSTM), and a novel mixed-state space model (Mixed-SSM). The mixed-SSM combines the state space model (SSM), multilayer perceptron (MLP), and multi-head self-attention mechanism (MHSA) to capture complementary temporal, nonlinear, and long-term features. Pearson and Spearman correlation analyses are used to select features strongly correlated with PV output, improving the prediction correlation coefficient (R2) by at least 0.87%. The K-Means++ algorithm further enhances input data features, achieving a maximum R2 of 86.9% and a positive R2 gain of 6.62%. Compared with BiTCN variants such as BiTCN-BiGRU, BiTCN-transformer, and BiTCN-LSTM, the proposed method delivers a mean absolute error (MAE) of 1.1%, root mean squared error (RMSE) of 1.2%, and an R2 of 89.1%. These results demonstrate the model’s effectiveness in forecasting PV power and supporting low-carbon, safe grid operation. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>Detailed diagram of the proposed BiTCN-MixedSSM model structure.</p>
Full article ">Figure 2
<p>Diagrammatic representation of the structure of the proposed novel Mixed-SSM module.</p>
Full article ">Figure 3
<p>Diagrammatic representation of the detailed structure of the proposed Mixed-SSM module with (<b>a</b>) state space model, (<b>b</b>) multilayer perceptron, and (<b>c</b>) multi-attention mechanism.</p>
Full article ">Figure 4
<p>Graphical representation of Pearson correlation analysis of PV power forecast data.</p>
Full article ">Figure 5
<p>Graphical representation of Spearman’s coefficient correlation analysis of PV power forecast data.</p>
Full article ">Figure 6
<p>The dataset of PV power generation in relation to meteorological parameters presents various curves, with the y-axis indicating the range of values corresponding to each curve. These curves include: (<b>a</b>) the temperature versus humidity relationship, (<b>b</b>) the dew point temperature plotted against wind speed, (<b>c</b>) maximum wind speed in relation to wind direction, and (<b>d</b>) precipitation versus barometric pressure for a specific day analyzed in the study.</p>
Full article ">Figure 7
<p>Graphical representation of the unstandardized PV forecast dataset.</p>
Full article ">Figure 8
<p>Graphical representation of a standardized PV power forecasting dataset.</p>
Full article ">Figure 9
<p>Comparative graphical representation of actual and predicted PV generation for BiTCN variants.</p>
Full article ">Figure 10
<p>Graphical representation of the raw dataset for PV power prediction with (<b>a</b>) wind speed data, (<b>b</b>) maximum wind speed data, and (<b>c</b>) wind direction data.</p>
Full article ">Figure 11
<p>Graphical representation of the data after fusion feature processing of the raw dataset for PV power generation prediction.</p>
Full article ">
27 pages, 920 KiB  
Article
AI-Generated Spam Review Detection Framework with Deep Learning Algorithms and Natural Language Processing
by Mudasir Ahmad Wani, Mohammed ElAffendi and Kashish Ara Shakil
Computers 2024, 13(10), 264; https://doi.org/10.3390/computers13100264 - 12 Oct 2024
Viewed by 365
Abstract
Spam reviews pose a significant challenge to the integrity of online platforms, misleading consumers and undermining the credibility of genuine feedback. This paper introduces an innovative AI-generated spam review detection framework that leverages Deep Learning algorithms and Natural Language Processing (NLP) techniques to [...] Read more.
Spam reviews pose a significant challenge to the integrity of online platforms, misleading consumers and undermining the credibility of genuine feedback. This paper introduces an innovative AI-generated spam review detection framework that leverages Deep Learning algorithms and Natural Language Processing (NLP) techniques to identify and mitigate spam reviews effectively. Our framework utilizes multiple Deep Learning models, including Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, Gated Recurrent Unit (GRU), and Bidirectional LSTM (BiLSTM), to capture intricate patterns in textual data. The system processes and analyzes large volumes of review content to detect deceptive patterns by utilizing advanced NLP and text embedding techniques such as One-Hot Encoding, Word2Vec, and Term Frequency-Inverse Document Frequency (TF-IDF). By combining three embedding techniques with four Deep Learning algorithms, a total of twelve exhaustive experiments were conducted to detect AI-generated spam reviews. The experimental results demonstrate that our approach outperforms the traditional machine learning models, offering a robust solution for ensuring the authenticity of online reviews. Among the models evaluated, those employing Word2Vec embeddings, particularly the BiLSTM_Word2Vec model, exhibited the strongest performance. The BiLSTM model with Word2Vec achieved the highest performance, with an exceptional accuracy of 98.46%, a precision of 0.98, a recall of 0.97, and an F1-score of 0.98, reflecting a near-perfect balance between precision and recall. Its high F2-score (0.9810) and F0.5-score (0.9857) further highlight its effectiveness in accurately detecting AI-generated spam while minimizing false positives, making it the most reliable option for this task. Similarly, the Word2Vec-based LSTM model also performed exceptionally well, with an accuracy of 97.58%, a precision of 0.97, a recall of 0.96, and an F1-score of 0.97. The CNN model with Word2Vec similarly delivered strong results, achieving an accuracy of 97.61%, a precision of 0.97, a recall of 0.96, and an F1-score of 0.97. This study is unique in its focus on detecting spam reviews specifically generated by AI-based tools rather than solely detecting spam reviews or AI-generated text. This research contributes to the field of spam detection by offering a scalable, efficient, and accurate framework that can be integrated into various online platforms, enhancing user trust and the decision-making processes. Full article
Show Figures

Figure 1

Figure 1
<p>Detailed data collection procedure.</p>
Full article ">Figure 2
<p>Generating AI-based spam/fake reviews based on human-authored samples.</p>
Full article ">Figure 3
<p>Check for the working of GPT Module.</p>
Full article ">Figure 4
<p>Data preparation and preprocessing with NLTK toolkit.</p>
Full article ">Figure 5
<p>Experimental setup and configuration.</p>
Full article ">Figure 6
<p>Performance of selected Deep Learning models on TF-IDF representation.</p>
Full article ">Figure 7
<p>Performance of selected Deep Learning models on Word2Vec feature representation.</p>
Full article ">Figure 8
<p>Performance of selected Deep Learning models on One-Hot Encoding.</p>
Full article ">Figure 9
<p>The radar plot showing proposed approaches. Particularly, Word2Vec-based BiLSTM outperformed the existing methods.</p>
Full article ">Figure 10
<p>Heptagon: seven ways to prevent abuse and ensure ethical use of AI-generated reviews.</p>
Full article ">
21 pages, 4502 KiB  
Article
An Analytical Approach for IGBT Life Prediction Using Successive Variational Mode Decomposition and Bidirectional Long Short-Term Memory Networks
by Kaitian Deng, Xianglian Xu, Fang Yuan, Tianyu Zhang, Yuli Xu, Tunzhen Xie, Yuanqing Song and Ruiqing Zhao
Electronics 2024, 13(20), 4002; https://doi.org/10.3390/electronics13204002 - 11 Oct 2024
Viewed by 403
Abstract
The precise estimation of the operational lifespan of insulated gate bipolar transistors (IGBT) holds paramount significance for ensuring the efficient and uncompromised safety of industrial equipment. However, numerous methodologies and models currently employed for this purpose often fall short of delivering highly accurate [...] Read more.
The precise estimation of the operational lifespan of insulated gate bipolar transistors (IGBT) holds paramount significance for ensuring the efficient and uncompromised safety of industrial equipment. However, numerous methodologies and models currently employed for this purpose often fall short of delivering highly accurate predictions. The analytical approach that combines the Pattern Optimization Algorithm (POA) with Successive Variational Mode Decomposition (SVMD) and Bidirectional Long Short-term Memory (BiLSTM) network is introduced. Firstly, SVMD is employed as an unsupervised feature learning method to partition the data into intrinsic modal functions (IMFs), which are used to eliminate noise and preserve the essential signal. Secondly, the BiLSTM network is integrated for supervised learning purposes, enabling the prediction of the decomposed sequence. Additionally, the hyperparameters of BiLSTM and the penalty coefficients of SVMD are optimized utilizing the POA technique. Subsequently, the various modal functions are predicted utilizing the trained prediction model, and the individual mode predictions are subsequently aggregated to yield the model’s definitive final life prediction. Through case studies involving IGBT aging datasets, the optimal prediction model was formulated and its lifespan prediction capability was validated. The superiority of the proposed method is demonstrated by comparing it with benchmark models and other state-of-the-art methods. Full article
Show Figures

Figure 1

Figure 1
<p>Structure diagram of LSTM.</p>
Full article ">Figure 2
<p>BiLSTM network structure diagram.</p>
Full article ">Figure 3
<p>Optimization process of POA.</p>
Full article ">Figure 4
<p>Diagram of the process of processing time series data.</p>
Full article ">Figure 5
<p>The flowchart of POA-SVMD-POA-BiLSTM IGBT life prediction.</p>
Full article ">Figure 6
<p>Fitness curve during decomposition: (<b>a</b>) fitness curve during optimization with SVMD; (<b>b</b>) fitness curve during optimization with POA-SVMD.</p>
Full article ">Figure 7
<p>The modal component obtained by decomposing the detected signal.</p>
Full article ">Figure 8
<p>RMSE, MAE, and MAPE for different models. (A: POA-SVMD-POA-BiLSTM; B: POA-BiLSTM; C: POA-LMD-POA-BiLSTM; D: POA-EEMD-POA-BiLSTM).</p>
Full article ">Figure 9
<p>Box diagram of forecasting error absolute values (A: POA-EEMD-POA-BiLSTM; B: POA-LMD-POA-BiLSTM; C: POA-SVMD-POA-BiLSTM; D: POA-BiLSTM).</p>
Full article ">Figure 10
<p>The life prediction results of each modal component: (<b>a</b>) predicted values of the components of the POA-SVMD-BiLSTM model; (<b>b</b>) predicted values of the components of the POA-SVMD-POA-BiLSTM model.</p>
Full article ">Figure 11
<p>Predictions compared to different models.</p>
Full article ">
Back to TopTop