[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,658)

Search Parameters:
Keywords = DNN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
0 pages, 680 KiB  
Article
Score-Based Black-Box Adversarial Attack on Time Series Using Simulated Annealing Classification and Post-Processing Based Defense
by Sichen Liu and Yuan Luo
Electronics 2024, 13(3), 650; https://doi.org/10.3390/electronics13030650 - 4 Feb 2024
Cited by 1 | Viewed by 959
Abstract
While deep neural networks (DNNs) have been widely and successfully used for time series classification (TSC) over the past decade, their vulnerability to adversarial attacks has received little attention. Most existing attack methods focus on white-box setups, which are unrealistic as attackers typically [...] Read more.
While deep neural networks (DNNs) have been widely and successfully used for time series classification (TSC) over the past decade, their vulnerability to adversarial attacks has received little attention. Most existing attack methods focus on white-box setups, which are unrealistic as attackers typically only have access to the model’s probability outputs. Defensive methods also have limitations, relying primarily on adversarial retraining which degrades classification accuracy and requires excessive training time. On top of that, we propose two new approaches in this paper: (1) A simulated annealing-based random search attack that finds adversarial examples without gradient estimation, searching only on the l-norm hypersphere of allowable perturbations. (2) A post-processing defense technique that periodically reverses the trend of corresponding loss values while maintaining the overall trend, using only the classifier’s confidence scores as input. Experiments applying these methods to InceptionNet models trained on the UCR dataset benchmarks demonstrate the effectiveness of the attack, achieving up to 100% success rates. The defense method provided protection against up to 91.24% of attacks while preserving prediction quality. Overall, this work addresses important gaps in adversarial TSC by introducing novel black-box attack and lightweight defense techniques. Full article
Show Figures

Figure 1

Figure 1
<p>Adversarial attack and corresponding defense on TSC.</p>
Full article ">Figure 2
<p>The original margin loss <math display="inline"> <semantics> <msub> <mi>l</mi> <mi>s</mi> </msub> </semantics> </math> and constructed margin loss <math display="inline"> <semantics> <msub> <mi>l</mi> <mi>d</mi> </msub> </semantics> </math> when <math display="inline"> <semantics> <mrow> <mi>t</mi> <mo>=</mo> <mn>0.2</mn><mo>,</mo> <mspace width="4pt"/> <mi>α</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics> </math>. The trends of <math display="inline"> <semantics> <msub> <mi>l</mi> <mi>s</mi> </msub> </semantics> </math> and <math display="inline"> <semantics> <msub> <mi>l</mi> <mi>d</mi> </msub> </semantics> </math> are the same globally, but they are opposite locally.</p>
Full article ">Figure 3
<p>ASR curve of different <math display="inline"> <semantics> <mi>ϵ</mi> </semantics> </math> values. Each figure shows the result of one dataset. (<b>a</b>) UWaveGestureLibraryAll; (<b>b</b>) OSU Leaf; (<b>c</b>) ECG5000; (<b>d</b>) ChlorineConcentration.</p>
Full article ">Figure 4
<p>AQT curve of different <math display="inline"> <semantics> <mi>ϵ</mi> </semantics> </math> values. Each figure shows the result of one dataset. (<b>a</b>) UWaveGestureLibraryAll; (<b>b</b>) OSU Leaf; (<b>c</b>) ECG5000; (<b>d</b>) ChlorineConcentration.</p>
Full article ">Figure 5
<p>DSR curve of different <math display="inline"> <semantics> <mi>ϵ</mi> </semantics> </math> values in four datasets: UWaveGestureLibraryAll, OSU Leaf, ECG5000 and ChlorineConcentration.</p>
Full article ">
14 pages, 7528 KiB  
Article
Optimal Power Allocation in Optical GEO Satellite Downlinks Using Model-Free Deep Learning Algorithms
by Theodore T. Kapsis, Nikolaos K. Lyras and Athanasios D. Panagopoulos
Electronics 2024, 13(3), 647; https://doi.org/10.3390/electronics13030647 - 4 Feb 2024
Viewed by 906
Abstract
Geostationary (GEO) satellites are employed in optical frequencies for a variety of satellite services providing wide coverage and connectivity. Multi-beam GEO high-throughput satellites offer Gbps broadband rates and, jointly with low-Earth-orbit mega-constellations, are anticipated to enable a large-scale free-space optical (FSO) network. In [...] Read more.
Geostationary (GEO) satellites are employed in optical frequencies for a variety of satellite services providing wide coverage and connectivity. Multi-beam GEO high-throughput satellites offer Gbps broadband rates and, jointly with low-Earth-orbit mega-constellations, are anticipated to enable a large-scale free-space optical (FSO) network. In this paper, a power allocation methodology based on deep reinforcement learning (DRL) is proposed for optical satellite systems disregarding any channel statistics knowledge requirements. An all-FSO, multi-aperture GEO-to-ground system is considered and an ergodic capacity optimization problem for the downlink is formulated with transmitted power constraints. A power allocation algorithm was developed, aided by a deep neural network (DNN) which is fed channel state information (CSI) observations and trained in a parameterized on-policy manner through a stochastic policy gradient approach. The proposed method does not require the channels’ transition models or fading distributions. To validate and test the proposed allocation scheme, experimental measurements from the European Space Agency’s ARTEMIS optical satellite campaign were utilized. It is demonstrated that the predicted average capacity greatly exceeds other baseline heuristic algorithms while strongly converging to the supervised, unparameterized approach. The predicted average channel powers differ only by 0.1 W from the reference ones, while the baselines differ significantly more, about 0.1–0.5 W. Full article
(This article belongs to the Special Issue New Advances of Microwave and Optical Communication)
Show Figures

Figure 1

Figure 1
<p>Normalized PDFs of experimental and synthesized data. (<b>a</b>) 10 September 2003 20:10–20:30. (<b>b</b>) 10 September 2003 00:30–00:50. (<b>c</b>) 12 September 2003 00:30–00:50. (<b>d</b>) 13 September 2003 23:30–23:50. (<b>e</b>) 16 September 2003 20:10–20:30.</p>
Full article ">Figure 2
<p>The loss function and predicted average capacity in terms of the (<b>a</b>) number of hidden layers (200 nodes/layer) and (<b>b</b>) number of neurons (2 hidden layers). <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>3.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 3
<p>Performance results for five optical satellite downlinks using three different policy distributions: truncated Normal <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>μ</mi> <mo>,</mo> <mi>σ</mi> <mo>,</mo> <mn>0</mn> <mo>,</mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>)</mo> </mrow> </semantics></math>, truncated Weibull <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>k</mi> <mo>,</mo> <mi>λ</mi> <mo>,</mo> <mn>0</mn> <mo>,</mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>)</mo> </mrow> </semantics></math>, and truncated Exponential <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>λ</mi> <mo>,</mo> <mn>0</mn> <mo>,</mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>)</mo> </mrow> </semantics></math>. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>2.5</mn> <mo> </mo> <mi mathvariant="normal">W</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 4
<p>Learning episodes of five different PA algorithms for five optical channels: (<b>a</b>) the average system capacity and (<b>b</b>) the constraint function. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>2.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 5
<p>The average channel powers of five different PA algorithms for five optical channels during 5000 episodes. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>2.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 6
<p>Learning episodes of five different PA algorithms for five optical channels: (<b>a</b>) the average system capacity and (<b>b</b>) the constraint function. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>3.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 7
<p>The average channel powers of five different PA algorithms for five optical channels during 5000 episodes. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>3.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 8
<p>Learning episodes of five different PA algorithms for five optical channels: (<b>a</b>) the average system capacity and (<b>b</b>) the constraint function. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>4.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p>The average channel powers of five different PA algorithms for five optical channels during 5000 episodes. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>4.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p>The mean and standard deviation of the TN distributions for two allocated channel powers corresponding to ch.0 and ch.3.</p>
Full article ">Figure 11
<p>Loss function across the training and validation sets for two hyperparameter <span class="html-italic">λ</span> values related to the total power constraint function. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>s</mi> </mrow> </msub> <mo>=</mo> <mn>1.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>,</mo> <mo> </mo> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mi>a</mi> <mi>v</mi> </mrow> </msub> <mo>=</mo> <mn>3.0</mn> <mo> </mo> <mi mathvariant="normal">W</mi> <mo>.</mo> </mrow> </semantics></math> The relaxed <span class="html-italic">λ</span> = 0.2 allows the power constraint to be violated, resulting in higher loss values, while the stricter <span class="html-italic">λ</span> = 1.0 yields lower loss values. No overfitting is observed.</p>
Full article ">
9 pages, 4388 KiB  
Article
Diffractive Deep-Neural-Network-Based Classifier for Holographic Memory
by Toshihiro Sakurai, Tomoyoshi Ito and Tomoyoshi Shimobaba
Photonics 2024, 11(2), 145; https://doi.org/10.3390/photonics11020145 - 4 Feb 2024
Viewed by 1201
Abstract
Holographic memory offers high-capacity optical storage with rapid data readout and long-term durability. Recently, read data pages have been classified using digital deep neural networks (DNNs). This approach is highly accurate, but the prediction time hinders the data readout throughput. This study presents [...] Read more.
Holographic memory offers high-capacity optical storage with rapid data readout and long-term durability. Recently, read data pages have been classified using digital deep neural networks (DNNs). This approach is highly accurate, but the prediction time hinders the data readout throughput. This study presents a diffractive DNN (D2NN)-based classifier for holographic memory. D2NNs have so far attracted a great deal of attention for object identification and image transformation at the speed of light. A D2NN, consisting of trainable diffractive layers and devoid of electronic devices, facilitates high-speed data readout. Furthermore, we numerically investigated the classification performance of a D2NN-based classifier. The classification accuracy of the D2NN was 99.7% on 4-bit symbols, exceeding that of the hard decision method. Full article
(This article belongs to the Special Issue Holographic Information Processing)
Show Figures

Figure 1

Figure 1
<p>Holographic memory with a D2NN classifier.</p>
Full article ">Figure 2
<p>Examples of (<b>a</b>) a data page with a reference pattern, (<b>b</b>) a reconstructed intensity data page, and (<b>c</b>) a reconstructed phase data page.</p>
Full article ">Figure 3
<p>Assignment of 16 symbol areas (<b>a</b>), label image showing the 13th label as an example (<b>b</b>), and D2NN output intensity (<b>c</b>).</p>
Full article ">Figure 4
<p>Structure of the D2NN with equal distances for the diffractive layers (<b>a</b>), and training and validation curves for the accuracy (<b>b</b>) and loss value (<b>c</b>).</p>
Full article ">Figure 5
<p>Confusion matrices: (<b>a</b>) before modifying the label images, (<b>b</b>) after modifying the label images.</p>
Full article ">Figure 6
<p>Examples of label images. The upper row corresponds to labels 1, 2, 4, and 8. The lower row corresponds to labels 7, 11, 13, and 14. The Hamming distance of the label images in each row is one.</p>
Full article ">Figure 7
<p>Modified label images. The upper row corresponds to labels 1, 2, 4, and 8. The label 0 area was supplemented with 50% brightness. The bottom row corresponds to labels 7, 11, 13, and 14. The label 15 area was supplemented with 50% brightness.</p>
Full article ">Figure 8
<p>Structure of the D2NN with TPE-optimized distance for the diffractive layers (<b>a</b>), and training and validation curves for the accuracy (<b>b</b>) and loss value (<b>c</b>).</p>
Full article ">
15 pages, 20626 KiB  
Article
CoDC: Accurate Learning with Noisy Labels via Disagreement and Consistency
by Yongfeng Dong, Jiawei Li, Zhen Wang and Wenyu Jia
Biomimetics 2024, 9(2), 92; https://doi.org/10.3390/biomimetics9020092 - 3 Feb 2024
Cited by 2 | Viewed by 1464
Abstract
Inspired by the biological nervous system, deep neural networks (DNNs) are able to achieve remarkable performance in various tasks. However, they struggle to handle label noise, which can poison the memorization effects of DNNs. Co-teaching-based methods are popular in learning with noisy labels. [...] Read more.
Inspired by the biological nervous system, deep neural networks (DNNs) are able to achieve remarkable performance in various tasks. However, they struggle to handle label noise, which can poison the memorization effects of DNNs. Co-teaching-based methods are popular in learning with noisy labels. These methods cross-train two DNNs based on the small-loss criterion and employ a strategy using either “disagreement” or “consistency” to obtain the divergence of the two networks. However, these methods are sample-inefficient for generalization in noisy scenarios. In this paper, we propose CoDC, a novel Co-teaching-basedmethod for accurate learning with label noise via both Disagreement and Consistency strategies. Specifically, CoDC maintains disagreement at the feature level and consistency at the prediction level using a balanced loss function. Additionally, a weighted cross-entropy loss is proposed based on information derived from the historical training process. Moreover, the valuable knowledge involved in “large-loss” samples is further developed and utilized by assigning pseudo-labels. Comprehensive experiments were conducted on both synthetic and real-world noise and under various noise types. CoDC achieved 72.81% accuracy on the Clothing1M dataset and 76.96% (Top1) accuracy on the WebVision1.0 dataset. These superior results demonstrate the effectiveness and robustness of learning with noisy labels. Full article
(This article belongs to the Special Issue Bioinspired Artificial Intelligence Applications)
Show Figures

Figure 1

Figure 1
<p>Sample selection of CoDC. For each sample, the classification loss, disagreement loss, and consistency loss are calculated and the small-loss samples are taken as the clean sample. In general, the essence of this method is to maintain a balance of disagreement and consistency between the two networks to achieve the best performance.</p>
Full article ">Figure 2
<p>Label transition matrix.</p>
Full article ">Figure 3
<p>Construction of two long-tailed distributed datasets.</p>
Full article ">Figure 4
<p>Test accuracy comparison for the number of epochs on the four long-tailed noisy datasets. Each experiment was repeated three times. The error bar for the standard deviation in each figure has been shaded.</p>
Full article ">
21 pages, 2951 KiB  
Article
Enhancing Day-Ahead Cooling Load Prediction in Tropical Commercial Buildings Using Advanced Deep Learning Models: A Case Study in Singapore
by Namitha Kondath, Aung Myat, Yong Loke Soh, Whye Loon Tung, Khoo Aik Min Eugene and Hui An
Buildings 2024, 14(2), 397; https://doi.org/10.3390/buildings14020397 - 1 Feb 2024
Cited by 1 | Viewed by 857
Abstract
Commercial buildings in hot and humid tropical climates rely significantly on cooling systems to maintain optimal occupant comfort. A well-accurate day-ahead load profile prediction plays a pivotal role in planning the energy requirements of cooling systems. Despite the pressing need for effective day-ahead [...] Read more.
Commercial buildings in hot and humid tropical climates rely significantly on cooling systems to maintain optimal occupant comfort. A well-accurate day-ahead load profile prediction plays a pivotal role in planning the energy requirements of cooling systems. Despite the pressing need for effective day-ahead cooling load predictions, current methodologies have not fully harnessed the potential of advanced deep-learning techniques. This paper aims to address this gap by investigating the application of innovative deep-learning models in day-ahead hourly cooling load prediction for commercial buildings in tropical climates. A range of multi-output deep learning techniques, including Deep Neural Networks (DNNs), Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Long Short-Term Memory networks (LSTMs), are employed to enhance prediction accuracy. Furthermore, these individual deep learning techniques are synergistically integrated to create hybrid models, such as CNN-LSTM and Sequence-to-Sequence models. Experiments are conducted to choose the time horizons from the past that can serve as input to the models. In addition, the influence of various categories of input parameters on prediction performance has been assessed. Historical cooling load, calendar features, and outdoor weather parameters are found in decreasing order of influence on prediction accuracy. This research focuses on buildings located in Singapore and presents a comprehensive case study to validate the proposed models and methodologies. The sequence-to-sequence model provided better performance than all the other models. It offered a CV-RMSE of 7.4%, 10%, and 6% for SIT@Dover, SIT@NYP, and the simulated datasets, which were 2.3%, 3%, and 1% less, respectively, than the base Deep Neural Network model. Full article
(This article belongs to the Special Issue Advancements in Adaptive, Inclusive, and Responsive Buildings)
Show Figures

Figure 1

Figure 1
<p>The overall research methodology.</p>
Full article ">Figure 2
<p>A cross-correlation coefficient was calculated for (<b>a</b>) the SIT@Dover dataset, (<b>b</b>) the SIT@NYP dataset, and (<b>c</b>) the simulated dataset.</p>
Full article ">Figure 3
<p>The basic architecture of RNN.</p>
Full article ">Figure 4
<p>Structure of CNN-LSTM hybrid model.</p>
Full article ">Figure 5
<p>Sequence-to-sequence model architecture.</p>
Full article ">Figure 6
<p>Comparison of performance for two different selections of time horizons for (<b>a</b>) SIT@Dover, (<b>b</b>) SIT@NYP, and (<b>c</b>) the simulated datasets.</p>
Full article ">Figure 7
<p>Example of a 12 h cooling load prediction by different models. (<b>a</b>) Prediction by DNN, CNN, and RNN. (<b>b</b>) Prediction by LSTM and hybrid models.</p>
Full article ">Figure 8
<p>Distribution of error in load predictions.</p>
Full article ">Figure 9
<p>RMSE and MAPE when various categories of parameters are excluded from the input. (<b>a</b>) SIT@Dover dataset, (<b>b</b>) SIT@NYP dataset, and (<b>c</b>) Simulated dataset.</p>
Full article ">
23 pages, 21681 KiB  
Article
LDnADMM-Net: A Denoising Unfolded Deep Neural Network for Direction-of-Arrival Estimations in A Low Signal-to-Noise Ratio
by Can Liang, Mingxuan Liu, Yang Li, Yanhua Wang and Xueyao Hu
Remote Sens. 2024, 16(3), 554; https://doi.org/10.3390/rs16030554 - 31 Jan 2024
Cited by 1 | Viewed by 1055
Abstract
In this paper, we explore the problem of direction-of-arrival (DOA) estimation for a non-uniform linear array (NULA) under strong noise. The compressed sensing (CS)-based methods are widely used in NULA DOA estimations. However, these methods commonly rely on the tuning of parameters, which [...] Read more.
In this paper, we explore the problem of direction-of-arrival (DOA) estimation for a non-uniform linear array (NULA) under strong noise. The compressed sensing (CS)-based methods are widely used in NULA DOA estimations. However, these methods commonly rely on the tuning of parameters, which are hard to fine-tune. Additionally, these methods lack robustness under strong noise. To address these issues, this paper proposes a novel DOA estimation approach using a deep neural network (DNN) for a NULA in a low SNR. The proposed network is designed based on the denoising convolutional neural network (DnCNN) and the alternating direction method of multipliers (ADMM), which is dubbed as LDnADMM-Net. First, we construct an unfolded DNN architecture that mimics the behavior of the iterative processing of an ADMM. In this way, the parameters of an ADMM can be transformed into the network weights, and thus we can adaptively optimize these parameters through network training. Then, we employ the DnCNN to develop a denoising module (DnM) and integrate it into the unfolded DNN. Using this DnM, we can enhance the anti-noise ability of the proposed network and obtain a robust DOA estimation in a low SNR. The simulation and experimental results show that the proposed LDnADMM-Net can obtain high-accuracy and super-resolution DOA estimations for a NULA with strong robustness in a low signal-to-noise ratio (SNR). Full article
(This article belongs to the Special Issue State-of-the-Art and Future Developments: Short-Range Radar)
Show Figures

Figure 1

Figure 1
<p>The non-uniform linear array with <span class="html-italic">N</span> antennas for the DOA estimation.</p>
Full article ">Figure 2
<p>The structure of the LDnADMM-Net.</p>
Full article ">Figure 3
<p>The structure of the denoising module.</p>
Full article ">Figure 4
<p>The structure of the sparse representation layer.</p>
Full article ">Figure 5
<p>The structure of the non-linear transform layer.</p>
Full article ">Figure 6
<p>The structure of the multiplier update layer.</p>
Full article ">Figure 7
<p>The NULA antenna pattern with the beam pointing at 0°.</p>
Full article ">Figure 8
<p>The denoising performance versus the target SNR in the scenario of one target. (<b>a</b>) The mean square error; (<b>b</b>) the output SNR.</p>
Full article ">Figure 9
<p>The DOA estimation results versus the target SNR in the scenario of one target. (<b>a</b>) The success rate; (<b>b</b>) the root mean square error.</p>
Full article ">Figure 10
<p>The denoising performance in the scenario of two targets. (<b>a</b>) The mean square error versus the DOA interval; (<b>b</b>) the mean square error versus the target SNR.</p>
Full article ">Figure 11
<p>The DOA estimation results versus the DOA interval in the scenario of two targets. (<b>a</b>) The probability of resolution; (<b>b</b>) the root mean square error.</p>
Full article ">Figure 12
<p>The DOA estimation results versus those of the target SNR in the scenario of two targets. (<b>a</b>) The probability of resolution; (<b>b</b>) the root mean square error.</p>
Full article ">Figure 13
<p>The comparison of antenna patterns between the experiment and the simulation.</p>
Full article ">Figure 14
<p>The test scenario in the anechoic chamber.</p>
Full article ">Figure 15
<p>The experimental results in the scenario of one target. (<b>a</b>) The OMP method; (<b>b</b>) the ADMM method; (<b>c</b>) the ADMM-Net method; (<b>d</b>) the LDnADMM-Net method.</p>
Full article ">Figure 16
<p>The experimental results in the scenario of two targets separated by 2°. (<b>a</b>) The OMP method; (<b>b</b>) the ADMM method; (<b>c</b>) the ADMM-Net method; (<b>d</b>) the LDnADMM-Net method.</p>
Full article ">Figure 17
<p>The experimental results in the scenario of two targets separated by 4°. (<b>a</b>) The OMP method; (<b>b</b>) the ADMM method; (<b>c</b>) the ADMM-Net method; (<b>d</b>) the LDnADMM-Net method.</p>
Full article ">
20 pages, 10265 KiB  
Article
A Comparison of Different Machine Learning Methods to Reconstruct Daily Evapotranspiration Time Series Estimated by Thermal–Infrared Remote Sensing
by Gengle Zhao, Lisheng Song, Long Zhao and Sinuo Tao
Remote Sens. 2024, 16(3), 509; https://doi.org/10.3390/rs16030509 - 29 Jan 2024
Cited by 3 | Viewed by 1248
Abstract
Remote sensing-based models usually have difficulty in generating spatio-temporally continuous terrestrial evapotranspiration (ET) due to cloud cover and model failures. To overcome this problem, machine learning methods have been widely used to reconstruct ET. Therefore, studies comparing and evaluating the accuracy and effectiveness [...] Read more.
Remote sensing-based models usually have difficulty in generating spatio-temporally continuous terrestrial evapotranspiration (ET) due to cloud cover and model failures. To overcome this problem, machine learning methods have been widely used to reconstruct ET. Therefore, studies comparing and evaluating the accuracy and effectiveness of reconstruction among different machine learning methods at the basin scale are necessary. In this study, four popular machine learning methods, including deep forest (DF), deep neural network (DNN), random forest (RF) and extreme gradient boosting (XGB), were used to reconstruct the ET product, addressing gaps resulting from cloud cover and model failure. The ET reconstructed by the four methods was evaluated and compared for Heihe River Basin. The results showed that the four methods performed well for Heihe River Basin, but the RF method was particularly robust. It not only performed well compared with ground measurements (R = 0.73) but also demonstrated the ability to fully reconstruct gaps generated by the TSEB model across the entire basin. Validation based on ground measurements showed that the DNN and XGB models performed well (R > 0.70). However, some gaps still existed in the desert after reconstruction using the DNN and XGB models, especially for the XGB model. The DF model filled these gaps throughout the basin, but this model had lower consistency compared with ground measurements (R = 0.66) and yielded many low values. The results of this study suggest that machine learning methods have considerable potential in the reconstruction of ET at the basin scale. Full article
Show Figures

Figure 1

Figure 1
<p>Study area and vegetation type map for Heihe River Basin and the location of EC sites in the upstream, midstream and downstream regions, along with the landscape around the EC sites.</p>
Full article ">Figure 2
<p>Flowchart of the estimation and reconstruction of the daily T and E based on TSEB and four machine learning methods. Note that the EC measurement data were used only for accuracy verification.</p>
Full article ">Figure 3
<p>Pearson correlation coefficient matrix of all parameters. The sample size for the calculation of correlation coefficients was 2189858. All <span class="html-italic">p</span>−values for the correlation coefficients (two−tailed) were less than 0.01.</p>
Full article ">Figure 4
<p>Average impact values of input parameters calculated by the SHAP method.</p>
Full article ">Figure 5
<p>Validation of generated daily ET (including TSEB−estimated and reconstructed daily ET) using DF at six EC sites. The dashed line is a 1:1 line.</p>
Full article ">Figure 6
<p>Validation of generated daily ET (including TSEB−estimated and reconstructed daily ET) using DNN at six EC sites.</p>
Full article ">Figure 7
<p>Validation of generated daily ET (including TSEB−estimated and reconstructed daily ET) using RF at six EC sites.</p>
Full article ">Figure 8
<p>Validation of generated daily ET (including TSEB−estimated and reconstructed daily ET) using XGB at six EC sites.</p>
Full article ">Figure 9
<p>Validation of reconstructed daily ET by (<b>a</b>) DF, (<b>b</b>) DNN, (<b>c</b>) RF and (<b>d</b>) XGB at EC sites. Only daily ET reconstructed by machine learning methods are considered.</p>
Full article ">Figure 10
<p>Spatial distribution of relative uncertainties of daily ET reconstructed by four machine learning methods over Heihe River Basin.</p>
Full article ">Figure 11
<p>Plot of cumulative distribution frequency curves vs. effective coverage percentage of the daily ET. The area of the curve on the X−axis in the figure represent the missing amounts. The blue line in the figure indicates that the coverage of daily ET reconstructed by RF or DF was always 100%.</p>
Full article ">Figure 12
<p>Temporal coverage of ET estimated from (<b>a</b>–<b>d</b>) different machine learning methods and (<b>e</b>) original TSEB model.</p>
Full article ">Figure 13
<p>Spatial patterns of (<b>a</b>) TSEB-estimated daily ET and daily ET reconstructed by (<b>b</b>) DF, (<b>c</b>) DNN, (<b>d</b>) XGB and (<b>e</b>) RF in different seasons. The white areas represent gaps.</p>
Full article ">
12 pages, 6203 KiB  
Article
Predictive Modeling of Renewable Energy Purchase Prices Using Deep Learning Based on Polish Power Grid Data for Small Hybrid PV Microinstallations
by Michał Pikus and Jarosław Wąs
Energies 2024, 17(3), 628; https://doi.org/10.3390/en17030628 - 28 Jan 2024
Viewed by 1209
Abstract
In the quest for sustainable energy solutions, predicting electricity prices for renewable energy sources plays a pivotal role in efficient resource allocation and decision making. This article presents a novel approach to forecasting electricity prices for renewable energy sources using deep learning models, [...] Read more.
In the quest for sustainable energy solutions, predicting electricity prices for renewable energy sources plays a pivotal role in efficient resource allocation and decision making. This article presents a novel approach to forecasting electricity prices for renewable energy sources using deep learning models, leveraging historical data from the power system operator (PSE). The proposed methodology encompasses data collection, preprocessing, feature engineering, model selection, training, and evaluation. By harnessing the power of recurrent neural networks (RNNs) and other advanced deep learning architectures, the model captures intricate temporal relationships, weather patterns, and demand fluctuations that impact renewable energy prices. The study demonstrates the applicability of this approach through empirical analysis, showcasing its potential to enhance energy market predictions and aid in the transition to more sustainable energy systems. The outcomes underscore the importance of accurate renewable energy price predictions in fostering informed decision making and facilitating the integration of renewable sources into the energy landscape. As governments worldwide prioritize renewable energy adoption, this research contributes to the arsenal of tools driving the evolution towards a cleaner and more resilient energy future. Full article
(This article belongs to the Special Issue Smart Energy Systems: Learning Methods for Control and Optimization)
Show Figures

Figure 1

Figure 1
<p>Multi-LSTM DNN model architecture.</p>
Full article ">Figure 2
<p>Multi-LSTM price prediction results.</p>
Full article ">Figure 3
<p>Triple Victron Quattro inverters (source <a href="https://enerp.pl/project/victron-ess-kamienica-30kva-404kwh" target="_blank">https://enerp.pl/project/victron-ess-kamienica-30kva-404kwh</a>), accessed on 2 January 2023.</p>
Full article ">Figure 4
<p>Example of batteries and inverter installation: (<b>a</b>) one-phase installation including 1 inverter and 6 batteries; (<b>b</b>) separate batteries mounted on a rack (source <a href="https://enerp.pl/project/victron-ess-kalisz-45kva-50kwh" target="_blank">https://enerp.pl/project/victron-ess-kalisz-45kva-50kwh</a>), accessed on 2 January 2023.</p>
Full article ">Figure 5
<p>Example installation (source <a href="https://enerp.pl" target="_blank">https://enerp.pl</a>), accessed on 2 January 2023.</p>
Full article ">Figure 6
<p>PV panel installation.</p>
Full article ">
18 pages, 3181 KiB  
Article
HAVANA: Hard Negative Sample-Aware Self-Supervised Contrastive Learning for Airborne Laser Scanning Point Cloud Semantic Segmentation
by Yunsheng Zhang, Jianguo Yao, Ruixiang Zhang, Xuying Wang, Siyang Chen and Han Fu
Remote Sens. 2024, 16(3), 485; https://doi.org/10.3390/rs16030485 - 26 Jan 2024
Viewed by 1207
Abstract
Deep Neural Network (DNN)-based point cloud semantic segmentation has presented significant breakthrough using large-scale labeled aerial laser point cloud datasets. However, annotating such large-scaled point clouds is time-consuming. Self-Supervised Learning (SSL) is a promising approach to this problem by pre-training a DNN model [...] Read more.
Deep Neural Network (DNN)-based point cloud semantic segmentation has presented significant breakthrough using large-scale labeled aerial laser point cloud datasets. However, annotating such large-scaled point clouds is time-consuming. Self-Supervised Learning (SSL) is a promising approach to this problem by pre-training a DNN model utilizing unlabeled samples followed by a fine-tuned downstream task involving very limited labels. The traditional contrastive learning for point clouds selects the hardest negative samples by solely relying on the distance between the embedded features derived from the learning process, potentially evolving some negative samples from the same classes to reduce the contrastive learning effectiveness. This work proposes a hard-negative sample-aware self-supervised contrastive learning algorithm to pre-train the model for semantic segmentation. We designed a k-means clustering-based Absolute Positive And Negative samples (AbsPAN) strategy to filter the possible false-negative samples. Experiments on two typical ALS benchmark datasets demonstrate that the proposed method is more appealing than supervised training schemes without pre-training. Especially when the labels are severely inadequate (10% of the ISPRS training set), the results obtained by the proposed HAVANA method still exceed 94% of the supervised paradigm performance with full training set. Full article
Show Figures

Figure 1

Figure 1
<p>The self-supervised learning paradigm for point cloud semantic segmentation.</p>
Full article ">Figure 2
<p>Illustration of contrastive learning. For the backbone network, the green bars are encoders and the purple bars are decoders. The circled part of the red line represents our “AbsPAN” strategy selecting both positive and negative samples and optimizing the loss function.</p>
Full article ">Figure 3
<p>Illustration of the backbone network. The encoder (the first five layers) contains two convolutional blocks, kernel point convolution (KPConv) and strided kernel point convolution. The decoder (the last four layers) uses the nearest upsampling, ensuring that the pointwise feature is added up. The features transferred from the encoder layer are concatenated to the upsampled ones by skip links. The concatenated features are processed by the unary convolution, which is the equivalent of shared multi-layer perceptron (MLP) in PointNet.</p>
Full article ">Figure 4
<p>Illustration of negative samples mining processing: (<b>a</b>) raw point clouds; (<b>b</b>) clustering results, (<b>c</b>) hardest negative candidates, (<b>d</b>) AbsPAN results. Among them, the orange points represent block 1, and the green points represent block 2; (<b>c</b>) is in the embedding feature space of the network; (<b>b</b>) is the clustering result by k-means in geometric feature space, the red points represent cluster 1, and the blue points represent cluster 2.</p>
Full article ">Figure 5
<p>DALES point distribution across object categories.</p>
Full article ">Figure 6
<p>ISPRS training set. Section A (<b>left</b>) is the full training set. Section B (<b>right</b>) shows the subsets of training data cropped from all training sets. The five subsets have the same category distribution.</p>
Full article ">Figure 7
<p>ISPRS Vaihingen test set.</p>
Full article ">Figure 8
<p>LASDU data set. Annotated dataset with points of different labels presented in different colors. The black border divides the entire area into four separate regions.</p>
Full article ">Figure 9
<p>Classification results. (<b>a</b>) Visualized classification results of our HAVANA on the ISPRS test set, (<b>b</b>) error map of our HAVANA on the Vaihingen 3D dataset.</p>
Full article ">Figure 10
<p>Overall accuracy of different subsets in PointNet++ framework.</p>
Full article ">Figure 11
<p>Overall accuracy of different subsets.</p>
Full article ">Figure 12
<p>Visualization of LASDU dataset classification results.</p>
Full article ">Figure 13
<p>Classification error map in LASDU dataset. The green and red points indicate the correct and wrong classification, respectively.</p>
Full article ">
13 pages, 2044 KiB  
Article
Deep Neural Network Confidence Calibration from Stochastic Weight Averaging
by Zongjing Cao, Yan Li, Dong-Ho Kim and Byeong-Seok Shin
Electronics 2024, 13(3), 503; https://doi.org/10.3390/electronics13030503 - 25 Jan 2024
Cited by 1 | Viewed by 1524
Abstract
Overconfidence in deep neural networks (DNN) reduces the model’s generalization performance and increases its risk. The deep ensemble method improves model robustness and generalization of the model by combining prediction results from multiple DNNs. However, training multiple DNNs for model averaging is a [...] Read more.
Overconfidence in deep neural networks (DNN) reduces the model’s generalization performance and increases its risk. The deep ensemble method improves model robustness and generalization of the model by combining prediction results from multiple DNNs. However, training multiple DNNs for model averaging is a time-consuming and resource-intensive process. Moreover, combining multiple base learners (also called inducers) is hard to master, and any wrong choice may result in lower prediction accuracy than from a single inducer. We propose an approximation method for deep ensembles that can obtain ensembles of multiple DNNs without any additional costs. Specifically, multiple local optimal parameters generated during the training phase are sampled and saved by using an intelligent strategy. We use cycle learning rates starting at 75% of the training process and save the weights associated with the minimum learning rate in every iteration. Saved sets of the multiple model parameters are used as weights for a new model to perform forward propagation during the testing phase. Experiments on benchmarks of two different modalities, static images and dynamic videos, show that our method not only reduces the calibration error of the model but also improves the accuracy of the model. Full article
Show Figures

Figure 1

Figure 1
<p>Illustration of the 3D loss landscape with SGD optimization of the DNN during the training phase.</p>
Full article ">Figure 2
<p>The data flow of fusion during the testing phase of our method, where <math display="inline"><semantics> <mrow> <msub> <mi>θ</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>θ</mi> <mn>2</mn> </msub> <mo>,</mo> <mo>…</mo> <mo>,</mo> <msub> <mi>θ</mi> <mi>m</mi> </msub> </mrow> </semantics></math> indicate the set of <span class="html-italic">m</span> base estimators, and <math display="inline"><semantics> <mover accent="true"> <mi>y</mi> <mo>^</mo> </mover> </semantics></math> represents the output of base estimator, <math display="inline"><semantics> <msub> <mi>θ</mi> <mi>m</mi> </msub> </semantics></math>, on sample <span class="html-italic">x</span>.</p>
Full article ">Figure 3
<p>Illustration of the learning rate schedule. During the initial <math display="inline"><semantics> <mrow> <mn>75</mn> <mo>%</mo> </mrow> </semantics></math> of training, a standard decaying schedule is employed, followed by a high constant value for the remaining <math display="inline"><semantics> <mrow> <mn>25</mn> <mo>%</mo> </mrow> </semantics></math>. The dots of different colors represent the weights of the model in different training epochs.</p>
Full article ">Figure 4
<p>Comparison of reliability diagram and confidence histogram on the Jester test set. The upper charts are visualizations of average confidence and accuracy, and below them are reliability diagrams from (<b>left</b>) the SGD optimization method and (<b>right</b>) from our method.</p>
Full article ">Figure 5
<p>Comparison of reliability diagram and confidence histogram from VGG-16 trained on the CINIC-10 test set. The upper charts are visualizations of average confidence and accuracy, and below them are reliability diagrams from (<b>left</b>) the SGD optimization method and (<b>right</b>) from our method.</p>
Full article ">
35 pages, 14873 KiB  
Article
Real-Time Movie Recommendation: Integrating Persona-Based User Modeling with NMF and Deep Neural Networks
by Hyun-Chul Lee, Yong-Seong Kim and Seong-Whan Kim
Appl. Sci. 2024, 14(3), 1014; https://doi.org/10.3390/app14031014 - 24 Jan 2024
Cited by 1 | Viewed by 1371
Abstract
The proliferation of uncategorized information on the Internet has intensified the need for effective recommender systems. Recommender systems have evolved from content-based filtering to collaborative filtering and, most recently, to deep learning-based and hybrid models. However, they often face challenges such as high [...] Read more.
The proliferation of uncategorized information on the Internet has intensified the need for effective recommender systems. Recommender systems have evolved from content-based filtering to collaborative filtering and, most recently, to deep learning-based and hybrid models. However, they often face challenges such as high computational costs, reduced reliability, and the Cold Start problem. We introduce a persona-based user modeling approach for real-time movie recommendations. Our system employs Non-negative Matrix Factorization (NMF) and Deep Learning algorithms to manage complex and sparse data types and to mitigate the Cold Start issue. Experimental results, based on criteria involving 50 topics and 35 personas, indicate a significant performance gain. Specifically, with 500 users, the precision@K for NMF was 86.01%, and for the Deep Neural Network (DNN), it was 92.67%. Tested with 900 users, the precision@K for NMF increased to 97.04%, and for DNN, it was 95.55%. These results represent an approximate 10% and 5% improvement in performance, respectively. The system not only delivers fast and accurate recommendations but also reduces computational overhead by updating the model only when user personas change. The generated user personas can be adapted for other recommendation services or large-scale data mining. Full article
Show Figures

Figure 1

Figure 1
<p>Existing Recommender System—The process of Existing recommending movies with a conventional recommender system.</p>
Full article ">Figure 2
<p>Non-negative Matrix Factorization (NMF). NMF decomposes the entire matrix V into W and H.</p>
Full article ">Figure 3
<p>Topic Modeling [<a href="#B14-applsci-14-01014" class="html-bibr">14</a>]—‘Seeking life’s bare (genetics) necessities’ refers to data analysis aimed at determining the number of genes necessary for an organism to survive within the framework of evolution. The words marked in blue, such as ‘computer’ and ‘prediction’, relate to data analysis. Those marked in pink, like ‘life’ and ‘organism’, pertain to evolutionary biology. The terms highlighted in yellow, ‘sequenced’ and ‘genes’, are associated with genetics.</p>
Full article ">Figure 4
<p>Topic LDA—The LDA model graphically represented with plate notation.</p>
Full article ">Figure 5
<p>Movie Recommender System—NMF and Deep Learning Recommendation System using three modules.</p>
Full article ">Figure 6
<p>Composition of Recommendation Algorithm— User persona generation from movie evaluation data.</p>
Full article ">Figure 7
<p>Topic Modeling of Movie Data.</p>
Full article ">Figure 8
<p>User data preprocessing.</p>
Full article ">Figure 9
<p>Coherence Score—determine the number of topics (60) for the final topic modeling.</p>
Full article ">Figure 10
<p>Topic Modeling Results—the circles on the top side represent the topics, while the down side shows the words (movies) associated with each topic. In the bar chart on the right, the sky-blue bars represent the total number of words (movies), and the red bars represent the number of movies in the respective topic. Ultimately, the movie corresponding to the longest red bar can be one of the main movies of the chosen topic and becomes the top recommendation.</p>
Full article ">Figure 10 Cont.
<p>Topic Modeling Results—the circles on the top side represent the topics, while the down side shows the words (movies) associated with each topic. In the bar chart on the right, the sky-blue bars represent the total number of words (movies), and the red bars represent the number of movies in the respective topic. Ultimately, the movie corresponding to the longest red bar can be one of the main movies of the chosen topic and becomes the top recommendation.</p>
Full article ">Figure 10 Cont.
<p>Topic Modeling Results—the circles on the top side represent the topics, while the down side shows the words (movies) associated with each topic. In the bar chart on the right, the sky-blue bars represent the total number of words (movies), and the red bars represent the number of movies in the respective topic. Ultimately, the movie corresponding to the longest red bar can be one of the main movies of the chosen topic and becomes the top recommendation.</p>
Full article ">Figure 11
<p>Optimal K choice from K-mean Clustering.</p>
Full article ">Figure 12
<p>User Persona Construction.</p>
Full article ">Figure 13
<p>Deep Neural Networks for YouTube Recommendation—Recommendation system architecture demonstrating the “funnel” where candidate videos are retrieved and ranked before presenting only a few to the user.</p>
Full article ">Figure 14
<p>User Persona Movie Matrix—visualizes the process of receiving a user or system request, recommending a movie to the user using the trained NMF model, and providing feedback.</p>
Full article ">Figure 15
<p>Recommender—NMF.</p>
Full article ">Figure 16
<p>Recommender—Deep Learning.</p>
Full article ">Figure 17
<p>Movie Recommendation—Data flow of movie recommendation.</p>
Full article ">Figure 18
<p>Persona update Rule.</p>
Full article ">Figure 19
<p>Movie Recommendation Process according to Changes in User Preferences.</p>
Full article ">Figure 20
<p>Precision@K Performance of NMF and DNN with 100 Users under the Conditions of 70 Topics and 35 Personas.</p>
Full article ">Figure 21
<p>Precision@K Performance of NMF and DNN with 200 Users under the Conditions of 70 Topics and 35 Personas.</p>
Full article ">Figure 22
<p>Precision@K Performance of NMF and DNN with 500 Users under the Conditions of 50 Topics and 45 Personas.</p>
Full article ">Figure 23
<p>Precision@K Performance of NMF and DNN with 900 Users under the Conditions of 50 Topics and 45 Personas.</p>
Full article ">Figure 24
<p>The overall Configuration of the Persona Recommender System.</p>
Full article ">
21 pages, 3891 KiB  
Article
Advancing Process Control in Fluidized Bed Biomass Gasification Using Model-Based Deep Reinforcement Learning
by Ibtihaj Khurram Faridi, Evangelos Tsotsas and Abdolreza Kharaghani
Processes 2024, 12(2), 254; https://doi.org/10.3390/pr12020254 - 24 Jan 2024
Viewed by 1132
Abstract
This study presents a model-based deep reinforcement learning (MB-DRL) controller for the fluidized bed biomass gasification (FBG) process. The MB-DRL controller integrates a deep neural network (DNN) model and a reinforcement learning-based optimizer. The DNN model is trained with operational data from a [...] Read more.
This study presents a model-based deep reinforcement learning (MB-DRL) controller for the fluidized bed biomass gasification (FBG) process. The MB-DRL controller integrates a deep neural network (DNN) model and a reinforcement learning-based optimizer. The DNN model is trained with operational data from a pilot-scale FBG plant to approximate FBG process dynamics. The reinforcement learning-based optimizer employs a specially designed reward function, determining optimal control policies for FBG. Moreover, the controller includes an online learning component, ensuring periodic updates to the DNN model training. The performance of the controller is evaluated by testing its control accuracy for regulating synthetic gas composition, flow rate, and CO concentration in the FBG. The evaluation also includes a comparison with a model predictive controller. The results demonstrate the superior control performance of MB-DRL, surpassing MPC by over 15% in regulating synthetic gas composition and flow rate, with similar effectiveness observed in synthetic gas temperature control. Additionally, this study also includes systematic investigations into factors like DNN layer count and learning update intervals to provide insights for the practical implementation of the controller. The results, presenting a 50% reduction in control error with the addition of a single layer to the DNN model, highlight the significance of optimizing MB-DRL for effective implementation. Full article
Show Figures

Figure 1

Figure 1
<p>Schematic of the fluidized bed biomass gasification plant along with key instrumentation and control elements.</p>
Full article ">Figure 2
<p>The graphical representation of the proposed model-based reinforcement learning controller and its implementation on the fluidized bed biomass gasification process.</p>
Full article ">Figure 3
<p>Convergence analysis: training and validation losses (measured with mean squared error) of deep neural network (DNN) model.</p>
Full article ">Figure 4
<p>The basic structure of a gated recurrent unit (GRU) cell at time step <math display="inline"><semantics> <mrow> <mi>t</mi> </mrow> </semantics></math>. <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>x</mi> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> is the current input, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>z</mi> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> is the update gate, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>r</mi> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> is the reset gate, <math display="inline"><semantics> <mrow> <msub> <mrow> <mover accent="true"> <mrow> <mi>h</mi> </mrow> <mo stretchy="false">~</mo> </mover> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> is the candidate hidden state, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>h</mi> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> is the current hidden state, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>x</mi> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> is the input for the current neural network, and <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>h</mi> </mrow> <mrow> <mi>t</mi> <mo>−</mo> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math> is the previous moment’s hidden state. The activation function <math display="inline"><semantics> <mrow> <mi>σ</mi> </mrow> </semantics></math> is employed, and <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>h</mi> </mrow> <mrow> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> captures essential information influenced by the reset gate and input data.</p>
Full article ">Figure 5
<p>Gated recurrent unit (GRU) neural network model structure to model FBG.</p>
Full article ">Figure 6
<p>Training assessment of the gated recurrent neural network-based process model of fluidized bed biomass gasification using mean absolute error.</p>
Full article ">Figure 7
<p>The control performance of the proposed model-based deep reinforcement learning controller (MB-DRL) to control (<b>a</b>) synthetic gas flow rate, (<b>b</b>) CO vol% in the synthetic gas, and (<b>c</b>) the synthetic gas temperature along comparison with the model predictive control (MPC) method. Ref represents the dynamic trajectory of the set point that controllers aim to track.</p>
Full article ">Figure 7 Cont.
<p>The control performance of the proposed model-based deep reinforcement learning controller (MB-DRL) to control (<b>a</b>) synthetic gas flow rate, (<b>b</b>) CO vol% in the synthetic gas, and (<b>c</b>) the synthetic gas temperature along comparison with the model predictive control (MPC) method. Ref represents the dynamic trajectory of the set point that controllers aim to track.</p>
Full article ">Figure 8
<p>Response time analysis of the model-based deep reinforcement learning (MB-DRL) controller, presenting the controller’s speed of computing control actions during each run of the closed-loop system simulation.</p>
Full article ">Figure 9
<p>Influence of deep neural network (DNN) model depth on (<b>a</b>) computational cost and (<b>b</b>) control error for model-based reinforcement learning (MB-DRL) controller and the model predictive control (MPC) method.</p>
Full article ">Figure 9 Cont.
<p>Influence of deep neural network (DNN) model depth on (<b>a</b>) computational cost and (<b>b</b>) control error for model-based reinforcement learning (MB-DRL) controller and the model predictive control (MPC) method.</p>
Full article ">Figure 10
<p>The impact of various learning update intervals on a model-based reinforcement learning controller for regulating syngas composition in a fluidized bed biomass gasifier, presented in terms of control error (MAPE%) (left <span class="html-italic">y</span>-axis) and computational cost (right <span class="html-italic">y</span>-axis).</p>
Full article ">
9 pages, 840 KiB  
Proceeding Paper
Multimodal Deep Learning in Early Autism Detection—Recent Advances and Challenges
by Sheril Sophia Dcouto and Jawahar Pradeepkandhasamy
Eng. Proc. 2023, 59(1), 205; https://doi.org/10.3390/engproc2023059205 - 23 Jan 2024
Cited by 1 | Viewed by 3613
Abstract
Autism spectrum disorder (ASD) is a global concern, with a prevalence rate of approximately 1 in 36 children according to estimates from the Centers for Disease Control and Prevention (CDC). Diagnosing ASD poses challenges due to the absence of a definitive medical test. [...] Read more.
Autism spectrum disorder (ASD) is a global concern, with a prevalence rate of approximately 1 in 36 children according to estimates from the Centers for Disease Control and Prevention (CDC). Diagnosing ASD poses challenges due to the absence of a definitive medical test. Instead, doctors rely on a comprehensive evaluation of a child’s developmental background and behavior to reach a diagnosis. Although ASD can occasionally be identified in children aged 18 months or younger, a reliable diagnosis by an experienced professional is typically made by the age of two. Early detection of ASD is crucial for timely interventions and improved outcomes. In recent years, the field of early diagnosis of ASD has been greatly impacted by the emergence of deep learning models, which have brought about a revolution by greatly improving the accuracy and efficiency of ASD detection. The objective of this review paper is to examine the recent progress in early ASD detection through the utilization of multimodal deep learning techniques. The analysis revealed that integrating multiple modalities, including neuroimaging, genetics, and behavioral data, is key to achieving higher accuracy in early ASD detection. It is also evident that, while neuroimaging data holds promise and has the potential to contribute to higher accuracy in ASD detection, it is most effective when combined with other modalities. Deep learning models, with their ability to analyze complex patterns and extract meaningful features from large datasets, offer great promise in addressing the challenge of early ASD detection. Among various models used, CNN, DNN, GCN, and hybrid models have exhibited encouraging outcomes in the early detection of ASD. The review highlights the significance of developing accurate and easily accessible tools that utilize artificial intelligence (AI) to aid healthcare professionals, parents, and caregivers in early ASD symptom recognition. These tools would enable timely interventions, ensuring that necessary actions are taken during the initial stages. Full article
(This article belongs to the Proceedings of Eng. Proc., 2023, RAiSE-2023)
Show Figures

Figure 1

Figure 1
<p>Behavioral issues in ASD children.</p>
Full article ">Figure 2
<p>Traditional clinical evaluation for ASD.</p>
Full article ">Figure 3
<p>PRISMA review methodology.</p>
Full article ">
15 pages, 6796 KiB  
Article
The Construction and Application of a Deep Learning-Based Primary Support Deformation Prediction Model for Large Cross-Section Tunnels
by Junling Zhang, Min Mei, Jun Wang, Guangpeng Shang, Xuefeng Hu, Jing Yan and Qian Fang
Appl. Sci. 2024, 14(2), 912; https://doi.org/10.3390/app14020912 - 21 Jan 2024
Cited by 2 | Viewed by 850
Abstract
The deformation of tunnel support structures during tunnel construction is influenced by geological factors, geometrical factors, support factors, and construction factors. Accurate prediction of tunnel support structure deformation is crucial for engineering safety and optimizing support parameters. Traditional methods for tunnel deformation prediction [...] Read more.
The deformation of tunnel support structures during tunnel construction is influenced by geological factors, geometrical factors, support factors, and construction factors. Accurate prediction of tunnel support structure deformation is crucial for engineering safety and optimizing support parameters. Traditional methods for tunnel deformation prediction have often relied on numerical simulations and model experiments, which may not always meet the time-sensitive requirements. In this study, we propose a fusion deep neural network (FDNN) model that combines multiple algorithms with a complementary tunnel information encoding method. The FDNN model utilizes Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) networks to extract features related to tunnel structural deformation. FDNN model is used to predict deformations in the Capital Ring Expressway, and the predictions align well with monitoring results. To demonstrate the superiority of the proposed model, we use four different performance evaluation metrics to analyze the predictive performance of FDNN, DNN, XGBoost, Decision Tree Regression (DTR), and Random Forest Regression (RFR) methods. The results indicate that FDNN exhibits high precision and robustness. To assess the impact of different data types on the predictive results, we use tunnel geometry data as the base and combine geological, support, and construction data. The analysis reveals that models trained on datasets comprising all four data types perform the best. Geological parameters have the most significant impact on the predictive performance of all models. The findings of this research guide predicting tunnel construction parameters, particularly in the dynamic design of support parameters. Full article
(This article belongs to the Special Issue Urban Underground Engineering: Excavation, Monitoring, and Control)
Show Figures

Figure 1

Figure 1
<p>Cross-section of the tunnel support system.</p>
Full article ">Figure 2
<p>Layout of monitoring points in a standard tunnel cross-section.</p>
Full article ">Figure 3
<p>Monitoring data curve over time.</p>
Full article ">Figure 4
<p>Deformation distribution of initial tunnel support.</p>
Full article ">Figure 5
<p>Tunnel data encoding format.</p>
Full article ">Figure 6
<p>DNN model composition.</p>
Full article ">Figure 7
<p>Neuron in neural network.</p>
Full article ">Figure 8
<p>Random forest framework.</p>
Full article ">Figure 9
<p>Detailed flowchart of proposed FDNN model.</p>
Full article ">Figure 10
<p>Change of loss function during training process in training and test datasets.</p>
Full article ">Figure 11
<p>Results of prediction and measurement under training and test datasets. (<b>a</b>) Training set. (<b>b</b>) Test set.</p>
Full article ">Figure 12
<p><span class="html-italic">MAE</span> of all models under different datasets.</p>
Full article ">Figure 13
<p><span class="html-italic">R</span><sup>2</sup> of all models under different datasets.</p>
Full article ">
16 pages, 1913 KiB  
Article
Dynamic Telemetry and Deep Neural Networks for Anomaly Detection in 6G Software-Defined Networks
by Grzegorz Rzym, Amadeusz Masny and Piotr Chołda
Electronics 2024, 13(2), 382; https://doi.org/10.3390/electronics13020382 - 17 Jan 2024
Cited by 1 | Viewed by 1244
Abstract
With the increasing availability of computational power, contemporary machine learning has undergone a paradigm shift, placing a heightened emphasis on deep learning methodologies. The pervasive automation of various processes necessitates a critical re-evaluation of contemporary network implementations, specifically concerning security protocols and the [...] Read more.
With the increasing availability of computational power, contemporary machine learning has undergone a paradigm shift, placing a heightened emphasis on deep learning methodologies. The pervasive automation of various processes necessitates a critical re-evaluation of contemporary network implementations, specifically concerning security protocols and the imperative need for swift, precise responses to system failures. This article introduces a meticulously crafted solution designed explicitly for 6G software-defined networks (SDNs). The approach employs deep neural networks for anomaly detection within network traffic, contributing to a more robust security framework. Furthermore, the paper delves into the realm of network monitoring automation by harnessing dynamic telemetry, providing a specialized and forward-looking strategy to tackle the distinctive challenges inherent in SDN environments. In essence, our proposed solution aims to elevate the security and responsiveness of 6G mobile networks. By addressing the intricate challenges posed by next-generation network architectures, it seeks to fortify these networks against emerging threats and dynamically adapt to the evolving landscape of next-generation technology. Full article
(This article belongs to the Special Issue Smart Communication and Networking in the 6G Era)
Show Figures

Figure 1

Figure 1
<p>Anomaly detection mechanism operation diagram.</p>
Full article ">Figure 2
<p>System architecture.</p>
Full article ">Figure 3
<p>Traffic distribution for a 100 s interval along with predicted sigma intervals.</p>
Full article ">Figure 4
<p>Traffic distribution for a 300 s interval along with predicted sigma intervals.</p>
Full article ">Figure 5
<p>Distribution of misclassified anomalies throughout the day for two cases.</p>
Full article ">Figure 6
<p>Predicted traffic for a 300 s interval (3 cases).</p>
Full article ">Figure 7
<p>Predicted traffic for a 100 s interval (3 cases).</p>
Full article ">
Back to TopTop