[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,263)

Search Parameters:
Keywords = deep multi-task learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 12240 KiB  
Article
DMCCT: Dual-Branch Multi-Granularity Convolutional Cross-Substitution Transformer for Hyperspectral Image Classification
by Laiying Fu, Xiaoyong Chen, Yanan Xu and Xiao Li
Appl. Sci. 2024, 14(20), 9499; https://doi.org/10.3390/app14209499 - 17 Oct 2024
Abstract
In the field of hyperspectral image classification, deep learning technology, especially convolutional neural networks, has achieved remarkable progress. However, convolutional neural network models encounter challenges in hyperspectral image classification due to limitations in their receptive fields. Conversely, the global modeling capability of Transformers [...] Read more.
In the field of hyperspectral image classification, deep learning technology, especially convolutional neural networks, has achieved remarkable progress. However, convolutional neural network models encounter challenges in hyperspectral image classification due to limitations in their receptive fields. Conversely, the global modeling capability of Transformers has garnered attention in hyperspectral image classification. Nevertheless, the high computational cost and inadequate local feature extraction hinder its widespread application. In this study, we propose a novel fusion model of convolutional neural networks and Transformers to enhance performance in hyperspectral image classification, namely the dual-branch multi-granularity convolutional cross-substitution Transformer (DMCCT). The proposed model adopts a dual-branch structure to separately extract spatial and spectral features, thereby mitigating mutual interference and information loss between spectral and spatial data during feature extraction. Moreover, a multi-granularity embedding module is introduced to facilitate multi-scale and multi-level local feature extraction for spatial and spectral information. In particular, the improved convolutional cross-substitution Transformer module effectively integrates convolution and Transformer, reducing the complexity of attention operations and enhancing the accuracy of hyperspectral image classification tasks. Subsequently, the proposed method is evaluated against existing approaches using three classical datasets, namely Pavia University, Kennedy Space Center, and Indian Pines. Experimental results demonstrate the efficacy of the proposed method, achieving significant classification results on these datasets with overall classification accuracies of 98.57%, 97.96%, and 96.59%, respectively. These results establish the superiority of the proposed method in the context of hyperspectral image classification under similar experimental conditions. Full article
18 pages, 9570 KiB  
Article
A Depth Awareness and Learnable Feature Fusion Network for Enhanced Geometric Perception in Semantic Correspondence
by Fazeng Li, Chunlong Zou, Juntong Yun, Li Huang, Ying Liu, Bo Tao and Yuanmin Xie
Sensors 2024, 24(20), 6680; https://doi.org/10.3390/s24206680 - 17 Oct 2024
Abstract
Deep learning is becoming the most widely used technology for multi-sensor data fusion. Semantic correspondence has recently emerged as a foundational task, enabling a range of downstream applications, such as style or appearance transfer, robot manipulation, and pose estimation, through its ability to [...] Read more.
Deep learning is becoming the most widely used technology for multi-sensor data fusion. Semantic correspondence has recently emerged as a foundational task, enabling a range of downstream applications, such as style or appearance transfer, robot manipulation, and pose estimation, through its ability to provide robust correspondence in RGB images with semantic information. However, current representations generated by self-supervised learning and generative models are often limited in their ability to capture and understand the geometric structure of objects, which is significant for matching the correct details in applications of semantic correspondence. Furthermore, efficiently fusing these two types of features presents an interesting challenge. Achieving harmonious integration of these features is crucial for improving the expressive power of models in various tasks. To tackle these issues, our key idea is to integrate depth information from depth estimation or depth sensors into feature maps and leverage learnable weights for feature fusion. First, depth information is used to model pixel-wise depth distributions, assigning relative depth weights to feature maps for perceiving an object’s structural information. Then, based on a contrastive learning optimization objective, a series of weights are optimized to leverage feature maps from self-supervised learning and generative models. Depth features are naturally embedded into feature maps, guiding the network to learn geometric structure information about objects and alleviating depth ambiguity issues. Experiments on the SPair-71K and AP-10K datasets show that the proposed method achieves scores of 81.8 and 83.3 on the percentage of correct keypoints (PCK) at the 0.1 level, respectively. Our approach not only demonstrates significant advantages in experimental results but also introduces the depth awareness module and a learnable feature fusion module, which enhances the understanding of object structures through depth information and fully utilizes features from various pre-trained models, offering new possibilities for the application of deep learning in RGB and depth data fusion technologies. We will also continue to focus on accelerating model inference and optimizing model lightweighting, enabling our model to operate at a faster speed. Full article
(This article belongs to the Special Issue Machine and Deep Learning in Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The previous work [<a href="#B39-sensors-24-06680" class="html-bibr">39</a>] (<b>a</b>) found it challenging to differentiate between the front and rear wheels of motorcycles, and our method (<b>b</b>) aids in alleviating this issue. Green lines represent correct matches, and red is incorrect.</p>
Full article ">Figure 2
<p>An overview of our method pipeline.</p>
Full article ">Figure 3
<p>Pipeline of latent depth awareness module.</p>
Full article ">Figure 4
<p>Comparison of PCA from the feature map before and after processing through this module. From left to right: original image, PCA of original feature map, deep feature information, and final result.</p>
Full article ">Figure 5
<p>Framework of the feature fusion module.</p>
Full article ">Figure 6
<p>Qualitative comparison of dog, horse and sheep categories. Green lines represent correct matches, and red is incorrect. (<b>a</b>) Result of CATs++ [<a href="#B58-sensors-24-06680" class="html-bibr">58</a>], (<b>b</b>) result of DHF [<a href="#B38-sensors-24-06680" class="html-bibr">38</a>], (<b>c</b>) result of SD+DINO [<a href="#B39-sensors-24-06680" class="html-bibr">39</a>], (<b>d</b>) our result. Green lines represent correct matches, and red is incorrect.</p>
Full article ">Figure 7
<p>Qualitative comparison of bus, car, and train categories. Green lines represent correct matches, and red is incorrect. (<b>a</b>) Result of CATs++ [<a href="#B58-sensors-24-06680" class="html-bibr">58</a>], (<b>b</b>) result of DHF [<a href="#B38-sensors-24-06680" class="html-bibr">38</a>], (<b>c</b>) result of SD+DINO [<a href="#B39-sensors-24-06680" class="html-bibr">39</a>], (<b>d</b>) our result. Green lines represent correct matches, and red is incorrect.</p>
Full article ">Figure 8
<p>Qualitative comparison of person and TV categories. Green lines represent correct matches, and red is incorrect. (<b>a</b>) Result of CATs++ [<a href="#B58-sensors-24-06680" class="html-bibr">58</a>], (<b>b</b>) result of DHF [<a href="#B38-sensors-24-06680" class="html-bibr">38</a>], (<b>c</b>) result of SD+DINO [<a href="#B39-sensors-24-06680" class="html-bibr">39</a>], (<b>d</b>) our result. Green lines represent correct matches, and red is incorrect.</p>
Full article ">Figure 9
<p>The limitation of scale differences.</p>
Full article ">
17 pages, 6407 KiB  
Article
Intelligent Semantic Communication System Based on Kolmogorov–Arnold Networks Driven by Dynamic Terminal-Side Computing Power Network
by Wang Liu, Qingtao Zeng, Likun Lu and Waheed Abdul
Electronics 2024, 13(20), 4076; https://doi.org/10.3390/electronics13204076 - 17 Oct 2024
Viewed by 173
Abstract
With the advent of the 6G era, the number of IoT devices has experienced explosive growth, leading to the generation of massive amounts of data at the network edge. Semantic communication, as an innovative solution to handling this data deluge, can significantly enhance [...] Read more.
With the advent of the 6G era, the number of IoT devices has experienced explosive growth, leading to the generation of massive amounts of data at the network edge. Semantic communication, as an innovative solution to handling this data deluge, can significantly enhance communication efficiency. However, the limited storage and computational resources of terminal devices constrain the widespread application of semantic communication in 6G networks. To address this issue, we propose a terminal-side-computing-driven intelligent semantic communication solution. Specifically, we introduce a semantic communication model based on Kolmogorov–Arnold Networks (KANs), named K-DeepSC. Using image-reconstruction tasks as an example, the proposed K-DeepSC reduces the number of model parameters by 44% compared to semantic communication models based on Multi-Layer Perceptrons (MLPs), while maintaining similar performance. Furthermore, to fully leverage idle terminal computing power for semantic tasks, we explore computation offloading in dynamic Terminal-Side Computing Power Networks. By optimizing task delay minimization, a deep reinforcement learning algorithm is employed to determine the optimal offloading strategy. Simulation results demonstrate that our proposed solution effectively reduces semantic task processing delay. Full article
(This article belongs to the Special Issue 5G/B5G/6G Wireless Communication and Its Applications)
Show Figures

Figure 1

Figure 1
<p>Proposed semantic communication model K-DeepSC.</p>
Full article ">Figure 2
<p>Dynamic terminal-side computing power network system model.</p>
Full article ">Figure 3
<p>DQN architecture.</p>
Full article ">Figure 4
<p>Performance comparison of K-DeepSC and four other systems under AWGN and Rayleigh Channels.</p>
Full article ">Figure 5
<p>Convergence of the DQN algorithm under different batch size and learning rate.</p>
Full article ">Figure 6
<p>Comparison of rewards among different algorithms.</p>
Full article ">Figure 7
<p>Comparison of rewards and task completion rates among different algorithms under different sizes of semantic tasks.</p>
Full article ">Figure 8
<p>Rewards comparison of different algorithms under different number of terminals.</p>
Full article ">
21 pages, 2380 KiB  
Article
Crack Detection, Classification, and Segmentation on Road Pavement Material Using Multi-Scale Feature Aggregation and Transformer-Based Attention Mechanisms
by Arselan Ashraf, Ali Sophian and Ali Aryo Bawono
Constr. Mater. 2024, 4(4), 655-675; https://doi.org/10.3390/constrmater4040036 - 16 Oct 2024
Viewed by 185
Abstract
This paper introduces a novel approach to pavement material crack detection, classification, and segmentation using advanced deep learning techniques, including multi-scale feature aggregation and transformer-based attention mechanisms. The proposed methodology significantly enhances the model’s ability to handle varying crack sizes, shapes, and complex [...] Read more.
This paper introduces a novel approach to pavement material crack detection, classification, and segmentation using advanced deep learning techniques, including multi-scale feature aggregation and transformer-based attention mechanisms. The proposed methodology significantly enhances the model’s ability to handle varying crack sizes, shapes, and complex pavement textures. Trained on a dataset of 10,000 images, the model achieved substantial performance improvements across all tasks after integrating transformer-based attention. Detection precision increased from 88.7% to 94.3%, and IoU improved from 78.8% to 93.2%. In classification, precision rose from 88.3% to 94.8%, and recall improved from 86.8% to 94.2%. For segmentation, the Dice Coefficient increased from 80.3% to 94.7%, and IoU for segmentation advanced from 74.2% to 92.3%. These results underscore the model’s robustness and accuracy in identifying pavement cracks in challenging real-world scenarios. This framework not only advances automated pavement maintenance but also provides a foundation for future research focused on optimizing real-time processing and extending the model’s applicability to more diverse pavement conditions. Full article
Show Figures

Figure 1

Figure 1
<p>Trends in pavement crack detection, classification, and segmentation.</p>
Full article ">Figure 2
<p>Crack image segmentation.</p>
Full article ">Figure 3
<p>Research methodology.</p>
Full article ">Figure 4
<p>Confusion Matrix Analysis for Crack Classification before Transformer-Based Attention.</p>
Full article ">Figure 5
<p>Confusion matrix analysis for crack classification after transformer-based attention.</p>
Full article ">Figure 6
<p>Visual representation of the model’s output, showing the detection, classification, and segmentation of pavement cracks.</p>
Full article ">
23 pages, 4654 KiB  
Article
Effective Acoustic Model-Based Beamforming Training for Static and Dynamic Hri Applications
by Alejandro Luzanto, Nicolás Bohmer, Rodrigo Mahu, Eduardo Alvarado, Richard M. Stern and Néstor Becerra Yoma
Sensors 2024, 24(20), 6644; https://doi.org/10.3390/s24206644 - 15 Oct 2024
Viewed by 289
Abstract
Human–robot collaboration will play an important role in the fourth industrial revolution in applications related to hostile environments, mining, industry, forestry, education, natural disaster and defense. Effective collaboration requires robots to understand human intentions and tasks, which involves advanced user profiling. Voice-based communication, [...] Read more.
Human–robot collaboration will play an important role in the fourth industrial revolution in applications related to hostile environments, mining, industry, forestry, education, natural disaster and defense. Effective collaboration requires robots to understand human intentions and tasks, which involves advanced user profiling. Voice-based communication, rich in complex information, is key to this. Beamforming, a technology that enhances speech signals, can help robots extract semantic, emotional, or health-related information from speech. This paper describes the implementation of a system that provides substantially improved signal-to-noise ratio (SNR) and speech recognition accuracy to a moving robotic platform for use in human–robot interaction (HRI) applications in static and dynamic contexts. This study focuses on training deep learning-based beamformers using acoustic model-based multi-style training with measured room impulse responses (RIRs). The results show that this approach outperforms training with simulated RIRs or matched measured RIRs, especially in dynamic conditions involving robot motion. The findings suggest that training with a broad range of measured RIRs is sufficient for effective HRI in various environments, making additional data recording or augmentation unnecessary. This research demonstrates that deep learning-based beamforming can significantly improve HRI performance, particularly in challenging acoustic environments, surpassing traditional beamforming methods. Full article
(This article belongs to the Special Issue Advanced Sensors and AI Integration for Human–Robot Teaming)
Show Figures

Figure 1

Figure 1
<p>SA-MVDR training and evaluation stages.</p>
Full article ">Figure 2
<p>SA-RNN beamformer architecture.</p>
Full article ">Figure 3
<p>Block diagram of the signal processing in the mobile HRI platform.</p>
Full article ">Figure 4
<p>(<b>a</b>) Key dimensions used in the mobile HRI experiments; (<b>b</b>) the spacing of the microphones in the Kinect microphone array; (<b>c</b>) rear view of the HRI platform.</p>
Full article ">Figure 4 Cont.
<p>(<b>a</b>) Key dimensions used in the mobile HRI experiments; (<b>b</b>) the spacing of the microphones in the Kinect microphone array; (<b>c</b>) rear view of the HRI platform.</p>
Full article ">Figure 5
<p>The PR2 robot’s head performs a periodic angular sweep between φ1 and φ2 at a given angular velocity for recording the testing datasets in dynamic conditions (see <a href="#sensors-24-06644-t001" class="html-table">Table 1</a>).</p>
Full article ">Figure 6
<p>(<b>a</b>) Set up for recording RIRs: (<b>a</b>) Meeting Room; (<b>b</b>) Classroom 1; (<b>c</b>) Classroom 2; (<b>d</b>) Seminary room; (<b>e</b>) Auditorium.</p>
Full article ">Figure 7
<p>Results of Measured Matched Acoustic Modeling (MMAM) and Measured Generic Acoustic Modeling (MGAM) under static conditions.</p>
Full article ">Figure 8
<p>Results of Measured Matched Acoustic Modeling (MMAM) and Measured Generic Acoustic Modeling (MGAM) under dynamic conditions.</p>
Full article ">Figure 9
<p>Results of Simulated Generic Acoustic Modeling (SGAM) and Measured Generic Acoustic Modeling (MGAM) under static conditions.</p>
Full article ">Figure 10
<p>Results of Simulated Generic Acoustic Modeling (SGAM) and Measured Generic Acoustic Modeling (MGAM) under dynamic conditions.</p>
Full article ">
18 pages, 1657 KiB  
Technical Note
Emitter Signal Deinterleaving Based on Single PDW with Modulation-Hypothesis-Augmented Transformer
by Huajun Liu, Longfei Wang and Gan Wang
Remote Sens. 2024, 16(20), 3830; https://doi.org/10.3390/rs16203830 (registering DOI) - 15 Oct 2024
Viewed by 247
Abstract
Radar emitter signal deinterleaving based on pulse description words (PDWs) is a challenging task in the field of electronic warfare because of the parameter sparsity and uncertainty of PDWs. In this paper, a modulation-hypothesis-augmented Transformer model is proposed to identify emitters from a [...] Read more.
Radar emitter signal deinterleaving based on pulse description words (PDWs) is a challenging task in the field of electronic warfare because of the parameter sparsity and uncertainty of PDWs. In this paper, a modulation-hypothesis-augmented Transformer model is proposed to identify emitters from a single PDW with an end-to-end manner. Firstly, the pulse features are enriched by the modulation hypothesis mechanism to generate I/Q complex signals from PDWs. Secondly, a multiple-parameter embedding method is proposed to expand the signal discriminative features and to enhance the identification capability of emitters. Moreover, a novel Transformer deep learning model, named PulseFormer and composed of spectral convolution, multi-layer perceptron, and self-attention based basic blocks, is proposed for discriminative feature extraction, emitter identification, and signal deinterleaving. Experimental results on synthesized PDW dataset show that the proposed method performs better on emitter signal deinterleaving in complex environments without relying on the pulse repetition interval (PRI). Compared with other deep learning methods, the PulseFormer performs better in noisy environments. Full article
Show Figures

Figure 1

Figure 1
<p>The diagram of Transformer-based signal deinterleaving.</p>
Full article ">Figure 2
<p>The architecture of the PulseFormer model.</p>
Full article ">Figure 3
<p>Components of the PulseFormer model: (<b>a</b>) SP-Conv; (<b>b</b>) MHSA; (<b>c</b>) MLP.</p>
Full article ">Figure 4
<p>Histogram visualization of different features in the first experiment: (<b>a</b>) PW; (<b>b</b>) CF; (<b>c</b>) PA; (<b>d</b>) DOA.</p>
Full article ">Figure 5
<p>Histogram visualization of different features in the second experiment: (<b>a</b>) PW; (<b>b</b>) CF; (<b>c</b>) PA; (<b>d</b>) DOA.</p>
Full article ">Figure 6
<p>The confusion matrix w/w.o. multiple-parameter embedding. (<b>a</b>) Modulation-hypothesis augmentation. (<b>b</b>) Modulation-hypothesis augmentation with multiple-parameter embedding.</p>
Full article ">Figure 7
<p>The t-SNE visualization shows the high-dimensional feature distribution predicted by our model under the modulation-hypothesis augmentation methods. (<b>a</b>) Without multiple-parameter embedding. (<b>b</b>) With multiple-parameter embedding.</p>
Full article ">Figure 8
<p>Performance comparison under different noise conditions.</p>
Full article ">Figure 9
<p>Performance comparison under pulse loss.</p>
Full article ">
13 pages, 1781 KiB  
Article
A3GT: An Adaptive Asynchronous Generalized Adversarial Training Method
by Zeyi He, Wanyi Liu, Zheng Huang, Yitian Chen and Shigeng Zhang
Electronics 2024, 13(20), 4052; https://doi.org/10.3390/electronics13204052 (registering DOI) - 15 Oct 2024
Viewed by 266
Abstract
Adversarial attack methods can significantly improve the classification accuracy of deep learning models, but research has found that although most deep learning models with defense methods still show good classification accuracy in the face of various adversarial attack attacks, the improved robust models [...] Read more.
Adversarial attack methods can significantly improve the classification accuracy of deep learning models, but research has found that although most deep learning models with defense methods still show good classification accuracy in the face of various adversarial attack attacks, the improved robust models have a significantly lower classification accuracy when facing clean samples compared to themselves without using defense methods. This means that while improving the model’s adversarial robustness, it is necessary to find a defense method to balance the accuracy of clean samples (clean accuracy) and the accuracy of adversarial samples (robust accuracy). Therefore, in this work, we propose an Adaptive Asynchronous Generalized Adversarial Training (A3GT) method, which is an improvement over the existing Generalist method. It employs an adaptive update strategy without the need for extensive experiments to determine the optimal starting iteration for global updates. The experimental results show that compared with other advanced methods, A3GT can achieve a balance between clean sample classification accuracy and robust classification accuracy while improving the model’s adversarial robustness. Full article
(This article belongs to the Special Issue Adversarial Attacks and Defenses in AI Safety/Reliability)
Show Figures

Figure 1

Figure 1
<p>The overall framework of the Adaptive Asynchronous Generalized Adversarial Training.</p>
Full article ">Figure 2
<p>Comparison of clean accuracy and robust accuracy (using AutoAttack adversarial attacks [<a href="#B25-electronics-13-04052" class="html-bibr">25</a>]) between the A3GT method and other adversarial training methods.</p>
Full article ">Figure 3
<p>CMMR scores and perturbation strength curve of A3GT method compared with other defense methods.</p>
Full article ">Figure 4
<p>Comparison of CMMR scores between A3GT method and other defense methods at perturbation strengths of <math display="inline"><semantics> <mrow> <mi>ϵ</mi> <mo>=</mo> <mn>0.04</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>ϵ</mi> <mo>=</mo> <mn>0.08</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>ϵ</mi> <mo>=</mo> <mn>0.12</mn> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <mi>ϵ</mi> <mo>=</mo> <mn>0.16</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 5
<p>Comparison of different global learner learning frequencies <span class="html-italic">c</span> and two basic learner mixing ratios <math display="inline"><semantics> <mi>γ</mi> </semantics></math> under the A3GT method. The experiments evaluated clean accuracy and robust accuracy against PGD, C&amp;W, and AutoAttack attacks using ResNet-18 as the base model.</p>
Full article ">
13 pages, 5647 KiB  
Article
ResNet Modeling for 12 nm FinFET Devices to Enhance DTCO Efficiency
by Yiming Huang, Bin Li, Zhaohui Wu and Wenchao Liu
Electronics 2024, 13(20), 4040; https://doi.org/10.3390/electronics13204040 - 14 Oct 2024
Viewed by 346
Abstract
In this paper, a deep learning-based device modeling framework for design-technology co-optimization (DTCO) is proposed. A ResNet surrogate model is utilized as an alternative to traditional compact models, demonstrating high accuracy in both single-task (I–V or C–V) and multi-task (I–V and C–V) device [...] Read more.
In this paper, a deep learning-based device modeling framework for design-technology co-optimization (DTCO) is proposed. A ResNet surrogate model is utilized as an alternative to traditional compact models, demonstrating high accuracy in both single-task (I–V or C–V) and multi-task (I–V and C–V) device modeling. Moreover, transfer learning is applied to the ResNet model, using the BSIM-CMG compact model for a 12 nm FinFET SPICE model as the pre-trained source. Through this approach, superior modeling accuracy and faster training speed are achieved compared to a ResNet surrogate model initialized with random weights, thereby meeting the rapid and efficient demands of the DTCO process. The effectiveness of the ResNet surrogate model in circuit simulation for 12 nm FinFET devices is demonstrated. Full article
Show Figures

Figure 1

Figure 1
<p>Preprocessing of the dataset.</p>
Full article ">Figure 2
<p>ResNet structure.</p>
Full article ">Figure 3
<p>(<b>a</b>) Lintanh curves. (<b>b</b>) Lintanh’s first derivatives curves. (<b>c</b>) Lintanh’s second derivatives.</p>
Full article ">Figure 4
<p>Training and prediction process of the ResNet model.</p>
Full article ">Figure 5
<p>Comparison of ResNet model performance in 12 nm FinFET single-task (I–V) and multi-task (I–V and C–V) modeling with (<b>a</b>) training loss (AdaptiveSmoothL1); (<b>b</b>) validation loss (MAPE); and (<b>c</b>) test-set accuracy (R<sup>2</sup>).</p>
Full article ">Figure 6
<p>Comparison of ANN and ResNet model performance in 12 nm FinFET single-task (I–V) and multi-task (I–V and C–V) modeling for (<b>a</b>) training loss (AdaptiveSmoothL1), (<b>b</b>) validation loss (MAPE), and (<b>c</b>) test-set accuracy (R<sup>2</sup>).</p>
Full article ">Figure 7
<p>Comparison of ResNet Predictions and SPICE Simulations for FinFET I–V and C–V Characteristics. (<b>a</b>) Ids–Vds, (<b>b</b>) Cgs–Vds, (<b>c</b>) Cgd–Vds, and (<b>d</b>) Cgb–Vds.</p>
Full article ">Figure 8
<p>ResNet predictions vs. SPICE simulations: Scatter plots of I–V and C–V characteristics of FinFET devices. (<b>a</b>) Ids, (<b>b</b>) Cgs, (<b>c</b>) Cgd, and (<b>d</b>) Cgb.</p>
Full article ">Figure 9
<p>Performance of ResNet model on FinFET generalization test set. (<b>a</b>) Single-task IV modeling, (<b>b</b>) single-task IV migration learning modeling, (<b>c</b>) multi-task IV-CV modeling, and (<b>d</b>) multi-task IV–CV migration learning modeling.</p>
Full article ">Figure 10
<p>RO simulation curve SPICE model vs. ResNet model.</p>
Full article ">
19 pages, 1236 KiB  
Article
Multi-Task Diffusion Learning for Time Series Classification
by Shaoqiu Zheng, Zhen Liu, Long Tian, Ling Ye, Shixin Zheng, Peng Peng and Wei Chu
Electronics 2024, 13(20), 4015; https://doi.org/10.3390/electronics13204015 - 12 Oct 2024
Viewed by 301
Abstract
Current deep learning models for time series often face challenges with generalizability in scenarios characterized by limited samples or inadequately labeled data. By tapping into the robust generative capabilities of diffusion models, which have shown success in computer vision and natural language processing, [...] Read more.
Current deep learning models for time series often face challenges with generalizability in scenarios characterized by limited samples or inadequately labeled data. By tapping into the robust generative capabilities of diffusion models, which have shown success in computer vision and natural language processing, we see potential for improving the adaptability of deep learning models. However, the specific application of diffusion models in generating samples for time series classification tasks remains underexplored. To bridge this gap, we introduce the MDGPS model, which incorporates multi-task diffusion learning and gradient-free patch search (MDGPS). Our methodology aims to bolster the generalizability of time series classification models confronted with restricted labeled samples. The multi-task diffusion learning module integrates frequency-domain classification with random masked patches diffusion learning, leveraging frequency-domain feature representations and patch observation distributions to improve the discriminative properties of generated samples. Furthermore, a gradient-free patch search module, utilizing the particle swarm optimization algorithm, refines time series for specific samples through a pre-trained multi-task diffusion model. This process aims to reduce classification errors caused by random patch masking. The experimental results on four time series datasets show that the proposed MDGPS model consistently surpasses other methods, achieving the highest classification accuracy and F1-score across all datasets: 95.81%, 87.64%, 82.31%, and 100% in accuracy; and 95.21%, 82.32%, 78.57%, and 100% in F1-Score for Epilepsy, FD-B, Gesture, and EMG, respectively. In addition, evaluations in a reinforcement learning scenario confirm MDGPS’s superior performance. Ablation and visualization experiments further validate the effectiveness of its individual components. Full article
(This article belongs to the Special Issue Advances in Algorithm Optimization and Computational Intelligence)
Show Figures

Figure 1

Figure 1
<p>The framework for the multi-task diffusion learning with gradient-free patch search for time series classification model.</p>
Full article ">Figure 2
<p>Flowchart depicting the operation of the gradient-free patch search module utilizing the particle swarm optimization algorithm.</p>
Full article ">Figure 3
<p>The training time (s) of different methods on the EMG dataset.</p>
Full article ">Figure 4
<p>The t-SNE visualization of the learned representations on the Epilepsy dataset.</p>
Full article ">Figure 5
<p>The t-SNE visualization of the learned representations on the FD-B dataset.</p>
Full article ">Figure 6
<p>The comparison results depict the ratio of defeated fighters between the red and blue sides in the intelligent agent and environment exchange reinforcement learning scenario.</p>
Full article ">Figure 7
<p>Comparative results on classification accuracy of reinforcement learning decisions based on the cart pole and the mountain car behavioural clones.</p>
Full article ">Figure 8
<p>Comparison results of reward values in behavioral cloning.</p>
Full article ">
18 pages, 4253 KiB  
Article
RSTSRN: Recursive Swin Transformer Super-Resolution Network for Mars Images
by Fanlu Wu, Xiaonan Jiang, Tianjiao Fu, Yao Fu, Dongdong Xu and Chunlei Zhao
Appl. Sci. 2024, 14(20), 9286; https://doi.org/10.3390/app14209286 - 12 Oct 2024
Viewed by 429
Abstract
High-resolution optical images will provide planetary geology researchers with finer and more microscopic image data information. In order to maximize scientific output, it is necessary to further increase the resolution of acquired images, so image super-resolution (SR) reconstruction techniques have become the best [...] Read more.
High-resolution optical images will provide planetary geology researchers with finer and more microscopic image data information. In order to maximize scientific output, it is necessary to further increase the resolution of acquired images, so image super-resolution (SR) reconstruction techniques have become the best choice. Aiming at the problems of large parameter quantity and high computational complexity in current deep learning-based image SR reconstruction methods, we propose a novel Recursive Swin Transformer Super-Resolution Network (RSTSRN) for SR applied to images. The RSTSRN improves upon the LapSRN, which we use as our backbone architecture. A Residual Swin Transformer Block (RSTB) is used for more efficient residual learning, which consists of stacked Swin Transformer Blocks (STBs) with a residual connection. Moreover, the idea of parameter sharing was introduced to reduce the number of parameters, and a multi-scale training strategy was designed to accelerate convergence speed. Experimental results show that the proposed RSTSRN achieves superior performance on 2×, 4× and 8×SR tasks to state-of-the-art methods with similar parameters. Especially on high-magnification SR tasks, the RSTSRN has great performance superiority. Compared to the LapSRN network, for 2×, 4× and 8× Mars image SR tasks, the RSTSRN network has increased PSNR values by 0.35 dB, 0.88 dB and 1.22 dB, and SSIM values by 0.0048, 0.0114 and 0.0311, respectively. Full article
(This article belongs to the Special Issue Advances in Image Recognition and Processing Technologies)
Show Figures

Figure 1

Figure 1
<p>RSTSRN network architecture.</p>
Full article ">Figure 2
<p>Multi-scale training process.</p>
Full article ">Figure 3
<p>Shifted window.</p>
Full article ">Figure 4
<p>Examples of HiRISE images.</p>
Full article ">Figure 5
<p>Visual comparison for 2×SR on the barbara.</p>
Full article ">Figure 6
<p>Visual comparison for 4× SR on the butterfly.</p>
Full article ">Figure 7
<p>Visual comparison for 8× SR on the img_040.</p>
Full article ">Figure 8
<p>The 2× SR results of the 54th Mars image.</p>
Full article ">Figure 9
<p>The 4× SR results of the 59th Mars image.</p>
Full article ">Figure 10
<p>The 8× SR results of the 26th Mars image.</p>
Full article ">
24 pages, 7029 KiB  
Article
Multi-UAV Cooperative Pursuit of a Fast-Moving Target UAV Based on the GM-TD3 Algorithm
by Yaozhong Zhang, Meiyan Ding, Yao Yuan, Jiandong Zhang, Qiming Yang, Guoqing Shi, Frank Jiang and Meiqu Lu
Drones 2024, 8(10), 557; https://doi.org/10.3390/drones8100557 - 8 Oct 2024
Viewed by 420
Abstract
Recently, developing multi-UAVs to cooperatively pursue a fast-moving target has become a research hotspot in the current world. Although deep reinforcement learning (DRL) has made a lot of achievements in the UAV pursuit game, there are still some problems such as high-dimensional parameter [...] Read more.
Recently, developing multi-UAVs to cooperatively pursue a fast-moving target has become a research hotspot in the current world. Although deep reinforcement learning (DRL) has made a lot of achievements in the UAV pursuit game, there are still some problems such as high-dimensional parameter space, the ease of falling into local optimization, the long training time, and the low task success rate. To solve the above-mentioned issues, we propose an improved twin delayed deep deterministic policy gradient algorithm combining the genetic algorithm and maximum mean discrepancy method (GM-TD3) for multi-UAV cooperative pursuit of high-speed targets. Firstly, this paper combines GA-based evolutionary strategies with TD3 to generate action networks. Then, in order to avoid local optimization in the algorithm training process, the maximum mean difference (MMD) method is used to increase the diversity of the policy population in the updating process of the population parameters. Finally, by setting the sensitivity weights of the genetic memory buffer of UAV individuals, the mutation operator is improved to enhance the stability of the algorithm. In addition, this paper designs a hybrid reward function to accelerate the convergence speed of training. Through simulation experiments, we have verified that the training efficiency of the improved algorithm has been greatly improved, which can achieve faster convergence; the successful rate of the task has reached 95%, and further validated UAVs can better cooperate to complete the pursuit game task. Full article
Show Figures

Figure 1

Figure 1
<p>A schematic diagram of the cooperative pursuit problem of multi-UAVs.</p>
Full article ">Figure 2
<p>A schematic diagram of the UAV motion control model.</p>
Full article ">Figure 3
<p>Schematic diagram of UAV radar detection and intra-swarm communication model.</p>
Full article ">Figure 4
<p>The effect of multi-UAVs cooperatively rounding up the target.</p>
Full article ">Figure 5
<p>Diagram of three UAVs completing pursuit task.</p>
Full article ">Figure 6
<p>Diagram of ERL basic framework.</p>
Full article ">Figure 7
<p>Diagram of basic flow of GA.</p>
Full article ">Figure 8
<p>Diagram of TD3 algorithm framework.</p>
Full article ">Figure 9
<p>Framework of GM-TD3 algorithm.</p>
Full article ">Figure 10
<p>Flowchart of GM-TD3 algorithm.</p>
Full article ">Figure 11
<p>Relationship between relative positions between UAVs.</p>
Full article ">Figure 12
<p>Mean and variance change curve of <span class="html-italic">actor_eval</span> neural network parameters during training.</p>
Full article ">Figure 13
<p>Reward curve during algorithm training.</p>
Full article ">Figure 14
<p>Snapshot of simulation effects for 3vs1 pursuit game task with NFZs.</p>
Full article ">Figure 15
<p>Snapshot of simulation effects for 4vs1 pursuit game task with 10 NFZs.</p>
Full article ">Figure 16
<p>Snapshot of simulation effects for 6vs1 pursuit game task with 10 NFZs.</p>
Full article ">Figure 16 Cont.
<p>Snapshot of simulation effects for 6vs1 pursuit game task with 10 NFZs.</p>
Full article ">Figure 17
<p>Snapshot of simulation effects for 3vs1 pursuit game task without NFZs.</p>
Full article ">Figure 18
<p>The global reward curve of the three algorithms.</p>
Full article ">Figure 19
<p>The task success rates of the three algorithms.</p>
Full article ">Figure 20
<p>Task success rate under different UAV maximum speed ratios.</p>
Full article ">
16 pages, 5803 KiB  
Article
A Comparative Study of Metaheuristic Feature Selection Algorithms for Respiratory Disease Classification
by Damla Gürkan Kuntalp, Nermin Özcan, Okan Düzyel, Fevzi Yasin Kababulut and Mehmet Kuntalp
Diagnostics 2024, 14(19), 2244; https://doi.org/10.3390/diagnostics14192244 - 8 Oct 2024
Viewed by 375
Abstract
The correct diagnosis and early treatment of respiratory diseases can significantly improve the health status of patients, reduce healthcare expenses, and enhance quality of life. Therefore, there has been extensive interest in developing automatic respiratory disease detection systems. Most recent methods for detecting [...] Read more.
The correct diagnosis and early treatment of respiratory diseases can significantly improve the health status of patients, reduce healthcare expenses, and enhance quality of life. Therefore, there has been extensive interest in developing automatic respiratory disease detection systems. Most recent methods for detecting respiratory disease use machine and deep learning algorithms. The success of these machine learning methods depends heavily on the selection of proper features to be used in the classifier. Although metaheuristic-based feature selection methods have been successful in addressing difficulties presented by high-dimensional medical data in various biomedical classification tasks, there is not much research on the utilization of metaheuristic methods in respiratory disease classification. This paper aims to conduct a detailed and comparative analysis of six widely used metaheuristic optimization methods using eight different transfer functions in respiratory disease classification. For this purpose, two different classification cases were examined: binary and multi-class. The findings demonstrate that metaheuristic algorithms using correct transfer functions could effectively reduce data dimensionality while enhancing classification accuracy. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

Figure 1
<p>V-shaped and S-shaped transfer functions.</p>
Full article ">Figure 2
<p>Fitness values of different transfer functions for each MHA FS method for Case 1.</p>
Full article ">Figure 3
<p>Fitness values of different transfer functions for each MHA FS method for Case 2.</p>
Full article ">Figure 4
<p>Comparison of MHA FS methods for individual transfer functions for Case 1.</p>
Full article ">Figure 5
<p>Comparison of MHA FS methods for individual transfer functions for Case 2.</p>
Full article ">Figure 6
<p>Selected feature sizes of MHA FS methods for S- and V-shaped transfer functions for Case 1.</p>
Full article ">Figure 7
<p>Selected feature sizes of MHA methods for S- and V-shaped transfer functions for Case 2.</p>
Full article ">
18 pages, 7988 KiB  
Article
Wind Turbine Bearing Failure Diagnosis Using Multi-Scale Feature Extraction and Residual Neural Networks with Block Attention
by Yuanqing Luo, Yuhang Yang, Shuang Kang, Xueyong Tian, Shiyue Liu and Feng Sun
Actuators 2024, 13(10), 401; https://doi.org/10.3390/act13100401 - 5 Oct 2024
Viewed by 403
Abstract
Wind turbine rolling bearings are crucial components for ensuring the reliability and stability of wind power systems. Their failure can lead to significant economic losses and equipment downtime. Therefore, the accurate diagnosis of bearing faults is of great importance. Although existing deep learning [...] Read more.
Wind turbine rolling bearings are crucial components for ensuring the reliability and stability of wind power systems. Their failure can lead to significant economic losses and equipment downtime. Therefore, the accurate diagnosis of bearing faults is of great importance. Although existing deep learning fault diagnosis methods have achieved certain results, they still face limitations such as inadequate feature extraction capabilities, insufficient generalization to complex working conditions, and ineffective multi-scale feature capture. To address these issues, this paper proposes an advanced fault diagnosis method named the two-stream feature fusion convolutional neural network (TSFFResNet-Net). Firstly, the proposed method combines the advantages of one-dimensional convolutional neural networks (1D-ResNet) and two-dimensional convolutional neural networks (2D-ResNet). It transforms one-dimensional vibration signals into two-dimensional images through the empirical wavelet transform (EWT) method. Then, parallel convolutional kernels in 1D-ResNet and 2D-ResNet are used to extract multi-scale features, respectively. Next, the Convolutional Block Attention Module (CBAM) is introduced to enhance the network’s ability to capture key features by focusing on important features in specific channels or spatial areas. After feature fusion, CBAM is introduced again to further enhance the effect of feature fusion, ensuring that the features extracted by different network branches can be effectively integrated, ultimately providing more accurate input features for the classification task of the fully connected layer. The experimental results demonstrate that the proposed method outperforms other traditional methods and advanced convolutional neural network models on different datasets. Compared with convolutional neural network models such as LeNet-5, AlexNet, and ResNet, the proposed method achieves a significantly higher accuracy on the test set, with a stable accuracy of over 99%. Compared with other models, it shows better generalization and stability, effectively improving the overall performance of rolling bearing vibration signal fault diagnosis. The method provides an effective solution for the intelligent fault diagnosis of wind turbine rolling bearings. Full article
(This article belongs to the Section Actuators for Manufacturing Systems)
Show Figures

Figure 1

Figure 1
<p>Convolutional neural network structure diagram.</p>
Full article ">Figure 2
<p>CBAM module flow.</p>
Full article ">Figure 3
<p>Multi-scale feature extraction module.</p>
Full article ">Figure 4
<p>TSFFResNet-Net structure model.</p>
Full article ">Figure 5
<p>CWRU bearing fault diagnosis testbench.</p>
Full article ">Figure 6
<p>Data input for different channels. (<b>a</b>) One-dimensional signal. (<b>b</b>) Two-dimensional signal.</p>
Full article ">Figure 7
<p>Accuracy and loss during training.</p>
Full article ">Figure 8
<p>Data visualization results.</p>
Full article ">Figure 9
<p>Precision comparison of the proposed method under different SNR.</p>
Full article ">Figure 10
<p>Comparison results of confusion matrix with SNR of 3dB.</p>
Full article ">Figure 11
<p>Accuracy curves of four algorithms.</p>
Full article ">Figure 12
<p>Bearing test platform.</p>
Full article ">Figure 13
<p>Experimental comparison results:(<b>a</b>) Comparative results of a single experiment. (<b>b</b>) The average performance change results of five experiments.</p>
Full article ">Figure 14
<p>Diagnostic accuracy of different methods.</p>
Full article ">Figure 15
<p>Comparison results of confusion matrix of different methods.</p>
Full article ">
20 pages, 3829 KiB  
Article
Beyond Granularity: Enhancing Continuous Sign Language Recognition with Granularity-Aware Feature Fusion and Attention Optimization
by Yao Du, Taiying Peng and Xiaohui Hu
Appl. Sci. 2024, 14(19), 8937; https://doi.org/10.3390/app14198937 - 4 Oct 2024
Viewed by 518
Abstract
The advancement of deep learning techniques has significantly propelled the development of the continuous sign language recognition (cSLR) task. However, the spatial feature extraction of sign language videos in the RGB space tends to focus on the overall image information while neglecting the [...] Read more.
The advancement of deep learning techniques has significantly propelled the development of the continuous sign language recognition (cSLR) task. However, the spatial feature extraction of sign language videos in the RGB space tends to focus on the overall image information while neglecting the perception of traits at different granularities, such as eye gaze and lip shape, which are more detailed, or posture and gestures, which are more macroscopic. Exploring the efficient fusion of visual information of different granularities is crucial for accurate sign language recognition. In addition, applying a vanilla Transformer to sequence modeling in cSLR exhibits weak performance because specific video frames could interfere with the attention mechanism. These limitations constrain the capability to understand potential semantic characteristics. We introduce a feature fusion method for integrating visual features of disparate granularities and refine the metric of attention to enhance the Transformer’s comprehension of video content. Specifically, we extract CNN feature maps with varying receptive fields and employ a self-attention mechanism to fuse feature maps of different granularities, thereby obtaining multi-scale spatial features of the sign language framework. As for video modeling, we first analyze why the vanilla Transformer failed in cSLR and observe that the magnitude of the feature vectors of video frames could interfere with the distribution of attention weights. Therefore, we utilize the Euclidean distance among vectors to measure the attention weights instead of scaled-dot to enhance dynamic temporal modeling capabilities. Finally, we integrate the two components to construct the model MSF-ET (Multi-Scaled feature Fusion–Euclidean Transformer) for cSLR and train the model end-to-end. We perform experiments on large-scale cSLR benchmarks—PHOENIX-2014 and Chinese Sign Language (CSL)—to validate the effectiveness. Full article
Show Figures

Figure 1

Figure 1
<p>Currently, most video spatial representation methods for cSLR extract features by pre-trained CNN backbone networks ((<b>left</b>) in the figure). Although the approach can extract high-level semantic information, it lacks perception of details, such as mouth shape and gaze, which are important for understanding sign language. We propose a multi-scale feature fusion method based on self-attention mechanism ((<b>right</b>) in figure), which enables more comprehensive extraction of semantic information.</p>
Full article ">Figure 2
<p>Overall model architecture. Our proposed MSF-ET model consists of three main components: spatial encoder, feature fusion module, and temporal encoder. The spatial encoder is composed of multiple 2D convolutional layers, followed by max-pooling to downsample the feature maps with different receptive fields. The feature fusion module uses a self-attention mechanism to fuse the multi-scaled features of the frames. The temporal encoder is composed of the encoder based on Euclidean distance self-attention model and local Transformer layer. The encoder learns the contextual information of the video and the local features for glosses alignment. Finally, connectionist temporal classification (CTC) is used to train the model and decode the gloss sequences.</p>
Full article ">Figure 3
<p>Multi-scaled features integration and fusion. The spatial encoder outputs feature maps of sizes 3 and 7, respectively. These feature maps are first flattened into 1D vectors. Then, a special token <math display="inline"><semantics> <mrow> <mo>[</mo> <mi>c</mi> <mi>l</mi> <mi>s</mi> <mo>]</mo> </mrow> </semantics></math> is added to the head of the vector, similar to ViT. Next, the flattened vectors are added with trainable position embedding and then utilize the Transformer encoder to obtain the global context information of both feature maps. Finally, the two <math display="inline"><semantics> <mrow> <mo>[</mo> <mi>c</mi> <mi>l</mi> <mi>s</mi> <mo>]</mo> </mrow> </semantics></math> are concatenated to achieve multi-scale feature fusion.</p>
Full article ">Figure 4
<p>The demo for the attention map of vanilla Transformer. The heatmap denotes the attention scores, and the bar above the heatmap is the magnitude of key vectors. The figure indicates that the distribution of attention weights is overly concentrated in regions where the key vectors have longer magnitudes, thereby drowning out information from other positions and hindering the Transformer’s ability to fully comprehend the global information within the sequence.</p>
Full article ">Figure 5
<p>The detail about self-attention with Euclidean distance and local window. We assume that the window size is 5. Therefore, every frame interacts with others by attention mechanism in the window centered on itself.</p>
Full article ">Figure 6
<p>The CAM visualization of attention weights corresponding to feature maps at different scales. We applied this visualization to videos of different sign language performers to demonstrate the generalizability of the results. ((<b>A</b>) is sourced from ‘01April_2010_Thursday_heute_default-1’ in the PHOENIX2014 validation set. (<b>B</b>) is sourced from ‘03November_2010_Wednesday_tagesschau_default-7’ in the PHOENIX-2014 validation set.)</p>
Full article ">Figure 7
<p>An example of attention weight visualization for a Transformer utilizing Euclidean distance-based metrics. The sample data used in this figure are consistent with those in <a href="#applsci-14-08937-f004" class="html-fig">Figure 4</a>. It is evident that the use of Euclidean distance significantly alleviates the phenomenon of attention sparsity.</p>
Full article ">Figure 8
<p>Relationship between inference time and video sequence length during model inference.</p>
Full article ">
13 pages, 2999 KiB  
Article
Generative AI-Driven Data Augmentation for Crack Detection in Physical Structures
by Jinwook Kim, Joonho Seon, Soohyun Kim, Youngghyu Sun, Seongwoo Lee, Jeongho Kim, Byungsun Hwang and Jinyoung Kim
Electronics 2024, 13(19), 3905; https://doi.org/10.3390/electronics13193905 - 2 Oct 2024
Viewed by 533
Abstract
The accurate segmentation of cracks in structural materials is crucial for assessing the safety and durability of infrastructure. Although conventional segmentation models based on deep learning techniques have shown impressive detection capabilities in these tasks, their performance can be restricted by small amounts [...] Read more.
The accurate segmentation of cracks in structural materials is crucial for assessing the safety and durability of infrastructure. Although conventional segmentation models based on deep learning techniques have shown impressive detection capabilities in these tasks, their performance can be restricted by small amounts of training data. Data augmentation techniques have been proposed to mitigate the data availability issue; however, these systems often have limitations in texture diversity, scalability over multiple physical structures, and the need for manual annotation. In this paper, a novel generative artificial intelligence (GAI)-driven data augmentation framework is proposed to overcome these limitations by integrating a projected generative adversarial network (ProjectedGAN) and a multi-crack texture transfer generative adversarial network (MCT2GAN). Additionally, a novel metric is proposed to evaluate the quality of the generated data. The proposed method is evaluated using three datasets: the bridge crack library (BCL), DeepCrack, and Volker. From the simulation results, it is confirmed that the segmentation performance can be improved by the proposed method in terms of intersection over union (IoU) and Dice scores across three datasets. Full article
(This article belongs to the Special Issue Generative AI and Its Transformative Potential)
Show Figures

Figure 1

Figure 1
<p>Process of generating a synthesized crack dataset in the proposed dataset synthesis framework.</p>
Full article ">Figure 2
<p>Structure of ProjectedGAN.</p>
Full article ">Figure 3
<p>Structure of MCT2GAN.</p>
Full article ">Figure 4
<p>Process of quality evaluation.</p>
Full article ">Figure 5
<p>Crack images generated from GAN-based methods.</p>
Full article ">Figure 6
<p>Performance metrics of generated crack image sample in training process.</p>
Full article ">Figure 7
<p>Samples of the generated crack images in the training process. The images in the first row are the synthesized crack images. The images in the second row are the overlapped images between the ground truth and the predicted mask of the synthesized image from the pre-trained FCN model.</p>
Full article ">
Back to TopTop