[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,472)

Search Parameters:
Keywords = ResNet

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 33522 KiB  
Article
Multiloss Joint Gradient Control Knowledge Distillation for Image Classification
by Wei He, Jianchen Pan, Jianyu Zhang, Xichuan Zhou, Jialong Liu, Xiaoyu Huang and Yingcheng Lin
Electronics 2024, 13(20), 4102; https://doi.org/10.3390/electronics13204102 (registering DOI) - 17 Oct 2024
Abstract
Knowledge distillation (KD) techniques aim to transfer knowledge from complex teacher neural networks to simpler student networks. In this study, we propose a novel knowledge distillation method called Multiloss Joint Gradient Control Knowledge Distillation (MJKD), which functions by effectively combining feature- and logit-based [...] Read more.
Knowledge distillation (KD) techniques aim to transfer knowledge from complex teacher neural networks to simpler student networks. In this study, we propose a novel knowledge distillation method called Multiloss Joint Gradient Control Knowledge Distillation (MJKD), which functions by effectively combining feature- and logit-based knowledge distillation methods with gradient control. The proposed knowledge distillation method discretely considers the gradients of the task loss (cross-entropy loss), feature distillation loss, and logit distillation loss. The experimental results suggest that logits may contain more information and should, consequently, be assigned greater weight during the gradient update process in this work. The empirical findings on the CIFAR-100 and Tiny-ImageNet datasets indicate that MJKD generally outperforms traditional knowledge distillation methods, significantly enhancing the generalization ability and classification accuracy of student networks. For instance, MJKD achieves a 63.53% accuracy on Tiny-ImageNet for the ResNet18 MobileNetV2 pair. Furthermore, we present visualizations and analyses to explore its potential working mechanisms. Full article
(This article belongs to the Special Issue Knowledge Information Extraction Research)
Show Figures

Figure 1

Figure 1
<p>The figure illustrates the concept of knowledge distillation [<a href="#B3-electronics-13-04102" class="html-bibr">3</a>] alongside our proposed Multiloss Joint Gradient Control Knowledge Distillation (MJKD) approach. In MJKD, the gradients associated with the task loss, logit distillation loss, and feature distillation loss are computed independently and subsequently utilized to update their respective momentum buffers.</p>
Full article ">Figure 2
<p>Task loss on CIFAR-100 (<b>a</b>) and Tiny-ImageNet (<b>b</b>).</p>
Full article ">Figure 3
<p>Distillation Loss on CIFAR-100 (<b>a</b>,<b>c</b>) and Tiny-ImageNet (<b>b</b>,<b>d</b>).</p>
Full article ">Figure 4
<p>Illustration of loss for weighing <math display="inline"><semantics> <mi>α</mi> </semantics></math> on CIFAR-100 and Tiny-ImageNet.</p>
Full article ">Figure 5
<p>Loss landscapes of (<b>a</b>) KD, (<b>b</b>) DKD, and (<b>c</b>) MJKD on the Tiny-ImageNet dataset.</p>
Full article ">Figure 6
<p>Difference in the correlation matrices of student and teacher logits on the Tiny-ImageNet dataset.</p>
Full article ">
20 pages, 2235 KiB  
Article
Efficient Ensemble Adversarial Attack for a Deep Neural Network (DNN)-Based Unmanned Aerial Vehicle (UAV) Vision System
by Zhun Zhang, Qihe Liu, Shijie Zhou, Wenqi Deng, Zhewei Wu and Shilin Qiu
Drones 2024, 8(10), 591; https://doi.org/10.3390/drones8100591 - 17 Oct 2024
Abstract
In recent years, unmanned aerial vehicles (UAVs) vision systems based on deep neural networks (DNNs) have made remarkable advancements, demonstrating impressive performance. However, due to the inherent characteristics of DNNs, these systems have become increasingly vulnerable to adversarial attacks. Traditional black-box attack methods [...] Read more.
In recent years, unmanned aerial vehicles (UAVs) vision systems based on deep neural networks (DNNs) have made remarkable advancements, demonstrating impressive performance. However, due to the inherent characteristics of DNNs, these systems have become increasingly vulnerable to adversarial attacks. Traditional black-box attack methods typically require a large number of queries to generate adversarial samples successfully. In this paper, we propose a novel adversarial attack technique designed to achieve efficient black-box attacks with a minimal number of queries. We define a perturbation generator that first decomposes the image into four frequency bands using wavelet decomposition and then searches for adversarial perturbations across these bands by minimizing a weighted loss function on a set of fixed surrogate models. For the target victim model, the perturbation images generated by the perturbation generator are used to query and update the weights in the loss function, as well as the weights for different frequency bands. Experimental results show that, compared to state-of-the-art methods on various image classifiers trained on ImageNet (such as VGG-19, DenseNet-121, and ResNext-50), our method achieves a success rate over 98% for targeted attacks and nearly a 100% success rate for non-targeted attacks with only 1–2 queries per image. Full article
26 pages, 1656 KiB  
Article
Secure and Transparent Lung and Colon Cancer Classification Using Blockchain and Microsoft Azure
by Entesar Hamed I. Eliwa, Amr Mohamed El Koshiry, Tarek Abd El-Hafeez and Ahmed Omar
Adv. Respir. Med. 2024, 92(5), 395-420; https://doi.org/10.3390/arm92050037 - 17 Oct 2024
Abstract
Background: The global healthcare system faces challenges in diagnosing and managing lung and colon cancers, which are significant health burdens. Traditional diagnostic methods are inefficient and prone to errors, while data privacy and security concerns persist. Objective: This study aims to develop a [...] Read more.
Background: The global healthcare system faces challenges in diagnosing and managing lung and colon cancers, which are significant health burdens. Traditional diagnostic methods are inefficient and prone to errors, while data privacy and security concerns persist. Objective: This study aims to develop a secure and transparent framework for remote consultation and classification of lung and colon cancer, leveraging blockchain technology and Microsoft Azure cloud services. Dataset and Features: The framework utilizes the LC25000 dataset, containing 25,000 histopathological images, for training and evaluating advanced machine learning models. Key features include secure data upload, anonymization, encryption, and controlled access via blockchain and Azure services. Methods: The proposed framework integrates Microsoft Azure’s cloud services with a permissioned blockchain network. Patients upload CT scans through a mobile app, which are then preprocessed, anonymized, and stored securely in Azure Blob Storage. Blockchain smart contracts manage data access, ensuring only authorized specialists can retrieve and analyze the scans. Azure Machine Learning is used to train and deploy state-of-the-art machine learning models for cancer classification. Evaluation Metrics: The framework’s performance is evaluated using metrics such as accuracy, precision, recall, and F1-score, demonstrating the effectiveness of the integrated approach in enhancing diagnostic accuracy and data security. Results: The proposed framework achieves an impressive accuracy of 100% for lung and colon cancer classification using DenseNet, ResNet50, and MobileNet models with different split ratios (70–30, 80–20, 90–10). The F1-score and k-fold cross-validation accuracy (5-fold and 10-fold) also demonstrate exceptional performance, with values exceeding 99.9%. Real-time notifications and secure remote consultations enhance the efficiency and transparency of the diagnostic process, contributing to better patient outcomes and streamlined cancer care management. Full article
18 pages, 4387 KiB  
Article
Enhanced Image-Based Malware Classification Using Transformer-Based Convolutional Neural Networks (CNNs)
by Moses Ashawa, Nsikak Owoh, Salaheddin Hosseinzadeh and Jude Osamor
Electronics 2024, 13(20), 4081; https://doi.org/10.3390/electronics13204081 - 17 Oct 2024
Abstract
As malware samples grow in complexity and employ advanced evasion techniques, traditional detection methods are insufficient for accurately classifying large volumes of sophisticated malware variants. To address this issue, image-based malware classification techniques leveraging machine learning algorithms have been developed as a more [...] Read more.
As malware samples grow in complexity and employ advanced evasion techniques, traditional detection methods are insufficient for accurately classifying large volumes of sophisticated malware variants. To address this issue, image-based malware classification techniques leveraging machine learning algorithms have been developed as a more optimal solution to this challenge. However, accurately classifying content distribution-based features with unique pixel intensities from grayscale images remains a challenge. This paper proposes an enhanced image-based malware classification system using convolutional neural networks (CNNs) using ResNet-152 and vision transformer (ViT). The two architectures are then compared to determine their classification abilities. A total of 6137 benign files and 9861 malicious executables are converted from text files to unsigned integers and then to images. The ViT examined unsigned integers as pixel values, while ResNet-152 converted the pixel values into floating points for classification. The result of the experiments demonstrates a high-performance accuracy of 99.62% with effective hyperparameters of 10-fold cross-validation. The findings indicate that the proposed model is capable of being implemented in dynamic and complex malware environments, achieving a practical computational efficiency of 47.2 s for the identification and classification of new malware samples. Full article
Show Figures

Figure 1

Figure 1
<p>Showing how the virtual machines are configured to store the executable files.</p>
Full article ">Figure 2
<p>Feature extraction and conversion process.</p>
Full article ">Figure 3
<p>State machine malware representation.</p>
Full article ">Figure 4
<p>Rendering of malware sample image. (<b>a</b>) Pictures integrated in the malware sample and (<b>b</b>) Images of malware that share similarities across various malware categories.</p>
Full article ">Figure 5
<p>The summary of the architectures of the proposed enhanced image-based malware classification model.</p>
Full article ">Figure 6
<p>Training at zero iterations.</p>
Full article ">Figure 7
<p>Image classification at different iterations. (<b>a</b>) 2000th count, (<b>b</b>) 4000th count (<b>c</b>) 6000th count, (<b>d</b>) 8000th count, (<b>e</b>) 10,000th count, (<b>f</b>) 12,000th count, (<b>g</b>) 14,000th count, (<b>h</b>) 16,000th count and (<b>i</b>) 18000th count.</p>
Full article ">Figure 7 Cont.
<p>Image classification at different iterations. (<b>a</b>) 2000th count, (<b>b</b>) 4000th count (<b>c</b>) 6000th count, (<b>d</b>) 8000th count, (<b>e</b>) 10,000th count, (<b>f</b>) 12,000th count, (<b>g</b>) 14,000th count, (<b>h</b>) 16,000th count and (<b>i</b>) 18000th count.</p>
Full article ">Figure 8
<p>Image SOM visualization. The gray color shows that there is no specific categorization of the image pixel intensity.</p>
Full article ">Figure 9
<p>Image sieve diagram visualization for the sample space showing benign and malicious classes.</p>
Full article ">Figure 10
<p>PCA dimensionality reduction. (<b>a</b>) First component, (<b>b</b>) second component, (<b>c</b>) third component, and (<b>d</b>) fourth component.</p>
Full article ">Figure 10 Cont.
<p>PCA dimensionality reduction. (<b>a</b>) First component, (<b>b</b>) second component, (<b>c</b>) third component, and (<b>d</b>) fourth component.</p>
Full article ">Figure 11
<p>Heatmap showing regions of interest as classified by the model. (<b>a</b>) Image cluster showing malware names and their textual cluster classifications. (<b>b</b>) Image cluster showing malware activities and their score clusters.</p>
Full article ">Figure 12
<p>Visualized malware images based on their malware families. (<b>a</b>) QakBot, (<b>b</b>) Gamarue, (<b>c</b>) Sodinokibi, and (<b>d</b>) Ryuk.</p>
Full article ">
21 pages, 8536 KiB  
Article
Early Detection of Lumpy Skin Disease in Cattle Using Deep Learning—A Comparative Analysis of Pretrained Models
by Chamirti Senthilkumar, Sindhu C, G. Vadivu and Suresh Neethirajan
Vet. Sci. 2024, 11(10), 510; https://doi.org/10.3390/vetsci11100510 - 17 Oct 2024
Viewed by 269
Abstract
Lumpy Skin Disease (LSD) poses a significant threat to agricultural economies, particularly in livestock-dependent countries like India, due to its high transmission rate leading to severe morbidity and mortality among cattle. This underscores the urgent need for early and accurate detection to effectively [...] Read more.
Lumpy Skin Disease (LSD) poses a significant threat to agricultural economies, particularly in livestock-dependent countries like India, due to its high transmission rate leading to severe morbidity and mortality among cattle. This underscores the urgent need for early and accurate detection to effectively manage and mitigate outbreaks. Leveraging advancements in computer vision and artificial intelligence, our research develops an automated system for LSD detection in cattle using deep learning techniques. We utilized two publicly available datasets comprising images of healthy cattle and those with LSD, including additional images of cattle affected by other diseases to enhance specificity and ensure the model detects LSD specifically rather than general illness signs. Our methodology involved preprocessing the images, applying data augmentation, and balancing the datasets to improve model generalizability. We evaluated over ten pretrained deep learning models—Xception, VGG16, VGG19, ResNet152V2, InceptionV3, MobileNetV2, DenseNet201, NASNetMobile, NASNetLarge, and EfficientNetV2S—using transfer learning. The models were rigorously trained and tested under diverse conditions, with performance assessed using metrics such as accuracy, sensitivity, specificity, precision, F1-score, and AUC-ROC. Notably, VGG16 and MobileNetV2 emerged as the most effective, achieving accuracies of 96.07% and 96.39%, sensitivities of 93.75% and 98.57%, and specificities of 97.14% and 94.59%, respectively. Our study critically highlights the strengths and limitations of each model, demonstrating that while high accuracy is achievable, sensitivity and specificity are crucial for clinical applicability. By meticulously detailing the performance characteristics and including images of cattle with other diseases, we ensured the robustness and reliability of the models. This comprehensive comparative analysis enriches our understanding of deep learning applications in veterinary diagnostics and makes a substantial contribution to the field of automated disease recognition in livestock farming. Our findings suggest that adopting such AI-driven diagnostic tools can enhance the early detection and control of LSD, ultimately benefiting animal health and the agricultural economy. Full article
Show Figures

Figure 1

Figure 1
<p>A cow exhibiting characteristic symptoms of Lumpy Skin Disease (LSD), including raised nodules on the skin. These visible signs are critical for early detection and diagnosis, which is further enhanced by the application of deep learning models in this study. The automated detection of such lesions through advanced image analysis techniques can significantly improve the accuracy and speed of LSD identification in bovine populations.</p>
Full article ">Figure 2
<p>Working principle of transfer learning in Lumpy Skin Disease detection.</p>
Full article ">Figure 3
<p>Accuracy and loss graphs for the Xception model on Dataset 1 (<b>left</b>) and Dataset 2 (<b>right</b>).</p>
Full article ">Figure 4
<p>Performance metrics for the VGG16 model on Dataset 1, including accuracy, loss, confusion matrix, and ROC curve.</p>
Full article ">Figure 5
<p>Performance metrics for the VGG16 model on Dataset 2, including accuracy, loss, confusion matrix, and ROC curve.</p>
Full article ">Figure 6
<p>Accuracy and loss graphs for the VGG19 model on Dataset 1 (<b>left</b>) and Dataset 2 (<b>right</b>).</p>
Full article ">Figure 7
<p>Accuracy and loss graphs for the ResNet152V2 model on Dataset 1 (<b>left</b>) and Dataset 2 (<b>right</b>).</p>
Full article ">Figure 8
<p>Accuracy and loss graphs for the InceptionV3 model on Dataset 1 (<b>left</b>) and Dataset 2 (<b>right</b>).</p>
Full article ">Figure 9
<p>Performance metrics for the MobileNetV2 model on Dataset 1, including accuracy, loss, confusion matrix, and ROC curve.</p>
Full article ">Figure 10
<p>Performance metrics for the MobileNetV2 model on Dataset 2, including accuracy, loss, confusion matrix, and ROC curve.</p>
Full article ">Figure 11
<p>Accuracy and loss graphs for the DenseNet201 model on Dataset 1 (<b>left</b>) and Dataset 2 (<b>right</b>).</p>
Full article ">Figure 12
<p>Accuracy and loss graphs for the NASNetMobile model on Dataset 1 (<b>left</b>) and Dataset 2 (<b>right</b>).</p>
Full article ">Figure 13
<p>Accuracy and loss graphs for the NASNetLarge model on Dataset 1 (<b>left</b>) and Dataset 2 (<b>right</b>).</p>
Full article ">Figure 14
<p>Accuracy and loss graphs for the EfficientNetV2S model on Dataset 1 (<b>left</b>) and Dataset 2 (<b>right</b>).</p>
Full article ">Figure 15
<p>Comparative performance analysis of all 10 models on Dataset 1.</p>
Full article ">Figure 16
<p>Comparative performance analysis of all 10 models on Dataset 2.</p>
Full article ">
18 pages, 5058 KiB  
Article
Measuring Student Engagement through Behavioral and Emotional Features Using Deep-Learning Models
by Nasir Mahmood, Sohail Masood Bhatti, Hussain Dawood, Manas Ranjan Pradhan and Haseeb Ahmad
Algorithms 2024, 17(10), 458; https://doi.org/10.3390/a17100458 (registering DOI) - 16 Oct 2024
Viewed by 280
Abstract
Students’ behavioral and emotional engagement in the classroom environment may reflect the students’ learning experience and subsequent educational outcomes. The existing research has overlooked the measurement of behavioral and emotional engagement in an offline classroom environment with more students, and it has not [...] Read more.
Students’ behavioral and emotional engagement in the classroom environment may reflect the students’ learning experience and subsequent educational outcomes. The existing research has overlooked the measurement of behavioral and emotional engagement in an offline classroom environment with more students, and it has not measured the student engagement level in an objective sense. This work aims to address the limitations of the existing research and presents an effective approach to measure students’ behavioral and emotional engagement and the student engagement level in an offline classroom environment during a lecture. More precisely, video data of 100 students during lectures in different offline classes were recorded and pre-processed to extract frames with individual students. For classification, convolutional-neural-network- and transfer-learning-based models including ResNet50, VGG16, and Inception V3 were trained, validated, and tested. First, behavioral engagement was computed using salient features, for which the self-trained CNN classifier outperformed with a 97%, 91%, and 83% training, validation, and testing accuracy, respectively. Subsequently, the emotional engagement of the behaviorally engaged students was computed, for which the ResNet50 model surpassed the others with a 95%, 90%, and 82% training, validation, and testing accuracy, respectively. Finally, a novel student engagement level metric is proposed that incorporates behavioral and emotional engagement. The proposed approach may provide support for improving students’ learning in an offline classroom environment and devising effective pedagogical policies. Full article
(This article belongs to the Special Issue Algorithms for Feature Selection (2nd Edition))
Show Figures

Figure 1

Figure 1
<p>Proposed methodology.</p>
Full article ">Figure 2
<p>Behavioral reflecting frames.</p>
Full article ">Figure 3
<p>Emotion-reflecting frames.</p>
Full article ">Figure 4
<p>Proposed CNN architecture for measuring behavior level.</p>
Full article ">Figure 5
<p>Proposed ResNet50 architecture for measuring emotion level.</p>
Full article ">Figure 6
<p>Intra-comparison of behavior detection models’ training and validation accuracies and training and validation losses: (<b>a</b>) CNN; (<b>b</b>) VGG16; (<b>c</b>) ResNet50; and (<b>d</b>) Inception V3.</p>
Full article ">Figure 7
<p>Inter-comparison of behavior detection models: (<b>a</b>) training accuracy; (<b>b</b>) training loss; (<b>c</b>) validation accuracy; and (<b>d</b>) validation loss.</p>
Full article ">Figure 8
<p>Intra-comparison of emotion detection models’ training and validation accuracies and training and validation losses: (<b>a</b>) ResNet50; (<b>b</b>) CNN; (<b>c</b>) VGG16; and (<b>d</b>) Inception V3.</p>
Full article ">Figure 9
<p>Inter-comparison of emotion detection models: (<b>a</b>) training accuracy; (<b>b</b>) training loss; (<b>c</b>) validation accuracy; and (<b>d</b>) validation loss.</p>
Full article ">Figure 10
<p>Student engagement level computation.</p>
Full article ">
22 pages, 1174 KiB  
Article
Dual Stream Encoder–Decoder Architecture with Feature Fusion Model for Underwater Object Detection
by Mehvish Nissar, Amit Kumar Mishra and Badri Narayan Subudhi
Mathematics 2024, 12(20), 3227; https://doi.org/10.3390/math12203227 (registering DOI) - 15 Oct 2024
Viewed by 343
Abstract
Underwater surveillance is an imminent and fascinating exploratory domain, particularly in monitoring aquatic ecosystems. This field offers valuable insights into underwater behavior and activities, which have broad applications across various domains. Specifically, underwater surveillance involves detecting and tracking moving objects within aquatic environments. [...] Read more.
Underwater surveillance is an imminent and fascinating exploratory domain, particularly in monitoring aquatic ecosystems. This field offers valuable insights into underwater behavior and activities, which have broad applications across various domains. Specifically, underwater surveillance involves detecting and tracking moving objects within aquatic environments. However, the complex properties of water make object detection a challenging task. Background subtraction is a commonly employed technique for detecting local changes in video scenes by segmenting images into the background and foreground to isolate the object of interest. Within this context, we propose an innovative dual-stream encoder–decoder framework based on the VGG-16 and ResNet-50 models for detecting moving objects in underwater frames. The network includes a feature fusion module that effectively extracts multiple-level features. Using a limited set of images and performing training in an end-to-end manner, the proposed framework yields accurate results without post-processing. The efficacy of the proposed technique is confirmed through visual and quantitative comparisons with eight cutting-edge methods using two standard databases. The first one employed in our experiments is the Underwater Change Detection Dataset, which includes five challenges, each challenge comprising approximately 1000 frames. The categories in this dataset were recorded under various underwater conditions. The second dataset used for practical analysis is the Fish4Knowledge dataset, where we considered five challenges. Each category, recorded in different aquatic settings, contains a varying number of frames, typically exceeding 1000 per category. Our proposed method surpasses all methods used for comparison by attaining an average F-measure of 0.98 on the Underwater Change Detection Dataset and 0.89 on the Fish4Knowledge dataset. Full article
(This article belongs to the Section Mathematics and Computer Science)
Show Figures

Figure 1

Figure 1
<p>Proposed architecture for underwater object detection.</p>
Full article ">Figure 2
<p>Framework of feature fusion module.</p>
Full article ">Figure 3
<p>Visual examples of different challenges present in Underwater Change Detection and Fish4Knowledge datasets.</p>
Full article ">Figure 4
<p>Visual analysis of object detection in Underwater Change Detection Dataset: (<b>a</b>) actual image (<b>b</b>) ground-truth, outcomes obtained by (<b>c</b>) GSMM [<a href="#B69-mathematics-12-03227" class="html-bibr">69</a>], (<b>d</b>) AGMM [<a href="#B9-mathematics-12-03227" class="html-bibr">9</a>], (<b>e</b>) ABMM [<a href="#B70-mathematics-12-03227" class="html-bibr">70</a>], (<b>f</b>) ADE [<a href="#B10-mathematics-12-03227" class="html-bibr">10</a>], (<b>g</b>) GWFT [<a href="#B71-mathematics-12-03227" class="html-bibr">71</a>], (<b>h</b>) MFPD [<a href="#B72-mathematics-12-03227" class="html-bibr">72</a>], (<b>i</b>) GAFP [<a href="#B39-mathematics-12-03227" class="html-bibr">39</a>], (<b>j</b>) HPPM [<a href="#B40-mathematics-12-03227" class="html-bibr">40</a>], and (<b>k</b>) proposed method (PM).</p>
Full article ">Figure 5
<p>Visual analysis of object detection in Fish4Knowledge dataset: (<b>a</b>) original image (<b>b</b>) ground-truth, results obtained by (<b>c</b>) GSMM [<a href="#B69-mathematics-12-03227" class="html-bibr">69</a>], (<b>d</b>) AGMM [<a href="#B9-mathematics-12-03227" class="html-bibr">9</a>], (<b>e</b>) ABMM [<a href="#B70-mathematics-12-03227" class="html-bibr">70</a>], (<b>f</b>) ADE [<a href="#B10-mathematics-12-03227" class="html-bibr">10</a>], (<b>g</b>) GWFT [<a href="#B71-mathematics-12-03227" class="html-bibr">71</a>], (<b>h</b>) MFPD [<a href="#B72-mathematics-12-03227" class="html-bibr">72</a>], (<b>i</b>) GAFP [<a href="#B39-mathematics-12-03227" class="html-bibr">39</a>], (<b>j</b>) HPPM [<a href="#B40-mathematics-12-03227" class="html-bibr">40</a>], and (<b>k</b>) proposed method (PM).</p>
Full article ">
16 pages, 1563 KiB  
Article
Tree Species Classification from UAV Canopy Images with Deep Learning Models
by Yunmei Huang, Botong Ou, Kexin Meng, Baijian Yang, Joshua Carpenter, Jinha Jung and Songlin Fei
Remote Sens. 2024, 16(20), 3836; https://doi.org/10.3390/rs16203836 (registering DOI) - 15 Oct 2024
Viewed by 335
Abstract
Forests play a critical role in the provision of ecosystem services, and understanding their compositions, especially tree species, is essential for effective ecosystem management and conservation. However, identifying tree species is challenging and time-consuming. Recently, unmanned aerial vehicles (UAVs) equipped with various sensors [...] Read more.
Forests play a critical role in the provision of ecosystem services, and understanding their compositions, especially tree species, is essential for effective ecosystem management and conservation. However, identifying tree species is challenging and time-consuming. Recently, unmanned aerial vehicles (UAVs) equipped with various sensors have emerged as a promising technology for species identification due to their relatively low cost and high spatial and temporal resolutions. Moreover, the advancement of various deep learning models makes remote sensing based species identification more a reality. However, three questions remain to be answered: first, which of the state-of-the-art models performs best for this task; second, which is the optimal season for tree species classification in a temperate forest; and third, whether a model trained in one season can be effectively transferred to another season. To address these questions, we focus on tree species classification by using five state-of-the-art deep learning models on UAV-based RGB images, and we explored the model transferability between seasons. Utilizing UAV images taken in the summer and fall, we captured 8799 crown images of eight species. We trained five models using summer and fall images and compared their performance on the same dataset. All models achieved high performances in species classification, with the best performance on summer images, with an average F1-score was 0.96. For the fall images, Vision Transformer (ViT), EfficientNetB0, and YOLOv5 achieved F1-scores greater than 0.9, outperforming both ResNet18 and DenseNet. On average, across the two seasons, ViT achieved the best accuracy. This study demonstrates the capability of deep learning models in forest inventory, particularly for tree species classification. While the choice of certain models may not significantly affect performance when using summer images, the advanced models prove to be a better choice for fall images. Given the limited transferability from one season to another, further research is required to overcome the challenge associated with transferability across seasons. Full article
(This article belongs to the Special Issue LiDAR Remote Sensing for Forest Mapping)
Show Figures

Figure 1

Figure 1
<p>Work pipeline for tree species classification with UAV images and deep learning models.</p>
Full article ">Figure 2
<p>Study area and label examples. (<b>a</b>) Martell Forest in Indiana, USA; (<b>b</b>) Canopy image of a black cherry (<span class="html-italic">Prunus serotina</span>) plantation; (<b>c</b>) Label examples of the black cherry plantation (all crowns were identified with bounding boxes).</p>
Full article ">Figure 3
<p>Examples of seasonal differences among eight species. These crown images are cropped from orthophotos of our study area and show the crown variation of the same trees.</p>
Full article ">Figure 4
<p>F1-scores of five models for summer and seasons.</p>
Full article ">Figure A1
<p>Number of images VS F1-score on summer datasets for four species with ResNet18. Due to the limits of numbers of images, we selected four species and the number of training images starts from 60 to 280, with 20 as an increment. In each training session, all four classes have an equal amount of training images and the same test dataset. Thus, we trained ResNet18 12 times with various numbers of images. When the number of images ranges from 60 to 180, the increment of accuracy is faster than the further part image numbers ranging from 200 to 280. For the experiments on images with the numbers 260 and 280, their change in accuracy was unremarkable. Hence, from our observation, the number of images impacts the model’s classification accuracy, and after training images reach a certain amount, the influences decrease.</p>
Full article ">Figure A2
<p>Number of images VS F1-scores with ResNet18 on two datasets for eight species. Different shapes of points stand for different seasons. The round shape points stand for the summer dataset, the squares belong to fall.</p>
Full article ">
19 pages, 4794 KiB  
Article
An Efficient Ensemble Approach for Brain Tumors Classification Using Magnetic Resonance Imaging
by Zubair Saeed, Tarraf Torfeh, Souha Aouadi, (Jim) Xiuquan Ji and Othmane Bouhali
Information 2024, 15(10), 641; https://doi.org/10.3390/info15100641 (registering DOI) - 15 Oct 2024
Viewed by 287
Abstract
Tumors in the brain can be life-threatening, making early and precise detection crucial for effective treatment and improved patient outcomes. Deep learning (DL) techniques have shown significant potential in automating the early diagnosis of brain tumors by analyzing magnetic resonance imaging (MRI), offering [...] Read more.
Tumors in the brain can be life-threatening, making early and precise detection crucial for effective treatment and improved patient outcomes. Deep learning (DL) techniques have shown significant potential in automating the early diagnosis of brain tumors by analyzing magnetic resonance imaging (MRI), offering a more efficient and accurate approach to classification. Deep convolutional neural networks (DCNNs), which are a sub-field of DL, have the potential to analyze rapidly and accurately MRI data and, as such, assist human radiologists, facilitating quicker diagnoses and earlier treatment initiation. This study presents an ensemble of three high-performing DCNN models, i.e., DenseNet169, EfficientNetB0, and ResNet50, for accurate classification of brain tumors and non-tumor MRI samples. Our proposed ensemble model demonstrates significant improvements over various evaluation parameters compared to individual state-of-the-art (SOTA) DCNN models. We implemented ten SOTA DCNN models, i.e., EfficientNetB0, ResNet50, DenseNet169, DenseNet121, SqueezeNet, ResNet34, ResNet18, VGG16, VGG19, and LeNet5, and provided a detailed performance comparison. We evaluated these models using two learning rates (LRs) of 0.001 and 0.0001 and two batch sizes (BSs) of 64 and 128 and identified the optimal hyperparameters for each model. Our findings indicate that the ensemble approach outperforms individual models, having 92% accuracy, 90% precision, 92% recall, and an F1 score of 91% at a 64 BS and 0.0001 LR. This study not only highlights the superior performance of the ensemble technique but also offers a comprehensive comparison with the latest research. Full article
(This article belongs to the Special Issue Detection and Modelling of Biosignals)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Block diagram of our proposed methodology.</p>
Full article ">Figure 2
<p>Dataset samples (<b>a</b>) No tumor, (<b>b</b>) Glioma, (<b>c</b>) Meningioma, and (<b>d</b>) Pituitary.</p>
Full article ">Figure 3
<p>Overall dataset distribution against each class.</p>
Full article ">Figure 4
<p>(<b>a</b>) Accuracy, (<b>b</b>) Precision, (<b>c</b>) Recall, (<b>d</b>) F1-Score of Proposed Ensemble Model using 64 BS and 0.0001 LS, EfficientNetB0 using 64 BS and 0.0001, ResNet59 using 64 BS and 0.0001, DenseNet169, DenseNet121 using 64 BS and 0.001, SqueezeNet using 64 BS and 0.001, ResNet34 using 64 BS and 0.001, ResNet18 using 64 BS and 0.001, VGG16 using 128 BS and 0.001, VGG18 using 128 BS and 0.001, LeNet5 using 64 BS and 0.001.</p>
Full article ">Figure 5
<p>Training and Validation plots of (<b>a</b>) Proposed Ensemble Model using 64 BS and 0.0001 LS, (<b>b</b>) EfficientNetB0 using 64 BS and 0.0001, (<b>c</b>) ResNet59 using 64 BS and 0.0001, (<b>d</b>) DenseNet169 using 64 BS and 0.001, (<b>e</b>) DenseNet121 using 64 BS and 0.001, (<b>f</b>) SqueezeNet using 64 BS and 0.001, (<b>g</b>) ResNet34 using 64 BS and 0.001, (<b>h</b>) ResNet18 using 64 BS and 0.001, (<b>i</b>) VGG16 using 128 BS and 0.001, (<b>j</b>) VGG18 using 128 BS and 0.001, (<b>k</b>) LeNet5 using 64 BS and 0.001.</p>
Full article ">Figure 6
<p>False predictions of Proposed Ensemble Technique.</p>
Full article ">
18 pages, 4549 KiB  
Article
A Transfer Learning-Based Framework for Classifying Lymph Node Metastasis in Prostate Cancer Patients
by Suryadipto Sarkar, Teresa Wu, Matthew Harwood and Alvin C. Silva
Biomedicines 2024, 12(10), 2345; https://doi.org/10.3390/biomedicines12102345 (registering DOI) - 15 Oct 2024
Viewed by 423
Abstract
Background: Prostate cancer is the second most common new cancer diagnosis in the United States. It is usually slow-growing, and when it is low-grade and confined to the prostate gland, it can be treated either conservatively (through active surveillance) or with surgery. However, [...] Read more.
Background: Prostate cancer is the second most common new cancer diagnosis in the United States. It is usually slow-growing, and when it is low-grade and confined to the prostate gland, it can be treated either conservatively (through active surveillance) or with surgery. However, if the cancer has spread beyond the prostate, such as to the lymph nodes, then that indicates a more aggressive cancer, and surgery may not be adequate. Methods: The challenge is that it is often difficult for radiologists reading prostate-specific imaging such as magnetic resonance images (MRIs) to differentiate malignant lymph nodes from non-malignant ones. An emerging field is the development of artificial intelligence (AI) models, including machine learning and deep learning, for medical imaging to assist in diagnostic tasks. Earlier research focused on implementing texture algorithms to extract imaging features used in classification models. More recently, researchers began studying the use of deep learning for both stand-alone feature extraction and end-to-end classification tasks. In order to tackle the challenges inherent in small datasets, this study was designed as a scalable hybrid framework utilizing pre-trained ResNet-18, a deep learning model, to extract features that were subsequently fed into a machine learning classifier to automatically identify malignant lymph nodes in patients with prostate cancer. For comparison, two texture algorithms were implemented, namely the gray-level co-occurrence matrix (GLCM) and Gabor. Results: Using an institutional prostate lymph node dataset (42 positives, 84 negatives), the proposed framework achieved an accuracy of 76.19%, a sensitivity of 79.76%, and a specificity of 69.05%. Using GLCM features, the classification achieved an accuracy of 61.90%, a sensitivity of 74.07%, and a specificity of 42.86%. Using Gabor features, the classification achieved an accuracy of 65.08%, a sensitivity of 73.47%, and a specificity of 52.50%. Conclusions: Our results demonstrate that a hybrid approach, i.e., using a pre-trainined deep learning model for feature extraction, followed by a machine learning classifier, is a viable solution. This hybrid approach is especially useful in medical-imaging-based applications with small datasets. Full article
(This article belongs to the Special Issue Advanced Cancer Diagnosis and Treatment: Second Edition)
Show Figures

Figure 1

Figure 1
<p>A schematic representation of the overall workflow: feature extraction, feature selection, and classification. It is to be noted that although five classifiers were initially used to differentiate “metastatic” and “normal” lymph nodes, only the decision tree classifier was retained for the final analyses because all of the other classifiers had sub-par performance. This makes sense because the feature selection framework utilized Random Forest, which is a boosted decision tree algorithm.</p>
Full article ">Figure 2
<p>ResNet18 model: it comprises a total of 71 layers, and the trained weights from the “average pooling” layers are used for classification.</p>
Full article ">Figure 3
<p>A pictorial explanation of the TNM cancer staging protocol as it pertains to prostate cancer.</p>
Full article ">Figure 4
<p>A prostate MRI from the same patient; four MRI sequences are shown.</p>
Full article ">Figure 5
<p>The distribution of normal and metastatic samples in each of the four MRI sequences, namely ADC, FRFSE, Water-GAD and Pelvis (T2 FatSat). The percentages in the inner circle represent the number of samples under each label–sequence combination as a percentage of the total number of samples present in the entire dataset, whereas the percentages in the outer circle represent the distribution of images across the four MRI sequences.</p>
Full article ">
13 pages, 5647 KiB  
Article
ResNet Modeling for 12 nm FinFET Devices to Enhance DTCO Efficiency
by Yiming Huang, Bin Li, Zhaohui Wu and Wenchao Liu
Electronics 2024, 13(20), 4040; https://doi.org/10.3390/electronics13204040 - 14 Oct 2024
Viewed by 346
Abstract
In this paper, a deep learning-based device modeling framework for design-technology co-optimization (DTCO) is proposed. A ResNet surrogate model is utilized as an alternative to traditional compact models, demonstrating high accuracy in both single-task (I–V or C–V) and multi-task (I–V and C–V) device [...] Read more.
In this paper, a deep learning-based device modeling framework for design-technology co-optimization (DTCO) is proposed. A ResNet surrogate model is utilized as an alternative to traditional compact models, demonstrating high accuracy in both single-task (I–V or C–V) and multi-task (I–V and C–V) device modeling. Moreover, transfer learning is applied to the ResNet model, using the BSIM-CMG compact model for a 12 nm FinFET SPICE model as the pre-trained source. Through this approach, superior modeling accuracy and faster training speed are achieved compared to a ResNet surrogate model initialized with random weights, thereby meeting the rapid and efficient demands of the DTCO process. The effectiveness of the ResNet surrogate model in circuit simulation for 12 nm FinFET devices is demonstrated. Full article
Show Figures

Figure 1

Figure 1
<p>Preprocessing of the dataset.</p>
Full article ">Figure 2
<p>ResNet structure.</p>
Full article ">Figure 3
<p>(<b>a</b>) Lintanh curves. (<b>b</b>) Lintanh’s first derivatives curves. (<b>c</b>) Lintanh’s second derivatives.</p>
Full article ">Figure 4
<p>Training and prediction process of the ResNet model.</p>
Full article ">Figure 5
<p>Comparison of ResNet model performance in 12 nm FinFET single-task (I–V) and multi-task (I–V and C–V) modeling with (<b>a</b>) training loss (AdaptiveSmoothL1); (<b>b</b>) validation loss (MAPE); and (<b>c</b>) test-set accuracy (R<sup>2</sup>).</p>
Full article ">Figure 6
<p>Comparison of ANN and ResNet model performance in 12 nm FinFET single-task (I–V) and multi-task (I–V and C–V) modeling for (<b>a</b>) training loss (AdaptiveSmoothL1), (<b>b</b>) validation loss (MAPE), and (<b>c</b>) test-set accuracy (R<sup>2</sup>).</p>
Full article ">Figure 7
<p>Comparison of ResNet Predictions and SPICE Simulations for FinFET I–V and C–V Characteristics. (<b>a</b>) Ids–Vds, (<b>b</b>) Cgs–Vds, (<b>c</b>) Cgd–Vds, and (<b>d</b>) Cgb–Vds.</p>
Full article ">Figure 8
<p>ResNet predictions vs. SPICE simulations: Scatter plots of I–V and C–V characteristics of FinFET devices. (<b>a</b>) Ids, (<b>b</b>) Cgs, (<b>c</b>) Cgd, and (<b>d</b>) Cgb.</p>
Full article ">Figure 9
<p>Performance of ResNet model on FinFET generalization test set. (<b>a</b>) Single-task IV modeling, (<b>b</b>) single-task IV migration learning modeling, (<b>c</b>) multi-task IV-CV modeling, and (<b>d</b>) multi-task IV–CV migration learning modeling.</p>
Full article ">Figure 10
<p>RO simulation curve SPICE model vs. ResNet model.</p>
Full article ">
18 pages, 8011 KiB  
Article
Intelligent Vision System with Pruning and Web Interface for Real-Time Defect Detection on African Plum Surfaces
by Arnaud Nguembang Fadja, Sain Rigobert Che and Marcellin Atemkemg
Information 2024, 15(10), 635; https://doi.org/10.3390/info15100635 (registering DOI) - 14 Oct 2024
Viewed by 439
Abstract
Agriculture stands as the cornerstone of Africa’s economy, supporting over 60% of the continent’s labor force. Despite its significance, the quality assessment of agricultural products remains a challenging task, particularly at a large scale, consuming valuable time and resources. The African plum is [...] Read more.
Agriculture stands as the cornerstone of Africa’s economy, supporting over 60% of the continent’s labor force. Despite its significance, the quality assessment of agricultural products remains a challenging task, particularly at a large scale, consuming valuable time and resources. The African plum is an agricultural fruit that is widely consumed across West and Central Africa but remains underrepresented in AI research. In this paper, we collected a dataset of 2892 African plum samples from fields in Cameroon representing the first dataset of its kind for training AI models. The dataset contains images of plums annotated with quality grades. We then trained and evaluated various state-of-the-art object detection and image classification models, including YOLOv5, YOLOv8, YOLOv9, Fast R-CNN, Mask R-CNN, VGG-16, DenseNet-121, MobileNet, and ResNet, on this African plum dataset. Our experimentation resulted in mean average precision scores ranging from 88.2% to 89.9% and accuracies between 86% and 91% for the object detection models and the classification models, respectively. We then performed model pruning to reduce model sizes while preserving performance, achieving up to 93.6% mean average precision and 99.09% accuracy after pruning YOLOv5, YOLOv8 and ResNet by 10–30%. We deployed the high-performing YOLOv8 system in a web application, offering an accessible AI-based quality assessment tool tailored for African plums. To the best of our knowledge, this represents the first such solution for assessing this underrepresented fruit, empowering farmers with efficient tools. Our approach integrates agriculture and AI to fill a key gap. Full article
Show Figures

Figure 1

Figure 1
<p>Sample images showcasing plum fruits on the fruit tree [<a href="#B27-information-15-00635" class="html-bibr">27</a>].</p>
Full article ">Figure 2
<p>The YOLOv5 is structured into three primary segments: the backbone, neck, and output [<a href="#B41-information-15-00635" class="html-bibr">41</a>].</p>
Full article ">Figure 3
<p>Overview of the key steps in our implementation. These structured steps ensure efficient implementation of the project.</p>
Full article ">Figure 4
<p>Sample images showcasing the labeling of good and defective plums with and without the background category. (<b>a</b>) Labeling of a good plum with the background class. (<b>b</b>) Labeling of a good plum. (<b>c</b>) Labeling of a defective plum with the background class. (<b>d</b>) Labeling of a defective plum.</p>
Full article ">Figure 5
<p>YOLOv5 training performance. This figure shows the training curves for the YOLOv5 object detection model. The top plot displays the loss function during the training process, which includes components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 6
<p>YOLOv8 training and evaluation. This figure presents the performance metrics for the YOLOv8 object detection model during the training and evaluation phases. The top plot shows the training loss, which is composed of components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 7
<p>YOLOv9 training and evaluation. This figure presents the performance metrics for the YOLOv9 object detection model during the training and evaluation phases. The top plot shows the training loss, which is composed of components for bounding box regression, object classification, and objectness prediction. The bottom plot displays the model’s mAP50 and mAP50-95 metrics on the validation dataset, which are key indicators of the model’s ability to accurately detect and classify objects.</p>
Full article ">Figure 8
<p>Fast R-CNN training and evaluation metrics. This figure shows the training and validation metrics for the Fast R-CNN object detection model. The blue line represents the overall training loss, which includes components for bounding box regression, object classification, and region proposal classification. The orange and green lines show the validation metrics for the classification loss and the regression loss, respectively. These metrics indicate the model’s performance in generating accurate region proposals and classifying/localizing detected objects.</p>
Full article ">Figure 9
<p>Mask R-CNN training and evaluation metrics. This figure presents the training and validation performance metrics for the Mask R-CNN instance segmentation model. The blue line represents the overall training loss, which includes components for bounding box regression, object classification, and region proposal classification. The orange and green lines show the validation metrics for the classification loss and the regression loss, respectively. These metrics indicate the model’s performance in generating accurate region proposals and classifying/localizing detected objects.</p>
Full article ">Figure 10
<p>Training and validation metrics for the VGG-16 model. The top curves represent training (green) and validation (red) accuracy, while the bottom curves depict training (green) and validation (red) loss. The model demonstrates rapid generalization from a strong initial point, as indicated by the swift convergence of accuracy and loss metrics.</p>
Full article ">Figure 11
<p>Training and validation metrics for the DenseNet-121 model. The top curves represent training (green) and validation (red) accuracy, while the bottom curves depict training (blue) and validation (yellow) loss. The model demonstrates rapid generalization from a strong initial point, as indicated by the swift convergence of accuracy and loss metrics.</p>
Full article ">Figure 12
<p>Model predictions with background class: YOLOv5, YOLOv8, and YOLOv9. (<b>a</b>) YOLOv5 good fruit prediction. (<b>b</b>) YOLOv8 good fruit prediction. (<b>c</b>) YOLOv9 good fruit prediction. (<b>d</b>) YOLOv5 bad fruit prediction. (<b>e</b>) YOLOv8 bad fruit prediction. (<b>f</b>) YOLOv9 bad fruit prediction.</p>
Full article ">
18 pages, 2803 KiB  
Article
Photosynthetic Traits of Quercus coccifera Green Fruits: A Comparison with Corresponding Leaves during Mediterranean Summer
by Dimitrios Kalachanis, Christos Chondrogiannis and Yiola Petropoulou
Plants 2024, 13(20), 2867; https://doi.org/10.3390/plants13202867 - 14 Oct 2024
Viewed by 386
Abstract
Fruit photosynthesis occurs in an internal microenvironment seldom encountered by a leaf (hypoxic and extremely CO2-enriched) due to its metabolic and anatomical features. In this study, the anatomical and photosynthetic traits of fully exposed green fruits of Quercus coccifera L. were [...] Read more.
Fruit photosynthesis occurs in an internal microenvironment seldom encountered by a leaf (hypoxic and extremely CO2-enriched) due to its metabolic and anatomical features. In this study, the anatomical and photosynthetic traits of fully exposed green fruits of Quercus coccifera L. were assessed during the period of fruit production (summer) and compared to their leaf counterparts. Our results indicate that leaf photosynthesis, transpiration and stomatal conductance drastically reduced during the summer drought, while they recovered significantly after the autumnal rainfalls. In acorns, gas exchange with the surrounding atmosphere is hindered by the complete absence of stomata; hence, credible CO2 uptake measurements could not be applied in the field. The linear electron transport rates (ETRs) in ambient air were similar in intact leaves and pericarps (i.e., when the physiological internal atmosphere of each tissue is maintained), while the leaf NPQ was significantly higher, indicating enhanced needs for harmless energy dissipation. The ETR measurements performed on leaf and pericarp discs at different CO2/O2 partial pressures in the supplied air mixture revealed that pericarps displayed significantly lower values at ambient gas levels, yet they increased by ~45% under high CO2/O2 ratios (i.e., at gas concentrations simulating the fruit’s interior). Concomitantly, NPQ declined gradually in both tissues as the CO2/O2 ratio increased, yet the decrease was more pronounced in pericarps. Furthermore, net CO2 assimilation rates for both leaf and pericarp segments were low in ambient air and increased almost equally at high CO2, while pericarps exhibited significantly higher respiration. It is suggested that during summer, when leaves suffer from photoinhibition, acorns could contribute to the overall carbon balance, through the re-assimilation of respiratory CO2, thereby reducing the reproductive cost. Full article
(This article belongs to the Section Plant Physiology and Metabolism)
Show Figures

Figure 1

Figure 1
<p>Fine structure of <span class="html-italic">Q. coccifera</span> leaves (left column) and pericarps (right column), as revealed by scanning electron microscope images (<b>A</b>,<b>B</b>) and light micrographs of cross sections (<b>C</b>,<b>D</b>). Stomata are indicated by arrows in the abaxial leaf surface (<b>A</b>), whereas no stomata could be found in pericarps (<b>B</b>). Samples were collected in August. In <a href="#plants-13-02867-f001" class="html-fig">Figure 1</a>A,B, bars = 20 μm; in <a href="#plants-13-02867-f001" class="html-fig">Figure 1</a>C,D. bars = 50 μm. Ch: chlorenchyma, Cu: cuticle, Ep: epidermis, PP: palisade parenchyma, SP: spongy parenchyma, Sc: sclerenchyma, St: stoma.</p>
Full article ">Figure 2
<p>Light (<b>A</b>) and epifluorescence (<b>B</b>) microscope images of pericarp cross sections. Pericarps were collected in August. Bars = 100 μm.</p>
Full article ">Figure 3
<p>Fast chlorophyll <span class="html-italic">a</span> fluorescence transients (OJIP) from intact leaves (open green circles) and pericarps (closed red circles) in summer. Transients are given on a logarithmic time scale and are expressed as relative variable fluorescence (V<sub>t</sub>), i.e., after double normalization at the F<sub>0</sub> and F<sub>P</sub> steps. Insert shows the I-P part of the transient on a linear time scale, double normalized at the F<sub>I</sub> and F<sub>P</sub> steps. Each curve is the average of 30 independent transients.</p>
Full article ">Figure 4
<p>Light response curves of PSII quantum yield (Φ<sub>PSII</sub>, <b>A</b>), linear electron transport rate (ETR, <b>B</b>) and non-photochemical quenching (NPQ, <b>C</b>) from intact leaves (open green circles) and pericarps (closed red circles) in summer. Values are means ± SD from 6 independent measurements. Asterisks denote statistically significant differences (<span class="html-italic">p</span> &lt; 0.05) between leaves and pericarps.</p>
Full article ">Figure 5
<p>Induction curves of electron transport rate (ETR, <b>A</b>) and non-photochemical quenching (NPQ, <b>B</b>) at 200 μmol m<sup>−2</sup> s<sup>−1</sup> from leaf (open green circles) and pericarp (closed red circles) discs under ambient O<sub>2</sub>/CO<sub>2</sub> concentrations. Subsequently, the samples were subjected to mutually varying external partial pressures of the interfering gases, i.e., a gradual CO<sub>2</sub> increase and a concurrent O<sub>2</sub> decrease, plus the reversion to ambient levels. Values are means ± SD from 8 independent measurements.</p>
Full article ">Figure 6
<p>Net CO<sub>2</sub> assimilation rate (<span class="html-italic">A</span>) at 200 μmol m<sup>−2</sup> s<sup>−1</sup>, under 400 and 2000 ppm CO<sub>2</sub> in the supplied air mixture, and dark respiration (<span class="html-italic">R</span><sub>d</sub>, at 400 ppm CO<sub>2</sub>) from leaf and pericarp segments in summer. Values are means ± SD from 6 independent measurements. Asterisks denote statistically significant differences (<span class="html-italic">p</span> &lt; 0.05) between leaves and pericarps.</p>
Full article ">Figure 7
<p>Net CO<sub>2</sub> assimilation (<span class="html-italic">A</span>) and transpiration (T<sub>r</sub>) rates and stomatal conductance (<span class="html-italic">g</span><sub>s</sub>) of leaves attached to the plant in August (green columns) and October (orange columns). PAR at 1420 μmol m<sup>−2</sup> s<sup>−1</sup>. Values are means ± SD from 24 independent measurements. Asterisks denote statistically significant differences (<span class="html-italic">p</span> &lt; 0.05) between summer and autumn for the indicated parameter.</p>
Full article ">
15 pages, 1641 KiB  
Article
Interactive Segmentation for Medical Images Using Spatial Modeling Mamba
by Yuxin Tang, Yu Li, Hua Zou and Xuedong Zhang
Information 2024, 15(10), 633; https://doi.org/10.3390/info15100633 (registering DOI) - 14 Oct 2024
Viewed by 519
Abstract
Interactive segmentation methods utilize user-provided positive and negative clicks to guide the model in accurately segmenting target objects. Compared to fully automatic medical image segmentation, these methods can achieve higher segmentation accuracy with limited image data, demonstrating significant potential in clinical applications. Typically, [...] Read more.
Interactive segmentation methods utilize user-provided positive and negative clicks to guide the model in accurately segmenting target objects. Compared to fully automatic medical image segmentation, these methods can achieve higher segmentation accuracy with limited image data, demonstrating significant potential in clinical applications. Typically, for each new click provided by the user, conventional interactive segmentation methods reprocess the entire network by re-inputting the click into the segmentation model, which greatly increases the user’s interaction burden and deviates from the intended goal of interactive segmentation tasks. To address this issue, we propose an efficient segmentation network, ESM-Net, for interactive medical image segmentation. It obtains high-quality segmentation masks based on the user’s initial clicks, reducing the complexity of subsequent refinement steps. Recent studies have demonstrated the strong performance of the Mamba model in various vision tasks; however, its application in interactive segmentation remains unexplored. In our study, we incorporate the Mamba module into our framework for the first time and enhance its spatial representation capabilities by developing a Spatial Augmented Convolution (SAC) module. These components are combined as the fundamental building blocks of our network. Furthermore, we designed a novel and efficient segmentation head to fuse multi-scale features extracted from the encoder, optimizing the generation of the predicted segmentation masks. Through comprehensive experiments, our method achieved state-of-the-art performance on three medical image datasets. Specifically, we achieved 1.43 NoC@90 on the Kvasir-SEG dataset, 1.57 NoC@90 on the CVC-ClinicDB polyp segmentation dataset, and 1.03 NoC@90 on the ADAM retinal disk segmentation dataset. The assessments on these three medical image datasets highlight the effectiveness of our approach in interactive medical image segmentation. Full article
(This article belongs to the Special Issue Applications of Deep Learning in Bioinformatics and Image Processing)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>ESM-Click Overview. Our model comprises two stages: preliminary segmentation and refinement segmentation. The encoded image and click features are fed into our proposed ESM-Net segmentation network to extract target-aware features and generate a coarse segmentation mask guided by the initial click. Starting from the second click, the new user-provided click is fed into the refinement network to optimize the details of the previously generated coarse mask. By iteratively executing the refinement network, a high-quality prediction mask is eventually produced.</p>
Full article ">Figure 2
<p>The Overall Architecture of ESM-Net integrates Spatial Augmented Convolution (SAC), Mamba modules, and MBConv for downsampling within the encoder module. (<b>a</b>) The Spatial Augmented Convolution Module enhances the spatial representation of features before input to the Mamba Module using a gate-like structure. (<b>b</b>) The Mamba Module transforms input features into feature sequences and processes them with SS2D to obtain comprehensive features from the merged sequences. (<b>c</b>) KAN SegHead receives multi-scale features from the encoder and utilizes KANLinear layers to output the final segmentation mask.</p>
Full article ">Figure 3
<p>The mean Intersection over Union (mIoU) and mean Dice coefficient (mDice) scores corresponding to the predictions obtained per click using different methods on the Kvasir-SEG and Clinic datasets.</p>
Full article ">Figure 4
<p>Qualitative results of ESM-Click. The first row illustrates example segmentations from the Kvasir-SEG dataset. The second row presents segmentation examples from the Clinic dataset with varying numbers of clicks. The third row showcases interactive segmentation cases from the ADAM dataset. Segmentation probability maps are depicted in blue; segmentation overlays on the original images are shown in red using the IoU evaluation metric. Green dots indicate positive clicks, while red dots indicate negative clicks.</p>
Full article ">
19 pages, 714 KiB  
Article
Enhanced COVID-19 Detection from X-ray Images with Convolutional Neural Network and Transfer Learning
by Qanita Bani Baker, Mahmoud Hammad, Mohammed Al-Smadi, Heba Al-Jarrah, Rahaf Al-Hamouri and Sa’ad A. Al-Zboon
J. Imaging 2024, 10(10), 250; https://doi.org/10.3390/jimaging10100250 - 13 Oct 2024
Viewed by 706
Abstract
The global spread of Coronavirus (COVID-19) has prompted imperative research into scalable and effective detection methods to curb its outbreak. The early diagnosis of COVID-19 patients has emerged as a pivotal strategy in mitigating the spread of the disease. Automated COVID-19 detection using [...] Read more.
The global spread of Coronavirus (COVID-19) has prompted imperative research into scalable and effective detection methods to curb its outbreak. The early diagnosis of COVID-19 patients has emerged as a pivotal strategy in mitigating the spread of the disease. Automated COVID-19 detection using Chest X-ray (CXR) imaging has significant potential for facilitating large-scale screening and epidemic control efforts. This paper introduces a novel approach that employs state-of-the-art Convolutional Neural Network models (CNNs) for accurate COVID-19 detection. The employed datasets each comprised 15,000 X-ray images. We addressed both binary (Normal vs. Abnormal) and multi-class (Normal, COVID-19, Pneumonia) classification tasks. Comprehensive evaluations were performed by utilizing six distinct CNN-based models (Xception, Inception-V3, ResNet50, VGG19, DenseNet201, and InceptionResNet-V2) for both tasks. As a result, the Xception model demonstrated exceptional performance, achieving 98.13% accuracy, 98.14% precision, 97.65% recall, and a 97.89% F1-score in binary classification, while in multi-classification it yielded 87.73% accuracy, 90.20% precision, 87.73% recall, and an 87.49% F1-score. Moreover, the other utilized models, such as ResNet50, demonstrated competitive performance compared with many recent works. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Figure 1
<p>General framework of the proposed approach.</p>
Full article ">Figure 2
<p>X-ray images of Normal, COVID-19, and Pneumonia used in binary classification task.</p>
Full article ">Figure 3
<p>X-ray images of Normal and Abnormal.</p>
Full article ">Figure 4
<p>Image-augmentation techniques.</p>
Full article ">Figure 5
<p>Proposed transfer learning-based technique.</p>
Full article ">
Back to TopTop