[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (705)

Search Parameters:
Keywords = interpretable AI

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
12 pages, 621 KiB  
Systematic Review
Systematic Review of AI-Assisted MRI in Prostate Cancer Diagnosis: Enhancing Accuracy Through Second Opinion Tools
by Saeed Alqahtani
Diagnostics 2024, 14(22), 2576; https://doi.org/10.3390/diagnostics14222576 (registering DOI) - 15 Nov 2024
Viewed by 434
Abstract
Background: Prostate cancer is a leading cause of cancer-related deaths in men worldwide, making accurate diagnosis critical for effective treatment. Recent advancements in artificial intelligence (AI) and machine learning (ML) have shown promise in improving the diagnostic accuracy of prostate cancer. Objectives: This [...] Read more.
Background: Prostate cancer is a leading cause of cancer-related deaths in men worldwide, making accurate diagnosis critical for effective treatment. Recent advancements in artificial intelligence (AI) and machine learning (ML) have shown promise in improving the diagnostic accuracy of prostate cancer. Objectives: This systematic review aims to evaluate the effectiveness of AI-based tools in diagnosing prostate cancer using MRI, with a focus on accuracy, specificity, sensitivity, and clinical utility compared to conventional diagnostic methods. Methods: A comprehensive search was conducted across PubMed, Embase, Ovid MEDLINE, Web of Science, Cochrane Library, and Institute of Electrical and Electronics Engineers (IEEE) Xplore for studies published between 2019 and 2024. Inclusion criteria focused on full-text, English-language studies involving AI for Magnetic Resonance Imaging (MRI) -based prostate cancer diagnosis. Diagnostic performance metrics such as area under curve (AUC), sensitivity, and specificity were analyzed, with risk of bias assessed using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) tool. Results: Seven studies met the inclusion criteria, employing various AI techniques, including deep learning and machine learning. These studies reported improved diagnostic accuracy (with AUC scores of up to 97%) and moderate sensitivity, with performance varying based on training data quality and lesion characteristics like Prostate Imaging Reporting and Data System (PI-RADS) scores. Conclusions: AI has significant potential to enhance prostate cancer diagnosis, particularly when used for second opinions in MRI interpretations. While these results are promising, further validation in diverse populations and clinical settings is necessary to fully integrate AI into standard practice. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

Figure 1
<p>Flowchart for search results.</p>
Full article ">
25 pages, 1114 KiB  
Article
Explainable Pre-Trained Language Models for Sentiment Analysis in Low-Resourced Languages
by Koena Ronny Mabokela, Mpho Primus and Turgay Celik
Big Data Cogn. Comput. 2024, 8(11), 160; https://doi.org/10.3390/bdcc8110160 (registering DOI) - 15 Nov 2024
Viewed by 163
Abstract
Sentiment analysis is a crucial tool for measuring public opinion and understanding human communication across digital social media platforms. However, due to linguistic complexities and limited data or computational resources, it is under-represented in many African languages. While state-of-the-art Afrocentric pre-trained language models [...] Read more.
Sentiment analysis is a crucial tool for measuring public opinion and understanding human communication across digital social media platforms. However, due to linguistic complexities and limited data or computational resources, it is under-represented in many African languages. While state-of-the-art Afrocentric pre-trained language models (PLMs) have been developed for various natural language processing (NLP) tasks, their applications in eXplainable Artificial Intelligence (XAI) remain largely unexplored. In this study, we propose a novel approach that combines Afrocentric PLMs with XAI techniques for sentiment analysis. We demonstrate the effectiveness of incorporating attention mechanisms and visualization techniques in improving the transparency, trustworthiness, and decision-making capabilities of transformer-based models when making sentiment predictions. To validate our approach, we employ the SAfriSenti corpus, a multilingual sentiment dataset for South African under-resourced languages, and perform a series of sentiment analysis experiments. These experiments enable comprehensive evaluations, comparing the performance of Afrocentric models against mainstream PLMs. Our results show that the Afro-XLMR model outperforms all other models, achieving an average F1-score of 71.04% across five tested languages, and the lowest error rate among the evaluated models. Additionally, we enhance the interpretability and explainability of the Afro-XLMR model using Local Interpretable Model-Agnostic Explanations (LIME) and Shapley Additive Explanations (SHAP). These XAI techniques ensure that sentiment predictions are not only accurate and interpretable but also understandable, fostering trust and reliability in AI-driven NLP technologies, particularly in the context of African languages. Full article
(This article belongs to the Special Issue Artificial Intelligence and Natural Language Processing)
18 pages, 746 KiB  
Article
Evaluating Anomaly Explanations Using Ground Truth
by Liat Antwarg Friedman, Chen Galed, Lior Rokach and Bracha Shapira
AI 2024, 5(4), 2375-2392; https://doi.org/10.3390/ai5040117 - 15 Nov 2024
Viewed by 292
Abstract
The widespread use of machine and deep learning algorithms for anomaly detection has created a critical need for robust explanations that can identify the features contributing to anomalies. However, effective evaluation methodologies for anomaly explanations are currently lacking, especially those that compare the [...] Read more.
The widespread use of machine and deep learning algorithms for anomaly detection has created a critical need for robust explanations that can identify the features contributing to anomalies. However, effective evaluation methodologies for anomaly explanations are currently lacking, especially those that compare the explanations against the true underlying causes, or ground truth. This paper aims to address this gap by introducing a rigorous, ground-truth-based framework for evaluating anomaly explanation methods, which enables the assessment of explanation correctness and robustness—key factors for actionable insights in anomaly detection. To achieve this, we present an innovative benchmark dataset of digital circuit truth tables with model-based anomalies, accompanied by local ground truth explanations. These explanations were generated using a novel algorithm designed to accurately identify influential features within each anomaly. Additionally, we propose an evaluation methodology based on correctness and robustness metrics, specifically tailored to quantify the reliability of anomaly explanations. This dataset and evaluation framework are publicly available to facilitate further research and standardize evaluation practices. Our experiments demonstrate the utility of this dataset and methodology by evaluating common model-agnostic explanation methods in an anomaly detection context. The results highlight the importance of ground-truth-based evaluation for reliable and interpretable anomaly explanations, advancing both theory and practical applications in explainable AI. This work establishes a foundation for rigorous, evidence-based assessments of anomaly explanations, fostering greater transparency and trust in AI-driven anomaly detection systems. Full article
(This article belongs to the Special Issue Interpretable and Explainable AI Applications)
Show Figures

Figure 1

Figure 1
<p>A flow chart describing two processes: (1) data generation (<b>above</b>), in which the data set with anomalies and the ground truth explanations are created, and (2) explanation evaluation (<b>below</b>), in which anomalies are detected and explained and the local explanations are evaluated against the ground truth.</p>
Full article ">Figure 2
<p><b>Left</b>—C17 original diagram, along with sample instances. <b>Right</b>—C17 anomalous diagram after changing gate z4 from NAND to AND, along with sample instances, and by doing that, we created an anomaly. The instances containing outputs marked in red are anomalous.</p>
Full article ">
16 pages, 3001 KiB  
Article
An Explainable Data-Driven Optimization Method for Unmanned Autonomous System Performance Assessment
by Hang Yi, Haisong Zhang, Hao Wang, Wenming Wang, Lixin Jia, Lihang Feng and Dong Wang
Electronics 2024, 13(22), 4469; https://doi.org/10.3390/electronics13224469 - 14 Nov 2024
Viewed by 251
Abstract
Unmanned autonomous systems (UASs), including drones and robotics, are widely employed across various fields. Despite significant advances in AI-enhanced intelligent systems, there remains a notable deficiency in the interpretability and comprehensive quantitative evaluation of these systems. The existing literature has primarily focused on [...] Read more.
Unmanned autonomous systems (UASs), including drones and robotics, are widely employed across various fields. Despite significant advances in AI-enhanced intelligent systems, there remains a notable deficiency in the interpretability and comprehensive quantitative evaluation of these systems. The existing literature has primarily focused on constructing evaluation frameworks and methods, but has often overlooked the rationality and reliability of these methods. To address these challenges, this paper proposes an innovative optimization evaluation method for data-driven unmanned autonomous systems. By optimizing the weights of existing indicators based on data distribution characteristics, this method enhances the stability and reliability of assessment outcomes. Furthermore, interpretability techniques such as Local Interpretable Model-agnostic Explanations (LIMEs) and Partial Dependence Plots (PDPs) were employed to verify the effectiveness of the designed evaluation indicators, thereby ensuring the robustness of the evaluation system. The experimental results validated the effectiveness of the proposed approach. Full article
(This article belongs to the Special Issue Advanced Control Strategies and Applications of Multi-Agent Systems)
Show Figures

Figure 1

Figure 1
<p>Flowchart of weight optimization method based on data feature consistency.</p>
Full article ">Figure 2
<p>Physical image of an unmanned field exploration robot.</p>
Full article ">Figure 3
<p>Evaluation framework for planning and decision-making capabilities of intelligent field detection robots.</p>
Full article ">Figure 4
<p>Perform weight iterative optimization solution process.</p>
Full article ">Figure 5
<p>The overall mean and variance distribution of each score before and after the optimization of the evaluation system.</p>
Full article ">Figure 6
<p>PDP interpretability results.</p>
Full article ">Figure 7
<p>ALE interpretability results.</p>
Full article ">Figure 8
<p>LIME interpretability results.</p>
Full article ">Figure 9
<p>SHAP interpretability results.</p>
Full article ">
11 pages, 919 KiB  
Article
Performance of a Deep Learning System and Performance of Optometrists for the Detection of Glaucomatous Optic Neuropathy Using Colour Retinal Photographs
by Catherine L. Jan, Algis Vingrys, Jacqueline Henwood, Xianwen Shang, Christian Davey, Peter van Wijngaarden, George Y. X. Kong, Jennifer C. Fan Gaskin, Bernardo P. Soares Bezerra, Randall S. Stafford and Mingguang He
Bioengineering 2024, 11(11), 1139; https://doi.org/10.3390/bioengineering11111139 - 13 Nov 2024
Viewed by 399
Abstract
Background/Objectives: Glaucoma is the leading cause of irreversible blindness, with a significant proportion of cases remaining undiagnosed globally. The interpretation of optic disc and retinal nerve fibre layer images poses challenges for optometrists and ophthalmologists, often leading to misdiagnosis. AI has the potential [...] Read more.
Background/Objectives: Glaucoma is the leading cause of irreversible blindness, with a significant proportion of cases remaining undiagnosed globally. The interpretation of optic disc and retinal nerve fibre layer images poses challenges for optometrists and ophthalmologists, often leading to misdiagnosis. AI has the potential to improve diagnosis. This study aims to validate an AI system (a convolutional neural network based on the Inception-v3 architecture) for detecting glaucomatous optic neuropathy (GON) using colour fundus photographs from a UK population and to compare its performance against Australian optometrists. Methods: A retrospective external validation study was conducted, comparing AI’s performance with that of 11 AHPRA-registered optometrists in Australia on colour retinal photographs, evaluated against a reference (gold) standard established by a panel of glaucoma specialists. Statistical analyses were performed using sensitivity, specificity, and area under the receiver operating characteristic curve (AUROC). Results: For referable GON, the sensitivity of the AI (33.3% [95%CI: 32.4–34.3) was significantly lower than that of optometrists (65.1% [95%CI: 64.1–66.0]), p < 0.0001, although with significantly higher specificity (AI: 97.4% [95%CI: 97.0–97.7]; optometrists: 85.5% [95%CI: 84.8–86.2], p < 0.0001). The optometrists demonstrated significantly higher AUROC (0.753 [95%CI: 0.744–0.762]) compared to AI (0.654 [95%CI: 0.645–0.662], p < 0.0001). Conclusion: The AI system exhibited lower performance than optometrists in detecting referable glaucoma. Our findings suggest that while AI can serve as a screening tool, both AI and optometrists have suboptimal performance for the nuanced diagnosis of glaucoma using fundus photographs alone. Enhanced training with diverse populations for AI is essential for improving GON detection and addressing the significant challenge of undiagnosed cases. Full article
(This article belongs to the Special Issue Artificial Intelligence Applications in Ophthalmology)
Show Figures

Figure 1

Figure 1
<p>Flow diagram of research process.</p>
Full article ">Figure 2
<p>(<b>a</b>) Features in false negative cases from AI and optometrist classifications for identifying ‘glaucoma certain’ cases when compared with specialist diagnosis. (<b>b</b>) Features in false negative cases from AI and optometrists’ classifications for identifying referral cases (certain plus suspect) when compared with specialist diagnosis.</p>
Full article ">
16 pages, 3729 KiB  
Article
Understanding Polymers Through Transfer Learning and Explainable AI
by Luis A. Miccio
Appl. Sci. 2024, 14(22), 10413; https://doi.org/10.3390/app142210413 - 12 Nov 2024
Viewed by 479
Abstract
In this work we study the use of artificial intelligence models, particularly focusing on transfer learning and interpretability, to predict polymer properties. Given the challenges imposed by data scarcity in polymer science, transfer learning offers a promising solution by using learnt features of [...] Read more.
In this work we study the use of artificial intelligence models, particularly focusing on transfer learning and interpretability, to predict polymer properties. Given the challenges imposed by data scarcity in polymer science, transfer learning offers a promising solution by using learnt features of models pre-trained on other datasets. We conducted a comparative analysis of direct modelling and transfer learning-based approaches using a polyacrylates’ glass transitions dataset as a proof-of-concept study. The AI models utilized tokenized SMILES strings to represent polymer structures, with convolutional neural networks processing these representations to predict Tg. To enhance model interpretability, Shapley value analysis was employed to assess the contribution of specific chemical groups to the predictions. The results indicate that while transfer learning provides robust predictive capabilities, direct modelling on polymer-specific data offers superior performance, particularly in capturing the complex interactions influencing Tg. This work highlights the importance of model interpretability and the limitations of applying molecular-level models to polymer systems. Full article
(This article belongs to the Special Issue Applications of Machine Learning with White-Boxing)
Show Figures

Figure 1

Figure 1
<p>Architecture of the neural network used for predicting the glass transition temperature of polymers. The model includes an embedding layer that converts tokenized SMILES strings into dense vector representations, followed by multiple 1D convolutional layers for feature extraction. The output of these layers is passed through fully connected layers, where the final activations are used as fingerprints. The final layer outputs a continuous value corresponding to the Tg. The shadowed area represents the “black box” within the model, highlighting the complexity and need for interpretability techniques such as Shapley value analysis.</p>
Full article ">Figure 2
<p>Schematic representation of the transfer learning-based model used for Tg prediction. The model is pre-trained on a molecular glass formers dataset and fine-tuned on a polymer-specific dataset (polyacrylates). The pre-trained model generates fingerprints based on learned molecular features, which are then used as input for the Tg prediction task. The black box indicates the fine-tuning process, emphasizing the transfer of knowledge from molecular systems to more complex polymer structures.</p>
Full article ">Figure 3
<p>Predicted vs. experimental Tg plots for best performing direct model at each data scarcity scenario. (30%) train on 93 samples, validate on 41 samples and test on 58 samples, and (50%) train on 67 samples, validate on 29 samples and test on 93 samples. Dashed lines are just an arbitrary guide that indicate the 25 and 50K difference regions (grey and red, respectively). The black arrows indicate these samples where the model large deviations.</p>
Full article ">Figure 4
<p>Plots showing the correlation between the observed Tg prediction deviations and molecular descriptors, including the number of atoms (#atoms) (<b>a</b>), the number of hydrogen bond acceptors (#acc) (<b>b</b>), and the observed difference distribution for samples with no H donors (<b>c</b>). The red lines represent the best linear fits, with shaded areas indicating confidence intervals. The deviations are approximated to a normal distribution, providing insights into the model’s ability to generalize across different polymer structures.</p>
Full article ">Figure 5
<p>Predicted vs. experimental Tg plots for best performing TL-based model at each data scarcity scenario. (30%) train on 93 samples, validate on 41 samples and test on 58 samples, and (50%) train on 67 samples, validate on 29 samples and test on 93 samples. Dashed lines are just an arbitrary guide that indicate the 25 and 50K difference regions (grey and red, respectively). The black arrows indicate these samples where the model large deviations. The results highlight the transfer learning model’s effectiveness, though with slightly higher deviations compared to direct modelling.</p>
Full article ">Figure 6
<p>Plots showing the correlation between the observed Tg prediction deviations and molecular descriptors, including the number of atoms (#atoms) (<b>a</b>), the number of hydrogen bond acceptors (#acc) (<b>b</b>), and the observed difference distribution for samples with no H donors (<b>c</b>). The red lines represent the best linear fits, with shaded areas indicating confidence intervals. The deviations are approximated to a normal distribution, providing insights into the model’s ability to generalize across different polymer structures.</p>
Full article ">Figure 7
<p>(<b>a</b>) A schematic of the glass transition process in polyacrylates, showing the relaxation of polymer chains within a characteristic length (Kuhn length, represented by dashed circles) where several factors, such as segment stiffness, covalent bonds, entanglements, and intermolecular forces, play crucial roles. (<b>b</b>) A schematic of the glass transition process in molecular systems, where the transition occurs across multiple molecules, with less emphasis on chain stiffness and entanglements. (<b>c</b>) The puzzle analogy representing the complete physics of the glass transition process, with the missing piece indicating irreducible error and missing physical phenomena in the molecular-based fingerprint.</p>
Full article ">Figure 8
<p>Comparison of Shapley value contributions between the direct and transfer learning-based models for samples with the largest deviations. The figure highlights the weighting of specific chemical groups, such as tert-butyl and methacrylate groups, where the transfer learning model shows more heterogeneous behaviour compared to the direct model. The results suggest that the transfer learning model underestimates the influence of short nonpolar chains, leading to poorer generalization to polymer chains.</p>
Full article ">Figure 9
<p>Shapley value contributions for an example where both the direct and transfer learning models slightly overestimate Tg. The analysis shows how the linear segment within the side chain (A) decreases the predicted Tg, while stiffer structures, such as phenyl groups, increase it (B). The direct model exhibits a better understanding of the impact of side chain structures on Tg, particularly when these structures significantly influence the polymer’s thermal behaviour.</p>
Full article ">Figure 10
<p>Shapley value contributions for two examples with complex side chains where the transfer learning model shows low deviations. The phenyl group and oxygen-containing segments (B) are shown to increase the predicted Tg due to their stiffness and influence on dipolar forces. The analysis highlights the transfer learning model’s ability to accurately predict Tg in polymers with more complex side chain structures, despite its general limitations.</p>
Full article ">
19 pages, 4719 KiB  
Article
Anomaly Detection and Analysis in Nuclear Power Plants
by Abhishek Chaudhary, Junseo Han, Seongah Kim, Aram Kim and Sunoh Choi
Electronics 2024, 13(22), 4428; https://doi.org/10.3390/electronics13224428 - 12 Nov 2024
Viewed by 350
Abstract
Industries are increasingly adopting digital systems to improve control and accessibility by providing real-time monitoring and early alerts for potential issues. While digital transformation fuels exponential growth, it exposes these industries to cyberattacks. For critical sectors such as nuclear power plants, a cyberattack [...] Read more.
Industries are increasingly adopting digital systems to improve control and accessibility by providing real-time monitoring and early alerts for potential issues. While digital transformation fuels exponential growth, it exposes these industries to cyberattacks. For critical sectors such as nuclear power plants, a cyberattack not only risks damaging the facility but also endangers human lives. In today’s digital world, enormous amounts of data are generated, and the analysis of these data can help ensure effectiveness, including security. In this study, we analyzed the data using a deep learning model for early detection of abnormal behavior. We first examined the Asherah Nuclear Power Plant simulator by initiating three different cyberattacks, each targeting a different system, thereby collecting and analyzing data from the simulator. Second, a Bi-LSTM model was used to detect anomalies in the simulator, which detected it before the plant’s protection system was activated in response to a threat. Finally, we applied explainable AI (XAI) to acquire insight into how distinctive features contribute to the detection of anomalies. XAI provides valuable explanations of model behavior by revealing how specific features influence anomaly detection during attacks. This research proposes an effective anomaly detection technique and interpretability to better understand counter-cyber threats in critical industries, such as nuclear plants. Full article
Show Figures

Figure 1

Figure 1
<p>Basic Asherah NPP simulator.</p>
Full article ">Figure 2
<p>Three different RPS states: (<b>a</b>) normal, (<b>b</b>) under attack with RPS disabled, and (<b>c</b>) under attack with RPS enabled.</p>
Full article ">Figure 3
<p>Attack scenarios in different systems.</p>
Full article ">Figure 4
<p>(<b>a</b>) CC Pump speed after the attack, (<b>b</b>) RPS activation point, and (<b>c</b>) shutdown of the reactor after RPS activation.</p>
Full article ">Figure 5
<p>(<b>a</b>) FW Pump flow after attack, (<b>b</b>) RPS activation point, and (<b>c</b>) shutdown of the reactor after RPS activation.</p>
Full article ">Figure 6
<p>(<b>a</b>) Pressurizer spray valve command after the attack, (<b>b</b>) RPS activation point, and (<b>c</b>) shutdown of the reactor after RPS activation.</p>
Full article ">Figure 7
<p>Proposed approach.</p>
Full article ">Figure 8
<p>(<b>a</b>) CC Pump attack detection and (<b>b</b>) log-scale plot of (<b>a</b>).</p>
Full article ">Figure 9
<p>(<b>a</b>) FW Pump attack detection and (<b>b</b>) log-scale plot of (<b>a</b>).</p>
Full article ">Figure 10
<p>(<b>a</b>) PZ spray valve attack detection and (<b>b</b>) log-scale plot of (<b>a</b>).</p>
Full article ">Figure 11
<p>(<b>a</b>) CC Pump attack at ‘Initial Detection’, (<b>b</b>) CC Pump attack after ‘RPS Activated’, (<b>c</b>) CC_PumpSpeed at initial detection, and (<b>d</b>) RX_ReactorPower contribution after ‘RPS Activated’.</p>
Full article ">Figure 12
<p>(<b>a</b>) FW Pump attack at initial detection, (<b>b</b>) FW attack after ‘RPS Activated’, (<b>c</b>) FW_Pump1Flow feature contribution by the FW Pump at initial detection, and (<b>d</b>) RX_ReactorPower feature contribution after RPS activation.</p>
Full article ">Figure 13
<p>PZ spray valve (<b>a</b>) at initial detection and (<b>b</b>) after RPS activation; (<b>c</b>) PZ_Press feature contribution at initial detection; (<b>d</b>) RX_ReactorPower feature contribution after RPS activation.</p>
Full article ">Figure 14
<p>Anomaly detection by TranAD on (<b>a</b>) CC Pump attack, (<b>b</b>) Feed Water attack, and (<b>c</b>) pressurizer spray valve attack.</p>
Full article ">
27 pages, 3088 KiB  
Article
Approaches to Identifying Emotions and Affections During the Museum Learning Experience in the Context of the Future Internet
by Iana Fominska, Stefano Di Tore, Michele Nappi, Gerardo Iovane, Maurizio Sibilio and Angela Gelo
Future Internet 2024, 16(11), 417; https://doi.org/10.3390/fi16110417 - 10 Nov 2024
Viewed by 782
Abstract
The Future Internet aims to revolutionize digital interaction by integrating advanced technologies like AI and IoT, enabling a dynamic and resilient network. It envisions emotionally intelligent systems that can interpret and respond to human feelings, creating immersive, empathy-driven learning experiences. This evolution aspires [...] Read more.
The Future Internet aims to revolutionize digital interaction by integrating advanced technologies like AI and IoT, enabling a dynamic and resilient network. It envisions emotionally intelligent systems that can interpret and respond to human feelings, creating immersive, empathy-driven learning experiences. This evolution aspires to form a responsive digital ecosystem that seamlessly connects technology and human emotion. This paper presents a computational model aimed at enhancing the emotional aspect of learning experiences within museum environments. The model is designed to represent and manage affective and emotional feedback, with a focus on how emotions can significantly impact the learning process in a museum context. The proposed model seeks to identify and quantify emotions during a visitor’s engagement with museum exhibits. To achieve this goal, we primarily explored the following: (i) methods and techniques for assessing and recognizing emotional responses in museum visitors, (ii) feedback management strategies based on the detection of visitors’ emotional states. Then, the methodology was tested on 1000 cases via specific questionnaire forms, along with the presentation of images and short videos, and the results of data analysis are reported. The findings contribute toward establishing a comprehensive methodology for the identification and quantification of the emotional state of museum visitors. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

Figure 1
<p>Emotion classes for emotivity.</p>
Full article ">Figure 2
<p>Emotion classes for affectivity (* means to distinguish the term when referring to a passionate person who has a connotation of sensual/sexual passion for the things synonymous with love).</p>
Full article ">Figure 3
<p>The distribution of respondents: states divided by gender, with strong orange for female and light orange for male.</p>
Full article ">Figure 4
<p>Distribution of responses to questionnaires with respect to the different emotions.</p>
Full article ">Figure 5
<p>Experimental results for emotive intensity with respect to the 36 questions in the questionnaires.</p>
Full article ">Figure 6
<p>Distribution of responses to questionnaires with respect to the different affections.</p>
Full article ">Figure 7
<p>Experimental results of affectivity intensity with respect to the 36 questions in the questionnaires.</p>
Full article ">Figure 8
<p>Experimental results for affectivity and emotivity.</p>
Full article ">
40 pages, 7132 KiB  
Review
AI in Cytopathology: A Narrative Umbrella Review on Innovations, Challenges, and Future Directions
by Daniele Giansanti
J. Clin. Med. 2024, 13(22), 6745; https://doi.org/10.3390/jcm13226745 - 9 Nov 2024
Viewed by 304
Abstract
The integration of artificial intelligence (AI) in cytopathology is an emerging field with transformative potential, aiming to enhance diagnostic precision and operational efficiency. This umbrella review seeks to identify prevailing themes, opportunities, challenges, and recommendations related to AI in cytopathology. Utilizing a standardized [...] Read more.
The integration of artificial intelligence (AI) in cytopathology is an emerging field with transformative potential, aiming to enhance diagnostic precision and operational efficiency. This umbrella review seeks to identify prevailing themes, opportunities, challenges, and recommendations related to AI in cytopathology. Utilizing a standardized checklist and quality control procedures, this review examines recent advancements and future implications of AI technologies in this domain. Twenty-one review studies were selected through a systematic process. AI has demonstrated promise in automating and refining diagnostic processes, potentially reducing errors and improving patient outcomes. However, several critical challenges need to be addressed to realize the benefits of AI fully. This review underscores the necessity for rigorous validation, ongoing empirical data on diagnostic accuracy, standardized protocols, and effective integration with existing clinical workflows. Ethical issues, including data privacy and algorithmic bias, must be managed to ensure responsible AI applications. Additionally, high costs and substantial training requirements present barriers to widespread AI adoption. Future directions highlight the importance of applying successful integration strategies from histopathology and radiology to cytopathology. Continuous research is needed to improve model interpretability, validation, and standardization. Developing effective strategies for incorporating AI into clinical practice and establishing comprehensive ethical and regulatory frameworks will be crucial for overcoming these challenges. In conclusion, while AI holds significant promise for advancing cytopathology, its full potential can only be achieved by addressing challenges related to validation, cost, and ethics. This review provides an overview of current advancements, identifies ongoing challenges, and offers a roadmap for the successful integration of AI into diagnostic cytopathology, informed by insights from related fields. Full article
Show Figures

Figure 1

Figure 1
<p>Details on the process of study selection.</p>
Full article ">Figure 2
<p>Synoptics diagram reporting a sketch of the the results.</p>
Full article ">Figure 3
<p>Temporal trend of studies focusing on the integration of AI with digital cytopathology.</p>
Full article ">Figure 4
<p>Studies in cytopathology that do not involve AI compared with studies in cytopathology focused on AI.</p>
Full article ">Figure 5
<p>Studies in cytopathology that do not involve AI compared with studies in cytopathology focused on AI with reference to the last five years.</p>
Full article ">Figure 6
<p>Proportion of studies in cytopathology and AI.</p>
Full article ">Figure 7
<p>Synoptic diagram reporting a sketch of the discussion.</p>
Full article ">Figure 8
<p>Temporal trends in scientific production in histopathology &amp;AI with details for the last five and ten years.</p>
Full article ">Figure 9
<p>Scientific production in histopathology &amp; AI (reviews and systematic reviews are illustrated separately.</p>
Full article ">Figure 10
<p>Ratio of AI-Focused to Non-AI-Focused Research in Histopathology.</p>
Full article ">Figure 11
<p>Ratio of AI-Focused to Non-AI-Focused Research in Histopathology in last five years.</p>
Full article ">Figure 12
<p>Temporal trends in scientific production in radiology &amp; AI with details for the last five and ten years.</p>
Full article ">Figure 13
<p>Scientific production in radiology &amp; AI (reviews and systematic reviews are illustrated).</p>
Full article ">Figure 14
<p>Ratio of AI-Focused to Non-AI-Focused Research in radiology.</p>
Full article ">Figure 15
<p>Ratio of AI-Focused to Non-AI-Focused Research in radiology in the last five years.</p>
Full article ">
36 pages, 11635 KiB  
Article
Reliable Autism Spectrum Disorder Diagnosis for Pediatrics Using Machine Learning and Explainable AI
by Insu Jeon, Minjoong Kim, Dayeong So, Eun Young Kim, Yunyoung Nam, Seungsoo Kim, Sehoon Shim, Joungmin Kim and Jihoon Moon
Diagnostics 2024, 14(22), 2504; https://doi.org/10.3390/diagnostics14222504 - 8 Nov 2024
Viewed by 481
Abstract
Background: As the demand for early and accurate diagnosis of autism spectrum disorder (ASD) increases, the integration of machine learning (ML) and explainable artificial intelligence (XAI) is emerging as a critical advancement that promises to revolutionize intervention strategies by improving both accuracy and [...] Read more.
Background: As the demand for early and accurate diagnosis of autism spectrum disorder (ASD) increases, the integration of machine learning (ML) and explainable artificial intelligence (XAI) is emerging as a critical advancement that promises to revolutionize intervention strategies by improving both accuracy and transparency. Methods: This paper presents a method that combines XAI techniques with a rigorous data-preprocessing pipeline to improve the accuracy and interpretability of ML-based diagnostic tools. Our preprocessing pipeline included outlier removal, missing data handling, and selecting pertinent features based on clinical expert advice. Using R and the caret package (version 6.0.94), we developed and compared several ML algorithms, validated using 10-fold cross-validation and optimized by grid search hyperparameter tuning. XAI techniques were employed to improve model transparency, offering insights into how features contribute to predictions, thereby enhancing clinician trust. Results: Rigorous data-preprocessing improved the models’ generalizability and real-world applicability across diverse clinical datasets, ensuring a robust performance. Neural networks and extreme gradient boosting models achieved the best performance in terms of accuracy, precision, and recall. XAI techniques demonstrated that behavioral features significantly influenced model predictions, leading to greater interpretability. Conclusions: This study successfully developed highly precise and interpretable ML models for ASD diagnosis, connecting advanced ML methods with practical clinical application and supporting the adoption of AI-driven diagnostic tools by healthcare professionals. This study’s findings contribute to personalized intervention strategies and early diagnostic practices, ultimately improving outcomes and quality of life for individuals with ASD. Full article
(This article belongs to the Special Issue Medical Data Processing and Analysis—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>ML framework and methodology flowchart.</p>
Full article ">Figure 2
<p>Flowchart of data selection and analysis process.</p>
Full article ">Figure 3
<p>Neural network (NNET) architecture for ASD diagnosis.</p>
Full article ">Figure 4
<p>Permutation feature importance (PFI) results: (<b>a</b>) GBM, (<b>b</b>) XGBoost, and (<b>c</b>) neural network.</p>
Full article ">Figure 4 Cont.
<p>Permutation feature importance (PFI) results: (<b>a</b>) GBM, (<b>b</b>) XGBoost, and (<b>c</b>) neural network.</p>
Full article ">Figure 5
<p>Local interpretable model-agnostic explanation (LIME) results in six different cases, using the NNET model.</p>
Full article ">Figure 5 Cont.
<p>Local interpretable model-agnostic explanation (LIME) results in six different cases, using the NNET model.</p>
Full article ">Figure 6
<p>LIME results in six different cases using the XGBoost model.</p>
Full article ">Figure 6 Cont.
<p>LIME results in six different cases using the XGBoost model.</p>
Full article ">Figure 7
<p>Shapley additive explanation (SHAP) absolute value bar plot. The <span class="html-italic">x</span>-axis represents the mean absolute SHAP value for each feature, showing how much each feature contributes to the model’s predictions on average. The <span class="html-italic">y</span>-axis lists the features in descending order of importance, with the most important features at the top.</p>
Full article ">Figure 8
<p>SHAP bee-swarm plot. The <span class="html-italic">x</span>-axis represents the SHAP value, which indicates the direction and magnitude of each feature’s effect on the prediction. The <span class="html-italic">y</span>-axis lists the features in order of importance, with the most influential features at the top. The color of the dots represents the value of the feature, from low (purple) to high (yellow).</p>
Full article ">Figure 9
<p>SHAP waterfall plot. The <span class="html-italic">x</span>-axis represents the cumulative contribution of each feature to the model’s prediction, starting from the base value to the final prediction. The <span class="html-italic">y</span>-axis lists the features in descending order based on their contribution to the prediction for a given instance.</p>
Full article ">Figure 10
<p>SHAP force plot. The <span class="html-italic">x</span>-axis represents the model’s prediction, showing how different features contribute to the final predicted value. The <span class="html-italic">y</span>-axis is not explicitly labeled, but the plot visually shows the positive and negative contributions of each feature, represented by the color and direction of the forces.</p>
Full article ">
23 pages, 1624 KiB  
Article
An Explainable Deep Learning-Enhanced IoMT Model for Effective Monitoring and Reduction of Maternal Mortality Risks
by Sherine Nagy Saleh, Mazen Nabil Elagamy, Yasmine N. M. Saleh and Radwa Ahmed Osman
Future Internet 2024, 16(11), 411; https://doi.org/10.3390/fi16110411 - 8 Nov 2024
Viewed by 695
Abstract
Maternal mortality (MM) is considered one of the major worldwide concerns. Despite the advances of artificial intelligence (AI) in healthcare, the lack of transparency in AI models leads to reluctance to adopt them. Employing explainable artificial intelligence (XAI) thus helps improve the transparency [...] Read more.
Maternal mortality (MM) is considered one of the major worldwide concerns. Despite the advances of artificial intelligence (AI) in healthcare, the lack of transparency in AI models leads to reluctance to adopt them. Employing explainable artificial intelligence (XAI) thus helps improve the transparency and effectiveness of AI-driven healthcare solutions. Accordingly, this article proposes a complete framework integrating an Internet of Medical Things (IoMT) architecture with an XAI-based deep learning model. The IoMT system continuously monitors pregnant women’s vital signs, while the XAI model analyzes the collected data to identify risk factors and generate actionable insights. Additionally, an efficient IoMT transmission model is developed to ensure reliable data transfer with the best-required system quality of service (QoS). Further analytics are performed on the data collected from different regions in a country to address high-risk cities. The experiments demonstrate the effectiveness of the proposed framework by achieving an accuracy of 80% for patients and 92.6% for regional risk prediction and providing interpretable explanations. The XAI-generated insights empower healthcare providers to make informed decisions and implement timely interventions. Furthermore, the IoMT transmission model ensures efficient and secure data transfer. Full article
(This article belongs to the Special Issue eHealth and mHealth)
Show Figures

Figure 1

Figure 1
<p>Proposed region monitoring scheme.</p>
Full article ">Figure 2
<p>Proposed patient monitoring scheme.</p>
Full article ">Figure 3
<p>Proposed deep learning model for risk classification.</p>
Full article ">Figure 4
<p>Proposed patient monitoring transmission scheme.</p>
Full article ">Figure 5
<p>Pearson correlation of the region-based MMR assessment using all 33 features. The naming labels denote only the odd features from the list.</p>
Full article ">Figure 6
<p>Accuracy and loss training and validation charts for the state dataset.</p>
Full article ">Figure 7
<p>Feature influence on the low-risk class (<b>a</b>) and the high-risk class (<b>b</b>) in the region dataset.</p>
Full article ">Figure 8
<p>Pearson correlation of the dataset features.</p>
Full article ">Figure 9
<p>Box plot showing the range of values for each feature and whether there are any outliers.</p>
Full article ">Figure 10
<p>SHAP summary showing feature influence for low-risk (<b>a</b>), medium-risk (<b>b</b>), and high-risk (<b>c</b>) patients.</p>
Full article ">Figure 11
<p>LIME analysis for three different patient records: one at low risk (<b>a</b>), another at a mid-level of risk (<b>b</b>), and the final at high risk (<b>c</b>).</p>
Full article ">Figure 12
<p>Required transmission power for the UP/DL communication (dBm) versus interfering devices’ transmission power (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>I</mi> </msub> </semantics></math>) (dBm).</p>
Full article ">Figure 13
<p>Overall energy efficiency for the UP/DL communication (EE) (bit/J) versus interfering devices’ transmission power <math display="inline"><semantics> <mrow> <mo>(</mo> <msub> <mi>P</mi> <mi>I</mi> </msub> <mo>)</mo> </mrow> </semantics></math> (dBm). Subfigures (<b>a</b>–<b>e</b>) correspond to distances between source S1 and destination D1 and between destinations D1 and D2 of 50 m, 100 m, 150 m, 200 m, and 250 m, respectively.</p>
Full article ">Figure 14
<p>(<b>a</b>) Required smart monitor device transmission power <math display="inline"><semantics> <mrow> <mo>(</mo> <msub> <mi>P</mi> <mrow> <mi>S</mi> <mi>r</mi> <mi>e</mi> <mi>q</mi> </mrow> </msub> <mo>)</mo> </mrow> </semantics></math> (dBm) versus the required uplink QoS <math display="inline"><semantics> <mrow> <mo>(</mo> <msub> <mi>Q</mi> <mrow> <mi>U</mi> <mi>L</mi> </mrow> </msub> <mo>)</mo> </mrow> </semantics></math>. (<b>b</b>) Required gateway transmission power <math display="inline"><semantics> <mrow> <mo>(</mo> <msub> <mi>P</mi> <mrow> <mi>S</mi> <mi>r</mi> <mi>e</mi> <mi>q</mi> </mrow> </msub> <mo>)</mo> </mrow> </semantics></math> (dBm) versus the required downlink QoS <math display="inline"><semantics> <mrow> <mo>(</mo> <msub> <mi>Q</mi> <mrow> <mi>D</mi> <mi>L</mi> </mrow> </msub> <mo>)</mo> </mrow> </semantics></math>.</p>
Full article ">
18 pages, 3457 KiB  
Article
Navigating the Spectrum: Assessing the Concordance of ML-Based AI Findings with Radiology in Chest X-Rays in Clinical Settings
by Marie-Luise Kromrey, Laura Steiner, Felix Schön, Julie Gamain, Christian Roller and Carolin Malsch
Healthcare 2024, 12(22), 2225; https://doi.org/10.3390/healthcare12222225 - 7 Nov 2024
Viewed by 530
Abstract
Background: The integration of artificial intelligence (AI) into radiology aims to improve diagnostic accuracy and efficiency, particularly in settings with limited access to expert radiologists and in times of personnel shortage. However, challenges such as insufficient validation in actual real-world settings or automation [...] Read more.
Background: The integration of artificial intelligence (AI) into radiology aims to improve diagnostic accuracy and efficiency, particularly in settings with limited access to expert radiologists and in times of personnel shortage. However, challenges such as insufficient validation in actual real-world settings or automation bias should be addressed before implementing AI software in clinical routine. Methods: This cross-sectional study in a maximum care hospital assesses the concordance between diagnoses made by a commercial AI-based software and conventional radiological methods augmented by AI for four major thoracic pathologies in chest X-ray: fracture, pleural effusion, pulmonary nodule and pneumonia. Chest radiographs of 1506 patients (median age 66 years, 56.5% men) consecutively obtained between January and August 2023 were re-evaluated by the AI software InferRead DR Chest®. Results: Overall, AI software detected thoracic pathologies more often than radiologists (18.5% vs. 11.1%). In detail, it detected fractures, pneumonia, and nodules more frequently than radiologists, while radiologists identified pleural effusions more often. Reliability was highest for pleural effusions (0.63, 95%-CI 0.58–0.69), indicating good agreement, and lowest for fractures (0.39, 95%-CI 0.32–0.45), indicating moderate agreement. Conclusions: The tested software shows a high detection rate, particularly for fractures, pneumonia, and nodules, but hereby produces a nonnegligible number of false positives. Thus, AI-based software shows promise in enhancing diagnostic accuracy; however, cautious interpretation and human oversight remain crucial. Full article
(This article belongs to the Section Artificial Intelligence in Medicine)
Show Figures

Figure 1

Figure 1
<p>Chest radiograph analyzed with AI. AI-generated report based on a routine chest radiograph marking two findings together with their precise locations: one lung abnormality (1) and one aortic abnormality (2). All findings are listed on the right side, along with an overall abnormality probability score of 60%. Below the list is a concise report created by the software, similar to a radiology report.</p>
Full article ">Figure 2
<p>Frequency of diagnoses by radiologist augmented by AI versus AI only.</p>
Full article ">Figure 3
<p>Frequency of positive diagnoses by radiologist augmented by AI (purple), AI software only (yellow) and agreeing diagnoses (intersection). Total sample size was 1506.</p>
Full article ">Figure 4
<p>Estimates for Cohen’s kappa, PABAK and Gwet’s AC1 coefficient with 95% bootstrap confidence intervals for the total sample and stratified by gender.</p>
Full article ">Figure 5
<p>Inconsistencies of fracture diagnoses: (<b>A</b>)—false-positive fracture of the right clavicle described by AI at the overlay of the clavicle and 2nd rib (pneumonia in the left basal lung was correctly identified); (<b>B</b>)—false-positive fracture description of the 10th rib on the right by AI due to overlay of external oxygen hose; (<b>C</b>)—false-negative bone status; fracture of the left clavicle was not described by AI (Note: other finding incorrectly labelling “aortic abnormality”); (<b>D</b>)—rib fracture on the right found by AI but overlooked by radiologists.</p>
Full article ">Figure 6
<p>Inconsistencies of pneumonia diagnoses: (<b>A</b>)—false-positive AI interpretation of dense breast parenchyma as pneumonia in a young woman (Note: other finding incorrectly labelling “aortic abnormality”); (<b>B</b>)—false-negative interpretation of pneumonia. X-ray shows pneumonic congestion in the right lower lobe not detected by AI (Note: other findings incorrectly labelling “aortic abnormality” and “pneumonia”).</p>
Full article ">Figure 7
<p>Inconsistencies of nodule diagnoses: (<b>A</b>)—false-positive AI interpretation of pulmonary mass in the lower lobe on the left as infiltration caused by pneumonia. (Note: Also false classification of Portcatheter as pneumothorax, again “aortic abnormality” and “rip fracture”); (<b>B</b>)—correctly identified nodule by AI overlooked by radiologist (Note: other finding incorrectly labelling “aortic abnormality”); (<b>C</b>)—False-positive nodule described by AI, really corresponding to the overlay of two ribs; (<b>D</b>)—False-positive nodule described by AI, really corresponding to external oxygen hose (Note: other findings incorrectly labelling “aortic abnormality” and “rib fracture”).</p>
Full article ">Figure 8
<p>Inconsistencies of pleural effusion diagnoses: (<b>A</b>,<b>B</b>)—AI missing the sulcal effusion described by the radiologist based on the lateral projection; (<b>C</b>,<b>D</b>)—Similar case with an even more prominent pleural effusion not detected by AI (Note: other findings incorrectly labelling “aortic abnormality”, “pneumonia” and “rib fracture”).</p>
Full article ">
27 pages, 15476 KiB  
Article
Explainable AI-Based Ensemble Clustering for Load Profiling and Demand Response
by Elissaios Sarmas, Afroditi Fragkiadaki and Vangelis Marinakis
Energies 2024, 17(22), 5559; https://doi.org/10.3390/en17225559 - 7 Nov 2024
Viewed by 379
Abstract
Smart meter data provide an in-depth perspective on household energy usage. This research leverages on such data to enhance demand response (DR) programs through a novel application of ensemble clustering. Despite its promising capabilities, our literature review identified a notable under-utilization of ensemble [...] Read more.
Smart meter data provide an in-depth perspective on household energy usage. This research leverages on such data to enhance demand response (DR) programs through a novel application of ensemble clustering. Despite its promising capabilities, our literature review identified a notable under-utilization of ensemble clustering in this domain. To address this shortcoming, we applied an advanced ensemble clustering method and compared its performance with traditional algorithms, namely, K-Means++, fuzzy K-Means, Hierarchical Agglomerative Clustering, Spectral Clustering, Gaussian Mixture Models (GMMs), BIRCH, and Self-Organizing Maps (SOMs), across a dataset of 5567 households for a range of cluster counts from three to nine. The performance of these algorithms was assessed using an extensive set of evaluation metrics, including the Silhouette Score, the Davies–Bouldin Score, the Calinski–Harabasz Score, and the Dunn Index. Notably, while ensemble clustering often ranked among the top performers, it did not consistently surpass all individual algorithms, indicating its potential for further optimization. Unlike approaches that seek the algorithmically optimal number of clusters, our method proposes a practical six-cluster solution designed to meet the operational needs of utility providers. For this case, the best performing algorithm according to the evaluation metrics was ensemble clustering. This study is further enhanced by integrating Explainable AI (xAI) techniques, which improve the interpretability and transparency of our clustering results. Full article
(This article belongs to the Special Issue Advances in Energy Market and Distributed Generation)
Show Figures

Figure 1

Figure 1
<p>Detailed overview of methodology.</p>
Full article ">Figure 2
<p>Snapshot of the dataset.</p>
Full article ">Figure 3
<p>Silhouette scores for cluster ranges from 3 to 9.</p>
Full article ">Figure 4
<p>Davies–Bouldin scores for cluster ranges from 3 to 9.</p>
Full article ">Figure 5
<p>Calinski–Harabasz scores for cluster ranges from 3 to 9.</p>
Full article ">Figure 6
<p>Dunn index for cluster ranges from 3 to 9.</p>
Full article ">Figure 7
<p>Non-normalized load profiles of all households in each cluster.</p>
Full article ">Figure 8
<p>Normalized load profiles of all households in each cluster.</p>
Full article ">Figure 9
<p>Weekend vs. weekday mean cluster profiles.</p>
Full article ">Figure 10
<p>SHAP values analysis for cluster 1.</p>
Full article ">Figure 11
<p>SHAP values analysis for cluster 2.</p>
Full article ">
30 pages, 2287 KiB  
Article
Exploring the Impact of Artificial Intelligence on Healthcare Management: A Combined Systematic Review and Machine-Learning Approach
by Vito Santamato, Caterina Tricase, Nicola Faccilongo, Massimo Iacoviello and Agostino Marengo
Appl. Sci. 2024, 14(22), 10144; https://doi.org/10.3390/app142210144 - 6 Nov 2024
Viewed by 1412
Abstract
The integration of artificial intelligence (AI) in healthcare management marks a significant advance in technological innovation, promising transformative effects on healthcare processes, patient care, and the efficacy of emergency responses. The scientific novelty of the study lies in its integrated approach, combining systematic [...] Read more.
The integration of artificial intelligence (AI) in healthcare management marks a significant advance in technological innovation, promising transformative effects on healthcare processes, patient care, and the efficacy of emergency responses. The scientific novelty of the study lies in its integrated approach, combining systematic review and predictive algorithms to provide a comprehensive understanding of AI’s role in improving healthcare management across different contexts. Covering the period between 2019 and 2023, which includes the global challenges posed by the COVID-19 pandemic, this research investigates the operational, strategic, and emergency response implications of AI adoption in the healthcare sector. It further examines how the impact of AI varies across temporal and geographical contexts. The study addresses two main research objectives: to explore how AI influences healthcare management in operational, strategic, and emergency response domains, and to identify variations in the impact of AI on healthcare management based on temporal and geographical contexts. Utilizing an integrated approach, we compared various prediction algorithms, including logistic regression, and interpreted the results through SHAP (SHapley Additive exPlanations) analysis. The findings reveal five key thematic areas: AI’s role in enhancing quality assurance, resource management, technological innovation, security, and the healthcare response to the COVID-19 pandemic. The study highlights AI’s positive influence on operational efficiency and strategic decision making, while also identifying challenges related to data privacy, ethical considerations, and the need for ongoing technological integration. These insights provide opportunities for targeted interventions to optimize AI’s impact in current and future healthcare landscapes. In conclusion, this work contributes to a deeper understanding of the role of AI in healthcare management and provides insights for policymakers, healthcare professionals, and researchers, offering a roadmap for addressing both the opportunities and challenges posed by AI integration in the healthcare sector. Full article
(This article belongs to the Special Issue Large Language Models: Transforming E-health)
Show Figures

Figure 1

Figure 1
<p>PRISMA flow chart for this study.</p>
Full article ">Figure 2
<p>Predictive models performances.</p>
Full article ">Figure 3
<p>Confusion matrix for logistic regression.</p>
Full article ">Figure 4
<p>Receiver Operating Characteristic (ROC) Curves for logistic regression across five target classes.</p>
Full article ">Figure 5
<p>SHAP value analysis for predictive model targets.</p>
Full article ">
19 pages, 2078 KiB  
Article
Enhancing Medical Image Classification with Unified Model Agnostic Computation and Explainable AI
by Elie Neghawi and Yan Liu
AI 2024, 5(4), 2260-2278; https://doi.org/10.3390/ai5040111 - 5 Nov 2024
Viewed by 614
Abstract
Background: Advances in medical image classification have recently benefited from general augmentation techniques. However, these methods often fall short in performance and interpretability. Objective: This paper applies the Unified Model Agnostic Computation (UMAC) framework specifically to the medical domain to demonstrate [...] Read more.
Background: Advances in medical image classification have recently benefited from general augmentation techniques. However, these methods often fall short in performance and interpretability. Objective: This paper applies the Unified Model Agnostic Computation (UMAC) framework specifically to the medical domain to demonstrate its utility in this critical area. Methods: UMAC is a model-agnostic methodology designed to develop machine learning approaches that integrate seamlessly with various paradigms, including self-supervised, semi-supervised, and supervised learning. By unifying and standardizing computational models and algorithms, UMAC ensures adaptability across different data types and computational environments while incorporating state-of-the-art methodologies. In this study, we integrate UMAC as a plug-and-play module within convolutional neural networks (CNNs) and Transformer architectures, enabling the generation of high-quality representations even with minimal data. Results: Our experiments across nine diverse 2D medical image datasets show that UMAC consistently outperforms traditional data augmentation methods, achieving a 1.89% improvement in classification accuracy. Conclusions: Additionally, by incorporating explainable AI (XAI) techniques, we enhance model transparency and reliability in decision-making. This study highlights UMAC’s potential as a powerful tool for improving both the performance and interpretability of medical image classification models. Full article
(This article belongs to the Topic Applications of NLP, AI, and ML in Software Engineering)
Show Figures

Figure 1

Figure 1
<p>Unified Agnostic Computation Process with self-supervised learning in the medical field. In this context, <math display="inline"><semantics> <mi>θ</mi> </semantics></math> and <math display="inline"><semantics> <mi>ξ</mi> </semantics></math> are parameters, while <math display="inline"><semantics> <mi>σ</mi> </semantics></math> and <math display="inline"><semantics> <msup> <mi>σ</mi> <mo>′</mo> </msup> </semantics></math> represent random parameters.</p>
Full article ">Figure 2
<p>Example of Augmentation Function applied to a DermaMNIST image, showcasing color shifts and spatial transformations.</p>
Full article ">Figure 3
<p>Sample images from the MedMNIST datasets, including examples from BreastMNIST, DermaMNIST, RetinaMNIST, ChestMNIST, and PneumoniaMNIST.</p>
Full article ">Figure 4
<p>UMAC training with Multiple MedMNIST2D Datasets.</p>
Full article ">
Back to TopTop