[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,231)

Search Parameters:
Keywords = NLP

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
11 pages, 858 KiB  
Article
Safety and Effectiveness of Oral Anticoagulants in Atrial Fibrillation: Real-World Insights Using Natural Language Processing and Machine Learning
by Juan Cosín-Sales, Manuel Anguita Sánchez, Carmen Suárez, Carlos Arias-Cabrales, Luisa Martínez-Sanchez, Savana Research Group, Daniel Arumi and Susana Fernández de Cabo
J. Clin. Med. 2024, 13(20), 6226; https://doi.org/10.3390/jcm13206226 (registering DOI) - 18 Oct 2024
Abstract
Background/Objectives: We assessed the effectiveness and safety of vitamin K antagonists (VKAs) versus direct oral anticoagulants (DOACs) in patients with atrial fibrillation (AF) using artificial intelligence techniques. Methods: This is a retrospective study in 15 Spanish hospitals (2014–2020), including adult AF patients with [...] Read more.
Background/Objectives: We assessed the effectiveness and safety of vitamin K antagonists (VKAs) versus direct oral anticoagulants (DOACs) in patients with atrial fibrillation (AF) using artificial intelligence techniques. Methods: This is a retrospective study in 15 Spanish hospitals (2014–2020), including adult AF patients with no history of anticoagulation, thrombosis events, rheumatic mitral valvular heart disease, mitral valve stenosis, or pregnancy. We employed EHRead® technology based on natural language processing (NLP) and machine learning (ML), along with SNOMED-CT terminology, to extract clinical data from electronic health records (EHRs). Using propensity score matching (PSM), the effectiveness, safety, and hospital mortality of VKAs versus DOACs were analyzed through Kaplan–Meier curves and Cox regression. Results: Out of 138,773,332 EHRs from 4.6 million individuals evaluated, 44,292 patients were included, 79.6% on VKAs and 20.4% on DOACs. Most patients were elderly [VKA 78 (70, 84) and DOAC 75 (66, 83) years], with numerous comorbidities (75.5% and 70.2% hypertension, 47.2% and 39.9% diabetes, and 40.3% and 34.8% heart failure, respectively). Additionally, 60.4% of VKA and 48.7% of DOAC users had a CHA2DS2-VASc Score ≥4. After PSM, 8929 patients per subgroup were selected. DOAC users showed a lower risk of thrombotic events [HR 0.81 (95% CI 0.70–0.94)], minor bleeding [HR 0.89 (95% CI 0.83–0.96)], and mortality [HR 0.80 (95% CI 0.69–0.92)]. Conclusions: Applying NLP and ML, we generated valuable real-world evidence on anticoagulated AF patients in Spain. Even in complex populations, DOACs have demonstrated a better safety and effectiveness profile than VKAs. Full article
Show Figures

Figure 1

Figure 1
<p>Flow chart with an overview of the patient selection methodology. The initial screening set for our study comprised over 4.6 million patients at participating hospitals, processing nearly 139 million EHRs. The application of exclusion filters (age &lt;18 years, rheumatic mitral valvopathy, mitral valve stenosis, venous thromboembolism, pregnancy, previous anticoagulant therapy, recent stroke, TIA or SE, no follow-up data, and the cutoff date to mitigate COVID-19 impacts) defined a study population of 44,292 patients with atrial fibrillation (AF), with 35,267 (79.6%) on VKAs and 9025 (20.4%) on DOACs. Patients on VKAs and DOACs were matched using propensity score matching (PSM) to ensure balanced groups for comparative analysis.</p>
Full article ">Figure 2
<p>Kaplan–Meier curves for PSM patients comparing DOACs and VKAs showing treatment outcomes for stroke/SE/TIA (<b>A</b>), death (<b>B</b>), major (<b>C</b>), and minor bleeding (<b>D</b>). Cox regression analysis was employed to compare survival across treatments, with VKAs serving as the reference. Significance was determined based on a <span class="html-italic">p</span>-value &lt; 0.05 and a 95% confidence interval. The proportion y-axis has been adjusted to 0.7 for better visualization of results. See <a href="#app1-jcm-13-06226" class="html-app">Supplemental results (Supplementary Figure S3)</a> for non-adjusted curves.</p>
Full article ">
34 pages, 5078 KiB  
Systematic Review
Context-Aware Embedding Techniques for Addressing Meaning Conflation Deficiency in Morphologically Rich Languages Word Embedding: A Systematic Review and Meta Analysis
by Mosima Anna Masethe, Hlaudi Daniel Masethe and Sunday O. Ojo
Computers 2024, 13(10), 271; https://doi.org/10.3390/computers13100271 - 17 Oct 2024
Viewed by 159
Abstract
This systematic literature review aims to evaluate and synthesize the effectiveness of various embedding techniques—word embeddings, contextual word embeddings, and context-aware embeddings—in addressing Meaning Conflation Deficiency (MCD). Using the PRISMA framework, this study assesses the current state of research and provides insights into [...] Read more.
This systematic literature review aims to evaluate and synthesize the effectiveness of various embedding techniques—word embeddings, contextual word embeddings, and context-aware embeddings—in addressing Meaning Conflation Deficiency (MCD). Using the PRISMA framework, this study assesses the current state of research and provides insights into the impact of these techniques on resolving meaning conflation issues. After a thorough literature search, 403 articles on the subject were found. A thorough screening and selection process resulted in the inclusion of 25 studies in the meta-analysis. The evaluation adhered to the PRISMA principles, guaranteeing a methodical and lucid process. To estimate effect sizes and evaluate heterogeneity and publication bias among the chosen papers, meta-analytic approaches were utilized such as the tau-squared (τ2) which represents a statistical parameter used in random-effects, H-squared (H2) is a statistic used to measure heterogeneity, and I-squared (I2) quantify the degree of heterogeneity. The meta-analysis demonstrated a high degree of variation in effect sizes among the studies, with a τ2 value of 8.8724. The significant degree of heterogeneity was further emphasized by the H2 score of 8.10 and the I2 value of 87.65%. A trim and fill analysis with a beta value of 5.95, a standard error of 4.767, a Z-value (or Z-score) of 1.25 which is a statistical term used to express the number of standard deviations a data point deviates from the established mean, and a p-value (probability value) of 0.2 was performed to account for publication bias which is one statistical tool that can be used to assess the importance of hypothesis test results. The results point to a sizable impact size, but the estimates are highly unclear, as evidenced by the huge standard error and non-significant p-value. The review concludes that although contextually aware embeddings have promise in treating Meaning Conflation Deficiency, there is a great deal of variability and uncertainty in the available data. The varied findings among studies are highlighted by the large τ2, I2, and H2 values, and the trim and fill analysis show that changes in publication bias do not alter the impact size’s non-significance. To generate more trustworthy insights, future research should concentrate on enhancing methodological consistency, investigating other embedding strategies, and extending analysis across various languages and contexts. Even though the results demonstrate a significant impact size in addressing MCD through sophisticated word embedding techniques, like context-aware embeddings, there is still a great deal of variability and uncertainty because of various factors, including the different languages studied, the sizes of the corpuses, and the embedding techniques used. These differences show how future research methods must be standardized to guarantee that study results can be compared to one another. The results emphasize how crucial it is to extend the linguistic scope to more morphologically rich and low-resource languages, where MCD is especially difficult. The creation of language-specific models for low-resource languages is one way to increase performance and consistency across Natural Language Processing (NLP) applications in a practical sense. By taking these actions, we can advance our understanding of MCD more thoroughly, which will ultimately improve the performance of NLP systems in a variety of language circumstances. Full article
(This article belongs to the Special Issue Natural Language Processing (NLP) and Large Language Modelling)
Show Figures

Figure 1

Figure 1
<p>PRISMA flow diagram for the CAWE research [<a href="#B30-computers-13-00271" class="html-bibr">30</a>].</p>
Full article ">Figure 2
<p>Proposed Embeddings Taxonomy.</p>
Full article ">Figure 3
<p>Forest plot for distribution of effect size of meta-analysis summary of CAWE in <a href="#computers-13-00271-t003" class="html-table">Table 3</a> [<a href="#B29-computers-13-00271" class="html-bibr">29</a>,<a href="#B33-computers-13-00271" class="html-bibr">33</a>,<a href="#B34-computers-13-00271" class="html-bibr">34</a>,<a href="#B35-computers-13-00271" class="html-bibr">35</a>,<a href="#B36-computers-13-00271" class="html-bibr">36</a>,<a href="#B37-computers-13-00271" class="html-bibr">37</a>,<a href="#B38-computers-13-00271" class="html-bibr">38</a>,<a href="#B39-computers-13-00271" class="html-bibr">39</a>,<a href="#B40-computers-13-00271" class="html-bibr">40</a>,<a href="#B41-computers-13-00271" class="html-bibr">41</a>,<a href="#B42-computers-13-00271" class="html-bibr">42</a>,<a href="#B43-computers-13-00271" class="html-bibr">43</a>,<a href="#B44-computers-13-00271" class="html-bibr">44</a>,<a href="#B45-computers-13-00271" class="html-bibr">45</a>,<a href="#B46-computers-13-00271" class="html-bibr">46</a>,<a href="#B47-computers-13-00271" class="html-bibr">47</a>,<a href="#B48-computers-13-00271" class="html-bibr">48</a>,<a href="#B49-computers-13-00271" class="html-bibr">49</a>,<a href="#B50-computers-13-00271" class="html-bibr">50</a>,<a href="#B51-computers-13-00271" class="html-bibr">51</a>,<a href="#B52-computers-13-00271" class="html-bibr">52</a>,<a href="#B53-computers-13-00271" class="html-bibr">53</a>,<a href="#B54-computers-13-00271" class="html-bibr">54</a>,<a href="#B55-computers-13-00271" class="html-bibr">55</a>,<a href="#B56-computers-13-00271" class="html-bibr">56</a>].</p>
Full article ">Figure 4
<p>Forest plot with subgroups for distribution of effect size of meta-analysis summary of CAWE in <a href="#computers-13-00271-t003" class="html-table">Table 3</a> [<a href="#B29-computers-13-00271" class="html-bibr">29</a>,<a href="#B33-computers-13-00271" class="html-bibr">33</a>,<a href="#B34-computers-13-00271" class="html-bibr">34</a>,<a href="#B35-computers-13-00271" class="html-bibr">35</a>,<a href="#B36-computers-13-00271" class="html-bibr">36</a>,<a href="#B37-computers-13-00271" class="html-bibr">37</a>,<a href="#B38-computers-13-00271" class="html-bibr">38</a>,<a href="#B39-computers-13-00271" class="html-bibr">39</a>,<a href="#B40-computers-13-00271" class="html-bibr">40</a>,<a href="#B41-computers-13-00271" class="html-bibr">41</a>,<a href="#B42-computers-13-00271" class="html-bibr">42</a>,<a href="#B43-computers-13-00271" class="html-bibr">43</a>,<a href="#B44-computers-13-00271" class="html-bibr">44</a>,<a href="#B45-computers-13-00271" class="html-bibr">45</a>,<a href="#B46-computers-13-00271" class="html-bibr">46</a>,<a href="#B47-computers-13-00271" class="html-bibr">47</a>,<a href="#B48-computers-13-00271" class="html-bibr">48</a>,<a href="#B49-computers-13-00271" class="html-bibr">49</a>,<a href="#B50-computers-13-00271" class="html-bibr">50</a>,<a href="#B51-computers-13-00271" class="html-bibr">51</a>,<a href="#B52-computers-13-00271" class="html-bibr">52</a>,<a href="#B53-computers-13-00271" class="html-bibr">53</a>,<a href="#B54-computers-13-00271" class="html-bibr">54</a>,<a href="#B55-computers-13-00271" class="html-bibr">55</a>,<a href="#B56-computers-13-00271" class="html-bibr">56</a>].</p>
Full article ">Figure 5
<p>Galbraith Plot.</p>
Full article ">Figure 6
<p>Funnel Plot.</p>
Full article ">Figure 7
<p>Bubble Plot per Publication year.</p>
Full article ">Figure 8
<p>Bubble Plot for Dataset.</p>
Full article ">Figure 9
<p>Publications by Year.</p>
Full article ">Figure 10
<p>Publication per Year by Source.</p>
Full article ">Figure 11
<p>MCD by Country.</p>
Full article ">Figure 12
<p>Bibliometric Analysis on Network Visualization per cited Authors.</p>
Full article ">Figure 13
<p>Approaches per year.</p>
Full article ">Figure 14
<p>Models per year.</p>
Full article ">Figure 15
<p>Publisher per citations.</p>
Full article ">
32 pages, 614 KiB  
Review
Automatic Generation of Medical Case-Based Multiple-Choice Questions (MCQs): A Review of Methodologies, Applications, Evaluation, and Future Directions
by Somaiya Al Shuraiqi, Abdulrahman Aal Abdulsalam, Ken Masters, Hamza Zidoum and Adhari AlZaabi
Big Data Cogn. Comput. 2024, 8(10), 139; https://doi.org/10.3390/bdcc8100139 - 17 Oct 2024
Viewed by 268
Abstract
This paper offers an in-depth review of the latest advancements in the automatic generation of medical case-based multiple-choice questions (MCQs). The automatic creation of educational materials, particularly MCQs, is pivotal in enhancing teaching effectiveness and student engagement in medical education. In this review, [...] Read more.
This paper offers an in-depth review of the latest advancements in the automatic generation of medical case-based multiple-choice questions (MCQs). The automatic creation of educational materials, particularly MCQs, is pivotal in enhancing teaching effectiveness and student engagement in medical education. In this review, we explore various algorithms and techniques that have been developed for generating MCQs from medical case studies. Recent innovations in natural language processing (NLP) and machine learning (ML) for automatic language generation have garnered considerable attention. Our analysis evaluates and categorizes the leading approaches, highlighting their generation capabilities and practical applications. Additionally, this paper synthesizes the existing evidence, detailing the strengths, limitations, and gaps in current practices. By contributing to the broader conversation on how technology can support medical education, this review not only assesses the present state but also suggests future directions for improvement. We advocate for the development of more advanced and adaptable mechanisms to enhance the automatic generation of MCQs, thereby supporting more effective learning experiences in medical education. Full article
(This article belongs to the Special Issue Advances in Natural Language Processing and Text Mining)
Show Figures

Figure 1

Figure 1
<p>An example of medical ontology on COVID-19.</p>
Full article ">
13 pages, 852 KiB  
Review
Enhancing Thoracic Surgery with AI: A Review of Current Practices and Emerging Trends
by Mohamed Umair Aleem, Jibran Ahmad Khan, Asser Younes, Belal Nedal Sabbah, Waleed Saleh and Marcello Migliore
Curr. Oncol. 2024, 31(10), 6232-6244; https://doi.org/10.3390/curroncol31100464 - 17 Oct 2024
Viewed by 451
Abstract
Artificial intelligence (AI) is increasingly becoming integral to medical practice, potentially enhancing outcomes in thoracic surgery. AI-driven models have shown significant accuracy in diagnosing non-small-cell lung cancer (NSCLC), predicting lymph node metastasis, and aiding in the efficient extraction of electronic medical record (EMR) [...] Read more.
Artificial intelligence (AI) is increasingly becoming integral to medical practice, potentially enhancing outcomes in thoracic surgery. AI-driven models have shown significant accuracy in diagnosing non-small-cell lung cancer (NSCLC), predicting lymph node metastasis, and aiding in the efficient extraction of electronic medical record (EMR) data. Moreover, AI applications in robotic-assisted thoracic surgery (RATS) and perioperative management reveal the potential to improve surgical precision, patient safety, and overall care efficiency. Despite these advancements, challenges such as data privacy, biases, and ethical concerns remain. This manuscript explores AI applications, particularly machine learning (ML) and natural language processing (NLP), in thoracic surgery, emphasizing their role in diagnosis and perioperative management. It also provides a comprehensive overview of the current state, benefits, and limitations of AI in thoracic surgery, highlighting future directions in the field. Full article
(This article belongs to the Special Issue Artificial Intelligence in Thoracic Surgery)
Show Figures

Figure 1

Figure 1
<p>Subcategories of artificial intelligence [<a href="#B5-curroncol-31-00464" class="html-bibr">5</a>].</p>
Full article ">Figure 2
<p>Process and steps in artificial intelligence [<a href="#B5-curroncol-31-00464" class="html-bibr">5</a>]. * <span class="html-italic">Unsupervised methods are used for pattern recognition and clustering rather than generating predictive models</span>.</p>
Full article ">
20 pages, 10806 KiB  
Article
Advancing Mental Health Care: Intelligent Assessments and Automated Generation of Personalized Advice via M.I.N.I and RoBERTa
by Yuezhong Wu, Huan Xie, Lin Gu, Rongrong Chen, Shanshan Chen, Fanglan Wang, Yiwen Liu, Lingjiao Chen and Jinsong Tang
Appl. Sci. 2024, 14(20), 9447; https://doi.org/10.3390/app14209447 - 16 Oct 2024
Viewed by 408
Abstract
As mental health issues become increasingly prominent, we are now facing challenges such as the severe unequal distribution of medical resources and low diagnostic efficiency. This paper integrates finite state machines, retrieval algorithms, semantic-matching models, and medical-knowledge graphs to design an innovative intelligent [...] Read more.
As mental health issues become increasingly prominent, we are now facing challenges such as the severe unequal distribution of medical resources and low diagnostic efficiency. This paper integrates finite state machines, retrieval algorithms, semantic-matching models, and medical-knowledge graphs to design an innovative intelligent auxiliary evaluation tool and a personalized medical-advice generation application, aiming to improve the efficiency of mental health assessments and the provision of personalized medical advice. The main contributions include the folowing: (1) Developing an auxiliary diagnostic tool that combines the Mini-International Neuropsychiatric Interview (M.I.N.I.) with finite state machines to systematically collect patient information for preliminary assessments; (2) Enhancing data processing by optimizing retrieval algorithms for efficient filtering and employing a fine-tuned RoBERTa model for deep semantic matching and analysis, ensuring accurate and personalized medical-advice generation; (3) Generating intelligent suggestions using NLP techniques; when semantic matching falls below a specific threshold, integrating medical-knowledge graphs to produce general medical advice. Experimental results show that this application achieves a semantic-matching degree of 0.9 and an accuracy of 0.87, significantly improving assessment accuracy and the ability to generate personalized medical advice. This optimizes the allocation of medical resources, enhances diagnostic efficiency, and provides a reference for advancing mental health care through artificial-intelligence technology. Full article
Show Figures

Figure 1

Figure 1
<p>Overall architecture.</p>
Full article ">Figure 2
<p>Auxiliary assessment effect.</p>
Full article ">Figure 3
<p>Working principle diagram of finite state machine.</p>
Full article ">Figure 4
<p>Program flowchart.</p>
Full article ">Figure 5
<p>Keyword-weight matching. Letters a–f represent different weight values.</p>
Full article ">Figure 6
<p>Cross-Encoder structure.</p>
Full article ">Figure 7
<p>RoBERTa model based on Cross-Encoder.</p>
Full article ">Figure 8
<p>Medical advice generation process flowchart. Letters a–e represent different keywords.</p>
Full article ">Figure 9
<p>Knowledge generation process in the mental and psychological health domain-knowledge graph. Different colors represent entity categories, and different letters represent subcategories of the same entity.</p>
Full article ">Figure 10
<p>Weight distribution.</p>
Full article ">Figure 11
<p>Weight-matching score.</p>
Full article ">Figure 12
<p>Manhattan distance score.</p>
Full article ">Figure 13
<p>Euclidean distance score.</p>
Full article ">Figure 14
<p>Cosine value score.</p>
Full article ">Figure 15
<p>RoBERTa vs. FT-RoBERTa.</p>
Full article ">Figure 16
<p>Mental and psychological health domain medical-knowledge graph.</p>
Full article ">Figure 17
<p>Entity-related info.</p>
Full article ">Figure 18
<p>Attribute-related info.</p>
Full article ">Figure 19
<p>Entity relationship-related Info.</p>
Full article ">Figure 20
<p>Depression knowledge generation.</p>
Full article ">Figure 21
<p>Mania knowledge generation.</p>
Full article ">
13 pages, 1731 KiB  
Article
A Study on the Identification of the Water Army to Improve the Helpfulness of Online Product Reviews
by Chuyang Li, Shijia Zhang and Xiangdong Liu
Mathematics 2024, 12(20), 3234; https://doi.org/10.3390/math12203234 (registering DOI) - 16 Oct 2024
Viewed by 280
Abstract
Based on the perspective of identifying the water army, this paper uses the methods of machine learning and data visualization to analyze the helpfulness of online produce reviews, portray product portraits, and provide real and helpful product reviews. In order to identify and [...] Read more.
Based on the perspective of identifying the water army, this paper uses the methods of machine learning and data visualization to analyze the helpfulness of online produce reviews, portray product portraits, and provide real and helpful product reviews. In order to identify and eliminate the water army, the Term Frequency-Inverse Document Frequency Model (TF-IDF) and Latent Semantic Index Model (LSI) are used. After eliminating the water army, three classification methods were selected to perform sentimental analysis, including logistics, SnowNLP, and Convolutional Neural Network for text(TextCNN). The TextCNN has the highest F1 score among the three classification methods. At the same time, the Latent Dirichlet Allocation mode (LDA) is used to extract the topics of various reviews. Finally, targeted countermeasures are proposed to manufacturers, consumers, and regulators. Full article
(This article belongs to the Special Issue Big Data Mining and Analytics with Applications)
Show Figures

Figure 1

Figure 1
<p>Sentiment analysis process.</p>
Full article ">Figure 2
<p>CNN training process.</p>
Full article ">Figure 3
<p>Sentiment analysis score chart. The horizontal axis represents the ID of 4561 users, and the vertical axis represents the score.</p>
Full article ">Figure 4
<p>Cosine similarity for both positive and negative reviews. The horizontal axis measures a number of topics, and the vertical axis measures cosine similarity.</p>
Full article ">
15 pages, 2908 KiB  
Study Protocol
Exploring Natural Language Processing through an Exemplar Using YouTube
by Joohyun Chung, Sangmin Song and Heesook Son
Int. J. Environ. Res. Public Health 2024, 21(10), 1357; https://doi.org/10.3390/ijerph21101357 - 15 Oct 2024
Viewed by 329
Abstract
There has been a growing emphasis on data across various health-related fields, not just in nursing research, due to the increasing volume of unstructured data in electronic health records (EHRs). Natural Language Processing (NLP) provides a solution by transforming this unstructured data into [...] Read more.
There has been a growing emphasis on data across various health-related fields, not just in nursing research, due to the increasing volume of unstructured data in electronic health records (EHRs). Natural Language Processing (NLP) provides a solution by transforming this unstructured data into structured formats, thereby facilitating valuable insights. This methodology paper explores the application of NLP in nursing, using an exemplar case study that analyzes YouTube data to investigate social phenomena among adults living alone. The methodology involves five steps: accessing data through YouTube’s API, data cleaning, preprocessing (tokenization, sentence segmentation, linguistic normalization), sentiment analysis using Python, and topic modeling. This study serves as a comprehensive guide for integrating NLP into nursing research, supplemented with digital content demonstrating each step. For successful implementation, nursing researchers must grasp the fundamental concepts and processes of NLP. The potential of NLP in nursing is significant, particularly in utilizing unstructured textual data from nursing documentation and social media. Its benefits include streamlining nursing documentation, enhancing patient communication, and improving data analysis. Full article
(This article belongs to the Section Health Care Sciences)
Show Figures

Figure 1

Figure 1
<p>The five steps of NLP.</p>
Full article ">Figure 2
<p>Conceptualization of Text representation.</p>
Full article ">Figure 3
<p>Visualization–Intertopic Distance Map.</p>
Full article ">Figure 4
<p>Visualization–Bar charts of selected topics.</p>
Full article ">Figure 5
<p>Intertopic distance map and topic probability for males.</p>
Full article ">Figure 6
<p>Intertopic distance map and topic probability for females.</p>
Full article ">
33 pages, 2537 KiB  
Review
AI Applications in Adult Stroke Recovery and Rehabilitation: A Scoping Review Using AI
by Isuru Senadheera, Prasad Hettiarachchi, Brendon Haslam, Rashmika Nawaratne, Jacinta Sheehan, Kylee J. Lockwood, Damminda Alahakoon and Leeanne M. Carey
Sensors 2024, 24(20), 6585; https://doi.org/10.3390/s24206585 - 12 Oct 2024
Viewed by 505
Abstract
Stroke is a leading cause of long-term disability worldwide. With the advancements in sensor technologies and data availability, artificial intelligence (AI) holds the promise of improving the amount, quality and efficiency of care and enhancing the precision of stroke rehabilitation. We aimed to [...] Read more.
Stroke is a leading cause of long-term disability worldwide. With the advancements in sensor technologies and data availability, artificial intelligence (AI) holds the promise of improving the amount, quality and efficiency of care and enhancing the precision of stroke rehabilitation. We aimed to identify and characterize the existing research on AI applications in stroke recovery and rehabilitation of adults, including categories of application and progression of technologies over time. Data were collected from peer-reviewed articles across various electronic databases up to January 2024. Insights were extracted using AI-enhanced multi-method, data-driven techniques, including clustering of themes and topics. This scoping review summarizes outcomes from 704 studies. Four common themes (impairment, assisted intervention, prediction and imaging, and neuroscience) were identified, in which time-linked patterns emerged. The impairment theme revealed a focus on motor function, gait and mobility, while the assisted intervention theme included applications of robotic and brain–computer interface (BCI) techniques. AI applications progressed over time, starting from conceptualization and then expanding to a broader range of techniques in supervised learning, artificial neural networks (ANN), natural language processing (NLP) and more. Applications focused on upper limb rehabilitation were reviewed in more detail, with machine learning (ML), deep learning techniques and sensors such as inertial measurement units (IMU) used for upper limb and functional movement analysis. AI applications have potential to facilitate tailored therapeutic delivery, thereby contributing to the optimization of rehabilitation outcomes and promoting sustained recovery from rehabilitation to real-world settings. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

Figure 1
<p>Screen captures of the interactive dashboard. Key data features depicted in the dashboard include: AI technology, clinical methodology, time-linked patterns and network analysis.</p>
Full article ">Figure 2
<p>Flow diagram of studies included in the scoping review.</p>
Full article ">Figure 3
<p>Topic clusters to characterize the research landscape of AI applications in adult stroke rehabilitation and recovery. Clusters are represented by different colors.</p>
Full article ">Figure 4
<p>Landscape of AI techniques in stroke rehabilitation and recovery. The thickness of the lines indicates the usage of major categories of AI techniques. Yellow boxes indicate AI technique categories, and red boxes indicate specific AI techniques.</p>
Full article ">Figure 5
<p>Progression of AI techniques applied in adult stroke rehabilitation and recovery research. Markers indicate the time of the first emergence of AI techniques.</p>
Full article ">Figure 6
<p>Timeline of AI applications in adult stroke rehabilitation compared to evolution of AI technologies. Yellow boxes (top) indicate the time of the invention of the AI technique. Green boxes (bottom) indicate the first application of AI technology in adult post-stroke rehabilitation and recovery.</p>
Full article ">
18 pages, 827 KiB  
Article
Zero-Shot Learning for Accurate Project Duration Prediction in Crowdsourcing Software Development
by Tahir Rashid, Inam Illahi, Qasim Umer, Muhammad Arfan Jaffar, Waheed Yousuf Ramay and Hanadi Hakami
Computers 2024, 13(10), 266; https://doi.org/10.3390/computers13100266 - 12 Oct 2024
Viewed by 290
Abstract
Crowdsourcing Software Development (CSD) platforms, i.e., TopCoder, function as intermediaries connecting clients with developers. Despite employing systematic methodologies, these platforms frequently encounter high task abandonment rates, with approximately 19% of projects failing to meet satisfactory outcomes. Although existing research has focused on task [...] Read more.
Crowdsourcing Software Development (CSD) platforms, i.e., TopCoder, function as intermediaries connecting clients with developers. Despite employing systematic methodologies, these platforms frequently encounter high task abandonment rates, with approximately 19% of projects failing to meet satisfactory outcomes. Although existing research has focused on task scheduling, developer recommendations, and reward mechanisms, there has been insufficient attention to the support of platform moderators, or copilots, who are essential to project success. A critical responsibility of copilots is estimating project duration; however, manual predictions often lead to inconsistencies and delays. This paper introduces an innovative machine learning approach designed to automate the prediction of project duration on CSD platforms. Utilizing historical data from TopCoder, the proposed method extracts pertinent project attributes and preprocesses textual data through Natural Language Processing (NLP). Bidirectional Encoder Representations from Transformers (BERT) are employed to convert textual information into vectors, which are then analyzed using various machine learning algorithms. Zero-shot learning algorithms exhibit superior performance, with an average accuracy of 92.76%, precision of 92.76%, recall of 99.33%, and an f-measure of 95.93%. The implementation of the proposed automated duration prediction model is crucial for enhancing the success rate of crowdsourcing projects, optimizing resource allocation, managing budgets effectively, and improving stakeholder satisfaction. Full article
(This article belongs to the Special Issue Best Practices, Challenges and Opportunities in Software Engineering)
Show Figures

Figure 1

Figure 1
<p>Success and failure distribution of CCSD projects.</p>
Full article ">Figure 2
<p>Proposed model.</p>
Full article ">Figure 3
<p>Spider Graph: Performance comparison against alternative approaches.</p>
Full article ">Figure 4
<p>Ridge Graph: Comparison against different machine learning algorithms.</p>
Full article ">
27 pages, 920 KiB  
Article
AI-Generated Spam Review Detection Framework with Deep Learning Algorithms and Natural Language Processing
by Mudasir Ahmad Wani, Mohammed ElAffendi and Kashish Ara Shakil
Computers 2024, 13(10), 264; https://doi.org/10.3390/computers13100264 - 12 Oct 2024
Viewed by 365
Abstract
Spam reviews pose a significant challenge to the integrity of online platforms, misleading consumers and undermining the credibility of genuine feedback. This paper introduces an innovative AI-generated spam review detection framework that leverages Deep Learning algorithms and Natural Language Processing (NLP) techniques to [...] Read more.
Spam reviews pose a significant challenge to the integrity of online platforms, misleading consumers and undermining the credibility of genuine feedback. This paper introduces an innovative AI-generated spam review detection framework that leverages Deep Learning algorithms and Natural Language Processing (NLP) techniques to identify and mitigate spam reviews effectively. Our framework utilizes multiple Deep Learning models, including Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, Gated Recurrent Unit (GRU), and Bidirectional LSTM (BiLSTM), to capture intricate patterns in textual data. The system processes and analyzes large volumes of review content to detect deceptive patterns by utilizing advanced NLP and text embedding techniques such as One-Hot Encoding, Word2Vec, and Term Frequency-Inverse Document Frequency (TF-IDF). By combining three embedding techniques with four Deep Learning algorithms, a total of twelve exhaustive experiments were conducted to detect AI-generated spam reviews. The experimental results demonstrate that our approach outperforms the traditional machine learning models, offering a robust solution for ensuring the authenticity of online reviews. Among the models evaluated, those employing Word2Vec embeddings, particularly the BiLSTM_Word2Vec model, exhibited the strongest performance. The BiLSTM model with Word2Vec achieved the highest performance, with an exceptional accuracy of 98.46%, a precision of 0.98, a recall of 0.97, and an F1-score of 0.98, reflecting a near-perfect balance between precision and recall. Its high F2-score (0.9810) and F0.5-score (0.9857) further highlight its effectiveness in accurately detecting AI-generated spam while minimizing false positives, making it the most reliable option for this task. Similarly, the Word2Vec-based LSTM model also performed exceptionally well, with an accuracy of 97.58%, a precision of 0.97, a recall of 0.96, and an F1-score of 0.97. The CNN model with Word2Vec similarly delivered strong results, achieving an accuracy of 97.61%, a precision of 0.97, a recall of 0.96, and an F1-score of 0.97. This study is unique in its focus on detecting spam reviews specifically generated by AI-based tools rather than solely detecting spam reviews or AI-generated text. This research contributes to the field of spam detection by offering a scalable, efficient, and accurate framework that can be integrated into various online platforms, enhancing user trust and the decision-making processes. Full article
Show Figures

Figure 1

Figure 1
<p>Detailed data collection procedure.</p>
Full article ">Figure 2
<p>Generating AI-based spam/fake reviews based on human-authored samples.</p>
Full article ">Figure 3
<p>Check for the working of GPT Module.</p>
Full article ">Figure 4
<p>Data preparation and preprocessing with NLTK toolkit.</p>
Full article ">Figure 5
<p>Experimental setup and configuration.</p>
Full article ">Figure 6
<p>Performance of selected Deep Learning models on TF-IDF representation.</p>
Full article ">Figure 7
<p>Performance of selected Deep Learning models on Word2Vec feature representation.</p>
Full article ">Figure 8
<p>Performance of selected Deep Learning models on One-Hot Encoding.</p>
Full article ">Figure 9
<p>The radar plot showing proposed approaches. Particularly, Word2Vec-based BiLSTM outperformed the existing methods.</p>
Full article ">Figure 10
<p>Heptagon: seven ways to prevent abuse and ensure ethical use of AI-generated reviews.</p>
Full article ">
14 pages, 563 KiB  
Article
An NLP-Based Perfume Note Estimation Based on Descriptive Sentences
by Jooyoung Kim, Kangrok Oh and Beom-Seok Oh
Appl. Sci. 2024, 14(20), 9293; https://doi.org/10.3390/app14209293 - 12 Oct 2024
Viewed by 421
Abstract
The perfume industry is a suitable candidate for applying advanced natural language processing techniques, yet most existing studies focus on developing fragrance design systems based on artificial intelligence advances. To meet the increasing demand for analyzing and exploiting descriptive sentences for the fragrance [...] Read more.
The perfume industry is a suitable candidate for applying advanced natural language processing techniques, yet most existing studies focus on developing fragrance design systems based on artificial intelligence advances. To meet the increasing demand for analyzing and exploiting descriptive sentences for the fragrance market, we investigate the relationship between descriptive sentences of perfumes and their notes in this paper. Our purpose for this investigation is to build a core idea for a perfume recommendation system of descriptive sentences. To accomplish this, we propose a system for perfume note estimation of descriptive sentences based on several sentence transformer models. In our leave-one-out cross-validation tests using our dataset containing 62 perfumes and 255 perfume notes, we achieved significant performance improvements (from a 37.1∼41.1% to 72.6∼79.0% hit rate with the top five items, and from a 22.1∼31.9% to a 57.3∼63.2% mean reciprocal rank) for perfume note estimation via our fine-tuning process. In addition, some qualitative examples, including query descriptions, estimated perfume notes, and the ground truth perfume notes, are presented. The proposed system improves the perfume note estimation performances using a fine-tuning process on a newly constructed dataset containing descriptive sentences of perfumes and their notes. Full article
(This article belongs to the Special Issue Natural Language Processing (NLP) and Applications—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Abstraction of the website.</p>
Full article ">Figure 2
<p>Count description of the perfume notes in the dataset.</p>
Full article ">Figure 3
<p>An overview of the proposed system with an example of perfume Barbae.</p>
Full article ">Figure 4
<p>CMC curves representing the Hit<math display="inline"><semantics> <mrow> <mo>@</mo> <mi>k</mi> </mrow> </semantics></math> performances of the investigated sentence embedding models with and without the proposed fine-tuning strategy.</p>
Full article ">
18 pages, 893 KiB  
Article
Temporal Relational Graph Convolutional Network Approach to Financial Performance Prediction
by Brindha Priyadarshini Jeyaraman, Bing Tian Dai and Yuan Fang
Mach. Learn. Knowl. Extr. 2024, 6(4), 2303-2320; https://doi.org/10.3390/make6040113 - 10 Oct 2024
Viewed by 575
Abstract
Accurately predicting financial entity performance remains a challenge due to the dynamic nature of financial markets and vast unstructured textual data. Financial knowledge graphs (FKGs) offer a structured representation for tackling this problem by representing complex financial relationships and concepts. However, constructing a [...] Read more.
Accurately predicting financial entity performance remains a challenge due to the dynamic nature of financial markets and vast unstructured textual data. Financial knowledge graphs (FKGs) offer a structured representation for tackling this problem by representing complex financial relationships and concepts. However, constructing a comprehensive and accurate financial knowledge graph that captures the temporal dynamics of financial entities is non-trivial. We introduce FintechKG, a comprehensive financial knowledge graph developed through a three-dimensional information extraction process that incorporates commercial entities and temporal dimensions and uses a financial concept taxonomy that ensures financial domain entity and relationship extraction. We propose a temporal and relational graph convolutional network (RGCN)-based representation for FintechKG data across multiple timesteps, which captures temporal dependencies. This representation is then combined with FinBERT embeddings through a projection layer, enabling a richer feature space. To demonstrate the efficacy of FintechKG, we evaluate its performance using the example task of financial performance prediction. A logistic regression model uses these combined features and social media embeddings for performance prediction. We classify whether the revenue will increase or decrease. This approach demonstrates the effectiveness of FintechKG combined with textual information for accurate financial forecasting. Our work contributes a systematic FKG construction method and a framework that utilizes both relational and textual embeddings for improved financial performance prediction. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

Figure 1
<p>FintechKG knowledge graph extraction pipeline.</p>
Full article ">Figure 2
<p>Taxonomy of income, liabilities, and assets.</p>
Full article ">Figure 3
<p>Relation types.</p>
Full article ">Figure 4
<p>BERT NER model.</p>
Full article ">Figure 5
<p>Knowledge graph for tweets: “In 2005 Cargotec’s net sales exceeded EUR 2.3 billion.”, “When this information was released on 5 September 2008, Nokia’s American Depositary shares fell by 8%”, “Orion Pharma ’s operating profit increased by 42.5% from 2004”. and “Nordea Group’s operating profit increased in 2010 by 18 percent year-on-year to 3.64 billion euros and total revenue by 3 percent to 9.33 billion euros”.</p>
Full article ">Figure 6
<p>RGCN architecture capturing temporal information.</p>
Full article ">Figure 7
<p>LSTM with RGCN embedding architecture.</p>
Full article ">
15 pages, 2357 KiB  
Article
Dynamic Multi-Granularity Translation System: DAG-Structured Multi-Granularity Representation and Self-Attention
by Shenrong Lv, Bo Yang, Ruiyang Wang, Siyu Lu, Jiawei Tian, Wenfeng Zheng, Xiaobing Chen and Lirong Yin
Systems 2024, 12(10), 420; https://doi.org/10.3390/systems12100420 - 9 Oct 2024
Viewed by 416
Abstract
In neural machine translation (NMT), the sophistication of word embeddings plays a pivotal role in the model’s ability to render accurate and contextually relevant translations. However, conventional models with single granularity of word segmentation cannot fully embed complex languages like Chinese, where the [...] Read more.
In neural machine translation (NMT), the sophistication of word embeddings plays a pivotal role in the model’s ability to render accurate and contextually relevant translations. However, conventional models with single granularity of word segmentation cannot fully embed complex languages like Chinese, where the granularity of segmentation significantly impacts understanding and translation fidelity. Addressing these challenges, our study introduces the Dynamic Multi-Granularity Translation System (DMGTS), an innovative approach that enhances the Transformer model by incorporating multi-granularity position encoding and multi-granularity self-attention mechanisms. Leveraging a Directed Acyclic Graph (DAG), the DMGTS utilizes four levels of word segmentation for multi-granularity position encoding. Dynamic word embeddings are also introduced to enhance the lexical representation by incorporating multi-granularity features. Multi-granularity self-attention mechanisms are applied to replace the conventional self-attention layers. We evaluate the DMGTS on multiple datasets, where our system demonstrates marked improvements. Notably, it achieves significant enhancements in translation quality, evidenced by increases of 1.16 and 1.55 in Bilingual Evaluation Understudy (BLEU) scores over traditional static embedding methods. These results underscore the efficacy of the DMGTS in refining NMT performance. Full article
(This article belongs to the Section Artificial Intelligence and Digital Systems Engineering)
Show Figures

Figure 1

Figure 1
<p>Architecture of the proposed DMGTS.</p>
Full article ">Figure 2
<p>Performance comparison of different word segmentation methods.</p>
Full article ">Figure 3
<p>Comparison of output encoding procedures with and without adding relative positions. (<b>a</b>) SA; (<b>b</b>) the output encoding of the first “I”; (<b>c</b>) the output encoding of the second “I”.</p>
Full article ">Figure 4
<p>Multi-granularity position encoding using DAG.</p>
Full article ">Figure 5
<p>Multi-granularity relative distance matrix.</p>
Full article ">
20 pages, 1853 KiB  
Article
Chinese Named Entity Recognition Based on Multi-Level Representation Learning
by Weijun Li, Jianping Ding, Shixia Liu, Xueyang Liu, Yilei Su and Ziyi Wang
Appl. Sci. 2024, 14(19), 9083; https://doi.org/10.3390/app14199083 - 8 Oct 2024
Viewed by 551
Abstract
Named Entity Recognition (NER) is a crucial component of Natural Language Processing (NLP). When dealing with the high diversity and complexity of the Chinese language, existing Chinese NER models face challenges in addressing word sense ambiguity, capturing long-range dependencies, and maintaining robustness, which [...] Read more.
Named Entity Recognition (NER) is a crucial component of Natural Language Processing (NLP). When dealing with the high diversity and complexity of the Chinese language, existing Chinese NER models face challenges in addressing word sense ambiguity, capturing long-range dependencies, and maintaining robustness, which hinders the accuracy of entity recognition. To this end, a Chinese NER model based on multi-level representation learning is proposed. The model leverages a pre-trained word-based embedding to capture contextual information. A linear layer adjusts dimensions to fit an Extended Long Short-Term Memory (XLSTM) network, enabling the capture of long-range dependencies and contextual information, and providing deeper representations. An adaptive multi-head attention mechanism is proposed to enhance the ability to capture global dependencies and comprehend deep semantic context. Additionally, GlobalPointer with rotational position encoding integrates global information for entity category prediction. Projected Gradient Descent (PGD) is incorporated, introducing perturbations in the embedding layer of the pre-trained model to enhance stability in noisy environments. The proposed model achieves F1-scores of 96.89%, 74.89%, 72.19%, and 80.96% on the Resume, Weibo, CMeEE, and CLUENER2020 datasets, respectively, demonstrating improvements over baseline and comparison models. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>WP-XAG Model.</p>
Full article ">Figure 2
<p>mLSTM block structure diagram.</p>
Full article ">Figure 3
<p>Adaptive multi-head attention structure diagram.</p>
Full article ">Figure 4
<p>GlobalPointer for entity tagging.</p>
Full article ">Figure 5
<p>Model iteration comparison.</p>
Full article ">
9 pages, 543 KiB  
Commentary
Qualitative Health-Related Quality of Life and Natural Language Processing: Characteristics, Implications, and Challenges
by Esther Lázaro and Vanessa Moscardó
Healthcare 2024, 12(19), 2008; https://doi.org/10.3390/healthcare12192008 - 8 Oct 2024
Viewed by 407
Abstract
Objectives: This article focuses on describing the main characteristics of the application of NLP in the qualitative assessment of quality of life, as well as its implications and challenges. Methods: The qualitative methodology allows analysing patient comments in unstructured free text and obtaining [...] Read more.
Objectives: This article focuses on describing the main characteristics of the application of NLP in the qualitative assessment of quality of life, as well as its implications and challenges. Methods: The qualitative methodology allows analysing patient comments in unstructured free text and obtaining valuable information through manual analysis of these data. However, large amounts of data are a healthcare challenge since it would require a high number of staff and time resources that are not available in most healthcare organizations. Results: One potential solution to mitigate the resource constraints of qualitative analysis is the use of machine learning and artificial intelligence, specifically methodologies based on natural language processing. Full article
Show Figures

Figure 1

Figure 1
<p>Phases in the process of developing the NLP-QoL tool. 1. Bibliography review; 2. Data sources; 3. Knowledge creation; 4. Knowledge translation; 5. Deployment for making-decision.</p>
Full article ">
Back to TopTop