[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (5,606)

Search Parameters:
Keywords = prediction task

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
30 pages, 2248 KiB  
Article
Novel Model for Pork Supply Prediction in China Based on Modified Self-Organizing Migrating Algorithm
by Haohao Song, Jiquan Wang, Gang Xu, Zhanwei Tian, Fei Xu and Hong Deng
Agriculture 2024, 14(9), 1592; https://doi.org/10.3390/agriculture14091592 - 12 Sep 2024
Abstract
Pork supply prediction is a challenging task of significant importance for pig producers and administrators, as it aids decision-making and maintains the pork supply–demand balance. Previous studies failed to consider impact factors like the month-age transfer principle of pigs, epidemic factors, and the [...] Read more.
Pork supply prediction is a challenging task of significant importance for pig producers and administrators, as it aids decision-making and maintains the pork supply–demand balance. Previous studies failed to consider impact factors like the month-age transfer principle of pigs, epidemic factors, and the simultaneous import and export volumes of pork, leading to the absence of a quantitative prediction model for pork supply. In this background, we proposed a novel quantitative prediction model of pork supply that incorporates pork production and pork import/export volumes. First, a prediction model for pork production that takes into account the month-age transfer principle of pigs and epidemic factors was presented, along with a recursive model of the pig-herd system. A novel method based on a modified self-organizing migrating algorithm (MSOMA) was proposed for calculating the quantity of monthly newly retained sows (NRS). Furthermore, the pork-production prediction model considered the epidemic factor as a random disturbance term (RDT), and a prediction method based on MSOMA and a back-propagation neural network (MSOMA-BPNN) was introduced to predict such disturbance terms. Second, the proposed MSOMA-BPNN was employed to predict pork import and export volumes. The pork supply was subsequently determined based on the predicted pork production, as well as the pork import and export volumes. The proposed pork supply prediction model was applied to forecast China’s pork supply from 2010 to 2023. The results validate the high effectiveness and reliability of the proposed model, providing valuable insights for decision makers. The empirical results demonstrate that the proposed model is a promising and effective tool for predicting the pork supply. To our knowledge, this is a novel tool for pork supply prediction, considering the pig-herd system and pork import and export volumes from a systemic perspective. These features allow for consideration of the scientific formulation of a pig production plan, the establishment of early warning mechanisms to deal with epidemic situations and emergencies, and the regulation of pork supply and demand balance. Full article
(This article belongs to the Section Agricultural Economics, Policies and Rural Management)
Show Figures

Figure 1

Figure 1
<p>The month-age transfer of pigs.</p>
Full article ">Figure 2
<p>The flowchart of MSOMA.</p>
Full article ">Figure 3
<p>The process of self-organizing.</p>
Full article ">Figure 4
<p>The structure of MSOMA-BPNN for the prediction of RDTs.</p>
Full article ">Figure 5
<p>The framework of proposed pork supply prediction model.</p>
Full article ">
31 pages, 5286 KiB  
Article
“It’s Different for Girls!” The Role of Anxiety, Physiological Arousal, and Subject Preferences in Primary School Children’s Math and Mental Rotation Performance
by Michelle Lennon-Maslin and Claudia Michaela Quaiser-Pohl
Behav. Sci. 2024, 14(9), 809; https://doi.org/10.3390/bs14090809 - 12 Sep 2024
Abstract
(1) Background: This study examines the role of subjective anxiety (mathematics and spatial anxiety), along with physiological responses, in mathematics or math and mental rotation performance in 131 German primary school students (65 girls, 66 boys; Mean age = 8.73 years). (2) Method: [...] Read more.
(1) Background: This study examines the role of subjective anxiety (mathematics and spatial anxiety), along with physiological responses, in mathematics or math and mental rotation performance in 131 German primary school students (65 girls, 66 boys; Mean age = 8.73 years). (2) Method: Students’ preference for math vs. German and their subjective anxiety were assessed using standardized questionnaires. Emotional reactivity was measured using the Galvanic Skin Response (GSR). Math performance was evaluated via percentage scored and completion times on number line estimation, word problems, and missing terms tasks. Spatial skills were assessed using a novel mental rotation task (nMRT) incorporating gender-congruent and -neutral stimuli. (3) Results: Girls outperformed boys on percentage scored on the math task but took longer to complete this. No gender differences were found in performance on the nMRT. Girls demonstrated higher math anxiety and were less likely to prefer math over German. Math anxiety predicted math scores and accuracy on the nMRT while gender predicted math performance and mental rotation response time. Subject preference was associated with longer completion times and emotional reactivity with longer response times. Girls’ preference for math and lower emotional reactivity were linked to shorter completion times, while lower math anxiety predicted higher scores. In contrast, these factors did not affect boys’ math performance. Additionally, subjective anxiety, emotional reactivity, or subject preference did not impact spatial performance for either gender. (4) Conclusions: Supporting mathematical self-efficacy and emotional regulation, especially in girls, is crucial for enhancing STEM outcomes in primary education. Gender-fair assessment in mental rotation reveals equitable spatial performance and reduces the impact of anxiety. Full article
Show Figures

Figure 1

Figure 1
<p>Examples of MRT items used with concrete objects: (<b>A</b>) animal stimulus and (<b>B</b>) letter stimulus [<a href="#B85-behavsci-14-00809" class="html-bibr">85</a>]; (<b>C</b>) female-stereotyped stimulus “Pram” [<a href="#B21-behavsci-14-00809" class="html-bibr">21</a>,<a href="#B23-behavsci-14-00809" class="html-bibr">23</a>]; and with an abstract object (<b>D</b>) “Cube” figure [<a href="#B22-behavsci-14-00809" class="html-bibr">22</a>].</p>
Full article ">Figure 2
<p>Gender (sex) differences in percentage scored (blue bar = Math_PT) and completion times (red bar = Math_CT) on the math task. Error bars represent the standard error of the mean (SEM). The error bars indicate similar variability in scores and completion times between boys and girls.</p>
Full article ">Figure 3
<p>Percentage of math vs. German as the preferred school subject by gender (sex).</p>
Full article ">Figure 4
<p>Effects of math anxiety on percentage scores and emotional reactivity (GSR) on completion times for girls. X-Axis = standardized scores; Y-Axis = percentage scores and completion times; blue circles: percentage scores (Math_PT) vs. math anxiety (MA); blue line: trend line for percentage scores (Math_PT); red circles: completion times (Math_CT) vs. emotional reactivity (ZGSR_Math); red line: trend line for completion times (Math_CT).</p>
Full article ">
17 pages, 4616 KiB  
Article
Machine Learning-Based Classification of Body Imbalance and Its Intensity Using Electromyogram and Ground Reaction Force in Immersive Environments
by Jahan Zeb Gul, Muhammad Omar Cheema, Zia Mohy Ud Din, Maryam Khan, Woo Young Kim and Muhammad Muqeet Rehman
Appl. Sci. 2024, 14(18), 8209; https://doi.org/10.3390/app14188209 - 12 Sep 2024
Abstract
Body balancing is a complex task that includes the coordination of muscles, tendons, bones, ears, eyes, and the brain. Imbalance or disequilibrium is the inability to maintain the center of gravity. Perpetuating body balance plays an important role in preventing us from falling [...] Read more.
Body balancing is a complex task that includes the coordination of muscles, tendons, bones, ears, eyes, and the brain. Imbalance or disequilibrium is the inability to maintain the center of gravity. Perpetuating body balance plays an important role in preventing us from falling or swaying. Biomechanical tests and video analysis can be performed to analyze body imbalance. The musculoskeletal system is one of the fundamental systems by which our balance or equilibrium is sustained and our upright posture is maintained. Electromyogram (EMG) and ground reaction force (GRF) monitoring can be utilized in cases where a rapid response to body imbalance is a necessity. Body balance also depends on visual stimuli that can be either real or virtual. Researchers have used virtual reality (VR) to predict motion sickness and analyze heart rate variability, as well as in rehabilitation. VR can also be used to induce body imbalance in a controlled way. In this research, body imbalance was induced in a controlled way by playing an Oculus game and, simultaneously, EMG and GRF were recorded. Features were extracted from the EMG and were then fed to a machine learning algorithm. Several machine learning algorithms were tested and upon 10-fold cross-validation; a minimum accuracy of 71% and maximum accuracy of 98% were achieved by Gaussian Naïve Bayes and Gradient Boosting classifiers, respectively, in the classification of imbalance and its intensities. This research can be incorporated into various rehabilitative and therapeutic systems. Full article
Show Figures

Figure 1

Figure 1
<p>Experimental setup and paradigm: The subject first stands on the force platform, keeping the distance equal by maintaining the center of gravity. Electrodes are placed on the vastus lateralis muscle of both legs to record EMG signals. Oculus Quest 2 is worn on the head of the subject and controllers are given in the hands of the subject. Controllers are used to control the roller coaster ride in a virtual environment; meanwhile, EMG and GRF are recorded simultaneously using the BIOPAC MP 36 (BIOPAC Systems, Inc., Goleta, CA, USA) and PASCO force platform (PASCO Scientific, Roseville, CA, USA), respectively.</p>
Full article ">Figure 2
<p>Complete graphical representation of this research: The whole paradigm of the study has been pictorially represented here. The subject wears the Oculus Quest 2 to play the <span class="html-italic">Epic Roller Coasters</span> game; meanwhile, his EMG and GRF signal for each leg are recorded using BIOPAC MP 36 and PASCO force platform. The acquired data are processed to extract features out of the signals and give them labels. The features and labels are fed to the machine learning classifiers, which classify the body imbalance intensities as high, medium, and low using the EMG and GRF signals. The performance of classifiers is evaluated and compared.</p>
Full article ">Figure 3
<p>Data acquisition and filtration: Digital filters of BIOPAC MP 36 are used to filter the EMG signals. The sampling rate of both the BIOPAC MP 36 and PASCO force platform is 1000 Hz.</p>
Full article ">Figure 4
<p>Virtual reality immersion: Oculus Quest 2 by Meta is used to create a virtual reality immersive environment for the subject. The game named <span class="html-italic">Epic Roller Coasters</span> is used to make the subject lose his balance in a controlled way. The mode of this game is Rock Falls.</p>
Full article ">Figure 5
<p>Changes observed in both the GRF and EMG: While playing the virtual reality game, changes in the GRF and EMG signals can be observed. It can also be observed that both signals vary when the subject sways in any direction, like forward or backward, or when they are still.</p>
Full article ">Figure 6
<p>Feature extraction and machine learning classification: Signals of both modalities, i.e., EMG and GRF, are first normalized. Features are extracted from both signals. Features extracted from EMG are the mean, standard deviation, skewness, and kurtosis. Features extracted from GRF are dominant frequency and force envelope. Features of both modalities are concatenated after extraction along with their labels, i.e., high, medium, and low intensities. Features and labels are then fed to the machine learning classifiers which classify the body imbalance intensities into high, medium, and low using the GRF and EMG signals of the subject.</p>
Full article ">Figure 7
<p>EMG acquisition and power spectral analysis: (<b>A</b>) The electrode placement at the vastus lateralis muscle to acquire EMG is illustrated. The ground electrode is placed near the ankle of the subject. (<b>B</b>) A small chunk of the acquired EMG signal is plotted. There is time in seconds on the x-axis and amplitude in millivolts on the y-axis. (<b>C</b>) The plot shows the power spectrum and its density mapped against the frequency values of the entire EMG signal. (<b>D</b>) This is a fast Fourier transform plot of the entire EMG signal showing the frequency components in the EMG signal and their magnitudes. (<b>E</b>) Short-Time Fourier Transform of the small chunk of EMG signal is illustrated in the figure. It has frequency and power spectral density plotted against the time in seconds. The high-power spectral density during the action of the vastus lateralis muscle is shown in red color and the low density is shown in blue color. (<b>F</b>) A plot without color mapping is shown in the figure, having the values of power per frequency component plotted against the frequency.</p>
Full article ">Figure 8
<p>Average 10-fold cross validation accuracies and error bars: (<b>A</b>) Average cross-validation accuracies of machine learning classifiers are illustrated in a bar chart. The highest 10-fold cross-validation accuracies are achieved by Decision Tree, Random Forest, Gradient Boosting, AdaBoost, and Bagging classifiers. (<b>B</b>) Average cross-validation accuracies are illustrated along with their error bars. Error is negligible in the case of Naïve Bayes, Decision Tree, Random Forest, Gradient Boosting, AdaBoost, and Bagging classifiers.</p>
Full article ">Figure 9
<p>Confusion matrices obtained after classification: Confusion matrices are illustrated in (<b>A</b>–<b>H</b>) subplots for different machine learning classifiers like k-Nearest Neighbors, Linear Discriminant Analysis, Gaussian Naïve Bayes, Decision Tree, Gradient Boosting, Multilayer Perceptron, AdaBoost and Bagging classifiers, respectively. The diagonal boxes in shades of dark and light blue color represent true values predicted correctly by the classifiers corresponding to the high, medium, and low classes.</p>
Full article ">Figure 10
<p>Evaluation metrics of classifiers upon classification: (<b>A</b>) Different performance metrics like accuracy, precision, recall, and F1 score are shown. The highest values of accuracy, precision, recall, and F1 score are achieved by Decision Tree, Random Forest, Gradient Boosting, Multilayer Perceptron, and Bagging classifiers, respectively, while they are lowest amongst others in the case of the Gaussian Naïve Bayes classifier. (<b>B</b>) There is overfitting in the case of the Random Forest classifier because it has a <span class="html-italic">p</span>-value less than 0.05 upon conducting a statistical test called the <span class="html-italic">t</span>-Test.</p>
Full article ">
27 pages, 3028 KiB  
Article
A New Predictive Method for Classification Tasks in Machine Learning: Multi-Class Multi-Label Logistic Model Tree (MMLMT)
by Bita Ghasemkhani, Kadriye Filiz Balbal and Derya Birant
Mathematics 2024, 12(18), 2825; https://doi.org/10.3390/math12182825 - 12 Sep 2024
Abstract
This paper introduces a novel classification method for multi-class multi-label datasets, named multi-class multi-label logistic model tree (MMLMT). Our approach supports multi-label learning to predict multiple class labels simultaneously, thereby enhancing the model’s capacity to capture complex relationships within the data. The primary [...] Read more.
This paper introduces a novel classification method for multi-class multi-label datasets, named multi-class multi-label logistic model tree (MMLMT). Our approach supports multi-label learning to predict multiple class labels simultaneously, thereby enhancing the model’s capacity to capture complex relationships within the data. The primary goal is to improve the accuracy of classification tasks involving multiple classes and labels. MMLMT integrates the logistic regression (LR) and decision tree (DT) algorithms, yielding interpretable models with high predictive performance. By combining the strengths of LR and DT, our method offers a flexible and powerful framework for handling multi-class multi-label data. Extensive experiments demonstrated the effectiveness of MMLMT across a range of well-known datasets with an average accuracy of 85.90%. Furthermore, our method achieved an average of 9.87% improvement compared to the results of state-of-the-art studies in the literature. These results highlight MMLMT’s potential as a valuable approach to multi-label learning. Full article
(This article belongs to the Special Issue Advances in Machine Learning and Applications)
Show Figures

Figure 1

Figure 1
<p>Timeline of related works for the MMLMT method across various domains, including (a) disease prediction [<a href="#B27-mathematics-12-02825" class="html-bibr">27</a>,<a href="#B28-mathematics-12-02825" class="html-bibr">28</a>], (b) drug repurposing [<a href="#B29-mathematics-12-02825" class="html-bibr">29</a>], (c) biomedical applications [<a href="#B30-mathematics-12-02825" class="html-bibr">30</a>], (d) image classification [<a href="#B31-mathematics-12-02825" class="html-bibr">31</a>,<a href="#B32-mathematics-12-02825" class="html-bibr">32</a>,<a href="#B33-mathematics-12-02825" class="html-bibr">33</a>,<a href="#B34-mathematics-12-02825" class="html-bibr">34</a>], (e) natural language processing [<a href="#B35-mathematics-12-02825" class="html-bibr">35</a>,<a href="#B36-mathematics-12-02825" class="html-bibr">36</a>,<a href="#B37-mathematics-12-02825" class="html-bibr">37</a>,<a href="#B38-mathematics-12-02825" class="html-bibr">38</a>], (f) education [<a href="#B39-mathematics-12-02825" class="html-bibr">39</a>], (g) industry [<a href="#B40-mathematics-12-02825" class="html-bibr">40</a>], (h) transportation [<a href="#B41-mathematics-12-02825" class="html-bibr">41</a>], (i) natural disaster management [<a href="#B42-mathematics-12-02825" class="html-bibr">42</a>], (j) ophthalmology [<a href="#B43-mathematics-12-02825" class="html-bibr">43</a>], (k) medical diagnosis [<a href="#B44-mathematics-12-02825" class="html-bibr">44</a>], (l) pedestrian attribute recognition [<a href="#B45-mathematics-12-02825" class="html-bibr">45</a>], (m) landslide susceptibility mapping [<a href="#B46-mathematics-12-02825" class="html-bibr">46</a>,<a href="#B47-mathematics-12-02825" class="html-bibr">47</a>], (n) flash flood susceptibility map-ping [<a href="#B48-mathematics-12-02825" class="html-bibr">48</a>], (o) underground column stability assessment [<a href="#B49-mathematics-12-02825" class="html-bibr">49</a>].</p>
Full article ">Figure 2
<p>An overview of the MMLMT method.</p>
Full article ">Figure 3
<p>The MMLMT performance over various datasets in the sensitivity metric.</p>
Full article ">Figure 4
<p>The MMLMT performance over various datasets in the PRC Area metric.</p>
Full article ">Figure 5
<p>An example LMT tree structure from the Thyroid-L7 dataset.</p>
Full article ">
18 pages, 9000 KiB  
Article
Multilevel Geometric Feature Embedding in Transformer Network for ALS Point Cloud Semantic Segmentation
by Zhuanxin Liang and Xudong Lai
Remote Sens. 2024, 16(18), 3386; https://doi.org/10.3390/rs16183386 - 12 Sep 2024
Viewed by 108
Abstract
Effective semantic segmentation of Airborne Laser Scanning (ALS) point clouds is a crucial field of study and influences subsequent point cloud application tasks. Transformer networks have made significant progress in 2D/3D computer vision tasks, exhibiting superior performance. We propose a multilevel geometric feature [...] Read more.
Effective semantic segmentation of Airborne Laser Scanning (ALS) point clouds is a crucial field of study and influences subsequent point cloud application tasks. Transformer networks have made significant progress in 2D/3D computer vision tasks, exhibiting superior performance. We propose a multilevel geometric feature embedding transformer network (MGFE-T), which aims to fully utilize the three-dimensional structural information carried by point clouds and enhance transformer performance in ALS point cloud semantic segmentation. In the encoding stage, compute the geometric features surrounding tee sampling points at each layer and embed them into the transformer workflow. To ensure that the receptive field of the self-attention mechanism and the geometric computation domain can maintain a consistent scale at each layer, we propose a fixed-radius dilated KNN (FR-DKNN) search method to address the limitation of traditional KNN search methods in considering domain radius. In the decoding stage, we aggregate prediction deviations at each level into a unified loss value, enabling multilevel supervision to improve the network’s feature learning ability at different levels. The MGFE-T network can predict the class label of each point in an end-to-end manner. Experiments were conducted on three widely used benchmark datasets. The results indicate that the MGFE-T network achieves superior OA and mF1 scores on the LASDU and DFC2019 datasets and performs well on the ISPRS dataset with imbalanced classes. Full article
Show Figures

Figure 1

Figure 1
<p>MGFE-T Semantic Segmentation Network Architecture.</p>
Full article ">Figure 2
<p>GFE-T/Transformer Block with Residual Connection.</p>
Full article ">Figure 3
<p>GFE-T Module Architecture.</p>
Full article ">Figure 4
<p>Comparison of FR-DKNN with other methods (<span class="html-italic">k</span> = 4, <span class="html-italic">d</span> = 2).</p>
Full article ">Figure 5
<p>Preview of the LASDU dataset.</p>
Full article ">Figure 6
<p>Preview of the DFC2019 dataset (3 of 110 files).</p>
Full article ">Figure 7
<p>Preview of the ISPRS dataset.</p>
Full article ">Figure 8
<p>Visualization of semantic segmentation results for some regions of the LASDU dataset (the first, second, and third columns are the ground truth, the results of the baseline, and the results of MGFE-T, respectively).</p>
Full article ">Figure 9
<p>Visualization of semantic segmentation results for some regions of the DFC2019 dataset (the first, second, and third columns are the ground truth, the results of the baseline, and the results of MGFE-T, respectively).</p>
Full article ">Figure 10
<p>Visualization of semantic segmentation results for some regions of the ISPRS dataset (the first, second, and third columns are the ground truth, the results of the baseline, and the results of MGFE-T, respectively).</p>
Full article ">Figure 11
<p>Comparison of experimental results for different radius percentiles.</p>
Full article ">
20 pages, 6757 KiB  
Article
A Task Offloading and Resource Allocation Strategy Based on Multi-Agent Reinforcement Learning in Mobile Edge Computing
by Guiwen Jiang, Rongxi Huang, Zhiming Bao and Gaocai Wang
Future Internet 2024, 16(9), 333; https://doi.org/10.3390/fi16090333 - 11 Sep 2024
Viewed by 271
Abstract
Task offloading and resource allocation is a research hotspot in cloud-edge collaborative computing. Many existing pieces of research adopted single-agent reinforcement learning to solve this problem, which has some defects such as low robustness, large decision space, and ignoring delayed rewards. In view [...] Read more.
Task offloading and resource allocation is a research hotspot in cloud-edge collaborative computing. Many existing pieces of research adopted single-agent reinforcement learning to solve this problem, which has some defects such as low robustness, large decision space, and ignoring delayed rewards. In view of the above deficiencies, this paper constructs a cloud-edge collaborative computing model, and related task queue, delay, and energy consumption model, and gives joint optimization problem modeling for task offloading and resource allocation with multiple constraints. Then, in order to solve the joint optimization problem, this paper designs a decentralized offloading and scheduling scheme based on “task-oriented” multi-agent reinforcement learning. In this scheme, we present information synchronization protocols and offloading scheduling rules and use edge servers as agents to construct a multi-agent system based on the Actor–Critic framework. In order to solve delayed rewards, this paper models the offloading and scheduling problem as a “task-oriented” Markov decision process. This process abandons the commonly used equidistant time slot model but uses dynamic and parallel slots in the step of task processing time. Finally, an offloading decision algorithm TOMAC-PPO is proposed. The algorithm applies the proximal policy optimization to the multi-agent system and combines the Transformer neural network model to realize the memory and prediction of network state information. Experimental results show that this algorithm has better convergence speed and can effectively reduce the service cost, energy consumption, and task drop rate under high load and high failure rates. For example, the proposed TOMAC-PPO can reduce the average cost by from 19.4% to 66.6% compared to other offloading schemes under the same network load. In addition, the drop rate of some baseline algorithms with 50 users can achieve 62.5% for critical tasks, while the proposed TOMAC-PPO only has 5.5%. Full article
(This article belongs to the Special Issue Convergence of Edge Computing and Next Generation Networking)
Show Figures

Figure 1

Figure 1
<p>Cloud-edge collaboration model for MEC network.</p>
Full article ">Figure 2
<p>Task computing and transmission queues.</p>
Full article ">Figure 3
<p>Synchronization timeout due to disconnection (The blue arrow is the direction of information transmission, and the red cross indicates that the sender and receiver are not connected).</p>
Full article ">Figure 4
<p>Offloading and scheduling when user moves across cell (The numbers ①–⑥ represent the order in which the scheduling rules are executed).</p>
Full article ">Figure 5
<p>Multi-agent system based on the Actor–Critic framework.</p>
Full article ">Figure 6
<p>Task-oriented Markov decision process.</p>
Full article ">Figure 7
<p>Policy network structure.</p>
Full article ">Figure 8
<p>Value network structure.</p>
Full article ">Figure 9
<p>Clipped summarize objective schematic diagram.</p>
Full article ">Figure 10
<p>Offloading and allocation algorithms conversion and cumulative rewards versus episodes.</p>
Full article ">Figure 11
<p>Average task cost versus user number.</p>
Full article ">Figure 12
<p>Average task cost versus failure probability.</p>
Full article ">Figure 13
<p>Optimization effect of schemes on the key performance metrics of different types of tasks.</p>
Full article ">
16 pages, 3585 KiB  
Article
Upper-Limb and Low-Back Load Analysis in Workers Performing an Actual Industrial Use-Case with and without a Dual-Arm Collaborative Robot
by Alessio Silvetti, Tiwana Varrecchia, Giorgia Chini, Sonny Tarbouriech, Benjamin Navarro, Andrea Cherubini, Francesco Draicchio and Alberto Ranavolo
Safety 2024, 10(3), 78; https://doi.org/10.3390/safety10030078 - 11 Sep 2024
Viewed by 126
Abstract
In the Industry 4.0 scenario, human–robot collaboration (HRC) plays a key role in factories to reduce costs, increase production, and help aged and/or sick workers maintain their job. The approaches of the ISO 11228 series commonly used for biomechanical risk assessments cannot be [...] Read more.
In the Industry 4.0 scenario, human–robot collaboration (HRC) plays a key role in factories to reduce costs, increase production, and help aged and/or sick workers maintain their job. The approaches of the ISO 11228 series commonly used for biomechanical risk assessments cannot be applied in Industry 4.0, as they do not involve interactions between workers and HRC technologies. The use of wearable sensor networks and software for biomechanical risk assessments could help us develop a more reliable idea about the effectiveness of collaborative robots (coBots) in reducing the biomechanical load for workers. The aim of the present study was to investigate some biomechanical parameters with the 3D Static Strength Prediction Program (3DSSPP) software v.7.1.3, on workers executing a practical manual material-handling task, by comparing a dual-arm coBot-assisted scenario with a no-coBot scenario. In this study, we calculated the mean and the standard deviation (SD) values from eleven participants for some 3DSSPP parameters. We considered the following parameters: the percentage of maximum voluntary contraction (%MVC), the maximum allowed static exertion time (MaxST), the low-back spine compression forces at the L4/L5 level (L4Ort), and the strength percent capable value (SPC). The advantages of introducing the coBot, according to our statistics, concerned trunk flexion (SPC from 85.8% without coBot to 95.2%; %MVC from 63.5% without coBot to 43.4%; MaxST from 33.9 s without coBot to 86.2 s), left shoulder abdo-adduction (%MVC from 46.1% without coBot to 32.6%; MaxST from 32.7 s without coBot to 65 s), and right shoulder abdo-adduction (%MVC from 43.9% without coBot to 30.0%; MaxST from 37.2 s without coBot to 70.7 s) in Phase 1, and right shoulder humeral rotation (%MVC from 68.4% without coBot to 7.4%; MaxST from 873.0 s without coBot to 125.2 s), right shoulder abdo-adduction (%MVC from 31.0% without coBot to 18.3%; MaxST from 60.3 s without coBot to 183.6 s), and right wrist flexion/extension rotation (%MVC from 50.2% without coBot to 3.0%; MaxST from 58.8 s without coBot to 1200.0 s) in Phase 2. Moreover, Phase 3, which consisted of another manual handling task, would be removed by using a coBot. In summary, using a coBot in this industrial scenario would reduce the biomechanical risk for workers, particularly for the trunk, both shoulders, and the right wrist. Finally, the 3DSSPP software could be an easy, fast, and costless tool for biomechanical risk assessments in an Industry 4.0 scenario where ISO 11228 series cannot be applied; it could be used by occupational medicine physicians and health and safety technicians, and could also help employers to justify a long-term investment. Full article
Show Figures

Figure 1

Figure 1
<p>Some 3DSSPP reconstructions of the three subtasks analyzed: Phase 1 with (<b>a1</b>) and without (<b>b1</b>) the coBot; Phase 2 with (<b>a2</b>) and without (<b>b2</b>) the coBot; and Phase 3 with (<b>a3</b>) and without (<b>b3</b>) the coBot.</p>
Full article ">Figure 2
<p>Mean and SD values for Phase 1, with Bazar (wB) in blue and without Bazar (woB) in red, for the investigated parameters (L4–L5 orthogonal forces, strength percent capable value, %MVC, and maximum holding time). An asterisk (*) over the bars shows statistical significance.</p>
Full article ">Figure 3
<p>Mean and SD values for Phase 2, with Bazar (wB) in blue and without Bazar (woB) in red, for the investigated parameters (L4–L5 orthogonal forces, strength percent capable value, %MVC, and maximum holding time). An asterisk (*) over the bars shows statistical significance.</p>
Full article ">Figure 4
<p>Mean and SD values for Phase 3 without Bazar (woB) in red, for the investigated parameters (L4–L5 orthogonal forces, strength percent capable value, %MVC, and maximum holding time). When using the Bazar coBot, this phase would be totally automatized, so we do not have values with the Bazar (wB).</p>
Full article ">
20 pages, 1425 KiB  
Article
Knowledge Graph Embedding Using a Multi-Channel Interactive Convolutional Neural Network with Triple Attention
by Lin Shi, Weitao Liu, Yafeng Wu, Chenxu Dai, Zhanlin Ji and Ivan Ganchev
Mathematics 2024, 12(18), 2821; https://doi.org/10.3390/math12182821 - 11 Sep 2024
Viewed by 136
Abstract
Knowledge graph embedding (KGE) has been identified as an effective method for link prediction, which involves predicting missing relations or entities based on existing entities or relations. KGE is an important method for implementing knowledge representation and, as such, has been widely used [...] Read more.
Knowledge graph embedding (KGE) has been identified as an effective method for link prediction, which involves predicting missing relations or entities based on existing entities or relations. KGE is an important method for implementing knowledge representation and, as such, has been widely used in driving intelligent applications w.r.t. question-answering systems, recommendation systems, and relationship extraction. Models based on convolutional neural networks (CNNs) have achieved good results in link prediction. However, as the coverage areas of knowledge graphs expand, the increasing volume of information significantly limits the performance of these models. This article introduces a triple-attention-based multi-channel CNN model, named ConvAMC, for the KGE task. In the embedding representation module, entities and relations are embedded into a complex space and the embeddings are performed in an alternating pattern. This approach helps in capturing richer semantic information and enhances the expressive power of the model. In the encoding module, a multi-channel approach is employed to extract more comprehensive interaction features. A triple attention mechanism and max pooling layers are used to ensure that interactions between spatial dimensions and output tensors are captured during the subsequent tensor concatenation and reshaping process, which allows preserving local and detailed information. Finally, feature vectors are transformed into prediction targets for embedding through the Hadamard product of feature mapping and reshaping matrices. Extensive experiments were conducted to evaluate the performance of ConvAMC on three benchmark datasets compared with state-of-the-art (SOTA) models, demonstrating that the proposed model outperforms all compared models across all evaluation metrics on two of the datasets, and achieves advanced link prediction results on most evaluation metrics on the third dataset. Full article
Show Figures

Figure 1

Figure 1
<p>The end-to-end structure of the proposed ConvAMC model (<b>e</b> denotes entity embedding; <b>r</b> denotes relation embedding; Reshape and Concat denote reshaping and splicing operations, respectively; conv denotes convolution operation; Chuck denotes segmentation operation; <span class="html-fig-inline" id="mathematics-12-02821-i001"><img alt="Mathematics 12 02821 i001" src="/mathematics/mathematics-12-02821/article_deploy/html/images/mathematics-12-02821-i001.png"/></span> denotes the Hadamard product operation).</p>
Full article ">Figure 2
<p>Two reshaping methods: (<b>a</b>) stacking; (<b>b</b>) alternating.</p>
Full article ">Figure 3
<p>The influence of the convolutional kernel size on ConvAMC’s link prediction performance. (<b>a</b>) WN18RR results; (<b>b</b>) FB15k-237 results.</p>
Full article ">
30 pages, 2615 KiB  
Article
Evaluation of the Monitoring Capabilities of Remote Sensing Satellites for Maritime Moving Targets
by Weiming Li, Zhiqiang Du, Li Wang and Tiancheng Zhou
ISPRS Int. J. Geo-Inf. 2024, 13(9), 325; https://doi.org/10.3390/ijgi13090325 - 11 Sep 2024
Viewed by 196
Abstract
Although an Automatic Identification System (AIS) can be used to monitor trajectories, it has become a reality for remote sensing satellite clusters to monitor maritime moving targets. The increasing demand for monitoring poses challenges for the construction of satellites, the monitoring capabilities of [...] Read more.
Although an Automatic Identification System (AIS) can be used to monitor trajectories, it has become a reality for remote sensing satellite clusters to monitor maritime moving targets. The increasing demand for monitoring poses challenges for the construction of satellites, the monitoring capabilities of which urgently need to be evaluated. Conventional evaluation methods focus on the spatial characteristics of monitoring; however, the temporal characteristics and the target’s kinematic characteristics are neglected. In this study, an evaluation method that integrates the spatial and temporal characteristics of monitoring along with the target’s kinematic characteristics is proposed. Firstly, a target motion prediction model for calculating the transfer probability and a satellite observation information calculation model for obtaining observation strips and time windows are established. Secondly, an index system is established, including the target detection capability, observation coverage capability, proportion of empty window, dispersion of observation window, and deviation of observation window. Thirdly, a comprehensive evaluation is completed through combining the analytic hierarchy process and entropy weight method to obtain the monitoring capability score. Finally, simulation experiments are conducted to evaluate the monitoring capabilities of satellites for ship trajectories. The results show that the method is effective when the grid size is between 1.6 and 1.8 times the target size and the task duration is approximately twice the time interval between trajectory points. Furthermore, the method is proven to be usable in various environments. Full article
Show Figures

Figure 1

Figure 1
<p>The relationship between the models and the index system.</p>
Full article ">Figure 2
<p>Diagram of the task area.</p>
Full article ">Figure 3
<p>Calculation of the coordinates in a three-dimensional spherical coordinate system.</p>
Full article ">Figure 4
<p>Distribution diagram of a target’s transfer probability.</p>
Full article ">Figure 5
<p>Scene of satellite side-looking imaging: (<b>a</b>) the process of satellite side-looking imaging; and (<b>b</b>) a cross-sectional view of the scene.</p>
Full article ">Figure 6
<p>Overlay diagram of the satellite observation strip and the covered task area, yellow represents the area that can be observed by satellites, and red circles represent the task area.</p>
Full article ">Figure 7
<p>The calculation process for the TDC.</p>
Full article ">Figure 8
<p>The process of merging time windows.</p>
Full article ">Figure 9
<p>The calculation process for the DOW: (<b>a</b>) the conversion for measuring the dispersion between periods; and (<b>b</b>) the calculation process for the DOW when there are only two observation windows.</p>
Full article ">Figure 10
<p>The model for evaluation with the analytic hierarchy process.</p>
Full article ">Figure 11
<p>Simulated Trajectory 1.</p>
Full article ">Figure 12
<p>The comparison results of the comprehensive evaluation scores and time consumption for different gird sizes: (<b>a</b>) the comprehensive evaluation score, blue line represents score trend, red dashed line represents trend boundary line; (<b>b</b>) time consumption comparison of the seven different gird sizes.</p>
Full article ">Figure 13
<p>The values of the TDC at different grid sizes.</p>
Full article ">Figure 14
<p>The comprehensive evaluation scores.</p>
Full article ">Figure 15
<p>The values of each index: (<b>a</b>) Index values for <math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>s</mi> <mi>k</mi> </mrow> </msub> </semantics></math> of <math display="inline"><semantics> <mrow> <mo>Δ</mo> <mi>T</mi> </mrow> </semantics></math>; (<b>b</b>) index values for <math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>s</mi> <mi>k</mi> </mrow> </msub> </semantics></math> of <math display="inline"><semantics> <mrow> <mn>1.5</mn> <mo>Δ</mo> <mi>T</mi> </mrow> </semantics></math>; (<b>c</b>) index values for <math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>s</mi> <mi>k</mi> </mrow> </msub> </semantics></math> of <math display="inline"><semantics> <mrow> <mn>2</mn> <mo>Δ</mo> <mi>T</mi> </mrow> </semantics></math>; (<b>d</b>) index values for <math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>s</mi> <mi>k</mi> </mrow> </msub> </semantics></math> of <math display="inline"><semantics> <mrow> <mn>2.5</mn> <mo>Δ</mo> <mi>T</mi> </mrow> </semantics></math>; and (<b>e</b>) index values for <math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>s</mi> <mi>k</mi> </mrow> </msub> </semantics></math> of <math display="inline"><semantics> <mrow> <mn>3</mn> <mo>Δ</mo> <mi>T</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 16
<p>Simulated trajectories: (<b>a</b>) Trajectory 2; (<b>b</b>) Trajectory 3; and (<b>c</b>) Trajectory 4.</p>
Full article ">Figure 17
<p>The comprehensive evaluation scores.</p>
Full article ">Figure 18
<p>The values of each index: (<b>a</b>) index values for Trajectory 1; (<b>b</b>) index values for Trajectory 2; and (<b>c</b>) index values for Trajectory 3.</p>
Full article ">Figure 19
<p>The comprehensive evaluation scores.</p>
Full article ">Figure 20
<p>The values of each index: (<b>a</b>) Index values for Night Trajectory 2; (<b>b</b>) Index values for Night Trajectory 3; and (<b>c</b>) Index values for Night Trajectory 4.</p>
Full article ">Figure 21
<p>The trajectory when covered by clouds; for example, <math display="inline"><semantics> <mrow> <mn>25</mn> <mo>%</mo> </mrow> </semantics></math>, the number represents the ID of the trajectory point.</p>
Full article ">Figure 22
<p>The comprehensive evaluation scores.</p>
Full article ">Figure 23
<p>The values of each index: (<b>a</b>) Index values with <math display="inline"><semantics> <mrow> <mn>25</mn> <mo>%</mo> </mrow> </semantics></math> cloud coverage; (<b>b</b>) Index values with <math display="inline"><semantics> <mrow> <mn>50</mn> <mo>%</mo> </mrow> </semantics></math> cloud coverage; and (<b>c</b>) Index values with <math display="inline"><semantics> <mrow> <mn>75</mn> <mo>%</mo> </mrow> </semantics></math> cloud coverage.</p>
Full article ">
20 pages, 3519 KiB  
Article
The Implementation of Multimodal Large Language Models for Hydrological Applications: A Comparative Study of GPT-4 Vision, Gemini, LLaVa, and Multimodal-GPT
by Likith Anoop Kadiyala, Omer Mermer, Dinesh Jackson Samuel, Yusuf Sermet and Ibrahim Demir
Hydrology 2024, 11(9), 148; https://doi.org/10.3390/hydrology11090148 - 11 Sep 2024
Viewed by 343
Abstract
Large Language Models (LLMs) combined with visual foundation models have demonstrated significant advancements, achieving intelligence levels comparable to human capabilities. This study analyzes the latest Multimodal LLMs (MLLMs), including Multimodal-GPT, GPT-4 Vision, Gemini, and LLaVa, with a focus on hydrological applications such as [...] Read more.
Large Language Models (LLMs) combined with visual foundation models have demonstrated significant advancements, achieving intelligence levels comparable to human capabilities. This study analyzes the latest Multimodal LLMs (MLLMs), including Multimodal-GPT, GPT-4 Vision, Gemini, and LLaVa, with a focus on hydrological applications such as flood management, water level monitoring, agricultural water discharge, and water pollution management. We evaluated these MLLMs on hydrology-specific tasks, testing their response generation and real-time suitability in complex real-world scenarios. Prompts were designed to enhance the models’ visual inference capabilities and contextual comprehension from images. Our findings reveal that GPT-4 Vision demonstrated exceptional proficiency in interpreting visual data, providing accurate assessments of flood severity and water quality. Additionally, MLLMs showed potential in various hydrological applications, including drought prediction, streamflow forecasting, groundwater management, and wetland conservation. These models can optimize water resource management by predicting rainfall, evaporation rates, and soil moisture levels, thereby promoting sustainable agricultural practices. This research provides valuable insights into the potential applications of advanced AI models in addressing complex hydrological challenges and improving real-time decision-making in water resource management Full article
Show Figures

Figure 1

Figure 1
<p>The workflow for MLLM benchmarking in hydrological tasks.</p>
Full article ">
4 pages, 720 KiB  
Proceeding Paper
Attributing Minimum Night Flow to Individual Pipes in Real-World Water Distribution Networks Using Machine Learning
by Matthew Hayslep, Edward Keedwell, Raziyeh Farmani and Joshua Pocock
Eng. Proc. 2024, 69(1), 112; https://doi.org/10.3390/engproc2024069112 - 10 Sep 2024
Viewed by 39
Abstract
This article introduces an explainable machine learning model for estimating the amount of flow that each pipe in a district metered area (DMA) contributes to the minimum night flow (MNF). This approach is validated using the MNF of DMAs and pipe failures, showing [...] Read more.
This article introduces an explainable machine learning model for estimating the amount of flow that each pipe in a district metered area (DMA) contributes to the minimum night flow (MNF). This approach is validated using the MNF of DMAs and pipe failures, showing good results for both tasks. The predictions from this model could be used to guide leak management or intervention strategies. In total, 800 DMAs ranging from rural to urban networks and representing nearly 12 million meters of pipe from a UK water company are used to train, validate, test, and evaluate the methodology. Full article
Show Figures

Figure 1

Figure 1
<p>Predictions for an example DMA. The sum of pipe MNF predictions (i.e., DMA-MNF) is at the top. A histogram of prediction values is shown at the bottom.</p>
Full article ">Figure 2
<p>Performance plots for the pipe MNF prediction model: (<b>a</b>) observed vs. predicted DMA–MNF values using validation method (1); (<b>b</b>) ROC curve for pipe failure classification, i.e., validation method (2).</p>
Full article ">
25 pages, 3537 KiB  
Article
A Complete EDA and DL Pipeline for Softwarized 5G Network Intrusion Detection
by Abdallah Moubayed
Future Internet 2024, 16(9), 331; https://doi.org/10.3390/fi16090331 - 10 Sep 2024
Viewed by 130
Abstract
The rise of 5G networks is driven by increasing deployments of IoT devices and expanding mobile and fixed broadband subscriptions. Concurrently, the deployment of 5G networks has led to a surge in network-related attacks, due to expanded attack surfaces. Machine learning (ML), particularly [...] Read more.
The rise of 5G networks is driven by increasing deployments of IoT devices and expanding mobile and fixed broadband subscriptions. Concurrently, the deployment of 5G networks has led to a surge in network-related attacks, due to expanded attack surfaces. Machine learning (ML), particularly deep learning (DL), has emerged as a promising tool for addressing these security challenges in 5G networks. To that end, this work proposed an exploratory data analysis (EDA) and DL-based framework designed for 5G network intrusion detection. The approach aimed to better understand dataset characteristics, implement a DL-based detection pipeline, and evaluate its performance against existing methodologies. Experimental results using the 5G-NIDD dataset showed that the proposed DL-based models had extremely high intrusion detection and attack identification capabilities (above 99.5% and outperforming other models from the literature), while having a reasonable prediction time. This highlights their effectiveness and efficiency for such tasks in softwarized 5G environments. Full article
(This article belongs to the Special Issue Advanced 5G and beyond Networks)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Proposed EDA and DL framework for softwarized 5G network intrusion detection.</p>
Full article ">Figure 2
<p>Number of benign and malicious attack instances—overview.</p>
Full article ">Figure 3
<p>Missing value analysis—the shade of blue represents the degree of correlation of missing values between columns.</p>
Full article ">Figure 4
<p>Number of benign and malicious attack instances—categorized per attack type.</p>
Full article ">Figure 5
<p>Attack type distribution.</p>
Full article ">Figure 6
<p>Mutual information—binary case: (<b>a</b>) Top 25 Features, (<b>b</b>) Remaining Features.</p>
Full article ">Figure 7
<p>Mutual information—multi-class case: (<b>a</b>) Top 25 Features, (<b>b</b>) Remaining Features.</p>
Full article ">Figure 8
<p>Principal component analysis—binary case.</p>
Full article ">Figure 9
<p>Principal component analysis—multi-class case.</p>
Full article ">Figure 10
<p>Principal component analysis—importance.</p>
Full article ">Figure 11
<p>Dense autoencoder neural network architecture.</p>
Full article ">Figure 12
<p>Convolutional neural network architecture.</p>
Full article ">Figure 13
<p>Recurrent neural network architecture.</p>
Full article ">Figure 14
<p>Dense autoencoder neural network confusion matrix—multi-class classification scenario.</p>
Full article ">Figure 15
<p>Convolutional neural network confusion matrix—multi-class classification scenario.</p>
Full article ">Figure 16
<p>Recurrent neural network confusion matrix—multi-class classification scenario.</p>
Full article ">
19 pages, 8192 KiB  
Article
Investigating the Relationship between Balanced Composition and Aesthetic Judgment through Computational Aesthetics and Neuroaesthetic Approaches
by Fangfu Lin, Wu Song, Yan Li and Wanni Xu
Symmetry 2024, 16(9), 1191; https://doi.org/10.3390/sym16091191 - 10 Sep 2024
Viewed by 203
Abstract
Background: Symmetry is a special kind of balance. This study aims to systematically explore and apply the role of balanced composition in aesthetic judgments by focusing on balanced composition features and employing research methods from computational aesthetics and neuroaesthetics. Methods: First, experimental materials [...] Read more.
Background: Symmetry is a special kind of balance. This study aims to systematically explore and apply the role of balanced composition in aesthetic judgments by focusing on balanced composition features and employing research methods from computational aesthetics and neuroaesthetics. Methods: First, experimental materials were classified by quantifying balanced composition using several indices, including symmetry, center of gravity, and negative space. An EEG experiment was conducted with 18 participants, who were asked to respond dichotomously to the same stimuli under different judgment tasks (balance and aesthetics), with both behavioral and EEG data being recorded and analyzed. Subsequently, participants’ data were combined with balanced composition indices to construct and analyze various SVM classification models. Results: Participants largely used balanced composition as a criterion for aesthetic evaluation. ERP data indicated that from 300–500 ms post-stimulus, brain activation was more significant in the aesthetic task, with unbeautiful and imbalanced stimuli eliciting larger frontal negative waves and occipital positive waves. From 600–1000 ms, beautiful stimuli caused smaller negative waves in the PZ channel. The results of the SVM models indicated that the model incorporating aesthetic subject data (ACC = 0.9989) outperforms the model using only balanced composition parameters of the aesthetic object (ACC = 0.7074). Conclusions: Balanced composition is a crucial indicator in aesthetics, with similar early processing stages in both balance and aesthetic judgments. Multi-modal data models validated the advantage of including human factors in aesthetic evaluation systems. This interdisciplinary approach not only enhances our understanding of the cognitive and emotional processes involved in aesthetic judgments but also enables the construction of more reasonable machine learning models to simulate and predict human aesthetic preferences. Full article
(This article belongs to the Section Life Sciences)
Show Figures

Figure 1

Figure 1
<p>Stimulus materials generated by Processing 5.0 software.</p>
Full article ">Figure 2
<p>An illustration of the experimental materials generated after adjustment.</p>
Full article ">Figure 3
<p>Decomposition of Negative Space using Quadtree.</p>
Full article ">Figure 4
<p>Pre-experiment: (<b>a</b>) Randomly Present the Printed Experimental Materials on the Desktop; (<b>b</b>) Participant Classifying Experimental Materials Based on Aesthetic Standards in the Pre-Experiment.</p>
Full article ">Figure 5
<p>Illustration of the stimulus paradigm applied.</p>
Full article ">Figure 6
<p>Main experiment: (<b>a</b>) Display of the experimental environment; (<b>b</b>) Participant Performing Binary Responses to Stimuli in the Main Experiment.</p>
Full article ">Figure 7
<p>EEG Analysis for task (aesthetics, balance) × answer (yes, no): (<b>a</b>) Grand–average event–related brain potentials; (<b>b</b>) Isopotential contour plot. (N = 16).</p>
Full article ">Figure 8
<p>Decision Boundary of Linear SVM for Multi–Class Data: (<b>a</b>) Scheme I; (<b>b</b>) Scheme II.</p>
Full article ">Figure 9
<p>Model results of the two schemes: (<b>a</b>) Scheme I; (<b>b</b>) Scheme II.</p>
Full article ">
20 pages, 14961 KiB  
Article
Data Augmentation in Histopathological Classification: An Analysis Exploring GANs with XAI and Vision Transformers
by Guilherme Botazzo Rozendo, Bianca Lançoni de Oliveira Garcia, Vinicius Augusto Toreli Borgue, Alessandra Lumini, Thaína Aparecida Azevedo Tosta, Marcelo Zanchetta do Nascimento and Leandro Alves Neves
Appl. Sci. 2024, 14(18), 8125; https://doi.org/10.3390/app14188125 - 10 Sep 2024
Viewed by 225
Abstract
Generative adversarial networks (GANs) create images by pitting a generator (G) against a discriminator (D) network, aiming to find a balance between the networks. However, achieving this balance is difficult because G is trained based on just one value [...] Read more.
Generative adversarial networks (GANs) create images by pitting a generator (G) against a discriminator (D) network, aiming to find a balance between the networks. However, achieving this balance is difficult because G is trained based on just one value representing D’s prediction, and only D can access image features. We introduce a novel approach for training GANs using explainable artificial intelligence (XAI) to enhance the quality and diversity of generated images in histopathological datasets. We leverage XAI to extract feature information from D and incorporate it into G via the loss function, a unique strategy not previously explored in this context. We demonstrate that this approach enriches the training with relevant information and promotes improved quality and more variability in the artificial images, decreasing the FID by up to 32.7% compared to traditional methods. In the data augmentation task, these images improve the classification accuracy of Transformer models by up to 3.81% compared to models without data augmentation and up to 3.01% compared to traditional GAN data augmentation. The Saliency method provides G with the most informative feature information. Overall, our work highlights the potential of XAI for enhancing GAN training and suggests avenues for further exploration in this field. Full article
Show Figures

Figure 1

Figure 1
<p>Schematic summary of the proposed model.</p>
Full article ">Figure 2
<p>Example images from the CR dataset: (<b>a</b>,<b>b</b>) benign tumors, (<b>c</b>,<b>d</b>) malignant tumors.</p>
Full article ">Figure 3
<p>Example images from the LA dataset: (<b>a</b>) 1 month, (<b>b</b>) 6 months, (<b>c</b>) 16 months, and (<b>d</b>) 24 months.</p>
Full article ">Figure 4
<p>Example images from the LG dataset: (<b>a</b>,<b>b</b>) male and (<b>c</b>,<b>d</b>) female.</p>
Full article ">Figure 5
<p>Example images from the UCSB dataset: (<b>a</b>,<b>b</b>) benign tumors, (<b>c</b>,<b>d</b>) malignant tumors.</p>
Full article ">Figure 6
<p>Schematic illustration of the classification evaluation process.</p>
Full article ">Figure 7
<p>Examples of generated images for the CR dataset.</p>
Full article ">Figure 8
<p>Examples of generated images for the LA dataset.</p>
Full article ">Figure 9
<p>Examples of generated images for the LG dataset.</p>
Full article ">Figure 10
<p>Examples of generated images for the UCSB dataset.</p>
Full article ">
21 pages, 3932 KiB  
Article
Multi-Step Passenger Flow Prediction for Urban Metro System Based on Spatial-Temporal Graph Neural Network
by Yuchen Chang, Mengya Zong, Yutian Dang and Kaiping Wang
Appl. Sci. 2024, 14(18), 8121; https://doi.org/10.3390/app14188121 - 10 Sep 2024
Viewed by 240
Abstract
Efficient operation of urban metro systems depends on accurate passenger flow predictions, a task complicated by intricate spatiotemporal correlations. This paper introduces a novel spatiotemporal graph neural network (STGNN) designed explicitly for predicting multistep passenger flow within metro stations. In the spatial dimension, [...] Read more.
Efficient operation of urban metro systems depends on accurate passenger flow predictions, a task complicated by intricate spatiotemporal correlations. This paper introduces a novel spatiotemporal graph neural network (STGNN) designed explicitly for predicting multistep passenger flow within metro stations. In the spatial dimension, previous research primarily focuses on local spatial dependencies, struggling to capture implicit global information. We propose a spatial modeling module that leverages a dynamic global attention network (DGAN) to capture dynamic global information from all-pair interactions, intricately fusing prior knowledge from the input graph with a graph convolutional network. In the temporal dimension, we design a temporal modeling module tailored to navigate the challenges of both long-term and recent-term temporal passenger flow patterns. This module consists of series decomposition blocks and locality-aware sparse attention (LSA) blocks to incorporate multiple local contexts and reduce computational complexities in long sequence modeling. Experiments conducted on both simulated and real-world datasets validate the exceptional predictive performance of our proposed model. Full article
Show Figures

Figure 1

Figure 1
<p>Input data of recent-term segment and long-term segment.</p>
Full article ">Figure 2
<p>The architecture of STGNN. The architecture adopts the end-to-end trainable network for passenger flow prediction. The pipeline of STGNN is mainly divided into two parts: spatial modeling and temporal modeling. Spatial modeling is primarily composed of the L-Spatial Encoder and the R-Spatial Encoder, which update the spatial features of nodes through their respective DAGN and GCN. Temporal modeling is primarily composed of series decomposition blocks, a Temporal Encoder, and a Temporal Decoder. The Temporal Encoder primarily updates the seasonal features of the passenger flow sequence, while the Temporal Decoder updates these seasonal features and uses the trend component of the passenger flow sequence to assist in prediction. Node embedding and temporal embedding are respectively achieved through MLP, implementing the embedding of node features and time point features.</p>
Full article ">Figure 3
<p>The architecture of locality-aware sparse attention. The input initially passes through a fully connected layer and Softmax to obtain values and a set of data-dependent weights. Subsequently, the input proceeds through two sets of convolutional layers with different kernel sizes (e.g., 1, 3). The results are then weighted by the obtained weights to generate queries and keys. Finally, the select function chooses the top u dominant queries to complete the attention operation with keys and values.</p>
Full article ">Figure 4
<p>Simulation layout in AnyLogic for metro station and Hangzhou Metro System. The red lines represent the edges connecting the nodes, while the blue dashed lines indicate an example of passenger flow direction. (<b>a</b>) Anylogic for Metro Station; (<b>b</b>) Hangzhou Metro System.</p>
Full article ">Figure 5
<p>Comparison with Ground Truth.</p>
Full article ">Figure 5 Cont.
<p>Comparison with Ground Truth.</p>
Full article ">Figure 6
<p>Network configuration analysis. These four figures demonstrate the impact of different model dimensions and attention head numbers on RMSE and MAE, respectively.</p>
Full article ">Figure 7
<p>Components analysis. These three figures demonstrate the impact of different model components on RMSE and MAE, respectively.</p>
Full article ">
Back to TopTop