[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (236)

Search Parameters:
Keywords = authentic task

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
30 pages, 6439 KiB  
Article
Adaptive Multi-Function Radar Temporal Behavior Analysis
by Zhenjia Xu, Qingsong Zhou, Zhihui Li, Jialong Qian, Yi Ding, Qinxian Chen and Qiyun Xu
Remote Sens. 2024, 16(22), 4131; https://doi.org/10.3390/rs16224131 - 6 Nov 2024
Viewed by 535
Abstract
The performance of radar mode recognition has been significantly enhanced by the various architectures of deep learning networks. However, these approaches often rely on supervised learning and are susceptible to overfitting on the same dataset. As a transitional phase towards Cognitive Multi-Functional Radar [...] Read more.
The performance of radar mode recognition has been significantly enhanced by the various architectures of deep learning networks. However, these approaches often rely on supervised learning and are susceptible to overfitting on the same dataset. As a transitional phase towards Cognitive Multi-Functional Radar (CMFR), Adaptive Multi-Function Radar (AMFR) possesses the capability to emit identical waveform signals across different working modes and states for task completion, with dynamically adjustable waveform parameters that adapt based on scene information. From a reconnaissance perspective, the valid signals received exhibit sparsity and localization in the time series. To address this challenge, we have redefined the reconnaissance-focused research priorities for radar systems to emphasize behavior analysis instead of pattern recognition. Based on our initial comprehensive digital system simulation model of a radar, we conducted reconnaissance and analysis from the perspective of the reconnaissance side, integrating both radar and reconnaissance aspects into environmental simulations to analyze radar behavior under realistic scenarios. Within the system, waveform parameters on the radar side vary according to unified rules, while resource management and task scheduling switch based on operational mechanisms. The target in the reconnaissance side maneuvers following authentic behavioral patterns while adjusting the electromagnetic space complexity in the environmental aspect as required. The simulation results indicate that temporal annotations in signal flow data play a crucial role in behavioral analysis from a reconnaissance perspective. This provides valuable insights for future radar behavior analysis incorporating temporal correlations and sequential dependencies. Full article
Show Figures

Figure 1

Figure 1
<p>Simulation system architecture.</p>
Full article ">Figure 2
<p>System-visualized operating interface.</p>
Full article ">Figure 3
<p>(<b>a</b>) The trend of <math display="inline"><semantics> <msub> <mfenced separators="" open="(" close=")"> <mrow> <mi>A</mi> <mo>·</mo> <mi>T</mi> </mrow> </mfenced> <mrow> <mi>r</mi> <mi>e</mi> <mi>q</mi> <mi>u</mi> <mi>i</mi> <mi>r</mi> <mi>e</mi> </mrow> </msub> </semantics></math> under a given <math display="inline"><semantics> <msub> <mi>P</mi> <mi>d</mi> </msub> </semantics></math> value. (<b>b</b>) The trend of <math display="inline"><semantics> <msub> <mfenced separators="" open="(" close=")"> <mrow> <mi>A</mi> <mo>·</mo> <mi>T</mi> </mrow> </mfenced> <mrow> <mi>r</mi> <mi>e</mi> <mi>q</mi> <mi>u</mi> <mi>i</mi> <mi>r</mi> <mi>e</mi> </mrow> </msub> </semantics></math> under a given <math display="inline"><semantics> <msub> <mi>P</mi> <mrow> <mi>f</mi> <mi>a</mi> </mrow> </msub> </semantics></math> value.</p>
Full article ">Figure 4
<p>Relationship curve between <math display="inline"><semantics> <mfenced separators="" open="(" close=")"> <mrow> <mi>A</mi> <mo>·</mo> <mi>T</mi> </mrow> </mfenced> </semantics></math> and <span class="html-italic">N</span>.</p>
Full article ">Figure 5
<p>Antenna array gain pattern.</p>
Full article ">Figure 6
<p>Spectrum diagram when MFR executes the S state.</p>
Full article ">Figure 7
<p>(<b>a</b>–<b>f</b>) Time-domain diagram of JAM signal interception.</p>
Full article ">Figure 8
<p>(<b>a</b>–<b>f</b>) Signal waveform parameter diagram.</p>
Full article ">Figure 9
<p>Relationship between <math display="inline"><semantics> <mrow> <mi>F</mi> <mi>a</mi> <mi>c</mi> <mi>t</mi> <mi>o</mi> <mi>r</mi> </mrow> </semantics></math> value and number of pulses <span class="html-italic">N</span> under given conditions.</p>
Full article ">Figure 10
<p>Trend of evaluation function value <math display="inline"><semantics> <mrow> <mi>E</mi> <mi>v</mi> <mi>a</mi> </mrow> </semantics></math> with maximum number of pulses <math display="inline"><semantics> <msub> <mi>N</mi> <mrow> <mi>c</mi> <mi>o</mi> <mi>n</mi> <mi>f</mi> <mi>i</mi> <mi>r</mi> <mi>m</mi> <mo>_</mo> <mi>m</mi> <mi>a</mi> <mi>x</mi> </mrow> </msub> </semantics></math>.</p>
Full article ">
19 pages, 1482 KiB  
Review
A Comprehensive Evaluation of Iris Segmentation on Benchmarking Datasets
by Mst Rumana Sumi, Priyanka Das, Afzal Hossain, Soumyabrata Dey and Stephanie Schuckers
Sensors 2024, 24(21), 7079; https://doi.org/10.3390/s24217079 - 3 Nov 2024
Viewed by 601
Abstract
Iris is one of the most widely used biometric modalities because of its uniqueness, high matching performance, and inherently secure nature. Iris segmentation is an essential preliminary step for iris-based biometric authentication. The authentication accuracy is directly connected with the iris segmentation accuracy. [...] Read more.
Iris is one of the most widely used biometric modalities because of its uniqueness, high matching performance, and inherently secure nature. Iris segmentation is an essential preliminary step for iris-based biometric authentication. The authentication accuracy is directly connected with the iris segmentation accuracy. In the last few years, deep-learning-based iris segmentation methodologies have increasingly been adopted because of their ability to handle challenging segmentation tasks and their advantages over traditional segmentation techniques. However, the biggest challenge to the biometric community is the scarcity of open-source resources for adoption for application and reproducibility. This review provides a comprehensive examination of available open-source iris segmentation resources, including datasets, algorithms, and tools. In the process, we designed three U-Net and U-Net++ architecture-influenced segmentation algorithms as standard benchmarks, trained them on a large composite dataset (>45K samples), and created 1K manually segmented ground truth masks. Overall, eleven state-of-the-art algorithms were benchmarked against five datasets encompassing multiple sensors, environmental conditions, demography, and illumination. This assessment highlights the strengths, limitations, and practical implications of each method and identifies gaps that future studies should address to improve segmentation accuracy and robustness. To foster future research, all resources developed during this work would be made publicly available. Full article
Show Figures

Figure 1

Figure 1
<p>User interface of the iris segmentation toolkit.</p>
Full article ">Figure 2
<p>Example of annotated ground truth using the toolkit.</p>
Full article ">Figure 3
<p>Architecture of the proposed U-Net benchmark model.</p>
Full article ">Figure 4
<p>Architecture of the proposed U-Net++ benchmark model.</p>
Full article ">Figure 5
<p>Examples of segmentation outputs from different models across various cases: (<b>A</b>) Iris without occlusions, (<b>B</b>) Iris with eyeglass occlusion, (<b>C</b>) Iris with eyelash occlusion, (<b>D</b>) Off-angle iris with eyeglass occlusion and specular reflection, (<b>E</b>) Eyelash occlusion with a dilated pupil, (<b>F</b>) Eyeglass occlusion and specular reflection, (<b>G</b>) Smaller iris area.</p>
Full article ">
19 pages, 510 KiB  
Article
Shaping Career Development Through College Readiness at the High School Level
by Wael Yousef
Educ. Sci. 2024, 14(11), 1190; https://doi.org/10.3390/educsci14111190 - 31 Oct 2024
Viewed by 703
Abstract
Career and College Readiness (CCR) generates higher levels of persistence, grit, motivation, and competencies in performing work- or school-related tasks among learners during post-secondary life. One primary limitation defining the present scholarship on CCR is the authentic analysis of the on the ground [...] Read more.
Career and College Readiness (CCR) generates higher levels of persistence, grit, motivation, and competencies in performing work- or school-related tasks among learners during post-secondary life. One primary limitation defining the present scholarship on CCR is the authentic analysis of the on the ground or field practices high schools perform to increase learners’ CCR competencies. The inadequate research on actual CCR practices in high schools motivated this study. To learn more about how high schools prepare students for post-secondary life, 16 principals from high schools provided detailed narratives on CCR practices in their institutions. Principals completed interviews of one to two hours using Microsoft Teams, supplying the researcher with specific information and examples of how their institutions equip students for future careers and higher education. Manual qualitative thematic analysis of the entire transcript guided the organization and interpretation of the findings, allowing the presentation of meaningful themes supported by a plethora of illustrations. Six themes representing 18 distinct CCR practices emerged, and the themes were rigorous curriculum, content knowledge, key academic behaviors, key cognitive strategies, multiculturalism, and citizenship development. Results partially supported the pervasive CCR model based on Conley’s readiness index. Increasingly, diversity and citizenship learning have defined high schools’ CCR work in preparing learners to become effective local and global citizens. The current investigation paves the way for future observational and field research uncovering whether schools truly prepare students or not, and such preparation varies across community, country, and institutional characteristics. Full article
Show Figures

Figure 1

Figure 1
<p>Illustration of Four Keys and Their Attributes in the Conley Readiness Index [<a href="#B1-education-14-01190" class="html-bibr">1</a>] (p. 3).</p>
Full article ">
20 pages, 5052 KiB  
Article
AIoT-Based Visual Anomaly Detection in Photovoltaic Sequence Data via Sequence Learning
by Qian Wei, Hongjun Sun, Jingjing Fan, Guojun Li and Zhiguang Zhou
Energies 2024, 17(21), 5369; https://doi.org/10.3390/en17215369 - 29 Oct 2024
Viewed by 636
Abstract
Anomaly detection is a common analytical task aimed at identifying rare cases that differ from the majority of typical cases in a dataset. In the management of photovoltaic (PV) power generation systems, it is essential for electric power companies to effectively detect anomalies [...] Read more.
Anomaly detection is a common analytical task aimed at identifying rare cases that differ from the majority of typical cases in a dataset. In the management of photovoltaic (PV) power generation systems, it is essential for electric power companies to effectively detect anomalies in PV sequence data, as this helps operators and experts understand and interpret anomalies within PV arrays when making response decisions. However, traditional methods that rely on manual labor and regular data collection are difficult to monitor in real time, resulting in delays in fault detection and localization. Traditional machine learning algorithms are slow and cumbersome in processing data, which affects the operational safety of PV plants. In this paper, we propose a visual analytic approach for detecting and exploring anomalous sequences in a PV sequence dataset via sequence learning. We first compare the sequences with their reconstructions through an unsupervised anomaly detection algorithm (Long Short-Term Memory) based on AutoEncoders to identify anomalies. To further enhance the accuracy of anomaly detection, we integrate the artificial intelligence of things (AIoT) technology with a strict time synchronization data collection and real-time processing algorithm. This integration ensures that data from multiple sensors are synchronized and processed in real time. Then, we analyze the characteristics of the anomalies based on the visual comparison of different PV sequences and explore the potential correlation factors to analyze the possible causes of the anomalies. Case studies based on authentic enterprise datasets demonstrate the effectiveness of our method in the anomaly detection and exploration of PV sequence data. Full article
(This article belongs to the Section A2: Solar Energy and Photovoltaic Systems)
Show Figures

Figure 1

Figure 1
<p>The Architecture of AIoT-Based PV Visual Anomaly Detection System.</p>
Full article ">Figure 2
<p>The results of the AIoT High Concurrency Performance Testing. This figure illustrates the success rate and response time analysis for each request, including the minimum, maximum, average response times, and standard deviation for the same request. When simulating 1.5 million concurrent real-time data requests, the success rate of each request exceeded 99.99%, and the average TPS reached 34,090.909 transactions per second. The minimum response time at this point is 1 ms, the maximum response time is 942 ms, and 99% of the response times are less than 638 ms, which means that 99% of the requests were successfully responded to within 638 ms. The request values counted here include successful requests and do not include failed requests.</p>
Full article ">Figure 3
<p>AIoT-Based Visual Anomaly Detection System in PV Sequence Data via Sequence Learning. <b>Section A</b> represents PV power generation equipment that includes the installation of smaller PV generation units at or near the point of consumption; <b>Section B</b> represents the Artificial Intelligence of Things (AIoT) system; and <b>Section C</b> represents the PV Visual Anomaly Detection system. In the <b>Section C</b>, (a) The control panel provides the interface for setting the parameters necessary for the identification of anomalies in sequence data. (b) The calendar view allows users to select the specific day for analysis. (c) The matrix view shows whether the power generated by each inverter is anomalous or not during the day. (d) The scatter view shows the distribution of the features of the sequence data. (e) The time analysis view shows specifically the differences in generation characteristics before and after the reconstruction of the sequence data. (f) The feature comparison view shows the differences in environmental factors between the selected data and its counterpart at the same time. (g) The feature matrix view compares the correlation between power generation and environmental variables for sequence data of the same category. (h) The information panel shows information about individual inverters.</p>
Full article ">Figure 4
<p>The workflow of our system for anomaly detection and analysis of PV sequence data.</p>
Full article ">Figure 5
<p>LSTM-AE with input sequence data of length = 5.</p>
Full article ">Figure 6
<p>The processing of anomaly detection of PV sequence data.</p>
Full article ">Figure 7
<p>Two different types of sequence data (sequence A and B) and their attribute positions in the scatter plot (scatter points A and B in the bottom right corner of the graph) were used to verify the accuracy of anomaly recognition.</p>
Full article ">Figure 8
<p>Three different types of sequence data (Sequence A, B and C) were used for comparative analysis of anomaly detection. Among them, Sequence A is an abnormal sequence marked in red, Sequence B is a normal sequence, and Sequence C is an abnormal sequence marked in blue.</p>
Full article ">Figure 9
<p>Illustration of the sequence segments comparison. Sequence segment to be framed for analysis (<b>a</b>) and those to be matched for comparison (<b>b</b>–<b>d</b>). Among them, (<b>a</b>,<b>b</b>,<b>d</b>) are both abnormal data marked in blue.</p>
Full article ">Figure 10
<p>Illustration of the causes analysis. (<b>a</b>) presents the anomalous sequence to be analyzed (Sequence C in case 2). (<b>b</b>,<b>c</b>) present the comparison of environmental factors before and after the anomalous sequence is framed respectively. (<b>d</b>) presents the category of the sequence analyzed and (<b>e</b>) is the corresponding feature matrix.</p>
Full article ">
27 pages, 920 KiB  
Article
AI-Generated Spam Review Detection Framework with Deep Learning Algorithms and Natural Language Processing
by Mudasir Ahmad Wani, Mohammed ElAffendi and Kashish Ara Shakil
Computers 2024, 13(10), 264; https://doi.org/10.3390/computers13100264 - 12 Oct 2024
Viewed by 708
Abstract
Spam reviews pose a significant challenge to the integrity of online platforms, misleading consumers and undermining the credibility of genuine feedback. This paper introduces an innovative AI-generated spam review detection framework that leverages Deep Learning algorithms and Natural Language Processing (NLP) techniques to [...] Read more.
Spam reviews pose a significant challenge to the integrity of online platforms, misleading consumers and undermining the credibility of genuine feedback. This paper introduces an innovative AI-generated spam review detection framework that leverages Deep Learning algorithms and Natural Language Processing (NLP) techniques to identify and mitigate spam reviews effectively. Our framework utilizes multiple Deep Learning models, including Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, Gated Recurrent Unit (GRU), and Bidirectional LSTM (BiLSTM), to capture intricate patterns in textual data. The system processes and analyzes large volumes of review content to detect deceptive patterns by utilizing advanced NLP and text embedding techniques such as One-Hot Encoding, Word2Vec, and Term Frequency-Inverse Document Frequency (TF-IDF). By combining three embedding techniques with four Deep Learning algorithms, a total of twelve exhaustive experiments were conducted to detect AI-generated spam reviews. The experimental results demonstrate that our approach outperforms the traditional machine learning models, offering a robust solution for ensuring the authenticity of online reviews. Among the models evaluated, those employing Word2Vec embeddings, particularly the BiLSTM_Word2Vec model, exhibited the strongest performance. The BiLSTM model with Word2Vec achieved the highest performance, with an exceptional accuracy of 98.46%, a precision of 0.98, a recall of 0.97, and an F1-score of 0.98, reflecting a near-perfect balance between precision and recall. Its high F2-score (0.9810) and F0.5-score (0.9857) further highlight its effectiveness in accurately detecting AI-generated spam while minimizing false positives, making it the most reliable option for this task. Similarly, the Word2Vec-based LSTM model also performed exceptionally well, with an accuracy of 97.58%, a precision of 0.97, a recall of 0.96, and an F1-score of 0.97. The CNN model with Word2Vec similarly delivered strong results, achieving an accuracy of 97.61%, a precision of 0.97, a recall of 0.96, and an F1-score of 0.97. This study is unique in its focus on detecting spam reviews specifically generated by AI-based tools rather than solely detecting spam reviews or AI-generated text. This research contributes to the field of spam detection by offering a scalable, efficient, and accurate framework that can be integrated into various online platforms, enhancing user trust and the decision-making processes. Full article
Show Figures

Figure 1

Figure 1
<p>Detailed data collection procedure.</p>
Full article ">Figure 2
<p>Generating AI-based spam/fake reviews based on human-authored samples.</p>
Full article ">Figure 3
<p>Check for the working of GPT Module.</p>
Full article ">Figure 4
<p>Data preparation and preprocessing with NLTK toolkit.</p>
Full article ">Figure 5
<p>Experimental setup and configuration.</p>
Full article ">Figure 6
<p>Performance of selected Deep Learning models on TF-IDF representation.</p>
Full article ">Figure 7
<p>Performance of selected Deep Learning models on Word2Vec feature representation.</p>
Full article ">Figure 8
<p>Performance of selected Deep Learning models on One-Hot Encoding.</p>
Full article ">Figure 9
<p>The radar plot showing proposed approaches. Particularly, Word2Vec-based BiLSTM outperformed the existing methods.</p>
Full article ">Figure 10
<p>Heptagon: seven ways to prevent abuse and ensure ethical use of AI-generated reviews.</p>
Full article ">
18 pages, 4451 KiB  
Article
A Biometric Identification for Multi-Modal Biomedical Signals in Geriatric Care
by Yue Che, Lingyan Du, Guozhi Tang and Shihai Ling
Sensors 2024, 24(20), 6558; https://doi.org/10.3390/s24206558 - 11 Oct 2024
Viewed by 600
Abstract
With the acceleration of global population aging, the elderly have an increasing demand for home care and nursing institutions, and the significance of health prevention and management of the elderly has become increasingly prominent. In this context, we propose a biometric recognition method [...] Read more.
With the acceleration of global population aging, the elderly have an increasing demand for home care and nursing institutions, and the significance of health prevention and management of the elderly has become increasingly prominent. In this context, we propose a biometric recognition method for multi-modal biomedical signals. This article focuses on three key signals that can be picked up by wearable devices: ECG, PPG, and breath (RESP). The RESP signal is introduced into the existing two-mode signal identification for multi-mode identification. Firstly, the features of the signal in the time–frequency domain are extracted. To represent deep features in a low-dimensional feature space and expedite authentication tasks, PCA and LDA are employed for dimensionality reduction. MCCA is used for feature fusion, and SVM is used for identification. The accuracy and performance of the system were evaluated using both public data sets and self-collected data sets, with an accuracy of more than 99.5%. The experimental data fully show that this method significantly improves the accuracy of identity recognition. In the future, combined with the signal monitoring function of wearable devices, it can quickly identify individual elderly people with abnormal conditions, provide safer and more efficient medical services for the elderly, and relieve the pressure on medical resources. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

Figure 1
<p>Block diagram of multi-modal identification.</p>
Full article ">Figure 2
<p>Biomedical signal acquisition experiment diagram.</p>
Full article ">Figure 3
<p>Block diagram of signal preprocessing.</p>
Full article ">Figure 4
<p>Comparison of biomedical signals before and after filtering.</p>
Full article ">Figure 5
<p>Energy spectrum diagram of (<b>a</b>) ECG, (<b>b</b>) PPG, and (<b>c</b>) RESP signals.</p>
Full article ">Figure 6
<p>Localized waves of (<b>a</b>) ECG, (<b>b</b>) PPG, and (<b>c</b>) RESP signals. The waveform in the red box in (<b>a</b>) is a complete single-period beat.</p>
Full article ">Figure 7
<p>Complex vectors of (<b>a</b>) ECG, (<b>b</b>) PPG, and (<b>c</b>) RESP signals.</p>
Full article ">Figure 8
<p>Block diagram of feature dimension reduction fusion.</p>
Full article ">Figure 9
<p>Principal component contribution rate and cumulative contribution rate of ECG. (<b>a</b>) shows the contribution rates of the first ten feature principal components for PCA reduction of ECG, and (<b>b</b>) shows the cumulative contribution rates of the first ten feature principal components.</p>
Full article ">Figure 10
<p>Principal component contribution rate and cumulative contribution rate of PPG. (<b>a</b>) shows the contribution rates of the first ten feature principal components for PCA reduction of PPG, and (<b>b</b>) shows the cumulative contribution rates of the first ten feature principal components.</p>
Full article ">Figure 11
<p>Principal component contribution rate and cumulative contribution rate of RESP. (<b>a</b>) shows the contribution rates of the first ten feature principal components for PCA reduction of RESP, and (<b>b</b>) shows the cumulative contribution rates of the first ten feature principal components.</p>
Full article ">Figure 12
<p>LDA performs secondary dimensionality reduction results.</p>
Full article ">Figure 13
<p>Individual identification results.</p>
Full article ">
18 pages, 4515 KiB  
Article
Historical Blurry Video-Based Face Recognition
by Lujun Zhai, Suxia Cui, Yonghui Wang, Song Wang, Jun Zhou and Greg Wilsbacher
J. Imaging 2024, 10(9), 236; https://doi.org/10.3390/jimaging10090236 - 20 Sep 2024
Viewed by 746
Abstract
Face recognition is a widely used computer vision, which plays an increasingly important role in user authentication systems, security systems, and consumer electronics. The models for most current applications are based on high-definition digital cameras. In this paper, we focus on digital images [...] Read more.
Face recognition is a widely used computer vision, which plays an increasingly important role in user authentication systems, security systems, and consumer electronics. The models for most current applications are based on high-definition digital cameras. In this paper, we focus on digital images derived from historical motion picture films. Historical motion picture films often have poorer resolution than modern digital imagery, making face detection a more challenging task. To approach this problem, we first propose a trunk–branch concatenated multi-task cascaded convolutional neural network (TB-MTCNN), which efficiently extracts facial features from blurry historical films by combining the trunk with branch networks and employing various sizes of kernels to enrich the multi-scale receptive field. Next, we build a deep neural network-integrated object-tracking algorithm to compensate for failed recognition over one or more video frames. The framework combines simple online and real-time tracking with deep data association (Deep SORT), and TB-MTCNN with the residual neural network (ResNet) model. Finally, a state-of-the-art image restoration method is employed to reduce the effect of noise and blurriness. The experimental results show that our proposed joint face recognition and tracking network can significantly reduce missed recognition in historical motion picture film frames. Full article
(This article belongs to the Section Computer Vision and Pattern Recognition)
Show Figures

Figure 1

Figure 1
<p>Face identification in historical motion picture video (President Johnson’s face can be identified, as indicated by the red boxes, with prediction probabilities in frames 000433 and 000435, while face identification failed in frame 000434).</p>
Full article ">Figure 2
<p>The architecture of the proposed model includes three components: a face detector (TB-MTCNN), a face tracker (Deep SORT), and a face classifier (ResNet18). First, TB-MTCNN processes the historical video frames to detect faces. Then, the Deep SORT algorithm utilizes face detection information from previous frames to track the face in the current frame if detection fails. Finally, the detection and tracking results are sent to the face classifier to determine whether the face belongs to President Johnson.</p>
Full article ">Figure 3
<p>TB-MTCNN structure. The blue arrow represents the MTCNN architecture, while the yellow arrow highlights the integrated branch into the TB-MTCNN.</p>
Full article ">Figure 4
<p>Performance comparison of face detection models on the Wider Face Medium dataset.</p>
Full article ">Figure 5
<p>Sample performance comparison of face detection models on historical videos.</p>
Full article ">Figure 6
<p>Effects of image restoration.</p>
Full article ">Figure 7
<p>The Deep SORT method recovered the missed detections (the red boxes are faces detected using the TB-MTCNN detector, and the yellow boxes are faces recovered using Deep SORT).</p>
Full article ">
18 pages, 59323 KiB  
Article
Method for Augmenting Side-Scan Sonar Seafloor Sediment Image Dataset Based on BCEL1-CBAM-INGAN
by Haixing Xia, Yang Cui, Shaohua Jin, Gang Bian, Wei Zhang and Chengyang Peng
J. Imaging 2024, 10(9), 233; https://doi.org/10.3390/jimaging10090233 - 20 Sep 2024
Viewed by 558
Abstract
In this paper, a method for augmenting samples of side-scan sonar seafloor sediment images based on CBAM-BCEL1-INGAN is proposed, aiming to address the difficulties in acquiring and labeling datasets, as well as the insufficient diversity and quantity of data samples. Firstly, a Convolutional [...] Read more.
In this paper, a method for augmenting samples of side-scan sonar seafloor sediment images based on CBAM-BCEL1-INGAN is proposed, aiming to address the difficulties in acquiring and labeling datasets, as well as the insufficient diversity and quantity of data samples. Firstly, a Convolutional Block Attention Module (CBAM) is integrated into the residual blocks of the INGAN generator to enhance the learning of specific attributes and improve the quality of the generated images. Secondly, a BCEL1 loss function (combining binary cross-entropy and L1 loss functions) is introduced into the discriminator, enabling it to focus on both global image consistency and finer distinctions for better generation results. Finally, augmented samples are input into an AlexNet classifier to verify their authenticity. Experimental results demonstrate the excellent performance of the method in generating images of coarse sand, gravel, and bedrock, as evidenced by significant improvements in the Frechet Inception Distance (FID) and Inception Score (IS). The introduction of the CBAM and BCEL1 loss function notably enhances the quality and details of the generated images. Moreover, classification experiments using the AlexNet classifier show an increase in the recognition rate from 90.5% using only INGAN-generated images of bedrock to 97.3% using images augmented using our method, marking a 6.8% improvement. Additionally, the classification accuracy of bedrock-type matrices is improved by 5.2% when images enhanced using the method presented in this paper are added to the training set, which is 2.7% higher than that of the simple method amplification. This validates the effectiveness of our method in the task of generating seafloor sediment images, partially alleviating the scarcity of side-scan sonar seafloor sediment image data. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

Figure 1
<p>CBAM-BCEL1-INGAN network flowchart.</p>
Full article ">Figure 2
<p>Generator structure diagram.</p>
Full article ">Figure 3
<p>Multiscale discriminator.</p>
Full article ">Figure 4
<p>Residual block structure based on CBAM.</p>
Full article ">Figure 5
<p>Parts of the samples in the dataset.</p>
Full article ">Figure 6
<p>Amplification of 4 groups of models, Figures (<b>a</b>–<b>d</b>) are partial amplifications generated by models 1–4 (The red circles are features of unnecessary detail).</p>
Full article ">Figure 7
<p>Structure of the AlexNet network model (The direction the arrow points in indicates the path of data from one layer to the next).</p>
Full article ">Figure 8
<p>Some images were generated using the sinGAN network during training.</p>
Full article ">
23 pages, 24285 KiB  
Article
Novel Hybrid Optimization Techniques for Enhanced Generalization and Faster Convergence in Deep Learning Models: The NestYogi Approach to Facial Biometrics
by Raoof Altaher and Hakan Koyuncu
Mathematics 2024, 12(18), 2919; https://doi.org/10.3390/math12182919 - 20 Sep 2024
Viewed by 943
Abstract
In the rapidly evolving field of biometric authentication, deep learning has become a cornerstone technology for face detection and recognition tasks. However, traditional optimizers often struggle with challenges such as overfitting, slow convergence, and limited generalization across diverse datasets. To address these issues, [...] Read more.
In the rapidly evolving field of biometric authentication, deep learning has become a cornerstone technology for face detection and recognition tasks. However, traditional optimizers often struggle with challenges such as overfitting, slow convergence, and limited generalization across diverse datasets. To address these issues, this paper introduces NestYogi, a novel hybrid optimization algorithm that integrates the adaptive learning capabilities of the Yogi optimizer, anticipatory updates of Nesterov momentum, and the generalization power of stochastic weight averaging (SWA). This combination significantly improves both the convergence rate and overall accuracy of deep neural networks, even when trained from scratch. Extensive data augmentation techniques, including noise and blur, were employed to ensure the robustness of the model across diverse conditions. NestYogi was rigorously evaluated on two benchmark datasets Labeled Faces in the Wild (LFW) and YouTube Faces (YTF), demonstrating superior performance with a detection accuracy reaching 98% and a recognition accuracy up to 98.6%, outperforming existing optimization strategies. These results emphasize NestYogi’s potential to overcome critical challenges in face detection and recognition, offering a robust solution for achieving state-of-the-art performance in real-world applications. Full article
Show Figures

Figure 1

Figure 1
<p>Position of optimization algorithms based on convergence rate and solution space exploration. The proposed NestYogi optimizer integrates Yogi, Nesterov momentum, and SWA, achieving superior performance in both dimensions.</p>
Full article ">Figure 2
<p>First-order optimization algorithms.</p>
Full article ">Figure 3
<p>Comparison of optimization algorithms based on convergence rate and solution space exploration on the MNIST dataset.</p>
Full article ">Figure 4
<p>SWA.</p>
Full article ">Figure 5
<p>Triplet loss architecture.</p>
Full article ">Figure 6
<p>Overview of the proposed model’s performance. (<b>a</b>) Proposed model face detection total loss. (<b>b</b>) Proposed model face detection precision. (<b>c</b>) Proposed model face detection recall. (<b>d</b>) Proposed model face detection mAP50.</p>
Full article ">Figure 7
<p>Overview of the proposed model’s face recognition loss and F1 score. (<b>a</b>) Proposed model’s face recognition loss. (<b>b</b>) Proposed model’s face recognition F1 score.</p>
Full article ">Figure 8
<p>Overview of the proposed model’s accuracy with various architectures. (<b>a</b>) Proposed face recognition model’s accuracy using VGG16. (<b>b</b>) Proposed face recognition model’s accuracy using InceptionV3. (<b>c</b>) Proposed face recognition model’s accuracy using ResNet50.</p>
Full article ">Figure 9
<p>Proposed face detection predictions.</p>
Full article ">Figure 10
<p>Model evaluation utilizing the final detection and recognition model.</p>
Full article ">
13 pages, 233 KiB  
Article
Bridging Teacher Knowledge and Practice: Exploring Authentic Assessment across Educational Levels
by Rachael Hains-Wesson and Sanri le Roux
Educ. Sci. 2024, 14(8), 894; https://doi.org/10.3390/educsci14080894 - 16 Aug 2024
Viewed by 1285
Abstract
As teachers, we are living and working in times of abundant challenge and change. These challenges transpire across different education levels and sectors, including K–12, vocational, tertiary, and adult learning. Within this vast education ecosystem, a major challenge for all teachers is to [...] Read more.
As teachers, we are living and working in times of abundant challenge and change. These challenges transpire across different education levels and sectors, including K–12, vocational, tertiary, and adult learning. Within this vast education ecosystem, a major challenge for all teachers is to allocate time, effort, and resources to ensure that their students receive a quality education with real-world implications, influencing soft-skill attainment, such as teamwork, communication, and critical thinking skills. In this article, the authors discuss, through a theoretical lens, the value of considering a national and universal approach to self- and peer-evaluations of authentic assessment tasks to improve teacher practice in Australia. Currently, there is modest opportunity amongst K–12 and tertiary teachers to comprehensively learn together, limiting cross-fertilisation of practice and interconnectedness, and as a national community of practice. The authors argue in this paper that offering an avenue to share knowledge and practice in authentic assessment design could potentially assist in addressing this challenge. Therefore, the article is dedicated to exploring the barriers and opportunities to advance a national and universal approach to transferable professional development in authentic assessment practice within the Australian education ecosystem. Full article
14 pages, 2317 KiB  
Article
Gender Prediction of Generated Tweets Using Generative AI
by Jalal S. Alowibdi
Information 2024, 15(8), 452; https://doi.org/10.3390/info15080452 - 1 Aug 2024
Viewed by 1079
Abstract
With the use of Generative AI (GenAI), Online Social Networks (OSNs) now generate a huge volume of content data. Yet, user-generated content on OSNs, aided by GenAI, presents challenges in analyzing and understanding its characteristics. In particular, tweets generated by GenAI at the [...] Read more.
With the use of Generative AI (GenAI), Online Social Networks (OSNs) now generate a huge volume of content data. Yet, user-generated content on OSNs, aided by GenAI, presents challenges in analyzing and understanding its characteristics. In particular, tweets generated by GenAI at the request of authentic human users present difficulties in determining the gendered variation of the content. The vast amount of data generated from tweets’ content necessitates a thorough investigation into the gender-specific language used in these tweets. This study explores the task of predicting the gender of text content in tweets generated by GenAI. Through our analysis and experimentation, we have achieved a remarkable 90% accuracy in attributing gender-specific language to these tweets. Our research not only highlights the potential of GenAI in gender prediction but also underscores the sophisticated techniques employed to decipher the refined linguistic cues that differentiate male and female language in GenAI-generated content. This advancement in understanding and predicting gender-specific language in GenAI-generated tweets covers the way for more refined and accurate content analysis in the evolving landscape of OSNs. Full article
(This article belongs to the Special Issue Second Edition of Predictive Analytics and Data Science)
Show Figures

Figure 1

Figure 1
<p>Flowchart of the proposed algorithm for gender prediction in GenAI-generated tweets.</p>
Full article ">Figure 2
<p>Performance metrics for Feature 500.</p>
Full article ">Figure 3
<p>Performance metrics for Feature 1000.</p>
Full article ">Figure 4
<p>Performance metrics for all features.</p>
Full article ">Figure 5
<p>Most-used words by GenAI-Male (<b>left</b>) and GenAI-Female (<b>right</b>).</p>
Full article ">Figure 6
<p>Least-used words by GenAI-Male (<b>left</b>) and GenAI-Female (<b>right</b>).</p>
Full article ">
29 pages, 1362 KiB  
Article
An Efficient Certificateless Anonymous Signcryption Scheme for WBAN
by Weifeng Long, Lunzhi Deng, Jiwen Zeng, Yan Gao and Tianxiu Lu
Sensors 2024, 24(15), 4899; https://doi.org/10.3390/s24154899 - 28 Jul 2024
Viewed by 587
Abstract
A Wireless Body Area Network (WBAN), introduced into the healthcare sector to improve patient care and enhance the efficiency of medical services, also brings the risk of the leakage of patients’ privacy. Therefore, maintaining the communication security of patients’ data has never been [...] Read more.
A Wireless Body Area Network (WBAN), introduced into the healthcare sector to improve patient care and enhance the efficiency of medical services, also brings the risk of the leakage of patients’ privacy. Therefore, maintaining the communication security of patients’ data has never been more important. However, WBAN faces issues such as open medium channels, resource constraints, and lack of infrastructure, which makes the task of designing a secure and economical communication scheme suitable for WBAN particularly challenging. Signcryption has garnered attention as a solution suitable for resource-constrained devices, offering a combination of authentication and confidentiality with low computational demands. Although the advantages offered by existing certificateless signcryption schemes are notable, most of them only have proven security within the random oracle model (ROM), lack public ciphertext authenticity, and have high computational overheads. To overcome these issues, we propose a certificateless anonymous signcryption (CL-ASC) scheme suitable for WBAN, featuring anonymity of the signcrypter, public verifiability, and public ciphertext authenticity. We prove its security in the standard model, including indistinguishability, unforgeability, anonymity of the signcrypter, and identity identifiability, and demonstrate its superiority over relevant schemes in terms of security, computational overheads, and storage costs. Full article
(This article belongs to the Section Wearables)
Show Figures

Figure 1

Figure 1
<p>WBAN Framework.</p>
Full article ">Figure 2
<p>Schematic of system model.</p>
Full article ">Figure 3
<p>Comparison of computational overheads [<a href="#B28-sensors-24-04899" class="html-bibr">28</a>,<a href="#B30-sensors-24-04899" class="html-bibr">30</a>,<a href="#B32-sensors-24-04899" class="html-bibr">32</a>,<a href="#B40-sensors-24-04899" class="html-bibr">40</a>,<a href="#B46-sensors-24-04899" class="html-bibr">46</a>,<a href="#B49-sensors-24-04899" class="html-bibr">49</a>,<a href="#B50-sensors-24-04899" class="html-bibr">50</a>,<a href="#B51-sensors-24-04899" class="html-bibr">51</a>,<a href="#B53-sensors-24-04899" class="html-bibr">53</a>].</p>
Full article ">Figure 4
<p>Comparison of storage costs [<a href="#B28-sensors-24-04899" class="html-bibr">28</a>,<a href="#B30-sensors-24-04899" class="html-bibr">30</a>,<a href="#B32-sensors-24-04899" class="html-bibr">32</a>,<a href="#B40-sensors-24-04899" class="html-bibr">40</a>,<a href="#B46-sensors-24-04899" class="html-bibr">46</a>,<a href="#B49-sensors-24-04899" class="html-bibr">49</a>,<a href="#B50-sensors-24-04899" class="html-bibr">50</a>,<a href="#B51-sensors-24-04899" class="html-bibr">51</a>,<a href="#B53-sensors-24-04899" class="html-bibr">53</a>].</p>
Full article ">
16 pages, 2449 KiB  
Article
Enhancing Knowledge-Concept Recommendations with Heterogeneous Graph-Contrastive Learning
by Liting Wei, Yun Li, Weiwei Wang and Yi Zhu
Mathematics 2024, 12(15), 2324; https://doi.org/10.3390/math12152324 - 25 Jul 2024
Viewed by 608
Abstract
With the implementation of conceptual labeling on online learning resources, knowledge-concept recommendations have been introduced to pinpoint concepts that learners may wish to delve into more deeply. As the core subject of learning, learners’ preferences in knowledge concepts should be given greater attention. [...] Read more.
With the implementation of conceptual labeling on online learning resources, knowledge-concept recommendations have been introduced to pinpoint concepts that learners may wish to delve into more deeply. As the core subject of learning, learners’ preferences in knowledge concepts should be given greater attention. Research indicates that learners’ preferences for knowledge concepts are influenced by the characteristics of their group structure. There is a high degree of homogeneity within a group, and notable distinctions exist between the internal and external configurations of a group. To strengthen the group-structure characteristics of learners’ behaviors, a multi-task strategy for knowledge-concept recommendations is proposed; this strategy is called Knowledge-Concept Recommendations with Heterogeneous Graph-Contrastive Learning. Specifically, due to the difficulty of accessing authentic social networks, learners and their structural neighbors are considered positive contrastive pairs to construct self-supervision signals on the predefined meta-path from heterogeneous information networks as auxiliary tasks, which capture the higher-order neighbors of learners by presenting different perspectives. Then, the Information Noise-Contrastive Estimation loss is regarded as the main training objective to increase the differentiation of learners from different professional backgrounds. Extensive experiments are constructed on MOOCCube, and we find that our proposed method outperforms the other state-of-the-art concept-recommendation methods, achieving 6.66% with HR@5, 8.85% with NDCG@5, and 8.68% with MRR. Full article
Show Figures

Figure 1

Figure 1
<p>MOOCCube dataset behavior heatmap showing probability as the proportion of learners’ clicks on the intersection and union of knowledge concepts.</p>
Full article ">Figure 2
<p>The method flow of the KCRHGCL framework, which consists of the following three main workflows: (i) the representation learning of the learners—in the constructed heterogeneous graph, appropriate meta-paths are selected, graph convolutional networks are used to extract features on these predefined meta-paths, and structural neighbors are introduced to enhance learner representations; (ii) the representation learning of the concepts—heterogeneous graph attention networks are used to capture the representations of concepts from different meta-paths; and (iii) the model optimization of the multi-task—negative sampling techniques and Information Noise Contrastive Estimation (InfoNCE) are utilized as the training objective loss to more effectively distinguish between differences in learners’ preferences across group structures.</p>
Full article ">Figure 3
<p>The results of the embedding dimension size <span class="html-italic">d</span>.</p>
Full article ">Figure 4
<p>The results of the regularization weights <math display="inline"><semantics> <mi>μ</mi> </semantics></math>.</p>
Full article ">Figure 5
<p>Convergence analysis for KCRHGCL.</p>
Full article ">
21 pages, 18061 KiB  
Article
SpaceLight: A Framework for Enhanced On-Orbit Navigation Imagery
by Zhang Zhang, Jiaqi Feng, Liang Chang, Lei Deng, Dong Li and Chaoming Si
Aerospace 2024, 11(7), 503; https://doi.org/10.3390/aerospace11070503 - 23 Jun 2024
Viewed by 678
Abstract
In the domain of space rendezvous and docking, visual navigation plays a crucial role. However, practical applications frequently encounter issues with poor image quality. Factors such as lighting uncertainties, spacecraft motion, uneven illumination, and excessively dark environments collectively pose significant challenges, rendering recognition [...] Read more.
In the domain of space rendezvous and docking, visual navigation plays a crucial role. However, practical applications frequently encounter issues with poor image quality. Factors such as lighting uncertainties, spacecraft motion, uneven illumination, and excessively dark environments collectively pose significant challenges, rendering recognition and measurement tasks during visual navigation nearly infeasible. The existing image enhancement methods, while visually appealing, compromise the authenticity of the original images. In the specific context of visual navigation, space image enhancement’s primary aim is the faithful restoration of the spacecraft’s mechanical structure with high-quality outcomes. To address these issues, our study introduces, for the first time, a dedicated unsupervised framework named SpaceLight for enhancing on-orbit navigation images. The framework integrates a spacecraft semantic parsing network, utilizing it to generate attention maps that pinpoint structural elements of spacecraft in poorly illuminated regions for subsequent enhancement. To more effectively recover fine structural details within these dark areas, we propose the definition of a global structure loss and the incorporation of a pre-enhancement module. The proposed SpaceLight framework adeptly restores structural details in extremely dark areas while distinguishing spacecraft structures from the deep-space background, demonstrating practical viability when applied to visual navigation. This paper is grounded in space on-orbit servicing engineering projects, aiming to address visual navigation practical issues. It pioneers the utilization of authentic on-orbit navigation images in the research, resulting in highly promising and unprecedented outcomes. Comprehensive experiments demonstrate SpaceLight’s superiority over state-of-the-art low-light enhancement algorithms, facilitating enhanced on-orbit navigation image quality. This advancement offers robust support for subsequent visual navigation. Full article
Show Figures

Figure 1

Figure 1
<p>Spacecraft Semantic Parsing Network Structure.</p>
Full article ">Figure 2
<p>Attention Map.</p>
Full article ">Figure 3
<p>On-Orbit Navigation Image Enhancement Network.</p>
Full article ">Figure 4
<p>Pre-Enhancement Laplacian Pyramid.</p>
Full article ">Figure 5
<p>Visual Outcomes of Different Image Enhancement Algorithms.</p>
Full article ">Figure 6
<p>Comprehensive Performance Comparison Across All Tests.</p>
Full article ">Figure 7
<p>Impact of Component Variations on Image Generation.</p>
Full article ">Figure 8
<p>Comparative Semantic Segmentation Results of Various Image Enhancement Algorithms.</p>
Full article ">
16 pages, 1194 KiB  
Article
CoreTemp: Coreset Sampled Templates for Multimodal Mobile Biometrics
by Jaeho Yoon, Jaewoo Park, Jungyun Kim and Andrew Beng Jin Teoh
Appl. Sci. 2024, 14(12), 5183; https://doi.org/10.3390/app14125183 - 14 Jun 2024
Viewed by 875
Abstract
Smart devices have become the core ingredient in maintaining human society, where their applications span basic telecommunication, entertainment, education, and even critical security tasks. However, smartphone security measures have not kept pace with their ubiquitousness and convenience, exposing users to potential security breaches. [...] Read more.
Smart devices have become the core ingredient in maintaining human society, where their applications span basic telecommunication, entertainment, education, and even critical security tasks. However, smartphone security measures have not kept pace with their ubiquitousness and convenience, exposing users to potential security breaches. Shading light on shortcomings of traditional security measures such as PINs gives rise to biometrics-based security measures. Open-set authentication with pretrained Transformers especially shows competitive performance in this context. Bringing this closer to practice, we propose CoreTemp, a greedy coreset sampled template, which offers substantially faster authentication speeds. In parallel with CoreTemp, we design a fast match algorithm where the combination shows robust performance in open-set mobile biometrics authentication. Designed to resemble the effects of ensembles with marginal increment in computation, we propose PIEformer+, where its application with CoreTemp has state-of-the-art performance. Benefiting from much more efficient authentication speeds to the best of our knowledge, we are the first to attempt identification in this context. Our proposed methodology achieves state-of-the-art results on HMOG and BBMAS datasets, particularly with much lower computational costs. In summary, this research introduces a novel integration of greedy coreset sampling with an advanced form of pretrained, implicitly ensembled Transformers (PIEformer+), greatly enhancing the speed and efficiency of mobile biometrics authentication, and also enabling identification, which sets a new benchmark in the relevant field. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>A pipeline of (<b>a</b>) authentication and (<b>b</b>) identification with CoreTemp.</p>
Full article ">Figure 2
<p>Illustration of PIEformer (<b>left</b>) and PIEformer+ (<b>right</b>). The red circles indicate attention mechanisms.</p>
Full article ">Figure 3
<p>The averaged ROC curve for intra-dataset experiments, as a more detailed presentation of <a href="#applsci-14-05183-t002" class="html-table">Table 2</a>. Best viewed in color.</p>
Full article ">Figure 4
<p>CMC curve of Transformer, PIEformer, and PIEformer+. Best viewed in color.</p>
Full article ">Figure 5
<p>EER and Rank1 values according to the sampling ratio. Points marked with “X” indicate an average of 10 templates per user, which we consider the lower bound, where its ratio <span class="html-italic">p</span> approximates the closest value.</p>
Full article ">
Back to TopTop