[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (530)

Search Parameters:
Keywords = virtual representation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 46116 KiB  
Article
Structural Complexity Significantly Impacts Canopy Reflectance Simulations as Revealed from Reconstructed and Sentinel-2-Monitored Scenes in a Temperate Deciduous Forest
by Yi Gan, Quan Wang and Guangman Song
Remote Sens. 2024, 16(22), 4296; https://doi.org/10.3390/rs16224296 - 18 Nov 2024
Viewed by 249
Abstract
Detailed three-dimensional (3D) radiative transfer models (RTMs) enable a clear understanding of the interactions between light, biochemistry, and canopy structure, but they are rarely explicitly evaluated due to the availability of 3D canopy structure data, leading to a lack of knowledge on how [...] Read more.
Detailed three-dimensional (3D) radiative transfer models (RTMs) enable a clear understanding of the interactions between light, biochemistry, and canopy structure, but they are rarely explicitly evaluated due to the availability of 3D canopy structure data, leading to a lack of knowledge on how canopy structure/leaf characteristics affect radiative transfer processes within forest ecosystems. In this study, the newly released 3D RTM Eradiate was extensively evaluated based on both virtual scenes reconstructed using the quantitative structure model (QSM) by adding leaves to point clouds generated from terrestrial laser scanning (TLS) data, and real scenes monitored by Sentinel-2 in a typical temperate deciduous forest. The effects of structural parameters on reflectance were investigated through sensitivity analysis, and the performance of the 3D model was compared with the 5-Scale and PROSAIL radiative transfer models. The results showed that the Eradiate-simulated reflectance achieved good agreement with the Sentinel-2 reflectance, especially in the visible and near-infrared spectral regions. Furthermore, the simulated reflectance, particularly in the blue and shortwave infrared spectral bands, was clearly shown to be influenced by canopy structure using the Eradiate model. This study demonstrated that the Eradiate RTM, based on the 3D explicit representation, is capable of providing accurate radiative transfer simulations in the temperate deciduous forest and hence provides a basis for understanding tree interactions and their effects on ecosystem structure and functions. Full article
Show Figures

Figure 1

Figure 1
<p>The workflow of this study.</p>
Full article ">Figure 2
<p>The locations of the seven TLS-measured points with a Sentinel-2 Level-2A MSI image (R: B4; G: B3; B: B2) as the base map (CRS: EPSG:6676—JGD2011/Japan Plane Rectangular CS VIII).</p>
Full article ">Figure 3
<p>The basic leaf shape and information used in the addition of leaves on the branches.</p>
Full article ">Figure 4
<p>The reconstruction processes of 3D forest scene based on TLS point clouds. (<b>a</b>) Raw plot-based point clouds after co-registration with multiple scans; (<b>b</b>) segmented vegetation points and ground points, colorized with black and red; (<b>c</b>) vegetation points and colorized seed clusters; (<b>d</b>) segmented and filtered tree points by Dijkstra segmentation algorithm; (<b>e</b>) reconstructed 3D tree quantitative structure models with TreeQSM approach; (<b>f</b>) generated virtual forest scene for RTM simulations, and 3D tree QSMs coupling with FaNNI foliage insertion algorithm.</p>
Full article ">Figure 5
<p>Seven reconstructed virtual forest scenes for RTM simulations in this study.</p>
Full article ">Figure 6
<p>Results of global sensitivity analysis for input parameters to the bidirectional reflectance factor (BRF) in the Eradiate radiative transfer model.</p>
Full article ">Figure 7
<p>Influence of leaf area index (LAI, from 1.0 to 7.0) on the simulated bidirectional reflectance factor (BRF) of the Eradiate (<b>a</b>–<b>j</b>), 5-Scale (<b>k</b>–<b>t</b>), and PROSAIL (<b>u</b>–<b>D</b>) radiative transfer models at the solar zenith angle (SZA) of 30° over different view zenith angles (VZAs).</p>
Full article ">Figure 8
<p>Relationships of the reflectance at different leaf area index (LAI) levels between Eradiate and 5-Scale as well as PROSAIL radiative transfer models (RTMs).</p>
Full article ">Figure 9
<p>Comparison of simulated reflectance from (<b>a</b>) Eradiate, (<b>b</b>) 5-Scale, and (<b>c</b>) PROSAIL with the reflectance extracted from Sentinel-2 MSI images. The line and shaded area depict the mean and standard deviation of reflectance.</p>
Full article ">Figure 10
<p>The performance ((<b>a</b>) RMSE; (<b>b</b>) MB; (<b>c</b>) MGE) of Eradiate, 5-Scale, and PROSAIL radiative transfer models (RTMs) for reflectance simulation vs. Sentinel-2-extracted reflectance.</p>
Full article ">
20 pages, 8781 KiB  
Article
A Virtual View Acquisition Technique for Complex Scenes of Monocular Images Based on Layered Depth Images
by Qi Wang and Yan Piao
Appl. Sci. 2024, 14(22), 10557; https://doi.org/10.3390/app142210557 - 15 Nov 2024
Viewed by 398
Abstract
With the rapid development of stereoscopic display technology, how to generate high-quality virtual view images has become the key in the applications of 3D video, 3D TV and virtual reality. The traditional virtual view rendering technology maps the reference view into the virtual [...] Read more.
With the rapid development of stereoscopic display technology, how to generate high-quality virtual view images has become the key in the applications of 3D video, 3D TV and virtual reality. The traditional virtual view rendering technology maps the reference view into the virtual view by means of 3D transformation, but when the background area is occluded by the foreground object, the content of the occluded area cannot be inferred. To solve this problem, we propose a virtual view acquisition technique for complex scenes of monocular images based on a layered depth image (LDI). Firstly, the depth discontinuities of the edge of the occluded area are reasonably grouped by using the multilayer representation of the LDI, and the depth edge of the occluded area is inpainted by the edge inpainting network. Then, the generative adversarial network (GAN) is used to fill the information of color and depth in the occluded area, and the inpainting virtual view is generated. Finally, GAN is used to optimize the color and depth of the virtual view, and the high-quality virtual view is generated. The effectiveness of the proposed method is proved by experiments, and it is also applicable to complex scenes. Full article
Show Figures

Figure 1

Figure 1
<p>The overall frame of virtual viewpoint image generation.</p>
Full article ">Figure 2
<p>Depth images of various types generated by the method proposed in this paper.</p>
Full article ">Figure 3
<p>Depth image preprocessing. (<b>a</b>) The input RGB image. (<b>b</b>) Depth image after filtering. (<b>c</b>) The enlarged image of the red box area in (<b>b</b>); (<b>d</b>) The preprocessed image of (<b>c</b>). (<b>e</b>) The image of lines with discontinuous depth.</p>
Full article ">Figure 4
<p>The area division of the input RGB image. (<b>a</b>) The input RBG image. (<b>b</b>) Generated virtual view image without inpainting. The pink area is the foreground area, the gray area is the occluded area, and the blue area is the background area.</p>
Full article ">Figure 5
<p>The framework of the edge inpainting network.</p>
Full article ">Figure 6
<p>The framework of virtual view optimization network.</p>
Full article ">Figure 7
<p>Virtual viewpoint images generated at different positions. (<b>a</b>) The images of the standard model established by C4D, and the model position is x = 0; (<b>b</b>) Viewpoint images of the model generated by C4D at x = −3; (<b>c</b>) The virtual viewpoint images of the model estimated by the method in this paper at x = −3; (<b>d</b>) Viewpoint images of the model generated by C4D at x = +3; (<b>e</b>) Virtual viewpoint images of the model estimated by the method in this paper at x = +3.</p>
Full article ">Figure 8
<p>Camera distributions.</p>
Full article ">Figure 9
<p>Generated virtual viewpoint images using the ballet image sequence. (<b>a</b>) The input image, which is the 10th frame taken by Cam4. (<b>b</b>) The 10th frame image taken by Cam3. (<b>c</b>) The 10th frame image taken by Cam5. (<b>d</b>) The 10th frame image of Cam3, which is generated by the method in this paper. (<b>e</b>) The 10th frame image of Cam5, which is generated by the method in this paper.</p>
Full article ">Figure 10
<p>Generated virtual viewpoint images using the breakdancers image sequence. (<b>a</b>) The input image, which is the 20th frame taken by Cam4. (<b>b</b>) The 20th frame image taken by Cam3. (<b>c</b>) The 20th frame image taken by Cam5. (<b>d</b>) The 20th frame image of Cam3, which is generated by the method in this paper. (<b>e</b>) The 20th frame image of Cam5, which is generated by the method in this paper.</p>
Full article ">Figure 11
<p>Different types of virtual viewpoint images rendered by the method in this paper.</p>
Full article ">Figure 12
<p>Rendered virtual viewpoint images. (<b>a</b>) Input images; (<b>b</b>) The partial enlarged images of the contents in red boxes in (<b>a</b>); (<b>c</b>) The true images of virtual viewpoint images (<b>d</b>,<b>e</b>); (<b>d</b>) Virtual viewpoint images generated by the method of [<a href="#B58-applsci-14-10557" class="html-bibr">58</a>]; (<b>e</b>) Virtual viewpoint images generated by the method in this paper.</p>
Full article ">
26 pages, 14206 KiB  
Article
The Heritage Building Information Modeling Methodology for Structural Diagnosis: An Integrated System of Digital Models for the Baptistery of San Giovanni in Pisa
by Giada Bartolini, Anna De Falco, Lorenzo Gianfranceschi, Massimiliano Martino and Laura Vignali
Heritage 2024, 7(11), 6366-6391; https://doi.org/10.3390/heritage7110298 (registering DOI) - 15 Nov 2024
Viewed by 251
Abstract
The structural diagnosis of monumental buildings necessitates organizing diverse cross-disciplinary data. The H-BIM procedure employs 3D digital models to create a comprehensive virtual repository, offering advantages in documentation access, interoperability, intervention design, cost evaluation, and maintenance management. This work proposes an approach to [...] Read more.
The structural diagnosis of monumental buildings necessitates organizing diverse cross-disciplinary data. The H-BIM procedure employs 3D digital models to create a comprehensive virtual repository, offering advantages in documentation access, interoperability, intervention design, cost evaluation, and maintenance management. This work proposes an approach to combining different models while addressing interoperability challenges by best exploiting their positive characteristics. After evaluating the advantages and limitations of textured-mesh and NURBS-based models, and virtual reality environments based on specific comparison criteria, an integrated system of these models within the H-BIM framework is proposed. The latter is applied to study the relevant case of the Baptistery of San Giovanni in Pisa, Italy. The integrated H-BIM model is designed primarily to facilitate the structural diagnosis of the monument, and illustrates how combining different 3D representations, each providing multiple information with different levels of detail, enhances its capabilities. This integration results in a more effective tool for the multidisciplinary conservation of cultural heritage, accommodating a wide range of data beyond structural aspects, thus fostering collaboration among professionals from various fields of expertise. Full article
Show Figures

Figure 1

Figure 1
<p>Workflow of the methodology proposed to create an integrated system of digital representations revolving around an H-BIM model.</p>
Full article ">Figure 2
<p>Textured-mesh model generative procedure.</p>
Full article ">Figure 3
<p>NURBS-based model generative procedure.</p>
Full article ">Figure 4
<p>Virtual reality environment generative procedure.</p>
Full article ">Figure 5
<p>Geometric level, information level, and connection level of the H-BIM model.</p>
Full article ">Figure 6
<p>(<b>a</b>) San Giovanni Baptistery in Cathedral Square (Pisa, Italy); (<b>b</b>) vertical section; (<b>c</b>) internal view; (<b>d</b>) ground floor plan view.</p>
Full article ">Figure 7
<p>(<b>a</b>) Textured-mesh model; (<b>b</b>) NURBS-based model.</p>
Full article ">Figure 8
<p>VR environment.</p>
Full article ">Figure 9
<p>Details of the VR environment.</p>
Full article ">Figure 10
<p>Grasshopper canvas: Morph-solid and Morph-settings.</p>
Full article ">Figure 11
<p>Symbols materializing measuring points. The material is available at <a href="http://github.com/giadabart/bapt" target="_blank">github.com/giadabart/bapt</a> (accessed on 9 November 2024) and <a href="http://github.com/carloresta/baptistery-data" target="_blank">github.com/carloresta/baptistery-data</a> (accessed on 9 November 2024).</p>
Full article ">Figure 12
<p>Example of property settings for the object group <span class="html-italic">Foundations</span>. The material is available at <a href="http://github.com/giadabart/bapt" target="_blank">github.com/giadabart/bapt</a> (accessed on 9 November 2024).</p>
Full article ">Figure 13
<p>Example of property settings for the object group <span class="html-italic">Cracks</span>. The material is available at <a href="http://maxmartino.it/battistero/index.htm" target="_blank">maxmartino.it/battistero/index.htm</a> (accessed on 9 November 2024).</p>
Full article ">Figure 14
<p>Example of property settings for the object group <span class="html-italic">Generic H-BIM object</span>. The material is available at <a href="http://maxmartino.it/battistero/index.htm" target="_blank">maxmartino.it/battistero/index.htm</a> (accessed on 9 November 2024) and at <a href="http://github.com/giadabart/bapt" target="_blank">github.com/giadabart/bapt</a> (accessed on 9 November 2024).</p>
Full article ">Figure 15
<p>Archicad Graphic override of the construction phases of the Baptistery.</p>
Full article ">Figure 16
<p>Archicad Graphic override of the building materials of the Baptistery.</p>
Full article ">
13 pages, 5412 KiB  
Article
Supervised Contrastive Learning for 3D Cross-Modal Retrieval
by Yeon-Seung Choo, Boeun Kim, Hyun-Sik Kim and Yong-Suk Park
Appl. Sci. 2024, 14(22), 10322; https://doi.org/10.3390/app142210322 - 10 Nov 2024
Viewed by 425
Abstract
Interoperability between different virtual platforms requires the ability to search and transfer digital assets across platforms. Digital assets in virtual platforms are represented in different forms or modalities, such as images, meshes, and point clouds. The cross-modal retrieval of three-dimensional (3D) object representations [...] Read more.
Interoperability between different virtual platforms requires the ability to search and transfer digital assets across platforms. Digital assets in virtual platforms are represented in different forms or modalities, such as images, meshes, and point clouds. The cross-modal retrieval of three-dimensional (3D) object representations is challenging due to data representation diversity, making common feature space discovery difficult. Recent studies have been focused on obtaining feature consistency within the same classes and modalities using cross-modal center loss. However, center features are sensitive to hyperparameter variations, making cross-modal center loss susceptible to performance degradation. This paper proposes a new 3D cross-modal retrieval method that uses cross-modal supervised contrastive learning (CSupCon) and the fixed projection head (FPH) strategy. Contrastive learning mitigates the influence of hyperparameters by maximizing feature distinctiveness. The FPH strategy prevents gradient updates in the projection network, enabling the focused training of the backbone networks. The proposed method shows a mean average precision (mAP) increase of 1.17 and 0.14 in 3D cross-modal object retrieval experiments using ModelNet10 and ModelNet40 datasets compared to state-of-the-art (SOTA) methods. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Traditional SimCLR [<a href="#B8-applsci-14-10322" class="html-bibr">8</a>] method using single modal data augmentations (<b>left</b>). The augmented data (marked with symbols ′ and ″) are adapted in supervised learning to aggregate representation features. Supervised contrastive learning (SupCon) [<a href="#B14-applsci-14-10322" class="html-bibr">14</a>] adapted SimCLR into supervised learning tasks in single modality (<b>middle</b>). Our method, CSupCon, applied contrastive learning to cross-modal tasks (<b>right</b>). The numbers represent different data instances. The rectangles and circles represent different modalities, and the different colors represent different classes. The blue and red lines indicate positive and negative instances.</p>
Full article ">Figure 2
<p>Overview of the proposed method. In the feature extraction stage, an augmented data instance <math display="inline"><semantics> <msup> <mi>x</mi> <mo>′</mo> </msup> </semantics></math> and another augmented data instance <math display="inline"><semantics> <msup> <mi>x</mi> <mrow> <mo>″</mo> </mrow> </msup> </semantics></math> from input <span class="html-italic">x</span>, embedding features <math display="inline"><semantics> <msup> <mi>v</mi> <mo>′</mo> </msup> </semantics></math> and <math display="inline"><semantics> <msup> <mi>v</mi> <mrow> <mo>″</mo> </mrow> </msup> </semantics></math>, are extracted from each modality using its corresponding backbone network. The proposed method, cross-modal supervised contrastive learning (CSupCon), pushes the features away for different classes and pulls the features towards each other for the same classes. On the other side, in the fixed projection head (FPH) strategy, the <math display="inline"><semantics> <msup> <mi>v</mi> <mo>′</mo> </msup> </semantics></math> features are used to predict semantic labels for classification.</p>
Full article ">Figure 3
<p>The visualization result of feature clustering from the ModelNet40 test data.</p>
Full article ">Figure 4
<p>The results of cross-modal retrieval using the proposed method from the ModelNet40 test data.</p>
Full article ">Figure 5
<p>The result of cross-modal retrieval on the ModelNet40 test data by class. The illustration depicts sorted classes based on the amount of training data and their corresponding mAPs. In general, results are not favorable for classes with a small (less than 200 in this example) number of training data (i.e., classes included in the blue dotted rectangle).</p>
Full article ">
18 pages, 7263 KiB  
Article
MPCTrans: Multi-Perspective Cue-Aware Joint Relationship Representation for 3D Hand Pose Estimation via Swin Transformer
by Xiangan Wan, Jianping Ju, Jianying Tang, Mingyu Lin, Ning Rao, Deng Chen, Tingting Liu, Jing Li, Fan Bian and Nicholas Xiong
Sensors 2024, 24(21), 7029; https://doi.org/10.3390/s24217029 - 31 Oct 2024
Viewed by 527
Abstract
The objective of 3D hand pose estimation (HPE) based on depth images is to accurately locate and predict keypoints of the hand. However, this task remains challenging because of the variations in hand appearance from different viewpoints and severe occlusions. To effectively address [...] Read more.
The objective of 3D hand pose estimation (HPE) based on depth images is to accurately locate and predict keypoints of the hand. However, this task remains challenging because of the variations in hand appearance from different viewpoints and severe occlusions. To effectively address these challenges, this study introduces a novel approach, called the multi-perspective cue-aware joint relationship representation for 3D HPE via the Swin Transformer (MPCTrans, for short). This approach is designed to learn multi-perspective cues and essential information from hand depth images. To achieve this goal, three novel modules are proposed to utilize features from multiple virtual views of the hand, namely, the adaptive virtual multi-viewpoint (AVM), hierarchy feature estimation (HFE), and virtual viewpoint evaluation (VVE) modules. The AVM module adaptively adjusts the angles of the virtual viewpoint and learns the ideal virtual viewpoint to generate informative multiple virtual views. The HFE module estimates hand keypoints through hierarchical feature extraction. The VVE module evaluates virtual viewpoints by using chained high-level functions from the HFE module. Transformer is used as a backbone to extract the long-range semantic joint relationships in hand depth images. Extensive experiments demonstrate that the MPCTrans model achieves state-of-the-art performance on four challenging benchmark datasets. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>Existing challenges in HPE: (<b>a</b>–<b>c</b>) serious occlusions and (<b>d</b>–<b>f</b>) variations in hand appearance from different viewpoints. Some of or even most parts of the hand are missing in these scenarios, resulting in difficulties in HPE. The top line shows RGB images, while the bottom line presents the corresponding depth images.</p>
Full article ">Figure 2
<p>Two key characteristics are revealed from the observation of hand images. Characteristic I: (<b>a</b>,<b>b</b>) represent the inherent relationships between hand joints. Characteristic II: (<b>c</b>,<b>d</b>) indicate the multi-perspective cues of the hand. These relationships and cues can be computed using a self-attention mechanism.</p>
Full article ">Figure 3
<p>Overview of the MPCTrans model. First, hand depth images are converted into 3D point clouds via the AVM module, which employs adaptive learning at virtual viewpoints to generate optimal virtual multi-view depth images. Second, these adaptive virtual multi-view depth images are partitioned into windows, with attention mechanisms applied only within these partitions. Linear projection and position embedding techniques are utilized to transform patches into 1D vectors. Third, three HFE modules leverage information from lower layers and the output features from the final stage to estimate hand poses for each view. Fourth, the feature maps from the three HFE modules are concatenated. Subsequently, the VVE module assesses this concatenated feature map to assign a score to each view. Finally, the pose estimates and results from each view are fused to produce the final hand pose prediction.</p>
Full article ">Figure 4
<p>Illustration of the AVM module. Original depth images may not always provide the best perspective for pose estimation. Our method adaptively learns <span class="html-italic">M</span> optimal virtual views from <span class="html-italic">M</span> initial virtual views, where <span class="html-italic">M</span> is set as 25.</p>
Full article ">Figure 5
<p>Attention visualization of different perspectives in the VVE module. Each perspective contributes differently to HPE, resulting in different attention visualizations in the multi-head attention of the VVE module. Consequently, each perspective has a different rating.</p>
Full article ">Figure 6
<p>Comparison of the MPCTrans model with state-of-the-art methods (DeepPrior++ [<a href="#B5-sensors-24-07029" class="html-bibr">5</a>], HandPointNet [<a href="#B16-sensors-24-07029" class="html-bibr">16</a>], DenseReg [<a href="#B6-sensors-24-07029" class="html-bibr">6</a>], Point-to-Point [<a href="#B10-sensors-24-07029" class="html-bibr">10</a>], A2J [<a href="#B7-sensors-24-07029" class="html-bibr">7</a>], V2V [<a href="#B11-sensors-24-07029" class="html-bibr">11</a>], VVS [<a href="#B20-sensors-24-07029" class="html-bibr">20</a>]) on the NYU and ICVL datasets. (<b>a</b>,<b>b</b>) Mean joint error per hand joint and percentage of successful frames over different error thresholds in the NYU dataset. (<b>c</b>,<b>d</b>) Mean joint error per hand joint and percentage of successful frames across various error thresholds in the ICVL dataset.</p>
Full article ">Figure 7
<p>Viewpoint initialization scheme. (<b>a</b>), (<b>b</b>), (<b>c</b>), and (<b>d</b>) represent the initial virtual viewpoint positions for 4, 9, 16, and 25 adaptive virtual multi-views, respectively.</p>
Full article ">Figure 8
<p>Visualization of different numbers of adapted virtual multi-views: (<b>a</b>), (<b>b</b>), (<b>c</b>), and (<b>d</b>) represent the 4, 9, 16, and 25 adapted virtual multi-views, respectively, learned from the initial virtual viewpoint positions shown in <a href="#sensors-24-07029-f007" class="html-fig">Figure 7</a>.</p>
Full article ">Figure 9
<p>Comparison of the visualization results of the MPCTrans model with different numbers of adaptive virtual multi-views on the ICVL dataset. “ours-4views”, “ours-9views”, “ours-16views”, and “ours-25views” represent the results of the model with 4, 9, 16, and 25 adaptive virtual multi-views, respectively.</p>
Full article ">Figure 10
<p>Comparison of the visualization results of the MPCTrans model with different numbers of adaptive virtual multi-views on the NYU dataset.</p>
Full article ">
23 pages, 11099 KiB  
Article
Creating Digital Twins of Robotic Stations Using a Laser Tracker
by Dariusz Szybicki, Magdalena Muszyńska, Paulina Pietruś, Andrzej Burghardt and Krzysztof Kurc
Electronics 2024, 13(21), 4271; https://doi.org/10.3390/electronics13214271 - 31 Oct 2024
Viewed by 446
Abstract
This article deals with the design and creation of digital twins of robotic stations. A literature review of digital twins, robot programming methods and laser tracker applications is presented. This paper shows that the construction of digital twins is closely related to one [...] Read more.
This article deals with the design and creation of digital twins of robotic stations. A literature review of digital twins, robot programming methods and laser tracker applications is presented. This paper shows that the construction of digital twins is closely related to one of the most popular methods of robot programming, i.e., off-line programming. In the case of digital twins of robotic stations, modeling accuracy and two-way communication with the real station proved to be crucial. The article proposes a methodology for solving the basic problem of off-line robot programming, i.e., the limited accuracy of the representation of the station and the details. The algorithm of proceeding in the case when the station already exists and its digital model is built and the case when the digital model is first created and the real solution is built on its basis is shown. According to the developed methodology, a digital twin of a real robotic station was created and the possibilities arising from the use of virtual tools were shown. The developed digital twin has the ability to communicate with advanced Matlab 2021-type tools, uses cloud solutions and virtual and augmented reality for training, simulates physical phenomena and provides the ability to accurately program robots off-line. Full article
Show Figures

Figure 1

Figure 1
<p>A digital model of a robotic station with the position of a component defined relative to the robot base frame.</p>
Full article ">Figure 2
<p>Models of retroreflectors with bases: (<b>a</b>) allowing measurement of planes; (<b>b</b>) edges; (<b>c</b>) hole positions.</p>
Full article ">Figure 3
<p>A diagram of the implementation of the methodology associated with the variant when the robotic station is physically built and its digital twin is created.</p>
Full article ">Figure 4
<p>Position of base frame WB, Leica tracker WL and element WE.</p>
Full article ">Figure 5
<p>The flow of the element’s position and orientation data into the CAD software.</p>
Full article ">Figure 6
<p>A CAD model of the robotic station with planes determining the positions of the components.</p>
Full article ">Figure 7
<p>A diagram of the implementation of the methodology associated with the variant when the robotic station is only at the design stage and first a digital model is created, on the basis of which the real station will be built.</p>
Full article ">Figure 8
<p>Effects of using the digital model: (<b>a</b>) the workspace of the manipulator considering the tool used; (<b>b</b>) a fragment of the detailed drawing of the pedestal.</p>
Full article ">Figure 9
<p>The assembly of components using a laser tracker and dedicated software.</p>
Full article ">Figure 10
<p>Station element model: (<b>a</b>) mounting hole position; (<b>b</b>) element with hole and retroreflector with base.</p>
Full article ">Figure 11
<p>CAD model of IRB2400 robot along with photos of real components.</p>
Full article ">Figure 12
<p>IRB 2400 robot model during design process.</p>
Full article ">Figure 13
<p>IRB2400 robot: (<b>a</b>) excerpt from manufacturer’s technical documentation; (<b>b</b>) developed 3D model.</p>
Full article ">Figure 14
<p>IRB2400 robot model during FEM analysis.</p>
Full article ">Figure 15
<p>Cabinet with automation components: (<b>a</b>) 3D model of cabinet; (<b>b</b>) photo of real cabinet.</p>
Full article ">Figure 16
<p>Graphic showing of RobotStudio software’s ability to communicate with other software.</p>
Full article ">Figure 17
<p>Most commonly used methods of exchanging data between digital model and real station.</p>
Full article ">Figure 18
<p>A diagram of the implementation of the concept of digital twins and Industry 4.0 solutions.</p>
Full article ">Figure 19
<p>The robotic station that was used to create the digital twin.</p>
Full article ">Figure 20
<p>Photos of station measurements using retroreflector: (<b>a</b>) pedestal position measurement; (<b>b</b>) floor position measurement; (<b>c</b>) controller position measurement.</p>
Full article ">Figure 21
<p>Robotic station: (<b>a</b>) real station; (<b>b</b>) digital model of station in RobotStudio software.</p>
Full article ">Figure 22
<p>The digital twin of the robotic station in training applications: (<b>a</b>) RobotStudio AR app on a smartphone with training on the station; (<b>b</b>) an excerpt of the training delivered using 3D glasses.</p>
Full article ">
22 pages, 6101 KiB  
Article
Connection-Aware Digital Twin for Mobile Adhoc Networks in the 5G Era
by Manuel Jesús-Azabal, Zheng Zhang, Bingxia Gao, Jing Yang and Vasco N. G. J. Soares
Future Internet 2024, 16(11), 399; https://doi.org/10.3390/fi16110399 - 30 Oct 2024
Viewed by 535
Abstract
5G Mobile Adhoc Networks (5G-MANETs) are a popular and agile solution for data transmission in local contexts while maintaining communication with remote entities via 5G. These characteristics have established 5G-MANETs as versatile communication infrastructures for deploying contextual applications, leveraging physical proximity while exploiting [...] Read more.
5G Mobile Adhoc Networks (5G-MANETs) are a popular and agile solution for data transmission in local contexts while maintaining communication with remote entities via 5G. These characteristics have established 5G-MANETs as versatile communication infrastructures for deploying contextual applications, leveraging physical proximity while exploiting the possibilities of the Internet. As a result, there is growing interest in exploring the potential of these networks and their performance in real-world scenarios. However, the management and monitoring of 5G-MANETs are challenging due to their inherent characteristics, such as highly variable topology, unstable connections, energy consumption of individual devices, message routing, and occasional inability to connect to 5G. Considering these challenges, the proposed work aims to address real-time monitoring of 5G-MANETs using a connection-aware Digital Twin (DT). The approach provides two main functions: offering a live virtual representation of the network, even in scenarios where multiple nodes lack 5G connectivity, and estimating the performance of the infrastructure, enabling the specification of customized conditions. To achieve this, a communication architecture is proposed, analyzing its components and defining the involved processes. The DT is implemented and evaluated in a laboratory setting, assessing its accuracy in representing the physical network under varying conditions of topology and Internet availability. The results show 100% accuracy for the DT in fully connected topologies, with ultra-low latency averaging under 80 ms, and suitable performance in partially connected contexts, with latency averages below 3000 ms. Full article
(This article belongs to the Special Issue Advanced 5G and Beyond Networks)
Show Figures

Figure 1

Figure 1
<p>Overview of the DT for 5G-MANETs.</p>
Full article ">Figure 2
<p>Architecture of the proposal, detailing the components for physical 5G-MANETs and the DT.</p>
Full article ">Figure 3
<p>Topologies of MANETs and the applied strategy according to the presence of online nodes in the network.</p>
Full article ">Figure 4
<p>Connection-aware strategy for partially-offline topology.</p>
Full article ">Figure 5
<p>Architecture of the DT in the proof-of-concept.</p>
Full article ">Figure 6
<p>Defined topologies to assess the accuracy of the DT.</p>
Full article ">Figure 7
<p>Result set obtained in the evaluation of the proposal. (<b>a</b>) Delivery probability of the control messages transmitted by the 5G-MANET to the DT, comparing the four different evaluation contexts. (<b>b</b>) Latency of the control messages transmitted by the 5G-MANET to the DT, comparing the four different evaluation contexts. (<b>c</b>) Comparison of delivery probability obtained in the 5G-MANET and in the DT, at the four different evaluation contexts. (<b>d</b>) Comparison of latency obtained in the 5G-MANET and in the DT, at the four different evaluation contexts. (<b>e</b>) Comparison of overhead obtained in the 5G-MANET and in the DT, at the four different evaluation contexts. (<b>f</b>) Comparison of hops count obtained in the physical network and in the DT, at the four different evaluation contexts.</p>
Full article ">Figure 7 Cont.
<p>Result set obtained in the evaluation of the proposal. (<b>a</b>) Delivery probability of the control messages transmitted by the 5G-MANET to the DT, comparing the four different evaluation contexts. (<b>b</b>) Latency of the control messages transmitted by the 5G-MANET to the DT, comparing the four different evaluation contexts. (<b>c</b>) Comparison of delivery probability obtained in the 5G-MANET and in the DT, at the four different evaluation contexts. (<b>d</b>) Comparison of latency obtained in the 5G-MANET and in the DT, at the four different evaluation contexts. (<b>e</b>) Comparison of overhead obtained in the 5G-MANET and in the DT, at the four different evaluation contexts. (<b>f</b>) Comparison of hops count obtained in the physical network and in the DT, at the four different evaluation contexts.</p>
Full article ">Figure 8
<p>Estimated cost of the computational complexity. (<b>a</b>) Average energy consumption (mAh) of the nodes involved in the 5G-MANET across the four different evaluation contexts. Node A (*) acts as the gateway device of the device. (<b>b</b>) Average CPU usage of the MQTT broker and The ONE Simulator across the four different evaluation contexts.</p>
Full article ">
18 pages, 9899 KiB  
Article
A Robotic Teleoperation System with Integrated Augmented Reality and Digital Twin Technologies for Disassembling End-of-Life Batteries
by Feifan Zhao, Wupeng Deng and Duc Truong Pham
Batteries 2024, 10(11), 382; https://doi.org/10.3390/batteries10110382 - 30 Oct 2024
Viewed by 711
Abstract
Disassembly is a key step in remanufacturing, especially for end-of-life (EoL) products such as electric vehicle (EV) batteries, which are challenging to dismantle due to uncertainties in their condition and potential risks of fire, fumes, explosions, and electrical shock. To address these challenges, [...] Read more.
Disassembly is a key step in remanufacturing, especially for end-of-life (EoL) products such as electric vehicle (EV) batteries, which are challenging to dismantle due to uncertainties in their condition and potential risks of fire, fumes, explosions, and electrical shock. To address these challenges, this paper presents a robotic teleoperation system that leverages augmented reality (AR) and digital twin (DT) technologies to enable a human operator to work away from the danger zone. By integrating AR and DTs, the system not only provides a real-time visual representation of the robot’s status but also enables remote control via gesture recognition. A bidirectional communication framework established within the system synchronises the virtual robot with its physical counterpart in an AR environment, which enhances the operator’s understanding of both the robot and task statuses. In the event of anomalies, the operator can interact with the virtual robot through intuitive gestures based on information displayed on the AR interface, thereby improving decision-making efficiency and operational safety. The application of this system is demonstrated through a case study involving the disassembly of a busbar from an EoL EV battery. Furthermore, the performance of the system in terms of task completion time and operator workload was evaluated and compared with that of AR-based control methods without informational cues and ‘smartpad’ controls. The findings indicate that the proposed system reduces operation time and enhances user experience, delivering its broad application potential in complex industrial settings. Full article
(This article belongs to the Section Battery Processing, Manufacturing and Recycling)
Show Figures

Figure 1

Figure 1
<p>Framework of the proposed system.</p>
Full article ">Figure 2
<p>The physical robot and virtual robot. (<b>a</b>) Physical robot in the physical environment; (<b>b</b>) virtual robot registered on the physical robot by using the Vuforia image target.</p>
Full article ">Figure 3
<p>Key information in the AR interface. (<b>a</b>) Display in the Unity environment; (<b>b</b>) display in the AR HMD devices.</p>
Full article ">Figure 4
<p>AR-based robot control method. (<b>a</b>) Gesture recognition; (<b>b</b>) AR robot control interface.</p>
Full article ">Figure 5
<p>Coordinate system transformation.</p>
Full article ">Figure 6
<p>Teleoperated human–robot collaborative disassembly platform.</p>
Full article ">Figure 7
<p>Workflow of busbar disassembly using the proposed system. (<b>a</b>) Robot initialisation and programme execution; (<b>b</b>) virtual robot synchronisation and key information display; (<b>c</b>) real-time monitoring of robot status; (<b>d</b>) manual control intervention; (<b>e</b>) execution of planned motions.</p>
Full article ">Figure 8
<p>Busbar disassembly. (<b>a</b>) Busbar jammed on studs; (<b>b</b>) successful disassembly of the busbar.</p>
Full article ">Figure 9
<p>Three robot teleoperation methods. (<b>a</b>) Proposed system; (<b>b</b>) AR-based control; (<b>c</b>) smartpad control.</p>
Full article ">Figure 10
<p>Disassembly times for different control methods.</p>
Full article ">Figure 11
<p>Average NASA RTLX scores for different control methods.</p>
Full article ">Figure 12
<p>NASA RTLX indicator scores for different control methods.</p>
Full article ">
17 pages, 759 KiB  
Review
Blending Tradition and Innovation: Student Opinions on Modern Anatomy Education
by Alina Maria Șișu, Emil Robert Stoicescu, Sorin Lucian Bolintineanu, Alexandra Corina Faur, Roxana Iacob, Delius Mario Ghenciu, Alexandra-Ioana Dănilă and Ovidiu Alin Hațegan
Educ. Sci. 2024, 14(11), 1150; https://doi.org/10.3390/educsci14111150 - 24 Oct 2024
Viewed by 684
Abstract
Anatomy education has evolved significantly with the introduction of diverse instructional techniques. This review evaluates these methods, including traditional cadaver dissection, three-dimensional (3D) model printing, virtual dissection using tools like the Anatomage table, problem-based learning (PBL), and the use of wax and plastinated [...] Read more.
Anatomy education has evolved significantly with the introduction of diverse instructional techniques. This review evaluates these methods, including traditional cadaver dissection, three-dimensional (3D) model printing, virtual dissection using tools like the Anatomage table, problem-based learning (PBL), and the use of wax and plastinated models. Each approach presents unique benefits and challenges. Cadaver dissection remains invaluable for providing hands-on experience and a deep understanding of anatomical structures, although it faces ethical, logistical, and financial constraints. Wax and plastinated models offer durable, precise representations of anatomical structures without the ethical concerns associated with cadavers. Additionally, 3D printing and virtual dissection have emerged as effective supplementary tools, enhancing spatial understanding and allowing repeated practice. PBL integrates anatomical knowledge with clinical reasoning, promoting critical thinking and problem-solving skills. The main aim of this study was to gather and analyze students’ opinions on various anatomy teaching methods, while a secondary objective was to review the literature on novel and traditional approaches in anatomy education. This review emphasizes the importance of incorporating a variety of teaching methods to create a dynamic and engaging anatomy curriculum, preparing students for clinical practice. Full article
Show Figures

Figure 1

Figure 1
<p>PRISMA diagram—the methodology used for the review.</p>
Full article ">Figure 2
<p>Integration of diverse teaching methods in anatomy. Created using BioRender.</p>
Full article ">
23 pages, 10682 KiB  
Article
VFLD: Voxelized Fractal Local Descriptor
by Francisco Gomez-Donoso, Felix Escalona, Florian Dargère and Miguel Cazorla
Appl. Sci. 2024, 14(20), 9414; https://doi.org/10.3390/app14209414 - 15 Oct 2024
Viewed by 485
Abstract
A variety of methods for 3D object recognition and registration based on a deep learning pipeline have recently emerged. Nonetheless, these methods require large amounts of data that are not easy to obtain, sometimes rendering them virtually useless in real-life scenarios due to [...] Read more.
A variety of methods for 3D object recognition and registration based on a deep learning pipeline have recently emerged. Nonetheless, these methods require large amounts of data that are not easy to obtain, sometimes rendering them virtually useless in real-life scenarios due to a lack of generalization capabilities. To counter this, we propose a novel local descriptor that takes advantage of the fractal dimension. For each 3D point, we create a descriptor by computing the fractal dimension of the neighbors at different radii. Our redmethod has many benefits, such as being agnostic to the sensor of choice and noise, up to a level, and having few parameters to tinker with. Furthermore, it requires no training and does not rely on semantic information. We test our descriptor using well-known datasets and it largely outperforms Fast Point Feature Histogram, which is the state-of-the-art descriptor for 3D data. We also apply our descriptor to a registration pipeline and achieve accurate three-dimensional representations of the scenes, which are captured with a commercial sensor. Full article
(This article belongs to the Special Issue Current Advances in 3D Scene Classification and Object Recognition)
Show Figures

Figure 1

Figure 1
<p>Visualization of the occupied boxes (blue) of a point cloud after applying the voxel grid. In black, we can see the main bounding box of the object that marks the size for the divisions.</p>
Full article ">Figure 2
<p>Effects of the <math display="inline"><semantics> <mrow> <mi>n</mi> <mi>I</mi> <mi>t</mi> <mi>e</mi> <mi>r</mi> <mi>s</mi> </mrow> </semantics></math> parameter in the box-counting process from the original point cloud (leftmost) to the generated grid with 100 iterations (rightmost).</p>
Full article ">Figure 3
<p>Plot of the computed fractal dimension for <math display="inline"><semantics> <mrow> <mi>n</mi> <mi>I</mi> <mi>t</mi> <mi>e</mi> <mi>r</mi> <mi>s</mi> <mo>=</mo> <mo>{</mo> <mn>3</mn> <mo>,</mo> <mn>5</mn> <mo>,</mo> <mn>7</mn> <mo>,</mo> <mn>15</mn> <mo>}</mo> </mrow> </semantics></math> for a random set of points. This is a log–log plot in which the X-axis refers to the inverse of the voxel size in the box-counting method, while the Y-axis is the number of occupied boxes. Note the difference in the slope (FD, fractal dimension) of the fitted line when <math display="inline"><semantics> <mrow> <mi>n</mi> <mi>I</mi> <mi>t</mi> <mi>e</mi> <mi>r</mi> <mi>s</mi> </mrow> </semantics></math> is set too high.</p>
Full article ">Figure 4
<p>Diagram of the VLFD generation process.</p>
Full article ">Figure 5
<p>Visualization of the steps that comprise the computation of the descriptor. (<b>a</b>) The surrounding points at different radii are obtained. Two radii are used in this example for visualization purposes. (<b>b</b>) Box-counting is used to obtain the leaf size and the occupied boxes on each subset. Four iterations are visualized. (<b>c</b>) The log–log curve is generated for the data obtained, and a line is fitted. Its slope is the fractal dimension.</p>
Full article ">Figure 6
<p>Random samples of the ModelNet10 dataset.</p>
Full article ">Figure 7
<p>Random samples of the Simple Figures dataset.</p>
Full article ">Figure 8
<p>Examples of ScanNet RGB-D scenes, viewed from above.</p>
Full article ">Figure 9
<p>Random examples of point clouds included in the ViDRILO dataset.</p>
Full article ">Figure 10
<p>Accuracy (top) and precision–recall (bottom) curves for different starting values for the search radii for ModelNet (leftmost) and Simple Figures (rightmost) datasets.</p>
Full article ">Figure 11
<p>Accuracy (top) and precision–recall (bottom) curves for different increment values for the search radii for ModelNet (leftmost) and Simple Figures (rightmost) datasets.</p>
Full article ">Figure 12
<p>Accuracy (top) and precision–recall (bottom) curves for different box-counting iterations for ModelNet (leftmost) and Simple Figures (rightmost) datasets.</p>
Full article ">Figure 13
<p>Accuracy (top) and precision–recall (bottom) curves for different amounts of the search radii for ModelNet (leftmost) and Simple Figures (rightmost) datasets.</p>
Full article ">Figure 14
<p>Accuracy (top) and precision–recall (bottom) curves for different densities of the sampling process for ModelNet (leftmost) and Simple Figures (rightmost) datasets.</p>
Full article ">Figure 15
<p>Result of applying different Gaussian noise levels <math display="inline"><semantics> <mrow> <mi>F</mi> <mo>=</mo> <mo>(</mo> <mn>0</mn> <mo>,</mo> <mn>0.5</mn> <mo>,</mo> <mn>1</mn> <mo>,</mo> <mn>3</mn> <mo>)</mo> </mrow> </semantics></math> to a random sample of the ModelNet10 dataset.</p>
Full article ">Figure 16
<p>Accuracy (top) and precision–recall (bottom) curves for different noise levels added to ModelNet (leftmost) and Simple Figures (rightmost) datasets.</p>
Full article ">Figure 17
<p>Accuracy (top) and precision–recall (bottom) curves for different state-of-the-art methods for ModelNet (leftmost) and Simple Figures (rightmost) datasets. (<b>a</b>) ModelNet accuracy; (<b>b</b>) Simple Figures accuracy; (<b>c</b>) ModelNet precision–recall; (<b>d</b>) Simple figures precision–recall.</p>
Full article ">Figure 18
<p>ScanNet scenes splitting steps. (<b>a</b>) The initial scene is downsampled into a point cloud of 12,000 points. (<b>b</b>) The 2D minimum area rectangle including all the points is obtained. (<b>c</b>) This two-dimensional rectangle is divided into four equal parts. (<b>d</b>) Three more clouds are created by iterating clockwise from the leftmost cloud.</p>
Full article ">Figure 19
<p>Two types of downsampling used for the evaluation protocol. (<b>a</b>) A ScanNet scene, obtained with uniform downsampling. (<b>b</b>) A ScanNet scene, obtained with voxelized downsampling.</p>
Full article ">Figure 20
<p>Amount of scenes (in ordinate) for each error rate interval (in abscissa) of VFLD (in blue) and FPFH (in red) for the registration evaluation protocol on the uniform environment.</p>
Full article ">Figure 21
<p>Number of scenes (in ordinate) for each error rate interval (in abscissa) of VFLD (in blue) and FPFH (in red) for the registration evaluation protocol on the voxelized environment.</p>
Full article ">Figure 22
<p>Registration results of two different environments using VFLD as the descriptor of choice for the feature-matching step. Four different clouds are shown for each example; each one is of a different color. (<b>a</b>–<b>d</b>,<b>h</b>–<b>k</b>) are color images from a sequences of the dataset, and (<b>e</b>–<b>g</b>,<b>l</b>–<b>n</b>) are the tridimensional reconstruction of the scene achieved using the proposed VFLD.</p>
Full article ">
16 pages, 653 KiB  
Article
Appraising Education 4.0 in Nigeria’s Higher Education Institutions: A Case Study of Built Environment Programmes
by Andrew Ebekozien, Mohamed Ahmed Hafez, Clinton Aigbavboa, Mohamad Shaharudin Samsurijan, Abubakar Zakariyya Al-Hasan and Angeline Ngozika Chibuike Nwaole
Sustainability 2024, 16(20), 8878; https://doi.org/10.3390/su16208878 - 14 Oct 2024
Viewed by 713
Abstract
In the era of digitalisation, the construction industry is fast embracing digital technology, which evolved from Industry 4.0 (fourth industrial revolution [4IR]). Built environment programmes (BEPs) are expected to meet the needs of the digitalisation era via Education 4.0. Education 4.0 aims to [...] Read more.
In the era of digitalisation, the construction industry is fast embracing digital technology, which evolved from Industry 4.0 (fourth industrial revolution [4IR]). Built environment programmes (BEPs) are expected to meet the needs of the digitalisation era via Education 4.0. Education 4.0 aims to transform education’s future with diverse digital automation and innovative paedagogical procedures. Studies concerning Education 4.0 in Nigeria’s BEPs are scarce. Hence, this study aims to appraise Education 4.0 and investigate the perceived issues facing implementing Education 4.0 in Higher Education Institutions (HEIs), using BEPs as a case study. The findings intend to improve Education 4.0 implementation practices in BEPs. Data were sourced from 40 participants across Nigeria for better coverage and representation via a semi-structured interview approach. The participants were knowledgeable about Education 4.0 and Nigeria’s BEPs. This study adopted a thematic analysis of the virtually collected data and presented the findings in themes. This study shows that Education 4.0 implementation in Nigeria’s BEPs is lax and should be overhauled to improve achieving Sustainable Development Goal 4 (SDG 4)—quality education—and other related SDGs. The findings reveal that improved Education 4.0 can enhance the achievement of SDG 4. The findings cluster the perceived 18 hindrances facing Education 4.0 implementation into three main groups. Also, the findings proffer feasible measures to improve Education 4.0 implementation in Nigeria’s HEIs, using BEPs as a case study, via improved transformative competencies, technological advancement, innovative paedagogical procedures, and stakeholders’ collaboration to improve achieving SDG 4. The proposed framework could assist in creating new values and transforming the students’ BEP competencies via stakeholder collaboration and Education 4.0 for the private sector (future talents’ beneficiary), thus fostering their employability. Full article
Show Figures

Figure 1

Figure 1
<p>Proposed framework to improve Education 4.0 implementation in Nigeria’s BEPs. <b>Source:</b> authors’ work.</p>
Full article ">Figure 2
<p>Developed framework to improve Education 4.0 implementation in Nigeria’s BEPs. <b>Source:</b> authors’ work.</p>
Full article ">
21 pages, 1550 KiB  
Article
Using 3D Hand Pose Data in Recognizing Human–Object Interaction and User Identification for Extended Reality Systems
by Danish Hamid, Muhammad Ehatisham Ul Haq, Amanullah Yasin, Fiza Murtaza and Muhammad Awais Azam
Information 2024, 15(10), 629; https://doi.org/10.3390/info15100629 - 12 Oct 2024
Viewed by 714
Abstract
Object detection and action/gesture recognition have become imperative in security and surveillance fields, finding extensive applications in everyday life. Advancement in such technologies will help in furthering cybersecurity and extended reality systems through the accurate identification of users and their interactions, which plays [...] Read more.
Object detection and action/gesture recognition have become imperative in security and surveillance fields, finding extensive applications in everyday life. Advancement in such technologies will help in furthering cybersecurity and extended reality systems through the accurate identification of users and their interactions, which plays a pivotal role in the security management of an entity and providing an immersive experience. Essentially, it enables the identification of human–object interaction to track actions and behaviors along with user identification. Yet, it is performed by traditional camera-based methods with high difficulties and challenges since occlusion, different camera viewpoints, and background noise lead to significant appearance variation. Deep learning techniques also demand large and labeled datasets and a large amount of computational power. In this paper, a novel approach to the recognition of human–object interactions and the identification of interacting users is proposed, based on three-dimensional hand pose data from an egocentric camera view. A multistage approach that integrates object detection with interaction recognition and user identification using the data from hand joints and vertices is proposed. Our approach uses a statistical attribute-based model for feature extraction and representation. The proposed technique is tested on the HOI4D dataset using the XGBoost classifier, achieving an average F1-score of 81% for human–object interaction and an average F1-score of 80% for user identification, hence proving to be effective. This technique is mostly targeted for extended reality systems, as proper interaction recognition and users identification are the keys to keeping systems secure and personalized. Its relevance extends into cybersecurity, augmented reality, virtual reality, and human–robot interactions, offering a potent solution for security enhancement along with enhancing interactivity in such systems. Full article
(This article belongs to the Special Issue Extended Reality and Cybersecurity)
Show Figures

Figure 1

Figure 1
<p>Multi-stage HOI recognition.</p>
Full article ">Figure 2
<p>Representation of set of 21 3D hand Landmarks and vertices.</p>
Full article ">Figure 3
<p>Confusion matrix for object recognition (hand joints).</p>
Full article ">Figure 4
<p>Confusion matrix for object recognition (aand vertices).</p>
Full article ">Figure 5
<p>Confusion matrix for object recognition (fusion concatenation).</p>
Full article ">Figure 6
<p>Object based F1-Score for interaction classification.</p>
Full article ">Figure 7
<p>User identification average F1-Score in object-wise interactions.</p>
Full article ">
19 pages, 2283 KiB  
Article
Affective Experiences of U.S. School Personnel in the Sociopolitical Context of 2021: Reflecting on the Past to Shape the Future
by Miranda Wood, Cheyeon Ha, Marc Brackett and Christina Cipriano
Educ. Sci. 2024, 14(10), 1093; https://doi.org/10.3390/educsci14101093 - 8 Oct 2024
Viewed by 554
Abstract
This study aims to explore the self-reported affective experiences of U.S. school personnel during 2021. This year found school personnel balancing remote learning, health emergencies, a global pandemic, and intense political upheaval. We contextualize school personnel experiences within the current sociopolitical context. In [...] Read more.
This study aims to explore the self-reported affective experiences of U.S. school personnel during 2021. This year found school personnel balancing remote learning, health emergencies, a global pandemic, and intense political upheaval. We contextualize school personnel experiences within the current sociopolitical context. In this sample, school personnel (n = 8052) represent all U.S. states and territories alongside representation of diverse racial and ethnic identities (n = 1901). Participants were surveyed before completing a free virtual course on emotion management. The survey included open-ended questions and scale items. Participants reported primary feelings, sources of stress and joy, and perceptions of personal and social and emotional support for themselves and students. Findings are presented in five cohorts of school personnel across the year. The primary feelings were being anxious, stressed, and overwhelmed, the stressors were lack of support, time, and resources, as well as COVID-19, and workload, and the sources of joy were students, coworkers, and teaching. Anxiety and gratitude decreased throughout the year while happiness increased. Responses differed across time and between racial groups, with Black and African American participants reporting the highest percentages of being stressed by COVID-19 and community fluctuated over time as a source of joy. Implications for the education system and opportunities for emotion management are discussed. Full article
Show Figures

Figure 1

Figure 1
<p>Prosocial Classroom Model.</p>
Full article ">Figure 2
<p>Affective experiences of school personnel across time.</p>
Full article ">Figure 3
<p>Sources of stress for school personnel across time.</p>
Full article ">Figure 4
<p>School personnel sources of joy across time.</p>
Full article ">
17 pages, 4904 KiB  
Article
Development of a Digital Twin Driven by a Deep Learning Model for Fault Diagnosis of Electro-Hydrostatic Actuators
by Roman Rodriguez-Aguilar, Jose-Antonio Marmolejo-Saucedo and Utku Köse
Mathematics 2024, 12(19), 3124; https://doi.org/10.3390/math12193124 - 6 Oct 2024
Viewed by 647
Abstract
The first quarter of the 21st century has witnessed many technological innovations in various sectors. Likewise, the COVID-19 pandemic triggered the acceleration of digital transformation in organizations driven by artificial intelligence and communication technologies in Industry 4.0 and Industry 5.0. Aiming at the [...] Read more.
The first quarter of the 21st century has witnessed many technological innovations in various sectors. Likewise, the COVID-19 pandemic triggered the acceleration of digital transformation in organizations driven by artificial intelligence and communication technologies in Industry 4.0 and Industry 5.0. Aiming at the construction of digital twins, virtual representations of a physical system allow real-time bidirectional communication. This will allow the monitoring of operations, identification of possible failures, and decision making based on technical evidence. In this study, a fault diagnosis solution is proposed, based on the construction of a digital twin, for a cloud-based Industrial Internet of Things (IIoT) system contemplating the control of electro-hydrostatic actuators (EHAs). The system was supported by a deep learning model using Long Short-Term Memory (LSTM) networks for an effective diagnostic approach. The implemented study considers data preparation and integration and system development and application to evaluate the performance against the fault diagnosis problem. According to the results obtained, positive results are shown in the construction of the digital twin using a deep learning model for the fault diagnosis problem of an active EHA-IIoT configuration. Full article
Show Figures

Figure 1

Figure 1
<p>A general scheme regarding the solution implemented.</p>
Full article ">Figure 2
<p>Architecture of Long-Short Term Memory (LSTM) [<a href="#B74-mathematics-12-03124" class="html-bibr">74</a>].</p>
Full article ">Figure 3
<p>A general scheme for the control of an electro-hydrostatic actuator (EHA) [<a href="#B29-mathematics-12-03124" class="html-bibr">29</a>].</p>
Full article ">Figure 4
<p>The general system structure: (<b>a</b>) IIoT approach, (<b>b</b>) LSTM localization, (<b>c</b>) digital twin user interface, (<b>d</b>) sample state graphics seen in real time for <math display="inline"><semantics> <mrow> <mi>I</mi> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </semantics></math> (green: normal state, blue; minor fault state, orange: medium fault state, red: mjor fault state), (<b>e</b>) optimization module, and (<b>f</b>) simulation module (AnyLogic<sup>®</sup>).</p>
Full article ">Figure 5
<p>Fault state values by training–testing data sets.</p>
Full article ">Figure 6
<p>LSTM model results (Data Set A). The green color corresponds to a correct classification and the red color to a classification in the incorrect category.</p>
Full article ">Figure 7
<p>LSTM model results (Data Set B). The green color corresponds to a correct classification and the red color to a classification in the incorrect category.</p>
Full article ">Figure 8
<p>LSTM model results (Data Set C). The green color corresponds to a correct classification and the red color to a classification in the incorrect category.</p>
Full article ">Figure 9
<p>LSTM model results (Data Set D). The green color corresponds to a correct classification and the red color to a classification in the incorrect category.</p>
Full article ">Figure 10
<p>LSTM model results (Data Set E). The green color corresponds to a correct classification and the red color to a classification in the incorrect category.</p>
Full article ">
13 pages, 358 KiB  
Article
Using a Multivariate Virtual Experiment for Uncertainty Evaluation with Unknown Variance
by Manuel Marschall, Finn Hughes, Gerd Wübbeler, Gertjan Kok, Marcel van Dijk and Clemens Elster
Metrology 2024, 4(4), 534-546; https://doi.org/10.3390/metrology4040033 - 1 Oct 2024
Viewed by 622
Abstract
Virtual experiments are a digital representation of a real measurement and play a crucial role in modern measurement sciences and metrology. Beyond their common usage as a modeling and validation tool, a virtual experiment may also be employed to perform a parameter sensitivity [...] Read more.
Virtual experiments are a digital representation of a real measurement and play a crucial role in modern measurement sciences and metrology. Beyond their common usage as a modeling and validation tool, a virtual experiment may also be employed to perform a parameter sensitivity analysis or to carry out a measurement uncertainty evaluation. For the latter to be compliant with statistical principles and metrological guidelines, the procedure to obtain an estimate and a corresponding measurement uncertainty requires careful consideration. We employ a Monte Carlo sampling procedure using a virtual experiment that allows one to perform a measurement uncertainty evaluation according to the Monte Carlo approach of JCGM-101 and JCGM-102, two widely applied guidelines for uncertainty evaluation in metrology. We extend and formalize a previously published approach for simple additive models to account for a large class of non-linear virtual experiments and measurement models for multidimensionality of the data and output quantities, and for the case of unknown variance of repeated measurements. With the algorithm developed here, a simple procedure for the evaluation of measurement uncertainty is provided that may be applied in various applications that admit a certain structure for their virtual experiment. Moreover, the measurement model commonly employed for uncertainty evaluation according to JCGM-101 and JCGM-102 is not required for this algorithm, and only evaluations of the virtual experiment are performed to obtain an estimate and an associated uncertainty of the measurand. We demonstrate the efficacy of the developed approach and the effect of the underlying assumptions for a generic polynomial regression example and an example of a simplified coordinate measuring machine and its virtual representation. The results of this work highlight that considerable effort, diligence, and statistical considerations need to be invested to make use of a virtual experiment for uncertainty evaluation in a way that ensures equivalence with the accepted guidelines. Full article
(This article belongs to the Collection Measurement Uncertainty)
Show Figures

Figure 1

Figure 1
<p>Experimental setting for the generic polynomial regression example of <a href="#sec4dot1-metrology-04-00033" class="html-sec">Section 4.1</a>. The dotted green line indicates the unknown polynomial, from which noisy observations (black crosses) can be conducted at the nominal measurement locations <math display="inline"><semantics> <mrow> <mo>{</mo> <mn>1</mn> <mo>,</mo> <mo>…</mo> <mo>,</mo> <mn>10</mn> <mo>}</mo> </mrow> </semantics></math>. The resulting (mean) estimates of the virtual experiment approach (solid red line) and the JCGM-102 Monte Carlo approach (dashed blue line) are shown.</p>
Full article ">Figure 2
<p>Marginal densities of the measurand, i.e., polynomial coefficients, of the generic polynomial regression example are shown. The solid red line represents the result of the application of Algorithm 2 using only the virtual experiment and its gradient. The dashed blue line is the result of the JCGM-102 Monte Carlo approach and Algorithm 1 using the corresponding measurement model (<a href="#FD13-metrology-04-00033" class="html-disp-formula">13</a>).</p>
Full article ">Figure 3
<p>Marginal densities of the measurand, in terms of circle radius and position, of the coordinate measuring machine example are shown. The solid red line represents the result of applying Algorithm 2 using only the virtual experiment and its gradient. The dashed blue line is the result of the JCGM-102 Monte Carlo approach and Algorithm 1 using the corresponding measurement model (<a href="#FD15-metrology-04-00033" class="html-disp-formula">15</a>), and the dotted green line corresponds to the JCGM-102 Monte Carlo approach and Algorithm 1 applied to the alternative measurement model (<a href="#FD19-metrology-04-00033" class="html-disp-formula">19</a>).</p>
Full article ">
Back to TopTop