[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (4)

Search Parameters:
Keywords = tennis movement recognition

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
14 pages, 1396 KiB  
Article
Can Perceivers Differentiate Intense Facial Expressions? Eye Movement Patterns
by Leyu Huang, Tongtong Zhu, Jiaotao Cai, Yan Sun and Yanmei Wang
Behav. Sci. 2024, 14(3), 185; https://doi.org/10.3390/bs14030185 - 26 Feb 2024
Viewed by 1200
Abstract
Recent research on intense real-life faces has shown that although there was an objective difference in facial activities between intense winning faces and losing faces, viewers failed to differentiate the valence of such expressions. In the present study, we explored whether participants could [...] Read more.
Recent research on intense real-life faces has shown that although there was an objective difference in facial activities between intense winning faces and losing faces, viewers failed to differentiate the valence of such expressions. In the present study, we explored whether participants could perceive the difference between intense positive facial expressions and intense negative facial expressions in a forced-choice response task using eye-tracking techniques. Behavioral results showed that the recognition accuracy rate for intense facial expressions was significantly above the chance level. For eye-movement patterns, the results indicated that participants gazed more and longer toward the upper facial region (eyes) than the lower region (mouth) for intense losing faces. However, the gaze patterns were reversed for intense winning faces. The eye movement pattern for successful differentiation trials did not differ from failed differentiation trials. These findings provided preliminary evidence that viewers can utilize intense facial expression information and perceive the difference between intense winning faces and intense losing faces produced by tennis players in a forced-choice response task. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>A</b>) An example of an experimental trial in a forced-choice response task. (<b>B</b>) Mean reaction times for valence judgment for losing task (black bar) and winning task (gray bar). (<b>C</b>) Mean accuracy of valence judgment for losing task (black bar) and winning task (gray bar). Error bars indicate standard errors of the mean.</p>
Full article ">Figure 2
<p>Bar graphs of simple main effect analyses of face type (winning faces vs. losing faces) and task type (choosing the winning face vs. choosing the losing face) on (<b>A</b>) average fixation counts for the eye region and the mouth region of intense faces; (<b>B</b>) average fixation durations for the eye region and the mouth region of intense faces; (<b>C</b>) average fixation counts for the eye region and the mouth region of moderate faces; (<b>D</b>) average fixation durations for the eye region and the mouth region of moderate faces. Intense eye region = the eye region of intense faces; intense mouth region = the mouth region of intense faces; moderate eye region = the eye region of moderate faces; moderate mouth region = the mouth region of moderate faces; win task = choosing the winning face from a pair of faces with different valence; lose task = choosing the losing face from a pair of faces with different valence; win face = the gaze behavior toward the winning face; lose face = the gaze behavior toward the losing face. Error bars indicate standard errors of the mean.</p>
Full article ">
16 pages, 2895 KiB  
Article
Temporal Pattern Attention for Multivariate Time Series of Tennis Strokes Classification
by Maria Skublewska-Paszkowska and Pawel Powroznik
Sensors 2023, 23(5), 2422; https://doi.org/10.3390/s23052422 - 22 Feb 2023
Cited by 6 | Viewed by 2060
Abstract
Human Action Recognition is a challenging task used in many applications. It interacts with many aspects of Computer Vision, Machine Learning, Deep Learning and Image Processing in order to understand human behaviours as well as identify them. It makes a significant contribution to [...] Read more.
Human Action Recognition is a challenging task used in many applications. It interacts with many aspects of Computer Vision, Machine Learning, Deep Learning and Image Processing in order to understand human behaviours as well as identify them. It makes a significant contribution to sport analysis, by indicating players’ performance level and training evaluation. The main purpose of this study is to investigate how the content of three-dimensional data influences on classification accuracy of four basic tennis strokes: forehand, backhand, volley forehand, and volley backhand. An entire player’s silhouette and its combination with a tennis racket were taken into consideration as input to the classifier. Three-dimensional data were recorded using the motion capture system (Vicon Oxford, UK). The Plug-in Gait model consisting of 39 retro-reflective markers was used for the player’s body acquisition. A seven-marker model was created for tennis racket capturing. The racket is represented in the form of a rigid body; therefore, all points associated with it changed their coordinates simultaneously. The Attention Temporal Graph Convolutional Network was applied for these sophisticated data. The highest accuracy, up to 93%, was achieved for the data of the whole player’s silhouette together with a tennis racket. The obtained results indicated that for dynamic movements, such as tennis strokes, it is necessary to analyze the position of the whole body of the player as well as the racket position. Full article
Show Figures

Figure 1

Figure 1
<p>Motion capture cameras arrangement, where <math display="inline"><semantics> <mi>α</mi> </semantics></math> is the angle in the <math display="inline"><semantics> <mrow> <mi>O</mi> <mi>X</mi> </mrow> </semantics></math> plane between the floor and the camera axis, <math display="inline"><semantics> <mi>β</mi> </semantics></math> is the angle in the <math display="inline"><semantics> <mrow> <mi>O</mi> <mi>Y</mi> </mrow> </semantics></math> plane between the camera axis perpendicular to the floor and the camera.</p>
Full article ">Figure 2
<p>An example of forehand and backhand strokes (<b>a</b>) forehand preparation phase (<b>b</b>) forehand shot (<b>c</b>) no shot (<b>d</b>) backhand preparation phase (<b>e</b>) backhand shot (<b>f</b>) no shot.</p>
Full article ">Figure 3
<p>Scheme of the used BiRNN network.</p>
Full article ">Figure 4
<p>Classification model for tennis data movements.</p>
Full article ">Figure 5
<p>Selected learning parameters. (<b>a</b>) Learning accuracy for input without the tennis racket. (<b>b</b>) Learning accuracy for input with the tennis racket. (<b>c</b>) Loss value for input without the tennis racket. (<b>d</b>) Loss value for input with the tennis racket.</p>
Full article ">Figure 6
<p>Confusion matrices (in %). (<b>a</b>) Matrix for input without the tennis racket. (<b>b</b>) Matrix for input with the tennis racket.</p>
Full article ">
17 pages, 1246 KiB  
Article
An Enhanced Joint Hilbert Embedding-Based Metric to Support Mocap Data Classification with Preserved Interpretability
by Cristian Kaori Valencia-Marin, Juan Diego Pulgarin-Giraldo, Luisa Fernanda Velasquez-Martinez, Andres Marino Alvarez-Meza and German Castellanos-Dominguez
Sensors 2021, 21(13), 4443; https://doi.org/10.3390/s21134443 - 29 Jun 2021
Cited by 7 | Viewed by 2793
Abstract
Motion capture (Mocap) data are widely used as time series to study human movement. Indeed, animation movies, video games, and biomechanical systems for rehabilitation are significant applications related to Mocap data. However, classifying multi-channel time series from Mocap requires coding the intrinsic dependencies [...] Read more.
Motion capture (Mocap) data are widely used as time series to study human movement. Indeed, animation movies, video games, and biomechanical systems for rehabilitation are significant applications related to Mocap data. However, classifying multi-channel time series from Mocap requires coding the intrinsic dependencies (even nonlinear relationships) between human body joints. Furthermore, the same human action may have variations because the individual alters their movement and therefore the inter/intraclass variability. Here, we introduce an enhanced Hilbert embedding-based approach from a cross-covariance operator, termed EHECCO, to map the input Mocap time series to a tensor space built from both 3D skeletal joints and a principal component analysis-based projection. Obtained results demonstrate how EHECCO represents and discriminates joint probability distributions as kernel-based evaluation of input time series within a tensor reproducing kernel Hilbert space (RKHS). Our approach achieves competitive classification results for style/subject and action recognition tasks on well-known publicly available databases. Moreover, EHECCO favors the interpretation of relevant anthropometric variables correlated with players’ expertise and acted movement on a Tennis-Mocap database (also publicly available with this work). Thereby, our EHECCO-based framework provides a unified representation (through the tensor RKHS) of the Mocap time series to compute linear correlations between a coded metric from joint distributions and player properties, i.e., age, body measurements, and sport movement (action class). Full article
(This article belongs to the Special Issue Sensors and Musculoskeletal Dynamics to Evaluate Human Movement)
Show Figures

Figure 1

Figure 1
<p>Schematic illustration of our EHECCO-based metric. Input spaces <math display="inline"><semantics> <mi mathvariant="script">X</mi> </semantics></math> and <math display="inline"><semantics> <mi mathvariant="script">Y</mi> </semantics></math> are mapped to RKHSs <math display="inline"><semantics> <mi mathvariant="script">H</mi> </semantics></math> and <math display="inline"><semantics> <mrow> <mi mathvariant="script">G</mi> <mo>,</mo> </mrow> </semantics></math> respectively. Then, the tensor space <math display="inline"><semantics> <mrow> <mi mathvariant="script">H</mi> <mo>⊗</mo> <mi mathvariant="script">G</mi> </mrow> </semantics></math> is built using a cross-covariance operator strategy.</p>
Full article ">Figure 2
<p>EHECCO-based Mocap data classification framework. Hip joint normalization and spectral clustering-based codebook generation are carried out to extract relevant skeletal poses. Then, 3D joint representation (<math display="inline"><semantics> <mi mathvariant="script">X</mi> </semantics></math>) and PCA-based latent projection ( <math display="inline"><semantics> <mi mathvariant="script">Y</mi> </semantics></math>) are used to support the EHECCO metric from joint probability. Lastly, an SVM classifier is trained from the EHECCO distance that also supports 2D data visualization.</p>
Full article ">Figure 3
<p>Illustrative results for codebook generation and latent space-based representation (HDM05 and CMU subset datasets). Top: Codebook generation for a Mocap video of the <span class="html-italic">throwing high with the right hand while standing</span> class (HDM05). Middle: Codebook generation for a Mocap record of <span class="html-italic">boxing</span> class (CMU subset). Bottom left: PCA-based latent space for HDM05 video. Bottom right: PCA-based latent space for CMU subset video. The first two components are shown for visualization purposes. Black markers represent the original input Mocap frames (time series). Color markers represent the chosen frames (codebook).</p>
Full article ">Figure 4
<p>EHECCO-based classification results for HDM05 and CMU subset databases. <b>Top left</b>: HDM05’s confusion matrix (style/subject recognition). <b>Top right</b>: HDM05 t-SNE-based 2D projection from EHECCO distance. <b>Bottom left</b>: CMU subset’s confusion matrix (action recognition). <b>Bottom right</b>: CMU subset t-SNE-based 2D projection from EHECCO distance.</p>
Full article ">Figure 5
<p>Illustrative results for codebook generation (Tennis-Mocap dataset). Top: <span class="html-italic">forehand</span>; Middle: <span class="html-italic">volley</span>; Bottom: <span class="html-italic">Smash</span>.</p>
Full article ">Figure 6
<p>EHECCO-based classification and anthropomorphic measurement results for Tennis-Mocap database. <b>Top left</b>: confusion matrix (action recognition). <b>Top right</b>: t-SNE-based 2D projection from EHECCO distance. <b>Bottom left</b>: Absolute value of the Pearson’s correlation coefficient between the EHECCO first t-SNE-based mean projection of each player’s videos and his/her anthropomorphic measurements. The most relevant correlations are shown.</p>
Full article ">
12 pages, 1476 KiB  
Article
Learning Three Dimensional Tennis Shots Using Graph Convolutional Networks
by Maria Skublewska-Paszkowska, Pawel Powroznik and Edyta Lukasik
Sensors 2020, 20(21), 6094; https://doi.org/10.3390/s20216094 - 27 Oct 2020
Cited by 18 | Viewed by 3899
Abstract
Human movement analysis is very often applied to sport, which has seen great achievements in assessing an athlete’s progress, giving further training tips and in movement recognition. In tennis, there are two basic shots: forehand and backhand, which are performed during all matches [...] Read more.
Human movement analysis is very often applied to sport, which has seen great achievements in assessing an athlete’s progress, giving further training tips and in movement recognition. In tennis, there are two basic shots: forehand and backhand, which are performed during all matches and training sessions. Recognition of these movements is important in the quantitative analysis of a tennis game. In this paper, the authors propose using Spatial-Temporal Graph Neural Networks (ST-GCN) to challenge the above task. Recognition of the shots is performed on the basis of images obtained from 3D tennis movements (forehands and backhands) recorded by the Vicon motion capture system (Oxford Metrics Ltd, Oxford, UK), where both the player and the racket were recorded. Two methods of putting data into the ST-GCN network were compared: with and without fuzzying of data. The obtained results confirm that the use of fuzzy input graphs for ST-GCNs is a better tool for recognition of forehand and backhand tennis shots relative to graphs without fuzzy input. Full article
(This article belongs to the Collection Sensor Technology for Sports Science)
Show Figures

Figure 1

Figure 1
<p>A spatial temporal graph of skeleton. Red dots represent joints and other characteristic points. Red lines represent connections between points within the lower limbs, blue—upper limbs, green—spine, yellow—tennis racket.</p>
Full article ">Figure 2
<p>An example of raw forehand shot phases. (<b>a</b>) beginning of the preparation phase, (<b>b</b>) end of the preparation phase (<b>c</b>) hitting the ball and (<b>d</b>) swinging the racket after the hit.</p>
Full article ">Figure 3
<p>Scheme of used classifier consisting of the ST-GCN part and the active features knowledge base.</p>
Full article ">Figure 4
<p>Efficiency plot of ST-GCN classifier.</p>
Full article ">Figure 5
<p>Efficiency of Fuzzy ST-GCN classifier.</p>
Full article ">
Back to TopTop