CN106778639A - A kind of exercise data search method based on the description of attitude relative space-time characteristic statisticses - Google Patents
A kind of exercise data search method based on the description of attitude relative space-time characteristic statisticses Download PDFInfo
- Publication number
- CN106778639A CN106778639A CN201611204708.4A CN201611204708A CN106778639A CN 106778639 A CN106778639 A CN 106778639A CN 201611204708 A CN201611204708 A CN 201611204708A CN 106778639 A CN106778639 A CN 106778639A
- Authority
- CN
- China
- Prior art keywords
- action
- feature
- relative
- retrieval
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/23—Recognition of whole body movements, e.g. for sport training
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/42—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- Processing Or Creating Images (AREA)
Abstract
本发明公开了一种基于姿态相对时空特征统计描述的运动数据检索方法,分以下下步骤:三维姿态相对时空特征提取,提取三维姿态中关节形成的点、线、面几何元素之间的相对空间位置及其变化的度量作为姿态的特征表示;动作统计描述生成,提取姿态相对时空特征的统计描述作为动作的特征向量表示,选择的统计量包括,均值、极差、方差以及偏态共4个统计变量;相似度匹配:采用欧式距离计算库中动作和检索动作的相似程度,对最终相似度由高到低排序后将候选结果动作返回;相关反馈,通过线性支持向量机(SVM‑Support Vector Machine)针对用户的反馈建立分类模型,从而优化特征子集及其权重组合,使得检索的性能达到最优。
The invention discloses a motion data retrieval method based on the statistical description of the relative spatio-temporal features of postures, which is divided into the following steps: extracting the relative spatio-temporal features of the three-dimensional postures, and extracting the relative space between the geometric elements of points, lines and planes formed by joints in the three-dimensional postures The measurement of the position and its change is used as the characteristic representation of the posture; the statistical description of the movement is generated, and the statistical description of the relative spatiotemporal characteristics of the posture is extracted as the characteristic vector representation of the movement. The selected statistics include mean, range, variance and skewness. Statistical variables; similarity matching: use the Euclidean distance to calculate the similarity between the actions in the library and the retrieval actions, sort the final similarity from high to low, and return the candidate result actions; related feedback, through the linear support vector machine (SVM‑Support Vector Machine) establishes a classification model based on user feedback, thereby optimizing feature subsets and their weight combinations, making retrieval performance optimal.
Description
技术领域technical field
本发明涉及人体运动数据检索方法,属于计算机三维动画技术及多媒体数据处理领域,具体地说,是一种基于姿态相对时空特征统计描述的人体运动数据检索方法。The invention relates to a human body motion data retrieval method, which belongs to the field of computer three-dimensional animation technology and multimedia data processing, and specifically relates to a human body motion data retrieval method based on statistical description of posture relative spatiotemporal features.
背景技术Background technique
由于各种应用的迫切需要以及商业捕获设备的广泛推广,目前已经出现了越来越多的大型三维人体动作库,如美国卡内基梅隆大学的人体动作库(http://mocap.cs.cmu.edu)等。如何从人体动作库中获取用户所需要的人体运动数据已经成为动作数据有效利用的关键问题,基于实例的检索技术采用特征表示运动的内容,通过特征之间的相似度度量实现内容相似运动的匹配,较好地克服了传统的基于文本标注检索方法的不足,已经成为运动捕获数据检索领域的研究热点。有效的特征表示及相应的相似度匹配机制是基于实例的检索的基本和关键问题。Due to the urgent needs of various applications and the widespread promotion of commercial capture equipment, there have been more and more large-scale 3D human motion libraries, such as the human motion library of Carnegie Mellon University (http://mocap.cs .cmu.edu), etc. How to obtain the human body motion data required by users from the human body motion library has become a key issue in the effective use of motion data. The instance-based retrieval technology uses features to represent the content of motion, and uses the similarity measurement between features to achieve the matching of motion with similar content. , overcomes the shortcomings of traditional text-based annotation retrieval methods, and has become a research hotspot in the field of motion capture data retrieval. Effective feature representation and corresponding similarity matching mechanism are the basic and key issues of instance-based retrieval.
特征要能够对运动的内容进行完整而有效的表示。在已有的姿态特征表示中,关节的三维坐标及其编码表示、四元数和欧拉角及其编码表示是常用的表示方法,然而研究表明它们只适用于数值相似的运动数据检索,并不能有效地表示由于风格、持续时间等差异形成的逻辑相似的运动。Müller等人在文献Müller M,et al.Efficient content basedretrieval of motion capture data.ACM Transactions on Graphics,2005,24(3):677-685)中提出布尔几何特征以实现逻辑相似的运动检索,然而有限的特征数量及二值布尔状态使得其对于细分的动作区分能力较差。潘红等人在文献潘红,肖俊,吴飞等.基于关键帧的三维人体运动检索[J],计算机辅助设计与图形学学报,2009,21(2):214-222[77]中提出以四肢骨骼和中心骨骼之间的夹角作为姿态特征,然而存在着众多和中心骨骼无关的局部动作。Tang等人在文献Tang J T,et al.Retrieval of logically relevant 3D humanmotions by Adaptive Feature Selection with Graded Relevance Feedback[J].Pattern Recognition Letters,2012,33(4):420-430中指出相对位置关系是提取逻辑含义的有效表示,并使用关节对之间的距离作为姿态的特征表示,然而其忽略了直线和平面这两个重要的几何元素,此外,该特征不能直接表达关节旋转内容。Chen等人在文献ChenC,et al.Perceptual 3D pose distance estimation by boosting relationalgeometric features[J].Computer Animation and Virtual Worlds,2009,20(2-3):267-277中定义了姿态相对几何特征集以评价感知上姿态的相似性,然而不同关节形成的点、线、面的任意组合,使得特征数量达到了56万多维,需要针对具体应用加以合理约束以降低特征维数。Features should be able to fully and effectively represent the content of the motion. In the existing gesture feature representations, the three-dimensional coordinates of joints and their coded representations, quaternions and Euler angles and their coded representations are commonly used representations, but studies have shown that they are only suitable for numerically similar motion data retrieval, and Cannot effectively represent logically similar movements due to differences in style, duration, etc. Müller et al proposed Boolean geometric features in the literature Müller M, et al. Efficient content based retrieval of motion capture data. ACM Transactions on Graphics, 2005, 24(3): 677-685) to achieve logically similar motion retrieval, but limited The number of features and the binary Boolean state make it less capable of distinguishing subdivided actions. Pan Hong et al. In the literature Pan Hong, Xiao Jun, Wu Fei et al. 3D Human Motion Retrieval Based on Key Frames [J], Journal of Computer-Aided Design and Graphics, 2009,21(2):214-222[77] It is proposed to use the angle between the limb bones and the central bone as the pose feature, but there are many local actions that have nothing to do with the central bone. Tang et al pointed out in the document Tang J T, et al. Retrieval of logically relevant 3D humanmotions by Adaptive Feature Selection with Graded Relevance Feedback [J]. Pattern Recognition Letters, 2012,33(4):420-430 that the relative positional relationship is the extraction It is an effective representation of logical meaning and uses the distance between joint pairs as the feature representation of the pose. However, it ignores the two important geometric elements of straight lines and planes. In addition, this feature cannot directly express the content of joint rotation. Chen et al. defined the pose relative geometric feature set in the document ChenC, et al. Perceptual 3D pose distance estimation by boosting relational geometric features [J]. Computer Animation and Virtual Worlds, 2009, 20(2-3): 267-277. Evaluate the similarity of posture in perception. However, any combination of points, lines, and surfaces formed by different joints makes the number of features reach more than 560,000 dimensions. Reasonable constraints need to be imposed on specific applications to reduce the number of feature dimensions.
姿态是运动的构成单元,一些研究者以姿态的特征表示为基础,通过降维以及聚类方法来获取姿态的低维表示;通过矩阵奇异值分解、球谐变换、张量代数子空间分解以及关键帧技术获取在更高层次上对运动片段的重新描述。然而在降维、聚类以及动作的二次特征提取距离计算中,反映运动内容的姿态局部特征均分配了特定的权重,这些局部特征的权重并不能够在检索的过程中随着检索实例的不同进行动态调整,导致了上述特征表示并不能有效地支持运动的局部空间检索。Attitude is the constituent unit of motion. Based on the feature representation of attitude, some researchers obtain the low-dimensional representation of attitude through dimensionality reduction and clustering methods; through matrix singular value decomposition, spherical harmonic transformation, tensor algebraic subspace decomposition and Keyframing techniques achieve a higher level re-description of motion clips. However, in dimensionality reduction, clustering, and calculation of the secondary feature extraction distance of actions, the local features of gestures that reflect the content of motion are all assigned specific weights, and the weights of these local features cannot be used in the retrieval process as the retrieval instance changes. Different dynamic adjustments lead to the fact that the above feature representations cannot effectively support the local space retrieval of motion.
综上所述,已有的运动内容特征表示尚不能够有效地支持运动数据检索,本发明提出了一种基于姿态相对时空特征统计描述的运动数据实例检索方法,在姿态时空相对特征内容表示的基础上,该方法进一步采用统计描述提取动作的二次特征向量,在检索的过程中,该方法还采用基于SVM相关反馈以提高检索性能。与本发明方法比较接近的方法是文献(陈松乐等.民族舞蹈运动数据的实例检索方法.计算机辅助设计与图形学学报,2014,26(2):198-210)和文献(Songle Chen,et al.Relevance Feedback for Human MotionRetrieval Using a Boosting Approach.Multimedia tools and applications,2016,75(2):787-817)中提出的方法,然而这两种方法均直接以姿态相对时空特征作为特征表示,并采用动态时间弯曲算法作为动作的相似度计算方法,计算复杂度太高,难以满足大规模运动捕获数据库检索的性能需求。To sum up, the existing feature representations of motion content cannot effectively support motion data retrieval. The present invention proposes a motion data instance retrieval method based on the statistical description of attitude relative spatiotemporal features. On the basis of this method, the method further adopts the secondary feature vector of the extracted action by statistical description. During the retrieval process, the method also adopts SVM-based correlation feedback to improve the retrieval performance. The method closer to the method of the present invention is literature (Chen Songle, etc. The example retrieval method of national dance data. Computer Aided Design and Graphics Journal, 2014, 26 (2): 198-210) and literature (Songle Chen, et al .Relevance Feedback for Human MotionRetrieval Using a Boosting Approach.Multimedia tools and applications, 2016,75(2):787-817), however, both methods directly use the attitude relative spatiotemporal features as feature representations, and use As a motion similarity calculation method, the dynamic time warping algorithm is too complex to meet the performance requirements of large-scale motion capture database retrieval.
发明内容Contents of the invention
发明目的:本发明所要解决的技术问题是针对现有技术的不足,提出了一种基于姿态相对时空特征统计描述的运动数据检索方法。Purpose of the invention: The technical problem to be solved by the present invention is to propose a motion data retrieval method based on the statistical description of the relative spatio-temporal features of postures in view of the deficiencies of the prior art.
技术方案:本发明公开的一种基于姿态相对时空特征统计描述的运动数据检索方法,包括以下步骤:Technical solution: The invention discloses a motion data retrieval method based on the statistical description of posture relative spatio-temporal features, comprising the following steps:
步骤1,三维姿态相对时空特征提取:三维姿态中关节形成的点、线、面几何元素集合是不同动作模式对应的局部区域的最小构成单元。本发明提取三维姿态中关节形成的点、线、面几何元素之间的相对空间位置及其变化的度量作为姿态的特征表示,通过不同局部区域包含的不同类型特征的权重组合,来表达广泛的姿态模式。对于库中所有动作包含的每个姿态,其相对时空特征提取具体步骤如下:Step 1, 3D pose relative spatio-temporal feature extraction: the set of point, line, and surface geometric elements formed by joints in 3D pose is the smallest constituent unit of the local area corresponding to different action modes. The present invention extracts the relative spatial position between the geometric elements of points, lines and surfaces formed by the joints in the three-dimensional posture and the measurement of their changes as the feature representation of the posture, and expresses a wide range of features through the weight combination of different types of features contained in different local areas. Attitude mode. For each gesture contained in all actions in the library, the specific steps of its relative spatiotemporal feature extraction are as follows:
1)定义三维人体关节模型,选择其中最重要的若干个关节作为三维姿态表示;1) Define the three-dimensional human joint model, and select the most important joints as the three-dimensional pose representation;
2)构建几何元素集合,选择的关节构成了几何元素集合中的点集,点集中任意2点形成直线,任意3点则构成平面;2) Construct a set of geometric elements. The selected joints constitute a point set in the set of geometric elements. Any 2 points in the point set form a straight line, and any 3 points form a plane;
3)提取每个三维姿态相对空间特征,包括关节对距离特征、关节与骨骼距离特征、关节与平面距离特征、骨骼对夹角特征、骨骼与平面夹角特征、平面与平面夹角特征、关节旋转特征;3) Extract the relative spatial features of each 3D pose, including joint pair distance feature, joint and bone distance feature, joint and plane distance feature, bone pair angle feature, bone and plane angle feature, plane and plane angle feature, joint rotation feature;
4)提取每个三维姿态相对时间特征,包括关节角速度与加速度特征。4) Extract relative time features of each three-dimensional attitude, including joint angular velocity and acceleration features.
步骤2,动作统计描述生成:动作由若干个姿态连续变化而成,本发明提取姿态相对时空特征的统计描述作为动作的特征向量表示,选择的统计量包括,均值、极差、方差以及偏态共4个统计变量。对于库中的每个动作,动作统计描述生成的具体步骤如下:Step 2, Action Statistical Description Generation: The action is formed by continuous changes of several postures. The present invention extracts the statistical description of the relative spatiotemporal characteristics of the posture as the feature vector representation of the action. The selected statistics include mean, range, variance and skewness There are 4 statistical variables in total. For each action in the library, the specific steps of action statistics description generation are as follows:
1)装载动作中包含的每个三维姿态的相对时空特征,并把每个特征作为一维随机变量,动作中的每个姿态对应了该随机变量的取值;1) Load the relative spatiotemporal features of each three-dimensional pose contained in the action, and use each feature as a one-dimensional random variable, and each pose in the action corresponds to the value of the random variable;
2)计算每个相对时空特征的均值;2) Calculate the mean of each relative spatiotemporal feature;
3)计算每个相对时空特征的极差;3) Calculate the range of each relative spatio-temporal feature;
4)计算每个相对时空特征的方差;4) Calculate the variance of each relative spatiotemporal feature;
5)计算每个相对时空特征的偏态;5) Calculate the skewness of each relative spatiotemporal feature;
6)将每个相对时空特征提取的均值、极差、方差、偏态归一化后按顺序进行排列并保存,形成每个动作的特征向量表示。6) The mean value, range, variance, and skewness of each relative spatiotemporal feature extraction are normalized and arranged in order and saved to form a feature vector representation of each action.
步骤3,相似度匹配:采用步骤1和步骤2提取用户提交的检索动作实例的特征描述,并进而采用欧式距离计算库中动作和检索动作的相似程度,对最终相似度由高到低排序后将结果动作返回。相似度匹配过程具体步骤如下:Step 3, similarity matching: use step 1 and step 2 to extract the feature description of the retrieval action instance submitted by the user, and then use the Euclidean distance to calculate the similarity between the action and the retrieval action in the database, and sort the final similarity from high to low Return the result action. The specific steps of the similarity matching process are as follows:
1)用户提交检索动作实例;1) The user submits a retrieval action instance;
2)采用步骤1和步骤2计算用户提交的检索动作的相对时空特征统计描述;2) Use steps 1 and 2 to calculate the statistical description of the relative spatiotemporal characteristics of the retrieval action submitted by the user;
3)对于库中的每个动作,基于相对时空特征统计描述,采用欧式距离计算和用户提交的检索动作的距离,每个特征的权重相同;3) For each action in the library, based on the statistical description of relative spatio-temporal features, the Euclidean distance is used to calculate the distance from the search action submitted by the user, and the weight of each feature is the same;
4)对库中每个动作和用户提交的检索动作的距离进行排序;4) Sort the distance between each action in the library and the retrieval action submitted by the user;
5)将距离最小的Top-20个动作返回给用户。5) Return the Top-20 actions with the smallest distance to the user.
步骤4,相关反馈:步骤3的相似度计算中每个特征都采用了相同的权重。实际上对于用户的每一次检索都存在着一个特定的特征子集及其权重组合,使得检索的性能达到最优。为了更好的捕捉用户的检索意图并逐步逼近最优的特征子集及其权重组合,本发明通过支持向量机(SVM-Support Vector Machine)相关反馈方法来优化检索结果,基于SVM的相关反馈具体步骤如下:Step 4, relevant feedback: In the similarity calculation of step 3, each feature uses the same weight. In fact, there is a specific feature subset and its weight combination for each user's retrieval, which makes the retrieval performance optimal. In order to better capture the user's retrieval intention and gradually approach the optimal feature subset and its weight combination, the present invention optimizes the retrieval results through the SVM-Support Vector Machine (SVM-Support Vector Machine) correlation feedback method. The correlation feedback based on SVM is specific Proceed as follows:
1)对于步骤3或者上一轮相关反馈的结果,用户标注返回实例和其提交的检索动作是否相关;1) For the result of step 3 or the previous round of relevant feedback, the user marks whether the returned instance is related to the search action submitted by him;
2)将用户标注的和其提交检索动作相关的样本作为正例样本,将用户标注的和其提交检索动作不相关的样本作为反例样本;2) The samples marked by the user that are related to the submitted retrieval action are taken as positive samples, and the samples marked by the user that are not related to the submitted search action are taken as negative examples;
3)选择线性核,采用SVM对正反例样本进行学习,得到SVM分类模型;3) Select the linear kernel, use SVM to learn positive and negative samples, and obtain the SVM classification model;
4)依次将库中的每个动作输入到SVM分类模型,计算得分;4) Input each action in the library to the SVM classification model in turn, and calculate the score;
5)对得分进行排序,将得分最高的Top-20个动作返回给用户;5) Sort the scores and return the Top-20 actions with the highest scores to the user;
6)如果用户对结果满意,则检索过程结束,否则返回步骤1),进行下一轮反馈。6) If the user is satisfied with the result, the retrieval process ends, otherwise return to step 1) for the next round of feedback.
有益效果Beneficial effect
1)本发明采用姿态时空相对特征作为三维姿态描述,该特征描述提取姿态中关节形成的点、线、面几何元素之间的相对空间位置及其变化的度量作为动作内容表示。关节形成的点、线、面几何元素集合是不同动作模式对应的局部区域的最小构成单元,而点、线、面几何元素之间的角度与距离等度量从不同方面反映了最小构成单元之间的相对空间位置关系,通过不同局部区域包含的不同类型特征的权重组合能够表达广泛的动作模式。1) The present invention adopts the spatial-temporal relative feature of the gesture as the three-dimensional gesture description, and the feature description extracts the relative spatial position between the geometric elements formed by the joints in the gesture, the line, and the surface and the measurement of the change thereof as the action content representation. The set of point, line, and surface geometric elements formed by joints is the smallest constituent unit of the local area corresponding to different action modes, and the measures such as angle and distance between point, line, and surface geometric elements reflect the relationship between the smallest constituent units from different aspects. A wide range of action patterns can be expressed through the weight combination of different types of features contained in different local regions.
2)本发明采用三维姿态时空相对特征的统计描述作为动作的特征表示,均值、极差、方差、偏态统计变量反映了每个特征在动作中的基准与变化情况,能够有效表达动作运动内容。同时,统计描述将时间上不同长度的动作统一转换为统一长度的特征向量,因此可以使用欧式距离以及线性SVM等方法来进行相似度计算和相关反馈,极大地提高了计算效率和检索的有效性。2) The present invention uses the statistical description of the spatial-temporal relative characteristics of three-dimensional postures as the feature representation of the action. The mean, range, variance, and skew statistical variables reflect the benchmark and change of each feature in the action, which can effectively express the content of the action . At the same time, the statistical description uniformly converts actions of different lengths in time into feature vectors of uniform length, so methods such as Euclidean distance and linear SVM can be used for similarity calculation and related feedback, which greatly improves the computational efficiency and retrieval effectiveness. .
3)本发明采用线性SVM进行相关反馈以进一步提高检索性能,SVM本身具有小样本学习泛化能力好的优点,因此可以克服由于用户标注的样本量的不足而导致的一般学习方法泛化能力差的问题。而采用线性核函数使得可以根据分类模型中每个特征的权重解释对于本次检索每个特征的不同重要程序,提高了检索结果的可解释性。3) The present invention uses linear SVM for correlation feedback to further improve retrieval performance. SVM itself has the advantage of good generalization ability of small sample learning, so it can overcome the poor generalization ability of general learning methods caused by the lack of sample size marked by users The problem. The use of the linear kernel function makes it possible to explain the different important programs for each feature in this retrieval according to the weight of each feature in the classification model, which improves the interpretability of the retrieval results.
附图说明Description of drawings
图1为本发明的方法流程图。Fig. 1 is a flow chart of the method of the present invention.
图2为本发明中使用的人体关节模型示意图。Fig. 2 is a schematic diagram of a human joint model used in the present invention.
图3为本发明定义的相对时空特征示意图。Fig. 3 is a schematic diagram of relative spatio-temporal features defined in the present invention.
图4为实施例通过相似度匹配获得的对检索输入动作“跳”的检索结果。Fig. 4 is the retrieval result of the retrieval input action "jump" obtained through similarity matching in the embodiment.
图5为实施例通过SVM相关反馈对检索动作“跳”进行检索4次迭代后的结果。Fig. 5 is the result of searching for the retrieval action "jump" for 4 iterations through SVM related feedback in the embodiment.
具体实施方式detailed description
下面结合附图和具体实施方式对本发明做更进一步的具体说明。The present invention will be further described in detail below in conjunction with the accompanying drawings and specific embodiments.
一种基于姿态相对时空特征统计描述的运动数据检索方法,包括以下步骤:A motion data retrieval method based on statistical description of posture relative spatiotemporal features, comprising the following steps:
步骤1,三维姿态相对时空特征提取:三维姿态中关节形成的点、线、面几何元素集合是不同动作模式对应的局部区域的最小构成单元。本发明提取三维姿态中关节形成的点、线、面几何元素之间的相对空间位置及其变化的度量作为姿态的特征表示,通过不同局部区域包含的不同类型特征的权重组合,姿态时空相对特征能够表达广泛的姿态模式。对于库中所有动作包含的每个姿态,其相对时空特征提取具体步骤如下:Step 1, 3D pose relative spatio-temporal feature extraction: the set of point, line, and surface geometric elements formed by joints in 3D pose is the smallest constituent unit of the local area corresponding to different action modes. The present invention extracts the relative spatial position between the geometric elements of points, lines and surfaces formed by the joints in the three-dimensional posture and the measurement of their changes as the characteristic representation of the posture. Capable of expressing a wide range of pose patterns. For each gesture contained in all actions in the library, the specific steps of its relative spatiotemporal feature extraction are as follows:
1)定义三维人体关节模型,本发明采用的三维人体关节模型如图2所示,共包含了18个关节。根据每个关节在采样时刻的具体平移或者旋转数值,计算出采样时刻每个关节的三维坐标(x,y,z);1) Define a three-dimensional human joint model. The three-dimensional human joint model adopted in the present invention is shown in FIG. 2 and includes 18 joints in total. According to the specific translation or rotation value of each joint at the sampling time, calculate the three-dimensional coordinates (x, y, z) of each joint at the sampling time;
2)构建几何元素集合,选择的关节构成了几何元素集合中的点集,点集中任意2点形成直线,任意3点则构成平面。本发明对直线和平面这两个几何元素进行合理的约束,只考虑相邻关节形成的直线与平面以约减姿态特征的维数,最终的姿态几何元素集合中共包含了17条骨骼形成的直线,以及相邻关节形成的10个平面;2) Construct a set of geometric elements. The selected joints constitute a point set in the set of geometric elements. Any 2 points in the point set form a straight line, and any 3 points form a plane. The present invention reasonably constrains the two geometric elements of straight lines and planes, and only considers the straight lines and planes formed by adjacent joints to reduce the dimension of posture features. The final set of posture geometric elements contains a total of 17 straight lines formed by bones. , and 10 planes formed by adjacent joints;
3)提取每个三维姿态相对空间特征,包括关节对距离特征、关节与骨骼距离特征、关节与平面距离特征、骨骼对夹角特征、骨骼与平面夹角特征、平面与平面夹角特征、关节旋转特征。提取的三维姿态相对空间特征如图3所示,具体计算过程如下;3) Extract the relative spatial features of each 3D pose, including joint pair distance feature, joint and bone distance feature, joint and plane distance feature, bone pair angle feature, bone and plane angle feature, plane and plane angle feature, joint Rotate feature. The extracted 3D posture relative spatial features are shown in Figure 3, and the specific calculation process is as follows;
a)关节对距离特征Fj,j,d,本发明使用欧氏距离计算几何元素集合中关节对之间的距离,设姿态中关节j1、j2的三维坐标分别为(x1,y1,z1)、(x2,y2,z2),则关节对之间的距离计算公式为:a) Joint pair distance feature F j, j, d , the present invention uses Euclidean distance to calculate the distance between joint pairs in the geometric element set, and the three-dimensional coordinates of joints j 1 and j 2 in the posture are respectively (x 1 , y 1 ,z 1 ), (x 2 ,y 2 ,z 2 ), the formula for calculating the distance between joint pairs is:
b)关节与骨骼距离特征Fj,l,d,关节到骨骼的距离通过三角形面积公式来计算,设d12、d13、d23分别为关节j1、j2、j3之间的距离,p=(d12+d23+d13)/2,则关节j1与关节j2、j3形成的直线之间的距离为:b) The distance feature F j,l,d between the joint and the bone. The distance from the joint to the bone is calculated by the triangle area formula. Let d 12 , d 13 , and d 23 be the distances between joints j 1 , j 2 , and j 3 respectively , p=(d 12 +d 23 +d 13 )/2, then the distance between joint j 1 and the line formed by joints j 2 and j 3 is:
Fj,l,d=2p(p-d12)(p-d13)(p-d23)/d23;F j,l,d = 2p(pd 12 )(pd 13 )(pd 23 )/d 23 ;
c)关节与平面距离特征Fj,p,d,关节到平面的距离通过关节和平面上任意一点形成的向量和平面法向量之间的点积求得,设n为关节j2、j3、j4形成的平面的法向量,v为关节j1、j3形成的向量,则关节j1到j2、j3、j4形成的平面的距离为:c) The distance feature F j,p,d between the joint and the plane. The distance from the joint to the plane is obtained by the dot product between the vector formed by the joint and any point on the plane and the normal vector of the plane. Let n be joints j 2 and j 3 , the normal vector of the plane formed by j 4 , v is the vector formed by joints j 1 and j 3 , then the distance from joint j 1 to the plane formed by j 2 , j 3 and j 4 is:
Fj,p,d=n·v/||n||;F j,p,d = n·v/||n||;
d)骨骼对夹角特征Fl,l,a,骨骼与骨骼的夹角通过向量点积公式来计算,若关节j1、j2形成向量va,关节j3、j4形成向量vb,则骨骼之间的夹角计算公式为:d) Bone pair angle feature F l,l,a , the angle between bones and bones is calculated by the vector dot product formula, if joints j 1 and j 2 form vector v a , joints j 3 and j 4 form vector v b , the formula for calculating the included angle between bones is:
Fl,l,a=arccos(va·vb/(||va||×||vb||));F l,l,a = arccos(v a ·v b /(||v a ||×||v b ||));
e)骨骼与平面夹角特征Fl,p,a,骨骼与平面之间的夹角通过骨骼与平面法向量的点积公式进行计算,设n为关节j3、j4、j5形成的平面P的法向量,v为关节j1、j2形成的向量,则关节j1、j2形成的骨骼与P的夹角为:e) The angle feature F l,p,a between the bone and the plane, the angle between the bone and the plane is calculated by the dot product formula of the normal vector of the bone and the plane, let n be the joint j 3 , j 4 , j 5 formed The normal vector of plane P, v is the vector formed by joints j 1 and j 2 , then the angle between the bone formed by joints j 1 and j 2 and P is:
Fl,p,a=arccos(n·v/(||n||×||v||));F l,p,a = arccos(n·v/(||n||×||v||));
f)平面与平面夹角特征Fp,p,a,平面与平面之间的夹角通过平面的法向量的点积公式进行计算,设n1为关节j1、j2、j3形成的平面P1的法向量,n2为关节j4、j5、j6形成的平面P2的法向量,则P1与P2的夹角为:f) The angle feature F p,p,a between the plane and the plane, the angle between the plane and the plane is calculated by the dot product formula of the normal vector of the plane, let n 1 be the joint j 1 , j 2 , j 3 formed The normal vector of the plane P 1 , n 2 is the normal vector of the plane P 2 formed by the joints j 4 , j 5 , j 6 , then the angle between P 1 and P 2 is:
Fp,p,a=arccos(n1·n2/(||n1||×||n2||));F p,p , a = arccos(n 1 ·n 2 /(||n 1 ||×||n 2 ||));
g)关节旋转特征Feuler,以上姿态空间特征只有一维信息,并不能反映三维的相邻关节旋转信息,本发明选用欧拉角表示相邻关节的旋转信息。g) Joint rotation feature Feuler , the above posture space features only have one-dimensional information, and cannot reflect three-dimensional adjacent joint rotation information. The present invention uses Euler angles to represent the adjacent joint rotation information.
4)提取每个三维姿态相对时间特征,包括关节角速度与加速度特征。本发明采用Kim等人在文献Kim T H,Park S I,Shin S Y.Rhythmic-motion synthesis based onmotion-beat analysis[J].Acm Transactions on Graphics,2003,22(3):392-401中提出的方法来计算关节的角速度与加速度。提取的角速度与加速度特征示意图如图3所示。假设关节j在时刻i-1和i的旋转用四元数分别表示为qj(i-1)和qj(i),采样间隔时间为Δt。4) Extract relative time features of each three-dimensional attitude, including joint angular velocity and acceleration features. The present invention adopts the method proposed in Kim TH, Park SI, Shin S Y. Rhythmic-motion synthesis based on motion-beat analysis [J]. Acm Transactions on Graphics, 2003,22 (3): 392-401 by Kim et al. To calculate the angular velocity and acceleration of the joint. The schematic diagram of the extracted angular velocity and acceleration features is shown in Figure 3. Assume that the rotations of joint j at time i-1 and i are represented by quaternions as q j (i-1) and q j (i) respectively, and the sampling interval is Δt.
a)关节j在时刻i的角速度为:a) The angular velocity of joint j at time i is:
b)在计算出角速度后,加速度可以根据角速度的变化求得,即:b) After calculating the angular velocity, the acceleration can be obtained according to the change of the angular velocity, namely:
步骤2,动作统计描述生成:动作由若干个姿态连续变化而成,本发明提取姿态相对时空特征的统计描述作为动作的特征向量表示,选择的统计量包括,均值、极差、方差以及偏态共4个统计变量。对于库中的每个动作,动作统计描述生成的具体步骤如下:Step 2, Action Statistical Description Generation: The action is formed by continuous changes of several postures. The present invention extracts the statistical description of the relative spatiotemporal characteristics of the posture as the feature vector representation of the action. The selected statistics include mean, range, variance and skewness There are 4 statistical variables in total. For each action in the library, the specific steps of action statistics description generation are as follows:
1)装载动作中包含的每个三维姿态的相对时空特征,并把每个特征作为一维随机变量,动作中的每个姿态对应了该随机变量的取值,n为动作中包含的姿态数,动作中包含的n个姿态关于一个姿态相对时空特征的取值为x1,x2,...,xn;1) Load the relative spatio-temporal features of each 3D pose included in the action, and use each feature as a 1D random variable, each pose in the action corresponds to the value of the random variable, n is the number of poses included in the action , the n gestures contained in the action have values of x 1 , x 2 ,...,x n with respect to the relative spatio-temporal features of a gesture;
2)计算每个相对时空特征的均值:2) Calculate the mean of each relative spatiotemporal feature:
3)计算每个相对时空特征的极差:3) Calculate the range of each relative spatiotemporal feature:
r=max(x1,x2,…,xn)-min(x1,x2,…,xn);r=max(x 1 , x 2 , . . . , x n )-min(x 1 , x 2 , . . . , x n );
4)计算每个相对时空特征的方差:4) Calculate the variance of each relative spatiotemporal feature:
5)计算每个相对时空特征的偏态,偏态是数据分布形状的度量,计算方法为:5) Calculate the skewness of each relative spatiotemporal feature, which is a measure of the shape of the data distribution, and the calculation method is:
6)将每个相对时空特征提取的均值、极差、方差、偏态,进行归一化处理,然后按顺序进行排列并保存,形成每个动作的特征向量表示。6) Normalize the mean value, range, variance, and skewness of each relative spatiotemporal feature extraction, and then arrange and save them in order to form a feature vector representation of each action.
步骤3,相似度匹配:采用步骤1和步骤2提取用户提交的检索动作实例的特征描述,并进而采用欧式距离计算库中动作和检索动作的相似程度,对最终相似度由高到低排序后将结果动作返回。相似度匹配过程具体步骤如下:Step 3, similarity matching: use step 1 and step 2 to extract the feature description of the retrieval action instance submitted by the user, and then use the Euclidean distance to calculate the similarity between the action and the retrieval action in the database, and sort the final similarity from high to low Return the result action. The specific steps of the similarity matching process are as follows:
1)用户提交检索动作实例;1) The user submits a retrieval action instance;
2)采用步骤1和步骤2计算用户提交的检索动作的相对时空特征统计描述;2) Use steps 1 and 2 to calculate the statistical description of the relative spatiotemporal characteristics of the retrieval action submitted by the user;
3)对于库中的每个动作,基于相对时空特征统计描述,采用欧式距离计算和用户提交的检索动作的距离,每个特征的权重相同。设用户的检索动作为i,库中的候选动作为j,Nf为三维姿态相对时空特征总数,fi,k,l和fj,k,l分别为动作i和动作j关于姿态相对时空特征f的统计描述k的取值,则检索动作为i和库中的候选动作为j之间的距离具体计算方法为:3) For each action in the library, based on the statistical description of the relative spatio-temporal features, the Euclidean distance is used to calculate the distance from the search action submitted by the user, and each feature has the same weight. Suppose the user’s retrieval action is i, the candidate action in the library is j, N f is the total number of 3D pose relative spatiotemporal features, f i,k,l and f j,k,l are action i and action j’s relative spatiotemporal features The statistics of the feature f describe the value of k, then the specific calculation method for the distance between the retrieval action i and the candidate action j in the library is:
4)对库中每个动作和用户提交的检索动作的距离进行排序;4) Sort the distance between each action in the library and the retrieval action submitted by the user;
5)将距离最小的Top-20个动作返回给用户。5) Return the Top-20 actions with the smallest distance to the user.
步骤4,相关反馈:步骤3的相似度计算中每个特征都采用了相同的权重。实际上对于用户的每一次检索都存在着一个特定的特征子集及其权重组合,使得检索的性能达到最优。为了更好的捕捉用户的检索意图并逐步逼近最优的特征子集及其权重组合,本发明通过支持向量机(SVM-Support Vector Machine)相关反馈方法来优化检索结果,基于SVM的相关反馈具体步骤如下:Step 4, relevant feedback: In the similarity calculation of step 3, each feature uses the same weight. In fact, there is a specific feature subset and its weight combination for each user's retrieval, which makes the retrieval performance optimal. In order to better capture the user's retrieval intention and gradually approach the optimal feature subset and its weight combination, the present invention optimizes the retrieval results through the SVM-Support Vector Machine (SVM-Support Vector Machine) correlation feedback method. The correlation feedback based on SVM is specific Proceed as follows:
1)对于步骤3或者上一轮相关反馈的结果,用户标注返回实例和其提交的检索动作是否相关;1) For the result of step 3 or the previous round of relevant feedback, the user marks whether the returned instance is related to the search action submitted by him;
2)将用户标注的和其提交检索动作相关的样本作为正例样本,将用户标注的和其提交检索动作不相关的样本作为反例样本;2) The samples marked by the user that are related to the submitted retrieval action are taken as positive samples, and the samples marked by the user that are not related to the submitted search action are taken as negative examples;
3)选择线性核,采用SVM对正反例样本进行学习,得到SVM分类模型;3) Select the linear kernel, use SVM to learn positive and negative samples, and obtain the SVM classification model;
4)依次将库中的每个动作输入到SVM分类模型,计算得分;4) Input each action in the library to the SVM classification model in turn, and calculate the score;
5)对得分进行排序,将得分最高的Top-20个动作返回给用户;5) Sort the scores and return the Top-20 actions with the highest scores to the user;
6)如果用户对结果满意,则检索过程结束,否则返回步骤1),进行下一轮反馈。6) If the user is satisfied with the result, the retrieval process ends, otherwise return to step 1) for the next round of feedback.
使用本方案实现的运动检索系统对“跳”动作进行检索的效果如图4和图5所示,图4为使用相似度匹配的结果,由于每个特征都使用相同的权重,特征的权重组合并没有经过优化,所以在首页显示的9个检索结果动作中,只有4个是相关的。图5为经过线性SVM相关反馈4次迭代后的结果,通过在线学习分类模型优化特征权重组合,检索性能得到了明显的提高,在首页显示9个检索结果动作全部为相关动作。Figure 4 and Figure 5 show the effect of the motion retrieval system implemented in this scheme on the "jump" action. Figure 4 shows the result of using similarity matching. Since each feature uses the same weight, the weight combination of features It is not optimized, so out of 9 search result actions displayed on the home page, only 4 are relevant. Figure 5 shows the results after 4 iterations of linear SVM correlation feedback. By optimizing the combination of feature weights through the online learning classification model, the retrieval performance has been significantly improved. All 9 retrieval result actions displayed on the home page are related actions.
本发明提供了一种基于姿态相对时空特征统计描述的运动数据检索方法,具体实现该技术方案的方法和途径很多,以上所述仅是本发明的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。本实施例中未明确的各组成部分均可用现有技术加以实现。The present invention provides a motion data retrieval method based on the statistical description of attitude relative to spatio-temporal features. There are many methods and approaches to specifically realize the technical solution. The above description is only a preferred embodiment of the present invention. Those of ordinary skill may make some improvements and modifications without departing from the principle of the present invention, and these improvements and modifications shall also be regarded as the protection scope of the present invention. All components that are not specified in this embodiment can be realized by existing technologies.
Claims (5)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201611204708.4A CN106778639A (en) | 2016-12-23 | 2016-12-23 | A kind of exercise data search method based on the description of attitude relative space-time characteristic statisticses |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201611204708.4A CN106778639A (en) | 2016-12-23 | 2016-12-23 | A kind of exercise data search method based on the description of attitude relative space-time characteristic statisticses |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN106778639A true CN106778639A (en) | 2017-05-31 |
Family
ID=58897723
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201611204708.4A Pending CN106778639A (en) | 2016-12-23 | 2016-12-23 | A kind of exercise data search method based on the description of attitude relative space-time characteristic statisticses |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN106778639A (en) |
Cited By (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109241909A (en) * | 2018-09-06 | 2019-01-18 | 闫维新 | A kind of long-range dance movement capture evaluating system based on intelligent terminal |
| CN109543054A (en) * | 2018-10-17 | 2019-03-29 | 天津大学 | A kind of Feature Dimension Reduction method for searching three-dimension model based on view |
| CN109993818A (en) * | 2017-12-31 | 2019-07-09 | 中国移动通信集团辽宁有限公司 | Method, device, equipment and medium for motion synthesis of three-dimensional human model |
| CN110362843A (en) * | 2018-11-20 | 2019-10-22 | 莆田学院 | A kind of visual human's entirety posture approximation generation method based on typical posture |
| CN114267087A (en) * | 2022-02-28 | 2022-04-01 | 成都考拉悠然科技有限公司 | An action registration method and system based on a small sample machine learning model |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102298649A (en) * | 2011-10-09 | 2011-12-28 | 南京大学 | Space trajectory retrieval method of body movement data |
| US20150186374A1 (en) * | 2013-12-27 | 2015-07-02 | Nuctech Company Limited | Retrieving system, retrieving method, and security inspection device based on contents of fluoroscopic images |
-
2016
- 2016-12-23 CN CN201611204708.4A patent/CN106778639A/en active Pending
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102298649A (en) * | 2011-10-09 | 2011-12-28 | 南京大学 | Space trajectory retrieval method of body movement data |
| US20150186374A1 (en) * | 2013-12-27 | 2015-07-02 | Nuctech Company Limited | Retrieving system, retrieving method, and security inspection device based on contents of fluoroscopic images |
Non-Patent Citations (4)
| Title |
|---|
| TANG J K T ET AL: "《Retrieval of logically relevant 3D human motions by Adaptive Feature Selection》", 《PATTERN RECOGNITION LETTERS》 * |
| XIANG SEAN ZHOU ET AL: "《relevance feedback in image retrieval:comprehensive review》", 《MULTIMEDIA SYSTEMS》 * |
| 胡莹等: "《一种改进的SVM相关反馈图像检索方法》", 《计算机应用研究》 * |
| 陈松乐等: "《民族舞蹈运动数据的实例检索方法》", 《计算机辅助设计与图形学学报》 * |
Cited By (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109993818A (en) * | 2017-12-31 | 2019-07-09 | 中国移动通信集团辽宁有限公司 | Method, device, equipment and medium for motion synthesis of three-dimensional human model |
| CN109993818B (en) * | 2017-12-31 | 2023-09-19 | 中国移动通信集团辽宁有限公司 | Method, device, equipment and medium for synthesizing motion of three-dimensional human body model |
| CN109241909A (en) * | 2018-09-06 | 2019-01-18 | 闫维新 | A kind of long-range dance movement capture evaluating system based on intelligent terminal |
| CN109543054A (en) * | 2018-10-17 | 2019-03-29 | 天津大学 | A kind of Feature Dimension Reduction method for searching three-dimension model based on view |
| CN109543054B (en) * | 2018-10-17 | 2022-12-09 | 天津大学 | View-based feature dimension reduction three-dimensional model retrieval method |
| CN110362843A (en) * | 2018-11-20 | 2019-10-22 | 莆田学院 | A kind of visual human's entirety posture approximation generation method based on typical posture |
| CN114267087A (en) * | 2022-02-28 | 2022-04-01 | 成都考拉悠然科技有限公司 | An action registration method and system based on a small sample machine learning model |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Tang et al. | Fast and robust dynamic hand gesture recognition via key frames extraction and feature fusion | |
| Cai et al. | Effective active skeleton representation for low latency human action recognition | |
| Zhang et al. | On geometric features for skeleton-based action recognition using multilayer lstm networks | |
| Wan et al. | Explore efficient local features from RGB-D data for one-shot learning gesture recognition | |
| Lin et al. | Human action recognition and retrieval using sole depth information | |
| CN102298649B (en) | Space trajectory retrieval method of body movement data | |
| Bu et al. | Learning high-level feature by deep belief networks for 3-D model retrieval and recognition | |
| Yang et al. | Articulated human detection with flexible mixtures of parts | |
| CN104616316B (en) | Personage's Activity recognition method based on threshold matrix and Fusion Features vision word | |
| Liu et al. | Robust 3D action recognition through sampling local appearances and global distributions | |
| CN103295025B (en) | A kind of automatic selecting method of three-dimensional model optimal view | |
| CN106778639A (en) | A kind of exercise data search method based on the description of attitude relative space-time characteristic statisticses | |
| Zhang et al. | Human pose estimation and tracking via parsing a tree structure based human model | |
| CN105868706A (en) | Method for identifying 3D model based on sparse coding | |
| Tang et al. | Retrieval of logically relevant 3D human motions by adaptive feature selection with graded relevance feedback | |
| CN103336835A (en) | Image retrieval method based on weight color-sift characteristic dictionary | |
| CN103294832A (en) | Motion capture data retrieval method based on feedback study | |
| Liu et al. | Efficient human motion retrieval via temporal adjacent bag of words and discriminative neighborhood preserving dictionary learning | |
| Valcik et al. | Assessing similarity models for human‐motion retrieval applications | |
| Xiao et al. | Sketch-based human motion retrieval via selected 2D geometric posture descriptor | |
| Zhang et al. | Boosted random contextual semantic space based representation for visual recognition | |
| Wang et al. | An eigen-based motion retrieval method for real-time animation | |
| CN101916284B (en) | Three-dimensional model searching method based on shape orientation multi-resolution analysis | |
| CN106649665A (en) | Object-level depth feature aggregation method for image retrieval | |
| Li et al. | Augmenting bag-of-words: a robust contextual representation of spatiotemporal interest points for action recognition |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| WD01 | Invention patent application deemed withdrawn after publication | ||
| WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20170531 |