[go: up one dir, main page]

CN106529441A - Fuzzy boundary fragmentation-based depth motion map human body action recognition method - Google Patents

Fuzzy boundary fragmentation-based depth motion map human body action recognition method Download PDF

Info

Publication number
CN106529441A
CN106529441A CN201610949051.8A CN201610949051A CN106529441A CN 106529441 A CN106529441 A CN 106529441A CN 201610949051 A CN201610949051 A CN 201610949051A CN 106529441 A CN106529441 A CN 106529441A
Authority
CN
China
Prior art keywords
depth
time slice
training
fuzzy
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610949051.8A
Other languages
Chinese (zh)
Other versions
CN106529441B (en
Inventor
蒋敏
金科
孔军
昝宝锋
胡珂杰
徐海洋
刘天山
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huirong Electronic System Engineering Ltd
Original Assignee
Jiangnan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangnan University filed Critical Jiangnan University
Priority to CN201610949051.8A priority Critical patent/CN106529441B/en
Publication of CN106529441A publication Critical patent/CN106529441A/en
Application granted granted Critical
Publication of CN106529441B publication Critical patent/CN106529441B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种基于模糊边界分片的深度动作图人体行为识别方法。模型训练方法包括以下步骤:将视频深度图序列分片并根据模糊参数α确定分片的模糊边界;对于每一个分片后的子序列分别计算它们主视图、左视图和俯视图的深度动作图DMM;利用插值法将这些深度动作图转换为固定的尺寸并归一化;将归一化后的每个视频序列子序列的深度动作图DMM串联,获得该视频序列的特征向量;采用概率协作表示分类器R‑ProCRC对特征进行分类,最终实现人体行为识别。本发明公开的人体行为识别方法,有效捕获了时域特征的变化规律,增强了行为特征对时域差异的抗干扰能力,能够实现对人体行为的鲁棒识别。

The invention discloses a human behavior recognition method based on fuzzy boundary slicing in depth motion graphs. The model training method includes the following steps: slice the video depth map sequence and determine the blurred boundary of the slice according to the blur parameter α; calculate the depth motion map DMM of the main view, left view and top view for each sub-sequence after fragmentation ; use interpolation to convert these depth motion maps to a fixed size and normalize them; concatenate the normalized depth motion maps DMM of each video sequence subsequence to obtain the feature vector of the video sequence; use probabilistic collaborative representation The classifier R‑ProCRC classifies the features and finally realizes human action recognition. The human body behavior recognition method disclosed by the invention effectively captures the change rule of the time domain feature, enhances the anti-interference ability of the behavior feature to the time domain difference, and can realize the robust recognition of the human body behavior.

Description

基于模糊边界分片的深度动作图人体行为识别方法Human action recognition method based on fuzzy boundary slices in deep motion map

技术领域:Technical field:

本发明属于机器视觉领域,特别涉及一种基于模糊边界分片的深度动作图人体行为识别方法。The invention belongs to the field of machine vision, and in particular relates to a human behavior recognition method based on a fuzzy boundary slice in a deep motion map.

背景技术:Background technique:

人体行为识别技术是通过处理人体行为的视频序列,从中提取行为特征,并通过这些特征识别动作的一种技术。Human behavior recognition technology is a technology that extracts behavioral features from video sequences of human behaviors and recognizes actions through these features.

在机器视觉和模式识别领域,人体行为识别现在已经成为一个非常活跃的分支。这种技术在人机交互领域有着许多潜在的应用,包括视频分析、监控系统,智能机器人等。前几年,人体行为识别方向的研究主要基于彩色摄像头采集的图像帧序列[1][2].这种数据有着固有的缺陷,它们对于光照、遮挡以及复杂的背景十分敏感,影响识别精度。如今深度摄像头得到了广泛的关注,这种传感器能够提供物体的3D结构信息,包括人体的形状以及动作信息。目前提取深度的特征方法主要有两个分支:基于估计的3D骨骼点的方法[3]和基于原始深度图像的方法。但是这种关节点估算有时会出现错误,特别是对于人体部分遮挡或者背景复杂的情况。C.Chen等[4]通过计算原始深度图像的深度动作图(DMM,DepthMotion Maps),并通过PCA降维后利用像素点作为特征得到了相对较好的结果。但这种方法将所有视频帧累加在一张DMM上,丢失了动作的时间信息。In the field of machine vision and pattern recognition, human behavior recognition has now become a very active branch. This technology has many potential applications in the field of human-computer interaction, including video analysis, surveillance systems, intelligent robots, etc. In the past few years, research on human behavior recognition was mainly based on image frame sequences collected by color cameras[1][2]. This kind of data has inherent defects, and they are very sensitive to illumination, occlusion, and complex backgrounds, which affect recognition accuracy. Nowadays, depth cameras have received widespread attention. This sensor can provide 3D structural information of objects, including the shape and motion information of the human body. At present, there are two main branches of the feature extraction method of depth: the method based on the estimated 3D skeleton point [3] and the method based on the original depth image. However, this kind of joint point estimation can sometimes be wrong, especially for cases where the human body is partially occluded or the background is complex. C. Chen et al. [4] obtained relatively good results by calculating the depth motion map (DMM, DepthMotion Maps) of the original depth image, and using pixels as features after PCA dimensionality reduction. However, this method accumulates all video frames on one DMM and loses the time information of the action.

本发明针对上述人体行为识别存在的问题,提出了一种基于模糊边界分片的深度动作图的人体行为识别方法,不仅吸取了原始深度图像信息相对于骨骼信息的优点,并且使用模糊边界分片的技术对DMM进行分片,提取了动作的时间信息,对光照、遮挡、复杂背景等因素具有较高鲁棒性。In view of the above-mentioned problems of human behavior recognition, the present invention proposes a human behavior recognition method based on the depth action map of the fuzzy boundary slice, which not only absorbs the advantages of the original depth image information relative to the skeleton information, but also uses the fuzzy boundary slice The technology slices the DMM, extracts the time information of the action, and has high robustness to factors such as illumination, occlusion, and complex background.

发明内容:Invention content:

本发明的主要目的是提出一种基于模糊边界分片的深度动作图人体行为识别方法,捕获动作的时间信息并使用鲁棒的概率协作表示分类器R-ProCRC(RobustProbabilistic Collaborative Representation based Classifier)[5]来进行分类,提高识别精度。The main purpose of the present invention is to propose a human behavior recognition method based on fuzzy boundary slices in depth action graphs, capture the time information of actions and use the robust probabilistic collaborative representation classifier R-ProCRC (Robust Probabilistic Collaborative Representation based Classifier) [5 ] to classify and improve the recognition accuracy.

为了实现上述目的,本发明提供如下技术方案,包含训练阶段和测试阶段。In order to achieve the above object, the present invention provides the following technical solutions, including a training phase and a testing phase.

基于模糊边界分片的深度动作图的人体行为识别方法训练阶段技术方案如下:The technical scheme of the training phase of the human action recognition method based on the depth action map of the fuzzy boundary fragmentation is as follows:

步骤一:给定人体行为视频序列的深度图样本的训练集合 其中X(k)表示第k个训练样本的深度图序列, 为第k个样本中第i帧的原始深度图像,Nk为第k个样本的总帧数;Y(k)表示第k个训练样本所属行为类别;M表示训练集中样本个数;Step 1: Given a training set of depth map samples of human behavior video sequences where X (k) represents the depth map sequence of the kth training sample, Be the original depth image of the i-th frame in the k-th sample, N k is the total number of frames of the k-th sample; Y (k) represents the behavior category to which the k-th training sample belongs; M represents the number of samples in the training set;

步骤二:将训练集中每个视频序列训练样本X(k)按时间轴直接划分为DIV个等长时间片,每个时间片长度为划分后的时间片表示为其中 Step 2: Divide each video sequence training sample X (k) in the training set directly into DIV equal time slices according to the time axis, and the length of each time slice is The divided time slice is expressed as in

步骤三:选择合适的模糊参数α,对每个时间片进行分片模糊处理;模糊时间片表示为为避免下标溢出,对于第一个时间片不做前向模糊处理,对最后一个时间片不做后向模糊处理;Step 3: Select the appropriate fuzzy parameter α, and perform slice fuzzy processing on each time slice; the fuzzy time slice is expressed as To avoid subscript overflow, forward blurring is not performed for the first time slice, and backward blurring is not performed for the last time slice;

步骤四:计算每个模糊时间片在三个不同投影方向的深度动作图其中v∈{f,s,t}分别表示视频序列投影的三视图(三个方向),主视图、左视图和俯视图;至此,计算获得所有训练样本对应的深度动作图集合 Step 4: Calculate each fuzzy time slice Depth motion maps in three different projection directions where v ∈ {f, s, t} respectively represent the three views (three directions) of the video sequence projection, the main view, the left view and the top view; so far, all training samples are calculated Corresponding set of deep motion maps

步骤五:利用双三次差值法将步骤四求解获得的深度动作图调整为一个相同的尺寸,并将这些深度动作图归一化为0到1之间,归一化后的记为 Step 5: Use the bicubic difference method to solve the depth action map obtained in step 4 Adjust to the same size, and normalize these depth action maps between 0 and 1, the normalized recorded as

步骤六:将任一训练样本X(k)对应深度动作图集合进行向量化,并将所有向量化后的序列动作图串联,完成训练样本X(k)对应的特征构建,该特征记为H(k),则所有样本的特征集合表示为{H(k)}k∈[1M]Step 6: Correspond any training sample X (k) to the set of depth motion maps Carry out vectorization, and concatenate all vectorized sequence action graphs to complete the feature construction corresponding to the training sample X (k) , which is recorded as H (k) , and the feature set of all samples is expressed as {H (k) } k∈[1M] ;

步骤七:将步骤六得到的所有样本的输出特征{H(k)}k∈[1M]通过PCA降维,并保存所有降维后的特征 Step 7: All the samples obtained in Step 6 The output features of {H (k) } k∈[1M] are dimensionally reduced by PCA, and all features after dimensionality reduction are saved

基于模糊边界分片的深度动作图的人体行为识别方法的测试阶段技术方案如下:The technical scheme of the testing phase of the human action recognition method based on the depth action map of the fuzzy boundary fragmentation is as follows:

步骤一:给定一个人体行为视频序列的深度图样本的测试样本(TestX,TestY),其中TestX表示测试样本的深度图序列,Xi为测试样本中第i帧的原始深度图像,NT为该测试样本的总帧数;TestY表示测试样本所属行为类别;Step 1: Given a test sample (TestX, TestY) of a depth map sample of a human behavior video sequence, where TestX represents the depth map sequence of the test sample, X i is the original depth image of the i-th frame in the test sample, NT is the total number of frames of the test sample; TestY indicates the behavior category to which the test sample belongs;

步骤二:将测试样本TestX按时间轴直接划分为DIV个等长时间片(分片方式与训练阶段相同),每个时间片长度为划分后的时间片表示为其中 Step 2: Divide the test sample TestX directly into DIV equal time slices according to the time axis (the segmentation method is the same as the training stage), and the length of each time slice is The divided time slice is expressed as in

步骤三:根据训练阶段采用的模糊参数α,对每个时间片进行分片模糊处理;模糊时间片表示为为避免下标溢出,对于第一个时间片不做前向模糊处理,对最后一个时间片不做后向模糊处理;Step 3: According to the fuzzy parameter α used in the training stage, perform slice fuzzy processing on each time slice; the fuzzy time slice is expressed as To avoid subscript overflow, forward blurring is not performed for the first time slice, and backward blurring is not performed for the last time slice;

步骤四:计算每个模糊时间片在三个不同投影方向的深度动作图DMMj,v,其中v∈{f,s,t}分别表示视频序列投影的三视图(三个方向),主视图、左视图和俯视图;至此,计算获得测试样本TestX对应的深度动作图集合{DMMj,v}j∈[1DIV],v∈[f,s,t]Step 4: Calculate each fuzzy time slice Depth motion map DMM j,v in three different projection directions, where v∈{f,s,t} respectively represent three views (three directions) of video sequence projection, front view, left view and top view; so far, calculate Obtain the depth motion map set {DMM j,v } j∈[1DIV],v∈[f,s,t] corresponding to the test sample TestX;

步骤五:利用双三次差值法将步骤四求解获得的测试样本深度动作图{DMMj,v}j∈[1DIV],v∈[f,s,t]调整为训练阶段同一尺寸,并根据训练阶段归一化方法,将这些深度动作图归一化为0到1之间,归一化后的DMMj,v记为 Step 5: Use the bicubic difference method to adjust the test sample depth action map {DMM j,v } j∈[1DIV],v∈[f,s,t] obtained in step 4 to the same size as the training stage, and according to The normalization method in the training phase normalizes these depth action maps to be between 0 and 1, and the normalized DMM j,v is recorded as

步骤六:将测试样本TestX对应的深度动作图集合进行向量化,并将所有向量化后的序列动作图串联,完成测试样本TestX对应的特征构建,该特征记为HTStep 6: Collect the depth action images corresponding to the test sample TestX Carry out vectorization and concatenate all vectorized sequence action graphs to complete the feature construction corresponding to the test sample TestX, which is denoted as H T ;

步骤七:将步骤六得到的测试样本TestX的输出特征HT通过PCA降维,得到降维后的特征PCA降维方法与训练阶段一致;Step 7: Reduce the dimensionality of the output feature H T of the test sample TestX obtained in Step 6 through PCA to obtain the dimensionality-reduced feature The PCA dimensionality reduction method is consistent with the training phase;

步骤八:然后将降维后的输出特征送入R-ProCRC[5]分类器,获得分类输出PridY;Step 8: Then the output features after dimensionality reduction Send it to the R-ProCRC[5] classifier to obtain the classification output PridY;

步骤九:比较PridY和TestY,若PridY=TestY,则识别正确,否则识别错误。Step 9: Compare PridY and TestY, if PridY=TestY, the recognition is correct, otherwise the recognition is wrong.

与现有的技术相比,本发明具有以下有益效果:Compared with the prior art, the present invention has the following beneficial effects:

1、本方法采用深度数据来进行人体行为识别,相比于传统的彩色视频数据,深度数据能够实现对人体的高效分割,同时保存了人体的形状和结构特征,有助于提高分类精度;1. This method uses depth data for human body behavior recognition. Compared with traditional color video data, depth data can achieve efficient segmentation of human body, while preserving the shape and structural characteristics of human body, which helps to improve classification accuracy;

2、传统的利用深度动作图DMM的特征提取方法将整个视频帧投影到一张DMM图上,丢失时间了信息。本方法提出的模糊边界分片的深度动作图人体行为识别方法,将深度图序列进行时间维度的分片,有效捕获了时域特征的变化规律;2. The traditional feature extraction method using the depth motion map DMM projects the entire video frame onto a DMM map, which loses time information. The human behavior recognition method of depth action map based on fuzzy boundary segmentation proposed by this method, slices the depth image sequence in the time dimension, and effectively captures the change law of time domain features;

3、针对人体行为时域差异性,本方法提出的模糊边界分片的深度动作图的人体行为识别方法,使用模糊参数α来控制分片之间的边界,使得相邻分片信息得到共享,进一步提高了特征对人体行为时域差异的鲁棒性,配合使用R-ProCRC分类器识别精度得到显著提高。3. Aiming at the time-domain differences of human behaviors, the human behavior recognition method of the depth motion map of the fuzzy boundary slices proposed by this method uses the fuzzy parameter α to control the boundaries between slices, so that the information of adjacent slices can be shared. The robustness of the feature to the time domain difference of human behavior is further improved, and the recognition accuracy of the R-ProCRC classifier is significantly improved.

附图说明:Description of drawings:

图1基于模糊边界分片的深度动作图的人体行为识别特征构建方法流程图;Fig. 1 is a flow chart of a method for constructing human behavior recognition features based on a depth action map of fuzzy boundary slices;

图2模糊边界分片策略示意图;Figure 2 Schematic diagram of fuzzy boundary fragmentation strategy;

图3深度动作图在三个视角下的投影实例图;Figure 3 is an example of the projection of the depth action image under three viewing angles;

具体实施方式detailed description

为了更好的说明本发明的目的、具体步骤以及特点,下面结合附图,以MSRAction3D数据集为例,对本发明作进一步详细的说明:In order to better illustrate the purpose, specific steps and characteristics of the present invention, the present invention will be further described in detail below in conjunction with the accompanying drawings, taking the MSRAction3D data set as an example:

本发明提出的模糊边界分片的深度动作图的人体行为识别方法,其中特征提取的流程图如图1所示。首先对样本进行确定边界的等量分片,然后根据参数α来确定边界的模糊程度,对于每一个分片后的视频子序列都计算它的深度动作图DMM,利用双三次差值法将所有样本的DMM固定到一个相同的尺寸,并归一化,串联向量化后得到子序列的特征,完成训练样本的输出特征的构建。The human behavior recognition method of the depth motion map of the fuzzy boundary segmentation proposed by the present invention, wherein the flow chart of feature extraction is shown in FIG. 1 . Firstly, the sample is divided into equal segments to determine the boundary, and then the blurring degree of the boundary is determined according to the parameter α. For each segmented video sub-sequence, its depth motion map DMM is calculated, and all The DMM of the sample is fixed to the same size and normalized, and the features of the subsequence are obtained after serial vectorization, and the construction of the output feature of the training sample is completed.

本发明提出的一种基于模糊边界分片的深度动作图的人体行为识别方法,包含训练阶段和测试阶段。A human behavior recognition method based on a deep motion graph of fuzzy boundary slices proposed by the present invention includes a training phase and a testing phase.

基于模糊边界分片的深度动作图的人体行为识别方法训练阶段技术方案如下:The technical scheme of the training phase of the human action recognition method based on the depth action map of the fuzzy boundary fragmentation is as follows:

步骤一:给定人体行为视频序列的深度图样本的训练集合 其中X(k)表示第k个训练样本的深度图序列, 为第k个样本中第i帧的原始深度图像,Nk为第k个样本的总帧数;Y(k)表示第k个训练样本所属行为类别;M表示训练集中样本个数;Step 1: Given a training set of depth map samples of human behavior video sequences where X (k) represents the depth map sequence of the kth training sample, Be the original depth image of the i-th frame in the k-th sample, N k is the total number of frames of the k-th sample; Y (k) represents the behavior category to which the k-th training sample belongs; M represents the number of samples in the training set;

步骤二:将训练集中每个视频序列训练样本X(k)按时间轴直接划分为DIV个等长时间片,每个时间片长度为划分后的时间片表示为其中 Step 2: Divide each video sequence training sample X (k) in the training set directly into DIV equal time slices according to the time axis, and the length of each time slice is The divided time slice is expressed as in

步骤三:选择合适的模糊参数α,对每个时间片进行分片模糊处理;模糊时间片表示为为避免下标溢出,对于第一个时间片不做前向模糊处理,对最后一个时间片不做后向模糊处理;Step 3: Select the appropriate fuzzy parameter α, and perform slice fuzzy processing on each time slice; the fuzzy time slice is expressed as To avoid subscript overflow, forward blurring is not performed for the first time slice, and backward blurring is not performed for the last time slice;

步骤四:计算每个模糊时间片在三个不同投影方向的深度动作图其中v∈{f,s,t}分别表示视频序列投影的三视图(三个方向),主视图、左视图和俯视图;至此,计算获得所有训练样本对应的深度动作图集合 Step 4: Calculate each fuzzy time slice Depth motion maps in three different projection directions where v ∈ {f, s, t} respectively represent the three views (three directions) of the video sequence projection, the main view, the left view and the top view; so far, all training samples are calculated Corresponding set of deep motion maps

步骤五:利用双三次差值法将步骤四求解获得的深度动作图调整为一个相同的尺寸,并将这些深度动作图归一化为0到1之间,归一化后的记为 Step 5: Use the bicubic difference method to solve the depth action map obtained in step 4 Adjust to the same size, and normalize these depth action maps between 0 and 1, the normalized recorded as

步骤六:将任一训练样本X(k)对应深度动作图集合进行向量化,并将所有向量化后的序列动作图串联,完成训练样本X(k)对应的特征构建,该特征记为H(k),则所有样本的特征集合表示为{H(k)}k∈[1M]Step 6: Correspond any training sample X (k) to the set of depth motion maps Carry out vectorization, and concatenate all vectorized sequence action graphs to complete the feature construction corresponding to the training sample X (k) , which is recorded as H (k) , and the feature set of all samples is expressed as {H (k) } k∈[1M] ;

步骤七:将步骤六得到的所有样本的输出特征{H(k)}k∈[1M]通过PCA降维,并保存所有降维后的特征 Step 7: All the samples obtained in Step 6 The output features of {H (k) } k∈[1M] are dimensionally reduced by PCA, and all features after dimensionality reduction are saved

基于模糊边界分片的深度动作图的人体行为识别方法的测试阶段技术方案如下:The technical scheme of the testing phase of the human action recognition method based on the depth action map of the fuzzy boundary fragmentation is as follows:

步骤一:给定一个人体行为视频序列的深度图样本的测试样本(TestX,TestY),其中TestX表示测试样本的深度图序列,Xi为测试样本中第i帧的原始深度图像,NT为该测试样本的总帧数;TestY表示测试样本所属行为类别;Step 1: Given a test sample (TestX, TestY) of a depth map sample of a human behavior video sequence, where TestX represents the depth map sequence of the test sample, X i is the original depth image of the i-th frame in the test sample, NT is the total number of frames of the test sample; TestY indicates the behavior category to which the test sample belongs;

步骤二:将测试样本TestX按时间轴直接划分为DIV个等长时间片(分片方式与训练阶段相同),每个时间片长度为划分后的时间片表示为其中 Step 2: Divide the test sample TestX directly into DIV equal time slices according to the time axis (the segmentation method is the same as the training stage), and the length of each time slice is The divided time slice is expressed as in

步骤三:根据训练阶段采用的模糊参数α,对每个时间片进行分片模糊处理;模糊时间片表示为为避免下标溢出,对于第一个时间片不做前向模糊处理,对最后一个时间片不做后向模糊处理;Step 3: According to the fuzzy parameter α used in the training stage, perform slice fuzzy processing on each time slice; the fuzzy time slice is expressed as To avoid subscript overflow, forward blurring is not performed for the first time slice, and backward blurring is not performed for the last time slice;

步骤四:计算每个模糊时间片在三个不同投影方向的深度动作图DMMj,v,其中v∈{f,s,t}分别表示视频序列投影的三视图(三个方向),主视图、左视图和俯视图;至此,计算获得测试样本TestX对应的深度动作图集合{DMMj,v}j∈[1DIV],v∈[f,s,t]Step 4: Calculate each fuzzy time slice Depth motion map DMM j,v in three different projection directions, where v∈{f,s,t} respectively represent three views (three directions) of video sequence projection, front view, left view and top view; so far, calculate Obtain the depth motion map set {DMM j,v } j∈[1DIV],v∈[f,s,t] corresponding to the test sample TestX;

步骤五:利用双三次差值法将步骤四求解获得的测试样本深度动作图{DMMj,v}j∈[1DIV],v∈[f,s,t]调整为训练阶段同一尺寸,并根据训练阶段归一化方法,将这些深度动作图归一化为0到1之间,归一化后的DMMj,v记为 Step 5: Use the bicubic difference method to adjust the test sample depth action map {DMM j,v } j∈[1DIV],v∈[f,s,t] obtained in step 4 to the same size as the training stage, and according to The normalization method in the training phase normalizes these depth action maps to be between 0 and 1, and the normalized DMM j,v is recorded as

步骤六:将测试样本TestX对应的深度动作图集合进行向量化,并将所有向量化后的序列动作图串联,完成测试样本TestX对应的特征构建,该特征记为HTStep 6: Collect the depth action images corresponding to the test sample TestX Carry out vectorization and concatenate all vectorized sequence action graphs to complete the feature construction corresponding to the test sample TestX, which is denoted as H T ;

步骤七:将步骤六得到的测试样本TestX的输出特征HT通过PCA降维,得到降维后的特征PCA降维方法与训练阶段一致;Step 7: Reduce the dimensionality of the output feature H T of the test sample TestX obtained in Step 6 through PCA to obtain the dimensionality-reduced feature The PCA dimensionality reduction method is consistent with the training phase;

步骤八:然后将降维后的输出特征送入R-ProCRC[5]分类器,获得分类输出PridY;Step 8: Then the output features after dimensionality reduction Send it to the R-ProCRC[5] classifier to obtain the classification output PridY;

步骤九:比较PridY和TestY,若PridY=TestY,则识别正确,否则识别错误。Step 9: Compare PridY and TestY, if PridY=TestY, the recognition is correct, otherwise the recognition is wrong.

上述技术方案中,训练阶段步骤二对视频序列进行等长时间片划分的方法中,分片数DIV的选择根据具体的人体行为样本数据集来确定最优分片数,以MSR Action3D数据集为例,DIV=3。In the above-mentioned technical scheme, in the method of dividing the video sequence into equal length slices in step 2 of the training phase, the selection of the number of slices DIV is based on the specific human behavior sample data set to determine the optimal number of slices, taking the MSR Action3D data set as For example, DIV=3.

上述技术方案中,训练阶段步骤二对视频序列进行等长时间片划分的方法中,每个时间片长度为时间片长度采用向下取整方式;最后一个时间片若长度不足则按实际长度选取;In the above-mentioned technical scheme, in the method for dividing the video sequence into equal time slices in step 2 of the training phase, the length of each time slice is The length of the time slice is rounded down; if the length of the last time slice is insufficient Then choose according to the actual length;

上述技术方案中,训练阶段步骤三对时间片的分片模糊处理中,模糊参数α的选择根据具体的人体行为样本数据集来确定最优参数,以MSR Action3D数据集为例,α=0.8。In the above technical solution, in step 3 of the training stage, in the segmental fuzzy processing of time slices, the selection of the fuzzy parameter α is based on the specific human behavior sample data set to determine the optimal parameter. Taking the MSR Action3D data set as an example, α=0.8.

上述技术方案中,训练阶段步骤三对时间片的分片模糊处理中,对每个样本的第一个时间片不做前向模糊处理;对每个样本的最后一个时间片不做后向模糊处理。In the above technical solution, in step three of the training phase, in the segmental fuzzy processing of time slices, no forward fuzzy processing is performed on the first time slice of each sample; no backward fuzzy processing is performed on the last time slice of each sample deal with.

上述技术方案中,训练阶段步骤二和步骤三共同完成了视频序列的模糊分片处理,如图2所示,模糊参数α控制了分片之间的边界,使得相邻分片信息得到共享,进一步提高了特征对人体行为时域差异的鲁棒性。In the above technical solution, step 2 and step 3 of the training stage jointly complete the fuzzy slice processing of the video sequence, as shown in Figure 2, the fuzzy parameter α controls the boundary between slices, so that adjacent slice information is shared, The robustness of features to temporal differences in human behavior is further improved.

上述技术方案中,训练阶段步骤四对各个模糊时间片在三个不同投影方向的深度动作图的计算采用视频帧绝对差叠加的方法,具体为:In the above technical scheme, step four of the training phase is for each fuzzy time slice Depth motion maps in three different projection directions The calculation of uses the video frame absolute difference superposition method, specifically:

对同一时间片内的视频帧做三个视觉方向的投影,获得各个视频的主视图、左视图和俯视图,接着对相邻帧的同一视角投影相减,相减后绝对值进行叠加,这样动作的轨迹就被保存了下来,具体公式如下:The video frames in the same time slice are projected in three visual directions to obtain the main view, left view and top view of each video, and then the projections of the same viewing angle of adjacent frames are subtracted, and the absolute values are superimposed after subtraction. The trajectory of is saved, the specific formula is as follows:

其中分别表示第k个样本中的第一个时间片和最后一个时间片,v∈{f,s,t}分别表示视频序列投影的三视图(三个方向),即主视图、左视图和俯视图;表示第k个样本中第i帧深度图的v方向投影;j∈(2,…,DIV-1),α∈(0,1);如图3所示,各投影方向的DMM图有效保存了人体行为序列的单个模糊分片在三个投影方向的轨迹信息。in with respectively represent the first time slice and the last time slice in the k-th sample, and v∈{f,s,t} respectively represent the three views (three directions) of the video sequence projection, that is, the front view, the left view and the top view ; Indicates the v-direction projection of the depth map of the i-th frame in the k-th sample; j∈(2,...,DIV-1), α∈(0,1); as shown in Figure 3, the DMM map of each projection direction is effectively saved Trajectory information of a single fuzzy slice of human behavior sequences in three projection directions is obtained.

上述技术方案中,训练阶段步骤五采用双三次差值法将深度动作图调整为同一尺寸,本专利采用的主视图、左视图和俯视图的尺寸分别定义为:50×25,50×40,40×20;在实际实施中,可以选择不同的插值法实现深度动作图尺寸的缩放,选择前提为尽量减少图像信息的损失。In the above technical scheme, step 5 of the training stage adopts the bicubic difference method to convert the depth action map to Adjusted to the same size, the dimensions of the front view, left view, and top view used in this patent are defined as: 50×25, 50×40, and 40×20; in actual implementation, different interpolation methods can be selected to achieve the size of the depth action map The scaling of , the selection premise is to minimize the loss of image information.

上述技术方案中,训练阶段步骤七对步骤六计算得到的所有样本 的输出特征{H(k)}k∈[1M]通过PCA降维,具体降维后的特征维数可根据训练样本的个数而定,在本专利实施实例中,若训练样本的个数为M,则最终的特征维数为(M-20)×1。In the above technical solution, all samples calculated in step 7 of the training phase against step 6 The output features of {H (k) } k∈[1M] are dimensionally reduced by PCA, and the specific feature dimension after dimensionality reduction can be determined according to the number of training samples. In the implementation example of this patent, if the number of training samples is M, the final feature dimension is (M-20)×1.

上述技术方案中,测试阶段步骤二至步骤六采用的特征构造方法及参数与训练阶段相同。In the above technical solution, the feature construction methods and parameters used in steps 2 to 6 of the test phase are the same as those used in the training phase.

上述技术方案中,测试阶段步骤七采对PCA对步骤六计算得到的测试样本TestX的输出特征HT降维,降维后的维数为(M-20)×1,M为训练阶段样本个数。In the above-mentioned technical scheme, step 7 of the test stage uses PCA to reduce the dimensionality of the output feature H T of the test sample TestX calculated in step 6, and the dimension after dimensionality reduction is (M-20)×1, and M is the number of samples in the training stage number.

上述技术方案中,测试阶段步骤八使用R-ProCRC分类对步骤七获得的降维后的行为特征进行分类的具体方法为:In the above technical solution, step 8 of the test phase uses R-ProCRC classification to classify the dimensionality-reduced behavioral features obtained in step 7 The specific methods for classification are:

(1)计算最优参数 (1) Calculate the optimal parameters

其中是降维后的训练特征,HT为测试样本特征,是属于类别c(c∈C)的所有输入特征向量的集合,‖C‖为总类别数,λ和γ是0到1之间的参数; 其中的构造方法为:首先将初始化为与字典相同尺寸的0矩阵,接着将分配到中,位置为中的相对位置,即可得到例如的值为: 是一个对角化的权重矩阵:in is the training feature after dimensionality reduction, H T is the test sample feature, Is the set of all input feature vectors belonging to category c (c∈C), ‖C‖ is the total number of categories, λ and γ are parameters between 0 and 1; in The construction method is as follows: firstly, the initialized as a dictionary with 0 matrix of the same size, then the Assigned to , the position is exist The relative position in can be obtained E.g The value is: is a diagonalized weight matrix:

这里,表示第i行的所有元素,表示测试样本特征向量的第i个值;here, Represents all elements of the i-th row, Indicates the i-th value of the test sample feature vector;

(2)估计测试样本输出特征属于类别c的概率(2) Estimate the output characteristics of the test sample Probability of belonging to class c

测试样本输出特征属于类别c的概率估计方法如下:Test sample output features The probability of belonging to category c is estimated as follows:

对于所有样本来说都是相同的,因此上式可以简化为:for all samples are the same, so the above formula can be simplified as:

由此可以得到特征所属的类别。From this we can get the characteristics category to which it belongs.

为验证本发明的有效性,本发明在著名的人体行为深度信息数据库MSRAction3D、Action Pair上先后进行了实验。表1给出了两种人体行为深度信息数据库的特性。In order to verify the validity of the present invention, the present invention has carried out experiments successively on the famous human body behavior depth information database MSRAction3D and Action Pair. Table 1 shows the characteristics of the two human behavior depth information databases.

表1:深度数据库特性描述Table 1: Depth Database Feature Description

如表2所示,在实验中我们将MSR Action3D数据库分成三个固定的子集,。每个子集都使用3种实验方式,Test One将每个subject的第一次演示行为作为训练集,剩余的作为测试集。Test Two将每个subject的前两次演示行为作为训练集,剩余的作为测试集。Cross Test使用1,3,5,7,9这些subject的所有视频序列作为训练集,其余用做测试。实验结果见表3,可见本发明的行为识别精度在大多数情况下都优于传统的DMM方法。As shown in Table 2, we divide the MSR Action3D database into three fixed subsets in the experiments. Each subset uses 3 experimental methods, Test One uses the first demonstration behavior of each subject as the training set, and the rest as the test set. Test Two uses the first two demonstration behaviors of each subject as the training set, and the rest as the test set. Cross Test uses all video sequences of 1, 3, 5, 7, and 9 subjects as the training set, and the rest are used for testing. The experimental results are shown in Table 3. It can be seen that the behavior recognition accuracy of the present invention is better than the traditional DMM method in most cases.

表2:MSR Action3D数据库子集Table 2: MSR Action3D database subset

表3:MSR Action3D数据库子集对比Table 3: Comparison of MSR Action3D database subsets

表4显示了本发明在Action Pair数据库上的识别率以及与DMM的对比。由于Action Pair数据库行为都是存在大量动作时序相反的行为,比如“捡起”和“放下”,“起立”和“坐下”等,因此对时间信息十分敏感。传统的DMM只有50.6%的识别率,而本发明的识别率达到了97.2%。Table 4 shows the recognition rate of the present invention on the Action Pair database and the comparison with DMM. Since the actions in the Action Pair database have a large number of actions with opposite timing, such as "pick up" and "put down", "stand up" and "sit down", they are very sensitive to time information. The recognition rate of traditional DMM is only 50.6%, but the recognition rate of the present invention reaches 97.2%.

表4:Action Pair不同算法识别率Table 4: Action Pair recognition rate of different algorithms

由于本发明采用深度数据来进行人体行为识别,相比于传统的彩色视频数据,深度数据能够快速精准地完成人体分隔,保存人体的形状和结构特征,利于提高精度,同时使用深度动作图DMM的处理方式,相对于基于骨骼跟踪技术估计获得的骨骼点有更好的鲁棒性;传统的利用深度动作图DMM的特征提取方法将整个视频帧投影到一张DMM图上,丢失时间了信息;本方法提出的模糊边界分片的深度动作图的人体行为识别,将现有的DMM进行分片,并使用模糊参数α来控制分片之间的边界,使得相邻分片信息得到共享,也使得DMM能够更好的捕获时间信息,配合使用R-ProCRC[5]分类器可以获得卓越的识别准确率。Since the present invention uses depth data for human body behavior recognition, compared with traditional color video data, depth data can quickly and accurately complete human body separation, preserve the shape and structural features of the human body, and help improve accuracy. The processing method is more robust than the bone points estimated based on bone tracking technology; the traditional feature extraction method using the depth motion map DMM projects the entire video frame onto a DMM map, which loses time information; The human behavior recognition of the depth action map of the fuzzy boundary segmentation proposed by this method divides the existing DMM into slices, and uses the fuzzy parameter α to control the boundaries between slices, so that the information of adjacent slices can be shared, and the It enables DMM to better capture time information, and the R-ProCRC [5] classifier can be used together to obtain excellent recognition accuracy.

上面结合附图对本发明的具体实施方式做了详细说明,但是本发明并不限于上述实施方式,在本领域普通技术人员所具备的知识范围内,还可以在不脱离本发明宗旨的前提下做出各种变化。The specific implementation of the present invention has been described in detail above in conjunction with the accompanying drawings, but the present invention is not limited to the above-mentioned implementation, and within the knowledge of those of ordinary skill in the art, it can also be done without departing from the gist of the present invention. Various changes.

参考文献references

[1].Bian W,Tao D,Rui Y.Cross-domain human action recognition.[J].IEEETransactions on Systems Man&Cybernetics Part B Cybernetics A Publication ofthe IEEE Systems Man&Cybernetics Society,2012,42(2):298-307.[1]. Bian W, Tao D, Rui Y. Cross-domain human action recognition. [J]. IEEE Transactions on Systems Man&Cybernetics Part B Cybernetics A Publication of the IEEE Systems Man&Cybernetics Society, 2012,42(2):298-307.

[2].Niebles J C,Wang H,Li F F.Unsupervised Learning of Human ActionCategories Using Spatial-Temporal Words[J].International Journal of ComputerVision,2008,79(3):299-318.[2].Niebles J C,Wang H,Li F F.Unsupervised Learning of Human Action Categories Using Spatial-Temporal Words[J].International Journal of Computer Vision,2008,79(3):299-318.

[3].Wang J,Liu Z,Wu Y,et al.Mining actionlet ensemble for actionrecognition with depth cameras[C]//IEEE Conference on Computer Vision andPattern Recognition.IEEE Computer Society,2012:1290-1297.[3].Wang J, Liu Z, Wu Y, et al.Mining actionlet ensemble for actionrecognition with depth cameras[C]//IEEE Conference on Computer Vision and Pattern Recognition.IEEE Computer Society,2012:1290-1297.

[4].Chen C,Liu K,Kehtarnavaz N.Real-time human action recognitionbased on depth motion maps[J].Journal of Real-Time Image Processing,2013:1-9.[4]. Chen C, Liu K, Kehtarnavaz N. Real-time human action recognition based on depth motion maps [J]. Journal of Real-Time Image Processing, 2013: 1-9.

[5].Sijia C,Lei Z,et al.A Probabilistic Collaborative Representationbased Approach for Pattern Classification.IEEE Trans.on Pattern Analysis andMachine Intelligence,2016.[5]. Sijia C, Lei Z, et al. A Probabilistic Collaborative Representation based Approach for Pattern Classification. IEEE Trans. on Pattern Analysis and Machine Intelligence, 2016.

[6].Xu H,Chen E,Liang C,et al.Spatio-Temporal Pyramid Model based ondepth maps for action recognition[C]//Mmsp2015 IEEE,International Workshop onMultimedia Signal Processing.2015。[6]. Xu H, Chen E, Liang C, et al.Spatio-Temporal Pyramid Model based ondepth maps for action recognition[C]//Mmsp2015 IEEE, International Workshop on Multimedia Signal Processing.2015.

Claims (7)

1.基于模糊边界分片的深度动作图的人体行为识别,其特征在于,包括训练阶段和测试阶段。1. The human action recognition based on the depth action graph of fuzzy boundary slice, is characterized in that, comprises training stage and testing stage. 2.根据权利要求1所述的基于模糊边界分片的深度动作图的人体行为识别,其特征在于,训练阶段包括下列步骤:2. the human body behavior recognition based on the depth action map of fuzzy boundary slice according to claim 1, is characterized in that, training stage comprises the following steps: 步骤一、给定人体行为视频序列的深度图样本的训练集合 其中X(k)表示第k个训练样本的深度图序列, 为第k个样本中第i帧的原始深度图像,Nk为第k个样本的总帧数;Y(k)表示第k个训练样本所属行为类别;M表示训练集中样本个数;Step 1. Given the training set of depth map samples of human behavior video sequences where X (k) represents the depth map sequence of the kth training sample, Be the original depth image of the i-th frame in the k-th sample, N k is the total number of frames of the k-th sample; Y (k) represents the behavior category to which the k-th training sample belongs; M represents the number of samples in the training set; 步骤二、将训练集中每个视频序列训练样本X(k)按时间轴直接划分为DIV个等长时间片,每个时间片长度为划分后的时间片表示为其中 Step 2, each video sequence training sample X (k) in the training set is directly divided into DIV equal time slices according to the time axis, and the length of each time slice is The divided time slice is expressed as in 步骤三、选择合适的模糊参数α,对每个时间片进行分片模糊处理;模糊时间片表示为为避免下标溢出,对于第一个时间片不做前向模糊处理,对最后一个时间片不做后向模糊处理;Step 3: Select the appropriate fuzzy parameter α, and perform slice fuzzy processing on each time slice; the fuzzy time slice is expressed as To avoid subscript overflow, forward blurring is not performed for the first time slice, and backward blurring is not performed for the last time slice; 步骤四、计算每个模糊时间片在三个不同投影方向的深度动作图其中v∈{f,s,t}分别表示视频序列投影的三视图(三个方向),主视图、左视图和俯视图;至此,计算获得所有训练样本对应的深度动作图集合 Step 4. Calculate each fuzzy time slice Depth motion maps in three different projection directions where v ∈ {f, s, t} respectively represent the three views (three directions) of the video sequence projection, the main view, the left view and the top view; so far, all training samples are calculated Corresponding set of deep motion maps 步骤五、利用双三次差值法将步骤四求解获得的深度动作图调整为一个相同的尺寸,并将这些深度动作图归一化为0到1之间,归一化后的记为 Step 5. Use the bicubic difference method to solve the depth action map obtained in step 4 Adjust to the same size, and normalize these depth action maps between 0 and 1, the normalized recorded as 步骤六、将任一训练样本X(k)对应深度动作图集合进行向量化,并将所有向量化后的序列动作图串联,完成训练样本X(k)对应的特征构建,该特征记为H(k),则所有样本的特征集合表示为{H(k)}k∈[1 M]Step 6. Correspond any training sample X (k) to the set of depth motion maps Carry out vectorization, and concatenate all vectorized sequence action graphs to complete the feature construction corresponding to the training sample X (k) , which is recorded as H (k) , and the feature set of all samples is expressed as {H (k) } k∈[ 1M] ; 步骤七、将步骤六得到的所有样本的输出特征{H(k)}k∈[1 M]通过PCA降维,并保存所有降维后的特征 Step 7, all the samples obtained in step 6 The output features of {H (k) } k∈[1 M] are dimensionally reduced by PCA, and all features after dimensionality reduction are saved 3.根据权利要求1所述的基于模糊边界分片的深度动作图的人体行为识别,其特征在于,测试阶段包括下列步骤:3. the human body behavior recognition based on the depth action map of fuzzy boundary slice according to claim 1, is characterized in that, testing stage comprises the following steps: 步骤一、给定一个人体行为视频序列的深度图样本的测试样本(TestX,TestY),其中TestX表示测试样本的深度图序列, Xi为测试样本中第i帧的原始深度图像,NT为该测试样本的总帧数;TestY表示测试样本所属行为类别;Step 1, given a test sample (TestX, TestY) of the depth map sample of a human behavior video sequence, wherein TestX represents the depth map sequence of the test sample, X i is the original depth image of the i-th frame in the test sample, NT is the total number of frames of the test sample; TestY indicates the behavior category to which the test sample belongs; 步骤二、将测试样本TestX按时间轴直接划分为DIV个等长时间片(分片方式与训练阶段相同),每个时间片长度为划分后的时间片表示为其中 Step 2. Divide the test sample TestX directly into DIV equal time slices according to the time axis (the slice method is the same as the training stage), and the length of each time slice is The divided time slice is expressed as in 步骤三、根据训练阶段采用的模糊参数α,对每个时间片进行分片模糊处理;模糊时间片记为为避免下标溢出,对于第一个时间片不做前向模糊处理,对最后一个时间片不做后向模糊处理;;Step 3. According to the fuzzy parameter α used in the training stage, perform fragmentation fuzzy processing on each time slice; the fuzzy time slice is denoted as To avoid subscript overflow, no forward blurring is done for the first time slice, and no backward blurring is done for the last time slice; 步骤四、计算每个模糊时间片在三个不同投影方向的深度动作图DMMj,v,其中v∈{f,s,t}分别表示视频序列投影的三视图(三个方向),主视图、左视图和俯视图;至此,计算获得测试样本TestX对应的深度动作图集合{DMMj,v}j∈[1 DIV],v∈[f,s,t]Step 4. Calculate each fuzzy time slice Depth motion map DMM j,v in three different projection directions, where v∈{f,s,t} respectively represent three views (three directions) of video sequence projection, front view, left view and top view; so far, calculate Obtain the depth motion map set {DMM j, v } j∈[1 DIV], v∈[f, s, t] corresponding to the test sample TestX; 步骤五、利用双三次差值法将步骤四求解获得的测试样本深度动作图{DMMj,v}j∈[1 DIV],v∈[f,s,t]调整为训练阶段同一尺寸,并根据训练阶段归一化方法,将这些深度动作图归一化为0到1之间,归一化后的DMMj,v记为 Step 5. Use the bicubic difference method to adjust the test sample depth motion map {DMM j,v } j∈[1 DIV],v∈[f,s,t] obtained in step 4 to the same size as the training stage, and According to the normalization method of the training stage, these depth action maps are normalized between 0 and 1, and the normalized DMM j,v is recorded as 步骤六、将测试样本TestX对应的深度动作图集合进行向量化,并将所有向量化后的序列动作图串联,完成测试样本TestX对应的特征构建,该特征记为HTStep 6: Collect the depth action images corresponding to the test sample TestX Carry out vectorization and concatenate all vectorized sequence action graphs to complete the feature construction corresponding to the test sample TestX, which is denoted as H T ; 步骤七、将步骤六得到的测试样本TestX的输出特征HT通过PCA降维,得到降维后的特征PCA降维方法与训练阶段一致;Step 7. The output feature H T of the test sample TestX obtained in step 6 is reduced by PCA to obtain the feature after dimension reduction The PCA dimensionality reduction method is consistent with the training phase; 步骤八、然后将降维后的输出特征送入R-ProCRC分类器,获得分类输出PridY;Step 8, and then the output features after dimensionality reduction Send it to the R-ProCRC classifier to obtain the classification output PridY; 步骤九、比较PridY和TestY,若PridY=TestY,则识别正确,否则识别错误。Step 9: Comparing PridY and TestY, if PridY=TestY, the recognition is correct, otherwise the recognition is wrong. 4.根据权利要求2所述的训练阶段步骤二对视频序列进行等长时间片划分的方法中,每个时间片长度为时间片长度采用向下取整方式;最后一个时间片若长度不足则按实际长度选取。4. in the method that step 2 of training stage according to claim 2 is carried out to video sequence in the method for equal time slice division, each time slice length is The length of the time slice is rounded down; if the length of the last time slice is insufficient Then choose according to the actual length. 5.根据权利要求2所述的训练阶段步骤三对时间片的分片模糊处理中,对每个样本的第一个时间片不做前向模糊处理;对每个样本的最后一个时间片不做后向模糊处理。5. in the fragmentation fuzzy processing of time slice according to training stage step 3 according to claim 2, the first time slice of each sample is not done forward fuzzy processing; The last time slice of each sample is not Do backward blurring. 6.根据权利要求2所述的训练阶段步骤四对各个模糊时间片在三个不同投影方向的深度动作图的计算采用视频帧绝对差叠加的方法,具体为:对同一时间片内的视频帧做三个视觉方向的投影,获得各个视频的主视图、左视图和俯视图,接着对相邻帧的同一视角投影相减,相减后绝对值进行叠加,这样动作的轨迹就被保存了下来,具体公式如下:6. training phase step four according to claim 2 to each fuzzy time slice Depth motion maps in three different projection directions The method of superimposing the absolute difference of video frames is used for the calculation of , specifically: projection in three visual directions on the video frames in the same time slice to obtain the main view, left view and top view of each video, and then the same viewing angle of adjacent frames The projection is subtracted, and the absolute value is superimposed after the subtraction, so that the trajectory of the action is saved. The specific formula is as follows: 其中分别表示第k个样本中的第一个时间片和最后一个时间片,v∈{f,s,t}分别表示视频序列投影的三视图(三个方向),即主视图、左视图和俯视图;表示第k个样本中第i帧深度图的v方向投影;j∈(2,…,DIV-1),α∈(0,1);各投影方向的DMM图有效保存了人体行为序列的单个模糊分片在三个投影方向的轨迹信息。in with respectively represent the first time slice and the last time slice in the k-th sample, and v∈{f,s,t} respectively represent the three views (three directions) of the video sequence projection, that is, the front view, the left view and the top view ; Represents the v-direction projection of the i-th frame depth map in the k-th sample; j∈(2,...,DIV-1), α∈(0,1); the DMM map of each projection direction effectively preserves a single human behavior sequence Trajectory information of fuzzy slices in three projection directions. 7.根据权利要求3所述的测试阶段步骤八使用R-ProCRC分类对测试步骤七获得的降维后的行为特征进行分类的具体方法为:7. according to claim 3, test phase step eight uses R-ProCRC classification to the behavioral characteristics after the dimensionality reduction that test step seven obtains The specific methods for classification are: (1)计算最优参数 (1) Calculate the optimal parameters 其中是降维后的训练特征,HT为测试样本特征,是属于类别c(c∈C)的所有输入特征向量的集合,‖C‖为总类别数,λ和γ是0到1之间的参数;其中的构造方法为:首先将初始化为与字典相同尺寸的0矩阵,接着将分配到中,位置为中的相对位置,即可得到例如的值为: 是一个对角化的权重矩阵:in is the training feature after dimensionality reduction, H T is the test sample feature, Is the set of all input feature vectors belonging to category c (c∈C), ‖C‖ is the total number of categories, λ and γ are parameters between 0 and 1; in The construction method is as follows: firstly, the initialized as a dictionary with 0 matrix of the same size, then the Assigned to , the position is exist The relative position in can be obtained E.g The value is: is a diagonalized weight matrix: 这里,表示第i行的所有元素,表示测试样本特征向量的第i个值;here, Represents all elements of the i-th row, Indicates the i-th value of the test sample feature vector; (2)估计测试样本输出特征属于类别c的概率(2) Estimate the output characteristics of the test sample Probability of belonging to class c 测试样本输出特征属于类别c的概率估计方法如下:Test sample output features The probability of belonging to category c is estimated as follows: 对于所有样本来说都是相同的,因此上式可以简化为:for all samples are the same, so the above formula can be simplified as: 由此可以得到特征所属的类别。From this we can get the characteristics category to which it belongs.
CN201610949051.8A 2016-10-26 2016-10-26 Depth motion figure Human bodys' response method based on smeared out boundary fragment Active CN106529441B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610949051.8A CN106529441B (en) 2016-10-26 2016-10-26 Depth motion figure Human bodys' response method based on smeared out boundary fragment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610949051.8A CN106529441B (en) 2016-10-26 2016-10-26 Depth motion figure Human bodys' response method based on smeared out boundary fragment

Publications (2)

Publication Number Publication Date
CN106529441A true CN106529441A (en) 2017-03-22
CN106529441B CN106529441B (en) 2019-04-05

Family

ID=58292902

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610949051.8A Active CN106529441B (en) 2016-10-26 2016-10-26 Depth motion figure Human bodys' response method based on smeared out boundary fragment

Country Status (1)

Country Link
CN (1) CN106529441B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108038420A (en) * 2017-11-21 2018-05-15 华中科技大学 A kind of Human bodys' response method based on deep video
CN108564043A (en) * 2018-04-17 2018-09-21 中国民航大学 A kind of Human bodys' response method based on time-space distribution graph
CN108573231A (en) * 2018-04-17 2018-09-25 中国民航大学 Human action recognition method based on deep motion map generated from motion history point cloud
CN109670401A (en) * 2018-11-15 2019-04-23 天津大学 A kind of action identification method based on skeleton motion figure
CN111950320A (en) * 2019-05-14 2020-11-17 四川大学 An Iris Recognition Method Based on Relative Total Variation and Probabilistic Co-Representation

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100034462A1 (en) * 2008-06-16 2010-02-11 University Of Southern California Automated Single Viewpoint Human Action Recognition by Matching Linked Sequences of Key Poses
CN104598890A (en) * 2015-01-30 2015-05-06 南京邮电大学 Human body behavior recognizing method based on RGB-D video
CN105912991A (en) * 2016-04-05 2016-08-31 湖南大学 Behavior identification method based on 3D point cloud and key bone nodes

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100034462A1 (en) * 2008-06-16 2010-02-11 University Of Southern California Automated Single Viewpoint Human Action Recognition by Matching Linked Sequences of Key Poses
CN104598890A (en) * 2015-01-30 2015-05-06 南京邮电大学 Human body behavior recognizing method based on RGB-D video
CN105912991A (en) * 2016-04-05 2016-08-31 湖南大学 Behavior identification method based on 3D point cloud and key bone nodes

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
CHEN CHEN ; ROOZBEH JAFARI ; NASSER KEHTARNAVAZ: "UTD-MHAD: A multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor", 《2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING》 *
CHRISTIAN SCHULDT;IVAN LAPTEV;BARBARA CAPUTO: "Recognizing Human Actions: A Local SVM Approach", 《ICPR "04 PROCEEDINGS OF THE PATTERN RECOGNITION, 17TH INTERNATIONAL CONFERENCE ON (ICPR"04) VOLUME 3》 *
VENNILA MEGAVANNAN ; BHUVNESH AGARWAL ; R. VENKATESH BABU: "Human action recognition using depth maps", 《2012 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108038420A (en) * 2017-11-21 2018-05-15 华中科技大学 A kind of Human bodys' response method based on deep video
CN108564043A (en) * 2018-04-17 2018-09-21 中国民航大学 A kind of Human bodys' response method based on time-space distribution graph
CN108573231A (en) * 2018-04-17 2018-09-25 中国民航大学 Human action recognition method based on deep motion map generated from motion history point cloud
CN108564043B (en) * 2018-04-17 2021-08-31 中国民航大学 A Human Behavior Recognition Method Based on Spatio-temporal Distribution Map
CN108573231B (en) * 2018-04-17 2021-08-31 中国民航大学 Human body behavior identification method of depth motion map generated based on motion history point cloud
CN109670401A (en) * 2018-11-15 2019-04-23 天津大学 A kind of action identification method based on skeleton motion figure
CN109670401B (en) * 2018-11-15 2022-09-20 天津大学 Action recognition method based on skeletal motion diagram
CN111950320A (en) * 2019-05-14 2020-11-17 四川大学 An Iris Recognition Method Based on Relative Total Variation and Probabilistic Co-Representation

Also Published As

Publication number Publication date
CN106529441B (en) 2019-04-05

Similar Documents

Publication Publication Date Title
Zheng et al. Deep learning for event-based vision: A comprehensive survey and benchmarks
CN111311666B (en) Monocular vision odometer method integrating edge features and deep learning
Li et al. In ictu oculi: Exposing ai created fake videos by detecting eye blinking
CN110490928B (en) Camera attitude estimation method based on deep neural network
CN109284738B (en) Irregular face correction method and system
CN108062525B (en) A deep learning hand detection method based on hand region prediction
WO2020108362A1 (en) Body posture detection method, apparatus and device, and storage medium
CN105069472B (en) A kind of vehicle checking method adaptive based on convolutional neural networks
CN103295242B (en) A kind of method for tracking target of multiple features combining rarefaction representation
CN109461172A (en) Manually with the united correlation filtering video adaptive tracking method of depth characteristic
CN104484890B (en) Video target tracking method based on compound sparse model
CN104574445A (en) Target tracking method and device
CN111901532B (en) Video stabilization method based on recurrent neural network iteration strategy
CN106529441A (en) Fuzzy boundary fragmentation-based depth motion map human body action recognition method
CN109146911A (en) A kind of method and device of target following
CN111428664A (en) Real-time multi-person posture estimation method based on artificial intelligence deep learning technology for computer vision
CN112801019B (en) Method and system for eliminating unsupervised vehicle re-identification bias based on synthetic data
Min et al. COEB-SLAM: A robust VSLAM in dynamic environments combined object detection, epipolar geometry constraint, and blur filtering
CN110322479B (en) Dual-core KCF target tracking method based on space-time significance
CN112329662A (en) Multi-view saliency estimation method based on unsupervised learning
CN112069943A (en) Online multi-person pose estimation and tracking method based on top-down framework
CN107292908A (en) Pedestrian tracting method based on KLT feature point tracking algorithms
Feng et al. Efficient deep learning for stereo matching with larger image patches
CN105930793A (en) Human body detection method based on SAE characteristic visual learning
CN106778576A (en) A kind of action identification method based on SEHM feature graphic sequences

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20221114

Address after: 201613 Room 302, 3 building, 299 Zhong Chen Road, Songjiang District, Shanghai.

Patentee after: HUIRONG ELECTRONIC SYSTEM ENGINEERING LTD.

Address before: No. 1800 road 214122 Jiangsu Lihu Binhu District City of Wuxi Province

Patentee before: Jiangnan University

PE01 Entry into force of the registration of the contract for pledge of patent right
PE01 Entry into force of the registration of the contract for pledge of patent right

Denomination of invention: Human behavior recognition method based on fuzzy boundary segmentation in deep action maps

Effective date of registration: 20230705

Granted publication date: 20190405

Pledgee: Agricultural Bank of China Limited Shanghai Songjiang Sub-branch

Pledgor: HUIRONG ELECTRONIC SYSTEM ENGINEERING LTD.

Registration number: Y2023980047238