[go: up one dir, main page]

CN103413145A - Articulation point positioning method based on depth image - Google Patents

Articulation point positioning method based on depth image Download PDF

Info

Publication number
CN103413145A
CN103413145A CN2013103742367A CN201310374236A CN103413145A CN 103413145 A CN103413145 A CN 103413145A CN 2013103742367 A CN2013103742367 A CN 2013103742367A CN 201310374236 A CN201310374236 A CN 201310374236A CN 103413145 A CN103413145 A CN 103413145A
Authority
CN
China
Prior art keywords
point
feature
node
depth
joint
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2013103742367A
Other languages
Chinese (zh)
Other versions
CN103413145B (en
Inventor
刘亚洲
张艳
孙权森
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Science and Technology
Original Assignee
Nanjing University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Science and Technology filed Critical Nanjing University of Science and Technology
Priority to CN201310374236.7A priority Critical patent/CN103413145B/en
Publication of CN103413145A publication Critical patent/CN103413145A/en
Application granted granted Critical
Publication of CN103413145B publication Critical patent/CN103413145B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses an articulation point positioning method based on a depth image. The method comprises a training process and a recognition process. The training process comprises the first step of calculating the random characters of a training sample, and the second step of training a decision tree classifier according to the characters. The recognition process comprises the third step of calculating the random characters of a test sample, the fourth step of classifying the pixel points of an object according to the decision tree classifier to obtain the portions of different classes of the object, and the fifth step of calculating the position of the articulation point of each portion. The method can reflect the local gradient information around pixels, is high in calculation efficiency, enhances the rotation invariance of the characters, and improves the accuracy of object recognition.

Description

Articulation point localization method based on depth image
Technical field
The present invention relates to computer vision, pattern-recognition and field of human-computer interaction, more particularly, relate to a kind of localization method of articulation point based on depth image.
Background technology
Articulation point localization method based on depth image refers to, comprises in the depth image of target at a width, determines the method for the articulation point position of target.The target here specifically refers to staff or human body.By determining the articulation point position of target, can judge skeleton structure, and then computing machine can realize the response for target, finally realize the purpose of man-machine interaction or the automatic processing and identification of computing machine.
Depth image take the form of a kind of two dimensional gray figure.But be different from traditional gray level image, on each pixel of depth image with message reflection be the distance of target object apart from video camera, so the pixel value of depth image is called depth value.Depth image has the following advantages: 1, can not run into the impact of the factors such as illumination, shade; 2, use depth image can directly utilize the three-dimensional information of target object, simplified greatly the problems such as three-dimensional reconstruction, identification, location of object.
The articulation point location comprises two committed steps: the study of sorter and the location of articulation point.At first the study of sorter depend on feature selecting, and for current goal, whether the feature of selection has very strong descriptive power, directly determined the success or not of target identification; Then, on the basis of fixed feature, determine a series of rules that can classify to current goal.The definite of articulation point refers to, after the sorter that utilizes study to obtain completes the Classification and Identification of target, at each position of target, finds the position of articulation point.
In the feature extraction of traditional visible images, Gradient Features is the common features of two large classes with putting feature.Gradient Features is such as the Canny operator, Laplace-Gaussian operator and histograms of oriented gradients HOG etc.For the first two operator, can reasonablely detect the point of all edges in image, but these two kinds of methods probably are divided into image several disconnected region units.HOG is method very classical in human detection and identification, and its advantage is that processing accuracy is high, detects effective.But it is high that shortcoming is dimension, and computing cost is large, therefore process in real time and be difficult to guarantee.On the other hand, common some feature such as angle point, round dot etc., although dimension is not high, but be difficult to adapt to the changeable form of human body in the situation that background is more in disorder, and the some feature also need to carry out the operations such as cluster, strengthened the difficulty of dealing with problems, caused the problem that Detection accuracy is low.Therefore, simple employing Gradient Features or some feature are not good solutions.
Summary of the invention
The technical problem to be solved in the present invention is, in above-mentioned target identification technology, the problem that real-time is poor or accuracy rate is low of the target identification that the employing single features causes as basis of characterization, propose a kind of method of extracting the random character of depth data and carried out training classifier, finally completed the method for articulation point location.
The technical solution that realizes the object of the invention is: the method comprises trains and identifies two processes,
Training process comprises the following steps:
1) random character of calculation training sample;
2) according to described features training decision tree classifier.
Identifying comprises the following steps:
3) calculate the random character of test sample book;
4) utilize decision tree classifier to classify to each pixel of target, obtain the different classes of position of target;
5) calculate the position of the articulation point at each position.
In said method, the training sample in described step 1) refers to the depth image that only retains target and mark through true value.
In said method, described step 1) comprises following concrete steps:
11) employing formula (1) is calculated the centre of form c (cx, cy) of target:
cx = 1 k Σ i = 1 k x i cy = 1 k Σ i = 1 k y i - - - ( 1 )
Wherein, k means the sum of pixel on target, (x i, y i) mean the coordinate of each pixel on target, i=1,2 ... k;
12) take mark point is starting point, generates with two different reference point that random vector is pointed, and wherein, the length rz=r1* α/valz of random vector, r1 are the random length generated, and α is coefficient, and valz is the depth value of starting point; Angle beta=the θ of random vector+ο, θ are the angles of starting point and straight line that the centre of form connects and horizontal axis, and ο is the random angle generated; If two reference point have one at least not on image, the value of the feature of starting point is 1; Otherwise calculate the depth difference of two reference point: if depth difference is greater than at self-defined threshold set
Figure BDA0000371471660000033
In optional one, the value of the feature of starting point is 1, otherwise the value of the feature of starting point is 0;
13) for each mark point, repeating step 12) fn feature of fn generation, the inferior ordered pair feature generated according to feature is simultaneously carried out the numbering of 1~fn.
In said method, described step 2) comprise following concrete steps:
21) using the root node of decision tree classifier as present node;
22) calculate the information gain of each feature of present node:
Gain ( ϵ ) = entroy ( T ) - Σ i = 1 m T i T entroy ( T i ) ,
Wherein, the numbering of ε representation feature, ε=1,2 ... fn, T mean that present node namely marks a sample set, T iMean the subset of sample set, m means the subset number, according to the value that is numbered the feature of ε, is 0 or 1 here, and will mark and a little be divided into two subsets is m=2, and entroy (T) means the information entropy of sample set, entroy ( T ) = - Σ j = 1 s p ( C j , T ) log 2 p ( C j , T ) , P(C j, T) mean to belong to classification C in sample set T jFrequency, s means the number of classification in T;
23) will have the numbering of feature of maximum information gain as the numbering of present node;
24) if it is 0 that the mark point is numbered the value of the feature of present node numbering, this mark point is divided into to the left branch node of present node, otherwise is divided into the right branch node;
25) using branch node as present node, if the information entropy of present node is less than the threshold value h τ of entropy, perhaps the number of plies of decision tree reaches maximum number of plies depth, perhaps the mark point number of present node is less than sample point minimal amount small, stop division, using present node as leaf node, otherwise repeating step 22)~25);
26) the mark point category distribution of leaf node is carried out to probability statistics.
In said method, in described step 3), test sample book is to have removed background, only retains the depth image of target.
In said method, described step 3) comprises following concrete steps:
31) employing formula (1) is calculated the centre of form of target;
The point of 32) take on target is starting point, generates with two different reference point that random vector is pointed, and the length rz=r1* α/valz of random vector wherein, r1 is the random length generated, and α is coefficient, and valz is the depth value of starting point; Angle beta=the θ of random vector+ο, θ are the angles of starting point and straight line that the centre of form connects and horizontal axis, and ο is the random angle generated; Two reference point have one at least not on image, and the value of the feature of starting point is 1; Otherwise calculate the depth difference of two reference point, if depth difference is greater than at self-defined threshold set
Figure BDA0000371471660000041
In optional one, the value of the feature of starting point is 1, otherwise the value of the feature of starting point is 0.
33) for each mark point, repeating step 32) fn feature of fn generation, the inferior ordered pair feature generated according to feature is simultaneously carried out the numbering of 1~fn.
In said method, described step 4) comprises following concrete steps:
41) using the root node of decision tree as present node;
42) if the value of the feature that is numbered the present node numbering of pixel is 0, is divided into the left branch node of present node, otherwise is divided into the right branch node;
43) branch node pixel is divided into is as present node, repeating step 42), 43), until pixel arrives leaf node;
44), if the maximum probability of all categories of leaf node is greater than probability threshold value p τ, judges classification with maximum probability classification as pixel, otherwise give up this pixel.
In said method, described step 5) comprises following concrete steps:
51) from identical category each the some q iSet out, find respectively corresponding articulation point position candidate p i, i=1,2 ... r, wherein, r is the sum of the point of identical category.
52) institute's related node position candidate is screened, find the position of articulation point.
In said method, described step 51) comprise following concrete steps:
511) with q iCentered by point, generating yardstick is the rectangular characteristic zone of w * h;
512) in employing formula (1) calculated characteristics zone with the centre of form of the generic point of central point;
513) distance of computing center's point and the centre of form;
514) if distance is not more than distance threshold d τ, using the centre of form as the articulation point position candidate, otherwise generate point centered by the centre of form, yardstick is the rectangular characteristic zone of w * h, repeating step 512)~514), if repeat abundant number of times, for example also do not find the articulation point position candidate 30 times, the centre of form that will obtain for the last time is as articulation point position candidate p i.
In said method, described step 52) comprise following concrete steps:
521) with p 1As the initial score object, according to p i, i=2,3 ..., the order of r, repeat next step;
522), according to the sequencing that becomes the score object, calculate each score object and p iDistance, as score object and a p iDistance be less than threshold value dis τ, the mark of this score object adds 1, no longer score object and p of calculated for subsequent iDistance.If all score objects and p iThe distance all be not less than dis τ, by p iAs next one score object;
523) select the score object with highest score tops, if tops is greater than score threshold sco τ, should score to liking the articulation point position; Otherwise reduce sco τ, until find the articulation point position.
The present invention compared with prior art, its remarkable advantage: the present invention utilizes the character of depth image, proposed to adopt the random character of the pixel depth difference of random 2 on every side as pixel, pixel local gradient information on every side be can reflect, a feature and Gradient Features good combination can be regarded as.This feature only relates to the simple arithmetic operations of pixel value, and counting yield is high, for processing advantage is provided in real time.In addition, in the random angle of this random character, add the deviation angle of the pixel of target for the target centre of form, strengthened the rotational invariance of feature, improved the accuracy rate of target identification.
The accompanying drawing explanation
Fig. 1 is based on the articulation point localization method process flow diagram of depth image.
Fig. 2 is the staff schematic diagram of mark true value.
Fig. 3 is the schematic diagram of generating reference point.
Fig. 4 is the schematic diagram that adopts decision tree classifier to classify to pixel.
Fig. 5 is the sorted position of staff schematic diagram.
Fig. 6 is the schematic diagram of staff articulation point position.
Embodiment
Integrated operation flow process of the present invention as shown in Figure 1.Below in conjunction with accompanying drawing, the specific embodiment of the present invention is described in further detail.The depth image of data source of the present invention for obtaining from image capture device, image capture device can be binocular vision device or structured light projecting device.This object point of the value representation of each pixel on depth image is to the distance of camera projection centre.By depth image, can obtain shape information and the three dimensional local information of object.
The present invention adopts and has removed background, and the depth map that only retains the staff target is made sample.On this basis, the depth map of process true value mark is as training sample, and the depth map of process true value mark is not as test sample book.11 articulation points of staff mark are respectively: the binding site of palm and wrist, and the finger tip of 5 fingers and finger root, 0~10 means the label of the mark of these 11 articulation points, to mean different classifications, each articulation point marks the point of sufficient amount.As shown in Figure 2.
Articulation point localization method of the present invention comprises that training classifier and application class device carry out two key steps of target identification.
The process of training classifier refers to by known training sample, and the process of the classifying rules of target is determined in study, comprises the random character of calculation training sample and according to two processes of described features training decision tree classifier.
Step 1: the random character of calculation training sample.
Step 11: the angle that calculates the centre of form, the every bit on target and straight line that the centre of form connects and the horizontal axis of target.
Employing formula (1) is calculated the centre of form c (cx, cy) of target:
cx = 1 k Σ i = 1 k x i cy = 1 k Σ i = 1 k y i - - - ( 1 )
Wherein, k means the sum of pixel on target, (x i, y i) mean the coordinate of each pixel on target, i=1,2 ... k.
Employing formula (2) is calculated the every bit p (x, y) of target and the angle that c (cx, cy) connects straight line and horizontal axis:
θ=arctan[(y-cy)/(x-cx)] (2)
Step 12: read the articulation point markup information of target, for each mark point l (x, y), number of computations is the feature of fn, can make that the fn span is 1000~10000, and wherein the computation process of each feature is as follows:
Step 121: the l (x, y) of take is starting point, generates at random two reference point.
The arbitrary reference point generated is designated as q (qx, qy), and q (qx, qy) take l (x, y) be starting point, take random-length as rz, angle is the vector of β point pointed at random, namely
qx = x + rz * cos β qy = y + rz * sin β - - - ( 3 )
Wherein, rz=r1* α/valz, r1 are the random length generated, and can make it get the arbitrary value in 0~1 at every turn, and α is coefficient, can make that its span is that 1000~10000, valz is the depth value of l (x, y); β=θ+ο, θ are the angles that l (x, y) and centre of form c (cx, cy) connect straight line and horizontal axis, and employing formula (2) is calculated, and ο is the random angle generated, and can make it get the arbitrary value in 0 °~360 ° at every turn.Fig. 3 is shown in by the reference point schematic diagram.
Step 122: the eigenwert of determining l (x, y) according to the situation of reference point.
If two reference point are all on image, their depth value is respectively d 1And d 2, the poor dd=d of compute depth 1-d 2, the value f of the feature of l (x, y) is determined by following formula:
f = 1 , dd > t 0 , dd < t - - - ( 4 )
Wherein, t is the random threshold value generated, and can be at self-defined threshold set
Figure BDA0000371471660000082
In optional one.Here can establish
Figure BDA0000371471660000083
If two reference point have one at least not on image, f=1.
Step 123: according to the order of calculated characteristics, be the ε time, determine l (x, y) feature be numbered ε.
Step 2: according to the features training decision tree classifier of step 1 extraction.
The corresponding training sample of the root node of decision tree classifier namely marks a sample set, and what branch node was corresponding is the subset of mark point sample set, using root node as present node, carries out following steps:
Step 21: adopt following formula to calculate the information gain of each feature of present node:
Gain ( &epsiv; ) = entroy ( T ) - &Sigma; i = 1 m T i T entroy ( T i ) - - - ( 5 )
Wherein, the numbering of ε representation feature, ε=1,2 ... fn, T mean that present node namely marks a sample set, T iMean the subset of sample set, m means the subset number, according to the value that is numbered the feature of ε, is 0 or 1 here, and will mark and a little be divided into two subsets is m=2, and entroy (T) means the information entropy of sample set, entroy ( T ) = - &Sigma; j = 1 s p ( C j , T ) log 2 p ( C j , T ) , P(C j, T) mean to belong to classification C in sample set T jFrequency, s means the number of classification in T, s=11 here.
Step 22: present node is split into to branch node according to the feature with maximum information gain.
The numbering that will have the feature of maximum information gain is numbered as present node, is 0 if the mark point is numbered the value of the feature of present node numbering, this mark point is divided into to the left branch node of present node, otherwise is divided into the right branch node.
Step 23: each branch node of present node, respectively as present node, is judged to the condition that it stops dividing below whether meeting:
A) information entropy of node is less than the threshold value h τ of entropy, and h τ can get 0.5 here;
B) number of plies of decision tree reaches maximum number of plies depth, can make here that the depth span is 10~30;
C) the mark point number of node is less than sample point minimal amount small, and the span that can make small here is 100~1000.
If present node does not meet the condition that stops dividing, repeating step 22~23; If present node meets the condition that stops dividing, present node is leaf node.Category distribution to the mark point of leaf node is carried out probability statistics.
If obtain k leaf node after the decision tree classifier training finishes, the mark point number of i leaf node is designated as n i, i=1,2 ... k.The label of i leaf node is that the mark point number of the classification of j is designated as n Ij, j=0,1 ... 10.The label of i leaf node is the probability of the classification of j
Figure BDA0000371471660000091
The maximum probability of the mark point classification of this leaf node is P (i, j max)=max{P Ij, j=0,1 ..., 10.Wherein, j maxMean to have the classification of maximum probability.
Next step, adopt decision tree classifier to carry out Classification and Identification to test sample book, and Classification and Identification comprises the random character that calculates test sample book, three of the positions process of utilizing decision tree classifier image to be classified and calculate articulation point.
Step 3: the random character that calculates test sample book.
Step 31: employing formula (1) is calculated the centre of form of target, and employing formula (2) is calculated each pixel of target and the angle of straight line that the centre of form connects and horizontal axis.
Step 32: for each pixel p (x, y) of target, number of computations is the feature of fn, and wherein the computation process of each feature is as follows:
Step 321: the p (x, y) of take is starting point, generates at random two reference point.
The arbitrary reference point generated is designated as q (qx, qy), q (qx, qy) be to take p (x, y) to be starting point, the random-length of take is the vector of β point pointed as rz, random angle, employing formula (3) is calculated, in formula, and rz=r1* α/valz, r1 is the random length generated, can make it get the arbitrary value in 0~1, α is coefficient at every turn, can make that its span is 1000~10000, valz is the depth value of p (x, y); β=θ+ο, θ are the angles that p (x, y) and centre of form c (cx, cy) connect straight line and horizontal axis, and employing formula (2) is calculated, and ο is the random angle generated, and can make it get the arbitrary value in 0 °~360 ° at every turn.
Step 322: according to the situation of reference point, determine the eigenwert of p (x, y), its method and step 122 are identical.
Step 323: according to the order of calculated characteristics, be the ε time, determine p (x, y) feature be numbered ε.
Step 4: utilize decision tree classifier to classify to each pixel p (x, y) of target, assorting process as shown in Figure 4.
Step 41: using the root node of decision tree as present node, carry out following steps.
Step 42: according to the value of the feature that is numbered present node numbering of pixel, be 0 or 1, be divided into the left or right branch node of present node, and using branch node as present node.
Step 43: repeating step 42, until present node is leaf node, if the P of this leaf node is (i, j max) be greater than probability threshold value p τ, j maxBe the class label of pixel, otherwise give up this pixel.
After for all pixels of target, classifying, belong to the position that other pixel of same class forms staff, pixel for the different parts intersection, its classification is usually clear and definite not, step 43 can arrange p τ >=0.7, can remove the pixel that the classification determinacy is lower like this, belongs to same classification at the pixel that guarantees to a greater extent same position, other pixel of same class is distributed in same position, for having simplified condition in the articulation point position of further determining each position.Staff station diagram after Fig. 5 presentation class, the black line of two position boundarys and do not have the part of digital label to mean the indefinite pixel be rejected of classification.
Step 5: the particular location that calculates the articulation point at each position.
Step 51: from each some q of identical category iSet out, find respectively corresponding articulation point position candidate p i, i=1,2 ... r, wherein, r is the sum of the point of identical category.
Step 511: with q iCentered by point, generating yardstick is the rectangular characteristic zone W (x, y, w, h) of w * h.
If W (x, y, w, h) not exclusively on image, with the lap of image and W (x, y, w, h) as characteristic area.
Step 512: in characteristic area, employing formula (1) is calculated all and central point q iThe centre of form c (cx, cy) of generic point.
Step 513: adopt following formula to calculate q iDistance with c (cx, cy): dis = ( cx - x ) 2 + ( cy - y ) 2 .
Step 514: if dis≤d τ, using c (cx, cy) as articulation point position candidate p iIf dis>d τ, with c (cx, cy) point centered by, generate rectangular characteristic zone W (cx, cy, w, h), repeating step 512)~514), for example also do not find the articulation point position candidate 30 times if repeat abundant number of times, the centre of form that will obtain for the last time is as articulation point position candidate p i.Here d τ is distance threshold, can make that its span is 0.1~0.3.
Step 52: to the related node position candidate p of institute iScreen, find the position of articulation point.
Step 521: with p 1As the initial score object, according to p i, i=2,3 ..., the order of r, repeat next step.
Step 522: according to the sequencing that becomes the score object, calculate each score object and p iDistance, as score object and a p iDistance be less than threshold value dis τ, the mark of this score object adds 1, no longer score object and p of calculated for subsequent iDistance.If all score objects and p iThe distance all be not less than dis τ, by p iAs next one score object.Here dis τ can get 2~4.
Step 523: select the score object with highest score tops, if tops is greater than score threshold sco τ, should score to liking the articulation point position; Otherwise reduce sco τ, until find the articulation point position.Here sco τ can get 2~4.
The position of the institute's related node finally, found as shown in Figure 6.

Claims (10)

1.一种基于深度图像的关节点定位方法,其特征在于包括训练过程和识别过程:1. A joint point localization method based on depth image, it is characterized in that comprising training process and recognition process: 训练过程的步骤如下:The steps of the training process are as follows: 1)计算训练样本的随机特征;1) Calculate the random features of the training samples; 2)根据随机特征训练决策树分类器;2) Train a decision tree classifier based on random features; 识别过程的步骤如下:The steps in the identification process are as follows: 3)计算测试样本的随机特征;3) Calculate the random features of the test samples; 4)利用决策树分类器对目标的每个像素点进行分类,得到目标的不同类别的部位;4) Use the decision tree classifier to classify each pixel of the target to obtain different types of parts of the target; 5)计算每个部位的关节点的位置。5) Calculate the position of the joint points of each part. 2.根据权利要求1所述的基于深度图像的关节点定位方法,其特征在于:所述步骤1)中的训练样本是指只保留目标且经过真值标注的深度图像的集合。2. The method for locating joint points based on depth image according to claim 1, characterized in that: the training samples in the step 1) refer to the set of depth images that only keep the target and have been marked with true values. 3.根据权利要求1或2所述的基于深度图像的关节点定位方法,其特征在于:所述步骤1)具体步骤如下:3. The joint point location method based on depth image according to claim 1 or 2, characterized in that: the step 1) specific steps are as follows: 11)采用式(1)计算目标的形心c(cx,cy):11) Use formula (1) to calculate the centroid c(cx, cy) of the target: cxcx == 11 kk &Sigma;&Sigma; ii == 11 kk xx ii cycy == 11 kk &Sigma;&Sigma; ii == 11 kk ythe y ii -- -- -- (( 11 )) 其中,k表示目标上像素点的总数,(xi,yi)表示目标上每个像素点的坐标,i=1,2,...k;Among them, k represents the total number of pixels on the target, ( xi , y i ) represents the coordinates of each pixel on the target, i=1, 2,...k; 12)以标注点为起点,生成以两个随机向量所指向的不同参考点,其中,随机向量的长度rz=r1*α/valz,r1是随机生成的长度,α是系数,valz是起点的深度值;随机向量的角度β=θ+ο,θ是起点与形心所连直线与水平坐标轴的夹角,ο是随机生成的角度;如果两个参考点至少有一个不在图像上,则起点的特征的值为1;两个参考点都在图像上,计算两个参考点的深度差,如果深度差大于在自定义阈值集
Figure FDA0000371471650000023
中任意选择的一个,则起点的特征的值为1,否则起点的特征的值为0;
12) Starting from the marked point, generate different reference points pointed by two random vectors, where the length of the random vector is rz=r1*α/valz, r1 is the randomly generated length, α is the coefficient, and valz is the starting point Depth value; the angle β=θ+ο of the random vector, θ is the angle between the line connecting the starting point and the centroid and the horizontal coordinate axis, and ο is the angle generated randomly; if at least one of the two reference points is not on the image, then The value of the feature of the starting point is 1; both reference points are on the image, calculate the depth difference between the two reference points, if the depth difference is greater than in the custom threshold set
Figure FDA0000371471650000023
Any one selected in , the value of the feature of the starting point is 1, otherwise the value of the feature of the starting point is 0;
13)对于每个标注点,重复步骤12)fn次生成fn个特征,同时根据特征生成的次序对特征进行1~fn的编号。13) For each marked point, repeat step 12) fn times to generate fn features, and number the features from 1 to fn according to the order of feature generation.
4.根据权利要求1所述的基于深度图像的关节点定位方法,其特征在于:所述步骤2)包括以下具体步骤:4. The method for locating joint points based on depth image according to claim 1, characterized in that: said step 2) comprises the following specific steps: 21)把决策树分类器的根节点作为当前节点;21) Take the root node of the decision tree classifier as the current node; 22)计算当前节点的每个特征的信息增益:22) Calculate the information gain of each feature of the current node: GainGain (( &epsiv;&epsiv; )) == entroyentroy (( TT )) -- &Sigma;&Sigma; ii == 11 mm TT ii TT entroyentroy (( TT ii )) ,, 其中,ε表示特征的编号,ε=1,2,...fn,T表示当前节点即标注点样本集,Ti表示样本集的子集,m表示子集个数,根据编号为ε的特征的值为0或1,将标注点划分到两个子集即m=2,entroy(T)表示样本集的信息熵, entroy ( T ) = - &Sigma; j = 1 s p ( C j , T ) log 2 p ( C j , T ) , p(Cj,T)表示样本集T中属于类别Cj的频率,s表示T中类别的个数;Among them, ε represents the serial number of the feature, ε=1, 2,...fn, T represents the current node is the sample set of label points, T i represents the subset of the sample set, m represents the number of subsets, according to the number ε The value of the feature is 0 or 1, and the marked points are divided into two subsets, that is, m=2, entroy(T) represents the information entropy of the sample set, entroy ( T ) = - &Sigma; j = 1 the s p ( C j , T ) log 2 p ( C j , T ) , p(C j , T) represents the frequency of category C j in the sample set T, and s represents the number of categories in T; 23)将具有最大信息增益的特征的编号作为当前节点的编号;23) Use the number of the feature with the largest information gain as the number of the current node; 24)标注点编号为当前节点编号的特征的值如果为0,将该标注点划分到当前节点的左分支节点,否则划分到右分支节点;24) If the value of the feature whose label point number is the current node number is 0, the label point is divided into the left branch node of the current node, otherwise it is divided into the right branch node; 25)将分支节点作为当前节点,如果当前节点的信息熵小于熵的阈值hτ,或者决策树的层数达到最大层数depth,或者当前节点的标注点个数小于样本点最小数目small,则停止分裂,把当前节点作为叶子节点,否则重复步骤22)~25);25) Take the branch node as the current node, if the information entropy of the current node is less than the entropy threshold hτ, or the number of layers of the decision tree reaches the maximum layer depth, or the number of labeled points of the current node is less than the minimum number of sample points small, then stop Split, take the current node as a leaf node, otherwise repeat steps 22) to 25); 26)对叶子节点的标注点类别分布进行概率统计。26) Perform probability statistics on the distribution of label point categories of leaf nodes. 5.根据权利要求1所述的基于深度图像的关节点定位方法,其特征在于:所述步骤3)中的测试样本是指只保留目标的深度图像的集合。5 . The method for locating joint points based on depth images according to claim 1 , wherein the test samples in the step 3) refer to a collection of depth images that only retain objects. 6 . 6.根据权利要求1或5所述的基于深度图像的关节点定位方法,其特征在于:所述步骤3)包括以下具体步骤:6. The method for locating joint points based on depth image according to claim 1 or 5, characterized in that: said step 3) includes the following specific steps: 31)采用式(1)计算目标的形心;31) Use formula (1) to calculate the centroid of the target; 32)以目标上的点为起点,生成以两个随机向量所指向的不同参考点,其中随机向量的长度rz=r1*α/valz,r1是随机生成的长度,α是系数,valz是起点的深度值;随机向量的角度β=θ+ο,θ是起点与形心所连直线与水平坐标轴的夹角,ο是随机生成的角度;如果两个参考点至少有一个不在图像上,则起点的特征的值为1;如果两个参考点都在图像上,计算两个参考点的深度差,如果深度差大于在自定义阈值集
Figure FDA0000371471650000031
中任意选择的一个,则起点的特征的值为1,否则起点的特征的值为0;
32) Use the point on the target as the starting point to generate different reference points pointed to by two random vectors, where the length of the random vector rz=r1*α/valz, r1 is the randomly generated length, α is the coefficient, and valz is the starting point The depth value; the angle β=θ+ο of the random vector, θ is the angle between the line connecting the starting point and the centroid and the horizontal coordinate axis, and ο is a randomly generated angle; if at least one of the two reference points is not on the image, Then the value of the feature of the starting point is 1; if both reference points are on the image, calculate the depth difference between the two reference points, if the depth difference is greater than the custom threshold set
Figure FDA0000371471650000031
Any one selected in , the value of the feature of the starting point is 1, otherwise the value of the feature of the starting point is 0;
33)对于每个标注点,重复步骤32)fn次生成fn个特征,同时根据特征生成的次序对特征进行1~fn的编号。33) For each marked point, repeat step 32) fn times to generate fn features, and number the features from 1 to fn according to the order of feature generation.
7.根据权利要求1所述的基于深度图像的关节点定位方法,其特征在于:所述步骤4)包括以下具体步骤:7. The joint point location method based on depth image according to claim 1, characterized in that: said step 4) comprises the following specific steps: 41)把决策树分类器的根节点作为当前节点;41) Take the root node of the decision tree classifier as the current node; 42)如果像素点的编号为当前节点编号的特征的值为0,将其划分到当前节点的左分支节点,否则划分到右分支节点;42) If the number of the pixel point is 0 for the characteristic value of the current node number, divide it into the left branch node of the current node, otherwise divide it into the right branch node; 43)将像素划分到的分支节点作为当前节点,重复步骤42)、43),直到像素到达叶子节点为止;43) Take the branch node where the pixel is divided into as the current node, and repeat steps 42), 43) until the pixel reaches the leaf node; 44)叶子节点的所有类别的最大概率如果大于概率阈值pτ,则判定具有最大概率的类别为像素的类别,否则舍弃该像素点。44) If the maximum probability of all categories of leaf nodes is greater than the probability threshold pτ, the category with the maximum probability is determined to be the category of the pixel, otherwise the pixel is discarded. 8.根据权利要求1所述的基于深度图像的关节点定位方法,其特征在于:所述步骤5)包括以下具体步骤:8. The method for locating joint points based on depth image according to claim 1, characterized in that: said step 5) comprises the following specific steps: 51)从相同类别的每个点qi出发,分别找到对应的关节点候选位置pi,i=1,2,...r,其中,r是相同类别的点的总数;51) Starting from each point q i of the same category, find the corresponding joint point candidate position p i , i=1, 2,...r, where r is the total number of points of the same category; 52)对所有关节点候选位置进行筛选,找到关节点的位置。52) Screen all the candidate positions of the joint nodes to find the position of the joint points. 9.根据权利要求1或8所述的基于深度图像的关节点定位方法,其特征在于:所述步骤51)包括以下具体步骤:9. The depth image-based joint point positioning method according to claim 1 or 8, characterized in that: the step 51) includes the following specific steps: 511)以qi为中心点,生成尺度为w×h的矩形特征区域;511) Using q i as the center point, generate a rectangular feature area with a scale of w×h; 512)采用式(1)计算特征区域内与中心点同类别的点的形心;512) Use formula (1) to calculate the centroid of points of the same category as the center point in the feature area; 513)计算中心点与形心的距离;513) Calculate the distance between the center point and the centroid; 514)如果距离不大于距离阈值dτ,则将形心作为关节点候选位置,否则生成以形心为中心点,尺度为w×h的矩形特征区域,重复步骤512)~514),如果重复足够多的次数还没有找到关节点候选位置,则将最后一次得到的形心作为关节点候选位置pi;所述重复足够多的次数为大于等于30次。514) If the distance is not greater than the distance threshold dτ, use the centroid as the candidate position of the joint point, otherwise generate a rectangular feature area with the centroid as the center point and a scale of w×h, repeat steps 512) to 514), if repeated enough If the joint point candidate position has not been found for too many times, the last obtained centroid is used as the joint point candidate position p i ; the sufficient number of repetitions is greater than or equal to 30 times. 10.根据权利要求1或8所述的基于深度图像的关节点定位方法,其特征在于:所述步骤52)包括以下具体步骤:10. The depth image-based joint point positioning method according to claim 1 or 8, characterized in that: the step 52) includes the following specific steps: 521)以p1作为初始计分对象,按照pi,i=2,3,...,r的顺序,重复执行下一步;521) Take p 1 as the initial scoring object, and perform the next step repeatedly in the order of p i , i=2, 3, ..., r; 522)按照成为计分对象的先后顺序,计算每个计分对象与pi的距离,当一个计分对象与pi的距离小于阈值disτ,则该计分对象的分数加1,不再计算后续计分对象与pi的距离;如果所有计分对象与pi的距离都不小于disτ,则将pi作为下一个计分对象;522) Calculate the distance between each scoring object and p i according to the order of scoring objects. When the distance between a scoring object and p i is less than the threshold disτ, add 1 to the score of the scoring object and no longer calculate The distance between the subsequent scoring object and pi ; if the distance between all scoring objects and pi is not less than disτ, then take pi as the next scoring object; 523)选出具有最高分数tops的计分对象,如果tops大于分数阈值scoτ,则该计分对象是关节点位置;否则减小scoτ,直到找到关节点位置为止。523) Select the scoring object with the highest score tops, if tops is greater than the score threshold scoτ, then the scoring object is the joint point position; otherwise, reduce scoτ until the joint point position is found.
CN201310374236.7A 2013-08-23 2013-08-23 Intra-articular irrigation method based on depth image Expired - Fee Related CN103413145B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310374236.7A CN103413145B (en) 2013-08-23 2013-08-23 Intra-articular irrigation method based on depth image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310374236.7A CN103413145B (en) 2013-08-23 2013-08-23 Intra-articular irrigation method based on depth image

Publications (2)

Publication Number Publication Date
CN103413145A true CN103413145A (en) 2013-11-27
CN103413145B CN103413145B (en) 2016-09-21

Family

ID=49606152

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310374236.7A Expired - Fee Related CN103413145B (en) 2013-08-23 2013-08-23 Intra-articular irrigation method based on depth image

Country Status (1)

Country Link
CN (1) CN103413145B (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104050460A (en) * 2014-06-30 2014-09-17 南京理工大学 Pedestrian detection method with multi-feature fusion
CN104091152A (en) * 2014-06-30 2014-10-08 南京理工大学 Method for detecting pedestrians in big data environment
CN105389569A (en) * 2015-11-17 2016-03-09 北京工业大学 Human body posture estimation method
CN105893970A (en) * 2016-03-31 2016-08-24 杭州电子科技大学 Nighttime road vehicle detection method based on luminance variance characteristics
CN106096551A (en) * 2016-06-14 2016-11-09 湖南拓视觉信息技术有限公司 The method and apparatus of face part Identification
CN106558071A (en) * 2016-11-10 2017-04-05 张昊华 A kind of method and terminal for obtaining human synovial information
CN106846403A (en) * 2017-01-04 2017-06-13 北京未动科技有限公司 The method of hand positioning, device and smart machine in a kind of three dimensions
CN107203756A (en) * 2016-06-06 2017-09-26 亮风台(上海)信息科技有限公司 A kind of method and apparatus for recognizing gesture
CN107436679A (en) * 2016-05-27 2017-12-05 富泰华工业(深圳)有限公司 Gestural control system and method
CN107766848A (en) * 2017-11-24 2018-03-06 广州鹰瞰信息科技有限公司 The pedestrian detection method and storage medium of vehicle front
CN108345869A (en) * 2018-03-09 2018-07-31 南京理工大学 Driver's gesture recognition method based on depth image and virtual data
CN109484935A (en) * 2017-09-13 2019-03-19 杭州海康威视数字技术股份有限公司 A kind of lift car monitoring method, apparatus and system
CN110598510A (en) * 2018-06-13 2019-12-20 周秦娜 Vehicle-mounted gesture interaction technology
CN114529567A (en) * 2022-01-25 2022-05-24 上海卫星工程研究所 Harris and decision tree based corner detection method and system
CN117011835A (en) * 2023-07-14 2023-11-07 东风柳州汽车有限公司 Method, device, equipment and storage medium for assisting child sleeping

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101359372A (en) * 2008-09-26 2009-02-04 腾讯科技(深圳)有限公司 Training method and device of classifier, and method apparatus for recognising sensitization picture
CN102411711A (en) * 2012-01-04 2012-04-11 山东大学 A Finger Vein Recognition Method Based on Personalized Weights

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101359372A (en) * 2008-09-26 2009-02-04 腾讯科技(深圳)有限公司 Training method and device of classifier, and method apparatus for recognising sensitization picture
CN102411711A (en) * 2012-01-04 2012-04-11 山东大学 A Finger Vein Recognition Method Based on Personalized Weights

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104091152A (en) * 2014-06-30 2014-10-08 南京理工大学 Method for detecting pedestrians in big data environment
CN104050460A (en) * 2014-06-30 2014-09-17 南京理工大学 Pedestrian detection method with multi-feature fusion
CN104050460B (en) * 2014-06-30 2017-08-04 南京理工大学 Pedestrian detection method based on multi-feature fusion
CN105389569B (en) * 2015-11-17 2019-03-26 北京工业大学 A kind of estimation method of human posture
CN105389569A (en) * 2015-11-17 2016-03-09 北京工业大学 Human body posture estimation method
CN105893970A (en) * 2016-03-31 2016-08-24 杭州电子科技大学 Nighttime road vehicle detection method based on luminance variance characteristics
CN107436679A (en) * 2016-05-27 2017-12-05 富泰华工业(深圳)有限公司 Gestural control system and method
CN107203756B (en) * 2016-06-06 2020-08-28 亮风台(上海)信息科技有限公司 Method and device for recognizing gestures
CN107203756A (en) * 2016-06-06 2017-09-26 亮风台(上海)信息科技有限公司 A kind of method and apparatus for recognizing gesture
CN106096551A (en) * 2016-06-14 2016-11-09 湖南拓视觉信息技术有限公司 The method and apparatus of face part Identification
CN106096551B (en) * 2016-06-14 2019-05-21 湖南拓视觉信息技术有限公司 The method and apparatus of face position identification
CN106558071A (en) * 2016-11-10 2017-04-05 张昊华 A kind of method and terminal for obtaining human synovial information
CN106558071B (en) * 2016-11-10 2019-04-23 张昊华 A method and terminal for acquiring human joint information
CN106846403B (en) * 2017-01-04 2020-03-27 北京未动科技有限公司 Method and device for positioning hand in three-dimensional space and intelligent equipment
CN106846403A (en) * 2017-01-04 2017-06-13 北京未动科技有限公司 The method of hand positioning, device and smart machine in a kind of three dimensions
CN109484935A (en) * 2017-09-13 2019-03-19 杭州海康威视数字技术股份有限公司 A kind of lift car monitoring method, apparatus and system
CN107766848A (en) * 2017-11-24 2018-03-06 广州鹰瞰信息科技有限公司 The pedestrian detection method and storage medium of vehicle front
CN108345869A (en) * 2018-03-09 2018-07-31 南京理工大学 Driver's gesture recognition method based on depth image and virtual data
CN110598510A (en) * 2018-06-13 2019-12-20 周秦娜 Vehicle-mounted gesture interaction technology
CN110598510B (en) * 2018-06-13 2023-07-04 深圳市点云智能科技有限公司 Vehicle-mounted gesture interaction technology
CN114529567A (en) * 2022-01-25 2022-05-24 上海卫星工程研究所 Harris and decision tree based corner detection method and system
CN117011835A (en) * 2023-07-14 2023-11-07 东风柳州汽车有限公司 Method, device, equipment and storage medium for assisting child sleeping

Also Published As

Publication number Publication date
CN103413145B (en) 2016-09-21

Similar Documents

Publication Publication Date Title
CN103413145A (en) Articulation point positioning method based on depth image
Zhao et al. Single image action recognition using semantic body part actions
CN109948425A (en) A pedestrian search method and device based on structure-aware self-attention and online instance aggregation matching
CN110349122A (en) A kind of pavement crack recognition methods based on depth convolution fused neural network
CN112016605B (en) A Target Detection Method Based on Bounding Box Corner Alignment and Boundary Matching
CN103473571B (en) Human detection method
CN109766868B (en) A real scene occluded pedestrian detection network and detection method based on body key point detection
CN104809481B (en) A kind of natural scene Method for text detection based on adaptive Color-based clustering
CN102842045B (en) A kind of pedestrian detection method based on assemblage characteristic
CN103971102A (en) Static gesture recognition method based on finger contour and decision-making trees
CN110969166A (en) A small target recognition method and system in an inspection scene
CN106845487A (en) A kind of licence plate recognition method end to end
CN107742102A (en) A kind of gesture identification method based on depth transducer
CN109063768A (en) Vehicle recognition methods, apparatus and system again
CN105718912B (en) A kind of vehicle characteristics object detecting method based on deep learning
CN103390164A (en) Object detection method based on depth image and implementing device thereof
CN107766791A (en) A kind of pedestrian based on global characteristics and coarseness local feature recognition methods and device again
CN109410238A (en) A kind of fructus lycii identification method of counting based on PointNet++ network
CN109242047A (en) Bank card number detection and recognition methods based on K-means++ cluster and residual error network class
CN106874901A (en) A kind of driving license recognition methods and device
CN110008900A (en) A Region-to-target Candidate Target Extraction Method for Visible Light Remote Sensing Images
CN103279760A (en) Real-time classifying method of plant quarantine larvae
CN107992783A (en) Face image processing process and device
CN107609464B (en) A kind of real-time face rapid detection method
CN106886757A (en) A kind of multiclass traffic lights detection method and system based on prior probability image

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20160921

Termination date: 20200823