[go: up one dir, main page]

CN101719015B - Method for positioning finger tips of directed gestures - Google Patents

Method for positioning finger tips of directed gestures Download PDF

Info

Publication number
CN101719015B
CN101719015B CN2009101981969A CN200910198196A CN101719015B CN 101719015 B CN101719015 B CN 101719015B CN 2009101981969 A CN2009101981969 A CN 2009101981969A CN 200910198196 A CN200910198196 A CN 200910198196A CN 101719015 B CN101719015 B CN 101719015B
Authority
CN
China
Prior art keywords
gesture
indication
image
indication gesture
region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN2009101981969A
Other languages
Chinese (zh)
Other versions
CN101719015A (en
Inventor
管业鹏
于蕴杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Shanghai for Science and Technology
Original Assignee
University of Shanghai for Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Shanghai for Science and Technology filed Critical University of Shanghai for Science and Technology
Priority to CN2009101981969A priority Critical patent/CN101719015B/en
Publication of CN101719015A publication Critical patent/CN101719015A/en
Application granted granted Critical
Publication of CN101719015B publication Critical patent/CN101719015B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

本发明涉及一种指示手势的手指尖定位方法。它是根据指示手势行为中的指示手势的手部特征,自动确定指示手势的手指尖位置。采用背景差分法,提取指示手势对象,运用肤色分割方法,提取指示手势的手部区域,根据指示手势的手指尖位于指示手势的手部边沿轮廓且距离指示手势的手指区域重心最远处,自动确定手指尖点。从而快速有效地定位指示手势的手指尖点位置,满足指示手势人机交互中的特征提取要求。

The invention relates to a fingertip positioning method for pointing gestures. It automatically determines the fingertip position of the pointing gesture according to the hand characteristics of the pointing gesture in the pointing gesture behavior. Use the background subtraction method to extract the pointing gesture object, and use the skin color segmentation method to extract the hand area of the pointing gesture. According to the fingertip of the pointing gesture is located on the edge contour of the hand of the pointing gesture and is farthest from the center of gravity of the finger area of the pointing gesture, automatically Identify fingertip points. Therefore, the position of the fingertip of the pointing gesture can be quickly and effectively located, and the feature extraction requirement in the human-computer interaction of the pointing gesture can be met.

Description

The finger tip localization method of indication gesture
Technical field
The present invention relates to a kind of finger tip localization method of indicating gesture, be used for video digital images analysis and understanding.Belong to the intelligent information processing technology field.
Background technology
Along with rapid development of computer technology; research meets the human novel human-computer interaction technology Showed Very Brisk that exchanges custom naturally; and human-computer interaction technology is from being that the center transfers to progressively that focus be put on man with the computing machine; multimedia user interface has then greatly enriched the form of expression of computerized information, the user can be replaced or utilizes a plurality of sensory channels simultaneously.Yet the man-machine interaction form of multimedia user interface still forces the user to use conventional input equipment (as keyboard, Genius mouse and touch-screen etc.) to import, and becomes the bottleneck of current man-machine interaction.Virtual reality can realize harmonious, man-machine interface focusing on people as a kind of novel human-machine interaction form.In virtual reality, if with staff directly as computer entry device, then can make full use of human daily technical ability, and not need special training or study, the communication of between humans and machines will no longer need intermediary.In with the field of tool of staff as natural interaction, indication gesture (pointing gesture) is explained easily.The indication gesture is to use the reflection of finger to the spatial impression targets of interest in people's daily life, is the important pioneer of human family of languages development and ontogeny, can disclose human society intelligence, is a kind of desirable natural interactive mode.
For realizing man-machine interaction based on the indication gesture, adopt based on the data helmet, data glove and body marker etc. at present, these class methods are intrusive mood, the user needs specialized training, operation inconvenience; For overcoming above-mentioned deficiency, adopt based on non-contact sensor (as video camera), obtain the characteristics of objects of indication gesture, determine to refer to the gesture extraterrestrial target, realize man-machine interaction, wherein, refer to the gesture extraterrestrial target, determine with the intersection point on plane, target place by the finger tip of indication gesture and the line of people's an eye line.
When obtaining the finger tip position of indication gesture, following method is arranged generally: the one, to wear the wireless sensing gloves or wear the gloves of special color, algorithm need be equipped with special instrument and equipment, operation inconvenience; The 2nd, the requirement background is simple, and indication gesture object is single, and this method is difficult to adapt to complicated scene change.
Summary of the invention
The objective of the invention is to finger tip extracting method at existing indication gesture, existence need or require background simple by contact external device (as gloves), indicate problems such as the gesture object is single, a kind of finger tip localization method of improved indication gesture is provided.It is to locate fast according to the feature of the indication gesture hand in the behavior of indication gesture to refer to the gesture finger tip, thereby improves the dirigibility and the simplicity of man-machine interaction.
For achieving the above object, design of the present invention is: adopt the background subtraction point-score, extract indication gesture object, utilization skin color segmentation method, extract the hand region of indication gesture, finger tip according to the indication gesture is positioned at the hand edge profile of indication gesture and the finger areas center of gravity farthest of distance indication gesture, determines the finger cusp automatically, thus the finger cusp position of the gesture of location indication fast and effeciently.
According to the foregoing invention design, the present invention adopts following technical proposals:
A kind of finger tip localization method of indicating gesture.It is the hand-characteristic according to the indication gesture in the behavior of indication gesture, determines the finger tip position of indication gesture automatically, and concrete steps are as follows:
1) starts indication images of gestures acquisition system: gather video image;
2) obtain background image
Continuous acquisition does not comprise the scene image of indication gesture target, when two image difference are less than certain setting threshold in certain setting-up time interval, a certain width of cloth image in then should the time interval is image as a setting, otherwise gather again, two image difference in the time interval of satisfying setting are less than certain setting threshold;
3) indication gesture object cuts apart
Current frame image and step 2 by camera acquisition) background image that obtains subtracts each other, and is partitioned into indication gesture subject area;
4) extract hand region;
5) hand region of definite indication gesture;
6) the finger tip position of definite indication gesture.
Above-mentioned steps 4) concrete operations step is as follows:
(1) color space conversion, calculate color-values Cr, Cb:, calculate the color-values Cr of YCbCr color space, Cb by the red R of RGB color space, green G, blue B three-component:
Cr=0.5×R-0.4187×G-0.0813×B
Cb=-0.1687×R-0.3313×G+0.5×B
(2) area of skin color extracts: determine color-values Cr, the threshold value T of Cb and Cr/Cb respectively 1, T 2, T 3, T 4, T 5, the zone with all pixels that satisfy following formula are formed is defined as area of skin color S
S=(Cr≥T 1∩Cr≤T 2)∩(Cb≥T 3∩Cb≤T 4)∩(Cr/Cb≥T 5)
Wherein, ∩ is " logical and " operational character;
(3) extract the area of skin color of possible indication gesture object: will satisfy the image-region of step 3) and step (2) simultaneously, as the area of skin color of possible indication gesture object;
(4) extract hand region: the bianry image to step (3) carries out the connected region search, calculate the ratio of high S1 of connected region and wide Sw, and the hole in the connected region is counted H and connected region size W, the zone that all pixels that satisfy following formula are formed is considered as non-hand region, rejects from the bianry image zone of step (3);
F=(S l/S w≥T 6∩S l/S w≤T 7)∩(H>1)∩W<T 8
Wherein, T 6, T 7, T 8Be threshold value.
Above-mentioned steps 5) method of the hand region of definite indication gesture is: calculate the bottom pixels i through the bianry image connected region of step (4) gained respectively, the ordinate By:By=max (j) of j and horizontal ordinate Bx:Bx=i (j=By), to comprise and have minimum By value and the corresponding with it connected region that Bx formed, be defined as indicating the hand region of gesture;
Above-mentioned steps 6) method of the finger tip position of definite indication gesture is: the hand region center of gravity Cx that calculates the indication gesture, Cy, and the pixel coordinate i of each point on the hand region outline line of indication gesture, j is to center of gravity Cx, the distance D of Cy, distance D had peaked pixel coordinate i, j is defined as indicating the finger tip position Px of gesture, Py, Px:Px=i (D=max (D)), Py:Py=j (D=max (D)).
Principle of the present invention is as follows:
In technical scheme of the present invention, can provide characteristic more completely according to the background subtraction point-score, all can be embodied in the variation of scene image sequence based on any perceptible target travel in the scene, utilize the difference between present image and the background image, from video image, detect indication gesture object.
If in the time interval Δ t, obtain t respectively N-1With t nThe two two field picture f (t in two moment N-1, x, y), f (t n, x y), asks difference with two width of cloth images by pixel, difference image Diff (x, y):
DiffR(x,y)=|fR(t n,x,y)-fR(t n-1,x,y)|
DiffG(x,y)=|fG(t n,x,y)-fG(t n-1,x,y)|
DiffB(x,y)=|fB(t n,x,y)-fB(t n-1,x,y)|
Wherein, DiffR, DiffG, DiffB is corresponding difference image red, green, blue three-component respectively, | f| is the absolute value of f.If two sequence image f (t in the time interval Δ t N-1, x, y), f (t n, x, difference DiffR y) (x, y)≤T|DiffG (x, y)≤T|DiffB (x, y)≤T, wherein, T is a threshold value, | be " logical OR " operational symbol, show not change object in the Δ t time interval, thus can be with t n~t N-1Between the image in a certain moment, image as a setting.
Utilize the gained background image,, adopt the background subtraction point-score, be partitioned into indication gesture subject area according to the current current frame image that obtains.
Simultaneously, although the colour of skin varies with each individual, and vary, in the color space of YCrCb, the human colour of skin presents good cluster characteristic, and insensitive to the attitude variation, can overcome variable effects such as rotation, expression, has strong robustness.In the YCbCr model, the monochrome information of Y representation in components color, Cr and Cb component are represented red and blue colourity respectively.The conversion formula that is transformed into the YCbCr space from rgb space is as follows:
Y Cb Cr = 0.2989 0.5866 0.1145 - 0.1688 - 0.3312 0.5000 0.5000 - 0.4183 - 0.0817 R G B
In the YCbCr space, the colour of skin is in the stable scope in the distribution of Cb, Cr and Cr/Cb,, in the current image that obtains, will satisfy (Cr 〉=T that is 1∩ Cr≤T 2) ∩ (Cb 〉=T 3∩ Cb≤T 4) ∩ (Cr/Cb 〉=T 5) pixel of condition, as the area of skin color of present image, wherein, ∩ is " logical and " operational character.For overcoming the influence of class colour of skin information (as timber floor, wooden cabinet etc.) in the present image, with satisfying the indication gesture subject area that area of skin color that this colour of skin condition extracted and above-mentioned employing background subtraction point-score are partitioned into simultaneously, as indication gesture subject area.
For extracting the finger areas of indication gesture, need from the indication gesture subject area of obtaining, get rid of the human face region that also satisfies above-mentioned two conditions simultaneously, because color characteristic areas such as human eye, eyebrow, lip are not in face complexion, therefore, in the human face region that extracts, will there be hole, and the height of face complexion area, wide ratio, be distributed in the stable scope, reject the influence that people's face extracts indication gesture finger areas in view of the above.
According to the behavioural characteristic of indication gesture and since the finger areas of the more non-indication gesture of finger areas of indication gesture in image on the upper side, therefore, the connected region that will comprise bottom vertical pixel coordinates value minimum of finger areas is defined as indicating the hand region of gesture.Because the forefinger zone of indication gesture is little than arm area in size, therefore, finger tip one is positioned from finger areas center of gravity farthest, thereby realizes the location of indication gesture finger tip.
The present invention compared with prior art, have following conspicuous outstanding substantive distinguishing features and remarkable advantage: the present invention is according to the hand-characteristic of the indication gesture in the behavior of indication gesture, though and the human colour of skin varies with each individual, but in the color space of YCrCb, present good cluster characteristic, in conjunction with the background subtraction separating method, extract the hand region of indication gesture object, and the finger tip of indication gesture is positioned at the hand edge profile of indication gesture and the finger areas center of gravity farthest of distance indication gesture, location finger tip point, computing is easy, flexibly, realize easily, solved when extracting the finger tip of indication gesture, need be simple by contact external device (as gloves) or context request, indication gesture object is single, and dynamic scene is changed responsive, noise is big, the deficiency of computing complexity; Improve the robustness that indication gesture finger tip extracts, can adapt to the automatic location of indication gesture finger tip under the complex background condition.Easy, flexible, the easy realization of method of the present invention.
Description of drawings
Fig. 1 is the flowsheet of the inventive method.
Fig. 2 is the original background image of one embodiment of the invention.
Fig. 3 is the current indication images of gestures of one embodiment of the invention.
Fig. 4 is the indication gesture object bianry image that is partitioned into based on the background difference of one embodiment of the invention.
Fig. 5 is the colour of skin bianry image that is partitioned into based on the YCbCr color space of one embodiment of the invention.
Fig. 6 is the area of skin color that possible indicate the gesture object of one embodiment of the invention
The hand region bianry image that Fig. 7 extracts.
Fig. 8 indicates gesture hand region bianry image.
Embodiment
A specific embodiment of the present invention is: running program as shown in Figure 1.This routine original background image as shown in Figure 2, current indication images of gestures such as Fig. 3 according to the hand-characteristic of the indication gesture in the indication gesture behavior, to coloured image shown in Figure 3, indicate the finger tip location of gesture.Concrete steps are as follows:
1) starts indication images of gestures acquisition system: gather video image;
2) obtain background image
Continuous acquisition does not comprise the scene image of indication gesture target, when two image difference are less than certain setting threshold in certain setting-up time interval, a certain width of cloth image in then should the time interval is image as a setting, otherwise gather again, two image difference in the time interval of satisfying setting are less than certain setting threshold;
3) indication gesture object cuts apart
Subtract each other by current indication images of gestures (Fig. 3) and original background image (Fig. 2), be partitioned into indication gesture subject area, as shown in Figure 4;
4) extract hand region
The concrete operations step is as follows:
(1) color space conversion, calculate color-values Cr, Cb:, calculate the color-values Cr of YCbCr color space, Cb by the red R of RGB color space, green G, blue B three-component:
Cr=0.5×R-0.4187×G-0.0813×B
Cb=-0.1687×R-0.3313×G+0.5×B
(2) area of skin color extracts: determine color-values Cr respectively, and the threshold range of Cb and Cr/Cb, the zone with all pixels that satisfy following formula are formed is defined as area of skin color S, as shown in Figure 5;
S=(Cr≥87∩Cr≤133)∩(Cb≥78∩Cb≤127)∩(Cr/Cb≥1.05)
Wherein, ∩ is " logical and " operational character.
(3) extract the area of skin color of possible indication gesture object: will satisfy the bianry image of step 3) and step (2) simultaneously, as the area of skin color of possible indication gesture object, as shown in Figure 6;
(4) extract hand region: to bianry image shown in Figure 6, carry out the connected region search, calculate the high S of connected region lWith wide S wRatio, and the hole in the connected region counts H and connected region size W, the zone that all pixels that satisfy following formula are formed is considered as non-hand region, rejects from bianry image zone shown in Figure 6, obtains hand region, as shown in Figure 7;
F=(S l/S w≥0.75∩S l/S w≤2.5)∩(H>1)∩W<300
5) hand region of definite indication gesture: the bottom pixels i that calculates bianry image connected region shown in Figure 7, the ordinate By:By=max (j) of j and horizontal ordinate Bx:Bx=i (j=By), to comprise and have minimum By value and the corresponding with it connected region that Bx formed, be defined as indicating the hand region of gesture, as shown in Figure 8;
6) the finger tip position of definite indication gesture: the hand region center of gravity (Cx that calculates the indication gesture, Cy), and the pixel coordinate i of each point on the hand region outline line of indication gesture, j is to center of gravity (Cx, Cy) distance D, distance D is had peaked pixel coordinate i, and j is defined as indicating the finger tip position (Px of gesture, Py), Px:Px=i (D=max (D)), Py:Py=j (D=max (D))), shown in spider among Fig. 8.

Claims (1)

1. a finger tip localization method of indicating gesture is characterized in that the hand-characteristic according to the indication gesture in the behavior of indication gesture, determines the finger cusp position of indication gesture automatically, and concrete steps are as follows:
1) starts indication images of gestures acquisition system: gather video image;
2) obtain background image
Continuous acquisition does not comprise the scene image of indication gesture target, when two image difference are less than certain setting threshold in certain setting-up time interval, a certain width of cloth image in then should the time interval is image as a setting, otherwise gather again, two image difference in the time interval of satisfying setting are less than certain setting threshold;
3) indication gesture object cuts apart
Current frame image and step 2 by camera acquisition) background image that obtains subtracts each other, and is partitioned into indication gesture subject area;
4) extract hand region;
5) hand region of definite indication gesture;
6) the finger tip position of definite indication gesture;
The concrete operations step that described step 4) is extracted hand region is as follows:
(1) color space conversion, calculate color-values Cr, Cb:, calculate the color-values Cr of YCbCr color space, Cb by the red R of RGB color space, green G, blue B three-component:
Cr=0.5×R-0.4187×G-0.0813×B
Cb=-0.1687×R-0.3313×G+0.5×B
(2) area of skin color extracts: determine color-values Cr, the threshold value T of Cb and Cr/Cb respectively 1, T 2, T 3, T 4, T 5, the zone with all pixels that satisfy following formula are formed is defined as area of skin color S
S=(Cr≥T 1∩Cr≤T 2)∩(Cb≥T 3∩Cb≤T 4)∩(Cr/Cb≥T 5)
Wherein, ∩ is " logical and " operational character;
(3) extract the area of skin color of possible indication gesture object: will satisfy the image-region of step 3) and step (2) simultaneously, as the area of skin color of possible indication gesture object;
(4) extract hand region: the bianry image to step (3) carries out the connected region search, calculates the high S of connected region lWith wide S wRatio, and the hole in the connected region counts H and connected region size W, the zone that all pixels that satisfy following formula are formed is considered as non-hand region, rejects from the bianry image zone of step (3);
F=(S l/S w≥T 6∩S l/S w≤T 7)∩(H>1)∩W<T 8
Wherein, T 6, T 7, T 8Be threshold value;
The concrete operations step of the hand region of the definite indication of described step 5) gesture is as follows: calculate the bottom pixels i through the bianry image connected region of step (4) gained respectively, ordinate By:By=max (j) and the horizontal ordinate Bx:Bx=i of j, j=By, to comprise and have minimum By value and the corresponding with it connected region that Bx formed, be defined as indicating the hand region of gesture;
Described step 6) determines that the method for the finger tip position of indication gesture is: the hand region center of gravity Cx that calculates the indication gesture, Cy, and the pixel coordinate i of each point on the hand region outline line of indication gesture, j is to center of gravity Cx, the distance D of Cy has peaked pixel coordinate i, j with distance D, be defined as indicating the finger tip position Px of gesture, Py.
CN2009101981969A 2009-11-03 2009-11-03 Method for positioning finger tips of directed gestures Expired - Fee Related CN101719015B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2009101981969A CN101719015B (en) 2009-11-03 2009-11-03 Method for positioning finger tips of directed gestures

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2009101981969A CN101719015B (en) 2009-11-03 2009-11-03 Method for positioning finger tips of directed gestures

Publications (2)

Publication Number Publication Date
CN101719015A CN101719015A (en) 2010-06-02
CN101719015B true CN101719015B (en) 2011-08-31

Family

ID=42433597

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2009101981969A Expired - Fee Related CN101719015B (en) 2009-11-03 2009-11-03 Method for positioning finger tips of directed gestures

Country Status (1)

Country Link
CN (1) CN101719015B (en)

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101872244B (en) * 2010-06-25 2011-12-21 中国科学院软件研究所 Method for human-computer interaction based on hand movement and color information of user
CN102402273A (en) * 2010-09-07 2012-04-04 原相科技股份有限公司 Interactive control system, dynamic detection method of object, host device and its control method
CN102063618B (en) * 2011-01-13 2012-10-31 中科芯集成电路股份有限公司 Dynamic gesture identification method in interactive system
US9171200B2 (en) 2011-03-04 2015-10-27 Hewlett-Packard Development Company, L.P. Gestural interaction identification
CN102142084B (en) * 2011-05-06 2012-12-26 北京网尚数字电影院线有限公司 Method for gesture recognition
CN102184021B (en) * 2011-05-27 2013-06-12 华南理工大学 Television man-machine interaction method based on handwriting input and fingertip mouse
CN102253715A (en) * 2011-07-20 2011-11-23 康佳集团股份有限公司 Intelligent image-text human-computer interaction method and intelligent image-text human-computer interaction system of mobile terminal
CN102324019B (en) * 2011-08-12 2014-03-05 浙江大学 Method and system for automatically extracting gesture candidate regions in a video sequence
WO2013056431A1 (en) * 2011-10-18 2013-04-25 Nokia Corporation Methods and apparatuses for gesture recognition
CN102592113B (en) * 2011-12-23 2014-07-30 哈尔滨工业大学深圳研究生院 Rapid identification method for static gestures based on apparent characteristics
CN104115097B (en) * 2012-02-15 2017-06-09 英特尔公司 The automation perceived quality assessment of touch panel device
CN102722701B (en) * 2012-05-24 2014-03-26 清华大学 Visual monitoring method and device in fingerprint collection process
US9213436B2 (en) * 2012-06-20 2015-12-15 Amazon Technologies, Inc. Fingertip location for gesture input
CN102981623B (en) * 2012-11-30 2016-05-18 深圳先进技术研究院 Trigger the method and system of input instruction
CN103218601B (en) * 2013-04-03 2016-08-03 华为技术有限公司 The method and device of detection gesture
CN103226708B (en) * 2013-04-07 2016-06-29 华南理工大学 A kind of multi-model fusion video hand division method based on Kinect
CN103207678A (en) * 2013-04-25 2013-07-17 深圳市中兴移动通信有限公司 Electronic equipment and unblocking method thereof
CN103324291A (en) * 2013-07-12 2013-09-25 安徽工业大学 Method for obtaining position of human body interesting area relative to screen window
CN104793731A (en) * 2015-01-04 2015-07-22 北京君正集成电路股份有限公司 Information input method for wearable device and wearable device
CN104809387B (en) * 2015-03-12 2017-08-29 山东大学 Contactless unlocking method and device based on video image gesture identification
CN105242776A (en) * 2015-09-07 2016-01-13 北京君正集成电路股份有限公司 Control method for intelligent glasses and intelligent glasses
CN106909884B (en) * 2017-01-17 2020-08-04 北京航空航天大学 Hand region detection method and device based on layered structure and deformable part model
CN107357422B (en) * 2017-06-28 2023-04-25 深圳先进技术研究院 Camera-projection interactive touch control method, device and computer readable storage medium
CN109376612B (en) * 2018-09-27 2022-04-22 广东小天才科技有限公司 A method and system for assisting localization learning based on gestures

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1059608A2 (en) * 1999-06-08 2000-12-13 Matsushita Electric Industrial Co., Ltd. Device using eigenspace method for recognizing hand shape and position
US6215890B1 (en) * 1997-09-26 2001-04-10 Matsushita Electric Industrial Co., Ltd. Hand gesture recognizing device
CN101344816A (en) * 2008-08-15 2009-01-14 华南理工大学 Human-computer interaction method and device based on gaze tracking and gesture recognition
CN101437124A (en) * 2008-12-17 2009-05-20 三星电子(中国)研发中心 Method for processing dynamic gesture identification signal facing (to)television set control
CN101567093A (en) * 2009-05-25 2009-10-28 济南大学 Method for initializing three-dimension gesture model

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6215890B1 (en) * 1997-09-26 2001-04-10 Matsushita Electric Industrial Co., Ltd. Hand gesture recognizing device
EP1059608A2 (en) * 1999-06-08 2000-12-13 Matsushita Electric Industrial Co., Ltd. Device using eigenspace method for recognizing hand shape and position
CN101344816A (en) * 2008-08-15 2009-01-14 华南理工大学 Human-computer interaction method and device based on gaze tracking and gesture recognition
CN101437124A (en) * 2008-12-17 2009-05-20 三星电子(中国)研发中心 Method for processing dynamic gesture identification signal facing (to)television set control
CN101567093A (en) * 2009-05-25 2009-10-28 济南大学 Method for initializing three-dimension gesture model

Also Published As

Publication number Publication date
CN101719015A (en) 2010-06-02

Similar Documents

Publication Publication Date Title
CN101719015B (en) Method for positioning finger tips of directed gestures
CN102184021B (en) Television man-machine interaction method based on handwriting input and fingertip mouse
CN103530613B (en) Target person hand gesture interaction method based on monocular video sequence
CN103984928B (en) Finger gesture recognition methods based on depth image
CN106598227B (en) Gesture identification method based on Leap Motion and Kinect
CN104407694B (en) The man-machine interaction method and device of a kind of combination face and gesture control
CN103472916B (en) A kind of man-machine interaction method based on human body gesture identification
CN106250867A (en) A kind of skeleton based on depth data follows the tracks of the implementation method of system
Chandel et al. Hand gesture recognition for sign language recognition: A review
CN103257713B (en) A kind of gesture control method
CN109684959A (en) The recognition methods of video gesture based on Face Detection and deep learning and device
CN107357428A (en) Man-machine interaction method and device based on gesture identification, system
CN112329646A (en) Hand gesture motion direction identification method based on mass center coordinates of hand
CN106200971A (en) Man-machine interactive system device based on gesture identification and operational approach
CN112199015B (en) Intelligent interaction all-in-one machine and writing method and device thereof
CN103995595A (en) Game somatosensory control method based on hand gestures
CN109359566B (en) A Gesture Recognition Method Using Finger Features for Hierarchical Classification
CN106909871A (en) Gesture instruction recognition methods
CN105335711A (en) Fingertip detection method in complex environment
CN106484108A (en) Chinese characters recognition method based on double vision point gesture identification
CN102073878B (en) Non-wearable finger pointing gesture visual identification method
CN107329564B (en) Man-machine finger guessing method based on gesture intelligent perception and man-machine cooperation mechanism
Yang et al. 3D character recognition using binocular camera for medical assist
CN109343694B (en) Gesture recognition system and method for finger-guessing and punching game
CN110134241A (en) Dynamic gesture exchange method based on monocular cam

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
EE01 Entry into force of recordation of patent licensing contract

Assignee: Shanghai Hantu Decoration Co., Ltd.

Assignor: Shanghai University

Contract record no.: 2011310000244

Denomination of invention: Method for positioning finger tips of directed gestures

Granted publication date: 20110831

License type: Exclusive License

Open date: 20100602

Record date: 20111206

CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20110831

Termination date: 20151103

EXPY Termination of patent right or utility model