[go: up one dir, main page]

CN108446032A - A kind of mouse gestures implementation method in projection interactive system - Google Patents

A kind of mouse gestures implementation method in projection interactive system Download PDF

Info

Publication number
CN108446032A
CN108446032A CN201711462719.7A CN201711462719A CN108446032A CN 108446032 A CN108446032 A CN 108446032A CN 201711462719 A CN201711462719 A CN 201711462719A CN 108446032 A CN108446032 A CN 108446032A
Authority
CN
China
Prior art keywords
image
track
line
shape
mouse gestures
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201711462719.7A
Other languages
Chinese (zh)
Other versions
CN108446032B (en
Inventor
邓宏平
汪俊锋
任玮蒙
薛东阳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Anhui Wisdom Gold Tong Technology Co Ltd
Original Assignee
Anhui Wisdom Gold Tong Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Anhui Wisdom Gold Tong Technology Co Ltd filed Critical Anhui Wisdom Gold Tong Technology Co Ltd
Priority to CN201711462719.7A priority Critical patent/CN108446032B/en
Publication of CN108446032A publication Critical patent/CN108446032A/en
Application granted granted Critical
Publication of CN108446032B publication Critical patent/CN108446032B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03547Touch pads, in which fingers can move on a surface
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03545Pens or stylus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/66Analysis of geometric attributes of image moments or centre of gravity
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Geometry (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The present invention relates to the mouse gestures implementation methods in a kind of projection interactive system, include the following steps:The picture of view field is obtained by infrared camera, acquires the image of each frame on picture, and by the image transmitting to computer;For each frame image, finger touch position is obtained using image processing algorithm;Illuminate finger position, by infrared camera capture infrared image, and to the image carry out binary conversion treatment, obtain finger for connected domain;The center point coordinate of connected domain is calculated, and stores finger position coordinate in each frame image;Finger position coordinate is stored in sequence, obtains finger position tracing point;Tracing point is collected, tracing point is analyzed, obtains the shape of track;The shape of track is detected, abnormal trajectory shape is excluded according to desired guiding trajectory shape, obtains mouse gestures.

Description

A kind of mouse gestures implementation method in projection interactive system
Technical field
The present invention relates to mouse gestures technical fields, and in particular to a kind of mouse gestures realization side projected in interactive system Method.
Background technology
It is combined to form projection interactive system using visual sensor and projecting apparatus, is a kind of new interactive mode.It should View field directly can be become touch screen and realize the function of interacting with a computer by mode.In projecting interactive system, projection The projected area of instrument often covers entire blackboard region.This interactive mode advantage is touched relative to traditional liquid crystal capacitance formula Device is controlled, interaction touch-control size is big, can be advantageously applied to teaching field.But in classroom, interactive whiteboard often compares Height, this leads to many teachers of short and small stature, can not touch the higher region of blank, therefore also causes to project interactive system Usage experience be affected.On many softwares, minimizes, maximizes, closes three buttons, be all to concentrate on upper right area Domain.At this time since the position of three buttons is higher, result in being difficult operation.
Invention content
The purpose of the present invention is to provide the mouse gestures implementation methods in a kind of projection interactive system, can be in screen Any one region is solved by drawing mouse gestures to realize the relevant operation to window because can not touch the upper right corner Region, can not be to window operation the problem of.
To achieve the above object, present invention employs following technical schemes:
A kind of mouse gestures implementation method in projection interactive system, includes the following steps:
(1) picture that view field is obtained by infrared camera, acquires the image of each frame on picture, and by the image It is transferred to computer;
(2) it is directed to each frame image, finger touch position is obtained using image processing algorithm;
(3) finger position is illuminated, infrared image is captured by infrared camera, and binary conversion treatment is carried out to the image, Obtain finger for connected domain;
(4) center point coordinate of connected domain is calculated, and stores finger position coordinate in each frame image;
(5) finger position coordinate is stored in sequence, obtains finger position tracing point;
(6) tracing point is collected, tracing point is analyzed, obtains the shape of track;
(7) shape of track is detected, abnormal trajectory shape is excluded according to desired guiding trajectory shape, obtains mouse hand Gesture.
Further, in the step (7), the shape of track is detected, including straight-line detection and non-rectilinear detection.
Further, the straight-line detection, specifically comprises the following steps:
(A1) all tracing points are connected, obtains a lines, by the coordinate value of the lines, obtained corresponding outer Connect rectangle;
(A2) starting point for finding traversal successively on the Far Left of the vertical direction of boundary rectangle, by the starting point according to A scan line is made in the direction parallel with rectangular horizontal side, and the length of the scanned copy line is longer than the length on rectangular horizontal side;
(A3) the intersection point number of every scan line and track is counted, if intersection point there are one every scan line and path lines, Then belong to normal trace shape, if the intersection point of every scan line and path line is more than one, belongs to abnormal trajectory shape.
Further, the lines detection further includes track Linearity surveying, is specifically included:
(A4) every normal trajectory shape first place is connected;
(A5) distance average of all tracing points and this path line first place line in every path line is calculated, if average Value is more than threshold values, then the degree of tracing point deviation straight line is excessive, then is not belonging to straight line, otherwise belongs to straight line.
Further, the straight-line detection further includes that course bearing judges, specifically includes following steps:
It obtains in the queue of track, the coordinate of beginning and end, according to the size of starting point coordinate and terminal point coordinate, calculates track Vertical and horizontal length judges whether mouse gestures operation is normal according to vertical and horizontal length.
Further, the non-rectilinear detection, specifically comprises the following steps:
(B1) line is carried out to all tracing points, track queue is plotted in image;
(B2) each pixel in trace image is traversed, judges to whether there is four foregrounds in the neighborhood territory pixel of some pixel Pixel, and if it exists, then belong to crosspoint;
(B3) number in crosspoint in entire trace image is counted;
(B4) if the number in crosspoint is one, loop configuration judgement is carried out, is otherwise filtered as noise;
Further, the loop configuration detection, includes the following steps:
(B5) using the corresponding coordinate position in current crosspoint, segmentation is scanned to three directions;
(B6) current track structure is divided into three parts, counts the branch of a curve of not torus knot by scanning segmentation image Length, width and the height in portion are judged by the length, width and height, if meeting threshold requirements, are tied for alpha Otherwise structure carries out noise filtering.
Further, the step (B5), three directions respectively level is to the right, straight up and straight down.
As shown from the above technical solution, the mouse gestures implementation method in projection interactive system of the present invention, passes through The mouse gestures system, user can directly utilize finger or light pen, various tracks be drawn on projection medium, to realize To the minimum of window, maximization, various operations.The present invention can be in any one region of screen, by drawing mouse hand Gesture solves to realize relevant operation to window because can not touch upper right comer region, can not be to window operation the problem of, and This mouse gestures is easy to operate, can preferably meet the needs of users
Description of the drawings
Fig. 1 is flow chart of the method for the present invention;
Fig. 2 is alpha segmentation of structures figure of the present invention.
Specific implementation mode
The present invention will be further described below in conjunction with the accompanying drawings:
As shown in Figure 1, the mouse gestures implementation method in a kind of projection interactive system of the present embodiment, specifically includes as follows Step:
S1:The picture of view field is obtained by infrared camera, acquires the image of each frame on picture, and by the image It is transferred to computer;
S2:For each frame image, finger touch position is obtained using image processing algorithm;
S3:Finger position is illuminated, infrared image is captured by infrared camera, and binary conversion treatment is carried out to the image, Obtain finger for connected domain:
Using the infrared laser on wall, finger is illuminated, then infrared camera captures infrared image, to the figure As carrying out binary conversion treatment, the corresponding connected domain of finger is then obtained, the center point coordinate of the connected domain, central point is calculated Calculation formula such as formula (1)-(3) shown in:Wherein, in formula (1), i and divide j do not indicate in connected domain abscissa and The position of ordinate, f (i, j) expressions are in connected domain, the corresponding pixel value of point (i, j).What wherein formula (2) indicated is center The abscissa positions of point, what formula (3) indicated is the position of central point ordinate.
S4:The center point coordinate of connected domain is calculated, and stores finger position coordinate in each frame image;
S5:Finger position coordinate is stored in sequence, obtains finger position tracing point;
S6:Tracing point is collected, tracing point is analyzed, obtains the shape of track:
Using projection interactive software, finger down, during the entire process of being lifted to finger, the seat of all touch points are collected Mark forms track (coordinate queue) sequence.Then, the tracing point queue, is analyzed being fed in subsequent algorithm, To analyze the shape of track, specific window control is gone out by shape analysis and is operated.
S7:The shape of track is detected, abnormal trajectory shape is excluded according to desired guiding trajectory shape, obtains mouse hand Gesture:Mainly distinguish 1 font, alpha shapes.
Using camera acquisition trajectories point image, tracing point is analyzed, obtains the boundary rectangle of track.Then it analyzes The lines of track, the specific method is as follows:
SA1:All tracing points are connected, a lines are obtained.The lines are analyzed and (find all the points The minimum value of x coordinate, maximum value, the minimum value of y, maximum value), it is corresponding outer to obtain its according to above-mentioned two pairs of coordinate points Connect rectangle.
SA2:Find the starting point of traversal successively on the Far Left of the vertical direction of boundary rectangle, adjacent starting point, It is divided into 2 pixels, the starting point of traversal is the top left corner apex of rectangle, and the end point of traversal is lower-left angular vertex, in the step Each starting point, make a scan line all in accordance with the direction parallel with matrix horizontal sides, the length of scan line will be longer than square The length of battle array horizontal sides.
SA3:The number for counting the intersection point of every scan line and track, for all scan lines, if every is swept Retouching line, only there are one intersection points with path line, then are normal, meet the requirement of lines.If intersection point number is more than one, It is unsatisfactory for lines.Since track is 1 word shape, then there are one its intersection points with every scan line.If all scanning Line, detection.For being unsatisfactory for the track of lines, the track Linearity surveying for carrying out next step is needed.
Track Linearity surveying is that have mostly directly to detect track on earth, if meets straight line, avoids equally being one Bar line but be not influence etc. of the line (for example tilt 45 degree straight line) of vertical direction to subsequent operation.It is straight if it is 45 degree Line may not be the meaning of the diminution either amplification window of expression, but if we do not carry out Linearity surveying to it, Then we can be mistakenly considered the operation of diminution or amplification.
SA4:All tracing points are connected;
SA5:Calculate the equation of 2 points of corresponding straight lines of head and the tail, it is known that the coordinate of 2 points of head and the tail is (x respectively1,y1) and (x2,y2), shown in the equation that straight line is met in cartesian coordinate system such as formula (4), by inciting somebody to action two pairs of coordinate points of head and the tail It is brought into formula (4), so as to seek out the linear equation corresponding to 2 points of head and the tail.
Y=kx+b (4)
All tracing points are calculated, the distance of the head and the tail line is arrived.Assuming that the coordinate of a certain tracing point is (xi,yi), then should Shown in distance such as formula (5) of the point to head and the tail line.
All tracing points are calculated to the average value of the distance of line, as shown in formula (6), wherein n is all tracing points Number.
If the average value is more than threshold value (30 pixel), the degree that tracing point deviates straight line is excessive, therefore thinks that it does not belong to In straight line;If belonging to straight line, the walking direction of next step can be entered;If being not belonging to straight line, handed over Crunode detects and loop configuration detection.
Walking direction:90 ° of bigger tracks are deviateed in direction, distinguishes, is as follows:
SA6:It obtains in the queue of track, the coordinate of beginning and end;
Compare starting point ordinate y_start, terminal ordinate y_end.If y_end-y_start > th, direction is Under.If y_start-y_end > th, then direction is upward.Wherein th=100, the longitudinal length of the track indicated herein, such as The longitudinal length of fruit track is too short, then it is assumed that it may not be a mouse gestures operation.
SA7:After obtaining the direction of queue, it is also necessary to judge the difference of abscissa.
Compare starting point abscissa x_start, terminal ordinate x_end.If | x_end-x_start | > th, it cannot Judged, this situation illustrates that line orientations are tilted, and needs setting-out again.Conversely, illustrating that current track is vertical The lines in direction.
To being not belonging to the track of vertical element, needs to judge whether it is alpha shapes, specifically comprise the following steps:
SB1:Line is carried out to all tracing points, track queue is plotted in image;
SB2:Each pixel in trace image is traversed, judges to whether there is four foregrounds in the neighborhood territory pixel of some pixel Pixel, and if it exists, then belong to crosspoint;
As shown in the table, as in the eight neighborhood of current pixel, if there are 4 foreground pixels, for pixel x, 8 neighborhoods Structure, pixel 1-8 are pixels in its 8 neighborhood, if in the eight neighborhood of some pixel, there are 4 pixels, then the point belongs to friendship Crunode.
1 2 3
4 x 5
6 7 8
SB3:Count the number in crosspoint in entire trace image;
SB4:If the number in crosspoint is one, loop configuration judgement is carried out, is otherwise filtered as noise;
Further, the loop configuration detection, includes the following steps:
SB5:Using the corresponding coordinate position in current crosspoint, it is scanned segmentation to three directions, i.e., level to the right, is erected Directly upwards and straight down, as shown in Figure 2;
SB6:Scanning segmentation image, three parts, i.e. a loop configuration and two curves are divided by current track structure Branch counts their length for two branch of a curve parts of not loop configuration.If the length of Liang Ge branches is less than Threshold value (such as 50), then it is assumed that the production length of its Liang Ge branch is too short, is unsatisfactory for alpha structures, to as noise It filters out, obtains the width and height of the boundary rectangle of the loop configuration.Wide and height (has if it is less than threshold value (such as 50) as noise This possible track is not that user is formed when doing mouse gestures, the track that other reasons are formed) filtering.If Liang Ge branches Length is less than threshold value (such as 50), then it is assumed that the production length of its Liang Ge branch is too short, is unsatisfactory for alpha structures, thus by it As noise filtering.
After abovementioned steps are completed to the identification of current track shape, so that it may with according to recognition result, to present procedure Carry out corresponding operation.
Embodiment described above is only that the preferred embodiment of the present invention is described, not to the model of the present invention It encloses and is defined, under the premise of not departing from design spirit of the present invention, technical side of the those of ordinary skill in the art to the present invention The various modifications and improvement that case is made should all be fallen into the protection domain of claims of the present invention determination.

Claims (8)

1. the mouse gestures implementation method in a kind of projection interactive system, which is characterized in that include the following steps:
(1)The picture of view field is obtained by infrared camera, acquires the image of each frame on picture, and by the image transmitting To computer;
(2)For each frame image, finger touch position is obtained using image processing algorithm;
(3)Finger position is illuminated, infrared image is captured by infrared camera, and binary conversion treatment is carried out to the image, obtained Finger for connected domain;
(4)The center point coordinate of connected domain is calculated, and stores finger position coordinate in each frame image;
(5)Finger position coordinate is stored in sequence, obtains finger position tracing point;
(6)Tracing point is collected, tracing point is analyzed, obtains the shape of track;
(7)The shape of track is detected, abnormal trajectory shape is excluded according to desired guiding trajectory shape, obtains mouse gestures.
2. the mouse gestures implementation method in projection interactive system according to claim 1, it is characterised in that:The step (7)In, the shape of track is detected, including straight-line detection and non-rectilinear detection.
3. the mouse gestures implementation method in projection interactive system according to claim 2, it is characterised in that:The straight line Detection, specifically comprises the following steps:
(A1)All tracing points are connected, a lines are obtained, by the coordinate value of the lines, obtain corresponding external square Shape;
(A2)The starting point for finding traversal successively on the Far Left of the vertical direction of boundary rectangle, by the starting point according to square A scan line is made in the parallel direction of shape horizontal sides, and the length of the scanned copy line is longer than the length on rectangular horizontal side;
(A3)Every scan line and the intersection point number of track are counted, only there are one intersection points if every scan line is with path line, belong to Belong to abnormal trajectory shape if the intersection point of every scan line and path line is more than one in normal trace shape.
4. the mouse gestures implementation method in projection interactive system according to claim 3, it is characterised in that:The lines Property detection further include track Linearity surveying, specifically include:
(A4)Every normal trajectory shape first place is connected;
(A5)The distance average for calculating all tracing points and this path line first place line in every path line, if average value is big In threshold values, then the degree of tracing point deviation straight line is excessive, then is not belonging to straight line, otherwise belongs to straight line.
5. the mouse gestures implementation method in projection interactive system according to claim 4, it is characterised in that:The straight line Detection further includes that course bearing judges, specifically includes following steps:
It obtains in the queue of track, the coordinate of beginning and end, according to the size of starting point coordinate and terminal point coordinate, it is longitudinal to calculate track And lateral length, judge whether mouse gestures operation is normal according to vertical and horizontal length.
6. the mouse gestures implementation method in projection interactive system according to claim 2, it is characterised in that:It is described non-straight Line detects, and specifically comprises the following steps:
(B1)Line is carried out to all tracing points, track queue is plotted in image;
(B2)Each pixel in trace image is traversed, judges to whether there is four foreground pictures in the neighborhood territory pixel of some pixel Element, and if it exists, then belong to crosspoint;
(B3)Count the number in crosspoint in entire trace image;
(B4)If the number in crosspoint is one, loop configuration judgement is carried out, is otherwise filtered as noise.
7. the mouse gestures implementation method in projection interactive system according to claim 6, it is characterised in that:The annular Structure detection includes the following steps:
(B5)Using the corresponding coordinate position in current crosspoint, segmentation is scanned to three directions;
(B6)Scanning segmentation image, three parts are divided by current track structure, the not no branch of a curve portion of torus knot of statistics Length, width and height are judged by the length, width and height, no for alpha structures if meeting threshold requirements Then carry out noise filtering.
8. the mouse gestures implementation method in projection interactive system according to claim 7, it is characterised in that:The step (B5), three directions respectively level is to the right, straight up and straight down.
CN201711462719.7A 2017-12-28 2017-12-28 Mouse gesture implementation method in projection interaction system Active CN108446032B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711462719.7A CN108446032B (en) 2017-12-28 2017-12-28 Mouse gesture implementation method in projection interaction system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711462719.7A CN108446032B (en) 2017-12-28 2017-12-28 Mouse gesture implementation method in projection interaction system

Publications (2)

Publication Number Publication Date
CN108446032A true CN108446032A (en) 2018-08-24
CN108446032B CN108446032B (en) 2022-03-08

Family

ID=63190829

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711462719.7A Active CN108446032B (en) 2017-12-28 2017-12-28 Mouse gesture implementation method in projection interaction system

Country Status (1)

Country Link
CN (1) CN108446032B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110633027A (en) * 2019-09-16 2019-12-31 广东小天才科技有限公司 Point reading implementation method, system, computer equipment and storage medium
CN110942092A (en) * 2019-11-21 2020-03-31 大庆思特传媒科技有限公司 Graphic image recognition method and recognition system
CN112115853A (en) * 2020-09-17 2020-12-22 西安羚控电子科技有限公司 Gesture recognition method and device, computer storage medium and electronic equipment
CN113610944A (en) * 2021-07-30 2021-11-05 新线科技有限公司 A line drawing method, device, equipment and storage medium
CN114460881A (en) * 2022-02-11 2022-05-10 广东好太太智能家居有限公司 Clothes drying equipment control device, method and laundry drying equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102236438A (en) * 2010-04-21 2011-11-09 上海三旗通信科技有限公司 Algorithm and operating mode for identifying track on touch pad of mobile terminal
CN103413137A (en) * 2013-08-05 2013-11-27 浙江大学 Interaction gesture motion trail partition method based on multiple rules
US20150160784A1 (en) * 2006-02-28 2015-06-11 Microsoft Corporation Compact Interactive Tabletop with Projection-Vision
CN104750250A (en) * 2015-03-04 2015-07-01 深圳市祈锦通信技术有限公司 Gesture-interaction-supported projection method and system
CN104850232A (en) * 2015-05-27 2015-08-19 济南大学 Method for acquiring remote gesture tracks under camera conditions
CN105260021A (en) * 2015-10-15 2016-01-20 深圳市祈锦通信技术有限公司 Intelligent interactive projection system

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150160784A1 (en) * 2006-02-28 2015-06-11 Microsoft Corporation Compact Interactive Tabletop with Projection-Vision
CN102236438A (en) * 2010-04-21 2011-11-09 上海三旗通信科技有限公司 Algorithm and operating mode for identifying track on touch pad of mobile terminal
CN103413137A (en) * 2013-08-05 2013-11-27 浙江大学 Interaction gesture motion trail partition method based on multiple rules
CN104750250A (en) * 2015-03-04 2015-07-01 深圳市祈锦通信技术有限公司 Gesture-interaction-supported projection method and system
CN104850232A (en) * 2015-05-27 2015-08-19 济南大学 Method for acquiring remote gesture tracks under camera conditions
CN105260021A (en) * 2015-10-15 2016-01-20 深圳市祈锦通信技术有限公司 Intelligent interactive projection system

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110633027A (en) * 2019-09-16 2019-12-31 广东小天才科技有限公司 Point reading implementation method, system, computer equipment and storage medium
CN110942092A (en) * 2019-11-21 2020-03-31 大庆思特传媒科技有限公司 Graphic image recognition method and recognition system
CN110942092B (en) * 2019-11-21 2023-08-08 大庆思特传媒科技有限公司 Graphic image recognition method and recognition system
CN112115853A (en) * 2020-09-17 2020-12-22 西安羚控电子科技有限公司 Gesture recognition method and device, computer storage medium and electronic equipment
CN113610944A (en) * 2021-07-30 2021-11-05 新线科技有限公司 A line drawing method, device, equipment and storage medium
EP4365847A4 (en) * 2021-07-30 2024-10-23 Shenzhen Hitevision Technology Co., Ltd. LINE DRAWING METHOD AND APPARATUS, ELECTRONIC DEVICE, AND COMPUTER READABLE STORAGE MEDIUM
CN114460881A (en) * 2022-02-11 2022-05-10 广东好太太智能家居有限公司 Clothes drying equipment control device, method and laundry drying equipment

Also Published As

Publication number Publication date
CN108446032B (en) 2022-03-08

Similar Documents

Publication Publication Date Title
CN108446032A (en) A kind of mouse gestures implementation method in projection interactive system
US10747988B2 (en) Method and device for face tracking and smart terminal
EP2365420B1 (en) System and method for hand gesture recognition for remote control of an internet protocol TV
US10437466B2 (en) Formula inputting method and apparatus
CN103150019A (en) Handwriting input system and method
CN104809387B (en) Contactless unlocking method and device based on video image gesture identification
CN107688779A (en) A kind of robot gesture interaction method and apparatus based on RGBD camera depth images
CN105528592A (en) Fingerprint scanning method and device and gesture recognition method and device
CN101661329A (en) Operating control method and device of intelligent terminal
US20130212511A1 (en) Apparatus and method for guiding handwriting input for handwriting recognition
CN105138990A (en) Single-camera-based gesture convex hull detection and palm positioning method
CN106874893A (en) A kind of statistical method and system of paper fraction
CN205563605U (en) Fingerprint scanning device and gesture recognition device
CN106104456A (en) Messaging device, information processing method and message handling program
US10551934B2 (en) Gesture recognition method, apparatus and device, computer program product therefor
CN105468200B (en) A kind of method, apparatus and fingerprint equipment of fingerprint touch-control
CN104166843A (en) Document image source judgment method based on linear continuity
CN109782996B (en) Three-finger coaxial splitting point merging method, touch device and touch display device
CN103744559B (en) The multipoint positioning method of infrared touch panel
JP2002515144A (en) Handwriting input stroke division method
CN106097281A (en) A kind of calibration maps for projecting interactive system and demarcate detection method
CN106648171B (en) A kind of interactive system and method based on lettering pen
CN104731396A (en) Magic pen and display control method thereof
CN114913527A (en) Calibration method and system for camera calibration of scanning translation pen
CN107506133A (en) Project the operation trace response method and system of touch-control system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 230000 Yafu Park, Juchao Economic Development Zone, Chaohu City, Hefei City, Anhui Province

Applicant after: ANHUI HUISHI JINTONG TECHNOLOGY Co.,Ltd.

Address before: 102, room 602, C District, Hefei National University, Mount Huangshan Road, 230000 Hefei Road, Anhui, China

Applicant before: ANHUI HUISHI JINTONG TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant