CN108446032A - A kind of mouse gestures implementation method in projection interactive system - Google Patents
A kind of mouse gestures implementation method in projection interactive system Download PDFInfo
- Publication number
- CN108446032A CN108446032A CN201711462719.7A CN201711462719A CN108446032A CN 108446032 A CN108446032 A CN 108446032A CN 201711462719 A CN201711462719 A CN 201711462719A CN 108446032 A CN108446032 A CN 108446032A
- Authority
- CN
- China
- Prior art keywords
- image
- track
- line
- shape
- mouse gestures
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000002452 interceptive effect Effects 0.000 title claims abstract description 23
- 238000000034 method Methods 0.000 title claims abstract description 20
- 230000002159 abnormal effect Effects 0.000 claims abstract description 6
- 238000006243 chemical reaction Methods 0.000 claims abstract description 5
- 238000001514 detection method Methods 0.000 claims description 16
- 230000011218 segmentation Effects 0.000 claims description 7
- 238000001914 filtration Methods 0.000 claims description 4
- 230000003321 amplification Effects 0.000 description 2
- 230000014509 gene expression Effects 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000003199 nucleic acid amplification method Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000012141 concentrate Substances 0.000 description 1
- 235000013399 edible fruits Nutrition 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03547—Touch pads, in which fingers can move on a surface
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03545—Pens or stylus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/042—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/66—Analysis of geometric attributes of image moments or centre of gravity
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30241—Trajectory
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Geometry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
The present invention relates to the mouse gestures implementation methods in a kind of projection interactive system, include the following steps:The picture of view field is obtained by infrared camera, acquires the image of each frame on picture, and by the image transmitting to computer;For each frame image, finger touch position is obtained using image processing algorithm;Illuminate finger position, by infrared camera capture infrared image, and to the image carry out binary conversion treatment, obtain finger for connected domain;The center point coordinate of connected domain is calculated, and stores finger position coordinate in each frame image;Finger position coordinate is stored in sequence, obtains finger position tracing point;Tracing point is collected, tracing point is analyzed, obtains the shape of track;The shape of track is detected, abnormal trajectory shape is excluded according to desired guiding trajectory shape, obtains mouse gestures.
Description
Technical field
The present invention relates to mouse gestures technical fields, and in particular to a kind of mouse gestures realization side projected in interactive system
Method.
Background technology
It is combined to form projection interactive system using visual sensor and projecting apparatus, is a kind of new interactive mode.It should
View field directly can be become touch screen and realize the function of interacting with a computer by mode.In projecting interactive system, projection
The projected area of instrument often covers entire blackboard region.This interactive mode advantage is touched relative to traditional liquid crystal capacitance formula
Device is controlled, interaction touch-control size is big, can be advantageously applied to teaching field.But in classroom, interactive whiteboard often compares
Height, this leads to many teachers of short and small stature, can not touch the higher region of blank, therefore also causes to project interactive system
Usage experience be affected.On many softwares, minimizes, maximizes, closes three buttons, be all to concentrate on upper right area
Domain.At this time since the position of three buttons is higher, result in being difficult operation.
Invention content
The purpose of the present invention is to provide the mouse gestures implementation methods in a kind of projection interactive system, can be in screen
Any one region is solved by drawing mouse gestures to realize the relevant operation to window because can not touch the upper right corner
Region, can not be to window operation the problem of.
To achieve the above object, present invention employs following technical schemes:
A kind of mouse gestures implementation method in projection interactive system, includes the following steps:
(1) picture that view field is obtained by infrared camera, acquires the image of each frame on picture, and by the image
It is transferred to computer;
(2) it is directed to each frame image, finger touch position is obtained using image processing algorithm;
(3) finger position is illuminated, infrared image is captured by infrared camera, and binary conversion treatment is carried out to the image,
Obtain finger for connected domain;
(4) center point coordinate of connected domain is calculated, and stores finger position coordinate in each frame image;
(5) finger position coordinate is stored in sequence, obtains finger position tracing point;
(6) tracing point is collected, tracing point is analyzed, obtains the shape of track;
(7) shape of track is detected, abnormal trajectory shape is excluded according to desired guiding trajectory shape, obtains mouse hand
Gesture.
Further, in the step (7), the shape of track is detected, including straight-line detection and non-rectilinear detection.
Further, the straight-line detection, specifically comprises the following steps:
(A1) all tracing points are connected, obtains a lines, by the coordinate value of the lines, obtained corresponding outer
Connect rectangle;
(A2) starting point for finding traversal successively on the Far Left of the vertical direction of boundary rectangle, by the starting point according to
A scan line is made in the direction parallel with rectangular horizontal side, and the length of the scanned copy line is longer than the length on rectangular horizontal side;
(A3) the intersection point number of every scan line and track is counted, if intersection point there are one every scan line and path lines,
Then belong to normal trace shape, if the intersection point of every scan line and path line is more than one, belongs to abnormal trajectory shape.
Further, the lines detection further includes track Linearity surveying, is specifically included:
(A4) every normal trajectory shape first place is connected;
(A5) distance average of all tracing points and this path line first place line in every path line is calculated, if average
Value is more than threshold values, then the degree of tracing point deviation straight line is excessive, then is not belonging to straight line, otherwise belongs to straight line.
Further, the straight-line detection further includes that course bearing judges, specifically includes following steps:
It obtains in the queue of track, the coordinate of beginning and end, according to the size of starting point coordinate and terminal point coordinate, calculates track
Vertical and horizontal length judges whether mouse gestures operation is normal according to vertical and horizontal length.
Further, the non-rectilinear detection, specifically comprises the following steps:
(B1) line is carried out to all tracing points, track queue is plotted in image;
(B2) each pixel in trace image is traversed, judges to whether there is four foregrounds in the neighborhood territory pixel of some pixel
Pixel, and if it exists, then belong to crosspoint;
(B3) number in crosspoint in entire trace image is counted;
(B4) if the number in crosspoint is one, loop configuration judgement is carried out, is otherwise filtered as noise;
Further, the loop configuration detection, includes the following steps:
(B5) using the corresponding coordinate position in current crosspoint, segmentation is scanned to three directions;
(B6) current track structure is divided into three parts, counts the branch of a curve of not torus knot by scanning segmentation image
Length, width and the height in portion are judged by the length, width and height, if meeting threshold requirements, are tied for alpha
Otherwise structure carries out noise filtering.
Further, the step (B5), three directions respectively level is to the right, straight up and straight down.
As shown from the above technical solution, the mouse gestures implementation method in projection interactive system of the present invention, passes through
The mouse gestures system, user can directly utilize finger or light pen, various tracks be drawn on projection medium, to realize
To the minimum of window, maximization, various operations.The present invention can be in any one region of screen, by drawing mouse hand
Gesture solves to realize relevant operation to window because can not touch upper right comer region, can not be to window operation the problem of, and
This mouse gestures is easy to operate, can preferably meet the needs of users
Description of the drawings
Fig. 1 is flow chart of the method for the present invention;
Fig. 2 is alpha segmentation of structures figure of the present invention.
Specific implementation mode
The present invention will be further described below in conjunction with the accompanying drawings:
As shown in Figure 1, the mouse gestures implementation method in a kind of projection interactive system of the present embodiment, specifically includes as follows
Step:
S1:The picture of view field is obtained by infrared camera, acquires the image of each frame on picture, and by the image
It is transferred to computer;
S2:For each frame image, finger touch position is obtained using image processing algorithm;
S3:Finger position is illuminated, infrared image is captured by infrared camera, and binary conversion treatment is carried out to the image,
Obtain finger for connected domain:
Using the infrared laser on wall, finger is illuminated, then infrared camera captures infrared image, to the figure
As carrying out binary conversion treatment, the corresponding connected domain of finger is then obtained, the center point coordinate of the connected domain, central point is calculated
Calculation formula such as formula (1)-(3) shown in:Wherein, in formula (1), i and divide j do not indicate in connected domain abscissa and
The position of ordinate, f (i, j) expressions are in connected domain, the corresponding pixel value of point (i, j).What wherein formula (2) indicated is center
The abscissa positions of point, what formula (3) indicated is the position of central point ordinate.
S4:The center point coordinate of connected domain is calculated, and stores finger position coordinate in each frame image;
S5:Finger position coordinate is stored in sequence, obtains finger position tracing point;
S6:Tracing point is collected, tracing point is analyzed, obtains the shape of track:
Using projection interactive software, finger down, during the entire process of being lifted to finger, the seat of all touch points are collected
Mark forms track (coordinate queue) sequence.Then, the tracing point queue, is analyzed being fed in subsequent algorithm,
To analyze the shape of track, specific window control is gone out by shape analysis and is operated.
S7:The shape of track is detected, abnormal trajectory shape is excluded according to desired guiding trajectory shape, obtains mouse hand
Gesture:Mainly distinguish 1 font, alpha shapes.
Using camera acquisition trajectories point image, tracing point is analyzed, obtains the boundary rectangle of track.Then it analyzes
The lines of track, the specific method is as follows:
SA1:All tracing points are connected, a lines are obtained.The lines are analyzed and (find all the points
The minimum value of x coordinate, maximum value, the minimum value of y, maximum value), it is corresponding outer to obtain its according to above-mentioned two pairs of coordinate points
Connect rectangle.
SA2:Find the starting point of traversal successively on the Far Left of the vertical direction of boundary rectangle, adjacent starting point,
It is divided into 2 pixels, the starting point of traversal is the top left corner apex of rectangle, and the end point of traversal is lower-left angular vertex, in the step
Each starting point, make a scan line all in accordance with the direction parallel with matrix horizontal sides, the length of scan line will be longer than square
The length of battle array horizontal sides.
SA3:The number for counting the intersection point of every scan line and track, for all scan lines, if every is swept
Retouching line, only there are one intersection points with path line, then are normal, meet the requirement of lines.If intersection point number is more than one,
It is unsatisfactory for lines.Since track is 1 word shape, then there are one its intersection points with every scan line.If all scanning
Line, detection.For being unsatisfactory for the track of lines, the track Linearity surveying for carrying out next step is needed.
Track Linearity surveying is that have mostly directly to detect track on earth, if meets straight line, avoids equally being one
Bar line but be not influence etc. of the line (for example tilt 45 degree straight line) of vertical direction to subsequent operation.It is straight if it is 45 degree
Line may not be the meaning of the diminution either amplification window of expression, but if we do not carry out Linearity surveying to it,
Then we can be mistakenly considered the operation of diminution or amplification.
SA4:All tracing points are connected;
SA5:Calculate the equation of 2 points of corresponding straight lines of head and the tail, it is known that the coordinate of 2 points of head and the tail is (x respectively1,y1) and
(x2,y2), shown in the equation that straight line is met in cartesian coordinate system such as formula (4), by inciting somebody to action two pairs of coordinate points of head and the tail
It is brought into formula (4), so as to seek out the linear equation corresponding to 2 points of head and the tail.
Y=kx+b (4)
All tracing points are calculated, the distance of the head and the tail line is arrived.Assuming that the coordinate of a certain tracing point is (xi,yi), then should
Shown in distance such as formula (5) of the point to head and the tail line.
All tracing points are calculated to the average value of the distance of line, as shown in formula (6), wherein n is all tracing points
Number.
If the average value is more than threshold value (30 pixel), the degree that tracing point deviates straight line is excessive, therefore thinks that it does not belong to
In straight line;If belonging to straight line, the walking direction of next step can be entered;If being not belonging to straight line, handed over
Crunode detects and loop configuration detection.
Walking direction:90 ° of bigger tracks are deviateed in direction, distinguishes, is as follows:
SA6:It obtains in the queue of track, the coordinate of beginning and end;
Compare starting point ordinate y_start, terminal ordinate y_end.If y_end-y_start > th, direction is
Under.If y_start-y_end > th, then direction is upward.Wherein th=100, the longitudinal length of the track indicated herein, such as
The longitudinal length of fruit track is too short, then it is assumed that it may not be a mouse gestures operation.
SA7:After obtaining the direction of queue, it is also necessary to judge the difference of abscissa.
Compare starting point abscissa x_start, terminal ordinate x_end.If | x_end-x_start | > th, it cannot
Judged, this situation illustrates that line orientations are tilted, and needs setting-out again.Conversely, illustrating that current track is vertical
The lines in direction.
To being not belonging to the track of vertical element, needs to judge whether it is alpha shapes, specifically comprise the following steps:
SB1:Line is carried out to all tracing points, track queue is plotted in image;
SB2:Each pixel in trace image is traversed, judges to whether there is four foregrounds in the neighborhood territory pixel of some pixel
Pixel, and if it exists, then belong to crosspoint;
As shown in the table, as in the eight neighborhood of current pixel, if there are 4 foreground pixels, for pixel x, 8 neighborhoods
Structure, pixel 1-8 are pixels in its 8 neighborhood, if in the eight neighborhood of some pixel, there are 4 pixels, then the point belongs to friendship
Crunode.
1 | 2 | 3 |
4 | x | 5 |
6 | 7 | 8 |
SB3:Count the number in crosspoint in entire trace image;
SB4:If the number in crosspoint is one, loop configuration judgement is carried out, is otherwise filtered as noise;
Further, the loop configuration detection, includes the following steps:
SB5:Using the corresponding coordinate position in current crosspoint, it is scanned segmentation to three directions, i.e., level to the right, is erected
Directly upwards and straight down, as shown in Figure 2;
SB6:Scanning segmentation image, three parts, i.e. a loop configuration and two curves are divided by current track structure
Branch counts their length for two branch of a curve parts of not loop configuration.If the length of Liang Ge branches is less than
Threshold value (such as 50), then it is assumed that the production length of its Liang Ge branch is too short, is unsatisfactory for alpha structures, to as noise
It filters out, obtains the width and height of the boundary rectangle of the loop configuration.Wide and height (has if it is less than threshold value (such as 50) as noise
This possible track is not that user is formed when doing mouse gestures, the track that other reasons are formed) filtering.If Liang Ge branches
Length is less than threshold value (such as 50), then it is assumed that the production length of its Liang Ge branch is too short, is unsatisfactory for alpha structures, thus by it
As noise filtering.
After abovementioned steps are completed to the identification of current track shape, so that it may with according to recognition result, to present procedure
Carry out corresponding operation.
Embodiment described above is only that the preferred embodiment of the present invention is described, not to the model of the present invention
It encloses and is defined, under the premise of not departing from design spirit of the present invention, technical side of the those of ordinary skill in the art to the present invention
The various modifications and improvement that case is made should all be fallen into the protection domain of claims of the present invention determination.
Claims (8)
1. the mouse gestures implementation method in a kind of projection interactive system, which is characterized in that include the following steps:
(1)The picture of view field is obtained by infrared camera, acquires the image of each frame on picture, and by the image transmitting
To computer;
(2)For each frame image, finger touch position is obtained using image processing algorithm;
(3)Finger position is illuminated, infrared image is captured by infrared camera, and binary conversion treatment is carried out to the image, obtained
Finger for connected domain;
(4)The center point coordinate of connected domain is calculated, and stores finger position coordinate in each frame image;
(5)Finger position coordinate is stored in sequence, obtains finger position tracing point;
(6)Tracing point is collected, tracing point is analyzed, obtains the shape of track;
(7)The shape of track is detected, abnormal trajectory shape is excluded according to desired guiding trajectory shape, obtains mouse gestures.
2. the mouse gestures implementation method in projection interactive system according to claim 1, it is characterised in that:The step
(7)In, the shape of track is detected, including straight-line detection and non-rectilinear detection.
3. the mouse gestures implementation method in projection interactive system according to claim 2, it is characterised in that:The straight line
Detection, specifically comprises the following steps:
(A1)All tracing points are connected, a lines are obtained, by the coordinate value of the lines, obtain corresponding external square
Shape;
(A2)The starting point for finding traversal successively on the Far Left of the vertical direction of boundary rectangle, by the starting point according to square
A scan line is made in the parallel direction of shape horizontal sides, and the length of the scanned copy line is longer than the length on rectangular horizontal side;
(A3)Every scan line and the intersection point number of track are counted, only there are one intersection points if every scan line is with path line, belong to
Belong to abnormal trajectory shape if the intersection point of every scan line and path line is more than one in normal trace shape.
4. the mouse gestures implementation method in projection interactive system according to claim 3, it is characterised in that:The lines
Property detection further include track Linearity surveying, specifically include:
(A4)Every normal trajectory shape first place is connected;
(A5)The distance average for calculating all tracing points and this path line first place line in every path line, if average value is big
In threshold values, then the degree of tracing point deviation straight line is excessive, then is not belonging to straight line, otherwise belongs to straight line.
5. the mouse gestures implementation method in projection interactive system according to claim 4, it is characterised in that:The straight line
Detection further includes that course bearing judges, specifically includes following steps:
It obtains in the queue of track, the coordinate of beginning and end, according to the size of starting point coordinate and terminal point coordinate, it is longitudinal to calculate track
And lateral length, judge whether mouse gestures operation is normal according to vertical and horizontal length.
6. the mouse gestures implementation method in projection interactive system according to claim 2, it is characterised in that:It is described non-straight
Line detects, and specifically comprises the following steps:
(B1)Line is carried out to all tracing points, track queue is plotted in image;
(B2)Each pixel in trace image is traversed, judges to whether there is four foreground pictures in the neighborhood territory pixel of some pixel
Element, and if it exists, then belong to crosspoint;
(B3)Count the number in crosspoint in entire trace image;
(B4)If the number in crosspoint is one, loop configuration judgement is carried out, is otherwise filtered as noise.
7. the mouse gestures implementation method in projection interactive system according to claim 6, it is characterised in that:The annular
Structure detection includes the following steps:
(B5)Using the corresponding coordinate position in current crosspoint, segmentation is scanned to three directions;
(B6)Scanning segmentation image, three parts are divided by current track structure, the not no branch of a curve portion of torus knot of statistics
Length, width and height are judged by the length, width and height, no for alpha structures if meeting threshold requirements
Then carry out noise filtering.
8. the mouse gestures implementation method in projection interactive system according to claim 7, it is characterised in that:The step
(B5), three directions respectively level is to the right, straight up and straight down.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711462719.7A CN108446032B (en) | 2017-12-28 | 2017-12-28 | Mouse gesture implementation method in projection interaction system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711462719.7A CN108446032B (en) | 2017-12-28 | 2017-12-28 | Mouse gesture implementation method in projection interaction system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108446032A true CN108446032A (en) | 2018-08-24 |
CN108446032B CN108446032B (en) | 2022-03-08 |
Family
ID=63190829
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711462719.7A Active CN108446032B (en) | 2017-12-28 | 2017-12-28 | Mouse gesture implementation method in projection interaction system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108446032B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110633027A (en) * | 2019-09-16 | 2019-12-31 | 广东小天才科技有限公司 | Point reading implementation method, system, computer equipment and storage medium |
CN110942092A (en) * | 2019-11-21 | 2020-03-31 | 大庆思特传媒科技有限公司 | Graphic image recognition method and recognition system |
CN112115853A (en) * | 2020-09-17 | 2020-12-22 | 西安羚控电子科技有限公司 | Gesture recognition method and device, computer storage medium and electronic equipment |
CN113610944A (en) * | 2021-07-30 | 2021-11-05 | 新线科技有限公司 | A line drawing method, device, equipment and storage medium |
CN114460881A (en) * | 2022-02-11 | 2022-05-10 | 广东好太太智能家居有限公司 | Clothes drying equipment control device, method and laundry drying equipment |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102236438A (en) * | 2010-04-21 | 2011-11-09 | 上海三旗通信科技有限公司 | Algorithm and operating mode for identifying track on touch pad of mobile terminal |
CN103413137A (en) * | 2013-08-05 | 2013-11-27 | 浙江大学 | Interaction gesture motion trail partition method based on multiple rules |
US20150160784A1 (en) * | 2006-02-28 | 2015-06-11 | Microsoft Corporation | Compact Interactive Tabletop with Projection-Vision |
CN104750250A (en) * | 2015-03-04 | 2015-07-01 | 深圳市祈锦通信技术有限公司 | Gesture-interaction-supported projection method and system |
CN104850232A (en) * | 2015-05-27 | 2015-08-19 | 济南大学 | Method for acquiring remote gesture tracks under camera conditions |
CN105260021A (en) * | 2015-10-15 | 2016-01-20 | 深圳市祈锦通信技术有限公司 | Intelligent interactive projection system |
-
2017
- 2017-12-28 CN CN201711462719.7A patent/CN108446032B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150160784A1 (en) * | 2006-02-28 | 2015-06-11 | Microsoft Corporation | Compact Interactive Tabletop with Projection-Vision |
CN102236438A (en) * | 2010-04-21 | 2011-11-09 | 上海三旗通信科技有限公司 | Algorithm and operating mode for identifying track on touch pad of mobile terminal |
CN103413137A (en) * | 2013-08-05 | 2013-11-27 | 浙江大学 | Interaction gesture motion trail partition method based on multiple rules |
CN104750250A (en) * | 2015-03-04 | 2015-07-01 | 深圳市祈锦通信技术有限公司 | Gesture-interaction-supported projection method and system |
CN104850232A (en) * | 2015-05-27 | 2015-08-19 | 济南大学 | Method for acquiring remote gesture tracks under camera conditions |
CN105260021A (en) * | 2015-10-15 | 2016-01-20 | 深圳市祈锦通信技术有限公司 | Intelligent interactive projection system |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110633027A (en) * | 2019-09-16 | 2019-12-31 | 广东小天才科技有限公司 | Point reading implementation method, system, computer equipment and storage medium |
CN110942092A (en) * | 2019-11-21 | 2020-03-31 | 大庆思特传媒科技有限公司 | Graphic image recognition method and recognition system |
CN110942092B (en) * | 2019-11-21 | 2023-08-08 | 大庆思特传媒科技有限公司 | Graphic image recognition method and recognition system |
CN112115853A (en) * | 2020-09-17 | 2020-12-22 | 西安羚控电子科技有限公司 | Gesture recognition method and device, computer storage medium and electronic equipment |
CN113610944A (en) * | 2021-07-30 | 2021-11-05 | 新线科技有限公司 | A line drawing method, device, equipment and storage medium |
EP4365847A4 (en) * | 2021-07-30 | 2024-10-23 | Shenzhen Hitevision Technology Co., Ltd. | LINE DRAWING METHOD AND APPARATUS, ELECTRONIC DEVICE, AND COMPUTER READABLE STORAGE MEDIUM |
CN114460881A (en) * | 2022-02-11 | 2022-05-10 | 广东好太太智能家居有限公司 | Clothes drying equipment control device, method and laundry drying equipment |
Also Published As
Publication number | Publication date |
---|---|
CN108446032B (en) | 2022-03-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108446032A (en) | A kind of mouse gestures implementation method in projection interactive system | |
US10747988B2 (en) | Method and device for face tracking and smart terminal | |
EP2365420B1 (en) | System and method for hand gesture recognition for remote control of an internet protocol TV | |
US10437466B2 (en) | Formula inputting method and apparatus | |
CN103150019A (en) | Handwriting input system and method | |
CN104809387B (en) | Contactless unlocking method and device based on video image gesture identification | |
CN107688779A (en) | A kind of robot gesture interaction method and apparatus based on RGBD camera depth images | |
CN105528592A (en) | Fingerprint scanning method and device and gesture recognition method and device | |
CN101661329A (en) | Operating control method and device of intelligent terminal | |
US20130212511A1 (en) | Apparatus and method for guiding handwriting input for handwriting recognition | |
CN105138990A (en) | Single-camera-based gesture convex hull detection and palm positioning method | |
CN106874893A (en) | A kind of statistical method and system of paper fraction | |
CN205563605U (en) | Fingerprint scanning device and gesture recognition device | |
CN106104456A (en) | Messaging device, information processing method and message handling program | |
US10551934B2 (en) | Gesture recognition method, apparatus and device, computer program product therefor | |
CN105468200B (en) | A kind of method, apparatus and fingerprint equipment of fingerprint touch-control | |
CN104166843A (en) | Document image source judgment method based on linear continuity | |
CN109782996B (en) | Three-finger coaxial splitting point merging method, touch device and touch display device | |
CN103744559B (en) | The multipoint positioning method of infrared touch panel | |
JP2002515144A (en) | Handwriting input stroke division method | |
CN106097281A (en) | A kind of calibration maps for projecting interactive system and demarcate detection method | |
CN106648171B (en) | A kind of interactive system and method based on lettering pen | |
CN104731396A (en) | Magic pen and display control method thereof | |
CN114913527A (en) | Calibration method and system for camera calibration of scanning translation pen | |
CN107506133A (en) | Project the operation trace response method and system of touch-control system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
CB02 | Change of applicant information | ||
CB02 | Change of applicant information |
Address after: 230000 Yafu Park, Juchao Economic Development Zone, Chaohu City, Hefei City, Anhui Province Applicant after: ANHUI HUISHI JINTONG TECHNOLOGY Co.,Ltd. Address before: 102, room 602, C District, Hefei National University, Mount Huangshan Road, 230000 Hefei Road, Anhui, China Applicant before: ANHUI HUISHI JINTONG TECHNOLOGY Co.,Ltd. |
|
GR01 | Patent grant | ||
GR01 | Patent grant |