[go: up one dir, main page]

CN104035552A - Three-dimensional operation method and three-dimensional operation device - Google Patents

Three-dimensional operation method and three-dimensional operation device Download PDF

Info

Publication number
CN104035552A
CN104035552A CN201310075672.4A CN201310075672A CN104035552A CN 104035552 A CN104035552 A CN 104035552A CN 201310075672 A CN201310075672 A CN 201310075672A CN 104035552 A CN104035552 A CN 104035552A
Authority
CN
China
Prior art keywords
user
dimensional
operation behavior
coordinate
behavior
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201310075672.4A
Other languages
Chinese (zh)
Other versions
CN104035552B (en
Inventor
李勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Guangzhou Mobile R&D Center
Samsung Electronics Co Ltd
Original Assignee
Samsung Guangzhou Mobile R&D Center
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Guangzhou Mobile R&D Center, Samsung Electronics Co Ltd filed Critical Samsung Guangzhou Mobile R&D Center
Priority to CN201310075672.4A priority Critical patent/CN104035552B/en
Publication of CN104035552A publication Critical patent/CN104035552A/en
Application granted granted Critical
Publication of CN104035552B publication Critical patent/CN104035552B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Position Input By Displaying (AREA)

Abstract

本发明提供一种三维操作方法和三维操作装置,本发明的三维操作方法包含步骤:基于分别设置在屏幕周边的距离传感器和角度传感器测量的信息分析和判断用户的操作行为;根据用户的操作行为的判断结果而判定操作类型,量化操作参数;根据操作类型和操作参数,在用户界面执行三维操作。

The present invention provides a three-dimensional operation method and a three-dimensional operation device. The three-dimensional operation method of the present invention includes the steps of: analyzing and judging the user's operation behavior based on the information measured by the distance sensor and the angle sensor respectively arranged around the screen; Determine the type of operation based on the result of the judgment and quantify the operating parameters; perform three-dimensional operations on the user interface according to the type of operation and the operating parameters.

Description

Three-dimensional manipulating method and three-dimensional manipulating device
Technical field
The present invention relates to have the field of mobile terminals of three dimensional display, relate in particular to method and the three-dimensional manipulating device that can on three dimensional display, realize three-dimensional manipulating.
Background technology
Along with development and the application of three-dimensional (3D) technology, on mobile phone, also there is game and the application of a lot of 3D effects thereupon, screen can be shown 3D effect figure, makes user conscientiously experience three-dimensional display effect.
But existing touch-screen is all to need object directly to touch on screen, could realize some operation.Such as, need finger touches to screen, on screen, complete one section of sliding trace, judge user's behavior according to sliding trace, and then realize operation and control to mobile phone.
Due to the limitation of touch-screen, prior art is only suitable for operation and the control under 2D pattern, cannot solve displaying and control in 3d space.In other words, the plane touch screen of current this simple touch manner cannot complete the control function that stereoeffect is shown.
Summary of the invention
The present invention proposes in view of completing three-dimensional manipulating on touch-screen, does not touch screen even if its object is to provide a kind of, also can complete three-dimensional manipulating method and the three-dimensional manipulating device of the three-dimensional manipulating to three dimensional display.
Three-dimensional manipulating method of the present invention, comprises step: based on the measured range information of detection part and angle information analysis with judge user's operation behavior; According to the judged result of user's operation behavior and decision type, quantization operation parameter; According to action type and operating parameter, carry out three-dimensional manipulating in user interface.
And described detection part comprises three range sensors and angular transducer.
And described detection part further comprises that a range sensor and angular transducer are as check sensor.
And described detection part comprises radar.
And, described operating parameter comprise movement velocity, direction of motion and with screen angulation.
And described user's operation behavior is divided into one-way movement, Multidirectional motion and rotation.
And the user's of described one-way movement operation behavior judges as follows: whether the coordinate that judges the upper each point of finger variation tendency is at a time identical; Judge that it is linear in three coordinate planes, whether having a coordinate projection in plane at least; User's operation behavior is packaged into class.
And the user's of described Multidirectional motion operation behavior judges as follows: judgement finger number; Simulate the trail change of finger in space; User's operation behavior is packaged into class.
And the user's of described rotation operation behavior judges as follows: whether the gesture that tentatively judges user rotates; Judgement finger number; The rotation trend of analysis user gesture; Analysis user is pointed the movement locus in space; User's operation behavior is packaged into class.
Three-dimensional manipulating device of the present invention, comprises three dimensional display, controller, storer and detection part.
And described multiple distances and angular transducer are arranged on the periphery of described three dimensional display.
And described controller is according to the information analysis of described multiple distances and angular transducer and judge user's operation behavior, and judged user's operation behavior is referred to a kind of action type.
And described three dimensional display shows three-dimensional picture according to described action type.
By three-dimensional manipulating method provided by the present invention and three-dimensional manipulating device, can utilize the gesture in space to remove to operate three-dimensional picture, provide perfect support to 3D effect, allow UI become aestheticism more, more perceptual.
Brief description of the drawings
Fig. 1 is the process flow diagram of three-dimensional manipulating method of the present invention.
Fig. 2 is the schematic diagram of mounting distance and angular transducer on three dimensional display.
Fig. 3 is the three-dimensional coordinate system for certain any coordinate of obtaining on finger is described.
Fig. 4 is the process flow diagram of the deterministic process of the operation behavior of the user for one-way movement is described.
Fig. 5 is the process flow diagram of the deterministic process of the operation behavior of the user for Multidirectional motion is described.
Fig. 6 is the process flow diagram of the deterministic process of the user's formation for rotation behavior is described.
Embodiment
Below, the embodiment that present invention will be described in detail with reference to the accompanying.For convenience of explanation, mobile terminal of the present invention describes as an example of mobile phone example.Mobile phone of the present invention at least comprises three dimensional display, controller, storer and display.
First, with reference to figure 1, three-dimensional manipulating method of the present invention is described.
As shown in Figure 1, three-dimensional manipulating method of the present invention comprises:
Step 101: based on being separately positioned on the range sensor of screen periphery and angular transducer measured range information and angle information analysis and judging user's operation behavior.
Step 102: according to the judged result of user's operation behavior, decision operation type.
At this, user's operation behavior is summarized as different action types by software interface, then by relevant software interface offer user interface for.Here, action type mainly contain move, amplify, dwindle, rotation etc.
Step 103: according to action type, realize 3-D display in user interface.
Illustrate the method for the operation behavior that judges user below.
User's three-dimensional manipulating behavior has a variety of, if carry out quantitative analysis from the angle of changes in coordinates merely, roughly can user's operation behavior be divided into this three class of one-way movement, Multidirectional motion and rotation according to changes in coordinates.
In order to judge and the operation behavior of analysis user, the present invention has installed distance and angular transducer at three dimensional display four jiaos.Fig. 1 is illustrated in the schematic diagram of mounting distance and angular transducer on three dimensional display.As shown in Figure 1, at four jiaos of mounting distances and the angular transducer of three dimensional display, by sensor can measure on finger certain a bit and line and the three dimensional display angulation of the air line distance of sensor and this point and sensor.Also, can measure the information such as distance and angle of finger and sensor by range sensor, then calculate the multiple spot relative position of selling with three dimensional display by solid geometry algorithm, according to these information, user's operation behavior be judged.
Taking single-point as example, concrete computing method are as follows.
First suppose that mobile phone screen is in xy coordinate, the length of screen is the wide W of being of L.Set up as shown in Figure 2 coordinate system, certain on the finger some relative coordinate in coordinate system is d(x, y, z).According to the measurement data of sensor, we can measure the distance l of d and screen upper sensor 1, l 2, l 3, and they and mobile phone screen angulation a, b, c.
Can list following equation according to above-mentioned measurement data:
z=l 1sina………………1)
x 2+y 2=(l 2cosb) 2………………2)
(W-x) 2+y 2=(l 2cosa) 2………………3)
(L-y) 2+x 2=(l 2cosc) 2………………4)
Just can calculate x according to aforesaid equation, y, the value of z is:
x = W 2 + ( l 2 cos b ) 2 - ( l 1 cos a ) 2 2 W - - - 5 )
y = L 2 + ( l 2 cos b ) 2 - ( l 3 cos c ) 2 2 L - - - 6 )
z=l 1sina………………7)
Point the range information of each point and angle information only with three range sensors and angular transducer although measure user, but because may being subject to noise, environment etc. on measuring, sensor disturbs the certain error of generation, therefore the present invention is also provided with the 4th sensor as check sensor, according to the return data of the 4th sensor, utilize above-mentioned formula to recalculate, choose and measure coordinate points accurately, reach the effect of denoising.
Just can calculate the relative coordinate value D of the upper each point of finger and mobile phone screen according to method above 1(x 1, y 1, z 1), D 2(x 2, y 2, z 2) ... D n(x n, y n, z n), more just can depict the gesture of user in this moment according to these coordinates.User's finger is constantly moving, and the data in each moment are all different, use time 1to time mrepresent not in the same time, and use D 1to D nrepresent all reference point that are selected, so just can obtain one group of coordinate data below:
? time 1 time 2 ... time M
D 1 x 11,y 11,z 11 x 12,y 12,z 12 ... x 1M,y 1M,z 1M
D 2 x 21,y 21,z 21 x 22,y 22,z 22 ... x 2M,y 2M,z 2M
D 3 x 31,y 31,z 31 x 31,y 31,z 31 ... x 3M,y 3M,z 3M
... ... ... ... ...
D N x N1,y N1,z N1 x N2,y N2,z N2 ... x NM,y NM,z NM
According to data above, just can carry out modeling to the coordinate figure in every bit or each moment, depict the variation track in space a little, then go to judge user's operation behavior according to the track of changes in coordinates.
Due to D 1to D ncoordinate be all three-dimensional spatial coordinate, in order to judge user's operation behavior according to changes in coordinates, coordinate need to be carried out to projection operation on three coordinate surfaces, the projection according to track on coordinate surface judges movement locus.Carry out quantitative analysis according to the different movement locus characteristics of motion separately again, finally determine user's behavior.
At this, D(x, y are found out in so-called projection exactly, z) to the vertical irradiation point of certain plane, for example, by D(x, y, z) three coordinate planes are carried out to projection operation, the subpoint coordinate obtaining is respectively d (x, y, 0), d (0, y, z), d (x, 0, z).
The method of the operation behavior that judges user is described with embodiment below.
Embodiment 1: one-way movement
One-way movement is the most a kind of movement tendency of application in existing touch screen technology, can be pointed the track sliding and completed by record the judgement of the operation behavior to user in existing touch screen technology on screen.But in the present invention, finger or palm are to slide in space, directly do not contact touch-screen, and existing judgment technology cannot realize.
In the present invention, the determination methods of the user's of one-way movement operation behavior as shown in Figure 4.
The feature of one-way movement is one or many situations of doing to be parallel to a certain coordinate surface or oblique movement operation with finger.In this case, all reference point are at time 1to time mmovement locus in time is unanimous on the whole.If to whole movement locus differential, choose the variation tendency at a time of coordinate a little carry out sample calculation, its rule meets following formula substantially:
x M + 1 - x M y M + 1 - y M ≈ x M - x M - 1 y M - y M - 1
x M + 1 - x M z M + 1 - z M ≈ x M - x M - 1 z M - z M - 1
y M + 1 - y M z M + 1 - z M ≈ y M - y M - 1 z M - z M - 1
Due to the speed difference that finger streaks, the variation tendency that can not ensure whole track is definite value, but differential is consistent to the variation tendency of single-point.
Therefore the coordinate that, first the determination methods of one-way movement of the present invention judges the upper each point of finger variation tendency whether identical (S401) at a time.
In addition, by D 1to D ncoordinate projection to three coordinate plane on time, on a coordinate plane, be at least linear.
Therefore, the determination methods of one-way movement of the present invention judges that in three coordinate planes, whether having a coordinate projection in plane at least is linear (S402) again.
Meet in the situation of above-mentioned 2 simultaneously, can judge that finger or palm are doing unidirectional linear movement, are packaged into class (S403) by user's operation behavior.
The present invention can also be according to the track of motion, the variation of coordinate, calculate the speed, direction of motion and with the operating parameters such as coordinate plane angulation, and the action type of the operation behavior corresponding to user and various operating parameter are offered to UI layer, show corresponding three-dimensional picture for UI layer.
Embodiment 2: Multidirectional motion
Multidirectional motion is also more common user's operation behavior, in existing touch screen technology, can judge by the track of the contact number in initial moment and touch the movement locus of hand, then according to track, user's behavior be judged.
But in three dimensions, the same with one-way movement, user's finger motion locus forms in three dimensions, does not touch touch-screen.
Fig. 5 is used for the process flow diagram of the deterministic process of the user's that Multidirectional motion is described operation behavior.
If judge user's operation behavior, first to judge and in the measurement range of screen, have several fingers (S501).
Judgement finger number can realize by the following method, that is, and and to time 1moment coordinate a little carry out modeling, by time 1the D in moment 1to D nall coordinates all project on three coordinate planes, the track now obtaining on each coordinate plane should be the linear track of many line-like, the number that the track number in the maximum plane of three coordinate plane neutral line tracks should be pointed exactly.Restore in space the position of finger in this moment by the track of subpoint again.
Then, simulate user and point the trail change (S502) in space.Can pass through time 1to time mthe D in moment 1to D ncoordinate carry out projection operation, the variation of coordinate points that just can be by coordinate plane, simulate the trail change in space.
Then, judge user's behavior according to trail change, as mode of motion such as finger polymerization, dispersions, user's operation behavior is packaged into class.
Embodiment 3: rotate
Rotation is the operation behavior of unexistent user in two-dimensional space, is the peculiar a kind of movement tendency of 3d space.Rotation is one changes in coordinates trend relatively freely.User's hand can rotate before screen with any posture taking wrist as axle.The movement locus of each reference point is similar one section of irregular arc all, cannot simply get on to judge that from single changes in coordinates trend hand is rotating.
Fig. 6 is the process flow diagram of the process for the rotation behavior that judges user's gesture is described.
First, the present invention tentatively judges user's gesture whether rotate (S601).Its determination methods is, to coordinate a little carry out Projection Analysis.Specifically, by time 1to time mthe D in moment 1to D ncoordinate (coordinate data a little) project to respectively on three coordinate surfaces, if now have at least a projected footprint in plane rounded or oval, just can tentatively judge user's the similar sphere of gesture, rotating.
Secondly, judgement finger number (S602).Utilize time 1moment coordinate a little, determine finger number and shape within the scope of screen.
Then, the rotation trend (S603) of analysis user gesture.In the time that user's hand is rotated, time m-1to time min this process, on every finger, the variation of the vector of each point should be consistent.Method of vector analysis is as follows:
To time 1to time mthe D in moment 1to D ncoordinate analyze, calculate time 1to time 2, time 2to time 3..., time m-1to time mthe vector value of each point.
Postulated point D 1and D 2on same finger, 2 at time 1to time 2the variation of moment vector is identical, both meets can obtain by that analogy formula:
D 1 M D 1 M - 1 → × D 2 M D 2 M - 1 → = 0
In order to reduce measuring error, can carry out vector cross product to the multiple points on same finger and obtain the result after matching.In section, the sampling point vector variation tendency on each finger is identical at one time, can determine that user's gesture is rotating.
Then, analyze the movement locus of every finger in space, calculate rotation parameter (S604).
According to data above, can regard approx the variation tendency of user's gesture as a kind of similar sphere rotatablely move.Can calculate respectively like this time 1to time mthe information such as the centre of sphere, radius and the turning axle of moment spherical rotary, and calculate the vectorial variation tendency of each moment centre of sphere and turning axle.
Taking single point in time as example, introduce concrete computing method below:
The computing method > of the < centre of sphere and radius
Known theoretically, as long as there are 4 points on same circle just can obtain coordinate and the radius of the centre of sphere.Suppose that the radius of a ball is R, the approximate value of sphere centre coordinate is O (x 0, y 0, z 0), the coordinate that the i on sphere is ordered is D i(x i, y i, z i), there is following relational expression in the coordinate of every bit and the radius of a ball:
(x 0-x i) 2+(y 0-y i) 2+(z 0-z i) 2=R 2………………8)
The coordinate of choosing like this four points both can calculate sphere centre coordinate value and radius.But this is theoretical value, because user's rotation action is similar spheric motion, the measurement of each point coordinate is chosen also certain error, so will carry out matching to the result of calculation of multiple points.
Here have a variety of methods available, for example, the most frequently used method is exactly exhaustive, all bring the coordinate of all observation point into equation and calculate multiple sphere centre coordinates, then carry out denoising matching, although this method is easily understood, but calculated amount is very large, affects handset capability.
Mainly introduce a kind of method of compensating computation below, in the hope of the reliable value of sphere centre coordinate and the radius of a ball.
To asking partial differential for sphere centre coordinate and the radius of a ball in formula above, and with its modified value δ x, δ y, δ zand δ rreplace differential, obtain:
2(x 0-x ix+2(y 0-y iy+2(z 0-z iz=2R 0δ R………………9)
At this, R 0for radius match value.
x 0 - x i R 0 &delta; x + y 0 - y i R 0 &delta; y + z 0 - z i R 0 &delta; z = &delta; R - - - 10 )
The observed reading R that the radius of a ball of supposing the approximate coordinate computation of the coordinate figure of ordering by i and the centre of sphere is the radius of a ball t, its corrected value is V rt,
R t = ( x 0 - x i ) 2 + ( y 0 - y i ) 2 + ( z 0 - z i ) 2 - - - 11 )
R=R t+V Rt=R 0R………………12)
V Rt=R 0-R tR………………13)
In above formula, δ rtry to achieve, can use formula substitution above, obtain the radius of a ball observed reading equation that i is ordered:
V Rt = R 0 - R t + x 0 - x i R 0 &delta; x + y 0 - y i R 0 &delta; y + z 0 - z i R 0 &delta; z - - - 14 )
Equational general type is:
V Rt=a tδ x+b tδ y+c tδ z+l t………………15)
By equation and 3 unknown number δ of m observed reading x, δ y, δ zthe equation of composition is
[aa]δ x+[ab]δ y+[ac]δ z+[al]=0………………16)
[ab]δ x+[bb]δ y+[bc]δ z+[bl]=0………………17)
[ac]δ x+[bc]δ y+[cc]δ z+[cl]=0………………18)
Managing equational association factor matrices is
Q = Q 11 Q 12 Q 13 Q 12 Q 22 Q 23 Q 13 Q 23 Q 33
The solution of unknown number is
δ x=-[al]Q 11-[bl]Q 12-[cl]Q 13………………19)
δ y=-[al]Q 12-[bl]Q 22-[cl]Q 23………………20)
δ z=-[al]Q 13-[bl]Q 23-[cl]Q 33………………21)
Coordinate after centre of sphere adjustment is:
x=x 0x………………22)
y=y 0y………………23)
z=z 0z………………24)
The radius of a ball after adjustment is:
R = R 0 - [ l ] m ………………25)
In formula, molecule is observed reading equation constant term sum, the number that denominator is observed reading.
According to said method, just can comparatively accurately obtain the centre of sphere and the radius of the spherical rotary of single point in time, certain computing method have a variety of, and the present invention is only as example.
< turning axle computing method >
The rotating shaft of sphere is the straight line through the centre of sphere, and this axis is parallel with the normal vector of Plane of rotation.In other words, only requiring to obtain the normal vector of Plane of rotation, then substitution sphere centre coordinate, can determine unique turning axle.Now, how the normal vector of Plane of rotation calculates the key that has just become problem.
Here, so-called Plane of rotation refers to, at time 1all point coordinate in moment revert to the shape that forms a similar sphere in coordinate system, choose the coordinate points D that approaches each finger fingertip 1to D tfor sampled point, what these sampled points can be similar to fits within same plane, and this plane is exactly the Plane of rotation in this moment.
Can determine straight line according to 2 in space, the straight line of two intersections just can be determined the principle of a plane, as long as there are not four points on same straight line just can determine a plane, and calculates his normal vector.
Supposing has D 1to D tindividual, be roughly dispersed in four to five regions, as long as choose respectively in zones of different, just can determine multiple Plane of rotations and calculate multiple normal vectors, these normal vectors are done to vectorial multiplication cross computing, if two planes are parallel, the result of vectorial multiplication cross is 0 again.In the present invention, can, to use the method for multidirectional amount matching, determine the normal vector of Plane of rotation.
Calculate after the information such as the centre of sphere, radius, turning axle of the simulation spheric motion of single point in time, just can use the same method to time 2to time mthe coordinate in moment carries out modeling, calculates the spherical rotary information in each moment, as shown in the table:
According to these information, just can further calculate the information such as centre of sphere change vector, the vector variation of turning axle, can do to the variation track of whole gesture the modeling analysis of a quantification.
Finally, user's operation behavior Information encapsulation is become to class, return to application programming interfaces.
Above-mentioned several users' operation behavior is just done simple a division from the angle of changes in coordinates, because the action of staff can be ever-changing, can also, according to the information such as speed and time of changes in coordinates, carry out finer analysis to movement locus.
In addition, although the present invention is illustrated as example to be arranged on distance on four jiaos of three dimensional displays and angular transducer, but the present invention is not limited thereto, the present invention can by be arranged on a radargrammetry user finger of mobile phone screen periphery certain a bit and the air line distance of radar and line and the three dimensional display angulation of this point and radar., the present invention can replace range sensor and angular transducer measuring distance information and angle information by radar.At this, the present invention does not limit the magnitude setting of radar, for example can two radars be set at mobile phone screen periphery, a radar is for measuring distance information and angle information, and another radar is for measured range information and the angle information accuracy of verification.
The present invention realizes the 3D of three dimensional display to control, and applies very extensive.Be used in the middle of the demonstration of 3D game and 3D effect figure, the operation behavior of the gesture analysis user by user, is packaged into unified UI interface, and application can utilize the UI style of these Interface realization different-effects, experiences to the diverse user of user.When the present invention is really applied on mobile phone, its application part is innumerable.For example:
1.3D cuts fruit: in space, do the action cutting fruit really by hand, complete game.
2. virtual magic square: user can directly do by gesture various rotation operations before mobile phone screen, and the magic square of directly controlling in screen forwards to, realizes 3D bandwagon effect.
3.TOM cat: user can do an action of drawing before mobile phone screen, removes to draw hair ear or the hair of TOM, to more expressing one's feelings and move TOM collocation in.
4.3D desktop: break original 2D desktop effect, the icon on desktop can be to make stereoeffect, and desk-top space also can become 3d space, by gesture, icon is drawn in, puts far, amplifies, and the operation such as dwindles.
The example of 3D operation is too numerous to enumerate, the invention has the advantages that and utilizes the gesture in space to go to judge user's behavior, is different from existing touch screen technology completely, and user just can complete operation without touching screen.And liberate finger, touching operation has been become to spatial movement behavioural analysis.User can have more selection, and more gesture removes to control mobile phone.Also for UI layer provides abundant class of operation, to the perfection support of 3D effect, allow UI become and obtain aestheticism more simultaneously, more perceptual.

Claims (13)

1.一种三维操作方法,其特征在于,包含步骤:1. A three-dimensional operation method, characterized in that, comprising steps: 基于检测部件所测量的距离信息和角度信息分析和判断用户的操作行为;Analyze and judge the user's operation behavior based on the distance information and angle information measured by the detection component; 根据用户的操作行为的判断结果而判定操作类型,量化操作参数;Determine the operation type and quantify the operation parameters according to the judgment result of the user's operation behavior; 根据操作类型和操作参数,在用户界面执行三维操作。Perform three-dimensional operations on the user interface according to the operation type and operation parameters. 2.根据权利要求1所述的三维操作方法,其特征在于,所述检测部件包括三个距离传感器和角度传感器。2. The three-dimensional operation method according to claim 1, characterized in that, the detection component comprises three distance sensors and an angle sensor. 3.根据权利要求2所述的三维操作方法,其特征在于,所述检测部件进一步包括一个距离传感器和角度传感器作为校验传感器。3. The three-dimensional operation method according to claim 2, characterized in that the detection component further comprises a distance sensor and an angle sensor as calibration sensors. 4.根据权利要求1所述的三维操作方法,其特征在于,所述检测部件包括雷达。4. The three-dimensional operation method according to claim 1, wherein the detection component comprises a radar. 5.根据权利要求1所述的三维操作方法,其特征在于,所述操作参数包含运动速度、运动方向以及与屏幕所成的角度。5 . The three-dimensional operation method according to claim 1 , wherein the operation parameters include movement speed, movement direction and angle with the screen. 6.根据权利要求1所述的三维操作方法,其特征在于,所述用户的操作行为分为单向运动、多向运动和转动。6 . The three-dimensional operation method according to claim 1 , wherein the user's operation behavior is divided into one-way movement, multi-way movement and rotation. 7.根据权利要求6所述的三维操作方法,其特征在于,所述单向运动的用户的操作行为通过如下步骤进行判断:7. The three-dimensional operation method according to claim 6, wherein the user's operation behavior of the one-way movement is judged by the following steps: 判断手指上各点的坐标在某一时刻的变化趋势是否相同;Judging whether the coordinates of each point on the finger have the same change trend at a certain moment; 判断三个坐标平面中是否至少有一个平面上的坐标投影为线形;Determine whether the coordinate projection on at least one of the three coordinate planes is linear; 将用户的操作行为封装成类。Encapsulate the user's operation behavior into a class. 8.根据权利要求6所述的三维操作方法,其特征在于,所述多向运动的用户的操作行为通过如下步骤进行判断:8. The three-dimensional operation method according to claim 6, characterized in that the user's operation behavior of the multi-directional movement is judged by the following steps: 判断手指个数;judge the number of fingers; 模拟出手指在空间的轨迹变化;Simulate the trajectory changes of fingers in space; 将用户的操作行为封装成类。Encapsulate the user's operation behavior into a class. 9.根据权利要求6所述的三维操作方法,其特征在于,所述转动的用户的操作行为通过如下步骤进行判断:9. The three-dimensional operation method according to claim 6, characterized in that the operation behavior of the rotating user is judged by the following steps: 初步判断用户的手势是否做旋转运动;Preliminarily judge whether the user's gesture is rotating; 判断手指个数;judge the number of fingers; 分析用户手势的旋转趋势;Analyze the rotation trend of user gestures; 分析用户手指在空间中的运动轨迹;Analyze the trajectory of the user's finger in space; 将用户的操作行为封装成类。Encapsulate the user's operation behavior into a class. 10.一种三维操作装置,其特征在于,包括三维显示器、控制器、存储器以及检测部件。10. A three-dimensional operation device, characterized by comprising a three-dimensional display, a controller, a memory, and a detection component. 11.根据权利要求10所述的三维操作装置,其特征在于,所述多个距离和角度传感器设置在所述三维显示器的周边。11. The three-dimensional operation device according to claim 10, wherein the plurality of distance and angle sensors are arranged around the three-dimensional display. 12.根据权利要求10所述的三维操作装置,其特征在于,所述控制器根据所述多个距离和角度传感器的信息分析和判断用户的操作行为,并将所判断的用户的操作行为归类到一种操作类型。12. The three-dimensional operation device according to claim 10, wherein the controller analyzes and judges the user's operation behavior according to the information of the plurality of distance and angle sensors, and returns the judged user's operation behavior to class to an operation type. 13.根据权利要求12所述的三维操作装置,其特征在于,所述三维显示器根据所述操作类型显示三维画面。13. The three-dimensional operation device according to claim 12, wherein the three-dimensional display displays a three-dimensional picture according to the operation type.
CN201310075672.4A 2013-03-08 2013-03-08 Three-dimensional manipulating method and three-dimensional manipulating device Expired - Fee Related CN104035552B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310075672.4A CN104035552B (en) 2013-03-08 2013-03-08 Three-dimensional manipulating method and three-dimensional manipulating device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310075672.4A CN104035552B (en) 2013-03-08 2013-03-08 Three-dimensional manipulating method and three-dimensional manipulating device

Publications (2)

Publication Number Publication Date
CN104035552A true CN104035552A (en) 2014-09-10
CN104035552B CN104035552B (en) 2017-08-25

Family

ID=51466354

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310075672.4A Expired - Fee Related CN104035552B (en) 2013-03-08 2013-03-08 Three-dimensional manipulating method and three-dimensional manipulating device

Country Status (1)

Country Link
CN (1) CN104035552B (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105487692A (en) * 2014-12-22 2016-04-13 哈尔滨安天科技股份有限公司 Controller switching method and system based on three-dimensional display space
CN106662639A (en) * 2014-10-02 2017-05-10 谷歌公司 Non-line-of-sight radar-based gesture recognition
US10310621B1 (en) 2015-10-06 2019-06-04 Google Llc Radar gesture sensing using existing data protocols
US10409385B2 (en) 2014-08-22 2019-09-10 Google Llc Occluded gesture recognition
US10496182B2 (en) 2015-04-30 2019-12-03 Google Llc Type-agnostic RF signal representations
US10642367B2 (en) 2014-08-07 2020-05-05 Google Llc Radar-based gesture sensing and data transmission
CN111198644A (en) * 2018-11-16 2020-05-26 西安易朴通讯技术有限公司 Method and system for identifying screen operation of intelligent terminal
US10817070B2 (en) 2015-04-30 2020-10-27 Google Llc RF-based micro-motion tracking for gesture tracking and recognition
US10936085B2 (en) 2015-05-27 2021-03-02 Google Llc Gesture detection and interactions
US10948996B2 (en) 2014-06-03 2021-03-16 Google Llc Radar-based gesture-recognition at a surface of an object
US11140787B2 (en) 2016-05-03 2021-10-05 Google Llc Connecting an electronic component to an interactive textile
US11169988B2 (en) 2014-08-22 2021-11-09 Google Llc Radar recognition-aided search

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11134089A (en) * 1997-10-29 1999-05-21 Takenaka Komuten Co Ltd Hand pointing device
CN102236414A (en) * 2011-05-24 2011-11-09 北京新岸线网络技术有限公司 Picture operation method and system in three-dimensional display space
CN102804786A (en) * 2009-06-16 2012-11-28 Lg电子株式会社 Viewing range notification method and TV receiver for implementing the same
CN102810015A (en) * 2011-05-31 2012-12-05 中兴通讯股份有限公司 Input method and terminal based on space motion

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11134089A (en) * 1997-10-29 1999-05-21 Takenaka Komuten Co Ltd Hand pointing device
CN102804786A (en) * 2009-06-16 2012-11-28 Lg电子株式会社 Viewing range notification method and TV receiver for implementing the same
CN102236414A (en) * 2011-05-24 2011-11-09 北京新岸线网络技术有限公司 Picture operation method and system in three-dimensional display space
CN102810015A (en) * 2011-05-31 2012-12-05 中兴通讯股份有限公司 Input method and terminal based on space motion

Cited By (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10948996B2 (en) 2014-06-03 2021-03-16 Google Llc Radar-based gesture-recognition at a surface of an object
US10642367B2 (en) 2014-08-07 2020-05-05 Google Llc Radar-based gesture sensing and data transmission
US11221682B2 (en) 2014-08-22 2022-01-11 Google Llc Occluded gesture recognition
US10936081B2 (en) 2014-08-22 2021-03-02 Google Llc Occluded gesture recognition
US11169988B2 (en) 2014-08-22 2021-11-09 Google Llc Radar recognition-aided search
US12153571B2 (en) 2014-08-22 2024-11-26 Google Llc Radar recognition-aided search
US11816101B2 (en) 2014-08-22 2023-11-14 Google Llc Radar recognition-aided search
US10409385B2 (en) 2014-08-22 2019-09-10 Google Llc Occluded gesture recognition
US10664059B2 (en) 2014-10-02 2020-05-26 Google Llc Non-line-of-sight radar-based gesture recognition
CN106662639B (en) * 2014-10-02 2019-07-26 谷歌有限责任公司 Non-Line-of-Sight Radar-Based Gesture Recognition
US11163371B2 (en) 2014-10-02 2021-11-02 Google Llc Non-line-of-sight radar-based gesture recognition
CN106662639A (en) * 2014-10-02 2017-05-10 谷歌公司 Non-line-of-sight radar-based gesture recognition
CN105487692A (en) * 2014-12-22 2016-04-13 哈尔滨安天科技股份有限公司 Controller switching method and system based on three-dimensional display space
US10496182B2 (en) 2015-04-30 2019-12-03 Google Llc Type-agnostic RF signal representations
US11709552B2 (en) 2015-04-30 2023-07-25 Google Llc RF-based micro-motion tracking for gesture tracking and recognition
US12340028B2 (en) 2015-04-30 2025-06-24 Google Llc RF-based micro-motion tracking for gesture tracking and recognition
US10817070B2 (en) 2015-04-30 2020-10-27 Google Llc RF-based micro-motion tracking for gesture tracking and recognition
US10936085B2 (en) 2015-05-27 2021-03-02 Google Llc Gesture detection and interactions
US10540001B1 (en) 2015-10-06 2020-01-21 Google Llc Fine-motion virtual-reality or augmented-reality control using radar
US11385721B2 (en) 2015-10-06 2022-07-12 Google Llc Application-based signal processing parameters in radar-based detection
US10823841B1 (en) 2015-10-06 2020-11-03 Google Llc Radar imaging on a mobile computing device
US10768712B2 (en) 2015-10-06 2020-09-08 Google Llc Gesture component with gesture library
US11132065B2 (en) 2015-10-06 2021-09-28 Google Llc Radar-enabled sensor fusion
US10310621B1 (en) 2015-10-06 2019-06-04 Google Llc Radar gesture sensing using existing data protocols
US10705185B1 (en) 2015-10-06 2020-07-07 Google Llc Application-based signal processing parameters in radar-based detection
US10379621B2 (en) 2015-10-06 2019-08-13 Google Llc Gesture component with gesture library
US11175743B2 (en) 2015-10-06 2021-11-16 Google Llc Gesture recognition using multiple antenna
US10503883B1 (en) 2015-10-06 2019-12-10 Google Llc Radar-based authentication
US11256335B2 (en) 2015-10-06 2022-02-22 Google Llc Fine-motion virtual-reality or augmented-reality control using radar
US10908696B2 (en) 2015-10-06 2021-02-02 Google Llc Advanced gaming and virtual reality control using radar
US11481040B2 (en) 2015-10-06 2022-10-25 Google Llc User-customizable machine-learning in radar-based gesture detection
US11656336B2 (en) 2015-10-06 2023-05-23 Google Llc Advanced gaming and virtual reality control using radar
US11693092B2 (en) 2015-10-06 2023-07-04 Google Llc Gesture recognition using multiple antenna
US11698439B2 (en) 2015-10-06 2023-07-11 Google Llc Gesture recognition using multiple antenna
US11698438B2 (en) 2015-10-06 2023-07-11 Google Llc Gesture recognition using multiple antenna
US10459080B1 (en) 2015-10-06 2019-10-29 Google Llc Radar-based object detection for vehicles
US12117560B2 (en) 2015-10-06 2024-10-15 Google Llc Radar-enabled sensor fusion
US10401490B2 (en) 2015-10-06 2019-09-03 Google Llc Radar-enabled sensor fusion
US12085670B2 (en) 2015-10-06 2024-09-10 Google Llc Advanced gaming and virtual reality control using radar
US11140787B2 (en) 2016-05-03 2021-10-05 Google Llc Connecting an electronic component to an interactive textile
CN111198644B (en) * 2018-11-16 2023-10-27 西安易朴通讯技术有限公司 Screen operation identification method and system of intelligent terminal
CN111198644A (en) * 2018-11-16 2020-05-26 西安易朴通讯技术有限公司 Method and system for identifying screen operation of intelligent terminal

Also Published As

Publication number Publication date
CN104035552B (en) 2017-08-25

Similar Documents

Publication Publication Date Title
CN104035552A (en) Three-dimensional operation method and three-dimensional operation device
Kwan et al. Mobi3dsketch: 3d sketching in mobile ar
CN102483674B (en) Methods and apparatus for pressure-based manipulation of content on a touch screen
CN105453016B (en) In response to the context-sensitive movement of touch input
KR101318244B1 (en) System and Method for Implemeting 3-Dimensional User Interface
CN102216879B (en) Method and device for inputting a user&#39;s instructions based on movement sensing
CN102479050B (en) Method and apparatus for moving object by drag operation on touch panel
TWI506243B (en) System and method for simulating a calibration path of a probe of a measuring machine
CN101632029A (en) Enhanced single-sensor position detection
CA2937956A1 (en) In-air ultrasound pen gestures
CN105302407A (en) Application icon display method and apparatus
CN104156144B (en) Electric document page-turning effect implementation method and device
CN103940442A (en) Location method and device adopting accelerating convergence algorithm
CN102622225A (en) Multipoint touch application program development method supporting user defined gestures
Kang et al. Editing 3D models on smart devices
CN103543885B (en) The multimedia spherical screen demonstration instrument of multi-point touch and multi-touch method thereof
CN107250956B (en) Method and system for binode sensing
CN103970328A (en) Touch or non-touch type multi-input-point control command detecting method and device
Pryss et al. Enabling tracks in location-based smart mobile augmented reality applications
JP2023511022A (en) Interactive 3D representation of objects
CN104567784A (en) Three-dimensional measurement simulation system and method
Baharav et al. Capacitive touch sensing: signal and image processing algorithms
WO2018098771A1 (en) Method for determining azimuthal angle or posture, touch input device, touch panel, and system
CN103513919B (en) A kind of touch control device and the display control method of 3D rotating disk, device
US20130249807A1 (en) Method and apparatus for three-dimensional image rotation on a touch screen

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20170825

Termination date: 20210308

CF01 Termination of patent right due to non-payment of annual fee