[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN104035552A - Three-dimensional operation method and three-dimensional operation device - Google Patents

Three-dimensional operation method and three-dimensional operation device Download PDF

Info

Publication number
CN104035552A
CN104035552A CN201310075672.4A CN201310075672A CN104035552A CN 104035552 A CN104035552 A CN 104035552A CN 201310075672 A CN201310075672 A CN 201310075672A CN 104035552 A CN104035552 A CN 104035552A
Authority
CN
China
Prior art keywords
user
dimensional
operation behavior
coordinate
dimensional manipulating
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201310075672.4A
Other languages
Chinese (zh)
Other versions
CN104035552B (en
Inventor
李勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Guangzhou Mobile R&D Center
Samsung Electronics Co Ltd
Original Assignee
Samsung Guangzhou Mobile R&D Center
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Guangzhou Mobile R&D Center, Samsung Electronics Co Ltd filed Critical Samsung Guangzhou Mobile R&D Center
Priority to CN201310075672.4A priority Critical patent/CN104035552B/en
Publication of CN104035552A publication Critical patent/CN104035552A/en
Application granted granted Critical
Publication of CN104035552B publication Critical patent/CN104035552B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Position Input By Displaying (AREA)

Abstract

The invention discloses a three-dimensional operation method and a three-dimensional operation device. The three-dimensional operation method includes that analyzing and judging the operation behavior of a user based on information measured by a distance sensor and an angle sensor arranged at the periphery of a screen; judging the operation type according to the judging result for the operation behavior of the user, and quantizing operation parameters; performing a three-dimensional operation on a user interface according to the operation type and operation parameters.

Description

Three-dimensional manipulating method and three-dimensional manipulating device
Technical field
The present invention relates to have the field of mobile terminals of three dimensional display, relate in particular to method and the three-dimensional manipulating device that can on three dimensional display, realize three-dimensional manipulating.
Background technology
Along with development and the application of three-dimensional (3D) technology, on mobile phone, also there is game and the application of a lot of 3D effects thereupon, screen can be shown 3D effect figure, makes user conscientiously experience three-dimensional display effect.
But existing touch-screen is all to need object directly to touch on screen, could realize some operation.Such as, need finger touches to screen, on screen, complete one section of sliding trace, judge user's behavior according to sliding trace, and then realize operation and control to mobile phone.
Due to the limitation of touch-screen, prior art is only suitable for operation and the control under 2D pattern, cannot solve displaying and control in 3d space.In other words, the plane touch screen of current this simple touch manner cannot complete the control function that stereoeffect is shown.
Summary of the invention
The present invention proposes in view of completing three-dimensional manipulating on touch-screen, does not touch screen even if its object is to provide a kind of, also can complete three-dimensional manipulating method and the three-dimensional manipulating device of the three-dimensional manipulating to three dimensional display.
Three-dimensional manipulating method of the present invention, comprises step: based on the measured range information of detection part and angle information analysis with judge user's operation behavior; According to the judged result of user's operation behavior and decision type, quantization operation parameter; According to action type and operating parameter, carry out three-dimensional manipulating in user interface.
And described detection part comprises three range sensors and angular transducer.
And described detection part further comprises that a range sensor and angular transducer are as check sensor.
And described detection part comprises radar.
And, described operating parameter comprise movement velocity, direction of motion and with screen angulation.
And described user's operation behavior is divided into one-way movement, Multidirectional motion and rotation.
And the user's of described one-way movement operation behavior judges as follows: whether the coordinate that judges the upper each point of finger variation tendency is at a time identical; Judge that it is linear in three coordinate planes, whether having a coordinate projection in plane at least; User's operation behavior is packaged into class.
And the user's of described Multidirectional motion operation behavior judges as follows: judgement finger number; Simulate the trail change of finger in space; User's operation behavior is packaged into class.
And the user's of described rotation operation behavior judges as follows: whether the gesture that tentatively judges user rotates; Judgement finger number; The rotation trend of analysis user gesture; Analysis user is pointed the movement locus in space; User's operation behavior is packaged into class.
Three-dimensional manipulating device of the present invention, comprises three dimensional display, controller, storer and detection part.
And described multiple distances and angular transducer are arranged on the periphery of described three dimensional display.
And described controller is according to the information analysis of described multiple distances and angular transducer and judge user's operation behavior, and judged user's operation behavior is referred to a kind of action type.
And described three dimensional display shows three-dimensional picture according to described action type.
By three-dimensional manipulating method provided by the present invention and three-dimensional manipulating device, can utilize the gesture in space to remove to operate three-dimensional picture, provide perfect support to 3D effect, allow UI become aestheticism more, more perceptual.
Brief description of the drawings
Fig. 1 is the process flow diagram of three-dimensional manipulating method of the present invention.
Fig. 2 is the schematic diagram of mounting distance and angular transducer on three dimensional display.
Fig. 3 is the three-dimensional coordinate system for certain any coordinate of obtaining on finger is described.
Fig. 4 is the process flow diagram of the deterministic process of the operation behavior of the user for one-way movement is described.
Fig. 5 is the process flow diagram of the deterministic process of the operation behavior of the user for Multidirectional motion is described.
Fig. 6 is the process flow diagram of the deterministic process of the user's formation for rotation behavior is described.
Embodiment
Below, the embodiment that present invention will be described in detail with reference to the accompanying.For convenience of explanation, mobile terminal of the present invention describes as an example of mobile phone example.Mobile phone of the present invention at least comprises three dimensional display, controller, storer and display.
First, with reference to figure 1, three-dimensional manipulating method of the present invention is described.
As shown in Figure 1, three-dimensional manipulating method of the present invention comprises:
Step 101: based on being separately positioned on the range sensor of screen periphery and angular transducer measured range information and angle information analysis and judging user's operation behavior.
Step 102: according to the judged result of user's operation behavior, decision operation type.
At this, user's operation behavior is summarized as different action types by software interface, then by relevant software interface offer user interface for.Here, action type mainly contain move, amplify, dwindle, rotation etc.
Step 103: according to action type, realize 3-D display in user interface.
Illustrate the method for the operation behavior that judges user below.
User's three-dimensional manipulating behavior has a variety of, if carry out quantitative analysis from the angle of changes in coordinates merely, roughly can user's operation behavior be divided into this three class of one-way movement, Multidirectional motion and rotation according to changes in coordinates.
In order to judge and the operation behavior of analysis user, the present invention has installed distance and angular transducer at three dimensional display four jiaos.Fig. 1 is illustrated in the schematic diagram of mounting distance and angular transducer on three dimensional display.As shown in Figure 1, at four jiaos of mounting distances and the angular transducer of three dimensional display, by sensor can measure on finger certain a bit and line and the three dimensional display angulation of the air line distance of sensor and this point and sensor.Also, can measure the information such as distance and angle of finger and sensor by range sensor, then calculate the multiple spot relative position of selling with three dimensional display by solid geometry algorithm, according to these information, user's operation behavior be judged.
Taking single-point as example, concrete computing method are as follows.
First suppose that mobile phone screen is in xy coordinate, the length of screen is the wide W of being of L.Set up as shown in Figure 2 coordinate system, certain on the finger some relative coordinate in coordinate system is d(x, y, z).According to the measurement data of sensor, we can measure the distance l of d and screen upper sensor 1, l 2, l 3, and they and mobile phone screen angulation a, b, c.
Can list following equation according to above-mentioned measurement data:
z=l 1sina………………1)
x 2+y 2=(l 2cosb) 2………………2)
(W-x) 2+y 2=(l 2cosa) 2………………3)
(L-y) 2+x 2=(l 2cosc) 2………………4)
Just can calculate x according to aforesaid equation, y, the value of z is:
x = W 2 + ( l 2 cos b ) 2 - ( l 1 cos a ) 2 2 W - - - 5 )
y = L 2 + ( l 2 cos b ) 2 - ( l 3 cos c ) 2 2 L - - - 6 )
z=l 1sina………………7)
Point the range information of each point and angle information only with three range sensors and angular transducer although measure user, but because may being subject to noise, environment etc. on measuring, sensor disturbs the certain error of generation, therefore the present invention is also provided with the 4th sensor as check sensor, according to the return data of the 4th sensor, utilize above-mentioned formula to recalculate, choose and measure coordinate points accurately, reach the effect of denoising.
Just can calculate the relative coordinate value D of the upper each point of finger and mobile phone screen according to method above 1(x 1, y 1, z 1), D 2(x 2, y 2, z 2) ... D n(x n, y n, z n), more just can depict the gesture of user in this moment according to these coordinates.User's finger is constantly moving, and the data in each moment are all different, use time 1to time mrepresent not in the same time, and use D 1to D nrepresent all reference point that are selected, so just can obtain one group of coordinate data below:
? time 1 time 2 ... time M
D 1 x 11,y 11,z 11 x 12,y 12,z 12 ... x 1M,y 1M,z 1M
D 2 x 21,y 21,z 21 x 22,y 22,z 22 ... x 2M,y 2M,z 2M
D 3 x 31,y 31,z 31 x 31,y 31,z 31 ... x 3M,y 3M,z 3M
... ... ... ... ...
D N x N1,y N1,z N1 x N2,y N2,z N2 ... x NM,y NM,z NM
According to data above, just can carry out modeling to the coordinate figure in every bit or each moment, depict the variation track in space a little, then go to judge user's operation behavior according to the track of changes in coordinates.
Due to D 1to D ncoordinate be all three-dimensional spatial coordinate, in order to judge user's operation behavior according to changes in coordinates, coordinate need to be carried out to projection operation on three coordinate surfaces, the projection according to track on coordinate surface judges movement locus.Carry out quantitative analysis according to the different movement locus characteristics of motion separately again, finally determine user's behavior.
At this, D(x, y are found out in so-called projection exactly, z) to the vertical irradiation point of certain plane, for example, by D(x, y, z) three coordinate planes are carried out to projection operation, the subpoint coordinate obtaining is respectively d (x, y, 0), d (0, y, z), d (x, 0, z).
The method of the operation behavior that judges user is described with embodiment below.
Embodiment 1: one-way movement
One-way movement is the most a kind of movement tendency of application in existing touch screen technology, can be pointed the track sliding and completed by record the judgement of the operation behavior to user in existing touch screen technology on screen.But in the present invention, finger or palm are to slide in space, directly do not contact touch-screen, and existing judgment technology cannot realize.
In the present invention, the determination methods of the user's of one-way movement operation behavior as shown in Figure 4.
The feature of one-way movement is one or many situations of doing to be parallel to a certain coordinate surface or oblique movement operation with finger.In this case, all reference point are at time 1to time mmovement locus in time is unanimous on the whole.If to whole movement locus differential, choose the variation tendency at a time of coordinate a little carry out sample calculation, its rule meets following formula substantially:
x M + 1 - x M y M + 1 - y M ≈ x M - x M - 1 y M - y M - 1
x M + 1 - x M z M + 1 - z M ≈ x M - x M - 1 z M - z M - 1
y M + 1 - y M z M + 1 - z M ≈ y M - y M - 1 z M - z M - 1
Due to the speed difference that finger streaks, the variation tendency that can not ensure whole track is definite value, but differential is consistent to the variation tendency of single-point.
Therefore the coordinate that, first the determination methods of one-way movement of the present invention judges the upper each point of finger variation tendency whether identical (S401) at a time.
In addition, by D 1to D ncoordinate projection to three coordinate plane on time, on a coordinate plane, be at least linear.
Therefore, the determination methods of one-way movement of the present invention judges that in three coordinate planes, whether having a coordinate projection in plane at least is linear (S402) again.
Meet in the situation of above-mentioned 2 simultaneously, can judge that finger or palm are doing unidirectional linear movement, are packaged into class (S403) by user's operation behavior.
The present invention can also be according to the track of motion, the variation of coordinate, calculate the speed, direction of motion and with the operating parameters such as coordinate plane angulation, and the action type of the operation behavior corresponding to user and various operating parameter are offered to UI layer, show corresponding three-dimensional picture for UI layer.
Embodiment 2: Multidirectional motion
Multidirectional motion is also more common user's operation behavior, in existing touch screen technology, can judge by the track of the contact number in initial moment and touch the movement locus of hand, then according to track, user's behavior be judged.
But in three dimensions, the same with one-way movement, user's finger motion locus forms in three dimensions, does not touch touch-screen.
Fig. 5 is used for the process flow diagram of the deterministic process of the user's that Multidirectional motion is described operation behavior.
If judge user's operation behavior, first to judge and in the measurement range of screen, have several fingers (S501).
Judgement finger number can realize by the following method, that is, and and to time 1moment coordinate a little carry out modeling, by time 1the D in moment 1to D nall coordinates all project on three coordinate planes, the track now obtaining on each coordinate plane should be the linear track of many line-like, the number that the track number in the maximum plane of three coordinate plane neutral line tracks should be pointed exactly.Restore in space the position of finger in this moment by the track of subpoint again.
Then, simulate user and point the trail change (S502) in space.Can pass through time 1to time mthe D in moment 1to D ncoordinate carry out projection operation, the variation of coordinate points that just can be by coordinate plane, simulate the trail change in space.
Then, judge user's behavior according to trail change, as mode of motion such as finger polymerization, dispersions, user's operation behavior is packaged into class.
Embodiment 3: rotate
Rotation is the operation behavior of unexistent user in two-dimensional space, is the peculiar a kind of movement tendency of 3d space.Rotation is one changes in coordinates trend relatively freely.User's hand can rotate before screen with any posture taking wrist as axle.The movement locus of each reference point is similar one section of irregular arc all, cannot simply get on to judge that from single changes in coordinates trend hand is rotating.
Fig. 6 is the process flow diagram of the process for the rotation behavior that judges user's gesture is described.
First, the present invention tentatively judges user's gesture whether rotate (S601).Its determination methods is, to coordinate a little carry out Projection Analysis.Specifically, by time 1to time mthe D in moment 1to D ncoordinate (coordinate data a little) project to respectively on three coordinate surfaces, if now have at least a projected footprint in plane rounded or oval, just can tentatively judge user's the similar sphere of gesture, rotating.
Secondly, judgement finger number (S602).Utilize time 1moment coordinate a little, determine finger number and shape within the scope of screen.
Then, the rotation trend (S603) of analysis user gesture.In the time that user's hand is rotated, time m-1to time min this process, on every finger, the variation of the vector of each point should be consistent.Method of vector analysis is as follows:
To time 1to time mthe D in moment 1to D ncoordinate analyze, calculate time 1to time 2, time 2to time 3..., time m-1to time mthe vector value of each point.
Postulated point D 1and D 2on same finger, 2 at time 1to time 2the variation of moment vector is identical, both meets can obtain by that analogy formula:
D 1 M D 1 M - 1 → × D 2 M D 2 M - 1 → = 0
In order to reduce measuring error, can carry out vector cross product to the multiple points on same finger and obtain the result after matching.In section, the sampling point vector variation tendency on each finger is identical at one time, can determine that user's gesture is rotating.
Then, analyze the movement locus of every finger in space, calculate rotation parameter (S604).
According to data above, can regard approx the variation tendency of user's gesture as a kind of similar sphere rotatablely move.Can calculate respectively like this time 1to time mthe information such as the centre of sphere, radius and the turning axle of moment spherical rotary, and calculate the vectorial variation tendency of each moment centre of sphere and turning axle.
Taking single point in time as example, introduce concrete computing method below:
The computing method > of the < centre of sphere and radius
Known theoretically, as long as there are 4 points on same circle just can obtain coordinate and the radius of the centre of sphere.Suppose that the radius of a ball is R, the approximate value of sphere centre coordinate is O (x 0, y 0, z 0), the coordinate that the i on sphere is ordered is D i(x i, y i, z i), there is following relational expression in the coordinate of every bit and the radius of a ball:
(x 0-x i) 2+(y 0-y i) 2+(z 0-z i) 2=R 2………………8)
The coordinate of choosing like this four points both can calculate sphere centre coordinate value and radius.But this is theoretical value, because user's rotation action is similar spheric motion, the measurement of each point coordinate is chosen also certain error, so will carry out matching to the result of calculation of multiple points.
Here have a variety of methods available, for example, the most frequently used method is exactly exhaustive, all bring the coordinate of all observation point into equation and calculate multiple sphere centre coordinates, then carry out denoising matching, although this method is easily understood, but calculated amount is very large, affects handset capability.
Mainly introduce a kind of method of compensating computation below, in the hope of the reliable value of sphere centre coordinate and the radius of a ball.
To asking partial differential for sphere centre coordinate and the radius of a ball in formula above, and with its modified value δ x, δ y, δ zand δ rreplace differential, obtain:
2(x 0-x ix+2(y 0-y iy+2(z 0-z iz=2R 0δ R………………9)
At this, R 0for radius match value.
x 0 - x i R 0 &delta; x + y 0 - y i R 0 &delta; y + z 0 - z i R 0 &delta; z = &delta; R - - - 10 )
The observed reading R that the radius of a ball of supposing the approximate coordinate computation of the coordinate figure of ordering by i and the centre of sphere is the radius of a ball t, its corrected value is V rt,
R t = ( x 0 - x i ) 2 + ( y 0 - y i ) 2 + ( z 0 - z i ) 2 - - - 11 )
R=R t+V Rt=R 0R………………12)
V Rt=R 0-R tR………………13)
In above formula, δ rtry to achieve, can use formula substitution above, obtain the radius of a ball observed reading equation that i is ordered:
V Rt = R 0 - R t + x 0 - x i R 0 &delta; x + y 0 - y i R 0 &delta; y + z 0 - z i R 0 &delta; z - - - 14 )
Equational general type is:
V Rt=a tδ x+b tδ y+c tδ z+l t………………15)
By equation and 3 unknown number δ of m observed reading x, δ y, δ zthe equation of composition is
[aa]δ x+[ab]δ y+[ac]δ z+[al]=0………………16)
[ab]δ x+[bb]δ y+[bc]δ z+[bl]=0………………17)
[ac]δ x+[bc]δ y+[cc]δ z+[cl]=0………………18)
Managing equational association factor matrices is
Q = Q 11 Q 12 Q 13 Q 12 Q 22 Q 23 Q 13 Q 23 Q 33
The solution of unknown number is
δ x=-[al]Q 11-[bl]Q 12-[cl]Q 13………………19)
δ y=-[al]Q 12-[bl]Q 22-[cl]Q 23………………20)
δ z=-[al]Q 13-[bl]Q 23-[cl]Q 33………………21)
Coordinate after centre of sphere adjustment is:
x=x 0x………………22)
y=y 0y………………23)
z=z 0z………………24)
The radius of a ball after adjustment is:
R = R 0 - [ l ] m ………………25)
In formula, molecule is observed reading equation constant term sum, the number that denominator is observed reading.
According to said method, just can comparatively accurately obtain the centre of sphere and the radius of the spherical rotary of single point in time, certain computing method have a variety of, and the present invention is only as example.
< turning axle computing method >
The rotating shaft of sphere is the straight line through the centre of sphere, and this axis is parallel with the normal vector of Plane of rotation.In other words, only requiring to obtain the normal vector of Plane of rotation, then substitution sphere centre coordinate, can determine unique turning axle.Now, how the normal vector of Plane of rotation calculates the key that has just become problem.
Here, so-called Plane of rotation refers to, at time 1all point coordinate in moment revert to the shape that forms a similar sphere in coordinate system, choose the coordinate points D that approaches each finger fingertip 1to D tfor sampled point, what these sampled points can be similar to fits within same plane, and this plane is exactly the Plane of rotation in this moment.
Can determine straight line according to 2 in space, the straight line of two intersections just can be determined the principle of a plane, as long as there are not four points on same straight line just can determine a plane, and calculates his normal vector.
Supposing has D 1to D tindividual, be roughly dispersed in four to five regions, as long as choose respectively in zones of different, just can determine multiple Plane of rotations and calculate multiple normal vectors, these normal vectors are done to vectorial multiplication cross computing, if two planes are parallel, the result of vectorial multiplication cross is 0 again.In the present invention, can, to use the method for multidirectional amount matching, determine the normal vector of Plane of rotation.
Calculate after the information such as the centre of sphere, radius, turning axle of the simulation spheric motion of single point in time, just can use the same method to time 2to time mthe coordinate in moment carries out modeling, calculates the spherical rotary information in each moment, as shown in the table:
According to these information, just can further calculate the information such as centre of sphere change vector, the vector variation of turning axle, can do to the variation track of whole gesture the modeling analysis of a quantification.
Finally, user's operation behavior Information encapsulation is become to class, return to application programming interfaces.
Above-mentioned several users' operation behavior is just done simple a division from the angle of changes in coordinates, because the action of staff can be ever-changing, can also, according to the information such as speed and time of changes in coordinates, carry out finer analysis to movement locus.
In addition, although the present invention is illustrated as example to be arranged on distance on four jiaos of three dimensional displays and angular transducer, but the present invention is not limited thereto, the present invention can by be arranged on a radargrammetry user finger of mobile phone screen periphery certain a bit and the air line distance of radar and line and the three dimensional display angulation of this point and radar., the present invention can replace range sensor and angular transducer measuring distance information and angle information by radar.At this, the present invention does not limit the magnitude setting of radar, for example can two radars be set at mobile phone screen periphery, a radar is for measuring distance information and angle information, and another radar is for measured range information and the angle information accuracy of verification.
The present invention realizes the 3D of three dimensional display to control, and applies very extensive.Be used in the middle of the demonstration of 3D game and 3D effect figure, the operation behavior of the gesture analysis user by user, is packaged into unified UI interface, and application can utilize the UI style of these Interface realization different-effects, experiences to the diverse user of user.When the present invention is really applied on mobile phone, its application part is innumerable.For example:
1.3D cuts fruit: in space, do the action cutting fruit really by hand, complete game.
2. virtual magic square: user can directly do by gesture various rotation operations before mobile phone screen, and the magic square of directly controlling in screen forwards to, realizes 3D bandwagon effect.
3.TOM cat: user can do an action of drawing before mobile phone screen, removes to draw hair ear or the hair of TOM, to more expressing one's feelings and move TOM collocation in.
4.3D desktop: break original 2D desktop effect, the icon on desktop can be to make stereoeffect, and desk-top space also can become 3d space, by gesture, icon is drawn in, puts far, amplifies, and the operation such as dwindles.
The example of 3D operation is too numerous to enumerate, the invention has the advantages that and utilizes the gesture in space to go to judge user's behavior, is different from existing touch screen technology completely, and user just can complete operation without touching screen.And liberate finger, touching operation has been become to spatial movement behavioural analysis.User can have more selection, and more gesture removes to control mobile phone.Also for UI layer provides abundant class of operation, to the perfection support of 3D effect, allow UI become and obtain aestheticism more simultaneously, more perceptual.

Claims (13)

1. a three-dimensional manipulating method, is characterized in that, comprises step:
Based on the measured range information of detection part and angle information analysis with judge user's operation behavior;
According to the judged result of user's operation behavior and decision type, quantization operation parameter;
According to action type and operating parameter, carry out three-dimensional manipulating in user interface.
2. three-dimensional manipulating method according to claim 1, is characterized in that, described detection part comprises three range sensors and angular transducer.
3. three-dimensional manipulating method according to claim 2, is characterized in that, described detection part further comprises that a range sensor and angular transducer are as check sensor.
4. three-dimensional manipulating method according to claim 1, is characterized in that, described detection part comprises radar.
5. three-dimensional manipulating method according to claim 1, is characterized in that, described operating parameter comprise movement velocity, direction of motion and with screen angulation.
6. three-dimensional manipulating method according to claim 1, is characterized in that, described user's operation behavior is divided into one-way movement, Multidirectional motion and rotation.
7. three-dimensional manipulating method according to claim 6, is characterized in that, the user's of described one-way movement operation behavior judges as follows:
Whether the coordinate that judges the upper each point of finger variation tendency is at a time identical;
Judge that it is linear in three coordinate planes, whether having a coordinate projection in plane at least;
User's operation behavior is packaged into class.
8. three-dimensional manipulating method according to claim 6, is characterized in that, the user's of described Multidirectional motion operation behavior judges as follows:
Judgement finger number;
Simulate the trail change of finger in space;
User's operation behavior is packaged into class.
9. three-dimensional manipulating method according to claim 6, is characterized in that, the user's of described rotation operation behavior judges as follows:
Tentatively judge whether user's gesture rotates;
Judgement finger number;
The rotation trend of analysis user gesture;
Analysis user is pointed the movement locus in space;
User's operation behavior is packaged into class.
10. a three-dimensional manipulating device, is characterized in that, comprises three dimensional display, controller, storer and detection part.
11. three-dimensional manipulating devices according to claim 10, is characterized in that, described multiple distances and angular transducer are arranged on the periphery of described three dimensional display.
12. three-dimensional manipulating devices according to claim 10, it is characterized in that, described controller is according to the information analysis of described multiple distances and angular transducer and judge user's operation behavior, and judged user's operation behavior is referred to a kind of action type.
13. three-dimensional manipulating devices according to claim 12, is characterized in that, described three dimensional display shows three-dimensional picture according to described action type.
CN201310075672.4A 2013-03-08 2013-03-08 Three-dimensional manipulating method and three-dimensional manipulating device Expired - Fee Related CN104035552B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310075672.4A CN104035552B (en) 2013-03-08 2013-03-08 Three-dimensional manipulating method and three-dimensional manipulating device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310075672.4A CN104035552B (en) 2013-03-08 2013-03-08 Three-dimensional manipulating method and three-dimensional manipulating device

Publications (2)

Publication Number Publication Date
CN104035552A true CN104035552A (en) 2014-09-10
CN104035552B CN104035552B (en) 2017-08-25

Family

ID=51466354

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310075672.4A Expired - Fee Related CN104035552B (en) 2013-03-08 2013-03-08 Three-dimensional manipulating method and three-dimensional manipulating device

Country Status (1)

Country Link
CN (1) CN104035552B (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105487692A (en) * 2014-12-22 2016-04-13 哈尔滨安天科技股份有限公司 Controller switching method and system based on three-dimensional display space
CN106662639A (en) * 2014-10-02 2017-05-10 谷歌公司 Non-line-of-sight radar-based gesture recognition
US10310621B1 (en) 2015-10-06 2019-06-04 Google Llc Radar gesture sensing using existing data protocols
US10409385B2 (en) 2014-08-22 2019-09-10 Google Llc Occluded gesture recognition
US10496182B2 (en) 2015-04-30 2019-12-03 Google Llc Type-agnostic RF signal representations
US10642367B2 (en) 2014-08-07 2020-05-05 Google Llc Radar-based gesture sensing and data transmission
CN111198644A (en) * 2018-11-16 2020-05-26 西安易朴通讯技术有限公司 Method and system for identifying screen operation of intelligent terminal
US10817070B2 (en) 2015-04-30 2020-10-27 Google Llc RF-based micro-motion tracking for gesture tracking and recognition
US10936085B2 (en) 2015-05-27 2021-03-02 Google Llc Gesture detection and interactions
US10948996B2 (en) 2014-06-03 2021-03-16 Google Llc Radar-based gesture-recognition at a surface of an object
US11140787B2 (en) 2016-05-03 2021-10-05 Google Llc Connecting an electronic component to an interactive textile
US11169988B2 (en) 2014-08-22 2021-11-09 Google Llc Radar recognition-aided search

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11134089A (en) * 1997-10-29 1999-05-21 Takenaka Komuten Co Ltd Hand pointing device
CN102236414A (en) * 2011-05-24 2011-11-09 北京新岸线网络技术有限公司 Picture operation method and system in three-dimensional display space
CN102804786A (en) * 2009-06-16 2012-11-28 Lg电子株式会社 Viewing range notification method and TV receiver for implementing the same
CN102810015A (en) * 2011-05-31 2012-12-05 中兴通讯股份有限公司 Input method and terminal based on space motion

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11134089A (en) * 1997-10-29 1999-05-21 Takenaka Komuten Co Ltd Hand pointing device
CN102804786A (en) * 2009-06-16 2012-11-28 Lg电子株式会社 Viewing range notification method and TV receiver for implementing the same
CN102236414A (en) * 2011-05-24 2011-11-09 北京新岸线网络技术有限公司 Picture operation method and system in three-dimensional display space
CN102810015A (en) * 2011-05-31 2012-12-05 中兴通讯股份有限公司 Input method and terminal based on space motion

Cited By (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10948996B2 (en) 2014-06-03 2021-03-16 Google Llc Radar-based gesture-recognition at a surface of an object
US10642367B2 (en) 2014-08-07 2020-05-05 Google Llc Radar-based gesture sensing and data transmission
US11169988B2 (en) 2014-08-22 2021-11-09 Google Llc Radar recognition-aided search
US10936081B2 (en) 2014-08-22 2021-03-02 Google Llc Occluded gesture recognition
US12153571B2 (en) 2014-08-22 2024-11-26 Google Llc Radar recognition-aided search
US11816101B2 (en) 2014-08-22 2023-11-14 Google Llc Radar recognition-aided search
US10409385B2 (en) 2014-08-22 2019-09-10 Google Llc Occluded gesture recognition
US11221682B2 (en) 2014-08-22 2022-01-11 Google Llc Occluded gesture recognition
US11163371B2 (en) 2014-10-02 2021-11-02 Google Llc Non-line-of-sight radar-based gesture recognition
US10664059B2 (en) 2014-10-02 2020-05-26 Google Llc Non-line-of-sight radar-based gesture recognition
CN106662639B (en) * 2014-10-02 2019-07-26 谷歌有限责任公司 Non-Line-of-Sight Radar-Based Gesture Recognition
CN106662639A (en) * 2014-10-02 2017-05-10 谷歌公司 Non-line-of-sight radar-based gesture recognition
CN105487692A (en) * 2014-12-22 2016-04-13 哈尔滨安天科技股份有限公司 Controller switching method and system based on three-dimensional display space
US10496182B2 (en) 2015-04-30 2019-12-03 Google Llc Type-agnostic RF signal representations
US11709552B2 (en) 2015-04-30 2023-07-25 Google Llc RF-based micro-motion tracking for gesture tracking and recognition
US10817070B2 (en) 2015-04-30 2020-10-27 Google Llc RF-based micro-motion tracking for gesture tracking and recognition
US10936085B2 (en) 2015-05-27 2021-03-02 Google Llc Gesture detection and interactions
US10540001B1 (en) 2015-10-06 2020-01-21 Google Llc Fine-motion virtual-reality or augmented-reality control using radar
US11698439B2 (en) 2015-10-06 2023-07-11 Google Llc Gesture recognition using multiple antenna
US10823841B1 (en) 2015-10-06 2020-11-03 Google Llc Radar imaging on a mobile computing device
US10768712B2 (en) 2015-10-06 2020-09-08 Google Llc Gesture component with gesture library
US10705185B1 (en) 2015-10-06 2020-07-07 Google Llc Application-based signal processing parameters in radar-based detection
US11132065B2 (en) 2015-10-06 2021-09-28 Google Llc Radar-enabled sensor fusion
US10310621B1 (en) 2015-10-06 2019-06-04 Google Llc Radar gesture sensing using existing data protocols
US12117560B2 (en) 2015-10-06 2024-10-15 Google Llc Radar-enabled sensor fusion
US10503883B1 (en) 2015-10-06 2019-12-10 Google Llc Radar-based authentication
US11175743B2 (en) 2015-10-06 2021-11-16 Google Llc Gesture recognition using multiple antenna
US10459080B1 (en) 2015-10-06 2019-10-29 Google Llc Radar-based object detection for vehicles
US11256335B2 (en) 2015-10-06 2022-02-22 Google Llc Fine-motion virtual-reality or augmented-reality control using radar
US11385721B2 (en) 2015-10-06 2022-07-12 Google Llc Application-based signal processing parameters in radar-based detection
US11481040B2 (en) 2015-10-06 2022-10-25 Google Llc User-customizable machine-learning in radar-based gesture detection
US11656336B2 (en) 2015-10-06 2023-05-23 Google Llc Advanced gaming and virtual reality control using radar
US11693092B2 (en) 2015-10-06 2023-07-04 Google Llc Gesture recognition using multiple antenna
US10908696B2 (en) 2015-10-06 2021-02-02 Google Llc Advanced gaming and virtual reality control using radar
US11698438B2 (en) 2015-10-06 2023-07-11 Google Llc Gesture recognition using multiple antenna
US10401490B2 (en) 2015-10-06 2019-09-03 Google Llc Radar-enabled sensor fusion
US12085670B2 (en) 2015-10-06 2024-09-10 Google Llc Advanced gaming and virtual reality control using radar
US10379621B2 (en) 2015-10-06 2019-08-13 Google Llc Gesture component with gesture library
US11140787B2 (en) 2016-05-03 2021-10-05 Google Llc Connecting an electronic component to an interactive textile
CN111198644B (en) * 2018-11-16 2023-10-27 西安易朴通讯技术有限公司 Screen operation identification method and system of intelligent terminal
CN111198644A (en) * 2018-11-16 2020-05-26 西安易朴通讯技术有限公司 Method and system for identifying screen operation of intelligent terminal

Also Published As

Publication number Publication date
CN104035552B (en) 2017-08-25

Similar Documents

Publication Publication Date Title
CN104035552A (en) Three-dimensional operation method and three-dimensional operation device
US11054912B2 (en) Three-dimensional graphical user interface for informational input in virtual reality environment
CN100483319C (en) Use of a two finger input on touch screens
CN102216879B (en) Method and device for inputting a user&#39;s instructions based on movement sensing
CN105453016B (en) In response to the context-sensitive movement of touch input
CN101819462B (en) Image texture haptic representation system based on force/haptic interaction equipment
KR101318244B1 (en) System and Method for Implemeting 3-Dimensional User Interface
TWI463390B (en) Positioning method for touch display panel
CN101632029A (en) Enhanced single-sensor position detection
CA2937956A1 (en) In-air ultrasound pen gestures
CN105302407A (en) Application icon display method and apparatus
CN102096548A (en) Method and system for duplicating an object using a touch-sensitive display
TWI506243B (en) System and method for simulating a calibration path of a probe of a measuring machine
US10895950B2 (en) Method and system for generating a holographic image having simulated physical properties
CN111665933A (en) Method and device for operating object in virtual or augmented reality
CN103940442A (en) Location method and device adopting accelerating convergence algorithm
US20110007007A1 (en) Touch control method
Pryss et al. Enabling tracks in location-based smart mobile augmented reality applications
CN103543885B (en) The multimedia spherical screen demonstration instrument of multi-point touch and multi-touch method thereof
Baharav et al. Capacitive touch sensing: signal and image processing algorithms
WO2018098771A1 (en) Method for determining azimuthal angle or posture, touch input device, touch panel, and system
CN103513919B (en) A kind of touch control device and the display control method of 3D rotating disk, device
US20130249807A1 (en) Method and apparatus for three-dimensional image rotation on a touch screen
CN104050721B (en) The method and device of smooth control three dimensional object
CN103593835A (en) An accurate measuring method for a three dimensional image reconstruction body

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20170825

Termination date: 20210308