US8957854B2 - Zero-click activation of an application - Google Patents
Zero-click activation of an application Download PDFInfo
- Publication number
- US8957854B2 US8957854B2 US13/447,856 US201213447856A US8957854B2 US 8957854 B2 US8957854 B2 US 8957854B2 US 201213447856 A US201213447856 A US 201213447856A US 8957854 B2 US8957854 B2 US 8957854B2
- Authority
- US
- United States
- Prior art keywords
- motion
- software
- jiggle
- computing device
- mouse
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/038—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
Definitions
- This disclosure relates to software applications on a personal computer or other computing device, and more particularly to activation of such applications using a directional pointing input device such as a mouse, trackball, stylus, etc.
- a user of a personal computer often wishes to perform an action, such as a Web search, from the desktop shell, or from within an application such as a browser, email reader or word processor. This generally requires multiple mouse clicks, targeting the mouse over a specific user-interface (UI) widget, or entering a key-chord sequence using a keyboard.
- UI user-interface
- initiating a Web search from within a browser generally requires activating an edit control in a toolbar, clicking the mouse to navigate to a search engine or to activate a context menu, or pressing a memorized key sequence such as Alt-S.
- Initiating such a search from outside a browser also requires first activating the preferred search application, which involves other multiple mouse movements or keystrokes, such as clicking on the browser icon on the desktop.
- Custom input devices and mouse gestures have been devised as an alternative to complicated click or keystroke sequences.
- An application action may be initiated when a detector application recognizes that the mouse has been moved in a predetermined manner. For example, drawing an “S” shape with the mouse could be configured to open the browser to a search engine site.
- This approach has a number of potential drawbacks: (1) It requires the user to manually activate the gesture recognizer, for if it runs all the time it can misinterpret normal mouse movement as a preconfigured gesture. (2) It requires the user to memorize the strokes of the various gesture commands as configured on a specific computer. (3) It requires the user to have sufficient dexterity and motor skills to articulate the gestures. Mouse gestures are therefore difficult for novice users or elderly users. Furthermore, typical mouse gestures often require a preactivation step performed with the mouse (e.g. holding down the right button before drawing the letter “S”).
- the present disclosure describes a simple, effective procedure (and for software to support such a procedure) for initiating a selected application, that does not rely on preactivation or memorizing procedures or require a specific level of manual dexterity on the part of the user.
- the present disclosure provides a method for activating a software application command or function without specific positioning of a mouse or clicking a mouse button (“zero-click activation”). This is done in at least one embodiment by evaluating motion of an input device with regard to predetermined criteria including at least one of distance, displacement, velocity, and/or acceleration so as to detect rapid, direction changing movement of the input device such as a shaking or jiggling up and down or side to side or angular motion or other seemingly erratic patterned movement of the input device; and initiating a preconfigured application in response to detection of said shaking motion.
- the input device is a mouse, and the method is performed without actuating a button on the mouse.
- the preconfigured application may be a search function, and in particular may be a Web search invoked when a shaking action is detected while the user is viewing a Web site.
- This shaking function unlike the shaking of a mouse to bring a computer out of a screen saver or suspend state, actually invokes or instantiates a software operation or function, with little effort on the part of a user.
- a mouse shake or jiggle may be used for immediately accessing a Web search, such as a Yahoo!® Web search, from any personal computing device or terminal with a pointing device, or for invoking software commands or functions in programs a user is interacting with.
- a Web search such as a Yahoo!® Web search
- FIG. 1A illustrates a mouse shake or “jiggle” action performed in accordance with an embodiment of the disclosure.
- FIG. 1B schematically illustrates a variety of motion patterns that may be interpreted as a “jiggle” in accordance with the disclosure.
- FIG. 2 is a flowchart of steps in a software-enabled procedure for activating a “Search” function, in accordance with an embodiment of the disclosure.
- FIG. 3 illustrates opening of a search query window in response to a mouse shake, in accordance with an embodiment of the disclosure.
- FIG. 4 is a flowchart of steps in a software-enabled procedure for activating a predetermined function in response to a shake or “jiggle” movement of an input device.
- FIG. 5 is a flowchart of steps in a first algorithm for interpreting a mouse movement as a “jiggle” in accordance with an embodiment of the disclosure.
- FIGS. 6A-6C illustrate coding of initialization, mouse movement, and timeout events, respectively, in the algorithm of FIG. 5 .
- FIG. 7A is a flowchart of steps in a second algorithm for interpreting a mouse movement as a “jiggle” in accordance with another embodiment of the disclosure.
- FIG. 7B is a flowchart giving details of a comparison step shown in FIG. 7A .
- a personal computer 10 is provided with an input device, in this embodiment mouse 1 , controlling a pointer or cursor 2 on a display screen 3 (see FIG. 1A ).
- mouse 1 an input device
- a user of the computer rapidly moves the mouse in changing directions, that is the user shakes (“jiggles”) the mouse to invoke a preconfigured application.
- motion of the mouse may be sideways (x-direction) or forward/reverse (y-direction), as shown in FIG. 1A .
- the rapid, multi-directional motion (shaking motion or “jiggle”) is not limited to the sideways zigzag motion 4 shown. As shown schematically in FIG.
- the motion may follow any of a variety of paths 41 : horizontal, vertical or diagonal zigzag paths, curved paths, or a combination thereof.
- Software-enabled algorithms for detecting a mouse movement and interpreting that movement as a “jiggle” may run constantly without interfering with normal mouse usage; accordingly, they do not need to be enabled using a mode switch, as is generally the case with conventional mouse gesture recognition software. Examples of detection algorithms which may be used in embodiments of the invention are discussed in more detail below.
- the preconfigured application is a “Search” function.
- a default assignment of a mouse shake to a standard “Search” action gives the mouse shake a specific utility and value for a large number of users, similar to the utility of the widely-known context menu associated with the right-click on a typical mouse.
- the action invoked by the mouse shake 4 may be user configurable or application designer configurable, as are most mouse gestures.
- Software interpreting and supporting a mouse shake may be configured, by way of non-limiting example, as a local system- or application- or browser-plugin or separate application that performs a method as diagrammed in the flowchart of FIG. 2 .
- the software detects whenever the user shakes the mouse from any application running on the system, e.g. from the desktop, from a webpage, editing a document, or reading an email (step 201 ).
- Motion of the mouse 1 is evaluated with respect to one or more criteria such as acceleration, velocity, total travel distance, directions traveled over a short time span, displacement or the like to distinguish a rapid up-and-down or side-to-side or other multi-directional shake or “jiggle” from more typical mouse movements encountered during application usage.
- directional changes in the mouse are detected; if such directional changes are within prescribed boundaries over a predetermined time period, the mouse movement is deemed to be a “jiggle.”
- a display screen of the computer is mapped onto a grid, and movement of the mouse between cells of the grid over a predetermined time period with a certain pattern of movement is deemed to be a “jiggle.”
- the system detects the application currently in use (step 202 ), to determine the predetermined type of action to be invoked. For example, in one embodiment, if the application in use is a Web browser, shaking the mouse will invoke a Web search utility whereupon a search query text entry box will appear. In another embodiment, if a word processing application is in use, shaking the mouse can invoke a dictionary lookup or spell check; if another offline application is currently being interacted with, shaking the mouse can invoke a local file search. In still another embodiment, if a word processing application is in use, shaking the mouse can invoke saving the document in a preselected or default folder.
- the appropriate Search function (or other predetermined function) is then started (step 203 ), and a window is opened on the display screen to accept the user's search query (step 204 ).
- the new window 30 including a text edit box 31 for the search query, may be located next to the pointer 2 , and may display the logo of the search function provider (e.g. Yahoo!®) in a portion 32 thereof.
- the search query term is then communicated to a search service provider, in one embodiment the provider of the zero-click activation application, so that the search can be performed in accordance with the query typed by the user in box 31 (step 205 ), and the results displayed on the screen (step 206 ).
- the action taken in response to the mouse shake may be programmed via a remote connection to a server. It will be appreciated that multiple actions may be taken in response to the mouse shake (e.g. audio feedback, visual effects displayed on the desktop, a change in the appearance of the cursor, etc.).
- the pointer 2 is displayed in a different color to alert the user that edit box 31 has been opened and is waiting for input.
- a shaking motion of the mouse may be used to invoke any desired function from any application.
- any user input device involving physical movement may be shaken or “jiggled,” and that motion may be interpreted as an instruction to go to a predetermined function.
- the movement of the device is detected (step 401 ), and that movement is evaluated and interpreted as a “jiggle” (step 402 ), as opposed to a normal input or pointing function.
- control is transferred to a predetermined function (step 403 ).
- the predetermined function may depend upon the application being interacted with at the time the shake or “jiggle” is performed.
- detecting a shake or “jiggle” involves recording movements of the input device, and evaluating those movements in accordance with an algorithm that applies specific criteria.
- the device is a mouse, and that movement of the mouse is captured by the operating system (e.g. by application programming interface “LowLevelMouseHook” in a Microsoft® Win32 operating system) regardless of which application is presently running.
- Software implementing an exemplary algorithm analyzes the mouse movement data to detect changes in the direction of movement.
- a mouse shake or “jiggle” is detected when a prescribed number of directional changes that are within prescribed boundaries is recorded within a specific period of time (the Timeout period).
- FIG. 5 A schematic flowchart for this algorithm is shown in FIG. 5 .
- the software sets the current direction to “Positive” and reads the current mouse x-y position (step 501 ).
- the software receives a signal that the mouse has moved (a “mouse move event”), it compares the current x-y position of the mouse against the previously received position value. This comparison is used to determine whether a directional change has occurred (step 502 ). If the current direction is “Positive,” and if the x value is less than in the previous measurement, then a change in direction to “Negative” has occurred.
- each period of time between direction changes is called a “stage.”
- the software also keeps a running total of the x and y displacements for each stage being currently measured (step 503 ).
- the x and y displacements are evaluated and compared with configurable minimum and maximum values. If the current displacement is greater than the minimum value and less than the maximum value (step 504 ), the stage is completed (step 505 ). If the displacement is out of range, the stage is discounted (step 506 ).
- a “jiggle event” is deemed to have occurred (step 508 ). If the time to complete the stage exceeds the movement timeout value, the entire algorithm resets, so that all stages and displacements are set to zero.
- a user may choose a sensitivity level for the algorithm—that is, the number of directional changes in the mouse movement required to conclude that a “jiggle event” has occurred.
- the default value is 5 directional changes, but a user may set the number to 3.
- the minimum displacement of the mouse is 10 pixels, the maximum displacement is 200 pixels and the Timeout period is 500 msec.
- FIG. 6A illustrates a block of pseudocode 601 for implementing initialization or reset (step 510 in FIG. 5 ) in the above-described algorithm.
- FIG. 6B illustrates a block of pseudocode 602 for detecting an evaluating a change in direction and the total x and y displacements, and for signaling a “jiggle event,” in accordance with this algorithm.
- FIG. 6C illustrates a block of pseudocode 603 for a timeout event, in which displacements are set to zero and the timer is restarted.
- movement of the mouse between cells of a grid comprising pixels of a display device is detected over a prescribed period of time, and a “jiggle event” is deemed to have occurred if the mouse movement fits a prescribed pattern.
- FIG. 7A A schematic flowchart for this algorithm is shown in FIG. 7A .
- the display screen is mapped (step 701 ) onto a grid made up of cells of a convenient size for tracking the mouse movement (as viewed by corresponding movement of a cursor or pointer on the display); for example, the cells may be square and 4 ⁇ 4 pixels in size.
- the mouse movement is then tracked (step 702 ) to determine which cells have been visited in the current movement.
- step 704 the mouse movement is compared with a movement pattern characterized (step 703 ) by a number of cells and the order in which the cells are visited.
- the prescribed movement pattern 750 involves 3 cells in a right-left-right sequence.
- the software determines (steps 751 - 753 ) whether the mouse has moved from a starting cell at least a minimum number N Min , but less than a maximum number N Max , of cells to the right; followed by movement of at least N Min but fewer than N Max cells to the left; and then movement of at least N Min but fewer than N Max cells to the right.
- the software also evaluates (step 706 ) whether the movement has occurred within a predetermined configurable time period (e.g. 500 msec).
- a predetermined configurable time period e.g. 500 msec.
- a shaking or “jiggle” motion of the mouse is assumed to be a side-to-side movement (that is, in the x-direction); any movement including a change of position in the y-direction of more than two cells is ignored (step 707 ).
- steps 704 - 707 are repeated for other variations in the order of visited cells (e.g. left-right-left instead of right-left-right), and in the number of visited cells (step 708 ); the mouse movement is re-evaluated in comparison with a revised movement pattern. If all the criteria are met, a “jiggle event” is judged to have occurred (step 709 ), and the mouse tracking and timer are reset (step 710 ).
- the software for the above-described zero-click activation process may also be built in any way now known or to become known, and, by way of non-limiting example, may be implemented as a browser extension, plug-in or into a browser script library, so as to only take effect during browsing of Web sites, or only certain Web sites, or to work on systems that do not have security permissions to install a local plugin.
- This approach may help raise awareness of the zero-click activation feature. For example, an animated banner advertisement may prompt the user to shake the mouse; when the user does so, the browser detects the shake, opens a search window, and shows an option to download the feature to the client.
- the zero-click activation of an application described above allows for a simple, immediate action to initiate a Web search or other predetermined software operation or function.
- Using the shaking action avoids the problem of training the user to articulate a specific mouse gesture or button sequence. For example, shaking the mouse is much easier to remember that “right button down, move right, move down, move left, right button up”.
- shaking a pointing device (such as a mouse) is not a typical pointing action and is likely to not be misinterpreted as a normal usage of the pointing device (mouse, trackball, stylus, etc.). Accordingly, a preactivation command (e.g. a right-click) is not necessary to first enable recognition or interpretation of the device movement.
- the shaking action as described herein requires no button click or specific level of manual dexterity, so that (unlike mouse gestures) it is appropriate for all level of user expertise and all age levels. It removes the need to carefully articulate the pointer over a small button, icon or UI widget, and is thus easier for a novice user to discover and activate.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
- Position Input By Displaying (AREA)
Abstract
Description
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/447,856 US8957854B2 (en) | 2006-09-27 | 2012-04-16 | Zero-click activation of an application |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/535,588 US8159457B2 (en) | 2006-09-27 | 2006-09-27 | Zero-click activation of an application |
US13/447,856 US8957854B2 (en) | 2006-09-27 | 2012-04-16 | Zero-click activation of an application |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/535,588 Continuation US8159457B2 (en) | 2006-09-27 | 2006-09-27 | Zero-click activation of an application |
Publications (2)
Publication Number | Publication Date |
---|---|
US20120194431A1 US20120194431A1 (en) | 2012-08-02 |
US8957854B2 true US8957854B2 (en) | 2015-02-17 |
Family
ID=39224418
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/535,588 Expired - Fee Related US8159457B2 (en) | 2006-09-27 | 2006-09-27 | Zero-click activation of an application |
US13/447,856 Active US8957854B2 (en) | 2006-09-27 | 2012-04-16 | Zero-click activation of an application |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/535,588 Expired - Fee Related US8159457B2 (en) | 2006-09-27 | 2006-09-27 | Zero-click activation of an application |
Country Status (1)
Country | Link |
---|---|
US (2) | US8159457B2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11126273B2 (en) | 2017-10-20 | 2021-09-21 | Razer (Asia-Pacific) Pte. Ltd. | User input devices and methods for identifying a user input in a user input device |
Families Citing this family (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4767200B2 (en) * | 2007-03-08 | 2011-09-07 | 任天堂株式会社 | Information selection device and information selection program |
US20090235149A1 (en) * | 2008-03-17 | 2009-09-17 | Robert Frohwein | Method and Apparatus to Operate Different Widgets From a Single Widget Controller |
US20090240658A1 (en) * | 2008-03-20 | 2009-09-24 | Alex You-Jen Chang | Search button on input device |
US20090327974A1 (en) * | 2008-06-26 | 2009-12-31 | Microsoft Corporation | User interface for gestural control |
KR101080526B1 (en) * | 2008-07-07 | 2011-11-04 | 장부중 | Method and apparatus for providing user interface |
US9720554B2 (en) * | 2008-07-23 | 2017-08-01 | Robert J. Frohwein | Method and apparatus to operate different widgets from a single widget controller |
US10019081B2 (en) * | 2009-01-15 | 2018-07-10 | International Business Machines Corporation | Functionality switching in pointer input devices |
US9454297B2 (en) * | 2009-06-12 | 2016-09-27 | Steelseries Aps | Navigation device and methods thereof |
KR101105973B1 (en) | 2009-09-03 | 2012-01-18 | 엔에이치엔(주) | System and method for providing web page including mouse gesture function |
US10409851B2 (en) | 2011-01-31 | 2019-09-10 | Microsoft Technology Licensing, Llc | Gesture-based search |
WO2013040761A1 (en) | 2011-09-21 | 2013-03-28 | Microsoft Corporation | Gesture based search |
US10444979B2 (en) | 2011-01-31 | 2019-10-15 | Microsoft Technology Licensing, Llc | Gesture-based search |
WO2012125990A2 (en) | 2011-03-17 | 2012-09-20 | Laubach Kevin | Input device user interface enhancements |
US20120287047A1 (en) * | 2011-05-09 | 2012-11-15 | White William L | Method and apparatus to control computer time out |
US9116558B2 (en) | 2011-10-28 | 2015-08-25 | Atmel Corporation | Executing gestures with active stylus |
KR101116552B1 (en) | 2011-11-21 | 2012-02-28 | 엔에이치엔(주) | System and method for providing web page including mouse gesture function |
DE102012019649A1 (en) * | 2012-10-06 | 2014-04-10 | Leopold Kostal Gmbh & Co. Kg | Method for detecting e.g. finger gesture formed in free space by optical sensor for controlling computer in motor vehicle, involves determining whether value of changing speed of angle exceeds threshold levels within predetermined interval |
US9292107B1 (en) * | 2012-10-29 | 2016-03-22 | Trend Micro Inc. | Mobile telephone as computer mouse |
CN104182404A (en) * | 2013-05-22 | 2014-12-03 | 腾讯科技(深圳)有限公司 | Method and device for realizing shortcut operations of browser, browser and mobile terminal |
US10481769B2 (en) | 2013-06-09 | 2019-11-19 | Apple Inc. | Device, method, and graphical user interface for providing navigation and search functionalities |
JP6314394B2 (en) * | 2013-09-13 | 2018-04-25 | 富士通株式会社 | Information processing apparatus, setting method, setting program, system, and management apparatus |
CN103472921A (en) * | 2013-09-22 | 2013-12-25 | 广东欧珀移动通信有限公司 | Method and device for controlling output of user input information in mobile terminal |
GB201320019D0 (en) | 2013-11-13 | 2013-12-25 | Univ Singapore | Method and hand held laboratory device to control screen navigation |
CN106325702B (en) * | 2015-07-08 | 2021-01-08 | 腾讯科技(深圳)有限公司 | Method and device for recognizing mouse gestures |
US10606457B2 (en) | 2016-10-11 | 2020-03-31 | Google Llc | Shake event detection system |
KR102605448B1 (en) * | 2018-10-30 | 2023-11-22 | 삼성에스디에스 주식회사 | Search method and apparatus thereof |
US11755194B2 (en) | 2020-10-06 | 2023-09-12 | Capital One Services, Llc | Interactive searching using gestures on any mobile search results page |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5598187A (en) * | 1993-05-13 | 1997-01-28 | Kabushiki Kaisha Toshiba | Spatial motion pattern input system and input method |
US6573883B1 (en) * | 1998-06-24 | 2003-06-03 | Hewlett Packard Development Company, L.P. | Method and apparatus for controlling a computing device with gestures |
US20040125073A1 (en) * | 2002-12-30 | 2004-07-01 | Scott Potter | Portable electronic apparatus and method employing motion sensor for function control |
US20050212766A1 (en) * | 2004-03-23 | 2005-09-29 | Reinhardt Albert H M | Translation controlled cursor |
US20060164382A1 (en) * | 2005-01-25 | 2006-07-27 | Technology Licensing Company, Inc. | Image manipulation in response to a movement of a display |
US20070002018A1 (en) * | 2005-06-30 | 2007-01-04 | Eigo Mori | Control of user interface of electronic device |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0736142B2 (en) * | 1991-10-10 | 1995-04-19 | インターナショナル・ビジネス・マシーンズ・コーポレイション | Method and information processing apparatus for recognizing movement stop of movement instruction means |
US5583984A (en) * | 1993-06-11 | 1996-12-10 | Apple Computer, Inc. | Computer system with graphical user interface including automated enclosures |
US6112239A (en) * | 1997-06-18 | 2000-08-29 | Intervu, Inc | System and method for server-side optimization of data delivery on a distributed computer network |
US7750891B2 (en) * | 2003-04-09 | 2010-07-06 | Tegic Communications, Inc. | Selective input system based on tracking of motion parameters of an input device |
JP2001306587A (en) * | 2000-04-27 | 2001-11-02 | Fujitsu Ltd | Device and method for retrieving information, and storage medium |
US6834249B2 (en) * | 2001-03-29 | 2004-12-21 | Arraycomm, Inc. | Method and apparatus for controlling a computing system |
US7256773B2 (en) * | 2003-06-09 | 2007-08-14 | Microsoft Corporation | Detection of a dwell gesture by examining parameters associated with pen motion |
US7477233B2 (en) * | 2005-03-16 | 2009-01-13 | Microsoft Corporation | Method and system for providing modifier key behavior through pen gestures |
-
2006
- 2006-09-27 US US11/535,588 patent/US8159457B2/en not_active Expired - Fee Related
-
2012
- 2012-04-16 US US13/447,856 patent/US8957854B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5598187A (en) * | 1993-05-13 | 1997-01-28 | Kabushiki Kaisha Toshiba | Spatial motion pattern input system and input method |
US6573883B1 (en) * | 1998-06-24 | 2003-06-03 | Hewlett Packard Development Company, L.P. | Method and apparatus for controlling a computing device with gestures |
US20040125073A1 (en) * | 2002-12-30 | 2004-07-01 | Scott Potter | Portable electronic apparatus and method employing motion sensor for function control |
US20050212766A1 (en) * | 2004-03-23 | 2005-09-29 | Reinhardt Albert H M | Translation controlled cursor |
US20060164382A1 (en) * | 2005-01-25 | 2006-07-27 | Technology Licensing Company, Inc. | Image manipulation in response to a movement of a display |
US20070002018A1 (en) * | 2005-06-30 | 2007-01-04 | Eigo Mori | Control of user interface of electronic device |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11126273B2 (en) | 2017-10-20 | 2021-09-21 | Razer (Asia-Pacific) Pte. Ltd. | User input devices and methods for identifying a user input in a user input device |
Also Published As
Publication number | Publication date |
---|---|
US20080074391A1 (en) | 2008-03-27 |
US8159457B2 (en) | 2012-04-17 |
US20120194431A1 (en) | 2012-08-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8957854B2 (en) | Zero-click activation of an application | |
JP5542774B2 (en) | Flick gesture | |
US8147248B2 (en) | Gesture training | |
US10228833B2 (en) | Input device user interface enhancements | |
US7966573B2 (en) | Method and system for improving interaction with a user interface | |
US8176438B2 (en) | Multi-modal interaction for a screen magnifier | |
EP2972727B1 (en) | Non-occluded display for hover interactions | |
US10331219B2 (en) | Identification and use of gestures in proximity to a sensor | |
KR101502659B1 (en) | Gesture recognition method and touch system incorporating the same | |
US6903730B2 (en) | In-air gestures for electromagnetic coordinate digitizers | |
US7849421B2 (en) | Virtual mouse driving apparatus and method using two-handed gestures | |
US20110221666A1 (en) | Methods and Apparatus For Gesture Recognition Mode Control | |
JP2005004732A (en) | Detection of dwell gesture by examining parameter associated with pen motion | |
US20200326841A1 (en) | Devices, methods, and systems for performing content manipulation operations | |
US11216181B2 (en) | Device, method, and graphical user interface for simulating and interacting with handwritten text | |
US20240004532A1 (en) | Interactions between an input device and an electronic device | |
CN107608550A (en) | Touch operation response method and device | |
KR102198596B1 (en) | Disambiguation of indirect input | |
WO2012054212A2 (en) | Scrubbing touch infotip | |
KR101844651B1 (en) | Mouse input device and method of mobile terminal using 3d touch input type in mobile cloud computing client environments | |
KR101692848B1 (en) | Control method of virtual touchpad using hovering and terminal performing the same | |
KR102205235B1 (en) | Control method of favorites mode and device including touch screen performing the same | |
KR20210029175A (en) | Control method of favorites mode and device including touch screen performing the same | |
CN116594533A (en) | Method, device, equipment and medium for processing movement of software interface mouse icon | |
Demski et al. | Context Sensitive Gestures |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: YAHOO HOLDINGS, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YAHOO! INC.;REEL/FRAME:042963/0211 Effective date: 20170613 |
|
AS | Assignment |
Owner name: OATH INC., NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YAHOO HOLDINGS, INC.;REEL/FRAME:045240/0310 Effective date: 20171231 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551) Year of fee payment: 4 |
|
AS | Assignment |
Owner name: VERIZON MEDIA INC., NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:OATH INC.;REEL/FRAME:054258/0635 Effective date: 20201005 |
|
AS | Assignment |
Owner name: VERIZON PATENT AND LICENSING INC., NEW JERSEY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VERIZON MEDIA INC.;REEL/FRAME:057453/0431 Effective date: 20210801 |
|
AS | Assignment |
Owner name: YAHOO! INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:COE, BRIAN;REEL/FRAME:058623/0483 Effective date: 20060926 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |