Ma et al., 2024 - Google Patents
Trigger motion and interface optimization of an eye-controlled human-computer interaction system based on voluntary eye blinksMa et al., 2024
- Document ID
- 7055957232841282581
- Author
- Ma G
- He J
- Chen C
- Niu Y
- Zhang L
- Zhou T
- Publication year
- Publication venue
- Human–Computer Interaction
External Links
Snippet
Eye-controlled human-computer interaction (ECHCI) attracts attention for its human- centered, natural and direct operation characteristics. The most common ECHCI trigger motion is “gazing,” which causes Midas Touch problems, lowering usability and operation …
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for entering handwritten data, e.g. gestures, text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterized by the transducing means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F19/00—Digital computing or data processing equipment or methods, specially adapted for specific applications
- G06F19/30—Medical informatics, i.e. computer-based analysis or dissemination of patient or disease data
- G06F19/34—Computer-assisted medical diagnosis or treatment, e.g. computerised prescription or delivery of medication or diets, computerised local control of medical devices, medical expert systems or telemedicine
- G06F19/3406—Local monitoring or local control of medical devices, e.g. configuration parameters, graphical user interfaces [GUI] or dedicated hardware interfaces
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F19/00—Digital computing or data processing equipment or methods, specially adapted for specific applications
- G06F19/30—Medical informatics, i.e. computer-based analysis or dissemination of patient or disease data
- G06F19/32—Medical data management, e.g. systems or protocols for archival or communication of medical images, computerised patient records or computerised general medical references
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Cecotti | Spelling with non-invasive Brain–Computer Interfaces–Current and future trends | |
Cecotti | A multimodal gaze-controlled virtual keyboard | |
Z. Allison et al. | A four-choice hybrid P300/SSVEP BCI for improved accuracy | |
Fager et al. | Patients' experiences with technology during inpatient rehabilitation: opportunities to support independence and therapeutic engagement | |
Ma et al. | Trigger motion and interface optimization of an eye-controlled human-computer interaction system based on voluntary eye blinks | |
KR20180083252A (en) | System and method for biomechanically based eye signals for interacting with real and virtual objects | |
Jantz et al. | A brain-computer interface for extended reality interfaces | |
De Paolis et al. | The effects of touchless interaction on usability and sense of presence in a virtual environment | |
Majaranta | Text entry by eye gaze | |
Meriño et al. | Asynchronous control of unmanned aerial vehicles using a steady-state visual evoked potential-based brain computer interface | |
Kim et al. | Exploration of the virtual reality teleportation methods using hand-tracking, eye-tracking, and eeg | |
Meng et al. | An exploration of hands-free text selection for virtual reality head-mounted displays | |
Okoe et al. | A gaze‐enabled graph visualization to improve graph reading tasks | |
Zuo et al. | Study on the brightness and graphical display object directions of the single-gaze-gesture user interface | |
Niu et al. | Enhancing user experience of eye-controlled systems: design recommendations on the optimal size, distance and shape of interactive components from the perspective of peripheral vision | |
Wang et al. | Research on a spatial–temporal characterisation of blink-triggered eye control interactions | |
Zhang et al. | A novel brain-controlled prosthetic hand method integrating AR-SSVEP augmentation, asynchronous control, and machine vision assistance | |
Skovsgaard | Noise challenges in monomodal gaze interaction | |
Shi et al. | HCI Research and Innovation in China: A 10-Year Perspective | |
Yan et al. | Interactive interface design of aging smart home products based on perception ability evaluation | |
Deng | Multimodal interactions in virtual environments using eye tracking and gesture control. | |
Guo et al. | Bimanual asymmetric coordination in a three-dimensional zooming task based on leap motion: Implications for upper limb rehabilitation | |
Lee | Bridging Divide with Innovative Media: Telexistence and Human Augmentation | |
Sun et al. | “Can It Be Customized According to My Motor Abilities?”: Toward Designing User-Defined Head Gestures for People with Dystonia | |
Mollenbach | Selection strategies in gaze interaction |