[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

Ma et al., 2024 - Google Patents

Trigger motion and interface optimization of an eye-controlled human-computer interaction system based on voluntary eye blinks

Ma et al., 2024

Document ID
7055957232841282581
Author
Ma G
He J
Chen C
Niu Y
Zhang L
Zhou T
Publication year
Publication venue
Human–Computer Interaction

External Links

Snippet

Eye-controlled human-computer interaction (ECHCI) attracts attention for its human- centered, natural and direct operation characteristics. The most common ECHCI trigger motion is “gazing,” which causes Midas Touch problems, lowering usability and operation …
Continue reading at www.tandfonline.com (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for entering handwritten data, e.g. gestures, text
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterized by the transducing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F19/00Digital computing or data processing equipment or methods, specially adapted for specific applications
    • G06F19/30Medical informatics, i.e. computer-based analysis or dissemination of patient or disease data
    • G06F19/34Computer-assisted medical diagnosis or treatment, e.g. computerised prescription or delivery of medication or diets, computerised local control of medical devices, medical expert systems or telemedicine
    • G06F19/3406Local monitoring or local control of medical devices, e.g. configuration parameters, graphical user interfaces [GUI] or dedicated hardware interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F19/00Digital computing or data processing equipment or methods, specially adapted for specific applications
    • G06F19/30Medical informatics, i.e. computer-based analysis or dissemination of patient or disease data
    • G06F19/32Medical data management, e.g. systems or protocols for archival or communication of medical images, computerised patient records or computerised general medical references
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048

Similar Documents

Publication Publication Date Title
Cecotti Spelling with non-invasive Brain–Computer Interfaces–Current and future trends
Cecotti A multimodal gaze-controlled virtual keyboard
Z. Allison et al. A four-choice hybrid P300/SSVEP BCI for improved accuracy
Fager et al. Patients' experiences with technology during inpatient rehabilitation: opportunities to support independence and therapeutic engagement
Ma et al. Trigger motion and interface optimization of an eye-controlled human-computer interaction system based on voluntary eye blinks
KR20180083252A (en) System and method for biomechanically based eye signals for interacting with real and virtual objects
Jantz et al. A brain-computer interface for extended reality interfaces
De Paolis et al. The effects of touchless interaction on usability and sense of presence in a virtual environment
Majaranta Text entry by eye gaze
Meriño et al. Asynchronous control of unmanned aerial vehicles using a steady-state visual evoked potential-based brain computer interface
Kim et al. Exploration of the virtual reality teleportation methods using hand-tracking, eye-tracking, and eeg
Meng et al. An exploration of hands-free text selection for virtual reality head-mounted displays
Okoe et al. A gaze‐enabled graph visualization to improve graph reading tasks
Zuo et al. Study on the brightness and graphical display object directions of the single-gaze-gesture user interface
Niu et al. Enhancing user experience of eye-controlled systems: design recommendations on the optimal size, distance and shape of interactive components from the perspective of peripheral vision
Wang et al. Research on a spatial–temporal characterisation of blink-triggered eye control interactions
Zhang et al. A novel brain-controlled prosthetic hand method integrating AR-SSVEP augmentation, asynchronous control, and machine vision assistance
Skovsgaard Noise challenges in monomodal gaze interaction
Shi et al. HCI Research and Innovation in China: A 10-Year Perspective
Yan et al. Interactive interface design of aging smart home products based on perception ability evaluation
Deng Multimodal interactions in virtual environments using eye tracking and gesture control.
Guo et al. Bimanual asymmetric coordination in a three-dimensional zooming task based on leap motion: Implications for upper limb rehabilitation
Lee Bridging Divide with Innovative Media: Telexistence and Human Augmentation
Sun et al. “Can It Be Customized According to My Motor Abilities?”: Toward Designing User-Defined Head Gestures for People with Dystonia
Mollenbach Selection strategies in gaze interaction