Zhou et al., 2023 - Google Patents
Cross-modal Task Understanding and Execution of Voice-fingertip Reading Instruction by Using Small Family Service RoboticZhou et al., 2023
- Document ID
- 860340990800674765
- Author
- Zhou Z
- Zhu S
- Zhu K
- Cheng C
- Gu J
- Publication year
- Publication venue
- 2022 IEEE International Conference on Cyborg and Bionic Systems (CBS)
External Links
Snippet
The correct understanding of human task instructions is an important basic condition for family service robots to carry out their work. In daily family scenarios, single-modal voice commands often have the problem of missing pronoun references, which makes robots …
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/36—Image preprocessing, i.e. processing the image information without deciding about the identity of the image
- G06K9/46—Extraction of features or characteristics of the image
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Matuszek et al. | Learning from unscripted deictic gesture and language for human-robot interactions | |
Liu et al. | Towards robust human-robot collaborative manufacturing: Multimodal fusion | |
Islam et al. | Understanding human motion and gestures for underwater human–robot collaboration | |
Han et al. | Investigating the use of non-verbal cues in human-robot interaction with a Nao robot | |
Chen et al. | Real-time multi-modal human–robot collaboration using gestures and speech | |
Loper et al. | Mobile human-robot teaming with environmental tolerance | |
Mi et al. | Intention-related natural language grounding via object affordance detection and intention semantic extraction | |
Lim et al. | Rich and robust human-robot interaction on gesture recognition for assembly tasks | |
Wang et al. | Multi-modal interaction with transformers: bridging robots and human with natural language | |
Aly et al. | Towards understanding object-directed actions: A generative model for grounding syntactic categories of speech through visual perception | |
Sowa et al. | Interpretation of shape-related iconic gestures in virtual environments | |
Fan et al. | A vision-language-guided robotic action planning approach for ambiguity mitigation in human–robot collaborative manufacturing | |
Zhou et al. | Cross-modal Task Understanding and Execution of Voice-fingertip Reading Instruction by Using Small Family Service Robotic | |
Kim et al. | Gvcci: Lifelong learning of visual grounding for language-guided robotic manipulation | |
Wang et al. | Design and implementation of virtual reality interactive product software based on artificial intelligence deep learning algorithm | |
Han et al. | Placing objects in gesture space: Toward incremental interpretation of multimodal spatial descriptions | |
Gandage et al. | Virtual Paint | |
Alba-Flores | UAVs control using 3D hand keypoint gestures | |
Halder et al. | Natural Interaction Modalities for Human-CPS Interaction in Construction Progress Monitoring | |
Kim et al. | Joint engagement classification using video augmentation techniques for multi-person HRI in the wild | |
André et al. | Multimodal fusion in human-agent dialogue | |
CN116070173B (en) | Finger reading method and system for cross-modal task instruction understanding | |
Wu et al. | Supervisory System for a Collaborative Robotic Cell Based on RGBD Camera | |
Liu et al. | Intelligent Robotics and Applications: 8th International Conference, ICIRA 2015, Portsmouth, UK, August 24-27, 2015, Proceedings, Part II | |
Kang et al. | IPS: Integrating Pose with Speech for enhancement of body pose estimation in VR remote collaboration |