Akalin et al., 2023 - Google Patents
A taxonomy of factors influencing perceived safety in human–robot interactionAkalin et al., 2023
View HTML- Document ID
- 16007930401221015620
- Author
- Akalin N
- Kiselev A
- Kristoffersson A
- Loutfi A
- Publication year
- Publication venue
- International Journal of Social Robotics
External Links
Snippet
Safety is a fundamental prerequisite that must be addressed before any interaction of robots with humans. Safety has been generally understood and studied as the physical safety of robots in human–robot interaction, whereas how humans perceive these robots has …
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06Q—DATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/10—Office automation, e.g. computer aided management of electronic mail or groupware; Time management, e.g. calendars, reminders, meetings or time accounting
- G06Q10/101—Collaborative creation of products or services
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06Q—DATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce, e.g. shopping or e-commerce
- G06Q30/02—Marketing, e.g. market research and analysis, surveying, promotions, advertising, buyer profiling, customer management or rewards; Price estimation or determination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06Q—DATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Systems or methods specially adapted for a specific business sector, e.g. utilities or tourism
- G06Q50/01—Social networking
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Hong et al. | Why is artificial intelligence blamed more? Analysis of faulting artificial intelligence for self-driving car accidents in experimental settings | |
Gillath et al. | Attachment and trust in artificial intelligence | |
Azevedo-Sa et al. | Real-time estimation of drivers’ trust in automated driving systems | |
Feigh et al. | Toward a characterization of adaptive systems: A framework for researchers and system designers | |
Aroyo et al. | Overtrusting robots: Setting a research agenda to mitigate overtrust in automation | |
Shahrdar et al. | A survey on trust in autonomous systems | |
Salvini et al. | Safety concerns emerging from robots navigating in crowded pedestrian areas | |
Hergeth et al. | Keep your scanners peeled: Gaze behavior as a measure of automation trust during highly automated driving | |
Gebru et al. | A review on human–machine trust evaluation: Human-centric and machine-centric perspectives | |
Huang et al. | Using gaze patterns to predict task intent in collaboration | |
Hoff et al. | Trust in automation: Integrating empirical evidence on factors that influence trust | |
Akalin et al. | A taxonomy of factors influencing perceived safety in human–robot interaction | |
Ganesh | The ironies of autonomy | |
Davoli et al. | On driver behavior recognition for increased safety: a roadmap | |
Nordhoff et al. | Perceived safety and trust in SAE Level 2 partially automated cars: Results from an online questionnaire | |
Zhang et al. | The effects of visual and cognitive distractions on operational and tactical driving behaviors | |
Wintersberger et al. | Trust in technology as a safety aspect in highly automated driving | |
Wintersberger et al. | Evaluating feedback requirements for trust calibration in automated vehicles | |
Delacroix et al. | Smart technologies and our sense of self: Going beyond epistemic counter-profiling | |
Tenhundfeld et al. | Assessment of trust in automation in the “real world”: Requirements for new trust in automation measurement techniques for use by practitioners | |
Sawyer et al. | Human factors and ergonomics in design of a 3: automation, autonomy, and artificial intelligence | |
Cao et al. | Investigating the role of multi-modal social cues in human-robot collaboration in industrial settings | |
Colley et al. | Effects of uncertain trajectory prediction visualization in highly automated vehicles on trust, situation awareness, and cognitive load | |
Zhao et al. | Emotion-matched in-vehicle assistants: an exploration in regulating drivers' incidental emotions and enhancing takeover performance and situational awareness | |
Story et al. | Evaluating the use of human aware navigation in industrial robot arms |