Abstract
This paper presents an integrated framework for real-time sensing and synchronization of both user’s moving speed with direction and viewing direction in walking-in-place experience for virtual training applications. The framework consists of two inertial measurement units (IMU) attached to each shank and a HMD made up of Android mobile device with 3-axis orientation sensor. Although there are several prior works to enable unconstrained omnidirectional walking through virtual environments, an implementation of the low cost interface solution using wearable devices is an important issue for virtual training systems. We provide a simplified technique for implementing ‘Walking in Virtual Reality’ without omnidirectional treadmill. In addition, this research aims to lightweight (in point of software) and portable (in point of hardware) solution to implement the Virtual Reality Walk-In-Place(VR WIP) interface for training applications.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Three-dimensional (3D) Virtual Reality (VR) technique has been widely used in many applications, including training, education and entertainment systems. However, VR applications based on Head Mount Displays (HMD) often have faced problems of navigation in the virtual environment due to the unrealistic sensing and synchronizing characteristics. Usually, conventional interaction techniques [1] depend only on the orientation of person’s moving body or direction of viewing scenes. The difference in the walking and viewing direction results in an error in navigating trajectory in the virtual environment, which makes the VR system unrealistic. It is inevitable to separate viewing direction and gait direction while exploring virtual environment, especially in VR training systems because the replication of the real situation is very important. In this paper, we propose an integrated framework for solving the problem using IMU attached to each shank and built-in gyro and acceleration sensors in Android mobile devices. In addition, this research aims to provide lightweight software as well as a wearable hardware solution to implement the Virtual Reality Walk-In-Place(VR WIP) [3] interface for training applications.
2 Configuration of the Proposed Framework
The user interface is composed of an Android mobile device having built-in 3-axis orientation sensor and a set of IMU attached to each shank. The IMUs detect gait with walking speed and direction, while, senors in the Android mobile device take information about the viewing orientation of the user’s head movement (Fig. 1).
The Android mobile device can provide 3D VR head-mounted display within a user’s field of view, enabling a user to view virtual objects in the user’s surroundings. Proposed system has applied to a VR training system of disaster rescue from fire in an underground subway station. Figure 2 shows the implemented configuration of the VR training system with the proposed Integrated Framework. VR Contents are played on the Android mobile device, which also provides Bluetooth wireless interface linking to IMUs. The Avatar’s motion is synchronized with user’s movement, and viewing scenery is synchronized to user’s viewing direction. The user needs to escape from the underground firing palace to save her/his life in a limited given time. The user needs to find the best way to escape during the training, where the exit signs among the objects in the user’s view indicate the rescue route.
3 Experimental Results and Discussions
We need to recognize both the speed and direction of user’s gait in order to synchronize with scene displayed in the HMD and motion of the Avatar in the virtual space. Figure 3 shows 3-axis coordinate system of IMUs attached at the shank.
Figures 4 and 5 show measured acceleration of x- and z-axis, and angular velocity of y-axis during walking-in-place actions of the user. We can find a repetitive pattern matching to the each step of gait in the measured data. Variation in y-axis acceleration is negligible because there is a negligible amount of movement in the y-axis direction during walking-in-place motions as shown in Fig. 3. The repetitive pattern in the measured data indicates the step-up, the highest reaching point of foot, and end point of each step. We analyzed these patterns and applied this to machine learning technique to recognize walking state [4–6]. By using learning data, our framework can detect gait speed in real-time.
The walking direction, yaw angle of the user’s leg, can be detected by integration of angular velocity of x-axis. In order to obtain a better estimation of the user’s orientation, we can consider tilt from the output of accelerometers [2, 7] (Fig. 6).
The sensors on Android head-mounted display (HMD) can detect a viewing direction in real-time, enabling to synchronize the virtual surroundings seen through the HMD [8]. Figure 7 shows logged moving trajectory of a Avatar with viewing direction in the disaster rescue VR training system. The scaling factor was used for each user to optimize the performance of the training system. These results indicate that developed algorithm and frame work can be applicable to the VR training system.
4 Conclusion
We present an integrated framework for real-time sensing and synchronization of both user’s moving speed with direction and viewing direction in walking-in-place experience for virtual training applications. The proposed solution can provide a simplified technique for implementing ‘Walking in Virtual Reality’ without omnidirectional treadmill. Experimental results indicate that the algorithm developed for Real-time Sensing of Gait and Viewing Direction can be favorably applicable to the VR training system.
References
Steinicke, F., et al.: Human Walking in Virtual Environments. Springer, New York (2013)
Doheny, E.P., Foran, T.G., Greene, B.R.: A single gyroscope method for spatial gait analysis. In: 32nd Annual International Conference of the IEEE EMBS (2010)
Slater, M., Usoh, M., Steed, A.: Taking steps: the influence of a walking technique on presence in virtual reality. ACM Trans. Comput. Hum. Interact. (TOCHI) 2(3), 201–219 (1995)
Kim, J., Gracanin, D., Quek, F.: Sensor-fusion walking-in-place interaction technique using mobile devices. In: Virtual Reality Short Papers and Posters (VRW) 2012. IEEE (2012)
Mannini, A., Sabatini, A.M.: Machine learning methods for classifying human physical activity from on-body accelerometers. Sens. 10(2), 1154–1175 (2010)
Jasiewicz, J.M., et al.: Gait event detection using linear accelerometers or angular velocity transducers in able-bodied and spinal-cord injured individuals. Gait Posture 24(4), 502–509 (2006)
Luinge, H.J., Veltink, P.H., Baten, C.T.M.: Estimating orientation with gyroscopes and accelerometers. Technol. Health Care 7(6), 455–459 (1999)
Meier, R.: Professional Android 4 Application Development. John Wiley & Sons, New York (2012)
Acknowledgement
This research was supported by the MSIP(Ministry of Science, ICT and Future Planning), Korea, under the “IT Consilience Creative Program” (NIPA-2014-H0201-14-1002) supervised by the NIPA(National IT Industry Promotion Agency).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Ha, G., Lee, S., Cha, J., Lee, H., Kim, T., Kim, S. (2015). A Real-Time Sensing of Gait and Viewing Direction for Human Interaction in Virtual Training Applications. In: Stephanidis, C. (eds) HCI International 2015 - Posters’ Extended Abstracts. HCI 2015. Communications in Computer and Information Science, vol 528. Springer, Cham. https://doi.org/10.1007/978-3-319-21380-4_82
Download citation
DOI: https://doi.org/10.1007/978-3-319-21380-4_82
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-21379-8
Online ISBN: 978-3-319-21380-4
eBook Packages: Computer ScienceComputer Science (R0)