Design of Desktop Audiovisual Entertainment System with Deep Learning and Haptic Sensations
"> Figure 1
<p>The system framework of Sense.</p> "> Figure 2
<p>(<b>a</b>) The appearance of the 4D sensation device; (<b>b</b>) the experience area of the 4D sensation device.</p> "> Figure 3
<p>Module installation.</p> "> Figure 4
<p>(<b>a</b>) Sense operating software; (<b>b</b>) control software of the 4D sensation device.</p> "> Figure 5
<p>Examples of detected results after applying Google Cloud Vision.</p> "> Figure 6
<p>Some animation images among the dataset used in this study.</p> "> Figure 7
<p>The actual trial scenarios.</p> "> Figure 8
<p>AUC results of some fire-related objects.</p> "> Figure 9
<p>Sixty-four test animation images.</p> "> Figure 10
<p>(<b>a</b>) The test animation image; (<b>b</b>) the testing result of the proposed SSD-based system; (<b>c</b>) the testing result of Google Cloud Vision.</p> ">
Abstract
:1. Introduction
2. Design of 4D Sensation Device
2.1. Haptic Modules of 4D Sensation Device
- A.
- Heat Module
- B.
- Hot Air Module
- C.
- Wet Module
- D.
- Wind Module
- E.
- Scent Module
- F.
- Vibration Module
2.2. Hardware Core of 4D Sensation Device
3. Scene Recognition System
3.1. Scene Recognition System Based on Google Cloud Vision
3.2. Scene Recognition System Based on SSD
4. Experiments and Field Trial
4.1. User Experience and Field Trials of 4D Haptic Sensation Device
- Two testers could not clearly smell the odor emitted by the scent module; therefore, we increased the concentration of essential oil in the scent module. The testers could clearly smell the odor in the retest.
- Simultaneously running two modules provided a highly intense haptic experience. Based on the combinations of various modules, this system could create multiple levels of haptic experiences.
- In two tests, one tester felt that the timing of the scent module seemed irrational and inappropriate. This was possibly because when the scent module was run, the background scene was a forest or a garden, but the user was focused on the story plot and therefore was not paying attention to the change in scenery. Additionally, because forest scenes appeared frequently throughout the video, causing the scent module to be run frequently, and owing to olfactory fatigue, the user was unable to clearly detect scent in the end.
- When scenes presented low frequencies and loud sounds (e.g., explosions), all users could feel the vibrating feeling generated by the vibration module. All testers agreed that the timing of the vibration module was the most accurate and that the experience was exhilarating.
4.2. Simulation Performance of the SSD-Based Animation Detection System
4.3. Comparison SSD with Google Cloud Vision
5. Conclusions
Author Contributions
Funding
Conflicts of Interest
Ethics Committee Approval
References
- Gonçalves, G.; Melo, M.; Vasconcelos-Raposo, J.; Bessa, M.E. Impact of different sensory stimuli on presence in credible virtual environments. IEEE Trans. Vis. Comput. Graph. 2019, 26, 3231–3240. [Google Scholar] [CrossRef]
- Chen, Y.-S.; Han, P.-H.; Hsiao, J.-C.; Lee, K.-C.; Hsieh, C.-E.; Lu, K.-Y.; Chou, C.-H.; Hung, Y.-P. SoEs: Attachable Augmented Haptic on Gaming Controller for Immersive Interaction. In Proceedings of the 29th Annual Symposium on User Interface Software and Technology, Tokyo, Japan, 16–19 October 2016; pp. 71–72. [Google Scholar]
- Han, P.H.; Chen, Y.S.; Yang, K.T.; Chuan, W.S.; Chang, Y.T.; Yang, T.M.; Lin, J.Y.; Lee, K.C.; Hsieh, C.E.; Lee, L.C.; et al. BoEs: Attachable Haptics Bits On Gaming Controller For Designing Interactive Gameplay. In Proceedings of the ACM SIGGRAPH Asia 2017 VR Showcase, Bangkok, Thailand, 27–30 November 2017; Article No. 3. pp. 1–2. [Google Scholar]
- Han, P.-H.; Hsieh, C.-E.; Chen, Y.-S.; Hsiao, J.-C.; Lee, K.-C.; Ko, S.F.; Chen, K.W.; Chou, C.-H.; Hung, Y.-P. AoEs: Enhancing teleportation experience in immersive environment with mid-air haptics. In Proceedings of the ACM SIGGRAPH 2017 Emerging Technologies, Los Angeles, CA, USA, 30 July–3 August 2017; Article No. 3. pp. 1–2. [Google Scholar]
- Günther, S.; Müller, F.; Schön, D.; Elmoghazy, O.; Mühlhäuser, M.; Schmitz, M. Therminator: Understanding the Interdependency of Visual and On-Body Thermal Feedback in Virtual Reality. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; pp. 1–14. [Google Scholar]
- Sasaki, T.; Liu, K.-H.; Hasegawa, T.; Hiyama, A.; Inami, M. Virtual Super-Leaping: Immersive Extreme Jumping in VR. In Proceedings of the 10th Augmented Human International Conference 2019 (AH2019), Reims, France, 11–12 March 2019; Article No. 18. pp. 1–8. [Google Scholar]
- TensorFlow Object Detection API. Available online: https://github.com/TensorFlow/models/tree/master/research/object_detection (accessed on 24 November 2019).
- Google Cloud Vision. Available online: https://cloud.google.com/vision/docs/2018 (accessed on 8 September 2018).
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. arXiv 2015, arXiv:1506.01497. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Bergman, D.L. Symmetry Constrained Machine Learning. In Proceedings of the 2019 Intelligent Systems and Applications Conference, Las Palmas, Spain, 7–12 January 2019; pp. 501–512. [Google Scholar]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
- Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Xu, K.; Wang, X.; Liu, X.; Cao, C.; Li, H.; Peng, H.; Wang, D. A dedicated hardware accelerator for real-time acceleration of YOLOv2. J. Real Time Image Process. 2020. [Google Scholar] [CrossRef]
- Maher, A.; Taha, H.; Zhang, B. Realtime multi-aircraft tracking in aerial scene with deep orientation network. J. Real Time Image Process. 2018, 15, 495–507. [Google Scholar] [CrossRef]
- Chen, T.; Li, M.; Li, Y.; Lin, M.; Wang, N.; Wang, M.; Xiao, T.; Xu, B.; Zhang, C.; Zhang, Z. MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous Distributed Systems. In Proceedings of the Advances in Neural Information Processing System, Workshop on Machine Learning System, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S. SSD: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
- Ko, S.F.; Lin, Y.H.; Han, P.H.; Chang, C.C.; Chou, C.H. Combining Deep Learning Algorithm with Scene Recognition and Haptic Feedback for 4D–VR Cinema. In Proceedings of the ACM SIGGRAPH Asia 2018, Tokyo, Japan, 4–7 December 2018; Article No. 18. pp. 1–2. [Google Scholar]
- Fusion 360. Available online: https://www.autodesk.com/products/fusion-360/overview (accessed on 19 October 2017).
- Operating Video of Our System Based on Google Could Vision. Available online: https://youtu.be/r-iIezN6quI (accessed on 25 August 2020).
- Operating Video of Our System Based on SSD Deep Learning Algorithm. Available online: https://youtu.be/2-XtgCypyYA (accessed on 25 August 2020).
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
- Su, Y.S.; Chou, C.H.; Chu, Y.L.; Yang, Z.F. A finger-worn device for exploring Chinese printed text with using CNN algorithm on a micro IoT processor. IEEE ACCESS 2019, 7, 116529–116541. [Google Scholar] [CrossRef]
- Su, Y.S.; Lin, C.L.; Chen, S.Y.; Lai, C.F. Bibliometric study of social network analysis literature. Libr. Hi Tech. 2019, 38, 420–433. [Google Scholar] [CrossRef]
- Su, Y.S.; Chen, H.R. Social Facebook with Big Six approaches for improved students’ learning performance and behavior: A case study of a project innovation and implementation course. Front. Psychol. 2020, 11, 1166. [Google Scholar] [CrossRef]
- Su, Y.S.; Ni, C.F.; Li, W.C.; Lee, I.H.; Lin, C.P. Applying deep learning algorithms to enhance simulations of large-scale groundwater flow in IoTs. Appl. Soft Comput. 2020, 92, 106298. [Google Scholar] [CrossRef]
Heat Module (Infrared Heat Lamp) | Hot Air Module (Mini Space Heater) | ||
Wet Module (Ultrasonic Humidifier Generator) | Wind Module (Fan) | ||
Scent Module (Ultrasonic Humidifier Mist Generator) | Vibration Module (Vibration Speaker) |
Objects | 4D Haptic Feedback Device | |||||
---|---|---|---|---|---|---|
Heat1 | Heat2 | Hot Air | Wet | Wind | Scent | |
Heat | √ | |||||
Wildfire | √ | |||||
Flame | √ | √ | ||||
Fire | √ | √ | ||||
Explosion | √ | √ | √ | |||
Waterfall | √ | |||||
Rain | √ | |||||
Snow | √ | √ | ||||
winter | √ | √ | ||||
Wind | √ | |||||
Wing | √ | |||||
Forest | √ | |||||
Tree | √ |
Module Type | Intensity of Feeling No Feeling (1) to Intense Feeling (5) | ||||
---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | |
Heat Module | 0 | 0 | 1 | 2 | 7 |
Hot Air Module | 0 | 2 | 1 | 4 | 3 |
Wind Module | 0 | 0 | 2 | 5 | 3 |
Wet Module | 0 | 0 | 3 | 1 | 6 |
Scent Module | 0 | 2 | 3 | 2 | 3 |
Vibration Module | 0 | 0 | 1 | 2 | 7 |
Wind & Wet Modules | 0 | 0 | 1 | 1 | 8 |
Hot Air & Vibration Module | 0 | 0 | 0 | 5 | 5 |
Module Type | Rational Time for Running Haptic Module Irrational (1) to Very Rational (5) | ||||
---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | |
Heat Module | 0 | 0 | 1 | 3 | 6 |
Hot Air Module | 0 | 0 | 1 | 3 | 6 |
Wind Module | 0 | 0 | 0 | 4 | 6 |
Wet Module | 0 | 0 | 0 | 4 | 6 |
Scent Module | 0 | 1 | 4 | 3 | 2 |
Vibration Module | 0 | 0 | 1 | 2 | 7 |
Overall Evaluation | 0 | 0 | 0 | 5 | 5 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chou, C.-H.; Su, Y.-S.; Hsu, C.-J.; Lee, K.-C.; Han, P.-H. Design of Desktop Audiovisual Entertainment System with Deep Learning and Haptic Sensations. Symmetry 2020, 12, 1718. https://doi.org/10.3390/sym12101718
Chou C-H, Su Y-S, Hsu C-J, Lee K-C, Han P-H. Design of Desktop Audiovisual Entertainment System with Deep Learning and Haptic Sensations. Symmetry. 2020; 12(10):1718. https://doi.org/10.3390/sym12101718
Chicago/Turabian StyleChou, Chien-Hsing, Yu-Sheng Su, Che-Ju Hsu, Kong-Chang Lee, and Ping-Hsuan Han. 2020. "Design of Desktop Audiovisual Entertainment System with Deep Learning and Haptic Sensations" Symmetry 12, no. 10: 1718. https://doi.org/10.3390/sym12101718
APA StyleChou, C. -H., Su, Y. -S., Hsu, C. -J., Lee, K. -C., & Han, P. -H. (2020). Design of Desktop Audiovisual Entertainment System with Deep Learning and Haptic Sensations. Symmetry, 12(10), 1718. https://doi.org/10.3390/sym12101718