[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3338286.3344396acmconferencesArticle/Chapter ViewAbstractPublication PagesmobilehciConference Proceedingsconference-collections
research-article

Semantic 3D gaze mapping for estimating focused objects

Published: 01 October 2019 Publication History

Abstract

Eye-trackers are expected to be used in portable daily-use devices. However, it must register object information and define a unified coordinate system in advance for human--computer interaction and quantitative analysis. Therefore, we propose a semantic 3D gaze mapping to collect gaze information from multiple people on the unified map and detect focused objects automatically. The semantic 3D map can be reconstructed using keyframe-based semantic segmentation and structure-from-motion, and the 3D point-of-gaze can also be computed on the map. We confirmed that the fixation time of the focused object can be calculated through an experiment without prior information.

References

[1]
Pablo F Alcantarilla and T Solutions. 2011. Fast explicit diffusion for accelerated features in nonlinear scale spaces. IEEE Trans. Patt. Anal. Mach. Intell 34, 7 (2011), 1281--1298.
[2]
Kai Essig, Daniel Dornbusch, Daniel Prinzhorn, Helge Ritter, Jonathan Maycock, and Thomas Schack. 2012. Automatic analysis of 3D gaze coordinates on scene objects using data from eye-tracking and motion-capture systems. In Proceedings of the Symposium on Eye Tracking Research and Applications. ACM, 37--44.
[3]
Kakeru Hagihara, Keiichiro Taniguchi, Irshad Abibouraguimane, Yuta Itoh, Keita Higuchi, Jiu Otsuka, Maki Sugimoto, and Yoichi Sato. 2018. Object-wise 3D Gaze Mapping in Physical Workspace. In Proceedings of the 9th Augmented Human International Conference. ACM, 25.
[4]
Michael Maurus, Jan Hendrik Hammer, and Jürgen Beyerer. 2014. Realistic heatmap visualization for interactive analysis of 3D gaze data. In Proceedings of the Symposium on Eye Tracking Research and Applications. ACM, 295--298.
[5]
Pierre Moulon, Pascal Monasse, Romuald Perrot, and Renaud Marlet. 2016. Openmvg: Open multiple view geometry. In International Workshop on Reproducible Research in Pattern Recognition. Springer, 60--74.
[6]
Susan M Munn and Jeff B Pelz. 2008. 3D point-of-regard, position and head orientation from a portable monocular video-based eye tracker. In Proceedings of the 2008 symposium on Eye tracking research & applications. ACM, 181--188.
[7]
Jeff B Pelz, Thomas B Kinsman, and Karen M Evans. 2011. Analyzing complex gaze behavior in the natural world. In Human Vision and Electronic Imaging XVI, Vol. 7865. International Society for Optics and Photonics, 78650Z.
[8]
Thies Pfeiffer. 2012. Measuring and visualizing attention in space with 3D attention volumes. In Proceedings of the Symposium on Eye Tracking Research and Applications. ACM, 29--36.
[9]
Thies Pfeiffer, Patrick Renner, and Nadine Pfeiffer-Leßmann. 2016. EyeSee3D 2.0: model-based real-time analysis of mobile eye-tracking in static and dynamic three-dimensional scenes. In Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications. ACM, 189--196.
[10]
Kentaro Takemura, Yuji Kohashi, Tsuyoshi Suenaga, Jun Takamatsu, and Tsukasa Ogasawara. 2010. Estimating 3D point-of-regard and visualizing gaze trajectories under natural head movements. In Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications. ACM, 157--160.
[11]
Kentaro Takemura, Kenji Takahashi, Jun Takamatsu, and Tsukasa Ogasawara. 2014. Estimating 3-D point-of-regard in a real environment using a head-mounted eye-tracking system. IEEE Trans. Human-Machine Syst. 44, 4 (2014), 531--536.
[12]
Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. 2017. Pyramid scene parsing network. In Proceedings of the IEEE conference on computer vision and pattern recognition. 2881--2890.

Cited By

View all
  • (2021)4D Attention: Comprehensive Framework for Spatio-Temporal Gaze MappingIEEE Robotics and Automation Letters10.1109/LRA.2021.30972746:4(7240-7247)Online publication date: Oct-2021

Recommendations

Reviews

Steve Carson

Knowing where users focus their gaze in real, virtual, augmented, or mixed environments is essential for creating immersive applications in areas like gaming, product design, psychology, and advertising. In multiuser environments, knowing the 3D point-of-gaze (PoG) of each user is a key enabler for immersive interaction. Creating such systems is difficult if they require detailed modeling of a part of the real world and associating its elements with a virtual world, making systems that construct their own models from sensory data highly desirable. Previous research has used a wide variety of hardware (virtual reality headsets with eye trackers, head/body position trackers, video cameras) and software techniques (structure from motion, simultaneous localization and mapping, and key frames) to solve aspects of determining 3D PoG focused objects at the expense of complex and time-consuming setup. The authors cleverly leverage this previous research, combining results from multiple open-source software packages to derive semantic (that is, user attention to objects in the local environment) 3D gaze mapping for estimating user focus without prior information. The main contributions of this paper are its excellent descriptive figures, well-defined algorithms, logical expository flow, and presentation of the mathematical details of computing 3D PoG from multiple 2D PoGs projected into key frames by Delaunay triangulation. The authors also conduct an experiment that shows their method can compute approximate user fixation time on objects, making it suitable for marketing and advertising research.

Access critical reviews of Computing literature here

Become a reviewer for Computing Reviews.

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
MobileHCI '19: Proceedings of the 21st International Conference on Human-Computer Interaction with Mobile Devices and Services
October 2019
646 pages
ISBN:9781450368254
DOI:10.1145/3338286
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 October 2019

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. 3D point-of-gaze
  2. Semantic 3D map
  3. Structure from Motion

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

MobileHCI '19
Sponsor:

Acceptance Rates

Overall Acceptance Rate 202 of 906 submissions, 22%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)8
  • Downloads (Last 6 weeks)0
Reflects downloads up to 05 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2021)4D Attention: Comprehensive Framework for Spatio-Temporal Gaze MappingIEEE Robotics and Automation Letters10.1109/LRA.2021.30972746:4(7240-7247)Online publication date: Oct-2021

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media