[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Pistol: Pupil Invisible Supportive Tool in the Wild

Published: 21 February 2024 Publication History

Abstract

This paper is an in the wild evaluation of the eye tracking tool Pistol. Pistol supports Pupil Invisible projects and other eye trackers (Dikablis, Emke GmbH, Look, Pupil, and many more) in offline mode. For all eye tracking recordings, Pistol is able to compute pupil and iris ellipse, eye aperture, eye lids, 3D eye ball, optical vector estimation, and eye movement types. For Pupil Invisible projects, it also supports the detection of a special marker and allows the 2D and 3D gaze estimation mapped on the scene camera. In this extended journal paper for Pistol we focus on challenging in the wild situations, such as reflections on glasses, eye tracking during sport, changing lightning conditions, and this for two eye trackers. The tool Pistol can be downloaded here Link

References

[1]
Gardony AL, Lindeman RW, Brunyé TT. Eye-tracking for human-centered mixed reality: promises and challenges, Vol. 11310, 113100T (International Society for Optics and Photonics, 2020).
[2]
Arslan O, Atik O, and Kahraman S Eye tracking in usability of electronic chart display and information system The Journal of Navigation 2021 74 594-604
[3]
Walton DR et al. Beyond blur: real-time ventral metamers for foveated rendering ACM Trans Graph (TOG) 2021 40 1-14
[4]
Meng X, Du R, and Varshney A Eye-dominance-guided foveated rendering IEEE Trans Visual Comput Graph 2020 26 1972-1980
[5]
Joseph AW, Murugesh R. Potential eye tracking metrics and indicators to measure cognitive load in human-computer interaction research. J Sci Res 2020;64.
[6]
Snell S, Bontempo D, Celine G, Anthonappa R. Assessment of medical practitioners’ knowledge about paediatric oral diagnosis and gaze patterns using eye tracking technology. International Journal of Paediatric Dentistry 2020.
[7]
Lev A, Braw Y, Elbaum T, Wagner M, Rassovsky Y. Eye tracking during a continuous performance test: Utility for assessing adhd patients. Journal of Attention Disorders 1087054720972786 2020.
[8]
Nesaratnam N, Thomas P, and Vivian A Stepping into the virtual unknown: feasibility study of a virtual reality-based test of ocular misalignment Eye 2017 31 1503-1506
[9]
Economides JR, Adams DL, Jocson CM, and Horton JC Ocular motor behavior in macaques with surgical exotropia J Neurophysiol 2007 98 3411-3422
[10]
Davis R and Sikorskii A Eye tracking analysis of visual cues during wayfinding in early stage alzheimer’s disease Dement Geriatr Cogn Disord 2020 49 91-7
[11]
Pavisic IM et al. Eye-tracking indices of impaired encoding of visual short-term memory in familial alzheimer’s disease Sci Rep 2021 11 1-14
[12]
Panchuk D, Vine S, Vickers JN. in Eye tracking methods in sport expertise 176–187 (Routledge, 2015).
[13]
Jermann P, Nüssli M-A, and Li W Using dual eye-tracking to unveil coordination and expertise in collaborative tetris Proc HCI 2010 2010 24 36-44
[14]
Liu C, et al. A gaze model improves autonomous driving, 2019;1–5.
[15]
Shinohara Y, Currano R, Ju W, Nishizaki Y. Visual attention during simulated autonomous driving in the us and japan, 2017;144–153.
[16]
Lepekhin A, Capo D, Levina A, Borremans A, Khasheva Z. Adoption of industrie 4.0 technologies in the manufacturing companies in russia, 2020;1–6.
[17]
Taylor MP. et al. Operator 4.0 or maker 1.0? exploring the implications of industrie 4.0 for innovation, safety and quality of work in small economies and enterprises. Computers & industrial engineering 2020;139:105486.
[18]
Jiang Z, Chang Y, and Liu X Design of software-defined gateway for industrial interconnection J Ind Inf Integr 2020 18
[19]
Hasselbring W, Carr L, Hettrick S, Packer H, and Tiropanis T Open source research software Computer 2020 53 84-88
[20]
Fuhl W. et al.Mam: Transfer learning for fully automatic video annotation and specialized detector creation, 2018;0–0.
[21]
Fuhl W, Weber D, Eivazi S. Pistol: Pupil invisible supportive tool to extract pupil, iris, eye opening, eye movements, pupil and iris gaze vector, and 2d as well as 3d gaze 2023.
[22]
Tobii. Tobii. https://www.tobii.com/ (2021). [Online; accessed 04-November-2021].
[23]
ASL. Asl. https://est-kl.com/ (2021). [Online; accessed 04-November-2021].
[24]
EyeTech. Eyetech. http://www.eyetec.com/ (2021). [Online; accessed 04-November-2021].
[25]
SRResearch. Srresearch. https://www.sr-research.com/ (2021). [Online; accessed 04-November-2021].
[26]
eyecomtec. eyecomtec. https://eyecomtec.com (2021). [Online; accessed 04-November-2021].
[27]
Ergoneers. Ergoneers. https://www.ergoneers.com (2021). [Online; accessed 04-November-2021].
[28]
Oculus. Oculus. https://www.oculus.com (2021). [Online; accessed 04-November-2021].
[29]
PupilLabs. Pupillabs. https://pupil-labs.com/ (2021). [Online; accessed 04-November-2021].
[30]
iMotions. imotions. https://imotions.com/ (2021). [Online; accessed 04-November-2021].
[31]
VIVE. Vive. https://www.vive.com/ (2021). [Online; accessed 04-November-2021].
[32]
Rakhmatulin I. A review of the low-cost eye-tracking systems for 2010-2020. arXiv preprint arXiv:2010.05480 2020.
[33]
Duchowski AT A breadth-first survey of eye-tracking applications Behavior Research Methods, Instruments, & Computers 2002 34 455-470
[34]
Park J and Yim K Technical survey on the real time eye-tracking pointing device as a smart medical equipment Smart Media Journal 2021 10 9-15
[35]
Mao R, Li G, Hildre HP, and Zhang H A survey of eye tracking in automobile and aviation studies: Implications for eye-tracking studies in marine operations IEEE Transactions on Human-Machine Systems 2021 51 87-98
[36]
Santini T, Fuhl W, Geisler D, Kasneci E. Eyerectoo: Open-source software for real-time pervasive head-mounted eye tracking., 2017;96–101.
[37]
Zandi B, Lode M, Herzog A, Sakas G, Khanh TQ. Pupilext: Flexible open-source platform for high-resolution pupillometry in vision research. Frontiers in neuroscience 2021;15.
[38]
Li D, Babcock J, Parkhurst DJ. openeyes: a low-cost head-mounted eye-tracking solution, 2006;95–100.
[39]
Niehorster DC, Hessels RS, and Benjamins JS Glassesviewer: Open-source software for viewing and analyzing data from the tobii pro glasses 2 eye tracker Behav Res Methods 2020 52 1244-1253
[40]
Jones PR. Myex: a matlab interface for the tobii eyex eye-tracker. Journal of Open Research Software 2018;6.
[41]
Phan TV. Development of a custom application for the Tobii Eye Tracker. Ph.D. thesis, Hochschule für angewandte Wissenschaften Hamburg 2011.
[42]
Hale ML Eyestream: An open websocket-based middleware for serializing and streaming eye tracker event data from gazepoint gp3 hd research hardware Journal of Open Source Software 2019 4 1620
[43]
Voßkühler A, Nordmeier V, Kuchinke L, and Jacobs AM Ogama (open gaze and mouse analyzer): open-source software designed to analyze eye and mouse movements in slideshow study designs Behav Res Methods 2008 40 1150-1162
[44]
Benjamins JS, Hessels RS, Hooge IT. Gazecode: Open-source software for manual mapping of mobile eye-tracking data, 2018;1–4.
[45]
Fuhl W. Maximum and leaky maximum propagation. arXiv preprint arXiv:2105.10277 2021.
[46]
Fuhl W. Tensor normalization and full distribution training. arXiv preprint arXiv:2109.02345 2021.
[47]
Fuhl W, Kasneci E. Learning to validate the quality of detected landmarks 2019.
[48]
Bradski G. The opencv library. Dr. Dobb’s Journal: Software Tools for the Professional Programmer 2000;25;120–123.
[49]
Fuhl W, Gao H, Kasneci E. Neural networks for optical vector and eye ball parameter estimation (ACM, 2020).
[50]
Fuhl W, Rong Y, Enkelejda K. Fully convolutional neural networks for raw eye tracking data segmentation, generation, and reconstruction, 2020;0–0.
[51]
Fuhl W, Kasneci G, Kasneci E. Teyed: Over 20 million real-world eye images with pupil, eyelid, and iris 2d and 3d segmentations, 2d and 3d landmarks, 3d eyeball, gaze vector, and eye movement types, 2021;367–375 (IEEE).
[52]
Fuhl W, Santini T, Kübler TC, Kasneci E. Else: Ellipse selection for robust pupil detection in real-world environments, 2016;123–130.
[53]
Fuhl W, Kübler TC, Sippel K, Rosenstiel W, Kasneci E. Excuse: Robust pupil detection in real-world scenarios 2015.
[54]
Fuhl W. et al.Cbf: Circular binary features for robust and real-time pupil center detection, 2018;1–6.
[55]
Santini T, Fuhl W, and Kasneci E Pure: Robust pupil detection for real-time pervasive eye tracking Comput Vis Image Underst 2018 170 40-50
[56]
Santini T, Fuhl W, Kasneci E. Purest: Robust pupil tracking for real-time pervasive eye tracking, 2018;1–5.
[57]
Fuhl W, et al.Bore: Boosted-oriented edge optimization for robust, real time remote pupil center detection, 2018;1–5.
[58]
Fuhl W, Gao H, Kasneci E. Tiny convolution, decision tree, and binary neuronal networks for robust and real time pupil outline estimation, 2020;1–5.
[59]
Fuhl W, Schneider J, Kasneci E. 1000 pupil segmentations in a second using haar like features and statistical learning, 2021;3466–3476.
[60]
Fuhl W, et al.Eyes wide open? eyelid location and eye aperture estimation for pervasive eye tracking in real-world scenarios 2016.
[61]
Fuhl W, Santini T, Kasneci E. Fast and robust eyelid outline and aperture detection in real-world scenarios 2017.
[62]
Kothari RS, Chaudhary AK, Bailey RJ, Pelz JB, and Diaz GJ Ellseg: An ellipse segmentation framework for robust gaze tracking IEEE Trans Visual Comput Graphics 2021 27 2757-2767
[63]
Yiu Y-H et al. Deepvog: Open-source pupil segmentation and gaze estimation in neuroscience using deep learning J Neurosci Methods 2019 324
[64]
Fuhl W, Rong Y, Kasneci E. Fully convolutional neural networks for raw eye tracking data segmentation, generation, and reconstruction, 2021;142–149 (IEEE).
[65]
Fuhl W, Castner N, Kasneci E. Rule-based learning for eye movement type detection, 2018;1–6.
[66]
Fuhl W, Castner N, Kasneci E. Histogram of oriented velocities for eye movement detection, 2018;1–6.
[67]
Fuhl W, Herrmann-Werner A, Nieselt K. The tiny eye movement transformer, 2023;1–2.

Index Terms

  1. Pistol: Pupil Invisible Supportive Tool in the Wild
      Index terms have been assigned to the content through auto-classification.

      Recommendations

      Comments

      Please enable JavaScript to view thecomments powered by Disqus.

      Information & Contributors

      Information

      Published In

      cover image SN Computer Science
      SN Computer Science  Volume 5, Issue 3
      Mar 2024
      750 pages

      Publisher

      Springer-Verlag

      Berlin, Heidelberg

      Publication History

      Published: 21 February 2024
      Accepted: 05 January 2024
      Received: 26 July 2023

      Author Tags

      1. Eye tracking
      2. Gaze
      3. Eye lid
      4. Pupil
      5. Iris
      6. Eye movements

      Qualifiers

      • Research-article

      Funding Sources

      • Eberhard Karls Universität Tübingen (1020)

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • 0
        Total Citations
      • 0
        Total Downloads
      • Downloads (Last 12 months)0
      • Downloads (Last 6 weeks)0
      Reflects downloads up to 21 Dec 2024

      Other Metrics

      Citations

      View Options

      View options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media