[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Study of Polynomial Mapping Functions in Video-Oculography Eye Trackers

Published: 01 July 2012 Publication History

Abstract

Gaze-tracking data have been used successfully in the design of new input devices and as an observational technique in usability studies. Polynomial-based Video-Oculography (VOG) systems are one of the most attractive gaze estimation methods thanks to their simplicity and ease of implementation. Although the functionality of these systems is generally acceptable, there has been no thorough comparative study to date of how the mapping equations affect the final system response. After developing a taxonomic classification of calibration functions, we examined over 400,000 models and evaluated the validity of several conventional assumptions. Our rigorous experimental procedure enabled us to optimize the calibration process for a real VOG gaze-tracking system and halve the calibration time while avoiding a detrimental effect on the accuracy or tolerance to head movement. Finally, a geometry-based method is implemented and tested. The results and performance is compared with those obtained by the general purpose expressions.

References

[1]
Anliker, J. 1976. Eye movements---On-line measurement, analysis, and control. In Eye Movements and Psychological Processes. 185--199.
[2]
Babcock, J. S. and Pelz, J. B. 2004. Building a lightweight eyetracking headgear. In Proceedings of the Eye Tracking Research and Applications Symposium (ETRA’04). ACM, New York, NY, 109--114.
[3]
Bailey, B. P. and Iqbal, S. T. 2008. Understanding changes in mental workload during execution of goal-directed tasks and its application for interruption management. ACM Trans. Comput.-Hum. Interact. 14, 21:1--21:28.
[4]
Barea, R., Boquete, L., Bergasa, L. M., Lopez, E., and Mazo, M. 2003. Electro-oculographic guidance of a wheelchair using eye movements codification. I. J. Robotic Res. 22, 7--8, 641--652.
[5]
Benel, D. C. R., Ottens, D., and Horst, R. 1991. Use of an eye tracking system in the usability laboratory. In Proceedings of the Human Factors Society 35th Annual Meeting.
[6]
Beymer, D. and Flickner, M. 2003. Eye gaze tracking using an active stereo head. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Vol. 2, 451--458.
[7]
Bolt, R. A. 1982. Eyes at the interface. In Proceedings of the Conference on Human Factors in Computing Systems (CHI’82). ACM, New York, NY, 360--362.
[8]
Brolly, X. L. C. and Mulligan, J. B. 2004. Implicit calibration of a remote gaze tracker. In Proceedings of the Conference on Computer Vision and Pattern Recognition Workshop. 134.
[9]
Burke, M., Hornof, A., Nilsen, E., and Gorman, N. 2005. High-cost banner blindness: Ads increase perceived workload, hinder visual search, and are forgotten. ACM Trans. Comput.-Hum. Interact. 12, 4, 423--445.
[10]
Cerrolaza, J. J., Villanueva, A., and Cabeza, R. 2008. Taxonomic study of polynomial regressions applied to the calibration of video-oculographic systems. In Proceedings of the Eye Tracking Research and Applications Symposium. 259--266.
[11]
Chen, J., Tong, Y., Gray, W., and Ji, Q. 2008. A robust 3D eye gaze tracking systemusing noise reduction. In Proceedings of the Eye Tracking Research and Applications Symposium.
[12]
Cherif, Z. R., Nat-Ali, A., Motsch, J. F., and Krebs, M. O. 2002. An adaptive calibration of an infrared ligth device used for gaze tracking. In Proceedings of the IEEE Instrumentation and Measurement Technology Conference. 1029--1033.
[13]
Draper, N. R. and Smith, H. 1981. Applied Regression Analysis. John Wiley and Sons.
[14]
Duchowski, A. T. 2007. Eye Tracking Methodology: Theory and Practice. Springer.
[15]
Ebisawa, Y. 1995. Unconstrained pupil detection technique using two light sources and the image difference method. In Visualization and Intelligent Design in Engineering and Architecture. 79--89.
[16]
Ellis, S., Candrea, R., Misner, J., Craig, C. S., and Lankford, C. P. 1998. Windows to the soul? What eye movements tell us about software usability. In Proceedings of the 7th Annual Conference of the Usability Professionals’ Association. 151--178.
[17]
Fitts, P. M., Jones, R. E., and Milton, J. L. 1950. Eye Movements of Aircraft Pilots during Instrument-Landing Approaches. Aeronautical Engin. Rev. 9, 2, 24--29.
[18]
Guestrin, E. and Eizenman, M. 2006. General theory of remote gaze estimation using pupil center and corneal reflections. IEEE Trans. Biomed. Engin. 53, 6, 1124--1133.
[19]
Guestrin, E. D. and Eizenman, M. 2008. Remote point-of-gaze estimation requiring a single-point calibration for applications with infants. In Proceedings of the Eye Tracking Research and Applications Symposium.
[20]
Hansen, D. W. and Ji, Q. 2010. In the eye of the beholder: A survey of models for eyes and gaze. IEEE Trans. Pattern Anal. Mach. Intell. 32, 3, 478--500.
[21]
Hansen, D. W., Agustin, J. S., and Villanueva, A. 2010. Homography normalization for robust gaze estimation in uncalibrated setups. In Proceedings of the Eye Tracking Research and Applications Symposium. 13--20.
[22]
Hartridge, H. and Thompson, L. 1948. Methods of investigating eye movements. British J. Ophthalmology.
[23]
Hennessey, C. and Lawrence, P. D. 2008. 3d point-of-gaze estimation on a volumetric display. In Proceedings of the Eye Tracking Research and Applications Symposium.
[24]
Hennessey, C., Noureddin, B., and Lawrence, P. 2006. A single camera eye-gaze tracking system with free head motion. In Proceedings of the Eye Tracking Research and Applications Symposium. ACM Press, New York, NY, 87--94.
[25]
Hornof, A. J. and Halverson, T. 2003. Cognitive strategies and eye movements for searching hierarchical computer displays. In Proceedings of the Conference on Human Factors in Computing Systems (CHI’03). ACM, New York, NY, 249--256.
[26]
Kang, J. J., Eizenman, M., Guestrin, E. D., and Eizenman, E. 2008. Investigation of the cross-ratios method for point-of-gaze estimation. IEEE Trans. Biomed. Engin. 55, 9, 2293--2302.
[27]
Karpala, F. and Jernigan, M. E. 1980. Compensation for distortion in eye-movement monitors. IEEE Trans. Biomed. Engin. 27, 3, 113--119.
[28]
Li, D., Winfield, D., and Parkhurst, D. J. 2005. Starburst: A hybrid algorithmfor video-based eye tracking combining feature-based and model-based approaches. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[29]
Massopust, P. 2010. Interpolation and Approximation with Splines and Fractals. Oxford University Press, Inc., New York, NY.
[30]
Merchant, J., Morrissette, R., and Porterfield, J. 1974. Remote measurement of eye direction allowing subject motion over one cubic foot of space. IEEE Trans. Biomed. Engin. 21, 4, 309--317.
[31]
Model, D. and Eizenman, M. 2010. An automatic personal calibration procedure for advanced gaze estimation systems. IEEE Trans. Biomed. Engin. 57, 5, 1031--1039.
[32]
Monty, R., and Senders, J. 1976. Eye Movements and Psychological Processes. Lawrence Erlbaum Associates.
[33]
Morimoto, C., Koons, D., Amir, A., and Flickner, M. 2000. Pupil detection and tracking using multiple light sources. Image Visi. Comput. 18, 4, 331--335.
[34]
Morimoto, C. H. and Mimica, M. R. M. 2005. Eye gaze tracking techniques for interactive applications. Comput. Vis. Image Underst. 98, 1, 4--24.
[35]
Navallas, J., Ariz, M., Villanueva, A., San Agustn, J., and Cabeza, R. 2011. Optimizing interoperability between video-oculographic and electromyographic systems. J Rehabil. Resear. Dev 48, 3, 253--266.
[36]
Ohno, T. and Mukawa, N. 2004. A free head, simple calibration, gaze tracking system that enables gaze-based interaction. In Proceedings of the Eye Tracking Research & Applications Symposium. 115--122.
[37]
Phillips, G. 2003. Interpolation and Approximation by Polynomials. CMS Books in Mathematics, Springer.
[38]
Piratla, N. M. and Jayasumana, A. P. 2002. A neural network based real-time gaze tracker. J. Netw. Comput. Appl. 25, 179--196.
[39]
Plumlee, M. D. and Ware, C. 2006. Zooming versus multiple window interfaces: Cognitive costs of visual comparisons. ACM Trans. Comput.-Hum. Interact. 13, 179--209.
[40]
Poole, A. and Ball, L. J. 2005. Eye tracking in human-computer interaction and usability research: Current status and future prospects. In Encyclopedia of Human Computer Interaction, C. Ghaoui Ed., IGI Global.
[41]
Ramanauskas, N. 2006. Calibration of video-oculographical eye-tracking system. Electron. Electric. Engin. 8, 65--68.
[42]
Shih, S. W. and Liu, J. 2004. A novel approach to 3-D gaze tracking using stereo cameras. IEEE Trans. Syst. Man Cyber. Part-B 34, 1, 234--245.
[43]
Stampe, D. M. 1993. Heuristic filtering and reliable calibration methods for video-based pupil-tracking systems. Behav. Resear. Meth. 25, 2, 137--142.
[44]
Starker, I. and Bolt, R. A. 1990. A gaze-responsive self-disclosing display. In Proceedings of the Conference on Human Factors in Computing Systems. ACM, New York, NY, 3--10.
[45]
Torricelli, D., Conforto, S., Schmid, M., and D’Alessio, T. 2008. A neural-based remote eye gaze tracker under natural head motion. Comput. Methods Prog. Biomed. 92, 66--78.
[46]
Underwood, G. 2005. Cognitive Process in Eye Guidance. Oxford University Press.
[47]
Vertegaal, R. 1999. The gaze groupware system: Mediating joint attention in multiparty communication and collaboration. In Proceedings of the Conference on Human Factors in Computing Systems (CHI’99). ACM, New York, NY, 294--301.
[48]
Villanueva, A. 2005. Mathematical models for video-oculography. Ph.D. thesis, Public University of Navarra.
[49]
Villanueva, A. and Cabeza, R. 2007. Models for gaze tracking systems. EURASIP J. Image Video Process.
[50]
Villanueva, A. and Cabeza, R. 2008. A novel gaze estimation system with one calibration point. IEEE Trans. Syst. Man Cyber. SMC-B 37, 4, 1123--1138.
[51]
White, K. P., Hutchinson, T. E., and Carley, J. M. 1993. Spatially dynamic calibration of an eyetracking system. IEEE Trans. Syst. Man, Cyber. 23, 4, 1162--1168.
[52]
Yoo, D. H. and Chung, M. J. 2005. A novel non-intrusive eye gaze estimation using cross-ratio under large head motion. Comput. Vis. Image Underst. 98, 1, 25--51.
[53]
Young, L. and Sheena, D. 1975. Survey of eye movement recording methods. Methods Designs, Behav. Resear. Methods Instrumen. 7, 5, 397--427.
[54]
Zhu, Z., Ji, Q., and Member, S. 2007. Novel eye gaze tracking techniques under natural head movement. IEEE Trans. Biomed. Engin. 54, 2246--2260.

Cited By

View all
  • (2024)Self-Calibrating Gaze Estimation via Matching Spatio-Temporal Reading Patterns and Eye-Feature PatternsIEEE Transactions on Industrial Informatics10.1109/TII.2024.341329220:10(11697-11707)Online publication date: Oct-2024
  • (2024)Eyeball Kinematics Informed Slippage Robust Gaze TrackingIEEE Sensors Journal10.1109/JSEN.2024.347500924:22(37620-37629)Online publication date: 15-Nov-2024
  • (2024)EEG, EOG, Likert Scale, and Interview Approaches for Assessing Stressful Hazard Perception ScenariosInternational Journal of Human–Computer Interaction10.1080/10447318.2024.2358461(1-26)Online publication date: 3-Jun-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Transactions on Computer-Human Interaction
ACM Transactions on Computer-Human Interaction  Volume 19, Issue 2
July 2012
226 pages
ISSN:1073-0516
EISSN:1557-7325
DOI:10.1145/2240156
Issue’s Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 July 2012
Accepted: 01 November 2011
Revised: 01 November 2011
Received: 01 July 2011
Published in TOCHI Volume 19, Issue 2

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Eye-tracking
  2. calibration
  3. gaze-tracking
  4. human computer interaction
  5. multiple linear regression
  6. video-oculography

Qualifiers

  • Research-article
  • Research
  • Refereed

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)34
  • Downloads (Last 6 weeks)2
Reflects downloads up to 21 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Self-Calibrating Gaze Estimation via Matching Spatio-Temporal Reading Patterns and Eye-Feature PatternsIEEE Transactions on Industrial Informatics10.1109/TII.2024.341329220:10(11697-11707)Online publication date: Oct-2024
  • (2024)Eyeball Kinematics Informed Slippage Robust Gaze TrackingIEEE Sensors Journal10.1109/JSEN.2024.347500924:22(37620-37629)Online publication date: 15-Nov-2024
  • (2024)EEG, EOG, Likert Scale, and Interview Approaches for Assessing Stressful Hazard Perception ScenariosInternational Journal of Human–Computer Interaction10.1080/10447318.2024.2358461(1-26)Online publication date: 3-Jun-2024
  • (2024)An analysis of perfume packaging designs on consumer’s cognitive and emotional behavior using eye gaze trackingMultimedia Tools and Applications10.1007/s11042-024-18715-w83:35(82563-82588)Online publication date: 14-Mar-2024
  • (2022)RETRACTED ARTICLE: Eye tracking: empirical foundations for a minimal reporting guidelineBehavior Research Methods10.3758/s13428-021-01762-855:1(364-416)Online publication date: 6-Apr-2022
  • (2022)High-Accuracy 3D Gaze Estimation with Efficient Recalibration for Head-Mounted Gaze Tracking SystemsSensors10.3390/s2212435722:12(4357)Online publication date: 8-Jun-2022
  • (2022)Gaze Tracking Based on Concatenating Spatial-Temporal FeaturesSensors10.3390/s2202054522:2(545)Online publication date: 11-Jan-2022
  • (2022)A Novel Integrated Eye-Tracking System with Stereo Stimuli for 3D Gaze EstimationIEEE Transactions on Instrumentation and Measurement10.1109/TIM.2022.3225009(1-1)Online publication date: 2022
  • (2022)Pupil-Contour-Based Gaze Estimation With Real Pupil Axes for Head-Mounted Eye TrackingIEEE Transactions on Industrial Informatics10.1109/TII.2021.311802218:6(3640-3650)Online publication date: Jun-2022
  • (2022)Real-time camera-based eye gaze tracking using convolutional neural network: a case study on social media websiteVirtual Reality10.1007/s10055-022-00642-626:4(1489-1506)Online publication date: 1-Dec-2022
  • Show More Cited By

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media