skip to main content
research-article

Applying computational tools to predict gaze direction in interactive visual environments

Authors Info & Claims
Published:19 May 2008Publication History
Skip Abstract Section

Abstract

Future interactive virtual environments will be “attention-aware,” capable of predicting, reacting to, and ultimately influencing the visual attention of their human operators. Before such environments can be realized, it is necessary to operationalize our understanding of the relevant aspects of visual perception, in the form of fully automated computational heuristics that can efficiently identify locations that would attract human gaze in complex dynamic environments. One promising approach to designing such heuristics draws on ideas from computational neuroscience. We compared several neurobiologically inspired heuristics with eye-movement recordings from five observers playing video games, and found that human gaze was better predicted by heuristics that detect outliers from the global distribution of visual features than by purely local heuristics. Heuristics sensitive to dynamic events performed best overall. Further, heuristic prediction power differed more between games than between different human observers. While other factors clearly also influence eye position, our findings suggest that simple neurally inspired algorithmic methods can account for a significant portion of human gaze behavior in a naturalistic, interactive setting. These algorithms may be useful in the implementation of interactive virtual environments, both to predict the cognitive state of human operators, as well as to effectively endow virtual agents in the system with humanlike visual behavior.

References

  1. Bailenson, J. and Yee, N. 2005. Digital chameleons: Automatic assimilation of nonverbal gestures in immersive virtual environments. Psychological Science 16, 814--819.Google ScholarGoogle ScholarCross RefCross Ref
  2. Ballard, D., Hayhoe, M., and Pelz, J. 1995. Memory representations in natural tasks. Journal of Cognitive Neuroscience 7, 1 (Winter), 66--80. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Burt, P. and Adelson, E. 1983. The laplacian pyramid as a compact image code. IEEE Transactions on Communications 31, 4, 532--540.Google ScholarGoogle ScholarCross RefCross Ref
  4. Carmi, R. and Itti, L. 2004. Bottom-up and top-down influences on attentional allocation in natural dyamic scenes. In Proc. Vision Science Society Annual Meeting (VSS04). 20.Google ScholarGoogle Scholar
  5. Finney, S. A. 2001. Real-time data collection in linux: A case study. Behavior Research Methods, Instruments, and Computers 33, 167--173.Google ScholarGoogle ScholarCross RefCross Ref
  6. Greenspan, H., Belongie, S., Goodman, R., Perona, P., Rakshit, S., and Anderson, C. 1994. Overcomplete steerable pyramid filters and rotation invariance. In Proceedings Computer Vision and Pattern Recognition. IEEE Computer Society, Los Alamitos, CA. 222--228.Google ScholarGoogle Scholar
  7. Hayhoe, M. 2004. Advances in relating eye movements and cognition. Infancy 6, 2, 267--274.Google ScholarGoogle ScholarCross RefCross Ref
  8. Hayhoe, M., Ballard, D., Triesch, J., and Shinoda, H. 2002. Vision in natural and virtual environments. In Proceedings of the Symposium on Eye Tracking Research & Applications (ETRA). 7--13. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Hayhoe, M., Shrivastava, A., Mruczek, R., and Pelz, J. 2003. Visual memory and motor planning in a natural task. Journal of Vision 3, 1, 49--63.Google ScholarGoogle ScholarCross RefCross Ref
  10. Henderson, J. M. and Hollingworth, A. 1999. High-level scene perception. Annual Review of Psychology 50, 243--271.Google ScholarGoogle ScholarCross RefCross Ref
  11. Itti, L. and Baldi, P. 2005. A principled approach to detecting surprising events in video. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). San Siego, CA. 631--637. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Itti, L. and Koch, C. 2001. Feature combination strategies for saliency-based visual attention systems. Journal of Electronic Imaging 10, 1 (Jan.), 161--169.Google ScholarGoogle ScholarCross RefCross Ref
  13. Itti, L., Koch, C., and Niebur, E. 1998. A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence 20, 11 (Nov.), 1254--1259. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Itti, L., Dhavale, N., and Pighin, F. 2003. Realistic avatar eye and head animation using a neurobiological model of visual attention. In Proceedings of SPIE 48th Annual International Symposium on Optical Science and Technology. 64--78.Google ScholarGoogle Scholar
  15. Koch, C. and Ullman, S. 1985. Shifts in selective visual-attention—towards the underlying neural circuitry. Human Neurobiology 4, 4, 219--227.Google ScholarGoogle Scholar
  16. Land, M. and Hayhoe, M. 2001. In what ways do eye movements contribute to everyday activities? Vision Research 41, 25--26, 3559--3565.Google ScholarGoogle Scholar
  17. Land, M., Mennie, N., and Rusted, J. 1999. The roles of vision and eye movements in the control of activities of daily living. Perception 28, 11, 1311--1328.Google ScholarGoogle ScholarCross RefCross Ref
  18. Mannan, S., Ruddock, K., and Wooding, D. 1997. Fixation patterns made during brief examination of two-dimensional images. Perception 26, 8, 1059--1072.Google ScholarGoogle ScholarCross RefCross Ref
  19. May, J., Dean, M., and Barnard, P. 2003. Using film cutting techniques in interface design. Human-Computer Interaction 18, 4, 325--372. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Najemnik, J. and Geisler, W. 2005. Optimal eye-movement strategies in visual search. Nature (London) 434, 7031 (Mar.), 387--391.Google ScholarGoogle ScholarCross RefCross Ref
  21. Navalpakkam, V. and Itti, L. 2005. Modeling the influence of task on attention. Vision Research 45, 2 (Jan.), 205--231.Google ScholarGoogle ScholarCross RefCross Ref
  22. Navalpakkam, V. and Itti, L. 2006. Optimal cue selection strategy. In Advances in Neural Information Processing Systems, Vol. 19 (NIPS*2005). MIT Press, Cambridge, MA. 1--8.Google ScholarGoogle Scholar
  23. Parker, R. 1978. Picture processing during recognition. Journal of Experimental Psychology—Human Perception and Performance 4, 2, 284--293.Google ScholarGoogle Scholar
  24. Parkhurst, D. and Niebur, E. 2004. Texture contrast attracts overt visual attention in natural scenes. European Journal of Neuroscience 19, 3 (Feb.), 783--789.Google ScholarGoogle ScholarCross RefCross Ref
  25. Parkhurst, D., Law, K., and Niebur, E. 2002. Modeling the role of salience in the allocation of overt visual attention. Vision Research 42, 1, 107--123.Google ScholarGoogle ScholarCross RefCross Ref
  26. Peli, V., Goldstein, T., and Woods, R. 2005. Scanpaths of motion sequences: Where people look when watching movies. In Proceedings of the Fourth Starkfest Conference on Vision and Movement in Man and Machines. School of Optometry, UC Berkeley, Berkeley, CA. 18--21.Google ScholarGoogle Scholar
  27. Peters, C. and O'Sullivan, C. 2003. Bottom-up visual attention for virtual human animation. In Computer Animation and Social Agents 2003. 111--117. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Peters, R., Iyer, A., Itti, L., and Koch, C. 2005. Components of bottom-up gaze allocation in natural images. Vision Research 45, 18, 2397--2416.Google ScholarGoogle ScholarCross RefCross Ref
  29. Pomplun, M. 2005. Saccadic selectivity in complex visual search displays. Vision Research 46, 12 (June), 1886--1900.Google ScholarGoogle Scholar
  30. Privitera, C. and Stark, L. 1998. Evaluating image processing algorithms that predict regions of interest. Pattern Recognition Letters 19, 11, 1037--1043. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Privitera, C. and Stark, L. 2000. Algorithms for defining visual regions-of-interest: Comparison with eye fixations. IEEE Transactions on Pattern Analysis and Machine Intelligence 22, 9, 970--982. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Rao, R., Zelinsky, G., Hayhoe, M., and Ballard, D. 2002. Eye movements in iconic visual search. Vision Research 42, 11 (May), 1447--1463.Google ScholarGoogle ScholarCross RefCross Ref
  33. Rayner, K. 1998. Eye movements in reading and information processing: 20 years of research. Psychological Bulletin 124, 3 (Nov.), 372--422.Google ScholarGoogle ScholarCross RefCross Ref
  34. Reinagel, P. and Zador, A. 1999. Natural scene statistics at the centre of gaze. Network-Computation in Neural Systems 10, 4 (Nov.), 341--350.Google ScholarGoogle ScholarCross RefCross Ref
  35. Rensink, R. A. 2000. The dynamic representation of scenes. Visual Cognition 7, 17--42.Google ScholarGoogle ScholarCross RefCross Ref
  36. Sodhi, M., Reimer, B., Cohen, J., Vastenburg, E., Kaars, R., and Kirschenbaum, S. 2002. On-road driver eye-movement tracking using head-mounted devices. In Proceedings of the Symposium on Eye Tracking Research & Applications (ETRA). 61--68. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Stampe, D. M. 1993. Heuristic filtering and reliable calibration methods for video based pupil tracking systems. Behavior Research Methods, Instruments, and Computers 25, 2, 137--142.Google ScholarGoogle ScholarCross RefCross Ref
  38. Terzopoulos, D. and Rabie, T. F. 1997. Animat vision: Active vision in artificial animals. Videre: Journal of Computer Vision Research 1, 1, 2--19.Google ScholarGoogle Scholar
  39. Toet, A. 2006. Gaze directed displays as an enabling technology for attention aware systems. Computers in Human Behavior 22, 4 (July), 615--647.Google ScholarGoogle ScholarCross RefCross Ref
  40. Torralba, A. 2003. Modeling global scene factors in attention. Journal of the Optical Society of America A-Optics Image Science and Vision 20, 7 (July), 1407--1418.Google ScholarGoogle ScholarCross RefCross Ref
  41. Tosi, V., Mecacci, L., and Pasquali, E. 1997. Scanning eye movements made when viewing film: Preliminary observations. International Journal of Neuroscience 92, 1--2, 47--52.Google ScholarGoogle ScholarCross RefCross Ref
  42. Treisman, A. and Gelade, G. 1980. A feature-integration theory of attention. Cognitive Psychology 12, 97--136.Google ScholarGoogle ScholarCross RefCross Ref
  43. Voge, S. 1999. Looking at paintings: Patterns of eye movements in artistically naive and sophisticated subjects. Leonardo 32, 4, 325--325.Google ScholarGoogle ScholarCross RefCross Ref
  44. Wolfe, J. and Horowitz, T. 2004. What attributes guide the deployment of visual attention and how do they do it? Nature Reviews Neuroscience 5, 6 (June), 495--501.Google ScholarGoogle Scholar
  45. Yarbus, A. 1967. Eye movements during perception of complex objects. In Eye Movements and Vision, L. Riggs, Ed. Plenum Press, New York.Google ScholarGoogle Scholar
  46. Zetzsche, C., Schill, K., Deubel, H., Krieger, G., Umkehrer, E., and Beinlich, S. 1998. Investigation of a sensorimotor system for saccadic scene analysis: an integrated approach. In From animals to animats, Proceedings of the Fifth International Conference on the Simulation of Adaptive Behavior. Vol. 5. 120--126. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Applying computational tools to predict gaze direction in interactive visual environments

            Recommendations

            Comments

            Login options

            Check if you have access through your login credentials or your institution to get full access on this article.

            Sign in

            Full Access

            • Published in

              cover image ACM Transactions on Applied Perception
              ACM Transactions on Applied Perception  Volume 5, Issue 2
              May 2008
              120 pages
              ISSN:1544-3558
              EISSN:1544-3965
              DOI:10.1145/1279920
              Issue’s Table of Contents

              Copyright © 2008 ACM

              Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

              Publisher

              Association for Computing Machinery

              New York, NY, United States

              Publication History

              • Published: 19 May 2008
              • Accepted: 1 May 2007
              • Revised: 1 April 2007
              • Received: 1 April 2006
              Published in tap Volume 5, Issue 2

              Permissions

              Request permissions about this article.

              Request Permissions

              Check for updates

              Qualifiers

              • research-article
              • Research
              • Refereed

            PDF Format

            View or Download as a PDF file.

            PDF

            eReader

            View online with eReader.

            eReader