People in conversation are highly sensitive to where others are looking. It has been argued that eye gaze is such a compelling social signal that following the gaze of another person is practically reflexive. Others have demonstrated that the role of gaze in interaction is more flexible. The direction of eye gaze can be informative about what a person is searching for, monitoring, orienting toward, referring to, deciding to choose, planning, or intending. But because it can signal all of these things (or none of them, as when attention is captured inadvertently by something sudden or salient), a look is often ambiguous. In this chapter, we consider the contributions of eye gaze to recognizing intentions and coordinating joint action in spatial contexts. To the extent that gazing at an object is instrumental to what the gazer is doing, a look can provide a window for an experimenter into the gazer’s cognitive processing, a cue for an interacting partner about the gazer’s intention, or a hint for a marketer about the gazer’s preference or indecision. Patterns of looks can be interpreted by people engaged in perspective taking, or by automated pattern recognizers engaged in mind-reading. Finally, eye gaze has the potential to be deployed communicatively, as when one person intends that another recognize what she is attending to and use this information to coordinate their behavior. We discuss evidence about how observers use gaze cues (whether interactively or non-interactively) in contexts where they can see the gazer’s face, as well as during remote electronic communication by way of a shared gaze system where each partner’s gaze is represented to the other by a dynamic cursor displayed over a shared screen.
(2004) Language mediated eye movements in the absence of a visual world: the blank screen paradigm. Cognition, 93, 79–87.
Altmann, & Kamide
(2003) The time-course of prediction in incremental sentence processing: evidence from anticipatory eye movements. Journal of Memory and Language, 49, 133–156.
Amati, F. & Brennan, S. E.
(2015) Intention recognition trumps reflexive gaze following. Poster presented at the 27th Annual Convention of the Association for Psychological Science, New York, NY.
Apperly, I. A., & Butterfill, S. A.
(2009) Do humans have two systems to track beliefs and belief-like states?Psychological Review, 116(4), 953.
Argyle, M., & Cook, M.
(1976) Gaze and mutual gaze. Oxford, England: Cambridge University Press.
Astington, J. W., & Gopnik, A.
(1991) Theoretical explanations of children’s understanding of the mind. British Journal of Developmental Psychology, 9, 7–31.
Atalay, A. S., Bodur, H. O., & Rasolofoarison, D.
(2012) Shining in the center: Central gaze cascade effect on product choice. Journal of Consumer Research, 39(4), 848–866.
Baillargeon R, Spelke E, Wasserman S.
(1985) Object permanence in 5-month-old infants. Cognition, 20, 191–208.
Baker, C. L., Saxe, R., & Tenenbaum, J. B.
(2009) Action understanding as inverse planning. Cognition, 113(3), 329–349.
Baldwin, D. A.
(1995) Understanding the link between joint attention and language. In Moore, C., & Dunham, P., (Eds.), Joint attention: Its origins and role in development. Lawrence Erlbaum Associates, Hillsdale, NJ.
Ballard, D. H., & Hayhoe, M. M.
(2009) Modelling the role of task in the control of gaze. Visual Cognition, 17(6–7), 1185–1204.
Baron-Cohen, S.
(1995) The eye direction detector (EDD) and the shared attention mechanism (SAM): Two cases for evolutionary psychology. In Moore, C., & Dunham, P., (Eds.), Joint attention: its role in development. Lawrence Erlbaum Associates, Hillsdale, NJ.
Batki, A., Baron-Cohen, S., Wheelwright, S., Connellan, J., & Ahluwalia, J.
(2000) Is there an innate gaze module? Evidence from human neonates. Infant Behavior and Development, 23(2), 223–229.
Bayliss, A. P., Schuch, S., & Tipper, S. P.
(2010) Gaze cueing elicited by emotional faces is influenced by affective context. Visual Cognition, 18(8), 1214–1232.
Bednarik, R., Vrzakova, H., & Hradis, M.
(2012, March). What do you want to do next: a novel approach for intent prediction in gaze-based interaction. In Proceedings of the Symposium on Eye Tracking Research and Applications (pp.83–90). ACM.
Bennewitz, M., Faber, F., Joho, D., & Behnke, S.
(2007) Fritz-A humanoid communication robot. In The 16th IEEE International Symposium on Robot and Human Interactive Communication, RO-MAN 2007 (pp.1072–1077).
Bee, N., Prendinger, H., André, E., & Ishizuka, M.
(2006) Automatic preference detection by analyzing the gaze ‘cascade effect’. In Electronic Proceedings of the 2nd conference on communication by gaze interaction, COGAIN (pp.61–64).
Böckler, A., & Sebanz, N.
(2013) Linking joint attention and joint action. In J. Metcalfe & H. Terrace (Eds.), Joint Attention and Metacognition. New York: Oxford University Press.
(1994) The effects of visibility on dialogue and performance in a cooperative problem solving task. Language and Speech, 37, 1–20.
Brennan, S. E.
(1990) Seeking and providing evidence for mutual understanding. Unpublished Ph.D. dissertation, Stanford University.
Brennan, S. E.
(2005) How conversation is shaped by visual and spoken evidence. Approaches to studying world-situated language use: Bridging the language-as-product and language-as-action traditions, 95–129.
Brennan, S. E., Chen, X., Dickinson, C., Neider, M., & Zelinsky, G.
(2008) Coordinating cognition: The costs and benefits of shared gaze during collaborative search. Cognition, 106, 1465–1477.
Brennan, S. E., & Hulteen, E.
(1995) Interaction and feedback in a spoken language system: A theoretical framework. Knowledge-Based Systems, 8, 143–151.
Brennan, S. E., & Williams, M.
(1995) The feeling of another’s knowing: Prosody and filled pauses as cues to listeners about the metacognitive states of speakers. Journal of Memory and Language, 34(3), 383–398
Carletta, J., Hill, R. L., Nicol, C., Taylor, T., De Ruiter, J. P., & Bard, E. G.
(2010) Eyetracking for two-person tasks with manipulation of a virtual world. Behavior Research Methods, 42(1), 254–265.
Castelhano, M. S., Mack, M. L., & Henderson, J. M.
(2009) Viewing task influences eye movement control during active scene perception. Journal of Vision, 9(3):6.1–15.
Chandon, P., Hutchinson, J. W., & Young, S. H.
(2001) Measuring the value of point-of-purchase marketing with commercial eye-tracking data. INSEAD.
Churilov, E.
(2014) Speaking, pointing, and gazing: Different communication strategies for different communicative constraints. Unpublished Master’s thesis, Stony Brook University.
Clark, H. H., & Brennan, S. E.
(1991) Grounding in communication. In L. B. Resnick, J. Levine, & S. D. Teasley (Eds.), Perspectives on socially shared cognition (pp.127–149). Washington, DC: APA. Reprinted in R. M. Baecker (Ed.), Groupware and computer-supported cooperative work: Assisting human-human collaboration (pp.222–233). San Mateo, CA: Morgan Kaufman Publishers, Inc.
Clark, H. H., & Krych, M. A.
(2004) Speaking while monitoring addressees for understanding. Journal of Memory and Language, 50(1), 62–81.
Cline, M. G.
(1967) The perception of where a person is looking. The American Journal of Psychology, 41–50.
Crosby, J. R.Monin, B., & Richardson, D.
(2008) Where do we look during potentially offensive behavior?Psychological Science, 19, 226–228).
DeAngelus, M., & Pelz, J. B.
(2009) Top-down control of eye movements: Yarbus revisited. Visual Cognition, 17(6–7), 790–811.
Dennett, D. C.
(1987) The intentional stance. Cambridge, MA: MIT press.
Dickie, C., Vertegaal, R., Sohn, C., & Cheng, D.
(2005) eyeLook: using attention to facilitate mobile media consumption. In Proceedings of the 18th annual ACM symposium on User interface software and technology (pp.103–106).
Doherty, M. J.
(2006) The development of mentalistic gaze understanding. Infant and Child Development, 15(2), 179–186.
Doshi, A., & Trivedi, M. M.
(2009) On the roles of eye gaze and head dynamics in predicting driver’s intent to change lanes. Intelligent Transportation Systems, IEEE Transactions on, 10(3), 453–462.
Drewes, H., De Luca, A., & Schmidt, A.
(2007) Eye-gaze interaction for mobile phones. In Proceedings of the 4th international conference on mobile technology, applications, and systems and the 1st international symposium on computer human interaction in mobile technology (pp.364–371). ACM.
Driver, J., Davis, G., Ricciardelli, P., Kidd, P., Maxwell, E., & Baron-Cohen, S.
(1965) Visual Perception from birth as shown by pattern selectivity. Annals of the New York Academy of Sciences, New Issues in Infant Development, 118, 793–814.
Ferreira, F., Apel, J., & Henderson, J. M.
(2008) Taking a new look at looking at nothing. Trends in Cognitive Sciences, 12(11), 405–410.
Foulsham, T., & Lock, M.
(2015) How the eyes tell lies: Social gaze during a preference task. Cognitive Science, 39, 1704–1726.
Friesen, C. K. & Kingstone, A.
(1998) The eyes have it! Reflexive orienting is triggered by nonpredictive gaze. Psychonomic Bulletin & Review, 5, 490–495.
Frischen, A., Bayliss, A. P., & Tipper, S. P.
(2007) Gaze cueing of attention: Visual attention, social cognition, and individual differences. Psychological Bulletin, 133(4), 694–724.
Frith, C. D., & Frith, U.
(2006) How we predict what other people are going to do. Brain Research, 1079(1), 36–46.
Fukusima, S. S., Loomis, J. M., & Da Silva, J. A.
(1997) Visual perception of egocentric distance as assessed by triangulation. Journal of Experimental Psychology: Human Perception and Performance, 23(1), 86.
Glaholt, M. G., & Reingold, E. M.
(2009) Stimulus exposure and gaze bias: A further test of the gaze cascade model. Attention, Perception, & Psychophysics, 71(3), 445–450.
Glaholt, M. G., & Reingold, E. M.
(2011) Eye movement monitoring as a process tracing methodology in decision making research. Journal of Neuroscience, Psychology, and Economics, 4(2), 125.
Glaholt, M. G., & Reingold, E. M.
(2012) Direct control of fixation times in scene viewing: Evidence from analysis of the distribution of first fixation duration. Visual Cognition, 20(6), 605–626.
Glaholt, M. G., Wu, M. C., & Reingold, E. M.
(2010) Evidence for top-down control of eye movements during visual decision making. Journal of Vision, 10(5), 15.
Goldberg, J. H., & Kotval, X. P.
(1999) Computer interface evaluation using eye movements: methods and constructs. International Journal of Industrial Ergonomics, 24(6), 631–645.
Goodwin, C.
(1981) Conversational organization: Interaction between speakers and hearers. New York, NY: Academic Press.
Gopnik, A., Meltzoff, A. N., & Bryant, P.
(1997) Words, thoughts, and theories (Vol. 1). Cambridge, MA: MIT Press
Greene, M. R., Liu, T., & Wolfe, J. M.
(2012) Reconsidering Yarbus: A failure to predict observers’ task from eye movement patterns. Vision Research, 62, 1–8.
(2014) An inverse Yarbus process: Predicting observers’ task from eye movement patterns. Vision Research, 103, 127–142.
Hanna, J. E., & Brennan, S. E.
(2007) Speakers’ eye gaze disambiguates referring expressions early during face-to-face conversation. Journal of Memory and Language, 57(4), 596–615.
Henderson, J. M.
(2003) Human gaze control during real-world scene perception. Trends in Cognitive Sciences, 7(11), 498–504.
Henderson, J. M., Shinkareva, S. V., Wang, J., Luke, S. G., & Olejarczyk, J.
(2013) Predicting cognitive state from eye movements. PloS One, 8(5), e64937.
Ho, S., Foulsham, T., & Kingstone, A.
(2015) Speaking and Listening with the Eyes: Gaze Signaling during Dyadic Interactions. PLoS One, 10(8), e0136905.
Hood, B. M., Willen, J. D., & Driver, J.
(1998) Adults’ eyes trigger shifts of visual attention in human infants. Psychological Science, 9(2), 131–134.
Houston-Price, C. & Nakai, S.
(2004) Distinguishing novelty and familiarity effects in infant preference procedures. Infant and Child Development, 13, 341–348.
Hunter, M. A. & Ames, E. W.
(1988) A multifactor model of infant preferences for novel and familiar stimuli. In Rovee-Collier C., Lipsitt L. P. (Eds.) Advances in Infancy Research, Vol. 5 (pp.69–95). Stamford, CT: Ablex.
Ishimaru, S., Kunze, K., Kise, K., Weppner, J., Dengel, A., Lukowicz, P., & Bulling, A.
(2014, March). In the blink of an eye: combining head motion and eye blink frequency for activity recognition with Google Glass. In Proceedings of the 5th augmented human international conference (p.15). ACM.
Jacob, R. J.
(1991) The use of eye movements in human-computer interaction techniques: What you look at is what you get. ACM Transactions on Information Systems (TOIS), 9(3), 152–169.
Jacob, R. J.
(1995) Eye tracking in advanced interface design. In W. Barfield & T. A. Furness (Eds.), Virtual environments and advanced interface design (pp.258–308). New York: OxfordUniversity Press.
Jermann, P., Nüssli, M. A., & Li, W.
(2010, September). Using dual eye-tracking to unveil coordination and expertise in collaborative Tetris. In Proceedings of the 24th BCS Interaction Specialist Group Conference (pp.36–44). British Computer Society.
Ji, Q., Zhu, Z. & Lan, P.
(2004) Real-Time Nonintrusive Monitoring and Prediction of Driver Fatigue. IEEE Transactions on Vehicular Technology, 53, 1052–1068.
Jusczyk, P., & Aslin, R.
(1995) Infants’ detection of the sound patterns of words in fluent speech. Cognitive Psychology 29, 1–23.
Kanan, C., Ray, N. A., Bseiso, D. N., Hsiao, J. H., & Cottrell, G. W.
(2014, March). Predicting an observer’s task using multi-fixation pattern analysis. In Proceedings of the Symposium on Eye Tracking Research and Applications (pp.287–290). ACM.
Krajbich, I., Armel, C., & Rangel, A.
(2010) Visual fixations and the computation and comparison of value in simple choice. Nature Neuroscience, 13(10), 1292–1298.
Krajbich, I., Lu, D., Camerer, C., & Rangel, A.
(2012) The attentional drift-diffusion model extends to simple purchasing decisions. Frontiers in Psychology, 3, 193.
Krajbich, I., & Rangel, A.
(2011) Multialternative drift-diffusion model predicts the relationship between visual fixations and choice in value-based decisions. Proceedings of the National Academy of Sciences, 108(33), 13852–13857.
Kuhn, G., Tatler, B. W., Findlay, J. M., & Cole, G. G.
(2008) Misdirection in magic: Implications for the relationship between eye gaze and attention. Visual Cognition, 16(2–3), 391–405.
Langton, S. R.
(2000) The mutual influence of gaze and head orientation in the analysis of social attention direction. The Quarterly Journal of Experimental Psychology: Section A, 53(3), 825–845.
Langton, S. R. H. & Bruce, V.
(2000) You must see the point: Automatic processing of cues to the direction of social attention. Journal of Experimental Psychology: Human Perception and Performance, 26, 747–757.
Liang, Y., Reyes, M. L., & Lee, J. D.
(2007) Real-time detection of driver cognitive distraction using support vector machines. Intelligent Transportation Systems, IEEE Transactions, 8(2), 340–350.
Liebling, D. J., & Preibusch, S.
(2014) Privacy considerations for a pervasive eye tracking world. In Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct Publication (pp.1169–1177). ACM.
Loomis, J. M., & Knapp, J. M.
(2003) Visual perception of egocentric distance in real and virtual environments. Virtual and Adaptive Environments, (11), 21–46.
Majaranta, P., & Räihä, K. J.
(2002, March). Twenty years of eye typing: systems and design issues. In Proceedings of the 2002 Symposium on Eye Tracking Research & Applications (pp.15–22). ACM.
Matlock, T., & Richardson, D. C.
(2004) Do eye movements go with fictive motion?Proceedings of the 26th Annual Conference of the Cognitive Science Society (pp.909–914). Mahwah, NJ: Lawrence Erlbaum.
Maughan, L., Gutnikov, S., & Stevens, R.
(2007) Like more, look more. look more, like more: the evidence from eye-tracking. Journal of Brand Management, 14(4), 335–342.
Mills, M., Hollingworth, A., Van der Stigchel, S., Hoffman, L., & Dodd, M. D.
(2011) Examining the influence of task set on eye movements and fixations. Journal of Vision, 11(8), 17.
Mundy, P., & Newell, L.
(2007) Attention, joint attention, and social cognition. Current Directions in Psychological Science, 16(5), 269–274.
Neider, M. B., Chen, X., Dickinson, C. A., Brennan, S. E., & Zelinsky, G. J.
(2008) Joint attention: Inferring what others perceive (and don’t perceive). Consciousness and Cognition, 17(1), 339–349.
Onishi, K. H., & Baillargeon, R.
(2005) Do 15-month-old infants understand false beliefs?Science, 308(5719), 255–258.
Patalano, A. L., Juhasz, B. J., & Dicke, J.
(2010) The relationship between indecisiveness and eye movement patterns in a decision making informational search task. Journal of Behavioral Decision Making, 23(4), 353–368.
Penn, D. C., & Povinelli, D. J.
(2007) On the lack of evidence that non-human animals possess anything remotely resembling a ‘theory of mind’. Philosophical Transactions of the Royal Society of London B: Biological Sciences, 362(1480), 731–744.
Pfeiffer, U. J., Vogeley, K., & Schilbach, L.
(2013) From gaze cueing to dual eye-tracking: novel approaches to investigate the neural correlates of gaze in social interaction. Neuroscience & Biobehavioral Reviews, 37, 2516–2528.
Pieters, R., & Warlop, L.
(1999) Visual attention during brand choice: The impact of time pressure and task motivation. International Journal of Research in Marketing, 16(1), 1–16.
Posner M. I.
(1980) Orienting of attention. Quarterly Journal of Experimental Psychology, 32, 3–25.
Posner, M. I. & Cohen, Y. A.
(1984) Components of visual orienting. In: Bouma, H.; Bouwhuis, D. G. (Eds.) Attention and performance XVII: Control of visual processing. Hillsdale, NJ: Erlbaum; 1984. pp.531–556.
Reutskaja, E., Nagel, R., Camerer, C. F., & Rangel, A.
(2011) Search dynamics in consumer choice under time pressure: An eye-tracking study. The American Economic Review, 900–926.
Ricciardelli, P., Carcagno, S., Vallar, G., & Bricolo, E.
(2013) Is gaze following purely reflexive or goal-directed instead? Revisiting the automaticity of orienting attention by gaze cues. Experimental Brain Research, 224(1), 93–106.
Richardson, D. C., & Dale, R.
(2005) Looking to understand: The coupling between speakers’ and listeners’ eye movements and its relationship to discourse comprehension. Cognitive Science, 29(6), 1045–1060.
Romani, M., Cesari, P., Urgesi, C., Facchini, S., & Aglioti, S. M.
(2003) Motor facilitation of the human cortico-spinal system during observation of bio-mechanically impossible movements. Neuroimage, 26(3), 755–763.
Russo, J. E., & Leclerc, F.
(1994) An eye-fixation analysis of choice processes for consumer nondurables. Journal of Consumer Research, 274–290.
Sacks, H., Schegloff, E. A., & Jefferson, G.
(1974) A simplest systematics for the organization of turn-taking in conversation. Language, 50, 696–735.
Sebanz, N., Bekkering, H., & Knoblich, G.
(2006) Joint action: bodies and minds moving together. Trends in Cognitive Sciences, 10(2), 70–76.
(2008) Gaze following in human infants depends on communicative signals. Current Biology, 18(9), 668–671.
Shiffrar, M., & Freyd, J. J.
(1990) Apparent motion of the human body. Psychological Science, 1(4), 257–264.
Shimojo, S., Simion, C., Shimojo, E., & Scheier, C.
(2003) Gaze bias both reflects and influences preference. Nature Neuroscience, 6(12), 1317–1322.
Simion, C., & Shimojo, S.
(2006) Early interactions between orienting, visual sampling and decision making in facial preference. Vision Research, 46(20), 3331–3335.
Smith, J. D., & Graham, T. C.
(2006, June). Use of eye movements for video game control. In Proceedings of the 2006 ACM SIGCHI International Conference on Advances in Computer Entertainment Technology (p.20). ACM.
Spelke, E. S.
(1990) Principles of object perception. Cognitive Science, 14(1), 29–56.
Spivey, M. J., & Geng, J. J.
(2001) Oculomotor mechanisms activated by imagery and memory: Eye movements to absent objects. Psychological Research, 65(4), 235–241.
Stein, R. & Brennan, S. E.
(2004) Another person’s eye gaze as a cue in solving programming problems. Proceedings, ICMI 2004, Sixth International Conference on Multimodal Interfaces (pp.9–15), Penn State University, State College, PA.
Stiefelhagen, R., & Zhu, J.
(2002, April). Head orientation and gaze direction in meetings. In CHI’02 Extended Abstracts on Human Factors in Computing Systems (pp.858–859). ACM.
Suri, R., & Monroe, K. B.
(2003) The effects of time constraints on consumers’ judgments of prices and products. Journal of Consumer Research, 30(1), 92–104
Tanenhaus, M. K., Spivey-Knowlton, M. J., Eberhard, K. M., & Sedivy, J. C.
(1995) Integration of visual and linguistic information in spoken language comprehension. Science, 268(5217), 1632–1634.
Teufel, C., Alexis, D. M., Clayton, N. S., & Davis, G.
(1995) Joint attention as social cognition. In C. Moore and P. J. Dunham (Eds.), Joint attention: Its origins and role in development, 103–130. Laurence Erlbaum Associates, Hillsdale, NJ.
Tomeo, E., Cesari, P., Aglioti, S. M., & Urgesi, C.
(2012) Fooling the kickers but not the goalkeepers: behavioral and neurophysiological correlates of fake action detection in soccer. Cerebral Cortex, 279, 1–14.
Trösterer, S., Gärtner, M., Wuchse, M., Maurer, B., Baumgartner, A., Meschtscherjakov, A., Tscheligi, M.
(2015) Four eyes see more than two: Shared gaze in the car. Proceedings, Interact 2015.
(1997, January). Towards gaze-mediated interaction: Collecting solutions of the “Midas touch problem”. In Human-Computer Interaction INTERACT’97 (pp.509–516). Springer US.
Wellman, H. M., & Woolley, J. D.
(1990) From simple desires to ordinary beliefs: The early development of everyday psychology. Cognition, 35, 245–275.
Wolpert, D. M., Doya, K., Kawato, M.
(2003) A unifying computational framework for motor control and social interaction. Philosophical Transactions of the Royal Society of London: Brain and Biological Sciences, 358 (1431), 593–602.
Wood, N.
(2014) Autocorrect awareness: categorizing changes and measuring authorial perceptions (unpublished honors thesis). Florida State University, 351, Tallahassee, FL.
Yarbus, A. L.
(1967) Eye movements during perception of complex objects. In Eye movements and vision (pp.171–211). Springer US.
Zelinsky, G. J., Dickinson, C. A., Chen, X., Neider, M. B., & Brennan, S. E.
(2005) Collaborative search using shared eye gaze. Abstract, Journal of Vision, 5. p.700.
Zelinsky, G. J., Peng, Y., & Samaras, D.
(2013) Eye can read your mind: Decoding gaze fixations to reveal categorical search targets. Journal of Vision, 13(14), 10.
Cited by (4)
Cited by 4 other publications
Härkki, Tellervo
2023. Mobile gaze tracking and an extended linkography for collaborative sketching and designing. International Journal of Technology and Design Education 33:2 ► pp. 379 ff.
Kiose, Maria I., Anna V. Leonteva, Olga V. Agafonova & Andrey A. Petrov
2023. Multimodal Communicative Moves in Expositive Dialogue: Common and Novel Topic Elaboration. RUDN Journal of Language Studies, Semiotics and Semantics 14:4 ► pp. 1013 ff.
2021. Early Responses: An Introduction. Discourse Processes 58:4 ► pp. 293 ff.
Wagemann, Johannes & Ulrich Weger
2021. Perceiving the Other Self: An Experimental First-Person Account of Nonverbal Social Interaction. The American Journal of Psychology 134:4 ► pp. 441 ff.
This list is based on CrossRef data as of 26 june 2024. Please note that it may not be complete. Sources presented here have been supplied by the respective publishers.
Any errors therein should be reported to them.