Skip to main content
Top
Gepubliceerd in: Psychological Research 6/2022

Open Access 02-12-2021 | Original Article

When eyes beat lips: speaker gaze affects audiovisual integration in the McGurk illusion

Auteurs: Basil Wahn, Laura Schmitz, Alan Kingstone, Anne Böckler-Raettig

Gepubliceerd in: Psychological Research | Uitgave 6/2022

Abstract

Eye contact is a dynamic social signal that captures attention and plays a critical role in human communication. In particular, direct gaze often accompanies communicative acts in an ostensive function: a speaker directs her gaze towards the addressee to highlight the fact that this message is being intentionally communicated to her. The addressee, in turn, integrates the speaker’s auditory and visual speech signals (i.e., her vocal sounds and lip movements) into a unitary percept. It is an open question whether the speaker’s gaze affects how the addressee integrates the speaker’s multisensory speech signals. We investigated this question using the classic McGurk illusion, an illusory percept created by presenting mismatching auditory (vocal sounds) and visual information (speaker’s lip movements). Specifically, we manipulated whether the speaker (a) moved his eyelids up/down (i.e., open/closed his eyes) prior to speaking or did not show any eye motion, and (b) spoke with open or closed eyes. When the speaker’s eyes moved (i.e., opened or closed) before an utterance, and when the speaker spoke with closed eyes, the McGurk illusion was weakened (i.e., addressees reported significantly fewer illusory percepts). In line with previous research, this suggests that motion (opening or closing), as well as the closed state of the speaker’s eyes, captured addressees’ attention, thereby reducing the influence of the speaker’s lip movements on the addressees’ audiovisual integration process. Our findings reaffirm the power of speaker gaze to guide attention, showing that its dynamics can modulate low-level processes such as the integration of multisensory speech signals.
Opmerkingen
Basil Wahn and Laura Schmitz shared first authorship.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Introduction

The role of gaze in social interaction

The eyes of another person are of fundamental importance to human behavior from early infancy on (Farroni et al., 2004; Striano & Reid, 2006), shaping attentional, perceptual, and affective processing. If someone directs their gaze towards us, our attention is immediately captured. Direct gaze (as opposed to averted gaze or closed eyes) is preferentially detected and processed (e.g., Baron-Cohen, 1995; Coelho et al., 2006; Senju & Hasegawa, 2005; von Grünau & Anston, 1995; also see Böckler et al., 2015). Being looked at by another person may elicit self-referential processing in the addressee (Conty et al., 2016; Hietanen & Hietanen, 2017) and boost prosocial behavior (Izuma et al., 2010, 2011) as well as positive appraisal of others (see Kleinke, 1986, for a review). Another’s eyes convey information about their affective and cognitive states (e.g., Emery, 2000; Kleinke, 1986; Schilbach, 2015) and can signal aggression (Nichols & Champness, 1971) as well as attraction (Mason et al., 2005). Critically, the eyes not only signal information to others but they simultaneously encode information from the environment, thus serving a dual function (Cañigueral & Hamilton, 2019; Gobel et al., 2015; Kendon, 1967; Risko et al., 2016).
Not surprisingly, eye contact plays a major role in human communication. In particular, direct gaze serves as an ostensive communicative signal (Csibra & Gergely, 2009; Sperber & Wilson, 1986): A speaker typically directs her gaze towards the addressee, thereby making it manifest to the addressee that she is the designated recipient of the upcoming message and that this message is being intentionally communicated to her (Csibra & Gergely, 2009; Lanthier et al., 2019, 2021). Moreover, the gaze is used in a conversation to control turn-taking behavior such that a speaker ends her turn with a direct gaze at the addressee and the addressee then begins her turn with averted gaze (Kendon, 1967; recently replicated by Ho et al., 2015). When responding to questions, preferred responses are produced while gazing at the questioner whereas the gaze is averted for dispreferred responses (Kendrick & Holler, 2017). To signal understanding in conversation, addressees systematically use eye blinking (Hömke et al., 2017). It has further been found that the stronger the coupling between speakers’ and addressee’s eye movements, the better the addressee’s comprehension (Richardson & Dale, 2005).

Multisensory integration in speech perception

While the processing of a speaker’s gaze is important in a conversation, the processing of the speaker’s speech is naturally paramount. During speech perception, addressees integrate the incoming auditory and visual signals (i.e., the speaker’s vocal sounds and lip movements) into a unitary percept through a process known as multisensory integration (Ernst & Bülthoff, 2004). Generally, whether our brain binds multiple incoming sensory signals together or processes them separately depends on whether these signals are perceived to be causally related. When we hear another person’s words while seeing their lips move, the visual signal (the lip movements) and the auditory signal (the vocal sounds) have the same causal origin (the other person) and are integrated into a single percept rather than perceived as two separate signals. Typically, the integration of signals from multiple sensory modalities boosts perceptual performance because several sources of redundant information are combined. For instance, speech in a noisy environment is understood better if the speaker’s lip movements can be observed by the addressee (Ma et al., 2009; MacLeod & Summerfield, 1987; Ross et al., 2007; also see Altieri et al., 2016).
If, however, the signals from multiple sensory modalities provide inconsistent information, the process of multisensory integration can impair or bias perceptual performance and can lead to sensory illusions (e.g., Shams et al., 2000). In the case of speech perception, if the visual and auditory signals do not match, this can give rise to auditory percepts that neither correspond to the visual nor the auditory signal. In particular, McGurk and MacDonald (1976) have shown that when participants were presented with an auditory syllable (e.g. “Ba”) while observing the speaker’s lips uttering a different syllable (e.g. “Ga”), they reported hearing an illusory syllable (e.g. “Da”). This striking multisensory illusion (often referred to as the “McGurk illusion” or “McGurk effect”) demonstrated, for the first time, the powerful influence of vision upon auditory speech perception. The McGurk illusion has since been extensively investigated with regard to the mechanisms underlying it and the factors influencing it (for a comprehensive review, see Alsius et al., 2018). Note that the illusion has also been reliably found in studies that were run online (e.g., Brown et al., 2018; Karas et al., 2019; Magnotti et al., 2018, 2020), with highly similar results between lab-based and online studies (Magnotti et al., 2018). Previous research suggests that the McGurk illusion can be explained using a causal inference model of multisensory perception (Magnotti & Beauchamp, 2017) that has also been applied successfully to a variety of other multisensory phenomena, such as the ventriloquist effect (Körding et al., 2007; Rohe & Noppeney, 2015) and the sound-induced flash illusion (Shams et al., 2005).
In terms of factors that influence the McGurk illusion, previous studies have shown that attentional demands play a critical role. In particular, if participants were presented with the typical McGurk stimuli and asked to indicate what they heard while at the same time performing a secondary (visual or auditory) task, the perception of the McGurk illusion decreased (Alsius et al., 2005). This finding indicates that audiovisual integration of speech is weakened under a high attentional load. Relatedly, another study (Munhall et al., 2009) showed that the McGurk illusion is perceived only if addressees consciously attend to the visual signal, i.e., the speaker’s lip movements. Together, these findings suggest that one needs to pay (a sufficient amount of) attention to the presented stimuli in order for the integration process to take place.
Whereas dual tasks like the above (Alsius et al., 2005) explicitly shift and divide participants’ attentional resources, a recent study by Gurler et al. (2015) investigated the natural (i.e., uninfluenced) distribution of people’s visual attention while they observed the face of the “McGurk speaker” uttering syllables. Via eye tracking, the authors measured which region on the speaker’s face participants tended to fixate on, using eye fixation as a proxy for attentional focus. The results showed that the distribution of participants’ eye fixations predicted the degree to which participants perceived the McGurk illusion. Specifically, the more participants tended to look at the speaker’s mouth, the more they perceived the McGurk illusion—presumably because they were more strongly influenced by the visual signal provided by the lip movements (for a replication and additional manipulations, see Stacey et al., 2020). Hence, observing the speaker’s lip movements—a strategy that would usually improve speech comprehension (Ma et al., 2009; MacLeod, & Summerfield, 1987; Ross et al., 2007)—has a negative and misleading effect because the visual signal coming from the lips is inconsistent with the auditory signal.
Interestingly, Gurler et al. (2015) also observed that there were several participants who did not look primarily at the speaker’s mouth, choosing instead to look at the speaker’s eyes or explore multiple regions of the face. These participants perceived the McGurk illusion to a lesser extent. Thus, this study suggests that there is no commonly shared focus of attention in the McGurk paradigm, i.e., participants tend to look at the mouth, or at the eyes, or at multiple face regions. Hence, one cannot predict in advance where a particular participant will focus her attention when confronted with the McGurk stimulus.
In sum, previous research by Alsius et al. (2005), Gurler et al. (2015), Munhall et al. (2009), and Stacey et al. (2020) suggests that the degree to which people perceive the McGurk illusion depends (1) on their attention in general (with less attention leading to a decrease of the illusion) and (2) on their attentional focus on the speaker’s mouth versus elsewhere (with a focus on the mouth leading to an increase of the illusion).

Does speaker gaze affect multisensory integration in speech perception?

The eyes of another person have—amongst other things—the function and the power to spontaneously direct, divert, or capture our attention. In particular, another’s direct eye gaze and motion onset (e.g., switching from direct to averted gaze, or from closed to open eyes) are two powerful cues that capture our attention (e.g., Abrams & Christ, 2003; Böckler et al., 2014; Senju & Hasegawa, 2005). In turn, what we attend (and to what extent we attend it) determines how we process incoming multisensory signals (e.g., Alsius et al., 2005; Munhall et al., 2009) – and thus affects whether we can be “tricked” by our senses and subjected to multisensory illusions (Gurler et al., 2015; Stacey et al., 2020).
During a conversation, a speaker’s eyes and the audiovisual signals coming from their lips both typically provide relevant information. The addressee processes the speaker’s gaze while at the same time processing her audiovisual speech signals. To date, it has not been systematically investigated whether these two processes interact, i.e., whether the speaker’s gaze affects how the addressee integrates the speaker’s audiovisual speech signals. In the present study, we aimed to address this question using the classic McGurk illusion. Building on previous research, we examined if a speaker’s gaze behavior (i.e., motion and eye contact) would dynamically capture an addressee’s attention and would thereby influence how the addressee integrates the speaker’s vocal sounds and lip movements.
To test the effect of motion, we manipulated whether the speaker moved his eyelids up/down (i.e., open/closed his eyes) prior to speaking or did not show any eye motion (factor “Motion”). To test the effect of eye contact, we manipulated whether the speaker spoke with open eyes or closed eyes (factor “Eyes”). The factor Motion was manipulated between-subjects, the factor Eyes was manipulated within-subjects.

Predictions for motion

If the sudden onset of motion in the eye region draws the addressee's attention to the speaker’s eyes (and away from the lips), we expect a reduced McGurk illusion when the speaker opens or closes his eyes prior to speaking compared to when there is no motion of the eyes. The misleading visual signal from the lips receives less attention in this case and thus influences the audiovisual integration process to a lesser extent, resulting in a more accurate perception of the auditory syllable (cf. Gurler et al., 2015; Munhall et al., 2009).

Predictions for eyes

Similarly, if a speaker’s direct gaze draws the addressee’s attention to the speaker’s eyes, as shown by previous research (e.g., Böckler et al., 2014, who used face pictures), we expect a reduced McGurk illusion when the speaker has open eyes as compared to closed eyes.
Alternatively, however, one could predict that a speaker’s direct, ostensive gaze leads to a general increase in the addressee’s attention because she feels personally addressed and is eager to understand the speaker’s message (cf. Csibra & Gergely, 2009; Lanthier et al., 2019, 2021). When the speaker speaks with closed eyes, the addressee might not consider herself the intended recipient of the message and thus not pay as much attention. Given that attention is essential for audiovisual integration to occur in the first place (Alsius et al., 2005; Munhall et al., 2009; Talsma et al., 2007) and that selective attention enhances the integration process (Talsma & Woldorff, 2005), this pattern of behavior would result in the addressee experiencing an enhanced McGurk illusion when the speaker has open eyes (addressee pays close attention) as compared to closed eyes (addressee pays reduced attention).
Two further considerations support the latter prediction of enhanced McGurk illusion when the speaker has open eyes, yet for different reasons. First, being looked at by another person increases self-referential processing (Conty et al., 2016; Hietanen & Hietanen, 2017), self-awareness (e.g., Baltazar et al., 2014; Hazem et al., 2017; Pönkänen et al., 2011), and arousal (Helminen et al., 2011; Hietanen et al., 2020). Accordingly, looking into a speaker’s open eyes is more demanding for the addressee than looking at a speaker’s closed eyes. Second, it is possible that a speaker’s closed eyes capture the addressee’s attention because closed eyes in a conversation are very unusual for a speaker and thus salient for an addressee. Together, these two points suggest that an addressee might focus more on a speaker’s eyes (and thus less on his lips) when they are closed as opposed to open. In this case, the addressee would be less susceptible to the McGurk illusion when the speaker’s eyes are closed. Thus, one should expect an enhanced McGurk illusion when the speaker has open eyes as compared to closed eyes.
In sum, the theoretical considerations spelled out above provide grounds for a bidirectional prediction for the factor Eyes. On the one hand, one can predict that a speaker’s open eyes will draw the addressee’s attention (away from the speaker’s lips), leading to a smaller McGurk illusion for open compared to closed eyes. On the other hand, one can predict that (1) a speaker’s open eyes will generally lead to higher levels of attention in the addressee and (2) a speaker’s closed eyes will draw the addressee’s attention (away from the speaker’s lips). Both (1) and (2) would lead to a larger McGurk illusion for open compared to closed eyes.

Methods

Participants

We determined our target sample size of 2 × 70 participants by running an a priori power analysis using G*Power (Faul et al., 2007, 2009) targeting moderately sized effects (Cohen’s d = 0.34 for paired samples t-tests; Cohen’s d = 0.48 for independent samples t-tests; alpha = 0.05, Power = 0.80). Thus, we recruited 70 participants for each level of our between-subjects factor Motion, i.e., 70 for the “Static” condition (motion absent) and 70 for the “Dynamic” condition (motion present). Data was collected through the online participant recruitment service Prolific (https://​www.​prolific.​co/​).1 All participants had normal or corrected-to-normal vision and hearing; they were between 18 and 45 years old and fluent in English.2 Only participants whose performance in previous Prolific studies had been reliable (approval rates of at least 75%) were admitted to this study. Moreover, participants who showed below 60% accuracy in an auditory baseline condition of our study were excluded because we needed to ensure intact hearing capabilities.
The participant samples for the Static and the Dynamic conditions consisted of 33 females, 36 males, and 1 other (M = 25.47 years, SD = 6.16 years), and of 21 females, 48 males, and 1 other (M = 24.30 years, SD = 6.33 years), respectively. All participants gave written informed consent and received monetary compensation for their participation (1.50 GBP for the Static condition (~ 12 min) and 1.88 GBP for the slightly longer Dynamic condition (~ 15 min)).

Apparatus and stimuli

Videos were recorded with a MacBook (early 2016) using its internal microphone and its 480p FaceTime camera. Each video showed a headshot of a man (see Fig. 1) uttering one out of five syllables (“Ba”, “Pa”, “Ga”, “Ka”, “Na”). The speaker’s eyes were either open (“Eyes open”) or closed (“Eyes closed”) while he was speaking. In the Dynamic condition, the speech act was preceded by motion: the speaker either opened or closed his eyes before uttering the syllable (with his eyes remaining open or closed). Each video clip lasted 3.5 s. To create a second version of the videos that did not contain the preceding motion (Static condition), the respective first part of each clip was cut off, resulting in a shorter (2 s) version of each clip. Thus, the long video clips showed the speaker first closing or opening his eyes and then uttering one of the syllables with either open or closed eyes (Dynamic) whereas the shorter clips only included the speech act without the preceding motion (Static), see Fig. 1.
These video clips naturally included matching auditory and visual signals (e.g., the speaker’s lips produced the word “Ga” and the auditory signal was “Ga”). In addition to these “congruent” versions, we created “incongruent” versions with mismatching auditory and visual signals. To this end, the actual sounds in the videos were muted and dubbed either with the auditory syllable “Ba” or “Pa” (using iMovie, version 10.1.9) such that, for example, the speaker’s lips produced the word “Ga” yet the auditory signal was “Ba” or “Pa”. The final set of audiovisual stimuli included 6 incongruent versions and 6 congruent versions (see Table 1 for an overview). Additionally, we used only the audio tracks for the syllables “Ba” and “Pa” as an auditory baseline (see “auditory only” in Table 1).
Table 1
Syllable combinations and AFC response options
(adopted from Stropahl et al., 2017)
Auditory–visual
4-AFC options (auditory, visual, Fusion1, Fusion2)
Trial type
Ba–Ga
Ba, Ga, Da, Ma
Incongruent
Ba–Ka
Ba, Ka, Ga, Da
Incongruent
Ba–Na
Ba, Na, Ga, Da
Incongruent
Pa–Ga
Pa, Da, Ka, Ta
Incongruent
Pa–Ka
Pa, Ka, Da, Ta
Incongruent
Pa–Na
Pa, Na, Ka, Ta
Incongruent
Ba–Ba
Ba, Ga, Da, Ma
Congruent
Ba–Ba
Ba, Ka, Ga, Da
Congruent
Ba–Ba
Ba, Na, Ga, Da
Congruent
Pa–Pa
Pa, Da, Ka, Ta
Congruent
Pa–Pa
Pa, Ka, Da, Ta
Congruent
Pa–Pa
Pa, Na, Ka, Ta
Congruent
Ba–None
Ba, Ga, Da, Ma
Auditory only
Ba–None
Ba, Ka, Ga, Da
Auditory only
Ba–None
Ba, Na, Ga, Da
Auditory only
Pa–None
Pa, Da, Ka, Ta
Auditory only
Pa–None
Pa, Ka, Da, Ta
Auditory only
Pa–None
Pa, Na, Ka, Ta
Auditory only
All video material was recorded in one continuous take to ensure similar intonation and loudness of the uttered syllables and to keep visual conditions constant. The speaker opened and closed his eyes and produced the syllables in sync with a metronome beat (40 bpm; not audible to participants) to ensure that the temporal sequences were comparable across videos. This way, we also attempted to exclude potential tempo differences between speaking with open vs. with closed eyes. All videos are publicly accessible (under CC-BY license) via the Open Science Framework (https://​osf.​io/​vjw6k/​?​view_​only=​7c4c9831aa6a4dbb​8fcfbd5a8ef7e501​).
The experiment was programmed in PsychoPy 3 (Peirce et al., 2019) and run online via Pavlovia (https://​pavlovia.​org/​).

Design and procedure

To test whether motion and eye contact affect multisensory speech processing, we used a 2 × 2 mixed factorial design. As a between-subjects factor, we varied whether motion (i.e., opening or closing of the speaker’s eyes) preceded the speech act or not (Motion: Dynamic/Static). As a within-subjects factor, we varied whether the speaker’s eyes were open or closed while he was uttering the syllable (Eyes: Open/Closed).
Participants were presented with three different trial types (incongruent, congruent, auditory only) in randomized order, see Table 1. In the incongruent trials, the syllable that the speaker produced with his lips (either “Ga”, “Ka”, or “Na”) did not match the auditory syllable that was presented (“Pa” or “Ba”). In the congruent trials, the syllable that the speaker produced matched the auditory syllable (“Pa” or “Ba”). In the auditory only trials, participants were presented with a black screen (instead of the speaker’s face) and either heard the syllable “Pa” or “Ba”. Note that we did not include “visual only” trials.
After stimulus presentation, participants were asked to indicate which syllable—out of four options presented on the screen—they heard (see Table 1, for a list of all syllable combinations and response options). The four response options were adopted from Stropahl et al. (2017) and comprised the presented auditory syllable, the presented visual syllable, and two fusion responses. The two fusion responses were the two responses that showed the highest fusion percentage (i.e., the most commonly reported illusory percepts resulting from the fusion of mismatching auditory and visual signals) in the original study by McGurk and MacDonald (1976). The response options were the same for all three trial types (see Table 1).
Participants responded by pressing the corresponding number of the response option (1, 2, 3 or 4) on the keyboard. There was no time limit for responses. Participants were told that if they were unsure about what the speaker said, they should simply choose the response option that seemed most likely to them. They were ensured that this task was not about accuracy but about their individual perception.
Participants performed a total of 120 trials. The order of response options for each trial type was randomly chosen out of two possible orders (either “Visual, Auditory, Fusion1, Fusion2” or “Fusion1, Fusion2, Visual, Auditory”). The options were presented in rows, one word beneath the other, and numbered consecutively from 1 to 4. The 120 trials were composed of 24 auditory only, 24 congruent, and 72 incongruent trials. Critically, in half of the congruent and in half of the incongruent trials, the speaker’s eyes were open (first and third row, Fig. 1); in the other half they were closed (second and fourth row, Fig. 1).
The only difference between the two experimental conditions was that in the Dynamic condition, the speaker either opened or closed his eyes prior to uttering the syllable (third and fourth row, Fig. 1). In the Static condition, there was no motion preceding the utterance (first and second row, Fig. 1). Exemplary trial sequences for all four-factor combinations (Static + Eyes open, Static + Eyes closed, Dynamic + Eyes open, Dynamic + Eyes closed) are shown in Fig. 1.
As part of the general study instructions, participants were asked to turn off all distractions (e.g., music, TV, phone) and to always look at the screen and listen to the voice. They were asked to concentrate fully on the task and not to perform any other tasks at the same time. The instructions pointed out that they should read the response options carefully because the available options and the order of options would change continuously. Finally, participants were asked to wear headphones, if possible, while completing the study.
Before starting the actual experiment, participants were familiarized with the trial logic by performing three training trials. The first two training trials were congruent trials and the third trial was an auditory-only trial.
As dependent variables, we recorded response accuracies (for congruent and auditory-only trials), response choices (for incongruent trials), and response times3 for all trials. Response accuracy was computed as the proportion of trials in which participants chose the auditory syllable (i.e., the vocal sound that is actually presented) out of the four response options. Response choice for incongruent trials (where auditory and visual signals mismatched) was computed as the proportion of trials in which participants chose the auditory syllable, the visual syllable (i.e., the sound originally produced by the lips), and the fused syllable (i.e., the illusory percept), respectively. Response time was computed as the time between stimulus offset (i.e., end of video and appearance of the response options) and response selection.

Data analysis

For statistical inference, we used permutation-based ANOVAs and post-hoc tests. That is, the null distribution of the test statistics was estimated by repeatedly sampling permutations of the actual data under the assumption that there are no differences between the levels of our experimental factors (Kherad-Pajouh & Renaud, 2015). All post-hoc tests were Bonferroni-corrected. As effect size measures, we report generalized eta squared (ηG2; Bakeman, 2005) for the ANOVAs and Cohen’s d for the post-hoc tests. Data were analyzed using customized R scripts.
All raw data are publicly available via the Open Science Framework (https://​osf.​io/​vjw6k/​?​view_​only=​7c4c9831aa6a4dbb​8fcfbd5a8ef7e501​).

Results

Auditory baseline and congruent trials

Response accuracy

First, we aimed to verify that participants demonstrated the expected highly accurate performance in auditory-only trials and in congruent trials. In these trials, the identification of the presented syllable should be straightforward as there is no mismatching information. We found that participants in both conditions were highly accurate (M = 85%) in auditory-only trials and almost reached ceiling performance (M = 95%) in congruent trials (see Fig. 2).
To test whether accuracy levels differed statistically in the auditory-only trials and congruent trials, we performed an analysis including the factor Motion (Static/Dynamic) as between-subjects factor and the factor Eyes (Auditory only, Eyes open (congruent), and Eyes closed (congruent)) as within-subjects factor. This resulted in a 2 (Motion: Static/Dynamic) × 3 (Eyes: Auditory only, Eyes closed, Eyes open) ANOVA. The results showed a significant main effect of Eyes (F(2,276) = 81.65, p < 0.001, ηG2 = 0.178) but no other significant effects (Motion: F(1,138) = 0.04, p = 0.840, ηG2 < 0.001; Eyes x Motion: F(2,276) = 1.16, p = 0.315, ηG2 = 0.003). We followed up the significant main effect of Eyes by performing pairwise comparisons between the three-factor levels using paired t-tests, separately for the Static and Dynamic condition. We found that participants showed a significantly lower accuracy in the auditory-only trials compared to the other two trial types in both the Static and the Dynamic condition (all corrected ps < 0.007; averaged Cohen’s d = 0.92); the other comparisons were not significant (all corrected ps > 0.199; averaged Cohen’s d = 0.14). In sum, these results show that participants could reliably identify the spoken syllables (“Ba” and “Pa”) in the auditory-only trials and that their performance was further boosted—as expected based on previous research (Ma et al., 2009; MacLeod, & Summerfield, 1987; Ross et al., 2007)—when compatible visual information (the speaker’s lip movements) was provided in the congruent trials. This performance boost occurred irrespective of whether the speaker’s eyes were closed or open and irrespective of whether prior motion occurred or not.

Response times

As a second step, we determined whether the differences in participants’ accuracy levels were reflected in their response times. Repeating the same 2 × 3 ANOVA as reported above with response times as dependent variable, we found no significant effects (Eyes: F(2,276) = 2.88, p = 0.058, ηG2 = 0.002; Motion: F(1,138) = 0.08, p = 0.775, ηG2 < 0.001; Eyes × Motion: F(2,276) = 0.29, p = 0.747, ηG2 < 0.001). These results indicate that findings in accuracies were not due to a speed-accuracy trade-off.

Incongruent trials

Response choices

To test our main research question of whether motion and eye contact affect multisensory speech processing, we analyzed participants’ responses in the incongruent trials where auditory and visual signals mismatched. On a descriptive level, we observed that participants chose the fusion response in a high proportion of trials in both conditions (M = 68%), indicating that they did indeed experience the McGurk illusion, as expected based on previous research (McGurk & MacDonald, 1976). Critically, the proportion of fusion responses (see “Fused” in Fig. 3) was smaller for Dynamic compared to Static and smaller for Eyes closed compared to Eyes open.
We tested whether these observations were statistically significant using a 2 (Motion: Static/Dynamic) × 2 (Eyes: Open/Closed) ANOVA with proportion of fusion responses as dependent variable. We found a significant main effect of Motion (F(1,138) = 7.98, p = 0.007, ηG2 = 0.053) and a significant main effect of Eyes (F(1,138) = 5.07, p = 0.026, ηG2 = 0.001), indicating that participants selected fewer fusion responses when motion preceded the speaker’s utterance and when the speaker’s eyes were closed (vs. open). The interaction effect was not significant (F(1,138) = 1.04, p = 0.308, ηG2 < 0.001). Nonetheless, to assess whether the main effect of Eyes was present in both Motion conditions, we ran two pairwise t-tests comparing Eyes open to Eyes closed, separately for each Motion condition. The effect was significant in the Static condition (t(69) = 2.76, p = 0.007, Cohen’s d = 0.33) yet it was not significant in the Dynamic condition (t(69) = 0.76, p = 0.447, Cohen’s d = 0.09).

Response times

We determined whether these differences in participants’ response selection were reflected in their response times. Repeating the same 2 × 2 ANOVA with response times as dependent variable, we found no significant effects (Motion: F(1,138) = 1.82, p = 0.177, ηG2 = 0.013; Eyes: F(1,138) = 1.42, p = 0.230, ηG2 < 0.001; Motion x Eyes: F(1,138) = 0.26, p = 0.613, ηG2 < 0.001); see Fig. 4.

Auditory and visual syllables

As the analysis of the fusion responses had shown that participants selected fewer fusion responses when eye motion preceded the speaker’s utterance and when the speaker’s eyes were closed, we aimed to find out which response (i.e., the auditory or the visual syllable) participants chose instead of the fusion response. To this end, we ran two further 2 × 2 ANOVAs, using the proportion of auditory syllables and the proportion of visual syllables as a dependent variable, respectively. These proportions indicate how often participants selected the auditory/visual syllable relative to the total number of responses.
For the proportion of auditory syllables, we found the mirror-inverted pattern of results as reported above for the proportion of fusion responses: a significant main effect of Motion (F(1,138) = 7.57, p = 0.008, ηG2 = 0.051) and a trend towards significance for Eyes (F(1,138) = 3.37, p = 0.068, ηG2 = 0.001). Again, the interaction effect was not significant (F(1,138) = 1.16, p = 0.275, ηG2 < 0.001). For the proportion of visual syllables, we found no significant effects (Motion: F(1,138) = 0.01, p = 0.927, ηG2 < 0.001; Eyes: F(1,138) = 0.72, p = 0.401, ηG2 = 0.001; Motion x Eyes: F(1,138) = 0.002, p = 0.969, ηG2 < 0.001). Taken together, these results indicate that participants selected the auditory syllable (what the speaker actually said) more often when eye motion preceded the speaker’s utterance and when the speaker’s eyes were closed (vs. open). Thus, under these conditions, participants selected the auditory syllable instead of the fusion response. This finding is in line with previous studies on the McGurk illusion which also showed that in incongruent trials where participants do not choose the fusion response, they typically choose the auditory rather than the visual syllable (e.g., Stropahl et al., 2017).

Control condition: static matched

Rationale

A shortcoming of the present study is the fact that the length of the videos that were shown in the two between-subject conditions were of different length. Participants in the Static condition saw videos lasting 2 s whereas participants in the Dynamic condition saw videos lasting 3.5 s. Hence, it is possible that the difference in behavior observed between the two conditions is not a result of our manipulation (motion absent vs. present) but rather an effect of the video length. To deal with this potential confound, we ran an additional control condition in which we replicated the Static condition, yet matched it in length to the Dynamic condition. This was done by showing a still frame of the speaker for the first 1.5 s of the video—instead of the motion onset that was shown in the Dynamic condition.

Methods

A sample of 70 participants (32 females, 37 males, 1 other; M = 26.13 years, SD = 7.37 years) took part in the control condition. Design, procedure, and data analysis were the same as in the main experiment.

Results

Participants were highly accurate (M = 83%) in auditory-only trials and almost reached ceiling performance (M = 96%) in congruent trials (see Fig. 2). To test whether accuracy levels differed statistically in the auditory-only trials and congruent trials, we performed a one-way ANOVA with the within-subjects factor Eyes (Auditory only, Eyes closed (congruent), Eyes open (congruent)). As before, there was a main effect of Eyes (F(2,138) = 74.86, p < 0.001, ηG2 = 0.343), showing that participants had a significantly lower accuracy in the auditory-only trials compared to the other two trial types (both corrected ps < 0.001; averaged Cohen’s d = 1.13). These differences in participants’ accuracy levels were also reflected in their response times (F(2,138) = 3.66, p = 0.030, ηG2 = 0.011), with slower responses in the auditory-only trials.
As before, participants chose the fusion response in a high proportion of trials (M = 55%), indicating that they experienced the McGurk illusion (see Fig. 3). The proportion of fusion responses was descriptively, yet not significantly, smaller for Eyes closed compared to Eyes open, as shown by a paired t-test (t(69) = 1.57, p = 0.122, Cohen’s d = 0.19). Again, this difference was reflected in a mirror-inverted pattern for the auditory syllables, which were selected significantly more often in the Eyes closed compared to the Eyes open condition, as shown by a paired t-test (t(69) = 2.10, p = 0.039, Cohen’s d = 0.23).
When integrating the latter result into the overall context, it seems that the effect of Eyes (i.e., a reduced McGurk illusion when the speaker’s eyes are closed) is significant in the Static condition only, yet fails to reach significance in the Static matched condition (see above) and Dynamic condition (see Sect. Incongruent trials - Response choices). However, when looking at the data more closely, it turns out that the effect of Eyes depends on the basic size of the McGurk illusion (i.e., the percentage of perceived fused responses). In particular, the reason for the absence of the effect of Eyes in the Static matched and Dynamic conditions might be the generally smaller McGurk illusion in these conditions compared to the Static condition (Static matched: 55%; Dynamic: 62%; Static: 74%; see Fig. 3).
As we noticed that our data for the Static matched and Dynamic conditions seemed to be bimodally distributed, we considered performing a Median split to gain a better understanding of participants' behavior. To this end, we first assessed the degree of bimodality by calculating a bimodality coefficient (Pfister et al., 2013). In line with Pfister and colleagues, we considered a coefficient larger than 0.55 as an indication for bimodality. For both conditions, the computed coefficients surpassed this reference value (Static matched: 0.58; Dynamic: 0.59), suggesting that a Median split is a reasonable approach.
We first conducted a Median split for the Static matched condition and analyzed the above-Median and below-Median data sets separately. The results showed that for the above-Median data set, the size of the McGurk illusion is 82% when the speaker’s eyes are open and 78% when the speaker’s eyes are closed, resulting in a significant difference (t(34) = 3.16, p = 0.003, Cohen’s d = 0.31). Note that this effect size is comparable to the effect size of the Static condition. In contrast, for the below-Median data set, the size of the McGurk illusion is 29% when the speaker’s eyes are open and 30% when the speaker’s eyes are closed, showing no significant difference (t(34) = − 0.61, p = 0.546, Cohen’s d = − 0.05). When conducting the same Median split for the Dynamic condition, we find the same pattern for eyes open vs. closed (above-Median: 85% vs. 82%; below-Median: 40% vs. 41%). The difference between eyes open vs. closed for the above-Median data set is close to significant (t(34) = 1.86, p = 0.069, Cohen’s d = 0.29), yet it is not significant for the below-Median data set (t(34) =  − 0.58, p = 0.564, Cohen’s d = − 0.05).
To sum up, the effect of Eyes can be detected only if participants reliably perceive the McGurk illusion. Thus, the effect of Eyes can only be seen in those participants showing a large McGurk illusion (i.e., in the above-Median data set) but not in those showing a small McGurk illusion (i.e., in the below-Median data set).
Critically, we also compared the proportion of fusion responses in the control condition (Static matched) with the Static and the Dynamic conditions from the main experiment by conducting a 2 × 3 ANOVA with the within-subjects factor Eyes (Open, Closed), the between-subjects factor Motion (Static, Dynamic, Static matched), and with the proportion of fusion responses as a dependent variable. We found a significant main effect of Eyes (F(1,207) = 7.51, p = 0.006, ηG2 = 0.001), indicating that participants selected fewer fusion responses when the speaker’s eyes were closed (vs. open). There was also a significant main effect of Motion (F(2,207) = 9.08, p < 0.001, ηG2 = 0.079). The interaction effect was not significant (F(2,207) = 0.54, p = 0.542, ηG2 < 0.001). We followed up the main effect of Motion with pairwise comparisons. There was a significant difference between Static and Static matched (t(138) = 4.22, p < 0.001, Cohen’s d = 0.71), with a higher proportion of fusion responses in Static. There was no significant difference between Dynamic and Static matched (t(138) = 1.49, p = 0.138, Cohen’s d = 0.25). On a descriptive level, however, there was a higher proportion of fusion responses in Dynamic. This result indicates that the extent to which participants in our control condition—which was identical to the Static condition yet matched in length to the Dynamic condition—experienced the McGurk illusion was more similar to the Dynamic condition than to the Static condition (see Fig. 3).
We performed the same 2 × 3 ANOVA with response times as the dependent variable. The pattern of results mirrored the analysis for fusion responses (see Fig. 4). There was a significant main effect of Eyes (F(1,207) = 5.35, p = 0.022, ηG2 = 0.001), indicating that participants were faster to select a response when the speaker’s eyes were closed (vs. open). This suggests that when the speaker’s eyes were closed, participants were more likely to choose the accurate auditory response (instead of the fusion response) and to make this response faster compared to when the speaker’s eyes were open. There was also a significant main effect of Motion (F(2,207) = 3.90 p = 0.022, ηG2 = 0.034). The interaction effect was not significant (F(2,207) = 1.30, p = 0.275, ηG2 < 0.001). We followed up the main effect of Motion with pairwise comparisons. There was a significant difference between Static and Static matched (t(138) = 2.90, p = 0.004, Cohen’s d = 0.49), with slower responses in Static matched. There was no significant difference between Dynamic and Static matched (t(138) = 1.40, p = 0.164, Cohen’s d = 0.24). On a descriptive level, however, responses were slower in Static matched. This result indicates that participants were fastest to respond in the Static condition, distinctively slower in the Dynamic condition, and again slightly slower in the Static matched control condition (see Fig. 4).

Discussion

In the present study, we investigated whether a speaker’s gaze behavior (i.e., motion and eye contact) dynamically captures an addressee’s attention and thereby influences how the addressee processes the speaker’s audiovisual speech signals.
To this end, we used the classic McGurk illusion and manipulated whether the speaker (a) moved his eyelids up/down (i.e., opened/closed his eyes) prior to speaking or did not show any eye motion, and (b) spoke with open or closed eyes. When the speaker’s eyes moved (i.e., opened or closed) before an utterance, and when the speaker spoke with closed eyes, the McGurk illusion was weakened (i.e., addressees reported significantly fewer illusory percepts). It seems that these two main effects can be traced back to two separate mechanisms, as outlined below.

Effect of motion

When looking at the main experiment, the effect of motion seems to highlight the power of motion cues to capture people’s attention, converging with previous research on the effects of sudden onset (eye) motion on attentional capture (e.g., Abrams & Christ, 2003; Böckler et al., 2014, 2015; van der Wel et al., 2018). In particular, the sudden opening/closing of the speaker’s eyes presumably directed participants’ attention to the eyes (and away from the lips). Thus, the misleading visual signal provided by the speaker’s lips had a smaller influence on the audiovisual integration process such that participants perceived fewer illusory percepts and instead perceived the actual auditory syllables. This finding is in line with previous research showing that the degree to which people perceive the McGurk illusion depends on their attentional focus on the speaker’s mouth vs. elsewhere, with a focus on the mouth leading to an increase of the illusion (Gurler et al., 2015; Stacey et al., 2020). It has also been shown that people only experience the McGurk illusion if they consciously attend to the speaker’s lips (Munhall et al., 2009). Thus, the effect of motion may be interpreted in terms of attentional capture.
However, when considering the control experiment (“Static matched”), this interpretation is called into question, as the results show that a static stimulus that matched the motion stimulus in length also led to a reduced McGurk illusion. From this result, one might conclude that the motion itself did not actually play the main role, but it was rather the video length that caused the difference between the Static (motion absent, shorter video) and the Dynamic condition (motion present, longer video) in the main experiment. Some additional differences between the Static matched and Dynamic conditions should be considered, however. Specifically, participants in the Static matched condition (1) experienced the McGurk illusion to an even smaller extent than participants in the Dynamic condition and (2) responded more slowly than in the Dynamic condition. These two aspects in combination suggest that different processes might be at work in the two conditions, yet leading to similar outcomes. In particular, it is possible that the reduced McGurk illusion in the Dynamic condition is caused, at least partially, by the motion cue functioning as attentional capture (as discussed above). In the Static matched condition, however, the reduced McGurk illusion might be caused by the fact that participants were generally less attentive because the still frame at the beginning of each video is rather boring. This general reduction of attention might lead to (1) a reduction in audiovisual integration (cf. Alsius et al., 2005; Munhall et al., 2009; Talsma & Woldorff, 2005) and (2) a slow-down of responses; just as observed in the Static matched condition.

Effect of eye contact

Our results suggest that the McGurk illusion is smaller when the eyes of the speaker were closed compared to open–provided that people reliably perceive the basic McGurk illusion (regarding the latter constraint, please see Sect. Control condition: Static matched - Results). First of all, it is possible that when the speaker spoke with closed eyes, participants did not consider themselves the intended recipient of the message and thus paid less attention overall to the stimuli. Since sufficient attention is a prerequisite for audiovisual integration to occur (Alsius et al., 2005; Munhall et al., 2009; Talsma et al., 2007), a general reduction of attention might have impaired the integration process, thus resulting in fewer illusory percepts when the speaker’s eyes were closed.
The effect of eye contact might also suggest that a speaker’s closed eyes capture the addressee’s attention more strongly than open eyes in the setting of the present study. One reason could be that closed eyes in a conversation are very unusual and thus salient for an addressee. Typically, interlocutors converse with open eyes following specific gaze patterns (Ho et al., 2015), yet the speaker in the present study spoke with closed eyes. Presumably, the fact that speaking with closed eyes is not consistent with common social norms led to increased salience and thus caught and captured participants’ attention. Apart from the fact that closed eyes might be more salient than open eyes in a conversation context, it is also possible that participants focused more on the speaker’s closed compared to his open eyes because feeling another’s direct gaze elicits self-referential processing (“Watching Eyes model”, see Conty et al., 2016; Hietanen & Hietanen, 2017) and self-awareness (e.g., Baltazar et al., 2014; Hazem et al., 2017; Pönkänen et al., 2011), increases arousal (Helminen et al., 2011), and invites for social interaction (Ho et al., 2015). Thus, in line with the finding that people avoid long eye contact with strangers (Ellsworth et al., 1972; Laidlaw et al., 2011) and look longer at faces with averted than direct gaze (Helminen et al., 2011), participants in the present study might have preferred to attend to the speaker’s eyes when those were closed and they were not feeling watched, as this creates less self-involvement. Thus, the effect of eye contact can be interpreted in terms of the saliency of closed eyes in a conversation context as well as in terms of the Watching Eyes model (cf. Conty et al., 2016), as both mechanisms would result in fewer illusory percepts when the speaker’s eyes were closed.
A future lab-based study could use eye tracking to measure gaze fixations as a proxy of overt attention and pupil size as a proxy of attentional processing to disentangle the two interpretations. Specifically, the interpretation that a speaker’s closed eyes reduced the illusion due to generally reduced attention would be supported by gaze patterns that reflect disengagement of the visual scene (e.g., fewer fixations on the face) in the Eyes closed condition. By contrast, the interpretation that the reduced illusion is due to attention capture by closed eyes would be supported by earlier, more and/or longer fixations at the eye region in the Eyes closed condition. Finally, it is also possible that both interpretations are partially correct, as they are not mutually exclusive.

Study limitations

Some of our findings warrant additional research before stronger conclusions can be drawn. Regarding our interpretation of the motion cue in terms of attentional capture, it is possible that apart from the eyes, other motion cues in the speaker’s face (e.g., wrinkling the forehead) might have similar effects. It has been shown, for instance, that an external moving object (i.e., a falling leaf in front of a speaker’s face) reduces the McGurk illusion (Tiippana et al., 2004). It is noteworthy that in this study, the leaf motion occurred during the speaker’s utterance whereas the motion in the present study occurred prior to the utterance. Future studies are needed to determine the specificity and boundary conditions of the motion effect, especially regarding the ‘identity’ and the timing of the motion.
More generally, regarding the reduced McGurk illusion in the Dynamic condition (motion present, longer video) compared to the Static condition (motion absent, shorter video), we cannot argue conclusively as to whether it was the motion or the length of the video that caused this effect. The results of the control condition (“Static matched”: motion absent, longer video) resembled those of the Dynamic condition, suggesting, at first glance, that the effect can be attributed to the video length. However, as pointed out above, it is possible that different processes were at work in the Dynamic condition (motion cue guides addressee’s attention to the speaker’s eyes) and the Static matched condition (still frame leads to addressee feeling bored and thus paying less attention overall). Both processes would result in an impaired integration and thus in fewer illusory percepts. Further control experiments with systematic and orthogonal manipulations of motion and video length are needed to resolve this issue.
The effect of eye contact (i.e., a smaller McGurk effect when the speaker’s eyes are closed) in the present study is rather small and not significant in all pairwise comparisons. However, the effect is significant when analysed with an ANOVA with higher statistical power (see Sect. Control condition: Static matched - Results) and the effect (or a tendency, at least) occurs in all of the three conditions, indicating that it is reliable. Future research could identify potential factors that might increase the size of the effect or abolish it altogether.
When it comes to the generalizability and applicability of findings from the McGurk illusion, one should bear in mind that the processing of the McGurk illusion does not necessarily generalize to natural forms of audiovisual speech processing (for reviews, see Alsius et al., 2018; Rosenblum, 2019) as, for instance, audiovisual sentence recognition abilities do not predict one’s susceptibility to the McGurk illusion (van Engen et al., 2017) and distinct brain regions are active during the perception of audiovisual speech and the McGurk illusion (Erickson et al., 2014). Further systematic studies are thus needed to investigate in how far findings from the McGurk illusion extend to audiovisual speech perception more generally.

Conclusion

The present findings contribute to an ongoing debate on whether multisensory integration is an automatic process or whether it can be affected by attentional processes (for reviews, see Talsma et al., 2010; Ten Oever et al., 2016; Spence & Frings, 2020). In line with earlier work (Alsius et al., 2005; Tiippana et al., 2004), our findings support the view that multisensory integration is susceptible to attentional manipulations in the case of audiovisual speech processing. In particular, our findings suggest that attentional focus on the speaker’s eyes affects the integration process in the McGurk illusion. Future studies could expand this work by investigating if attention-capturing stimuli in the auditory domain (e.g., a word spoken prior to the McGurk syllable) also affects the integration process or if the observed effects are specific to the visual modality.
In sum, the present study shows that speaker gaze affects audiovisual integration in the McGurk illusion. Our results thereby demonstrate that a speaker’s gaze behavior can dynamically capture an addressee’s attention, influencing how the addressee processes the speaker’s audiovisual speech signals. These findings reaffirm the power of speaker gaze to guide attention, showing that its dynamics can modulate low-level processes such as the integration of multisensory speech signals.

Declarations

Conflict of interest

The authors declare that there is no conflict of interest.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Onze productaanbevelingen

BSL Psychologie Totaal

Met BSL Psychologie Totaal blijf je als professional steeds op de hoogte van de nieuwste ontwikkelingen binnen jouw vak. Met het online abonnement heb je toegang tot een groot aantal boeken, protocollen, vaktijdschriften en e-learnings op het gebied van psychologie en psychiatrie. Zo kun je op je gemak en wanneer het jou het beste uitkomt verdiepen in jouw vakgebied.

BSL Academy Accare GGZ collective

Voetnoten
1
Based on previous studies showing that the McGurk illusion can be reliably found in online studies, we were confident that conducting our study online should not be problematic, as we should be able to replicate the McGurk illusion.
 
2
Due to the fact that participants were recruited online, we could only collect basic demographic information. As a result, we do not know for sure whether English was participants’ native language or whether they learned English as a second (or third) language. We do know, however, that between 20 and 30% of our participants originated from the US or UK and hence should be native English speakers; most of the other remaining 70–80% originated from European countries where English is not the first language. (In this respect, the samples in our three conditions did not differ noticeably.) This aspect could be relevant to the present study because phonetic perception is influenced by one’s native language. We thus cannot exclude the possibility that differences in our participants’ first language might have affected the extent to which they perceived the McGurk illusion. Moreover, we could not collect information regarding participants’ sociocultural background and can thus not exclude the possibility that differences in our participants’ sociocultural norms might have affected how they interpret eye contact in communicative contexts. We acknowledge these shortcomings in our method, yet believe that our rather large sample size should have helped to average out potential inter-individual differences.
 
3
It is important to consider that temporal precision in online studies is not as good as in lab-based studies. However, a recent study comparing the performance of different popular packages for online studies (Bridges et al., 2020) showed that PsychoPy (which we used for the present study) achieves a precision under 3.5 ms for response times in all browsers, making it the most precise among all packages tested. We consider this level of precision sufficient for the present study in which response times lie in a range of 1.4–1.6 s and differences in response times between conditions are around 100 ms.
 
Literatuur
go back to reference Abrams, R. A., & Christ, S. E. (2003). Motion onset captures attention. Psychological Science, 14(5), 427–432.PubMedCrossRef Abrams, R. A., & Christ, S. E. (2003). Motion onset captures attention. Psychological Science, 14(5), 427–432.PubMedCrossRef
go back to reference Alsius, A., Navarra, J., Campbell, R., & Soto-Faraco, S. (2005). Audiovisual integration of speech falters under high attention demands. Current Biology, 15(9), 839–843.PubMedCrossRef Alsius, A., Navarra, J., Campbell, R., & Soto-Faraco, S. (2005). Audiovisual integration of speech falters under high attention demands. Current Biology, 15(9), 839–843.PubMedCrossRef
go back to reference Alsius, A., Paré, M., & Munhall, K. G. (2018). Forty years after hearing lips and seeing voices: The McGurk effect revisited. Multisensory Research, 31(1–2), 111–144.PubMedCrossRef Alsius, A., Paré, M., & Munhall, K. G. (2018). Forty years after hearing lips and seeing voices: The McGurk effect revisited. Multisensory Research, 31(1–2), 111–144.PubMedCrossRef
go back to reference Altieri, N., Lentz, J. J., Townsend, J. T., & Wenger, M. J. (2016). The McGurk effect: An investigation of attentional capacity employing response times. Attention, Perception, & Psychophysics, 78(6), 1712–1727.CrossRef Altieri, N., Lentz, J. J., Townsend, J. T., & Wenger, M. J. (2016). The McGurk effect: An investigation of attentional capacity employing response times. Attention, Perception, & Psychophysics, 78(6), 1712–1727.CrossRef
go back to reference Bakeman, R. (2005). Recommended effect size statistics for repeated measures designs. Behavior Research Methods, 37(3), 379–384.PubMedCrossRef Bakeman, R. (2005). Recommended effect size statistics for repeated measures designs. Behavior Research Methods, 37(3), 379–384.PubMedCrossRef
go back to reference Baltazar, M., Hazem, N., Vilarem, E., Beaucousin, V., Picq, J. L., & Conty, L. (2014). Eye contact elicits bodily self-awareness in human adults. Cognition, 133(1), 120–127.PubMedCrossRef Baltazar, M., Hazem, N., Vilarem, E., Beaucousin, V., Picq, J. L., & Conty, L. (2014). Eye contact elicits bodily self-awareness in human adults. Cognition, 133(1), 120–127.PubMedCrossRef
go back to reference Baron-Cohen, S. (1995). The eye direction detector (EDD) and the shared attention mechanism (SAM): Two cases for evolutionary psychology. In C. Moore & P. J. Dunham (Eds.), Joint attention: Its origins and role in development (pp. 41–59). Erlbaum. Baron-Cohen, S. (1995). The eye direction detector (EDD) and the shared attention mechanism (SAM): Two cases for evolutionary psychology. In C. Moore & P. J. Dunham (Eds.), Joint attention: Its origins and role in development (pp. 41–59). Erlbaum.
go back to reference Böckler, A., van der Wel, R. P., & Welsh, T. N. (2014). Catching eyes: Effects of social and nonsocial cues on attention capture. Psychological Science, 25(3), 720–727.PubMedCrossRef Böckler, A., van der Wel, R. P., & Welsh, T. N. (2014). Catching eyes: Effects of social and nonsocial cues on attention capture. Psychological Science, 25(3), 720–727.PubMedCrossRef
go back to reference Böckler, A., van der Wel, R. P., & Welsh, T. N. (2015). Eyes only? Perceiving eye contact is neither sufficient nor necessary for attentional capture by face direction. Acta Psychologica, 160, 134–140.PubMedCrossRef Böckler, A., van der Wel, R. P., & Welsh, T. N. (2015). Eyes only? Perceiving eye contact is neither sufficient nor necessary for attentional capture by face direction. Acta Psychologica, 160, 134–140.PubMedCrossRef
go back to reference Bridges, D., Pitiot, A., MacAskill, M. R., & Peirce, J. W. (2020). The timing mega-study: comparing a range of experiment generators, both lab-based and online. PeerJ, 8, e9414.PubMedPubMedCentralCrossRef Bridges, D., Pitiot, A., MacAskill, M. R., & Peirce, J. W. (2020). The timing mega-study: comparing a range of experiment generators, both lab-based and online. PeerJ, 8, e9414.PubMedPubMedCentralCrossRef
go back to reference Brown, V. A., Hedayati, M., Zanger, A., Mayn, S., Ray, L., Dillman-Hasso, N., & Strand, J. F. (2018). What accounts for individual differences in susceptibility to the McGurk effect? PLoS ONE, 13(11), e0207160.PubMedPubMedCentralCrossRef Brown, V. A., Hedayati, M., Zanger, A., Mayn, S., Ray, L., Dillman-Hasso, N., & Strand, J. F. (2018). What accounts for individual differences in susceptibility to the McGurk effect? PLoS ONE, 13(11), e0207160.PubMedPubMedCentralCrossRef
go back to reference Cañigueral, R., & Hamilton, A. F. D. C. (2019). The role of eye gaze during natural social interactions in typical and autistic people. Frontiers in Psychology, 10, 560.PubMedPubMedCentralCrossRef Cañigueral, R., & Hamilton, A. F. D. C. (2019). The role of eye gaze during natural social interactions in typical and autistic people. Frontiers in Psychology, 10, 560.PubMedPubMedCentralCrossRef
go back to reference Coelho, E., George, N., Conty, L., Hugueville, L., & Tijus, C. (2006). Searching for asymmetries in the detection of gaze contact versus averted gaze under different head views: A behavioural study. Spatial Vision, 19(6), 529–545.PubMedCrossRef Coelho, E., George, N., Conty, L., Hugueville, L., & Tijus, C. (2006). Searching for asymmetries in the detection of gaze contact versus averted gaze under different head views: A behavioural study. Spatial Vision, 19(6), 529–545.PubMedCrossRef
go back to reference Conty, L., George, N., & Hietanen, J. K. (2016). Watching Eyes effects: When others meet the self. Consciousness and Cognition, 45, 184–197.PubMedCrossRef Conty, L., George, N., & Hietanen, J. K. (2016). Watching Eyes effects: When others meet the self. Consciousness and Cognition, 45, 184–197.PubMedCrossRef
go back to reference Csibra, G., & Gergely, G. (2009). Natural pedagogy. Trends in Cognitive Sciences, 13(4), 148–153.PubMedCrossRef Csibra, G., & Gergely, G. (2009). Natural pedagogy. Trends in Cognitive Sciences, 13(4), 148–153.PubMedCrossRef
go back to reference Ellsworth, P. C., Carlsmith, J. M., & Henson, A. (1972). The stare as a stimulus to flight in human subjects: A series of field experiments. Journal of Personality and Social Psychology, 21(3), 302.PubMedCrossRef Ellsworth, P. C., Carlsmith, J. M., & Henson, A. (1972). The stare as a stimulus to flight in human subjects: A series of field experiments. Journal of Personality and Social Psychology, 21(3), 302.PubMedCrossRef
go back to reference Emery, N. J. (2000). The eyes have it: The neuroethology, function and evolution of social gaze. Neuroscience & Biobehavioral Reviews, 24(6), 581–604.CrossRef Emery, N. J. (2000). The eyes have it: The neuroethology, function and evolution of social gaze. Neuroscience & Biobehavioral Reviews, 24(6), 581–604.CrossRef
go back to reference Erickson, L. C., Zielinski, B. A., Zielinski, J. E., Liu, G., Turkeltaub, P. E., Leaver, A. M., & Rauschecker, J. P. (2014). Distinct cortical locations for integration of audiovisual speech and the McGurk effect. Frontiers in Psychology, 5, 534.PubMedPubMedCentralCrossRef Erickson, L. C., Zielinski, B. A., Zielinski, J. E., Liu, G., Turkeltaub, P. E., Leaver, A. M., & Rauschecker, J. P. (2014). Distinct cortical locations for integration of audiovisual speech and the McGurk effect. Frontiers in Psychology, 5, 534.PubMedPubMedCentralCrossRef
go back to reference Ernst, M. O., & Bülthoff, H. H. (2004). Merging the senses into a robust percept. Trends in Cognitive Sciences, 8(4), 162–169.PubMedCrossRef Ernst, M. O., & Bülthoff, H. H. (2004). Merging the senses into a robust percept. Trends in Cognitive Sciences, 8(4), 162–169.PubMedCrossRef
go back to reference Farroni, T., Johnson, M. H., & Csibra, G. (2004). Mechanisms of eye gaze perception during infancy. Journal of Cognitive Neuroscience, 16(8), 1320–1326.PubMedCrossRef Farroni, T., Johnson, M. H., & Csibra, G. (2004). Mechanisms of eye gaze perception during infancy. Journal of Cognitive Neuroscience, 16(8), 1320–1326.PubMedCrossRef
go back to reference Faul, F., Erdfelder, E., Buchner, A., & Lang, A. G. (2009). Statistical power analyses using g* power 3.1: Tests for correlation and regression analyses. Behavior Research Methods, 41, 1149–1160.PubMedCrossRef Faul, F., Erdfelder, E., Buchner, A., & Lang, A. G. (2009). Statistical power analyses using g* power 3.1: Tests for correlation and regression analyses. Behavior Research Methods, 41, 1149–1160.PubMedCrossRef
go back to reference Faul, F., Erdfelder, E., Lang, A. G., & Buchner, A. (2007). G* power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods, 39, 175–191.PubMedCrossRef Faul, F., Erdfelder, E., Lang, A. G., & Buchner, A. (2007). G* power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods, 39, 175–191.PubMedCrossRef
go back to reference Gobel, M. S., Kim, H. S., & Richardson, D. C. (2015). The dual function of social gaze. Cognition, 136, 359–364.PubMedCrossRef Gobel, M. S., Kim, H. S., & Richardson, D. C. (2015). The dual function of social gaze. Cognition, 136, 359–364.PubMedCrossRef
go back to reference Gurler, D., Doyle, N., Walker, E., Magnotti, J., & Beauchamp, M. (2015). A link between individual differences in multisensory speech perception and eye movements. Attention, Perception, & Psychophysics, 77(4), 1333–1341.CrossRef Gurler, D., Doyle, N., Walker, E., Magnotti, J., & Beauchamp, M. (2015). A link between individual differences in multisensory speech perception and eye movements. Attention, Perception, & Psychophysics, 77(4), 1333–1341.CrossRef
go back to reference Hazem, N., George, N., Baltazar, M., & Conty, L. (2017). I know you can see me: Social attention influences bodily self-awareness. Biological Psychology, 124, 21–29.PubMedCrossRef Hazem, N., George, N., Baltazar, M., & Conty, L. (2017). I know you can see me: Social attention influences bodily self-awareness. Biological Psychology, 124, 21–29.PubMedCrossRef
go back to reference Helminen, T. M., Kaasinen, S. M., & Hietanen, J. K. (2011). Eye contact and arousal: The effects of stimulus duration. Biological Psychology, 88(1), 124–130.PubMedCrossRef Helminen, T. M., Kaasinen, S. M., & Hietanen, J. K. (2011). Eye contact and arousal: The effects of stimulus duration. Biological Psychology, 88(1), 124–130.PubMedCrossRef
go back to reference Hietanen, J. O., & Hietanen, J. K. (2017). Genuine eye contact elicits self-referential processing. Consciousness and Cognition, 51, 100–115.PubMedCrossRef Hietanen, J. O., & Hietanen, J. K. (2017). Genuine eye contact elicits self-referential processing. Consciousness and Cognition, 51, 100–115.PubMedCrossRef
go back to reference Hietanen, J. O., Peltola, M. J., & Hietanen, J. K. (2020). Psychophysiological responses to eye contact in a live interaction and in video call. Psychophysiology, 57(6), e13587.PubMedCrossRef Hietanen, J. O., Peltola, M. J., & Hietanen, J. K. (2020). Psychophysiological responses to eye contact in a live interaction and in video call. Psychophysiology, 57(6), e13587.PubMedCrossRef
go back to reference Ho, S., Foulsham, T., & Kingstone, A. (2015). Speaking and listening with the eyes: Gaze signaling during dyadic interactions. PLoS One, 10(8), e0136905.PubMedPubMedCentralCrossRef Ho, S., Foulsham, T., & Kingstone, A. (2015). Speaking and listening with the eyes: Gaze signaling during dyadic interactions. PLoS One, 10(8), e0136905.PubMedPubMedCentralCrossRef
go back to reference Hömke, P., Holler, J., & Levinson, S. C. (2017). Eye blinking as addressee feedback in face-to-face conversation. Research on Language and Social Interaction, 50(1), 54–70.CrossRef Hömke, P., Holler, J., & Levinson, S. C. (2017). Eye blinking as addressee feedback in face-to-face conversation. Research on Language and Social Interaction, 50(1), 54–70.CrossRef
go back to reference Izuma, K., Matsumoto, K., Camerer, C. F., & Adolphs, R. (2011). Insensitivity to social reputation in autism. Proceedings of the National Academy of Sciences, 108(42), 17302–17307.CrossRef Izuma, K., Matsumoto, K., Camerer, C. F., & Adolphs, R. (2011). Insensitivity to social reputation in autism. Proceedings of the National Academy of Sciences, 108(42), 17302–17307.CrossRef
go back to reference Izuma, K., Saito, D. N., & Sadato, N. (2010). Processing of the incentive for social approval in the ventral striatum during charitable donation. Journal of Cognitive Neuroscience, 22(4), 621–631.PubMedCrossRef Izuma, K., Saito, D. N., & Sadato, N. (2010). Processing of the incentive for social approval in the ventral striatum during charitable donation. Journal of Cognitive Neuroscience, 22(4), 621–631.PubMedCrossRef
go back to reference Karas, P. J., Magnotti, J. F., Metzger, B. A., Zhu, L. L., Smith, K. B., Yoshor, D., & Beauchamp, M. S. (2019). The visual speech head start improves perception and reduces superior temporal cortex responses to auditory speech. eLife, 8, e48116.PubMedPubMedCentralCrossRef Karas, P. J., Magnotti, J. F., Metzger, B. A., Zhu, L. L., Smith, K. B., Yoshor, D., & Beauchamp, M. S. (2019). The visual speech head start improves perception and reduces superior temporal cortex responses to auditory speech. eLife, 8, e48116.PubMedPubMedCentralCrossRef
go back to reference Kendon, A. (1967). Some functions of gaze-direction in social interaction. Acta Psychologica, 26, 22–63.PubMedCrossRef Kendon, A. (1967). Some functions of gaze-direction in social interaction. Acta Psychologica, 26, 22–63.PubMedCrossRef
go back to reference Kendrick, K. H., & Holler, J. (2017). Gaze direction signals response preference in conversation. Research on Language and Social Interaction, 50(1), 12–32.CrossRef Kendrick, K. H., & Holler, J. (2017). Gaze direction signals response preference in conversation. Research on Language and Social Interaction, 50(1), 12–32.CrossRef
go back to reference Kherad-Pajouh, S., & Renaud, O. (2015). A general permutation approach for analyzing repeated measures ANOVA and mixed-model designs. Statistical Papers, 56(4), 947–967.CrossRef Kherad-Pajouh, S., & Renaud, O. (2015). A general permutation approach for analyzing repeated measures ANOVA and mixed-model designs. Statistical Papers, 56(4), 947–967.CrossRef
go back to reference Kleinke, C. L. (1986). Gaze and eye contact: A research review. Psychological Bulletin, 100(1), 78.PubMedCrossRef Kleinke, C. L. (1986). Gaze and eye contact: A research review. Psychological Bulletin, 100(1), 78.PubMedCrossRef
go back to reference Körding, K. P., Beierholm, U., Ma, W. J., Quartz, S., Tenenbaum, J. B., & Shams, L. (2007). Causal inference in multisensory perception. PLoS One, 2(9), e943.PubMedPubMedCentralCrossRef Körding, K. P., Beierholm, U., Ma, W. J., Quartz, S., Tenenbaum, J. B., & Shams, L. (2007). Causal inference in multisensory perception. PLoS One, 2(9), e943.PubMedPubMedCentralCrossRef
go back to reference Laidlaw, K. E., Foulsham, T., Kuhn, G., & Kingstone, A. (2011). Potential social interactions are important to social attention. Proceedings of the National Academy of Sciences, 108(14), 5548–5553.CrossRef Laidlaw, K. E., Foulsham, T., Kuhn, G., & Kingstone, A. (2011). Potential social interactions are important to social attention. Proceedings of the National Academy of Sciences, 108(14), 5548–5553.CrossRef
go back to reference Lanthier, S. N., Jarick, M., Zhu, M. J. H., Byun, C. S. J., & Kingstone, A. (2019). Socially communicative eye contact and gender affect memory. Frontiers Psychology, 10, 1128.CrossRef Lanthier, S. N., Jarick, M., Zhu, M. J. H., Byun, C. S. J., & Kingstone, A. (2019). Socially communicative eye contact and gender affect memory. Frontiers Psychology, 10, 1128.CrossRef
go back to reference Ma, W. J., Zhou, X., Ross, L. A., Foxe, J. J., & Parra, L. C. (2009). Lip-reading aids word recognition most in moderate noise: a Bayesian explanation using high-dimensional feature space. PLoS One, 4(3), e4638.PubMedPubMedCentralCrossRef Ma, W. J., Zhou, X., Ross, L. A., Foxe, J. J., & Parra, L. C. (2009). Lip-reading aids word recognition most in moderate noise: a Bayesian explanation using high-dimensional feature space. PLoS One, 4(3), e4638.PubMedPubMedCentralCrossRef
go back to reference MacLeod, A., & Summerfield, Q. (1987). Quantifying the contribution of vision to speech perception in noise. British Journal of Audiology, 21(2), 131–141.PubMedCrossRef MacLeod, A., & Summerfield, Q. (1987). Quantifying the contribution of vision to speech perception in noise. British Journal of Audiology, 21(2), 131–141.PubMedCrossRef
go back to reference Magnotti, J. F., & Beauchamp, M. S. (2017). A causal inference model explains perception of the McGurk effect and other incongruent audiovisual speech. PLoS Computational Biology, 13(2), e1005229.PubMedPubMedCentralCrossRef Magnotti, J. F., & Beauchamp, M. S. (2017). A causal inference model explains perception of the McGurk effect and other incongruent audiovisual speech. PLoS Computational Biology, 13(2), e1005229.PubMedPubMedCentralCrossRef
go back to reference Magnotti, J. F., Dzeda, K. B., Wegner-Clemens, K., Rennig, J., & Beauchamp, M. S. (2020). Weak observer–level correlation and strong stimulus-level correlation between the McGurk effect and audiovisual speech-in-noise: A causal inference explanation. Cortex, 133, 371–383.PubMedPubMedCentralCrossRef Magnotti, J. F., Dzeda, K. B., Wegner-Clemens, K., Rennig, J., & Beauchamp, M. S. (2020). Weak observer–level correlation and strong stimulus-level correlation between the McGurk effect and audiovisual speech-in-noise: A causal inference explanation. Cortex, 133, 371–383.PubMedPubMedCentralCrossRef
go back to reference Magnotti, J. F., Smith, K. B., Salinas, M., Mays, J., Zhu, L. L., & Beauchamp, M. S. (2018). A causal inference explanation for enhancement of multisensory integration by co-articulation. Scientific Reports, 8(1), 1–10.CrossRef Magnotti, J. F., Smith, K. B., Salinas, M., Mays, J., Zhu, L. L., & Beauchamp, M. S. (2018). A causal inference explanation for enhancement of multisensory integration by co-articulation. Scientific Reports, 8(1), 1–10.CrossRef
go back to reference Mason, M. F., Tatkow, E. P., & Macrae, C. N. (2005). The look of love: Gaze shifts and person perception. Psychological Science, 16, 236–239.PubMedCrossRef Mason, M. F., Tatkow, E. P., & Macrae, C. N. (2005). The look of love: Gaze shifts and person perception. Psychological Science, 16, 236–239.PubMedCrossRef
go back to reference Munhall, K. G., Ten Hove, M. W., Brammer, M., & Paré, M. (2009). Audiovisual integration of speech in a bistable illusion. Current Biology, 19(9), 735–739.PubMedCrossRef Munhall, K. G., Ten Hove, M. W., Brammer, M., & Paré, M. (2009). Audiovisual integration of speech in a bistable illusion. Current Biology, 19(9), 735–739.PubMedCrossRef
go back to reference Nichols, K. A., & Champness, B. G. (1971). Eye gaze and the GSR. Journal of Experimental Social Psychology, 7, 623–626.CrossRef Nichols, K. A., & Champness, B. G. (1971). Eye gaze and the GSR. Journal of Experimental Social Psychology, 7, 623–626.CrossRef
go back to reference Peirce, J., Gray, J. R., Simpson, S., MacAskill, M., Höchenberger, R., Sogo, H., Kastman, E., & Lindeløv, J. K. (2019). PsychoPy2: Experiments in behavior made easy. Behavior Research Methods, 51(1), 195–203.PubMedPubMedCentralCrossRef Peirce, J., Gray, J. R., Simpson, S., MacAskill, M., Höchenberger, R., Sogo, H., Kastman, E., & Lindeløv, J. K. (2019). PsychoPy2: Experiments in behavior made easy. Behavior Research Methods, 51(1), 195–203.PubMedPubMedCentralCrossRef
go back to reference Pfister, R., Schwarz, K. A., Janczyk, M., Dale, R., & Freeman, J. (2013). Good things peak in pairs: A note on the bimodality coefficient. Frontiers in Psychology, 4, 700.PubMedPubMedCentralCrossRef Pfister, R., Schwarz, K. A., Janczyk, M., Dale, R., & Freeman, J. (2013). Good things peak in pairs: A note on the bimodality coefficient. Frontiers in Psychology, 4, 700.PubMedPubMedCentralCrossRef
go back to reference Pönkänen, L. M., Alhoniemi, A., Leppänen, J. M., & Hietanen, J. K. (2011). Does it make a difference if I have an eye contact with you or with your picture? An ERP study. Social Cognitive and Affective Neuroscience, 6(4), 486–494.PubMedCrossRef Pönkänen, L. M., Alhoniemi, A., Leppänen, J. M., & Hietanen, J. K. (2011). Does it make a difference if I have an eye contact with you or with your picture? An ERP study. Social Cognitive and Affective Neuroscience, 6(4), 486–494.PubMedCrossRef
go back to reference Richardson, D. C., & Dale, R. (2005). Looking to understand: The coupling between speakers’ and listeners’ eye movements and its relationship to discourse comprehension. Cognitive Science, 29(6), 1045–1060.PubMedCrossRef Richardson, D. C., & Dale, R. (2005). Looking to understand: The coupling between speakers’ and listeners’ eye movements and its relationship to discourse comprehension. Cognitive Science, 29(6), 1045–1060.PubMedCrossRef
go back to reference Risko, E. F., Richardson, D. C., & Kingstone, A. (2016). Breaking the fourth wall of cognitive science: Real-world social attention and the dual function of gaze. Current Directions in Psychological Science, 25(1), 70–74.CrossRef Risko, E. F., Richardson, D. C., & Kingstone, A. (2016). Breaking the fourth wall of cognitive science: Real-world social attention and the dual function of gaze. Current Directions in Psychological Science, 25(1), 70–74.CrossRef
go back to reference Rosenblum, L. D. (2019). Audiovisual speech perception and the McGurk effect. Oxford Research Encyclopedia of Linguistics.CrossRef Rosenblum, L. D. (2019). Audiovisual speech perception and the McGurk effect. Oxford Research Encyclopedia of Linguistics.CrossRef
go back to reference Ross, L. A., Saint-Amour, D., Leavitt, V. M., Javitt, D. C., & Foxe, J. J. (2007). Do you see what I am saying? Exploring visual enhancement of speech comprehension in noisy environments. Cerebral Cortex, 17(5), 1147–1153.PubMedCrossRef Ross, L. A., Saint-Amour, D., Leavitt, V. M., Javitt, D. C., & Foxe, J. J. (2007). Do you see what I am saying? Exploring visual enhancement of speech comprehension in noisy environments. Cerebral Cortex, 17(5), 1147–1153.PubMedCrossRef
go back to reference Schilbach, L. (2015). Eye to eye, face to face and brain to brain: Novel approaches to study the behavioral dynamics and neural mechanisms of social interactions. Current Opinion in Behavioral Sciences, 3, 130–135.CrossRef Schilbach, L. (2015). Eye to eye, face to face and brain to brain: Novel approaches to study the behavioral dynamics and neural mechanisms of social interactions. Current Opinion in Behavioral Sciences, 3, 130–135.CrossRef
go back to reference Senju, A., & Hasegawa, T. (2005). Direct gaze captures visuospatial attention. Visual Cognition, 12(1), 127–144.CrossRef Senju, A., & Hasegawa, T. (2005). Direct gaze captures visuospatial attention. Visual Cognition, 12(1), 127–144.CrossRef
go back to reference Shams, L., Kamitani, Y., & Shimojo, S. (2000). What you see is what you hear. Nature, 408(6814), 788–788.PubMedCrossRef Shams, L., Kamitani, Y., & Shimojo, S. (2000). What you see is what you hear. Nature, 408(6814), 788–788.PubMedCrossRef
go back to reference Shams, L., Ma, W. J., & Beierholm, U. (2005). Sound-induced flash illusion as an optimal percept. NeuroReport, 16, 1923–1927.PubMedCrossRef Shams, L., Ma, W. J., & Beierholm, U. (2005). Sound-induced flash illusion as an optimal percept. NeuroReport, 16, 1923–1927.PubMedCrossRef
go back to reference Spence, C., & Frings, C. (2020). Multisensory feature integration in (and out) of the focus of spatial attention. Attention, Perception, & Psychophysics, 82(1), 363–376.CrossRef Spence, C., & Frings, C. (2020). Multisensory feature integration in (and out) of the focus of spatial attention. Attention, Perception, & Psychophysics, 82(1), 363–376.CrossRef
go back to reference Sperber, D., & Wilson, D. (1986). Relevance: Communication and cognition. Blackwell. Sperber, D., & Wilson, D. (1986). Relevance: Communication and cognition. Blackwell.
go back to reference Striano, T., & Reid, V. M. (2006). Social cognition in the first year. Trends in Cognitive Sciences, 10, 471–476.PubMedCrossRef Striano, T., & Reid, V. M. (2006). Social cognition in the first year. Trends in Cognitive Sciences, 10, 471–476.PubMedCrossRef
go back to reference Stropahl, M., Schellhardt, S., & Debener, S. (2017). McGurk stimuli for the investigation of multisensory integration in cochlear implant users: The Oldenburg Audio Visual Speech Stimuli (OLAVS). Psychonomic Bulletin & Review, 24(3), 863–872.CrossRef Stropahl, M., Schellhardt, S., & Debener, S. (2017). McGurk stimuli for the investigation of multisensory integration in cochlear implant users: The Oldenburg Audio Visual Speech Stimuli (OLAVS). Psychonomic Bulletin & Review, 24(3), 863–872.CrossRef
go back to reference Talsma, D., Doty, T. J., & Woldorff, M. G. (2007). Selective attention and audiovisual integration: Is attending to both modalities a prerequisite for early integration? Cerebral Cortex, 17(3), 679–690.PubMedCrossRef Talsma, D., Doty, T. J., & Woldorff, M. G. (2007). Selective attention and audiovisual integration: Is attending to both modalities a prerequisite for early integration? Cerebral Cortex, 17(3), 679–690.PubMedCrossRef
go back to reference Talsma, D., Senkowski, D., Soto-Faraco, S., & Woldorff, M. G. (2010). The multifaceted interplay between attention and multisensory integration. Trends in Cognitive Sciences, 14(9), 400–410.PubMedPubMedCentralCrossRef Talsma, D., Senkowski, D., Soto-Faraco, S., & Woldorff, M. G. (2010). The multifaceted interplay between attention and multisensory integration. Trends in Cognitive Sciences, 14(9), 400–410.PubMedPubMedCentralCrossRef
go back to reference Talsma, D., & Woldorff, M. G. (2005). Selective attention and multisensory integration: Multiple phases of effects on the evoked brain activity. Journal of Cognitive Neuroscience, 17(7), 1098–1114.PubMedCrossRef Talsma, D., & Woldorff, M. G. (2005). Selective attention and multisensory integration: Multiple phases of effects on the evoked brain activity. Journal of Cognitive Neuroscience, 17(7), 1098–1114.PubMedCrossRef
go back to reference Ten Oever, S., Romei, V., van Atteveldt, N., Soto-Faraco, S., Murray, M. M., & Matusz, P. J. (2016). The COGs (context, object, and goals) in multisensory processing. Experimental Brain Research, 234(5), 1307–1323.PubMedCrossRef Ten Oever, S., Romei, V., van Atteveldt, N., Soto-Faraco, S., Murray, M. M., & Matusz, P. J. (2016). The COGs (context, object, and goals) in multisensory processing. Experimental Brain Research, 234(5), 1307–1323.PubMedCrossRef
go back to reference Tiippana, K., Andersen, T. S., & Sams, M. (2004). Visual attention modulates audiovisual speech perception. European Journal of Cognitive Psychology, 16(3), 457–472.CrossRef Tiippana, K., Andersen, T. S., & Sams, M. (2004). Visual attention modulates audiovisual speech perception. European Journal of Cognitive Psychology, 16(3), 457–472.CrossRef
go back to reference van der Wel, R. P., Welsh, T., & Böckler, A. (2018). Talking heads or talking eyes? Effects of head orientation and sudden onset gaze cues on attention capture. Attention, Perception, & Psychophysics, 80(1), 1–6.CrossRef van der Wel, R. P., Welsh, T., & Böckler, A. (2018). Talking heads or talking eyes? Effects of head orientation and sudden onset gaze cues on attention capture. Attention, Perception, & Psychophysics, 80(1), 1–6.CrossRef
go back to reference van Engen, K. J., Xie, Z., & Chandrasekaran, B. (2017). Audiovisual sentence recognition not predicted by susceptibility to the McGurk effect. Attention, Perception, & Psychophysics, 79(2), 396–403.CrossRef van Engen, K. J., Xie, Z., & Chandrasekaran, B. (2017). Audiovisual sentence recognition not predicted by susceptibility to the McGurk effect. Attention, Perception, & Psychophysics, 79(2), 396–403.CrossRef
go back to reference von Grünau, M., & Anston, C. (1995). The detection of gaze direction: A stare-in-the-crowd effect. Perception, 24(11), 1297–1313.CrossRef von Grünau, M., & Anston, C. (1995). The detection of gaze direction: A stare-in-the-crowd effect. Perception, 24(11), 1297–1313.CrossRef
Metagegevens
Titel
When eyes beat lips: speaker gaze affects audiovisual integration in the McGurk illusion
Auteurs
Basil Wahn
Laura Schmitz
Alan Kingstone
Anne Böckler-Raettig
Publicatiedatum
02-12-2021
Uitgeverij
Springer Berlin Heidelberg
Gepubliceerd in
Psychological Research / Uitgave 6/2022
Print ISSN: 0340-0727
Elektronisch ISSN: 1430-2772
DOI
https://doi.org/10.1007/s00426-021-01618-y

Andere artikelen Uitgave 6/2022

Psychological Research 6/2022 Naar de uitgave