Verbally induced action control
In this section we summarize two research areas (i.e., implementation intentions and instruction implementation) that provide evidence that verbal information can influence subsequent action, potentially mediated by stimulus–response learning. Based on this evidence, we will then argue that verbal information about an action and an effect might also lead to action–effect learning.
The theory of implementation intentions suggests that behavior can be strategically controlled by forming a verbal plan in an
if–then format (Gollwitzer,
1999). According to this theory, if–then planning creates direct perception–action links between the anticipated situation (critical cue) and the intended behavior (action). For instance, after forming the plan “If I pass a supermarket, then I will buy fruit,” the situation (supermarket) serves as a critical cue that triggers the planned action (buying fruit). Empirical laboratory tests of this idea are similar to the previously described action–effect learning procedure, except that during the learning phase, participants form specific verbal if–then action plans instead of actually enacting responses. In one example of such a test (Cohen et al.,
2008, Exp. 2), participants memorized “If I hear the low tone on the left side, then I will press the right button especially fast.” In the test phase, participants were asked to perform a two-alternative forced-choice task (i.e., if the tone was high, they pressed the left button; if the tone was low, they pressed the right button). The results showed a response/verbal plan compatibility effect: required responses to the critical stimulus (if-part) were facilitated if they overlapped with the responses specified in the then-part of the plan. These and other similar results demonstrate that verbal (stimulus–response) planning influences subsequent behavioral responses (Cohen et al.,
2008; Martiny-Huenger et al.,
2017; Miles & Proctor,
2008).
Research from an instruction-based perspective provides additional evidence that instructions in the form of stimulus–response mappings can influence performance. The basic design of this type of research also involves a learning phase (verbal instructions) and a test phase. However, in many studies, the test phase is split into a
diagnostic task and an
inducer task (e.g., Liefooghe et al.,
2012)
. The given instructions are relevant for the inducer task but irrelevant for the diagnostic task. For instance, the instructions for the inducer task might read, “if you see ‘cat’, press left; if you see ‘dog’, press right.” However, before completing the inducer task, a preceding diagnostic task is introduced that shares both the stimuli (i.e., words ‘cat’ and ‘dog’) and responses (left/right button press) with the inducer task, but has different task instructions (e.g., to press the right or left button if the words are italicized or upright, respectively). Using this design, studies have demonstrated the presence of an instruction-based compatibility effect in the diagnostic task when the required response and the stimulus match the instructions given for the inducer task (e.g., when “cat” was italicized and required the left key response; for a review see Brass et al.,
2017).
One of the fundamental differences between implementation intentions and instruction implementation research is that critical if–then sentences in implementation intention research are strongly highlighted and repeated as a central, important sentence to encode and remember (reviewed by Gollwitzer & Sheeran,
2006). Instruction-based research does not include such emphasis on a single sentence. The critical “if–then” instructions are just a part of the typical task instructions (Liefooghe & De Houwer,
2018; Liefooghe et al.,
2012). Another central difference in these approaches is in the delay between reading the verbal plans/instructions and tests of their effects. If–then plans’ effects are tested minutes (in laboratory settings, e.g., Cohen et al.,
2008) or even days or weeks later (e.g., Conner & Higgins,
2010; Papies et al.,
2009). The effects of “instructions” are tested only seconds later (Brass et al.,
2017). These time differences are relevant to the present research, and we will continue to discuss them later. In general, however, the two approaches share many similarities. For example, the verbal information in both cases typically includes a stimulus–response contingency. Both imply that verbally presented stimulus–response (perception–action) contingencies influence subsequent behavior.
The present experiments
In the present research, we tested whether verbal action–effect instructions lead to associations between an action and an effect that are automatically activated upon perceiving the effect even if instructions and test are separated by more than a few seconds. We asked participants to memorize a specific verbal instruction that contained information about an action–effect relation (“To make the screen blue, I have to press the left key”). Afterward, participants performed a vowel-consonant categorization task. Although the task was unrelated to the action–effect instructions, responses in the categorization task overlapped with the responses specified in the action–effect sentence (i.e., left/right key). Importantly, on some trials, the screen background color turned blue (i.e., effect). This aspect was irrelevant for the categorization task and participants were instructed to ignore it. However, the presented blue background visually primed the effect from the action–effect instructions. We hypothesized that the priming of the effect would result in facilitated action–effect-compatible responses (i.e., categorization responses that align with the action–effect instructions) and/or in impaired incompatible responses (i.e., categorization responses that are different from those specified in the action–effect instructions).
While conceptually related to Theeuwes et. al. (
2015), our present studies go beyond their evidence that action–effect instructions influence subsequent actions. We separated the processing of the instructions from the performance in the diagnostic task. To do this, we presented one action–effect instruction at the beginning of the experiment instead of continuously updating the instructions every 4, 6, or 16 trials. Thus, whereas Theeuwes et al. observed effects of instructions that participants read a few seconds earlier, we tested the effects of a single action–effect instruction presented to the participants a few minutes earlier (before reading other information like the categorization task instructions). Second, in the case of Theeuwes et. al. (
2015), participants continuously performed inducer-task trials in the test phase, where the action–effect instructions were relevant after every 4, 6, or 16 trials. In the present work, participants were also told that the verbal action–effect instructions would be relevant at some point during the experiment. However, this information served only as a cover story and the participants never actually had to implement the instructions.
In sum, the effects of instructions on subsequent responses in Theeuwes et. al. (
2015) were observed with instructions processed only seconds prior to testing their effects and in a context, where the participants were aware that the instructions were relevant just a few seconds later. In contrast, we tested effects with a longer time interval and in a context, where the instructions never had to be implemented and thus there were no explicit reminders of the action–effect instructions during the test phase. We conducted two online experiments. In the first experiment, we tested verbal action–effect instructions in an
effect–
action order. The central focus of the second experiment was to provide a direct replication of Experiment 1 with an increased sample size. In addition, we added an exploratory part in which we reversed the order of the instructions (
action–
effect order).
Experiment 2
The first experiment provides initial evidence that the action–effect sentence influenced subsequent performance in response to the priming of the effect. However, we found this evidence only in response errors. Furthermore, although the separate analyses of the critical and control trials provided a clear picture, the overall three-way interaction effect was only marginally significant. A sensitivity analysis of the Experiment 1 data suggests that a mixed-design ANOVA with 41 participants across four within-conditions within two groups would be sensitive to an effect of ηp2 = 0.21 with 80% power (α = 0.05). Given that the observed effect size was ηp2 = 0.08, we conclude that the first experiment was underpowered. Therefore, we conducted a second experiment with the central focus of providing a higher powered exact replication of Experiment 1. If the result pattern found in Experiment 1 was due to chance, it is unlikely that a second, higher powered, independent replication would produce such a specific pattern again.
In addition to the central aim of replicating Experiment 1, we added an exploratory examination of an action–effect sentence formulated in a more typical
action–
effect order (“
I will press the left key to make the screen blue”). This formulation of the instruction reflects the theoretical assumption of the action–effect principle that the associations resulting from action–effect learning are bidirectional; even if learning occurs in an action–then–effect order, encountering the effect first should trigger the response (e.g., Elsner & Hommel,
2001). Thus, Experiment 2 includes one part that is an exact replication of Experiment 1 with the effect (stimulus)–action (response) order format. Our hypotheses for this replication were the same as in Experiment 1: required responses that are incompatible (compatible) with the verbally linked, primed effect should be impaired (facilitated). The exploratory second part differed only in the order of the components (i.e., action [response]–effect [stimulus]). We had no specific hypotheses for this exploratory analysis. Whereas verbal if (stimulus)–then (response) planning research represents the order of presenting the verbal information as relevant, prior verbal action–effect studies have also found significant effects with an action (response)–effect (stimulus) order (Theeuwes et al.,
2015). Whereas Experiment 1 participants consisted mainly of students from Norway (mean age 24.1), Experiment 2 participants were recruited from the general population of the United Kingdom (mean age 41.4).
General discussion
In the present experiments we examined whether verbal action–effect instructions led to associations between perception (effect) and action that are automatically (i.e., unintentionally) activated upon encountering the effect. We tested this activation in behavioral responses in a speeded categorization task, where the effect was included as a task-irrelevant prime. Although some of the main findings were only marginally significant, the two experiments in combination revealed consistent evidence that the action–effect instructions (in an effect–action order) in combination with the effect prime influenced the accuracy of participants’ responses (with no evidence of a speed–accuracy trade-off). If the action effect prime was present, required responses that were incompatible with the instructed response showed more errors than when the responses were compatible with the previous action–effect instructions. Whereas Experiment 1 was underpowered, the replication in Experiment 2 (with four times the sample size) supported the results from Experiment 1. Why this increased sample size did not result in a clearer effect may be explained by the sample characteristics. There may have been increased random error variance from the significantly older, non-student sample in Experiment 2.
The result patterns could be interpreted as showing an interference effect in the effect-prime trials in which the previously verbally linked response was incompatible with the required response in the respective trial. However, facilitation from compatible response activation or interference from incompatible response activation can only be evaluated in comparison to an adequately similar control condition. The control condition in the present studies differed in terms of the critical priming factor (i.e., it did not include distracting sudden background-color changes). Assuming that the background color change negatively influenced responses in the prime/color-change trials, the absolute differences between critical and neutral trials are not comparable as we cannot estimate the size of that negative influence of the prime (i.e., prime main effect). Depending on the size of the prime/color-change induced interference, all combinations—only facilitation, facilitation and interference, or only interference—are possible. Investigating this would require a control condition that includes the same background-color change without including any (verbal) links of that color to a response. In such a condition, we could observe the consequences for responses induced merely by the sudden background-color change. Importantly, however, this limitation of not knowing whether facilitation, interference, or both caused the effect, does not reduce the informative value of the observed interaction effects, indicating that the verbal information systematically influenced the responses.
The absence of the hypothesized interaction effect in reaction times maybe explained by the response deadline. Response deadlines (i.e., forcing participants to emphasize speed over accuracy) typically leads to a reduced variability in response times and diminished power to detect reaction time effects (for a similar argument and findings in accuracy vs. reaction time measures, see Mekawi & Bresin,
2015). In sum, for the effect–action order formulation, we provide evidence that the verbally formulated perception–action relation—that was never directly experienced or executed—resulted in an association that was automatically reactivated upon perceiving the effect.
Our results align with previous research showing that imagining an effect while actually performing a response can lead to action–effect bindings (Cochrane & Milliken,
2019; Pfister et al.,
2014). However, in the present research, participants did not previously experience the effect or response, but processed them merely as verbal action–effect instructions.
Eder and Dignath (
2017) also showed action–effect learning from verbal instructions. However, in their test phase, participants experienced the previously instructed action–effect associations with each response. Therefore, it is not clear whether the observed effects were the direct effect of the instructions or some conflict between the instructions and the instruction-incompatible experiences. In our present experiments, participants never directly experienced the previously instructed action–effect contingency in the test phase. Thus, our study focused more narrowly on response priming from an instructed, verbal action–effect contingency. Finally, in contrast to the previously introduced research by Theeuwes et. al. (
2015) in which instructions were likely to be kept in working memory (i.e., with responses given within a short interval after instructions were given), the present results indicate that the impact of instructions can have a longer lasting effect (beyond seconds and with processing other information in between), in line with the findings from implementation intention studies (Gollwitzer,
2014; Webb & Sheeran,
2008).
Martiny-Huenger et. al. () suggested a possible mechanism for this effect. According to their theoretical framework, verbal instructions that include a perceivable effect and executable action may work similarly to associative learning from direct processing and execution of the perception and action. This idea is based on theories of simulation and embodied cognition (Barsalou,
1999,
2010; Hesslow,
2012) and past findings that language comprehension of concrete concepts overlaps with sensorimotor areas activity of the brain (e.g., Arbib,
2008; Gallese & Lakoff,
2005; Pulvermüller & Fadiga,
2010). From this perspective, comprehension of verbal information activates the same sensorimotor brain areas that are involved during actual perception and behavior. Verbally processing a stimulus–response or action–effect contingency can thus result in the formation of specific associations between them—associations that are unintentionally activated upon encountering the perception (e.g., visual action effect) as suggested by our present experiments.
Studies on action–effect learning from direct experiences usually appear to form bi-directional links between action and effect, because the learning order (action, then effect) is reversed in the test phase (effect presentation, then action, e.g., Elsner & Hommel,
2001) However, the results from the second experiment indicated that the effect of the instruction sentence was only observed in the condition when the action–effect sentence was formulated in an effect–action direction (i.e., perception, then action: “
To make the screen blue, I will press the left key”). In the action–effect order (action, then perception: “
I will press the left key to make the screen blue”), the effect of the instructed sentence was not observed. These findings are not in line with the results of Theeuwes et. al. (
2015), who only used the action–effect order and found effects of these instructions. If the present results prove to be robust in subsequent replications, a potential explanation could be in the differences of the procedure. Participants in the studies by Theeuwes et. al. (
2015) were more likely to have kept the action–effect relation active in working memory. Thus, the order of the relation may be less important when the components are active in working memory. However, with the delay between processing the verbal instructions and executing the responses, whatever memory processes mediated the effects (e.g., associative learning), they may be sensitive to the order in which the components were processed before.
The statistically weak evidence for a difference between the two instruction component orders prohibits us from drawing strong conclusions about potential differences between the order of processing the action–effect components. However, our results are in line with a previous study by McCrea et. al. (
2014), who investigated the consequences of differently formulated self-regulation instructions before doing a prospective memory task. Although the authors modeled instructions to fit different theoretical concepts, one of the instructions included a stimulus–response order that was similar to our effect–action order (“Whenever I see the red circle, then I will immediately press the spacebar”). The other two formulations included a response–stimulus order (e.g., “I will immediately press the spacebar when I see the red circle!”) similar to our action–effect order. Like our findings, only the stimulus–response order (i.e., perception–action) was effective in their study (McCrea et al.,
2014). More anecdotally, in the initial publications of if–then planning research, the strategy was sometimes presented in a response–stimulus format (e.g., “I intend to do y when situation z is encountered”; Gollwitzer,
1993; Gollwitzer & Brandstätter,
1997). However, at some point, this changed, and subsequent publications almost exclusively used the if (stimulus)–then (response) order (e.g., “When situation x arises, I will perform response y!”; Gollwitzer,
1999). This could have been the result of a mere refinement of the concept, or as a result of practical experience that the reversed order (response–situation) is less effective.
Why might the perception–then–action order be more effective than the action–then–perception order (at least in measures after a few seconds)? Disregarding the rich subjective experiences that we associate with language in general and discussing it from the perspective of simulation accounts of cognition alone might provide an interesting answer. As argued previously, repeating the presented instructions in the presented form may act as a placeholder for the actual experiences. From this actual-experience perspective and the fact that reading is sequential—the order of the components in the instructions results in differences in whether the perception (e.g., effect) is predictive of a response or not. In our effect–action order and McCrea et. al.’s (
2014) stimulus–response order, the perception (effect/stimulus) is followed by the response; the perception part is thus predictive of the action part. During the test phase, the perception is there first (effect prime, blue screen) and the perception, therefore, biased actual responses in line with the prior learning. In contrast, in the action–effect order and McCrea et. al.’s (
2014) response–stimulus order, the perception of the effect/stimulus was not predictive for the action, because in this case, the action preceded the perception. Thus, when the perception occurred in the test phase, it did not have any systematic predictive value and thus did not bias the subsequent responses.
Whereas the evidence we present for such an order effect in the present research is weak, it lines up with other prior evidence (e.g., McCrea et al.,
2014; if–then planning research in general). Furthermore, where it conflicts with prior evidence (e.g., Theeuwes et al.,
2015), it can easily be reconciled with differences in the procedures (i.e., instructions kept in working memory for a few seconds vs. effects that could not have been kept in working memory). More research is needed to support the reliability of a systematic difference between the component order. In addition to the new theoretical questions about action–perception learning raised by these findings, the present study contributes to the idea that language is intertwined with action control (Perlovsky & Sakai,
2014) and can be strategically used to control our behavior (Gollwitzer,
1999; Martiny-Huenger et al.,
2017).
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.