Despite the variability in physical appearance of a written word (e.g., house, HOUSE, house), skilled readers are able to access the appropriate lexical entry in a few hundreds of milliseconds. When and how the stimulus features are coded in an abstract manner are the main questions in this article. These are not trivial issues, since this process of abstraction might take place at an individual-letter level, at a graphemic level, at a whole-word level, or even at a semantic level. To make matters even more complex, feedforward and feedback connections might also exist among the levels.

There is some consensus among researchers that a word’s constituent letters are coded in an abstract form that is independent of its physical features (see Bowers, 2000, and Thompson, 2009, for reviews). Although initially the word-processing system is sensitive to differences in the visual features of stimuli, these differences are quickly diffused by mapping these features onto unique orthographic abstract representations (see Dehaene, Cohen, Sigman, & Vinckier, 2005, and Grainger, Rey, & Dufau, 2008, for neurally motivated accounts of this phenomenon). The assumption that abstract representations are rapidly accessible during visual-word recognition has come mainly from previous experimental evidence using the masked-priming technique (Forster & Davis, 1984; see Grainger, 2008, for a review; see Dehaene et al., 2001, 2004, for the neuroanatomic signature of masked priming). In the present experiment, we aimed to analyze the time course of the abstract orthographic level of representation (independent of visual feature mapping) through a matched- versus mismatched-case masked-priming experiment with words and pseudowords.Footnote 1

In an elegant masked-priming lexical decision experiment, Jacobs, Grainger, and Ferrand (1995) compared matched-case versus mismatched-case identity word pairs (YEUX–####–YEUX vs. yeux–####–YEUX [eyes]) and found that, despite the greater visual similarity for the matched-case identity pairs, word identification times were virtually the same in the two conditions. Perea, Jiménez, and Gómez (2014) replicated the Jacobs et al. (1995) pattern in words with high degree of cross-case visual similarity (e.g., city–###–CITY = CITY–###–CITY) and in words with low degree of cross-case visual similarity (e.g., edge–###–EDGE = EDGE–###–EDGE; see also Perea, Jiménez, & Gomez, 2015, for evidence with developing readers). Furthermore, the magnitude of the repetition priming effect (i.e., identity condition vs. unrelated condition) was similar in both types of stimuli, thus replicating Bowers, Vigliocco, and Haan (1998). Taken together, these findings are consistent with the accounts that assume that there is fast access to abstract representations. In order to disentangle whether the functional and structural specialization on abstract orthographic coding begins at a letter or at a lexical level of processing, it is relevant to examine whether this pattern of results holds for pseudowords (i.e., orthographic codes with no lexical entry).

Importantly, the response times (RTs) for pseudowords were shorter in the matched-case identity PRIME–TARGET condition than in the mismatched-case identity prime–TARGET condition (e.g., CURDE–####–CURDE faster than curde–#####–CURDE; Jacobs et al., 1995; see Perea et al., 2014, for parallel evidence in French, Spanish, and English). To explain this dissociation, Perea et al. (2014) argued that, for words, fast access to abstract lexical representations makes the visual dissimilarity between prime and target irrelevant for further processes. In contrast, for pseudowords, there is no abstract lexical representation that could cancel out the facilitative effect of case-specific identity priming. This means that the null differences between matched- versus mismatched-case identity priming for words occur as a consequence of top-down effects from the lexical to the orthographic/letter level, thus posing problems for purely feedforward accounts of letter/word processing (e.g., the Bayesian reader model; Norris & Kinoshita, 2008). Indeed, interactive models of visual-word recognition assume that visual information continuously cascades throughout the entire orthographic–phonological–lexical–semantic network (see Carreiras, Armstrong, Perea, & Frost, 2014, for a recent review), and the interaction between different levels of processing is what consolidates the orthographic units of the target word.

To tease apart the different components during lexical processing, an excellent approach is to collect event-related potentials (ERPs). Previous research using masked repetition priming in combination with ERPs has defined a series of components during visual-word recognition (see Grainger & Holcomb, 2009, for a review). Two of these ERP components are of specific interest for our present study: the N/P150 and the N250. First, the N/P150 is a bipolar ERP component peaking between 80 and 150 ms approximately. It reaches positive (P) values over more anterior sites, and negative (N) values on more occipital sites. Importantly, its amplitude is modulated by physical-feature overlap across the prime and target, which may reflect processing at the level of size-invariant visual features (Holcomb & Grainger, 2006). Second, the N250 is an ERP component that starts around 150 ms and peaks around 300 ms. This component shows a widespread scalp distribution (larger over anterior scalp areas), and its amplitude is a function of the prime–target orthographic overlap (e.g., larger for porch–TABLE than for teble–TABLE, which in turn is larger than for table–TABLE; HoIcomb & Grainger, 2006; Kiyonaga, Grainger, Midgley, & Holcomb, 2007). Of particular relevance here are the findings of Chauncey, Holcomb, and Grainger (2008) in a semantic categorization task with word targets, in which they manipulated changes in size (e.g., table–table) and font (e.g., table–table) between the prime and target. ERP effects were restricted to an N/P150 amplitude modulation by the font manipulation, whereas null effects were observed in the N250 component. Hence, when the N250 component appears, size and shape invariance have already been achieved, and this can be interpreted as further evidence of access to abstract orthographic representations. Recent research has shown that the N250 amplitude can also be modulated by lexical factors (Duñabeitia, Molinaro, Laka, Estévez, & Carreiras, 2009; Massol, Midgley, Holcomb, & Grainger, 2011; Morris, Franck, Grainger, & Holcomb, 2007), reflecting the interaction between prelexical bottom-up and lexical-semantic top-down representations (Morris et al., 2007).

Taken together, there is substantial evidence for early effects of physical feature processing leading to later abstract representations. There are still, however, unanswered questions related to the degree of feedback in the system. The main aim of the present experiment was to examine how abstract orthographic representations are computed during visual-word recognition. In particular, we tracked the time course of the electrophysiological response to words and pseudowords preceded by an identity prime that either matched or mismatched in case (e.g., words: ALTAR–ALTAR vs. altar–ALTAR; pseudowords: CURDE–CURDE vs. curde–CURDE). To avoid the differential impact of any low-level characteristics of the letters that composed the words and pseudowords (e.g., in terms of cross-case visual overlap; cf. a–A vs. u–U), each letter occurred approximately the same number of times in words and pseudowords.Footnote 2 For consistency with previous research on this issue (e.g., Jacobs et al., 1995; Perea et al., 2014), we also included unrelated words and pseudowords as primes—half of the unrelated primes in lowercase and the other half in uppercase.

At early orthographical encoding stages, no difference in the ERP waves should be obtained between words and pseudowords regarding the case-match identity manipulation: Any difference should emerge as a consequence of top-down effects from the lexical to the abstract orthographic level of encoding. Therefore, the latency of this dissociation would reveal feedback from lexical to orthographic processing. Specifically, we predicted an N/P150 effect of case-match identity masked priming for both the word and pseudoword targets, since the visual features of prime and target are different in the mismatched-case identity condition (as compared to the matched-case condition). This effect was expected to disappear/attenuate as soon as shape invariance was achieved during orthographic processing (around 200–300 ms poststimulus: N250: Carreiras, Perea, Gil-López, Abu Mallouh, & Salillas, 2013; Chauncey et al., 2008; Petit, Midgley, Holcomb & Grainger, 2006). That is, if prelexical orthographic representations are tapped into relatively early in processing, then N/P150 differences due to prime–target visual dissimilarity would be expected to dissipate in later epochs (N250) not only for words, but also for pseudowords. However, if the early difference due to visual dissimilarity does not dissipate for pseudowords, as would be inferred from previous behavioral evidence (Jacobs et al., 1995; Perea et al., 2014), then this would imply that the N250 component reflects lexical rather than (abstract) orthographic effects. Importantly, a word–pseudoword dissociation in the N250 component would also imply that the orthographic (abstract) encoding of words is modulated by lexical factors, thus posing strong problems for purely feedforward accounts of letter–word processing.

Method

Participants

A group of 24 undergraduate students of the University of Valencia (14 women, 10 men) participated in the experiment in exchange for course credit or for a small gift. All of them were native Spanish speakers with no history of neurological or psychiatric impairment, and with normal (or corrected-to-normal) vision. Their ages ranged from 19 to 30 years (mean = 22.5 years, SD = 3.4). All participants were right-handed, as assessed with a Spanish abridged version of the Edinburgh Handedness Inventory (Oldfield, 1971). The data from two participants were discarded because of noisy electroencephalogram (EEG) data.

Materials

We selected a set of 160 five-letter words from the B-Pal Spanish database (Davis & Perea, 2005). The mean frequency per million was 18.6 (range: 10–37.7), and the mean number of orthographic neighbors was 1.9 (range: 0–4). A matched set of 160 pseudowords was created using the Wuggy package (Keuleers & Brysbaert, 2010). The set of words and pseudowords was carefully matched in terms of the distributional properties of the letters [i.e., a given letter (a, b, c, etc.) occurred equally frequently in the experimental set of words and pseudowords: χ 2(22) = 4.25, p > .95]. The list of words/pseudowords is presented in Appendix A. All targets (words or pseudowords) were presented in uppercase and were preceded by a prime that was (i) the same as the target, including the case (matched-case identity condition; e.g., ALTAR–ALTAR); (ii) the same as the target, but in a different case (mismatched-case identity condition; e.g., altar–ALTAR); (iii) an unrelated word prime (half in lowercase, half in uppercase); and (iv) an unrelated pseudoword prime (half in lowercase, half in uppercase). Four counterbalanced lists were created so that each target stimulus was rotated across the different conditions.

Procedure

Participants were seated comfortably in a dimly lit and sound-attenuated chamber. All stimuli were presented on a high-resolution monitor that was positioned at eye level one meter in front of the participant. The stimuli were displayed in white lowercase Courier 24-pt font against a dark-gray background. Participants performed a lexical decision task: they had to decide as accurately and rapidly as possible whether or not the stimulus was a Spanish word. They pressed one of two response buttons (YES/NO). The hand used for each type of response was counterbalanced across subjects. RTs were measured from target onset until the participant’s response.

The sequence of events in each trial was as follows: A forward pattern mask (i.e., a series of #s) of 500-ms duration preceded the prime stimulus, which was shown for 33 ms (i.e., two refresh cycles at 60 Hz), which in turn was replaced by a 16.7-ms pattern mask (i.e., one refresh cycle). Then, the target stimulus was presented in the same spatial location as the prime until the participant responded or 1,500 ms had elapsed. A blank screen of random duration (range: 700–1,000 ms) was presented after the response (see Fig. 1).

Fig. 1
figure 1

Depiction of events within a trial

Sixteen practice trials preceded the experimental phase, and brief, 10-s breaks occurred every 60 trials. Every 120 trials, there was a brief pause for resting and impedance checking. To minimize participant-generated artifacts in the EEG signal during the presentation of the experimental stimuli, participants were asked to refrain from blinking and making eye movements from the onset of each trial to the response. Each participant received the stimuli in a different random order. The whole experimental session lasted approximately 45 min.

EEG recording and analyses

The EEG was recorded from 32 electrodes mounted in an elastic cap, and referenced to the right mastoid. The EEG recording was re-referenced offline to an average of the left and right mastoids. Impedances were kept below 5 kΩ. The EEG signal was band-pass filtered between 0.01 and 20 Hz and sampled at 250 Hz. All single-trial waveforms were screened offline for amplifier blocking, drift, muscle artifacts, eye movements, and blinks. This was done for a 550-ms epoch with a 150-ms prestimulus baseline. Trials containing artifacts and/or trials with incorrect lexical decision responses were not included in the average ERPs or in the statistical analyses. These processes led to an average rejection rate of 12 % of all trials (8.3 % due to artifact rejection; 3.7 % due to incorrect responses). An ANOVA on the number of included trials per condition showed no difference between conditions (Fs < 1). Importantly, at least 30 trials were included for each condition in the average ERP data from each participant. ERPs were averaged separately for each of the experimental conditions, each of the participants, and each of the electrode sites.

Statistical analyses were performed on the mean ERP values in three contiguous time windows (N/P150: 80–150 ms; N250: 250–350 ms; N400: 400–500 ms). This was done for the four experimental conditions defined by the combination of the factors Lexicality (words, pseudowords) and Case (matched, mismatched). The selection of these epochs was motivated by our aim to track the time course of the potential differences between experimental conditions, and was determined by visual inspection and on the basis of previous studies (Chauncey et al., 2008; Grainger & Holcomb, 2009; Holcomb & Grainger, 2006). Following a similar strategy in the related literature (see, e.g., Vergara-Martínez, Perea, Gómez, & Swaab, 2013; Vergara-Martínez & Swaab 2011), we analyzed the topographical distribution of the ERP results by including the averaged amplitude values across three electrodes of four representative scalp areas that resulted from the factorial combination of the factors Hemisphere (left, right) and Anterior–Posterior (AP) Distribution (anterior, posterior): left anterior (F3, FC5, FC1), left posterior (CP5, CP1, P3), right anterior (F4, FC2, FC6), and right posterior (CP2, CP6, P4) (see Fig. 2).

Fig. 2
figure 2

Schematic representation of the electrode montage

This strategy was applied in each ERP analysis of the present experiment. For each time window, a separate repeated measures analysis of variance (ANOVA) was performed, including the factors Hemisphere, AP Distribution, Lexicality, and Case. In all analyses, List (1–4) was included as a between-subjects factor in order to extract the variance that was due to the counterbalanced lists (Pollatsek & Well, 1995). Main effects of lexicality are reported when they are relevant for the interpretation of the results. Effects of the AP Distribution or Hemisphere factor are reported when they interact with the experimental manipulations. Interactions between factors were followed up with simple-effect tests.

Results

Behavioral results

Incorrect responses and lexical decision times less than 250 ms or greater than 1,500 ms were excluded from the latency analyses. The mean lexical decision times and the error rates per condition are displayed in Table 1. The mean lexical decision times and percentages of errors were submitted to separate ANOVAs with a 2 (Lexicality: word, pseudoword) × 2 (Case: matched, mismatched) design. As we indicated above, List was also included in the analyses as a dummy factor in the design (four levels: Lists 1–4). In the behavioral analyses, we computed the F ratios by participants (F 1) and by items (F 2). We examined the usual “repetition-priming” effect (i.e., identity vs. unrelated priming conditions), with results presented in Appendix B (see Perea et al., 2014, for a similar approach). To summarize the findings reported in the appendix, we replicated the same pattern from the earlier experiments: We found a substantial repetition-priming effect for words, and a rather weak effect for pseudowords, at both the behavioral level (see Jacobs et al., 1995; Perea et al., 2014) and the electrophysiological level (e.g., see Kiyonaga et al., 2007, for smaller masked repetition effects in pseudowords than in words on the N250 component, and for null effects on subsequent time windows for pseudoword relative to word stimuli; see also Figs. 6b and 8a in Grainger & Holcomb, 2009).

Table 1 Mean lexical decision times (RTs, in milliseconds) and percentages of errors (ERs) for word and pseudoword targets in the different priming conditions (matched vs. mismatched case, identity vs. unrelated)

The ANOVA on the latency data revealed that, unsurprisingly, words were responded to faster than the pseudowords (579 vs. 682 ms, respectively), F 1(1, 18) = 184.07, p < .001; F 2(1, 312) = 480.23, p < .001. In addition, the target stimuli were responded to faster when they were preceded by a matched-case identity prime rather than a mismatched-case identity prime (622 vs. 639 ms), F 1(1, 18) = 7.56, p = .013; F 2(1, 312) = 13.2, p < .001. More importantly, the effect of case was qualified by a significant interaction between the two factors, F 1(1, 18) = 7.26, p = .015; F 2(1, 312) = 7.44, p = .007. This interaction reflected that, for pseudoword targets, responses were, on average, 27 ms faster when the pseudoword was preceded by a matched-case rather than a mismatched-case identity prime, F 1(1, 18) = 10.41, p = .005; F 2(1, 156) = 18.01, p < .001, whereas, for word targets, no signs of a parallel effect were apparent (i.e., a nonsignificant 6-ms difference, both Fs < 1).

The ANOVA on the error data failed to reveal any main effects of lexicality or case (all Fs < 1), but we did find a significant interaction between the two factors, F 1(1, 18) = 11.3, p = .003; F 2(1, 312) = 14.82, p < .001. This interaction reflected that, for pseudoword targets, participants made more errors when they were preceded by a mismatched-case identity prime than when they were preceded by a matched-case identity prime (5.1 % vs. 3.1 %, respectively), F 1(1, 18) = 2.64, p = .12; F 2(1, 156) = 4.56, p = .034. For word targets, participants made fewer errors when the words were preceded by a mismatched-case identity prime than when they were preceded by a matched-case identity prime (2.0 % vs. 4.9 %, respectively; this corresponds to 0.8 vs. 1.9 errors per participant), F 1(1, 18) = 22.30, p < .001; F 2(1, 156) = 11.05, p = .001.Footnote 3

Taken together, the behavioral data replicated previous research (Jacobs et al., 1995; Perea et al., 2014): For pseudoword targets, lexical decision times were shorter when the target was preceded by a matched-case identity prime than when it was preceded by a mismatched-case identity prime (e.g., CURDE–CURDE < curde–CURDE), whereas for word targets, the case of the identity prime did not matter (e.g., ALTAR–ALTAR = altar–ALTAR). Note, however, that a small speed–accuracy trade-off occurred in the word data: Participants were slightly faster (6 ms) when the prime had the same case as the target, but they also made more errors (1.1, on average). This marginal trade-off is likely to be an empirical anomaly, since Perea et al. (2014, 2015) reported a null effect for word targets in the RTs and accuracy for these same conditions.

ERP results

Figure 3 shows the ERP waves for words and pseudowords (matched vs. mismatched case) in four representative electrodes (the occipital electrodes are also displayed, to show the bipolar nature of the N/P150). The ERPs in the target epoch produced an initial small negative potential peaking around 50 ms, which was followed by a much larger and slower positivity (P2) ranging between 100 and 250 ms (see the anterior locations; note that the opposite pattern is shown in the occipital electrodes). At posterior sites, in this same time frame, the positivity is smaller and faster, peaks earlier (before 150 ms), and is followed by two subsequent negativities (at 180 and 250 ms, approx.). Following these early potentials, a large and slow negativity peaking around 350 ms can be seen at both anterior and posterior areas (N400). Following the N400 component, the waves remain positive until the end of the epoch (550 ms).

Fig. 3
figure 3

Grand average event-related potentials to words and pseudowords in the two matched-case conditions (see legend), in four representative electrodes from the four areas of interest. Electrodes O1 and O2 are presented at the bottom to show the mirror nature of the N/P150 in different areas. The bar chart represents the effect of case (matched minus mismatch) in the left anterior area of interest

The first ERP component to show different amplitudes is the N/P150, a positive potential (over anterior sites) reaching its maximum at around 150 ms poststimulus: The matched-case condition showed larger positive values than the mismatched-case condition (for both words and pseudowords). Case effects were also observed between 150 and 200 ms. Importantly, in the following time epochs, case effects vanished for the word stimuli, whereas they were sizeable for pseudowords (around 250 ms and also shown in the N400): The mismatched-case identity condition showed larger negative values than the matched-case identity condition. To capture the dissociation between words and pseudowords regarding the effect of case, we conducted an ANOVA that included the factors Time Epoch (80–150, 250–350, and 400–500 ms), Case (matched, mismatched), Lexicality (words, pseudowords), Hemisphere (left, right), and AP Distribution (anterior, posterior). This initial analysis revealed an interaction of time epoch, case, and lexicality: F(2, 36) = 3.87, p = .045. We then conducted separate ANOVAs for each time epoch. The results of the ANOVAs for each epoch (80–150, 250–350, and 400–500 ms) are shown below.

80- to 150-ms epoch

The ANOVA revealed a main effect of case [F(1, 18) = 4.6, p < .05] that was modulated by an interaction with AP distribution [F(1, 18) = 22.56, p < .001]: Larger positive values were observed for the mismatched- over the matched-case identity condition over frontal areas of the scalp [F(1, 18) = 13.40, p < .005; posterior areas: F < 1]. No differences were apparent between words and pseudowords (F < 1).Footnote 4

250- to 350-ms epoch

The ANOVA showed a main effect of lexicality [F(1, 18) = 17.1, p < .001]: Larger negativities were observed for pseudowords than for words. The interaction between lexicality and AP distribution [F(1, 18) = 10.29, p < .01] revealed that the lexicality effect was much larger over frontal (1.4 μV) [F(1, 18) = 20,14, p < .01] than over posterior (0.75 μV) [F(1, 18) = 9.4, p < .01] scalp areas. In addition, we found an interaction between case and lexicality [F(1, 18) = 14.96, p < .001]: Larger negativities were observed for the mismatched- than for the matched-case identity condition for pseudowords [F(1, 18) = 7.6, p < .02], but not for words (F < 1). Again, the interaction between case and AP distribution [F(1, 18) = 9.29, p < .01] revealed that the effect of case was located over frontal scalp areas [F(1, 18) = 4.01, p = .06; posterior: F < 1].

400- to 500-ms epoch

The ANOVA revealed larger negativities for pseudowords than for words [main effect of lexicality: F(1, 18) = 62.4, p < .001], as well as an interaction between lexicality and AP distribution [F(1, 18) = 17.90, p < .01]: In contrast to the previous epoch, the lexicality effect was larger over posterior (3.8 μV) [F(1, 18) = 77.97, p < .001] than over anterior (2.7 μV) [F(1, 18) = 39.16, p < .001] scalp areas. As in the previous time epoch, the interaction between case and lexicality was significant [F(1, 18) = 10.61, p < .005]: Larger negativities were observed for the mismatched- than for the matched-case condition for pseudowords [F(1, 18) = 13.14, p < .005], but not for words [F(1, 18) = 1.8, p = .19].

Discussion

In the present ERP experiment, we examined the temporal course of the effect of case (lowercase vs. uppercase) in a masked-priming paradigm with matched-case PRIME–TARGET identity pairs and mismatched-case prime–TARGET identity pairs. The behavioral data and the ERP data provided converging evidence for the dissociating role of lexicality in this effect: The matched-case identity condition showed an advantage over the mismatched-case identity condition for pseudoword targets (e.g., CURDE–CURDE faster than curde–CURDE), but not for word targets (e.g., similar RTs to ALTAR–ALTAR and altar–ALTAR; see also Jacobs et al., 1995; Perea et al., 2014). The ERP data revealed case-match effects at the earliest time epoch of analysis (N/P150: 80–150 ms). The N/P150 is driven by featural overlap across the prime and target, and it has been suggested to reflect processing at the level of size-invariant visual features (see Petit et al., 2006). At this very early processing stage, visual features are subject to a low-level analysis resulting in differences for upper- and lowercase versions of the same word or pseudoword (note that in the matched-case condition, the prime and target stimuli fully overlap in their physical features). Importantly, this takes places at a prelexical level of processing, and consistent with this interpretation, the lexicality of the letter string did not have an impact on the mapping of features onto case-specific letter representations. We would point out, however, that the effects were slightly larger for pseudowords than for words (see Fig. 4); thus, it is possible that even at this early stage of processing, there are some subtle modulations from higher-order lexical information, as has been previously reported by Hauk et al. (2006; see also Assadollahi & Pulvermüller, 2003; Dambacher, Kliegl, Hofmann, & Jacobs, 2006). The effect of case match lasted until approximately 200 ms, as revealed by a complementary analysis performed on a following time epoch [150–200 ms: F(1, 18) = 6.00, p < .05].

Fig. 4
figure 4

Topographic distributions of the case effect (calculated as the difference in voltage amplitudes between the event-related potential responses to matched- vs. mismatched-case identity priming) for words and pseudowords in the three time windows of the analysis

The effect of case match interacted with lexicality in the second time epoch (N250: 250–350 ms): Case-match effects were dramatically attenuated for words in this time window. For pseudowords, however, case-match effects were sizeable in this time epoch, as well as in the following time epoch. The lack of an N250 effect of case for word stimuli is in line with previous studies in which the manipulations on size and font across prime and target word stimuli did not have any impact on the N250 component (Chauncey et al., 2008). The idea is that by this time in processing, size and shape invariance have already been achieved. This could be taken as further evidence of access to orthographic representations that are insensitive to changes in physical parameters of letters, like font, size, or case (Cohen & Dehaene, 2004). Importantly, although this was true for words, it was not for pseudowords. Therefore, orthographic abstract retrieval is achieved in the context of the activation of lexical entries, as revealed by the lack of N250 effects for words. In pseudowords, visual dissimilarity between prime and target is not cancelled out as no common abstract lexical representation is tapped into. That is, the abstract orthographic coding of a pseudoword is not attained, as different electrophysiological and behavioral responses are obtained for the various allographic forms of the same letters.

As Fig. 4 shows, the dissociation between words and pseudowords regarding the effects of case starts around 200 ms and is present further on: Null N400 effects were observed for the match-case manipulation for word targets (but not for pseudoword targets), thus revealing that the mapping between whole-word and semantic representations is not sensitive to visual feature differences between prime and target (Grainger & Holcomb, 2009; see Kutas & Federmeier, 2011, for a review on the N400 component). We are treating the effects obtained in these two separate time windows (N250 [250–350 ms] and N400 [400–500 ms]) as pertaining to two different functional ERPs, as we take into account the different scalp distribution for each of them (N250 [anterior] vs. N400 [central–posterior]).Footnote 5

What are the implications of these data for models of visual-word recognition? The present results challenge the theoretical accounts that assume that there is automatic activation of abstract orthographic codes during the process of visual-word recognition in a purely feedforward manner. Most hierarchical (and, at a certain point, “encapsulated”) models of orthographic processing assume that letters are initially processed by detectors at a retinotopic level (i.e., the letter detector for “a” would be activated by the letter “a” but not by the letter “A), whereas later in processing, “invariant to visual-features” detectors respond to multiple versions of the same letter (at an abstract, orthographic level—i.e., the detectors for the abstract representation of “a” would react similarly upon the presentation of “a” or “A”; see Dehaene et al., 2005; see also Grainger et al., 2008, for a neural account of abstract letter identification). Likewise, some computational models of visual-word recognition have provided a purely feedforward account of sublexical orthographic processing (e.g., Bayesian reader model: Norris & Kinoshita, 2008).

However, the present ERP data suggest that abstract orthographic processing is modulated by lexical processing: When no lexical representations are available (as is the case for pseudowords), there are processing differences between matched- and mismatched-case identity pairs in the N250 and N400 components. This divergence occurs because there is no feedback from higher-level processes, because pseudowords do not have lexical representations. Therefore, the present pattern of ERP and behavioral data favors the view that fully interactive processes are involved in orthographic encoding during word reading at relatively early stages of processing (see Carreiras et al., 2014, for a recent review of neural accounts of visual-word recognition). The type of top-down feedback that we propose may be neurobiologically instantiated in terms of frontal–occipito-temporal connections. Consistent with our interpretation, Woodhead et al. (2012) found higher activation in the left inferior frontal gyrus for words (as compared to false fonts), together with feedback connections from the inferior frontal gyrus to the left ventral occipito-temporal cortex within the first 200 ms of stimulus processing (see also Thesen et al., 2012, for converging evidence). Likewise, Twomey, Kawabata Duncan, Price, and Devlin (2011) found that the activation of the left ventral occipito-temporal cortex, a brain structure involved in early stage processing during visual word recognition, was modulated differently by emphasizing phonological versus orthographical criteria in the lexical decision task. Further evidence along the same lines was provided by Cornelissen et al. (2009), who used magnetoencephalography (MEG) measures to show very early coactivation (within 200 ms) of speech motor areas and fusiform gyrus (orthographic word-form processing) during the presentation of words, consonant strings, and unfamiliar faces (see also Cai, Paulignan, Brysbaert, Ibarrola, & Nazir, 2010; Devlin, Jamison, Gonnerman, & Matthews, 2006; Kherif, Josse, & Price, 2011; Pammer et al., 2004).Footnote 6

The present data apparently are at odds with the evidence obtained with single letters as stimuli, which has shown N/P150 effects of feature-level processing, whereas more abstract relatedness effects are shown later in time (Carreiras et al., 2013; Petit et al., 2006). However, the differences may be more apparent than real. The retrieval of abstract orthographic representations (or abstract letter representations; Brunsdon, Coltheart, & Nickels, 2006; Jackson & Coltheart, 2001; Rothlein & Rapp, 2014) may originate in modality-specific (visual, phonological, or even motor) representations. Although the Petit et al. (2006) and the Carreiras et al. (2013) experiments controlled for the visual similarity between cross-case letter pairs (a–A, b–B, c–C . . .), the obtained priming effects could have originated at a phonological or motor level (letter-name or motor representations shared by cross-case pairs; see Rothlein & Rapp, 2014). As an illustration, in the Carreiras et al. (2013) cross-case same–different experiment, the stimuli “a” and “A” corresponded to “same” responses. Leaving aside that the cross-case same–different letter task has an inherent letter level of processing (i.e., a prelexical level of processing) rather than a lexical level of processing, this may have resulted in the preactivation of any other type of representation common to both allographs (not only orthographic, but phonological or motoric as well). What we should note here is that during normal reading—or when performing a lexical decision task—the activated codes should be lexical (rather than prelexical) in nature, and in this scenario, fine-grained decomposing processes may be overrun by whole-word processing.

In conclusion, we have demonstrated that different-case word forms (lower- and uppercase) tap into the same abstract representation, leading to prime–target integration very early in processing, as we deduced from similar ERP waves for the N250 and N400 components. Importantly, this process does not occur for pseudowords. This poses some problems for accounts that assume that an abstract orthographic code is automatically attained during word processing in the absence of lexical feedback, and favors fully interactive models of visual-word recognition. Further research should be devoted to exploring how these abstract representations are constructed (and retrieved) in the process of learning to read (see Polk et al., 2009).