1,648
Views
0
CrossRef citations to date
0
Altmetric
Articles

Intact reading ability despite lacking a canonical visual word form area in an individual born without the left superior temporal lobe

ORCID Icon, , ORCID Icon &
Pages 249-275 | Received 29 Mar 2022, Accepted 26 Dec 2022, Published online: 18 Jan 2023

ABSTRACT

The visual word form area (VWFA), a region canonically located within left ventral temporal cortex (VTC), is specialized for orthography in literate adults presumbly due to its connectivity with frontotemporal language regions. But is a typical, left-lateralized language network critical for the VWFA’s emergence? We investigated this question in an individual (EG) born without the left superior temporal lobe but who has normal reading ability. EG showed canonical typical face-selectivity bilateraly but no wordselectivity either in right VWFA or in the spared left VWFA. Moreover, in contrast with the idea that the VWFA is simply part of the language network, no part of EG’s VTC showed selectivity to higher-level linguistic processing. Interestingly, EG’s VWFA showed reliable multivariate patterns that distinguished words from other categories. These results suggest that a typical left-hemisphere language network is necessary for acanonical VWFA, and that orthographic processing can otherwise be supported by a distributed neural code.

Introduction

In the past two decades, numerous regions in the ventral temporal cortex (VTC) have been identified and characterized that respond selectively to different high-level visual categories (e.g., faces: Kanwisher et al., Citation1997; places: Epstein & Kanwisher, Citation1998; bodies: Downing et al., Citation2001; for review, see Kanwisher, Citation2010). What are the origins of these specialized regions? How do human brains develop this functional topography? Some have hypothesized that the functional organization of the VTC may be innate and related to the evolutionary importance of certain high-level visual categories. Indeed, face perception and recognition abilities appear to be heritable (Wilmer et al., Citation2010; Zhu et al., Citation2010), and 2–9 month-old infants already show face- and place-responsive areas within expected locations in the VTC (Deen et al., Citation2017; Kosakowski et al., Citation2022). A related hypothesis is that pre-existing biases for certain visual attributes (e.g., retinotopy, Hasson et al., Citation2002; Malach et al., Citation2002; rectilinearity, Nasr et al., Citation2014) or perceptual dimensions (e.g., real-world size and animacy, Konkle & Caramazza, Citation2013) may predispose a brain region to become selective for particular visual categories. However, evolutionary pressures cannot explain the existence of a brain region that specializes for orthography (Hannagan et al., Citation2015)—the visual word form area (VWFA). The VWFA is a small region in the left lateral VTC that shows strong selectivity for visual words and letter strings in literate individuals (e.g., Baker et al., Citation2007; Cohen et al., Citation2003; Dehaene & Cohen, Citation2011; Hamamé et al., Citation2013). Perhaps surprisingly—given its late emergence—the VWFA is located in approximately the same location across individuals and scripts (Baker et al., Citation2007). What sets word forms apart from other visual categories and why does the VWFA land in this canonical location?

One compelling possibility is that the specialization of category-selective regions in the VTC is constrained by their differential connectivity to the rest of the brain (the Connectivity Hypothesis; Mahon & Caramazza, Citation2011; Martin, Citation2006). Indeed, previous work showed that category-selective responses can be predicted from both structural and functional connectivity (Osher et al., Citation2015; Saygin et al., Citation2012); further, these distinct connectivity patterns may already exist at birth and may drive future functional specialization (e.g., newborns show functional connectivity differences between lateral VTC which houses the VWFA vs. medial VTC: Barttfeld et al., Citation2018). Thus, written words may be processed in a stereotyped region within the left VTC due to this region’s pre-existing connectivity with the left-lateralized language network (e.g., Behrmann & Plaut, Citation2013; Dehaene et al., Citation2015; Martin, Citation2006). This network consists of left lateral frontal and lateral temporal areas and selectively supports language comprehension and production (e.g., Fedorenko et al., Citation2011).

Consistent with this idea, a number of studies have reported both anatomical and functional connections between the VWFA and the language network in neurotypical adults. For example, compared to the adjacent FFA, the VWFA shows stronger anatomical connectivity to the left superior temporal, anterior temporal, and inferior frontal areas (perisylvian putative language regions) (Bouhali et al., Citation2014). Several candidate white matter fascicles may serve to connect the VWFA with frontal and superior temporal language cortex (Wandell et al., Citation2012; Yeatman & Feldman, Citation2013). Similarly, using resting-state functional connectivity, Stevens et al. (Citation2017) found that the individually defined VWFA connects to the posterior left inferior frontal gyrus as well as the left planum temporale, both part of the distributed left-hemisphere language network.

Moreover, a longitudinal study in children showed that the location of the VWFA could be successfully predicted by its connectivity patterns at the pre-reading stage (before any visual word selectivity within the VTC is observed), and that the most predictive connections of the relevant area of the VTC were with the frontal and temporal language areas (Saygin et al., Citation2016). Even more strikingly, this pattern of preferential connectivity appears to be already present in neonates (Li et al., Citation2020). These connections between the VWFA and the frontotemporal language network appear to be functionally important such that their damage or stimulation leads to reading difficulties. For example, a case study of a child with a missing arcuate fasciculus (AF, presumably connecting the VTC and other parts of the temporal cortex to parietal and frontal areas; Wandell et al., Citation2012) found impaired reading ability (Rauschecker et al., Citation2009). Similarly, lesions in the occipito-temporal white matter result in reading impairment (Cohen et al., Citation2016) or alexia (Greenblatt, Citation1976), presumably due to disruptions of the VWFA’s connections with temporal language areas.

A different way to assess the importance of a language network in developing visual word selectivity is to ask whether language regions and the VWFA occupy the same hemisphere. In the majority of individuals, the language regions and the VWFA co-lateralize to the left hemisphere (LH) (e.g., Cai et al., Citation2010; Gerrits et al., Citation2019). In rare instances where neurotypical individuals show right-hemispheric (RH) language dominance, VTC activation during reading tasks also tends to be right-lateralized (e.g., Cai et al., Citation2008; Van der Haegen et al., Citation2012). Another population where the language network is right-lateralized are individuals with congenital or early left hemisphere (LH) damage (e.g., Asaridou et al., Citation2020). In the presence of early LH damage, linguistic abilities tend to develop normally (e.g., Newport et al., Citation2017; Staudt, Citation2010; Staudt et al., Citation2001; see François et al., Citation2021 for a review). However, little is known about the effects of early LH damage on reading ability and on the neural architecture of visual word processing. In particular, if the left VTC is completely deafferentated from the downstream LH language cortex at birth, does the VWFA emerge in the right VTC when the language network has no choice but to develop in the right hemisphere, or does it still emerge in the left VTC, due to some pre-existing bias (e.g., innate connectivity with any spared LH cortex)? Indeed, some previous studies show that language processing and reading can engage opposite hemispheres (e.g., Van der Haegen et al., Citation2012). Or—perhaps even more importantly—whatever hemisphere it emerges in, does the VWFA look atypical (e.g., less functionally selective or integrated into the language network, which in neurotypical individuals responds to linguistic demands across reading and listening modalities; e.g., Fedorenko et al., Citation2010; Regev et al., Citation2013; Vagharchakian et al., Citation2012)? Do these potential differences affect selectivity for other high-level visual categories in the VTC?

Moreover, if the VWFA manifests atypically but reading ability is normal, what does this tell us about how orthographic processing is implemented in the brain (and the flexibility of this implementation in the case of brain lesion)? What could be a potential neural mechanism that might support normal reading ability? Canonical category-selective regions, like VWFA, are defined by traditional univariate analyses which test for the visual category that each voxel responds most to (or whether there exists a set of voxels (i.e., functionally specialized region) that respond more (indicated by overall mean activation) to the category of interest as compared to control categories). Although this is the classic approach to demonstrate category selectivity in the VTC, it may be insensitive to potentially meaningful distributed representation patterns in suprathreshold and/or subthreshold voxels. Using multivariate pattern analysis (MVPA), previous studies show that categorical information can be reliably decoded by comparing within-category versus between-category correlations in the VTC (e.g., Haxby et al., Citation2001), suggesting that a more distributed code at voxel level (even outside of canonical category-selective areas that are identified by univariate analysis) can be used to represent each visual category. Moreover, a recent study found mature representational similarity structures via multivariate patterns in children who showed immature/no univariate selectivity (Cohen et al., Citation2019). Therefore, representational structures based on distributed response patterns across a larger set of voxels may developmentally precede category selectivity and/or the response preferences of a small patch of cortex to a particular category.

Here we investigate possible functional reorganization of the visual word selectivity in the absence of a typical left-lateralized language network. We have a unique opportunity to examine fMRI responses to stimuli from different visual categories in an individual (EG) born without the left superior temporal lobe (likely due to pre/perinatal stroke) but with lVTC largely intact. EG’s frontotemporal language network is completely remapped to the right hemisphere; no language-related responses, as assessed with fMRI, were observed in the remaining parts of EG’s left hemisphere (Tuckute et al., Citation2022). EG’s reading abilities (as well as other linguistic abilities) are intact. We here investigated whether (a) in the presence of a right-lateralized language network, a typical VWFA (showing significantly higher activation to orthographic stimuli than to other visual categories (e.g., faces, objects) would emerge in the right VTC either in its canonical location along lateral fusiform gyrus or perhaps in other parts of the right VTC; (b) visual word selectivity is observed in the (spared) left VTC; and (c) visual word processing could be taken over by brain regions that support general linguistic processing. Thereafter in this manuscript, we used “word(s)” to refer to written words and “word selectivity” to refer to visual word selectivity unless otherwise noted. To foreshadow the results, no word selectivity was observed in EG’s right or left VTC, despite typical selectivity for other visual categories; and brain regions that support high-level language processing did not distinguish between visual words and other visual categories, ruling out the possibility that univariate visual word processing is taking place within this high-level language network or within VTC. We then explored the possibility that orthographic processing is supported by still selective but more distributed neural populations using MVPA. MVPA is an analysis that by design, involves the pattern of activation across a set of voxels: instead of looking at individual voxels and e.g., averaging activation strength across them, MVPA is based on the pattern of responses across voxels—the neural code is distributed across voxels. Indeed, we observed such “multivariate selectivity” bilaterally, but manifesting more strongly in the right VTC.

Methods

Participants

Brain-lesioned participant

The participant EG (fake initials; right-handed female with an advanced professional degree, 54 years old at the time of testing) contacted Dr. Fedorenko’s lab to participate in brain research studies. Based on her own report, the lack of the left superior temporal lobe (; Supplementary Figure S1) was discovered when she was 25 years old (in her first MRI scan in 1987) and being treated for depression. Specifically, the entire left superior temporal cortex (based on the Desikan-Killiany parcellation (Desikan et al., Citation2006), including both superior temporal gyrus and sulcus) is missing, and the lesion extended to anterior middle- and inferiortemporal cortex and further to the anterior fusiform gyrus. No known head traumas or injuries were reported as a child or adult. Several medical MRI scans were performed in subsequent years (1988, 1998, and 2013) and no change was observed compared to the initial scans. Importantly, EG did not report any difficulties in reading or general language abilities (see details below). She had also acquired fluency in a second language (Russian). EG was invited to participate in a series of behavioural and fMRI assessments at MIT. With respect to testing relevant to the current study, EG completed five runs of the VWFA localizer (see The VWFA localizer task section below) in October 2016 (session 1), and four runs of the same VWFA localizer in November 2021 (session 2). Our main analysis focused on session 1 (see Data acquisition section below), and we invited EG back for the second session to replicate results and also more critically, to search potential word-selective responses outside the VTC with whole-brain coverage (see Supplementary Methods: Data acquisition parameters for the VWFA localizer with a whole-brain coverage (from the second session)). Written informed consent was obtained from EG, and the study was approved by MIT’s Committee on the Use of Humans as Experimental Subjects (COUHES).

Neurotypical controls

Twenty-five adults (11 female, mean age = 23.6 years old; age range 18–38 years; standard deviation 5.21 years) from The Ohio State University (OSU) and the surrounding community were included in the present study. As part of ongoing projects exploring the relationship between brain function and connectivity, all participants completed a battery of fMRI tasks, including, critically, the same VWFA localizer task that EG completed (see The VWFA localizer task section below). All participants had normal or corrected-to-normal vision, and reported no neurological, neuropsychological, or developmental diagnoses. Written informed consent was obtained from all participants and the study was approved by Institutional Review Board at OSU. It is worth noting that although the control group participants were younger than EG, further examination revealed no significant correlation between age and word selectivity (see Definition of functional regions of interest and univariate analyses section below) in our control group (r = 0.148, p = 0.763; p-value was obtained by a permutation test).

Reading assessment (EG only)

To formally evaluate EG’s linguistic abilities, five standardized language assessment tasks were administered: (i) an electronic version of the Peabody Picture Vocabulary Test (PPVT-IV) (Dunn & Dunn, Citation2007); (ii) an electronic version of the Test for Reception of Grammar (TROG-2) (Bishop, Citation2003); (iii) the Western Aphasia Battery-Revised (WAB-R) (Kertesz, Citation2006); (iv) the reading and spelling components of PALPA (Kay et al., Citation1992); and (v) an electronic version of the verbal components of the Kaufman Brief Intelligence Test (KBIT-2) (Kaufman & Kaufman, Citation2004). PPVT- IV and TROG-2 target receptive vocabulary and grammar, respectively. In these tasks, the participant is shown sets of four pictures accompanied by a word (PPVT-IV, 72 trials) or sentence (TROG-2, 80 trials) and has to choose the picture that corresponds to the word/sentence by clicking on it. WAB-R (Kertesz, Citation2006) is a more general language assessment developed for persons with aphasia. It consists of 9 subscales, assessing (1) spontaneous speech, (2) auditory verbal comprehension, (3) repetition, (4) naming and word finding, (5) reading, (6) writing, (7) apraxia, (8) construction, visuospatial, and calculation tasks, and (9) supplementary writing and reading tasks. Three composite scores (language, cortical, and aphasia quotients) were calculated from the subscales (the criterion cut-off score for diagnosis of aphasia is an aphasia quotient of 93.8). The verbal components of KBIT-2 include (1) the Verbal Knowledge subtest, which consists of 60 items measuring receptive vocabulary and general information about the world, and (2) the Riddles subtest consists of 48 items measuring verbal comprehension, reasoning, and vocabulary knowledge. Most relevant to the current investigation, the reading component of WAB-R includes comprehension of written sentences and reading commands; the supplementary reading tasks include reading of irregular words and nonwords. The reading component of PALPA (Kay et al., Citation1992) consists of 21 tasks (tasks #18 through 38), which focus on single letter recognition, visual lexical decision, out-loud reading (of words with diverse lexical properties and sentences), and homophone definition (see Supplementary Table S2 for details).

To further investigate any reading or visual word processing deficiencies which may not be apparent in the standard linguistic and reading measures reported above, we also had EG complete a speeded reading task and compared her performance to that in an independent large sample of neurotypical adults (see Ryskin et al., in preparation, for details). Briefly, twelve-word-long sentences were presented word-by-word at varying speeds. The original presentation speed was based on the natural out-loud reading speed (as recorded by a female native English speaker): each word was visually presented for the number of ms that it took the reader to say the word. The speed was manipulated by compressing the sentence presentations to 80%, 60%, 50%, 45%, 40%, 35% and 30% of the original speed (100%). Participants were asked to type in as many words as they were able to discern after each sentence, and the accuracy of participants’ responses (how many words of the sentence they typed in correctly, not taking into account minor typos) was recorded.

Data acquisition

EG’s data were collected on a 3 Tesla Siemens Trio scanner with a 32-channel head coil at the Athinoula A. Martinos Imaging Center at the McGovern Institute for Brain Research at MIT. Data of the control group participants were acquired on a 3 Tesla Siemens Prisma scanner with a 32-channel head coil at the Center for Cognitive and Behavioral Brain Imaging at OSU. To ensure that any differences between EG and our control group were not due to scanner differences, we compared word selectivity in the current control group to a smaller group of adults (N = 14) recruited at MIT and scanned with the same scanner and protocols as EG; comparable response levels (activation in the lVWFA and lFFA) were found between the OSU and MIT data (Supplementary Table S1). For both EG and controls, a whole-head, high resolution T1-weighted magnetization-prepared rapid acquisition with gradient echo (MPRAGE) scan was acquired (EG: repetition time (TR) = 2530 ms, echo time (TE) = 3.48 ms; voxel resolution = 1.0 mm3; the control group: TR = 1390 ms, TE = 4.62 ms, voxel resolution = 1.0 mm3). Functional images for the VWFA task were acquired with the same echo-planar imaging (EPI) sequence for both EG (session 1, the first visit) and controls: TR = 2000ms, TE = 30 ms, 172 TRs, 100 × 100 base resolution, voxel resolution = 2.0 mm3, field of view (FOV) = 200 mm; 25 slices approximately parallel to the base of the temporal lobe to cover the entire ventral temporal cortex and a small portion of the inferior frontal cortex. Unless otherwise noted, results presented are from EG’s first visit where the acquisition parameters were identical for EG and controls. To search for potential word selectivity outside the VTC, we invited EG back recently and collected data from the same VWFA task with a slightly different protocol to get whole-brain coverage (see Supplementary Methods: Data acquisition parameters for the VWFA localizer with a whole-brain coverage (from the second session)). EG also completed a language localizer task during session 1: EPI sequence with TR = 2000ms and TE = 30 ms, 227 TRs, 96 × 96 base resolution, voxel resolution  = 2.1 × 2.1 × 4.4 mm3, FOV = 200 mm, 31 near-axial slices acquired in the interleaved order.

The VWFA localizer task

A VWFA localizer was used to define high-level category-selective regions and to measure category-selective responses (see Saygin et al., Citation2016, for details). Briefly, black and white line drawings of words, scrambled words, objects, and faces, along with the fixation condition were shown in a blocked design. A grid was overlaid on top of the stimuli so that all stimulus types (not just scrambled words) had edges. Each stimulus was presented for 500 ms (ISI = 0.193s) and overlaid on a different single-color background, and 26 stimuli (including 2 repetitions) were presented in each block. Each run consisted of 19 blocks (4 blocks per condition and 3 fixation blocks), and participants performed a one-back task. The stimuli are available for download at http://www.zeynepsaygin.com/ZlabResources.html (examples of the stimuli were shown in Supplementary Methods: Stimuli and paradigm for the VWFA localizer). EG completed 5 runs, and participants in the control group completed 2 runs. Note that previous studies using the same task indicated that 2 runs of data are sufficient to successfully identify the VWFA in a neurotypical population (Saygin et al., Citation2016); here, we acquired more runs for EG to ensure that we had sufficient power and that the results obtained for EG were stable across runs (see Definition of functional regions of interest and univariate analyses section for details).

The language localizer task

During session 1, EG also completed a language localizer that was adapted from Fedorenko et al. (Citation2010). Briefly, there were four types of stimuli: English sentences, scrambled sentences (i.e., word lists), jabberwocky sentences (sentences where function words and bound morphemes remained and while all the content words are replaced by pronounceable nonwords, like, for example, “florped” or “blay”), and nonword sequences (scrambled jabberwocky sentences) (see Fedorenko et al., Citation2010 for details). Stimuli were presented in both visual and auditory blocks, each lasting 22s and 26s respectively. Each run consisted of 20 blocks (4 blocks per stimulus type × 2 modalities, and 4 fixation blocks), and the order was counterbalanced across runs. Note in visual blocks, stimuli were presented sequentially, that is, one word/nonword at a time. To control attention demands, a probe word/nonword was presented at the end of each trial, and subjects had to decide whether the probe appeared in the immediately preceding stimulus. The English sentences (En) and Nonword sequences (Ns) were the critical conditions to identify language-selective responses (i.e., high-level linguistic information: lexico-semantic and syntactic).

Preprocessing and fMRI analysis

Data were analyzed with Freesurfer v.6.0.0 (http://surfer.nmr.mgh. harvard.edu/), FsFast (https://surfer.nmr.mgh.harvard.edu/fswiki/FsFast), FSL (https://fsl.fmrib.ox.ac.uk/fsl/fslwiki/FLIRT) and custom MatLab code. All structural MRI data were processed using a semiautomated processing stream with default parameters (recon-all function in Freesurfer: https://surfer.nmr.mgh.harvard.edu/fswiki/recon-all/), which includes the following major steps: intensity correction, skull strip, surface co-registration, spatial smoothing, white matter and subcortical segmentation, and cortical parcellation. Cortical grey matter and ventral temporal cortex masks (comprising inferiortemporal and fusiform cortex) were created based on the Desikan-Killiany (Desikan et al., Citation2006) parcellation in native anatomy for each subject. The success of cortical reconstruction and segmentation was visually inspected for EG.

Functional images were motion-corrected (time points where the difference in total vector motion from the previous time point exceeded 1 mm were excluded), data from each run were registered to each individual’s anatomical brain image using bbregister, and resampled to 1.0 × 1.0 × 1.0mm3. For EG, instead of registering the functional image of each run to the anatomical brain separately, we aligned the functional images of the first four runs to the last run (which had successful functional-anatomical cross-modal registration with bbregister) with linear affine transformation (FLIRT); then the functional-anatomical transformation for the last run was applied to all functional runs and was visually inspected (tkregisterfv) for functional to anatomical alignment.

The preprocessed functional data were then entered into a first-level analysis. Specifically, data were detrended, smoothed (3 mm FWHM kernel), and the regressor for each experimental condition (Words, Scrambled Words, Objects, and Faces) was defined as a boxcar function (i.e., events on/off) convolved with a canonical hemodynamic response function (a standard gamma function (d = 2.25 and t = 1.25)). Orthogonalized motion measures from the preprocessing stage were used as nuisance regressors for the GLM. Resulting beta estimates for each condition and contrasts of interest between categories (i.e., Words > Others categories and Faces > Others categories) were used in further analyses. For multivariate analyses, no spatial smoothing was applied.

Definition of functional regions of interest and univariate analyses

The subject-specific functional regions of interests (fROIs) were defined with a group-constrained subject-specific (GcSS) approach (Fedorenko et al., Citation2010). In this approach, individual fROIs are defined by intersecting participant-specific fMRI contrast maps for the contrast of interest (e.g., Words > Other conditions) with some spatial constraint(s) or “parcel(s)”, which denote the area(s) in the brain within which most individuals show responses to the relevant contrast (based on past studies with large numbers of participants). In the present study, the left VWFA (lVWFA) and the right FFA (rFFA) parcels were derived from previous studies (FFA, Julian et al., Citation2012; VWFA, Saygin et al., Citation2016) (). These parcels were generated based on probabilistic maps of functional activation for the relevant contrasts in independent groups of participants. To provide a relatively symmetrical search space, we flipped the lVWFA and rFFA to get the functional parcel on the contralateral side (results using the original lFFA parcel from Julian et al. (Citation2012) are reported in the Supplementary Table S3). We then registered these parcels to our participants’ own anatomy using the combined volume and surface-based (CVS) non-linear registration method (mri_cvs_register; Postelnicu et al., Citation2009). After mapping the functional parcels to each participant’s brain, we defined the VWFA fROI by selecting within the VWFA parcel the top 10% of most active voxels for the contrast Words > Other conditions (i.e., scrambled words, objects, and faces). Similarly, we defined the FFA by selecting within the FFA parcel the top 10% of most active voxels for the contrast Faces > Other conditions. For the control participants, we used run 1 to define the fROIs, and the run 2 to extract percent signal changes (PSCs; beta estimates divided by baseline activation were averaged across voxels within fROI (defined with run 1 data)) for each of four experimental conditions; the same procedure was repeated with run 2 to define the fROIs and run 1 to extract PSCs. The average result from two directions (defined fROIs with run 1 and extracted with run 2 and then vice versa) for each subject was used in further analyses. For EG, who had 5 runs of data, this procedure was performed iteratively for every 2-run combination (e.g., defining fROIs with run 1 and extracting PSCs from run 2, then defining with run 2 and extract from run 1, and results were averaged across two directions); the same procedure were repeated for all 2-run combinations (e.g., 1&3, 1&4, etc.; 10 combinations in total). When comparing EG to controls, results from ten run combinations were averaged to derive a single estimate per condition per fROI. Note that although the parcels are relatively large, by design (in order to accommodate inter-individual variation in the precise location of the functional region), and can overlap, the resulting fROIs within an individual are small and do not overlap (Saygin et al., Citation2016). Additionally, this method identified the most responsive voxels to the condition of interests at the expected locations (parcels), but it does not require contiguity of selected voxels within the parcel (Supplementary Methods: Example fROIs for EG and controls). We further calculated the selectivity indices for words and faces with the following formula: (PSC to the condition of interest – average PSC of the remaining conditions)/(summed PSC for all four conditions); note that when calculating selectivity, we adjusted for baseline activation following previous studies (Simmons et al., Citation2007; Szwed et al., Citation2011) in order to correct for potential bias induced by negative activation that are sometimes observed in fMRI studies. Briefly, for each fROI, we identified the minimal response across the four conditions that was negative and added that absolute value to all conditions to make the smallest response across the categories equal to 0.

Multivariate analyses: split-half correlations

To further examine whether visual words may be represented and processed in a spatially distributed manner, we performed a multivariate pattern analysis (MVPA) to measure distinctive activation patterns for different conditions. The analyses were performed with CoSMoMVPA toolbox (Oosterhof et al., Citation2016) (https://www.cosmomvpa.org/). In line with the approach introduced in Haxby et al. (Citation2001), we examined split-half within-category and between-category correlations. In particular, a searchlight (radius = 3 voxels, ∼90 voxels for each searchlight) was created for each voxel within the VTC, and response patterns (i.e., beta estimates for each of the four conditions, normalized by subtracting mean responses across all conditions) were extracted from each searchlight.

Before performing the critical analysis, we asked whether the overall multivariate representation structure of the VTC is typical in EG. To do so, we constructed a representational similarity matrix (RSM) from pairwise similarities (i.e., correlations) based on the voxel-wise response patterns in a given searchlight area (defined above) to different conditions (e.g., the correlation between activation patterns across voxels to Words and Faces in a given searchlight area). Correlations of all searchlights within the VTC were then Fisher z-transformed and averaged. This resulted in one 4 × 4 RSM from two runs of data for each participant in the control group; for EG, RSMs from ten run combinations were averaged to get a single RSM. We then calculated RSM similarity (correlation) between EG and controls, and tested whether this correlation was different from the correlations between any given control individual and the rest of the control group (within-controls correlations; similarity of RSM of each subject to the average RSM of the remaining control subjects).

Then, similarity of response patterns within a category (e.g., Words-Words or Faces-Faces) vs. between categories (e.g., Words-Faces) was calculated within the VWFA and FFA parcel boundaries by Pearson’s correlation in a pair-wise manner, and then Fisher z-transformed. We identified voxels of interest that satisfied the following two criteria: (1) voxels whose local neighbourhoods showed higher within- than between-category correlations (and which therefore represent categories distinctively); and (2) voxels whose local neighbourhoods showed higher within-condition correlations for a particular category (e.g., Words-Words) than within-condition correlations for all other categories (e.g., Faces-Faces). The second criterion identified voxels which represent a particular category (e.g., visual words) in a more selective fashion. We refer to such voxels that satisfied both criteria as multivariate-selective voxels.

To examine hemispheric differences, we computed the number of searchlight centroids that exhibited multivariate selectivity for words or faces in the two hemispheres (within the relevant parcels). We then calculated laterality with the following formula: (number of searchlights on the left – number of searchlights on the right)/(summed searchlights on left and right).

Statistical analyses

Paired t-tests were used for comparisons between conditions for EG (across ten run split combinations) and within the control group. For all analyses where we compared EG’s response to the control group, we used a Crawford-Howell modified t-test (Crawford & Howell, Citation1998), which is widely used in single-case studies because it accounts for the variance of the control group, and the percentage of false positives remains the same for various sizes of the control sample (N = 5–50) (Crawford et al., Citation2009). This frequentist approach provided the point estimate (p-value) for the proportion of the control population that will obtain a score more extreme than the brain-lesioned participant’s score. In addition, we computed the Bayesian 95% credible interval (a Bayesian alternative to the modified t-test; Crawford & Garthwaite, Citation2007) to demonstrate the range of p-values based on 10,000 iterations. Two-tailed tests were reported except for the test of the word or face selectivity indices, which we only tested if they were larger than zero. Multiple-comparison correction were applied with Holm–Bonferroni method (reported as pbh in the Results section) and p-values were corrected by the number of comparisons performed in each analysis. To assess the significance of RSM correlations between EG and controls, we generated a null distribution of correlation values by shuffling the matrix of EG and controls (i.e., randomizing the labels of values in the RSMs) and then correlating the new shuffled matrices. This procedure was repeated 10,000 times to create the null distribution of the correlation values. The p-value was calculated by counting the number of correlations in the null-distribution that were higher than the correlation value based on the correct category labels, and then divided by 10,000.

Results

Does EG have normal reading ability?

In line with her self-report, EG performed within normal range on all language assessment tasks. Her accuracy was 90% on PPVT, 99% on TROG-2, and she obtained the scores of 97.6, 98.6, and 98.4 on the aphasia, language, and cortical quotients of the WAB-R respectively, and a score of 130 (98th percentile) on the KBIT-verbal. EG’s performance was therefore not distinguishable from the performance of neurotypical controls. With respect to the reading assessments, EG made no errors on the main reading component of WAB-R, no errors in the reading of irregular words, and one error in the reading of nonwords (reading “gobter” instead of “globter”). For the PALPA tasks, she made no errors on tasks that focus on single letters (tasks #18–23), no errors on the visual decision tasks (tasks #24–27), no errors on the out-loud reading tasks (tasks #29–37), and no errors on the homophone definition task (task #38). For the homophone decision task (task #28), EG made three errors (out of the 60 trials; all were made on nonword pairs: she did not judge the following pairs as sounding the same: heem-heam, byme-bime, and phex-feks). This performance is on par with neurotypical controls.

Consistent with the results from these standard reading measures, behavioural performance during the 1-back VWFA localizer task in the scanner also showed that EG’s response accuracy (minimal p-value = 0.305) and response time (minimal p-value = 0.099) were not different from controls for all visual categories (A). Critically, EG’s performance on the speeded reading task was also within the range of the control distribution (minimal p-value = 0.286) (see Methods for details) even for the fastest presentation rates (B). This result demonstrates that not only does EG perform within the typical range on temporally unconstrained/self-paced reading assessments, but her reading mechanisms are not compromised in terms of their speed. Altogether, EG appears to have intact linguistic, reading, and visual word recognition ability.

Figure 1. Comparing EG’s reading performance with neurotypical controls. (A) Accuracy (left) and response time (right) during the 1-back VWFA localizer task. (B) Proportion of words typed that matched the words in the target sentence in different speed conditions.

Figure 1. Comparing EG’s reading performance with neurotypical controls. (A) Accuracy (left) and response time (right) during the 1-back VWFA localizer task. (B) Proportion of words typed that matched the words in the target sentence in different speed conditions.

Is word selectivity observed in the right hemisphere when the language network remaps to the right hemisphere due to early left-hemisphere damage?

After confirming normal reading ability in EG, we moved on to our main analysis to examine univariate word selectivity in EG’s rVWFA. With the VWFA localizer, we defined the individual-specific rVWFA fROIs, in EG and controls, within a spatial constraint (rVWFA parcel; see and Methods) by contrasting Words versus all other categories (Words > Others). We then extracted the activation to all four conditions from independent data (see Methods for details). In neurotypical literate individuals, word selectivity is strongly left-lateralized (McCandliss et al., Citation2003); selective responses to visual words in the right homotope of lVWFA are less frequently observed and are less spatially consistent across individuals. Results from our control group are in line with this picture: we found no selective activation to Words as compared to other visual categories in the rVWFA (Words vs. Scrambled Words: t(24) = −1.090, p = 0.287; Words vs. Objects: t(24) = −0.133, p = 0.895; Words vs. Faces: t(24) = 0.026, p = 0.979; A).

Figure 2. EG’s MRI showing the missing cortex and the parcels for the right and left VWFA and FFA. Top, T1-weighted images in coronal, sagittal, and axial views. Bottom, the VWFA (purple) and FFA (blue) parcels are projected on EG’s reconstructed surface. The parcels served as spatial constraints in defining the fROIs (see Methods), but we also explored the entire VTC for category selectivity (outlined with black solid lines). By design, the parcels are relatively large (to accommodate inter-individual variability in the precise locations of these areas) and therefore can overlap, but the individual fROIs are small and do not overlap. Note that even though part of the anterior lVTC is missing in EG, the stereotypical locations for both the VWFA and FFA are spared.

Figure 2. EG’s MRI showing the missing cortex and the parcels for the right and left VWFA and FFA. Top, T1-weighted images in coronal, sagittal, and axial views. Bottom, the VWFA (purple) and FFA (blue) parcels are projected on EG’s reconstructed surface. The parcels served as spatial constraints in defining the fROIs (see Methods), but we also explored the entire VTC for category selectivity (outlined with black solid lines). By design, the parcels are relatively large (to accommodate inter-individual variability in the precise locations of these areas) and therefore can overlap, but the individual fROIs are small and do not overlap. Note that even though part of the anterior lVTC is missing in EG, the stereotypical locations for both the VWFA and FFA are spared.

Figure 3. Responses to four conditions in the rVWFA and rFFA for EG and the control group. (A) Bar plots show mean PSCs to the four conditions estimated in independent data within individually defined rVWFA fROIs (i.e., top 10% word-selective voxels within the rVWFA parcel) for the control group. (B) Mean PSCs to the four conditions estimated in independent data within the individually defined rVWFA fROI for EG. Here and in E, the results are averaged across run combinations. (C) Parametrically decreasing the threshold for defining the rVWFA fROI from the top 5% to 50% word-selective voxels within the rVWFA parcel. Number of voxels in the rVWFA fROIs: 5% = 269 voxels, 50% = 2690 voxels. Here and in F, average PSCs across run combinations are shown for each threshold. (D) Mean PSCs to the four conditions estimated in independent data within individually defined rFFA fROIs for the control group. (E) Mean PSCs to the four conditions estimated in independent data within the individually defined rFFA fROI for EG. (F) Parametrically decreasing the threshold for defining the rFFA fROI from the top 5% to 50% face-selective voxels within the rFFA parcel. Number of voxels in the defined rFFA fROIs: 5% = 67 voxels, 50% = 672 voxels. In the bar plots, dots correspond to individual data points for each condition (controls: n = 25 participants; EG: n = 10 run combinations, from ten iterations). Horizontal bars reflect significant paired t-tests p < 0.05. Error bars in both the bar and line plots denote standard errors of the mean by participants (for the control group) and by run combinations (for EG). Words = Written Words; ScrW = Scrambled Words.

Figure 3. Responses to four conditions in the rVWFA and rFFA for EG and the control group. (A) Bar plots show mean PSCs to the four conditions estimated in independent data within individually defined rVWFA fROIs (i.e., top 10% word-selective voxels within the rVWFA parcel) for the control group. (B) Mean PSCs to the four conditions estimated in independent data within the individually defined rVWFA fROI for EG. Here and in E, the results are averaged across run combinations. (C) Parametrically decreasing the threshold for defining the rVWFA fROI from the top 5% to 50% word-selective voxels within the rVWFA parcel. Number of voxels in the rVWFA fROIs: 5% = 269 voxels, 50% = 2690 voxels. Here and in F, average PSCs across run combinations are shown for each threshold. (D) Mean PSCs to the four conditions estimated in independent data within individually defined rFFA fROIs for the control group. (E) Mean PSCs to the four conditions estimated in independent data within the individually defined rFFA fROI for EG. (F) Parametrically decreasing the threshold for defining the rFFA fROI from the top 5% to 50% face-selective voxels within the rFFA parcel. Number of voxels in the defined rFFA fROIs: 5% = 67 voxels, 50% = 672 voxels. In the bar plots, dots correspond to individual data points for each condition (controls: n = 25 participants; EG: n = 10 run combinations, from ten iterations). Horizontal bars reflect significant paired t-tests p < 0.05. Error bars in both the bar and line plots denote standard errors of the mean by participants (for the control group) and by run combinations (for EG). Words = Written Words; ScrW = Scrambled Words.

Critically, in EG, whose language network is located in the right hemisphere, with no language responses anywhere in the left hemisphere (Tuckute et al., Citation2022), we asked whether word selectivity is also observed in the right hemisphere. Surprisingly, no word selectivity was observed in EG’s rVWFA: activation to Words in her rVWFA fROI did not significantly differ from other categories (Words vs. Scrambled Words: t(9) = 0.056, p = 0.956; Words vs. Faces: t(9) = −1.107, p = 0.297; Words vs. Objects: t(9) = −1.661, p = 0.131) (B). This lack of visual word-selectivity in rVWFA was comparable between EG and controls: we found that EG’s response to words in the rVWFA did not differ from controls (t(24) = 0.915, p = 0.369, modified t-test; 95% Bayesian CI [0.076, 0.313]). Further, to ensure that this result is not due to the choice of a particular threshold (i.e., top 10%) that we used to define the rVWFA fROI, we performed the same analysis as above at a range of thresholds (5%−50%). The lack of word selectivity was stable across thresholds (C).

To test whether selectivity for other visual categories in the right VTC is typical in EG, we examined face selectivity in EG’s rFFA, which is spatially proximal to the rVWFA. Similar to the rVWFA analysis above, we defined the rFFA fROI by contrasting Faces versus other categories (Faces > Others), and extracted the activation to the four conditions from independent data (see Methods for details). EG’s face selectivity remained intact (E) and did not differ from that of the control group (D; t(24) = 0.012, p = 0.990 modified t-test, Bayesian 95% CI [0.349 0.500]). EG’s rFFA fROI showed significantly higher responses to Faces than to other conditions (Faces vs. Words: t(9) = 20.816, pbh = 1.914 × 10−8; Faces vs. Scrambled Words: t(9) = 12.390, pbh = 5.862 × 10−7; Faces vs. Objects: t(9) = 17.723, pbh = 5.258 × 10−8; E), just like what we observed in the control group (Faces vs. Words: t(24) = 16.204, pbh = 5.995 × 10−14; Faces vs. Scrambled Words: t(24) = 15.299, pbh = 7.055 × 10−14; Faces vs. Objects: t(24) = 14.475, pbh = 2.352 × 10−13; D). Moreover, the selective face responses in EG’s rFFA was observed across all thresholds used to define the rFFA (F). We further calculated the strength of selectivity (selectivity index; Simmons et al., Citation2007) for both Words and Faces within rVWFA and rFFA respectively, by taking the difference between the condition of interests and the rest conditions and divided by the sum of all conditions (see Methods for details). We found that, consistent with previous observations of rVWFA, eleven out of 25 controls showed lack of visual word selectivity (A), just like EG. Both the control group and EG showed no word selectivity in the rVWFA (control: t(24) =  0.957, p = 0.174; EG: t(9) = −0.893, p = 0.802) but showed significant face selectivity in rFFA (controls: t(24) =  10.648, pbh = 2.844 × 10−10; EG: t(9) = 11.729, pbh = 1.403 × 10−6). Importantly, EG’s selectivity index was not different from controls for both Words and Faces (Words: t(24) =  −0.362, p = 0.721, modified t-test, Bayesian 95% CI [0.240 0.499]; Faces: t(24) =  −0.741, p = 0.466, modified t-test, Bayesian 95% CI [0.107 0.369]). We found that EG’s selectivity values are among those of typical controls (A, B which illustrates EG’s word selectivity in the rVWFA and face selectivity in the rFFA among the distribution of controls), again indicating the lack of word selectivity in the rVWFA and normal face selectivity in the rFFA in EG.

Figure 4. Selectivity indices for EG and controls. Histogram was used to show the visual word selectivity (A, C) and face selectivity (B, D) indices both controls and EG. Each bar represents one participant and participants are ordered by their selectivity indices from lowest to highest.

Figure 4. Selectivity indices for EG and controls. Histogram was used to show the visual word selectivity (A, C) and face selectivity (B, D) indices both controls and EG. Each bar represents one participant and participants are ordered by their selectivity indices from lowest to highest.

Note that lacking expected selectivity might be because of the search space we used, which is a probabilistic map that captures not all but the majority (e.g., 60%) of the activation across subjects. To ensure that we did not miss any possible word-selective voxels by applying a predefined spatial constraint (i.e., VWFA parcel) and to account for the possibility that EG’s VWFA may be located in a different part of the visual cortex, we searched for word selectivity within the entire rVTC mask for EG. Specifically, different thresholds from top 1% to top 10% were used to define the most word-selective voxels (Words > Others). Even within this broad mask, no word-selective responses were observed in independent data across all thresholds; in fact, the responses were lowest to words than the other three conditions (A). In contrast, robust face-selective responses were observed in independent data across all thresholds when searching for face-selective voxels (Faces > Others) (B). We also performed an analysis where we restricted the search space only to rVTC voxels that significantly and consistently responded to visual stimuli. Similar results were observed (Supplementary Figure S2C) where no word selectivity was observed in EG’s right visual cortex.

Figure 5. Mean PSCs in the rVTC and lVTC at different thresholds for EG. (A-B) Parametrically decreasing the threshold for defining word-selective (Words > Others) and face-selective (Faces > Others) voxels from the top 1% to 10% within the rVTC. Mean PSCs across run combinations (from 10 iterations) are shown for each threshold. (C-D) Parametrically decreasing the threshold for defining word-selective and face-selective voxels from the top 1% to 10% within the lVTC. Mean PSCs across run combinations are shown for each threshold. Number of selected voxels: rVTC: 1% = 216 voxels, 10% = 2164 voxels; lVTC: 1% = 154 voxels, 10% = 1536 voxels. Words = Written Words; ScrW = Scrambled Words

Figure 5. Mean PSCs in the rVTC and lVTC at different thresholds for EG. (A-B) Parametrically decreasing the threshold for defining word-selective (Words > Others) and face-selective (Faces > Others) voxels from the top 1% to 10% within the rVTC. Mean PSCs across run combinations (from 10 iterations) are shown for each threshold. (C-D) Parametrically decreasing the threshold for defining word-selective and face-selective voxels from the top 1% to 10% within the lVTC. Mean PSCs across run combinations are shown for each threshold. Number of selected voxels: rVTC: 1% = 216 voxels, 10% = 2164 voxels; lVTC: 1% = 154 voxels, 10% = 1536 voxels. Words = Written Words; ScrW = Scrambled Words

Moreover, we supplemented our main analyses, which rely on the Words > Others contrast to define the VWFA, as is commonly done in the literature (e.g., Dehaene-Lambertz et al., Citation2018; Rosenke et al., Citation2021), with another analysis that relies on a less stringent localizer contrast: Words vs. Scrambled Words (Glezer et al., Citation2009; Lerma-Usabiaga et al., Citation2018). Unlike the Words > Others contrast, this contrast does not control for semantics or visual stimulus complexity. Even with this broader contrast, we found no word selectivity within EG’s rVTC across thresholds (Supplementary Figure S2A).

Further, we considered the possibility that while EG lacks a right-lateralized VWFA that is selective to visual words over other visual categories, perhaps a part of her right VTC would respond to linguistic stimuli in general (amodal properties i.e., selectivity for sentences presented either visually or auditorily) and that maybe this putative language region in the VTC supports EG’s normal reading ability. To explore this possibility, we implemented two analyses. First, we examined the same fROI as above (using the visual Words > Others contrast to define the rVTC fROI at different thresholds) to see if it exhibits language selectivity (as defined by higher responses to meaningful English sentences (En) than to sequences of Nonwords (Ns)) (Sentences > Nonword) presented either auditorily or visually (see Methods). We found no preferential activation to high-level linguistic information (i.e., no significant Sentences > Nonwords effect for sentences presented either auditorily or visually; Fedorenko et al., Citation2010) in the rVTC (Supplementary Figure S3A). Further, neural responses to the four visual categories (from the VWFA localizer) were not distinguishable from those for the conditions of the auditory language task (either meaningful English sentences or nonword sequences), suggesting that EG has no univariate response selectivity in rVTC to either visual words or linguistic stimuli in general (i.e., rVTC fROI defined by Words > Others does not show selectivity to words, visual stimuli in general, or either visual or auditory language; Supplementary Figure S3A). In contrast, the face-selective rVTC fROI showed consistently higher activation to faces than to all conditions in the language (and VWFA) localizer across all thresholds (Supplementary Figure S3B), suggesting that this result was specific to word selectivity and not visual category-selective areas in general.

Second, we explored the possibility that there exists an amodal language region somewhere in the VTC, outside the boundaries of the VWFA parcel, possibly more anterior than the stereotypical location of the VWFA based on previous observations (e.g., Mummery et al., Citation2000; Schwartz et al., Citation2009); if so, perhaps this region would show selectivity to visual words as compared to other visual categories (note that this analysis is more of a reality check because such a region would have been picked out in the analysis searching for word selectivity across the VTC). In sum, we did not find a language-selective rVTC fROI that showed consistently higher activation to Sentences than to Nonword sequences (visually or auditorily presented) (Supplementary Figure S3C) even in the anterior part of the rVTC in EG. So, although EG’s language network was right-lateralized and showed typical selectivity to linguistic stimuli (Tuckute et al., Citation2022), we found no evidence of an inferior language region (i.e., within VTC) that showed visual word selectivity.

Is there any word selectivity in the spared left VTC?

Because we did not observe a right-lateralized VWFA and because the left VTC was largely intact, we also asked whether a canonical (univariate) lVWFA may have developed in EG (perhaps due to some specific visual features of word forms that are better represented in the left hemisphere/right visual field (e.g., Hsiao & Lam, Citation2013; Seghier & Price, Citation2011; Tadros et al., Citation2013). In the control group, we observed robust visual word selectivity in the lVWFA (as expected from previous studies), with Words eliciting greater activation than each of the other conditions: Words vs. Scrambled Words: t(24) = 7.302, pbh = 4.577 × 10−7; Words vs. Objects: t(24) = 6.213, pbh = 2.024 × 10−6; Words vs. Faces: t(24) = 6.870, pbh = 8.364 × 10−7; A). In contrast, we found no visual word selectivity in EG’s lVWFA (B): activation to Words was around baseline and lower than the response to other categories, although the differences did not reach significance (Words vs. Scrambled Words: t(9) = −1.509, p = 0.165; Words vs. Objects: t(9) =  −1.192, p = 0.264; Words vs. Faces: t(9) = −1.493, p = 0.170). Moreover, EG’s activation to Words was significantly lower than the control group’s (t(24) = −2.342, pbh = 5.558 × 10−2, modified t-test; 95% Bayesian CI [7.985 × 10−5, 0.042]). We did not observe word selectivity in EG across various thresholds used to define the lVWFA fROI (C), and we did not observe word selectivity in EG when we searched across the entire lVTC with varying thresholds (C). We also did not observe word selectivity when we used a less stringent contrast to define the lVWFA (Words vs. Scrambled Words; Supplementary Figure S2B) or even when we looked only within visually-responsive voxels in the lVTC (Supplementary Figure S2D). Finally, we performed the same analysis as above where we explored linguistic selectivity using a language localizer, and found no evidence of an amodal language fROI within the lVTC that might show visual word selectivity (Supplementary Figure S3D-F).

Figure 6. Responses to four conditions in the lVWFA and lFFA for EG and the control group. (A), Bar plots show mean PSCs to four conditions estimated in independent data within individually defined lVWFA fROIs (i.e., top 10% word-selective voxels within lVWFA parcel) for the control group. (B), Mean PSCs to the four conditions estimated in independent data within individually defined lVWFA fROIs for EG. Here and in E, the results are averaged across run combinations. (C), Parametrically decreasing the threshold for defining the lVWFA fROIs from the top 5% to 50% word-selective voxels within the lVWFA parcel for EG. Number of voxels in the defined lVWFA fROIs: 5% = 295 voxels, 50% = 2952 voxels. Here and in F, average PSCs across run combinations are shown for each threshold. (D), Mean PSCs across participants to the four conditions estimated in independent data within individually defined lFFA fROIs for the control group. €, Mean PSCs to four conditions estimated in independent data within individually defined lFFA fROIs for EG. (F), Parametrically decreasing the threshold for defining lFFA fROI from the top 5% to 50% face-selective voxels within the lFFA parcel in EG. Number of voxels in the defined lFFA fROIs: 5% = 70 voxels, 50% = 695 voxels. In the bar plots, dots correspond to individual data points (controls: n = 25 subjects; EG: n = 10 run combinations, from ten iterations). Horizontal bars reflect significant paired t-tests p < 0.05. Error bars in both the bar and line plots denote standard errors of the mean by participants (for the control group) and by run combinations (for EG). Words = Written Words; ScrW = Scrambled Words.

Figure 6. Responses to four conditions in the lVWFA and lFFA for EG and the control group. (A), Bar plots show mean PSCs to four conditions estimated in independent data within individually defined lVWFA fROIs (i.e., top 10% word-selective voxels within lVWFA parcel) for the control group. (B), Mean PSCs to the four conditions estimated in independent data within individually defined lVWFA fROIs for EG. Here and in E, the results are averaged across run combinations. (C), Parametrically decreasing the threshold for defining the lVWFA fROIs from the top 5% to 50% word-selective voxels within the lVWFA parcel for EG. Number of voxels in the defined lVWFA fROIs: 5% = 295 voxels, 50% = 2952 voxels. Here and in F, average PSCs across run combinations are shown for each threshold. (D), Mean PSCs across participants to the four conditions estimated in independent data within individually defined lFFA fROIs for the control group. €, Mean PSCs to four conditions estimated in independent data within individually defined lFFA fROIs for EG. (F), Parametrically decreasing the threshold for defining lFFA fROI from the top 5% to 50% face-selective voxels within the lFFA parcel in EG. Number of voxels in the defined lFFA fROIs: 5% = 70 voxels, 50% = 695 voxels. In the bar plots, dots correspond to individual data points (controls: n = 25 subjects; EG: n = 10 run combinations, from ten iterations). Horizontal bars reflect significant paired t-tests p < 0.05. Error bars in both the bar and line plots denote standard errors of the mean by participants (for the control group) and by run combinations (for EG). Words = Written Words; ScrW = Scrambled Words.

Similar to the analyses we performed for the rVTC, to test whether selectivity for other visual categories in the left VTC is typical in EG, we examined face selectivity in EG’s lFFA. EG’s face selectivity remained intact (E) and did not differ from that of the control group (D; t(24) = −0.481, p = 0.635, modified t-test; 95% Bayesian CI, [0.178 0.462]). EG’s lFFA fROI showed significantly higher responses to Faces than to other conditions (Faces vs. Words: t(9) = 40.347, pbh = 5.270 × 10−11; Faces vs. Scrambled Words: t(9) = 19.702, pbh = 1.036 × 10−8; Faces vs. Objects: t(9) = 29.689, pbh = 5.450 × 10−10; E), just like what we observed in the control group (Faces vs. Words: t(24) = 6.531, pbh = 2.814 × 10−6; Faces vs. Scrambled Words: t(24) = 6.529, pbh = 2.814 × 10−6; Faces vs. Objects: t(24) = 6.223, pbh = 1.975 × 10−6; D). As was the case for the rFFA, face selectivity was observed across all thresholds used to define the lFFA (F). When examining the strength of selectivity, we found that controls showed significant (compared to zero) word selectivity in the lVWFA (t(24) =  7.413, pbh = 1.774 × 10−7) and face selectivity in the lFFA (t(24) =  6.462, pbh = 1.107 × 10−6). Critically, when we compared the strength of the selectivity, we found that EG’s word selectivity in the lVWFA was lower than zero (although did not reach significant level: t(9) = −1.153, p = 0.861) and significantly lower than controls (t(24) =  −2.073, p = 0.049, modified t-test; 95% Bayesian CI, [7.918 × 10−5 0.066]). C shows that EG’s word selectivity in the lVWFA is at the tail of the control distribution and only one control did not show word selectivity in lVWFA (but unlike EG, this control showed no face selectivity either). Interestingly, even though EG’s face selectivity in the lFFA is not statistically higher than the average of controls (t(24) =  1.171, p = 0.253, modified t-test; 95% Bayesian CI, [0.039 0.236]), it was at the higher end of the control’s distribution with only three controls showing higher face selectivity than EG (D).

Altogether, the examination of EG’s right and left VTC suggests that without the typical left-hemisphere frontotemporal language network from birth—and presumably without the necessary connections between these areas and parts of the VTC—a canonical VWFA, a word-selective area, does not develop in either hemisphere.

Does the frontotemporal language network support visual word processing?

Finally, it is possible that while EG’s right or left VTC lacked visual word selectivity or general linguistic selectivity, perhaps parts of her amodal frontotemporal language network show canonical univariate selectivity to visual words. We invited EG back and collected fMRI data for the VWFA localizer again but with whole-brain coverage (see Supplementary Methods: Data acquisition parameters for the VWFA localizer with a whole-brain coverage (from the second session) for details). We first replicated our main results: we observed no canonical univariate VWFA in the left or right VTC but found a typical FFA in both hemispheres (Supplementary Figure S4). Critically, whole-brain coverage allowed us to ask whether there is evidence of visual word selectivity in the frontotemporal language network.

Using the language parcels (Fedorenko et al., Citation2010) as our search space, we identified voxels that showed higher activation to written Words > Others at different thresholds (from top 1% to top 10%). Then in independent runs, we extracted activation to conditions in both the VWFA and visual and auditory language localizers. We failed to find any voxels in the frontotemporal language regions that showed visual Word-selective responses like those typically observed in the lVWFA (A-C): even though voxels showed higher activation to written Words than to Objects, activation to written Words was not differentiated from Faces and Scrambled Words at all thresholds; moreover, activation for even auditorily presented stimuli were higher than the activation to visual Words (even though the fROIs were chosen with the visual Word contrast) again suggesting that these language regions did not take on the properties of a typical VWFA. We also identified language-selective fROIs by contrasting visually presented English sentences vs. Nonword sequences (e.g., Fedorenko et al., Citation2010; Tuckute et al., Citation2022); consistent with Tuckute et al. (Citation2022), we also found that EG’s language fROIs was reorganized to the RH and showed significantly higher activation to English sentences than to Nonword sequences in both temporal and frontal cortex. Specifically, the right temporal and frontal language-selective fROIs (D, 7E) showed preferred activation to high-level linguistic information, indicated by higher responses to English sentences (solid lines) than to Nonword sequences (dash lines) in either visual (yellow) or auditory (orange) modality. Additionally, they did not show distinct activation to visual Words vs. other visual categories suggesting that these language fROIs are indeed selective to high-level linguistic information rather than orthographic information and do not take the place of a typical lVWFA.

Figure 7. Responses to conditions in both the VWFA and language localizers in the frontotemporal language network for EG. (A-C), Mean PSCs in word-selective voxels (Words > Others) at different thresholds in the right temporal (A), right frontal (B) and left frontal language parcels (C). (D-F), Mean PSCs in language-selective voxels (Words > Others) at different thresholds in the right temporal (D), right frontal (E) and left frontal language parcels (F). Parametrically decreasing the threshold from the top 1% to 10% within each language parcel (i.e., the search space). Mean PSCs across run combinations (from 10 iterations for the VWFA task and 6 iterations for the language task) are shown for each threshold. Error bars denote standard errors of the mean by run combinations for EG. Words = Written Words; ScrW = Scrambled Words. Evis, visually presented English sentences; Nvis, visually presented Nonword sequences; Eaud, auditorily presented English sentences; Naud, auditorily presented Nonword sequences.

Figure 7. Responses to conditions in both the VWFA and language localizers in the frontotemporal language network for EG. (A-C), Mean PSCs in word-selective voxels (Words > Others) at different thresholds in the right temporal (A), right frontal (B) and left frontal language parcels (C). (D-F), Mean PSCs in language-selective voxels (Words > Others) at different thresholds in the right temporal (D), right frontal (E) and left frontal language parcels (F). Parametrically decreasing the threshold from the top 1% to 10% within each language parcel (i.e., the search space). Mean PSCs across run combinations (from 10 iterations for the VWFA task and 6 iterations for the language task) are shown for each threshold. Error bars denote standard errors of the mean by run combinations for EG. Words = Written Words; ScrW = Scrambled Words. Evis, visually presented English sentences; Nvis, visually presented Nonword sequences; Eaud, auditorily presented English sentences; Naud, auditorily presented Nonword sequences.

Distributed neural representation of visual words: multivariate pattern analysis (MVPA)

Finally, to explore whether words may be represented in a distributed fashion in EG, we performed a series of multivariate pattern analyses (MVPA). We first examined the representational similarity matrices (RSMs) in the entire VTC (see Methods) to investigate whether multivariate representational structure for visual categories was preserved in EG. Indeed, we found that EG’s rVTC RSM was strongly and significantly correlated with that of the control group (r = 0.937, p = 9.100 × 10−3 (permutation test); Supplementary Figure S5). We tested whether this correlation between EG and the control group was different from the correlations between any given control individual and the rest of the control group (see Methods). Single case comparisons showed that the RSM correlation for EG vs. controls did not significantly differ from the within-controls correlations (t(24) = 0.821, p = 0.420, Bayesian 95% CI [0.092 0.345]). Similar results were found for the lVTC: the RSMs of EG and the control group were strongly and significantly correlated (r = 0.846, p = 9.300 × 10−3), and the correlation for EG vs. controls did not significantly differ from the within-controls correlations (t(24) = 0.676, p = 0.506, Bayesian 95% CI [0.121 0.388]).

We then asked whether EG’s “VWFA” may contain voxels that show distinct distributed activation patterns to visual words. Specifically, as described in Methods, we searched for voxels that satisfied the following criteria: the searchlight around a given voxel should show (1) distinctive response patterns to Words vs. other categories (e.g., the Words-Words correlation should be higher than the Words-Faces correlation); and (2) stronger within-category correlations for the preferred category (e.g., the Words-Words correlation should be higher than the Faces-Faces, Objects-Objects, and Scrambled Words-Scrambled Words correlations). Previous studies have shown distributed representations within category-selective regions (e.g., the FFA) of non-preferred categories (e.g., places), and debate is ongoing over whether this information has functional relevance (e.g., Kanwisher, Citation2010). Our second criterion was included to identify voxels that show more stable multivariate representations for the category of interest (e.g., Words) compared to other categories. Indeed, we identified a set of voxels within the rVWFA and lVWFA parcels that showed a reliable distributed code for words in both controls and EG (Supplementary Figure S5A, S5C; Supplementary Table S4). In addition, mirroring the univariate analyses, we also identified a set of voxels that showed a reliable distributed code for faces within the rFFA and lFFA parcels in both controls and EG (Supplementary Figure S5B, S5D; Supplementary Table S4).

Critically, to test whether these distributed responses to visual words differ between EG and the controls, we subtracted the average within-category correlations for all non-word categories from the within-category correlations for Words. This difference tells us how much stronger the within-category correlation is for the preferred compared to the non-preferred categories. EG showed comparable within-category correlation differences to the controls in both the rVWFA (t(24) = 0.986, p = 0.334, modified t-test, 95% Bayesian CI [0.062, 0.287]) and lVWFA (t(24) = 0.255, p = 0.801, modified t-test, 95% Bayesian CI [0.276, 0.500]) (A). Similar within-category correlation differences were also observed between EG and the controls in the rFFA (t(24) = −0.907, p = 0.373; 95% Bayesian CI [0.075, 0.311]) and lFFA (t(24) = 0.546, p = 0.590; 95% Bayesian CI [0.159, 0.439]) (B).

Finally, we explored potential hemispheric asymmetries with respect to multivariate selectivity for words and faces. Left-hemisphere dominance for words and right-hemisphere dominance for faces are well established with univariate measures. Do multivariate patterns also show these asymmetries? We calculated the laterality for words and faces for VWFA and FFA respectively (see Methods). In controls, mirroring univariate results found here and in previous literature, we observed a larger number of multivariate searchlight centroids that showed greater multivariate Word selectivity in the left than in the right VWFA (the laterality index was significantly larger than 0: t(24) = 3.442, pbh = 0.003) (C). And only five out of 25 controls showed right-lateralized multivariate word representation (D). Importantly, in the case of EG, where the univariate preference for words is lacking on the left hemisphere, the opposite pattern was observed, with a larger number of searchlight centroids showing multivariate Word selectivity in the right than the left (laterality of Word selectivity within the VWFA was significantly smaller than 0, i.e., right-lateralized (t(9) = −2.460, pbh = 0.018)) (C). For Faces, consistent with previously reported univariate observations, we found that controls’ multivariate face selectivity is right lateralized (compared to 0: controls, t(24) = −2.794, pbh = 0.010) (C); EG, however, displayed the opposite pattern: significantly more multivariate face-selective centroids in the left FFA (t(9) = 6.634, pbh = 1.909 × 10−4) (C); and interestingly, even though eight out of 25 controls showed left-lateralized multivariate face selectivity, EG’s FFA laterality index was at the tail of the control distribution (E).

Figure 8. Results of the multivariate pattern analysis (MVPA). (A) Within-category correlation differences between preferred (i.e., Words-Words) and non-preferred (i.e., average within-category correlation of non-word conditions) conditions for EG and the controls in the rVWFA and lVWFA parcels. Here and in B, correlation values were Fisher’s z-transformed. (B) Within-category correlation differences between preferred (i.e., Faces-Faces) and non-preferred (i.e., average within-category correlation of non-face conditions) conditions for EG and the controls in the rFFA and lFFA parcels. (C) Laterality index for the number of multivariate searchlights for VWFA and FFA for both controls and EG. Asterisk (*) denotes significant laterality index (compared to zero). (D-E), Histogram was used to show the word and face laterality indices for both controls and EG. Each bar represents one participant and participants are ordered by their selectivity indices from lowest to highest.

Figure 8. Results of the multivariate pattern analysis (MVPA). (A) Within-category correlation differences between preferred (i.e., Words-Words) and non-preferred (i.e., average within-category correlation of non-word conditions) conditions for EG and the controls in the rVWFA and lVWFA parcels. Here and in B, correlation values were Fisher’s z-transformed. (B) Within-category correlation differences between preferred (i.e., Faces-Faces) and non-preferred (i.e., average within-category correlation of non-face conditions) conditions for EG and the controls in the rFFA and lFFA parcels. (C) Laterality index for the number of multivariate searchlights for VWFA and FFA for both controls and EG. Asterisk (*) denotes significant laterality index (compared to zero). (D-E), Histogram was used to show the word and face laterality indices for both controls and EG. Each bar represents one participant and participants are ordered by their selectivity indices from lowest to highest.

Discussion

Case studies provide valuable insight for understanding the functional organization of the human brain and the patterns of reorganization that follow neurological damage (Vaidya et al., Citation2019). In the current study, we had a unique opportunity to collect fMRI data from an individual (EG) born without her left superior temporal lobe in order to ask whether category-selective responses to visual words within the ventral temporal cortex would be affected. Specifically, we asked: in the absence of a typical left-hemisphere language network, does word selectivity emerge, and if so, where? Does the VWFA emerge in the right VTC given that EG’s language network is located in the right hemisphere (Tuckute et al., Citation2022)? Or do we instead, or additionally, observe word selectivity within the spared lVTC? Surprisingly, we found no univariate word selectivity within the rVWFA, the right homotope of the canonical lVWFA, and even in an expanded search space of the entire rVTC. We also did not observe univariate word-selective responses in the lVWFA or in any other spared parts of the lVTC. Moreover, the response magnitude to Words (relative to baseline) was significantly lower in EG’s lVWFA than that of controls. Importantly, this lack of category selectivity was specific to Words: selective responses to Faces remained intact in EG’s rFFA and lFFA.

Absent univariate word selectivity, we then explored multivariate representations of visual words in EG. We found that EG’s VTC showed an overall similar representational structure to that of the control group, and that, like the control group, EG had a set of voxels whose local neighbourhoods robustly differentiated between Words and other visual categories in their patterns of response. Critically, these voxels also showed a higher within-category correlation (across runs) for Words compared to the within-category correlations for other categories, and the degree of this “multivariate selectivity” was similar between EG and the controls. Interestingly, however, in contrast to the typically observed left-hemispheric dominance for words and right-hemispheric dominance for faces, EG had a larger proportion of voxels that showed multivariate selectivity for words in her right than her left VWFA, and a larger proportion of voxels that showed multivariate selectivity for faces in her left FFA than her right FFA.

Altogether, the current study suggests that in the absence of a typical left-lateralized language network—at least in cases of early left-hemispheric damage, when language processing has no choice but to localize to the right hemisphere—neural selectivity for visual words is, or at least can be, highly atypical. In our participant with extensive left-hemispheric temporal damage and a right-lateralized language network, we observed no word selectivity in the univariate analyses in either the right or the spared left VTC, replicated across scanning sessions (5 years apart). The absence of such selectivity, combined with EG’s intact reading ability, suggests that successful orthographic processing may depend on a distributed and more right-lateralized neural representation than that observed in typical literate adults.

Canonical univariate word selectivity may not emerge when a left-hemisphere language network cannot develop normally

The interesting case of EG allowed us to investigate how visual word processing within the VTC can be affected by a congenital or early left-hemisphere lesion outside of visual cortex. Our results provide the first evidence of atypical visual word selectivity in the VTC when the left-lateralized language cortex is missing, and when the language network consequently develops in the right hemisphere (Tuckute et al., Citation2022) during the early stages of language learning and prior to learning to read. Our study also suggests that even with some remaining anatomical connections between the spared lVTC and the frontal and temporal areas (presumably via local U-fibres or remaining long-range fibres), the lVWFA will not emerge at the stereotypical location when the left hemisphere does not support language processing.

How does a left language network contribute to the development of the VWFA? One possibility is that in a typical brain, the site of the putative VWFA is predisposed to written scripts via intrinsic co-activation with speech and/or language/semantic areas in the left temporal (and maybe frontal) cortex. The processing of speech sounds is tightly linked to reading, and impaired coding of speech sounds (e.g., phonological awareness) is often considered a precursor of dyslexia (e.g., Pennington & Bishop, Citation2008; Shaywitz et al., Citation2002). In the absence of regions that typically support speech processing (within the left superior and middle temporal gyri (STG and MTG); Raschle et al., Citation2012), EG’s left VTC lacked early interactions with these regions. Further, a typical lVWFA communicates visual orthographic information to higher-level left temporal (and maybe frontal) regions that integrate visual and auditory information, like the amodal language regions that process lexical-semantic and combinatorial linguistic information (e.g., Fedorenko et al., Citation2020) and the regions that support abstract conceptual processing (e.g., Binder et al., Citation2009; Patterson et al., Citation2007)—both sets of areas that were also missing in EG’s left hemisphere. Consistent with the idea that connections between the left VTC and the ipsilateral high-level language areas may be critical for the emergence of the lVWFA, a recent study found that in newborns, the putative site of the lVWFA already shows preferential functional connectivity to the areas that will later respond to high-level linguistic information, compared to adjacent regions (Li et al., Citation2020); this pre-existing functional coupling may further strengthen during language and reading acquisition. Because EG was missing both i) speech-responsive areas, and ii) higher-level language/conceptual areas in her left hemisphere, we cannot evaluate the relative contributions of these two sets of areas and their connections with the VTC to the emergence of a canonical (univariate) VWFA. We speculate that both may be important.

Right-hemispheric neural correlates that support visual word processing

Because EG’s language network resides in the right hemisphere (Tuckute et al., Citation2022), we expected to find a VWFA in the right VTC. Surprisingly, no univariate word selectivity was observed in EG’s VTC, and this result could not be explained by either a lack of functional specialization anywhere in the rVTC (intact face selectivity was observed), or our particular procedure for defining the VWFA (the results held across a range of thresholds, when using a larger search space, and across different contrasts). Interestingly, neurotypical individuals with right-dominant language activation sometimes also lack a VWFA in the right VTC (e.g., Cai et al., Citation2008; Van der Haegen et al., Citation2012); specifically, in these individuals lVTC appears to be engaged during word recognition when the language region is right dominant compared (i.e., VWFA stays in the left even if language is on the right), presumably due to stronger frontotemporal anatomical connections in the LH than in the RH. Supporting this observation, previous studies have found that both the volume and number of streamlines were left lateralized (i.e., larger on the left) for the long segment of the AF (De Schotten et al., Citation2011) that connects ventral temporal cortex and the frontal cortex. Thus, any language activation on the left (even if it is non-dominant) would engage the lVTC for reading. In the case of EG, her left-hemisphere language network was altogether lacking and her right-hemisphere language network may have lacked critical early interactions with the rVTC (perhaps due to weaker and/or fewer anatomical connectivity between frontal and temporal cortex on the right as previously reported (Powell et al., Citation2006)) which may have resulted in the atypical word selectivity (in both right and left) that we observed here.

Some have also argued that the development of word-selective cortex directly competes with face-selective cortex for neural resources, thus contributing to right-hemispheric dominance for face processing and left-hemispheric dominance for word processing in neurotypical individuals (e.g., Behrmann & Plaut, Citation2015; Dehaene & Cohen, Citation2007). Interestingly, EG showed typical face selectivity in the right FFA, not different from controls. It is therefore possible that a functionally specialized word-selective area identified via univariate analysis failed to emerge in EG’s rVTC because the relevant cortical tissue had already been “assigned” to face processing, perhaps due to some specific visual features of faces that are better represented in the right hemisphere/left visual field (e.g., holistic face processing; Li et al., Citation2017; Rossion et al., Citation2000). Interestingly, although the strength of the face selectivity did not differ between EG and controls in the rFFA, EG showed significantly higher face selectivity in the lFFA. In line with this finding, more multivariate-selective face voxels were present in EG’s lFFA, compared to her rFFA, in sharp contrast to controls. The functional significance of the latter is at present unclear, but can be explored in future work. On the other hand, EG had more multivariate-selective word voxels in her rVWFA than her lVWFA, presumably related to the fact that her language network resides in the right hemisphere. These data suggest that the hemispheric dominance of the language network drives the laterality of visual processing in the VTC (be it implemented focally at voxel level, or in a distributed fashion), at least for words, but perhaps also for faces (Behrmann & Plaut, Citation2020; Dehaene & Cohen, Citation2007).

Further, we found no evidence in support of theories that propose that the VWFA does not process orthography (despite evidence in our neurotypical control population) and that it instead processes linguistic stimuli in much the same way as the lateral frontotemporal amodal language network (see Büchel et al., Citation1998; Price & Devlin Citation2011). We investigated multiple ways of defining Word or Language-selective fROIs in the right and left VTC, and found no evidence that EG’s VTC is selective to either general linguistic stimuli or visual Word stimuli despite normal reading performance.

No amodal linguistic processing in the VTC that supports reading and no word-selective response observed outside the VTC (i.e., in frontotemporal language areas)

Some have claimed that even in neurotypical individuals, responses to visual words in the VTC do not reflect orthographic processing (Price & Devlin, Citation2003) but that, instead, the VWFA is part of the language network (Price & Devlin, Citation2011) and thus should show selectivity for linguistic stimuli in general (e.g., respond to nonvisual abstract meaning). For example, studies have shown that the VWFA could be activated/modulated by auditory speech due to top-down cross-model processing (e.g., Yoncheva et al., Citation2010; Cohen et al., Citation2021; Dehaene et al., Citation2010). Note that neither of these studies directly compared the activation strength of visual words vs. non-linguistic stimuli, thus the VWFA may still be functionally distinct from frontotemporal language regions but its response to nonvisual linguistic stimuli could simply be driven by communication/co-activation between VWFA and language regions. However, perhaps for populations with differently organized brains, this hypothesis has merit. For example, Kim et al. (Citation2017) found that in congenitally blind individuals, whose visual cortex has long been known to show reorganization (e.g., Lane et al., Citation2015; Röder et al., Citation2002; Striem-Amit & Amedi, Citation2014), the anatomical location of the (putative) VWFA responded to both Braille and to a grammatical complexity manipulation for auditorily presented sentences. However, we did not find language-selective voxels within the VTC (Supplementary Figure S3C,F).

Further, although we show that normal reading ability is possible in the absence of focal selectivity at voxel level for word processing in the VTC, there may be other pathways or neural structures outside of the VTC that are important for EG’s reading ability. For example, Seghier et al. (Citation2012) reported a patient who acquired dyslexia following extensive left ventral occipitotemporal cortex (LvOT) resection, but then regained reading ability; they provided evidence to suggest that the patient’s reading ability was now supported via a direct connection between the occipital visual cortex and the left superior temporal sulcus (STS) without the involvement of the ventral visual stream. It is therefore possible that the STS in the language-dominant hemisphere (i.e., the right hemisphere in EG) may support visual word processing through its connections with the occipital cortex. Thus, to supplement our main analyses, which were restricted to the ventral temporal cortex, we invited EG back and obtain the fMRI data from the same VWFA task with a whole-brain coverage. We found no evidence of significantly greater responses to visual Words as compared to other visual stimuli in the lateral temporal or frontal cortex, either when searching for word-selective voxels directly (by Words > Others) or looking for word selectivity in language fROIs (canonically defined with English sentences > Nonword sequences). These results further confirmed the lack of univariate Word-selective responses in EG.

Multivariate responses to words

Despite the absence of typical univariate word selectivity, EG’s VTC was similar in its overall representation similarity structure—across different visual stimulus classes—to that of the control group. Consistent with our results, Liu et al. (Citation2019) also found typical representational structure in category-selective regions after resections within or outside VTC; in addition, a recent study found mature representational similarity structures in children (5–7 year-olds) with no univariate selectivity, suggesting that distributed representations precede category selectivity (M. A. Cohen et al., Citation2019). Our results provide another case where a typical multivariate representational structure is observed in the absence of univariate selectivity for words. Moreover, our results showed that EG (like typical adults) displayed distinctive activation patterns within the general vicinity of both right and left VWFA that could distinguish visual words from other conditions, with higher within- than between-category correlations (Haxby et al., Citation2001) and more robust (consistent over time) representations of Words than other conditions. Supporting the idea that multivariate selectivity for words may be functionally useful, Stevens et al. (Citation2017) found that the VWFA (defined in individual participants using a standard univariate contrast) discriminated words from pseudowords using MVPA, and did so more strongly than other control regions (e.g., the FFA), suggesting that MVPA analyses may reflect similar underlying selectivity to univariate analyses.

It remains unknown what the relationship is between univariate and multivariate representations. Univariate selectivity may partially depend on a connectivity scaffold or other genetically defined instructions to determine the location of functional specialized areas (Deen et al., Citation2017; Kamps et al., Citation2020; Li et al., Citation2020). In contrast, distributed representations like those found in EG may rely less on a connectivity scaffold. Do distributed multivariate patterns reflect a fundamentally different neural code from that associated with univariate activations in a functionally specialized region? Some have argued that the answer is no, and that multivariate analyses may simply be more sensitive than univariate approaches given that they consider the heterogeneity of response across voxels within a region as well as potential subthreshold voxels (Davis et al., Citation2014). On the other hand, multivariate representations may be truly distinct and reflect information derived from bottom-up visual statistics to a greater extent. Thus perhaps in typical development, representational similarity structures precede univariate selectivity (M. A. Cohen et al., Citation2019) with an early landscape of broadly-tuned neural populations giving rise to the later emergence of univariate selectivity- but only after experience-dependent interactions with higher-level areas (e.g., the language cortex in case of the VWFA). Consistent with this idea, in our controls, even though no reliable univariate word selectivity in the rVWFA was observed, we nevertheless found multivariate representation for words within the rVWFA. A few controls also showed right-lateralized multivariate word representations or left-lateralized face representations, similar to what we have observed in EG. Critically however, typical literate adults showed univariate word selectivity in either the left or right VWFA (in addition to multivariate selectivity), whereas in EG, we observed only multivariate selectivity for words. This result, along with previous developmental work showing multivariate selectivity in the absence of univariate responses (Cohen et al., Citation2019; Dehaene-Lambertz et al., Citation2018), suggests that it is perhaps possible that strong univariate category selectivity follows initial multivariate representations in visual cortex.

Nevertheless, it remains unknown whether this multivariate representation in the rVWFA, or multivariate patterns in general, contributes to reading behaviour in controls, or is just epiphenomenal; however, even if multivariate representations do not contribute to reading behaviour in neurotypical adults, these representations may be an alternative mechanism that supports this behaviour in cases like EG where univariate categorical selectivity never developed. Perhaps due to weaker connectivity between language and visual cortex on the right than on the left (cf. the left hemisphere; e.g., Li et al., Citation2020; Powell et al., Citation2006), EG’s right visual cortex lacked critical early interaction with her still-developing right language cortex, and could not hone univariate word selectivity within a spatially specific location, resulting in a more distributed representation for visual words.

Limitations and future directions

Overall, our study provides unique evidence that in the absence of a typical left-hemisphere language network—at least in cases of early left-hemisphere damage—the canonical word-selective lVWFA as identified via univariate methods may not develop. Some limitations are worth noting. First, our results raise an interesting question about the behavioural relevance of category-selective regions. Previous studies have found that the strength of univariate category-selective selectivity (or lack thereof) accounted for performance differences in various object recognition tasks (e.g., Furl et al., Citation2011; Grill-Spector et al., Citation2004; Huang et al., Citation2014; Song et al., Citation2015). The VWFA exists due to reading and literacy and is not observed in illiterate individuals (e.g., Dehaene et al., Citation2015; Saygin et al., Citation2016) and disrupted activity in the VWFA is observed in dyslexic individuals (e.g., Centanni et al. Citation2019; Maisog et al., Citation2008; Paulesu et al., Citation2014; Richlan et al., Citation2011). Additionally, lesions or disruption of the VWFA with brain stimulation, results in failures in word recognition (Hirshorn et al., Citation2016). All this evidence underscores the behavioural importance of the VWFA in reading. The unique case of EG, who has normal reading ability without univariate word selectivity, suggests an alternative mechanism that might support orthographic processing. Here we examined one possibility that visual word processing might be represented in a distributed manner. Future work can further investigate this possibility. For example, it would be interesting to explore how multivariate representations are read out by downstream brain regions for further processing (Williams et al., Citation2007), or to further test the link between behavioural performance and multivariate representations. A previous study showed that performance in a visual search task among different pairs of object categories could be predicted from underlying neural similarity structures (Cohen et al., Citation2016). At this point, it is unclear how multivariate representations for words contribute to EG’s reading ability, and whether the right-hemisphere representations are more important than the left-hemisphere ones—questions that will be a focus of future studies. Second, EG’s missing left superior temporal cortex led to right-lateralized speech and language processing. As discussed earlier, it remains unclear whether speech and higher-level language areas contribute equally to the emergence of the lVWFA, and whether temporal language regions may be more important than the frontal ones. Third, we could not identify a univariate visual word region (i.e., a VWFA-like region, showing characteristic visual word selectivity) in the right frontotemporal cortex as well as the left frontal cortex in EG; moreover, EG’s language-selective cortex, identified by contrasting English sentences verses nonsense word strings, did not show preferential responses to visual words compared to other visual categories. It would be interesting to look at these patterns in neurotypical populations. For example, how does language cortex respond to visual words compared to other visual stimuli (with semantic meaning), nonword stimuli or sentences? Future studies can explore this question with an experimental design better tailored to ask these questions. Finally, here we reported a single case that sheds new light on the role of speech and language areas in developing a VWFA. However, conclusions that are drawn based on single case study need to take various factors into consideration. For example, do the results for the case study reflect the lesion itself or are we observing typical individual variability despite the lesion? Similar cases and data from other methods like noninvasive brain stimulation will help us better understand the causal role of language cortex in developing visual word selectivity.

Acknowledgements

We would like to thank our participant – EG – for her time and willingness to participate in brain research. We thank Greta Tuckute, Alexander Paunov, and Benjamin Lipkin for their help in coordinating the collaboration as well as their suggestions for data analysis. We also thank Idan Blank and Zachary Mineroff for help with EG’s fMRI and behavioural data collection, Rachel Ryskin for sharing the data from the speeded reading task, and members of Saygin Developmental Cognitive Neuroscience Lab for helping with control data collection. We would like to acknowledge the Athinoula A. Martinos Imaging Center at the McGovern Institute for Brain Research, MIT and the support from Center for Cognitive and Behavioral Brain Imaging (CCBBI) and Ohio Supercomputer Center (OSC). E.F. was supported by NIH awards R01 DC016607, R01 DC016950, and research funds from the McGovern Institute for Brain Research, the Department of Brain and Cognitive Sciences, and the Simons Center for the Social Brain at MIT. Z.M.S. was supported by the Alfred P. Sloan Foundation and OSU’s Arts & Sciences College and Chronic Brain Injury initiative.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data and code availability statement

The data and code that support the findings of this study are available from the corresponding author upon request.

Additional information

Funding

This work was supported by Alfred P. Sloan Research Fellowship; National Institutes of Health: [Grant Number R01 DC016950,R01-DC016607].

References

  • Asaridou, S. S., Demir-Lira, Ö. E., Goldin-Meadow, S., Levine, S. C., & Small, S. L. (2020). Language development and brain reorganization in a child born without the left hemisphere. Cortex, 127, 290–312. http://doi.org/10.1016/j.cortex.2020.02.006
  • Baker, C., Liu, J., Wald, L., Kwong, K. K., Benner, T., & Kanwisher, N. (2007). Visual word processing and experiential origins of functional selectivity in human extrastriate cortex. Proceedings of the National Academy of Sciences, 104(21), 9087–9092. https://www.pnas.org/content/104/21/9087.
  • Barttfeld, P., Abboud, S., Lagercrantz, H., Adén, U., Padilla, N., Edwards, A. D., Cohen, L., Sigman, M., Dehaene, S., & Dehaene-Lambertz, G. (2018). A lateral-to-mesial organization of human ventral visual cortex at birth. Brain Structure and Function, 223(7), 3107–3119. https://doi.org/10.1007/s00429-018-1676-3
  • Behrmann, M., & Plaut, D. C. (2013). Distributed circuits, not circumscribed centers, mediate visual recognition. Trends in Cognitive Sciences, 17(5), 210–219. https://doi.org/10.1016/j.tics.2013.03.007
  • Behrmann, M., & Plaut, D. C. (2015). A vision of graded hemispheric specialization: Graded hemispheric specialization. Annals of the New York Academy of Sciences, 1359(1), 30–46. https://doi.org/10.1111/nyas.12833
  • Behrmann, M., & Plaut, D. C. (2020). Hemispheric organization for visual object recognition: A theoretical account and empirical evidence. Perception, 49(4), 373–404. https://doi.org/10.1177/0301006619899049
  • Binder, J. R., Desai, R. H., Graves, W. W., & Conant, L. L. (2009). Where is the semantic system? A critical review and meta-analysis of 120 functional neuroimaging studies. Cerebral Cortex, 19(12), 2767–2796. https://doi.org/10.1093/cercor/bhp055
  • Bishop, D. (2003). Test for Reception of Grammar version 2 (TROG-2).
  • Bouhali, F., Thiebaut de Schotten, M., Pinel, P., Poupon, C., Mangin, J.-F., Dehaene, S., & Cohen, L. (2014). Anatomical connections of the visual word form area. Journal of Neuroscience, 34(46), 15402–15414. https://doi.org/10.1523/JNEUROSCI.4918-13.2014
  • Büchel, C., Price, C., & Friston, K. (1998). A multimodal language region in the ventral visual pathway. Nature, 394(6690), 274–277. http://doi.org/10.1038/28389.
  • Cai, Q., Lavidor, M., Brysbaert, M., Paulignan, Y., & Nazir, T. A. (2008). Cerebral lateralization of frontal lobe language processes and lateralization of the posterior visual word processing system. Journal of Cognitive Neuroscience, 20(4), 672–681. https://doi.org/10.1162/jocn.2008.20043
  • Cai, Q., Paulignan, Y., Brysbaert, M., Ibarrola, D., & Nazir, T. A. (2010). The left ventral occipito-temporal response to words depends on language lateralization but not on visual familiarity. Cerebral Cortex, 20(5), 1153–1163. https://doi.org/10.1093/cercor/bhp175
  • Centanni, T. M., Norton, E. S., Ozernov-Palchik, O., Park, A., Beach, S. D., Halverson, K., Gaab, N., & Gabrieli, J. D. E. (2019). Disrupted left fusiform response to print in beginning kindergartners is associated with subsequent reading. NeuroImage: Clinical, 22, 101715. http://doi.org/10.1016/j.nicl.2019.101715
  • Cohen, L., Dehaene, S., McCormick, S., Durant, S., & Zanker, J. M. (2016). Brain mechanisms of recovery from pure alexia: A single case study with multiple longitudinal scans. Neuropsychologia, 91, 36–49. https://doi.org/10.1016/j.neuropsychologia.2016.07.009
  • Cohen, L., Martinaud, O., Lemer, C., Lehéricy, S., Samson, Y., Obadia, M., Slachevsky, A., & Dehaene, S. (2003). Visual word recognition in the left and right hemispheres: Anatomical and functional correlates of peripheral alexias. Cerebral Cortex, 13(12), 1313–1333. https://doi.org/10.1093/cercor/bhg079
  • Cohen, L., Salondy, P., Pallier, C., & Dehaene, S. (2021). How does inattention affect written and spoken language processing? Cortex, 138, 212–227. https://doi.org/10.1016/j.cortex.2021.02.007
  • Cohen, M. A., Alvarez, G. A., Nakayama, K., & Konkle, T. (2016). Visual search for object categories is predicted by the representational architecture of high-level visual cortex. Journal of Neurophysiology, 117(1), 388–402. https://doi.org/10.1152/jn.00569.2016
  • Cohen, M. A., Dilks, D. D., Koldewyn, K., Weigelt, S., Feather, J., Kell, A. J. E., Keil, B., Fischl, B., Zöllei, L., Wald, L., Saxe, R., & Kanwisher, N. (2019). Representational similarity precedes category selectivity in the developing ventral visual pathway. NeuroImage, 197, 565–574. https://doi.org/10.1016/j.neuroimage.2019.05.010
  • Crawford, J. R., & Garthwaite, P. H. (2007). Comparison of a single case to a control or normative sample in neuropsychology: Development of a Bayesian approach. Cognitive Neuropsychology, 24(4), 343–372. https://doi.org/10.1080/02643290701290146
  • Crawford, J. R., Garthwaite, P. H., & Howell, D. C. (2009). On comparing a single case with a control sample: An alternative perspective. Neuropsychologia, 47(13), 2690–2695. https://doi.org/10.1016/j.neuropsychologia.2009.04.011
  • Crawford, J. R., & Howell, D. C. (1998). Comparing an individual’s test score against norms derived from small samples. The Clinical Neuropsychologist, 12(4), 482–486. https://doi.org/10.1076/clin.12.4.482.7241
  • Davis, T., LaRocque, K. F., Mumford, J. A., Norman, K. A., Wagner, A. D., & Poldrack, R. A. (2014). What do differences between multi-voxel and univariate analysis mean? How subject-, voxel-, and trial-level variance impact fMRI analysis. NeuroImage, 97, 271–283. https://doi.org/10.1016/j.neuroimage.2014.04.037
  • Deen, B., Richardson, H., Dilks, D. D., Takahashi, A., Keil, B., Wald, L. L., Kanwisher, N., & Saxe, R. (2017). Organization of high-level visual cortex in human infants. Nature Communications, 8(1), 13995. https://doi.org/10.1038/ncomms13995
  • Dehaene, S., & Cohen, L. (2007). Cultural recycling of cortical maps. Neuron, 56(2), 384–398. https://doi.org/10.1016/j.neuron.2007.10.004
  • Dehaene, S., & Cohen, L. (2011). The unique role of the visual word form area in reading. Trends in Cognitive Sciences, 15(6), 254–262. https://doi.org/10.1016/j.tics.2011.04.003
  • Dehaene, S., Cohen, L., Morais, J., & Kolinsky, R. (2015). Illiterate to literate: Behavioural and cerebral changes induced by reading acquisition. Nature Reviews Neuroscience, 16(4), 234–244. https://doi.org/10.1038/nrn3924
  • Dehaene, S., Pegado, F., Braga, L. W., Ventura, P., Filho, G. N., Jobert, A., … Cohen, L. (2010). How learning to read changes the cortical networks for vision and language. Science, 330(6009), 1359–1364. https://doi.org/10.1126/science.1194140
  • Dehaene-Lambertz, G., Monzalvo, K., & Dehaene, S. (2018). The emergence of the visual word form: Longitudinal evolution of category-specific ventral visual areas during reading acquisition. PLOS Biology, 16(3), e2004103. https://doi.org/10.1371/journal.pbio.2004103
  • De Schotten, M. T., Bizzi, A., Dell'Acqua, F., Allin, M., Walshe, M., Murray, R., … Catani, M. (2011). Atlasing location, asymmetry and inter-subject variability of white matter tracts in the human brain with MR diffusion tractography. Neuroimage, 54(1), 49–59. https://doi.org/10.1016/j.neuroimage.2010.07.055
  • Desikan, R. S., Ségonne, F., Fischl, B., Quinn, B. T., Dickerson, B. C., Blacker, D., Buckner, R. L., Dale, A. M., Maguire, R. P., Hyman, B. T., Albert, M. S., & Killiany, R. J. (2006). An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest. NeuroImage, 31(3), 968–980. https://doi.org/10.1016/j.neuroimage.2006.01.021
  • Downing, P. E., Jiang, Y., Shuman, M., & Kanwisher, N. (2001). A cortical area selective for visual processing of the human body. Science, 293(5539), 2470–2473. https://doi.org/10.1126/science.1063414
  • Dunn, L. M., & Dunn, D. M. (2007). The peabody picture vocabulary test, fourth edition. NCS Pearson, Inc.
  • Epstein, R., & Kanwisher, N. (1998). A cortical representation of the local visual environment. Nature, 392(6676), 598–601. https://doi.org/10.1038/33402
  • Fedorenko, E., Behr, M. K., & Kanwisher, N. (2011). Functional specificity for high-level linguistic processing in the human brain. Proceedings of the National Academy of Sciences, 108(39), 16428–16433. https://doi.org/10.1073/pnas.1112937108
  • Fedorenko, E., Blank, I. A., Siegelman, M., & Mineroff, Z. (2020). Lack of selectivity for syntax relative to word meanings throughout the language network. Cognition, 203, 104348. http://doi.org/10.1016/j.cognition.2020.104348
  • Fedorenko, E., Hsieh, P. J., Nieto-Castañón, A., Whitfield-Gabrieli, S., & Kanwisher, N. (2010). New method for fMRI investigations of language: Defining ROIs functionally in individual subjects. Journal of Neurophysiology, 104(2), 1177–1194. https://doi.org/10.1152/jn.00032.2010
  • François, C., Garcia-Alix, A., Bosch, L., & Rodriguez-Fornells, A. (2021). Signatures of brain plasticity supporting language recovery after perinatal arterial ischemic stroke. Brain and Language, 212, 104880. https://doi.org/10.1016/j.bandl.2020.104880
  • Furl, N., Garrido, L., Dolan, R. J., Driver, J., & Duchaine, B. (2011). Fusiform gyrus face selectivity relates to individual differences in facial recognition ability. Journal of Cognitive Neuroscience, 23(7), 1723–1740. https://doi.org/10.1162/jocn.2010.21545
  • Gerrits, R., Van der Haegen, L., Brysbaert, M., & Vingerhoets, G. (2019). Laterality for recognizing written words and faces in the fusiform gyrus covaries with language dominance. Cortex, 117, 196–204. https://doi.org/10.1016/j.cortex.2019.03.010
  • Glezer, L. S., Jiang, X., & Riesenhuber, M. (2009). Evidence for highly selective neuronal tuning to whole words in the “Visual Word Form Area”. Neuron, 62(2), 199–204. https://doi.org/10.1016/j.neuron.2009.03.017
  • Greenblatt, S. H. (1976). Subangular alexia without agraphia or hemianopsia. Brain and Language, 3(2), 229–245. https://doi.org/10.1016/0093-934X(76)90019-5
  • Grill-Spector, K., Knouf, N., & Kanwisher, N. (2004). The fusiform face area subserves face perception, not generic within-category identification. Nature Neuroscience, 7(5), 555–562. https://doi.org/10.1038/nn1224
  • Hamamé, C. M., Szwed, M., Sharman, M., Vidal, J. R., Perrone-Bertolotti, M., Kahane, P., Bertrand, O., & Lachaux, J.-P. (2013). Dejerine’s reading area revisited with intracranial EEG: Selective responses to letter strings. Neurology, 80(6), 602–603. https://doi.org/10.1212/WNL.0b013e31828154d9
  • Hannagan, T., Amedi, A., Cohen, L., Dehaene-Lambertz, G., & Dehaene, S. (2015). Origins of the specialization for letters and numbers in ventral occipitotemporal cortex. Trends in Cognitive Sciences, 19(7), 374–382. https://doi.org/10.1016/j.tics.2015.05.006
  • Hasson, U., Levy, I., Behrmann, M., Hendler, T., & Malach, R. (2002). Eccentricity bias as an organizing principle for human high-order object areas. Neuron, 34(3), 479–490. https://doi.org/10.1016/S0896-6273(02)00662-1
  • Haxby, J. V., Gobbini, M. I., Furey, M. L., Ishai, A., Schouten, J. L., & Pietrini, P. (2001). Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science, 293(5539), 2425–2430. https://doi.org/10.1126/science.1063736
  • Hirshorn, E. A., Li, Y., Ward, M. J., Richardson, R. M., Fiez, J. A., & Ghuman, A. S. (2016). Decoding and disrupting left midfusiform gyrus activity during word Reading. Proceedings of the National Academy of Sciences, 113(29), 8162–8167. https://doi.org/10.1073/pnas.1604126113
  • Hsiao, J. H., & Lam, S. M. (2013). The modulation of visual and task characteristics of a writing system on hemispheric lateralization in visual word recognition—a computational exploration. Cognitive Science, 37(5), 861–890. https://doi.org/10.1111/cogs.12033
  • Huang, L., Song, Y., Li, J., Zhen, Z., Yang, Z., & Liu, J. (2014). Individual differences in cortical face selectivity predict behavioral performance in face recognition. Frontiers in Human Neuroscience, 8. https://doi.org/10.3389/fnhum.2014.00483
  • Julian, J. B., Fedorenko, E., Webster, J., & Kanwisher, N. (2012). An algorithmic method for functionally defining regions of interest in the ventral visual pathway. NeuroImage, 60(4), 2357–2364. https://doi.org/10.1016/j.neuroimage.2012.02.055
  • Kamps, F. S., Hendrix, C. L., Brennan, P. A., & Dilks, D. D. (2020). Connectivity at the origins of domain specificity in the cortical face and place networks. Proceedings of the National Academy of Sciences, 117(11), 6163–6169. https://doi.org/10.1073/pnas.1911359117
  • Kanwisher, N. (2010). Functional specificity in the human brain: A window into the functional architecture of the mind. Proceedings of the National Academy of Sciences, 107(25), 11163–11170. https://doi.org/10.1073/pnas.1005062107
  • Kanwisher, N., McDermott, J., & Chun, M. M. (1997). The fusiform face area: A module in human extrastriate cortex specialized for face perception. The Journal of Neuroscience, 17(11), 4302–4311. https://doi.org/10.1523/JNEUROSCI.17-11-04302.1997
  • Kaufman, A. S., & Kaufman, N. L. (2004). Kaufman brief intelligence test (2nd ed.). American Guidance Service.
  • Kay, J., Coltheart, M., & Lesser, R. (1992). Psycholinguistic assessments of language processing in aphasia (PALPA) (1st edition). Psychology Press.
  • Kertesz, A. (2006). Western aphasia battery – revised (WAB-R) [Database record]. APA PsycTests. https://doi.org/10.1037/t15168-000.
  • Kim, J. S., Kanjlia, S., Merabet, L. B., & Bedny, M. (2017). Development of the visual word form area requires visual experience: Evidence from blind braille readers. Journal of Neuroscience, 37(47), 11495–11504. https://doi.org/10.1523/JNEUROSCI.0997-17.2017
  • Konkle, T., & Caramazza, A. (2013). Tripartite organization of the ventral stream by animacy and object size. Journal of Neuroscience, 33(25), 10235–10242. https://doi.org/10.1523/JNEUROSCI.0983-13.2013
  • Kosakowski, H. L., Cohen, M. A., Takahashi, A., Keil, B., Kanwisher, N., & Saxe, R. (2022). Selective responses to faces, scenes, and bodies in the ventral visual pathway of infants. Current Biology, 32(2), 265–274. https://doi.org/10.1016/j.cub.2021.10.064
  • Lane, C., Kanjlia, S., Omaki, A., & Bedny, M. (2015). “Visual” cortex of congenitally blind adults responds to syntactic movement. Journal of Neuroscience, 35(37), 12859–12868. https://doi.org/10.1523/JNEUROSCI.1256-15.2015
  • Lerma-Usabiaga, G., Carreiras, M., & Paz-Alonso, P. M. (2018). Converging evidence for functional and structural segregation within the left ventral occipitotemporal cortex in reading. Proceedings of the National Academy of Sciences, 115(42), E9981–E9990. https://doi.org/10.1073/pnas.1803003115
  • Li, J., Huang, L., Song, Y., & Liu, J. (2017). Dissociated neural basis of two behavioral hallmarks of holistic face processing: The whole-part effect and composite-face effect. Neuropsychologia, 102, 52–60. https://doi.org/10.1016/j.neuropsychologia.2017.05.026
  • Li, J., Osher, D. E., Hansen, H. A., & Saygin, Z. M. (2020). Innate connectivity patterns drive the development of the visual word form area. Scientific Reports, 10(1), 18039. https://doi.org/10.1038/s41598-020-75015-7
  • Liu, T. T., Freud, E., Patterson, C., & Behrmann, M. (2019). Perceptual function and category-selective neural organization in children with resections of visual cortex. Journal of Neuroscience, 39(32), 6299–6314. https://doi.org/10.1523/JNEUROSCI.3160-18.2019
  • Mahon, B. Z., & Caramazza, A. (2011). What drives the organization of object knowledge in the brain? Trends in Cognitive Sciences, 15(3), 97–103. https://doi.org/10.1016/j.tics.2011.01.004
  • Maisog, J. M., Einbinder, E. R., Flowers, D. L., Turkeltaub, P. E., & Eden, G. F. (2008). A meta-analysis of functional neuroimaging studies of dyslexia. Annals of the New York Academy of Sciences, 1145(1), 237–259. https://doi.org/10.1196/annals.1416.024
  • Malach, R., Levy, I., & Hasson, U. (2002). The topography of high-order human object areas. Trends in Cognitive Sciences, 6(4), 176–184. https://doi.org/10.1016/S1364-6613(02)01870-3
  • Martin, A. (2006). Shades of Déjerine—forging a causal link between the visual word form area and reading. Neuron, 50(2), 173–175. https://doi.org/10.1016/j.neuron.2006.04.004
  • McCandliss, B. D., Cohen, L., & Dehaene, S. (2003). The visual word form area: Expertise for reading in the fusiform gyrus. Trends in Cognitive Sciences, 7(7), 293–299. https://doi.org/10.1016/S1364-6613(03)00134-7
  • Mummery, C. J., Patterson, K., Price, C. J., Ashburner, J., Frackowiak, R. S., & Hodges, J. R. (2000). A voxel-based morphometry study of semantic dementia: Relationship between temporal lobe atrophy and semantic memory. Annals of Neurology, 47(1), 36–45. https://doi.org/10.1002/1531-8249(200001)47:1<36::AID-ANA8>3.0.CO;2-L
  • Nasr, S., Echavarria, C. E., & Tootell, R. B. H. (2014). Thinking outside the Box: Rectilinear shapes selectively activate scene-selective cortex. Journal of Neuroscience, 34(20), 6721–6735. https://doi.org/10.1523/JNEUROSCI.4802-13.2014
  • Newport, E. L., Landau, B., Seydell-Greenwald, A., Turkeltaub, P. E., Chambers, C. E., Dromerick, A. W., Carpenter, J., Berl, M. M., & Gaillard, W. D. (2017). Revisiting lenneberg’s hypotheses about early developmental plasticity: Language organization after left-hemisphere perinatal stroke. Biolinguistics, 11, 407–422. https://doi.org/10.5964/bioling.9105
  • Oosterhof, N. N., Connolly, A. C., & Haxby, J. V. (2016). CoSMoMVPA: Multi-modal multivariate pattern analysis of neuroimaging data in Matlab/GNU Octave. Frontiers in Neuroinformatics, 10), https://doi.org/10.3389/fninf.2016.00027
  • Osher, D. E., Saxe, R., Koldewyn, K., Gabrieli, J. D. E., Kanwisher, N., & Saygin, Z. M. (2015). Structural connectivity fingerprints predict cortical selectivity for multiple visual categories across cortex. Cerebral Cortex, 26(4), 1668–1683. https://doi.org/10.1093/cercor/bhu303
  • Patterson, K., Nestor, P. J., & Rogers, T. T. (2007). Where do you know what you know? The representation of semantic knowledge in the human brain. Nature Reviews Neuroscience, 8(12), 976–987. https://doi.org/10.1038/nrn2277
  • Paulesu, E., Danelli, L., & Berlingeri, M. (2014). Reading the dyslexic brain: Multiple dysfunctional routes revealed by a new meta-analysis of PET and fMRI activation studies. Frontiers in Human Neuroscience, 8, 830. https://doi.org/10.3389/fnhum.2014.00830
  • Pennington, B. F., & Bishop, D. V. M. (2008). Relations among speech, language, and Reading disorders. Annual Review of Psychology, 60(1), 283–306. https://doi.org/10.1146/annurev.psych.60.110707.163548
  • Postelnicu, G., Zollei, L., & Fischl, B. (2009). Combined volumetric and surface registration. IEEE Transactions on Medical Imaging, 28(4), 508–522. https://doi.org/10.1109/TMI.2008.2004426
  • Powell, H. W. R., Parker, G. J. M., Alexander, D. C., Symms, M. R., Boulby, P. A., Wheeler-Kingshott, C. A. M., Barker, G. J., Noppeney, U., Koepp, M. J., & Duncan, J. S. (2006). Hemispheric asymmetries in language-related pathways: A combined functional MRI and tractography study. NeuroImage, 32(1), 388–399. https://doi.org/10.1016/j.neuroimage.2006.03.011
  • Price, C. J., & Devlin, J. T. (2003). The myth of the visual word form area. Neuroimage, 19(3), 473–481. https://doi.org/10.1016/S1053-8119(03)00084-3
  • Price, C. J., & Devlin, J. T. (2011). The interactive account of ventral occipitotemporal contributions to Reading. Trends in Cognitive Sciences, 15(6), 246–253. https://doi.org/10.1016/j.tics.2011.04.001
  • Raschle, N. M., Zuk, J., & Gaab, N. (2012). Functional characteristics of developmental dyslexia in left-hemispheric posterior brain regions predate reading onset. Proceedings of the National Academy of Sciences, 109(6), 2156–2161. https://doi.org/10.1073/pnas.1107721109
  • Rauschecker, A. M., Deutsch, G. K., Ben-Shachar, M., Schwartzman, A., Perry, L. M., & Dougherty, R. F. (2009). Reading impairment in a patient with missing arcuate fasciculus. Neuropsychologia, 47(1), 180–194. https://doi.org/10.1016/j.neuropsychologia.2008.08.011
  • Regev, M., Honey, C. J., Simony, E., & Hasson, U. (2013). Selective and invariant neural responses to spoken and written narratives. Journal of Neuroscience, 33(40), 15978–15988. https://doi.org/10.1523/JNEUROSCI.1580-13.2013
  • Richlan, F., Kronbichler, M., & Wimmer, H. (2011). Meta-analyzing brain dysfunctions in dyslexic children and adults. Neuroimage, 56(3), 1735–1742. https://doi.org/10.1016/j.neuroimage.2011.02.040
  • Röder, B., Stock, O., Bien, S., Neville, H., & Rösler, F. (2002). Speech processing activates visual cortex in congenitally blind humans. European Journal of Neuroscience, 16(5), 930–936. https://doi.org/10.1046/j.1460-9568.2002.02147.x
  • Rosenke, M., van Hoof, R., van den Hurk, J., Grill-Spector, K., & Goebel, R. (2021). A probabilistic functional atlas of human occipito-temporal visual cortex. Cerebral Cortex, 31(1), 603–619. https://doi.org/10.1093/cercor/bhaa246
  • Rossion, B., Dricot, L., Devolder, A., Bodart, J.-M., Crommelinck, M., Gelder, B. d., & Zoontjes, R. (2000). Hemispheric asymmetries for whole-based and part-based face processing in the human fusiform gyrus. Journal of Cognitive Neuroscience, 12(5), 793–802. https://doi.org/10.1162/089892900562606
  • Saygin, Z. M., Osher, D. E., Koldewyn, K., Reynolds, G., Gabrieli, J. D. E., & Saxe, R. (2012). Anatomical connectivity patterns predict face selectivity in the fusiform gyrus. Nature Neuroscience, 15(2), 321–327. https://doi.org/10.1038/nn.3001
  • Saygin, Z. M., Osher, D. E., Norton, E. S., Youssoufian, D. A., Beach, S. D., Feather, J., Gaab, N., Gabrieli, J. D. E., & Kanwisher, N. (2016). Connectivity precedes function in the development of the visual word form area. Nature Neuroscience, 19(9), 1250–1255. https://doi.org/10.1038/nn.4354
  • Schwartz, M. F., Kimberg, D. Y., Walker, G. M., Faseyitan, O., Brecher, A., Dell, G. S., & Coslett, H. B. (2009). Anterior temporal involvement in semantic word retrieval: Voxel-based lesion-symptom mapping evidence from aphasia. Brain, 132(12), 3411–3427. https://doi.org/10.1093/brain/awp284
  • Seghier, M. L., Neufeld, N. H., Zeidman, P., Leff, A. P., Mechelli, A., Nagendran, A., Riddoch, J. M., Humphreys, G. W., & Price, C. J. (2012). Reading without the left ventral occipito-temporal cortex. Neuropsychologia, 50(14), 3621–3635. https://doi.org/10.1016/j.neuropsychologia.2012.09.030
  • Seghier, M. L., & Price, C. J. (2011). Explaining left lateralization for words in the ventral occipitotemporal cortex. Journal of Neuroscience, 31(41), 14745–14753. https://doi.org/10.1523/JNEUROSCI.2238-11.2011
  • Shaywitz, B. A., Shaywitz, S. E., Pugh, K. R., Mencl, W. E., Fulbright, R. K., Skudlarski, P., Constable, R. T., Marchione, K. E., Fletcher, J. M., Lyon, G. R., & Gore, J. C. (2002). Disruption of posterior brain systems for reading in children with developmental dyslexia. Biological Psychiatry, 52(2), 101–110. https://doi.org/10.1016/S0006-3223(02)01365-3
  • Simmons, W. K., Bellgowan, P. S., & Martin, A. (2007). Measuring selectivity in fMRI data. Nature Neuroscience, 10(1), 4–5. https://doi.org/10.1038/nn0107-4
  • Song, Y., Zhu, Q., Li, J., Wang, X., & Liu, J. (2015). Typical and atypical development of functional connectivity in the face network. Journal of Neuroscience, 35(43), 14624–14635. https://doi.org/10.1523/JNEUROSCI.0969-15.2015
  • Staudt, M. (2010). Reorganization after pre- and perinatal brain lesions*. Journal of Anatomy, 217(4), 469–474. https://doi.org/10.1111/j.1469-7580.2010.01262.x
  • Staudt, M., Grodd, W., Niemann, G., Wildgruber, D., Erb, M., & Krägeloh–Mann, I. (2001). Early left periventricular brain lesions induce right hemispheric organization of speech. Neurology, 57(1), 122–125. https://doi.org/10.1212/WNL.57.1.122
  • Stevens, W. D., Kravitz, D. J., Peng, C. S., Tessler, M. H., & Martin, A. (2017). Privileged functional connectivity between the visual word form area and the language system. The Journal of Neuroscience, 37(21), 5288–5297. https://doi.org/10.1523/JNEUROSCI.0138-17.2017
  • Striem-Amit, E., & Amedi, A. (2014). Visual cortex extrastriate body-selective area activation in congenitally blind people “seeing” by using sounds. Current Biology, 24(6), 687–692. https://doi.org/10.1016/j.cub.2014.02.010
  • Szwed, M., Dehaene, S., Kleinschmidt, A., Eger, E., Valabrègue, R., Amadon, A., & Cohen, L. (2011). Specialization for written words over objects in the visual cortex. Neuroimage, 56(1), 330–344. https://doi.org/10.1016/j.neuroimage.2011.01.073
  • Tadros, K., Dupuis-Roy, N., Fiset, D., Arguin, M., & Gosselin, F. (2013). Reading laterally: The cerebral hemispheric use of spatial frequencies in visual word recognition. Journal of Vision, 13(1), 4–4. https://doi.org/10.1167/13.1.4
  • Tuckute, G., Paunov, A., Kean, H., Small, H., Mineroff, Z., Blank, I., & Fedorenko, E. (2022). Frontal language areas do not emerge in the absence of temporal language areas: A case study of an individual born without a left temporal lobe. Neuropsychologia, 169, 108184. https://doi.org/10.1016/j.neuropsychologia.2022.108184
  • Vagharchakian, L., Dehaene-Lambertz, G., Pallier, C., & Dehaene, S. (2012). A temporal bottleneck in the language comprehension network. Journal of Neuroscience, 32(26), 9089–9102. https://doi.org/10.1523/JNEUROSCI.5685-11.2012
  • Vaidya, A. R., Pujara, M. S., Petrides, M., Murray, E. A., & Fellows, L. K. (2019). Lesion studies in contemporary neuroscience. Trends in Cognitive Sciences, 23(8), 653–671. https://doi.org/10.1016/j.tics.2019.05.009
  • Van der Haegen, L., Cai, Q., & Brysbaert, M. (2012). Colateralization of Broca’s area and the visual word form area in left-handers: FMRI evidence. Brain and Language, 122(3), 171–178. https://doi.org/10.1016/j.bandl.2011.11.004
  • Wandell, B. A., Rauschecker, A. M., & Yeatman, J. D. (2012). Learning to see words. Annual Review of Psychology, 63(1), 31–53. https://doi.org/10.1146/annurev-psych-120710-100434
  • Williams, M. A., Dang, S., & Kanwisher, N. G. (2007). Only some spatial patterns of fMRI response are read out in task performance. Nature Neuroscience, 10(6), 685–686. https://doi.org/10.1038/nn1900
  • Wilmer, J. B., Germine, L., Chabris, C. F., Chatterjee, G., Williams, M., Loken, E., Nakayama, K., & Duchaine, B. (2010). Human face recognition ability is specific and highly heritable. Proceedings of the National Academy of Sciences, 107(11), 5238–5241. https://doi.org/10.1073/pnas.0913053107
  • Yeatman, J. D., & Feldman, H. M. (2013). Neural plasticity after pre-linguistic injury to the arcuate and superior longitudinal fasciculi. Cortex, 49(1), 301–311. https://doi.org/10.1016/j.cortex.2011.08.006
  • Yoncheva, Y. N., Zevin, J. D., Maurer, U., & McCandliss, B. D. (2010). Auditory selective attention to speech modulates activity in the visual word form area. Cerebral Cortex, 20(3), 622–632. https://doi.org/10.1093/cercor/bhp129
  • Zhu, Q., Song, Y., Hu, S., Li, X., Tian, M., Zhen, Z., Dong, Q., Kanwisher, N., & Liu, J. (2010). Heritability of the specific cognitive ability of face perception. Current Biology, 20(2), 137–142. https://doi.org/10.1016/j.cub.2009.11.067