Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

# Longitudinal changes in cortical responses to letter-speech sound stimuli in 8–11 year-old children

## Abstract

While children are able to name letters fairly quickly, the automatisation of letter-speech sound mappings continues over the first years of reading development. In the current longitudinal fMRI study, we explored developmental changes in cortical responses to letters and speech sounds across 3 yearly measurements in a sample of 18 8–11 year old children. We employed a text-based recalibration paradigm in which combined exposure to text and ambiguous speech sounds shifts participants’ later perception of the ambiguous sounds towards the text. Our results showed that activity of the left superior temporal and lateral inferior precentral gyri followed a non-linear developmental pattern across the measurement sessions. This pattern is reminiscent of previously reported inverted-u-shape developmental trajectories in children’s visual cortical responses to text. Our findings suggest that the processing of letters and speech sounds involves non-linear changes in the brain’s spoken language network possibly related to progressive automatisation of reading skills.

## Introduction

Reading is an evolutionary novel, culturally acquired skill that requires us to associate speech with visual symbols. The initial mappings between letters and speech sounds are typically acquired within the first months of reading instruction1,2, particularly in orthographically transparent languages with relatively consistent letter-speech sound associations such as Dutch, German, Italian or Hungarian3. However, the fine-tuning and automatisation of letter-speech sound integration continues while children strengthen their word decoding skills over the first years of reading development1,4,5,6.

The link between speech and text processing has typically been illustrated by modulations in cortical activation in response to matching (congruent) vs. nonmatching (incongruent) audio-visual letter-speech sound stimuli. Functional magnetic resonance imaging (fMRI) studies employing such congruency manipulations have shown increased activation of speech sensitive areas within the superior temporal cortex (STC) in response to congruent letter-speech sound pairs compared to incongruent pairs and speech alone26,27,28. Furthermore, the strength of STC responses to congruent audio-visual stimuli has been found to scale with individual differences in phonological and reading skills29,30 and with reading problems in dyslexic readers31,32,33,34. However, congruency manipulation paradigms rely on learnt letter-speech sound mappings and involve additional cognitive and verbal working memory processes that may vary with participants’ reading35 and selective attention36 skills.

Here we employed the text-based recalibration paradigm to investigate longitudinal developmental changes in the posterior STC and the broader audio-visual reading network in 8–11 year old children who were scanned yearly over a 3 year period. This study is part of a larger longitudinal fMRI project investigating letter-speech sound mapping in children with and without developmental dyslexia. We have previously reported group differences in brain activation during the audio-visual exposure blocks of the text-based recalibration task in the first fMRI session. Namely, despite comparable behavioural recalibration performance, children with and without dyslexia showed differences in ventral occipito-temporal cortex activation46. Here we investigated longitudinal changes in this audio-visual activation across the three sessions in a group of children without dyslexia. During the audio-visual exposure blocks, the children were simultaneously presented with the ambiguous /a?a/ sound and disambiguating text. Children’s responses during the subsequent auditory-only post-test trials were used to assess the magnitude of the recalibration effect behaviourally. Our text-based recalibration paradigm was specifically designed to investigate reading-induced changes in the left posterior STC42. We thus expected to find longitudinal changes in auditory cortical activation. To also investigate possible changes across other regions of the brain’s audio-visual reading network (see e.g.46), we did not restrict our analyses to the auditory cortex but performed whole brain analyses. In these analyses, we were particularly interested in possible links between children’s cortical responses to audio-visual letter-speech sound stimuli, their recalibration task performance, and individual differences in phonological perception and reading skills.

## Results

### Pre-test results

Prior to the MRI, the children performed a pre-test during a mock scanner session (see Methods section) in order to determine the individual, most ambiguous sound (/a?a/) to be used during the text-based recalibration task. During the pre-test, the children heard nine sounds tokens along an /aba/–/ada/ continuum and for each sound indicated whether they perceived it as /aba/ or /ada/. The overall /aba/ response proportions to each of the nine sound tokens indicate that all participants perceived the sound tokens closer to the /aba/ end of the continuum as /aba/ and the sounds closer to the /ada/ end of the continuum as /ada/ (see /aba/ response proportions for tokens 1–3 and 7–9, Fig. 1). This was confirmed in a 3 (session) × 9 (sound) RM ANCOVA analyses of the /aba/ response proportions to each of the nine sounds, which revealed a main effect of sound [F(2,24) = 4.087, p < 0.05, Greenhouse-Geisser corrected] and an interaction between sound and children’s age in months during session 1 (baseline age) [F(2,24) = 5.629, p < 0.01, Greenhouse-Geisser corrected]. No other main- or interaction effects reached statistical significance (all F ≤ 2.97). This indicates that the average /aba/ response proportions to each of the nine sounds along the /aba/-/ada/ continuum did not significantly change across the three measurement sessions but that younger children on average showed a different /aba/ response proportion bias. Namely, their responses were biased more towards /aba/.

Children’s responses to the auditory-only post-test sounds showed a change in /aba/ response proportions across the ambiguous sound and its two flanking sounds along the /aba/-/ada/ continuum (Fig. 2). There were more /aba/ responses for the sound /a?a/-1 closer to the clear /aba/ token, as well as for the ambiguous /a?a/ sound, compared to the /a?a/+1 sound that is closer to the clear /ada/ token. However, a 3 (session) × 2 (“aba” / “ada” exposure block) × 3 (post-test sounds) RM ANCOVA did not show significant main- or interaction effects (all F ≤ 2.19), indicating a lack of a significant behavioural recalibration effect across participants.

### fMRI activity during the exposure blocks

The text and speech sound stimuli presented during exposure blocks of the recalibration task elicited significant activation in bilateral superior temporal gyri (STG), visual and parietal cortices, as well as in the left frontal cortex (Fig. 3; whole-brain FDR threshold q < 0.05). Talairach coordinates of activation clusters for each individual measurement session are reported in Table 1. At the given (FDR) threshold, the resulting maps appear to indicate a general broadening of this network from sessions 1 to 3. However, RM ANOVA analyses investigating changes in cortical activation across sessions, revealed significant clusters only in the left STG and lateral inferior precentral gyrus (pre-CG; Fig. 4; whole-brain FDR threshold q < 0.05). Plotting the average individual beta values per participant per session for visualisation purposes, indicates non-linear changes in cortical activation across sessions in both clusters (Fig. 4). Using a stringent correction for multiple comparisons (FDR correction using the Benjamini and Hochberg procedure), pairwise session comparisons only yielded a significant difference between sessions 1 and 2 in the lateral inferior pre-CG cluster (t(17) = −3.038, p < 0.01; q = 0.04), while at uncorrected level, activation in this cluster also differed between sessions 1 and 3 (t(17) = −2.571, p < 0.05; q = 0.06) and sessions 1 and 2 in the left STG cluster (t(17) = −2.211, p < 0.05; q = 0.08).

### Mixed regression results

To explore how children’s age, reading fluency and phoneme perception (pre-test slope steepness) may have influenced the change in left STG and lateral inferior pre-CG activity across sessions, we further assessed the individual beta values for each session with post-hoc mixed regression analyses. The analyses showed significant interactions between session and children’s baseline age [F(2,13) = 8.702, p < 0.005] and between session and the recalibration effect [F(2,17) = 5.808, p < 0.05] in the left STG cluster. Estimated marginal means comparisons showed that younger children showed higher left STG activation during session 2 compared to session 3 (mean difference = 0.596, t(11) = 2.950, p < 0.05), whereas older children showed more STG activation in session 3 compared to session 2 (mean difference = 0.502, t(10) = 2.918, p < 0.05). These analyses also revealed that, regardless of age, children who show a (stronger) recalibration effect across measurement sessions, show more activation in the left STG cluster during session 2 compared to session 3 (mean difference = 0.603, t(12) = 2.791, p < 0.05). None of the interaction effects reached significance for the left lateral inferior pre-CG (all F ≤ 4.030).

## Discussion

In this longitudinal fMRI study, we investigated developmental changes in cortical responses to audio-visual letter-speech sound stimuli using a text-based recalibration paradigm in children aged 8–9 to 10–11 years across three measurement time points spaced ~1 year apart. Our fMRI results showed a significant developmental change in cortical activation in left hemisphere STG and lateral inferior pre-CG regions. The changes in cortical activation across sessions within the left STG cluster were associated with children’s age during session 1 as well as the strength of the behavioural recalibration effect. No links with children’s reading fluency or phonological processing (pre-test slope steepness) were observed.

All children successfully performed the text-based recalibration task, as reflected by the different /aba/ response proportions to the ambiguous post-test sounds following each type of exposure block. However, as in our previous study in 8–10 year-old children, these behavioural recalibration effects did not reach significance in the MRI scanner environment46, while children in this age range did show significant text-based recalibration in an offline behavioural study44. This discrepancy can likely be attributed to the acoustic noise, and/or altered (visual) attentional focus in MRI environment47,48. By comparison, typically reading adults have been reported to show reliable recalibration effects both in and out of the MRI scanner42,49,50. Similarly to our previous fMRI study46, a sub-group of children in the current sample did show recalibration in the MRI scanner (9 in session 1, 8 in session 2 and 11 in session 3). It is therefore likely that the recalibration effect may emerge in the MRI environment as children get older, although this would need to be confirmed in future studies.

To investigate cortical activation associated with audio-visual processing of letters and speech sounds and its potential links to behavioural and reading measures, we focused our fMRI analyses on the audio-visual exposure blocks of the recalibration task. Our results showed bilateral cortical activation patterns across the sessions in brain areas previously linked to audio-visual integration and reading7,28,29,31,32,51, as well as to text-based recalibration specifically42,46. These brain areas included bilateral STG, occipital and parietal areas as well as the left frontal cortex. Comparisons of brain activation between sessions revealed longitudinal changes in activation in left hemisphere STG and lateral inferior pre-CG clusters. Plotting children’s average individual brain activation across sessions within these clusters revealed substantial variability with mostly non-linear activation changes. Post-hoc mixed regression analysis showed that activation in the left STG cluster was associated with children’s baseline age. Namely, children who were younger during session 1 showed higher left STG activation during session 2 compared to session 3, whereas older children showed more activation in session 3 compared to session 2. Visual inspection of changes in activation across sessions for the younger vs. the older children, suggested that younger children, aged 8 years during the first measurement session, were more likely to show a non-linear, inverted-u-type activity change across sessions. Older children, aged 9 years during the first session, instead showed a more individually variable, but on average smaller linear change in activation. These results will need to be verified in a larger sample of children.

Non-linear (e.g. inverted-u) developmental changes are reminiscent of cross-sectional results reported in adults and children of different age groups in previous electroencephalography (EEG) studies employing an audio-visual oddball paradigm with letters and speech sounds. These studies have indicated audio-visual deviancy responses in two time windows—a mismatch negativity response (MMN) around 100–250 ms, associated with automatic processing of audio-visual letter-speech sound stimuli5,52,53 and a sustained late negativity (LN) response associated with more “effortful” and deliberate integration mechanisms4,5,19. Using this paradigm, 8–9 year-old typically reading children were observed to have a relatively wide audio-visual integration window and more pronounced audio-visual MMN and LN responses5 compared to both younger and older children4 as well as adults52. That is, the findings reveal an apparent evolution from an LN-only response in 7/8-year olds to both MMN and LN responses and a broad audio-visual integration window in the 8/9-year olds, followed by increasing selectivity across 11-year olds and adults. This pattern of results arguably reflects a switch from more effortful to automatised audio-visual processing, with fully automatised letter-speech sound processing in adults. The observed longitudinal changes in left STG activation to letter-speech sound pairs across sessions may represent a different neural correlate of the same developmental trajectory from effortful letter-for-letter mapping of text and speech towards quick and automatic processing of over-learnt letter-speech sound associations.

Studies investigating text specialisation and sensitivity using both EEG and fMRI have shown similar non-linear (inverted-u) activation patterns of developmental change within ventral occipito-temporal cortex and the VWFA in particular14,16,18,54,55, paralleled by a change in grey matter volume of the left occipito-temporal sulcus12. The fact that we did not observe significant session differences in ventral occipito-temporal cortex activation in the current sample may be due to task characteristics or age group differences. The recalibration task employs only two types of simple visual text stimuli that do not require extensive processing or explicit judgment (“aba” and “ada”). It is therefore likely that this paradigm does not provide optimal engagement of the VWFA, making it less suitable for exploration of VWFA activation changes over time. The exposure blocks of text-based recalibration combine these clear visual text stimuli with an ambiguous speech sound. The combination of the two modalities likely produces different auditory percepts across trials. This could result in an enhancement of activation in speech sensitive STG regions, that have been shown to encode these type of perceptual changes42,49,56,57, especially at the age where letter-sound associations are in the process of becoming automatised. In line with this interpretation, our mixed regression analysis indicated that, regardless of age, children who did show (stronger) recalibration effects in the MRI scanner across sessions, showed more left STG activation during session 2 compared to session 3. It is likely that this finding is linked to an emergence and/or strengthening of the effect across sessions, as 9 children showed recalibration in session 1, 8 children in session 2 and 11 children in session 3. Our previous fMRI study using the text-based recalibration paradigm in children with and without dyslexia showed a correlation between letter-speech sound identification scores and left STG activation across children, with lower activation in the region linked to better task performance46. The observed decrease in left STG activation between session 2 and 3 in children who show a recalibration effect may similarly represent successful mapping of the text and ambiguous speech stimuli leading to (stronger) behavioural recalibration.

Beyond changes specific to audio-visual integration of letters and speech sounds, reading and spoken language development may mutually lead to the refinement of cortical responses to speech during reading development. It has been proposed that learning to read in an alphabetic script in particular, alters phonological processing by magnifying the phonological representations of the smallest-possible visual unit (often a single letter) and is accompanied by a developmental increase in STG responses observed during auditory rhyme judgement tasks in alphabetic compared to logographic orthographies58. These reading-induced changes in speech perception may further affect categorical perception of phonemes59 and interact with reading skills60. Developmentally, categorical perception of phonemes has been associated with left-lateralized activation of posterior STG that scales with increasingly distinct definition of phoneme categories in 7–12 year-old children60. Furthermore, the STG shows distinct phoneme-category specific activation patterns following categorical learning and subsequent perception of novel speech sounds61, as well as task-dependent56,57 and recalibration-induced42,49 changes in speech sound representation.

The emergence of the left lateral inferior pre-CG as one of the regions showing significant session effects is in line with this interpretation. Previous research has shown cortical activation in frontal/motor speech production areas during speech perception tasks (e.g.62,63,64,65) and audio-visual speech disambiguation66, particularly when using noisy/degraded auditory stimuli67,68,69,70. The lateral inferior pre-CG cluster in the current study overlaps with the somatosensory lip and tongue representation71. An fMRI study investigating the presence of articulatory information during auditory-only speech perception has shown that activity patterns within this brain area, albeit in the right hemisphere, encode information on the place of articulation of speech sounds72. The developmental change in cortical activation in this region may thus result from the fact that children differentially engaged the lateral pre-CG to help disambiguate the ambiguous speech stimuli, which in fact were taken from an /aba/-/ada/ continuum involving a change in place of articulation. On a more general level, this activity could relate to changes in subvocal rehearsal of the perceived /aba/ vs. /ada/ sounds.

As children grow and learn, the exact timing of developmental turning points related to e.g. learning phases in acquiring solid letter-speech sound mappings, refinement of speech perception and visual text processing, likely varies per child. It has been shown that children reach developmental milestones at different times73,74. Changes in cortical activation associated with these transitions may vary per brain region or involve different brain networks depending on the individual and behavioural goal. This may explain the observed individual differences in cortical activation changes across sessions within the left STG and lateral inferior pre-CG (grey lines, Fig. 4). To further understand these individual differences, it will be important to extend and replicate the current developmental findings in future longitudinal studies including larger sample sizes and, for example, multiple reading/language tasks. These studies may reveal task- and region-dependent developmental trajectories that, ultimately, will explain (a)typical variability in children’s reading development.

In terms of underlying developmental mechanisms of change, non-linear, inverted-u-type trajectories are reminiscent of those associated with children’s skill learning. When learning a skill, children benefit from both active strategies (e.g. explicitly mapping letters and speech sounds) as well as passive mechanisms such as statistical and associative learning75,76,77. The use of active strategies during the early stages of learning serves as input for the passive mechanisms and the dynamics of the two change over the course of skill acquisition. Namely, with increased proficiency, children begin to rely less on the explicit mechanisms and instead switch to fast, automatised processing supported by passive mechanisms such as associative learning78. In fact, skill learning in general is thought to reflect an expansion-renormalisation pattern, with an initial increase in cortical activation/volume in the key areas involved in the acquisition of a specific skill, followed by a decrease once proficiency is reached79,80. Because reading is a culturally acquired skill, similar mechanisms may underlie the non-linear developmental patterns in our left STG and lateral inferior pre-CG results. In particular, they may reflect changes during the proposed prolonged time-course of letter-speech sound automatisation1,4, which in case of the lateral inferior pre-CG may specifically be related to the differential engagement of the region to disambiguate the audio-visual text-ambiguous speech stimuli used in our task.

In conclusion, the observed longitudinal changes in 8-11 year-old children’s left STG and lateral inferior pre-CG responses to speech sounds and letters, together with previous findings in higher order visual areas, point to regionally specific functional changes in the developing reading and speech perception networks. Our findings show non-linear developmental trajectories in a speech (and text) sensitive auditory cortical region, and may reflect the fine-tuning of letter-speech sound mappings representing a gradual switch from more effortful to automatised processing. The observed changes in lateral inferior pre-CG activation may in turn point to a specific role of motor articulatory functions in audio-visual speech disambiguation during text-based recalibration. Future longitudinal studies with larger sample sizes are needed to verify the observed pattern of responses within these areas and the reading network as a whole and to establish their link to individual differences in typical as well as atypical reading development. These studies could also use tailored fMRI designs to investigate developmental changes in text-induced shifts in auditory cortical representations of the ambiguous post-tests sounds using a multivariate decoding approach (see42). Relating individual decoding accuracies of /aba/ vs. /ada/ percepts to children’s reading skills could provide further insights into the links between reading gains and reading-induced changes in the auditory cortical representation of speech sounds.

## Methods

### Participants

A total of 43 children (mean age at time-point 1: 8.9 ± 0.7 years; 24 females) were recruited from local elementary schools in the Maastricht area in the Netherlands. We included children of various reading levels ranging from poor to excellent. None of the children had persistent reading problems that would meet the criteria for a dyslexia diagnosis. Of the 43 children, 7 completed only the first measurement, 16 completed only the first 2 measurements, and the remaining 20 participants participated in all 3 measurements of the given study. Note that the relatively large number of children who dropped out after the 2nd measurement was due to an almost 4 month closure of the MRI research facilities during the first Covid-19 lock-down (March–June 2020). Of the 20 participants, one was excluded from the analyses due to poor data quality in session 3, and one participant had undergone remediation for dyslexia. The final sample consisted of 18 children who participated in all three longitudinal measurements (mean age at time-point 1: 8.7 ± 0.6 years; 11 females; 1 left-handed). All fMRI and post-hoc analyses were conducted on this sample of 18 children. The average time between the first two measurements was 13 months (±1.7 months; range 11–17 months; mean age at time-point 2: 9.8 ± 0.6 years) and the average time between measurements 2 and 3 was 10.4 months (±1.9 months; range 6–15 months; mean age time-point 3: 10.6 ± 0.59 years). The shorter time period between measurements 2 and 3 was driven by 5 participants who were scanned on average 8 months after the second measurement (range 6–9 months) due to planned orthodontic treatment that would have resulted in discontinuation of research participation. All children were native Dutch speakers, had no reported hearing impairments, normal or corrected-to normal vision and no history of developmental or neurological disorders. All children received a present (in session 1) or a gift card (sessions 2 and 3) for participation, along with a small, customised gift per scanning session. Parents provided written informed consent prior to each measurement in accordance with the declaration of Helsinki. The study was approved by the ethics committee of the Faculty of Psychology and Neuroscience, Maastricht University.

### Literacy and cognitive skills

After each scanning session, the children completed computerised tasks of the 3DM test battery (Dyslexia Differential Diagnosis81; assessing reading and phonological skills. In addition, at first and last measurements, the children completed two sub-tests of the Dutch version of the Wechsler Intelligence Scale for Children-III (WISC-III-NL82; –verbal (similarities) and non-verbal (block design). The subsequent sub-tests of the 3DM were used in the present study: reading, letter-speech sound identification and phoneme deletion (elision).

### Stimuli

The auditory stimuli employed in the recalibration task consisted of 650 ms recordings of the speech sounds /aba/ and /ada/ spoken by a male native Dutch speaker (see37 for a detailed description). The speech sounds were used to create a nine-token sound continuum ranging from a clear /aba/ sound to a clear /ada/ sound, with 7 ambiguous sounds in between, by changing the second formant (F2) in eight steps of 39 Mel using PRAAT software83. The visual stimuli consisted of corresponding “aba” and “ada” text presented in white at the centre of a black screen in ‘Times New Roman’ font (font size 50). The auditory and visual stimuli were presented using Presentation software (Version 18.1, Neurobehavioral Systems, Inc., Berkeley, CA, United States).

### Pre-test

During the mock scanner session, children completed a pre-test, in which they heard each of the 9 sound tokens along the /aba/-/ada/ continuum a total of 98 times in a randomised order, with the middle sounds along the continuum presented more frequently than the two clear tokens37,49,84. The children were instructed to listen carefully to each sound and to indicate if they perceived it as /aba/ or as /ada/ by pressing the left or right innermost button of the MR compatible button box with their left/right index finger following a response cue. The response cues consisted of text “aba” (left) and “ada” (right), held up by cartoon monsters created using the Monster Workshop content pack of the iClone 6 software (https://www.reallusion.com/). During the presentation of the speech sounds, the children viewed a black screen with a white fixation cross followed by the response cue 1 s later. Each trial was terminated after the child provided a response, triggering the presentation of the subsequent speech sound after 2 s (Fig. 5). The total duration of the pre-test was ~5 min.

The results of the pre-test were used to individually determine the most ambiguous speech sound along the /aba/–/ada/ continuum for each participant, based on the proportion of /aba/ responses to each sound token. The most ambiguous sound was identified as the token with an /aba/ vs. /ada/ response proportion closest to 0.5, representing the phoneme boundary44,84. This individually determined most ambiguous sound was used in the audio-visual exposure blocks and post-test trials of the recalibration task. In addition to the most ambiguous /a?a/ sound, the post-test trials also included its two flanking sounds along the /aba/-/ada/ continuum, namely /a?a/+1 and /a?a/-1. In addition, the resulting sigmoidal curve showing the response proportions to each sound token, served as an indicator of children’s categorical perception of the phonemes /aba/ and /ada/, with a steeper slope representing a sharper distinction between phoneme categories.

The recalibration task consisted of audio-visual exposure blocks followed by auditory post-test trials (Fig. 6). During each exposure block, the children were presented with either the text “aba” or “ada” in combination with the individually determined most ambiguous speech sound /a?a/ 8 times. The audio-visual stimuli were presented simultaneously (relative SOA of 0 ms), the duration of the auditory stimuli was 650 ms and visual text was presented for 1 s. The inter-trial interval between subsequent audio-visual exposure trials was set to 2 s (1 TR). The /aba/ and /ada/ exposure blocks were presented in a pseudo-randomised order, making sure that the same type of exposure block was not repeated more than twice in a row. During the audio-visual exposure blocks, the children were instructed to pay close attention to the speech sounds and text without providing a response.

Each exposure block was followed by four auditory-only post-test trials, the onset of which was jittered to 10 s (4–6 TR). These jittered periods served as the baseline condition, during which the children fixated on a white cross in the middle of a black screen. The post-test trials were presented in a randomised order with the most ambiguous sound /a?a/ presented twice, and each of the flanking sounds /a?a/+1 and /a?a/-1 on the /aba/-/ada/ continuum presented once. Following each post-test sound, a response cue containing cartoon monsters appeared on the screen. The duration of the response cue was fixed to 3 s and the onset was jittered 2,5–3 s with respect to the post-test sound. The subsequent post-test trial was presented 3–3,5 s following the response cue. The total inter trial interval (ITI) between post-test trials was 6 s (3 TR). Children were instructed to listen carefully to each post-test sound and to respond whether they perceived it as /aba/ or as /ada/ upon the presentation of the response cue using the MR-compatible button boxes. The responses were made by pressing the innermost button of the button box with the left/right index finger, as practiced in the mock scanner.

Children completed a total of four runs of the recalibration task, corresponding to 24 audio-visual exposure blocks (12 with “aba” text and 12 with “ada” text) and 4*24 post-test trials. The recalibration effect was calculated as the difference in average /aba/ response proportions across the post-test sounds following the two types of exposure blocks (i.e. total /aba/ response proportion after an “aba” exposure block minus total /aba/ response proportion following an “ada” exposure block). All auditory and audio-visual stimuli were presented during a 900 ms silent gap between volume acquisitions.

### Statistical analyses behavioural data

Children’s performance on the pre-test and recalibration tasks was analysed using repeated measures (RM) ANCOVA. The RM ANCOVA analyses for the pre-test included within-subject factors Session (1, 2, 3) and sound (1–9) and baseline age as a covariate. The RM ANCOVA for the Recalibration task was performed on the /aba/ response proportions following each of the three post-test sounds and consisted of within-subject factors: session (1, 2, 3), exposure type (“aba” vs. “ada”) and post-test sounds (/a?a/-1, /a?a/, /a?a/+1) with baseline age as a covariate. The degrees of freedom were adjusted using the Greenhouse-Geisser correction for conditions, which violated the sphericity assumption. Prior to running the analyses, the data were assessed for outliers using boxplots. This amounted to four observations for the pre-test (two outliers—lower quartile plus 1.5 times inter-quartile range and two extreme outliers—lower quartile plus 3 times inter-quartile range) and one observation for the recalibration analyses (upper quartile plus 1.5 times inter-quartile range). For both analyses, the observations identified as outliers and extreme outliers were excluded.

We used a custom logistic function described in44 (Eq. (1)) and the Slope Fitting tool in MATLAB 2019a (The MathWorks, Inc., Natick, MA, United States) to assess the fit of each subject-specific pre-test sigmoidal curve resulting from children’s /aba/ response proportions to each of the 9 sounds along the /aba/-/ada/ continuum (see Fig. 1). This was done to obtain the slope value (c in eq. (1)) of the sigmoidal response curve as it represents the distinction between the /aba/ and /ada/ phoneme categories. Namely, the closer this value was to 0, the steeper the slope and sharper the phoneme category distinction. The pre-test slope value providing the best fit to the data was obtained using the MATLAB least squares solver. To optimise the outcome, the variables in eq. (1) were restricted to the following values: 0 ≤ a ≥ 10, −10 ≤ b ≥ 10, −10 ≤ c ≥ 10, −9 ≤ d ≥ 18 and the best fit was determined after 30 iterations of the procedure.

$${{{\mathrm{y}}}} = \frac{{{{\mathrm{a}}}}}{{1 + e^{\frac{{ - \left( {{{{\mathrm{x}}}} - {{{\mathrm{d}}}}} \right)}}{{{{\mathrm{c}}}}}}}} + {{{\mathrm{b}}}}{{{\mathrm{.}}}}$$

Equation 1: a = amplitude; b = lowest asymptote of y-axis; c = slope of the function; d = location of the category boundary.

### (f)MRI measurements

Brain Imaging was performed with a Siemens Prisma 3 T MRI scanner (Siemens Medical Systems, Erlangen, Germany) using a 64-channel head–neck coil. Five functional runs were acquired (2.5 mm × 2.5 mm × 2.5 mm resolution) with a multi-band factor of five echoplanar-imaging sequence (repetition time [TR] = 2000 ms, acquisition time [TA] = 1100 ms, field of view = 210 mm × 210 mm, echo time [TE] = 35.8 ms). Each volume consisted of 50 slices (no gap), covering the whole brain. The recalibration task consisted of four 5 min runs, followed by a 7 min localiser task not included in the current analyses. The speech stimuli were presented binaurally at a comfortable listening level via MR compatible headphones (Sensimetrics, model S14, www.sens.com), in the 900 ms silent gap between two consecutive volume acquisitions. In addition, a high-resolution structural scan (1 mm × 1 mm × 1 mm) using a T1-weighted three-dimensional MPRAGE sequence ([TR] = 2300 ms, [TE] = 2.98 ms, 192 sagittal slices) was acquired.

### (f)MRI pre-processing

Data pre-processing and analyses were performed using BrainVoyager QX version 2.8 and BrainVoyager versions 20.6 and 21.4 (Brain Innovation, Maastricht, The Netherlands) as well as custom MATLAB routines (The MathWorks, Inc., Natick, MA, United States). The functional data were aligned to the first volume of the first functional run and underwent 3D motion correction (trilinear sinc interpolation), slice scan time correction and high pass temporal filtering (five cycles per time course). The anatomical data underwent manual inhomogeneity correction to improve white matter-grey matter boundary segmentation and was transformed into Talairach space85. To ensure all anatomical runs were well aligned across sessions, the native space anatomical data of sessions 2 and 3 was coregistered to the ACPC transformed session 1 anatomy using vmr-vmr co-registration in Brainvoyager QX. The resulting transformation file was subsequently applied to the anatomical runs of sessions 2 and 3 resulting in well aligned anatomical data. Transformation to Talairach space was performed using the transformation file of session 1.

The functional data were subsequently aligned across sessions by co-registration to the Talairach transformed anatomical data of session 1, re-sampled to 3 mm iso-voxel resolution and spatially smoothed using a 6 mm FWHM Gaussian kernel. Volumes of functional runs affected by excessive head motion (≥3 mm/degree translation/rotation in any direction) were removed from the run. If the number of affected volumes exceeded 20%, the run was excluded from further analyses. This amounted to a total of five runs in the second session and six runs in the third session. In addition, not all children completed all four recalibration runs due to time constraints. The total number of runs not acquired across all three sessions was 8 (2 in session 1 and 3 in sessions 2 and 3). The final number of runs included in the analyses was 197: 70 session 1, 64 session 2, 63 session 3.

Individual cortical surface representations of session 1 anatomical scans were automatically constructed for each participant based on the white matter-grey matter boundary. The boundary was then manually adjusted, and aligned using cortex based alignment employing a moving-target group average based on curvature information, resulting in an anatomically-aligned group-average 3D cortical representation86. Each participant’s functional data were projected onto their cortical surface to create surface-based time courses. All functional data were subsequently analysed per hemisphere at the surface level using the group-aligned average cortical surfaces.

### Whole brain univariate fMRI analysis

Cortical activation across sessions was assessed using a random effects (RFX) general linear model (GLM) approach based on the individual surface-based time courses. The GLM included a predictor for each type of exposure block (“aba” and “ada”), predictors for each post-test sound (4 predictors) and z-transformed motion predictors as variables of no interest to improve the signal-to-noise ratio in the data. Cortical activation in response to the audio-visual letter-ambiguous speech stimuli for each measurement session was assessed using contrast maps (t-statistics) based on the results of the GLM model. The contrast maps compared cortical activation during the exposure blocks (“aba” + “ada” exposure blocks; EXP) and the fixation cross baseline and were corrected for multiple comparisons using a whole-brain FDR threshold of q < 0.05. To explore the longitudinal changes in cortical activation across sessions, in a next step we performed a RM ANOVA with session as the within-subject factor. This yielded a cortical map (F-statistic) of areas that show significant changes in activation across the three measurement sessions at a whole-brain FDR threshold of q < 0.05. To limit our results to brain regions that showed significant activation during the audio-visual exposure blocks, the resulting F-map was masked with the combined activation map (EXP > baseline; FDR q < 0.05) across the three sessions. Thus, only the regions that showed overlap with the mask were explored further with post-hoc paired-samples t-tests for pairwise activation differences (individual average beta values EXP > baseline) between sessions 1 & 2, 2 & 3 and 1 & 3. An FDR correction for multiple comparisons (using the Benjamini and Hochberg procedure) was performed on the p values of the post-hoc paired-samples t-tests using MATLAB.

### Multiple regression analysis

Post-hoc mixed regression analyses were performed to explore possible relations between behavioural measures and the observed changes (main effect of Session) in cortical activation in the left STG and lateral inferior precentral gyrus (SPSS v26, IBM Corp., Armonk, NY, United States). We opted for this approach due to its reported suitability for accelerated longitudinal designs and tolerance for handling missing data87,88. We constructed two marginal models including each child’s average beta values for each cluster as the dependent variable, within-subject factor session, and within-subject covariates: baseline age, reading fluency, recalibration effect and pre-test slope value, as well as interactions between session and each of the covariates. To avoid circularity, we do not interpret the main effect of session, as this is already established by the RM ANOVA results, and only interpret the interactions between session and the covariates. The letter-speech sound identification and phoneme deletion fluency scores were not included in the analysis as they did not show much variation across sessions (see Table 2). Prior to running the model, the data were assessed for outliers using boxplots. Six observations in total were identified as outliers/extreme outliers and excluded from the analyses. Four observations were outliers—two upper quartile plus 1.5 times inter-quartile range (one value for reading fluency, one for the recalibration effect) and two lower quartile plus 1.5 times inter-quartile range (both pre-test slope values). Two additional pre-test slope values were categorised as extreme outliers, i.e. lower quartile plus 3 times inter-quartile range. In addition, four observations had missing data—two pre-test slope values and two reading fluency measures. The final models reported were Fixed-effects models estimated using Restricted Maximum Likelihood and an unstructured covariance matrix for RM. Model residuals had a normal distribution and showed no heteroscedasticity.

### Reporting summary

Further information on research design is available in the Nature Research Reporting Summary linked to this article.

## Data availability

Anonymised raw fMRI and behavioural data of the children for whom parents have given (anonymised) data sharing consent are available from the corresponding author upon reasonable request.

## Code availability

All code used for task presentation and data analysis is available from the corresponding author.

## References

1. Blomert, L. The neural signature of orthographic-phonological binding in successful and failing reading development. NeuroImage 57, 695–703 (2011).

2. Perfetti, C. A. The Universal Grammar of Reading. Sci. Stud. Read. 7, 3–24 (2003).

3. Borgwaldt, S. R., Hellwig, F. M. & De Groot, A. M. B. Onset entropy matters—Letter-to-phoneme mappings in seven languages. Read. Writ. 18, 211–229 (2005).

4. Froyen, D., Bonte, M. L., van Atteveldt, N. & Blomert, L. The long road to automation: neurocognitive development of letter-speech sound processing. J. Cogn. Neurosci. 21, 567–580 (2009).

5. Žarić, G. et al. Reduced neural integration of letters and speech sounds in dyslexic children scales with individual differences in reading fluency. PLoS ONE 9, e110337 (2014).

6. Ehri, L. C. Development of Sight Word Reading: Phases and Findings. in The Science of Reading: A Handbook (eds. Snowling, M. J. & Hulme, C.) 135–154 (Blackwell Publishing Ltd, 2005). https://doi.org/10.1002/9780470757642.

7. Dehaene, S., Cohen, L., Morais, J. & Kolinsky, R. Illiterate to literate: behavioural and cerebral changes induced by reading acquisition. Nat. Rev. Neurosci. 16, 234–244 (2015).

8. Schlaggar, B. L. & McCandliss, B. D. Development of neural systems for reading. Annu. Rev. Neurosci. 30, 475–503 (2007).

9. Yeatman, J. D., Dougherty, R. F., Ben-Shachar, M. & Wandell, B. A. Development of white matter and reading skills. Proc. Natl Acad. Sci. USA 109, E3045–E3053 (2012).

10. Vandermosten, M., Boets, B., Wouters, J. & Ghesquière, P. A qualitative and quantitative review of diffusion tensor imaging studies in reading and dyslexia. Neurosci. Biobehav. Rev. 36, 1532–1552 (2012).

11. Gullick, M. M. & Booth, J. R. Individual Differences in Crossmodal Brain Activity Predict Arcuate Fasciculus Connectivity in Developing. J. Cogn. Neurosci. 26, 1331–1346 (2014).

12. Ben-Shachar, M., Dougherty, R. F., Deutsch, G. K. & Wandell, B. A. The development of cortical sensitivity to visual word forms. J. Cogn. Neurosci. 23, 2387–2399 (2011).

13. Brem, S. et al. Tuning of the visual word processing system: distinct developmental ERP and fMRI effects. Hum. Brain Mapp. 30, 1833–1844 (2009).

14. Maurer, U. et al. Coarse neural tuning for print peaks when children learn to read. NeuroImage 33, 749–758 (2006).

15. Dehaene-Lambertz, G., Monzalvo, K. & Dehaene, S. The emergence of the visual word form: longitudinal evolution of category-specific ventral visual areas during reading acquisition. PLOS Biol. 16, e2004103 (2018).

16. Price, C. J. & Devlin, J. T. The Interactive Account of ventral occipitotemporal contributions to reading. Trends Cogn. Sci. 15, 246–253 (2011).

17. Fraga-Gonzalez, G. et al. The rise and fall of rapid occipito-temporal sensitivity to letters: transient specialization through elementary school. Dev. Cogn. Neurosci. 49, 100958 (2021).

18. Fraga González, G., Zaric, G., Tijms, J., Bonte, M. & van der Molen, M. W. Contributions of letter-speech sound learning and visual print tuning to reading improvement: evidence from brain potential and dyslexia training studies. Brain Sci. 7, 7010010 (2017).

19. Žarić, G. et al. Crossmodal deficit in dyslexic children: practice affects the neural timing of letter-speech sound integration. Front. Hum. Neurosci. 9, 369 (2015).

20. Bonte, M., Ley, A., Scharke, W. & Formisano, E. Developmental refinement of cortical systems for speech and voice processing. NeuroImage 128, 373–384 (2016).

21. Sowell, E. R. et al. Mapping cortical change across the human life span. Nat. Neurosci. 6, 309–315 (2003).

22. Giedd, J. N. et al. Brain development during childhood and adolescence: a longitudinal MRI study. Nat. Neurosci. 2, 861–863 (1999).

23. Gogtay, N. et al. Dynamic mapping of human cortical development during childhood through early adulthood. Proc. Natl Acad. Sci. USA 101, 8174–8179 (2004).

24. Hedman, A. M., van Haren, N. E. M., Schnack, H. G., Kahn, R. S. & Hulshoff Pol, H. E. Human brain changes across the life span: a review of 56 longitudinal magnetic resonance imaging studies. Hum. Brain Mapp. 33, 1987–2002 (2012).

25. Mills, K. L. et al. Structural brain development between childhood and adulthood: convergence across four longitudinal samples. NeuroImage 141, 273–281 (2016).

26. Blau, V., Van Atteveldt, N., Formisano, E., Goebel, R. & Blomert, L. Task-irrelevant visual letters interact with the processing of speech sounds in heteromodal and unimodal cortex. Eur. J. Neurosci. 28, 500–509 (2008).

27. Raij, T., Uutela, K. & Hari, R. Audiovisual integration of letters in the human brain. Neuron 28, 617–625 (2000).

28. van Atteveldt, N., Formisano, E., Goebel, R. & Blomert, L. Integration of letters and speech sounds in the human brain. Neuron 43, 271–282 (2004).

29. McNorgan, C., Randazzo-Wagner, M. & Booth, J. R. Cross-modal integration in the brain is related to phonological awareness only in typical readers, not in those with reading difficulty. Front. Hum. Neurosci. 7, 1–12 (2013).

30. McNorgan, C., Awati, N., Desroches, A. S. & Booth, J. R. Multimodal lexical processing in auditory cortex is literacy skill dependent. Cereb. Cortex 24, 2464–2475 (2014).

31. Blau, V., van Atteveldt, N., Ekkebus, M., Goebel, R. & Blomert, L. Reduced Neural Integration of Letters and Speech Sounds Links Phonological and Reading Deficits in Adult Dyslexia. Curr. Biol. 19, 503–508 (2009).

32. Blau, V. et al. Deviant processing of letters and speech sounds as proximate cause of reading failure: A functional magnetic resonance imaging study of dyslexic children. Brain 133, 868–879 (2010).

33. Kronschnabel, J., Brem, S., Maurer, U. & Brandeis, D. The level of audiovisual print-speech integration deficits in dyslexia. Neuropsychologia 62, 245–261 (2014).

34. Ye, Z., Rüsseler, J., Gerth, I. & Münte, T. F. Audiovisual speech integration in the superior temporal region is dysfunctional in dyslexia. Neuroscience 356, 1–10 (2017).

35. Plewko, J. et al. Letter and speech sound association in emerging readers with familial risk of dyslexia. Front. Hum. Neurosci. 12, 393 (2018).

36. Talsma, D. & Woldorff, M. G. Selective attention and multisensory integration: multiple phases of effects on the evoked brain activity. J. Cogn. Neurosci. 17, 1098–1114 (2005).

37. Bertelson, P., Vroomen, J. & De Gelder, B. Visual Recalibration of Auditory Speech Identification: a McGurk Aftereffect. Psychological Sci. 14, 592–597 (2003).

38. Vroomen, J. & Baart. Phonetic Recalibration in Audiovisual Speech. in The Neural Bases of Multisensory Processes (ed. Micah M. Murray, M. T. W.) 363–380 (CRC Press, 2012). https://doi.org/10.1201/9781439812174-24.

39. Norris, D., McQueen, J. M. & Cutler, A. Perceptual learning in speech. Cogn. Psychol. 47, 204–238 (2003).

40. Ullas, S., Formisano, E., Eisner, F. & Cutler, A. Interleaved lexical and audiovisual information can retune phoneme boundaries. Atten. Percept. Psychophys. 82, 2018–2026 (2020).

41. Keetels, M., Bonte, M. & Vroomen, J. A Selective Deficit in Phonetic Recalibration by Text in Developmental Dyslexia. Front. Psychol. 9, 1–11 (2018).

42. Bonte, M., Correia, J., Keetels, M., Vroomen, J. & Formisano, E. Reading-induced shifts of perceptual speech representations in auditory cortex. Sci. Rep. 7, 5143 (2017).

43. Keetels, M., Schakel, L., Bonte, M. & Vroomen, J. Phonetic recalibration of speech by text. Atten. Percept. Psychophys. 78, 938–945 (2016).

44. Romanovska, L., Janssen, R. & Bonte, M. Reading-induced shifts in speech perception in dyslexic and typically reading children. Front. Psychol. 10, 221 (2019).

45. Scott, M. Speech imagery recalibrates speech-perception boundaries. Atten. Percept. Psychophys. 1496–1511 https://doi.org/10.3758/s13414-016-1087-6 (2016).

46. Romanovska, L., Janssen, R. & Bonte, M. Cortical responses to letters and ambiguous speech vary with reading skills in dyslexic and typically reading children. NeuroImage: Clin. 30, 102588 (2021).

47. van Maanen, L., Forstmann, B. U., Keuken, M. C., Wagenmakers, E. J. & Heathcote, A. The impact of MRI scanner environment on perceptual decision-making. Behav. Res. Methods 48, 184–200 (2016).

48. Kobald, S. O., Getzmann, S., Beste, C. & Wascher, E. The impact of simulated MRI scanner background noise on visual attention processes as measured by the EEG. Sci. Rep. 6, 1–10 (2016).

49. Kilian-Hütten, N., Valente, G., Vroomen, J. & Formisano, E. Auditory cortex encodes the perceptual interpretation of ambiguous sound. J. Neurosci.: Off. J. Soc. Neurosci. 31, 1715–1720 (2011).

50. Ullas, S., Hausfeld, L., Cutler, A., Eisner, F. & Formisano, E. Neural Correlates of Phonetic Adaptation as Induced by Lexical and Audiovisual Context. J. Cogn. Neurosci. 1–14 https://doi.org/10.1162/jocn_a_01608 (2020).

51. Chyl, K. et al. Prereader to beginning reader: changes induced by reading acquisition in print and speech brain networks. J. Child Psychol. Psychiatry 59, 76–87 (2017).

52. Froyen, D., Van Atteveldt, N., Bonte, M. & Blomert, L. Cross-modal enhancement of the MMN to speech-sounds indicates early and automatic integration of letters and speech-sounds. Neurosci. Lett. 430, 23–28 (2008).

53. Näätänen, R. The perception of speech sounds by the human brain as reflected by the mismatch negativity brain response. Psychophysiology 38, 1–21 (2001).

54. Maurer, U., Zevin, J. D. & McCandliss, B. D. Left-lateralized N170 Effects of Visual Expertise in Reading: Evidence from Japanese Syllabic and Logographic Scripts. J. Cogn. Neurosci. 20, 1878–1891 (2008).

55. Maurer, U. et al. The development of print tuning in children with dyslexia: evidence from longitudinal ERP data supported by fMRI. NeuroImage 57, 714–722 (2011).

56. Bonte, M., Hausfeld, L., Scharke, W., Valente, G. & Formisano, E. Task-Dependent Decoding of Speaker and Vowel Identity from Auditory Cortical Response Patterns. J. Neurosci. 34, 4548–4557 (2014).

57. Yi, H. G., Leonard, M. K. & Chang, E. F. The Encoding of Speech Sounds in the Superior Temporal Gyrus. Neuron 102, 1096–1110 (2019).

58. Brennan, C., Cao, F., Pedroarena-Leal, N., McNorgan, C. & Booth, J. R. Reading acquisition reorganizes the phonological awareness network only in alphabetic writing systems. Hum. Brain Mapp. 34, 3354–3368 (2013).

59. Harm, M. W. & Seidenberg, M. S. Phonology, reading acquisition, and dyslexia: Insights from connectionist models. Psychological Rev. 106, 491–528 (1999).

60. Conant, L. L., Liebenthal, E., Desai, A. & Binder, J. R. FMRI of phonemic perception and its relationship to reading development in elementary-to middle-school-age children. NeuroImage 89, 192–202 (2014).

61. Ley, A. et al. Learning of New Sound Categories Shapes Neural Response Patterns in Human Auditory Cortex. J. Neurosci. 32, 13273–13280 (2012).

62. D’Ausilio, A. et al. The Motor Somatotopy of Speech Perception. Curr. Biol. 19, 381–385 (2009).

63. Pulvermüller, F. et al. Motor cortex maps articulatory features of speech sounds. Proc. Natl Acad. Sci. USA 103, 7865–7870 (2006).

64. Watkins, K. E., Strafella, A. P. & Paus, T. Seeing and hearing speech excites the motor system involved in speech production. Neuropsychologia 41, 989–994 (2003).

65. Wilson, S. M., Saygin, A. P., Sereno, M. I. & Iacoboni, M. Listening to speech activates motor areas involved in speech production. Nat. Neurosci. 7, 701–702 (2004).

66. Skipper, J. I., Van Wassenhove, V., Nusbaum, H. C. & Small, S. L. Hearing lips and seeing voices: How cortical areas supporting speech production mediate audiovisual speech perception. Cereb. Cortex 17, 2387–2399 (2007).

67. Jääskeläinen, I. P. The Role of Speech Production System in Audiovisual Speech Perception. Open Neuroimaging J. 4, 30–36 (2010).

68. Ma, W. J., Zhou, X., Ross, L. A., Foxe, J. J. & Parra, L. C. Lip-reading aids word recognition most in moderate noise: a Bayesian explanation using high-dimensional feature space. PLoS ONE 4, e4638 (2009).

69. Ross, L. A., Saint-Amour, D., Leavitt, V. M., Javitt, D. C. & Foxe, J. J. Do you see what I am saying? Exploring visual enhancement of speech comprehension in noisy environments. Cereb. Cortex 17, 1147–1153 (2007).

70. Sumby, W. H. & Pollack, I. Visual Contribution to Speech Intelligibility in Noise. J. Acoustical Soc. Am. 26, 212–215 (1954).

71. Penfield, W. & Boldrey, E. Somatic motor and sensory representation in the cerebral cortex of man as studied by electrical stimulation. Brain 60, 389–443 (1937).

72. Correia, J. M., Jansma, B. M. B. & Bonte, M. Decoding Articulatory Features from fMRI Responses in Dorsal Speech Regions. J. Neurosci. 35, 15015–15025 (2015).

73. Taanila, A., Murray, G. K., Jokelainen, J., Isohanni, M. & Rantakallio, P. Infant developmental milestones: a 31-year follow-up. Dev. Med. Child Neurol. 47, 581–586 (2005).

74. Flensborg-Madsen, T. & Mortensen, E. L. Developmental milestones during the first three years as precursors of adult intelligence. Dev. Psychol. 54, 1434–1444 (2018).

75. Siegler, R. S. Children’s learning. Am. Psychol. 60, 769–778 (2005).

76. Shrager, J. & Siegler, R. S. SCADS: a Model of Children’s Strategy Choices and Strategy Discoveries. Psychol. Sci. 9, 405–410 (1998).

77. Siegler, R. & Araya, R. A computational model of conscious and conscious strategy discovery. Adv. Child Dev. Behav. 33, 1–42 (2005).

78. Crowley, K., Shrager, J. & Siegler, R. S. Strategy Discovery as a Competitive Negotiation between Metacognitive and Associative Mechanisms. Dev. Rev. 17, 462–489 (1997).

79. Lövdén, M., Garzón, B. & Lindenberger, U. Human skill learning: expansion, exploration, selection, and refinement. Curr. Opin. Behav. Sci. 36, 163–168 (2020).

80. Wenger, E., Brozzoli, C., Lindenberger, U. & Lövdén, M. Expansion and Renormalization of Human Brain Structure During Skill Acquisition. Trends Cogn. Sci. 21, 930–939 (2017).

81. Blomert, L. & Vaessen, A. 3DM Differential Diagnostics for Dyslexia: cognitive analysis of reading and spelling. (Boom Test Publishers, 2009).

82. Kort, W. et al. WISC-III-NL. (Harcourt Test Publishers, 2005).

83. Boersma, P. & Weenink, D. J. M. PRAAT: Doing phonetics by computer. Glot Int. 5, 341–347 (2001).

84. Vroomen, J., Van Linden, S., Keetels, M., De Gelder, B. & Bertelson, P. Selective adaptation and recalibration of auditory speech by lipread information: Dissipation. Speech Commun. 44, 55–61 (2004).

85. Talairach & Tournoux. Co-Planar Stereotaxic Atlas of the Human Brain. (Thieme, 1988).

86. Frost, M. A. & Goebel, R. Measuring structural-functional correspondence: spatial variability of specialised brain regions after macro-anatomical alignment. NeuroImage 59, 1369–1381 (2012).

87. Tamnes, C. K., Bos, M. G. N., van de Kamp, F. C., Peters, S. & Crone, E. A. Longitudinal development of hippocampal subregions from childhood to adulthood. Dev. Cogn. Neurosci. 30, 212–222 (2018).

88. Vijayakumar, N., Mills, K. L., Alexander-Bloch, A., Tamnes, C. K. & Whittle, S. Structural brain development: a review of methodological approaches and best practices. Dev. Cogn. Neurosci. 33, 129–148 (2018).

## Acknowledgements

We thank Federico de Martino and Valentin Kemper for their help in setting up the MRI sequences. We are thankful to all the children and their families who took time to participate in our research and to the local primary schools and regional dyslexia institute for their help in recruiting participants. We would also like to thank Gerard van Breukelen for his advice on conducting the mixed regression analyses and Giancarlo Valente, Armin Heinecke, Agustin Lage Castellanos and Elia Formisano for their input on statistical and fMRI analyses. This research was funded by The Netherlands Organization for Scientific Research (Vidi-Grant 452-16-004 to M.B.).

## Author information

Authors

### Contributions

M.B. and L.R. designed the experiment. L.R. and R.J. collected and analysed the data. L.R. and M.B. wrote the paper.

### Corresponding author

Correspondence to Linda Romanovska.

## Ethics declarations

### Competing interests

The authors declare no competing interests.

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## Rights and permissions

Reprints and Permissions

Romanovska, L., Janssen, R. & Bonte, M. Longitudinal changes in cortical responses to letter-speech sound stimuli in 8–11 year-old children. npj Sci. Learn. 7, 2 (2022). https://doi.org/10.1038/s41539-021-00118-3

• Accepted:

• Published:

• DOI: https://doi.org/10.1038/s41539-021-00118-3