Abstract
Memory is fleeting. To avoid information loss, humans need to recode verbal stimuli into chunks of limited duration, each containing multiple words. Chunk duration may also be limited neurally by the wavelength of periodic brain activity, so-called neural oscillations. While both cognitive and neural constraints predict some degree of behavioral regularity in processing, this remains to be shown. Our analysis of self-paced reading data from 181 participants reveals periodic patterns at a frequency of \(\sim\) 2 Hz. We defined multi-word chunks by using a computational formalization based on dependency annotations and part-of-speech tags. Potential chunk outputs were first generated from the computational formalization and the final chunk outputs were selected based on normalized pointwise mutual information. We show that behavioral periodicity is time-aligned to multi-word chunks, suggesting that the multi-word chunks generated from local dependency clusters may minimize memory demands. This is the first evidence that sentence processing behavior is periodic, consistent with a role of both memory constraints and endogenous electrophysiological rhythms in the formation of chunks during language comprehension.
Similar content being viewed by others
Introduction
Language comprehension has its limits: In order to understand speech, we must link words together—yet we cannot do so across sequences of arbitrary duration because memory contents progressively deteriorate with time1,2. Earlier work has shown that temporal integration of events into larger units is restricted to a window of 3 s3. For instance, performance on sequence reproduction tasks is high for sequences of up to 3 s4. Accordingly, electroencephalography research has observed that the contingent negative variation, a potential associated with the anticipation of events (such as their expectancy or duration), accompanies the reproduction of stimuli of 1–3 s, but reduces for intervals beyond 4 s5. Auditory short-term memory is limited to a similar interval of 2–3 s6. For language, a proposed window of 6 words7 translates to 2.4 s when assuming a rate of 150 words per minute8. Likewise, the duration of single utterances in speech approaches a median of 2.6 s9 (see also10).
More recently, it has been suggested that the pace of electrophysiological activity could explain such timing constraints. In particular, cycles of slow-frequency activity play a role in the cognitive formation of multi-word units. A seminal study presented native speakers of Mandarin and an English-speaking control group with isochronous 4-syllable sentences while recording their magnetoencephalogram11. Two syllables would always form a two-word phrase. Two phrases would always form a four-word sentence. The authors report spectral peaks at the rates of phrases (2 Hz) and sentences (1 Hz) in native speakers only, suggesting that the peaks reflect the formation of cognitive units during comprehension (see also12). In natural non-isochronous stimuli, phase angles of oscillatory activity in the delta band (< 4 Hz) predict the offsets of multi-word chunks13, in particular when chunks last for 2.7 s14 independent of acoustic boundary markings. Beyond speech, delta-band tracking was also reported for visual processing of sign language15 and lip movements16. This suggests that neural activity might impose an endogenous rhythm onto processing across domains.
Regardless of whether memory limitations or the wavelength of periodic activity are behind timing constraint on multi-word chunking, both would predict that behavioral data recorded during sentence comprehension should contain regular temporal patterns—although the text input does not contain any temporal markings. In particular, periodic behavioral events should align with chunk boundaries. This alignment may be reflected in the wrap-up effects, which are reading slowdowns that occur at the endings of clauses17,18 and implicit prosodic phrases19, implying that they reflect the cognitive formation of multi-word units20. Indeed, it was recently shown that eye movements during naturalistic sentence reading exhibit rhythmicity around 1 Hz that show coherence with the electroencephalogram21. However, naturalistic reading allows for word skipping and backward regressions22. It is, therefore, challenging to relate the observed periodicity to cognitive units and its behavioral relevance for language processing remains obscure.
In the current study, we report for the first time that self-paced reading (SPR) data indeed also contain periodic patterns at a frequency that is consistent with periodic neuronal processes previously associated with chunking. Specifically, we show that these patterns align with chunk boundaries defined by our computational formalization. There are fruitful approaches for defining multi-word chunks through computational formalization. Different methods adopted for word tagging in an information extraction system might generate inconsistent output chunks23,24. Here we define multi-word chunks independently by using a computational formalism based on dependency annotations25,26,27, combining with the classic approach of a word-tagging system with the tagset of bio, where b means the beginning of a chunk, i means inside a chunk, and o means outside of a chunk28. This approach yields chunks that for the most part align with major syntactic boundaries—as exemplified in Fig. 1 (see Chunking Algorithm for details). These linguistically grounded chunk boundaries are established by finding the optimal set of sub-trees in a dependency tree. Often sentences can be chunked in more than one way, so we use an information theory process based on dependency relations and part-of-speech tags to prioritize more likely chunk candidates. Fundamentally, this means that the more often part-of-speech tags are connected to one another via specific dependency relations in a corpus, the more likely they will form a chunk in a given sentence.
Our findings provide the first behavioral evidence that reading behavior is regular at a slow time scale, consistent with both memory constraints on multi-word chunking and an involvement of rhythmic electrophysiological processes in the generation of multi-word chunks. Particularly, this periodic behavior seems to be relevant for the cognitive formation of multi-word units during higher-level language processing and may minimize memory demands.
Results
SPR times are periodic < 4 Hz
We applied frequency-domain time-series analysis to wrap-up effects29 in N = 181 openly accessible SPR data sets30. To highlight wrap-up effects, we differenced the raw SPR time series, effectively amplifying transitions from slowdown to speedup across subsequent words (see Fig. 8B). For statistical analysis, we took a permutation approach (see Data Analysis). This revealed a peak around 2 Hz (see Fig. 2). Specifically, at 1.75, 2, and 2.25 Hz, the t-value of the one-sample t-test on the observed power estimates exceeded the 950th entry of the sorted distribution of t-values from tests on 1000 PSD spectra resulting from permutations of the differenced data, corresponding to an uncorrected one-tailed \(p < 0.05\). After Bonferroni-correction for the 100 query frequencies, this remained significant (\(p < 0.001\), corrected) at 2 Hz. These results suggest that natural, unconstrained reading slows down and then speeds up at a period of 0.5 s.
Periodicity relates to chunking
To obtain chunks, we employ a computational model that defines them as sequences of words and bound morphemes that allow for all local dependencies to be established25 (for an example, see Fig. 7). The recognition of a chunk boundary has been shown psychologically real through different experimental paradigms (e.g. a click paradigm that has participants listen to sentences with clicks and indicate where the clicks are31,32,33) and analysis techniques (e.g. a hierarchical clustering scheme that data can be grouped by the measures of relatedness and then map them onto the hierarchical structure34,35). The approach we adopted here is analogous to formal linguistic definitions26 and resonates with classical phrase-structural approaches to chunking27. Note that the locality of these chunks implicitly minimizes memory demands, which are widely viewed as a key constraint on dependency processing36.
To link periodic slowdown-speedup transitions to chunking, we first detected positive peaks in the differenced time series. These differences mark major transitions from slow to fast reading times. We then performed mixed-effects logistic regression analyses to assess whether the occurrence of a turning point depended on the presence of a chunk boundary. To stay consistent with prior literature on wrap-up effects, our boundary factor included not only a level for chunk boundaries, but also a level sentence for sentence boundaries; for comparison, a level non-boundary marked words that did not occur at either type of boundary. We first fitted a baseline model including an intercept, fixed effects of word frequency and word form surprisal, and random effects of subject and story. The baseline model was then compared to a model adding the boundary factor. Inclusion significantly improved model fit above baseline (\(\chi ^{2}(2) = 588.82, p~<~0.001\)). Analogous comparisons for subsets revealed significant model improvement within all condition pairs (sentence and chunk: \(\chi ^{2}(1) = 371.30, p~<~001\); sentence and non-boundary: \(\chi ^{2}(1) = 597.01, p < 0.001\); chunk and non-boundary: \(\chi ^{2}(1) = 19.18, p < 0.001\); Fig. 3). This means that the transitions from slow to fast reading times occurred more often at sentence boundaries relative to both chunk boundaries and non-boundary words, and more often for chunk boundaries relative to non-boundary words.
SPR slows down within chunks
To substantiate the relevance of slowdown-speedups for chunking, we further assessed the progression of reading times within chunks, with the hypothesis that reading times increase gradually as readers approach the end of a chunk. To this end, within each chunk, we fitted a linear model that predicted reading time from word position. We then extracted the slope for each chunk and entered all slopes as a dependent measure into a new model with an intercept only, plus random factors for subject and story. There was a significant positive effect of the model intercept (\(t(11.98) = 4.34, p < 0.001\), Satterthwaite-approximated Degrees of Freedom; Fig. 4). This suggests that reading times increase across word positions within chunk.
Discussion
Our analyses provide the first behavioral evidence that higher-level language comprehension—specifically, the formation of multi-word memory chunks—is a periodic behavior. Previous work in cognitive neuroscience has linked slow periodic neural activity to eye movements during reading21 and the formation of multi-word chunks during language comprehension11,13,14. The current results suggest that this is indeed behaviorally relevant for language processing. Readers slow down and then speed up roughly every 0.5 s—mostly at sentence and chunk boundaries. They also show a gradual increase in SPR times from chunk onset to offset, which may indicate the incremental integration of words into a multi-word unit that progressively increases in size, consistent with neurophysiological evidence for a gradual increase of electrophysiological activity towards the end of each multi-word unit within a sentence37. Periodicity may thus reflect chunking. Classical wrap-up effects could reflect the periodic formation of multi-word chunks29.
The tendency of SPR transitions to occur at the boundaries of sentences and chunks links rather well to prior psycholinguistic proposals. In particular, it has been argued that memory constraints limit the distance of dependencies between the words and bound morphemes of sentences36,38. Here, we provide a complimentary hypothesis inspired by electrophysiology: If the wavelength of neural oscillations limits the duration of chunks, it would implicitly enforce short dependency distances to allow for dependency formation within the current memory chunk1,27. Nevertheless, the current data cannot dissociate this syntactic approach from perceptual notions of chunking. Likely, many of the chunk boundaries as defined here align with implicit prosodic boundaries. In spoken language, there is a strong alignment between syntactic and prosodic boundaries39,40,41. In the absence of prosodic markings, both listeners and readers generate implicit prosodic structure to guide perceptual sampling42. Moreover, implicit prosody is also reflected in periodic brain activity at delta-band frequency43. We embrace the classical view that perceptual sampling in time windows that cover multiple words and the formation of fine-grained dependency structure amongst these words go hand in hand7. Future research needs to investigate how such a staged architecture maps onto periodic neural and cognitive processes.
The current findings could provide an initial hint at a possible relationship between periodic slowdowns in reading and the periodicity of the electrophysiology of chunking. M/EEG studies have argued that delta-band oscillations (< 4 Hz) reflect the grouping of words into larger units11,13,14. Consistent with this, we observe a spectral peak at 2 Hz in reading times. Strikingly, the SPR data analyzed here do not contain any physical rhythm or boundaries. As chunk boundaries are not marked visually, they must be set by some cognitive heuristic31,44. Yet, given that the current study did not assess concurrent M/EEG in addition to behavioral responses, it remains to be shown that behavioral periodicity indeed stems from endogenous neuronal rhythms that synchronize with higher-level linguistic information45,46,47.
The current chunking formalism operates within the framework of dependency grammar48, which does not explicitly assume a hierarchical syntactic structure. Different types of cognitive units above the single-word level have been linked to periodic brain activity (for discussion, see49,50,51—some hierarchical, some not11,13,14). In principle, from the current results, we may only claim that the size of chunks may relate to the size of a neural processing window in the delta-band. The microstructure of syntax and syntactic processing within chunks is beyond the scope of the current work.
Conclusion
Readers speed up and slow down periodically at a period of 0.5 s. These transitions may indicate the formation of multi-word chunks that allow for establishing all dependencies amongst the words and bound morphemes held in working memory at a time. Multi-word chunking is a periodic behavior, possibly mirroring underlying rhythmic neuronal processes.
Methods
Data
We analyzed a set of openly-accessible self-paced reading (SPR) data from 181 native speakers of English30. Participants had been instructed to read 10 stories from the Natural Stories Corpus word by word, advancing through button press. The reading time was measured for each word from the presentation onset to the button-press. Each story includes roughly 1000 words, which results in 10,245 words and 485 sentences in total. The text was automatically parsed using the Stanford Parser52. The output from the parser was manually corrected and automatically converted to the annotations for the Universal Dependencies (UD) by Futrell et al.30, so the data has high-quality human-verified UD annotations.
Chunking algorithm
The processes described below are applied to the human-verified UD annotations of the Natural Stories Corpus. The chunks require no generalization as they, and the statistics used to derive them, are drawn directly from the UD annotations. The algorithm fundamentally gives a solution to finding base-level subtrees when more than one solution exists.
We define chunks as sequences of words and bound morphemes that form saturated local dependency clusters25,26,27. The chunking algorithm employs dependency annotations and part-of-speech tags48. Specifically, chunks are considered base-level subtrees, allowing for a language-agnostic definition and annotation. This means the core algorithm is based on subtrees with a depth of 1. However, this restriction is softened to allow for chunks with a depth of 2 to minimize unitary chunks using a simple heuristic as described below.
As a first step, potential candidate chunks are extracted. For a given sentence and its corresponding tree, the span between each node n at position x and its corresponding head h at position k (where k can be greater than or less than x) is considered a candidate chunk if the nodes between position x and \(k-1\) (if \(k>x\)) or between \(k+1\) and x (if \(k<x\)) all have the same head h. This process results in potentially overlapping chunks (e.g., the head of one chunk could be a dependent in another). To select the optimal chunk annotation for a given tree, each chunk is scored based on normalized pointwise mutual information (NPMI)53. We use the NPMI between the Universal part-of-speech (UPOS) tag of a node (t) and the tuple of the UPOS of the head of that node (ht) and the relation between the node and its head (rel). Such that for a given node:
and the average NPMI of a potential chunk is:
where N is the number of nodes in a phrase and d is a dependent in a phrase C. The potential chunks in a given tree are then selected greedily. That is the potential chunks are ordered based on their NPMI and the highest is selected first resulting in any conflicting chunk annotations being removed. This is repeated until no potential chunk labels are left.
This process results in a large number of unitary chunks (i.e., chunks with only one node) which is unlikely to echo the multi-word units of natural language. In order to rectify this, two simple heuristics were applied. The first removes superfluous punctuation (superfluous with respect to the syntactic tree). Punctuation is only removed if a node has a UPOS tag of PUNCT and has no dependents. An example is shown in Fig. 5.
The second heuristic attaches floating unitary nodes to chunks. This in effect removes the single depth restrictions of chunks, which was only introduced to simplify the original engineering use of this method. If a unitary chunk occurs at the boundary of a multi-token chunk and is syntactically linked to any element in that chunk (i.e., is the head or dependent of a node in the chunk), it is included in that chunk and the annotation is updated. Punctuation is treated slightly different. Any punctuation nodes that remain after applying the first heuristic is considered part of a chunk if it satisfies the boundary condition (with the syntactic criterion ignored), as the punctuation does not impact the analysis. An example is shown Fig. 6. Then the derived chunks are viewed as components, which can be a word or a sequence of words that takes into account inter-word relationships such as precedence and dominance. The overall process of generating chunk outputs is summarized in Fig. 7.
Data analysis
Preprocessing and spectral analysis were performed in MatLab® (The MathWorks, Inc., US); statistical analysis was performed in R54. The authors of the original corpus suggested trimming reading times outside a range of 100–3000 ms30. Because the complete removal of word reading times would have disrupted the spectral analysis, and thereby, the actual pace of reading, we kept the original latency of each button press relative to story onset and data values outside the range of 100–3000 ms were replaced with a median value. The median was calculated within subject and story. Imputation affected 6 % of data values. Reading times were then log-transformed to achieve a normal distribution (Fig. 8A). For highlighting chunking-related slowdown–speedup transitions in the data, we performed differencing on the imputed vector of reading times (Fig. 8B). This decision was based on prior evidence for reading-time slowdowns at the end of clauses and sentences55,56 (for review, see29) and independent evidence from visual chunking in non-human primates that observed changes in reaction times at chunk boundaries57. In the differenced vector, local maxima reflect transitions from slowdowns to speedups between adjacent words (Fig. 8C). After differencing, data were converted to a time series sampled at 1000 Hz. The original latency of each button press in milliseconds relative to story onset served as index, the log-transformed reading time served as value.
The time series within subject and story then underwent short-term Fourier transform using Welsh’s power spectral density (PSD) estimation (window length = 4168 samples, overlap = 2084 sample, frequency resolution = 0.1 Hz); PSD was converted to power (see Fig. 2 for results). Statistical analysis employed a permutation approach: First, observed spectra were averaged within subject across stories and a one-sample t-test was performed across subjects within frequency bin; the t-statistic was noted. Second, a distribution of estimates for comparison was generated: Within story and subject, 1000 random time series were generated by randomly permuting the differenced values and inserting them at the observed indices. Spectra were averaged within permutation run and subject across stories; within run, a one-sample t-test was performed across subjects within frequency bin. Third, within frequency bin, we sorted the test statistics from the permuted data and assessed whether the observed statistic would surpass the \(950^{\textrm{th}}\) value, corresponding to one-tailed \(p~<~0.05\)58, and then Bonferroni-corrected for the 100 query frequencies.
To relate slowdown–speedup transitions to chunking, we performed mixed-effects logistic regression analyses using the lme4 package59 in R. Words at sentence boundaries were defined as words followed by a period, question/exclamation mark, comma, or (semi)colon. Words at chunk boundaries were defined by the chunker. This means that sentence boundaries and chunk boundaries were mutually exclusive. Non-boundary words were all remaining words (Fig. 8D). Baseline and improvement models were compared using Analysis of Variance. Word frequency was determined with the wordfreq module in Python; word form surprisal was calculated using the minicons module in Python, based on GPT260. Frequency and surprisal were included as nuisance regressors in all models because of their well-known influence on processing effort (for review, see61,62). Before inclusion, frequency and surprisal were scaled and centered.
As a second strategy for relating slowdown–speedup transition to chunking, we assessed the progression of reading times within chunks. We took a two-level approach: First, within chunk, we linearly regressed reading time on word position. Second, regression slopes (i.e., \(\beta\) coefficients within chunk) were entered as dependent measure into a linear mixed model, fixed effect being only an intercept, random effects being subject, story, and chunk length (i.e., number of words within chunk); note that a random-slope model failed to converge.
Data availability
All data generated or analyzed during this study are included in Futrell et al.30.
References
Christiansen, M. H. & Chater, N. The now-or-never bottleneck: A fundamental constraint on language. Behav. Brain Sci. 39, e62. https://doi.org/10.1017/S0140525X1500031X (2016).
Vasishth, S., Nicenboim, B., Engelmann, F. & Burchert, F. Computational models of retrieval processes in sentence processing. Trends Cogn. Sci. 23, 968–982. https://doi.org/10.1016/j.tics.2019.09.003 (2019).
Pöppel, E. A hierarchical model of temporal perception. Trends Cogn. Sci. 1, 56–61. https://doi.org/10.1016/S1364-6613(97)01008-5 (1997).
Pöppel, E. Oscillations as possible basis for time perception. In The Study of Time (eds Fraser, J. et al.) 565–589 (Springer, 1972).
Elbert, T., Ulrich, R., Rockstroh, B. & Lutzenberger, W. The processing of temporal intervals reflected by CNV-like brain potentials. Psychophysiology 28, 648–655. https://doi.org/10.1111/j.1469-8986.1991.tb01009.x (1991).
Baddeley, A. D., Thomson, N. & Buchanan, M. Word length and the structure of short-term memory. J. Verbal Learn. Verbal Behav. 14, 575–589. https://doi.org/10.1016/S0022-5371(75)80045-4 (1975).
Frazier, L. & Fodor, J. D. The sausage machine: A new two-stage parsing model. Cognition 6, 291–325. https://doi.org/10.1016/0010-0277(78)90002-1 (1978).
Tauroza, S. & Allison, D. Speech rates in British English. Appl. Linguist. 11, 90–105. https://doi.org/10.1093/applin/11.1.90 (1990).
Vollrath, M., Kazenwadel, J. & Krüger, H. P. A universal constant in temporal segmentation of human speech. A reply to Schleidt and Feldhütter (1989). Naturwissenschaften 79, 479–480. https://doi.org/10.1007/BF01139205 (1992).
Vetchinnikova, S., Konina, A., Williams, N., Mikušová, N. & Mauranen, A. Chunking up speech in real time: Linguistic predictors and cognitive constraints. Lang. Cogn.https://doi.org/10.1017/langcog.2023.8 (2023).
Ding, N., Melloni, L., Zhang, H., Tian, X. & Poeppel, D. Cortical tracking of hierarchical linguistic structures in connected speech. Nat. Neurosci. 19, 158–164. https://doi.org/10.1038/nn.4186 (2016).
Ding, N. et al. Characterizing neural entrainment to hierarchical linguistic units using electroencephalography (EEG). Front. Hum. Neurosci. 11, 481. https://doi.org/10.3389/fnhum.2017.00481 (2017).
Meyer, L., Henry, M. J., Gaston, P., Schmuck, N. & Friederici, A. D. Linguistic bias modulates interpretation of speech via neural delta-band oscillations. Cereb. Cortex 27, 4293–4302. https://doi.org/10.1093/cercor/bhw228 (2016).
Henke, L. & Meyer, L. Endogenous oscillations time-constrain linguistic segmentation: Cycling the garden path. Cereb. Cortex 31, 4289–4299. https://doi.org/10.1093/cercor/bhab086 (2021).
Brookshire, G., Lu, J., Nusbaum, H. C., Goldin-Meadow, S. & Casasanto, D. Visual cortex entrains to sign language. Proc. Natl. Acad. Sci. U.S. A. 114, 6352–6357. https://doi.org/10.1073/pnas.1620350114 (2017).
Bourguignon, M., Baart, M., Kapnoula, E. C. & Molinaro, X. Lip-reading enables the brain to synthesize auditory features of unknown silent speech. J. Neurosci. 40, 1053–1065. https://doi.org/10.1523/JNEUROSCI.1101-19.2019 (2020).
Just, M. A. & Carpenter, P. A. A theory of reading: From eye fixations to comprehension. Psychol. Rev. 87, 329–354. https://doi.org/10.1037/0033-295X.87.4.329 (1980).
Rayner, K., Kambe, G. & Duffy, S. A. The effect of clause wrap-up on eye movements during reading. Q. J. Exp. Psychol. 53, 1061–1080. https://doi.org/10.1080/713755934 (2000).
Hirotani, M., Frazier, L. & Rayner, K. Punctuation and intonation effects on clause and sentence wrap-up: Evidence from eye movements. J. Mem. Lang. 54, 425–443. https://doi.org/10.1016/j.jml.2005.12.001 (2006).
Clifton, C., Carlson, K. & Frazier, L. Informative prosodic boundaries. Lang. Speech 45, 87–114. https://doi.org/10.1177/00238309020450020101 (2002).
Henke, L., Lewis, A. G. & Meyer, L. Fast and slow rhythms of naturalistic reading revealed by combined eye-tracking and electroencephalography. J. Neurosci. 43, 4461–4469. https://doi.org/10.1523/JNEUROSCI.1849-22.2023 (2023).
Rayner, K. Eye movements in reading and information processing: 20 years of research. Psychol. Bull. 124, 372–422. https://doi.org/10.1037/0033-2909.124.3.372 (1998).
Nugues, P. M. Language Processing with Perl and Prolog 2nd edn. (Springer, 2014).
Abney, S. Semisupervised Learning for Computational Linguistics 1st edn. (Taylor & Francis Inc., London, 2007).
Anderson, M., Vilares, D. & Gómez-Rodríguez, C. Artificially evolved chunks for morphosyntactic analysis. In Proceedings of the 18th International Workshop on Treebanks and Linguistic Theories (TLT, SyntaxFest 2019) 133–143. https://doi.org/10.18653/v1/W19-7815 (Association for Computational Linguistics, Paris, France, 2019).
de Marneffe, M.-C. & Nivre, J. Dependency grammar. Annu. Rev. Linguist. 5, 197–218. https://doi.org/10.1146/annurev-linguistics-011718-011842 (2019).
Abney, S. Parsing by chunks. In Principle-Based Parsing Studies in Linguistics and Philosophy Vol. 44 (eds Berwick, R., Abney, S. & Tenny, C.) (Springer, 1991).
Ramshaw, L. & Marcus, M. Text chunking using transformation-based learning. In Third Workshop on Very Large Corpora (1995).
Stowe, L. A., Kaan, E., Sabourin, L. & Taylor, R. C. The sentence wrap-up dogma. Cognition 176, 232–247. https://doi.org/10.1016/j.cognition.2018.03.011 (2018).
Futrell, R. et al. The natural stories corpus: A reading-time corpus of English texts containing rare syntactic constructions. Lang. Resour. Eval. 55, 63–77. https://doi.org/10.1007/s10579-020-09503-7 (2021).
Fodor, J. & Bever, T. The psychological reality of linguistic segments. J. Verbal Learn. Verbal Behav. 4, 414–420. https://doi.org/10.1016/S0022-5371(65)80081-0 (1965).
Holmes, V. M. & Forster, K. I. Detection of extraneous signals during sentence recognition. Percept. Psychophys. 7, 297–301. https://doi.org/10.3758/BF03210171 (1970).
Holmes, V. M. & Forster, K. I. Click location and syntactic structure. Percept. Psychophys. 12, 9–15. https://doi.org/10.3758/BF03212836 (1972).
Johnson, S. C. Hierarchical clustering schemes. Psychometrika 32, 241–254. https://doi.org/10.1007/BF02289588 (1967).
Levelt, W. Hierarchial chunking in sentence processing. Percept. Psychophys. 8, 99–103. https://doi.org/10.3758/BF03210182 (1970).
Futrell, R., Mahowald, K. & Gibson, E. Large-scale evidence of dependency length minimization in 37 languages. Proc. Natl. Acad. Sci. U. S. A. 112, 10336–10341. https://doi.org/10.1073/pnas.1502134112 (2015).
Nelson, M. J. et al. Neurophysiological dynamics of phrase-structure building during sentence processing. Proc. Natl. Acad. Sci. U. S. A. 114, E3669–E3678. https://doi.org/10.1073/pnas.1701590114 (2017).
Petrini, S. & Ferrer-i-Cancho, R. The distribution of syntactic dependency distances. arXiv:2211.14620 (2022).
Wagner, M. & Watson, D. G. Experimental and theoretical advances in prosody: A review. Lang. Cogn. Process. 25, 905–945. https://doi.org/10.1080/01690961003589492 (2010).
Grosjean, F., Grosjean, L. & Lane, H. The patterns of silence: Performance structures in sentence production. Cogn. Psychol. 11, 58–81. https://doi.org/10.1016/0010-0285(79)90004-5 (1979).
Truckenbrodt, H. On the relation between syntactic phrases and phonological phrases. Linguist. Inq. 30, 219–255 (1999).
Breen, M. Empirical investigations of the role of implicit prosody in sentence processing. Lang. Linguist. Compass 8, 37–50. https://doi.org/10.1111/lnc3.12061 (2014).
Glushko, A., Poeppel, D. & Steinhauer, K. Overt and implicit prosody contribute to neurophysiological responses previously attributed to grammatical processing. Sci. Rep. 12, 14759. https://doi.org/10.1038/s41598-022-18162-3 (2022).
Buxó-Lugo, A. & Watson, D. G. Evidence for the influence of syntax on prosodic parsing. J. Mem. Lang. 90, 1–13. https://doi.org/10.1016/j.jml.2016.03.001 (2016).
Meyer, L., Sun, Y. & Martin, A. E. Synchronous, but not entrained: Exogenous and endogenous cortical rhythms of speech and language processing. Lang. Cognit. Neurosci. 35, 1089–1099. https://doi.org/10.1080/23273798.2019.1693050 (2020).
Rimmele, J. M., Morillon, B., Poeppel, D. & Arnal, L. H. Proactive sensing of periodic and aperiodic auditory patterns. Trends Cogn. Sci. 22, 870–882. https://doi.org/10.1016/j.tics.2018.08.003 (2018).
Giraud, A.-L. Oscillations for all\(^-\backslash\)_ (-) _/\(^-\)? a commentary on Meyer, Sun & Martin (2020). Lang. Cogn. Neurosci. 35, 1106–1113. https://doi.org/10.1080/23273798.2020.1764990 (2020).
Nivre, J. et al. Universal dependencies v1: A multilingual treebank collection. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16) 1659–1666 (2016).
Kazanina, N. & Tavano, A. What neural oscillations can and cannot do for syntactic structure building. Nat. Rev. Neurosci. 24, 1–16. https://doi.org/10.1038/s41583-022-00659-5 (2023).
Lo, C.-W., Henke, L., Martorell, J. & Meyer, L. When linguistic dogma rejects a neuroscientific hypothesis. Nat. Rev. Neurosci.https://doi.org/10.1038/s41583-023-00738-1 (2023).
Kazanina, N. & Tavano, A. Reply to ‘when linguistic dogma rejects a neuroscientific hypothesis’. Nat. Rev. Neurosci.https://doi.org/10.1038/s41583-023-00739-0 (2023).
Klein, D. & Manning, C. D. Accurate unlexicalized parsing. In Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics 423–430. https://doi.org/10.3115/1075096.1075150 (Association for Computational Linguistics, Sapporo, Japan, 2003).
Bouma, G. Normalized (pointwise) mutual information in collocation extraction. In From Form to Meaning: Processing Texts Automatically, Proceedings of the Biennial GSCL Conference 2009 (2009).
R Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria (2017).
Mitchell, D. C. & Green, D. W. The effects of context and content on immediate processing in reading. Q. J. Exp. Psychol. 30, 609–636. https://doi.org/10.1080/14640747808400689 (1978).
Hill, R. L. & Murray, W. S. Commas and spaces: Effects of punctuation on eye movements and sentence parsing. In Reading as a Perceptual Process 565–589 (Elsevier, 2000).
Tosatto, L., Fagot, J., Nemeth, D. & Rey, A. The evolution of chunks in sequence learning. Cogn. Sci. 46, e13124. https://doi.org/10.1111/cogs.13124 (2022).
Maris, E. & Oostenveld, R. Nonparametric statistical testing of EEG-and MEG-data. J. Neurosci. Methods 164, 177–190. https://doi.org/10.1016/j.jneumeth.2007.03.024 (2007).
Bates, D., Mächler, M., Bolker, B. & Walker, S. Fitting linear mixed-effects models using lme4. J. Stat. Softw. 67, 1–48. https://doi.org/10.18637/jss.v067.i01 (2015).
Radford, A. et al. Language models are unsupervised multitask learners. OpenAI Blog 1, 9 (2019).
Hale, J. Information-theoretical complexity metrics. Lang. Linguist. Compass 10, 397–412. https://doi.org/10.1111/lnc3.12196 (2016).
Sassenhagen, J. How to analyse electrophysiological responses to naturalistic language with time-resolved multiple regression. Lang. Cogn. Neurosci.https://doi.org/10.1080/23273798.2018.1502458 (2019).
Acknowledgements
Research was supported by the Max Planck Society through the award of the Max Planck Research Group Language Cycles to Lars Meyer.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Contributions
L.H. & L.M.: designed research; C.-W.L. & M.A.: performed research; L.M.: analyzed data; C.-W.L., M.A., & L.M.: wrote the paper. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lo, CW., Anderson, M., Henke, L. et al. Periodic fluctuations in reading times reflect multi-word-chunking. Sci Rep 13, 18522 (2023). https://doi.org/10.1038/s41598-023-45536-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-023-45536-y