Abstract
When we are exposed to a novel stimulus sequence, we can learn the sequence by extracting a statistical structure that is potentially embedded in the sequence. This mechanism is called statistical learning, and is considered a fundamental and domain-general process that is innate in humans. In the real-world environment, humans are inevitably exposed to auditory sequences that often overlap with one another, such as speech sound streams from multiple speakers or entangled melody lines generated by multiple instruments. The present study investigated how single and dual attention modulates brain activity, reflecting statistical learning when two auditory sequences were presented simultaneously. The results demonstrated that the effect of statistical learning had more pronounced neural activity when listeners paid attention to only one sequence and ignored the other, rather than paying attention to both sequences. Biased attention may thus be an essential strategy when learners are exposed to multiple information streams.
Similar content being viewed by others
Introduction
The brain is a learning system that adapts to multiple stimuli in a living environment. Learners can acquire a great deal of information attentionally and nonattentionally1, 2. Statistical learning, which is a learning system of transitional probabilities in sequential information such as music and language, has been hypothesised to be a fundamental, domain-general, and automatic process regardless of attention3,4,5,6,7,8,9. Several studies demonstrated that the statistical learning of tone sequences could be reflected in neural responses elicited in the auditory cortex10,11,12,13,14,15,16,17; auditory responses to tones that appeared with a higher transitional probability were reduced compared to those with a lower transitional probability, while listeners acquired statistical knowledge from the exposed tone sequence. However, a previous study suggests that, based on the implicitness of statistical learning in humans, we cannot delineate learned statistical knowledge, even when we tried to “intentionally” learn statistical knowledge13. Although humans are unaware of the learned statistical structure of auditory sequences, they can, instead, recognize sequences with at least nine tones transitioned with high probability as familiar tone sequences13, 14, 16, 17. Thus, statistical knowledge itself is implicit and does not reach the level of explicit awareness, whereas it can be alternatively expressed via an abstract medium such as a musical melody.
Using a Markov chain11, 13, 14, 16, 17 or a word segmentation task10, 12, 15, 17, previous studies demonstrated that auditory statistical learning could be reflected in auditory event-related responses, such as P1/P1 m12, 16, N1/N1 m or mismatch negativity/mismatch field10,11,12,13,14,15, 17, and P2/P2 m11, peaking in a latency range from 50 to 200 ms after stimulus onset. Some studies suggest that the learning effect relationship with P1 m involves music expertise and specialised training experience12, 16. When pure tone sequences were presented, learning effects on P1 m, but not N1 m, were larger in musicians compared with non-musicians12. In our previous study, the statistical learning of atonal chord sequences of augmented triads was reflected in P1 m16. However, when auditory sequences of complex tones were presented, statistical learning was reflected in N1 m13, 14, 17. Thus, the type of the stimulus, the structural character of the stream, attention to the stimuli, and the instructions given to the listeners may modulate neural responses for statistical learning in the auditory cortex. However, the relationships between event-related responses elicited in the auditory cortex and statistical learning effect is still controversial.
Despite nonattentional sensitivity to statistics in humans, we often attentionally access linguistic and musical knowledge that potentially involves statistical information. A previous neurophysiological study on auditory statistical learning demonstrated that attentional statistical learning could be more effective than nonattentional statistical learning13. This suggests that attention to the auditory stimuli modulates the neural basis underlying statistical learning in the auditory cortex. According to the predictive coding theory18, the brain constantly generates probabilistic predictions based on learned knowledge. When a predicted tone is presented, there is a suppression of activities in the primary auditory cortex. In the previous studies on auditory statistical learning, auditory responses to predictable tones were reduced compared to unpredictable tones. Thus, there might be relationships between attention and prediction in auditory mechanisms underlying statistical learning. It is necessary to verify how neural responses in the auditory cortex are modulated by attention, in relation to prediction during auditory statistical learning.
Given the superiority of the attentional strategy over the nonattentional strategy for statistical learning, it is not known why we possess nonattentional learning ability. It might be difficult to pay attention to every detail of multiple information streams that can occur concurrently as a result of cognitive capacity limitations in humans19. Our living environment, however, is rich in overlapping sound streams, such as human voices, musical instruments, water, and wind. In general, learners can acquire a great deal of information through both attentional and nonattentional processes1, 2. A previous study behaviourally demonstrated that attentional and nonattentional learning operated independently and in parallel when learners were presented with two simultaneous streams of stimuli20. Thus, nonattentional learning may be necessary when learners are simultaneously exposed to multiple pieces of auditory information. Few studies have neurophysiologically investigated nonattentional and attentional statistical learning when humans are simultaneously exposed to multiple auditory streams of statistically structured sequences. In the present study, we used magnetoencephalography (MEG) to investigate how attentional and nonattentional learning were reflected in neurological responses in the auditory cortex when participants were presented with two simultaneous streams of tones.
Results
Experimental procedure
The experiment consisted of two sessions. In each session, MEG was carried out on the participants while they were listening to a dyad (two-note chord) sequence, and learning achievement was evaluated using behavioural tests (Fig. 1). The dyad sequences can be separated into high- and low-voiced concurrent tone sequences, within which the intervals were separated by more than an octave and presented every 0.5 s. The tone transition in high- and low-voice sequences was independently ruled by distinct second-order Markov chains, such that a forthcoming tone was statistically defined by the latest two successive tones in each voice (Fig. 2). Hereafter, the tones that appeared with higher and lower transitional probabilities are termed frequent and rare tones, respectively.
In the single attention session, participants were instructed to listen to either the high- or low-voice sequence (attended sequence) and ignore the other sequence (ignored sequence). After MEG measurement, participants completed an interview in which they were presented with 30 series of 8 tones. Participants then reported whether each 8-tone series sounded familiar. The 30 series of 8 tones were categorised into three types, and the presentation order was randomised. In the 10 series, tones were ordered based on the same constraint as the ignored sequence (tone series I). In an additional 10 series, tones were ordered based on the same constraint as the attended sequence (tone series A). In the remaining 10 series, tones were pseudo-randomly ordered (random tone series). The target of analysis was the MEG responses to the dyads, which were categorised into four groups: 2 (frequent, rare) × 2 (attended and ignored sequences).
In the dual attention session, participants were instructed to listen to both the high- and low-voice sequences. After MEG measurement, participants completed an interview in which they were presented with 30 series of 8 tones. Participants then reported whether each 8-tone series sounded familiar. The 30 series of 8 tones were categorised into three types, and the presentation order was randomised. In the 10 series, tones were ordered based on the same constraint as the high-voice sequence (high-voice series). In an additional 10 series, tones were ordered based on the same constraint as the low-voice sequences (low-voice series). In the remaining 10 series, tones were pseudo-randomly ordered (random tone series). The target of analysis was the MEG responses to the dyads, which were categorised into four groups: 2 (frequent, rare) × 2 (high- and low-voiced).
Behavioural results
In the single attention session, the results of two-tailed t-tests indicated that the familiarity ratios were significantly above chance level in both tone series A and I (tone series A: t[14] = 2.74, p = 0.016; tone series I: t[14] = 2.72, p = 0.017) (Fig. 3). The analysis of variance (ANOVA) for the single attention session detected a significant difference (F[2, 28] = 4.79, p = 0.016). The Bonferroni-corrected post-hoc test revealed that there was no significant difference. In the dual attention session, the results of two-tailed t-tests indicated that the familiarity ratios were not significantly different from the level of chance in any type of tone series. No other significant results were detected in the behavioural tests.
Magnetoencephalographic results
The averaged amplitudes and latencies of P1 m are shown in Fig. 4. Because there were insufficient samples of N1 m and P2 m for statistical analysis based on the criteria of equivalent current dipole (ECD) estimation with a goodness-of-fit above 80%, N1 m and P2 m components were excluded from further analysis. In the single attention session, the main stimulus effect on the P1 m peak amplitudes and latencies were significant (amplitudes F[3, 33] = 5.62, p = 0.0030; latencies F[3, 33] = 4.52, p = 0.0092). The P1 m peak amplitudes for the dyads that consisted of two rare tones in both the ignored and attended sequences were significantly increased compared with those for dyads that consisted of a rare tone in the ignored sequence and a frequent tone in the attended sequence (p = 0.010), and also those for the dyads that consisted of two frequent tones in both the ignored and attended sequences (p = 0.0037; Fig. 4). The P1 m peak latencies for the dyads that consisted of a frequent tone in the ignored sequence and a rare tone in the attended sequence were significantly longer compared with those for the dyads that consisted of a rare tone in the ignored sequence and a frequent tone in the attended sequence (p = 0.042), and also those for the dyads that consisted of two frequent tones in both the ignored and attended sequences (p = 0.042; Fig. 4). The hemisphere-tone interactions of the P1 m peak latencies were significant (F[3, 33] = 3.83, p = 0.019). In the dyads that consisted of a rare tone in the ignored sequence and a frequent tone in the attended sequence, the P1 m peak latencies were significantly shorter in the left than the right hemispheres (p = 0.022). In the left hemisphere, the P1 m peak latencies for the dyads that consisted of a frequent tone in the ignored sequence and a rare tone in the attended sequence were significantly longer compared with those for dyads that consisted of a rare tone in the ignored sequence and a frequent tone in the attended sequence (p = 0.014), and also those for the dyads that consisted of two frequent tones in both the ignored and attended sequences (p = 0.023). In the dual attention session, no other significant differences were detected.
Discussion
By learning statistics of transitional probabilities embedded in tone sequences, learners can predict a tone that will follow certain preceding tones in the sequence. With this prediction for upcoming tones, tones with higher transitional probability (i.e., more predictable tones) lead to a decrease in amplitude and shortening of latencies in neural responses. In contrast, tones with lower transitional probability (i.e., less predictable tones) lead to an increased neural response amplitude10,11,12,13,14,15,16,17. In the present study, participants were presented with two simultaneous tone sequences that had tones with higher and lower transitional probabilities (i.e., frequent and rare tones, respectively). Based on the combinations of frequent and rare tones in the two simultaneous tone sequences, there were four types of dyads: a dyad that consisted of two frequent tones in both sequences, a dyad that consisted of two rare tones in both sequences, a dyad that consisted of a frequent tone in a sequence and a rare tone in the other sequence, and vice versa. If participants could acquire statistical knowledge in the two tone sequences, the dyad that consisted of two frequent tones should have the lowest amplitudes, and those consisting of two rare tones should have the highest amplitudes. In contrast, the increase in responses to the dyad that consisted of a frequent tone and a rare tone can be interpreted as a statistical learning effect of a sequence with a rare tone.
In the single attention session, the participants were instructed to listen to one of the two simultaneous sequences and ignore the other sequence. In the dual attention session, the participants were instructed to listen to both of the two simultaneous sequences. As a result, in the single attention session, the neural responses to the dyad that consisted of two rare tones in both the attended and ignored sequences were significantly increased compared with those that consisted of two frequent tones in both the attended and ignored sequences. The chord that consisted of two rare tones in both the attended and ignored sequences evoked the highest amplitude of the four types of chords. The chord that consisted of a rare tone in the attended sequence and a frequent tone in the ignored sequence evoked the second highest amplitude of the four. The chord that consisted of a rare tone in the ignored sequence and a frequent tone in the attended sequence evoked the third highest amplitude of the four. The chord that consisted of two frequent tones in both the attended and ignored sequences evoked the lowest amplitude of the four. The peak response latency for the chords that consisted of a rare tone in the ignored sequence and a frequent tone in the attended sequence were significantly shorter compared with those for the dyads that consisted of a frequent tone in the ignored sequence and a rare tone in the attended sequence. These results suggest that statistical learning was facilitated in attentional learning but not in nonattentional learning, which is in agreement with the results of our previous study13.
In contrast, no significant statistical learning effect was detected in the dual attention session. The findings in the single and dual attention sessions were also consistent with the behavioural data. Our results suggest that the statistical learning of two simultaneous auditory sequences may be facilitated by paying attention only to one sequence (i.e., attentional learning) and ignoring the other sequence (i.e., nonattentional learning), whereas the learning effect could not be detected when paying attention to both sequences. This hypothesis might be consistent with previous studies19. Attentional access to much of the information that occurs concurrently could interfere with the acquisition of this information because cognitive capacity is limited in humans. In natural auditory environments, however, learners can concurrently acquire a great deal of information through both attentional and nonattentional processes. Earlier studies suggested that the brain regions and activation patterns engaged during attentional and nonattentional learning might be partially distinct21,22,23,24,25. It has been reported that the neural basis for the spatial dichotomy that underlies attentional and nonattentional learning predominantly depends on the frontal lobe and the striatum, respectively26,27,28,29. In other words, there might be a specific cognitive capacity underlying nonattentional learning that is independent of the capacity underlying attentional learning. A previous study behaviourally demonstrated that attentional and nonattentional learning operates independently and in parallel when learners were presented with two simultaneous streams of stimuli19. Our neurophysiological findings are consistent with those of the previous study. The attentional and nonattentional statistical learning of two simultaneous auditory sequences was reflected in P1 responses, which have been considered to be generated in the vicinity of the primary auditory cortex. When humans have learned transitional probabilities in an auditory sequence, they can predict a forthcoming tone that will frequently follow preceding tones in the sequence. Prediction of forthcoming tones that will appear with higher transitional probability reduces P1 responses in the auditory cortex. Through biased attention in the present study, participants might be able to clearly distinguish and better predict each sequence. We could not, however, demonstrate a difference in the neural basis underlying nonattentional and attentional statistical learning because of the methodological limitations of this study. Further research is needed to clarify the neural substrates for nonattentional and attentional learning.
Previous studies suggest that statistical learning can be reflected in the late components such as N1 and P2 and also in the earlier component, P110,11,12,13,14,15, 17. Previous studies reported that learning effects on P1 were not correlated with the other event-related responses30,31,32. Some studies suggest that the learning effect relationship with P1 involves music expertise and specialised training experience33, 34. Paraskevopoulos et al. demonstrated that, in the initial phase of statistical learning, learning effects on P1, but not N1, were larger in musicians compared with non-musicians12. In our previous study, the statistical learning of chord sequences was reflected in P116. Another study reported that, in learning the chord progression with conditional probability, the learning effects on the later responses such as early anterior negativity (EAN: 150–250 ms)35, 36 were facilitated by musical training37. Especially in the initial learning phase such as statistical learning, earlier responses of the P1 may be more associated with perception of musical sequences compared with other components.
In the neurophysiological studies using the paired-click paradigm, the P1 responses are measured as a marker of sensory gaiting function38. Compared to single sine tones, the click tones contain a broad frequency spectrum and recruit more neural activity. In the present study, the complex tones may have helped elicit the P1 responses, otherwise showing relatively high inter-individual variability among other components. In addition, infinite averaging of continuous data at every SOA of 0.5 s eliminates signals below a frequency of 2 Hz. The relatively short SOA of 0.5 s corresponding to the applied high-pass filter of 2 Hz may be critical to refractory recovery of the late components such as P2.
Neuroimaging studies have shown that the P1 and N1 components are generated in the auditory cortex with different topographies39, 40. P1 and N1 are generated in the lateral part of the primary auditory cortex and the secondary auditory cortex, respectively. Neurophysiological effects of statistical learning on neural responses in the auditory cortex can also be explained in the framework of predictive coding in a top-down manner18. The brain constantly generates probabilistic predictions of what is going to happen. The auditory input is compared with the expected tone and produces a signal that codes a prediction error. The lower the probability of the sensory input, the greater the prediction error and reaction to the stimulus. When no error occurs, there is a suppression of responses encoding prediction error in the primary auditory cortex. According to previous studies, earlier auditory responses that peaked at 20–80 ms, which is around P1 latency, were attributed to parallel thalamo-cortical connections, or to cortico-cortical connections between the primary auditory cortex and the superior temporal gyrus38. Thus, an early component of auditory responses in lower cortical areas can be interpreted as the transient expression of prediction error that is suppressed by predictions from higher cortical areas in a top-down connection18. This suppression may be compromised if the sequences have not been learned previously. The difference in the behaviour of the P1 and N1 responses in statistical learning suggests that the neural basis of the P1 and N1 components reflecting auditory statistical learning is at least partially different. Further studies are needed to clarify the specific attributes of P1 in statistical learning.
In conclusion, we demonstrated that the statistical learning of two simultaneous auditory sequences might be facilitated by paying attention to only one sequence and ignoring the other sequence, whereas the learning effect could not be detected when paying attention to both sequences. Our results suggest that there could be a partially distinct neural basis underlying nonattentional and attentional statistical learning. Biased attention may be an essential strategy under conditions where learners are exposed to multiple information streams.
Experimental procedure
Participants
Fifteen right-handed (Edinburgh handedness questionnaires; laterality quotient ranged from 57.9–100)41 healthy participants with no history of neurological or audiological disorders were included (9 males, 6 females; age range, 24–36 years). No participants had experience living abroad, and no participants possessed absolute pitch. This study was approved by the Ethics Committee of The University of Tokyo and performed in accordance with the guidelines and regulations. All participants were well informed of the purpose, safety, and protection of personal data in this experiment, and they provided written informed consent for this study.
Stimuli
Tones
Using a cascade-Klatt-type formant synthesizer42 HLsyn (Sensimetrics Corporation, Malden, MA, USA), we generated eight complex tones that consisted of fundamental frequencies (F0) in a five-tone equal temperament (F0 = 100 × 2(n−1)/5 Hz). The nine tones consisted of 4 low and high pitches each (low pitch: n = 1–4; 100, 115, 132, and 152 Hz, high pitch: n = 11–14; 400, 459, 528, and 606 Hz). Only F0 s were variable; all of the other parameters were constant (duration, 350 ms; rise/fall, 10/150 ms; binaural presentation, 80 dBSPL intensity).
Sequences
The auditory stimulus sequence was 728 repetitions of two-tone chords (dyads), each of which consisted of a low and high pitch. Within each chord, the intervals were separated by more than an octave, and they were presented with a stimulus onset asynchrony (SOA) of 500 ms. The order of low and high voice in the dyads was defined according to second-order Markov processes with the constraint that the probability of a forthcoming tone was statistically defined (80% for a tone; 6.67% for the other three tones) by the last two successive tones (Fig. 2). The distinct two of four Markov chains shown in Fig. 2 were used in each of the low and high voices, and the use of Markov chains was counterbalanced across participants. The dyad sequences can also be interpreted as two simultaneous sequences that consisted of low- and high-voice sequences (Fig. 1).
Experimental protocol
Participants completed two sessions: single and dual attention sessions. In each session, exposure to the sequence during MEG measurement was preceded by a behavioural test. The order of the two sessions was counterbalanced across participants to ensure that specific transitional patterns did not interfere with learning in adjacent experimental sessions.
Single attention session
Participants were instructed to listen to a sequence (attended sequence) and ignore the other sequence (ignored sequence). The use of attended and ignored sequences in two simultaneous sequences was counterbalanced across participants. To distinguish between ignored and attended conditions, there was a 500 ms silent period that was pseudo-randomly inserted (i.e., SOA 1000 ms) within every set of 40 successive tones in the attended sequence only. Before the session, participants were instructed to raise their right hand at every silent period in the attended sequence and ignore the other sequence. Using these approaches, we confirmed that all of the participants correctly raised their right hand at every silent period in the attended sequences, and that they were continuing to pay attention only to that sequence.
After measuring MEG, participants completed an interview in which they were presented with 30 series of 8 single tones. Participants then reported whether each 8-tone series sounded familiar. The 30 series of 8 tones were categorised into three types, and the presentation order was randomised. In the 10 series, tones were ordered based on the same constraint as the ignored sequence (tone series I). In an additional 10 series, tones were ordered based on the same constraint as the attended sequence (tone series A). In the remaining 10 series, tones were pseudo-randomly ordered (random tone series).
Dual attention session
Participants were instructed to listen to both of the two simultaneous sequences. To distinguish between two attended conditions, a 500-ms silent period was pseudo-randomly inserted (i.e., SOA 1000 ms) within every set of 40 successive tones independently in each sequence. Before the session, participants were instructed to raise their right and left hands at every silent period in a sequence and at the other sequences of the two simultaneous sequences, respectively. Thus, we could confirm that all participants correctly raised their hands at every silent period in both sequences, and that they continued to pay attention to each sequence independently.
After the MEG measurement, participants completed an interview in which they were presented with 30 series of 8 single tones. Participants then reported whether each 8-tone series sounded familiar. The 30 series of 8 tones could be categorised into three types, and the presentation order was randomised. In the 10 series, tones were ordered based on the same constraint as the high-voice sequence (high-voice series). In an additional 10 series, tones were ordered based on the same constraint as lower-pitch sequences (low-voice series). In the remaining 10 series, tones were pseudo-randomly ordered (random tone series).
Measurement
Measurement and analysis were performed as in our previous studies13, 14, 16, 17. We recorded MEG signals from participants while they listened to the two simultaneous sequences. Auditory stimuli were sequenced using the STIM2 system (Compumedics Neuroscan, El Paso, TX, USA), and were binaurally delivered to participant’s ears at 80 dBSPL through ER-3 A earphones (Etymotic Research, Elk Grove Village, IL, USA). MEG signals were recorded in a magnetically shielded room, using a 306-channel neuromagnetometer system (Elekta Neuromag Oy, Helsinki, Finland) with 204 planar first-order gradiometers and 102 magnetometers at 102 measuring sites on a helmet-shaped surface that covers the entire scalp. Auditory stimulus-triggered epochs were filtered online with a 0.1 to 200 Hz band-pass filter and were then recorded at a sampling rate of 600 Hz.
Data analysis
Epochs with artefacts that exceeded 3 pT/cm or 3 pT for any MEG channel were excluded from analyses. Contamination from environmental noise was reduced using the temporally extended signal space separation method with a buffer length of 10 s and a correlation limit of 0.98043. To extract learning effects from the neuromagnetic response series, neuromagnetic responses were selectively averaged from the beginning of the latter half of the sequence until the average number was reached twice for each dyad, and was distinguished according to transitional probabilities. In addition to selective averaging, all responses to dyads in the two sequences were averaged for each participant to evaluate the reliability of the evoked response individual components. The averaged responses were filtered offline with a 2 to 40 Hz band-pass. The baseline for magnetic signals in each MEG channel was defined by the mean amplitude in the pre-stimulus period from −100 to 0 ms. The analysis window was defined as 0 to 500 ms.
The P1 m, N1 m, and P2 m responses were separately modelled as single attention ECDs in each hemisphere. The ECDs for the P1 m, N1 m, and P2 m responses to all dyads were separately estimated at the peak latency using 66 temporal channels for each participant (Fig. 5a). Participants who demonstrated poor ECD estimation, with a goodness-of-fit below 80% in either the left or right hemisphere, were not used in further analysis. The number of participants who demonstrated ECD estimation with a goodness-of-fit above 80% for the P1 m, N1 m, and P2 m components was 12, 5, and 9, respectively. Because there were insufficient samples of N1 m and P2 m for statistical analysis, these components were excluded from further analysis.
The source-strength waveforms for P1 m in each hemisphere were calculated using the ECDs as templates (Fig. 5). Then, in the single attention session, we performed a 2 (hemisphere: right and left) × 4 (stimulus: chord consisting of two frequent tones both in the attended and ignored sequences; chord consisting of two rare tones in both the attended and ignored sequences; chord consisting of a frequent tone in the attended sequence and a rare tone in the ignored sequence; and chord consisting of a rare tone in the attended sequence and a frequent tone in the ignored sequence) repeated-measures ANOVA with peak amplitude and latency of source-strength of P1 m in the time window of 30 ms–90 ms. When we detected significant effects, Bonferroni-corrected post-hoc tests were conducted for further analysis. In the dual attention session, we performed a 2 (hemisphere: right and left) × 4 (stimulus: chord consisting of two frequent tones in both high- and low-voice sequences; chord consisting of two rare tones in both the high- and low-voice sequences; chord consisting of a rare tone in the high-voice sequence and a frequent tone in the low-voice sequence; and chord consisting of a rare tone in the low-voice sequence and a frequent tone in the high-voice sequence) repeated-measures ANOVA with peak amplitude and latency of source-strength of P1 m. When we detected significant effects, a Bonferroni-corrected post-hoc test was conducted for further analysis. In the behavioural test, logit transformation was applied to normalize the familiarity ratios (ratios of answering that the tone series sounded familiar to the number of tone series). We performed an ANOVA using the logit values of the familiarity ratios for each series. When we detected significant effects, Bonferroni-corrected post-hoc tests were conducted for further analysis. Statistical significance levels were set at p = 0.05. We could not include between-session factors (i.e., single and dual attention sessions) in the statistical analysis of MEG and behavioural responses, because simultaneous sequences were divided into attended and ignored sequences in the single attention session, and high- and low-voiced sequences in the dual attention session.
References
Jimenez, L. & Castor, M. Which Attention Is Needed for Implicit Sequence Learning? Journal of Experimental Psychology: Learning, Memory, and Cognition. 25, 236–259 (1999).
Aizenstein, H. J. et al. Regional brain activation during concurrent implicit and explicit sequence learning. Cerebral Cortex. 14, 199–208 (2004).
Saffran, J. R., Aslin, R. N. & Newport, E. L. Statistical learning by 8-month-old infants. Science. 274, 1926–1928 (1996).
Perruchet, P. & Pacton, S. Implicit learning and statistical learning: one phenomenon, two approaches. Trends Cogn Sci. 10, 233–238, doi:10.1016/j.tics.2006.03.006 (2006).
Frost, R. L. & Monaghan, P. Simultaneous segmentation and generalisation of nonadjacent dependencies from continuous speech. Cognition. 147, 70–74 (2016).
Ettlinger, M., Margulis, E. H. & Wong, P. C. Implicit memory in music and language. Frontiers in Psychology 2, 211, doi:10.3389/fpsyg.2011.00211 (2011).
Francois, C. & Schön, D. Musical expertise boosts implicit learning of both musical and linguistic structures. Cerebral Cortex 21, 2357–2365 (2011).
Cleeremans, A., Destrebecqz, A. & Boyer, M. Implicit learning: news from the front. Trends in Cognitive Sciences 2, 406–416 (1998).
Tillmann, B., Bharucha, J. J. & Bigand, E. Implicit learning of tonality: a self-organizing approach. Psychological Review 107, 885–913 (2000).
Abla, D., Katahira, K. & Okanoya, K. On-line Assessment of Statistical Learning by Event-related Potentials. Journal of Cognitive Neuroscience 20, 952–964 (2008).
Furl, N. et al. Neural prediction of higher-order auditory sequence statistics. Neuroimage 54, 2267–2277 (2011).
Paraskevopoulos, E., Kuchenbuch, A., Herholz, S. C. & Pantev, C. Statistical learning effects in musicians and non-musicians: an MEG study. Neuropsychologia 50, 341–349 (2012).
Daikoku, T., Yatomi, Y. & Yumoto, M. Implicit and explicit statistical learning of tone sequences across spectral shifts. Neuropsychologia. 63, 194–204 (2014).
Daikoku, T., Yatomi, Y. & Yumoto, M. Statistical learning of music- and language-like sequences and tolerance for spectral shifts. Neurobiology of Learning and Memory 118, 8–19 (2015).
Koelsch, S., Busch, T., Jentschke, S. & Rohrmeier, M. Under the hood of statistical learning: A statistical MMN reflects the magnitude of transitional probabilities in auditory sequences. Scientific Reports 6, 19741 (2016).
Daikoku, T., Yatomi, Y. & Yumoto, M. Pitch-class distribution modulates the statistical learning of atonal chord sequences. Brain and Cognition 108, 1–10 (2016).
Daikoku, T., Yatomi, Y. & Yumoto, M. Statistical learning of an auditory sequence and reorganization of acquired knowledge: A time course of word segmentation and ordering. Neuropsychologia 95, 1–10 (2017).
Friston, K. A theory of cortical responses. Phil. Trans. R. Soc. B 360, 815–836 (2005).
Watanabe, K. & Funahashi, S. Neural mechanisms of dual-task interference and cognitive capacity limitation in the prefrontal cortex. Nature Neuroscience 17, 601–611 (2014).
Curran, T. & Keele, S. W. Attentional and non-attentional forms of sequence learning. Journal of Experimental Psychology: Learning, Memory and Cognition 19, 189–202 (1993).
Rauch, S. L. et al. A PET investigation of implicit and explicit sequence learning. Human Brain Mapping 3, 271–286 (1995).
Poldrack, R. A., Clark, J., PareÂ-Blagoev, E. J. & Shohamy, D. Interactivememory systems in the human brain. Nature 414, 546–550 (2001).
Reber, P. J. & Squire, L. R. Encapsulation of implicit and explicit memory in sequence learning. Journal of Cognitive Neuroscience 10, 248–263 (1998).
Paradis, M. A Neurolinguistic Theory of Bilingualism. Amsterdam : John Benjamins (2004).
Destrebecqz, A. et al. The neural correlates of implicit and explicit sequence learning: Interacting networks revealed by the process dissociation procedure. Learning & Memory 12, 480–490 (2005).
Miller, E. K. & Cohen, J. D. An integrative theory of prefrontal cortex function. Annual Review of Neuroscience. 24, 167–202 (2001).
Corbetta, M. & Shulman, G. L. Control of goal-directed and stimulus-driven attention in the brain. Nature Reviews Neuroscience 3, 201–215 (2002).
Stevens, D. J., Arciuli, J. & Anderson, D. I. Concurrent Movement Impairs Incidental But Not Intentional Statistical Learning. Cognitive Science. 39, 1–18 (2014).
Daikoku, T., Takahashi, Y., Futagami, H., Tarumoto, N. & Yasuda, H. Physical fitness modulates incidental but not intentional statistical learning of simultaneous auditory sequences during concurrent physical exercise. Neurol Res. 39, 107–116 (2017).
Boutros, N. N. & Belger, A. Midlatency evoked potentials attenuation and augmentation reflect different aspects of sensory gating. Biological Psychiatry 45, 917–922 (1999).
Boutros, N. N. et al. The P50 evoked potential component and mismatch detection in normal volunteers: Implications for the study of sensory gating. Psychiatry Research 57, 83–88 (1995).
Kisley, M. A., Noecker, T. L. & Guinther, P. M. Comparison of sensory gating to mismatch negativity and self-reported perceptual phenomena in healthy adults. Psychophysiology 41, 604–612 (2004).
Kizkin, S., Karlidag, R., Ozcan, C. & Ozisik, H. I. Reduced P50 auditory sensory gating response in professional musicians. Brain and Cognition 61, 249–254 (2006).
Wang, W., Staffaroni, L., Reid, E., Steinschneider, M. & Sussman, E. Effects of musical training on sound pattern processing in high-school students. International Journal of Pediatric Otorhinolaryngology 73, 751–755 (2009).
Koelsch, S., Gunter, T., Friederici, A. D. & Schroger, E. Brain indices of music processing: “Non-musicians” are musical. Journal of Cognitive Neuroscience 12, 520–541 (2000).
Koelsch, S. Music-syntactic processing and auditory memory: Similarities and differences between ERAN and MMN. Psychophysiology 46, 179–190 (2009).
Kim, S. G., Kim, J. S. & Chung, C. K. The effect of conditional probability of chord progression on brain response: An MEG study. PLoS One 6, doi:10.1371/journal.pone.0017337 (2011).
Adler, L. E. et al. Neurophysiological evidence for a defect in neuronal mechanisms involved in sensory gating in schizophrenia. Biological Psychiatry 17, 639–654 (1982).
Liégeois-Chauvel, C., Musolino, A., Badier, J. M., Marquis, P. & Chauvel, P. Evoked potentials recorded from the auditory cortex in man: Evaluation and topography of the middle latency components. Electroencephalography and Clinical Neurophysiology 92, 204–214 (1994).
Yvert, B., Crouzeix, A., Bertrand, O., Seither-Preisler, A. & Pantev, C. Multiple supratemporal sources of magnetic and electric auditory evoked middle latency components in humans. Cerebral Cortex 11, 411–423 (2001).
Oldfield, R. C. The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia 9, 97–113 (1971).
Klatt, D. H. Software for a Cascade/Parallel Formant Synthesizer. Journal of the Acoustical Society of America 67, 971–995 (1980).
Taulu, S. & Hari, R. Removal of magnetoencephalographic artifacts with temporal signal-space separation: demonstration with single-trial auditory-evoked responses. Human Brain Mapping 30, 1524–1534 (2009).
Acknowledgements
This work was supported by Grant-in-Aid for Scientific Research in The Kao foundation for Arts and Sciences, and Kawai Foundation for sound technology & music. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Author information
Authors and Affiliations
Contributions
The experimental paradigms of the present study were considered by both of the authors. Tatsuya Daikoku made the paradigms, recruit participants, and collected the data. Masato Yumoto proposed methodologies of MEG/behavioural data analyses and figures in the manuscript. Tatsuya Daikoku analyzed all of the data. Both of the authors discussed how the results could neurophysiologically and psychologically be interpreted. Tatsuya Daikoku prepared the figures, and wrote the main manuscript text. Then, both of the authors reviewed and revised the manuscript.
Corresponding author
Ethics declarations
Competing Interests
The authors declare that they have no competing interests.
Additional information
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Daikoku, T., Yumoto, M. Single, but not dual, attention facilitates statistical learning of two concurrent auditory sequences. Sci Rep 7, 10108 (2017). https://doi.org/10.1038/s41598-017-10476-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-017-10476-x
This article is cited by
-
Seeking the neural representation of statistical properties in print during implicit processing of visual words
npj Science of Learning (2023)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.