Behavioral rhythms synchronize between humans for communication; however, the relationship of brain rhythm synchronization during speech rhythm synchronization between individuals remains unclear. Here, we conducted alternating speech tasks in which two subjects alternately pronounced letters of the alphabet during hyperscanning electroencephalography. Twenty pairs of subjects performed the task before and after each subject individually performed the task with a machine that pronounced letters at almost constant intervals. Speech rhythms were more likely to become synchronized in human–human tasks than human–machine tasks. Moreover, theta/alpha (6–12 Hz) amplitudes synchronized in the same temporal and lateral-parietal regions in each pair. Behavioral and inter-brain synchronizations were enhanced after human–machine tasks. These results indicate that inter-brain synchronizations are tightly linked to speech synchronizations between subjects. Furthermore, theta/alpha inter-brain synchronizations were also found in subjects while they observed human–machine tasks, which suggests that the inter-brain synchronization might reflect empathy for others' speech rhythms.
Individual human behavioral rhythms in nature are independent but can be spontaneously synchronized and entrained to become a shared rhythm through interactions with others (i.e., social interactions) by both verbal and nonverbal communication1,2. We experience daily synchronizations with others, such as with hand clapping or foot tapping and incidental coordination of speech frequencies in conversations3,4,5,6. This unconscious, shared rhythm brings individuals close to each other, generates empathy, and coordinates performance7,8. Indeed, mother-infant rhythmic coupling through the imitation of movement supports social and cognitive development, such as language acquisition and learning9.
Coordinated rhythms have been observed in the brain not only as local- and distant-regional synchronizations within an individual brain10,11 but also as inter-brain synchronizations between individuals12. Human brain rhythms in specific electroencephalography (EEG) frequency bands have increased not only while performing synchronized behaviors but also while observing such behaviors13,14. These brain activities are related to the understanding of others' intentions (i.e., social cognition), in this case, the understanding of others' behavioral rhythms15. Moreover, recent EEG studies that scanned multiple brains at the same time have shown phase synchronizations between individuals during the imitation of hand movements16,17,18,19 or with participation in cooperative games and actions20,21,22. Synchronized changes between individual brains have also been reported in functional magnetic resonance imaging (fMRI; 23–26) and near-infrared spectroscopy (fNIRS; 27, 28) studies.
Although there is considerable evidence of behavioral synchronization and brain synchronization between individuals during social interactions through nonverbal communication, these relationships are not as clear during verbal communication. One EEG hyperscanning study revealed correlated brain activities between the speakers and listeners during verbal communication31, but this study did not address the behavioral rhythm (i.e., the speech rhythm) itself. Furthermore, it is poorly understood whether such synchronizations include both simultaneous common movement and turn-taking, the latter of which is typically non-simultaneous and requires unconscious alternation of behavior, such as between speakers during a conversation29. Turn-taking requires interpersonal synchronization of speech rhythms, including timing, duration, interval, and speed of speech, along with the content and context of the conversation itself30.
In the current study, we attempted to address whether inter-brain synchronizations on an EEG appear when speech rhythms are synchronized between two subjects in verbal communication. We conducted alternating speech tasks in which two subjects alternately and sequentially pronounced the alphabet during hyperscanning EEG recordings (human–human, Fig. 1A). The behavioral and brain synchronization between the subjects were evaluated in terms of the correlation of speech rhythms (duration and interval of pronunciation) and brain rhythms (EEG oscillatory amplitudes in specific frequency bands), respectively.
In addition, each subject participated in alternating speech tasks with a machine (human–machine, Fig. 1A), and these results were compared with the results of the human–human tasks to address the following questions: (1) How does inter-brain synchronization change after the two subjects' behavioral rhythms have been coordinated to common rhythms (i.e., the machine's rhythms), and (2) how does inter-brain synchronization between a subject performing and a subject observing the task with a machine (social cognition) differ from the inter-brain synchronization between subjects in the human condition (social interaction)?
Speech rhythms: voice sounds
Each subject performed two types of alternating speech tasks: human–human tasks and human–machine tasks (Fig. 1A). They completed 14 sessions consisting of two pre-machine human–human sessions, 10 successive human–machine sessions [five voices (electronic, male, female, partner's, and subject's) at two paces (fixed and random)], and two post-machine human–human sessions (Fig. 1B). We recorded the sounds of human–human alternating speech tasks. We dissociated the durations of voices and intervals between the voices of the pair of subjects using short-time Fourier transformations (Fig. 2A).
During the human–human tasks, the durations and intervals of speech were correlated between the pair of subjects and not significantly different among the subjects (Fig. 3C, D). Notably, compared with the pre-machine tasks using one-factor repeated-measures ANOVA (pre-machine vs. post-machine), the pair-averaged correlations for the post-machine tasks were significantly higher (Fig. 3D; duration, F1, 104 = 4.50, P = 0.036 interval, F1, 104 = 6.19, P = 0.014) and the differences were significantly lower (Fig. 3E; duration, F1, 104 = 5.49, P = 0.021; interval, F1, 104 = 9.30, P = 0.003). Pre-machine, the durations and intervals were not correlated between the two subjects in six pairs. Moreover, the durations were significantly different between the two subjects in four pairs, and the intervals were significantly different between the two subjects in three pairs. Post-machine, 17 of 18 pairs showed significant correlations in the duration and interval of the speech (example shown in Fig. 3B, C).
For human–machine tasks, the correlations between subject pairs were lower, and the differences between subject pairs were higher than those for the human–human tasks [Fig. 3C, D; correlation of duration, F1, 338 = 58.36, P = 0.001; correlation of interval, F1, 338 = 127.01, P = 0.001; difference of duration, F1, 338 = 8.09, P = 0.004; difference of interval, F1, 338 = 9.04, P = 0.002; one-factor repeated-measures ANOVA (human vs. machine conditions)].
Brain rhythms: EEG oscillations
To characterize brain oscillatory activity, we conducted wavelet analysis on the collected EEG data28,29. The subject-averaged amplitudes of each frequency (ranging from 4 to 28 Hz) for performing the alternating speech tasks in both human–human and human–machine conditions were significantly higher than those for subjects who were observing the human–machine tasks [Fig. 4A; P < 0.01; one-factor repeated-measures ANOVA (human vs. machine conditions)]. Frequency ranges (4–28 Hz) were divided into two parts: theta/alpha bands (6–12 Hz), which showed higher amplitudes for human than machine conditions, and beta bands (20–28 Hz), which showed the opposite (machine > human).
The theta/alpha band amplitudes increased in the frontal regions during both human–human and human–machine tasks and extended to the central and parietal regions during human–human tasks (Fig. 4B). The theta/alpha amplitudes during human–human tasks were significantly higher than those during the human–machine tasks in the central and parietal regions (electrodes C3, T7, CP1, and CP2). For the human–human tasks, the post-machine activities were significantly higher than pre-machine activities in the Fp2, C3, and P4 electrodes.
Similar to the theta/alpha amplitudes, the beta band amplitudes showed enhancement in the frontal regions but extended to the temporal and occipital regions under both human and machine conditions. However, there was no significant difference between the conditions for any electrode. Moreover, the beta activities showed no significant difference between the post-machine and pre-machine human–human tasks.
To investigate inter-brain synchronization during alternating speech, we conducted cross-correlation analyses of the theta/alpha amplitudes from each pair of subjects' EEG data because these amplitudes were significantly different during the human–human and human–machine tasks. An example of the time course of a theta/alpha amplitude and the cross-correlation coefficients for one subject pair during a human–human alternating speech task is shown in Fig. 5A (left). In pair-averaged correlations of the theta/alpha amplitudes, high-peaked correlations were distributed in the temporal regions (electrodes F7, FC5, T7, T8, CP1, and CP2) for both pre- and post-machine using cross-correlation analyses that included both the pre- and post-machine conditions (Fig. 5B). For most of the electrodes, these values were significantly higher for the post-machine conditions [Fig. 5D; P < 0.05; one-factor repeated-measures ANOVA (pre-machine vs. post-machine) with post-hoc analyses (Bonferroni correction)].
The theta/alpha cross-correlation coefficients for the temporal and parietal electrodes were significantly correlated with the individual behavioral synchronizations between the subjects; in other words, there were high correlations between the speech durations and intervals between the two subjects (Fig. 5E; electrode measuring the peak statistic value, T7; correlation of duration, r = 0.49, P = 0.046; correlation of interval, r = 0.44, P = 0.071).
Finally, we investigated theta/alpha inter-brain synchronization between subjects who participated in alternating speech tasks with a machine and the subjects and observed the human–machine tasks (an example is shown in Fig. 5A, right). High-peaked correlations partially overlapped in the temporal regions, the same area in which significant correlations were seen in subjects during human–human tasks (compare Fig. 5B and Fig. 5C; electrodes F7, T7, CP1, and CP2). However, the theta/alpha cross-correlation coefficients for the temporal regions were not significantly correlated with the individual behavioral synchronizations between subjects (Fig. 5E; electrode measuring the peak statistic value, T7; correlation of duration, r = −0.25, P = 0.320; correlation of interval, r = −0.14, P = 0.591).
After each alternating speech session, the subjects were asked to rate the following factors using a 5-point scale: “comfort,” “synchrony,” “speed,” “initiative,” and “humanity” (only for the human–machine tasks). For all of the factors except for humanity, ratings for the human partners were significantly higher than those for the machine partner [comfort, F1, 502 = 82.54, P = 0.001; synchrony, F1, 502 = 80.75, P = 0.001; speed, F1, 502 = 7.59, P = 0.006; initiative, F1, 502 = 92.30, P = 0.001; one-factor repeated-measures ANOVA (human vs. machine conditions)]. These results suggested that the subjects felt more comfortable, had better and faster synchronization, and had a higher initiative when performing alternating speech with a human partner than with the machine partner.
For the human–human tasks, post-machine ratings were significantly higher than pre-machine ratings for comfort (F1, 142 = 30.71, P = 0.001), synchrony (F1, 142 = 4.45, P = 0.036), and speed (F1, 142 = 17.65, P = 0.001), but the post-machine initiative was rated slightly lower (in favor of the partner's initiative rather than the subject's own initiative; F1, 142 = 2.47, P = 0.118).
Effects of the machine's voice
Next, we examined the effects of machine voice type on the behavioral and EEG rhythms in human subjects participating in human–machine alternating speech tasks. In the analyses of speech rhythms (i.e., the correlations and differences of the durations and intervals), we conducted two-factor repeated-measures ANOVA for voice types and paces. For correlation of voices, the ANOVA showed a main effect for voice type (correlation of duration, F4, 222 = 3.77, P = 0.005; correlation of interval, F4, 222 = 13.00, P = 0.001) but no effect for pace (correlation of duration, F1, 222 = 0.25, P = 0.62; correlation of interval, F1, 222 = 0.29, P = 0.59). In contrast, for the difference in the voices, the ANOVA showed no effect of voice (difference of duration, F4, 222 = 2.23, P = 0.066; difference of interval, F4, 222 = 2.35, P = 0.055) or pace (difference of duration, F1, 222 = 0.01, P = 0.920; difference of interval, F1, 222 = 0.04, P = 0.841). Post-hoc analyses (Bonferroni correction) showed significant differences between voice types (Fig. 6A; P < 0.05; correlation of duration, male and female < electronic; difference of duration, self < male; correlation of interval, male < other voices; difference of interval, female and self < electronic and male).
The EEG analyses revealed theta/alpha activities that showed a main effect of voice but no effect of pace in some of the electrodes using a two-factor repeated-measures ANOVA. Post-hoc analyses revealed that the activities were higher with female and the partner's voices than for the other voices (P < 0.05; Bonferroni correction). Moreover, only temporal electrodes registered significant correlations between the speakers and observers in response to the machine's voices according to a two-factor repeated-measures ANOVA (Fig. 6B; T7 electrode vs. CP2 electrode). On the temporal electrodes, inter-brain correlations were higher for the female and partner's voices than for the electronic and male voices (P < 0.05).
Along with the behavioral and EEG results, the subjective ratings showed significant differences based on voice type: for comfort, electronic < all other voices; for synchrony, electronic < self and female; for initiative, electronic and male < self; and for humanity, electronic < self and partner (P < 0.05; Bonferroni correction).
There were no differences in any of the subjective emotions, speech, or EEG rhythms between the condition with a fixed pace of machine speech and that with a random pace. This finding might have occurred because of the small range used for the random pace (400–600 msec) and its similarity to the fixed pace (500 msec). Indeed, most of the subjects did not notice the difference between the paces used. Although this study revealed no differences between the fixed and random paces of the machine's speech, the possible effect of adaptation in human and machine communications will need to be clarified in future studies.
The current study is the first to describe inter-brain synchronization along with synchronization of speech rhythms between subjects during an alternating speech task. Individual speech rhythms (i.e., duration of speech and intervals between two voices) were more likely to synchronize (high correlations and small differences), which is similar to the spontaneous synchronizations that were observed between humans in previous nonverbal studies2,3,4,13,14,15,16. This phenomenon is specific to human–human communication because the same behavioral synchronizations were not found during the human–machine alternating speech tasks with the machine pronouncing letters at a fixed interval. The behavioral synchronization was also correlated with the theta/alpha oscillatory amplitudes, which were enhanced and synchronized between two subjects in the same bilateral temporal and lateral parietal regions during the human–human tasks. This result suggests that inter-brain synchronization tightly links to speech synchronization between subjects. Indeed, the inter-brain synchronization was enhanced as the speech rhythm coordination developed, with previous human–human and human–machine interactions strengthening subsequent human–human behavioral and brain coordination (comparing pre- and post-machine conditions).
Inter-brain synchronizations are of great importance to our cooperative communication. In previous studies19, inter-brain synchronizations have been reported in nonverbal communication, such as the imitation of hand movements16,17,18 or during participation in cooperative games and actions20,21,22, and in verbal communication between listeners and speakers31. Such imitations, games, and speeches included the context of the subjects' behaviors. The present study indicated that inter-brain synchronizations are also found during simple communication, without the context of the subjects' behavior.
Numerous EEG studies have proposed that theta/alpha activities are involved in high cognitive functions, such as working memory11,32,33,34. For example, theta and alpha oscillations in the temporal and parietal brain regions are modulated during auditory and visual working memory tasks, respectively11. In this study, the theta/alpha modulation in the temporal regions during auditory communication is similarly related to the auditory working memory. Therefore, we propose that social interactions that lead to successful turn-taking require a working memory of others' speech rhythms. Similarly, previous studies have shown that alpha inter-brain synchronizations are also used in tasks that require working memory for others' behavior, such as imitation tasks16 and cooperative tapping tasks14.
In this study, theta/alpha inter-brain synchronization was also observed in the temporal and lateral parietal regions between speakers and observers (i.e., between subjects who participate in and who observe the human–machine tasks). These regions, which are in the vicinity of the temporal parietal junction (TPJ), have been proposed to be related to social cognition, such as understanding others' intentions, emotions, and behavior, including the theory of mind and the mentalizing process33,34,35,36,37,38. The TPJ is associated not only with the observation of human behaviors but also with the observation of machine (i.e., robotic) behavior39,40,41. Moreover, the theta/alpha amplitude modulation has been reported to be related to the coordination of self-behaviors for the observation of others' behaviors in tapping tasks13,14. Thus, there are spatial and frequency overlaps between the neural mechanisms for social interaction and social cognition15. It should be noted, however, that the neural mechanisms for social cognition are enhanced during social interaction because the lateral parietal theta/alpha amplitudes during human–human tasks were higher and their synchronizations were greater than those observed during human–machine tasks.
There is a possibility that inter-brain synchronizations are related to jaw movements causing muscle artifacts in EEG data, especially within the temporal regions42. However, the theta/alpha enhancements in the temporal regions were not observed in human–machine tasks, although the tasks required subjects to pronounce letters in the same manner as in the human–human tasks. Moreover, in the human–human tasks, when one subject pronounced letters, the other subject did not pronounce letters. In other words, the two subjects did not move their jaws simultaneously. Finally, the human–machine tasks did not ask one subject (i.e., the observer) to pronounce letters when another subject (i.e., the speaker) participated in the task. Therefore, we concluded that the inter-brain connectivity was not due to the muscle movement artifacts.
In addition, our findings reveal the importance of the familiarity of the machine's voice (e.g., a Japanese-native voice for Japanese subjects) in human–machine speech coordination. Use of the subjects' voices, partners' voices, or a Japanese-native female voice showed higher correlations and smaller differences in the speech rhythms and higher theta/alpha synchronizations in the EEG than did other types of voices. Consistent with both behavioral and brain synchronizations, the use of familiar voices and human-like machine voices made the subjects feel comfortable and synchronized. Typically, the subjects' speech rhythms were more likely to be synchronized when the voices used were familiar, with positive subjective impressions, which underscores important differences in language and learning43,44.
Interestingly, the effects of the machine's voice on the inter-brain synchronizations between the speakers and observers demonstrated functional dissociations within the temporal-parietal regions. The inter-brain synchronization of the temporal regions is susceptible to the machine's voice; greater synchronization was observed with the use of familiar voices. This finding might be due to the sensitivity of the primary auditory areas in the temporal brain regions to auditory perceptual properties45,46. However, the lateral parietal regions usually showed inter-brain synchronization, regardless of the machine's voice.
The existing relationships between the subjects might also affect the degree of empathy and their coordination of behavioral and brain rhythms, although there were no differences between the genders, ages, and relationships of the subject pairs who were acquaintances or strangers at the time of the study (e.g., more than half of the desynchronized pairs were already acquaintances). Future studies should address how behavioral and brain synchronization are influenced by personal relationships (e.g., mother-infant, husband-wife, boss-subordinate, teacher-student, etc.).
Twenty pairs (N = 40) of healthy right-handed Japanese volunteers (nine female-female, seven male-male, and four female-male pairs; mean age = 21.57 ± 0.84 years, range 18–40 years) with normal or corrected-to-normal visual acuity, normal hearing acuity, and normal motor performance participated in the experiment. Six pairs met for the first time at the experiment, and 14 pairs were acquaintances. All of the subjects provided written informed consent prior to participation in the study. The study protocol was approved by the Ethical Committee of RIKEN (in accordance with the Declaration of Helsinki). One female-female pair was excluded from the analyses because of a lack of subjective ratings. Each subject completed a pretest, during which their pronunciation of the alphabet from “A” to “G” was recorded to program the machine's voice and define the spectrogram of their pronunciation.
Alternating speech tasks
Two types of tasks were performed: human–human tasks and human–machine tasks.
The human–human alternating speech task was performed by two human subjects, who were asked to pronounce the alphabet from “A” to “G” sequentially and alternately; for example, one subject said “A,” and then the other subject said “B”, and so on. Speech rhythms were not instructed. Each alternating speech session was conducted for 70 seconds. The interval between sessions was at least 10 seconds. The average frequency of loops (i.e., the number of times from “A” to “G”) was no fewer than 20 per 70-second session.
In the human–machine alternating speech tasks, one human subject participated in alternating speech with an anthropomorphic robot (PoCoBot; Business Design Company, Japan). The machine used five voices, electronic, male, female, the human subject's partner's voice, and the human subject's own voice, each of which was set at two different paces between pronunciations (e.g., between "A" and "C"): fixed intervals (1000 ms) and random intervals (800 to 1200 ms) .
Each subject completed 14 sessions (Fig. 1B). First, the two human subjects performed the human–human task twice, with each subject starting the alternating speech. Next, each subject participated in 10 human–machine tasks (five types of voices × two voice paces). Lastly, the two human subjects repeated the task twice, again with each subject starting the alternating speech.
After each session, each subject evaluated the subjective qualities of “comfort,” “synchrony,” “speed,” and “initiative” for the human–human and human–machine alternating speech tasks and “humanity” for the human–machine tasks using a 5-point scale.
Each subject completed 14 sessions and underwent EEG recording, voice recording, and video capture during each session. The tasks were conducted in an electronic- and sound-shielded room.
The series of sessions were conducted as follows (Fig. 1B). First, the two human subjects performed the task twice, with each subject starting the alternating speech (human–human). The subjects were seated at a distance of 150 cm and faced each other eye-to-eye by placing their chins in chin rests. During the sessions, the subjects were asked to keep their chins in the chin-rests, lay their hands on the desk, keep their bodies still, and fix their eyes on a cross that was located on a clear partition placed between them (Fig. 1A, left).
Next, each human subject of the pair performed the speech task with an anthropomorphic robot (PoCoBot; Business Design Company, Japan) (human–machine). The machine remained stationary at a distance of 50 cm from the subject during the sessions, although it was able to make motions with its head and hands between tasks (Fig. 1A, right). The machine used five voices, electronic, male, female, the human subject's partner's voice, and the human subject's own voice, each of which was set at two different paces: fixed intervals and random intervals. The electronic voice included embedded machine sounds. The male voice was an English native speaker. The female voice was a Japanese native speaker. The subject's partner's voice and the subject's own voice were recorded during the pretest and filtered to remove background noise. The fixed intervals were set at 1,000 ms between pronunciations. The random intervals were shuffled from 800 to 1,200 ms between pronunciations, with a uniform distribution. Each subject alternately completed 10 human–machine sessions (five voices × two paces) in two batches. One batch included five voices and alternated between the fixed and random paces. The orders of the conditions were counter-balanced across the subjects.
Lastly, the two human subjects repeated the task twice, again with each subject starting the alternating speech (Fig. 1B).
In our EEG and behavioral measurements, the subject who started the alternating speech was asked to press one key at the onset of each session to trigger a signal to the EEG systems. Then, 70 seconds later, another trigger signal was sent to the EEG systems that indicated the offset of each session. The experimenters then opened the door of the shielded room, and the subjects stopped the task and rested. We analyzed the EEG and speech data in the temporal windows between the onset trigger signal and offset trigger signal.
After each session, each subject used a 5-point scale to report their subjective ratings of “comfort” (5 was “very comfortable,” and 1 was “not comfortable”); “synchrony” (5 was “strong feelings,” and 1 was “no feelings” of synchronization with partner); “speed” (5 was “fast,” and 1 was “slow”); “initiative for alternating speech” (5 was “subject had higher initiative,” and 1 was “partner had higher initiative”), and “humanity” of the machine's behavior (5 was “human-like,” and 1 was “machine-like”).
Voice sound recording and analyses
Voice sounds were recorded from a linear stereo PCM recorder using a PCM-D50 (SONY, Japan; sampling 96.00 kHz, 16 bit), which was placed centrally between the two human subjects or one human subject and one machine. Its right and left microphones were directed at each subject. We filtered out the background noise from the right and left recorded voice sounds (from 20,000 to 44,000 Hz). We first calculated the spectrogram of the sound data that were recorded during the pretest with short-time Fourier transforms and then identified the frequency ranges of each letter spoken by each subject. Referencing the frequency ranges of each letter, we applied short-time Fourier transforms for the right and left recorded data during the alternating speech tasks and dissociated the voices and intervals of the two subjects or one subject and one machine (Fig. 2).
The durations of pronunciation of each letter and the intervals between each letter were averaged for each session. For both durations and intervals for each session of each subject pair, we calculated the correlations and differences between the two subjects or one subject and one machine by using Pearson's correlation analyses and ANOVA with post-hoc analysis (Bonferroni correction), respectively.
EEG recordings and analyses
The individual brain activities of the subjects in each pair were independently recorded by two EEG systems with 27 active scalp electrodes embedded in an electro cap (actiCAP) and BrainAmp ExG MR equipment (Brain Products, Germany). The sampling rate was 1,000 Hz. Reference electrodes were placed on the right and left ear lobes. Electrooculography (EOG) was performed with electrodes that were placed above and below the left eye to monitor eye blinks and vertical eye movements and electrodes that were placed 1 cm from the right and left eyes to monitor horizontal eye movements. To reduce or eliminate artifacts, we conducted infomax independent components analysis (ICA) on the EEG data. The ICA components with the most significant correlations with the vertical and horizontal EOG were rejected, and the ICA-corrected data were recalculated by regression of the remaining components47.
To identify the time-frequency amplitudes during the alternating speech tasks, we applied wavelet transforms assuming that Morlet's wavelets have a Gaussian shape in the time domain (SD σt) and the frequency domain (SD σf) around its center frequency (f)48. The TF amplitude E(t, f) for each time point of each session was calculated as the squared norm of the results of the convolution of the original EEG signals s(t) with the complex Morlet's wavelet function w(t, f):
where σf = 1/(2πσt). The wavelet that we used was characterized by a constant ratio (f/σf = 7), with f ranging from 0.5 to 40 Hz in 0.5-Hz steps. The amplitude was calculated by subtracting the baseline data measure in the ITI for each frequency band. The oscillatory amplitude of each frequency was averaged across times from the onset to the offset of each session for each condition.
Inter-brain synchronizations were evaluated by cross-correlation analyses of convolutions between the two subjects' oscillatory amplitudes at each frequency during a 2-second window. The window was moved from the onset to the offset of each session with a 1-second overlap. The calculated cross-correlation coefficients were averaged among all of the windows for each session, and then, the maximum values were defined.
This research was supported by a Grant-in-Aid for Scientific Research on Innovative Areas (22118510 & 21120005). The authors would like to thank Naoko Tsukakoshi and Yuuki Koizumi for their support in the data acquisition and analysis and Hideyuki Takahashi for his support with the robotic design.
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareALike 3.0 Unported License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc-sa/3.0/