Abstract
Music induces people to coordinate with one another. Here, we conduct two experiments to examine the underlying mechanism of the interbrain synchronization (IBS) that is induced by interpersonal coordination when people are exposed to musical beat and meter. In experiment 1, brain signals at the frontal cortex were recorded simultaneously from two participants of a dyad by using functional near-infrared spectroscopy (fNIRS) hyperscanning, while each tapped their fingers to aural feedback from their partner (coordination task) or from themselves (independence task) with and without the musical meter. The results showed enhanced IBS at the left-middle frontal cortex in case of the coordination task with musical beat and meter. The IBS was significantly correlated with the participants performance in terms of coordination. In experiment 2, we further examined the IBS while the participants coordinated their behaviors in various metrical contexts, such as strong and weak meters (i.e., high/low loudness of acoustically accenting beats). The results showed that strong meters elicited higher IBS at the middle frontal cortex than weak meters. These findings reveal that the musical beat and meter can affect brain-to-brain coupling in action coordination between people, and provide insights into the interbrain mechanism underlying the effects of music on cooperation.
Significance Statement
This study reveals enhanced interbrain synchronization approximately at the left-middle frontal cortex, during interpersonal coordination in the presence of musical beat and meter. The behavioral performance of the participants of a dyad during metrical coordination was also predicted. Further, the interbrain synchronization was modulated by the accent of the musical meter, with higher interbrain synchronization in case of strong meters than weak ones. These results suggest that music enhances interbrain synchronization to promote interpersonal coordination.
Introduction
Music moves us to coordinate with others (Keller et al., 2014). When people get together for activities involving music, they show better interpersonal coordination than otherwise across populations (e.g., infants, children, and adults; Kirschner and Tomasello, 2010; Gerry et al., 2012; Lang et al., 2016), population sizes (e.g., pair, group; Wiltermuth and Heath, 2009; Codrons et al., 2014), and types of collaboration (e.g., intentional, unintentional; Gioia, 2006; Demos et al., 2012). However, the brain mechanism underlying the effects of music on interpersonal coordination remains incompletely understood.
Research on brain imaging has mapped distributed brain networks that are engaged in rhythmic interpersonal coordination, including the activation of the auditory cortex, parietal cortex, thalamus, and caudate in synchronized drumming (Kokal et al., 2011), and the activation of the motor cortex as well as the cerebellum in rhythmic coordination (Mayville et al., 2002). These results have been obtained from comparisons between synchronized and unsynchronized conditions, or between coordinated and baseline conditions, rather than through a direct comparison between conditions involving the presence and absence of music. Further, previous studies on neuroimaging have focused on rhythmic interpersonal coordination at the level of an individual’s brain activity. Social interactions, which include interpersonal coordination, have recently been considered as a feedback loop in a multibrain system (Kingsbury and Hong, 2020). That is, interpersonal coordination requires at least two brains to communicate with each other. Therefore, previous findings are not necessarily informative regarding the brain mechanism underlying the effect of music on interpersonal coordination.
Recent research on brain imaging has used the hyperscanning approach to simultaneously measure brain signals from two or more individuals (Montague, 2002) to explore the neural communication between the interactive partners (Dumas et al., 2010; Yun et al., 2012). A growing body of studies on interpersonal coordination based on functional near-infrared spectroscopy (fNIRS)/electroencephalogram (EEG) hyperscanning has revealed interbrain synchronization (IBS) during various forms of interpersonal coordination, such as the cooperative key-press (Cui et al., 2012), joint drawing (Cheng et al., 2022), synchronized movement (Hu et al., 2017), and action imitation (Miyata et al., 2021). Moreover, interpersonal coordination in musical contexts has been found to be accompanied by IBS, including rhythmic finger-tapping (Konvalinka et al., 2014), rhythmic arm swinging (Nozawa et al., 2019), rhythmic group-walking (Ikeda et al., 2017), groups playing drums (Liu et al., 2021), playing guitars in a quartet (Müller et al., 2018), and singing and humming (Osaka et al., 2015; Müller et al., 2019). IBS in the frontal and parietal cortices has been mapped with and without music (Cheng et al., 2019; Feng et al., 2020). It is always positively correlated with behavioral coordination (Mu et al., 2016), where this is interpreted as a correlate of behavioral or cognitive alignment (Kingsbury and Hong, 2020). Taken together, this body of research suggests that IBS plays a key role in interpersonal coordination regardless of whether the subjects are exposed to music.
Nevertheless, the diversity of IBS between conditions with and without music remains unclear. Based on evidence whereby music enhances behavioral coordination, we predicted in past work that there is greater IBS during interpersonal coordination when the subjects are exposed to music than when they are not. Specifically, we focused on the abstract temporal information of music—musical beat and meter—and hierarchically grouped downbeats and upbeats (Cooper and Meyer, 1960). The reason for this focus was that musical beat and meter play an important role in our understanding of the effects of music, especially the temporal coordination of actions. That is, the understanding of music is inseparable from that of the beat and meter, and our interaction with music is the ability to identify and move in time with every beat while accenting downbeats compared with upbeats (Dixon and Cambouropoulos, 2000; Snyder and Krumhansl, 2001).
In this study, we directly compare IBS in interpersonal coordination between metrical and no-metrical contexts, and examine how it changes in various metrical contexts (i.e., different accents/frequencies of occurrence of downbeats and upbeats). Accent and the frequency of occurrence are key regularities needed to perceive musical meter (Palmer and Krumhansl, 1990). The former is one whereby downbeats last longer, are louder and higher in pitch, or are positioned at points of change in a melody. The latter is typically displayed as a march (perceived as one-two-one …) or a waltz (perceived as one-two-three-one …). We hypothesize that there is higher IBS underlying interpersonal coordination in case of exposure to musical meter (vs no musical meter) that is enhanced when highlighting the meter. To test our hypothesis, we conducted two fNIRS hyperscanning experiments. fNIRS technology was used because of its loose constraints on measurements in relatively natural settings (Egetemeir et al., 2011). Brain signals were recorded in an extended area of the frontal cortex (including the motor and premotor areas), based on the location of the IBS in previous studies. In experiment 1, dyads of participants were asked to tap their fingers to auditory feedback from their partner or themselves after listening to meter and no-meter stimuli. We anticipated higher IBS and better behavioral coordination when the participants tapped together after listening to the metrical stimuli than to no-metrical stimuli. In experiment 2, the participants were required to tap their fingers to auditory feedback from their partner after listening to different metrical stimuli, such as meters with different accents (i.e., strong vs weak) and frequencies of occurrence (i.e., duple vs triple). We anticipated enhanced IBS and better behavioral coordination in the participants when they listened to outstanding meter stimuli (e.g., strong meters) because stronger meters are known to elicit higher brain activity and better coordination than weak meters (J.L. Chen et al., 2008).
Experiment 1
Materials and Methods
Participants
Forty graduate and undergraduate female college students (mean age = 21.820 years, range: 18–25 years) were recruited for experiment 1 in randomly matched dyads in exchange for monetary compensation. Only female participants were recruited as gender is known to have an effect in the context of music (Christenson and Peterson, 1988) and IBS (Cheng et al., 2015; M. Chen et al., 2020). Members of a given dyad had not seen or known each other before. They were right-handed, and had normal or corrected-to-normal vision and hearing. They had either not studied music or had studied it for fewer than three years. Ethics approval was obtained from the University Committee on Human Research Protection of Author University.
Stimuli and apparatus
The auditory stimuli consisted of 440-Hz pure tones that lasted for 660 ms. They were created by MuseScore, a free music composition and notation software (https://musescore.org/en). The auditory stimuli consisted of two types of tone sequences. One was a meter tone sequence while the other was a no-meter tone sequence. Each tone sequence lasted for 12 s and was composed of 12 tones at an interval of 500–1000 ms between them. The total duration of the tone sequence for each trial was constant (i.e., 12 s). For the sequence of meter tones, the first tone (vs the second tone) was acoustically accented (+6 dB) to create the pattern of downbeats and upbeats, and this pattern was looped six times in each tone sequence. For the no-meter sequence, the tones were unaccented and had equal intensity (i.e., 40 dB above the threshold of individual sensation, collected before the experiment task). The tone sequences were varied across trials to avoid the practice effect, but were identical between the participants in a dyad.
The auditory stimuli were given to participants through two pairs of headphones (Philips) that were controlled by the same server with two 19-inch computer monitors. The monitors were placed on the middle of a table (110 × 80 cm) and were equipped with keyboards. Two participants of the dyad were seated at the table across from each other, each with her own computer monitor and keyboard. They each wore a pair of headphones to receive the auditory stimuli. They were also separated by a piece of white cardboard (110 × 80 cm) to block any visual information that might be used for communication.
Task and procedure
The participants of the dyads were asked to complete the finger-tapping task, that is, to tap their fingers to the auditory stimulus that they had just heard. Moreover, they were required to finger-tap along with their partner (the coordination task) after listening to the meter and the no-meter stimuli. During the coordination task, each participant received the auditory feedback of her partner’s response, and was asked to try her best to response synchronously with the partner. As a control, the participants were also instructed to tap their fingers with the computer (the independence task) after listening to the meter or no-meter stimuli. For the independence task, both participants received the auditory feedback of their own responses, and were asked to response synchronously with the auditory stimulus as precisely as possible.
The participants were first instructed on the experimental tasks and given several practice trials. They were then given a 20-s resting state, during which they were required to remain as motionless as possible and relax. The experimental tasks consisted of four blocks corresponding to the four conditions in this study (e.g., the participant heard the tap of her partner and was required to tap along with her), with a 30-s rest between blocks. The order of the four blocks was counterbalanced. Before each experimental task, one piece of instruction was presented on the display for 3 s to remind the participants of the task. Each block (i.e., performing the coordination/independence task with the meter/no-meter stimuli) consisted of 15 trials, and thus 60 trials were held in the entire experiment. For each trial, the participants first heard the meter/no-meter stimuli (i.e., 12 s), followed by a sound (262 Hz, 1000 ms) that served as a cue to remind them to start tapping their fingers. They were instructed to reproduce the stimuli that they had heard before by tapping their right index finger on the keyboard (participant #1: “f”; participant #2: “j”). During the experiment, the participants of each dyad were not allowed to communicate with each other through language or movement. We used the E-Prime software (Psychology Software Tools), which can provide the input and output of stimuli with millisecond-level precision. Specifically, the participants’ taps were collected through the response box (i.e., “f” and “j” on the keyboard device), and the feedback regarding the taps (i.e., the drip sound) were given to participants through SoundOut objects (that support audio file outputs) in E-prime.
Behavioral analysis
All trials were used in the behavioral and fNIRS analysis. To quantify behavioral coordination, we calculated the timing data of each participant’s taps during the coordination task in meter and no-meter conditions. We focused on the coordination task given our hypothesis (i.e., there is better behavioral coordination in the meter condition than in the no-meter condition). The onset time of the participants’ taps (i.e., the timestamp at which the participants began to tap) were extracted and then computed for the interpersonal time lag using the follow formula (Mu et al., 2016):
To further estimate behavioral coordination, the trend and fluctuation of interpersonal time lags across taps were analyzed. Specifically, the interpersonal time lags were entered into detrended fluctuation analysis to estimate the Hurst exponent H (Kantelhardt et al., 2001). The value of H ranges from zero to one. If H is 0.5, the correlations are completely absent, and if H > 0.5, this implies a positive correlation. On the contrary, if H < 0.5, this implies a negative correlation. The interpersonal time lags were also transformed into the log scale because they followed a non-normal distribution, and were then regressed into the mixed linear regression model with taps and individual differences as the fixed factors, and the dyad as the random effect. The significance of the model was assessed using the likelihood ratio test.
fNIRS data recordings
fNIRS data were simultaneously recorded from both members of each dyad by using the ETG-7100 optical topography system (Hitachi Medical Corporation). The values of oxyHb and deoxyHb were obtained at a sampling rate of 10 Hz. One optode probe patch (3 × 5 setup) was placed on the head of each participant, and contained eight emitters and seven detectors that formed 22 measurement channels with a 3-cm separation between optodes. The middle optode of the second row of probes of the patch was placed at FCz (Fig. 1B), following the international 10–20 system (Okamoto et al., 2004). The correspondence between the channels of fNIRS and the measurement points on the cerebral cortex was determined by using the virtual registration method (Tsuzuki et al., 2007), which has been validated by a multisubject study on anatomic craniocerebral correlation. A 3D digitizer was used to measure the positions at which the channels existed on the head (Xiao et al., 2017), and the NIRS-SPM software for MATLAB was used to validate the standard brain model and data from the 3D digitizer (Ye et al., 2009). The possible MNI coordinates were thus obtained for each channel. Finally, the brain regions of the channels were determined based on the Automated Anatomical Labeling (AAL) atlas (Tzourio-Mazoyer et al., 2002) and checked through data from Neurosynth (a platform for large-scale and automated synthesis of fMRI data).
fNIRS data analysis
The raw data for each participant were preprocessed by using the correlation-based signal improvement method (Cui et al., 2010), which is based on the observation that the values of oxyHb and deoxyHb are usually negatively correlated but the motion of the head induces a positive correlation between them. The preprocessed signals were used to calculate the IBS for each dyad through the wavelet transform coherence (WTC) method, which can assess the cross-correlation between brain signals on the time–frequency plane (Grinsted et al., 2004).
To determine the frequency band of interest, the linear mixed model was used to compare IBS between experimental conditions across the full range of frequency (0.015–0.1 Hz). The frequencies above 0.1 Hz were removed as they might have been correlated with non-neuronal components, such as Mayer waves (around 0.1 Hz), respiration (0.2–0.6 Hz), and cardiac pulsation (above 0.7 Hz). The frequencies below 0.015 Hz were also removed because IBS has often been observed only at frequencies above 0.015 in fNIRS hyperscanning studies (Dai et al., 2018; Wang et al., 2019; Nguyen et al., 2021). A cluster-based permutation test was used, in the same manner as in a recent fNIRS hyperscanning study by us (Pan et al., 2022; Zhu et al., 2022). This method is a nonparametric statistical test that allows for multiple comparisons of multichannel (i.e., 22 channels) and multifrequency (i.e., 30 frequency bins) data, rather than familiar methods of correction (e.g., Bonferroni, FDR) based on a parametric statistical framework. It provides a straightforward way to solve the problem of multiple comparisons (Maris and Oostenveld, 2007). Nonparametric statistical tests offer complete freedom to use any test statistic one considers appropriate. This allowed us to solve the problem of multiple comparisons in a simple way.
The cluster-based permutation test consisted of six steps as follows. First, the IBS for each frequency and each channel was calculated by using the WTC for each experimental condition. Second, the IBS between meter and no-meter conditions was compared by using the frequency-by-frequency and channel-by-channel mixed linear model. Such model included the fixed effect of condition (meter vs no-meter) and individual differences (the IBS in the independence task), and the random effect of the dyad. The meter and no-meter conditions were compared directly (instead of relying on the task vs rest comparison) because of our hypothesis (i.e., there is enhanced IBS in interpersonal coordination in the presence of the meter relative to its absence). Moreover, it has been reported that an active task condition can offer a better baseline than the rest condition (Stark and Squire, 2001; Reindl et al., 2018). Third, the channels and frequencies that exhibited significant effect because of the condition (i.e., the meter condition > the no-meter condition, p < 0.05) were identified. Fourth, clusters with neighboring frequency bins (N ≥ 2) were formed. The statistics of each cluster were computed by averaging all the F values. Fifth, the above steps were repeated 1000 times by permuting the data. The permutation was constructed by randomly pairing the data for a participant from one dyad with that of another from a different dyad. Finally, the significant levels (p < 0.05) were calculated by comparing the cluster statistic obtained from the dyads over 1000 permutations of randomized dyads.
Results
Behavioral coordination
The results of the linear mixed model showed that during the coordination task, the mean interpersonal time lag in the meter condition (mean ± deviation: 0.379 ± 0.071) was significantly shorter than in the no-meter condition [0.408 ± 0.060, F = 5.086, p = 0.037, β = −0.231, SE = 0.094, 95% confidence interval (CI) = −0.399 to −0.028; Fig. 2A]. These findings indicate better behavioral coordination of the dyads when exposed to the beat and the meter.
The detrended fluctuation analysis showed an insignificant difference between the meter (0.200 ± 0.091) and no-meter conditions (0.175 ± 0.062, F = 1.294, p = 0.270, β = 0.104, SE = 0.091, 95% CI = −0.076–0.282; Fig. 2B). Here, the values of H for each dyad were all <0.5, implying anticorrelations between interpersonal time lags across task trials, that is, a long interpersonal time lag would probably be followed by a short interpersonal time lag, and a short interpersonal time lag would probably be followed by a long interpersonal time lag.
In contrast, the results of linear regression analysis revealed that the interpersonal time lag significantly decreased across taps in the meter condition (F = 4.168, p = 0.041, β = −0.001, SE = 0.0003, 95% CI = −0.001 to −0.00003), and changed insignificantly in the no-meter condition (F = 1.116, p = 0.291, β = −0.0003, SE = 0.0003, 95% CI = −0.0003–0.001; Fig. 2C). These results indicate an improvement of coordination between interactors in the metrical context.
Enhanced IBS in interpersonal coordination with meter
The cluster-based permutation test identified one channel–frequency cluster that reached significance: the cluster in channel 6 in the range of frequencies of 0.026–0.030 Hz. These frequency bands roughly corresponded to the duration of one task trial in experiment 1 (i.e., around 33 s based on the participants’ pragmatic response). In this range of frequencies (i.e., 0.026–0.030 Hz), the IBS in channel 6 was significantly greater in the meter condition (0.388 ± 0.111) than the no-meter condition (0.293 ± 0.075; Fig. 3A). According to the AAL atlas and data of Neurosynth, channel 6 was approximately located in the left-middle frontal cortex (MFC).
IBS is associated with behavioral coordination
The correlation analysis revealed that the IBS at channel 6 in the coordination task was negatively correlated with the mean interpersonal time lag in the meter condition (r = −0.462, p = 0.041; Fig. 4). This correlation was not significant in the no-meter condition (r = −0.053, p = 0.825). These results indicate that the enhanced IBS at the middle frontal cortex was positively associated with behavioral coordination in case of exposure to the musical meter.
Experiment 2
Materials and Methods
The methods in experiment 2 were the same as in experiment 1, with the exceptions below.
Participants
A new group of 32 right-handed female students (mean age = 20.640 years, range: 18–28 years), comprising 16 random dyads, participated in experiment 2 in exchange for monetary compensation. The members of each dyad had not seen or known each other before.
Stimuli and apparatus
The stimuli were four kinds of meters generated by manipulating the accent (i.e., strong vs weak) and the frequency of occurrence (i.e., duple vs triple). For each tone sequence, there were 12 40-Hz pure tones at an interval of ∼500 ms between them. The total duration of the tone sequence for each trial was constant (i.e., 6 s). For the duple meters, every second tone was reduced in intensity by 10 or 2 dB, corresponding to strong and weak meters, respectively. For the triple meters, the intensities of every second and third tones were reduced by 10 dB (strong meters) or 2 dB (weak meters). These difference in loudness (i.e., 10 vs 2 dB) have been found in past work to induce significantly different responses (Ozertem and Erdogmus, 2008).
Task and procedure
Before the experimental task, the participants were asked to rest for 2 min to obtain more stable resting-state data from them (Lu et al., 2010). They completed the coordination task in the experiment, consisting of four blocks of eight trials yielded by crossing the accent (i.e., strong vs weak) and frequency of occurrence (i.e., duple vs triple). The order of the four blocks was counterbalanced.
Behavioral analysis
Only the mean interpersonal time lag was calculated to quantify the behavioral synchronization of the dyads in experiment 2, as insignificant values for the other behavioral measures (e.g., the value of H, the probability density) were recorded in the coordination task in experiment 1. Note that all trials were entered into the behavioral and fNIRS analysis.
NIRS data recordings
To obtain the brain signals of a more extended frontal cortex, two 3 × 5 patches were used in experiment 2. The first probe patch was set as in experiment 1, and generated 22 measurement channels (channels 1–22), while the second one covered the prefrontal area to form another 22 measurement channels (channels 23–44). The bottom row of the second patch was placed on top of the participant’ eyebrows, with the middle optrode at Fpz (Fig. 1C).
fNIRS data analysis
The same procedures (e.g., data preprocessing, the calculation of IBS, and the cluster-based permutation test) were used in experiment 2 as in experiment 1. But there was one exception: in the second step of the cluster-based permutation test, IBS between experimental conditions (i.e., strong vs weak) was compared by using the mixed linear model, in which only included the fixed effect of condition and the random effect of the dyad.
Results
Behavioral coordination
To explore the effects of the frequency of occurrence and accent on behavioral coordination, the mixed linear regression model was used. The results showed a significant main effect of the accent (F = 6.954, p = 0.011, β = −0.199, SE = 0.076, 95% CI = −0.347 to −0.051), with a shorter mean interpersonal time lag in case of strong meters (0.273 ± 0.101) than weak meters (0.305 ± 0.088; Fig. 5). The main effect of the frequency of occurrence, and the interaction between the frequency of occurrence and the accent were insignificant (Fs < 0.370, ps > 0.546). To clarify how accent affected behavioral coordination, the interpersonal time lags of the accented tones were extracted and compared between the strong (i.e., +10 dB) and weak (+2 dB) tones by using the mixed linear regression model. The results revealed significantly shorter interpersonal time lags for strong tones (0.280 ± 0.093) than for weak tones (0.306 ± 0.082; F = 4.851, p = 0.044, β = −0.149, SE = 0.068, 95% CI = −0.286 to −0.013). These results suggest that accents promoted behavioral coordination during the coordination task in metrical contexts, with better behavioral synchronization in case of strong meters and tones.
Increased IBS because of accent
The IBS was compared between the conditions (strong vs weak) by using the paired t test because we observed a significant effect of the accent on behavioral coordination. The cluster-based permutation test revealed two significant channel–frequency clusters: cluster 1 (frequencies 0.060–0.065 Hz, channel 5) and cluster 2 (frequencies 0.060–0.065 Hz, channel 10). The significant frequency bands were also in line with the duration of a task trial (i.e., around 15 s in experiment 2). For the significant frequencies (i.e., 0.060–0.065 Hz), the IBS at channels 5 and 10 was higher in case of a strong meter (channel 5: 0.448 ± 0.132; channel 10: 0.451 ± 0.082) than a weak meter (channel 5: 0.341 ± 0.092; channel 10: 0.345 ± 0.105; Fig. 3F). Both channels 5 and 10 were roughly located at the left-middle frontal cortex according to the results of fNIRS localization. Because these two channels belonged to the same region of the brain, they were averaged into one IBS. This IBS was negatively correlated with the mean interpersonal time lag in case of a strong meter (r = −0.569, p = 0.022; Fig. 6). However, the corresponding correlation was not significant in case of a weak meter (r = 0.256, p = 0.338).
Discussion
This study investigated the functional role of IBS, in the presence of musical meter, on interpersonal coordination through two fNIRS hyperscanning experiments. Experiment 1 revealed better behavioral coordination when the subjects were exposed to musical meter, as indexed by shorter mean interpersonal time lags of taps in the meter condition than the no-meter condition. Over time, the coordination of dyads improved when they were exposed to musical meter. These findings suggest that people dynamically adapt with others during interpersonal coordination. When musical meter is involved, their mutual adaption begins to be more effective and this leads to good behavioral coordination. Experiment 1 also revealed greater IBS at the left middle frontal cortex (MFC) when the dyads coordinated in a metrical context than in a no-metrical context. The MFC-IBS was positively correlated with behavioral coordination in the metrical context. Our fNIRS results are consistent with previous findings, in which the MFC-IBS was observed during scenarios involving social interaction, including decision-making (Zhang et al., 2017; Xie et al., 2022), cooperation (Hu et al., 2017; Feng et al., 2020), and communication (Balconi et al., 2021). In particular in various cooperative activities, pairs of participants showed enhanced MFC-IBS in case of specific populations and tasks, such as professional athletes (Li et al., 2020), expert teachers (Sun et al., 2020), and less creative individuals (Xue et al., 2018), and providing positive feedback (Lu et al., 2019a) and demanding divergent thinking (Lu et al., 2019b). Moreover, such enhanced MFC-IBS has been reported to be correlated with behavioral cooperation (Lu and Hao, 2019).
Why does music increase our coordination with others? Several theoretical frameworks have been proposed to explain this phenomenon. One possible explanation is that music and interpersonal coordination share neural resources (e.g., the premotor and supplementary motor areas, which belong to the mirror system) at the level of predicting stimuli (Novembre et al., 2012) and information integration (Loehr et al., 2013). Exposure to music can thus facilitate the general processing in case of interpersonal coordination (Keller et al., 2014). Accordingly, the IBS should be located in the area of the mirror system, but this is not supported by the findings of this study. Another explanation emphasizes that music regulates interpersonal coordination through endogenic rhythm (Phillips-Silver et al., 2010). Musical beat and meter are like the natural rhythm of our bodies, like breathing and heartbeat, and relate naturally to motion. A direct causal link has been reported between endogenous rhythms and interpersonal synchrony in a music performance task (Zamm et al., 2016). However, the IBS in our study might be uncorrelated with the endogenic rhythm (e.g., breathing and heartbeats) because of the auditory stimuli used here. Our fNIRS findings can be understood within the view that music functions as a kind of a “social glue” that brings people together (Demos et al., 2012). The MFC-IBS has been suggested to be related to our understanding of others’ mental states and behaviors. For example, the perspective-taking scores of the participants were positively correlated with the MFC-IBS during cooperation (Sun et al., 2020). Another study found a negative correlation between the MFC-IBS and the participants’ agreeableness-related trait when they made defection decisions (Zhang et al., 2021). Thus, the increased MFC-IBS here may be related to the process whereby participants pay more attention to their partners’ actions, and make a greater effort to understand and coordinate with them in musical contexts. Our fNIRS results provide a fundamental neural mechanism for the promotional effect of music on interpersonal coordination.
Notably, experiment 2 showed that the MFC-IBS and behavioral coordination were enhanced by strong meters relative to weak meters. These results are in line with previous findings, where greater active brain activity and accurate behavioral performance were observed when processing strong meters over weak meters (Patel et al., 2005; Kung et al., 2013). These results can be explained by signal detection theory, in which the intensity of the stimulus plays an important role in information processing (Green and Swets, 1966). In case of competition between sensory signals for limited channel capacity, the intensity of each stimulus becomes paramount and the accented stimulus is a priority (Macmillan and Creelman, 2004). Indeed, strong meters are known to more easily attract the attention of listeners relative to weak meters (Kantardjieff and Jones, 1997). Compared with soft music, music with accented beats and meters can facilitate greater dissociation from the internal sensations of fatigue (Hutchinson and Sherman, 2014). Thus, our results extend research on the benefits of accented stimuli for brain-to-brain coupling during interpersonal coordination in the presence of musical meter.
With regard to the frequency of occurrence of meters, the results in experiment 2 indicated an insignificant difference in IBS and behavioral coordination between duple and triple meters. This appears to be inconsistent with previous studies, in which the frequency of occurrence of meters was found to influence human cognition and behavior (Grahn and Brett, 2007; Grahn and Rowe, 2009). For instance, duple meters are easier to coordinate as they are “more natural” than triple meters as the former are similar to the endogenous rhythms of humans (i.e., pulse, heartbeat). Previous EEG studies have reported that compared with duple meters, triple meters awakened a larger amplitude of P300 (Jongsma et al., 2004) that was associated with a wider range of sensorimotor and frontoparietal areas (Fujioka et al., 2015). This suggests that triple meters might require more processing capacities. Although it is difficult to interpret a null finding, it should be noted that the bias between duple and triple meters can be learned, and is not universal across individuals (Hannon and Trainor, 2007; Trainor and Hannon, 2013). Future studies should examine the brain-to-brain coupling in the presence of duple and triple meters during interpersonal coordination with music while controlling for prior knowledge and individual differences.
In this study, we used a beat reproduction task in which the participants were asked to reproduce a sequence of the stimuli after having heard it (Grahn and Brett, 2007). This reproduction task is different from the beat tapping task (e.g., a finger-tapping task in synchrony with a given beat) because the former task also requires a memory component (Steinbrink et al., 2019). The significant IBS reported here might have thus been obtained because of the difference between the memory required for metered/accented tones and that for no-metered tones. However, a significant MFC-IBS was not observed when comparing the meter and no-meter conditions on the independence task by using the paired t test. In addition, a previous fMRI study showed that the temporal lobe was correlated with the capacity of the auditory working memory during a rhythm reproduction task (Grahn and Schuit, 2012). Thus, the enhanced IBS reported in this study seems to be independent of the memory component. To confirm our findings, future work should explore the significant IBS reported here through the beat tapping task.
Several limitations of this work should be addressed in future research. First, the experiments reported here involved only female participants. As mentioned before, musical meters are not perceived in the same way by all populations, and a range of interindividual differences are generally obtained (Schaefer and Overy, 2015). There is also evidence that the IBS during interpersonal coordination differs in males and females, with greater IBS in female-female pairs than male-male pairs (Mu et al., 2016), and stronger IBS between male-female pairs than same gender pairs (i.e., male-female pairs, female-female pairs; Cheng et al., 2015). It would be interesting to explore the IBS in dyads with partners with opposite gender (i.e., male-female pairs) during interpersonal coordination with meters in the future study. Second, the stimuli of meters used here were simple, and did not contain temporal information (i.e., pitch, volume, and timbre) as such meters can move participants to internally feel the meter (Grahn and Brett, 2007). In everyday life, however, we also experience compound meters that have time signatures indicating the number of beats as a multiple of three. Previous neuroimaging studies have shown that simple and compound meters elicit different levels of brain activity. For example, greater activity in the putamen was identified for simple meters than compound meters and the absence of meter (Teki et al., 2011). Future work should examine the IBS during interpersonal coordination with compound meters. Finally, the IBS reported here was found to be approximately located at the MFC in different channels in our two fNIRS hyperscanning experiments (i.e., channel 6 in experiment 1, and channels 5 and 10 in experiment 2). Although fNIRS is an ideal choice for neuroscience research in a natural environment, it has a relatively poor spatial resolution (∼1–3 cm; Boas et al., 2004). Future work should investigate brain-to-brain coupling during interpersonal metrical coordination by using other neural technologies with a high spatial resolution, such as fMRI hyperscanning.
In conclusion, this study provided direct evidence for enhanced IBS during interpersonal coordination in the presence of the musical meter, where this provides a potential neural marker of interpersonal metrical coordination. We have argued that the enhanced IBS at the middle frontal cortex reflects greater attention to and understanding of the partner’s actions. This extends our knowledge of the brain mechanism underlying the effects of music on interpersonal coordination.
Acknowledgments
Acknowledgements: We thank Yujiao Zhu and Tian Meng for their contributions to data collection and Bei Song and Yi Zhu for their insightful suggestions.
Footnotes
The authors declare no competing financial interests.
This work was supported by the National Natural Science Foundation of China Grant 31872783 (to Y.H.) and 71942001 (to Y.H.), the Science and Technology Innovation 2030- Major Project Grant 2021ZD0200535 (to Y.H.), and the Jiangsu Province Social Science Fund Grant 20JYD004 (to M.Z.).
This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.