Research Article | Open Access
Hui-Min Wang, Sheng-Chieh Huang, "Musical Rhythms Affect Heart Rate Variability: Algorithm and Models", Advances in Electrical Engineering, vol. 2014, Article ID 851796, 14 pages, 2014. https://doi.org/10.1155/2014/851796
Musical Rhythms Affect Heart Rate Variability: Algorithm and Models
There were a lot of psychological music experiments and models but there were few psychological rhythm experiments and models. There were a lot of physiological music experiments but there were few physiological music models. There were few physiological rhythm experiments but there was no physiological rhythm model. We proposed a physiological rhythm model to fill this gap. Twenty-two participants, 4 drum loops as stimuli, and electrocardiogram (ECG) were employed in this work. We designed an algorithm to map tempo, complexity, and energy into two heart rate variability (HRV) measures, the standard deviation of normal-to-normal heartbeats (SDNN) and the ratio of low- and high-frequency powers (LF/HF); these two measures form the physiological valence/arousal plane. There were four major findings. Initially, simple and loud rhythms enhanced arousal. Secondly, the removal of fast and loud rhythms decreased arousal. Thirdly, fast rhythms increased valence. Finally, the removal of fast and quiet rhythms increased valence. Our work extended the psychological model to the physiological model and deepened the musical model into the rhythmic model. Moreover, this model could be the rules of automatic music generating systems.
The relation of music to emotion has been studied for decades and the literature is fruitful . There exist a lot of psychological models between music and emotion , but the physiological models between music and emotion are limited . One of the physiological actions, heart rate variability (HRV), which is controlled by the autonomic nervous system (ANS), is tightly connected with emotions . Previously, we had analyzed the relationship between musical rhythms and HRV  and built two heuristic models [6, 7]. In this paper, a systematic algorithm is proposed to construct new models.
Musical emotions change with psychophysiological measures and musical features , whilst three basic questions are highlighted : how do musical features evoke emotions; how do actions involved in musical emotions progress; and which actions and brain processes are involved in musical emotions. Basically, people feel what music expresses but need not be always; in a simple case, only 61% of 45 participants felt what they perceived . More particularly, a stronger correlation is suggested by a recently developed theory that the aesthetic awe accompanies by being moved (cognitive), emotions (psychological), and thrills (physiological) in the same time . The three associated levels of musical response should be analyzed individually and can be discussed together. In the research about music psychology, sometimes the perceived and felt emotions are examined separately; sometimes the music clips are just labeled artificially; whether they are perceived or felt had not been mentioned yet . The mappings between music space and emotion space [13, 14] employ the following synonyms: music mood detection , music emotion measurement , characterization , recognition [16, 17], classification [12, 18], predicting , or modeling ; the review articles demonstrate the fruits of experts’ interests [16–20].
In a valuable model, four elementary properties are goodness-of-fit  (generalization ), simplicity [21, 22], predictability [21, 22], and the relation to existing theories . Beyond cognitive reactions, the prediction of emotional reactions is also important in music psychology [22, 23]. In addition to an existing review , a survey of the models [25–45] for emotional responses is given in this paper. The acoustic features, emotion spaces, and methods to generate these models are listed in Table 1 with four major observations. Initially, regression analysis [30, 31, 36, 38–40, 42, 43, 45] is widely employed, while some soft computing methods (fuzzy ; support vector machine [34, 37]; neural network ; K-nearest neighborhood ; Gaussian mixture model ) are also used. Secondly, beyond the enviable valence and arousal, tension [29, 32, 36, 40, 44] is another common dimension too. Thirdly, the felt emotion [28, 35] is infrequently examined compared to the perceived emotion. Finally, most works employ a wide range of acoustic features, while the interest in a single feature moves from pitch [40, 44] to timbre  and rhythm . For the emotion space, Russell  proposed the valence/arousal space and Thayer  reduced it to four labels; the following synonyms are commonly used: valence as pleasantness ; arousal as activity [36, 45]; tension [29, 32, 36, 40, 44] as interest , expectancy , strength , potency , and resonance .
|Emotion space: V: valence; A: arousal; T: tension; D11: 11 dimensions.|
Method: GMM: Gaussian mixture model; C5: classifiers 5; KNN: K-nearest neighborhood; C4: classifiers 4; MPW: moving perceptual window; ANOVA: analysis of variance; CRF: conditional random field; SVM: support vector machine; NN: neural network.
Whether perceived or felt emotion, the three commonly used methods, including valence/arousal dimensions, lists of basic emotions, and diverse emotion inventories, are not complete in the musical emotion study . To have a more comprehensive overview , the underlying mechanisms of the central nervous system (CNS) [46, 47] and the ANS [48–50] should be also deliberated. In the most essential sense, the acoustic properties can be perceived by the nervous system and evoke psychophysiological responses. For example, fast and loud voices usually cause emotional arousal and increased respiratory and heart rates through the auditory and limbic systems [46, 47]. Some related physiological reactions coincided with the emotion expressed in music  and rated valence/arousal level . Therefore a valence/arousal model  of musical emotion had been built based on the associated physiological measures, including electromyogram, electrocardiogram, skin conductivity, and respiration changes. Extended linear discriminant analysis (pLDA) was employed in the classification of musical emotions.
Since emotion correlates with physiological responses [48–50], too, the models of musical emotions were developed by integrating the acoustic features and physiological measures [51–53]. Despite combining with physiological measures, these models [51–53] tend to be similar to the psychological models mentioned above [25–45] in spite of the same perceived or felt emotion space. In addition, using acoustic features to model  the physiological responses seems more radical if the underlying mechanisms [46, 47] are considered. The model  employs 11 musical characteristics and the conclusion is that rhythmic features are the major factors of the physiological responses (respiration, skin conductance, and heart rate) to music. It also points out a limitation that some acoustic features would correlate with each other. This makes it difficult to discriminate their relative contributions to the detected relationships.
The rest of this paper is organized as follows. In Section 2, the rhythmic features for modeling and the HRV features for the physiological emotion space are introduced. In Section 3, the experiment and developed algorithm are presented. In Section 4, the models of the relationships between the rhythmic and the HRV features are illustrated in figures, and the equations and tables of statistics are provided. In Section 5, how tempo, complexity, and energy work in the psychological experiments, psychological models, physiological experiments, and physiological models are reviewed and compared. Finally, in Section 6, we summarize the contributions made in this study and suggest the directions for the further research.
2.1. Why Are Physiological Models Necessary?
To study the relation of musical features to emotion, most psychological models focus on the perceived emotion, as Table 1 demonstrated. However, the perceived emotion need not be equal to the felt emotion [10, 54], and the felt emotion may not have the related physiological responses neither . If the music clips are selected by the participants themselves, the perceived emotion has no statistical significance with the felt emotion in the ranks. Otherwise the differences are statistically significant . Moreover, some patterns of the physiological responses appear while there is no related self-report . Since the perceived emotion, felt emotion, and physiological actions play different roles, a physiological model is necessary beyond the psychological models.
2.2. Valence/Arousal Model
Although four dimensions are necessary for emotion spaces , it is difficult to realize a four-dimensional model. Three dimensions are also employed in the psychological models [27, 36, 38, 43]. Some work  announced that three dimensions could be reduced into two, without significant loss of goodness-of-fit; it also shows that the dimensional model is better than the discrete (categorical) model in resolution. Thus Russell’s two-dimensional valence/arousal model  states its merit, within the ebb and flow of relevant woks, in the discipline of psychology.
2.3. A Physiological Valence/Arousal Model
SDNN and LF/HF, two measures of HRV , are employed as two dimensions in our physiological valence/arousal model. SDNN is the standard deviation of normal-to-normal heartbeat intervals in time domain and LF/HF is the ratio of low- and high-frequency powers after the fast Fourier transform (FFT); SDNN presents the variation of the circulatory system and LF/HF presents the balance of the sympathetic and parasympathetic nervous systems . HRV is highly correlated with emotion [60–64] and some evidence [65–70] reveals that SDNN is a good indicator of valence in the physiological perspective. In general cases, the normal subjects’ SDNNs are higher than the depression subjects’ [65–67]. For the normal subjects’ case, the SDNN levels of subjects with positive mood are higher than the negative mood . For the depression subjects’ case, the SDNN levels of low-depression patients’ are higher than the high-depression patients’ [69, 70]. All of these findings reach a statistical significance level. Hence SDNN could be a proper indicator as the physiological valence. In addition, increased LF/HF values denote that the sympathetic nerve activity tended to be strong; thus LF/HF could be used as a physiological indicator of arousal levels .
2.4. Musical Features Employed in the Model
The three most important features of music are rhythm, pitch, and timbre [71–76]. Although pitch  and timbre  have been employed to recognize emotion in music, the role of rhythm seems more rudimentary. In fact, the rhythmic features  had been used in modeling the emotional responses and acquired a reasonable performance.
An all-encompassing representation of the rhythm may be the temporal organization of sound, tightly connected with meter, which refers to the periodic composition in music . Musical rhythm may have its roots in the motor rhythms controlling heart rate, breathing, and locomotion , dominated by brain stem the ancient structure ; fast, loud sounds produce an increased activation of the brain  and some evidence suggests that the musical rhythm can regulate emotion and motivational states .
Rhythm is less tractable than pitch and timbre in localizing its specific neural substances . However, many cultures emphasize the rhythm, with two others playing a less crucial role . To study the neural basis of rhythm, brain imaging, psychophysical, and computational approaches have been employed . Beat and meter induction are the fundamental elements of cognitive mechanisms , while the representations of metric structure and neural markers of expectation of the beat have been found in both electroencephalography (EEG) and magnetoencephalography (MEG) . Beat perception is innate; newborn infants expect downbeats (onsets of rhythmic cycles), even unmarked by stress or the spectral features . Infants also engage in the rhythmic movement to rhythmically regular sounds and the faster movement tempo is associated with the faster auditory tempo . Although the beat perception is innate, the ability to detect rhythmic changes is more developed in adults than infants  and the trained musicians than the untrained individuals .
There were 22 healthy subjects, 15 males and 7 females, engaged in the experiment. The average age was 23. None of them had been professionally trained in music.
3.1.2. Musical Stimuli
The ECG signal was captured by a 3-channel portable device (MSI E3-80, FDA 510(k) K071085) at 500 Hz sampling rate from the chest surface of the body. Only the channel-1 data was taken to be analyzed.
All experiments were carried out in moderate temperature, humidity, and light with subjects sitting and wearing headphones (eyes closed) in a quiet room. Each subject accepted 4 rounds of experiment in different days. Each round took 15 minutes, separated as 5 stages, as Figure 2 illustrated. Stage 1 had 5 minutes to let the subjects calm down. Stage 2 had 3 minutes of rest, as the baseline for the responses of the drum loops. Stage 3 had 2 minutes, as the baseline for the responses after the drum loops. Stage 4 had 3 minutes of stimulus of some drum loop. Stage 5 had 2 minutes of rest. The ECG signal from stage 2 to stage 5 was recorded and separated as epoch 1 (E1) to epoch 4 (E4). Comparison 1 (C1) was the difference of E1 and E3 and comparison 2 (C2) was the difference of E2 and E4.
3.2. Signal Processing
3.2.1. Musical Features
Musical rhythm is expressed by the successive notes that record the relating temporal information. Tempo () could be decided as the unit, beats per minute (bpm) .
The other characteristic, perceptual complexity (), was obtained by asking the human subjects to judge the complexity of the rhythms they had listened to. It was assessed by the subjects using a subjective rating of 1 to 4 on a Likert Scale (4 being the most complex) .
3.2.2. HRV Features
To acquire the measures of HRV features, QRS detection was the first step [96–98], where denotes the peak in a heartbeat signal. After the abnormal beats were rejected , the mean of R-R intervals (MRR), standard deviation of normal-to-normal R-R intervals (SDNN), and root of mean of sum of square of differences of adjacent R-R intervals (RMSSD) were measured in time domain . After interpolation  (prepared for FFT) and detrending  (to filter the respiratory signal), the FFT  was applied to calculate the low- (LF) and high-frequency (HF) powers and their ratio (LF/HF). The results of SDNN and LF/HF are listed in Table 2. Four groups of data, SDNN C1, SDNN C2, LF/HF C1, and LF/HF C2, were observed in our analysis.
For modeling the responses of some HRV measure and the related musical rhythms, our algorithm included 3 steps. Initially, all possible combinations of the rhythmic features were explored. Secondly, the values of the combinations were linearly transformed. Finally, the coefficients of the linear transformations were calculated such that the Euclidian metric between the results of step 2 and the related HRV responses is the minimum.
The stimuli are 4 drum loops (rhythm), named as to here:
Each rhythm relates to some HRV measure :
For each rhythm, there are three rhythmic features, , , and , named as to :
Since a musical stimulus contains multiple combinations and interactions of various features, it is difficult to realize which feature is contributing to the perceived emotion , or physiological responses. Our solution is considering all possible combinations: the influence of each feature is linear (order 1), of no effect (order 0), or inverse (order −1). Although the higher orders are plausible, order one is still the most suitable to construct a simplified model for understanding the relation of the musical rhythms to their related HRV measures:
Linear transformation is necessary because the units of rhythmic and HRV features are not uniform. All we need are the related correspondences. Consider
Thus for some subject’s HRV responses to (e.g., SDNN) to rhythms to , some combination of rhythmic features can model the relation if the metric between and is the minimum. Euclidean metric [27, 103] is employed in this work:
After squaring each side of (6), we can acquire the coefficients of the linear transformation to get the minimum metric if the partial derivatives of (with respect to and ) are both zero:
The judgment and removal of outliers are fundamental in the processing of experimental data . In our study, the experimental data was separated into three groups by the minimum metric mentioned in Section 3.3 and the group with larger metric was eliminated.
The basic idea is to rank a group of data by the algorithm we proposed and make the data with larger metric obsolete. There were four groups of data, SDNN (C1), SDNN (C2), LF/HF (C1), and LF/HF (C2). Each group had 22 records of the 22 participants. Each record had four subrecords of drum loops L1 to L4. First, we used the algorithm in each group and ranked the 22 records as 1 to 22 by their minimum metric. After summing the ranks of the four groups of these 22 participants, these participants were partitioned into three levels: small, medium, and large metric (7, 7, and 8 participants). We defined the predictable class with the small and medium metrics and the nonpredictable class with the large metric. The nonpredictable class was excluded from our experimental data. After removing the nonpredictable class, the rhythmic features and the four averages of these four groups of predictable class were modeled by our algorithm again.
3.4.2. ANOVA and -Test
Repeated measurements have four major advantages: obtaining the individual patterns of change, less subjects, serving as subjects’ own controls, and reliability; the disadvantages are the complication by the dependence among repeated observations and less control of the circumstances . In our experiment, the repeated measurements were employed.
Two basic methods of ANOVA are one-way between-groups and one-way within-groups. A more complicated method is two-way factorial ANOVA with one between-groups and one within-groups factor . The repeated-measures ANOVA can be considered as a special case of two-way ANOVA . The formula of repeated measures ANOVA with one within-groups factor and one between-groups factor () is listed as 
For each HRV measure, there are two ANOVA tables. is the participant. If the epoch is fixed, is the rhythm. If the rhythm is fixed, is the epoch. Then the Tukey honest significant difference test was used to acquire pair-by-pair comparisons  for each between-groups pair.
Table 2 collected the musical features, factors of models, and HRV data in this study. Valence C1 and arousal C1 reveal how rhythmic features influenced the HRV responses while listening to music, and valence C2 and arousal C2 reveal how rhythmic features influenced the HRV responses after listening to music. The values of modified combinations of the rhythmic and HRV features were illustrated in Figures 3(a) and 3(b). Furthermore, (9) to (12) demonstrated the relationships.
Fast tempo enhanced SDNN; that is, people prefer faster tempi:
High intensity and low complexity enhanced LF/HF, the physiological arousal:
People prefer faster tempi. However, the sound should not be too loud if we hope the effect can remain after the music stops: If fast and loud music stops, people feel relaxed:
Table 3 is the statistical results. The SDNN (C2) of the four rhythms had a significance value 0.01. The LF/HF of L3 had a significance value 0.03 while and after listening to music.
|The values of HRV measures in C1 are the comparisons (differences) of values of HRV measures in epochs E1 and E3. |
The values of HRV measures in C2 are the comparisons (differences) of values of HRV measures in epochs E2 and E4.
L1-L2: drum loops 1 and 2; L1~L4: drum loops 1 to 4.
Significant codes: 0 “***” 0.001 “**” 0.01 “*” 0.05 “.” 0.1 “ ” 1.
5.1. Advantages of the Proposed Model
There are four major advantages in this work. Initially, the psychological music models are many [2, 25–45], but the physiological music models are rare . Secondly, the psychological music models [2, 25–45] and physiological musical experiments [1, 111] are many, but studies of the psychological rhythm model  and the physiological rhythmic experiment  are rare. Thirdly, almost all studies concern the responses during the music, but discussions about the responses after the music are limited . Finally, there are many music studies about HRV [1, 111], but there is no model. Our work is the first one.
5.2. Comparison of the Models
There are six levels of musical model illustrated in Figure 4: (a) the model from the results of the psychological experiments, (b) the model of the perceived emotion, (c) the model of the felt emotion, (d) the model from the physiological results, (e) the proposed physiological valence/arousal model while listening to the musical rhythms, and (f) the proposed physiological valence/arousal model after listening to the musical rhythms.
The six levels of model need not reach a consensus. There are four reasons. Initially, the perceived emotion is not always the same as the felt emotion [10, 54]. These two emotions are closer if the music clips are chosen by the participants themselves instead of others. Secondly, some measured physiological responses might not have been reported by self-report . Thirdly, if the music clips are simplified to the form of rhythm or tempo, the responses are changed . Finally, the responses while and after listening to music need not be the same obviously.
Figure 4(a) shows that and dominate arousal and dominates valence in a survey of psychological experiments (pp. 383–392) . The effects of and on arousal are clear. As for valence, the regular and varied rhythms derive positive emotions and the irregular rhythms derive negative emotions (pp. 383–392) . Hence valence is negatively correlated to .
Figure 4(b) shows that [26, 31, 45] and [26, 27, 31, 45] dominate arousal and  and  dominate valence in the perceived models. The results of arousal for the perceived models are the same as the psychological experiments. But and are viewed as the valence-based features. Whether they are positively correlated to valence or not has not been highlighted in the paper .
Figure 4(c) shows that  and  dominate arousal and  dominates valence in the felt model. The results of felt and perceived models are similar. But the valence-based feature is limited at in the felt model .
Figure 4(d) shows that [55, 113, 114] and  dominate arousal and  dominates valence in the physiological experiments. Both psychological studies (Figures 4(a), 4(b), and 4(c)) and physiological studies (Figure 4(d)) have the same results about arousal. But the faster tempi are found to decrease SDNN, the physiological valence .
Figure 4(e) shows that and dominate arousal and dominates valence. is directly proportional to arousal and is inversely proportional to arousal. As for valence, faster tempi increase SDNN, the physiological valence. This relation is also illustrated in the left part of Figure 3(a).
Figure 4(f) shows that and dominate both arousal and valence after the stimuli. The arousal is negatively correlated to . Hence if a fast and loud rhythm is removed, the subjects will feel relaxed. Fast tempi increase valence in Figure 4(e), the responses during music. This effect remains after the music stops if the intensity of the music is low.
In the valence perspective, and are considered as two major rhythmic parameters . As for , there exist opposite comments. is positively correlated with valence in the felt models  but is negatively correlated with physiological valence since the faster tempo decreased SDNN . The proper conclusion is that the role of is dependent on the context: both slow and fast tempi can derive different emotions (pp. 383–392) . For example, both happy and angry music have fast tempi. As for , its contribution is not defined in the perceived model . Although is negatively correlated with valence in the psychological experiments (pp. 383–392) , the same definition has opposite results: the firm rhythm derives positive valence (pp. 383–392)  and the firm rhythm derives negative valence . Instead of music, if only the rhythm is considered, higher SDNN is observed with faster tempi in our work, as Figure 4(e) illustrated; and affects the responses of physiological valence after the music stops, as Figure 4(f) illustrated.
In the arousal perspective, and are two of the most dominant features. Our findings also show that if a fast, loud sound (drum loop 3) was removed, LF/HF would decrease, as illustrated in the right part of Figure 3(b). Although is usually considered as the most important factor in all features (pp. 383–392) , another research indicates is more important than in arousal ; our findings support this opinion, as illustrated in Figure 4(e). is considered as a valence parameter in Figures 4(a) and 4(b). But our results show that simple rhythms enhanced the physiological arousal.
5.3. Statistical Results
Table 3 collects all values and there are two significant results: the ANOVA of SDNN (C2) for L1~L4 and the -test of LF/HF (C1, C2) for L3. Although the significant value is more than 0.05, the left part of Figure 3(a) still shows that SDNN, the physiological valence, is positively correlated with the tempi of the drum loops. The right part of Figure 3(a) shows that SDNN is positively correlated with (). Both L2 and L4 have faster tempi and smaller energies. The left part of Figure 3(b) shows that LF/HF, the physiological arousal, is positively correlated with . L3, a fast drum loop with a very simple complexity, enhances arousal in the largest degree. The right part of Figure 3(b) shows that after the music stops, the fast and loud rhythms make the participants relaxed. Although there is no statistical significance in Figure 3(b), the -test result of L3, while and after listening to the music clip, reaches statistical significance ().
5.4.1. The Three Rhythmic Parameters
Of the three elementary rhythmic features, tempo is the most definite . is positively correlated with SDNN, the physiological valence (Figure 4(e)). In the other way, is negatively correlated with SDNN . To integrate these two opposite findings, the psychological perspectives (Figures 4(a), 4(b), and 4(c)) are more suitable to explain this dilemma. That is, is a feature of arousal; both positive and negative valence can be measured; it depends on the context (pp. 383–392) .
To measure the complexity of a music or rhythm clip, mathematical quantities are better than adjectives. For example, the firm rhythms can derive both positive (pp. 383–392)  and negative  valence. We can acquire a number from Likert Scale ; however, it is not proper as an engineering reference. Of the mathematical studies about rhythmic complexity [115–117], oddity  is the closest method for measuring the perceptual complexity. There is a more fundamental issue; two rhythms with different responses may have the same complexity measure.
The last parameter, the total energy of a waveform, is defined as , the summation of square of , where is the amplitude of the signal [94, 95]. But the intensity within the waveform may not be fixed: large or small variations suggest fear or happiness; the rapid or few changes in intensity may be associated with pleading or sadness (pp. 383–392) . If the waveform is partitioned into small segments , the energy feature will be not only an arousal, but also a valence parameter (pp. 383–392) . And the analytic resolution of the musical emotion will increase.
5.4.2. Nonlinearity of the Responses
The linear and inverse functions are simple, intuitive, and useful in a general overview. However, some studies assume that the liking (or valence) of music is an inverted-U curve, dependent on arousal (tempo and energy) [119–121] and complexity [119, 120]. The inverted-U phenomenon may disappear with the professional musical expertise . There is also a plausible transform of the inverted-U curve: although 120 bpm may be a preferred tempo for common people , the twin-peak curve was found in both a popular music database and physiological experiments . Finally, the more complex curves are possible .
5.5. The Modified Model
and correlate with arousal, as Figures 4(a)–4(d) and Figure 4(f) illustrated, but dominates the arousal , as Figure 4(e) showed. To integrate these models, let be the average energy of all beats. Thus is the product of and , and (9)–(12) can be modified as (13)–(16). This provides us with an advanced aspect.
The valence of musical responses keeps the same:
Both tempo and energy are arousal parameters now. Moreover, our model demonstrates that the complexity of rhythm is also an arousal parameter:
The valence after the musical stimuli is influenced by the energy:
Finally, both tempo and energy influence the arousal after musical stimuli, but tempo dominates the effect:
To integrate all psychological and physiological music theories, we summarize briefly that tempo, energy, and complexity are both valence and arousal parameters. The inverted-U curve [119–121] is suggested for valence. Considering the valence after music, people prefer the music of low intensity. For the responses after the musical stimuli, the arousal correlates negatively with tempo and energy whereas tempo is the dominant parameter.
5.6. Criteria for Good Models
There are four criteria for a valuable model . They are generalization (goodness-of-fit ) and predictive power, simplicity, and its relation to existing theories. In plain words, a model should be able to explain experimental data used in the model and not used in the model for verification. The parameters should be small relative to the amount of the data. The model should be able to unify two formerly unrelated theories. It will help to understand the underlying phenomenon instead of an input-output routine.
Our proposed model satisfies generalization, simplicity, and the relation to other theories. In our study, the selected model has the least metric (error) among all possible models. It has only three parameters (, , and ). And it fills the gap among the psychological and physiological experiments and the models of music and rhythm. The proposed model does not have enough predictive power. It was built for the subjects whose responses are able to be modeled; the nonpredictable class, one-third of all subjects, was excluded. There was no test data in this study. However, the model derived from the experimental data can be integrated well with the literature in this field, as Section 5.5 mentioned.
6. Conclusion and Future Work
A novel experiment, a novel algorithm, and a novel model have been proposed in this study. The experiment explored how rhythm, the fundamental feature of music, influenced the ANS and HRV. And the relationships between musical features and physiological features were derived from the algorithm. Moreover, the model demonstrated how the rhythmic features were mapped to the physiological valence/arousal plane.
The equations in our model could be the rules for music generating systems [125–127]. The model could also be fine-tuned if the rhythmic oddity  (for complexity), small segments  (for energy), and various curves [3, 119–121, 124] are considered.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
The authors thank Mr. Shih-Hsiang Lin heartily for his support in both program design and data analysis. The authors also thank Professor Shao-Yi Chien and Dr. Ming-Yie Jan for related discussion.
- P. N. Juslin and J. A. Sloboda, Eds., Handbook of Music and Emotion: Theory, Research, Applications, Oxford University Press, 2010.
- Y. H. Yang and H. H. Chen, Music Emotion Recognition, CRC Press, 2011.
- P. Gomez and B. Danuser, “Relationships between musical structure and psychophysiological measures of emotion,” Emotion, vol. 7, no. 2, pp. 377–387, 2007.
- G. Cervellin and G. Lippi, “From music-beat to heart-beat: a journey in the complex interactions between music, brain and heart,” European Journal of Internal Medicine, vol. 22, no. 4, pp. 371–374, 2011.
- S.-H. Lin, Y.-C. Huang, C.-Y. Chien, L.-C. Chou, S.-C. Huang, and M.-Y. Jan, “A study of the relationship between two musical rhythm characteristics and heart rate variability (HRV),” in Proceedings of the IEEE International Conference on BioMedical Engineering and Informatics, pp. 344–347, 2008.
- H.-M. Wang, S.-H. Lin, Y.-C. Huang et al., “A computational model of the relationship between musical rhythm and heart rhythm,” in Proceeding of the IEEE International Symposium on Circuits and Systems (ISCAS '09), pp. 3102–3105, Taipei, Taiwan, May 2009.
- H.-M. Wang, Y.-C. Lee, B. S. Yen, C.-Y. Wang, S.-C. Huang, and K.-T. Tang, “A physiological valence/arousal model from musical rhythm to heart rhythm,” in Proceedings of the IEEE International Symposium of Circuits and Systems (ISCAS '11), pp. 1013–1016, Rio de Janeiro, Brazil, May 2011.
- C. L. Krumhansl, “Music: a link between cognition and emotion,” Current Directions in Psychological Science, vol. 11, no. 2, pp. 45–50, 2002.
- M. Pearce and M. Rohrmeier, “Music cognition and the cognitive sciences,” Topics in Cognitive Science, vol. 4, no. 4, pp. 468–484, 2012.
- P. Evans and E. Schubert, “Relationships between expressed and felt emotions in music,” Musicae Scientiae, vol. 12, no. 1, pp. 75–99, 2008.
- V. J. Koneĉni, “Does music induce emotion? A theoretical and methodological analysis,” Psychology of Aesthetics, Creativity, and the Arts, vol. 2, no. 2, pp. 115–129, 2008.
- M. Zentner, D. Grandjean, and K. R. Scherer, “Emotions evoked by the sound of music: characterization, classification, and measurement,” Emotion, vol. 8, no. 4, pp. 494–521, 2008.
- J. A. Russell, “A circumplex model of affect,” Journal of Personality and Social Psychology, vol. 39, no. 6, pp. 1161–1178, 1980.
- R. E. Thayer, The Biopsychology of Mood and Arousal, Oxford University Press on Demand, 1989.
- D. Liu, L. Lu, and H.-J. Zhang, “Automatic mood detection from acoustic music data,” in Proceedings of the International Symposium on Music Information Retrieval, pp. 81–87, 2003.
- Y. E. Kim, E. M. Schmidt, R. Migneco et al., “Music emotion recognition: a state of the art review,” in Proceedings of the International Symposium on Music Information Retrieval, pp. 255–266, 2010.
- Y.-H. Yang and H. H. Chen, “Machine recognition of music emotion: a review,” ACM Transactions on Intelligent Systems and Technology, vol. 3, no. 3, article 40, 2012.
- K. Trohidis, G. Tsoumakas, G. Kalliris, and I. Vlahavas, “Multi-label classification of music into emotions,” in Proceedings of the International Symposium on Music Information Retrieval, pp. 325–330, September 2008.
- M. A. Rohrmeier and S. Koelsch, “Predictive information processing in music cognition. A critical review,” International Journal of Psychophysiology, vol. 83, no. 2, pp. 164–175, 2012.
- G. Widmer and W. Goebl, “Computational models of expressive music performance: the state of the art,” Journal of New Music Research, vol. 33, pp. 203–216, 2004.
- H. Honing, “Computational modeling of music cognition: a case study on model selection,” Music Perception, vol. 23, no. 5, pp. 365–376, 2006.
- H. Purwins, P. Herrera, M. Grachten, A. Hazan, R. Marxer, and X. Serra, “Computational models of music perception and cognition I: the perceptual and cognitive processing chain,” Physics of Life Reviews, vol. 5, no. 3, pp. 151–168, 2008.
- H. Purwins, M. Grachten, P. Herrera, A. Hazan, R. Marxer, and X. Serra, “Computational models of music perception and cognition II: domain-specific music processing,” Physics of Life Reviews, vol. 5, no. 3, pp. 169–182, 2008.
- T. Eerola, “Modeling listeners’ emotional response to music,” Topics in Cognitive Science, vol. 4, no. 4, pp. 607–624, 2012.
- J.-C. Wang, Y.-H. Yang, H.-M. Wang, and S.-K. Jeng, “The acoustic emotion gaussians model for emotion-based music annotation and retrieval,” in Proceedings of the 20th ACM International Conference on Multimedia (MM '12), pp. 89–98, November 2012.
- J. Cu, R. Cabredo, R. Legaspi, and M. T. Suarez, “On modelling emotional responses to rhythm features,” in Proceedings of the Trends in Artificial Intelligence (PRICAI '12), pp. 857–860, Springer, 2012.
- J. J. Deng and C. Leung, “Music emotion retrieval based on acoustic features,” in Advances in Electric and Electronics, vol. 155 of Lecture Notes in Electrical Engineering, pp. 169–177, 2012.
- H. G. Avisado, J. V. Cocjin, J. A. Gaverza, R. Cabredo, J. Cu, and M. Suarez, “Analysis of music timbre features for the construction of user-specific affect model,” in Theory and Practice of Computation, pp. 28–35, Springer, Berlin, Germany, 2012.
- M. M. Farbood, “A parametric, temporal model of musical tension,” Music Perception, vol. 29, no. 4, pp. 387–428, 2012.
- J. Albrecht, “A model of perceived musical affect accurately predicts self-report ratings,” in Proceedings of the International Conference on Music Perception, pp. 35–43, 2012.
- Y.-H. Yang and H. H. Chen, “Prediction of the distribution of perceived music emotions using discrete samples,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 19, pp. 2184–2196, 2011.
- R. Y. Granot and Z. Eitan, “Musical tension and the interaction of dynamic auditory parameters,” Music Perception, vol. 28, no. 3, pp. 219–246, 2011.
- E. M. Schmidt and Y. E. Kim, “Modeling musical emotion dynamics with conditional random fields,” in Proceedings of the 12th International Society for Music Information Retrieval Conference (ISMIR '11), pp. 777–782, October 2011.
- B. Schuller, J. Dorfner, and G. Rigoll, “Determination of nonprototypical valence and arousal in popular music: features and performances,” EURASIP Journal on Audio, Speech, and Music Processing, vol. 2010, Article ID 735854, 2010.
- E. Coutinho and A. Cangelosi, “The use of spatio-temporal connectionist models in psychological studies of musical emotions,” Music Perception, vol. 27, no. 1, pp. 1–15, 2009.
- T. Eerola, O. Lartillot, and P. Toiviainen, “Prediction of multidimensional emotional ratings in music from audio using multivariate regression models,” in Proceedings of the International Symposium on Music Information Retrieval, pp. 621–626, 2009.
- T.-L. Wu and S.-K. Jeng, “Probabilistic estimation of a novel music emotion model,” in Advances in Multimedia Modeling, pp. 487–497, Springer, 2008.
- G. Luck, P. Toiviainen, J. Erkkilä et al., “Modelling the relationships between emotional responses to, and musical content of, music therapy improvisations,” Psychology of Music, vol. 36, no. 1, pp. 25–45, 2008.
- Y.-H. Yang, Y.-C. Lin, Y.-F. Su, and H. H. Chen, “A regression approach to music emotion recognition,” IEEE Transactions on Audio, Speech and Language Processing, vol. 16, no. 2, pp. 448–457, 2008.
- F. Lerdahl and C. L. Krumhansl, “Modeling tonal tension,” Music Perception, vol. 24, no. 4, pp. 329–366, 2007.
- Y.-H. Yang, C.-C. Liu, and H. H. Chen, “Music emotion classification: a fuzzy approach,” in Proceeding of the Annual ACM International Conference on Multimedia (MM '06), pp. 81–84, New York, NY, USA, October 2006.
- M. D. Korhonen, D. A. Clausi, and M. E. Jernigan, “Modeling emotional content of music using system identification,” IEEE Transactions on Systems, Man, and Cybernetics B: Cybernetics, vol. 36, no. 3, pp. 588–599, 2006.
- M. Leman, V. Vermeulen, L. de Voogdt, D. Moelants, and M. Lesaffre, “Prediction of musical affect using a combination of acoustic structural cues,” Journal of New Music Research, vol. 34, no. 1, pp. 39–67, 2005.
- E. H. Margulis, “A model of melodic expectation,” Music Perception, vol. 22, no. 4, pp. 663–714, 2005.
- E. Schubert, “Modeling perceived emotion with continuous musical features,” Music Perception, vol. 21, pp. 561–585, 2004.
- K. R. Scherer, “Which emotions can be induced by music? What are the underlying mechanisms? And how can we measure them?” Journal of New Music Research, vol. 33, pp. 239–251, 2004.
- P. N. Juslin and D. Västfjäll, “Emotional responses to music: the need to consider underlying mechanisms,” Behavioral and Brain Sciences, vol. 31, no. 5, pp. 559–621, 2008.
- L.-O. Lundqvist, F. Carlsson, P. Hilmersson, and P. N. Juslin, “Emotional responses to music: experience, expression, and physiology,” Psychology of Music, vol. 37, no. 1, pp. 61–90, 2009.
- N. S. Rickard, “Intense emotional responses to music: a test of the physiological arousal hypothesis,” Psychology of Music, vol. 32, no. 4, pp. 371–388, 2004.
- J. Kim and E. André, “Emotion recognition based on physiological changes in music listening,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30, no. 12, pp. 2067–2083, 2008.
- E. Coutinho and A. Cangelosi, Computational and Psycho-Physiological Investigations of Musical Emotions, University of Plymouth, 2008.
- E. Coutinho and A. Cangelosi, “Musical emotions: predicting second-by-second subjective feelings of emotion from low-level psychoacoustic features and physiological measurements,” Emotion, vol. 11, no. 4, pp. 921–937, 2011.
- K. Trochidis, D. Sears, D. L. Tran, and S. McAdams, “Psychophysiological measures of emotional response to Romantic orchestral music and their musical and acoustic correlates,” in Proceedings of the International Symposium on Computer Music Modelling and Retrieval, pp. 45–52, 2012.
- V. N. Salimpoor, M. Benovoy, G. Longo, J. R. Cooperstock, and R. J. Zatorre, “The rewarding aspects of music listening are related to degree of emotional arousal,” PLoS ONE, vol. 4, no. 10, Article ID e7487, 2009.
- M. D. van der Zwaag, J. H. D. M. Westerink, and E. L. van den Broek, “Emotional and psychophysiological responses to tempo, mode, and percussiveness,” Musicae Scientiae, vol. 15, no. 2, pp. 250–269, 2011.
- J. R. J. Fontaine, K. R. Scherer, E. B. Roesch, and P. C. Ellsworth, “The world of emotions is not two-dimensional,” Psychological Science, vol. 18, no. 12, pp. 1050–1057, 2007.
- T. Eerola and J. K. Vuoskoski, “A comparison of the discrete and dimensional models of emotion in music,” Psychology of Music, vol. 39, no. 1, pp. 18–49, 2011.
- U. R. Acharya, K. P. Joseph, N. Kannathal, C. M. Lim, and J. S. Suri, “Heart rate variability: a review,” Medical and Biological Engineering and Computing, vol. 44, no. 12, pp. 1031–1051, 2006.
- “Heart rate variability: standards of measurement, physiological interpretation and clinical use. Task Force of the European Society of Cardiology and the North American Society of Pacing and Electrophysiology,” Circulation, vol. 93, no. 5, pp. 1043–1065, 1996.
- R. D. Lane, K. McRae, E. M. Reiman, K. Chen, G. L. Ahern, and J. F. Thayer, “Neural correlates of heart rate variability during emotion,” NeuroImage, vol. 44, no. 1, pp. 213–222, 2009.
- P. Rainville, A. Bechara, N. Naqvi, and A. R. Damasio, “Basic emotions are associated with distinct patterns of cardiorespiratory activity,” International Journal of Psychophysiology, vol. 61, no. 1, pp. 5–18, 2006.
- B. M. Appelhans and L. J. Luecken, “Heart rate variability as an index of regulated emotional responding,” Review of General Psychology, vol. 10, no. 3, pp. 229–240, 2006.
- G. H. E. Gendolla and J. Krüsken, “Mood state, task demand, and effort-related cardiovascular response,” Cognition and Emotion, vol. 16, no. 5, pp. 577–603, 2002.
- R. McCraty, M. Atkinson, W. A. Tiller, G. Rein, and A. D. Watkins, “The effects of emotions on short-term power spectrum analysis of heart rate variability,” The American Journal of Cardiology, vol. 76, no. 14, pp. 1089–1093, 1995.
- A. H. Kemp, D. S. Quintana, and M. A. Gray, “Is heart rate variability reduced in depression without cardiovascular disease?” Biological Psychiatry, vol. 69, no. 4, pp. e3–e4, 2011.
- C. M. M. Licht, E. J. C. De Geus, F. G. Zitman, W. J. G. Hoogendijk, R. Van Dyck, and B. W. J. H. Penninx, “Association between major depressive disorder and heart rate variability in the Netherlands study of depression and anxiety (NESDA),” Archives of General Psychiatry, vol. 65, no. 12, pp. 1358–1367, 2008.
- R. M. Carney, R. D. Saunders, K. E. Freedland, P. Stein, M. W. Rich, and A. S. Jaffe, “Association of depression with reduced heart rate variability in coronary artery disease,” The American Journal of Cardiology, vol. 76, no. 8, pp. 562–564, 1995.
- F. Geisler, N. Vennewald, T. Kubiak, and H. Weber, “The impact of heart rate variability on subjective well-being is mediated by emotion regulation,” Personality and Individual Differences, vol. 49, no. 7, pp. 723–728, 2010.
- S. Koelsch, A. Remppis, D. Sammler et al., “A cardiac signature of emotionality,” European Journal of Neuroscience, vol. 26, no. 11, pp. 3328–3338, 2007.
- R. Krittayaphong, W. E. Cascio, K. C. Light et al., “Heart rate variability in patients with coronary artery disease: differences in patients with higher and lower depression scores,” Psychosomatic Medicine, vol. 59, no. 3, pp. 231–235, 1997.
- M. Müller, D. P. W. Ellis, A. Klapuri, and G. Richard, “Signal processing for music analysis,” IEEE Journal on Selected Topics in Signal Processing, vol. 5, no. 6, pp. 1088–1110, 2011.
- K. Jensen, “Multiple scale music segmentation using rhythm, timbre, and harmony,” Eurasip Journal on Advances in Signal Processing, vol. 2007, Article ID 73205, 2007.
- N. Scaringella, G. Zoia, and D. Mlynek, “Automatic genre classification of music content: a survey,” IEEE Signal Processing Magazine, vol. 23, no. 2, pp. 133–141, 2006.
- J.-J. Aucouturier and F. Pachet, “Representing musical genre: a state of the art,” Journal of New Music Research, vol. 32, pp. 83–93, 2003.
- T. Li and M. Ogihara, “Detecting emotion in music,” in Proceedings of the International Symposium on Music Information Retrieval, pp. 239–240, 2003.
- G. Tzanetakis and P. Cook, “Musical genre classification of audio signals,” IEEE Transactions on Speech and Audio Processing, vol. 10, no. 5, pp. 293–302, 2002.
- L. Jaquet, B. Danuser, and P. Gomez, “Music and felt emotions: how systematic pitch level variations affect the experience of pleasantness and arousal,” Psychology of Music, 2012.
- J. C. Hailstone, R. Omar, S. M. D. Henley, C. Frost, M. G. Kenward, and J. D. Warren, “It's not what you play, it's how you play it: timbre affects perception of emotion in music,” The Quarterly Journal of Experimental Psychology, vol. 62, no. 11, pp. 2141–2155, 2009.
- L. Trainor, “Science & music: the neural roots of music,” Nature, vol. 453, no. 7195, pp. 598–599, 2008.
- J. Bispham, “Rhythm in Music: what is it? Who has it? And why?” Music Perception, vol. 24, no. 2, pp. 125–134, 2006.
- J. A. Grahn, “Neuroscientific investigations of musical rhythm: recent advances and future challenges,” Contemporary Music Review, vol. 28, no. 3, pp. 251–277, 2009.
- K. Overy and R. Turner, “The rhythmic brain,” Cortex, vol. 45, no. 1, pp. 1–3, 2009.
- J. S. Snyder, E. W. Large, and V. Penhune, “Rhythms in the brain: basic science and clinical perspectives,” Annals of the New York Academy of Sciences, vol. 1169, pp. 13–14, 2009.
- H. Honing, “Without it no music: beat induction as a fundamental musical trait,” Annals of the New York Academy of Sciences, vol. 1252, no. 1, pp. 85–91, 2012.
- J. A. Grahn, “Neural mechanisms of rhythm perception: current findings and future perspectives,” Topics in Cognitive Science, vol. 4, no. 4, pp. 585–606, 2012.
- I. Winkler, G. P. Háden, O. Ladinig, I. Sziller, and H. Honing, “Newborn infants detect the beat in music,” Proceedings of the National Academy of Sciences, vol. 106, pp. 2468–2471, 2009.
- M. Zentner and T. Eerola, “Rhythmic engagement with music in infancy,” Proceedings of the National Academy of Sciences of the United States of America, vol. 107, no. 13, pp. 5768–5773, 2010.
- S. E. Trehub and E. E. Hannon, “Conventional rhythms enhance infants' and adults' perception of musical patterns,” Cortex, vol. 45, no. 1, pp. 110–118, 2009.
- E. Geiser, E. Ziegler, L. Jancke, and M. Meyer, “Early electrophysiological correlates of meter and rhythm processing in music perception,” Cortex, vol. 45, no. 1, pp. 93–102, 2009.
- C.-H. Yeh, H.-H. Lin, and H.-T. Chang, “An efficient emotion detection scheme for popular music,” in Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS '09), pp. 1799–1802, Taipei, Taiwan, May 2009.
- L. Lu, D. Liu, and H. J. Zhang, “Automatic mood detection and tracking of music audio signals,” IEEE Transactions on Audio, Speech and Language Processing, vol. 14, no. 1, pp. 5–18, 2006.
- S. Dixon, “Automatic extraction of tempo and beat from expressive performances,” Journal of New Music Research, vol. 30, pp. 39–58, 2001.
- R. Likert, “A technique for the measurement of attitudes,” Archives of Psychology, vol. 22, no. 140, pp. 1–55, 1932.
- W. A. Sethares, R. D. Morris, and J. C. Sethares, “Beat tracking of musical performances using low-level audio features,” IEEE Transactions on Speech and Audio Processing, vol. 13, no. 2, pp. 275–285, 2005.
- J. P. Bello, C. Duxbury, M. Davies, and M. Sandler, “On the use of phase and energy for musical onset detection in the complex domain,” IEEE Signal Processing Letters, vol. 11, no. 6, pp. 553–556, 2004.
- W. Wei, C. Zhang, and W. Lin, “A QRS wave detection algorithm based on complex wavelet transform,” Applied Mechanics and Materials, vol. 239-240, pp. 1284–1288, 2013.
- B. U. Köhler, C. Hennig, and R. Orglmeister, “The principles of software QRS detection,” IEEE Engineering in Medicine and Biology Magazine, vol. 21, no. 1, pp. 42–57, 2002.
- J. Pan and W. J. Tompkins, “A real-time QRS detection algorithm,” IEEE Transactions on Biomedical Engineering, vol. 32, no. 3, pp. 230–236, 1985.
- J. Mateo and P. Laguna, “Analysis of heart rate variability in the presence of ectopic beats using the heart timing signal,” IEEE Transactions on Biomedical Engineering, vol. 50, no. 3, pp. 334–343, 2003.
- S. McKinley and M. Levine, “Cubic spline interpolation,” http://online.redwoods.edu/instruct/darnold/laproj/fall98/skymeg/proj.pdf.
- M. P. Tarvainen, P. O. Ranta-aho, and P. A. Karjalainen, “An advanced detrending method with application to HRV analysis,” IEEE Transactions on Biomedical Engineering, vol. 49, no. 2, pp. 172–175, 2002.
- P. Welch, “The use of fast Fourier transform for the estimation of power spectra: a method based on time averaging over short, modified periodograms,” IEEE Transactions on Audio and Electroacoustics, vol. 15, pp. 70–73, 1967.
- T. Warren Liao, “Clustering of time series data: a survey,” Pattern Recognition, vol. 38, no. 11, pp. 1857–1874, 2005.
- V. Barnett and T. Lewis, Outliers in Statistical Data, vol. 1, John Wiley & Sons, 1984.
- C. S. Davis, Statistical Methods for the Analysis of Repeated Measurements, Springer, New York, NY, USA, 2002.
- R. Kabacoff, R in Action, Manning Publications, 2011.
- L. Pace, Beginning R: An Introduction to Statistical Programming, Apress, 2012.
- J. Albert and M. Rizzo, R by Example, Springer, 2012.
- M. Gardener, Beginning R: The Statistical Programming Language, Wrox, 2012.
- J. Adler and J. Beyer, R in a Nutshell, O'Reilly, Frankfurt, Germany, 2010.
- S. D. Kreibig, “Autonomic nervous system activity in emotion: a review,” Biological Psychology, vol. 84, no. 3, pp. 394–421, 2010.
- S. Khalfa, M. Roy, P. Rainville, S. Dalla Bella, and I. Peretz, “Role of tempo entrainment in psychophysiological differentiation of happy and sad music?” International Journal of Psychophysiology, vol. 68, no. 1, pp. 17–26, 2008.
- L. Bernardi, C. Porta, and P. Sleight, “Cardiovascular, cerebrovascular, and respiratory changes induced by different types of music in musicians and non-musicians: the importance of silence,” Heart, vol. 92, no. 4, pp. 445–452, 2006.
- M. Iwanaga, A. Kobayashi, and C. Kawasaki, “Heart rate variability with repetitive exposure to music,” Biological Psychology, vol. 70, no. 1, pp. 61–66, 2005.
- E. Thul and G. T. Toussaint, “Analysis of musical rhythm complexity measures in a cultural context,” in Proceedings of the C3S2E Conference (C3S2E '08), pp. 7–9, May 2008.
- E. Thul and G. T. Toussaint, “On the relation between rhythm complexity measures and human rhythmic performance,” in Proceeding of the C3S2E Conference (C3S2E '08), pp. 199–204, May 2008.
- E. Thul and G. T. Toussaint, “Rhythm complexity measures: a comparison of mathematical models of human perception and performance,” in Proceedings of the International Symposium on Music Information Retrieval, pp. 14–18, 2008.
- Z. Xiao, E. Dellandrea, W. Dou, and L. Chen, “What is the best segment duration for music mood analysis ?” in Proceedings of the International Workshop on Content-Based Multimedia Indexing (CBMI '08), pp. 17–24, London, UK, June 2008.
- A. Lamont and R. Webb, “Short- and long-term musical preferences: what makes a favourite piece of music?” Psychology of Music, vol. 38, no. 2, pp. 222–241, 2010.
- A. C. North and D. J. Hargreaves, “Liking, arousal potential, and the emotions expressed by music,” Scandinavian Journal of Psychology, vol. 38, no. 1, pp. 45–53, 1997.
- S. Abdallah and M. Plumbley, “Information dynamics: patterns of expectation and surprise in the perception of music,” Connection Science, vol. 21, no. 2-3, pp. 89–117, 2009.
- M. G. Orr and S. Ohlsson, “Relationship between complexity and liking as a function of expertise,” Music Perception, vol. 22, no. 4, pp. 583–611, 2005.
- D. Moelants, “Preferred tempo reconsidered,” in Proceedings of the International Conference on Music Perception and Cognition, pp. 580–583, 2002.
- Y.-S. Shih, W.-C. Zhang, Y.-C. Lee et al., “Twin-peak effect in both cardiac response and tempo of popular music,” in Proceedings of the International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 1705–1708, 2011.
- A. P. Oliveira and A. Cardoso, “A musical system for emotional expression,” Knowledge-Based Systems, vol. 23, no. 8, pp. 901–913, 2010.
- I. Wallis, T. Ingalls, and E. Campana, “Computer-generating emotional music: the design of an affective music algorithm,” in Proceedings of the 11th International Conference on Digital Audio Effects (DAFx '08), pp. 7–12, September 2008.
- S. R. Livingstone, R. Mühlberger, A. R. Brown, and A. Loch, “Controlling musical emotionality: an affective computational architecture for influencing musical emotions,” Digital Creativity, vol. 18, no. 1, pp. 43–53, 2007.
Copyright © 2014 Hui-Min Wang and Sheng-Chieh Huang. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.