Researches of the Relationship of Music and Language on Three Levels: Syntax, Semantics and Perception of Pitch and Rhythm

Journal of Psychological Science ›› 2013, Vol. 36 ›› Issue (5) : 1078-1084.

PDF(438 KB)
PDF(438 KB)
Journal of Psychological Science ›› 2013, Vol. 36 ›› Issue (5) : 1078-1084.

Researches of the Relationship of Music and Language on Three Levels: Syntax, Semantics and Perception of Pitch and Rhythm

Author information +
History +

Abstract

Since accumulated evidences have revealed that music and language do have some relationship, this field of study has attracted more and more interest. As scientists have addressed, music and language have many features in common. They are both structured sequences which are arranged according to some rules, which is syntax. Also, they can both express meanings and emotions, which can be called semantics. Thirdly, they both have elements like pitch and rhythm. And the last, music and language are both human-unique and universal involved. To unveil the relationship of music and language, a lot of researches have been conducted to investigate the neural basis of them. Whether shared neural basis is involved in music and language has become the focus of these researches. Based on the common features of music and language, studies in this field are of 3 levels, syntax, semantics and the perception of pitch and rhythm. On the level of syntax, shared syntactic integration resource hypothesis (SSIRH) was raised in 2003. It suggests that the brain regions for syntax processing can be divided into processing regions and representation regions, and music and language may share the processing regions. Comparative studies were done and most of them well supported SSIRH. In these studies, several distinctive ERP components were found such as ERAN and N500, which later became the standard indicator of music incongruousness and were applied to many comparative studies. However, SSIRH has not been proved yet. More evidences are needed to demonstrate the accurate regions and the processing mechanism in detail. On the level of semantics, researchers found that music, even single chord could elicit a very similar N400 effect to that of language in an ERP experiment. These findings suggest that music is able to convey conceptual information like language, and they probably share some semantic processing resources. Since the processing mechanism and its location in brain remain unknown, further studies are required to illuminate the relationship of music and language in semantic level. On the level of pitch and rhythm perception, available works are inferior to that of above two levels in both amount and quality. Studies on rhythm perception in music and language are especially rare due to lack of quantitative criteria with few ones implying that there’s some connection between music and language. On the other hand, studies on pitch perception are relatively more. However, most of them put emphasis on the correlation of pitch detect ability between music and language. The limited comparative studies on processing of pitch perception in music and language merely discovered the differences between them. If music and language indeed have relationship in pitch perception, more comparative studies should be undertaken to clarify not only their different but also their common neural basis. All together, this review has tried to inform the reader with researches on the relationship of music and language processing and help to establish a comprehensive map in this field.

Key words

music / language / syntax / semantics / pitch or rhythm

Cite this article

Download Citations
Researches of the Relationship of Music and Language on Three Levels: Syntax, Semantics and Perception of Pitch and Rhythm[J]. Journal of Psychological Science. 2013, 36(5): 1078-1084

References

Besson, M., Sch?n, D., Moreno, S., Santos, A., & Magne, C. (2007). Influence of musical expertise and musical training on pitch processing in music and language. Restorative Neurology and Neuroscience, 25(3), 399-410.
Daltrozzo, J., & Sch?n, D. (2009a). Conceptual Processing in Music as Revealed by N400 Effects on Words and Musical Targets. Journal of Cognitive Neuroscience, 21(10), 1882-1892.
Daltrozzo, J., & Sch?n, D. (2009b). Is conceptual processing in music automatic? An electrophysiological approach. Brain Research, 1270, 88-94.
Fedorenko, E., Patel, A., Casasanto, D., Winawer, J., & Gibson, E. (2009). Structural integration in language and music: Evidence for a shared system. Memory & Cognition, 37(1), 1-9.
Hannon, E. E. (2009). Perceiving speech rhythm in music: Listeners classify instrumental songs according to language of origin. Cognition, 111(3), 403-409.
Jentschke, S., Koelsch, S., Sallat, S., & Friederici, A. D. (2008). Children with Specific Language Impairment Also Show Impairment of Music-syntactic Processing. Journal of Cognitive Neuroscience, 20(11), 1940-1951.
Kn?sche, T. R., Neuhaus, C., Haueisen, J., Alter, K., Maess, B., Witte, O. W., et al. (2005). Perception of phrase structure in music. Human Brain Mapping, 24(4), 259-273.
Koelsch, S. (2005). Neural substrates of processing syntax and semantics in music. Current Opinion in Neurobiology, 15(2), 207-212.
Koelsch, S. (2011). Towards a neural basis of processing musical semantics. Physics of Life Reviews, 8(2), 89-105.
Koelsch, S., Grossmann, T., Gunter, T. C., Hahne, A., Schr?ger, E., & Friederici, A. D. (2003). Children Processing Music: Electric Brain Responses Reveal Musical Competence and Gender Differences. Journal of Cognitive Neuroscience, 15(5), 683-693.
Koelsch, S., Gunter, T., Friederici, A. D., & Schr?ger, E. (2000). Brain Indices of Music Processing: “Nonmusicians” are Musical. Journal of Cognitive Neuroscience, 12(3), 520-541.
Koelsch, S., Gunter, T. C., v. Cramon, D. Y., Zysset, S., Lohmann, G., & Friederici, A. D. (2002). Bach Speaks: A Cortical "Language-Network" Serves the Processing of Music. NeuroImage, 17(2), 956-966.
Koelsch, S., Gunter, T. C., Wittfoth, M., & Sammler, D. (2005). Interaction between Syntax Processing in Language and in Music: An ERP Study. Journal of Cognitive Neuroscience, 17(10), 1565-1577.
Koelsch, S., Kasper, E., Sammler, D., Schulze, K., Gunter, T., & Friederici, A. D. (2004). Music, language and meaning: brain signatures of semantic processing. Nature neuroscience, 7(3), 302-307.
Koelsch, S., & Siebel, W. A. (2005). Towards a neural basis of music perception. Trends in Cognitive Sciences, 9(12), 578-584.
Krumhansl, C. L. (2000). Rhythm and pitch in music cognition. Psychological Bulletin, 126(1), 159-179.
Lerdahl, F. (1983). A generative theory of tonal music: Cambridge, Mass. : MIT Press.
Maess, B., Koelsch, S., Gunter, T. C., & Friederici, A. D. (2001). Musical syntax is processed in Broca's area: an MEG study. [10.1038/87502]. Nat Neurosci, 4(5), 540-545.
Magne, C., Sch?n, D., & Besson, M. (2006). Musician Children Detect Pitch Violations in Both Music and Language Better than Nonmusician Children: Behavioral and Electrophysiological Approaches. J. Cognitive Neuroscience, 18(2), 199-211.
Marques, C., Moreno, S., Luís Castro, S., & Besson, M. (2007). Musicians Detect Pitch Violation in a Foreign Language Better Than Nonmusicians: Behavioral and Electrophysiological Evidence. Journal of Cognitive Neuroscience, 19(9), 1453-1463.
Milovanov, R., Huotilainen, M., Valimaki, V., Esquef, P., & Tervaniemi, M. (2008). Musical aptitude and second language pronunciation skills in school-aged children: Neural and behavioral evidence. Brain Research, 1194, 81-89.
Nan, Y., Friederici, A. D., Shu, H., & Luo, Y.-j. (2009). Dissociable pitch processing mechanisms in lexical and melodic contexts revealed by ERPs. Brain Research, 1263, 104-113.
Patel, A. D. (1998). Syntactic Processing in Language and Music: Different Cognitive Operations, Similar Neural Resources? Music Perception: An Interdisciplinary Journal, 16(1), 27-42.
Patel, A. D. (2003). Language, music, syntax and the brain. [10.1038/nn1082]. Nat Neurosci, 6(7), 674-681.
Patel, A. D. (2008). Music, language, and the brain: New York, NY, US: Oxford University Press.
Patel, A. D., & Daniele, J. R. (2003). An empirical comparison of rhythm in language and music. Cognition, 87(1), B35-B45.
Patel, A. D., Gibson, E., Ratner, J., Besson, M., & Holcomb, P. J. (1998). Processing Syntactic Relations in Language and Music: An Event-Related Potential Study. Journal of Cognitive Neuroscience, 10(6), 717-733.
Patel, A. D., Iversen, J. R., Wassenaar, M., & Hagoort, P. (2008). Musical syntactic processing in agrammatic Broca's aphasia.
Peretz, I., & Zatorre, R. (2003). The Cognitive Neuroscience of Music: {Oxford University Press, USA}.
Sammler, D., Koelsch, S., Ball, T., Brandt, A., Elger, C. E., Friederici, A. D., et al. (2009). Overlap of Musical and Linguistic Syntax Processing: Intracranial ERP Evidence. Annals of the New York Academy of Sciences, 1169(1), 494-498.
Sch?n, D., Magne, C., & Besson, M. (2004). The music of speech: Music training facilitates pitch processing in both music and language. Psychophysiology, 41(3), 341-349.
Steinbeis, N., & Koelsch, S. (2007). Shared Neural Resources between Music and Language Indicate Semantic Processing of Musical Tension-Resolution Patterns. Cerebral Cortex, 18(5), 1169-1178.
Steinbeis, N., & Koelsch, S. (2008). Comparing the Processing of Music and Language Meaning Using EEG and fMRI Provides Evidence for Similar and Distinct Neural Representations. PLoS ONE, 3(5), e2226.
Steinbeis, N., & Koelsch, S. (2009). Affective Priming Effects of Musical Sounds on the Processing of Word Meaning. Journal of Cognitive Neuroscience, 23(3), 604-621.
Steinhauer, K., Alter, K., & Friederici, A. D. (1999). Brain potentials indicate immediate use of prosodic cues in natural speech processing. [10.1038/5757]. Nat Neurosci, 2(2), 191-196.
PDF(438 KB)

Accesses

Citation

Detail

Sections
Recommended

/