配色: 字号:
EMR000065a_Khetrapal
2023-03-20 | 阅:  转:  |  分享 
  
Empirical Musicology Reviewa Vol. 4, No, 1, 2009



11



Why does Music Therapy help in Autism?

NEHA KHETRAPAL[1]

University of Bielefeld, Germany

ABSTRACT: Music therapy is shown to be an effective intervention for emotional

recognition deficits in autism. However, researchers to date have yet to propose a

model that accounts for the neurobiological and cognitive components that are

responsible for such improvements. The current paper outlines a model whereby

the encoding of tonal pitch is proposed as the underlying mechanism. Accurate

tonal pitch perception is important for recognizing emotions like happiness and

sadness in the auditory domain. Once acquired, the ability to perceive tonal pitch

functions as a domain-specific module that proves beneficial for music cognition.

There is biological preparedness for the development of such a module and it is

hypothesized to be preserved in autism. The current paper reinforces the need to

build intervention programs based on this preserved module in autism, and

proposes that this module may form the basis for a range of benefits related to

music therapy. Possible brain areas associated with this module are suggested.

Submitted 2008 November 7; accepted 2009 January 12.

KEYWORDS: emotional processing, rehabilitation, music cognition, neurobiology



AUTISM

KANNER in 1943 was the first to describe and label autistic disorder. Autistic disorder is a

neurodevelopmental condition marked by social and communication problems as well as restricted

interests and behaviors (American Psychiatric Association [APA], 1994). It falls under the rubric of

autistic spectrum disorder (ASD). Autism is usually diagnosed before three years of age and has a

lifelong persistence. Males are usually more vulnerable with a ratio of 4 to 1 (Fombonne, 2005a). The

prevalence of the disorder is approximately 13 in every 10,000 births (Fombonne, 2005b). The

diagnostic and statistical manual, 4th edition, (DSM-IV; American Psychiatric Association) treats this

disorder as a part of the category of Pervasive Developmental Disorder, which is characterized by three

important disabilities: failure to develop normal social interaction; language delay and communication

disability; and restricted, repetitive and stereotyped behaviors. Roughly 10% of ASD individuals show

good rote memory and possess isolated savant skills like exceptional musical or drawing abilities or

calendar calculations (Volkmar & Pauls, 2003).

Music Therapy as Treatment

Individuals with autism show poor perception of affective cues within the social domain and

experimental evidence has also shown that such individuals often fail to interpret and recognize vocal

and facial expressions of emotions (Capps, Yirmiya & Sigman; 1992; Fein, Lucci, Braverman &

Waterhouse, 1992). Because the accurate perception and interpretation of such affective cues is so

important for adaptive living, rehabilitation efforts have focused on ways in which these deficits can be

alleviated. Music therapy, as a result, has been employed with success in autism (Kaplan & Steele,

2005). This is likely because individuals with autism show no deficits in processing musical affect

(Heaton, Hermelin & Pring, 1999).

There are two major goals when music therapy is employed for individuals with autism:

improving communication/language and improving socio-behavioral skills (Kaplan & Steele, 2005).

For instance, Edgerton (1994) investigated the effect of improvisational music therapy on

communicative behaviors and found gains for musical communicative modalities (tempo, rhythm, form

Empirical Musicology Reviewa Vol. 4, No, 1, 2009



12



and pitch) across ten music therapy sessions. Furthermore, a significant positive relationship was

obtained between musical vocal skills and nonmusical speech production skills, implying that

communication through music bypasses speech and language barriers in individuals with autism. Other

than improvisational music therapy, the use of rhythmic patterns and melodies has been found to be

beneficial for aiding learning and memory in autism. Buday (1995) observed that the number of signed

and spoken words correctly imitated by children with autism was higher using sung rather than spoken

text.

Music as a mode of intervention has yielded beneficial effects for the modification of

behavioral problems. Burleson, Center, and Reeves (1989) obtained results to show that background

music reduced off-task behaviors and facilitated task related performance by children. Orr, Myles, and

Carlson (1998) investigated the effect of rhythmic entrainment on erratic classroom behaviors in their

case study of a girl with autism. Rhythmic entrainment involves the use of music to aid relaxation by

the introduction of externally produced rhythms, designed specifically to re-entrain the body to its

natural rhythmic patterns. They found that rhythmic entrainment helped in reducing problematic

classroom behavior. Focusing on the role of synchronization, Wimpory, Chadwick, and Nash (1995)

used musical interaction therapy in order to synchronize music to interactions between a 3 year old girl

with autism and her mother. The girl showed improvements in eye contact and other forms of social

acknowledgement, and she initiated more episodes of interactive involvement with her mother.

Moreover, a two-year follow up showed that these benefits were sustained.

Social stories are also often employed as a method of choice for individuals with autism.

These short stories are designed to describe social situations extensively so as to provide relevant social

cues that children with autism often fail to pick up in their natural environments (Kuoch & Mirenda,

2003). The aim is to improve awareness and understanding of various social situations and the

appropriate responses to be elicited in such complex situations. Brownell (2002) presented social story

information either musically or by simply reading the story contents to four first- and second-grade

students with the aim of modifying behaviors. The results showed that musically presenting social

stories was a better treatment alternative.

The above mentioned studies show that music therapy is an effective treatment choice for

improving communicative abilities in autistic individuals. This improvement in communication has

been brought about by attempts to improve emotion recognition, eye contact, behavior modification

and so on. Even though music therapy has been used with success, there is still a shortage of theoretical

papers that deal with understanding the neural substrates and cognitive mechanisms underlying the

improvement in such adaptive skills. Towards this end, the current paper will discuss recent findings on

the neuropsychology and neural bases of music to help make the picture clearer. The focus here will be

on understanding why music therapy leads to an improvement of emotion recognition as part of

communicative skills.

A COGNITIVE MODEL OF MUSIC THERAPY FOR ASD

Peretz (2001a, 2001b; Peretz & Coltheart, 2003), Dowling (2001) and Lerdahl and Jackendoff (1983)

have proposed domain specific mechanisms and specialized neural networks devoted to music

processing components principally involving pitch based computations (for instance, tonal encoding of

pitch). Peretz & Coltheart (2003) have emphasized the functional modular nature for tonal encoding of

pitch in music. This module may minimally overlap with language according to Peretz (2006), which is

consistent with a previous finding that pitch variations generate a determinate scale in music but not in

speech intonation contours (Balzano, 1982). Peretz (2006) further proposed that this particular module

in music processing may be innately constrained. The module for tonal encoding of pitch is also

hypothesized to be responsible for the perception of emotions in music (Peretz, 2002). Supporting this

claim, Khalfa and colleagues (2008) showed that tonal variations of pitch support distinctions between

happy and sad musical emotions. They used psychophysiological and psychometric indices of emotions

to investigate the proposed hypothesis. Musical excerpts were created either by removing pitch

variations (the rhythmic version) or by removing both pitch and temporal variations (the beat alone

condition). The results showed that reliable happy/sad distinctions require tonal variations; the

rhythmic version or tempo alone conditions did not support this emotional distinction. The authors

further reasoned that tempo entrainment might exist in tempo alone conditions, but the effects might

disappear when tempo is embedded within a musical context or with rhythm. The results of this study

Empirical Musicology Reviewa Vol. 4, No, 1, 2009



13



are consistent with various other findings that show an advantage for decoding happiness and sadness

over fear and anger in musical stimuli (Bunt & Pavlicevic, 2001; Terwogt & Van Grinsven, 1991).

Initially, it was hypothesized that since there were emotion processing deficits in autism, an

emotionally responsive system for music might also fail to develop for individuals with the disorder

(Peretz, 2001c). However, recent developments in cognitive science have changed the initial theorizing

about the emotionally responsive system for music in autism. Individuals with autism are shown to be

good at detecting pitch from musical tones (Mottron, Peretz & Me′nard, 2000). As previously stated,

individuals with autism also show no deficits in processing affect in music (Heaton, Hermelin & Pring,

1999). Since Peretz (2002) hypothesized that the module for tonal encoding of pitch is responsible for

the perception of emotions in music, it is likely that the ability of individuals with autism to recognize

musical affect is based on the module for tonal encoding of pitch. Although scientific efforts to confirm

this hypothesis have not been initiated, the claim is consistent with evidence that there are separate

emotion recognition modules (Chuang & Wu, 2004), including neuroscientific evidence (Hornak, Rolls

& Wade, 1996). Thus, music can be employed to alleviate emotion recognition deficits in individuals

with autism because tonal pitch differentiation is the primary means by which emotions are perceived.

Some theoretical positions might argue that music therapy works effectively for individuals

with autism due to other benefits that the therapy might confer such as increased attention from the

class teacher, opportunities to use musical instruments (Kern and Aldrige, 2006) or the structure and

predictability of therapeutic sessions (Wigram & Gold, 2006) [2]. Therefore, it is important to

emphasize here that tonal encoding of pitch in musical sessions for autism might have different

advantages. Firstly, Peretz (2002) argues that the synchronization involved in musical activity is an

effective way to promote simultaneous dancing and singing (Brown, 2000). Synchronization in turn

targets mirror neurons (Berrol, 2006) that are also found to be functioning atypically in autism

(Williams, Whiten, Suddendorf & Perrett, 2001). Thus, there might be a different mechanism through

which the tonal variations in pitch bring about effective results. Regardless, the main aim of the current

article is to emphasize that the emotion recognition/distinctions (happy and sad) might be dependent

upon sensitivity to tonal variation.

Along similar lines, Gagnon & Peretz (2000) showed that tonality could be put to different

uses depending upon the task at hand. They obtained results to show dissociation between affective and

non-affective judgments. Their experimental participants showed no hemispheric differences while

judging melodies to be correct or incorrect (non-affective task). On the other hand, when required to

judge the melodies as pleasant or non-pleasant (affective task), the participants did show hemispheric

differences, where pleasant responses were indicative of a left hemisphere predominance and vice

versa. Therefore, the use of tonality is made differently for affective or non-affective purposes.

However, this result is not consistent with the results obtained by Bryden et al. (1982), which did not

reveal hemispheric differences for positive and negative judgments. It is important to note that the

study by Bryden and colleagues involved the manipulation of minor and major modes to yield pleasant

or non-pleasant melodies, whereas the study by Gagnon & Peretz (2000) manipulated tonality.

Therefore, it is quite possible that tonality and other musical cues that are responsible for emotional

identification involve different mechanisms. Understanding the nature of such mechanisms is an

important challenge for future research.

Future research efforts will yield valuable insights if geared towards investigating the

mechanisms through which music therapy brings about desirable results in autism, especially if the

fundamental goal is to improve the recognition of basic emotions such as happiness or sadness. The

current literature identifies tonal variations in pitch as a potential mechanism to investigate. Cognitive

and behavioral studies could be combined with brain imaging tools to understand the brain basis of

such mechanisms, as well as to throw light on the neural processes that undergo changes following the

application of musical intervention.

Formulating a Rehabilitation Program Based on Music

Understanding the cognitive and brain basis of the tonal encoding hypothesis will not be enough for

planning rehabilitation programs unless this effort is combined with the aim of generalizing the results

obtained in the laboratory. But what would be the basis of such generalization? As stated above, Peretz

(2006) hypothesized that the module for tonal encoding of pitch in music may minimally overlap with

language. If this is the case, then can we expect a generalization of effective music lessons to language

in day-to-day functioning? Given that individuals with autism fail to interpret and recognize vocal

Empirical Musicology Reviewa Vol. 4, No, 1, 2009



14



expressions of emotions, a reasonable goal of any intervention would be to improve recognition of

emotions in daily functioning. Then how successful can an intervention program be if it is based on a

module that minimally overlaps with language?

A look at the brain imaging literature may shed light on this issue. Patel (2003) posits

significant overlap between modules for language and music processing and even points to the overlap

in neural structures supporting language and music processing as evidence for the domain

nonspecificity of music and language (Koelsch et al., 2002; Maess et al., 2001). For example, BA 47

has been shown to be associated with the processing of both linguistic and musical structure (Levitin

and Menon, 2003), while BA 44 (Broca’s area) and its right hemisphere homologue have been found to

be associated with the processing of musical syntax (Maess et al., 2001). But it is important to

emphasize here that the existence of overlap between language and music for syntactic processing need

not imply the non-domain specificity of other specific modules in music processing. Consistent with

this claim, Peretz & Coltheart (2003) reviewed an imaging study that showed that regions such as the

middle temporal gyrus, occipital lobe, and parietal lobe were activated during music perception. These

regions are not typically activated in language processing tasks, supporting the claim that music

processing involves various modules, some of which are domain specific for music (Schmithorst,

2005). It should be noted, however, that the significance of these activated neural regions during music

perception is not clear.

At this stage, it might not be possible to discuss generalizations by considering overlaps

between language and music. Generalizations might become more feasible if prosody was brought into

the picture. Speech prosody refers to the musical aspects of speech that are used to convey emotions

(Juslin & Laukka, 2003) and is used in a similar manner across cultures to convey emotions (Bolinger,

1978). Therefore, verbal comprehension or semantics are not necessary to understand emotions

conveyed through speech, thereby facilitating the decoding of emotions even in a foreign language

(Thompson, Schellenberg & Husain, 2004). Even evolutionary theorizing suggests that music, speech,

prosody and facial expressions share common ancestry as spatial-temporal patterns (Dissanayake,

2000) perhaps laying the common ground for a single intervention attempt with favorable effects for

alleviation of emotional processing deficits from speech, prosody and facial expressions.

Pitch and rhythmic information are identified as critical dimensions in both music and prosody

(Patel et al., 1998). Both pitch and rhythmic cues retain their identities in musical tunes across various

manipulations of pitch levels and tempo. In speech, pitch variations provide an important source of

emotional information consistent with the results of Khalfa and colleagues (described above).

Therefore, if music lessons help in decoding emotions from speech prosody then music interventions

could greatly aid in adaptive living.

Thompson, Schellenberg & Husain (2004) investigated the effect of music lessons on the

recognition of emotions from speech prosody. In their study, 6 year old children underwent musical

(keyboard) training for a period of 1 year. After the end of the training period, these children were

better at identifying happy and sad emotions from speech than children who underwent drama, singing

or no art lessons at all for the same time period. The music group was better at the happy-sad

distinction as compared to the fearful-angry distinction. These results imply striking parallels to the

theorizing developed in the present paper. Firstly, the keyboard group was better at the identification of

happy and sad emotions; these two emotions are hypothesized to be dependent upon the tonal encoding

of pitch. Secondly, the keyboard group performed better than the groups taking other forms of training.

These other types of training like drama and singing could also be considered forms of music therapy

but are not as effective as keyboard lessons. Therefore, it is likely that tonal encoding of pitch underlies

the ability to distinguish between happiness and sadness in music. Another noteworthy result of the

study was the equivalent performance by keyboard and drama groups on fearful-angry comparisons.

According to Thompson et al. this may occur because the drama lessons typically involve speech

training and the use of prosody. Thus, it might also be possible that different sorts of music therapy

could be employed for alleviating deficits for different emotional categories, and the tonal variations of

pitch might be a better choice of intervention for focusing on the basic category of happy-sad

expressions; a question that can only be answered with further investigations.

What remains to be seen with future studies on autism is whether music lessons in autism aids

in recognition of emotions from speech prosody for one’s own and foreign language. If taking music

lessons enhances the recognition of emotions by bypassing the language barriers then this will provide

stronger support for the proposal outlined here. Thus, to comprehensively understand the nature of

music therapy and its effectiveness for autism it is important to focus well coordinated research efforts

Empirical Musicology Reviewa Vol. 4, No, 1, 2009



15



in two distinct directions. Research in the cognitive domain should progress with parallel efforts to

evaluate whether the findings of such research can be generalized to various domains of daily

functioning. Unless such endeavors are well planned and coordinated, attempts at applying the findings

of basic research studies will not be beneficial. This holistic approach will not only improve our

understanding of music therapy and musical emotions, but will also yield positive results for planning

rehabilitation programs for children afflicted by autism. Therefore, the need of the hour is greater

cooperation between researchers, parents, policy makers and the community at large. It is hoped that

this manuscript will be a source of inspiration towards this direction.

CONCLUSION

The current article describes the cognitive basis of emotion recognition through music. According to

the model developed here, tonal encoding of pitch provides the basis for recognition of emotions. This

hypothesis explains why individuals with autism lacking the ability to perceive and interpret affective

cues within the social domain are still able to process affect from music. Since individuals with autism

are able to recognize emotions from music but are unable to process emotional information from voice

and faces, it is worth mentioning that a global emotionally responsive system might be possible,

consistent with the findings that there can be separate emotion recognition modules. The tonal

encoding of pitch is further hypothesized to be domain specific and functionally modular in nature,

even though specialized neural networks supporting it have not yet been explored. The present article

lays emphasis on the need to build a rehabilitation program for promoting recognition of emotions in

autism based on intact music modules, and also helps in elucidating the basis of generalization of

results to everyday living.

Several questions await the results of future research and until such answers are forthcoming it

is difficult to come up with a thorough rehabilitation program based on music therapy. First, there have

been few studies in autism that show the effectiveness of music therapy. Second, there is a general lack

of understanding about the mechanisms in music that promote emotion recognition in autism. Future

research efforts that focus on tonal encoding of pitch may help to address such challenges. It is quite

possible that this basic module greatly helps in the recognition of happy and sad emotions in the

musical context. It might not be very effective for promoting the recognition of other emotions and

therefore other types of musical interventions might need to be included as therapy progresses. Tonal

encoding of pitch might have several other advantages rather than just promoting recognition of

emotions. There might be long term effects by the promotion of synchronization that targets the mirror

neurons functioning sub-optimally in autism. Towards this end, it might be helpful to employ brain

imaging tools to study the effect of music therapy on the mirror neurons in the brain. But we also need

to combine it with strong cognitive theorizing.

The next fruitful direction will be to understand the basis of generalizing the effective results

obtained in the therapy sessions within the lab to everyday functioning. One likely candidate in this

regard is to understand the effect of music therapy on the recognition of emotions from speech prosody.

Studies with children have shown that music sessions are effective at promoting recognition of

happiness and sadness as compared to other emotions from speech prosody. The future studies should

concentrate on understanding the mechanisms of such generalizations and it might be likely that, again,

the tonal variations of pitch are responsible for such effects. If such effects are obtained with prosody,

and music lessons are shown to aid in the recognition of happy and sad emotions from prosody in the

context of foreign languages for autism, we could be more confident about this potential mechanism.

Once generalizations for prosody are understood, other elements could be brought in to understand how

benefits obtained in the laboratory could be generalized to facial expressions, for example. Here again

we need combinations of cognitive theories with brain imaging tools to help us understand such

interplays. As music treatment with autism progresses, additional music tools such as group singing

could be incorporated to target the improvement in the recognition of other emotions and the basis of

further generalizations.

NOTES

[1] Email: nehakhetrapal@gmail.com

[2] I thank the editor for bringing up this point.

Empirical Musicology Reviewa Vol. 4, No, 1, 2009



16



REFERENCES

American Psychological Association. (1994). Publication Manual of the American Psychological

Association. 4th edition. Washington, DC: American Psychological Association.

Balzano, G. (1982). The pitch set as a level of description for studying musical pitch perception. In M.

Clynes (Ed.), Music, mind and brain. New York: Plenum Press.

Berrol, C. F. (2006). Neuroscience meets dance/movement therapy: Mirror neurons, the therapeutic

process and empathy. The Arts in Psychotherapy, 33, 302–315.

Bolinger, D. (1978). Intonation across languages. In J. Greenberg (Ed.), Universals in human language,

Vol. II: Phonology. Palo Alto, CA: Stanford University Press, pp. 472–524.

Brown, S. (2000). The “musilanguage” model of music evolution. In N. Wallin, B. Merker & S. Brown

(Ed.), The origins of music. Cambridge (MA): MIT Press, pp. 271–300.

Brownell, M. D. (2002). Musically adapted social stories to modify behaviors in students with autism:

Four case studies. Journal of Music Therapy, 39(2), 117-144.

Bryden, M. P., Ley, R. G. & Sugarman, J. H. (1982). A left-ear advantage for identifying the emotional

quality of tonal sequences. Neuropsychologia, 20, 83-87.

Buday, E. M. (1995). The effects of signed and spoken words taught with music on sign and speech

imitation by children with autism. Journal of Music Therapy, 32(3), 189-202.

Bunt, L. & Pavlicevic, M. (2001). Music and emotion: Perspectives from music therapy. In P. N. Juslin

& J. A. Sloboda (Eds.), Music and emotion: Theory and research. Oxford, England: Oxford University

Press, pp. 181–201.

Burleson, S. J., Center, D. B. & Reeves, H. (1989). The effect of background music on task

performance in psychotic children. Journal of Music Therapy, 26(4), 198-205.

Capps, L., Yirmiya, N. & Sigman, M. (1992). Understanding of simple and complex emotions in non-

retarded children with autism. Journal of Child Psychology and Psychiatry, 33, 1169 – 1182.

Chuang, Z. J. & Wu, C. H. (2004). Multi-modal emotion recognition from speech and text.

Computational Linguistics & Chinese Language Processing, 9, 45-62.

Dissanayake, E. (2000). Antecedents of the temporal arts in early mother–infant interactions. In N. L.

Wallin, B. Merker & S. Brown (Eds.), The origins of music. Cambridge, MA: MIT Press, pp. 388–410.

Dowling, W. J. (2001). Perception of music. In E. B. Goldstein (Ed.), Blackwell Handbook of

Perception. Oxford: Blackwell.

Edgerton, C. L. (1994). The effect of improvisational music therapy on the communicative behaviors of

autistic children. Journal of Music Therapy, 31(1), 31-62.

Fein, D., Lucci, D., Braverman, M. & Waterhouse, L. (1992). Comprehension of affect in context in

children with pervasive developmental disorders. Journal of Child Psychology and Psychiatry, 33,

1157—67.

Empirical Musicology Reviewa Vol. 4, No, 1, 2009



17



Fombonne, E. (2005a). The changing epidemiology of autism. Journal of Applied Research in

Intellectual Disabilities, 18, 281-294.

Fombonne, E. (2005b). Epidemiology of autistic disorder and other pervasive developmental disorders.

Journal of Clinical Psychiatry, 66, 3-8.

Gagnon, L. & Peretz, I. (2000). Laterality effects in processing tonal and atonal melodies with affective

and nonaffective task instructions. Brain and Cognition, 43, 206-210.

Heaton, P., Hermelin, B. & Pring L. (1999). Can children with autistic spectrum disorders perceive

affect in music? An experimental investigation. Psychological Medicine, 29, 1405-410.

Hornak, J., Rolls, E. T. & Wade, D. (1996). Face and voice expression identification in patients with

emotional and behavioural changes following ventral frontal lobe damage. Neuropsychologia, 34, 247-

261.

Juslin, P. N. & Laukka, P. (2003). Communication of emotions in vocal expression and music

performance: Different channels, same code? Psychological Bulletin, 129, 770–814.

Kanner, L. (1943). Autistic disturbances of affective contact. The Nervous Child, 2, 217–250.

Kaplan, R. S. & Steele, A. L. (2005). An analysis of music therapy program goals and outcomes for

clients with diagnoses on the autism spectrum. Journal of Music Therapy, 42, 2-19.

Kern, P. & Aldrige D. (2006). Using embedded music therapy interventions to support outdoor play of

young children with autism in an inclusive community-based child care program. Journal of Music

Therapy, 43(4), 270-294.

Khalfa, S., Roy, M., Rainville, P., Bella, S. D. & Peretz I. (2008). Role of tempo entrainment in

psychophysical differentiation of happy and sad music? International Journal of Psychophysiology, 68,

17-26.

Koelsch, S., Gunter, T. C., Cramon, D. Y. V., Zysset, S., Lohmann, G. & Friederici, A. D. (2002).

Bach speaks: A cortical “Language-Network” serves the processing of music. NeuroImage, 17, 956–

966.

Kuoch, H. & Mirenda, P. (2003). Social story interventions for young children with autism spectrum

disorders. Focus on Autism and Other Developmental Disabilities, 18(4), 219-227.

Lerdahl, F. & Jackendoff, R. (1983). A generative theory of tonal music. Cambridge, MA: MIT Press.

Levitin, D. J. & Menon, V. (2003). Musical structure is processed in “language” areas of the brain: A

possible role for Brodmann Area 47 in temporal coherence. NeuroImage, 20, 2142–2152.

Maess, B., Koelsch, S., Gunter, T. C. & Friederici, A. D. (2001). Musical syntax is processed in

Broca’s area: an MEG study. Nature Neuroscience, 4, 540-545.

Mottron, L., Peretz, I. & Me′nard, E. (2000). Local and global processing of music in high-functioning

persons with autism: Beyond central coherence? Journal of Child Psychology and Psychiatry, 41, 1057

–1165.

Orr, T. J., Myles, B. S. & Carlson, J. K. (1998). The impact of rhythmic entrainment on a person with

autism. Focus on Autism and Other Developmental Disabilities, 13(3), 163-166.

Empirical Musicology Reviewa Vol. 4, No, 1, 2009



18



Patel, A. D. (2003). Language, music, syntax and the brain. Nature Neuroscience, 6, 674–681.

Patel, A. D., Peretz, I., Tramo, M. & Labrecque, R. (1998). Processing prosodic and music patterns: A

neuropsychological investigation. Brain and Language, 61, 123–144.

Peretz, I. (2001a). The biological foundations of music. In E. Dupoux (Ed.), Language Brain Cognitive

Development: Essays in honor of Jacques Mehler. Cambridge: The Bradford Book: MIT Press.

Peretz, I. (2001b). Brain specialization for music. New evidence from congenital amusia. Annals of the

New York Academy of Sciences, 930, 153–165.

Peretz, I. (2001c). Listen to the brain: a biological perspective on musical emotions. In P. N. Juslin & J.

A. Sloboda (Ed.), Music and Emotion, Theory and Research. Oxford University Press.

Peretz, I. (2002). Brain specialization for music. The Neuroscientist, 8, 374-382.

Peretz , I. (2006). The nature of music from a biological perspective. Cognition, 100, 1-32.

Peretz, I. & Coltheart, M. (2003). Modularity of music processing. Nature Neuroscience, 6, 688–691.

Schmithorst, V. J. (2005). Separate cortical networks involved in music perception: preliminary

functional MRI evidence for modularity of music processing. NeuroImage, 25, 444– 451.

Terwogt, M. M. & Van Grinsven, F. (1991). Musical expression of moodstates. Psychology of Music,

19, 99-109.

Thompson, W. F., Schellenberg, E. G. & Husain, G. (2004). Decoding speech prosody: Do music

lessons help? Emotion, 4(1), 46-64.

Volkmar, F. R. & Pauls, D. (2003). ‘Autism.’ The Lancet, 362, 1133–1141.

Williams, J .H. G., Whiten, A., Suddendorf, T. & Perrett, D. I. (2001). Imitation, mirror neurons and

autism. Neuroscience and Biobehavioral Reviews, 25, 287-295.

Wigram, T & Gold, C. (2006). Music therapy in the assessment and treatment of autistic spectrum

disorder: clinical application and research evidence. Child: Care, Health and Development, 32(5), 535-

542.

Wimpory, D., Chadwick, P. & Nash, S. (1995). Brief report: Musical interaction therapy for children

with autism: An evaluative case study with two-year follow-up. Journal of Autism and Developmental

Disorders, 25(5), 541-552.



献花(0)
+1
(本文系mc_eastian首藏)