Singing the Words: A Multidisciplinary Investigation of the Musical and Linguistic Melody Junior Independent Project Singing the Words: A Multidisciplinary Investigation of the Musical and Linguistic Melody Hana Shin Department of Music Princeton University This paper represents my own work in accordance with University regulations. A Multidisciplinary Investigation of the Musical and Linguistic Melody 1 Introduction “The sounds as they appear to you are not only different from those that are really present, but they sometimes behave so strangely, as to seem quite impossible” (Deutsch, 2003). These are the words from a famous demonstration by psychologist Diana Deutsch, which repeats the phrase, “sometimes behave so strangely” over and over again until it indeed starts to behave strangely. After hearing it a few times in isolation, this speech phrase magically becomes a melody, and one can never again hear it the same way. In fact, utterances in our daily lives seem to possess some “musical” qualities, a certain “melody of speech.” People have given this phenomenon various names, from “melody of language” and “Sprachmelodie,” to “chanson de la parole” and many others (Monrad-Krohn, 1957, p. 326). Researchers have only started systematically investigating the science behind it in recent years, although various traditions in the arts – poetry, songwriting, music- have for a long time explored it under the name “prosody.” There seems to be something captivating about the rhythmic flow, stress patterns, and rhymes of speech, streams of vocal sounds nicely organized into orderly units and structures, that is both utilitarian and aesthetically pleasing. In fact, we experience the influence of prosody everyday, as we engage in and expect to hear from others certain rhythmic and intonational ways of speaking, the unspoken rules which are nonetheless essential in conveying proper meanings and creating a rich, “normal” conversational context that we often take for granted. Beyond these artistic and linguistic uses of prosody, a number of musicologists and researchers have gone so far as to claim that the prosody of the language they speak can influence composers’ musical styles. For example, musicologist Ralph Kirkpatrick A Multidisciplinary Investigation of the Musical and Linguistic Melody 2 has remarked that “[b]oth Couperin and Rameau, like Fauré and Debussy, are thoroughly conditioned by the nuances and inflections of spoken French” (as cited in Abraham, 1974, p. 83). Neuroscientist Aniruddh Patel (2005) has recently demonstrated that this claim may not be so outrageous, as his recent study using the “prosogram” model, which allows tracing of the melodic contour of sentences, has shown that linguistic prosody is often reflected in the musical melodies of at least two cultures, English and French. At the same time, music therapists have long intuited about the benefits of singing in improving speech. Not only have people believed for hundreds of years that singing could improve stammering, pronunciation, and quality of voice in speech production, but the Renaissance composer William Byrd wrote in his 1588 edition of Psalms, Sonnets, and Songs a preface section advocating the use of music “to strengthen the respiratory system, reduce fluency and voice disorders, and improve articulation” (as cited in Thaut, 2005, p. 165). Since melodies in songs can bring out and exaggerate the prosodic elements of speech, it seems not too much of a stretch to imagine that there is an intimate, biological link between the musical melody and the linguistic melody, which may indeed allow music to influence speech output. Exploring this mysterious relationship between prosody and melody is the topic of the current study, and there is much to be uncovered. Where exactly does the boundary lie? With how much confidence can we say that prosody, which does display certain “musical” qualities like rhythm and intonation, is the linguistic equivalent of the musical melody? Are they, in fact, sufficiently associated at the biological level as to produce any synergistic therapeutic effects? To address these questions, the elements, functions, and neurological basis of prosody and melody will be investigated, synthesizing current A Multidisciplinary Investigation of the Musical and Linguistic Melody 3 knowledge and debates in musicology, psycholinguistics, and neuroscience. A closer examination of prosody and melody in context of related disorders such as aphasia, amusia, and aprosody, will further clarify the complex biological interactions between the two and bring us a step closer to demystifying the magic of music as a therapeutic tool. Defining the Terms: What Are Melody and Prosody? Before continuing with the task at hand, some terminology needs to be clarified, and some concepts operationalized. The words “music” and “language” are often used, especially in scientific investigations, either without reference to any particular culture or time, or under the assumption of a vastly simplified “Western” culture. This study will be adopting an approach closer to the former, focusing on the concept or phenomenon of music or language, especially the moment-to-moment temporal, performative aspect that sets them apart from, for example, the static notated form. The words “melody” and “prosody” will be operating within this framework as well, referring to sequences of pitches and rhythms occurring in time, which are meaningful within the context of music and language, respectively. The dictionary definition of “melody” is “pitched sounds arranged in musical time in accordance with given cultural conventions and constraints” (Ringer, n.d.). This definition is broad enough to account for the vast diversity of musical styles found across different cultures, although for the purposes of the current investigation we need a bit more precise characterization. What are the elements of melody, and where do they fit in in the overall picture of music as a whole? Literature on melody, especially in the field of music psychology, utilizes several different systems for organizing components of music, and having a closer look at some of these systems provides some insight into how one A Multidisciplinary Investigation of the Musical and Linguistic Melody 4 might go about defining melody. One such approach is the physiological comparison model used by Schneck and Berger (2006) in their book, The Music Effect. Here, they draw attention to the parallels between elements of music and physiological mechanisms that perceive, process, and are influenced by those elements. Naturally, their categories are largely technical and devoid of aesthetic or philosophical judgments. These components are: periodicity, melody, harmony, dynamics, timbre, and form. An interesting feature of this system is the grouping of pitch and rhythm into the same category of “periodicity,” defined as the “tendency of an event to recur at regular intervals,” or well-defined cycles. This naturalistic view emphasizes the repetitive and pattern-derived nature of rhythm and frequencies of pitch as the basis of our physiological interaction with music. Such perspective sets apart “melody” as an inherently different concept. It may contain some periodic elements like recurring phrase structures or patterns, but at least in Schneck and Berger’s definition, it concerns, very specifically, the relations between the pitches and rhythms of a sound sequence, as well as the overall shape they create together, rather than the nature of each individual tone. It unfolds and transforms linearly over time and operates in a way fundamentally different from that of pitch and rhythm. In Music, Language, and the Brain, on the other hand, Patel (2008) divides music in a more hierarchical fashion, starting his discussion with the most concrete sound components (pitch and timbre) and rhythm and eventually moving towards broader concepts like meaning and evolutionary significance. In this view, melody fits in the middle of this spectrum, encompassing the sound elements and rhythm but not quite as broad as musical syntax, meaning, and evolutionary significance. Here, too, melody is A Multidisciplinary Investigation of the Musical and Linguistic Melody 5 considered separate from the elements it is composed of, and its global nature –the idea of being the relationship between discrete elements- is again emphasized. Patel characterizes prosody in a similar vein, as encompassing the relations between pitch and rhythmic elements in speech, and thus draws a direct parallel between melody and prosody. Monrad-Krohn (1957), one of the first researchers to explore the neuroscience of prosody, also mentioned in his early article that pitch, rhythm, and stress patterns are the three elements of prosody (p. 326). A few decades later, Ross and Medulam came up with the following definition for prosody: “the melodic line produced by the variation of pitch, rhythm and stress of pronunciation that bestows certain semantic and emotional meaning to speech” (Ross & Medulam 1979, p. 146). The “variation of pitch” is also referred to as “intonation” and seems to be a direct counterpart to pitch variation in melodies. “Variation of rhythm,” likewise, can be described in both melody and prosody as the “grouping,” or “clustering of adjacent elements with respect to temporal proximity into larger units,” and “meter,” or the periodic temporal-accentual scheme” (Patel, Peretz, Tramo, & Labreque, 1998, p. 125). “Stress of pronunciation” is not explicitly addressed in usual discussions about music, perhaps because the acoustic features involved in differentiation of stress patterns already include both pitch and rhythmic elements, as well as other elements like dynamics (putting a stress on a syllable may raise the pitch, lengthen the duration, and increase the volume, all at the same time). Therefore, “stress” seems to be a somewhat redundant feature, and may be best characterized as adding more articulatory elements to the existing pitch and rhythm. A Multidisciplinary Investigation of the Musical and Linguistic Melody 6 Synthesizing ideas from the three distinct approaches, the common elements of melody and prosody can be characterized as follows: “the pitch and rhythmic relations between sounds in sequence, as well as the global contour produced by those relations, which is further characterized by articulation, dynamics, and phrasing structure.” This partial definition is by no means exhaustive, and it immediately calls for clarification of words like “sequence” and “sounds.” For example, would a random sequence of pitches, or a sequence of noises, count as melodies? Again, aesthetic and stylistic judgments will be minimized here, although a common solution to this question seems to be to assume the conventional rules of stepwise motion, variety in direction, tension and release, etcetera, as characterizing “good” melodies. An entire literature could be devoted to exploring what makes a “good” melody, as it is a question composers and theorists have been grappling with for the past centuries. Since such discussion is beyond the scope of this essay, suffice it to say that something like Mary Had a Little Lamb or the opening theme of Mozart’s Symphony No.40 are closer to what people would generally understand as “melodies,” whereas a random computer-generated sequence of pitches within a range of several octaves would be less likely to be so. In addition, discussions in the following sections will be involving melody and prosody more as forms of acoustic phenomena, to be examined mostly in conceptual, acoustical, and biological, but not aesthetic, terms. Characterizing the Functions: What Do Melody and Prosody Do? Perhaps the phrase “melody of speech” makes intuitive sense because of these basic similarities in what makes up melody and prosody. In fact, in the broad acoustic spectrum, both are in a special place, tightly linked to the concept of time. By nature, A Multidisciplinary Investigation of the Musical and Linguistic Melody 7 both melody and prosody are crucially grounded on the idea of transience and progression over time. Of course, other acoustic features are also fundamentally dependent on time, by virtue of the fact that acoustic information is presented within a certain time frame. However, melody and prosody are especially closely tied to it, as their essential feature is the relations, or transitions that unfold and transform over time. Shortterm memory, which allows one to hold the old information while receiving new information so that they can be synthesized to produce a perception of a global structure, plays an important role for mental representation of both melody and prosody, which is to be discussed in due course. However, there is an important, functional distinction between melody and prosody. That is, “melodies are designed to be aesthetically interesting objects” while “spoken pitch patterns perform [linguistic] functions without themselves becoming a focus of attention for ordinary listeners” (Patel, 2003). One can easily see that “it is common to find oneself humming a catchy musical tune, while one rarely notices or explicitly remembers spoken pitch patterns” (Patel, 2006). Melody is arguably the most salient feature in many musical contexts, “the tune” of a piece (this is not to overlook pieces that instead seek to explore timbre, instrumentation, and other aspects of the sound world, like Schoenberg’s Farben; but the very existence and rarity of such pieces testify to the melody-oriented conception of music that still prevails in many musical cultures, at least in the West). It is hard to say what the function of melody is in the context of music, though it surely can be used to create tension and release, a sense of direction, and a driving momentum, contributing to the expressive quality of a piece. In other words, A Multidisciplinary Investigation of the Musical and Linguistic Melody 8 through its ups and downs, and syncopations and pauses, it can express its own rhetoric and purpose. On the other hand, prosody is easier to describe in terms of its function in aiding the communication of meaning. Monrad-Krohn (1957) characterized prosody as the “third element of speech” along with vocabulary and grammar, suggesting its role in conveying extra-linguistic meanings (p. 327). He and his contemporaries, as well as some later researchers, described the function of prosody in communication as “marking the boundaries of structural units, distinguishing pragmatic categories of utterance, and signaling focus” (Lehiste, 1973; Beckman & Pierrehumbert, 1986; Bolinger, 1989; Price, Ostenodrf, Shattuck-Hufnagel, & Fong, 1991), although more recent studies have shown that semantics and syntax both rely so heavily on prosody that pseudowords, delexicalized sentences, hummed intonations, computer muffled tones, and even pitch contours of speech are often enough to convey meaning (Pannekamp, Toepel, Alter, Hahne, & Friederici, 2005; Thompson, 2008). In addition, prosody provides rich information about the context in which speech takes place, including the language or dialect, gender, age, and emotion of the speaker (Bolinger, 1989). Even infants have been found to babble with the prosody of their native language (Hallé, de Boysson-Bardies, & Vihman, 1991). Therefore, prosody is undoubtedly an important feature of verbal communication. Once again, though, it conveys meanings for language. As demonstrated above, intonation alone may be enough to deliver the meaning of speech, but it is exactly this linguistic content that it delivers, not a meaning of its own, derived from its pitches and rhythms. Now, it may be said that prosody does have a capacity to convey emotional A Multidisciplinary Investigation of the Musical and Linguistic Melody 9 contents independently of its linguistic relevance. Monrad-Krohn (1957) believed that the origin of the use of prosody in communication probably goes back to the primitive times, as even animals seem to make use of acoustically distinct features to display a wide range of emotions: Even if one does not accept Schielderup-Ebbe’s postulate that the dog has nine and the hen ten different kinds of expressive sounds, anybody who is sufficiently acquainted with dogs and chickens will know that they can express themselves to a certain extent by prosodic means (cf. the dog’s high-pitched bark as a sign of glee and the low-pitched bark as a sign of hostility –and the cock’s (or the chicken-mother’s) syncopated and staccato calling to the hens (or the chicken) when a nice bit of food has been found- both definitely semantic (or propositional) expressions). (p. 327) In addition, researchers have pointed out the parallels between affective prosodic cues in speech and similar cues in music (Huron, 2011). For example, music and speech that are commonly judged as representing “sadness” both tend to display lower pitch, smaller pitch movement, lower volume, slower speed, more mumbled articulation, and darker timbre, and indeed some argue that the reason these features can convey sadness through music in the first place is because they mimic the appropriate physical responses, including the prosodic elements, that a person feeling the emotion would be likely to display (Huron, 2011). Prosody alone, then, seems able deliver affective meanings to a certain extent. However, even then, it is often so tightly tied to the linguistic content it carries along with the affective content, that it is hard to say prosody has the same degree of autonomy as musical melody in conveying a meaning of its own. A Multidisciplinary Investigation of the Musical and Linguistic Melody 10 Uncovering the Mechanisms: How Is Melody Processed? We have seen that melody and prosody share the basic “musical” elements yet differ in their role in conveying meaning. How do they, then, relate in biological terms? First, we will examine the neurological basis of melody. At the broad level, Schneck and Berger’s description of the auditory scanning and tracking processes seems to provide a conceptual understanding of how the auditory apparatus generally functions. Based on this model, auditory scanning is “brain and auditory system’s ability to accurately scan the sequential flow of melodic syntax in order to receive, perceive, and discriminate the acoustic information in a one-to-one correspondence with the way it is being generated and transmitted” (Schneck & Berger, 2006, p. 178). Our ears are constantly scanning the “acoustic landscape” (think of a dog’s ears), resting on certain “focal points,” which are then held in short-term memory while the next part of the stream of sounds is continuously received (Schneck & Berger). Auditory tracking, primarily a responsibility of the hippocampus, is then able to tag the auditory input in time to keep track of their spatial and temporal identity. Schneck and Berger remark that, “[i]n effect, the brain of the music listener is actually re-composing a composition the auditory system...and brain have essentially de-composed into component parts” like frequencies, amplitudes, and other acoustic features (p. 180). The information thus scanned and tracked is dispatched to “information-specific sites” to be further processed based on the sensory modality (is it auditory or visual information?) as well as temporal and spatial order. The brain then integrates all of this information, creating a whole, unified auditory sensation. The various steps involved in the process all happen very quickly, but subtle differences in speed, short-term memory capacity, and other physiological factors can result in A Multidisciplinary Investigation of the Musical and Linguistic Melody 11 discrepancies in the way individuals perceive the same stream of sounds, and in more extreme cases, an inability or lag in processing auditory input. This also provides some support for the aforementioned idea that Mary Had a Little Lamb makes a “better” melody than a random sequence of notes, since the tracking system can function more efficiently when it is not jumping all over the place, trying to chase after the notes. We will next consider Stewart’s (2011) model, which proposes, in more biological detail, a series of steps involved in the perception and cognition of melody. At the lowest level, the ascending auditory pathway, which starts at the choclea in the ears and culminates in the auditory cortex of the brain, encodes the information on individual pitches and differences between adjacent tones (Plack, 2005). The right secondary auditory cortex is then thought to work in conjunction with the short-term memory system to extract the contour of the incoming pitch stream (Stewart, 2011). How exactly contour is represented in this system is not clear, although information on pitch direction and intervals, represented by frequency ratios between successive notes, appear to help provide information on contour (Zatorre & Berlin, 2001). Evidence also points to the existence of a higher-level control that assesses the global shape of the stream, beyond the local one-to-one comparisons of individual notes (Stewart, 2011). Even though it is widely believed that neural processing of music is largely a responsibility of the right hemisphere (the non-dominant hemisphere in right-handed individuals) of the brain, and language is governed by the dominant, left hemisphere, both right and left sides of the temporal cortex seem to be involved in evaluating information on pitch intervals. Here, we begin to see a possible neurological connection between melody and speech processing. This link is made even stronger by the fact that the inferior frontal gyrus A Multidisciplinary Investigation of the Musical and Linguistic Melody 12 (IFG) in the right hemisphere, whose counterpart in the left hemisphere is essential in language comprehension and production, seems to be involved in the “anchoring” process, which allows comparing of the incoming pitches to a stable tonal reference, drawn from stored knowledge of hierarchical pitch relationships in long-term memory (Janata et al., 2002). Note that this process is essential in musical “comprehension,” or making sense of the relations between the pitches of an input sound stream, serving as a counterpart to the roughly equivalent function in the language domain. This hemispheric parallel will soon be discussed in more detail. A common method used in neuroscience to identify the specific brain regions associated with particular functions is to look at brain damage data and related deficits. Under this paradigm, amusia, a musical deficit, provides valuable insight into the neural mechanisms involved in melody processing. Amusia is generally defined as “disorder in the perception and production of musical tones” (Jiang, Hamm, Lim, Kirk, & Yang, 2010), and can be divided into the congenital type and the acquired type. Congenital amusia, also commonly referred to as “tone-deafness,” affects about 4% of the population (Peretz & Hyde, 2003) and seems to be influenced at least partially by genetic inheritance, as suggested by twin, pedigree, and DNA studies (Stewart, 2011). Acquired amusia occurs as a result of brain damage, and is especially common after strokes (Sarkamo et al., 2009). Both types of amusia can widely vary in intensity and symptoms, but the general form of the disorder involves impairment in pitch discrimination, resulting in the inability to recognize familiar melodies, detect pitch violations in melodies, judge whether two melodies are the same, and many other manifestations (Phillips-Silver et al., 2011). It could involve the receptive domain, responsible for the A Multidisciplinary Investigation of the Musical and Linguistic Melody 13 perception and recognition of music (or even musical notation) (Bautista & Clampetti, 2003), or the motor or expressive domain, associated with the production of music either through instruments or voice, or through notation (Sacks, 2007; Bautista & Clampetti, 2003), or both. Classification of symptoms in amusia is further complicated by the fact that there is a degree of dissociation in the processing of pitch, rhythm, and emotional content in music (Sacks, 2007), and amusia may include impairment in any combination of these, as well as added deficits in musical memory and recognition (Pearce, 2005). Naturally, current neurological data on amusia are very complicated and sometimes conflicting, but the vast majority of the literature seems to agree on the crucial involvement of the inferior frontal gyrus (IFG) (mentioned briefly above in context of its role in tonal referencing and a corresponding linguistic region in the left hemisphere) and the superior temporal gyrus (STG). In addition, it has been shown that unilateral lesion (damage in one of the two hemispheres) is sufficient to lead to amusia, and it is widely believed that the hemisphere opposite to the one that serves language (again, generally the left hemisphere in right-handed individuals) is involved in music processing (Piccirilli, Sciarma, & Luzzi, 2000). Even though some evidence to the contrary has been suggested (Piccirilli, Sciarm, & Luzzi, 2000), the current model as it stands proposes that language and music are processed at least somewhat independently in their respective hemispheres, though in roughly symmetrical regions like the IFG and the STG. One caveat to this is that rhythmic aspects of music seem to rely less on these regions, and it is in fact possible to have selective impairment in pitch processing without rhythmic deficits (Murayama, Kashiwagi, Kashiwagi, & Mimura, 2004) and vice versa (Phillips-Silver et al., 2011). This further complicates our model, as it has been shown that the right A Multidisciplinary Investigation of the Musical and Linguistic Melody 14 temporal auditory cortex is responsible for temporal segmenting of ongoing sequence of music, whereas the left temporal auditory cortex is involved in temporal grouping of the segments to understand the overarching beat or meter structure (Di Pietro, Leganaro, Leeman, & Schnider, 2004). In fact, rhythm seems to draw from a wide variety of other structures as well, including the cerebellum (Sakai et al., 1999) and the motor cortex (which accounts for the coordinated nature of rhythm) suggesting that the neurological processing of “melody” as we have defined it, involves complex interactions between many parts of the brain in both hemispheres. At least in terms of pitch contour processing, though, the IFG and STG regions really do seem to play an important role. Abnormal STG is associated with deficit in pitch perception and other musical characteristics, whereas damage to the IFG is associated with deficits in more cognitive aspects of processing such as memory, which is essential in holding information on one tone long enough to compare it to the next one, as well as internally representing the overall contour and structure (Sarkamo et al., 2009). Recent studies have suggested that the integration of information from the two regions (and more broadly, the frontal and temporal lobes where they reside), is the key to successful processing of melody. This is accomplished through the arcuate fasciculus (AF), a large fiber bundle that connects the two areas. This area seems to be reduced in volume in amusic brains (Stewart, 2011). Furthermore, it has also been shown that amusic brains display abnormal cortical thickness in the IFG, which may have compromised normal development of the fronto-temporal pathway, leading to inadequate communication between the two lobes (Hyde et al., 2007). Some researchers have even proposed that it is not the perception of the pitches themselves that is impaired in amusic A Multidisciplinary Investigation of the Musical and Linguistic Melody 15 patients (with some evidence that amusic brains can actually track fine pitch difference, as small as quarter-tones), but rather the integration of perceived pitches with pitch knowledge, which accounts for the patients’ lack of confidence in trusting their own perceptual experiences and results in a behavioral failure to discriminate pitch (Peretz, Brattico, Jarvenpaa, & Tervaniemi, 2009). Overall, neurological processing of melody, especially in terms of pitch, is largely attributable to the IFG and STG regions of the nondominant right hemisphere, as well as a proper communication between them. Uncovering the Mechanisms: How Is Prosody Processed? How does the science behind prosody compare to that of melody? Before delving into the neurological mechanisms of prosody processing in the brain, it is worth revisiting and elaborating on the general characteristics of prosody. Although, at the basic level, prosody has the pitch relations and rhythmic components that correspond to those of melody, it displays some fundamental differences in terms of its function and usage. Most notably, as mentioned above, since the primary significance of prosody is in facilitating the delivery of something other than itself, the organization of its pitch and rhythmic information is more loosely constructed. In other words, in contrast to musical melodies, which are objects of intrinsic aesthetic value, prosody in normal speech is “aesthetically inert” (Patel, 2008, p. 183) and therefore does not operate under any particular scales, tonal centers, meters, or other unifying musical structures. In fact, relative terms like “high,” “middle,” and “low” often suffice to describe the intonational patterns of a sentence, as in the IPO approach utilized in much of the research in the field, which takes the fundamental frequencies of speech and simplifies, or “standardizes” them into the three categories (Patel, 2008, p. 213-4). Whether our mental representation of A Multidisciplinary Investigation of the Musical and Linguistic Melody 16 prosody makes use of a similar approach is a question to be explored further, but it is clear that prosody and melody differ in terms of their principles of organizing the raw sound materials. At the perceptual level, too, prosody displays several unique features that are not found in the domain of music perception. One such phenomenon is “declination,” “a gradual lowering of the baseline pitch and narrowing of pitch range over the course of an utterance” (Patel, 2008, p. 184). It may have to do with the physical decline in air pressure that vibrates the vocal folds, and thus does not seem to have a direct equivalent in the perception of production musical melodies. Interestingly, listeners of speech seem to take this expectation into account when they are perceiving speech, adjusting their judgment of tone equivalence depending on the temporal position of the sound they are listening to (Patel, 2008, p.184). This not only testifies to the extent to which our sensations can already be “tinkered with” at the perceptual level, but also demonstrates how differences in function can be served by differences in the way melody and prosody are perceived. Another interesting example is the “perceptual transformations” that underlie our perception of speech. That is, there is a significant degree of simplification that takes place between the input of raw fundamental frequencies and complicated, continuous fluctuations among them (traceable and recordable thanks to modern technology), and the relatively discrete tones that we perceive as making up the tones of prosody. Details of these transformations are beyond the scope of this essay, but together they set up a series of thresholds that determine: a) how a stream of continuous sound is to be segregated into syllables, and b) how much fluctuation of difference in pitch and duration is needed before certain tones are perceived as one discrete tone as opposed to A Multidisciplinary Investigation of the Musical and Linguistic Melody 17 another discrete tone (Patel, 2008, p. 214-5). The result is something resembling more of a musical melody, often with nice stepwise motions, although again, prosody operates under a completely different set of rules, if any, that do not rely on the same scalar and tonal principles that melodies generally follow. Initial insights into the neurological basis of prosody processing have come from a series of anecdotes and early observational studies. Most notably, Ross and Medulam reported in their 1979 article the case of two patients with lesions in the right hemisphere, who could speak but only with monotonous voices “that were devoid of the prosodicaffective qualities of speech” (Carroll, 1996, p. 6). From this early study, it was concluded that the right hemisphere is key to properly functioning prosody in speech, and its possible link to music, another right-hemisphere function, was suggested. More recent studies have qualified this initial claim, characterizing the neurology of prosody as “a mosaic of multiple local asymmetries” and “a large-scale spatially distributed network in both hemispheres” (Tong et al., 2005). Prosody presents an interesting tension between two different forces, as it is a linguistic function yet maintains some ties to music. It has been suggested that the right hemisphere is responsible for the emotional interpretation of speech phrases, which may have closer links to music, and the left hemisphere is in charge of linguistic understanding of prosody (Pell, 2006). It is also believed that the pitch contour information of prosody is processed by the left hemisphere along with other language functions, in contrast to the pitch contour information in musical melodies, which, as we have seen, is largely controlled by the right hemisphere (Tong et al., 2005). But as in the case of melody, proper perception and production of prosody requires collaboration of both hemispheres. (Mitchell & Ross 2008). A Multidisciplinary Investigation of the Musical and Linguistic Melody 18 As with amusia, cases of aprosody (or aprosodia), defined as the “absence, in speech, of rhythm and the normal variations in pitch and stress” (MediLexicon, n.d.), provides valuable insight into the neurological mechanisms governing prosody processing. Aprosody was originally thought to be a deficit of affective perception, more closely linked to emotions, although recent studies have demonstrated some fundamental deficits in pitch and temporal elements of speech that seem to underlie the disorder (Van Lancker & Sidtis, 1992). Again, aprosody can involve damage to either of the two hemispheres, (Patel et al., 1998) and as expected, emotional aprosody seems to involve a more consistent pattern of damage to the right hemisphere, especially in the distributed fronto-temporal-parietal network, while the lateralization pattern for linguistic prosody is less clear (Rohrer et al., 2010). The vast complexity of neural patterns, as well as the multi-dimensional nature of prosody (with associations in the emotional, linguistic, and musical domains) suggests that its underlying neurological mechanism may best be characterized in context of other disorders such as amusia and aphasia. In fact, it is not uncommon for aprosody to accompany aphasia, a deficit in language functions. Like amusia, and perhaps even more so, aphasia displays a complicated pattern of symptoms and neurological associations, and its general definition is all-encompassing: “impaired or absent comprehension or production of, or communication by, speech, reading, writing, or signs, caused by an acquired lesion of the dominant cerebral hemisphere” (MediLexicon, n.d.). About 1 in 272 Americans are said to suffer from some form of aphasia (Norton, Zipse, Marchina, & Schlaug, 2009), although only 60% of patients fall under any existing classifications (Kolb & Whishaw, 2003). These categories are commonly distinguished by symptom A Multidisciplinary Investigation of the Musical and Linguistic Melody 19 types or affected brain areas, and two of these categories, Broca’s aphasia and Wernicke’s aphasia, will be discussed briefly in the following section. The former is a type of nonfluent, expressive aphasia, or deficits in the productive aspect of speech. In fact, Broca’s aphasia is characterized by “slow, laborious, non-fluent speech” but often good auditory verbal comprehension (American Speech-Language-Hearing Association, 2012). It is named after an important language region in the brain, called “Broca’s area,” which is located in the IFG of the dominant hemisphere (again, this means left hemisphere in most right-handed individuals, and the opposite side of the region responsible for melody processing). This region is also part of the motor association cortex, and is said to be involved in the dorsal “sound to action” pathway in language processing, suggesting an important role of precise motor control in speech production. In contrast, Wernicke’s aphasia is often characterized as fluent and receptive, because patients with this deficit seem to display fluent yet nonsensical speech. They also have difficulty comprehending others’ speech as well as mistakes in their own speaking. This form of aphasia is associated with Wernicke’s area in the posterior section of the STG of the dominant hemisphere and is thought to be part of the ventral, “sound to meaning” pathway (Johansson, 2010). Aprosody is more commonly associated with Broca’s aphasia, as important aspects of prosody, such as speech timing, are influenced by motor control (Marotta, Barbera, & Bongioanni, 2008). In addition, Broca’s aphasics have often been observed with dysprosodic speech output, characterized by impairment in melodic modulation, isochrony of syllables, and alteration of speech timing (Marotta et al., 2008). In particular, “a general flattening of the pitch contour” is found in these aphasic patients, as A Multidisciplinary Investigation of the Musical and Linguistic Melody 20 well as a reduction in the melodic range, resulting in overall monotonous speech (Marotta et al., 2008). In addition, an abnormal rising at the end of utterances is also found, which is thought to derive from a general difficulty in programming and coordinating linguistic performance (Sarno, 1998). An interesting effect of rhythm in particular, is an increase in the number and duration of pauses, resulting in hyper-segmentation of speech and a mismatch between intonational phrases and syntactic or pragmatic groupings that would normally be synchronized (Marotta et al., 2008) Therefore, these rhythmic and intonational impairments can work hand in hand to exacerbate the deficits already faced by Broca’s aphasics. Cross-Relationships: How Are Melody and Prosody Related? We have seen that the biological basis of prosody is most clearly seen through aprosody and its cousin aphasia, and similarly, the mechanisms underlying melody processing is largely revealed by examining amusic brains. However, some interesting cross-relationships exist between these seemingly separate entities, and they provide a clearer picture of how melody and prosody might be related at the neurological level. The first such relationship to be examined is that of amusia and aprosody, and the prosodic function in amusic patients in general. Although amusia and aprosody are separate disorders with distinct sets of characteristic symptoms, the two have often been observed in conjunction. However, the pattern and degree of such concurrence vary widely across patients, and a study by Patel et al. (1998) made a side-by-side comparison of two such cases. One amusic patient displayed extensive damage to the left STG and some damage to the right STG, and showed deficits in both prosodic and musical discrimination tasks. On the other hand, the second patient had less extensive, partial bilateral lesions in the A Multidisciplinary Investigation of the Musical and Linguistic Melody 21 STG and showed preserved abilities in both domains. The latter patient was diagnosed as amusic primarily due to her deficits in long-term memory, such as in recalling a tune and perceiving tonality, but apparently displayed normal ability to detect both pitch and rhythmic changes in melodies, based on the test measures used in the study. What is interesting is that she also displayed competency in detecting similar changes in prosody, whereas the other patient failed at both tasks. This pattern suggests that perception of both musical and linguistic melody may rely on some common neural resources, and that these resources are dissociable from those used for long-term musical recognition. It is also notable that the patient with deficits in both areas had more extensive lesions in the STG of the right hemisphere than the left hemisphere, and these lesions were also larger than those of the other patient. Other studies have reported a variety of patterns in the concurrence of amusia and aprosody, ranging from simultaneous progressive development of amusia and aprosody in a patient with right frontal and temporal damage (Confavreux et al., 1992), to concurrence of expressive amusia and expressive aprosody with preserved receptive function in a patient with a lesion in the right temporo-occipital region (Bautista & Ciampetti, 2003), to a case of “pure” amusia with no observed deficits in prosody function (Peretz, 2006). Evidently, data on this topic are scant, scattered and often not suitable for direct comparison. We can broadly infer that the right hemisphere appears to be involved more extensively than the left in prosody processing, although lack of a systematic approach in the field limits what we can conclude from these preliminary, observational studies. A Multidisciplinary Investigation of the Musical and Linguistic Melody 22 A number of researchers have also examined prosody perception in amusic individuals via more experimental approaches, and these, too, have created some debate over the relationship between prosody and melody processing. Several studies have found that amusic individuals do not show impairment in speech prosody perception (Ayotte et al., 2002; Peretz et al., 2002). At the same time, a similar study looking at speakers of Mandarin, a tonal language, has suggested that, although speaking a tonal language does not seem to prevent or compensate for amusia, further supporting the view that musical contour and speech intonation are processed separately, the Chinese amusic participants did show deficits in intonation discrimination tasks for both speech and non-speech stimuli (Jiang et al., 2010). Moreover, even in the context of non-tonal languages like English, it has been shown that amusic individuals have impairment in the ability to discriminate intonation contours extracted from speech (Patel, Foxton & Griffiths, 2005), and that some amusic individuals perform worse than controls on tasks measuring sensitivity to emotional prosody (Thompson, 2007). At the same time, it is true that few of these patients, in both tonal and non-tonal linguistic cultures, report inconveniences or self-perceived deficits in everyday speech (Jiang et al., 2010; Stewart, 2011). These discrepant results bring to attention some possible flaws in experimental design, as they may not actually be measuring the degree to which prosody perception is spared in the case of musical deficit, but simply the fundamental differences in the contexts in which prosody and melody perceptions take place. That is, linguistic stimuli are equipped with extra cues, such as the actual linguistic content (the words and syllables) which may be more easily “tagged” and tracked even in the absence of information on intonational contour (as may be the case for amusic individuals). Even A Multidisciplinary Investigation of the Musical and Linguistic Melody 23 without this information, they may still be able to follow the speech using the other cues. When these cues are removed, as in the case of musical melodies or wordless speech contours, the amusic patients could no longer rely on the linguistic cues, and would inevitably score worse on the tests, but these results would not necessarily mean that the participants were better at detecting prosodic contours than melodic contours (Patel, 2008). In addition, conversations often involve visual and contextual clues that may compensate for the lack of intonational information, and this may explain why so few amusic patients actually experience inconveniences in their daily lives even as they score lower on laboratory test measures of prosody perception (Stewart, 2011). Some researchers have also suggested that the pitch differences used in speech, especially those distinguishing between a statement and a question, are larger (5 to 6 semitones) than those in melodies, which tend to utilize smaller step-wise motions (Peretz, 2006). Considering that amusia is characterized by deficits in perceiving fine-grained pitch differences, it may well be the case that amusic patients fail to discriminate the small pitch differences presented in the standard tests used in the laboratory setting, but have little problem in their daily use of prosody in conversations. The debate over the degree of dissociation or overlap between prosody and melody processing in amusic individuals remains unresolved. Some confounding variables would have to be removed and a more systematic method would need to be devised in future research in order to acquire more precise data. Nonetheless, researchers at this point could probably agree on a general model in which melody and prosody processing mechanisms draw from at least some common resources, including the choclea and the initial detectors of acoustic information. Reliance on short-term memory A Multidisciplinary Investigation of the Musical and Linguistic Melody 24 for holding pitch and temporal patterns (Patel et al., 1998, p. 137), as well as brain regions in the right hemisphere, also seem to be shared in the processing of linguistic and musical melodies. At the same time, they do not rely entirely on the same set of mechanisms, nor do they operate under the same circumstances in our daily lives. Specifically, prosody processing seems to involve both the dominant and the nondominant hemispheres to a larger extent and rely on a wider range of non-acoustical cues, while melody processing seems to lie a little more deeply and uniquely in the nondominant hemisphere. The reverse relationship, that of melody processing in aprosody or aphasia, also calls for further research, although it is quite common for Broca’s aphasia to be accompanied by expressive amusia (Hebert, Racette, Gagnon, & Peretz, 2003). Since lesion in the left hemisphere is the major cause of aphasia in such patients, these cases support the idea that both hemispheres are involved in vocal production for speaking and singing (Schlaug, Marchina, & Norton, 2009). Therefore, even though much of music processing is believed to be a responsibility of the right hemisphere, the ability to express the music may be dependent on Broca’s region in the left hemisphere, much like the motor control over speech is. However, it is also possible that the neurological cause of amusia in such cases does not lie in Broca’s region but some other region in great proximity, since cases of “pure” aphasia without amusia have also been reported (Yamadori, Osumi, Masuhara, & Okubo, 1977). Patient C.C., for example, could produce the musical parts of songs normally, although his ability to repeat the words of the songs, either spoken or sung, was impaired (Hebert et al., 2003). This shows that speech and music can be dissociable even in songs, at least at the level of production, and again A Multidisciplinary Investigation of the Musical and Linguistic Melody 25 testifies to the complexity that underlies the relationship between language and music. Some researchers have pointed out, however, that most other cases of aphasia without amusia involve composers, conductors, or other high-level professional musicians, whose brains have been shown to be different in many ways from the brains of non-musicians (Patel, 2005, p. 65). Again, further research is needed to isolate the true effect of aphasic brain on melody processing. Putting It All Together for a Good Cause: The Melodic Intonation Therapy We have looked at the relationship between melody and prosody from a number of different angles, and evidence from these different approaches collectively suggest that musical and linguistic processing in the brain have both unique and overlapping mechanisms and resources, with the former mainly served by the non-dominant, usually right hemisphere, and the latter served by the dominant, left hemisphere. In fact, fMRI studies have revealed that singing and speaking share some bilateral fronto-temporal neural correlates, although singing or intoned speaking additionally activate the right more than left superior temporal regions, compared to the speaking condition (Vines, Norton, & Schlaug, 2011). Additionally, it is known that there is a “bihemispheric network for vocal production,” regardless of whether the output is intoned or spoken (Kalisch, Tegenthoff, & Dinse, 2008). That is, both musical and linguistic melodies involve at least some parts of both hemispheres. At the same time, this means that both hemispheres are capable, to a certain extent, of producing either melody or prosody, and might be able to compensate for each other in case of damage. In fact, this hypothesis is the basis for several music therapy methods, including an approach known as Melodic Intonation Therapy (MIT), designed to help aphasic patients regain speech. A Multidisciplinary Investigation of the Musical and Linguistic Melody 26 The MIT method was developed in 1973 by Sparks, Helm, and Albert, who were working with aphasic patients in the Aphasia Research Unit in the Boston VA hospital (Marshall & Holtzapple, 1976). The initial inspiration for development of the method was the observation that aphasic patients could often produce “well-articulated, linguistically accurate words while singing, but not during speech” (Schlaug et al., 2009). The preliminary results, involving success cases of three patients, were promising. All three were severely aphasic, unable to produce more than grunts or a few repeated phonemes despite reasonably good comprehension. Other therapeutic methods available at the time had not worked, but a few weeks of MIT was sufficient to bring out of these patients varying degrees of grammatically correct sentences and response to questions, as well as much more facilitated conversations (Albert, Sparks, & Helm, 1973). The method involves a repetition of a series of “intoned” phrases and sentences as well as left-hand tapping for each syllable, exaggerating both the pitch and rhythmic aspects of speech. The therapy is rather intensive and takes place 1.5 hours per day for 5 days a week until the patient has mastered all three levels, which usually takes 75-80 or more sessions (Schlaug et al., 2009). Each level consists of 20 high-probability words and social phrases like “water” and “I love you.” (Norton et al., 2009). The therapist sits across a table from the patient and introduces each word or phrase with a visual stimulus (Norton et al., 2009). The standard method utilizes only two pitches, high and low. For example, the phrase “Thank you” is presented one syllable at a time, with the first, accented syllable in the “high” pitch (comfortably within the patient’s vocal range) and the second, unaccented syllable a minor third below (Norton et al., 2009). The therapist and the patient might sing in unison or call and respond, and later the therapist’s voice A Multidisciplinary Investigation of the Musical and Linguistic Melody 27 fades out midway as the patient continues to sing the entire phrase. Eventually the therapist asks questions to induce the learned phrases from the patients, who then attempt to respond in the “song” they have learned, and finally verbally without the music (Thaut, 2005, p. 167). There are three levels in the method, increasing in difficulty from 2-3 syllable phrases to 5 or more syllables (Norton et al., 2009). There are considerable variations in practice, with some therapists using as many as 7 or 8 pitches instead of 2 pitches, using larger intervals, longer durations (Laughlin, Naeser, & Gordon, 1979), and utilizing instruments like the piano to aid the process (Norton et al., 2009). Several studies have attempted to explain the mechanisms of MIT. Broadly, it has been observed that rhythm and accent patterns of speech –that is, prosodic elements- are uttered correctly before proper articulation of the actual words –the linguistic content- is achieved (Thaut, 2005, p. 167). Also, focusing on the melodic line, rhythm and points of stress in the intoned verbal speech stimuli seem to help patients process the structural aspect of the speech better (Sparks & Holland, 1976), highlighting the role of these prosodic elements in providing “the foundation or structural support for the organization of speech communication (Leung, 1985). MIT, by amplifying the magnitude of such elements, seems to stimulate appropriate prosody production that prepares the patient for real speech. At the same time, it does not seem to be relying on the patient’s knowledge of familiar tunes, which was commonly believed in music therapy as a way to take advantage of the “automatic ability present” (Rogers & Felming 1981, p. 34) and induce a reflex-like verbal output facilitated by this capacity. In contrast, it has often been observed in MIT that use of familiar songs often interferes with associating the new A Multidisciplinary Investigation of the Musical and Linguistic Melody 28 words with the melodies and does not help the patients overcome their deficit in producing meaningful and communicative speech (Carroll, 1996, p. 9). Instead, in their original proposal of the method, Sparks et al. (1973) offer a hemispheric argument, in which MIT “facilitates use of language by the nondominant right hemisphere, which had been suppressed by the dominant left hemisphere, even [after] the dominant hemisphere was damaged” (p. 131). This claim is supported by the fact that the best candidates for MIT, those that show the most improvement over the course of the therapy, are Broca’s aphasics, usually those with no additional lesions in Wernicke’s area, and notably, the right hemisphere (Naesser & Helm-Estabrooks, 1985). More recent neurological research has revealed that recovery from aphasia can be accomplished through recruitment of either the regions around the site of damage in the affected hemisphere, or homologous regions in the unaffected hemisphere, and that “for patients with large left-hemisphere lesions, recovery through the right hemisphere may be the only possible path” (Schlaug et al., 2009). Specific regions believed to participate in this recovery process have also been identified, including the superior temporal lobe, the posterior IFG, and the primary motor cortex, connected via the arcuate fasciculus (AF) (Schlaug et al., 2009). The AF, in particular, is thought to be a structural link between Wernicke’s area and Broca’s area in the left hemisphere, as well as a functional unit that facilitates speech production mechanisms (recall that it plays an important role in the right hemisphere as well, connecting the IFG and the STG together), and increased amounts AF fibers have been detected in the right hemisphere of patients who showed more improvement after MIT (Schlaug et al., 2009). A Multidisciplinary Investigation of the Musical and Linguistic Melody 29 In addition to possibly inducing these structural changes, MIT seems to take advantage of what the right hemisphere is specialized for, which is processing global features, such as melodic contour and overall meter structures. Many aspects of MIT focus on the globality of the melodic shape of speech, as well as exaggerated contrast and more “slowly modulated signals” (Norton et al., 2009), which all seem to be suitable for the right hemisphere’s global processing strategy. This, perhaps, allows a hemispheric substitution in processing of speech. In addition, the left-hand tapping has received more attention in recent research as a possible trigger for engaging the sensorimotor network in the right hemisphere, which would facilitate articulatory movements (Norton et al., 2009; Schlaug et al., 2009). In support of all this, an experiment involving transcranial direct current stimulation (tDCS) has shown that excitatory stimulation of the right IFG during MIT sessions temporarily augments the speech output as much as a direct stimulation of Broca’s area does (Vines, Norton, & Schlaug, 2011). In addition, the therapy sessions themselves may play a role, as the sessions naturally lead to incorporation of proper breathing techniques, increased motivation, and a feeling of competence that builds up over time, especially for successful patients, which may all contribute to the general recovery process (Hebert et al., 2003). Although MIT works best for a select population of aphasic patients and even then doesn’t seem to work all the time (Hebert et al., 2003), it has received attention in recent years as one of the few effective methods available for treatment for nonfluent aphasia, and has shown some reliable and replicable results in various other populations, including children with Down syndrome (Carroll, 1996) and non-verbal autistic children (Miller & Toca, 1979), as well as Romanian aphasic patients with buccolingual apraxia (motor deficit in the cheeks and A Multidisciplinary Investigation of the Musical and Linguistic Melody 30 the tongue) (Popovici & Mihailescu, 1992). Its success in the West has also inspired a Japanese version of the therapy (Seki & Sugishita, 1983). Most notably, Harvey Alter, now the president and founder of the International Aphasia Movement, gave an inspiring speech at the 2008 Music Has Power Awards Benefit (the YouTube video is worth taking a look), describing, in his lilting but clearly recognizable speech, the struggles he faced after a sudden stroke that had left him speechless for years, and the power of music that has allowed him to regain his voice, once lost deep in the “land of aphasia” (Alter, 2008; Barrow, 2008). Conclusion Let us now revisit the questions we had posed at the beginning of our investigation. The boundary between linguistic prosody and musical melody, an ambiguous one in the aesthetic world, seems at least as complex in the neurological domain. This question has received attention of the scientific community only in the recent years, and as we have seen, the field is still at its infancy. However, combining insights drawn from the multi-dimensional approaches taken in this study, we can derive a general model of how the two are related. At the most fundamental level, both are concerned with the global structure of a stream of acoustical information, the relations between the individual units it consists of. Their functions clearly diverge, though, as a musical melody is an aesthetic object in and of itself whereas prosody serves to facilitate transfer of linguistic meaning. In biological terms, they both draw from the same basic auditory mechanisms that allow us to hear, with some modifications geared towards the specific functions they serve. However, melody processing, and musical functions in general, tends to be lateralized to the right hemisphere, and language processing generally A Multidisciplinary Investigation of the Musical and Linguistic Melody relies on the left hemisphere. Prosody, possessing characteristics of both music and language, seems to depend on both hemispheres to a larger extent. It is precisely this coexistence of overlaps and dissociations, the incredible plasticity of our brain and the fuzzy boundary between language and music at the deepest level, that continues to fascinate artists, researchers, and therapists alike. 31 A Multidisciplinary Investigation of the Musical and Linguistic Melody 32 References Abraham, G. (1974). The tradition of Western music. Berkeley: University of California Press. Albert, M. L., Sparks, R. W., & Helm, N. A. (1973). Melodic intonation therapy for aphasia. Archives of Neurology, 29, 130-131. Alter, H. (2008). Speech at the Institute for Music and Neurologic Functions’s 2008 Music Has Power Awards Benefit. [video file]. Retrieved from http://www.youtube.com/watch?v=F_5verI-bj8. American Speech-Language-Hearing Association. (2012). Aphasia. Retrieved from http://www.asha.org/public/speech/disorders/Aphasia/. Ayotte, J., Peretz, I., Hyde, K. (200). Conganital amusai: A group study of adults afflicted with a music-specific disorder. Brain, 125, 238-251. Barrow, K. (2008, April 22). At 60, he learned to sing so he could learn to talk. The New York Times. Retrieved from http://www.nytimes.com/2008/04/22/health/22stro.html. Bautista, R., & Clampetti, M. (2003). Expressive aprosody and amusia as a manifestation of right hemisphere seizures. Epilepsia, 44, 466-467. Beckman, M., & Pierrehumbert, J. (1986). Intonational structure in Japanese and English. Phonology Yearbook,3, 255-309. Bolinger, D. (1989). Intonation and its uses: Melody in grammar and discourse. Stanford: Stanford Univ. Press. Carroll, D. (1996). A study of the effectiveness of an adaptation of melodic intonation therapy in increasing the communicative speech of young children with Down A Multidisciplinary Investigation of the Musical and Linguistic Melody 33 syndrome. (Unpublished doctoral dissertation). McGill University, Canada. Confavreux, C., Croisile, B., Garassus, P., Aimard, G., & Trillet, M. (1992). Progressive amusia and aprosody. Archives of Neurology, 49, 971-976. Deutsch, D. (2003). But they sometimes behave so strangely. On Phantom words and other curiosities [CD]. La Jolla, CA, USA: Philomel Records, Inc. Di Pietro, M., Laganaro, M., Leeman, B., Schinder, A. (2004). Receptive amusia: temporal auditory deficit in a professional musician following a left temporoparietal lesion. Neuropscyghologia, 42, 868-977. Hallé, P. A., de Boysson-Bardies, B., & Vihman, M. (1991). Beginnings of prosdic organization: Intonation and duration patterns of disyllables produced by Japanese and French infants. Langauge and Speech, 34, 299-318. Hebert, S., Racette, A., Gagnon, L., Peretz, I. (2003). Revisiting the dissociation between singing and speaking in expressive aphasia. Brain, 126, 1838-1850. Huron, D. (2011). Why is sad music pleasurable? A possible role for prolactin. Musicae Scientiae, 15, 146-159. Hyde, K. L., Lerch, J. P., Zatorre, R., Griffiths, T. D., Evans, A. C., & Peretz, I. (2007). Cortical thickness in congenital amusia: When less is better than more. The Journal of Neuroscience, 27, 13028-13032. Janata, P., Birk, J., Van Horn, J., Leman, M., Tillmann, B., & Bharucha, J. (2002). The cortical topography of tonal structures underlying Western music. Science, 298, 2167-2170. Jiang, C., Hamm, J. P., Lim, V. K., Kirk, I. J., Yang, Y. (2010) Processing melodic A Multidisciplinary Investigation of the Musical and Linguistic Melody 34 contour and speech intonation in congenital amusics with Mandarin Chinese. Neuropsychologia, 48, 2630-2639. Johansson, B. B. (2010). Current trends in stroke rehabilitation. A review with focus on brain plasticity. Acta Neurologica Scandinavica, 123, 147-159. Kalisch, T., Tegenthoff, M., & Dinse, H. R. (2008). Improvement of sensorimotor functions in old age by passive sensory stimulation. Journal of Clinical Interventions in Aging, 3, 637-90. Kolb, B., & Whishaw, I. Q. (2003). Fundamentals of human neuropsychology. New york: Worth Publishers. Laughlin, S. A., Naeser, M. A., & Gordon, W. P. (1979). Effects of three syllable durations using the melodic intonation therapy technique. Journal of Speech, Language, and Hearing Research, 22, 311-320. Lehiste, I. (1973). Phonetic disambiguation of syntactic ambiguity. Glossa, 7, 107-121. Leung, K. (1985). Enhancing the speech and language development of communicatively disordered children through music and movement. Paper presented at the annual convention of the Council for Exceptional Children, Anaheim, CA, April 1985. Marotta, G., Barbera, M., & Bongioanni, P. (2008). Prosody and Broca’s aphasia: An acoustic analysis. Studi Linguistici e Filologici Online, 6, 79-98. Marshall, N., & Holtzapple, P. (1976). Melodic intonation therapy: Variations on a theme. In R. H. Brookshire (Ed.), Clinical Aphasiology Conference, vol.6 (pp.115-141). Minneapolis: BRK Publishers. MediLexicon. (n. d.). Aphasia. Retrieved from http://www.medilexicon.com/medicaldictionary.php?t=5407. A Multidisciplinary Investigation of the Musical and Linguistic Melody 35 MediLexicon. (n. d.). Aprosody. Retrieved from http://www.medilexicon.com/medicaldictionary.php?t=5808. Miller, S. B., & Toca, J. M. (1979). Adapted melodic intonation therapy: a case study of an experimental language program for an autistic child. Journal of Clinical Psychiatry, 40, 201-203. Mitchell, R., & Ross, E. (2008). fMRI evidence for the effect of verbal complexicity on lateralisation of the neural response associated with decoding prosodic emotion. Neuropsychologia, 46, 2880-2887. Monrad-Krohn, G. H. (1957). The third element of speech: Prosody in the neuropsychiatric clinic. The British Journal of Psychiatry, 103, 326-331.. Murayama, J., Kashiwagi, T., Kahiwagi, A., Mimura, M. (2004). Impaired pitch production and preserved rhythm production in a right brain-damaged patient with amusia. Brain and Cognition, 58, 36-42. Naeser, M. A., & Helm-Estabrooks, N. (1985). CT scan lesion localization and response to melodic intonation therapy with nonfluent aphasia cases. Cortex, 21, 203-223. Norton, A., Zipse, L., Marchina, S., & Schlaug, G. (2009). Melodic intonation therapy: Shared insights on how it is done and why it might help. Annals of the New York Academy of Science, 1169, 431-436. Pannekamp, A., Toepel, U., Alter, K., Hahne, A., & Friederici, A. D. (2005). Prosodydriven sentence processing: An event-related brain potential study. Journal of Cognitive Neuroscience, 17, 407-421. Patel, A. D. (2003). A new approach to the cognitive neuroscience of melody. In I. Peretz A Multidisciplinary Investigation of the Musical and Linguistic Melody 36 & R. Zatorre (Eds.), The Cognitive Neuroscience of Music (pp. 325-345). Oxford: Oxford Univ. Press. Patel, A. D. (2005). The relationship of music to the melody of speech and to syntactic processing disorders in aphasia. Annals of the New York Academy of Science, 1060, 59-70. Patel, A. D. (2006). An empirical method for comparing pitch patterns in spoken and musical melodies: A comment on J.G.S. Pearl’s “Eavesdropping with a master: Leos Janá ek and the music of speech. Empirical Musicology Review, 1, 166-169. Patel, A. D. (2008). Music, language, and the brain. New York: Oxford University Press. Patel, A. D., Foxton, J. M., & Griffiths, T. D. (2005). Musically tone-deaf individuals have difficulty discriminating intonation contour extracted from speech. Brain and Cognition, 59, 310-313. Patel, A. D., Peretz, I., Tramo, M., & Labreque, R. (1998). Processing prosodic and musical patterns: A neuropsychological investigation. Brain and Language, 61, 123-144. Pearce, J. M. S. (2005). Selected observations on amusia. European Neurology, 54, 145148. Pell, M. (2006). Cerebral mechanisms for understanding emotional prosody in speech. Brain and Langauge, 96, 221-234. Peretz, I. (2006). Brain specialization for music. Annals of the New York Academy of Sciences, 930, 153-165. Peretz, I., Ayotte, J., Zatorre, R.J., Mehler, J., Ahadm P., Penhune, V. B., et al. (2002). A Multidisciplinary Investigation of the Musical and Linguistic Melody 37 Congenital amusia: A disorder of fine-grained pitch discrimination. Neuron, 33, 185-191. Peretz, I., Brattico, E. Jarvenpaa, M., & Tervaniemi, M. (2009). The amusic brain: In tune, out of key, and unaware. Brain, 132, 1277-1286. Peretz, I., & Hyde, K. L. (2003). What is specific to music processing? Insights from congenital amusia. Trends in Cognitive Sciences, 7, 362-367. Phillips-Silver, J., Toiviainen, P., Gosselin, N., Piche, O., Nozarada, S., Palmer, C., & Paretz, I. (2011). Born to dance but beat deaf: A new form of congenital amusia. Neuropsychologia, 49, 961-969. Piccirilli, M., Sciarma, T., Luzzi, S. (2000). Modularity of music: Evidence from a case of pure amusia. Journal of Neurology, Neurosurgery & Psychiatry, 69, 541-545. Plack, C. (2005). The sense of hearing. Mahwah, NJ: Lawrence Erlbaum Associates; Popovici, M., & Mihailescu, L. (1992). Melodic intonation in the the rehabilitation of Romanian aphasics with bucco-lingual apraxia. Romanian Journal of Neurology and Psychiatry, 30, 99-113. Price, P. J., Ostendorf, M., Shattuck-Hufnagel, S., & Fong, G. (1991). The use of prosody in syntactic disambiguation. Journal of the Acoustical Society of America, 90, 2956-2970. Ringer, A. L. (n. d.). Melody. In Grove Music Online. Retrieved from http://www.oxfordmusiconline.com/subscriber/article/grove/music/18357?q=mel ody&search=quick&pos=1&_start=1#firsthit. Rohrer, J., Sauter, D., Scott, S., Rossor, M., & Warren, J. (2010). Receptive prosody in nonfluent primary progressive aphasias. Cortex, 48, 308-316. A Multidisciplinary Investigation of the Musical and Linguistic Melody 38 Rogers, A., & Fleming, P. L. (1981). Rhythm and melody in speech therapy for the neurologically impaired. Music Therapy, 1, 33-38. Ross, E. D., Mesulam, M. (1979). Dominant language functions of the right hemisphere? Prosody and emotional gesturing. Archives in Neurology, 36, 144-148. Sacks, O. (2007). Musicophilia. New York: Random House. Sakai, K., Hikosaka, O., Miyauchi, S., Takino, R., Tamada, T., Iwata, N. K., & Nielsen, M. (1999) Neural representation of a rhythm depends on its interval ratio. Journal of Neuroscience, 19, 10074-10081. Sarkamo, T., Tervaniemi, M., Soinila, S., Autti, T., Silvennoinen, H. M., Laine, M., et al. (2009). Cognitive deficits associated with acquired amusia after stroke: A neuropsychological follow-up study. Neuropsychologica, 47, 2642-2651. Sarno, M. T. (1998). Acquired aphasia. San Diego: Academic Press. Schlaug, G., Marchina, S., & Norton, A. (2009). Evidence for plasticity in white matter tracts of chronic aphasic patients undergoing intense intonation-based speech tharpy. Annals of the New York Academy of Sciences, 1169, 385-384. Schneck, D. J., & Berger D. S. (2006). The music effect: Music physiology and clinical applications. London: Jessica Kingsley Publishers. Seki, K., & Sugishita, M. (1983). Japanese-applied meodic intonation therapy for Broca aphasia. No To Shinkei, 35, 1031-1037. Sparks, R. W., & Holland, A. L. (1976). Method: Melodic intonation therapy for aphasia. Journal of Speech and Hearing Disorders, 41, 287-297. Stewart, L. (2011). Characterizing congenital amusia. The Quarterly Journal of Experiemtnal Psychology, 64, 625-638. A Multidisciplinary Investigation of the Musical and Linguistic Melody 39 Thaut, M. H. (2005). Rhythm, music, and the brain: Scientific foundations and clinical applications. New York: Routeledge. Thompson, W. F. (2007). Exploring variants of amusia: Tone deafness, rhythm impairment, and intonation insensitivity. Paper presented at the International Conference on Music Communication Science. Thompson, W. F. (2008). Impairments of emotional prosody among individuals with amusia. Neurosciences and Music III: Disorders and Plasticity, Montreal. Tong, Y., Gandour, J., Talavage, T., Wong, D., Dzemidzic, M., Xu, Y., Li, X., & Lowe, M. (2005). Neural circuitry underlying sentence-level linguistic prosody. NeuroImage, 28, 417-428. Van Lancker, D., & Sidtis, J. J. (1992). The identification of affective-prosodic stimuli by left- and right-hemisphere-damaged subjects: All errors are not created equal. Journal of Speech and Hearing Research, 35, 963-970. Vines, B. W., Norton, A. C., Schlaug, G. (2011). Non-invasive brain stimulation enhances the effects of melodic intonation therapy. Frontiers in Psychology, 2, 110. Yamadori, A., Osumi, Y., Masuhara, S., & Okubo, M. (1977). Preservation of singing in Broca’s aphasia. Journal of Neurology, Neurosurgery, and Psychiatry, 40, 221224. Zatorre, R. J., Berlin, P. (2001). Spectral and temporal processing in human auditory cortex. Cerebral Cortex, 11, 946-953.