Sunday, September 14, 2008

Music and Language Perception - Part 1 Commonalities

The functional similarities as well as the differences between the two domains of music and language have been topics of discussion among researchers for quite some time. The similarities between the two are seemingly obvious. Both share a hierarchical and temporal structure. Both share a syntax (although the perception of syntax is not as explicit in the case of music as it is for language...in addition to being subjective to each listener depending on his or her level of training and exposure). Time is critical to the perception and recognition of both spoken words as well as melodies.

Given these similarities, cognitive researchers have been tempted to believe that perception and recognition of both melodies as well as spoken words involve the same functional processes: the need for working memory to temporarily store sequential information and consolidate this information into a higher-order percept, the need for long term memory to influence perception through top-down contextual feedback, and a mechanism that enables the integration of these two processes.

In addition, both music and language share the same neural correlates for subcortical processing (from the outer ear upto the thalamus and in some cases upto the primary auditory cortex). They also share the same neural correlates for a large amount of cortical processing that includes areas such as the primary and secondary auditory cortices and perisylvian language areas (Broca's area and Wernicke's area). Also, musical syntax-based violations (both in melodic and in harmonic contexts) show activation of areas associated with syntax violations in language.

Tuesday, August 5, 2008

Amusia...continued

Previous studies pertaining to British and French-Canadian amusics showed that about 30% of amusics have difficulty discriminating between statements and questions depending on the intonation based on the final rising (question) or falling (statement) pitch glide .

Patel et al. re-examined prosodic judgments in British and French-Canadian amusics by asking participants to make same vs. different judgments for two different kinds of stimuli. The first set of stimuli consisted of statement-question pairs, where the two sentences were exactly similar except for the last part where the intonation changed depending on whether the sentence was a statement or a question. So, a question would have a rising pitch glide on the last word (e.g. He likes to drive fast cars?) whereas a statement would have a slight dip (e.g. He likes to drive fast cars.). The second set of stimuli consisted of focus-shift pairs, where although the two sentences were similar, the intonation differed depending on the location of the salient pitch accent (e.g. Go in front of the bank I said. vs. Go in front of the bank I said.).

Their results indicated that some amusics had difficulty discriminating statement-question pairs. However, these amusics did not have any difficulty discriminating between two sentences in a focus-shift pair. This suggested that while amusics could detect pitch movements in speech, they had difficulty detecting the direction of pitch movements (based on their difficulty in judging the rise or fall in pitch glide on the same word in the statement-question pair).

These results are also consistent with results from a previous study conducted by Foxton et al., where British amusics were asked to judge the direction of a pure tone pitch glide. Foxton et al. manipulated the glide rate by keeping the glide duration constant and by increasing or decreasing the size of the glide. They found that amusics had difficulty when the pitch glides were smaller (the glide rate was lower). On average, the threshold for accurate direction judgment in amusics was about 2.2 semitones compared to 0.1 semitones in controls. When expressed as glide rates, these numbers correspond to 22 semitones/second vs. 1 semitone/second. These results offer various avenues for experimental and computational exploration.

Currently, work is being done on amusic prosody judgments in tonal languages such as Mandarin Chinese. It will be extremely interesting to see how work in this area progresses.

Friday, July 25, 2008

Amusia

One highly exciting topic in music cognition is that of amusia and the perception of prosody in amusics. In simplistic terms, amusia refers to tone deafness. "Tone deafness" here does not refer to our subjective use of that phrase in everyday life, such as when applying it to a poor singer (with normal music perception abilities). Its application is much more limited, and is determined objectively by means of tests such as the MBEA (Montreal Battery for the Evaluation of Amusia). These tests involve same v different judgment tasks w.r.t. melody and rhythm. A lot of significant work in this area has already been done by Isabelle Peretz and her colleagues. Peretz, Ayotte, Hyde, Zatorre, Penhune, Patel were some of the initial researchers who successfully narrowed down the problem of amusia to a pitch-perception related disorder.

Amusia in a lot of cases is congenital (suggesting a genetic basis) and is estimated to affect about 4% of the population (a surprisingly high figure!). Research seems to indicate that it is not an all-or-nothing disorder. Amusia can impair music perception abilities from a low to a high degree, and might be existing within a continuum. Amusia can also be acquired on account of brain damage. What makes amusia an interesting area of study is the nature of evidence from amusics showing partial impairment of linguistic prosody judgments in some cases, and the lack of it in some cases. This inconsistency with regard to the accompaniment of a linguistic prosody disorder along with impaired musical abilities has urged researchers to explore this area in further detail. Current research findings were presented in a talk given by Aniruddh Patel at the Neurosciences and Music conference in Montreal. A poster related to judgment of prosody in tonal languages was presented by Sebastien Nguyen et al. I will talk about some of these research findings in greater detail in my next blog.

Thursday, July 10, 2008

The Neurosciences and Music III conference - Part 3

Friday, June 27th was a much better day. The day started off with a two and a half hour poster session. Steven Mithen's keynote lecture was refreshing. It was nice to hear from an established archaeologist's perspective about how humans could have evolved as a musical species. The gist of his talk can be summarized by a few lines from his abstract. "New research in the study of hominin fossils, archaeological remains as well as in the fields of neuroscience, developmental psychology and musicology are allowing" insights into music's evolutionary history, suggesting that "communication by variations in pitch and rhythm most likely preceded that of speech, and may indeed have provided the foundation from which spoken language evolved."

Symposium 5 dealt with the topic of emotions and music. Maria Saccuman gave an interesting talk about musicality in infants. Saccuman et al. subjected 18 two-day old infants to an fMRI study, with three kinds of stimuli involving western tonal music excerpts; alterations of these excerpts; and excerpts with violations of tonal syntax. Their results indicated the existence of dedicated systems for processing music-like stimuli and sensitivity to syntactic violations of these stimuli. Daniel Levitin presented results from his lab suggesting how adolescents with autistic spectrum disorders (ASD) are more sensitive to structural features of music than to the expressive (emotive) features. It was interesting to compare this with Levitin's previous work on music perception in individuals with Williams syndrome who, unlike ASD individuals, display a strong sensitivity to music's expressive features.

While there were several more talks on Friday as well as on Saturday, which I am sure were of interest to people in various domains of music cognition, I would like to focus on the one talk that impacted me the most on Saturday, June 28th: Gottfried Schlaug's work on how singing helps aphasic patients, as a part of Symposium 6. This was by far the BEST talk of the entire conference for me. His presentation and results were evidence of music's multimodal nature, and its ability to recover pathways in the homologous language regions of the right hemisphere, in aphasic patients with left hemispheric lesions (left frontal and superior temporal lobes). His talk illustrated how a form of treatment through melodic intonation (Melodic Intonation training) can considerably improve speech production in aphasic patients. Each patient underwent 75 treatment sessions or so, of MIT, which enabled the individual to produce/articulate speech by actively engaging the language regions of the right hemisphere through musical elements such as intonation, rhythmic tapping, and continuous voicing.

Thursday, July 3, 2008

The Neurosciences and Music III conference - Part 2

Thursday, June 26th was a packed day with sixteen talks in total!! I made the mistake of attending all sixteen talks, and should have left out a few. So, as one can imagine, my brains were completely saturated by the end of the tenth talk.

The first symposium which consisted of five talks, was on "Rhythms in the brain: Basic science and clinical perspectives." I found Chen et al.'s work on the importance of the premotor cortex in music production, to be the most interesting of the five. Chen et al. subjected participants to various musical rhythm-related tasks that included passive listening, anticipating prior to a motor act, and committing a motor act. Their fMRI results suggested that in addition to using motor areas and the cerebellum for sequencing rhythmic actions, musicians use the prefrontal cortex to a greater extent (their hypothesis: prefrontal activity in musicians is related to their superior ability to organize musical rhythms wrt working memory). Their data also indicated that the posterior STG (superior temporal gyrus) and the premortor cortex are important mediating nodes for transforming auditory information into motor activity. In addition, their data suggested a direct mapping of auditory information to the motor system through auditory links to the ventral premotor cortex. However, the dorsal premotor cortex seemed to have indirect links, for processing higher-order information pertaining to musical rhythm.

The second symposium was short consisting of only two talks, on normal and impaired singing. I liked the second talk by Steven Brown. The first talk on poor pitch singing didn't really offer any new insights. By the end of the talk I had specific questions, but unfortunately, did not receive satisfactory answers. The questions I had were along the lines of: 1) Where do you draw the line between a poor singer and a tone deaf (amusic) person? ( I believe the Montreal Battery of Evaluation of Amusia, could provide answers to this question). 2) Assuming that a person with normal music perception/recognition abilities, is a good musician (plays an instrument at a fairly accomplished level), and a poor singer, what differentiates this person from a poor singer with music cognition deficiencies? Steven Brown's talk was compelling because it answered my second question by narrowing down the reasons for poor pitch singing (in people with normal music cognition abilities) to deficient or anomalous activation in the larynx motor cortex (in addition to other areas).

The third symposium was on musical training and induced cortical plasticity. All the speakers presented research findings which were generally in support of the notion that musical training and performance induce changes in the brain within limits. The fourth symposium was on music and memory. Emmanuel Bigand's talk was interesting. The attempt of his research was to find the minimum time necessary for activating musical and linguistic memories. Minimum time could also be interpreted as minimum amount of information necessary, in the time course of music recognition. His research suggested that even a slice of information as small as 50ms was enough to "bootstrap" memory for music.

The most interesting talk in this session was that of Isabelle Peretz. Peretz et al. attempted to find out neural correlates of the musical lexicon (storage areas for familiar melodies, or melodies stored in long-term memory). Participants listened to familiar melodies, unfamiliar melodies, and scrambled melodies. Subtraction from fMRI imaging data between familiar melody listening and unfamiliar melody listening suggested two things: 1) the supplementary motor area in the left hemisphere might be involved in "inner singing" or emulation, 2) the right superior temporal sulcus may be involved in the retrieval of information from the musical lexicon.

Tuesday, July 1, 2008

The Neurosciences and Music III conference - Part 1

I just got back from the the 3-day Neurosciences and Music-III conference in Montreal, June 26th-28th, 2008. Held, once every three years, this is THE conference to go to for researchers in music cognition interested in the neuroscience aspect of the domain. The conference was packed with seven symposia and two poster sessions. The most current research in various domains of music cognition and neuroscience was presented in these sessions by some of its biggest scientists.

The good part about attending a conference like this is the fact that you get to witness quality research directly from its source; without merely relying on published papers. In addition, every conference provides us with an opportunity to network with peers, and established researchers, thereby sowing the seeds for possible collaborative efforts. Now for the bad part....and this probably applies to most conferences. Too much information is presented in too short a time, hardly leaving the audience with enough time to process all the information. On the one hand I hated skipping some of the talks...especially after having traveled all the way to Montreal (and traveling to these conferences isn't cheap for a graduate student), but on the other hand I saw no other solution after having crossed the threshold beyond the point of information overload. So the conference always puts you in the position of having to choose the talks you want to listen to wisely.

I will attempt to summarize my version of the events on all 3 days of the conference in my next few blogs.

Monday, June 16, 2008

Review of "This is your brain on music" by Daniel Levitin - Part 4

I found chapters 4, 5, and 6 to be the most interesting parts of the book. Without giving much away, I will try to provide my concise review of these chapters. Some of the topics covered in chapter 4 are the role of functional units (called schemata) in our long-term memory stores which help us in perception and enable anticipation of incoming information and how composers violate them to create a sense of novelty in the listener; the role of neurotransmitters and receptors in providing the listener with emotions w.r.t expectation and satisfaction or violation of expectations; hemispheric specializations and their functional roles in the context of music and language; and the effect of musical training on hemispheric specialization. At the end of this chapter, Levitin provides us with a summarized, high-level hypothetical picture of the neural organization of the brain for music and speech.

Chapter 5 deals almost entirely with functional processes pertaining to memory and categorization. Cognitive science students/researchers should find this chapter appealing. In addition, frequency effects of melodies, and melodic invariance are also briefly mentioned in this chapter.

I found chapter 6 to be the most enjoyable chapter of the book. I hope that other readers would find the remaining chapters equally interesting unlike me, because herein lies the problem. After reading chapter 6, I felt like a sky diver who completed a jump in the middle of the day.....experienced an intense adrenaline rush....after which he/she had nothing better to look forward to for the rest of the day, and had to experience the remaining part of the day in lethargy. To do justice to the book, I will end my review with chapter 6, and hope that someone more deserving will inform readers about the remaining chapters. Hopefully other readers will hit their high notes at later points in the book. Music, to most people, is an emotional activity, and this chapter highlights it beautifully, while informing us about the neural correlates involved in causing those emotions. Cognitive psychologists have studied various cognitive processes, but most have shied away from studying emotion. I am glad that the role of emotion was treated on par with other cognitive processes in this book, more so in the context of a routine yet wonderful activity such as listening to music.

Sunday, June 8, 2008

Review of "This is your brain on music" by Daniel Levitin - Part 3

Chapter 3 is where Levitin starts involving the reader in the neural correlates of music cognition. The gross functional generalizations of the four lobes of the brain (frontal, parietal, temporal, and occipital) are mentioned, along with the older/reptilian brain. It is nice reading about the several functional processes involved in different kinds of musical activity - perception, memory (both working memory and long-term memory), attention, motor control during performance and listening, and of course...emotion. Levitin highlights the fact that almost every neural center in the brain has a part to play for perceiving, recognizing, reading and performing music....vocal or instrumental. Another interesting thing to note is the similarity between music and language in the way they are organized and perceived, as well as the neural correlates they share such as Broca's area and Wernicke's area.

Two final points of interest in this chapter are 1) perception is the end product of a long chain of neural events, and is not instantaneous; 2) perception need not always be correct....our brain can actually misinform us and fool us into believing something.
It is just amazing thinking about how complex an activity "perception" is, and how our brains have evolved to respond to the requirements of perception with incredible efficiency.....something that most of us take for granted. However, in order to achieve this kind of efficiency, there has to be some kind of trade-off in accuracy. Our brain uses stored information, does some kind of probabilistic computation to complete/fill in perceived bottom-up information with top-down context. This is what enables us to "expect" things....in a lot of cases subconsciously.

This brings us to the interesting topic of auditory as well as musical illusions analogous to visual illusions such as the Ponzo and Kaniza illusions. The book led me to investigate Sardinian acapella music for the first time. (Unfortunately, I could not decipher/perceive the illusion, although I really liked the music). But I came across other auditory illusions with respect to melody, where the pitches were segregated into different ordered patterns for the right and left ear involving high and low pitches. But when heard, we perceive the segregation different from the actual order (we tend to group lower pitches as perceived by one ear, and higher pitches as perceived by another). This definitely needs to be tried out with earphones.

I will finish my review of "This is your brain on music" within my next two blogs. If you still haven't purchased the book, or at least checked it out from your local library.......you are seriously missing out!

Tuesday, June 3, 2008

Review of "This is your brain on music" by Daniel Levitin - Part 2

In Chapter 1, What is Music?: From Pitch to Timbre, Levitin provides an excellent, concise yet lucid overview of the basic elements of sound pertaining to music (tone, pitch, rhythm, tempo, contour, timbre, loudness, spatial location, reverberation etc.), from a psychoacoustics perspective, useful to both musicians and non-musicians. He then proceeds to define the higher order elements made of these basic elements, such as meter, key, melody and harmony. The higher levels of our brain process the relationships between these elements in a coherent and cohesive way. The relationships are important not only from the standpoint of how notes feature in music, but also how they do not. To illustrate this point, Levitin uses a visual art metaphor through an example. “Miles Davis….described his improvisational technique as parallel to the way that Picasso described his use of canvas: The most critical aspect of the work, both artists said, was not the objects themselves, but the space between the objects.” According to Levitin, the most important part of Miles’s work was the empty space between the notes. Miles’ genius lay in his ability to know precisely when to hit the next note, giving the listener enough time to anticipate.

I couldn't help but think of Jeff Beck in this context. An extremely underrated guitar player who superbly uses space to emote. I found his guitar playing constantly evolve over time in comparison to his contemporary old school blues-based guitarists such as Jimmy Page, and Clapton.

Levitin interestingly approaches sound as a relative, non-absolute phenomenon. According to Levitin, pitch is an internal image of the end product of a chain of mechanical and neurochemical events caused by sound waves impinging on the eardrums and pinnae. The difference between Levitin's view from other physicists is that while most physicists agree that "pitch" is a mental percept, they consider sound to be absolute on account of vibrating molecules, irrespective of whether it is perceived or not.

Levitin talks about melody as an auditory object that remains invariant in spite of transformations along some of the basic elements of sound, similar to the invariance of objects in the visual domain. Although the areas of the brain that respond to individual pitches have been mapped, we do not know the neurological basis for key invariance, where transposed melodies sound perceptually equivalent.

In Chapter 2, Foot Tapping: Discerning Rhythm, Loudness, and Harmony; Levitin explains how we use various areas of our brain to discern rhythm, and to perform rhythmically. These areas include the cerebellum and basal ganglia, higher cognitive regions such as the motor cortex, and the frontal lobes for planning. He explains the difference between rhythm, meter, and tempo – related but often confused concepts, and how we use our brain to perceive these concepts. I was thrilled to see him spend some time talking about musicians' use of syncopation to play with listener's emotions and expectations. Our amazing ability to segregate sounds and group sounds based on the situation (The Cocktail party example), for performing auditory scene analysis is mentioned.

Monday, May 26, 2008

Review of "This is your brain on music" by Daniel Levitin - Part 1

Although, I was already intimately involved with music both academically (through research and modeling of certain aspects of specific performers, by way of case studies), and as a guitar player in various bands, the book that really turned me on to the neural correlates of music cognition was Daniel Levitin's, "This is Your Brain on Music - The Science of a Human Obsession."

While Levitin's book is intended for a general audience, it provides several avenues of interest for the serious music cognition researcher as well. What makes Levitin's style unique, yet personal and thoroughly engaging is his immensely, well-rounded experience in music as a performer/musician, recording engineer in the music industry, as well as researcher and academician. To make his point, he provides several relevant examples from mainstream genres such as classic rock, R&B, soul, and pop.

Another point worth mentioning is the fact that Levitin's interest is in cognitive systems and not just in neuroscience itself. So his research is in no way reductionist. He makes his intentions very clear by stating that his interest in neuroscience is to understand the functional aspects behind cognitive processes. This view is reflected well throughout the book when he consistently connects functional processes pertaining to memory, categorization, emotion, attention etc. to existing paradigms in experimental psychology as well as various neural correlates.

The book also provides a window to the music cognition research community. It helps acquaint even the general reader to several key researchers such as Peretz, Zatorre, Janata, Tillman, Trehub, in a non-pedantic way. In my next few blogs I will provide a deeper overview of some of the chapters in this book.

If you are even remotely interested in music in any way (avid listener, performer, teacher, researcher), you should have purchased this book already!!!

Friday, May 16, 2008

Welcome to the musical brain!

Given that music is a temporal activity, the perception and recognition of music elicit strong comparisons with spoken-word recognition. The difference between the two however is that spoken-word recognition has been an active area of study since the late 70s with highly influential psychological models published by scientists such as McClelland, and Marslen-Wilson, while music as a separate field of psychological study is just about gaining ground.

I find Marslen-Wilson's initial and revised versions of the cohort model appealing because of their explicit way of separating bottom-up sensory information from top-down context related effects, and providing specific rules for constraining the influence of context during recognition. Spoken-word recognition has several examples of context influencing recognition, such as the phoneme restoration effect. Intuitively, recognizing melodies follows a similar set of functional processes as speech. I know based on anecdotal experience that partial melodies are enough to enable mentally filling in and humming the rest of the melody. However, one of the problems I have faced is that of finding examples in melody recognition suggesting contextual effects consistent with the phoneme restoration effect in speech. On a positive note, experimental evidence based on the gated presentation of melodies (Dalla Bella et al.) does suggest that the recognition of melodies involves combination of bottom-up pitch-related information with top-down context.