The American Psychiatric Association (APA) has updated its Privacy Policy and Terms of Use, including with new information specifically addressed to individuals in the European Economic Area. As described in the Privacy Policy and Terms of Use, this website utilizes cookies, including for the purpose of offering an optimal online experience and services tailored to your preferences.

Please read the entire Privacy Policy and Terms of Use. By closing this message, browsing this website, continuing the navigation, or otherwise continuing to use the APA's websites, you confirm that you understand and accept the terms of the Privacy Policy and Terms of Use, including the utilization of cookies.

×
LettersFull Access

Prosodic Impairment in Alzheimer's Disease: Assessment and Clinical Relevance

Published Online:

To the Editor: Prosody refers to the patterns of stress and intonation in language. A few studies suggest that Alzheimer's disease (AD) patients have deficits in production, repetition, and comprehension of emotional prosody. The authors present the case of an AD patient with prosodic speech alteration; to test such impairment, we administered an extensive test battery investigating emotional comprehension and production. The patient was tested for visual channel information by an Ekman and Friesen series. The authors found that prosody impairment can early occur in the disease course, and, for many primary emotions, impairment in the auditory channel seems to mirror the visual channel. Prosody is a term used to cover features of speech such as the emphasis placed on certain syllables, changes in tempo or timing, and differences in pitch and intonation. Ross and Mesulam1 demonstrated that prosody can be impaired independently of other language functions. Most studies that have examined emotional prosody (the nonverbal affective components of language) have used normal control subjects and patients with unilateral brain damage. These studies suggest that the integrity of the right hemisphere is important for both expression and comprehension of emotional prosody. Ross2 proposed that the right hemisphere is organized for affective language expression and comprehension in a manner that is comparable to the organization of propositional language in the left hemisphere. This difference (supralexical semantic information, such as interrogative or assertive mode, versus emotional connotation, such as happiness, anger, sadness, etc.) has been investigated by a few fMRI studies that have reported activation in the right hemisphere, especially in the right inferior frontal area,3 during emotional detection versus verbal detection. This finding is remarkably congruent with neuropsychological studies of aprosodias. In AD, damage to the right hemisphere structures occurs as part of the widespread progressive degeneration, suggesting that emotional prosody and other areas of affective communication may be compromised. A study of linguistic functioning in AD patients4 reported that transcortical sensory and sensory aphasias were prominent in the early stages of AD, with global aphasia appearing at the end stage. The progression of prosodic alteration seems to mirror the model proposed by Appell. In fact, Roberts et al.2 suggested that AD patients have deficits in production, repetition, and comprehension of emotional prosody (Table 1). Deficits in comprehension of emotional prosody seem to appear relatively early in the course of AD and are relatively resistant to further decline throughout the mild and moderate stages of the disease. Repetition and production of emotional prosody are relatively preserved early in the course of the illness, whereas a significant impairment emerges in the moderate stage of the disease.

We present the case of an AD patient who, in spite of the relative early stage of his disease, suffered from an evident impairment of prosodic comprehension and production: caregivers indicated the patient's speech alteration, described as “flattened,” “without emotions,” as one of the first hints of the disease. In order to test such prosodic impairment the patient underwent an extensive battery investigating comprehension (auditory emotional decoding), production (auditory emotional coding), as well as emotional decoding through the visual channel (images taken by Ekman and Friesen series, 1976).

Case Report

The patient, a 55-year-old man, was diagnosed with early-onset AD. The first symptoms (reported by the patient and caregivers) were memory impairment and anomia, followed by apathy, depression, and irritability, as well as the reported speech alteration. Neurological exam was normal as well as Mini-Mental State Exam score (29/30). MRI showed cortical/subcortical atrophy in fronto-parietal areas; the 18-FDG PET scan evidenced hypometabolism in the right frontal-temporal-parietal areas as well as left posterior parietal areas and basal ganglia. The patient underwent to an extensive neuropsychological test battery assessing episodic memory (both verbal and visuospatial), constructional abilities, attention-shifting abilities, abstract thinking, verbal fluency, and naming. A widespread, mild cognitive impairment, more pronounced on tasks sensitive for both right-hemisphere and executive functions congruent with PET result was reported. Also, in order to evaluate the patient's speech, Melodic Alteration—an experimental protocol—was designed:

Emotional Decoding

Auditory verbal task (Task 1): Twelve recorded sentences with emotionally neutral content (e.g., I bake a cake in the oven.) to be presented to the subject, conveyed by a professional actor, with surprise, sadness, anger, and happy emotions, plus the neutral intonation as a control, in a randomized order (total stimuli: 60). The results were compared with those of a control group of 10 healthy subjects, matched for age, education, and sex (mean age: 54.5; mean education: 14 years). Visual task (Task 2): Thirty facial expressions to be presented in six blocks, each of them showing the same five emotions of Task 1 in a randomized order. The patient was asked to indicate each emotion on a separate sheet, to avoid any naming difficulty.

Emotional Coding

Repetition task (Task 3): This was a reproduction of the emotional prosody of the first six-sentence-block of Task 1, to be presented in a randomized order; performance was recorded and then reviewed by two independent raters, who evaluated the similarity between the patient's and the actor's intonation. Production task (Task 4): This was an evaluation of the patient's ability to spontaneously produce emotional prosody, by presenting the second six-sentence block of Task 1, given in a written form. The patient was asked to convey the sentences with surprise, sadness, anger, happy, and neutral intonation. This performance was also recorded and submitted to the same raters who assigned an emotion to each sentence.

Task 1: The patient only recognized 31 out of 60 sentences (versus 58 [SD: 2] of the control group). A prevalent recognition impairment was evident for happiness (67% failed recognition), sadness (50%), and neutral intonation (87.5%). Remarkably, these three emotions were constantly confused with each other (Table 1). Task 2: The patient recognized 73% of facial expressions (versus 92% of control group) with a prevalent impairment of “neutral,” which was detected as “sadness” in 67% of cases (Table 2). Task 3: Two independent raters reported sentence repetition as “wrong” in 33% of cases, with a high level of concordance (K index=0.86; IC 95%: 1.9–25.2). Because of this concordance, only the first rater's results are shown (Table 3). Task 4: Spontaneous production of emotional intonation also showed a consistent deficit, with a major impairment on Anger (83% failed) and a complete impairment in Happiness (Table 4; K index=0.93; IC 95%: 0.01–0.59).

TABLE 1. Auditory Emotional Recognition
TABLE 1. Auditory Emotional Recognition
Enlarge table
TABLE 2. Visual Emotional Recognition
TABLE 2. Visual Emotional Recognition
Enlarge table
TABLE 3. Emotional Repetition
TABLE 3. Emotional Repetition
Enlarge table
TABLE 4. Emotional Coding
TABLE 4. Emotional Coding
Enlarge table

Discussion

We present a patient showing prominent prosodic deficits occurring in the early stage of AD, with regard to both visual and auditory emotional coding and decoding. In Task 1, prevalent recognition impairment was evident for Happiness (confused with Sadness and Surprise), Sadness (constantly confused with Neutral intonation and Happiness), and, finally, Neutral intonation (confused with Sadness and Happiness). In Task 2, Neutral expression was constantly confused with Sadness. These results could be explained by the association of “neutral” intonation (as well as “neutral” facial expression) with a reduction of melodic intonation (as well as reduction of facial expression), both typical of sadness also. On the other hand, misdiagnosing of surprise and happiness is evident in the auditory tasks, probably for the positive valence of both intonations, whereas, in the visual task, the act of “smiling” well distinguishes happiness from surprise. Task 3 showed deficits in repetition in all tasks except for Surprise, probably because of the strong pitch and intonation of such emotion, easier to repeat for the patient. Finally, Task 4 also showed a complete impairment on Happiness (100%) and a clear deficit to conveying Anger intonation: these deficits appear to be related to the difficulties in production of emotions that require strong intonation and tempo changes. In fact, in this task, only Neutral coding is correctly performed, obviously the only one that conveys absence of emotional sharing. A few studies, to our knowledge, have investigated the ability of emotional decoding through both auditory and visual modalities in AD; most of them lack the support of functional neuroimaging. Data in the literature are dissonant about facial expression recognition in AD6,7; the present case, instead, demonstrates that such impairment can occur early in the disease course, along with deficits in recognition of prosody; and shows that impairments in the auditory channel for many primary emotions seems to mirror impairments in the visual channel. Such conditions should be seriously considered by therapists and caregivers in order to improve communication and relationships with AD patients.

Memory Clinic Department of Neurological Science La Sapienza University Rome, Italy

1. Ross ED , Mesulam MM : Dominant language functions of the right hemisphere? prosody and emotional gesturing. Arch Neurol 1979; 36:144–148Crossref, MedlineGoogle Scholar

2. Roberts V , Ingram S , Lamar M , et al.: Prosody impairment and associated affective and behavioral disturbances in Alzheimer's disease. Neurology 1996; 47:1482–148Crossref, MedlineGoogle Scholar

3. Ross E : The aprosodias. functional-anatomic organization of the affective components of language in the right hemisphere. Arch Neurol 1981; 38:561–569Crossref, MedlineGoogle Scholar

4. Appell J , Kertesz A , Fisman M : A study of language functioning in Alzheimer patients. Brain Lang 1982; 17:73–91Crossref, MedlineGoogle Scholar

5. Buchanan TW , Lutz K , Mirzazade S , et al.: Recognition of emotional prosody and verbal components of spoken language: an fMRI study. Cogn Brain Res 2000; 9:227–238Crossref, MedlineGoogle Scholar

6. Burnham H , Hogervorst E : Recognition of facial expressions of emotion by patients with dementia of the Alzheimer type. Dement Geriatr Cogn Disord 2004; 18:75–79Crossref, MedlineGoogle Scholar

7. Weiss EM , Kohler CG , Vonbank J , et al.: Impairment in emotion recognition abilities in patients with mild cognitive impairment, early, and moderate Alzheimer disease compared with healthy comparison subjects. Am J Geriatr Psychiatry 2008; 16:974–980Crossref, MedlineGoogle Scholar