818
Views
16
CrossRef citations to date
0
Altmetric
Letter to the Editor

A resource of validated digital audio recordings to assess identification of emotion in spoken language after a brain injury

, &
Pages 248-250 | Published online: 05 Feb 2013

Background

A recent paper Citation[1] presented a set of validated and linguistically equated lexical sentences that can test for the ability of persons with brain injury to process emotions in the lexical content. Here is presented a set of validated digital audio recordings of these sentences (see Electronic Appendix), spoken in various prosodies. This set can be used to separate the impact of lexical content and prosody on the processing of emotion in speech in persons with brain injury. This set is made available to researchers and clinicians.

A reduced ability to recognize emotional cues following an acquired brain injury (ABI) has a detrimental impact on rehabilitation. Specifically, emotion-identification difficulties (EID) in speech following ABI are found to be widespread Citation[2] and related to poor quality-of-life Citation[3]. As a consequence, the past decade shows a growing interest in both the research and clinical community to correctly assess EID as a path for rehabilitation. Yet, existing tools are scarce and may not fit an ABI population Citation[1], as they cannot provide a complete picture of the specific roles of prosody (tone of speech) and lexical content in EID. For example, when one hears a sentence that carries a happy lexical content—‘I feel wonderful today’— spoken in an angry prosody, what emotion was conveyed by the speaker: Anger, happiness or a combination of both? EID in ABI may result from an impaired identification of prosodic auditory cues, of the lexical content of a spoken utterance, or from an impaired ability to selectively attend to one dimension when the other also conveys emotional information.

In our previous work Citation[1], we presented a set of 50 lexical sentences, with 10 exemplars in each of the following five categories: Anger, Fear, Happiness, Sadness and Neutral. These lexical sentences were distinctive in conveying their corresponding emotions and they were equated on main linguistic characteristics (e.g. frequency of usage, phonologic neighbourhood, sentence length) across the five emotion categories. This way, this lexical set controls for well-known biases on word and emotion identification. The next step in EID assessment is presented here, with a set of digital audio recordings of these lexical sentences spoken in five prosodies corresponding to the emotional categories mentioned above.

Stimuli selection

The 50 lexical sentences were recorded by a native English-speaking (Ontario, Canada) actress. Each sentence was recorded three times in each of the five different prosodies (Anger, Fear, Happiness, Sadness and Neutral) to generate a set of 750 recorded sentences. Digital audio files were equated with respect to root mean square amplitude. From these spoken sentences, a sub-set of 50 was selected such that no lexical sentence is repeated and each lexical emotional category is represented twice in each of the tested emotional prosodies, generating a 5 × 5 matrix (). These spoken sentences were chosen based on high ratings from a group of six young adults, reflecting good exemplars of their respective emotional prosodies. This set of 50 spoken sentences is available in an Electronic Appendix.

Table I.  The 50 spoken sentences, as distributed across emotions expressed in prosody and lexical content

Stimuli testing

To validate the prosody without the influence of the lexical content, two methods of digital acoustic filtering were employed: (a) prosody unveiling restricted representation (PURR Citation[4]); and (b) low-pass filtering (LPF500 Citation[5]). The former involves eliminating all frequencies above the third harmonic and aperiodic signals, while the latter simply removes all frequencies above 500Hz. Both are reported in the literature as effective techniques to de-lexicalize recorded speech, while leaving enough acoustic information for the identification of prosodic cues.

Participants

Twenty-eight students from the University of Toronto (18–28 years old) participated in this study. All were native English speakers as assessed by a self-report and a vocabulary test (Mill-Hill), had pure-tone air-conduction thresholds within clinically normal limits in the 0.25–3 kHz range in both ears (≤20 dB HL) and no indication of head trauma or CNS diseases. Randomly, 12 participants were assigned to the PURR and 16 to the LPF500 condition.

Procedure

In both conditions, participants were tested individually in a single-walled sound-attenuating booth. The 50 acoustically manipulated files (PURR or LPF500) were presented binaurally via headphones, once in each of four rating blocks (in a pseudo-random fashion), corresponding to Anger-, Fear-, Sadness- and Happiness-rating blocks, making for 200 trials. In each rating block, participants were asked to rate how much they agree that the speaker identifies with the pre-defined emotion (based on the prosody) using a 6-point Likert scale ranging from strongly disagree (1) to strongly agree (6).

Results

Across emotional-rating blocks, sentences spoken in the corresponding rated-prosody (means, LPF500: 4.4; PURR: 4.1) received scores higher than the average values of other emotional prosodies, while neutral-prosody sentences (means, LPF500: 1.7; PURR: 2.6) received lower ratings than the average (linear trend,- LPF500: F(1, 16) = 365.5, p < 0.001,  = 0.96; PURR: F(1, 11) = 80.5, p < 0.001,  = 0.88). For example, in the anger-rating block ratings given to acoustically manipulated anger-prosody sentences (means, LPF500: 4.5; PURR: 4.4) were higher than average values for Fear, Happy and Sad prosodies and ratings given to acoustically manipulated neutral-prosody sentences (means, LPF500: 1.9; PURR: 2.2) were lower than the average. In sum, these results confirm that listeners identified the intended emotional valence of emotional-prosody sentences and did not perceive any emotional valence in neutral-prosody sentences.

Discussion

In this letter we argue that, in order to obtain a more complete picture on EID, it is necessary to carefully investigate the specific influences of prosody and lexical content on the comprehension of spoken language. To this end, the Electronic Appendix includes a set of 50 validated and controlled, emotional and neutral, spoken sentences, available for use in research on emotion processing. Furthermore, since EID has significant impacts on the rehabilitation of ABI, we believe having access to this set can be useful to improve the assessment and treatment of communication problems related to emotion processing in this population. Given its general nature, this set can also be employed with other populations (e.g. ageing, dementia) characterized by impaired communication skills with an underlying emotional processing deficit component.

Declaration of interest: The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper.

Supplemental material

Supplementary Material

Download WAV Audio (802.9 KB)

Supplementary Material

Download WAV Audio (845.6 KB)

Supplementary Material

Download WAV Audio (895.2 KB)

Supplementary Material

Download WAV Audio (815 KB)

Supplementary Material

Download WAV Audio (795.7 KB)

Supplementary Material

Download WAV Audio (796.2 KB)

Supplementary Material

Download WAV Audio (799.5 KB)

Supplementary Material

Download WAV Audio (779.5 KB)

Supplementary Material

Download WAV Audio (1 MB)

Supplementary Material

Download WAV Audio (1.1 MB)

Supplementary Material

Download WAV Audio (741.7 KB)

Supplementary Material

Download WAV Audio (744.3 KB)

Supplementary Material

Download WAV Audio (792.1 KB)

Supplementary Material

Download WAV Audio (606.8 KB)

Supplementary Material

Download WAV Audio (820.3 KB)

Supplementary Material

Download WAV Audio (892.3 KB)

Supplementary Material

Download WAV Audio (793.6 KB)

Supplementary Material

Download WAV Audio (789.1 KB)

Supplementary Material

Download WAV Audio (1.1 MB)

Supplementary Material

Download WAV Audio (957.2 KB)

Supplementary Material

Download WAV Audio (653.4 KB)

Supplementary Material

Download WAV Audio (664.4 KB)

Supplementary Material

Download WAV Audio (725.6 KB)

Supplementary Material

Download WAV Audio (701.7 KB)

Supplementary Material

Download WAV Audio (733 KB)

Supplementary Material

Download WAV Audio (761.2 KB)

Supplementary Material

Download WAV Audio (686.9 KB)

Supplementary Material

Download WAV Audio (713.7 KB)

Supplementary Material

Download WAV Audio (806.4 KB)

Supplementary Material

Download WAV Audio (881.8 KB)

Supplementary Material

Download WAV Audio (971.5 KB)

Supplementary Material

Download WAV Audio (575.9 KB)

Supplementary Material

Download WAV Audio (741.3 KB)

Supplementary Material

Download WAV Audio (648.3 KB)

Supplementary Material

Download WAV Audio (665.9 KB)

Supplementary Material

Download WAV Audio (697.8 KB)

Supplementary Material

Download WAV Audio (912.7 KB)

Supplementary Material

Download WAV Audio (898.2 KB)

Supplementary Material

Download WAV Audio (894.8 KB)

Supplementary Material

Download WAV Audio (872.3 KB)

Supplementary Material

Download WAV Audio (646.5 KB)

Supplementary Material

Download WAV Audio (592.3 KB)

Supplementary Material

Download WAV Audio (740.4 KB)

Supplementary Material

Download WAV Audio (770.1 KB)

Supplementary Material

Download WAV Audio (685 KB)

Supplementary Material

Download WAV Audio (621.8 KB)

Supplementary Material

Download WAV Audio (889.6 KB)

Supplementary Material

Download WAV Audio (876.8 KB)

Supplementary Material

Download WAV Audio (809.7 KB)

Supplementary Material

Download WAV Audio (911 KB)

References

  • Ben-David BM, Van Lieshout PHHM, Leszcz T. A resource of validated affective and neutral sentences to assess identification of emotion in spoken language after a brain injury. Brain Injury 2011; 25: 206–220
  • Dimoska A, McDonald S, Pell M, Tate R, James C. Recognizing vocal expressions of emotion in patients with social skills deficits following traumatic brain injury. Journal of the International Neuropsychological Society 2010; 16: 369–382
  • Henry JD, Phillips LH, Crawford JR, Theodorou G, Summers F. Cognitive and psychosocial correlates of alexithymia following traumatic brain injury. Neuropsychologia 2006; 44: 62–72
  • Sonntag GP, Portele T. PURR – A method for prosody evaluation and investigation. Computer Speech and Language 1998; 12: 437–451
  • Knoll AM, Uther M, Costall A. Effects of low-pass filtering on the judgment of vocal affect in speech directed to infants, adults, and foreigners. Speech Communication 2009; 51: 210–216

Electronic Appendix

A folder containing 50 digital audio files (WAV) is available online. File names are compatible with .

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.