1,373
Views
3
CrossRef citations to date
0
Altmetric
Editorial

Bringing advanced speech processing technology to the clinical management of speech disorders

&
Pages 581-582 | Received 01 Jul 2018, Accepted 06 Aug 2018, Published online: 09 Jan 2019

Research to date has made it clear that many speech disorders will not be fully remediated without intensive and long-term intervention. At the same time, practitioners of speech-language pathology grapple with large caseloads and heavy documentation requirements that may leave them unable to provide an adequate duration of treatment for each client. Advanced speech processing technologies have the potential to act as an extender of a clinician’s services by facilitating practice in between treatment sessions, or to enhance the diagnostic accuracy of their evaluations. While there has been a proliferation of apps and software to support speech intervention, few have the capacity for automated or semi-automated assessment of the quality or accuracy of the user’s speech outputs. This special issue highlights current work that applies automatic speech recognition (ASR) or other speech processing technologies to the clinical assessment and/or treatment of disorders affecting speech production. It also discusses current limitations and barriers and identifies potential future directions for technological enhancement of the clinical management of speech disorders.

In the first contribution, McKechnie et al. (Citation2018) provide a systematic review of literature investigating automated analysis tools for the diagnosis and treatment of speech production disorders. While providing an overview of diverse previous studies, their paper also highlights the limited scope of existing work and calls for expanded focus on this topic. In the second paper, Keshet (Citation2018) offers an overview of the architecture of a modern ASR system, aimed at readers without technical expertise in this area. By familiarising the reader with the basic characteristics of ASR and its associated evidence base, these two papers lay a foundation for the advances described in the rest of the issue.

The next two papers explore the possibility of integrating automated speech analysis into mobile apps for use in the treatment of paediatric speech disorders. Ahmed et al. (Citation2018) investigate how children and speech-language pathologists responded to several tablet-based games that had been modified so that gameplay is controlled by the accuracy of the user’s speech. Campbell, Harel, Hitchcock, and McAllister Byun (Citation2017) lay groundwork for an automated extension of an existing app to provide visual biofeedback for the treatment of rhotic misarticulation. Despite the preliminary nature of these studies, both take a positive view of the potential for apps to serve as a source of feedback and motivation and thereby increase treatment doses delivered to children in speech therapy.

The next paper, by Niziolek and Kiran (Citation2018), describes a semi-automated analysis of variability in vowel formants produced by people with aphasia. They use these acoustic findings to make an argument about the nature of speech-motor control in this population. Similarly, Buz, Buchwald, and Keshet (Citation2018) argue that measurements of voice onset time (VOT) can provide insight into the level at which speech production is disrupted in patients with aphasia and/or apraxia of speech, and they report a procedure to automatically extract VOT measurements from large samples of either typical or clinical speech.

The final two papers focus on the population of individuals with dysarthria secondary to amyotrophic lateral sclerosis (ALS). As elicitation of extended speech samples is fatiguing and thus undesirable in the context of progressive speech disorders, it would be beneficial if automated analyses can enhance the diagnostic accuracy of brief samples of speech. Wang et al. (Citation2018) argue that a single acoustic and articulatory sample can be used to predict the severity of speech impairment in individuals with ALS. Likewise, Rong, Yunusova, Richburg, and Green (Citation2018) offer evidence that articulatory data from an alternating motion rate (AMR) task can be used to identify early signs of speech-motor degeneration in this population. Jointly, these papers show that automatically extracted acoustic and articulatory features can be leveraged to improve the process of diagnosing and monitoring speech deficits.

These contributions, along with similar recent research, point to a promising future in which technologies for automated speech processing can improve the diagnosis and treatment of speech disorders. At the same time, this line of inquiry is in a very early stage, and our contributors are uniform in calling for increased interdisciplinary collaboration to advance this important field. Keeping pace with advances in speech recognition technology will allow the discipline of speech-language pathology to stay relevant while also helping to secure improved quality of life for patients with communication deficits.

TARA MCALLISTER
Department of Communicative Sciences and Disorders, New York University, New York, NY, USA
[email protected]

KIRRIE J. BALLARD
Faculty of Health Sciences, The University of Sydney, Sydney, Australia

References

  • Ahmed, B., Monroe, P., Hair, A., Tan, C.T., Gutierrez-Osuna, R., & Ballard, K.J. (2018). Speech-driven mobile games for speech therapy. International Journal of Speech-Language Pathology, 20(5).
  • Buz, E., Buchwald, A., & Keshet, J. (2018). Assessing automatic phonetic annotation tools using unimpaired and impaired speech. International Journal of Speech-Language Pathology, 20(5).
  • Campbell, H., Harel, D., Hitchcock, E., & McAllister Byun, T. (2017). Selecting an acoustic correlate for automated measurement of /r/ production in children. International Journal of Speech-Language Pathology, 20(5). doi:10.1080/17549507.2017.1359334
  • Keshet, J., (2018). Automatic speech recognition: A primer for speech pathology researchers. International Journal of Speech-Language Pathology, 20(5). doi:10.1080/17549507.2018.1462851
  • McKechnie, J., Ballard, K.J., McCabe, P., Monroe, P., Gutierrez-Osuna, R., & Ahmed, B. (2018). Automated speech analysis tools for assessment, diagnosis and/or modification of speech production: A systematic search and review. International Journal of Speech-Language Pathology, 20(5). doi:10.1080/17549507.2018.1477991
  • Niziolek, C.A., & Kiran, S. (2018). Assessing speech correction abilities with acoustic analyses: Evidence of preserved online correction in persons with aphasia. International Journal of Speech-Language Pathology, 20(5).
  • Rong, P., Yunusova, Y., Richburg, B., & Green, J. (2018). A diagnostic tool for assessing articulatory involvement in ALS: Automatic extraction of abnormal speech movement features from the alternating motion rate (AMR) task. International Journal of Speech-Language Pathology, 20(5).
  • Wang, J., Kothalkar, P., Kim, M., Cao, B., Yunusova, Y., Campbell, T., … Green, J. (2018). Automatic speech severity prediction for individuals with ALS from a single speech acoustic and articulatory sample. International Journal of Speech-Language Pathology, 20(5). doi:10.1080/17549507.2018.1408855

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.