Machine learning techniques for semantic analysis of dysarthric speech: An experimental study

Vladimir Despotovic*, Oliver Walter, Reinhold Haeb-Umbach

*Corresponding author for this work

Research output: Contribution to journalArticleResearchpeer-review

13 Citations (Scopus)

Abstract

We present an experimental comparison of seven state-of-the-art machine learning algorithms for the task of semantic analysis of spoken input, with a special emphasis on applications for dysarthric speech. Dysarthria is a motor speech disorder, which is characterized by poor articulation of phonemes. In order to cater for these non-canonical phoneme realizations, we employed an unsupervised learning approach to estimate the acoustic models for speech recognition, which does not require a literal transcription of the training data. Even for the subsequent task of semantic analysis, only weak supervision is employed, whereby the training utterance is accompanied by a semantic label only, rather than a literal transcription. Results on two databases, one of them containing dysarthric speech, are presented showing that Markov logic networks and conditional random fields substantially outperform other machine learning approaches. Markov logic networks have proved to be especially robust to recognition errors, which are caused by imprecise articulation in dysarthric speech.

Original languageEnglish
Pages (from-to)242-251
Number of pages10
JournalSpeech Communication
Volume99
DOIs
Publication statusPublished - May 2018
Externally publishedYes

Keywords

  • Acoustic units
  • Dysarthric speech
  • Machine learning
  • Semantic analysis
  • Spoken language understanding

Fingerprint

Dive into the research topics of 'Machine learning techniques for semantic analysis of dysarthric speech: An experimental study'. Together they form a unique fingerprint.

Cite this