AI-Based Automated Speech Therapy Tools for persons with Speech Sound Disorders: A Systematic Literature Review

DOI: https://doi.org/10.21203/rs.3.rs-1517404/v1

Abstract

This paper presents a systematic literature review of published studies on AI-based automated speech therapy tools for persons with speech sound disorders (SSD). The COVID-19 pandemic has initiated the requirement for automated speech therapy tools for persons with SSD making speech therapy accessible and affordable. However, there are no guidelines for designing such automated tools and their required degree of automation compared to human experts. In this systematic review, we followed the PRISMA framework to address four research questions: 1) what types of SSD do AI-based automated speech therapy tools address, 2) what is the level of autonomy achieved by such tools, 3) what are the different modes of intervention, and 4) how effective are such tools in comparison with human experts. An extensive search was conducted on digital libraries to find research papers relevant to our study from 2007 to 2022. The results show that AI-based automated speech therapy tools for persons with SSD are increasingly gaining attention among researchers. Articulation disorders were the most frequently addressed SSD based on the reviewed papers. Further, our analysis shows that most researchers proposed fully automated tools without considering the role of other stakeholders. Our review indicates that mobile-based and gamified applications were the most frequent mode of intervention. The results further show that only a few studies compared the effectiveness of such tools compared to expert Speech-Language Pathologists (SLP). Our paper presents the state-of- the-art in the field, contributes significant insights based on the research questions, and provides suggestions for future research directions.

1 Introduction

Speech Sound Disorder (SSD) refers to difficulties with perception, motor pro- duction, phonological representation of speech sounds, and speech segments. In short, a person with SSD finds it difficult to produce or use some sounds correctly. According to the American Speech-Language-Hearing Association (ASHA), SSD can be organic and functional [1]. While organic SSD result from an underlying motor/neurological, structural, or sensory/perceptual cause, there is no known cause for functional SSD [1](see Fig. 1). The prevalence of SSD varies significantly according to different studies; however, these stud- ies reflect the magnitude of the problem. Multiple studies have estimated that residual or persistent speech errors occur in 1–2% of older children and adults [2]. In various studies overall, 2.3–24.6% of school-aged children were estimated to have speech delay or speech sound disorders [3, 4]. In a 2012 survey, National Center for Health Statistics found that 48.1% of 3 to 10- year-old children and 24.44% of 11 to 17-year-old children had speech sound problems [3]. According to the 2011 census, in India, hearing impairment (18.9%) was the second leading disability, and speech impairment (7.5%) was the fifth-highest disability [5]. In another survey conducted in India’s rural popu- lation, the researchers of All India Institute of Speech and Hearing (AIISH), Mysore, found that around 6.07% were at risk of communication disorders, including speech sound disorder [6].

In addressing such speech impairments, Speech-Language Pathologists (SLPs) play a significant role in the screening, assessment, diagnosis, and treat- ment of persons with SSD. Personalized speech therapy and practice monitored by SLPs can improve the acquisition of speech skills [7]. However, the acces- sibility of SLPs is crucial for such intervention. A report suggests that up to 70% of SLPs have waiting lists, which indicates a shortage in the workforce [7, 8]. Furthermore, according to United Nations Children’s Fund (UNICEF), there are not adequate speech-language therapy services for children with com- munication disorders and disabilities [9]. Moreover, speech therapy involves extended interactions and multiple sessions with SLPs. Such therapy requires extensive time, making it expensive and inaccessible for persons living in impoverished and rural areas. Addressing these issues of accessibility and expensiveness, many researchers have proposed AI-based automated tools for providing speech therapy autonomously to persons with SSD. With the advent of improved ASR tools, impaired speech datasets, and AI-based techniques, it is now possible to build such autonomous tools for speech therapy. These autonomous tools embedded in mobile devices or provided as cloud services can be revolutionary in making speech therapy accessible and affordable.

This paper reviews the research involving AI-based automated speech ther- apy tools for persons with SSD during the last 15 years. In the next section 2, we present the methodology adopted in this study. Section 3 reports the results and Section 4 presents the discussion. Finally, we conclude the study in Section 5.

2 Methodology

We followed the PRISMA protocol to perform the systematic literature review to achieve higher transparency and reliability [10]. We included studies that can cover one of the following questions we framed for this systematic literature review.

2.1 Eligibility Criteria

2.1.1 Types of Studies

We considered full articles, review papers, and short papers which proposed automated speech therapy tools using AI techniques such as machine learning and deep learning. The studies were restricted to articles written in English and published from 2007 to 2022 (research carried out over the last 15 years). The objective is to provide a snapshot of the research domain. We carried out the final search on February 4th, 2022.

2.1.2 Types of Participants

We included participants of any age and gender with SSD, such as articulation disorder, phonological disorder, apraxia, dysarthria, cleft palate, and hearing impairment. However, we excluded studies specifically addressing cognitive conditions such as Alzheimer’s disease, Down’s syndrome, Parkinson’s disease, and Autism Spectrum Disorders.

2.1.3 Types of Intervention

We didn’t restrict on types of intervention and included all studies that pro- posed automated speech therapy using different intervention methods. Studies included robotics-based, mobile-based, computer-based interventions along with gamified and story-based intervention methods.

2.2 Information Sources

The studies were identified by searching electronic databases using the search term generated using keywords from the research questions. The search string was applied to Scopus, IEEE Xplore, and ACM Digital Library electronic databases. The results from the databases were extracted, and inclusion/ex- clusion criteria were applied to find relevant studies.

2.3 Search Terms

The following keywords were used to search all the databases: speech, lan- guage, disorder, impairment, assessment, therapy, rehabilitation, treatment, AI, artificial intelligence, automated, automatic. Boolean operators were used to combine the terms as: (”AI” OR ”Artificial Intelligence” OR ”automa*”) AND (”speech” OR ”language”) AND (”disorder” OR ”impairment”) AND (”assessment” OR ”therapy” OR ”rehabilitation” OR ”treatment”).

2.4 Study Selection and Data Collection

We found a total of 763 research studies from individual databases, i.e., 635 from Scopus, 72 from IEEE Xplore, and 56   from  the  ACM   Digital  Library. Then we removed duplicates and corrupt entries to find 678 papers for the screening phase. We performed the screening of the studies in three stages. At first, two authors screened the titles, which resulted in 238 research studies. Two authors reviewed abstracts in the next screening stage, which resulted in 94 research studies. Finally, all four authors reviewed the full texts of the 94 articles. After applying the inclusion and exclusion criteria for this systematic literature review, we selected 24 research studies. Disagreements during the screening process were resolved by discussion and voting by all the authors. Figure 2 shows all the phases: ”identification,” ”screening,” and ”included” according to the PRISMA protocol [10].

3 Results

We report our results in five different sections. In the first section 3.1, we analyzed the included papers with respect to paper counts, authors, regions, languages, and venues. In the following sections (see 3.2,3.3,3.4,3.5), we present our findings based on the research questions addressed in this systematic literature review.

2.1 Paper Counts, Authors, Regions, Languages, and Venues

The final papers selected after the full review were 24 papers from 23 different venues. The number of studies on AI-Based automated speech therapy tools for persons with SSD published shows an upward trend over the years (see Figure 3). Out of 24 papers, we can observe that 20 articles were published during the last 6-7 years.

The majority of the paper included in this study were published in journals (see figure 4). Additionally, there were ten papers published in conference proceedings and two book chapters among the 24 included studies. However, we could not find any eligible studies published in a magazine.

There were 91 unique authors identified from the included studies. The VOSviewer software was used to calculate the most impactful authors, generate co-authorship clusters, and perform co-occurrences of keyword analysis [11]. All the authors were counted irrespective of the authorship order with the same weightage applied to all the authors. However, high weightage was attributed to authors publishing more articles. In addition, to find the list of most impactful authors, their collaborating links were also considered, along with the number of published documents. The top ten most impactful authors are listed in the table 1. The most significant cluster of authors based on the number of articles and collaborative link strength was found, as shown in the figure 5. It is worth noting that 79 authors (86.81 %) contributed to only one paper in the included studies, i.e., have only one work relating to AI-based automated speech therapy in the last 15 years. Moreover, after analyzing the author’s keywords of the included studies, the most significant cluster of linked and co- occurred keywords was found as shown in the figure 6. The most significant keyword was ASR(Automatic Speech Recognition).

Table 1 Top ten most impactful authors

Author

Documents

Total Link Strength

Lopez-Nores, M.

4

18

Pazos-arias, J.

4

18

Robles-Bykbaev, V.

4

18

Guaman-Heredia, M

2

11

Quasi-peralta, D.

2

11

Lee, T.

2

10

Ng, C.W.Y.

2

10

Ng, S.I.

2

10

Wang, J.

2

10

Garcia-duque, Jorge

2

9

We further report the geographical distribution of the included studies based on the location of the study indicated in the paper (see figure 7). We looked at the author’s affiliation and funding agency when required. Most papers reported on studies which were conducted in Europe (11 papers) and North America (6 papers). Studies conducted in Europe include four studies from Spain and one study each from Germany, Hungary, Romania, Portugal, the Czech Republic, and Italy. On the other hand, studies from North America include four studies from the USA, one collaborative study between Panama and Nicaragua, and another study from Mexico. Moreover, five papers reported  studies in Asia, which includes China (2 studies), India (1 study), Taiwan (1 study), and the Philippines (1 study). However, other continents are heavily underrepresented; Africa and Oceanic each have one study conducted. Finally, we could not find any eligible studies meeting our selection criteria which were conducted in South America.

We presented the language distribution of the papers based on the language addressed by the AI-based automated speech therapy tools as reported in the studies (see figure 8). The most addressed languages were English (10 studies) and Spanish (4 studies). Furthermore, two studies addressed the Cantonese language, and there was only one study each for Punjabi, German, Hungarian, Romanian, Portuguese, Italian, Arabic, and Mandarin. The studies were drawn from 23 unique venues. We could observe that the vast majority of venues from which papers were chosen (95.65%) were represented by only one article. Only one venue, i.e., ”Studies in Health Technology and Informatics,” had published two papers included in this review.

3.2 Speech Sound Disorders (RQ1)

We found that researchers have addressed multiple types of SSD in the liter- ature. However, 12 studies out of 24 studies did not address any specific SSD (see figure 9). These studies proposed automated tools for a generalized SSD population and experimented without specifying any particular SSD [8, 1222]. Researchers have also specifically worked and devised AI-based tools for per- sons with hearing impairment [23, 24]. A novel tongue-based Human-Computer interaction tool [25] and gamified AI-based tool [7] for persons with motor speech disorder have been proposed.

Moreover, Frieg et al. proposed a digital training system for dysarthric patients [26]. In another similar study, Saz et al. devised ASR-based tools and technologies and conducted user studies specifically for dysarthric patients [27]. Singh et al. and Chen et al. developed and assessed automatic AI-based speech therapy tools for articulation disorder in Punjabi and Mandarin, respectively [28,29]. Ballard et al. conducted a feasibility study of a tablet-based automated feedback tool for apraxia patients [30]. On the other hand, Ramamurthy et al. developed a novel companion robot, ”Buddy,” for cleft lip and palate disorder children  [31]. In another study, Rivas et al. proposed using a virtual world to provide speech therapy for children with dyslalia [32]. It is worth noting that only one study was related to speech data collection for Cantonese to perform phonology and articulation assessment [13]. The figure 10 shows the distribution of papers addressing specific SSD.

3.3 Level of Autonomy (RQ2)

Researchers worldwide have amplified the debate between autonomy vs. human control due to the risks and concerns associated with AI and large-scale automation [33]. In this area concerning automation and AI in speech ther- apy, we studied the level of autonomy achieved by AI-based automated speech therapy tools. In many studies, researchers build fully automated AI-based speech therapy tools without considering the role of parents, SLPs, and other stakeholders. While Desolda et al. emphasized the role of caregivers and SLP in the design of a remote therapy tool, ”Pronuntia” [12], Ng et al. proposed a fully automated assessment tool using the CUChild 127 speech corpus in Cantonese [14]. In another study, Bilkova et al. developed a novel lip, tongue, and teeth detection system using Convolutional Neural Network (CNN) and Augmented Reality (AR) for supporting the automatic evaluation of speech therapy exercises [25].

Furthermore, Sztaho et al. proposed a fully automated speech therapy tool by displaying visual feedback on intensity(accent), intonation, and rhythm to children with hearing impairments [23]. In another similar study, Hern´andez et al. developed a serious game with an automatic feedback feature for hearing- impaired children [24]. Ballard et al. performed a feasibility study of their tablet-based, fully automated therapy tool for children with apraxia without any role of SLP and other stakeholders [30]. Moreover, V. Robles-Bykbaev et al. proposed a framework imitating the main functionality of SLP along with a robotic assistant motivating children in therapy activity and automatically giving real-time feedback [8, 1719]. In another similar study, Ramamurthy et al. proposed a companion robot, ”Buddy,” which automatically evaluates speech exercises of children with CL/P disorders with the feature of monitoring by SLPs [31].

3.4 Modes of Intervention (RQ3)

Researchers have adopted different modes of intervention while implementing AI-based automated speech therapy tools for persons with SSD. As these thera- pies are often targeted at children, researchers emphasize developing tools that trigger excitement and build companionship. Desolda et al. proposed a web application for children, SLPs, and caregivers, allowing SLP to assign therapy exercises to children with SSD [12]. The system automatically evaluates the correctness of the exercises and gives real-time feedback. On the other hand,

Ballard et al. proposed a tablet-based therapy tool for children with apraxia [30]. Furthermore, Ng et al. and Sztaho et al. proposed a computer-based prosody teaching system for children with hearing impairment and a computer- based visual feedback system for the hearing impaired, respectively [14, 23]. Bykbaev et al. proposed a novel robotic assistant along with a fully automatic framework imitating the work of SLP [8]. In another similar study, Rama- murthy et al. proposed a therapy robot, ”Buddy,” allowing children to practice assigned exercises at home [31]. Many studies have incorporated serious games as an intervention tool for automatic speech therapy [7, 21, 25, 31, 34]. One of the studies incorporated augmented reality to build a serious game using tongue detection [25].

3.5 Effectiveness (RQ4)

The effectiveness of AI-based automated speech therapy tools depends on their performance compared to SLPs. Moreover, automated speech therapy tools providing wrong feedback can be disastrous to children’s speech improve- ment. Few studies (4 out of 24) compared the results of their automated tool with human experts (SLPs) (see figure 12). Ballard et al. conducted an inter- rater agreement test between their ASR tool and SLPs and found ASR-human agreement averaged 80% [30]. In another study, Sztaho et al. found that their automated tool scores correspond to the subjective evaluation by SLPs [23]. Bykbaev et al. found that over 90% of the therapy plans generated automat- ically by their expert ”Spelta” were ”better than” or ”as good as ” what the SLPs would have created manually [18]. Moreover, in the study by Saz et al., their Automatic Speech Recognition  (ASR)   and  Pronunciation  Verifica- tion (PV) modules based on impaired speech utterances provided performance similar to SLPs [27].

4 Discussion

We conducted this systematic literature review based on a sample of 24 out of 678 research papers deriving from Scopus, IEEEXplore, and ACM DL databases. Exciting insights and trends emerged from our analysis of these papers. In recent years, we observed an increasing interest in AI-based auto- mated speech therapy tools. This growing interest can be due to the recent advancement in ASR technology and its improved accuracy. Surprisingly, 79 authors (86.81%) out of 91 unique authors have only one work on AI-based automated speech therapy in the last 15 years. This data suggests that a sig- nificant amount of research in this field is ”one-off” by authors. Most authors explored the research area with one idea and did not develop or evaluate it further. We found that ”automatic speech recognition” is the most empha- sized keyword by the authors. This finding is consistent with the notion that ASR is the core of AI-based automated speech therapy tools. The majority of studies were from European, North American, and Asian countries, and the most prevalent language targeted by the included studies was English. This finding is in line with the fact that English is the most widely adopted lan- guage for ASR technologies [35]. However, we can also observe that researchers have attempted to build AI-based automated speech therapy tools in other languages.

Furthermore, we found that articulation disorder was the most frequent disorder addressed by the included studies, with three studies dedicated to them. This may be due to the fact that articulation disorder are commonly found in persons with other SSD [2]. The results show that most studies aimed at developing fully automated speech therapy tools without considering the role of other stakeholders such as speech-language pathologists, caretakers, parents, and family members. This finding is in line with the widespread belief that researchers traditionally follow the one-dimensional framework of levels of automation by Sheridon and Verplank, which suggests that more automation leads to less human control and vice versa [36]. Moreover, a fully automated system may bring multiple concerns, such as biased data, privacy, replacement of jobs, and extreme automation may lead to disastrous consequences [33]. In the case of AI-based automated speech therapy, many concerns arise, including biased speech data, replacement of SLPs, and privacy of children’s speech data. We further found that mobile-based deployment of AI-based automated speech therapy was more common among the included studies. A possible explanation is that researchers are more interested in building affordable and accessible automated speech therapy tools. Another significant issue we observed is that few studies compared their automated tools’ results with human experts such as SLPs. This considerable insight questions the effectiveness of automated AI-based speech therapy tools compared to expert SLPs.

There are some limitations in our study which is worth mentioning. We relied on three databases: ACM DL, IEEE Xplore, and Scopus; therefore, we may have missed relevant papers published in other databases. Another limi- tation is the in-applicability of quality appraisal methods such as the ”Risk of Bias Assessment” in our study, as in the case of health sciences. Furthermore, our study restricts papers addressing SSD defined by ASHA and excluded studies addressing other related disorders.

5 Conclusion

This systematic literature review was based on the PRISMA Statement to analyze papers on AI-based automated speech therapy tools for persons with SSD. We extracted relevant data from the included articles based on four predefined research questions: Types of SSD addressed; Level of autonomy achieved by such tools; Modes of interventions and Effectiveness of such tools. Our study answers all the predefined research questions providing a snap- shot of the research carried out in the domain. We found that articulation disorder, hearing impairment, dysarthria, and motor speech were the most frequently studied disorders, addressed in three, two, two, and two studies (research question 1). However, 50% of the studies did not address any specific SSD. Concerning the level of autonomy (research question 2), almost all stud- ies proposed fully-automated AI-based speech therapy tools suggesting that researchers did not emphasize the role of caretaker, parents, family members, and SLPs. Addressing the modes of intervention (research question 3), most researchers proposed mobile-based and computer-based applications. Finally, the analysis of the effectiveness (research question 4) of such AI-based speech therapy tools provides us the insights that very few studies have compared their proposed system’s effectiveness with expert SLPs.

Based on the findings and insights from our research questions, we pro- pose the following directions for future research on AI-based automated speech therapy tools.

i.e., involving all stakeholders in the design process instead of focussing on developing fully automated tools.

modes of delivery (mobile device, computer, companion robots) and different modes of presentation (gamified content, storytelling).

Declarations

Conflict of interest

On behalf of all authors, the corresponding author states that there is no conflict of interest

References

  1. Speech Sound Disorder. https://www.asha.org/public/speech/disorders/ speech-sound-disorders/. Accessed: 2022-02-28
  2. Flipsen Jr, P.: Emergence and prevalence of persistent and residual speech errors. In: Seminars in Speech and Language, vol. 36, pp. 217–223 (2015). Thieme Medical Publishers
  3. Black, L.I., Vahratian, A., Hoffman, H.J.: Communication disorders and use of intervention services among children aged 3–17 years: United states, 2012. nchs data brief. number 205. Centers for Disease Control and Prevention (2015)
  4. Wren, Y., Miller, L.L., Peters, T.J., Emond, A., Roulstone, S.: Preva- lence and predictors of persistent speech sound disorder at eight years old: Findings from a population cohort study. Journal of Speech, Language, and Hearing Research 59(4), 647–673 (2016)
  5. Velayutham, B., Kangusamy, B., Joshua, V., Mehendale, S.: The preva- lence of disability in elderly in india–analysis of 2011 census data. Disability and health journal 9(4), 584–592 (2016)
  6. Sreeraj Konadath, S.C., Jayaram, G., Sandeep, M., Mahima, G., Shreyank, P.: Prevalence of communication disorders in a rural population of india. Journal of Hearing Science 3(2) (2013)
  7. Duval, J., Rubin, Z., Segura, E.M., Friedman, N., Zlatanov, M., Yang, L., Kurniawan, S.: Spokeit: building a mobile speech therapy experience. In: Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services, pp. 1–12 (2018)
  8. Robles-Bykbaev, V., Guam´an-Heredia, M., Robles-Bykbaev, Y., Lojano- Redrov´an, J., Pes´antez-Avil´es, F., Quisi-Peralta, D., L´opez-Nores, M., Pazos-Arias, J.: Onto-speltra: A robotic assistant based on ontologies and agglomerative clustering to support speech-language therapy for children with disabilities. In: Colombian Conference on Computing, pp. 343–357 (2017). Springer
  9. Groce, N., Deluca, M., Colde, E., Berman-Bieler, R., Mitra, G., Farkas, A., Sabbe, L., Burlyaeva-Norman, A., Lansdown, G.: Children and young people with disabilities: Fact sheet (UNICEF). https://citeseerx.ist. psu.edu/viewdoc/download?doi = 10.1.1.799.7313&rep = rep1&type = pdf. Accessed: 2022-02-28 (2013)
  10. Shamseer, L., Moher, D., Clarke, M., Ghersi, D., Liberati, A., Petticrew, M., Shekelle, P., Stewart, L.A.: Preferred reporting items for system- atic review and meta-analysis protocols (prisma-p) 2015: elaboration and explanation. Bmj 349 (2015)
  11. Van Eck, N.J., Waltman, L.: VOSviewer Manual. https://www.vosviewer. com. Accessed: 2022-02-28
  12. Desolda, G., Lanzilotti, R., Piccinno, A., Rossano, V.: A system to support children in speech therapies at home. In: CHItaly 2021: 14th Biannual Conference of the Italian SIGCHI Chapter, pp. 1–5 (2021)
  13. Ng, S.-I., Ng, C.W.-Y., Wang, J., Lee, T., Lee, K.Y.-S., Tong, M.C.-F.: Cuchild: A large-scale cantonese corpus of child speech for phonology and articulation assessment. arXiv preprint arXiv:2008.03188 (2020)
  14. Ng, S.I., Tao, D., Wang, J., Jiang, Y., Ng, W.Y., Lee, T.: An automated assessment tool for child speech disorders. In: 2018 11th International Symposium on Chinese Spoken Language Processing (ISCSLP), pp. 493– 494 (2018). IEEE
  15. Mahmut, E.E., Nicola, S., Stoicu-Tivadar, V.: A computer-based speech sound disorder screening system architecture. Data, Informatics and Technology: An Inspiration for Improved Healthcare, 39–42 (2018)
  16. Das, M., Saha, A.: An automated speech-language therapy tool with inter- active virtual agent and peer-to-peer feedback. In: 2017 4th International Conference on Advances in Electrical Engineering (ICAEE), pp. 510–515 (2017). IEEE
  17. Robles-Bykbaev, V.E., Guam´an-Murillo, W., Quisi-Peralta, D., L´opez- Nores, M., Pazos-Arias, J.J., Garc´ıa-Duque, J.: An ontology-based expert system to generate therapy plans for children with disabilities and com- munication disorders. In: 2016 IEEE Ecuador Technical Chapters Meeting (ETCM), pp. 1–6 (2016). IEEE
  18. Robles-Bykbaev, V., L´opez-Nores, M., Garc´ıa-Duque, J., Pazos-Arias, J.J., Ar´evalo-Lucero, D.: Evaluation of an expert system for the gener- ation of speech and language therapy plans. JMIR medical informatics 4(3), 5660 (2016)
  19. Robles-Bykbaev, V.E., L´opez-Nores, M., Pazos-Arias, J.J., Ar´evalo- Lucero, D.: Spelta: An expert system to generate therapy plans for speech and language disorders. Expert Systems with Applications 42(21), 7641–7651 (2015)
  20. Seddik, A.F., El Adawy, M., Shahin, A.I.: A computer-aided speech dis- orders correction system for arabic language. In: 2013 2nd International Conference on Advances in Biomedical Engineering, pp. 18–21 (2013). IEEE
  21. Duval, J.: Approaches for creating therapy games. ACM SIGACCESS Accessibility and Computing (126), 1–1 (2020)
  22. Samonte, M.J.C., Bahia, R.J.D., Forlaje, S.B.A., Del Monte, J.G.J., Gon- zales, J.A.J., Sultan, M.V.: Assistive mobile app for children with hearing & speech impairment using character and speech recognition. In: Pro- ceedings of the 4th International Conference on Industrial and Business Engineering, pp. 265–270 (2018)
  23. Sztah´o, D., Kiss, G., Vicsi, K.: Computer based speech prosody teaching system. Computer Speech & Language 50, 126–140 (2018)
  24. C´espedes-Hern´andez, D., P´erez-Medina, J.L., Gonz´alez-Calleros, J.M., Rodr´ıguez, F.J.A´., Mun˜oz-Arteaga, J.: Sega-arm: A metamodel for the design of serious games to support auditory rehabilitation. In: Proceedings of the XVI International Conference on Human Computer Interaction, pp. 1–8 (2015)
  25. B´ılkov´a, Z., Novoz´amsky`, A., Bartoˇs, M., Dom´ınec, A., Greˇsko, Sˇ., Zitov´a, B., Paroubkov´a, M., Flusser, J.: Human computer interface based on tongue and lips movements and its application for speech therapy system. Electronic Imaging 2020(1), 389–1 (2020)
  26. Frieg, H., Muehlhaus, J., Ritterfeld, U., Bilda, K.: Isi-speech: A digital training system for acquired dysarthria. In: AAATE Conf., pp. 330–334 (2017)
  27. Saz, O., Yin, S.-C., Lleida, E., Rose, R., Vaquero, C., Rodr´ıguez, W.R.: Tools and technologies for computer-aided speech and language therapy. Speech Communication 51(10), 948–967 (2009)
  28. Singh, S., Thakur, A., Vir, D.: Automatic articulation error detection tool for punjabi language with aid for hearing impaired people. International Journal of Speech Technology 18(2), 143–156 (2015)
  29. Chen, Y.-J., Huang, J.-W., Yang, H.-M., Lin, Y.-H., Wu, J.-L.: Devel- opment of articulation assessment and training system with speech recognition and articulation training strategies selection. In: 2007 IEEE International Conference on Acoustics, Speech and Signal Processing-ICASSP’07, vol. 4, p. 209 (2007). IEEE
  30. Ballard, K.J., Etter, N.M., Shen, S., Monroe, P., Tien Tan, C.: Feasibility of automatic speech recognition for providing feedback during tablet- based treatment for apraxia of speech plus aphasia. American journal of speech-language pathology 28(2S), 818–834 (2019)
  31. Ramamurthy, P., Li, T.: Buddy: a speech therapy robot companion for children with cleft lip and palate (cl/p) disorder. In: Companion of the 2018 ACM/IEEE International Conference on Human-Robot Interaction, pp. 359–360 (2018)
  32. Rivas, E.Q., Molina, E.S.: A proposal for a virtual world that supports theraphy of dyslalia. In: Proceedings of the 6th Euro American Conference on Telematics and Information Systems, pp. 371–374 (2012)
  33. Shneiderman, B.: Human-centered artificial intelligence: Reliable, safe & trustworthy. International Journal of Human–Computer Interaction 36(6), 495–504 (2020)
  34. Anjos, I., Grilo, M., Ascens˜ao, M., Guimar˜aes, I., Magalh˜aes, J., Cavaco, S.: A serious mobile game with visual feedback for training sibilant consonants. In: International Conference on Advances in Computer Entertainment, pp. 430–450 (2017). Springer
  35. Benzeghiba, M., De Mori, R., Deroo, O., Dupont, S., Erbes, T., Jouvet, D., Fissore, L., Laface, P., Mertins, A., Ris, C., et al.: Automatic speech recog- nition and speech variability: A review. Speech communication 49(10–11), 763–786 (2007)
  36. Sheridan, T.B., Verplank, W.L.: Human and computer control of undersea teleoperators. Technical report, Massachusetts Inst of Tech Cambridge Man-Machine Systems Lab (1978)