AI-Assisted Treatment for Residual Speech Sound Disorders

NCT ID: NCT05988515

Last Updated: 2026-01-08

Study Results

Results pending

The study team has not published outcome measurements, participant flow, or safety data for this trial yet. Check back later for updates.

Basic Information

Get a concise snapshot of the trial, including recruitment status, study phase, enrollment targets, and key timeline milestones.

Recruitment Status

RECRUITING

Clinical Phase

NA

Total Enrollment

26 participants

Study Classification

INTERVENTIONAL

Study Start Date

2024-09-05

Study Completion Date

2027-12-31

Brief Summary

Review the sponsor-provided synopsis that highlights what the study is about and why it is being conducted.

The goal of this randomized-controlled trial is to determine how artificial intelligence-assisted home practice may enhance speech learning of the "r" sound in school-age children with residual speech sound disorders. All child participants will receive 1 speech lesson per week, via telepractice, for 5 weeks with a human speech-language clinician. Some participants will receive 3 speech sessions per week with an Artificial Intelligence (AI)-clinician during the same 5 weeks as the human clinician sessions (CONCURRENT treatment order group), whereas others will receive 3 speech sessions per week with an AI-clinician after the human clinician sessions end (SEQUENTIAL treatment order group.

Detailed Description

Dive into the extended narrative that explains the scientific background, objectives, and procedures in greater depth.

Artificial Intelligence-assisted treatment that detects mispronunciations within an evidence-based motor learning framework could increase access to sufficiently intense, efficacious treatment despite provider shortages. A successful Artificial intelligencesystem that can predict the clinical gold standard of trained listeners' perceptions could not only improve access to clinical care but also mitigate known confounds to accurate clinical feedback, including clinical experience and drift due to increasing familiarity between the speaker and listener. The Artificial intelligence tool used in this study includes a speech classifier trained to predict clinician judgment of American English "r" that is integrated into an existing evidence-based treatment software called Speech Motor Chaining.

Conditions

See the medical conditions and disease areas that this research is targeting or investigating.

Speech Sound Disorder

Study Design

Understand how the trial is structured, including allocation methods, masking strategies, primary purpose, and other design elements.

Allocation Method

RANDOMIZED

Intervention Model

PARALLEL

Participants will be randomized across two conditions, CONCURRENT and SEQUENTIAL

Participants randomized to receive CONCURRENT treatment will receive speech lessons with a human clinician once per week for five weeks, along with 3 sessions per week of parent-supervised Chaining-AI (15 Chaining-AI sessions). Therefore, clinician-led and Chaining-AI are completed concurrently.

Participants randomized to receive SEQUENTIAL treatment will receive speech lessons with a human clinician once per week for five weeks, followed by three sessions per week of parent-supervised Chaining-AI for the NEXT five weeks. Therefore, clinician-led and Chaining-AI are completed sequentially.
Primary Study Purpose

TREATMENT

Blinding Strategy

SINGLE

Outcome Assessors
All perceptual ratings will be obtained from trained listeners who are blinded to treatment group and to timepoint. Binary rating of accuracy of the "r" sound (i.e., judgments of correct or incorrect) will be obtained from three listeners per token.

Study Groups

Review each arm or cohort in the study, along with the interventions and objectives associated with them.

CONCURRENT treatment order

* 5 speech lessons with a human speech-language clinician: 1 time per week for 5 weeks.
* 15 speech lessons with an AI clinician (supervised by the caregiver), 3 times per week DURING the same 5 weeks as the human clinician sessions.

Group Type EXPERIMENTAL

Speech-Language Pathologist-led Speech Motor Chaining

Intervention Type BEHAVIORAL

Sessions begin with Pre-practice to elicit the /r/ sound. During Structured Practice, the same utterance is practiced several times in a row (with systematic increases in difficulty based on performance). Our web-based software manipulates the principles of motor learning, including feedback prompts for the clinician, the complexity of the utterance, and the variability in the practice trial; the software will analyze the clinician's rating to increase the difficulty of practice when the child is more accurate. Randomized Practice will also be guided by the software and includes all linguistic levels that were produced correctly during Structured Practice, with items presented in random order. A trained speech-language pathologist is involved in all practice trials to provide feedback throughout the session.

Artificial Intelligence-led Speech Motor Chaining (CHAINING-AI)

Intervention Type BEHAVIORAL

Sessions include Structured Practice and Randomized Practice using our web-based software with an Artificial Intelligence clinician to address the /r/ sound. Within a practice session, participants speak into a microphone, and the audio file is sent to a server to be analyzed by a classifier, which returns a binary accurate/inaccurate rating of productions in a fashion similar to SLP judgment. Our web-based software manipulates the principles of motor learning, including feedback prompts, the complexity of the utterance, and the variability in the practice trial. The software will analyze the child's accuracy as determined by the classifier to increase the difficulty of practice when the child is more accurate.

SEQUENTIAL treatment order

* 5 speech lessons with a human speech-language clinician: 1 time per week for 5 weeks.
* 15 speech lessons with an AI clinician (supervised by the caregiver), 3 times per week for the 5 weeks AFTER the human clinician sessions end.

Group Type EXPERIMENTAL

Speech-Language Pathologist-led Speech Motor Chaining

Intervention Type BEHAVIORAL

Sessions begin with Pre-practice to elicit the /r/ sound. During Structured Practice, the same utterance is practiced several times in a row (with systematic increases in difficulty based on performance). Our web-based software manipulates the principles of motor learning, including feedback prompts for the clinician, the complexity of the utterance, and the variability in the practice trial; the software will analyze the clinician's rating to increase the difficulty of practice when the child is more accurate. Randomized Practice will also be guided by the software and includes all linguistic levels that were produced correctly during Structured Practice, with items presented in random order. A trained speech-language pathologist is involved in all practice trials to provide feedback throughout the session.

Artificial Intelligence-led Speech Motor Chaining (CHAINING-AI)

Intervention Type BEHAVIORAL

Sessions include Structured Practice and Randomized Practice using our web-based software with an Artificial Intelligence clinician to address the /r/ sound. Within a practice session, participants speak into a microphone, and the audio file is sent to a server to be analyzed by a classifier, which returns a binary accurate/inaccurate rating of productions in a fashion similar to SLP judgment. Our web-based software manipulates the principles of motor learning, including feedback prompts, the complexity of the utterance, and the variability in the practice trial. The software will analyze the child's accuracy as determined by the classifier to increase the difficulty of practice when the child is more accurate.

Interventions

Learn about the drugs, procedures, or behavioral strategies being tested and how they are applied within this trial.

Speech-Language Pathologist-led Speech Motor Chaining

Sessions begin with Pre-practice to elicit the /r/ sound. During Structured Practice, the same utterance is practiced several times in a row (with systematic increases in difficulty based on performance). Our web-based software manipulates the principles of motor learning, including feedback prompts for the clinician, the complexity of the utterance, and the variability in the practice trial; the software will analyze the clinician's rating to increase the difficulty of practice when the child is more accurate. Randomized Practice will also be guided by the software and includes all linguistic levels that were produced correctly during Structured Practice, with items presented in random order. A trained speech-language pathologist is involved in all practice trials to provide feedback throughout the session.

Intervention Type BEHAVIORAL

Artificial Intelligence-led Speech Motor Chaining (CHAINING-AI)

Sessions include Structured Practice and Randomized Practice using our web-based software with an Artificial Intelligence clinician to address the /r/ sound. Within a practice session, participants speak into a microphone, and the audio file is sent to a server to be analyzed by a classifier, which returns a binary accurate/inaccurate rating of productions in a fashion similar to SLP judgment. Our web-based software manipulates the principles of motor learning, including feedback prompts, the complexity of the utterance, and the variability in the practice trial. The software will analyze the child's accuracy as determined by the classifier to increase the difficulty of practice when the child is more accurate.

Intervention Type BEHAVIORAL

Eligibility Criteria

Check the participation requirements, including inclusion and exclusion rules, age limits, and whether healthy volunteers are accepted.

Inclusion Criteria

* Must speak a rhotic dialect of American English as a dominant language.
* Must have begun learning English by at least the age of 3 years.
* Must be between 9;0 to 17;11 years of age.
* Must have reported difficulty with /ɹ/ production.
* Must have reported hearing within normal limits.
* Must receive a Scaled Score of 5 or above on both the Listening Comprehension and Story Retelling subtests from the Test of Integrated Language \& Literacy Skills (TILLS).
* Must receive a percentile score of 8 or below on the Goldman-Fristoe Test of Articulation-3 (GFTA-3) Sounds in Words subtest.
* Must have 1 scorable response with 5+ consecutive correct /pataka/ with \> 3.4 syllables per second in the MRR-Tri task of the Maximum Performance Tasks OR must demonstrate no childhood apraxia of speech (CAS-only) features in BOTH articulatory and rate/prosody domains of the ProCAD.
* Must score \<40% accurate based on word-level items from our /ɹ/ probe list.
* Must score \>=15% accuracy on /ɹ/ on 45 syllables following Dynamic Assessment.
* Must express interest in changing their /ɹ/ production.
* Must have oral structure and function that are appropriate for /ɹ/ production.
* Must have access to broadband internet with videoconferencing capabilities

Exclusion Criteria

* Must have no known history of autism spectrum disorder, Down Syndrome, cerebral palsy, intellectual disability, permanent hearing loss, epilepsy/antiepileptic medication, or brain injury/neurosurgery/stroke.
* Must not have diagnosis of attention deficit disorder, attention deficit hyperactivity disorder, Tourette's, or Obsessive-compulsive disorder.
* Must have no orthodontic appliances that block the roof of the mouth (e.g., palate expanders).
* Must not have current cleft palate, fluency disorder, or voice disorder.
* Must not demonstrate childhood apraxia of speech (CAS-only) features in BOTH articulatory and rate/prosody domains of the ProCAD.
Minimum Eligible Age

9 Years

Maximum Eligible Age

17 Years

Eligible Sex

ALL

Accepts Healthy Volunteers

Yes

Sponsors

Meet the organizations funding or collaborating on the study and learn about their roles.

National Institutes of Health (NIH)

NIH

Sponsor Role collaborator

State University of New York - Upstate Medical University

OTHER

Sponsor Role collaborator

National Institute on Deafness and Other Communication Disorders (NIDCD)

NIH

Sponsor Role collaborator

Syracuse University

OTHER

Sponsor Role lead

Responsible Party

Identify the individual or organization who holds primary responsibility for the study information submitted to regulators.

Responsibility Role SPONSOR

Locations

Explore where the study is taking place and check the recruitment status at each participating site.

Syracuse University

Syracuse, New York, United States

Site Status RECRUITING

Countries

Review the countries where the study has at least one active or historical site.

United States

Central Contacts

Reach out to these primary contacts for questions about participation or study logistics.

Jonathan Preston, PhD

Role: CONTACT

315-443-1351

Nina Benway, PhD

Role: CONTACT

Facility Contacts

Find local site contact details for specific facilities participating in the trial.

Jonathan Preston, PhD

Role: primary

315-443-3143

Nicole Caballero, MS

Role: backup

315-443-1185

Related Links

Access external resources that provide additional context or updates about the study.

https://speechproductionlab.syr.edu/research/

Lab website with recruitment information

Other Identifiers

Review additional registry numbers or institutional identifiers associated with this trial.

1R01DC020959-01

Identifier Type: NIH

Identifier Source: secondary_id

View Link

22-426b

Identifier Type: -

Identifier Source: org_study_id

More Related Trials

Additional clinical trials that may be relevant based on similarity analysis.

Neural Prediction to Enhance Language
NCT05962359 RECRUITING NA
Stimulus-Stimulus Pairing Study
NCT02464527 COMPLETED NA