3 Sources
[1]
Automating speech screening for children through AI innovation
Acoustical Society of AmericaMay 19 2025 Speech and language impairments affect over a million children every year, and identifying and treating these conditions early is key to helping these children overcome them. Clinicians struggling with time, resources, and access are in desperate need of tools to make diagnosing speech impairments faster and more accurate. Marisha Speights, assistant professor at Northwestern University, built a data pipeline to train clinical artificial intelligence tools for childhood speech screening. She will present her work Monday, May 19, at 8:20 a.m. CT as part of the joint 188th Meeting of the Acoustical Society of America and 25th International Congress on Acoustics, running May 18-23. AI-based speech recognition and clinical diagnostic tools have been in use for years, but these tools are typically trained and used exclusively on adult speech. That makes them unsuitable for clinical work involving children. New AI tools must be developed, but there are no large datasets of recorded child speech for these tools to be trained on, in part because building these datasets is uniquely challenging. There's a common misconception that collecting speech from children is as straightforward as it is with adults - but in reality, it requires a much more controlled and developmentally sensitive process. Unlike adult speech, child speech is highly variable, acoustically distinct, and underrepresented in most training corpora." Marisha Speights, Assistant Professor, Northwestern University To remedy this, Speights and her colleagues began collecting and analyzing large volumes of child speech recordings to build such a dataset. However, they quickly realized a problem: The collection, processing, and annotation of thousands of speech samples is difficult without exactly the kind of automated tools they were trying to build. "It's a bit of a catch-22," said Speights. "We need automated tools to scale data collection, but we need large datasets to train those tools." In response, the researchers built a computational pipeline to turn raw speech data into a useful dataset for training AI tools. They collected a representative sample of speech from children across the country, verified transcripts and enhanced audio quality using their custom software, and provided a platform that will enable detailed annotation by experts. The result is a high-quality dataset that can be used to train clinical AI, giving experts access to a powerful set of tools to make diagnosing speech impairments much easier. "Speech-language pathologists, health care clinicians and educators will be able to use AI-powered systems to flag speech-language concerns earlier, especially in places where access to specialists is limited," said Speights. Acoustical Society of America
[2]
Laying the groundwork to diagnose speech impairments in children with clinical AI tool
Speech and language impairments affect over a million children every year, and identifying and treating these conditions early is key to helping these children overcome them. Clinicians struggling with time, resources, and access are in desperate need of tools to make diagnosing speech impairments faster and more accurate. Marisha Speights, assistant professor at Northwestern University, built a data pipeline to train clinical artificial intelligence tools for childhood speech screening. She presented her work at the joint 188th Meeting of the Acoustical Society of America and 25th International Congress on Acoustics, held May 18-23. AI-based speech recognition and clinical diagnostic tools have been in use for years, but these tools are typically trained and used exclusively on adult speech. That makes them unsuitable for clinical work involving children. New AI tools must be developed, but there are no large datasets of recorded child speech for these tools to be trained on, in part because building these datasets is uniquely challenging. "There's a common misconception that collecting speech from children is as straightforward as it is with adults -- but in reality, it requires a much more controlled and developmentally sensitive process," said Speights. "Unlike adult speech, child speech is highly variable, acoustically distinct, and underrepresented in most training corpora." To remedy this, Speights and her colleagues began collecting and analyzing large volumes of child speech recordings to build such a dataset. However, they quickly realized a problem: The collection, processing, and annotation of thousands of speech samples is difficult without exactly the kind of automated tools they were trying to build. "It's a bit of a catch-22," said Speights. "We need automated tools to scale data collection, but we need large datasets to train those tools." In response, the researchers built a computational pipeline to turn raw speech data into a useful dataset for training AI tools. They collected a representative sample of speech from children across the country, verified transcripts and enhanced audio quality using their custom software, and provided a platform that will enable detailed annotation by experts. The result is a high-quality dataset that can be used to train clinical AI, giving experts access to a powerful set of tools to make diagnosing speech impairments much easier. "Speech-language pathologists, health care clinicians and educators will be able to use AI-powered systems to flag speech-language concerns earlier, especially in places where access to specialists is limited," said Speights.
[3]
Laying the Groundwork to Diagnose Speech Impairments in Children with Clinical AI | Newswise
Newswise -- NEW ORLEANS, May 19, 2025 - Speech and language impairments affect over a million children every year, and identifying and treating these conditions early is key to helping these children overcome them. Clinicians struggling with time, resources, and access are in desperate need of tools to make diagnosing speech impairments faster and more accurate. Marisha Speights, assistant professor at Northwestern University, built a data pipeline to train clinical artificial intelligence tools for childhood speech screening. She will present her work Monday, May 19, at 8:20 a.m. CT as part of the joint 188th Meeting of the Acoustical Society of America and 25th International Congress on Acoustics, running May 18-23. AI-based speech recognition and clinical diagnostic tools have been in use for years, but these tools are typically trained and used exclusively on adult speech. That makes them unsuitable for clinical work involving children. New AI tools must be developed, but there are no large datasets of recorded child speech for these tools to be trained on, in part because building these datasets is uniquely challenging. "There's a common misconception that collecting speech from children is as straightforward as it is with adults -- but in reality, it requires a much more controlled and developmentally sensitive process," said Speights. "Unlike adult speech, child speech is highly variable, acoustically distinct, and underrepresented in most training corpora." To remedy this, Speights and her colleagues began collecting and analyzing large volumes of child speech recordings to build such a dataset. However, they quickly realized a problem: The collection, processing, and annotation of thousands of speech samples is difficult without exactly the kind of automated tools they were trying to build. "It's a bit of a catch-22," said Speights. "We need automated tools to scale data collection, but we need large datasets to train those tools." In response, the researchers built a computational pipeline to turn raw speech data into a useful dataset for training AI tools. They collected a representative sample of speech from children across the country, verified transcripts and enhanced audio quality using their custom software, and provided a platform that will enable detailed annotation by experts. The result is a high-quality dataset that can be used to train clinical AI, giving experts access to a powerful set of tools to make diagnosing speech impairments much easier. "Speech-language pathologists, health care clinicians and educators will be able to use AI-powered systems to flag speech-language concerns earlier, especially in places where access to specialists is limited," said Speights. In the coming weeks, ASA's Press Room will be updated with newsworthy stories and the press conference schedule at https://acoustics.org/asa-press-room/. LAY LANGUAGE PAPERS ASA will also share dozens of lay language papers about topics covered at the conference. Lay language papers are summaries (300-500 words) of presentations written by scientists for a general audience. They will be accompanied by photos, audio, and video. Learn more at https://acoustics.org/lay-language-papers/. PRESS REGISTRATION ASA will grant free registration to credentialed and professional freelance journalists. If you are a reporter and would like to attend the in-person meeting or virtual press conferences, contact AIP Media Services at [email protected]. For urgent requests, AIP staff can also help with setting up interviews and obtaining images, sound clips, or background information. ABOUT THE ACOUSTICAL SOCIETY OF AMERICA The Acoustical Society of America is the premier international scientific society in acoustics devoted to the science and technology of sound. Its 7,000 members worldwide represent a broad spectrum of the study of acoustics. ASA publications include The Journal of the Acoustical Society of America (the world's leading journal on acoustics), JASA Express Letters, Proceedings of Meetings on Acoustics, Acoustics Today magazine, books, and standards on acoustics. The society also holds two major scientific meetings each year. See https://acousticalsociety.org/. ABOUT THE INTERNATIONAL COMMISSION FOR ACOUSTICS The purpose of the International Commission for Acoustics (ICA) is to promote international development and collaboration in all fields of acoustics including research, development, education, and standardization. ICA's mission is to be the reference point for the acoustic community, becoming more inclusive and proactive in our global outreach, increasing coordination and support for the growing international interest and activity in acoustics. Learn more at https://www.icacommission.org/.
Share
Copy Link
Researchers at Northwestern University develop a data pipeline to train AI tools for childhood speech screening, addressing the unique challenges of collecting and analyzing child speech data.
Speech and language impairments affect over a million children annually, with early identification and treatment being crucial for overcoming these challenges. Clinicians face significant hurdles in diagnosing speech impairments due to limited time, resources, and access to specialized tools. To address this pressing issue, researchers at Northwestern University have developed an innovative approach to automate speech screening for children using artificial intelligence (AI) 123.
Marisha Speights, an assistant professor at Northwestern University, has spearheaded the development of a data pipeline to train clinical AI tools specifically for childhood speech screening. This groundbreaking work was presented at the joint 188th Meeting of the Acoustical Society of America and 25th International Congress on Acoustics in May 2025 123.
While AI-based speech recognition and clinical diagnostic tools have been in use for years, they have primarily focused on adult speech, making them unsuitable for pediatric applications. Speights explains the unique challenges in collecting child speech data:
"There's a common misconception that collecting speech from children is as straightforward as it is with adults -- but in reality, it requires a much more controlled and developmentally sensitive process. Unlike adult speech, child speech is highly variable, acoustically distinct, and underrepresented in most training corpora." 123
The research team faced a significant hurdle in their efforts to build a comprehensive dataset of child speech recordings. They encountered a "catch-22" situation where automated tools were needed to scale data collection, but large datasets were required to train those very tools 123.
To overcome this challenge, Speights and her colleagues developed a computational pipeline that transforms raw speech data into a useful dataset for training AI tools. Their approach involved:
The resulting high-quality dataset has the potential to revolutionize the field of speech-language pathology and pediatric healthcare. By training clinical AI tools with this comprehensive child speech data, experts will gain access to powerful diagnostic tools that can significantly improve the efficiency and accuracy of speech impairment diagnoses 123.
Speights highlights the far-reaching impact of this innovation:
"Speech-language pathologists, healthcare clinicians, and educators will be able to use AI-powered systems to flag speech-language concerns earlier, especially in places where access to specialists is limited." 123
This advancement in AI-powered speech screening for children has the potential to transform early intervention strategies, improve access to specialized care, and ultimately enhance the quality of life for millions of children affected by speech and language impairments.
Databricks raises $1 billion in a new funding round, valuing the company at over $100 billion. The data analytics firm plans to invest in AI database technology and an AI agent platform, positioning itself for growth in the evolving AI market.
11 Sources
Business
14 hrs ago
11 Sources
Business
14 hrs ago
SoftBank makes a significant $2 billion investment in Intel, boosting the chipmaker's efforts to regain its competitive edge in the AI semiconductor market.
22 Sources
Business
22 hrs ago
22 Sources
Business
22 hrs ago
OpenAI introduces ChatGPT Go, a new subscription plan priced at ₹399 ($4.60) per month exclusively for Indian users, offering enhanced features and affordability to capture a larger market share.
15 Sources
Technology
22 hrs ago
15 Sources
Technology
22 hrs ago
Microsoft introduces a new AI-powered 'COPILOT' function in Excel, allowing users to perform complex data analysis and content generation using natural language prompts within spreadsheet cells.
8 Sources
Technology
14 hrs ago
8 Sources
Technology
14 hrs ago
Adobe launches Acrobat Studio, integrating AI assistants and PDF Spaces to transform document management and collaboration, marking a significant evolution in PDF technology.
10 Sources
Technology
13 hrs ago
10 Sources
Technology
13 hrs ago