Neuroscience News: AI Speech Model detection…

AI Speech Model Detects Neurological Disorders With 92% Accuracy

FeaturedNeurologyNeuroscience

August 29, 2025

Summary: A new AI framework can detect neurological disorders by analyzing speech with over 90% accuracy. The model, called CTCAIT, captures subtle patterns in voice that may indicate early symptoms of diseases like Parkinson’s, Huntington’s, and Wilson disease.

Unlike traditional methods, it integrates multi-scale temporal features and attention mechanisms, making it both highly accurate and interpretable. The findings highlight speech as a promising tool for non-invasive, accessible early diagnosis and monitoring of neurological conditions.

Key Facts

  • High Accuracy: 92.06% accuracy in Mandarin, 87.73% in English datasets.
  • Non-Invasive Biomarker: Speech abnormalities can reveal early neurodegenerative changes.
  • Broad Potential: Could be used for screening and monitoring across multiple neurological diseases.

Source: Chinese Academy of Science

Recently, the research team led by Prof. LI Hai at the Institute of Health and Medical Technology, the Hefei Institutes of Physical Science of the Chinese Academy of Sciences, has developed a novel deep learning framework that significantly improves the accuracy and interpretability of detecting neurological disorders through speech. 

“A slight change in the way we speak might be more than just a slip of the tongue—it could be a warning sign from the brain,” said Prof. LI Hai, who led the team, “Our new model can detect early symptoms of neurological diseases like Parkinson’ s, Huntington’ s, and Wilson disease—by analyzing voice recordings.”

This shows a brain and sound waves.
The method achieved a detection accuracy of 92.06% on a Mandarin Chinese dataset and 87.73% on an external English dataset, demonstrating strong cross-linguistic generalizability. Credit: Neuroscience News

The study was recently published in Neurocomputing.

Dysarthria is a common early symptom of various neurological disorders. Given that these speech abnormalities often reflect underlying neurodegenerative processes, voice signals have emerged as promising non-invasive biomarkers for early screening and continuous monitoring of such conditions. Automated speech analysis offers high efficiency, low cost, and non-invasiveness.

However, current mainstream methods often suffer from over-reliance on handcrafted features, limited capacity to model temporal-variable interactions, and poor interpretability.

To address these challenges, the team proposed Cross-Time and Cross-Axis Interactive Transformer (CTCAIT) for multivariate time series analysis. This framework first employs a large-scale audio model to extract high-dimensional temporal features from speech, representing them as multidimensional embeddings along time and feature axes.

It then leverages the Inception Time network to capture multi-scale and multi-level patterns within the time series. By integrating cross-time and cross-channel multi-head attention mechanisms, CTCAIT effectively captures pathological speech signatures embedded across different dimensions.

The method achieved a detection accuracy of 92.06% on a Mandarin Chinese dataset and 87.73% on an external English dataset, demonstrating strong cross-linguistic generalizability.

Furthermore, the team conducted interpretability analyses of the model’s internal decision-making processes and systematically compared the effectiveness of different speech tasks, offering valuable insights for its potential clinical deployment.

These efforts provide important guidance for potential clinical applications of the method in early diagnosis and monitoring of neurological disorders.

About this AI and neurology research news

Author: Weiwei Zhao
Source: Chinese Academy of Science
Contact: Weiwei Zhao – Chinese Academy of Science
Image: The image is credited to Neuroscience News

FeaturedNeurologyNeuroscience

·August 29, 2025

Summary: A new AI framework can detect neurological disorders by analyzing speech with over 90% accuracy. The model, called CTCAIT, captures subtle patterns in voice that may indicate early symptoms of diseases like Parkinson’s, Huntington’s, and Wilson disease.

Unlike traditional methods, it integrates multi-scale temporal features and attention mechanisms, making it both highly accurate and interpretable. The findings highlight speech as a promising tool for non-invasive, accessible early diagnosis and monitoring of neurological conditions.

Key Facts

  • High Accuracy: 92.06% accuracy in Mandarin, 87.73% in English datasets.
  • Non-Invasive Biomarker: Speech abnormalities can reveal early neurodegenerative changes.
  • Broad Potential: Could be used for screening and monitoring across multiple neurological diseases.

Source: Chinese Academy of Science

Recently, the research team led by Prof. LI Hai at the Institute of Health and Medical Technology, the Hefei Institutes of Physical Science of the Chinese Academy of Sciences, has developed a novel deep learning framework that significantly improves the accuracy and interpretability of detecting neurological disorders through speech. 

“A slight change in the way we speak might be more than just a slip of the tongue—it could be a warning sign from the brain,” said Prof. LI Hai, who led the team, “Our new model can detect early symptoms of neurological diseases like Parkinson’ s, Huntington’ s, and Wilson disease—by analyzing voice recordings.”

This shows a brain and sound waves.
The method achieved a detection accuracy of 92.06% on a Mandarin Chinese dataset and 87.73% on an external English dataset, demonstrating strong cross-linguistic generalizability. Credit: Neuroscience News

The study was recently published in Neurocomputing.

Dysarthria is a common early symptom of various neurological disorders. Given that these speech abnormalities often reflect underlying neurodegenerative processes, voice signals have emerged as promising non-invasive biomarkers for early screening and continuous monitoring of such conditions. Automated speech analysis offers high efficiency, low cost, and non-invasiveness.

However, current mainstream methods often suffer from over-reliance on handcrafted features, limited capacity to model temporal-variable interactions, and poor interpretability.

To address these challenges, the team proposed Cross-Time and Cross-Axis Interactive Transformer (CTCAIT) for multivariate time series analysis. This framework first employs a large-scale audio model to extract high-dimensional temporal features from speech, representing them as multidimensional embeddings along time and feature axes.

It then leverages the Inception Time network to capture multi-scale and multi-level patterns within the time series. By integrating cross-time and cross-channel multi-head attention mechanisms, CTCAIT effectively captures pathological speech signatures embedded across different dimensions.

The method achieved a detection accuracy of 92.06% on a Mandarin Chinese dataset and 87.73% on an external English dataset, demonstrating strong cross-linguistic generalizability.

Furthermore, the team conducted interpretability analyses of the model’s internal decision-making processes and systematically compared the effectiveness of different speech tasks, offering valuable insights for its potential clinical deployment.

These efforts provide important guidance for potential clinical applications of the method in early diagnosis and monitoring of neurological disorders.

About this AI and neurology research news

Author: Weiwei Zhao
Source: Chinese Academy of Science
Contact: Weiwei Zhao – Chinese Academy of Science
Image: The image is credited to Neuroscience News

Unknown's avatar

About michelleclarke2015

Life event that changes all: Horse riding accident in Zimbabwe in 1993, a fractured skull et al including bipolar anxiety, chronic fatigue …. co-morbidities (Nietzche 'He who has the reason why can deal with any how' details my health history from 1993 to date). 17th 2017 August operation for breast cancer (no indications just an appointment came from BreastCheck through the Post). Trinity College Dublin Business Economics and Social Studies (but no degree) 1997-2003; UCD 1997/1998 night classes) essays, projects, writings. Trinity Horizon Programme 1997/98 (Centre for Women Studies Trinity College Dublin/St. Patrick's Foundation (Professor McKeon) EU Horizon funded: research study of 15 women (I was one of this group and it became the cornerstone of my journey to now 2017) over 9 mth period diagnosed with depression and their reintegration into society, with special emphasis on work, arts, further education; Notes from time at Trinity Horizon Project 1997/98; Articles written for Irishhealth.com 2003/2004; St Patricks Foundation monthly lecture notes for a specific period in time; Selection of Poetry including poems written by people I know; Quotations 1998-2017; other writings mainly with theme of social justice under the heading Citizen Journalism Ireland. Letters written to friends about life in Zimbabwe; Family history including Michael Comyn KC, my grandfather, my grandmother's family, the O'Donnellan ffrench Blake-Forsters; Moral wrong: An acrimonious divorce but the real injustice was the Catholic Church granting an annulment – you can read it and make your own judgment, I have mine. Topics I have written about include annual Brain Awareness week, Mashonaland Irish Associataion in Zimbabwe, Suicide (a life sentence to those left behind); Nostalgia: Tara Hill, Co. Meath.
This entry was posted in Uncategorized and tagged , , , , . Bookmark the permalink.

Leave a comment