3 Sources
[1]
AI model can diagnose depression via speech and brain neural activity
Depression is one of the most common mental illnesses. As many as 280 million people worldwide are affected by this disease, which is why researchers at Kaunas University of Technology (KTU) have developed an artificial intelligence (AI) model that helps to identify depression based on both speech and brain neural activity. This multimodal approach, combining two different data sources, allows a more accurate and objective analysis of a person's emotional state, opening the door to a new phase of depression diagnosis. "Depression is one of the most common mental disorders, with devastating consequences for both the individual and society, so we are developing a new, more objective diagnostic method that could become accessible to everyone in the future," says Rytis Maskeliūnas, a professor at KTU and one of the authors of the invention. Scientists argue that while most diagnostic research for depression has traditionally relied on a single type of data, the new multimodal approach can provide better information about a person's emotional state. The article, "Multimodal Fusion of EEG and Audio Spectrogram for Major Depressive Disorder Recognition Using Modified DenseNet121," was published in Brain Sciences Journal. Impressive accuracy using voice and brain activity data This combination of speech and brain activity data achieved an impressive 97.53% accuracy in diagnosing depression, significantly outperforming alternative methods. "This is because the voice adds data to the study that we cannot yet extract from the brain," explains Maskeliūnas. According to Musyyab Yousufi, KTU Ph.D. student who contributed to the invention, the choice of data was carefully considered. "While it is believed that facial expressions might reveal more about a person's psychological state, this is quite easily falsifiable data. We chose voice because it can subtly reveal an emotional state, with the diagnosis affecting the pace of speech, intonation, and overall energy." In addition, unlike electrical brain activity (EEG) or voice data, the face can directly identify a person's state of severity up to a certain extent. "But we cannot violate patients' privacy, and also, collecting and combining data from several sources is more promising for further use," says the professor at KTU Faculty of Informatics (IF). Maskeliūnas emphasizes that the used EEG dataset was obtained from the Multimodal Open Dataset for Mental Disorder Analysis (MODMA), as the KTU research group represents computer science and not the medical science field. MODMA EEG data was collected and recorded for five minutes while participants were awake, at rest, and with their eyes closed. In the audio part of the experiment, the patients participated in a question-and-answer session and several activities focused on reading and describing pictures to capture their natural language and cognitive state. AI will need to learn how to justify the diagnosis The collected EEG and audio signals were transformed into spectrograms, allowing the data to be visualized. Special noise filters and pre-processing methods were applied to make the data noise free and comparable, and a modified DenseNet-121 deep-learning model was used to identify signs of depression in the images. Each image reflected signal changes over time. The EEG showed waveforms of brain activity, and the sound showed frequency and intensity distributions. The model included a custom classification layer trained to split the data into classes of healthy or depressed people. Successful classification was evaluated and then the accuracy of the application was assessed. In the future, this AI model could speed up the diagnosis of depression, or even make it remote, and reduce the risk of subjective evaluations. This requires further clinical trials and improvements to the program. However, Maskeliūnas adds, that the latter aspect of the research might raise some challenges. "The main problem with these studies is the lack of data because people tend to remain private about their mental health matters," he says. Another important aspect mentioned by the professor of the KTU Department of Multimedia Engineering is that the algorithm needs to be improved in such a way that it is not only accurate but also provides information to the medical professional on what led to this diagnostic result. "The algorithm still has to learn how to explain the diagnosis in a comprehensible way," says Maskeliūnas. According to a KTU professor, due to the growing demand for AI solutions that directly affect people in areas such as health care, finance, and the legal system, similar requirements are becoming common. This is why explainable artificial intelligence (XAI), which aims to explain to the user why the model makes certain decisions and to increase their trust in the AI, is now gaining momentum.
[2]
Researchers use artificial intelligence to diagnose depression
Depression is one of the most common mental illnesses. As many as 280 million people worldwide are affected by this disease, which is why researchers at Kaunas University of Technology (KTU) have developed an artificial intelligence (AI) model that helps to identify depression based on both speech and brain neural activity. This multimodal approach, combining two different data sources, allows a more accurate and objective analysis of a person's emotional state, opening the door to a new phase of depression diagnosis. "Depression is one of the most common mental disorders, with devastating consequences for both the individual and society, so we are developing a new, more objective diagnostic method that could become accessible to everyone in the future," says Rytis Maskeliūnas, a professor at KTU and one of the authors of the invention. Scientists argue that while most diagnostic research for depression has traditionally relied on a single type of data, the new multimodal approach can provide better information about a person's emotional state. Impressive accuracy using voice and brain activity data This combination of speech and brain activity data achieved an impressive 97.53 per cent accuracy in diagnosing depression, significantly outperforming alternative methods. "This is because the voice adds data to the study that we cannot yet extract from the brain," explains Maskeliūnas. According to Musyyab Yousufi, KTU PhD student who contributed to the invention, the choice of data was carefully considered: "While it is believed that, facial expressions might reveal more about a person's psychological state, but this is quite easily falsifiable data. We chose voice because it can subtly reveal an emotional state, with the diagnosis affecting the pace of speech, intonation, and overall energy." In addition, unlike electrical brain activity (EEG) or voice data, the face can directly identify a person's state of severity up to certain extent. "But we cannot violate patients' privacy, and also, collecting and combining data from several sources is more promising for further use," says the professor at KTU Faculty of Informatics (IF). Maskeliūnas emphasises that the used EEG dataset was obtained from the Multimodal Open Dataset for Mental Disorder Analysis (MODMA), as the KTU research group represents computer science and not the medical science field. MODMA EEG data was collected and recorded for five minutes while participants were awake, at rest, and with their eyes closed. In the audio part of the experiment, the patients participated in a question-and-answer session and several activities focused on reading and describing pictures to capture their natural language and cognitive state. AI will need to learn how to justify the diagnosis The collected EEG and audio signals were transformed into spectrograms, allowing the data to be visualised. Special noise filters and pre-processing methods were applied to make the data noise free and comparable, and a modified DenseNet-121 deep-learning model was used to identify signs of depression in the images. Each image reflected signal changes over time. The EEG showed waveforms of brain activity, and the sound showed frequency and intensity distributions. The model included a custom classification layer trained to split the data into classes of healthy or depressed people. Successful classification was evaluated and then the accuracy of the application was assessed. In the future, this AI model could speed up the diagnosis of depression, or even make it remote, and reduce the risk of subjective evaluations. This requires further clinical trials and improvements to the programme. However, Maskeliūnas adds, that the latter aspect of the research might raise some challenges. "The main problem with these studies is the lack of data because people tend to remain private about their mental health matters," he says. Another important aspect mentioned by the professor of the KTU Department of Multimedia Engineering is that the algorithm needs to be improved in such a way that it is not only accurate but also provides information to the medical professional on what led to this diagnostic result. "The algorithm still has to learn how to explain the diagnosis in a comprehensible way," says Maskeliūnas. According to a KTU professor, due to the growing demand for AI solutions that directly affect people in areas such as healthcare, finance, and the legal system, similar requirements are becoming common. This is why explainable artificial intelligence(XAI), which aims to explain to the user why the model makes certain decisions and to increase their trust in the AI, is now gaining momentum.
[3]
AI model combines speech and brain activity to diagnose depression
Kaunas University of TechnologyNov 20 2024 Depression is one of the most common mental illnesses. As many as 280 million people worldwide are affected by this disease, which is why researchers at Kaunas University of Technology (KTU) have developed an artificial intelligence (AI) model that helps to identify depression based on both speech and brain neural activity. This multimodal approach, combining two different data sources, allows a more accurate and objective analysis of a person's emotional state, opening the door to a new phase of depression diagnosis. Depression is one of the most common mental disorders, with devastating consequences for both the individual and society, so we are developing a new, more objective diagnostic method that could become accessible to everyone in the future." Rytis Maskeliūnas, professor at KTU and one of the authors of the invention Scientists argue that while most diagnostic research for depression has traditionally relied on a single type of data, the new multimodal approach can provide better information about a person's emotional state. Impressive accuracy using voice and brain activity data This combination of speech and brain activity data achieved an impressive 97.53 per cent accuracy in diagnosing depression, significantly outperforming alternative methods. "This is because the voice adds data to the study that we cannot yet extract from the brain," explains Maskeliūnas. According to Musyyab Yousufi, KTU PhD student who contributed to the invention, the choice of data was carefully considered: "While it is believed that, facial expressions might reveal more about a person's psychological state, but this is quite easily falsifiable data. We chose voice because it can subtly reveal an emotional state, with the diagnosis affecting the pace of speech, intonation, and overall energy". In addition, unlike electrical brain activity (EEG) or voice data, the face can directly identify a person's state of severity up to certain extent. "But we cannot violate patients' privacy, and also, collecting and combining data from several sources is more promising for further use," says the professor at KTU Faculty of Informatics (IF). Maskeliūnas emphasises that the used EEG dataset was obtained from the Multimodal Open Dataset for Mental Disorder Analysis (MODMA), as the KTU research group represents computer science and not the medical science field. MODMA EEG data was collected and recorded for five minutes while participants were awake, at rest, and with their eyes closed. In the audio part of the experiment, the patients participated in a question-and-answer session and several activities focused on reading and describing pictures to capture their natural language and cognitive state. AI will need to learn how to justify the diagnosis The collected EEG and audio signals were transformed into spectrograms, allowing the data to be visualised. Special noise filters and pre-processing methods were applied to make the data noise free and comparable, and a modified DenseNet-121 deep-learning model was used to identify signs of depression in the images. Each image reflected signal changes over time. The EEG showed waveforms of brain activity, and the sound showed frequency and intensity distributions. The model included a custom classification layer trained to split the data into classes of healthy or depressed people. Successful classification was evaluated and then the accuracy of the application was assessed. In the future, this AI model could speed up the diagnosis of depression, or even make it remote, and reduce the risk of subjective evaluations. This requires further clinical trials and improvements to the programme. However, Maskeliūnas adds, that the latter aspect of the research might raise some challenges. "The main problem with these studies is the lack of data because people tend to remain private about their mental health matters," he says. Another important aspect mentioned by the professor of the KTU Department of Multimedia Engineering is that the algorithm needs to be improved in such a way that it is not only accurate but also provides information to the medical professional on what led to this diagnostic result. "The algorithm still has to learn how to explain the diagnosis in a comprehensible way," says Maskeliūnas. According to a KTU professor, due to the growing demand for AI solutions that directly affect people in areas such as healthcare, finance, and the legal system, similar requirements are becoming common. This is why explainable artificial intelligence (XAI), which aims to explain to the user why the model makes certain decisions and to increase their trust in the AI, is now gaining momentum. Kaunas University of Technology Journal reference: Yousufi, M., et al. (2024). Multimodal Fusion of EEG and Audio Spectrogram for Major Depressive Disorder Recognition Using Modified DenseNet121. Brain Sciences. doi.org/10.3390/brainsci14101018.
Share
Copy Link
Researchers at Kaunas University of Technology have developed an AI model that combines speech and brain neural activity data to diagnose depression with high accuracy, potentially revolutionizing mental health diagnostics.
Researchers at Kaunas University of Technology (KTU) have developed a groundbreaking artificial intelligence model that combines speech and brain neural activity data to diagnose depression with remarkable accuracy. This multimodal approach represents a significant advancement in mental health diagnostics, potentially offering a more objective and accessible method for identifying one of the world's most common mental illnesses 1.
The AI model utilizes a combination of speech patterns and electroencephalogram (EEG) data to analyze a person's emotional state. This dual-data approach achieved an impressive 97.53% accuracy in diagnosing depression, surpassing traditional single-data methods 2.
Professor Rytis Maskeliūnas, one of the inventors, explains: "The voice adds data to the study that we cannot yet extract from the brain," highlighting the synergistic effect of combining these two data sources 3.
The research team used EEG data from the Multimodal Open Dataset for Mental Disorder Analysis (MODMA). Participants' brain activity was recorded for five minutes while they were awake, at rest, and with their eyes closed. For the audio component, subjects engaged in question-and-answer sessions and activities designed to capture their natural language and cognitive state 1.
Both EEG and audio signals were transformed into spectrograms for visualization. The team applied noise filters and pre-processing methods to ensure data quality. A modified DenseNet-121 deep-learning model was then used to identify depression indicators in these images 2.
This AI model has the potential to expedite depression diagnosis, possibly even enabling remote assessments, and reducing the risk of subjective evaluations. However, the researchers acknowledge that further clinical trials and improvements are necessary before widespread implementation 3.
One significant challenge highlighted by Professor Maskeliūnas is the scarcity of data, as people tend to be private about their mental health. Additionally, the team emphasizes the need for the AI to provide explanations for its diagnoses, aligning with the growing field of explainable artificial intelligence (XAI) 1.
This research reflects a growing trend in AI applications for healthcare, finance, and legal systems. The development of such AI models that directly impact people's lives underscores the importance of transparency and trust in AI decision-making processes 2.
NVIDIA announces significant upgrades to its GeForce NOW cloud gaming service, including RTX 5080-class performance, improved streaming quality, and an expanded game library, set to launch in September 2025.
9 Sources
Technology
8 hrs ago
9 Sources
Technology
8 hrs ago
Google's Made by Google 2025 event showcases the Pixel 10 series, featuring advanced AI capabilities, improved hardware, and ecosystem integrations. The launch includes new smartphones, wearables, and AI-driven features, positioning Google as a strong competitor in the premium device market.
4 Sources
Technology
8 hrs ago
4 Sources
Technology
8 hrs ago
Palo Alto Networks reports impressive Q4 results and forecasts robust growth for fiscal 2026, driven by AI-powered cybersecurity solutions and the strategic acquisition of CyberArk.
6 Sources
Technology
8 hrs ago
6 Sources
Technology
8 hrs ago
OpenAI updates GPT-5 to make it more approachable following user feedback, sparking debate about AI personality and user preferences.
6 Sources
Technology
16 hrs ago
6 Sources
Technology
16 hrs ago
President Trump's plan to deregulate AI development in the US faces a significant challenge from the European Union's comprehensive AI regulations, which could influence global standards and affect American tech companies' operations worldwide.
2 Sources
Policy
29 mins ago
2 Sources
Policy
29 mins ago