2 Sources
[1]
AI overconfidence mirrors human brain condition
Agents, chatbots and other tools based on artificial intelligence (AI) are increasingly used in everyday life by many. So-called large language model (LLM)-based agents, such as ChatGPT and Llama, have become impressively fluent in the responses they form, but quite often provide convincing yet incorrect information. Researchers at the University of Tokyo draw parallels between this issue and a human language disorder known as aphasia, where sufferers may speak fluently but make meaningless or hard-to-understand statements. This similarity could point toward better forms of diagnosis for aphasia, and even provide insight to AI engineers seeking to improve LLM-based agents. This article was written by a human being, but the use of text-generating AI is on the rise in many areas. As more and more people come to use and rely on such things, there's an ever-increasing need to make sure that these tools deliver correct and coherent responses and information to their users. Many familiar tools, including ChatGPT and others, appear very fluent in whatever they deliver. But their responses cannot always be relied upon due to the amount of essentially made-up content they produce. If the user is not sufficiently knowledgeable about the subject area in question, they can easily fall foul of assuming this information is right, especially given the high degree of confidence ChatGPT and others show. "You can't fail to notice how some AI systems can appear articulate while still producing often significant errors," said Professor Takamitsu Watanabe from the International Research Center for Neurointelligence (WPI-IRCN) at the University of Tokyo. "But what struck my team and I was a similarity between this behavior and that of people with Wernicke's aphasia, where such people speak fluently but don't always make much sense. That prompted us to wonder if the internal mechanisms of these AI systems could be similar to those of the human brain affected by aphasia, and if so, what the implications might be." To explore this idea, the team used a method called energy landscape analysis, a technique originally developed by physicists seeking to visualize energy states in magnetic metal, but which was recently adapted for neuroscience. They examined patterns in resting brain activity from people with different types of aphasia and compared them to internal data from several publicly available LLMs. And in their analysis, the team did discover some striking similarities. The way digital information or signals are moved around and manipulated within these AI models closely matched the way some brain signals behaved in the brains of people with certain types of aphasia, including Wernicke's aphasia. "You can imagine the energy landscape as a surface with a ball on it. When there's a curve, the ball may roll down and come to rest, but when the curves are shallow, the ball may roll around chaotically," said Watanabe. "In aphasia, the ball represents the person's brain state. In LLMs, it represents the continuing signal pattern in the model based on its instructions and internal dataset." The research has several implications. For neuroscience, it offers a possible new way to classify and monitor conditions like aphasia based on internal brain activity rather than just external symptoms. For AI, it could lead to better diagnostic tools that help engineers improve the architecture of AI systems from the inside out. Though, despite the similarities the researchers discovered, they urge caution not to make too many assumptions. "We're not saying chatbots have brain damage," said Watanabe. "But they may be locked into a kind of rigid internal pattern that limits how flexibly they can draw on stored knowledge, just like in receptive aphasia. Whether future models can overcome this limitation remains to be seen, but understanding these internal parallels may be the first step toward smarter, more trustworthy AI too."
[2]
AI Chatbots Mirror a Human Brain Disorder - Neuroscience News
Summary: Researchers have found a surprising similarity between the way large language models (LLMs) like ChatGPT process information and how the brains of people with Wernicke's aphasia function. In both cases, fluent but often incoherent output is produced, suggesting rigid internal processing patterns that can distort meaning. By applying energy landscape analysis to brain scans and AI model data, scientists observed shared dynamics in signal flow, hinting at deeper structural parallels. These findings could help improve both aphasia diagnosis and AI design, revealing how internal limitations affect language clarity. Agents, chatbots and other tools based on artificial intelligence (AI) are increasingly used in everyday life by many. So-called large language model (LLM)-based agents, such as ChatGPT and Llama, have become impressively fluent in the responses they form, but quite often provide convincing yet incorrect information. Researchers at the University of Tokyo draw parallels between this issue and a human language disorder known as aphasia, where sufferers may speak fluently but make meaningless or hard-to-understand statements. This similarity could point toward better forms of diagnosis for aphasia, and even provide insight to AI engineers seeking to improve LLM-based agents. This article was written by a human being, but the use of text-generating AI is on the rise in many areas. As more and more people come to use and rely on such things, there's an ever-increasing need to make sure that these tools deliver correct and coherent responses and information to their users. Many familiar tools, including ChatGPT and others, appear very fluent in whatever they deliver. But their responses cannot always be relied upon due to the amount of essentially made-up content they produce. If the user is not sufficiently knowledgeable about the subject area in question, they can easily fall foul of assuming this information is right, especially given the high degree of confidence ChatGPT and others show. "You can't fail to notice how some AI systems can appear articulate while still producing often significant errors," said Professor Takamitsu Watanabe from the International Research Center for Neurointelligence (WPI-IRCN) at the University of Tokyo. "But what struck my team and I was a similarity between this behavior and that of people with Wernicke's aphasia, where such people speak fluently but don't always make much sense. "That prompted us to wonder if the internal mechanisms of these AI systems could be similar to those of the human brain affected by aphasia, and if so, what the implications might be." To explore this idea, the team used a method called energy landscape analysis, a technique originally developed by physicists seeking to visualize energy states in magnetic metal, but which was recently adapted for neuroscience. They examined patterns in resting brain activity from people with different types of aphasia and compared them to internal data from several publicly available LLMs. And in their analysis, the team did discover some striking similarities. The way digital information or signals are moved around and manipulated within these AI models closely matched the way some brain signals behaved in the brains of people with certain types of aphasia, including Wernicke's aphasia. "You can imagine the energy landscape as a surface with a ball on it. When there's a curve, the ball may roll down and come to rest, but when the curves are shallow, the ball may roll around chaotically," said Watanabe. "In aphasia, the ball represents the person's brain state. In LLMs, it represents the continuing signal pattern in the model based on its instructions and internal dataset." The research has several implications. For neuroscience, it offers a possible new way to classify and monitor conditions like aphasia based on internal brain activity rather than just external symptoms. For AI, it could lead to better diagnostic tools that help engineers improve the architecture of AI systems from the inside out. Though, despite the similarities the researchers discovered, they urge caution not to make too many assumptions. "We're not saying chatbots have brain damage," said Watanabe. "But they may be locked into a kind of rigid internal pattern that limits how flexibly they can draw on stored knowledge, just like in receptive aphasia. "Whether future models can overcome this limitation remains to be seen, but understanding these internal parallels may be the first step toward smarter, more trustworthy AI too." Funding: This work was supported by Grant-in-aid for Research Activity from Japan Society for Promotion of Sciences (19H03535, 21H05679, 23H04217, JP20H05921), The University of Tokyo Excellent Young Researcher Project, Showa University Medical Institute of Developmental Disabilities Research, JST Moonshot R&D Program (JPMJMS2021), JST FOREST Program (24012854), Institute of AI and Beyond of UTokyo, Cross-ministerial Strategic Innovation Promotion Program (SIP) on "Integrated Health Care System" (JPJ012425). Comparison of large language model with aphasia Large language models (LLMs) respond fluently but often inaccurately, which resembles aphasia in humans. Does this behavioral similarity indicate any resemblance in internal information processing between LLMs and aphasic humans? Here, we address this question by comparing the network dynamics between LLMs -- ALBERT, GPT-2, Llama-3.1 and one Japanese variant of Llama -- and various aphasic brains. Using energy landscape analysis, we quantify how frequently the network activity pattern is likely to move from one state to another (transition frequency) and how long it tends to dwell in each state (dwelling time). First, by investigating the frequency spectrums of these two indices for brain dynamics, we find that the degrees of the polarization of the transition frequency and dwelling time enable accurate classification of receptive aphasia, expressive aphasia and controls: receptive aphasia shows the bimodal distributions for both indices, whereas expressive aphasia exhibits the most uniform distributions. In parallel, we identify highly polarized distributions in both transition frequency and dwelling time in the network dynamics in the four LLMs. These findings indicate the similarity in internal information processing between LLMs and receptive aphasia, and the current approach can provide a novel diagnosis and classification tool for LLMs and help their performance improve.
Share
Copy Link
Researchers at the University of Tokyo have discovered striking similarities between the behavior of large language models (LLMs) and individuals with Wernicke's aphasia, potentially offering insights for both AI development and neurological research.
Researchers at the University of Tokyo have uncovered intriguing parallels between the behavior of large language models (LLMs) and a human language disorder known as Wernicke's aphasia. This discovery could have significant implications for both artificial intelligence development and neurological research 1.
LLM-based agents like ChatGPT and Llama have become increasingly prevalent in everyday life, impressing users with their fluent responses. However, these AI tools often provide convincing yet incorrect information. This phenomenon bears a striking resemblance to Wernicke's aphasia, a condition where individuals speak fluently but produce meaningless or incomprehensible statements 2.
Professor Takamitsu Watanabe from the International Research Center for Neurointelligence (WPI-IRCN) at the University of Tokyo noted, "You can't fail to notice how some AI systems can appear articulate while still producing often significant errors" 1.
To explore this similarity, the research team employed a technique called energy landscape analysis. Originally developed for visualizing energy states in magnetic metal, this method has been adapted for neuroscience applications 1.
The team examined patterns in resting brain activity from individuals with different types of aphasia and compared them to internal data from several publicly available LLMs, including ALBERT, GPT-2, Llama-3, and a Japanese variant of Llama 2.
The analysis revealed remarkable similarities between the internal processes of LLMs and the brain activity of individuals with certain types of aphasia, particularly Wernicke's aphasia. The way digital information or signals are manipulated within these AI models closely matched the behavior of brain signals in aphasia patients 1.
Professor Watanabe explained the concept using an analogy: "You can imagine the energy landscape as a surface with a ball on it. When there's a curve, the ball may roll down and come to rest, but when the curves are shallow, the ball may roll around chaotically" 2.
This research has significant implications for both neuroscience and artificial intelligence:
Neuroscience: The findings offer a potential new approach to classifying and monitoring aphasia based on internal brain activity rather than just external symptoms 1.
AI Development: The research could lead to improved diagnostic tools for engineers, helping them enhance the architecture of AI systems from the inside out 2.
While the similarities are striking, the researchers urge caution in drawing conclusions. "We're not saying chatbots have brain damage," Watanabe clarified. "But they may be locked into a kind of rigid internal pattern that limits how flexibly they can draw on stored knowledge, just like in receptive aphasia" 1.
As AI continues to integrate into daily life, understanding these internal parallels could be the first step toward developing smarter, more trustworthy AI systems while also advancing our understanding of human brain disorders.
Databricks raises $1 billion in a new funding round, valuing the company at over $100 billion. The data analytics firm plans to invest in AI database technology and an AI agent platform, positioning itself for growth in the evolving AI market.
11 Sources
Business
14 hrs ago
11 Sources
Business
14 hrs ago
SoftBank makes a significant $2 billion investment in Intel, boosting the chipmaker's efforts to regain its competitive edge in the AI semiconductor market.
22 Sources
Business
22 hrs ago
22 Sources
Business
22 hrs ago
OpenAI introduces ChatGPT Go, a new subscription plan priced at ₹399 ($4.60) per month exclusively for Indian users, offering enhanced features and affordability to capture a larger market share.
15 Sources
Technology
22 hrs ago
15 Sources
Technology
22 hrs ago
Microsoft introduces a new AI-powered 'COPILOT' function in Excel, allowing users to perform complex data analysis and content generation using natural language prompts within spreadsheet cells.
8 Sources
Technology
14 hrs ago
8 Sources
Technology
14 hrs ago
Adobe launches Acrobat Studio, integrating AI assistants and PDF Spaces to transform document management and collaboration, marking a significant evolution in PDF technology.
10 Sources
Technology
13 hrs ago
10 Sources
Technology
13 hrs ago