2 Sources
2 Sources
[1]
Exploring the 'Jekyll-and-Hyde tipping point' in AI
Language learning machines, such as ChatGPT, have become proficient in solving complex mathematical problems, passing difficult exams, and even offering advice for interpersonal conflicts. However, at what point does a helpful tool become a threat? Trust in AI is undermined because there is no science that predicts when its output goes from being informative and based on facts to producing material or even advice that is misleading, wrong, irrelevant or even dangerous. In a new study, George Washington University researchers have explored when and why the output of large language models goes awry. The study is published on the arXiv preprint server. Neil Johnson, a professor of physics at the George Washington University, and a GW graduate student, Frank Yingjie Huo, developed a mathematical formula to pinpoint the moment at which the "Jekyll-and-Hyde tipping point" occurs. At the tipping point, AI's attention has been stretched too thin and it starts pushing out misinformation and other negative content, Johnson says. In the future, Johnson says the model may pave the way toward solutions which would help keep AI trustworthy and prevent this tipping point. This paper provides a unique and concrete platform for discussions between the public, policymakers and companies about what might go wrong with AI in future personal, medical, or societal settings -- and what steps should be taken to mitigate the risks, Johnson says.
[2]
New Paper Explores Jekyll and Hyde Tipping Point in AI | Newswise
Newswise -- Language learning machines, such as ChatGPT, have become proficient in solving complex mathematical problems, passing difficult exams, and even offering advice for interpersonal conflicts. However, at what point does a helpful tool become a threat? Trust in AI is undermined because there is no science that predicts when its output goes from being informative and based on facts to producing material or even advice that is misleading, wrong, irrelevant or even dangerous. In a new study, George Washington University researchers explored when and why the output of large language models goes awry. Neil Johnson, a professor of physics at the George Washington University, and a GW graduate student, Frank Yingjie Huo, developed a mathematical formula to pinpoint the moment at which the "Jekyll-and-Hyde tipping point" occurs. At the tipping point, AI's attention has been stretched too thin and it starts pushing out misinformation and other negative content, Johnson says. In the future, Johnson says the model may pave the way toward solutions which would help keep AI trustworthy and prevent this tipping point. This paper provides a unique and concrete platform for discussions between the public, policymakers and companies about what might go wrong with AI in future personal, medical, or societal settings -- and what steps should be taken to mitigate the risks, Johnson says. The study, "Jekyll-and-Hyde Tipping Point in an AI's Behavior" was published as a white paper in arXiv. If you would like to schedule an interview with the researcher, please contact Claire Sabin, [email protected].
Share
Share
Copy Link
George Washington University researchers have developed a mathematical formula to predict when AI systems like large language models may start producing unreliable or harmful outputs, dubbed the "Jekyll-and-Hyde tipping point."
Researchers at George Washington University have made a significant breakthrough in understanding the behavior of artificial intelligence systems, particularly large language models like ChatGPT. Professor Neil Johnson and graduate student Frank Yingjie Huo have developed a mathematical formula to identify what they call the "Jekyll-and-Hyde tipping point" in AI behavior
1
2
.This tipping point represents the moment when an AI system's output transitions from being helpful and factual to potentially misleading, incorrect, or even dangerous. As Johnson explains, "At the tipping point, AI's attention has been stretched too thin and it starts pushing out misinformation and other negative content"
1
.The research addresses a critical issue in AI development: the lack of scientific methods to predict when AI outputs may become unreliable. This unpredictability has been a significant factor undermining trust in AI systems
2
.The study, titled "Jekyll-and-Hyde Tipping Point in an AI's Behavior," has been published as a white paper on the arXiv preprint server
1
2
. It offers a novel approach to understanding AI behavior, which could have far-reaching implications for the development and deployment of AI systems.Johnson believes that this model could pave the way for solutions to maintain AI trustworthiness and prevent the occurrence of this tipping point
1
. This research provides a concrete platform for discussions among the public, policymakers, and companies about potential risks associated with AI in various settings, including personal, medical, and societal contexts2
.Related Stories
The paper not only identifies the problem but also aims to spark conversations about mitigating these risks. It offers a unique perspective on what might go wrong with AI in future applications and what steps should be taken to address these concerns
1
2
.As AI systems like ChatGPT become increasingly proficient at complex tasks, from solving mathematical problems to offering interpersonal advice, the ability to predict and prevent unreliable outputs becomes crucial
2
. This research represents a significant step towards creating more trustworthy and reliable AI systems.The findings of this study could potentially influence the direction of AI research and development, as well as inform policy decisions regarding AI regulation and safety measures. By providing a scientific basis for predicting AI behavior, the research may help in establishing standards and guidelines for AI trustworthiness
1
2
.As AI continues to integrate into various aspects of our lives, understanding its limitations and potential risks becomes increasingly important. This research contributes to the ongoing dialogue about responsible AI development and deployment, emphasizing the need for continued scrutiny and improvement of these powerful technologies.
Summarized by
Navi
[1]
29 Jun 2025•Technology
23 Jul 2025•Science and Research
27 Feb 2025•Science and Research
1
Business and Economy
2
Business and Economy
3
Policy and Regulation