2 Sources
2 Sources
[1]
'Probably' doesn't mean the same thing to your AI as it does to you
When a human says an event is "probable" or "likely," people generally have a shared, if fuzzy, understanding of what that means. But when an AI chatbot like ChatGPT uses the same word, it's not assessing the odds the way we do, my colleagues and I found. We recently published a study in the journal NPJ Complexity that suggests that, while large language model AIs excel at conversation, they often fail to align with humans when communicating uncertainty. The research focused on words of estimative probability, which include terms like "maybe," "probably" and "almost certain." By comparing how AI models and humans map these words to numerical percentages, we uncovered significant gaps between humans and large language models. While the models do tend to agree with humans on extremes like "impossible," they diverge sharply on hedge words like "maybe." For example, a model might use the word "likely" to represent an 80% probability, while a human reader assumes it means closer to 65%. This could be because humans can interpret words such as "likely" and "probable" based more on contextual cues and personal experiences. In contrast, large language models may be averaging over conflicting usages of those words in their training data, leading to divergences with human interpretations. Our study also found that large language models are sensitive to gendered language and the specific language used for prompting. When a prompt changed from "he" to "she," the AI's probability estimates often became more rigid, reflecting biases embedded in its training data. When a prompt changed from English to Chinese, the AI's probability estimates often shifted, possibly due to differences between English and Chinese in how people express and understand uncertainty. Why it matters Far from being a linguistic quirk, this misalignment is a fundamental challenge for AI safety and human-AI interaction. As large language models are increasingly used in high-stakes fields like health care, government policy and scientific reporting, the way they communicate risk becomes a matter of public trust. If an AI assistant helping a doctor, for instance, describes a side effect as "unlikely," but the model's internal calculation of "unlikely" is much higher than the doctor's interpretation, the resulting decision could be flawed. What other research is being done Scientists have studied how humans quantify uncertainty since the 1960s, a field pioneered by CIA analysts to improve intelligence reporting. More recently, there has been an explosion in large language model literature seeking to look under the hood of neural networks to better understand their "behaviors" and linguistic patterns. Our study adds a layer of complexity by treating the interaction between humans and artificial intelligence as a biological-like system where meaning can degrade. It moves beyond simply measuring if an AI is "smart" and instead asks if it is aligned. Other researchers are currently exploring whether so-called chain-of-thought prompting - asking the AI to show its work - can fix these errors. However, our study found that even advanced reasoning doesn't always bridge the gap between statistical data and verbal labels. What's next A goal for future AI development is to create models that don't just predict the next likely word but actually understand the weight of the uncertainty they are conveying. Researchers are calling for more robust consistency metrics to ensure that if a model sees a 10% chance in the data, it chooses the same word every time. As we move toward a world where AI summarizes scientific papers and manages people's schedules, making sure that "probably" means "probably" is a vital step in making these systems reliable partners rather than just sophisticated parrots. The Research Brief is a short take on interesting academic work.
[2]
We studied chatbots and language and saw a huge problem: They mean 80% when they say 'likely' but humans hear 65% | Fortune
When a human says an event is "probable" or "likely," people generally have a shared, if fuzzy, understanding of what that means. But when an AI chatbot like ChatGPT uses the same word, it's not assessing the odds the way we do, my colleagues and I found. We recently published a study in the journal NPJ Complexity that suggests that, while large language model AIs excel at conversation, they often fail to align with humans when communicating uncertainty. The research focused on words of estimative probability, which include terms like "maybe," "probably" and "almost certain." By comparing how AI models and humans map these words to numerical percentages, we uncovered significant gaps between humans and large language models. While the models do tend to agree with humans on extremes like "impossible," they diverge sharply on hedge words like "maybe." For example, a model might use the word "likely" to represent an 80% probability, while a human reader assumes it means closer to 65%. This could be because humans can interpret words such as "likely" and "probable" based more on contextual cues and personal experiences. In contrast, large language models may be averaging over conflicting usages of those words in their training data, leading to divergences with human interpretations. Our study also found that large language models are sensitive to gendered language and the specific language used for prompting. When a prompt changed from "he" to "she," the AI's probability estimates often became more rigid, reflecting biases embedded in its training data. When a prompt changed from English to Chinese, the AI's probability estimates often shifted, possibly due to differences between English and Chinese in how people express and understand uncertainty.
Share
Share
Copy Link
A new study published in NPJ Complexity reveals that AI chatbots and humans interpret probability words differently, with language models assigning 'likely' to 80% while humans assume 65%. This probability misalignment poses risks in high-stakes fields like healthcare and government policy, where miscommunication about uncertainty could lead to flawed decisions.
When an AI chatbot like ChatGPT describes something as "likely" or "probable," it's not communicating the same odds that humans understand. A recent NPJ Complexity study reveals a critical gap in AI and human interpretation of uncertainty, showing that language models assign "likely" to an 80% probability while humans typically interpret it as closer to 65%
1
2
. This probability misalignment extends beyond simple miscommunication, representing a fundamental challenge for human-AI interaction in critical applications.
Source: The Conversation
The research focused on words of estimative probability, including terms like "maybe," "probably," and "almost certain." While AI models and humans tend to agree on extremes like "impossible," they diverge sharply on hedge words. Humans interpret these terms based on contextual cues and personal experiences, drawing from real-world situations to assess likelihood
1
. Language models, however, appear to average over conflicting usages in their training data, leading to interpretations that don't align with human understanding of uncertainty.The study uncovered additional layers of complexity in how AI chatbot systems process probability language. When prompts shifted from "he" to "she," the models' probability estimates became more rigid, exposing biases embedded in training data
2
. Language choice also matters significantly. When researchers changed prompts from English to Chinese, probability estimates shifted, possibly reflecting cultural differences in how people express and understand uncertainty across languages1
. These findings suggest that gendered language and linguistic context introduce systematic variations in how models communicate risk.This misalignment poses serious implications for AI safety as language models expand into healthcare, government policy, and scientific reporting. If an AI assistant helping a doctor describes a side effect as "unlikely," but the model's internal calculation differs significantly from the doctor's interpretation, the resulting decision could be flawed
1
. The disconnect becomes a matter of public trust when AI systems summarize medical research or inform policy decisions where understanding uncertainty accurately is essential.Researchers have studied how humans quantify uncertainty since the 1960s, when CIA analysts pioneered methods to improve intelligence reporting. The current study treats the interaction between humans and AI as a biological-like system where meaning can degrade, moving beyond measuring whether an AI is "smart" to asking if it is aligned
1
. Other researchers are exploring whether chain-of-thought prompting can fix these errors, but the study found that even advanced reasoning doesn't always bridge the gap between statistical data and verbal labels.Related Stories
Looking ahead, developers aim to create models that don't just predict the next likely word but actually understand the weight of uncertainty they convey. Researchers are calling for more robust consistency metrics to ensure that if a model sees a 10% chance in the data, it chooses the same word every time
1
. As AI systems increasingly manage schedules and summarize scientific papers, ensuring that "probably" means "probably" is a vital step in making these systems reliable partners rather than sophisticated parrots. The path forward requires addressing not just technical performance but fundamental alignment in how machines and humans communicate risk and uncertainty.
Source: Fortune
Summarized by
Navi
[1]
16 Jul 2025•Science and Research

28 Aug 2025•Science and Research

26 Sept 2024

1
Technology

2
Technology

3
Policy and Regulation
