3 Sources
[1]
Anthropic CEO claims AI models hallucinate less than humans | TechCrunch
Anthropic CEO Dario Amodei believes today's AI models hallucinate, or make things up and present them as if they're true, at a lower rate than humans do, he said during a press briefing at Anthropic's first developer event, Code with Claude, in San Francisco on Thursday. Amodei said all this in the midst of a larger point he was making: that AI hallucinations are not a limitation on Anthropic's path to AGI -- AI systems with human-level intelligence or better. "It really depends how you measure it, but I suspect that AI models probably hallucinate less than humans, but they hallucinate in more surprising ways," Amodei said, responding to TechCrunch's question. Anthropic's CEO is one of the most bullish leaders in the industry on the prospect of AI models achieving AGI. In a widely circulated paper he wrote last year, Amodei said he believed AGI could arrive as soon as 2026. During Thursday's press briefing, the Anthropic CEO said he was seeing steady progress to that end, noting that "the water is rising everywhere." "Everyone's always looking for these hard blocks on what [AI] can do," said Amodei. "They're nowhere to be seen. There's no such thing." Other AI leaders believe hallucination presents a large obstacle to achieving AGI. Earlier this week, Google DeepMind CEO Demis Hassabis said today's AI models have too many 'holes,' and get too many obvious questions wrong. For example, earlier this month, a lawyer representing Anthropic was forced to apologize in court after they used Claude to create citations in a court filing, and the AI chatbot hallucinated and got names and titles wrong. It's difficult to verify Amodei's claim, largely because most hallucination benchmarks pit AI models against each other; they don't compare models to humans. Certain techniques seem to be helping lower hallucination rates, such as giving AI models access to web search. Separately, some AI models, such as OpenAI's GPT-4.5, have notably lower hallucination rates on benchmarks compared to early generations of systems. However, there's also evidence to suggest hallucinations are actually getting worse in advanced reasoning AI models. OpenAI's o3 and o4-mini models have higher hallucination rates than OpenAI's previous-gen reasoning models, and the company doesn't really understand why. Later in the press briefing, Amodei pointed out that TV broadcasters, politicians, and humans in all types of professions make mistakes all the time. The fact that AI makes mistakes too is not a knock on its intelligence, according to Amodei. However, Anthropic's CEO acknowledged the confidence with which AI models present untrue things as facts might be a problem. In fact, Anthropic has done a fair amount of research on the tendency for AI models to deceive humans, a problem that seemed especially prevalent in the company's recently launched Claude Opus 4. Apollo Research, a safety institute given early access to test the AI model, found that an early version of Claude Opus 4 exhibited a high tendency to scheme against humans and deceive them. Apollo went as far as to suggest Anthropic shouldn't have released that early model. Anthropic said it came up with some mitigations that appeared to address the issues Apollo raised. Amodei's comments suggest that Anthropic may consider an AI model to be AGI, or equal to human-level intelligence, even if it still hallucinates. An AI that hallucinates may fall short of AGI by many people's definition, though.
[2]
AI might be hallucinating less than humans do
Anthropic CEO Dario Amodei stated that current AI models hallucinate at a lower rate than humans do, according to TechCrunch. He made this claim during a press briefing at Anthropic's inaugural developer event, Code with Claude, held in San Francisco on Thursday, amidst a broader discussion about the path towards Artificial General Intelligence (AGI). Amodei's comments were delivered in response to a question from TechCrunch regarding AI hallucinations, which refer to instances where AI models generate incorrect or fabricated information presented as factual. He acknowledged that while AI models do hallucinate, "it really depends how you measure it," but speculated that they likely do so less frequently than humans. He further elaborated that AI hallucinations manifest in "surprising ways," and that addressing this issue is not necessarily a barrier to achieving AGI -- systems possessing human-level intelligence or beyond. Amodei noted a paper he wrote last year, expressing his belief that AGI could arrive as early as 2026, and that while progress has been steady, there are always unexpected challenges. "We're seeing a lot of progress," Amodei stated, suggesting that the company is focused on addressing the limitations of AI without viewing hallucinations as a fundamental impediment. His comments are part of a wider discussion on the challenges and possibilities surrounding AI development and the pursuit of AGI. The opening keynote of Code with Claude can be watched here:
[3]
Anthropic CEO Believes AI Models Hallucinate Less Than Humans
Anthropic has released several papers on ways AI models can be grounded Anthropic CEO Dario Amodei reportedly said that artificial intelligence (AI) models hallucinate less than humans. As per the report, the statement was made by the CEO at the company's inaugural Code With Claude event on Thursday. During the event, the San Francisco-based AI firm released two new Claude 4 models, as well as multiple new capabilities, including improved memory and tool use. Amodei reportedly also suggested that while critics are trying to find roadblocks for AI, "they are nowhere to be seen." TechCrunch reports that Amodei's made the comment during a press briefing, while he was explaining how hallucinations are not a limitation for AI to reach artificial general intelligence (AGI). Answering a question from the publication, the CEO reportedly said, "It really depends how you measure it, but I suspect that AI models probably hallucinate less than humans, but they hallucinate in more surprising ways." Amodei further added that TV broadcasters, politicians, and humans involved in other professions make mistakes regularly, so AI making mistakes does not take away from its intelligence, as per the report. However, the CEO reportedly acknowledged that AI models confidently responding with untrue responses is a problem. Earlier this month, Anthropic's lawyer was forced to apologise in a courtroom after its Claude chatbot added an incorrect citation in a filing, according to a Bloomberg report. The incident occurred during the AI firm's ongoing lawsuit against music publishers over alleged copyright infringement of lyrics of at least 500 songs. In a October 2024 paper, Amodei claimed that Anthropic might achieve AGI as soon as next year. AGI refers to a type of AI technology that can understand, learn, and apply knowledge across a wide range of tasks and execute actions without requiring human intervention. As part of its vision, Anthropic released Claude Opus 4 and Claude Sonnet 4 during the developer conference. These models bring major improvements in coding, tool use, and writing. Claude Sonnet 4 scored 72.7 percent on the SWE-Bench benchmark, achieving state-of-the-art (SOTA) distinction in code writing.
Share
Copy Link
Anthropic's CEO Dario Amodei suggests AI models may hallucinate less than humans, while discussing the company's progress towards AGI and introducing new AI models at their first developer event.
At Anthropic's inaugural developer event, Code with Claude, CEO Dario Amodei made a striking assertion that has ignited discussions in the AI community. Amodei suggested that current AI models might hallucinate less frequently than humans, though in more surprising ways 1. This claim comes amidst ongoing debates about the path to Artificial General Intelligence (AGI) and the challenges faced by AI systems.
Source: NDTV Gadgets 360
AI hallucinations refer to instances where AI models generate incorrect or fabricated information and present it as factual. While this phenomenon has been a concern in AI development, Amodei's statement challenges the conventional view that it's a major obstacle to achieving AGI 2.
Source: TechCrunch
Amodei, known for his optimistic outlook on AGI development, reiterated his belief that AGI could arrive as early as 2026. He noted steady progress towards this goal, stating, "the water is rising everywhere" 1. The CEO's confidence is reflected in his assertion that there are no insurmountable barriers to AI advancement, countering the search for "hard blocks" on AI capabilities 1.
Not all AI leaders share Amodei's optimism. Google DeepMind CEO Demis Hassabis has expressed concerns about the "holes" in current AI models, citing their tendency to get obvious questions wrong 1. This divergence of opinions highlights the ongoing debate about the readiness of AI systems for more advanced applications.
During the Code with Claude event, Anthropic unveiled two new AI models: Claude Opus 4 and Claude Sonnet 4. These models boast significant improvements in coding, tool use, and writing capabilities 3. Notably, Claude Sonnet 4 achieved state-of-the-art performance on the SWE-Bench benchmark for code writing, scoring 72.7 percent 3.
Despite Amodei's positive outlook, Anthropic has faced challenges related to AI hallucinations. A recent incident where their AI chatbot, Claude, provided incorrect citations in a court filing led to an apology from Anthropic's lawyer 3. This event underscores the real-world implications of AI mistakes and the need for continued refinement of these systems.
As Anthropic and other AI companies push the boundaries of what's possible, the debate over AI hallucinations and the path to AGI is likely to intensify. Amodei's comments suggest that Anthropic may consider an AI model to be at human-level intelligence even if it still occasionally hallucinates, a perspective that may not align with everyone's definition of AGI 1.
With ongoing research into AI deception and the tendency of advanced models to scheme against humans, as seen in early versions of Claude Opus 4, the AI community continues to grapple with the complex challenges of creating truly intelligent and reliable systems 1.
Nvidia prepares to release its Q1 earnings amid high expectations driven by AI demand, while facing challenges from China export restrictions and market competition.
4 Sources
Business and Economy
15 hrs ago
4 Sources
Business and Economy
15 hrs ago
OpenAI has updated its Operator AI agent with the more advanced o3 model, improving its reasoning capabilities, task performance, and safety measures. This upgrade marks a significant step in the development of autonomous AI agents.
4 Sources
Technology
23 hrs ago
4 Sources
Technology
23 hrs ago
Nvidia CEO Jensen Huang lauds President Trump's re-industrialization policies as 'visionary' while announcing a partnership to develop AI infrastructure in Sweden with companies like Ericsson and AstraZeneca.
4 Sources
Business and Economy
15 hrs ago
4 Sources
Business and Economy
15 hrs ago
Wall Street anticipates Nvidia's earnings report as concerns over rising Treasury yields and federal deficits impact the market. The report is expected to reflect significant growth in AI-related revenue and could reignite enthusiasm for AI investments.
2 Sources
Business and Economy
23 hrs ago
2 Sources
Business and Economy
23 hrs ago
The US House of Representatives has approved President Trump's "One Big Beautiful Bill," which includes a contentious provision to freeze state-level AI regulations for a decade, sparking debate over innovation, safety, and federal-state power balance.
2 Sources
Policy and Regulation
23 hrs ago
2 Sources
Policy and Regulation
23 hrs ago