Ontario Audit Reveals AI Medical Scribes Hallucinate Patient Info, Miss Critical Health Details

3 Sources

Share

An Ontario audit found all 20 government-approved AI medical scribes generated incorrect or fabricated information during testing. Nine systems hallucinated nonexistent patient details, 12 recorded wrong medication names, and 17 missed mental health issues—raising serious concerns about patient safety despite 5,000 doctors already using these tools.

AI Medical Scribes Show Widespread Accuracy Problems in Government Testing

The Office of the Auditor General of Ontario has uncovered troubling flaws in AI medical scribes used by thousands of physicians across the province. According to a recent report on artificial intelligence use in Ontario's government, all 20 AI note-taking systems approved by provincial authorities demonstrated significant problems with accuracy during standardized testing

1

. The findings raise urgent questions about patient safety as these systems handle sensitive medical information and help shape treatment decisions.

During procurement evaluations, medical professionals reviewed simulated doctor-patient recordings alongside AI-generated notes to assess their reliability. What they discovered was alarming: nine out of 20 systems fabricated information and made suggestions to patients' treatment plans that were never discussed in the recordings

2

. These AI hallucinations included nonexistent referrals for blood tests or therapy, and false claims that no masses were found during examinations

1

.

Source: Futurism

Source: Futurism

Critical Medication Errors and Missing Mental Health Information

The Ontario audit revealed that 12 of the 20 evaluated systems inserted incorrect drug information into patient notes, potentially leading to dangerous prescription errors

2

. Perhaps most concerning, 17 systems missed key details about patients' mental health issues that were clearly discussed in the recordings, with six missing these issues fully or partially

2

. The auditor general warned that such inaccuracies in medical notes could potentially result in inadequate or harmful patient treatment plans that may impact health outcomes

1

.

Source: The Register

Source: The Register

Despite these serious issues with medical note accuracy, the AI scribe program continues to operate. More than 5,000 physicians in Ontario are currently participating in the initiative launched by the Ministry of Health for doctors, nurse practitioners, and other healthcare professionals

2

. A Ministry spokesperson told CBC that there have been no known reports of patient harms associated with the technology, though auditor general Shelley Spence has advised patients to ask their doctors to review AI-generated transcripts carefully

3

.

Flawed Vendor Evaluation Process Prioritized Location Over Accuracy

The report identified fundamental problems with how the provincial government assessed these AI systems. Medical note accuracy contributed only 4 percent to vendors' total evaluation scores, while having a domestic presence in Ontario accounted for 30 percent

1

. Bias controls represented just 2 percent of the score, and security/privacy assessments added another 2 percent

2

. This skewed weighting meant vendors could achieve approval even with a zero score on accuracy metrics.

Across all approved vendors, the average AI scribe scored only 12 out of 20 on the accuracy portion of Supply Ontario's evaluation rubric

1

. The auditor general concluded these systems were not evaluated adequately and recommended that IT departments force doctors to confirm their review of AI-produced notes before committing them to patient logs .

Growing Concerns About AI in Healthcare Extend Beyond Ontario

While Minister Stephen Crawford clarified that the documented hallucinations occurred during testing rather than operational use with doctors, the fact remains that these inaccurate and fabricated information-prone systems are now handling real patient data

3

. OntarioMD, which supports physicians in adopting new technologies and was involved in the procurement process, has recommended manual review of AI notes, but no mandatory attestation feature exists in any approved system

2

.

The Ontario findings align with broader concerns about AI in healthcare. In the United States, OpenEvidence faces scrutiny for drawing overly strong conclusions from medical studies with small sample sizes, according to physicians interviewed by NBC News

3

. Many overworked doctors have embraced these tools to reduce administrative tasks, but the technology's reliability remains questionable. Healthcare providers are not required to use government-approved systems and may purchase from non-approved vendors, creating additional oversight challenges

1

. As adoption grows, medical professionals and patients alike must remain vigilant about verifying AI-generated documentation to protect patient safety.

Source: Ars Technica

Source: Ars Technica

Today's Top Stories

TheOutpost.ai

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Instagram logo
LinkedIn logo
Youtube logo
© 2026 TheOutpost.AI All rights reserved