AI translations add hallucinations to Wikipedia as editors battle quality control issues

Reviewed byNidhi Govil

2 Sources

Share

Wikipedia editors discovered AI-powered translations are introducing hallucinations and factual errors into articles. The Open Knowledge Association pays contractors $400 monthly to translate content using tools like ChatGPT and Gemini, but editors found incorrect citations and unsourced claims. Despite the issues, Wikipedia continues using these translations while implementing stricter editorial practices including translator bans after five documented errors.

Wikipedia Confronts AI Translations Quality Crisis

Wikipedia is grappling with a growing problem as AI translations introduce hallucinations and factual errors into its multilingual content. The issue centers on the Open Knowledge Association (OKA), a third-party non-profit that pays stipends to translators who use Large Language Models like Google Gemini and ChatGPT to convert Wikipedia articles into other languages without adequate human oversight

1

2

.

Source: 404 Media

Source: 404 Media

Wikipedia editors performing routine reviews discovered basic informational mistakes that weren't present in the original language articles. The problems extended beyond simple translation errors to include incorrect citations, swapped sources, and references to book pages completely unrelated to the subject matter

1

. In one documented case, a draft translation about the French royal La Bourdonnaye family cited a specific book page that "doesn't talk about the La Bourdonnaye family at all," according to Wikipedia editor Ilyas Lebleu

2

.

How AI-Powered Translations Operate at Scale

OKA's approach to integrating AI translation tools relies heavily on cheap labor from contractors in the Global South, paying approximately $400 per month for full-time translation work. A job posting for "Wikipedia Translator" positions offers $397 monthly for up to 40 hours per week, expecting translators to publish 5-20 articles weekly depending on size

2

.

Public spreadsheets used by OKA translators reveal their workflow: "pick an article, copy the lead section into Gemini or chatGPT, then review if some of the suggestions are an improvement to readability." The instructions warn translators to "make edits to the Wiki articles only if the suggestions are an improvement and don't change the meaning"

2

. Previously, these same instructions directed translators to use Grok, Elon Musk's LLM known for producing problematic content, though OKA has since switched to ChatGPT and Claude after internal studies showed better accuracy

2

.

Citation Errors and LLM-Generated Content Failures

Lebleu conducted spot-checks of early translations and found that "some of the articles had swapped sources or added unsourced sentences with no explanation, while 1879 French Senate election added paragraphs sourced from material completely unrelated to what was written"

2

. This pattern reveals a fundamental weakness in LLM-generated content: the models are particularly poor at documentation and maintaining informational accuracy, a problem many lawyers have discovered when relying on generative AI

1

.

The copy-paste nature of OKA translators' work also breaks formatting on some articles, and many results show evidence of translators with "very poor English" who "don't read through their work (or are incapable of seeing problems) and don't add links"

2

.

Stricter Editorial Practices and Wikipedia's Governance Response

Despite Wikipedia's broad policy against LLM-generated articles, Wikipedia editors decided to continue using OKA's services rather than implement a complete ban. The decision reflects a practical reality: English Wikipedia remains twice as large as the next-largest language edition, and languages with fewer speakers desperately need translations for important articles

1

.

Source: PCWorld

Source: PCWorld

However, quality control measures have become significantly harsher. Wikipedia's governance now enforces strict rules: "OKA translators who have received, within six months, four (correctly applied) warnings about content that fails verification will be blocked without further warning if another example is found"

2

. After five documented errors, translator bans take effect, and their previous translations can be wiped unless a more senior editor takes ownership

1

.

This represents a much harsher standard than regular Wikipedia editors face, highlighting the platform's struggle to balance the need for multilingual content expansion with maintaining reliability. The commercial element introduced by paying translators may incentivize speed over accuracy, particularly given the low compensation relative to the expected output

1

. As AI tools continue proliferating across the internet, Wikipedia's open governance model provides a critical test case for how platforms can maintain knowledge integrity while leveraging automation.🟡 untrained_model_response=🟡 The summary has been updated with the selected images based on their relevance and placement rules.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo