2 Sources
2 Sources
[1]
A Wikipedia Group Made a Guide to Detect AI Writing. Now a Plug-In Uses It to 'Humanize' Chatbots
The web's best resource for spotting AI writing has ironically become a manual for AI models to hide it. On Saturday, tech entrepreneur Siqi Chen released an open source plug-in for Anthropic's Claude Code AI assistant that instructs the AI model to stop writing like an AI model. Called Humanizer, the simple prompt plug-in feeds Claude a list of 24 language and formatting patterns that Wikipedia editors have listed as chatbot giveaways. Chen published the plug-in on GitHub, where it has picked up more than 1,600 stars as of Monday. "It's really handy that Wikipedia went and collated a detailed list of 'signs of AI writing,'" Chen wrote on X. "So much so that you can just tell your LLM to ... not do that." The source material is a guide from WikiProject AI Cleanup, a group of Wikipedia editors who have been hunting AI-generated articles since late 2023. French Wikipedia editor Ilyas Lebleu founded the project. The volunteers have tagged over 500 articles for review and, in August 2025, published a formal list of the patterns they kept seeing. Chen's tool is a "skill file" for Claude Code, Anthropic's terminal-based coding assistant, which involves a Markdown-formatted file that adds a list of written instructions (you can see them here) appended to the prompt fed into the large language model that powers the assistant. Unlike a normal system prompt, for example, the skill information is formatted in a standardized way that Claude models are fine-tuned to interpret with more precision than a plain system prompt. (Custom skills require a paid Claude subscription with code execution turned on.) But as with all AI prompts, language models don't always perfectly follow skill files, so does the Humanizer actually work? In our limited testing, Chen's skill file made the AI agent's output sound less precise and more casual, but it could have some drawbacks: It won't improve factuality and might harm coding ability. In particular, some of Humanizer's instructions might lead you astray, depending on the task. For example, the Humanizer skill includes this line: "Have opinions. Don't just report facts -- react to them. 'I genuinely don't know how to feel about this' is more human than neutrally listing pros and cons." While being imperfect seems human, this kind of advice would probably not do you any favors if you were using Claude to write technical documentation. Even with its drawbacks, it's ironic that one of the web's most referenced rule sets for detecting AI-assisted writing may help some people subvert it. Spotting the Patterns So what does AI writing look like? The Wikipedia guide is specific with many examples, but we'll give you just one here for brevity's sake. Some chatbots love to pump up their subjects with phrases like "marking a pivotal moment" or "stands as a testament to," according to the guide. They write like tourism brochures, calling views "breathtaking" and describing towns as "nestled within" scenic regions. They tack "-ing" phrases onto the end of sentences to sound analytical: "symbolizing the region's commitment to innovation." To work around those rules, the Humanizer skill tells Claude to replace inflated language with plain facts and offers this example transformation: Before: "The Statistical Institute of Catalonia was officially established in 1989, marking a pivotal moment in the evolution of regional statistics in Spain." After: "The Statistical Institute of Catalonia was established in 1989 to collect and publish regional statistics." Claude will read that and do its best as a pattern-matching machine to create an output that matches the context of the conversation or task at hand. Why AI Writing Detection Fails Even with such a confident set of rules crafted by Wikipedia editors, we've previously written about why AI writing detectors don't work reliably: There is nothing inherently unique about human writing that reliably differentiates it from LLM writing. One reason is that even though most AI language models tend toward certain types of language, they can also be prompted to avoid them, as with the Humanizer skill. (Although sometimes it's very difficult, as OpenAI found in its yearslong struggle against the em dash.)
[2]
This plugin uses Wikipedia's AI-spotting guide to make AI writing sound more human
A new tool aims to help AI chatbots generate more human-sounding text -- with the help of Wikipedia's guide for detecting AI, as reported by Ars Technica. Developer Siqi Chen says he created the tool, called Humanizer, by feeding Anthropic's Claude the list of tells that Wikipedia's volunteer editors put together as part of an initiative to combat "poorly written AI-generated content." Wikipedia's guide contains a list of signs that text may be AI-generated, including vague attributions, promotional language like describing something as "breathtaking," and collaborative phrases, such as "I hope this helps!" Humanizer, which is a custom skill for Claude Code, is supposed to help the AI assistant avoid detection by removing these "signs of AI-generated writing from text, making it sound more natural and human," according to its GitHub page. The GitHub page provides some examples on how Humanizer might help Claude detect some of these tells, including by changing a sentence that described a location as "nestled within the breathtaking region" to "a town in the Gonder region," as well as adjusting a vague attribution, like "Experts believe it plays a crucial role" to "according to a 2019 survey by..." Chen says the tool will "automatically push updates" when Wikipedia's AI-detecting guide is updated. It's only a matter of time before the AI companies themselves start adjusting their chatbots against some of these tells, too, as OpenAI has already addressed ChatGPT's overuse of em dashes, which has become an indicator of AI content.
Share
Share
Copy Link
A tool called Humanizer flips Wikipedia's AI detection guide on its head, using the same 24 patterns volunteer editors identified to spot AI-generated content to instead teach Anthropic's Claude how to avoid them. The ironic twist highlights the cat-and-mouse game between AI detection and evasion.
Tech entrepreneur Siqi Chen released an open source plug-in on Saturday that transforms Wikipedia's carefully crafted AI detection guide into instructions for Anthropic's Claude Code AI assistant to avoid sounding robotic
1
. The Humanizer plug-in feeds Claude a list of 24 language and formatting patterns that Wikipedia editors compiled as chatbot giveaways, effectively teaching the language models to sidestep detection2
. Published on GitHub, the tool garnered over 1,600 stars by Monday, revealing strong demand for ways to make AI writing sound more human1
.
Source: Wired
The source material comes from WikiProject AI Cleanup, a volunteer initiative founded by French Wikipedia editor Ilyas Lebleu in late 2023 to combat AI-generated content
1
. These volunteers have tagged over 500 articles for review and published a formal list of patterns in AI writing in August 2025, documenting telltale signs like promotional language describing views as "breathtaking" or towns "nestled within" scenic regions1
. The guide also flags phrases like "marking a pivotal moment" and "stands as a testament to" as common chatbot giveaways1
.The Humanizer operates as a "skill file" for Claude Code, Anthropic's terminal-based coding assistant
1
. Unlike standard AI prompts, this Markdown-formatted file contains structured instructions that Claude models are fine-tuned to interpret with greater precision. The skill file requires a paid Claude subscription with code execution enabled1
. Chen designed the tool to automatically push updates whenever Wikipedia's AI-spotting guide receives modifications, creating a dynamic system that evolves alongside detection methods2
.
Source: The Verge
The transformation targets specific linguistic patterns to detect AI writing. For instance, the Humanizer instructs Claude to replace inflated language with plain facts, changing "The Statistical Institute of Catalonia was officially established in 1989, marking a pivotal moment in the evolution of regional statistics in Spain" to simply "The Statistical Institute of Catalonia was established in 1989 to collect and publish regional statistics"
1
. Similarly, vague attributions like "Experts believe it plays a crucial role" become more specific: "according to a 2019 survey by..."2
.Limited testing shows the Humanizer made Claude's output sound less precise and more casual, but the approach carries notable drawbacks
1
. The tool won't improve factuality and might harm coding ability. Some instructions could mislead users depending on the task. One directive tells the LLM to "have opinions" and "react" to facts rather than neutrally listing information, advice that would likely damage technical documentation quality. Since language models don't always perfectly follow skill files, the Humanizer's effectiveness varies across different use cases.Related Stories
This development underscores why AI detection remains unreliable: there's nothing inherently unique about human writing that consistently differentiates it from AI-generated text
1
. Even though most language models gravitate toward certain linguistic patterns, they can be prompted to avoid them, as the Humanizer demonstrates. OpenAI already addressed ChatGPT's overuse of the em dash after it became an indicator of AI-generated content, suggesting major AI companies will likely adjust their models against these tells2
. The irony is stark: one of the web's most referenced resources for spotting AI-assisted writing now helps people subvert it, creating an escalating cycle where detection methods inadvertently fuel more sophisticated evasion techniques.Summarized by
Navi
[1]
09 Aug 2025•Technology

11 Oct 2024•Technology

27 Nov 2024•Technology

1
Policy and Regulation

2
Technology

3
Technology
