{"id":23370,"date":"2026-05-01T13:11:00","date_gmt":"2026-05-01T13:11:00","guid":{"rendered":"https:\/\/sandbox.hbmadvisory.com\/amplify\/russian-developers-ai-humaniser-struggles-to-bypass-automated-moderation-on-habr\/"},"modified":"2026-05-01T13:26:20","modified_gmt":"2026-05-01T13:26:20","slug":"russian-developers-ai-humaniser-struggles-to-bypass-automated-moderation-on-habr","status":"publish","type":"post","link":"https:\/\/sandbox.hbmadvisory.com\/amplify\/russian-developers-ai-humaniser-struggles-to-bypass-automated-moderation-on-habr\/","title":{"rendered":"Russian developer\u2019s AI humaniser struggles to bypass automated moderation on Habr"},"content":{"rendered":"<p><\/p>\n<div>\n<p>A Russian open-source developer\u2019s attempt to conceal AI-generated content using a new &#8216;humanising&#8217; tool reveals the limitations of current document-level detection systems, highlighting challenges unique to Russian language and stylistic patterns.<\/p>\n<\/div>\n<div>\n<p>A Russian open-source developer says his own tool for &#8220;humanising&#8221; AI-written text failed on a Habr article about AI-style writing, with the platform&#8217;s automated moderation rejecting the post on 27 April as likely machine-generated. The episode has highlighted a larger problem for Russian-language publishing: tools that smooth out individual phrases can still miss the document-level patterns that automated classifiers are trained to spot.<\/p>\n<p>The author said he had submitted a piece about Russian AI writing patterns after first running a draft through his own humaniser, a skill for Claude Code and OpenCode called humanizer-ru. He later rewrote one awkward repetition he had spotted himself, then noted the change in a postscript. Days later, Habr&#8217;s auto-moderator responded that the publication could not pass moderation because most of the text was highly likely to have been created with a generative AI model.<\/p>\n<p>That irony became the starting point for a broader audit of what the tool actually does and does not do. The developer says humanizer-ru was built to catch Russian bureaucratic prose, heavy nominalisations, repetitive genitive chains, overuse of &#8220;\u044f\u0432\u043b\u044f\u0435\u0442\u0441\u044f&#8221;, stale corporate phrasing and literal calques from English. He says only a small minority of rules from an earlier English-language humaniser transferred cleanly to Russian, forcing most of the logic to be rewritten for Russian morphology and style.<\/p>\n<p>The article also argues that Russian AI markers differ in important ways from English ones. It points to administrative phrasing built around nouns rather than verbs, long genitive chains, repetitive use of &#8220;\u044f\u0432\u043b\u044f\u0435\u0442\u0441\u044f&#8221; as a crutch, bureaucratic pronouns such as &#8220;\u0434\u0430\u043d\u043d\u044b\u0439&#8221; and &#8220;\u0443\u043a\u0430\u0437\u0430\u043d\u043d\u044b\u0439&#8221;, and phrase-level translations of English idioms that sound unnatural in Russian. In the author&#8217;s view, these are not merely stylistic quirks but statistical signals that can expose machine-generated text.<\/p>\n<p>After the ban, he says he analysed his own rejected post using five simple metrics that any classifier could measure. The piece, he says, showed a highly regular listicle structure, heavy use of em dashes, a large share of quoted AI-style examples, and paragraph lengths that barely varied. In other words, even after phrase-level cleanup, the document still looked like a template-heavy AI essay rather than a human draft with irregular rhythm and digressions.<\/p>\n<p>That distinction matters because, as the author frames it, most humaniser tools work at sentence level while moderation systems judge the shape of the whole document. Local substitutions can make phrases sound more natural, but they do little if the article still has repeating section structures, uniformly sized paragraphs and a predictable information cadence. The piece argues that some subjects, especially explainers about AI patterns themselves, are structurally difficult to disguise because they must include examples of the very language a detector is looking for.<\/p>\n<p>He says the episode led to a second version of the tool, v0.2, built around three layers: phrase-level rewriting, a document-level audit mode and a stored &#8220;voice passport&#8221; designed to preserve a writer&#8217;s individual style across sessions. The new release also adds genre presets for places such as Habr, Telegram, email, vc.ru, LinkedIn and technical documentation. The developer presents that as a shift away from producing a generic &#8220;human&#8221; voice towards helping a tool adapt to a specific author and publication context.<\/p>\n<p>The broader backdrop is a Russian-language AI ecosystem that is becoming more contested. Research from RuATD 2022 showed that Russian automatic text detection is a serious and active field, while other studies have found that people are often poor at recognising AI-generated self-presentations in professional and social contexts. At the same time, Russian regulators have been moving towards tighter control of foreign AI systems, and reporting in 2025 and 2026 has described both Habr&#8217;s own hard line on AI-assisted posts and broader efforts to shape what kinds of machine-generated language are acceptable online.<\/p>\n<h3>Source Reference Map<\/h3>\n<p><strong>Inspired by headline at:<\/strong> <sup><a target=\"_blank\" rel=\"nofollow noopener noreferrer\" href=\"https:\/\/habr.com\/ru\/articles\/1030384\/?utm_source=habrahabr&amp;utm_medium=rss&amp;utm_campaign=1030384\">[1]<\/a><\/sup><\/p>\n<p><strong>Sources by paragraph:<\/strong><\/p>\n<p>Source: <a target=\"_blank\" rel=\"nofollow noopener noreferrer\" href=\"https:\/\/www.noahwire.com\">Noah Wire Services<\/a><\/p>\n<\/p><\/div>\n<div>\n<h3 class=\"mt-0\">Noah Fact Check Pro<\/h3>\n<p class=\"text-sm sans\">The draft above was created using the information available at the time the story first<br \/>\n        emerged. We\u2019ve since applied our fact-checking process to the final narrative, based on the criteria listed<br \/>\n        below. The results are intended to help you assess the credibility of the piece and highlight any areas that may<br \/>\n        warrant further investigation.<\/p>\n<h3 class=\"mt-3 mb-1 font-semibold text-base\">Freshness check<\/h3>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Score:<br \/>\n        <\/span>8<\/p>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Notes:<br \/>\n        <\/span>The article was published on April 27, 2025, and discusses a recent incident involving Habr&#8217;s automated moderation system rejecting an AI-generated text. The content appears original and timely, with no evidence of prior publication or significant recycling from other sources. However, the article&#8217;s focus on a specific event may limit its broader applicability.<\/p>\n<h3 class=\"mt-3 mb-1 font-semibold text-base\">Quotes check<\/h3>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Score:<br \/>\n        <\/span>7<\/p>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Notes:<br \/>\n        <\/span>The article includes direct quotes from the author regarding the moderation incident. These quotes are consistent with the author&#8217;s previous statements on the topic. However, without independent verification of the author&#8217;s claims, the authenticity of these quotes cannot be fully confirmed.<\/p>\n<h3 class=\"mt-3 mb-1 font-semibold text-base\">Source reliability<\/h3>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Score:<br \/>\n        <\/span>6<\/p>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Notes:<br \/>\n        <\/span>The article originates from Habr, a Russian-language platform known for user-generated content. While Habr is a reputable source within its niche, its content is user-generated and may lack the editorial oversight found in major news organisations. The author&#8217;s personal involvement in the incident may also introduce bias.<\/p>\n<h3 class=\"mt-3 mb-1 font-semibold text-base\">Plausibility check<\/h3>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Score:<br \/>\n        <\/span>7<\/p>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Notes:<br \/>\n    <\/span>The article presents a plausible account of the author&#8217;s experience with Habr&#8217;s moderation system. The technical details about AI-generated text detection and the author&#8217;s humaniser tool are consistent with known challenges in the field. However, the lack of independent verification of the incident raises some concerns about the accuracy of the account.<\/p>\n<h3 class=\"mt-3 mb-1 font-semibold text-base\">Overall assessment<\/h3>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Verdict<\/span> (FAIL, OPEN, PASS): <span class=\"font-bold\">FAIL<\/span><\/p>\n<p class=\"text-sm pt-0 sans\"><span class=\"font-bold\">Confidence<\/span> (LOW, MEDIUM, HIGH): <span class=\"font-bold\">MEDIUM<\/span><\/p>\n<p class=\"text-sm mb-3 pt-0 sans\"><span class=\"font-bold\">Summary:<br \/>\n        <\/span>The article presents a timely and original account of an incident involving Habr&#8217;s moderation system and the author&#8217;s AI-humaniser tool. However, the lack of independent verification, reliance on the author&#8217;s personal account, and potential bias due to the author&#8217;s involvement in the incident raise significant concerns about the reliability and objectivity of the content. Given these issues, the article does not meet the necessary standards for publication under our editorial guidelines.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>A Russian open-source developer\u2019s attempt to conceal AI-generated content using a new &#8216;humanising&#8217; tool reveals the limitations of current document-level detection systems, highlighting challenges unique to Russian language and stylistic patterns. A Russian open-source developer says his own tool for &#8220;humanising&#8221; AI-written text failed on a Habr article about AI-style writing, with the platform&#8217;s automated<\/p>\n","protected":false},"author":1,"featured_media":23371,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[40],"tags":[],"class_list":{"0":"post-23370","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-london-news"},"amp_enabled":true,"_links":{"self":[{"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/posts\/23370","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/comments?post=23370"}],"version-history":[{"count":1,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/posts\/23370\/revisions"}],"predecessor-version":[{"id":23372,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/posts\/23370\/revisions\/23372"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/media\/23371"}],"wp:attachment":[{"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/media?parent=23370"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/categories?post=23370"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/sandbox.hbmadvisory.com\/amplify\/wp-json\/wp\/v2\/tags?post=23370"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}