Google Gemini evades AI detectors more effectively than ChatGPT, study finds
Google’s Gemini model proved hardest for AI‑detection tools to flag, according to a study released by Open Resource Applications (ORA). The research pitted a dozen widely used generative AI systems against three popular detectors – Grammarly, QuillBot and GPTZero – by assigning each model the same long‑form writing task.
When the pieces were run through the detectors, Gemini’s output was flagged far less often than that of its competitors. Grammarly failed to identify any Gemini‑generated text, while QuillBot marked none as AI‑written. GPTZero, the most stringent of the three, still recognized most machine‑generated content but recorded a detection rate just under 98 percent, far higher than the other tools.
ChatGPT, the most familiar AI writer, performed poorly in the test. Its text triggered detection alerts across all three platforms, reinforcing the notion that widespread exposure to its style gives detectors a reliable template. Grok, another contender, fell somewhere in the middle, with detection rates higher than Gemini but lower than ChatGPT.
ORA’s spokesperson explained that Gemini’s advantage stems from its varied sentence structures and less predictable phrasing. Detection algorithms often look for repetitive patterns or familiar linguistic rhythms; Gemini’s more fluid approach makes those patterns harder to spot. “Tools like GPTZero flag predictability and overall structure,” the spokesperson said. “A model that reasons through ideas rather than recycling familiar phrases is a lot harder to catch.”
The disparity among detectors also raises practical concerns. A student’s essay could pass one tool and fail another, while a freelance writer might see their work labeled as AI‑generated depending on the software their client prefers. For organizations that filter content for authenticity, the inconsistency complicates enforcement.
Industry observers note that the study reflects a broader shift in AI‑generated content. As models diversify their styles, the once‑clear line between human and machine writing blurs. Some estimates suggest that roughly half of online content now contains AI‑generated elements, prompting platforms to develop automated filters. Yet those filters rely on detection tools that, as the ORA data show, vary widely in accuracy.
Google’s Gemini may have a temporary edge, but the race is far from over. Detection services are already updating their algorithms to recognize the newer patterns Gemini introduces. Meanwhile, other AI developers are likely to adopt similar techniques, potentially narrowing the gap.
For newsrooms and content teams, the takeaway is practical: reliance on a single detection platform is risky. Incorporating multiple tools—or developing in‑house verification processes—could provide a more reliable safety net as AI writing continues to evolve.
As AI content generation becomes a staple of the modern newsroom, tools that automate story creation and optimize SEO, such as AI news platforms and content‑management systems, must also grapple with authenticity concerns. The ORA findings underscore the need for robust, adaptable detection strategies in an era where the line between human and machine prose is increasingly indistinct.
Used: News Factory APP - news discovery and automation - ChatGPT for Business