← Humdraft Blog

Why AI Detectors Flag Your Writing (Even When You Wrote It)

Here’s a thing nobody tells you about AI detectors: they don’t actually read your writing. They don’t care whether your argument is sharp, whether your conclusion is yours, whether you spent four hours rewriting that opening paragraph. They measure rhythm. And rhythm, it turns out, is something honest writers get wrong all the time.

We hear this every week from Humdraft users — graduate students, freelance writers, ESL professionals, marketers — who write their own drafts, who never even open ChatGPT for the piece in question, and who still get flagged at 80, 90, 99 percent AI. Their professor is suspicious. Their editor is mad. Their client wants a refund. And the writer is left holding a screenshot of a number that supposedly proves they’re a robot.

So let’s actually open the hood. Here’s how AI detectors work, why honest writers trip them, and what to do about it before you submit.

How detectors actually work (the two-minute version)

Every major AI detector — GPTZero, Turnitin, Originality.ai, Copyleaks, ZeroGPT, Winston — fundamentally measures two things: perplexity and burstiness.

Perplexity is a measure of how surprising your next word is, given the previous words. Language models are trained to produce the most likely next word at every step. So when a detector reads your text and predicts your next word easily, your text looks AI-generated. When your next word is unexpected — a weird verb choice, an odd transition, a sudden contraction — perplexity goes up, and your text looks human.

Burstiness is the variation in sentence length and complexity across your draft. Humans hum. We write a 28-word sentence, then a 4-word one, then a fragment. Then a question? Then we keep going for another long stretch because we got into our groove. AI is much smoother — uniform sentences, even pacing, predictable rhythm. The variance is the tell.

Some detectors layer on classifiers trained specifically on GPT-4, Claude, and Gemini outputs. Some look at vocabulary entropy, syntactic similarity to known model outputs, or stylometric fingerprints. But underneath, it’s the same core idea: does this read like a confident model that picked the most likely word at every step?

Why honest writers get false-flagged

Here’s the cruel part. The same things that make writing “professional” — the things you were taught to do in school — happen to be the exact things that make perplexity go down and burstiness flatten out.

  • You write carefully. You revise sentence-level pacing. You smooth transitions. You use a thesaurus and pick the “right” word every time. Congratulations: you sound like a model that picked the most likely word every time.
  • You’re a non-native speaker. ESL writers tend to use more textbook-correct grammar, fewer idioms, and a narrower vocabulary range. Detectors read this as low entropy and flag accordingly. There are studies showing GPTZero misclassifies up to 60% of ESL student essays as AI.
  • You write in a specialized register. Legal, academic, medical, technical writing all converge on conventions: passive voice, formal connectors, longer sentences. That convergence reads as model-like uniformity.
  • You used Grammarly. Or any other rewriting assistant. Once you let a tool smooth your prose, you’ve flattened the burstiness signature.
  • You wrote a short piece. Detectors get less reliable with less text. A 200-word email or paragraph is statistically too small to score confidently — but the detector will give you a number anyway.

The detectors all behave differently

It’s also worth knowing that detectors don’t agree with each other. Same draft, three tools, three verdicts. Here’s the rough shape of what we see in practice:

GPTZero

Tends to be the most aggressive on academic and ESL writing. Its “sentence-level” mode highlights specific sentences as AI, which is helpful for editing — but the per-sentence judgements are noisy on short or careful prose. False-positive rate on human academic writing is the highest of the major detectors.

Turnitin

Used in roughly 16,000 institutions, so it’s the one most students encounter. Turnitin will only return a number to instructors (not students), and that number is opaque — there’s no breakdown. It tends to be conservative on short submissions and more confident on long-form pieces. Its bigger problem is that students never see the score until after they submit, so they can’t self-check.

Originality.ai

Designed for SEO and content marketing, not academia. It scores at a paragraph level, gives you a clean “Original / AI” verdict, and tends to be the most lenient on careful long-form prose. But it’s very strict on listicles, FAQ pages, and product copy — basically anything with structural repetition.

Copyleaks & Winston AI

Both came up newer, both lean strict, and both are aggressively used by enterprise content platforms doing batch scans. Copyleaks reports paragraph-level confidence; Winston layers on plagiarism detection.

What to actually do about it

If you’ve been flagged or you’re about to submit something high-stakes, the workflow is shorter than you think. We built Humdraft around this exact two-step routine:

  1. Check before you submit. Use an AI checker to see your scores across multiple detectors at once — not just the one your professor or editor uses, because you don’t always know which one. If three detectors say “human” and one says “AI,” you have an outlier problem, not a writing problem. If all four say “AI,” you have a rhythm problem.
  2. Polish the rhythm. Drop the same draft into Humdraft and run a humanize pass. The rewrite increases burstiness, varies word choice, breaks up the smoothest cadences, and keeps your meaning intact. Multi-detector scores come back with the output, so you see the result before you ship.

That’s the whole shape of it. Detectors don’t go away — they’re going to be deployed at every level of academic, professional, and content workflows for the foreseeable future. The skill is no longer “don’t use AI.” The skill is reading your own rhythm and knowing when your draft has become smoother than yourself.

We think honest writers deserve to see their own scores. We think the verdict shouldn’t come from one black box. And we think that if you’re going to be judged by a robot for the rhythm of your prose, you should at least get to see the receipt before it’s sent.

Try Humdraft free — 500 words, no signup.
Paste your draft, hit Humdraft, and watch four detector scores drop in under three seconds.
Humdraft your draft

Honest notes on AI writing, detector quirks, and product updates worth your inbox. No spam, ever.