Technical primer

How AI content detection works.

AI content detectors are trying to answer a deceptively simple question: does this text read more like a human or a language model wrote it? The short answer is that they look for statistical patterns in word choice, sentence rhythm, and predictability that humans and LLMs leave behind in different shapes. Below is how it actually works, why no detector is perfect, and how to use AI scores responsibly.

What is AI text detection?

AI text detection is the task of estimating the probability that a given passage of text was generated by a large language model (LLM) like ChatGPT, Claude, or Gemini, rather than written by a human. The output is usually a percentage — say, 87% likely AI — paired with a sentence-level breakdown that points to the specific lines that triggered the score.

It is fundamentally a probabilistic problem. AI detectors do not read text the way you do. They convert each word into numbers, run those numbers through statistical tests, and report a confidence interval. The closer to 100%, the more confident the detector is that the text is AI. The closer to 0%, the more confident it is that the text is human.

Three signals every detector uses.

Modern AI content detectors stack three complementary signals. Each one alone is fooled easily; together, they are much harder to escape.

1. Perplexity — how predictable is this text?

Perplexity measures how surprised a reference language model is by the next word in a sequence. AI-generated text is, by construction, low-perplexity: the model picked each word because it was statistically the most probable next token. Human writing is messier — we use unusual word combinations, idioms, and tangents that look unlikely to a language model. A passage with consistently low perplexity reads as “AI-shaped” even if you cannot articulate why.

2. Burstiness — does the rhythm vary?

Burstiness captures variation in sentence length, structure, and cadence across a passage. Human writing tends to be bursty: a long winding sentence followed by a short punchy one. AI tends toward uniform, mid-length sentences with similar grammatical structure — the literary equivalent of a metronome. Burstiness is a strong signal because it is hard to fake without manually rewriting almost every sentence.

3. Lexical fingerprinting — does this match a known model?

Each major LLM has subtle preferences: GPT-4 over-uses certain hedging phrases, Claude tilts toward parenthetical asides, Gemini has its own vocabulary fingerprint. Lexical fingerprinting compares your text against a catalog of known model outputs and reports the closest match. This is what lets a good detector say not just “AI-likely” but this looks like GPT-4 or Claude.

Why detection is hard.

AI detection is an arms race. Three things make it genuinely difficult:

  • Paraphrasing. Tools that rewrite AI text to sound more human can lower perplexity scores. Multi-signal detectors handle this reasonably well, but heavy paraphrasing remains the biggest weakness.
  • Mixed authorship. Real-world text is often a blend: a human writes a draft, asks ChatGPT to clean it up, and edits the result. The signal becomes mushy. The right answer is a sentence-level breakdown, not a single percentage.
  • Evolving models. Every new model release shifts the statistical fingerprint. Detectors that do not update keep flagging the wrong things — or miss new generations entirely. AI Checker retrains on every major model release.

How AI Checker handles each challenge.

We combine perplexity, burstiness, and lexical fingerprinting, weighted differently per language and content type. We retrain continuously against output from GPT-4, Claude 3, Gemini, Llama, Mistral, and several open-source models so that the fingerprinting signal stays current. And we surface a sentence-by-sentence breakdown in every result, because mixed authorship is the rule, not the exception. You can see exactly which sentences pushed the score up — and decide whether the call is fair.

For privacy: text is processed in memory and is never used to train models. We do not log the contents of submissions on the free tier by default, and you can opt out of all logging in settings.

Frequently asked technical questions.

What is perplexity in AI detection?

Perplexity is a measure of how 'surprised' a language model is by a sequence of text. AI-written text tends to be highly predictable — and therefore low in perplexity — because it was generated to maximize probability at every step. Human writing is messier and produces higher perplexity scores.

What is burstiness?

Burstiness measures the variation in sentence length, structure, and rhythm across a passage. Humans naturally write in bursts: a long winding sentence, then a short punchy one. AI tends to produce text with much more uniform sentence-level cadence, which is a strong giveaway.

Can AI detection be fooled?

Sometimes — but it gets harder as detectors evolve. Heavy paraphrasing, manual editing, or running text through 'humanizer' tools can lower a score, but they rarely zero it out. Combining multiple signals (perplexity, burstiness, fingerprinting) makes evasion much more expensive than just using AI in the first place.

Why do detectors disagree?

Each detector weighs signals differently, was trained on different model outputs, and uses different thresholds. A 60% score on one tool can be 40% on another. The honest answer is that no single number is authoritative — the right approach is to look at the sentence-level breakdown and the patterns, not just the headline percentage.

Is AI detection ever 100% accurate?

No, and any tool that claims so is overselling. State-of-the-art detectors hover around 95–98% on unedited AI text. False positives on real human writing are rare but real — about 1–2% in practice. That's why we always recommend AI scores be one input in a human review process, never the final verdict.

When to use AI detection.

AI detection is a screening tool, not a verdict. The right time to use it:

  • Educators reviewing essays, lab reports, and admissions material — pair the score with classroom context and version history.
  • Editors vetting freelance submissions — use it to decide where to focus your line edit, not whether to reject.
  • Recruiters screening cover letters and take-home tasks — flag for follow-up, do not auto-reject.
  • SEO and content teams auditing agency deliverables for E-E-A-T compliance.

Ready to try it? Run a check on the AI Checker home page — no signup required.

Try AI Checker on your own text.

Free, instant, and private. No signup. No credit card.

Check text now