What the detector
is actually doing.
AI Slop Detector uses a weighted scoring pipeline rather than a one-rule shortcut. It combines transformer inference, phrase-density heuristics, sentence-level inspection, and explicit short-sample calibration to estimate whether a passage behaves more like AI-generated prose or ordinary human writing.
The transformer model is the primary classifier. Heuristics act as a smaller correction layer, sentence scoring makes the output inspectable, and low-word-count samples are intentionally compressed toward neutral because short text has high variance. The system is built to produce an evidence-weighted signal, not a proof of authorship.
Every input is normalized, split into chunks for long passages, passed through a RoBERTa-based detector, scanned for tracked phrase patterns, and then combined into a final scalar score on the interval [0, 1], where higher values indicate stronger AI-like behavior.
For longer passages, the transformer model dominates the blend because it captures broader language patterns better than surface-level rules.
Phrase heuristics stay intentionally lighter. They are useful for explainability and nudging borderline samples, but they should not overpower the model.
h = min(((Σ cjwj) / (word_count / 100)) / 12, 1)
long-text score = 0.82m + 0.18h
In the heuristic term above, cj is the match count for tracked phrase j and wj is that phrase's configured weight. The density is normalized by words-per-100 so a long article is not penalized just for having more total tokens.
Short text is handled differently because one or two sentences produce unstable model variance. For samples under 120 words, the detector adds a sentence-average term. For samples under 50 words, it also applies an explicit trust factor that shrinks the score back toward a baseline.
if w ≤ 60: b = 0.55m + 0.25s + 0.20h
if 60 < w ≤ 120: b = 0.62m + 0.23s + 0.15h
if w < 50: final = baseline + (b - baseline)t
The trust term t is bounded so very short text cannot swing too far in either direction. That is why a 20-word sample may return a low-reliability or unclear reading even when the raw model score is more extreme.
The app surfaces more than a raw number. It also ranks individual sentences and exposes weighted phrase matches so the output can be inspected, not just accepted at face value.
Very short text is the hardest case for AI detection. A 20-word sample simply does not contain enough syntactic, lexical, and distributional structure for a robust verdict, so the detector compresses those scores back toward neutral and may mark them as low reliability.
t = clamp(0.08, 0.32, ((w - 50) / 30) × 0.24 + 0.08)
score = baseline + (b - baseline)t
No detector can prove authorship. Heavily edited AI writing can look human. Highly structured human writing can sometimes look model-like. Scraped text can also be noisy if the page extractor pulls boilerplate, disclaimers, or navigation text into the sample.