User Rating 0.0 โ˜…โ˜…โ˜…โ˜…โ˜…
Total Usage 0 times
0 words
Paste text and click "Analyze Text" to see results
Is this tool helpful?

Your feedback helps us improve.

โ˜… โ˜… โ˜… โ˜… โ˜…

About

Large language models produce text with statistically detectable patterns. The average sentence length variance (ฯƒlen) in AI-generated prose is typically 30 - 50% lower than in human writing. AI models optimize for coherence, which flattens the natural "burstiness" of human cognition - the tendency to alternate between short, punchy sentences and longer, complex ones. This tool quantifies that signal alongside 8 other linguistic features: Type-Token Ratio, hapax legomena density, transition word overuse, punctuation diversity, syllable distribution, readability index, and n-gram repetition patterns. Each metric is scored independently, then combined into a weighted composite.

No detector is infallible. Heavily edited AI text or formulaic human writing (legal documents, technical manuals) can produce false results. The composite score assumes standard English prose of at least 50 words. Accuracy degrades below that threshold. Mixed-origin text (human-written with AI-assisted edits) will score in the ambiguous middle range, which is the correct result. Treat scores below 30 or above 70 as directional signals, not verdicts.

ai detection text analysis ai content detector human vs ai writing analysis linguistic analysis burstiness perplexity

Formulas

The composite human-likeness score H is computed as a weighted sum of normalized sub-scores:

H = nโˆ‘i=1 wi โ‹… si

Where si is the normalized score (range 0 - 1) for each metric and wi is its weight such that nโˆ‘i=1 wi = 1.

The Burstiness Index B is defined as:

B = ฯƒlen2L

Where ฯƒlen2 is the variance of sentence lengths and L is the mean sentence length in words.

The Flesch Reading Ease score:

FRE = 206.835 โˆ’ 1.015 ร— WS โˆ’ 84.6 ร— YW

Where W = total words, S = total sentences, Y = total syllables.

Type-Token Ratio:

TTR = |V|N

Where |V| is the count of unique word types and N is the total number of word tokens.

Reference Data

MetricWhat It MeasuresTypical Human RangeTypical AI RangeWeight
Sentence Length Variance (ฯƒlen)Standard deviation of words per sentence6 - 142 - 620%
Burstiness IndexRatio of variance to mean in sentence lengths0.5 - 1.50.1 - 0.415%
Type-Token Ratio (TTR)Lexical diversity: unique words รท total words0.55 - 0.800.40 - 0.6012%
Hapax Legomena RatioWords used exactly once รท total words0.40 - 0.650.25 - 0.4010%
Transition Word DensityDiscourse markers per 100 words0.5 - 2.02.5 - 5.012%
Punctuation DiversityUnique punctuation types รท total punctuation0.30 - 0.700.15 - 0.308%
Flesch Reading EaseReadability based on syllables and sentence length40 - 80 (varies)50 - 65 (narrow band)8%
Readability VariancePer-sentence Flesch score standard deviation15 - 355 - 158%
Bigram Repetition RateRepeated word pairs รท total bigrams0.02 - 0.080.08 - 0.187%
Average Syllables per WordVocabulary complexity indicator1.3 - 1.81.5 - 1.7 (consistent) -
Paragraph Length VarianceConsistency of paragraph sizingHigh (irregular)Low (uniform) -
Contraction Usage RateContractions per 100 words1.0 - 4.00.1 - 0.8 -
Unique Sentence StartersVariety in first words of sentences60 - 90%35 - 55% -

Frequently Asked Questions

Language models optimize for fluency by maintaining consistent sentence structures. The standard deviation of sentence word counts (ฯƒlen) in AI text typically falls between 2 and 6, while human writers naturally produce values of 6 to 14. Humans unconsciously vary rhythm: a 5-word declarative followed by a 35-word compound sentence. AI smooths this variation. The burstiness metric quantifies this pattern as variance divided by mean.
Statistical metrics like Type-Token Ratio and sentence length variance require sufficient sample size to stabilize. Below 50 words, there may only be 2 - 4 sentences, which is insufficient to compute meaningful variance. TTR is also length-dependent: shorter texts naturally produce higher ratios regardless of origin. The tool will still run on shorter text but flags the result as low-confidence.
Yes. If a human rewrites AI output by varying sentence lengths, adding contractions, inserting parenthetical asides, and breaking uniform paragraph structure, the statistical fingerprint shifts toward human-like values. The tool measures the text as-is, not its origin process. Conversely, human-written text that follows strict templates (legal briefs, military reports) may score as AI-like due to enforced structural uniformity.
The tool checks against a list of 80+ discourse markers ("however", "moreover", "furthermore", "in addition", "consequently"). The count is normalized per 100 words. AI models overuse transitions because training rewards text coherence. Human writers often rely on implicit logical connections or paragraph breaks instead. A density above 2.5 per 100 words is a moderate AI signal.
The syllable counting algorithm and transition word list are calibrated for English. Sentence splitting (period/exclamation/question mark) works cross-linguistically, so metrics like sentence length variance and burstiness remain valid. However, TTR, hapax ratio, and readability scores will be inaccurate for non-Latin scripts or agglutinative languages like Finnish or Turkish where word boundaries differ.
Hapax legomena are words that appear exactly once in a text. Human writers tend to produce more hapax because they draw from personal vocabulary idiosyncrasies, rare word choices, and context-specific terminology. AI models favor statistically common words from training distributions. A hapax ratio below 0.30 in a 300+ word sample is a moderate indicator of machine generation.