AI Detection Glossary 2026 — 60 Essential Terms
Comprehensive vocabulary for educators, hiring managers, publishers, and policymakers evaluating AI-generated content. Covers detection algorithms, watermarking standards, evasion techniques, fairness considerations, and the 2026 tool ecosystem.
Navigate by category below. All terms include category tag, definition, and source where applicable.
Browse by Category (12)
Detection Math
Perplexity
A measurement of how surprised a language model is by a piece of text. AI-generated text typically has lower perplexity (more predictable) than human writing (more variation). Detectors compare a candidate text’s perplexity to a reference distribution.
Source: GPT-2 paper, Radford et al. 2019
Burstiness
Variation in sentence length and structure across a passage. Human writing is bursty (short sentences mixed with long); AI tends toward uniform length. Burstiness + perplexity together form the foundation of detectors like GPTZero.
Cross-Entropy
Information-theoretic distance between two probability distributions. Used by detectors to compare token probabilities under candidate text vs typical AI distributions.
Token-level probability
The likelihood a language model assigns to each individual token (word fragment) in a passage. Detectors visualize per-token probability heat maps to surface AI patterns.
Logits
Raw, unnormalized scores a language model produces before they are converted to probabilities. Some detectors use logit distributions for white-box detection of specific model families.
Entropy
Measure of randomness/unpredictability. Low-entropy passages (highly predictable) raise the AI-likelihood score; high-entropy passages (creative, unpredictable) lower it.
Detector Types
Black-box detector
A detector that has no access to the model weights that may have produced the text. Uses surface features (perplexity, burstiness, n-gram patterns). Examples: GPTZero, Originality.ai.
White-box detector
A detector with access to the originating model’s logits or probabilities. Higher accuracy but only works against specific known models. Rare in commercial tools.
Watermark detector
A detector that looks for invisible statistical signatures intentionally embedded by the generating model (e.g. Google SynthID for text). Highest accuracy but only works on watermark-aware models.
Stylometric detector
A detector that analyzes writing style features (function-word frequency, punctuation patterns, sentence parsers) rather than token statistics. More robust against paraphrasing.
Output Classes
False positive
(FP)Human-written text incorrectly flagged as AI-generated. The most damaging detector error in academic and hiring contexts. Rates vary 0.5%-12% across major detectors and content types.
False negative
(FN)AI-generated text incorrectly classified as human. The detection-evasion outcome targeted by paraphrase tools and humanizers.
True positive
(TP)AI-generated text correctly identified as AI.
True negative
(TN)Human-written text correctly classified as human.
Confidence score
A 0-100% value indicating the detector’s certainty. Most detectors are NOT calibrated probabilities — a "75% AI" score does not mean 75% chance the passage is AI.
Adversarial
Humanizer
Software that paraphrases AI-generated text to evade detection. Effective against perplexity-based detectors but defeated by stylometric and watermark methods.
Paraphrase attack
Detection-evasion technique using a second language model to rewrite a passage. Reduces detection accuracy by 30-60% across most commercial tools.
Adversarial example
Input crafted specifically to fool a detector while preserving meaning. Often involves character substitutions (Cyrillic ‘а’ for Latin ‘a’) or whitespace insertion.
Watermark stripping
Process of removing or weakening an invisible watermark via paraphrase, translation round-trip, or word-substitution attacks.
Round-trip translation
Translating AI text into another language and back. Disrupts detection signals but degrades fluency, leaving secondary detectable artifacts.
Watermarking
SynthID Text
Google DeepMind’s text watermarking system released 2024. Modifies token sampling probabilities in a way detectable with high accuracy by paired detector.
C2PA
(Coalition for Content Provenance and Authenticity)Industry standard for cryptographic content credentials. Embeds signed metadata about how an asset was created and modified. Adoption: Adobe, Microsoft, OpenAI, Meta as of 2025.
Content Credentials
Consumer-friendly name for C2PA metadata. Visible as a small "CR" icon on supporting platforms; click to see provenance chain.
IPTC Metadata
International Press Telecommunications Council standard for embedding rights, source, and provenance data in news media. Updated 2024 to align with C2PA.
ISCC
(International Standard Content Code)ISO 24138:2024 standard for content identification using perceptual hashing. Identifies content even after format conversion.
Statistical watermark
Watermarking method that biases token sampling toward a pseudorandomly-determined "green list" without changing meaning. Detectable via statistical test on token sequence.
Cryptographic signature
A digital signature attached to a content asset that proves authorship and detects tampering. Used in C2PA Content Credentials.
Generation
Large Language Model
(LLM)A neural network trained on text to predict the next token. Modern LLMs (GPT-5, Claude 4.7, Gemini 2.5, Llama 4) range from 7B to 1T+ parameters.
Foundation model
A large model trained on broad data, intended to be adapted to many downstream tasks. The pre-fine-tune state of an LLM.
Fine-tuning
Additional training of a foundation model on task-specific data. A fine-tuned model produces outputs that may be harder to detect than base-model outputs.
RLHF
(Reinforcement Learning from Human Feedback)Training method where humans rate model outputs and the model is reinforced to produce preferred responses. Source of much "AI assistant style" detected by tools.
Temperature
A sampling parameter (typically 0.0-2.0) controlling randomness. Low temperature produces predictable, easily-detected text; high temperature produces creative but error-prone text.
Top-k / Top-p sampling
Decoding strategies that restrict token selection to the most likely candidates. Strongly affects output detectability; nucleus sampling (top-p) is the most common in 2026.
Industry
Turnitin AI Writing Detection
AI detection feature integrated into Turnitin academic plagiarism platform. Used by 16,000+ institutions; precision/recall published only at aggregated level.
Generative AI policy
Institutional rules governing how AI tools may be used in submissions, hiring, or publication. Common 2026 categories: prohibited, disclosed-use, allowed-with-citation, fully-allowed.
AI disclosure
Requirement that authors declare AI assistance in a piece of work. Practiced by many academic journals and several major newspapers as of 2025.
Hiring screening
Use of AI detectors on resumes, cover letters, and writing samples. Legal exposure rising in 2025-2026 as false-positive impact disproportionate on non-native English speakers.
EU AI Act
European Union regulation effective 2025-2027 that requires AI-generated content to be labeled and watermarked under certain conditions.
Metrics
Precision
TP / (TP + FP). Of items the detector labeled AI, what fraction actually were AI. High precision = few false alarms.
Recall
TP / (TP + FN). Of all the actual AI content, what fraction the detector caught. High recall = few missed AI items.
F1 score
Harmonic mean of precision and recall: 2 * (P * R) / (P + R). Single-number summary of detector quality.
AUC-ROC
Area Under the Receiver Operating Characteristic curve. Measures detector quality across all confidence thresholds.
Calibration
Whether a detector’s confidence scores reflect actual probabilities. Most commercial AI detectors are poorly calibrated.
Benchmark dataset
A held-out collection of human-written and AI-written passages used to evaluate detectors. Common 2026 benchmarks: GPABench, RAID, AI Text Detect Challenge.
Content Types
Burstiness floor
The minimum burstiness expected from genuine human writing. Some technical genres (legal contracts, code documentation) naturally have low burstiness — leading to high false-positive rates.
Code detection
AI detection on programming source code. Far less reliable than text detection because programming languages are syntactically constrained.
Mixed-authorship
(Hybrid content)Content with both human and AI sections. Hardest case for detectors; often requires sentence-level rather than document-level analysis.
Image-detection
Detection of AI-generated images (Midjourney, DALL-E, Stable Diffusion). Different methodology from text — relies on artifact detection, prompt forensics, and metadata.
Voice / audio detection
Detection of synthetic speech and deepfake audio. Spectral analysis, formant patterns, ad-hoc breath signatures.
Fairness
Non-native English bias
Documented pattern where AI detectors flag writing by non-native English speakers as AI 5x-10x more often than equivalent native-speaker writing. Stanford 2023 paper.
Genre bias
Detectors over-flag certain content types (formal academic writing, technical documentation) and under-flag others (informal social media, dialogue). Source of disparate-impact concerns.
Adversarial robustness
A detector’s ability to maintain accuracy under deliberate evasion attempts.
Tools
GPTZero
Consumer-focused AI detector launched 2023 by Edward Tian; combines perplexity, burstiness, and stylometric features.
Originality.ai
Commercial AI detector integrated with content marketing workflows. Higher claimed accuracy than free tools but mixed independent benchmark results.
Copyleaks
Plagiarism + AI detection platform widely used by enterprises. AI module released 2023, frequently updated.
Winston AI
AI detector with strong performance on academic content per benchmarks; integrates with several university systems.
AI Humanizer tools
Software that processes AI text to evade detection. Examples: Undetectable.ai, Quillbot, Stealth Writer. Effectiveness varies by detector.
Emerging
Provenance graph
A directed graph representing the chain of edits and authorship for a piece of content, intended to be queryable in 2026 platforms.
Model attribution
Identifying which specific model (GPT-5, Claude 4.7, Gemini 2.5) produced a piece of content. Active research area; commercial tools approach 60-75% accuracy in 2026.
Distillation detection
Detecting whether a model was trained on outputs of another model (e.g. Llama distilled from GPT-4). Important for IP and licensing compliance.
Active provenance
Real-time embedding of provenance metadata as content is generated, vs post-hoc detection. The direction industry standards are moving in 2026.
Related Reading
Side-by-side accuracy benchmarks
AI Detector False Positive Rates by Content TypeAcademic, blog, code — independent measurements
C2PA Content Credentials 2026Cryptographic provenance adoption status
Perplexity and Burstiness ExplainedThe math behind AI text detection
AI Detection in Schools — Turnitin Policy 2026Educator-focused policy guide
AI Coding Assistants 2026 (Bytepane)Cursor, Copilot, Claude Code compared