The Deepfakes We Missed

We Built Detection Systems for a Threat Model That Never Fully Materialized
Shaina Raza
Vector Institute for Artificial Intelligence

Abstract

Deepfake detection research has largely focused on a threat model inherited from the 2017–2019 wave of public concern: face-swap and talking-head manipulation targeting politicians, celebrities, and public figures. This paper argues that the dominant harms that ultimately emerged between 2022 and 2026 differ substantially from those assumptions. Real-world incidents are now concentrated around peer-generated non-consensual intimate imagery (NCII), voice-clone scam calls, emotional-manipulation fraud, and private messaging-based distribution. Meanwhile, benchmark design, datasets, and detection systems remain heavily concentrated on public-figure video manipulation. We present a large-scale classification of 438 papers published between 2017 and 2025 across five threat categories and compare research allocation against observed harm distributions. Our analysis suggests that the primary bottleneck in practical deepfake defense is no longer model capability alone, but a persistent mismatch between research priorities and deployed harms. We further identify structural causes behind this misalignment and outline three concrete research agendas for under-defended categories.

Core Argument. The main limitation in real-world deepfake defense is increasingly a mismatch between the threat models prioritized by the research community and the harms observed in practice. Future work should place greater emphasis on telecommunications-scale voice-clone detection, privacy-preserving NCII protection, and messaging-layer defenses for peer-distributed synthetic media.


Key Findings
71%
Research Focus
Papers targeting public-figure video manipulation
260×
Growth
Increase in reported AI-CSAM videos from 2024 to 2025
0 / 13
Benchmarks
Benchmarks directly evaluating T2, T4, or T5 threats
438
Papers
Detection papers categorized from 2017–2025

Research Allocation vs. Observed Harm
75
T1 Papers
(2024)
3,443
Reported AI-CSAM
Videos
22,364
IC3 Complaints
(2025)
Figure. Comparison between research allocation and observed harm indicators. Public-figure video detection dominates the literature, while rapidly growing harm categories remain comparatively under-explored.

Threat Taxonomy
ID Threat Category Corpus Share
T1 Public-figure face-swap and talking-head video 71.0%
T3 Audio and voice-clone detection 28.5%
T2 Peer-generated NCII Minimal representation
T5 Messaging-layer and peer-distributed manipulation Minimal representation
T4 Real-time and live-stream manipulation No dedicated papers identified

Main Contributions
  1. A classification of 438 deepfake detection papers published between 2017 and 2025 across five threat categories.
  2. An empirical comparison between research allocation and observed real-world harms using evidence synthesized from IC3, IWF, AIID, and victim-centered reporting systems.
  3. A structural analysis explaining why research misalignment persists, including benchmark inheritance, dataset asymmetry, and visibility-driven prioritization.
  4. Three research agendas targeting under-defended harm categories with strong societal relevance and technical feasibility.

Research Agendas
I. Real-Time Voice-Clone Detection for Telecommunications
Development of streaming synthesis detection systems robust to telephony codecs, packet loss, and low-bandwidth environments while maintaining sub-second latency constraints.
II. Privacy-Preserving NCII Detection
On-device and federated approaches for victim-centered verification workflows where sensitive candidate images remain local to the user device.
III. Messaging-Layer Defenses
Detection systems designed for peer-distributed synthetic media, including lightweight encrypted-platform detectors and distribution-pattern analysis under privacy constraints.

BibTeX
@inproceedings{raza2026deepfakeswemissed,
  title  = {The Deepfakes We Missed:
            We Built Detection Systems
            for a Threat Model That Never Fully Materialized},
  author = {Raza, Shaina},
  year   = {2026}
}