AI Term 3 min read

Deepfake

Deepfakes are AI-generated synthetic media—images, video, or audio—in which a person's likeness or voice is convincingly replaced with someone else's using deep learning techniques.


A deepfake is a piece of synthetic media created by deep learning models—most commonly generative adversarial networks (GANs) or diffusion models—that realistically replaces or manipulates a person's face, voice, or body in images, video, or audio. The term combines "deep learning" and "fake" and was popularized around 2017 when face-swap techniques became accessible to non-experts.

How Deepfakes Are Created

Face Swapping
The original and most common technique:

  • Encoder-decoder autoencoders: Two autoencoders share a common encoder but have separate decoders trained on each subject; at inference time the source face is encoded and decoded with the target decoder.
  • GANs: A generator produces synthetic frames while a discriminator tries to detect fakes, forcing increasingly realistic output.
  • Diffusion models: Newer approaches use diffusion-based inpainting to blend faces with higher fidelity and fewer artefacts.

Voice Cloning
Audio deepfakes replicate a person's voice:

  • Neural text-to-speech (TTS): Models like VITS or Tortoise-TTS reproduce speech style and timbre from a short recording.
  • Voice conversion: Real-time pitch and prosody transformation to match a target speaker's voice.

Full-Body Synthesis

  • Pose-driven video synthesis: Models such as DensePose or ControlNet generate realistic body movement from skeleton or depth input.
  • Talking-head models: Systems like SadTalker animate a still photo using an audio clip.

Detection Methods

Passive Detection
Automated analysis without watermarking:

  • Frequency artefacts: Deepfake generators often leave characteristic patterns in the frequency domain detectable by CNNs.
  • Biological signals: Subtle inconsistencies in blinking, eye reflections, skin texture, or pulse signals detectable via rPPG.
  • Temporal inconsistency: Frame-level classifiers look for unnatural flickering or blending boundaries across time.
  • Foundation model detectors: Models such as UniversalFakeDetect or Grounding DINO fine-tuned on deepfake datasets.

Active Provenance
Embedding authenticity evidence at creation:

  • Cryptographic signing: Camera manufacturers or platforms sign frames at capture time (C2PA standard).
  • Invisible watermarks: Perturbations embedded in genuine media that survive re-encoding, enabling origin verification.

Risks and Misuse

  • Non-consensual intimate imagery (NCII): The most prevalent harmful use, disproportionately targeting women.
  • Political disinformation: Fake video or audio of public figures spreading false statements.
  • Fraud and social engineering: Impersonating executives in video calls to authorize wire transfers (CEO fraud).
  • Identity theft: Bypassing facial-recognition authentication systems.

Legitimate Applications

  • Film and post-production: De-aging actors, stunt replacement, and recreating deceased performers.
  • Localisation: Dubbing with lip-sync that matches the target language.
  • Accessibility: Generating sign-language avatars or synthetic training data.
  • Education and simulation: Historical reconstructions, medical simulation.

Regulatory Landscape

Several jurisdictions have introduced or proposed laws specifically targeting malicious deepfakes, including the US DEFIANCE Act (2024), the EU AI Act's transparency requirements for AI-generated content, and a growing number of state-level statutes criminalizing NCII deepfakes.

Deepfakes represent one of the most concrete examples of the dual-use nature of generative AI: the same technology that enables powerful creative tools can cause serious harm when misused, making detection, provenance, and regulation active areas of research and policy.

← Back to Glossary
EU Made in Europe

Chat with 100+ AI Models in one App.

Use Claude, ChatGPT, Gemini alongside with EU-Hosted Models like Deepseek, GLM-5, Kimi K2.5 and many more.

Customer Support