Skip to content

ishitaakolkar/Deepfake_Detection

Repository files navigation

Deepfake Detection

This repository contains the implementation and experimentation of deepfake video classification using deep learning techniques such as Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) networks. Deepfakes—synthetically altered media generated using AI—pose increasing threats to authenticity and digital integrity. This project evaluates multiple model architectures to distinguish between real and fake facial video data.


Dataset used:

  1. DFDC: https://www.kaggle.com/competitions/deepfake-detection-challenge
  2. Celeb-DF: https://www.kaggle.com/datasets/reubensuju/celeb-df-v2

Work By:

  1. Ishita Akolkar
  2. Akshat Chhatriwala
  3. Dr Shankar Parmar

System Architecture

This model adopts a multi-stream architecture, where three parallel branches process distinct modalities of information:

  1. Spatial Stream

a. Inputs: RGB video frames b. Model: Lightweight CNN (e.g., MobileNet, ResNet-18) c. Role: Detect artifacts and texture-level anomalies within each frame

  1. Temporal Stream

  2. Inputs: Consecutive frame pairs

  3. Model: Optical flow extraction + CNN

  4. Role: Detect motion-level artifacts introduced by frame-to-frame inconsistencies

  5. Landmark Stream

a. Inputs: Facial landmark sequences over time b. Model: GCN / LSTM-based architectures c. Role: Detect unnatural geometric deformations or keypoint instability

Each stream produces independent predictions or feature embeddings. A late fusion module combines them via ensemble logic or learned fusion (e.g., MLP or attention-based layer) to yield the final classification.


Features

  • Face Extraction: Automatically detects and crops face regions from video frames for consistent input processing.
  • Deep Learning-Based Classifier: Custom CNN and LSTM-based models trained to learn spatial and temporal features indicative of manipulations.
  • Binary Classification: Classifies video input as real or fake with associated confidence scores.
  • Evaluation Metrics: Includes Accuracy, Precision, Recall, and Confusion Matrix for comprehensive model assessment.
  • Modular Architecture: Components for data preprocessing, training, evaluation, and visualization are modular and extensible.
  • Multi-Dataset Support: Easily adaptable to various benchmark datasets with flexible data loaders and configuration files.

Datasets

The models have been evaluated on publicly available deepfake datasets, including:

  • Deepfake Detection Challenge (DFDC)
  • Celeb-DF (v1 and v2)

Due to size constraints, datasets are not bundled with this repository.


Model Architectures

The system includes the following model variations:

  • CNN-Based Classifier: Learns spatial-level forgery artifacts on a frame-by-frame basis.
  • CNN + LSTM Hybrid: Utilizes temporal sequences of facial frames to capture inconsistencies across time.
  • Custom Feature Extractors: Tailored lightweight CNNs for faster training and generalization.

About

A multi-stream deepfake detection model based on spatial, temporal, and facial landmark features, designed to generalize across manipulation types through modular and complementary visual analysis.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors