Deepfake Detector - from crawler to cloud interface

TL;DR

We designed and shipped an end-to-end pipeline: custom short-video crawler → cloud storage → face-aware frame sampling → multi-model voting → auditable API and web UI. It’s pragmatic, scalable, and easy to extend with better models or new sources.


Introduction

This started as a uni project that we treated like a product. Goal: ingest short-form videos, analyze them with multiple ML models, and expose trustworthy results through an API and a small web app. Human reviewers can verify detections to keep the system honest and continuously improve it.

What made it hard

Architecture at a glance

deepfakedetector-schema-light.png

High level flow: 1. Crawler discovers and downloads videos, uploads them to object storage, and creates metadata records. 2. Analyzer samples frames, detects faces, runs deepfake models, aggregates results, and writes back a verdict. 3. API + Web expose users, videos, detections, and analytics with role-based access.

Pipeline, in practice

Crawler

Analyzer

API & Web

Reliability, ops, and cost

Results that mattered

Engineering notes

What’s next