Papers
arxiv:2406.19568

What Matters in Detecting AI-Generated Videos like Sora?

Published on Jun 27
· Submitted by JustinSheung on Jul 3
Authors:
,
,

Abstract

Recent advancements in diffusion-based video generation have showcased remarkable results, yet the gap between synthetic and real-world videos remains under-explored. In this study, we examine this gap from three fundamental perspectives: appearance, motion, and geometry, comparing real-world videos with those generated by a state-of-the-art AI model, Stable Video Diffusion. To achieve this, we train three classifiers using 3D convolutional networks, each targeting distinct aspects: vision foundation model features for appearance, optical flow for motion, and monocular depth for geometry. Each classifier exhibits strong performance in fake video detection, both qualitatively and quantitatively. This indicates that AI-generated videos are still easily detectable, and a significant gap between real and fake videos persists. Furthermore, utilizing the Grad-CAM, we pinpoint systematic failures of AI-generated videos in appearance, motion, and geometry. Finally, we propose an Ensemble-of-Experts model that integrates appearance, optical flow, and depth information for fake video detection, resulting in enhanced robustness and generalization ability. Our model is capable of detecting videos generated by Sora with high accuracy, even without exposure to any Sora videos during training. This suggests that the gap between real and fake videos can be generalized across various video generative models. Project page: https://justin-crchang.github.io/3DCNNDetection.github.io/

Community

Paper author Paper submitter

Hi @JustinSheung congrats on this work.

Are you planning to share models on the hub? If yes, here's how to do that: https://huggingface.co/docs/hub/models-uploading#upload-a-pytorch-model-using-huggingfacehub. They can then also be linked to this paper page as explained here.

·
Paper author

Hiiiii @nielsr

Thank you for sharing the information. We are considering making them available in the near future. I'll follow this guide when moving forward with that process.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2406.19568 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2406.19568 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2406.19568 in a Space README.md to link it from this page.

Collections including this paper 1