Abstract
Recent advances in generative AI have enabled multiple pathways for high‑fidelity video synthesis; text‑to‑video, image‑to‑video animation, and video out painting. Empirical side-by-side evaluations of how untrained viewers perceive and distinguish these outputs from real footage remain scarce. In this study, we systematically compare human detection accuracy across these three AI generation techniques within three thematic contexts: historical footage, film and media content, and natural environments. We constructed a balanced stimulus set comprising equal amounts of real and AI- generated videos (18 each). The AI clips were evenly distributed across the three generation methods using Google’s Veo 3, Lightricks LTX Video, and Wan VACE. All videos were produced and standardized within the ComfyUI framework to ensure consistent quality and duration. 87 participants judged every clip in a binary forced-choice task (“real” vs. “AI-generated”). Participants correctly identified videos 60% of the time on average. Image-to-Video clips were recognized most accurately (79%), followed by real footage (64%), out painting (49%), and text-to-video (43%). Accuracy also varied by theme: film and historical scenes yielded higher detection rates than environmental clips, which were frequently mistaken for AI. Logistic regression confirmed significant effects of both technique and theme as well as their interaction (p < 0.001), indicating that detection success depends jointly on how the content was generated and what it depicts. Findings reveal a consistent bias toward assuming synthetic origins and highlight that perceptual realism in AI video is shaped more by context than by model type, underscoring the importance of media-literacy approaches and context-aware evaluation tools for navigating increasingly synthetic visual media.
Keywords
Generative AI; AI-generated video; Human perception; Media literacy
DOI
https://doi.org/10.21606/iasdr.2025.1050
Citation
Sarfraz, D., Karim, R.M.,and Kim, K.(2025) Synthetic Realities: Evaluating Human Ability to Distinguish AI-Generated Videos from Real Footage, in Chang, C.-Y., and Hsu, Y. (eds.), IASDR 2025: Design Next, 02-05 December, Taiwan. https://doi.org/10.21606/iasdr.2025.1050
Creative Commons License

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License
Conference Track
Track 4 - Human-Centered AI
Synthetic Realities: Evaluating Human Ability to Distinguish AI-Generated Videos from Real Footage
Recent advances in generative AI have enabled multiple pathways for high‑fidelity video synthesis; text‑to‑video, image‑to‑video animation, and video out painting. Empirical side-by-side evaluations of how untrained viewers perceive and distinguish these outputs from real footage remain scarce. In this study, we systematically compare human detection accuracy across these three AI generation techniques within three thematic contexts: historical footage, film and media content, and natural environments. We constructed a balanced stimulus set comprising equal amounts of real and AI- generated videos (18 each). The AI clips were evenly distributed across the three generation methods using Google’s Veo 3, Lightricks LTX Video, and Wan VACE. All videos were produced and standardized within the ComfyUI framework to ensure consistent quality and duration. 87 participants judged every clip in a binary forced-choice task (“real” vs. “AI-generated”). Participants correctly identified videos 60% of the time on average. Image-to-Video clips were recognized most accurately (79%), followed by real footage (64%), out painting (49%), and text-to-video (43%). Accuracy also varied by theme: film and historical scenes yielded higher detection rates than environmental clips, which were frequently mistaken for AI. Logistic regression confirmed significant effects of both technique and theme as well as their interaction (p < 0.001), indicating that detection success depends jointly on how the content was generated and what it depicts. Findings reveal a consistent bias toward assuming synthetic origins and highlight that perceptual realism in AI video is shaped more by context than by model type, underscoring the importance of media-literacy approaches and context-aware evaluation tools for navigating increasingly synthetic visual media.