-
[pdf]
[supp]
[arXiv]
[bibtex]@InProceedings{Elmoghany_2025_ICCV, author = {Elmoghany, Mohamed and Rossi, Ryan and Yoon, Seunghyun and Mukherjee, Subhojyoti and Bakr, Eslam Mohamed and Mathur, Puneet and Wu, Gang and Lai, Viet Dac and Lipka, Nedim and Zhang, Ruiyi and Manjunatha, Varun and Van Nguyen, Chien and Dangi, Daksh and Salinas, Abel and Chen, Hongjie and Huang, Xiaolei and Barrow, Joe and Ahmed, Nesreen and Eldardiry, Hoda and Park, Namyong and Wang, Yu and Tu, Zhengzhong and Nguyen, Thien Huu and Manocha, Dinesh and Elhoseiny, Mohamed and Dernoncourt, Franck}, title = {A Survey on Long-Video Storytelling Generation: Architectures, Consistency, and Cinematic Quality}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops}, month = {October}, year = {2025}, pages = {7082-7094} }
A Survey on Long-Video Storytelling Generation: Architectures, Consistency, and Cinematic Quality
Abstract
Despite the recent progress in video generative models, existing state-of-the-art methods can only produce videos lasting 5-16 seconds, often labeled "long-form videos". Furthermore, videos exceeding 16 seconds struggle to maintain consistent character appearances and scene layouts throughout the narrative. In particular, multi-subject long videos still fail to preserve character consistency and motion coherence. While some methods can generate videos up to 150 seconds long, they often suffer from frame redundancy and low temporal diversity. Recent work has attempted to produce long-form videos featuring multiple characters, narrative coherence, and high-fidelity detail. We studied 32 papers on long-video generation to identify key architectural components and training strategies that consistently yield these qualities. We also construct a comprehensive novel taxonomy of existing methods and present comparative tables that categorize papers by their architectural designs and performance characteristics.
Related Material
