AIGC-VQA: A Holistic Perception Metric for AIGC Video Quality Assessment

Yiting Lu, Xin Li, Bingchen Li, Zihao Yu, Fengbin Guan, Xinrui Wang, Ruling Liao, Yan Ye, Zhibo Chen; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2024, pp. 6384-6394

Abstract


With the development of generative models such as the diffusion model and auto-regressive model AI-generated content (AIGC) is experiencing an explosive growth. Moreover existing quality metrics extracted from fixed pre-trained models struggle to align accurately with human perception. There is an urgent need for an adaptive metric capable of gauging the multiple critical factors (i.e. technical quality aesthetic quality and video-text alignment) related to quality within AIGC videos to provide quality assessment and guide optimization of generative models. In this work we propose a holistic metric for AIGC video quality assessment termed AIGC-VQA which contains three functional branches for the cooperation on technical aesthetic and video-text alignment aspects in AIGC videos. Specifically to efficiently transfer the knowledge of image-text alignment to the video-text alignment we introduce the spatial-temporal adapter to exploit the pre-trained prior from a large-scale image-text model and achieve the temporal knowledge adaptation. Besides we propose a divide-and-conquer training strategy for progressive cooperation on multiple branches. Due to the holistic perception ability our proposed AIGC-VQA obtains state-of-the-art results on the T2VQA-DB dataset.

Related Material


[pdf]
[bibtex]
@InProceedings{Lu_2024_CVPR, author = {Lu, Yiting and Li, Xin and Li, Bingchen and Yu, Zihao and Guan, Fengbin and Wang, Xinrui and Liao, Ruling and Ye, Yan and Chen, Zhibo}, title = {AIGC-VQA: A Holistic Perception Metric for AIGC Video Quality Assessment}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2024}, pages = {6384-6394} }