TY - GEN
T1 - An Empirical Study for Efficient Video Quality Assessment
AU - Sun, Wei
AU - Fu, Kang
AU - Cao, Linhan
AU - Zhu, Dandan
AU - Zhang, Kaiwei
AU - Zhu, Yucheng
AU - Zhang, Zicheng
AU - Hu, Menghan
AU - Min, Xiongkuo
AU - Zhai, Guangtao
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Video quality assessment (VQA) plays a critical role in optimizing various video processing systems, yet achieving both high accuracy and computational efficiency remains a challenging task. Recent advances in deep neural network (DNN)-based VQA models have led to notable performance improvements, but often at the cost of high computational complexity and memory consumption, limiting their applicability in resource-constrained scenarios. In this paper, we empirically investigate a set of good practices for building efficient yet effective VQA models. Specifically, we decompose the VQA training pipeline into three components: video preprocessing, quality-aware feature extraction, and optimization techniques. For each component, we identify and validate effective practices using the KVQ dataset - a user-generated content (UGC) VQA dataset that includes both in-the-wild distortions and processing-induced artifacts such as compression and enhancement. Based on these findings, we propose E-VQA, an efficient VQA model that combines the best-performing practices. Experiments conducted on the KVQ dataset, as well as the large-scale UGC VQA dataset LSVQ, demonstrate that E-VQA achieves competitive performance while significantly reducing computational complexity. Furthermore, E-VQA ranked third in the NTIRE 2025 Short-form UGC Video Quality Assessment Challenge, highlighting its practical effectiveness. The code is available at https://github.com/sunwei925/E-VQA.
AB - Video quality assessment (VQA) plays a critical role in optimizing various video processing systems, yet achieving both high accuracy and computational efficiency remains a challenging task. Recent advances in deep neural network (DNN)-based VQA models have led to notable performance improvements, but often at the cost of high computational complexity and memory consumption, limiting their applicability in resource-constrained scenarios. In this paper, we empirically investigate a set of good practices for building efficient yet effective VQA models. Specifically, we decompose the VQA training pipeline into three components: video preprocessing, quality-aware feature extraction, and optimization techniques. For each component, we identify and validate effective practices using the KVQ dataset - a user-generated content (UGC) VQA dataset that includes both in-the-wild distortions and processing-induced artifacts such as compression and enhancement. Based on these findings, we propose E-VQA, an efficient VQA model that combines the best-performing practices. Experiments conducted on the KVQ dataset, as well as the large-scale UGC VQA dataset LSVQ, demonstrate that E-VQA achieves competitive performance while significantly reducing computational complexity. Furthermore, E-VQA ranked third in the NTIRE 2025 Short-form UGC Video Quality Assessment Challenge, highlighting its practical effectiveness. The code is available at https://github.com/sunwei925/E-VQA.
UR - https://www.scopus.com/pages/publications/105017846671
U2 - 10.1109/CVPRW67362.2025.00129
DO - 10.1109/CVPRW67362.2025.00129
M3 - 会议稿件
AN - SCOPUS:105017846671
T3 - IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
SP - 1394
EP - 1404
BT - Proceedings - 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2025
PB - IEEE Computer Society
T2 - 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2025
Y2 - 11 June 2025 through 12 June 2025
ER -