Hierarchical NeuroSymbolic Approach for Comprehensive and Explainable Action Quality Assessment

Lauren Okamoto, Paritosh Parmar; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2024, pp. 3204-3213

Abstract


Action quality assessment (AQA) applies computer vision to quantitatively assess the performance or execution of a human action. Current AQA approaches are end-to-end neural models which lack transparency and tend to be biased because they are trained on subjective human judgements as ground-truth. To address these issues we introduce a neuro-symbolic paradigm for AQA which uses neural networks to abstract interpretable symbols from video data and makes quality assessments by applying rules to those symbols. We take diving as the case study. We found that domain experts prefer our system and find it more informative than purely neural approaches to AQA in diving. Our system also achieves state-of-the-art action recognition and temporal segmentation and automatically generates a detailed report that breaks the dive down into its elements and provides objective scoring with visual evidence. As verified by a group of domain experts this report may be used to assist judges in scoring help train judges and provide feedback to divers. Annotated training data and code: https://github.com/laurenok24/NSAQA.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Okamoto_2024_CVPR, author = {Okamoto, Lauren and Parmar, Paritosh}, title = {Hierarchical NeuroSymbolic Approach for Comprehensive and Explainable Action Quality Assessment}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2024}, pages = {3204-3213} }