Latent Multitask Learning for View-Invariant Action Recognition

Behrooz Mahasseni, Sinisa Todorovic; The IEEE International Conference on Computer Vision (ICCV), 2013, pp. 3128-3135


This paper presents an approach to view-invariant action recognition, where human poses and motions exhibit large variations across different camera viewpoints. When each viewpoint of a given set of action classes is specified as a learning task then multitask learning appears suitable for achieving view invariance in recognition. We extend the standard multitask learning to allow identifying: (1) latent groupings of action views (i.e., tasks), and (2) discriminative action parts, along with joint learning of all tasks. This is because it seems reasonable to expect that certain distinct views are more correlated than some others, and thus identifying correlated views could improve recognition. Also, part-based modeling is expected to improve robustness against self-occlusion when actors are imaged from different views. Results on the benchmark datasets show that we outperform standard multitask learning by 21.9%, and the state-of-the-art alternatives by 4.5-6%.

Related Material

author = {Mahasseni, Behrooz and Todorovic, Sinisa},
title = {Latent Multitask Learning for View-Invariant Action Recognition},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
month = {December},
year = {2013}