Inertial-Vision: Cross-Domain Knowledge Transfer for Wearable Sensors

Girmaw Abebe, Andrea Cavallaro; Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017, pp. 1392-1400

Abstract


Multi-modal ego-centric data from inertial measurement units (IMU) and first-person videos (FPV) can be effectively fused to recognise proprioceptive activities. Existing IMU-based approaches mostly employ cascades of handcrafted triaxial motion features or deep frameworks trained on limited data. FPV approaches generally encode scene dynamics with motion and pooled appearance features. In this paper, we propose a multi-modal ego-centric proprioceptive activity recognition that uses a convolutional neural network (CNN) followed by a long short-term memory (LSTM) network, transfer learning and a merit-based fusion of IMU and/or FPV streams. The CNN encodes short-term temporal dynamics of the ego-motion and the LSTM exploits the long-term temporal dependency among activities. The merit of a stream is evaluated with a sparsity measure of its initial classification output. We validate the proposed framework on multiple visual and inertial datasets.

Related Material


[pdf]
[bibtex]
@InProceedings{Abebe_2017_ICCV,
author = {Abebe, Girmaw and Cavallaro, Andrea},
title = {Inertial-Vision: Cross-Domain Knowledge Transfer for Wearable Sensors},
booktitle = {Proceedings of the IEEE International Conference on Computer Vision (ICCV) Workshops},
month = {Oct},
year = {2017}
}