Real-Time Sign Language Recognition Using a Consumer Depth Camera

Alina Kuznetsova, Laura Leal-Taixe, Bodo Rosenhahn; Proceedings of the IEEE International Conference on Computer Vision (ICCV) Workshops, 2013, pp. 83-90


Gesture recognition remains a very challenging task in the field of computer vision and human computer interaction (HCI). A decade ago the task seemed to be almost unsolvable with the data provided by a single RGB camera. Due to recent advances in sensing technologies, such as time-of-flight and structured light cameras, there are new data sources available, which make hand gesture recognition more feasible. In this work, we propose a highly precise method to recognize static gestures from a depth data, provided from one of the above mentioned devices. The depth images are used to derive rotation-, translationand scaleinvariant features. A multi-layered random forest (MLRF) is then trained to classify the feature vectors, which yields to the recognition of the hand signs. The training time and memory required by MLRF are much smaller, compared to a simple random forest with equivalent precision. This allows to repeat the training procedure of MLRF without significant effort. To show the advantages of our technique, we evaluate our algorithm on synthetic data, on publicly available dataset, containing 24 signs from American Sign Language(ASL) and on a new dataset, collected using recently appeared Intel Creative Gesture Camera.

Related Material

author = {Alina Kuznetsova and Laura Leal-Taixe and Bodo Rosenhahn},
title = {Real-Time Sign Language Recognition Using a Consumer Depth Camera},
booktitle = {Proceedings of the IEEE International Conference on Computer Vision (ICCV) Workshops},
month = {June},
year = {2013}