Diabetes60 - Inferring Bread Units From Food Images Using Fully Convolutional Neural Networks

Patrick Ferdinand Christ, Sebastian Schlecht, Florian Ettlinger, Felix Grun, Christoph Heinle, Sunil Tatavatry, Seyed-Ahmad Ahmadi, Klaus Diepold, Bjoern H. Menze; Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017, pp. 1526-1535

Abstract


In this paper we propose a challenging new computer vision task of inferring Bread Units (BUs) from food images. Assessing nutritional information and nutrient volume from a meal is an important task for diabetes patients. At the moment, diabetes patients learn the assessment of BUs on a scale of one to ten, by learning correspondence of BU and meals from textbooks. We introduce a large scale data set of around 9k different RGB-D images of 60 western dishes acquired using a Microsoft Kinect v2 sensor. We recruited 20 diabetes patients to give expert assessments of BU values to each dish based on several images. For this task, we set a challenging baseline using state-of-the-art CNNs and evaluated it against the performance of human annotators. In our work we present a CNN architecture to infer the depth from RGB-only food images to be used in BU regression such that the pipeline can operate on RGB data only and compare its performance to RGB-D input data.

Related Material


[pdf]
[bibtex]
@InProceedings{Christ_2017_ICCV,
author = {Ferdinand Christ, Patrick and Schlecht, Sebastian and Ettlinger, Florian and Grun, Felix and Heinle, Christoph and Tatavatry, Sunil and Ahmadi, Seyed-Ahmad and Diepold, Klaus and Menze, Bjoern H.},
title = {Diabetes60 - Inferring Bread Units From Food Images Using Fully Convolutional Neural Networks},
booktitle = {Proceedings of the IEEE International Conference on Computer Vision (ICCV) Workshops},
month = {Oct},
year = {2017}
}