Learning Time/Memory-Efficient Deep Architectures With Budgeted Super Networks

Tom Véniat, Ludovic Denoyer; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 3492-3500

Abstract


We propose to focus on the problem of discovering neural network architectures efficient in terms of both prediction quality and cost. For instance, our approach is able to solve the following tasks: learn a neural network able to predict well in less than 100 milliseconds or learn an efficient model that fits in a 50 Mb memory. Our contribution is a novel family of models called Budgeted Super Networks (BSN). They are learned using gradient descent techniques applied on a budgeted learning objective function which integrates a maximum authorized cost, while making no assumption on the nature of this cost. We present a set of experiments on computer vision problems and analyze the ability of our technique to deal with three different costs: the computation cost, the memory consumption cost and a distributed computation cost. We particularly show that our model can discover neural network architectures that have a better accuracy than the ResNet and Convolutional Neural Fabrics architectures on CIFAR-10 and CIFAR-100, at a lower cost.

Related Material


[pdf] [Supp] [arXiv]
[bibtex]
@InProceedings{Véniat_2018_CVPR,
author = {Véniat, Tom and Denoyer, Ludovic},
title = {Learning Time/Memory-Efficient Deep Architectures With Budgeted Super Networks},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2018}
}