Adaptive Posit: Parameter Aware Numerical Format for Deep Learning Inference on the Edge

Hamed F. Langroudi, Vedant Karia, John L. Gustafson, Dhireesha Kudithipudi; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2020, pp. 726-727

Abstract


Ultra low-precision (<8-bit width) arithmetic is a discernible approach to deploy deep learning networks on to edge devices. Recent findings show that posit with linear quantization has a similar dynamic range as the weight and activation values across the deep neural network layers. This characteristic can benefit the data representation of deep neural networks without impacting the overall accuracy. When capturing the full dynamic range of weights and activations, posit with mixed precision or linear quantization leads to a surge in hardware resource requirements. We propose adaptive posit, which has the ability to capture the non-homogeneous dynamic range of weights and activations across the deep neural network layers. A fine granular control is achieved by embedding the hyperparameters in the numerical format. To evaluate the overall system efficiency, we design a parameterized ASIC softcore for the adaptive posit encoder and decoder. Benchmarking and evaluation of the adaptive posit are performed on three datasets: Fashion-MNIST, CIFAR-10, and ImageNet. Results assert that on average the performance on inference with<8-bitadaptive posits surpasses (2% to 10%) that of posit.

Related Material


[pdf]
[bibtex]
@InProceedings{Langroudi_2020_CVPR_Workshops,
author = {Langroudi, Hamed F. and Karia, Vedant and Gustafson, John L. and Kudithipudi, Dhireesha},
title = {Adaptive Posit: Parameter Aware Numerical Format for Deep Learning Inference on the Edge},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2020}
}