Monte Carlo Gradient Quantization

Goncalo Mordido, Matthijs Van Keirsbilck, Alexander Keller; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2020, pp. 718-719

Abstract


We propose Monte Carlo methods to leverage both sparsity and quantization to compress gradients of neural networks throughout training. On top of reducing the communication exchanged between multiple workers in a distributed setting, we also improve the computational efficiency of each worker. Our method, called Monte Carlo Gradient Quantization (MCGQ), shows faster convergence and higher performance than existing quantization methods on image classification and language modeling. Using both low-bit-width-quantization and high sparsity levels, our method more than doubles the rates of existing compression methods from 200xto 520xand 462xto more than 1200xon different language modeling tasks.

Related Material


[pdf]
[bibtex]
@InProceedings{Mordido_2020_CVPR_Workshops,
author = {Mordido, Goncalo and Van Keirsbilck, Matthijs and Keller, Alexander},
title = {Monte Carlo Gradient Quantization},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2020}
}