Learning Deep Latent Variable Models by Short-Run MCMC Inference With Optimal Transport Correction

Dongsheng An, Jianwen Xie, Ping Li; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 15415-15424

Abstract


Learning latent variable models with deep top-down architectures typically requires inferring the latent variables for each training example based on the posterior distribution of these latent variables. The inference step typically relies on either time-consuming long run Markov chain Monte Caro (MCMC) or a separate inference model for variational learning. In this paper, we propose to use short run MCMC, such as Langevin dynamics, as an approximate inference engine, where the bias existing in the output distribution of the short run Langevin dynamics is corrected by optimal transport, which aims at minimizing the Wasserstein distance between the biased distribution produced by the finite step Langevin dynamics and the prior distribution. Our experiments show that the proposed strategy outperforms the variational auto-encoder (VAE) and alternating back-propagation algorithm (ABP) in terms of reconstruction error and synthesis quality.

Related Material


[pdf] [supp]
[bibtex]
@InProceedings{An_2021_CVPR, author = {An, Dongsheng and Xie, Jianwen and Li, Ping}, title = {Learning Deep Latent Variable Models by Short-Run MCMC Inference With Optimal Transport Correction}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)}, month = {June}, year = {2021}, pages = {15415-15424} }