LatentPaint: Image Inpainting in Latent Space With Diffusion Models

Ciprian Corneanu, Raghudeep Gadde, Aleix M. Martinez; Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2024, pp. 4334-4343

Abstract


Image inpainting is generally done using either a domain-specific (preconditioned) model or a generic model that is postconditioned at inference time. Preconditioned models are fast at inference time but extremely costly to train, requiring training on each domain they are applied to. Postconditioned models do not require any domain-specific training but are slow during inference, requiring multiple forward and backward passes to converge to a desirable solution. Here, we derive an approach that does not require any domain specific training, yet is fast at inference time. To solve the costly inference computational time, we perform the forward-backward fusion step on a latent space rather than the image space. This is solved with a newly proposed propagation module in the diffusion process. Experiments on a number of domains demonstrate our approach attains or improves state-of-the-art results with the advantages of preconditioned and postconditioned models and none of their disadvantages.

Related Material


[pdf]
[bibtex]
@InProceedings{Corneanu_2024_WACV, author = {Corneanu, Ciprian and Gadde, Raghudeep and Martinez, Aleix M.}, title = {LatentPaint: Image Inpainting in Latent Space With Diffusion Models}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2024}, pages = {4334-4343} }