-
[pdf]
[supp]
[bibtex]@InProceedings{Gim_2024_ACCV, author = {Gim, Jongmin and Park, Jihun and Lee, Kyoungmin and Im, Sunghoon}, title = {Content-Adaptive Style Transfer: A Training-Free Approach with VQ Autoencoders}, booktitle = {Proceedings of the Asian Conference on Computer Vision (ACCV)}, month = {December}, year = {2024}, pages = {2337-2353} }
Content-Adaptive Style Transfer: A Training-Free Approach with VQ Autoencoders
Abstract
We introduce Content-Adaptive Style Transfer (CAST), a novel training-free approach for arbitrary style transfer that enhances visual fidelity using vector quantized-based pretrained autoencoder. Our method systematically applies coherent stylization to corresponding content regions. It starts by capturing the global structure of images through vector quantization, then refines local details using our style-injected decoder. CAST consists of three main components: a content-consistent style injection module, which tailors stylization to unique image regions; an adaptive style refinement module, which fine-tunes stylization intensity; and a content refinement module, which ensures content integrity through interpolation and feature distribution maintenance. Experimental results indicate that CAST outperforms existing generative-based and traditional style transfer models in both quantitative and qualitative measures.
Related Material