-
[pdf]
[supp]
[arXiv]
[bibtex]@InProceedings{Ganeshan_2025_CVPR, author = {Ganeshan, Aditya and Groueix, Thibault and Guerrero, Paul and Mech, Radomir and Fisher, Matthew and Ritchie, Daniel}, title = {Pattern Analogies: Learning to Perform Programmatic Image Edits by Analogy}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {2025}, pages = {28715-28725} }
Pattern Analogies: Learning to Perform Programmatic Image Edits by Analogy
Abstract
Pattern images are everywhere in the digital and physical worlds, and tools to edit them are valuable. But editing pattern images is tricky: desired edits are often *programmatic*: structure-aware edits that alter the underlying program which generates the pattern. One could attempt to infer this underlying program, but current methods for doing so struggle with complex images and produce unorganized programs that make editing tedious. In this work, we introduce a novel approach to perform programmatic edits on pattern images. By using a *pattern analogy*—a pair of simple patterns to demonstrate the intended edit—and a learning-based generative model to execute these edits, our method allows users to intuitively edit patterns. To enable this paradigm, we introduce **SplitWeave**, a domain-specific language that, combined with a framework for sampling synthetic pattern analogies, enables the creation of a large, high-quality synthetic training dataset. We also present **TriFuser**, a Latent Diffusion Model (LDM) designed to overcome critical issues that arise when naively deploying LDMs to this task. Extensive experiments on real-world, artist-sourced patterns reveals that our method faithfully performs the demonstrated edit while also generalizing to related pattern styles beyond its training distribution.
Related Material