-
[pdf]
[supp]
[arXiv]
[bibtex]@InProceedings{Liu_2025_ICCV, author = {Liu, Shengqi and Cheng, Yuhao and Chen, Zhuo and Ren, Xingyu and Zhu, Wenhan and Li, Lincheng and Bi, Mengxiao and Yang, Xiaokang and Yan, Yichao}, title = {Multimodal Latent Diffusion Model for Complex Sewing Pattern Generation}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2025}, pages = {17640-17650} }
Multimodal Latent Diffusion Model for Complex Sewing Pattern Generation
Abstract
Generating sewing patterns in garment design is receiving increasing attention due to its CG-friendly and flexible-editing nature. Previous sewing pattern generation methods have been able to produce exquisite clothing, but struggle to design complex garments with detailed control. To address these issues, we propose **SewingLDM**, a multi-modal generative model that generates sewing patterns controlled by text prompts, body shapes, and garment sketches. Initially, we extend the original vector of sewing patterns into a more comprehensive representation to cover more intricate details and then compress them into a compact latent space. To learn the sewing pattern distribution in the latent space, we design a two-step training strategy to inject the multi-modal conditions, i.e., body shapes, text prompts, and garment sketches, into a diffusion model, ensuring the generated garments are body-suited and detail-controlled. Comprehensive qualitative and quantitative experiments show the effectiveness of our proposed method, significantly surpassing previous approaches in terms of complex garment design and various body adaptability.
Related Material
