GAN-based Detailed Clothing Generation System
DOI:
https://doi.org/10.52731/liir.v003.067Abstract
The implementation of Generative Adversarial Networks (GAN) in the fashion domain has been researched for various applications such as virtual-try-on, fashion item recommendation, and design generation. In this paper, we propose a GAN-based fashion design generation system that reduces the workload of the labor-intensive design creation task. Our system consists of two generative models: one that produces images of fashion items without any clothing patterns using conditioned StyleGAN2-ADA, and one that is a style transfer model reflecting the fine texture of the fashion item. The system also allows users to edit images of garments by manipulating the latent code of the generator. We demonstrate through qualitative and quantitative experiments that the proposed system trained on a dataset of real clothing inventory images can generate realistic and diverse images that reflect the input conditions in detail.
References
Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil
Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Z. Ghahramani,
M. Welling, C. Cortes, N. Lawrence, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc., 2014.
Tero Karras, Miika Aittala, Janne Hellsten, Samuli Laine, Jaakko Lehtinen, and Timo Aila.
Training generative adversarial networks with limited data. In H. Larochelle, M. Ranzato, R.
Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 12104–12114. Curran Associates, Inc., 2020.
Taesung Park, Jun-Yan Zhu, Oliver Wang, Jingwan Lu, Eli Shechtman, Alexei Efros, and
Richard Zhang. Swapping autoencoder for deep image manipulation. In H. Larochelle, M.
Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 7198–7211. Curran Associates, Inc., 2020.
Yujun Shen and Bolei Zhou. Closed-form factorization of latent semantics in gans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
(CVPR), pages 1532–1540, June 2021.
Xintong Han, Zuxuan Wu, Zhe Wu, Ruichi Yu, and Larry S. Davis. VITON: an image-based
virtual try-on network. CoRR, abs/1711.08447, 2017.
Wang-Cheng Kang, Chen Fang, Zhaowen Wang, and Julian J. McAuley. Visuallyaware
fashion recommendation and design with generative image models. CoRR, abs/1711.02231,
Sudhir Kumar and Mithun Das Gupta. c+gan: Complementary fashion item recommendation. CoRR, abs/1906.05596, 2019.
Lele Chen, Justin Tian, Guo Li, Cheng-Haw Wu, Erh-Kan King, Kuan-Ting Chen,
Shao-Hang Hsieh, and Chenliang Xu. Tailorgan: Making user-defined fashion designs.
CoRR, abs/2001.06427, 2020.
Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision
and Pattern Recognition (CVPR), June 2019.
Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of GANs
for improved quality, stability, and variation. In International Conference on Learning Representations, 2018.
Xun Huang and Serge Belongie. Arbitrary style transfer in real-time with adaptive instance
normalization. In Proceedings of the IEEE International Conference on Computer Vision
(ICCV), Oct 2017.
Kathleen M. Lewis, Srivatsan Varadharajan, and Ira Kemelmacher-Shlizerman. VOGUE:
try-on by stylegan interpolation optimization. CoRR, abs/2101.02285, 2021.
Gokhan Yildirim, Nikolay Jetchev, Roland Vollgraf, and Urs Bergmann. Generating
high-resolution fashion model images wearing custom outfits. CoRR, abs/1908.08847, 2019.
Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila.
Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
Mehdi Mirza and Simon Osindero. Conditional generative adversarial nets. CoRR,
abs/1411.1784, 2014.
Wenqi Xian, Patsorn Sangkloy, Varun Agrawal, Amit Raj, Jingwan Lu, Chen Fang, Fisher
Yu, and James Hays. Texturegan: Controlling deep image synthesis with texture patches. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR),
June 2018.
Edo Collins, Raja Bala, Bob Price, and Sabine Susstrunk. Editing in style: Uncovering the
local semantics of gans. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
Kurt Hornik, Ingo Feinerer, Martin Kober, and Christian Buchta. Spherical k-means clustering. Journal of Statistical Software, 50(10):1–22, 2012.
Erik Hark ¨ onen, Aaron Hertzmann, Jaakko Lehtinen, and Sylvain Paris. Ganspace: Discovering interpretable GAN controls. CoRR, abs/2004.02546, 2020.
Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp
Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium.
In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R.
Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran
Associates, Inc., 2017.
Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. InYoshua
Bengio and Yann LeCun, editors, 3rd International Conference on Learning Representations,
ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.