GAN-based Detailed Clothing Generation System

Authors

  • Ryozo Masukawa Meiji University
  • Shosuke Haji
  • Tomohiro Takagi
  • Taiga Matsui
  • Keita Ishikawa
  • Masane Fuchi
  • Kazuki Yamaji

DOI:

https://doi.org/10.52731/liir.v003.067

Abstract

The implementation of Generative Adversarial Networks (GAN) in the fashion domain has been researched for various applications such as virtual-try-on, fashion item recommendation, and design generation. In this paper, we propose a GAN-based fashion design generation system that reduces the workload of the labor-intensive design creation task. Our system consists of two generative models: one that produces images of fashion items without any clothing patterns using conditioned StyleGAN2-ADA, and one that is a style transfer model reflecting the fine texture of the fashion item. The system also allows users to edit images of garments by manipulating the latent code of the generator. We demonstrate through qualitative and quantitative experiments that the proposed system trained on a dataset of real clothing inventory images can generate realistic and diverse images that reflect the input conditions in detail.

References

Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil

Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Z. Ghahramani,

M. Welling, C. Cortes, N. Lawrence, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc., 2014.

Tero Karras, Miika Aittala, Janne Hellsten, Samuli Laine, Jaakko Lehtinen, and Timo Aila.

Training generative adversarial networks with limited data. In H. Larochelle, M. Ranzato, R.

Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 12104–12114. Curran Associates, Inc., 2020.

Taesung Park, Jun-Yan Zhu, Oliver Wang, Jingwan Lu, Eli Shechtman, Alexei Efros, and

Richard Zhang. Swapping autoencoder for deep image manipulation. In H. Larochelle, M.

Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 7198–7211. Curran Associates, Inc., 2020.

Yujun Shen and Bolei Zhou. Closed-form factorization of latent semantics in gans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition

(CVPR), pages 1532–1540, June 2021.

Xintong Han, Zuxuan Wu, Zhe Wu, Ruichi Yu, and Larry S. Davis. VITON: an image-based

virtual try-on network. CoRR, abs/1711.08447, 2017.

Wang-Cheng Kang, Chen Fang, Zhaowen Wang, and Julian J. McAuley. Visuallyaware

fashion recommendation and design with generative image models. CoRR, abs/1711.02231,

Sudhir Kumar and Mithun Das Gupta. c+gan: Complementary fashion item recommendation. CoRR, abs/1906.05596, 2019.

Lele Chen, Justin Tian, Guo Li, Cheng-Haw Wu, Erh-Kan King, Kuan-Ting Chen,

Shao-Hang Hsieh, and Chenliang Xu. Tailorgan: Making user-defined fashion designs.

CoRR, abs/2001.06427, 2020.

Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision

and Pattern Recognition (CVPR), June 2019.

Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of GANs

for improved quality, stability, and variation. In International Conference on Learning Representations, 2018.

Xun Huang and Serge Belongie. Arbitrary style transfer in real-time with adaptive instance

normalization. In Proceedings of the IEEE International Conference on Computer Vision

(ICCV), Oct 2017.

Kathleen M. Lewis, Srivatsan Varadharajan, and Ira Kemelmacher-Shlizerman. VOGUE:

try-on by stylegan interpolation optimization. CoRR, abs/2101.02285, 2021.

Gokhan Yildirim, Nikolay Jetchev, Roland Vollgraf, and Urs Bergmann. Generating

high-resolution fashion model images wearing custom outfits. CoRR, abs/1908.08847, 2019.

Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila.

Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF

Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.

Mehdi Mirza and Simon Osindero. Conditional generative adversarial nets. CoRR,

abs/1411.1784, 2014.

Wenqi Xian, Patsorn Sangkloy, Varun Agrawal, Amit Raj, Jingwan Lu, Chen Fang, Fisher

Yu, and James Hays. Texturegan: Controlling deep image synthesis with texture patches. In

Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR),

June 2018.

Edo Collins, Raja Bala, Bob Price, and Sabine Susstrunk. Editing in style: Uncovering the

local semantics of gans. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.

Kurt Hornik, Ingo Feinerer, Martin Kober, and Christian Buchta. Spherical k-means clustering. Journal of Statistical Software, 50(10):1–22, 2012.

Erik Hark ¨ onen, Aaron Hertzmann, Jaakko Lehtinen, and Sylvain Paris. Ganspace: Discovering interpretable GAN controls. CoRR, abs/2004.02546, 2020.

Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp

Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium.

In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R.

Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran

Associates, Inc., 2017.

Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. InYoshua

Bengio and Yann LeCun, editors, 3rd International Conference on Learning Representations,

ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.

Downloads

Published

2023-02-17