Loading [MathJax]/extensions/MathMenu.js
Texture Brush for Fashion Inspiration Transfer: A Generative Adversarial Network With Heatmap-Guided Semantic Disentanglement | IEEE Journals & Magazine | IEEE Xplore

Texture Brush for Fashion Inspiration Transfer: A Generative Adversarial Network With Heatmap-Guided Semantic Disentanglement


Abstract:

Automatically accomplishing intelligent fashion design with certain ‘inspiration’ images can greatly facilitate a designer’s design process, as well as allow users to int...Show More

Abstract:

Automatically accomplishing intelligent fashion design with certain ‘inspiration’ images can greatly facilitate a designer’s design process, as well as allow users to interactively participate in the process. In this research, we propose a generative adversarial network with heatmap-guided semantic disentanglement (HSD-GAN) to perform an ‘intelligent’ design with ‘inspiration’ transfer. Our model aims to learn how to integrate the feature representations, from the styles of both source fashion items and target fashion items, in an unsupervised manner. Specifically, a semantic disentanglement attention-based encoder is proposed to capture the most discriminative regions of different input fashion items and disentangle the features into two key factors: attribute and texture. A generator is then developed to synthesize mixed-style fashion items by utilizing the two factors. In addition, a heatmap-based patch loss is introduced to evaluate the visual-semantic matching degree between the texture of the generated fashion items and the input texture information. Extensive experimental results show that our proposed HSD-GAN consistently achieves superior performance, compared to other state-of-the-art methods.
Page(s): 2381 - 2395
Date of Publication: 23 November 2022

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

For a fashion designer, sources of inspiration can be anything visual, from the buildings of Ancient Rome to a plate of baked beans. Experienced designers seek intricate details and textures from these visual objects that can then be altered, collated, and brought up-to-date to sit within design collections. Raster or vector software, e.g., Photoshop or Adobe Illustrator, are commonly utilized to assist designers in accomplishing their working process. However, these traditional tools can only produce astonishing and realistic imagery if guided by experienced designers; they are unable to automatically create certain ‘inspiration’ fashion designs from visual objects. Therefore, it is highly desirable to devise an automatic fashion design method to assist fashion designers in completing the ‘inspiration’ design process. Fortunately, recent advances in big data analysis and deep learning techniques have provided powerful tools for making designs with this ‘inspiration’ transfer feasible.

Select All
1.
J. Cheng, F. Wu, Y. Tian, L. Wang and D. Tao, "RiFeGAN2: Rich feature generation for text-to-image synthesis from constrained prior knowledge", IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 8, pp. 5187-5200, Aug. 2022.
2.
P. Zhang, L. Yang, X. Xie and J. Lai, "Lightweight texture correlation network for pose guided person image generation", IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 7, pp. 4584-4598, Jul. 2022.
3.
W. Xian et al., "TextureGAN: Controlling deep image synthesis with texture patches", Proc. IEEE CVPR, pp. 8456-8465, Jun. 2018.
4.
Y. R. Cui et al., "FashionGAN: Display your fashion design using conditional generative adversarial nets" in Computer Graphics Forum, Hoboken, NJ, USA:Wiley, vol. 37, no. 7, pp. 109-119, 2018.
5.
S. Zhu, S. Fidler, R. Urtasun, D. Lin and C. C. Loy, "Be your own Prada: Fashion synthesis with structural coherence", Proc. IEEE Int. Conf. Comput. Vis. (ICCV), pp. 1680-1688, Oct. 2017.
6.
Y. Choi, Y. Uh, J. Yoo and J.-W. Ha, "StarGAN v2: Diverse image synthesis for multiple domains", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 8188-8197, Jun. 2020.
7.
X. Huang, M.-Y. Liu, S. Belongie and J. Kautz, "Multimodal unsupervised image-to-image translation", Proc. Eur. Conf. Comput. Vis., pp. 172-189, 2018.
8.
W. Cho, S. Choi, D. K. Park, I. Shin and J. Choo, "Image-to-image translation via group-wise deep whitening-and-coloring transformation", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 10639-10647, Jun. 2019.
9.
Y. Xia, W. Zheng, Y. Wang, H. Yu, J. Dong and F.-Y. Wang, "Local and global perception generative adversarial network for facial expression synthesis", IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 3, pp. 1443-1452, Mar. 2022.
10.
X. Huang and S. Belongie, "Arbitrary style transfer in real-time with adaptive instance normalization", Proc. IEEE Int. Conf. Comput. Vis. (ICCV), pp. 1501-1510, Oct. 2017.
11.
S. Yang et al., "Awesome typography: Statistics-based text effects transfer", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 2886-2895, Jun. 2017.
12.
S. Yang, J. Liu, W. Yang and Z. Guo, "Context-aware text-based binary image stylization and synthesis", IEEE Trans. Image Process., vol. 28, no. 2, pp. 952-964, Feb. 2019.
13.
Y. Deng, F. Tang, W. Dong, W. Sun, F. Huang and C. Xu, "Arbitrary style transfer via multi-adaptation network", Proc. 28th ACM Int. Conf. Multimedia, pp. 2719-2727, Oct. 2020.
14.
Y. Zhang et al., "Domain enhanced arbitrary image style transfer via contrastive learning", Proc. ACM SIGGRAPH, pp. 1-8, 2022.
15.
L. A. Gatys, A. S. Ecker and M. Bethge, "Image style transfer using convolutional neural networks", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 2414-2423, Jun. 2016.
16.
Y. Li et al., "Universal style transfer via feature transforms", Proc. Adv. Neural Inf. Process. Syst., pp. 386-396, 2017.
17.
J. Johnson, A. Alahi and L. Fei-Fei, "Perceptual losses for real-time style transfer and super-resolution", Proc. ECCV, pp. 694-711, 2016.
18.
C. Li and M. Wand, "Precomputed real-time texture synthesis with Markovian generative adversarial networks", Proc. ECCV, pp. 702-716, 2016.
19.
Y. Gao et al., "Wallpaper texture generation and style transfer based on multi-label semantics", IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 3, pp. 1552-1563, Mar. 2022.
20.
S. Xie, H. Hu and Y. Chen, "Facial expression recognition with two-branch disentangled generative adversarial network", IEEE Trans. Circuits Syst. Video Technol., vol. 31, no. 6, pp. 2359-2371, Jun. 2021.
21.
H. Yan et al., "Toward intelligent design: An AI-based fashion designer using generative adversarial networks aided by sketch and rendering generators" in IEEE Trans. Multimedia, Jan. 2022.
22.
P. Isola, J.-Y. Zhu, T. Zhou and A. A. Efros, "Image-to-image translation with conditional adversarial networks", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 1125-1134, Jul. 2017.
23.
S. Jiang, J. Li and Y. Fu, "Deep learning for fashion style generation", IEEE Trans. Neural Netw. Learn. Syst., vol. 33, no. 9, pp. 4538-4550, Sep. 2022.
24.
Z. Zhang et al., "M6-UFC: Unifying multi-modal controls for conditional image synthesis via non-autoregressive generative transformers" in arXiv:2105.14211, 2021.
25.
S. Zhang, Z. Song, X. Cao, H. Zhang and J. Zhou, "Task-aware attention model for clothing attribute prediction", IEEE Trans. Circuits Syst. Video Technol., vol. 30, no. 4, pp. 1051-1064, Apr. 2020.
26.
B. Zhou, A. Khosla, A. Lapedriza, A. Oliva and A. Torralba, "Learning deep features for discriminative localization", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 2921-2929, Jun. 2016.
27.
K. He, X. Zhang, S. Ren and J. Sun, "Deep residual learning for image recognition", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 770-778, Jun. 2016.
28.
K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition" in arXiv:1409.1556, 2015.
29.
T. Karras, S. Laine, M. Aittala, J. Hellsten, J. Lehtinen and T. Aila, "Analyzing and improving the image quality of StyleGAN", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 8110-8119, Jun. 2020.
30.
R. Zhang, P. Isola, A. A. Efros, E. Shechtman and O. Wang, "The unreasonable effectiveness of deep features as a perceptual metric", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 586-595, Jun. 2018.
Contact IEEE to Subscribe

References

References is not available for this document.