TY - JOUR
T1 - DeepPortraitDrawing
T2 - Generating human body images from freehand sketches
AU - Wu, Xian
AU - Wang, Chen
AU - Fu, Hongbo
AU - Shamir, Ariel
AU - Zhang, Song Hai
N1 - Publisher Copyright:
© 2023 Elsevier Ltd
PY - 2023/11
Y1 - 2023/11
N2 - Various methods for generating realistic images of objects and human faces from freehand sketches have been explored. However, generating realistic human body images from sketches is still a challenging problem. It is, first because of the sensitivity to human shapes, second because of the complexity of human images caused by body shape and pose changes, and third because of the domain gap between realistic images and freehand sketches. In this work, we present DeepPortraitDrawing, a deep generative framework for converting roughly drawn sketches to realistic human body images. To encode complicated body shapes under various poses, we take a local-to-global approach. Locally, we employ semantic part auto-encoders to construct part-level shape spaces, which are useful for refining the geometry of an input pre-segmented hand-drawn sketch. Globally, we employ a cascaded spatial transformer network to refine the structure of body parts by adjusting their spatial locations and relative proportions. Finally, we use a style-based generator as the global synthesis network for the sketch-to-image translation task which is modulated by segmentation maps for semantic preservation. Extensive experiments have shown that given roughly sketched human portraits, our method produces more realistic images than the state-of-the-art sketch-to-image synthesis techniques.
AB - Various methods for generating realistic images of objects and human faces from freehand sketches have been explored. However, generating realistic human body images from sketches is still a challenging problem. It is, first because of the sensitivity to human shapes, second because of the complexity of human images caused by body shape and pose changes, and third because of the domain gap between realistic images and freehand sketches. In this work, we present DeepPortraitDrawing, a deep generative framework for converting roughly drawn sketches to realistic human body images. To encode complicated body shapes under various poses, we take a local-to-global approach. Locally, we employ semantic part auto-encoders to construct part-level shape spaces, which are useful for refining the geometry of an input pre-segmented hand-drawn sketch. Globally, we employ a cascaded spatial transformer network to refine the structure of body parts by adjusting their spatial locations and relative proportions. Finally, we use a style-based generator as the global synthesis network for the sketch-to-image translation task which is modulated by segmentation maps for semantic preservation. Extensive experiments have shown that given roughly sketched human portraits, our method produces more realistic images than the state-of-the-art sketch-to-image synthesis techniques.
KW - Generative adversarial networks
KW - Image-to-image generation
KW - StyleGAN
UR - https://www.webofscience.com/wos/woscc/full-record/WOS:001061666600001
UR - https://openalex.org/W4385652581
UR - https://www.scopus.com/pages/publications/85168560868
U2 - 10.1016/j.cag.2023.08.005
DO - 10.1016/j.cag.2023.08.005
M3 - Journal Article
SN - 0097-8493
VL - 116
SP - 73
EP - 81
JO - Computers and Graphics (Pergamon)
JF - Computers and Graphics (Pergamon)
ER -