TY - JOUR
T1 - A U-Net Architecture for Inpainting Lightstage Normal Maps
AU - Zuo, Hancheng
AU - Tiddeman, Bernard
N1 - Publisher Copyright:
© 2024 by the authors.
PY - 2024/2/19
Y1 - 2024/2/19
N2 - In this paper, we investigate the
inpainting of normal maps that were captured from a lightstage.
Occlusion of parts of the face during performance capture can be caused
by the movement of, e.g., arms, hair, or props. Inpainting is the
process of interpolating missing areas of an image with plausible data.
We build on previous works about general image inpainting that use
generative adversarial networks (GANs). We extend our previous work on
normal map inpainting to use a U-Net structured generator network. Our
method takes into account the nature of the normal map data and so
requires modification of the loss function. We use a cosine loss rather
than the more common mean squared error loss when training the
generator. Due to the small amount of training data available, even when
using synthetic datasets, we require significant augmentation, which
also needs to take account of the particular nature of the input data.
Image flipping and inplane rotations need to properly flip and rotate
the normal vectors. During training, we monitor key performance metrics
including the average loss, structural similarity index measure (SSIM),
and peak signal-to-noise ratio (PSNR) of the generator, alongside the
average loss and accuracy of the discriminator. Our analysis reveals
that the proposed model generates high-quality, realistic inpainted
normal maps, demonstrating the potential for application to performance
capture. The results of this investigation provide a baseline on which
future researchers can build with more advanced networks and comparison
with inpainting of the source images used to generate the normal maps.
AB - In this paper, we investigate the
inpainting of normal maps that were captured from a lightstage.
Occlusion of parts of the face during performance capture can be caused
by the movement of, e.g., arms, hair, or props. Inpainting is the
process of interpolating missing areas of an image with plausible data.
We build on previous works about general image inpainting that use
generative adversarial networks (GANs). We extend our previous work on
normal map inpainting to use a U-Net structured generator network. Our
method takes into account the nature of the normal map data and so
requires modification of the loss function. We use a cosine loss rather
than the more common mean squared error loss when training the
generator. Due to the small amount of training data available, even when
using synthetic datasets, we require significant augmentation, which
also needs to take account of the particular nature of the input data.
Image flipping and inplane rotations need to properly flip and rotate
the normal vectors. During training, we monitor key performance metrics
including the average loss, structural similarity index measure (SSIM),
and peak signal-to-noise ratio (PSNR) of the generator, alongside the
average loss and accuracy of the discriminator. Our analysis reveals
that the proposed model generates high-quality, realistic inpainted
normal maps, demonstrating the potential for application to performance
capture. The results of this investigation provide a baseline on which
future researchers can build with more advanced networks and comparison
with inpainting of the source images used to generate the normal maps.
KW - Computer Networks and Communications
KW - Human-Computer Interaction
KW - inpainting
KW - lightstage
KW - U-Net
UR - http://www.scopus.com/inward/record.url?scp=85187250715&partnerID=8YFLogxK
U2 - 10.3390/computers13020056
DO - 10.3390/computers13020056
M3 - Article
SN - 2073-431X
VL - 13
JO - Computers
JF - Computers
IS - 2
M1 - 56
ER -