TY - JOUR
T1 - VD-GAN
T2 - A Unified Framework for Joint Prototype and Representation Learning from Contaminated Single Sample per Person
AU - Pang, Meng
AU - Wang, Binghui
AU - Cheung, Yiu Ming
AU - Chen, Yiran
AU - Wen, Bihan
N1 - Funding Information:
Manuscript received August 6, 2020; revised November 17, 2020 and December 23, 2020; accepted December 28, 2020. Date of publication January 8, 2021; date of current version February 9, 2021. The work of Yiu-ming Cheung was supported in part by the NSFC under Grant 61672444, in part by the Hong Kong Baptist University (HKBU) under Grant RC-FNRA-IG/18-19/SCI/03 and Grant RC-IRCMs/18-19/SCI/01, and in part by the Innovation and Technology Fund (ITF) of the Innovation and Technology Commission (ITC) of the Government of the Hong Kong, SAR, under Project ITS/339/18. The work of Bihan Wen was supported in part by the Ministry of Education, Republic of Singapore, under the Start-Up Grant; and in part by the National Research Foundation (NRF), Singapore, through the Singapore Cybersecurity Consortium (SGCSC) Grant Office, under Grant SGCSC_Grant_2019-S01. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Domingo Mery. (Corresponding author: Bihan Wen.) Meng Pang and Bihan Wen are with the School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore 639798 (e-mail: [email protected]; [email protected]).
PY - 2021/1/8
Y1 - 2021/1/8
N2 - Single sample per person (SSPP) face recognition with a contaminated
biometric enrolment database (SSPP-ce FR) is an emerging practical FR
problem, where the SSPP in the enrolment database is no longer standard
but contaminated by nuisance facial variations such as expression,
lighting, pose, and disguise. In this case, the conventional SSPP FR
methods, including the patch-based and generic learning methods, will
suffer from serious performance degradation. Few recent methods were
proposed to tackle SSPP-ce FR by either performing prototype learning on
the contaminated enrolment database or learning discriminative
representations that are robust against variation. Despite that, most of
these approaches can only handle a specified single variation, e.g.,
pose, but cannot be extended to multiple variations. To address these
two limitations, we propose a novel Variation Disentangling Generative
Adversarial Network (VDGAN) to jointly perform prototype learning and
representation learning in a unified framework. The proposed VD-GAN
consists of an encoder-decoder structural generator and a multi-task
discriminator to handle universal variations including single, multiple,
and even mixed variations in practice. The generator and discriminator
play an adversarial game such that the generator learns a discriminative
identity representation and generates an identity-preserved prototype
for each face image, while the discriminator aims to predict face
identity label, distinguish real vs. fake prototype, and disentangle
target variations from the learned representations. Qualitative and
quantitative evaluations on various real-world face datasets containing
single/multiple and mixed variations demonstrate the effectiveness of
VD-GAN.
AB - Single sample per person (SSPP) face recognition with a contaminated
biometric enrolment database (SSPP-ce FR) is an emerging practical FR
problem, where the SSPP in the enrolment database is no longer standard
but contaminated by nuisance facial variations such as expression,
lighting, pose, and disguise. In this case, the conventional SSPP FR
methods, including the patch-based and generic learning methods, will
suffer from serious performance degradation. Few recent methods were
proposed to tackle SSPP-ce FR by either performing prototype learning on
the contaminated enrolment database or learning discriminative
representations that are robust against variation. Despite that, most of
these approaches can only handle a specified single variation, e.g.,
pose, but cannot be extended to multiple variations. To address these
two limitations, we propose a novel Variation Disentangling Generative
Adversarial Network (VDGAN) to jointly perform prototype learning and
representation learning in a unified framework. The proposed VD-GAN
consists of an encoder-decoder structural generator and a multi-task
discriminator to handle universal variations including single, multiple,
and even mixed variations in practice. The generator and discriminator
play an adversarial game such that the generator learns a discriminative
identity representation and generates an identity-preserved prototype
for each face image, while the discriminator aims to predict face
identity label, distinguish real vs. fake prototype, and disentangle
target variations from the learned representations. Qualitative and
quantitative evaluations on various real-world face datasets containing
single/multiple and mixed variations demonstrate the effectiveness of
VD-GAN.
KW - generative adversarial network
KW - prototype learning
KW - representation learning
KW - Single sample per person
UR - http://www.scopus.com/inward/record.url?scp=85099589039&partnerID=8YFLogxK
U2 - 10.1109/TIFS.2021.3050055
DO - 10.1109/TIFS.2021.3050055
M3 - Journal article
AN - SCOPUS:85099589039
SN - 1556-6013
VL - 16
SP - 2246
EP - 2259
JO - IEEE Transactions on Information Forensics and Security
JF - IEEE Transactions on Information Forensics and Security
ER -