TY - JOUR
T1 - 3FabRec
T2 - 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020
AU - Browatzki, Bjorn
AU - Wallraven, Christian
N1 - Funding Information:
Acknowledgements This work was supported by Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korean government (MSIT) (No. 2019-0-00079, Department of Artificial Intelligence, Korea University)
Funding Information:
This work was supported by Institute of Information &Communications Technology Planning &Evaluation (IITP) grant funded by the Korean government (MSIT) (No. 2019-0-00079, Department of Artificial Intelligence, Korea University)
Publisher Copyright:
© 2020 IEEE.
PY - 2020
Y1 - 2020
N2 - Current supervised methods for facial landmark detection require a large amount of training data and may suffer from overfitting to specific datasets due to the massive number of parameters. We introduce a semi-supervised method in which the crucial idea is to first generate implicit face knowledge from the large amounts of unlabeled images of faces available today. In a first, completely unsupervised stage, we train an adversarial autoencoder to reconstruct faces via a low-dimensional face embedding. In a second, supervised stage, we interleave the decoder with transfer layers to retask the generation of color images to the prediction of landmark heatmaps. Our framework (3FabRec) achieves state-of-the-art performance on several common benchmarks and, most importantly, is able to maintain impressive accuracy on extremely small training sets down to as few as 10 images. As the interleaved layers only add a low amount of parameters to the decoder, inference runs at several hundred FPS on a GPU.
AB - Current supervised methods for facial landmark detection require a large amount of training data and may suffer from overfitting to specific datasets due to the massive number of parameters. We introduce a semi-supervised method in which the crucial idea is to first generate implicit face knowledge from the large amounts of unlabeled images of faces available today. In a first, completely unsupervised stage, we train an adversarial autoencoder to reconstruct faces via a low-dimensional face embedding. In a second, supervised stage, we interleave the decoder with transfer layers to retask the generation of color images to the prediction of landmark heatmaps. Our framework (3FabRec) achieves state-of-the-art performance on several common benchmarks and, most importantly, is able to maintain impressive accuracy on extremely small training sets down to as few as 10 images. As the interleaved layers only add a low amount of parameters to the decoder, inference runs at several hundred FPS on a GPU.
UR - http://www.scopus.com/inward/record.url?scp=85094650827&partnerID=8YFLogxK
U2 - 10.1109/CVPR42600.2020.00615
DO - 10.1109/CVPR42600.2020.00615
M3 - Conference article
AN - SCOPUS:85094650827
SN - 1063-6919
SP - 6109
EP - 6119
JO - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
JF - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
M1 - 9157661
Y2 - 14 June 2020 through 19 June 2020
ER -