Face Detection, Bounding Box Aggregation and Pose Estimation for

11 downloads 0 Views 6MB Size Report
four stages including face detection using multiple face de- tectors, face bounding box aggregation, pose estimation and facial landmark localisation, as shown ...
Face Detection, Bounding Box Aggregation and Pose Estimation for Robust Facial Landmark Localisation in the Wild

1

Zhen-Hua Feng1,2 Josef Kittler1 Muhammad Awais1 Patrik Huber1 Xiao-Jun Wu2 Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford GU2 7XH, UK 2 School of Internet of Things Engineering, Jiangnan University, Wuxi 214122, China {z.feng, j.kittler, m.a.rana, p.huber}@surrey.ac.uk, wu [email protected]

Abstract We present a framework for robust face detection and landmark localisation of faces in the wild, which has been evaluated as part of ‘the 2nd Facial Landmark Localisation Competition’. The framework has four stages: face detection, bounding box aggregation, pose estimation and landmark localisation. To achieve a high detection rate, we use two publicly available CNN-based face detectors and two proprietary detectors. We aggregate the detected face bounding boxes of each input image to reduce false positives and improve face detection accuracy. A cascaded shape regressor, trained using faces with a variety of pose variations, is then employed for pose estimation and image pre-processing. Last, we train the final cascaded shape regressor for fine-grained landmark localisation, using a large number of training samples with limited pose variations. The experimental results obtained on the 300W and Menpo benchmarks demonstrate the superiority of our framework over state-of-the-art methods.

1. Introduction Facial landmarks provide important information for face image analysis such as face recognition [4, 43, 45, 46], expression analysis [14, 15, 24] and 3D face reconstruction [26, 27, 30, 36, 60]. Given an input face image, the task of facial landmark localisation is to obtain the coordinates of a set of pre-defined facial landmarks. These facial landmarks usually have specific semantic meanings, such as the nose tip and eye corners, and are instrumental in enabling the subsequent face image analysis. The most well-known facial landmark localisation algorithms, which emerged from the early research efforts on this topic, are based on Active Shape Model (ASM) [7], Active Appearance Model (AAM) [8] and Constrained Local Model (CLM) [11]. These algorithms perform well in constrained scenarios, but cannot cope reliably with faces

obtained in unconstrained conditions. The current research attempts to redress this situation. Its aim is to design algorithms for robust and accurate face landmarking of unconstrained faces in the presence of extreme appearance variations in pose, expression, illumination, makeup, occlusion etc. To this end, the recent trend has been to resort to discriminative approaches such as Cascaded Shape Regression (CSR) [6, 13, 17, 18, 19, 50] and Deep Neural Networks (DNN) [44, 47, 49, 55, 56]. Both have shown promising landmarking results for faces in the wild. To evaluate the performance of facial landmark localisation algorithms, a number of face datasets have been collected or annotated, such as the IMM [35], BioID [29], XM2VTS [34], LFPW [3], COFW [5], AFLW [31], HELEN [32] and 300W [39] datasets. However, their key characteristics such as dataset size, appearance variation types, number of landmarks and landmark quality are not always all satisfactory or well defined. To comprehensively evaluate and compare facial landmark localisation algorithms, the Menpo benchmark has been conceived [54, 47]. It contains 8979 training images and 7281 test images, which have been semi-automatically annotated with 68 facial landmarks for semi-frontal faces and 39 landmarks for profile faces. The Menpo images were selected from the AFLW [31] and FDDB [28] datasets, exhibiting a wide range of appearance variations. In contrast to the previous evaluation campaigns, the Menpo benchmark does not provide face bounding boxes for landmark initialisation, which better reflects practical applications. However, the accuracy of landmark localisation algorithms highly relies on the consistency of the detected face bounding boxes for every input image. To achieve robust face detection and landmark localisation, we develop a coarse-to-fine framework designed to enhance the consistency of detected face designation as well as the accuracy of facial landmarking. The proposed framework has four stages including face detection using multiple face detectors, face bounding box aggregation, pose estimation and facial landmark localisation, as shown in Fig. 1. The key in-

Face Detection

Bounding Box Aggregation

Pose Estimation

Landmark Localisation

Figure 1. The pipeline of our proposed facial landmark localisation framework.

novation of the proposed framework is twofold: 1) we propose the use of multiple face detection experts to provide input to a bounding box aggregation strategy to improve the accuracy of face detection; 2) we divide the original CSR method into a number of coarse-to-fine steps that further improve the accuracy of facial landmark localisation. The paper is organised as follows. We first overview the related literature and introduce the basic cascaded shape regression method in Sections 2 and 3. Then the proposed coarse-to-fine framework and its four main stages are presented in Section 4. In Section 5, we report the experimental results obtained on the 300W and Menpo benchmarks. Conclusions are drawn in Section 6.

2. Related Work The early advances in facial landmark localisation were mainly driven by ASM, AAM and CLM, in which a generative PCA-based shape model is used for face shape representation. Typically in these methods, the statistical distribution of the samples used for the shape model training provides a prior for the model fitting process. To fit the shape model to an input image, an objective function is designed to optimise the model parameters. For example, AAM uses the intensity difference between an input image and a texture model as the loss for model optimisation. These models and their extensions have achieved great success in facial landmarking for faces with controlled appearance variations [1, 9, 20, 21, 23, 33, 40]. However, for faces in the wild, the PCA-based shape model may miss some shape details and in consequence it is not able to represent complex shapes faithfully. To address this issue, cuttingedge techniques use learning-based discriminative methods that employ the non-parametric Point Distribution Model (PDM) [10] for shape representation. Given an input face image and an initial shape estimate, a discriminative model learns a mapping function from the shape-related features to a shape update, driving the landmarks of the initial shape estimate to their optimal locations. In recent years, discriminative models, in particular

CSR-based methods, have become the state-of-the-art in robust facial landmark localisation of unconstrained faces. The key to the success of CSR is the architecture cascading multiple weak regressors, which greatly improves the generalisation capacity and accuracy of a single regression model. To form a CSR, linear regression [18, 50, 52], random forests or ferns [6, 37] and deep neural networks [44, 47, 57] have been used as weak regressors. To further improve landmark localisation accuracy of CSR-based approaches, new architectures have been proposed. For example, Feng et al. proposed to fuse multiple CSRs to improve the accuracy of a single CSR model [18]. Xiong et al. proposed the global supervised descent method using multiple view-based CSRs to deal with the difficulties posed by extreme appearance variations [51]. Moreover, data augmentation [17, 19, 58] and 3D-based approaches [58] have also been used to enhance existing CSR-based facial landmark localisation approaches.

3. Cascaded Shape Regression Given a face image I, the face shape is represented in the form of a vector consisting of the coordinates of L pre-defined 2D facial landmarks, i.e. s = [x1 , ..., xL , y1 , ..., yL ]T . To automatically obtain the face shape for the input image, CSR is used in our proposed framework. In fact, CSR is a strong regressor, Φ, formed by a set of weak regressors: Φ = {φ1 , ..., φM },

(1)

where φm is the mth weak regressor and M is the number of weak regressors. It should be noted that any regression method can be used as a weak regressor in CSR. In this paper, we adopted ridge regression for each weak regressor. Suppose we have N training samples, {In , s∗n }N n=1 , where In is the nth training image and s∗n is the ground truth face shape. We construct a CSR by progressively training a set of cascaded weak regressors. To this end, we first initialise the face shape estimate, s0n , for each training image

1

2 3

4

5 6

input : image I and a trained CSR model Φ = {φ1 , ..., φM } output: facial landmarks s0 initialise the current face shape, s0 , using the detected face bounding box; for m ← 1 to M do extract shape-related features, f (I, s0 ), from the image using the current shape estimate; apply the mth weak regressor to obtain the shape update δs using Eq. (3); update the current shape estimate s0 ← s0 + δs; end

Algorithm 1: Cascaded shape regression for facial landmark localisation. by scaling and translating the mean shape so that it perfectly fits into the detected face bounding box. Next, the shape difference between the current shape estimate and the ground truth shape of each training sample is calculated, i.e. δs∗n = s∗n − s0n . Then the first weak regressor, φ1 = {A1 , e1 }, is obtained by solving the optimisation problem: arg min A1 ,e1

N X

||A1 ·f (In , s0n )+e1 −δs∗n ||22 +λ||A1 ||2F , (2)

n=1

where A1 ∈ R2L×Nf is the trained projection matrix for the first weak regressor, Nf is the dimensionality of the extracted shape-related features using f () and e1 ∈ R2L is the offset. f (I, s0 ) is a mapping function that extracts local features, e.g. HOG, LBP or SIFT, from the neighbourhood around each landmark of the current shape estimate, s0 , and concatenates them into a long vector. Solving Eq. (2) is a typical least square estimation problem that has a closedform solution. Once we obtain the first weak regressor, we use it to predict the shape update, δs = A · f (I, s0 ) + e,

(3)

for each training sample. Last, the current shape estimates of all the training samples are updated for the second weak regressor training, i.e. s0n ← s0n + δsn (n = 1, ..., N ). We repeat this process several times until all the M weak regressors are learnt. For a test image, we first initialise the current shape estimate using the detected face bounding box as mentioned above. Then the pre-trained weak regressors are progressively applied to the current shape estimate for shape update, as summarised in Algorithm 1.

4. The Proposed Framework The proposed facial landmark localisation framework has four main stages, as shown in Fig. 1. Because the

Menpo benchmark does not provide face bounding boxes for landmark initialisation, we first perform face detection for a given image. To improve the detection rate, we use four different face detectors. Second, the detected face bounding boxes are filtered and fused using an aggregation algorithm to reduce false positives and improve the accuracy. Third, we predict the pose of the face image for preprocessing, e.g. image rotation and flipping. Last, cascaded shape regression is used to obtain the final facial landmark localisation result.

4.1. Face Detection As aforementioned, our face detection stage uses four face detectors. The first two are public available face detectors: the Convolutional Neural Network (CNN-) based dlib face detector 1 and the Multi-Task CNN (MTCNN) face detector [56]. In addition, we train two proprietary face detectors to further improve the detection rate. One of them is a general deep-learning-based face detector and the other one is a regression-based face detector specifically tailored for the Menpo benchmark. Recently, the state of the art in deep-learning-based object detection advanced significantly [41, 25, 38]. Regionbased methods like Faster Region-based CNN (Faster RCNN) [38] and its variants emerged as state-of-the-art in detection. Besides the two publicly available CNN-based face detectors, we train a new deep-learning-based face detector using Faster R-CNN. Faster R-CNN consists of two main components: Region Proposal Network (RPN) and Fast RCNN [22]. RPN takes an image as input and generates rectangular object proposals, each with probability of belonging to foreground objects (objectiveness) vs background. The RPN proposals are passed through non-maximum suppression and then sorted out with respect to objectiveness. Then top-N rectangular proposals are fed into Fast R-CNN for classification into different categories and refinement of the bounding box for these proposals. For face detection, Fast R-CNN is trained to detect only one class, i.e. human faces. We use VGG 16 layers network [42] as shared convolutional feature maps and use approximate joint training to update parameters. We train our face detector using the WIDER FACE dataset [53]. In order to adapt to the characteristics of the challenge, we remove all the tiny faces (faces consisting of less that 80 pixels area) from the training set of WIDER FACE. Furthermore, the trained face detector is fine-tuned on the training set of the Menpo benchmark. In the challenge, a face covers a significant area of the image, therefore, we use regions with intersection-over-union (IoU) of more than 0.75 with the ground truth box as positive samples, and regions with IoU of less than 0.35 with the ground truth as negative samples in the fine tuning stage. Our 1 http://dlib.net

Figure 2. Face bounding box aggregation: the first row shows the initial face detection results using the dlib (green), MTCNN (magenta), our own trained Faster R-CNN (red) and regression-based (blue) face detectors; the second row shows the final aggregated face bounding box for each image.

Faster-R-CNN-based face detector has achieved very competitive results on the well-known Face Detection Dataset and Benchmark (FDDB) [28]. When testing on the training set of the Menpo benchmark, even using these three deep-learning-based face detectors, it is not always possible to detect all the faces in an input image. To address this issue, we develop a 4th face detector optimised for the Menpo benchmark by leveraging the benchmark rules, namely that ‘each image contains a single annotated face and it is the one that is closer to the centre of the image’. This face detector is regressionbased, which is similar to the face bounding box refinement step used in the Dynamic Attention-Controlled CSR (DACCSR) [19]. The difference is that DAC-CSR uses only one regression model for bounding box refinement, but we cascade multiple weak regressors for face bounding box regression. The training of this cascaded bounding box regression, Φb = {φb,1 , ..., φb,M },

(4)

is similar to the training of a CSR introduced in Section 3. Here, each weak regressor updates the current face bounding box estimate, b = [x, y, w, h]0 , instead of the current face shape, s, for an image, where (x, y), w and h are the coordinates of the left-upper corner, width and hight of a face bounding box. For face bounding box initialisation, we simply use the bounding box covering the whole region of an input image. The ground truth bounding box of a face is obtained by using the one that exactly encloses all the landmarks of the face. As features f (I, b), we extract multiscale HOG and LBP descriptors from the image patch inside the current face bounding box. This regression-based face detector is trained on the training set of the Menpo benchmark using 2 weak regressors.

4.2. Face Bounding Box Aggregation The use of multiple face detectors results in a number of face bounding boxes as output, including false positives. However, the Menpo benchmark protocol specifies the face closer to the centre of an input image as the correct one. To improve the accuracy of the face detection module and reduce false positives, we perform face bounding box aggregation of all the detected faces of an input image. Fig. 2 shows some examples of our initially detected faces and the final bounding boxes based on our aggregation approach. To perform face bounding box aggregation, we first filter the face bounding boxes detected by the first three deeplearning-based face detectors. In this step, we eliminate all the detected face bounding boxes that satisfy one of the following conditions: 1) the detection confidence score is lower than 0.85; 2) the bounding box height is smaller than 1/5 of the image height; and 3) the bounding box does not include the image centre. After that, we perform face bounding box refinement for the remaining face bounding boxes output by the first three face detectors. The face bounding box refinement step is the same as the one used in DAC-CSR [19]. However, we train three of them for the three deep-learning-based face detectors separately. The training is similar to that of our 4th regression-based face detector, as introduced in the last section. However, here we use the detected face bounding box of a face detector, rather than the one covering the whole image, for bounding box initialisation. The refined bounding boxes are averaged to obtain the final result. If all the detected face bounding boxes are eliminated during the filtering stage, we use the one with the highest score among all the original bounding boxes output by the dlib and MTCNN face detectors for bounding box refinement. Last, if dlib and MTCNN have not detected any faces, we just simply use the bounding box

Because a semi-frontal face has 68 landmarks and a profile face has only 39 landmarks, we train a separate CSRbased facial landmark localiser for each of the scenarios. To prepare the training data for the semi-frontal subset, we randomly rotate all the training images between [0◦ , 360◦ ]. For the profile subset, we flip each training image from left to right and randomly rotate it between [−30◦ , 30◦ ]. For feature extraction, we adopt both HOG and LBP descriptors computed in the neighbourhoods of all the landmarks of the current shape estimate and concatenate them into a long vector. In addition, we also extract dense local features from the image patch enclosing all the landmarks as the contextual information, which has also been used in DAC-CSR [19].

4.4. Facial Landmark Localisation

Figure 3. Pose estimation and image pre-processing for semifrontal (first row) and profile (second row) faces.

output from our 4th regression-based face detector as the final face detection result.

4.3. Pose Estimation The Menpo benchmark has two test sets, semi-frontal and profile, that are evaluated separately. These two test sets consist of face images with a variety of in-the-plane pose rotations. In addition, the profile subset has either leftor right-rotated face images with respect to yaw angle rotation. Based on the face bounding boxes output from the first two stages of our framework, we could train a model using a dataset with a wide range of in-the-plane and out-of-plane pose variations for facial landmark localisation. The resulting model would cope well with pose variations due to the property of the training data. However, the accuracy of facial landmark localisation using a model trained on such an extreme range of face poses is limited. To mitigate this problem, we estimate the face pose in each input image for image pre-processing before the last facial landmark localisation stage. In the pose estimation stage, we first perform rough facial landmark localisation by cascading only 2 weak regressors. Then, for a semi-frontal face image, we use two landmarks (ID-28 and ID-34) to determine its in-the-plane rotation. For a profile face, we use the landmarks with ID-3 and ID-20 to determine the yaw rotation. Once the pose is estimated, we use this information to apply an appropriate pose normalisation. We rotate the image when the estimated pose is larger than 45◦ away from upright for semi-frontal faces, and flip right-rotated profile face images from right to left to obtain left-rotated faces. Two examples are shown in Fig. 3.

In the last stage, we perform facial landmark localisation for the face detected in an input image, using a pre-trained CSR model. We train separate CSR models for the semifrontal and profile subsets, using datasets with limited pose variations. For the semi-frontal subset, we flip an original training image and perform random image rotation between [−30◦ , 30◦ ]. For the profile training faces, we flip all the right-rotated faces in yaw from right to left to construct a training set with only left-rotated profile faces. We also perform in-the-plane rotation to all the profile faces for data augmentation. The training and testing phases of these two CSR models are described in Section 3. To further improve the robustness of the final localised facial landmarks, we perform random perturbation to a detected face bounding box by randomly translating the left-upper and right-bottom corners inside 5% of the width and height of the bounding box. We apply a pre-trained CSR to each randomly perturbed face bounding box to obtain multiple facial landmark localisation results and use the mean as the final result. In this stage, we cascade 6 weak regressors in a CSR. For shape-related feature extraction, i.e. f (I, s), we use the same technique as described in Section 4.3. It should be noted that, for images that have been rotated or flipped in the pose estimation stage, the localised facial landmarks have to be transformed back to their original coordinate systems.

5. Experimental Results 5.1. Implementation Details To evaluate the performance of our proposed framework, the 300W and Menpo benchmarks were used [54]2 . The latter one was used in the 2nd facial landmark localisation competition. For training, the 300W benchmark provides a number of annotated face datasets including XM2VTS [34], LFPW [3], HELEN [32] and AFW [59]. For evaluation, 2 https://ibug.doc.ic.ac.uk/resources

(a) Semi-frontal

(a) 68 points

(b) Profile

Figure 5. A comparison between our proposed framework and the APS method on (a) the semi-frontal, and (b) the profile subsets of the Menpo benchmark.

5.2. Evaluation on 300W

(b) 51 points

Figure 4. A comparison of our proposed framework with stateof-the-art methods on the 300W dataset (indoor + outdoor). The performance was evaluated in terms of the normalised RMS error over (a) 68 and (b) 51 facial points.

300W provides 300 indoor and 300 outdoor face images. Each 300W face image has 68 facial landmarks that were semi-automatically annotated. The Menpo benchmark has 8979 training images (6679 semi-frontal and 2300 profile faces) and 7281 test images (5335 semi-frontal and 1946 profile faces). For each Menpo semi-frontal/profile face, 68/39 landmarks were annotated. The 4th face detector and the face bounding box refinement models used in the aggregation stage were trained using all the training images of the Menpo benchmark. We used the LFPW, AFW and HELEN datasets and the training set of the Menpo benchmark for the CSR training in our pose estimation and facial landmark localisation stages. To extract HOG and LBP features, the VLFeat toolbox was used [48].

We first evaluate the accuracy of our proposed framework on the 300W dataset using the protocol of the second run of the 300W challenge that was completed at the beginning of 2015 [39]. In general, the protocol of a facial landmark localisation benchmarking dataset provides face bounding boxes for initialisation. The re-run of the 300W challenge is the only one that has the same protocol as the Menpo benchmark, i.e. a participant has to perform face detection and landmark localisation jointly. We compare our method with all the participants’ approaches from the second evaluation campaign of the 300W challenge in terms of accuracy [39]. The evaluation results are shown in Fig. 4. The proposed framework outperforms the algorithms of all the participants in the second evaluation campaign of the 300W challenge. Among these algorithms, Deng [12] and Fan [16] are the academia and industry winners of the challenge.

5.3. Evaluation on the Menpo Benchmark The final facial landmark localisation results on the test set of the Menpo benchmark are shown in Fig. 5. The figure compares our proposed framework with the baseline method, i.e. Viola-Jones face detector and Active Pictorial Structures (APS) [2]. The proposed framework significantly outperforms the

(a) Semi-frontal

(b) Profile

Figure 6. Some landmark localisation results of the proposed method on the Menpo benchmark.

baseline method in terms of accuracy. It is clear that the face detection and bounding box aggregation techniques used in our approach successfully detected the majority of the test faces. Moreover, the use of pose estimation and cascaded shape regression helps to achieve accurate facial landmark localisation results. Some examples are shown in Fig. 6.

6. Conclusion We have presented a coarse-to-fine facial landmark localisation framework for the Menpo benchmark competi-

tion. The proposed framework applied four face detectors and a bounding box aggregation method for robust and accurate face detection. Then a cascaded shape regression model trained using a number of training samples with a wide spectrum of pose variations was used for rough facial landmark localisation and pose estimation. Last, we rotated or flipped the test images for accurate facial landmark localisation using a fine-grained cascaded shape regression model trained using a face dataset with limited pose variations. The experimental results carried out on the 300W and

Menpo benchmarks demonstrate the superiority of the proposed framework. The key to the success of the proposed framework is the splitting of the original cascaded shape regression process into a number of coarse-to-fine steps. In addition, the use of an ensemble of multiple face detectors greatly improves the accuracy of the face detection step.

Acknowledgements This work was supported in part by the EPSRC Programme Grant ‘FACER2VM’ (EP/N007743/1), the National Natural Science Foundation of China (61373055, 61672265) and the Natural Science Foundation of Jiangsu Province (BK20140419, BK20161135).

References [1] J. Alabort-i-Medina and S. Zafeiriou. A Unified Framework for Compositional Fitting of Active Appearance Models. International Journal of Computer Vision, pages 1–39, 2016. [2] E. Antonakos, J. Alabort-i-Medina, and S. Zafeiriou. Active pictorial structures. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5435–5444, 2015. [3] P. N. Belhumeur, D. W. Jacobs, D. Kriegman, and N. Kumar. Localizing parts of faces using a consensus of exemplars. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 545–552, 2011. [4] J. R. Beveridge, H. Zhang, B. A. Draper, P. J. Flynn, Z.-H. Feng, P. Huber, J. Kittler, Z. Huang, S. Li, Y. Li, et al. Report on the FG 2015 video person recognition evaluation. In IEEE International Conference on Automatic Face and Gesture Recognition (FG), volume 1, pages 1–8. IEEE, 2015. [5] X. P. Burgos-Artizzu, P. Perona, and P. Doll´ar. Robust face landmark estimation under occlusion. In International Conference on Computer Vision, 2013. [6] X. Cao, Y. Wei, F. Wen, and J. Sun. Face alignment by explicit shape regression. International Journal of Computer Vision, 107(2):177–190, 2014. [7] T. Cootes, C. Taylor, D. Cooper, J. Graham, et al. Active shape models-their training and application. Computer Vision and Image Understanding, 61(1):38–59, 1995. [8] T. F. Cootes, G. Edwards, and C. J. Taylor. Active appearance models. In European Conference on Computer Vision (ECCV), volume 1407, pages 484–498, 1998. [9] T. F. Cootes, M. C. Ionita, C. Lindner, and P. Sauer. Robust and accurate shape model fitting using random forest regression voting. In European Conference on Computer Vision (ECCV), pages 278–291. Springer, 2012. [10] T. F. Cootes and C. J. Taylor. Combining point distribution models with shape models based on finite element analysis. Image and Vision Computing, 13(5):403–409, 1995. [11] D. Cristinacce and T. F. Cootes. Feature Detection and Tracking with Constrained Local Models. In British Mahine Vision Conference (BMVC), volume 3, pages 929–938, 2006. [12] J. Deng, Q. Liu, J. Yang, and D. Tao. M3CSR: Multi-view, multi-scale and multi-component cascade shape regression. Image and Vision Computing, 47:19–26, 2016.

[13] P. Doll´ar, P. Welinder, and P. Perona. Cascaded pose regression. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1078–1085. IEEE, 2010. [14] S. Eleftheriadis, O. Rudovic, M. P. Deisenroth, and M. Pantic. Variational gaussian process auto-encoder for ordinal prediction of facial action units. In Asian Conference on Computer Vision (ACCV), Taipei, Taiwan. Oral, November 2016. [15] S. Eleftheriadis, O. Rudovic, and M. Pantic. Discriminative shared gaussian processes for multiview and view-invariant facial expression recognition. IEEE transactions on Image Processing, 24(1):189–204, 2015. [16] H. Fan and E. Zhou. Approaching human level facial landmark localization by deep learning. Image and Vision Computing, 47:27–35, 2016. [17] Z.-H. Feng, G. Hu, J. Kittler, W. Christmas, and X.-J. Wu. Cascaded collaborative regression for robust facial landmark detection trained using a mixture of synthetic and real images with dynamic weighting. IEEE Transactions on Image Processing, 24(11):3425–3440, 2015. [18] Z.-H. Feng, P. Huber, J. Kittler, W. Christmas, and X. Wu. Random Cascaded-Regression Copse for Robust Facial Landmark Detection. IEEE Signal Processing Letters, 22(1):76–80, Jan 2015. [19] Z.-H. Feng, J. Kittler, W. Christmas, P. Huber, and X.-J. Wu. Dynamic Attention-controlled Cascaded Shape Regression Exploiting Training Data Augmentation and Fuzzy-set Sample Weighting. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017. [20] Z.-H. Feng, J. Kittler, W. Christmas, and X.-J. Wu. A unified tensor-based active appearance face model. arXiv preprint arXiv:1612.09548, 2016. [21] Z.-H. Feng, J. Kittler, W. Christmas, X.-J. Wu, and S. Pfeiffer. Automatic face annotation by multilinear AAM with missing values. In International Conference on Pattern Recognition (ICPR), pages 2586–2589. IEEE, 2012. [22] R. Girshick. Fast R-CNN. In IEEE International Conference on Computer Vision (ICCV), pages 1440–1448, 2015. [23] J. Gonzalez-Mora, F. De la Torre, R. Murthi, N. Guil, and E. L. Zapata. Bilinear Active Appearance Models. In IEEE International Conference on Computer Vision (ICCV), pages 1–8, 2007. [24] S. Happy and A. Routray. Automatic facial expression recognition using features of salient facial patches. IEEE transactions on Affective Computing, 6(1):1–12, 2015. [25] K. He, X. Zhang, S. Ren, and J. Sun. Spatial pyramid pooling in deep convolutional networks for visual recognition. In European Conference on Computer Vision (ECCV), pages 346–361. Springer, 2014. [26] G. Hu, F. Yan, J. Kittler, W. Christmas, C.-H. Chan, Z.-H. Feng, and P. Huber. Efficient 3D Morphable Face Model Fitting. Pattern Recognition, 67:366–379, 2017. [27] P. Huber, Z.-H. Feng, W. Christmas, J. Kittler, and M. R¨atsch. Fitting 3D Morphable Face Models using local features. In IEEE International Conference on Image Processing (ICIP), pages 1195–1199. IEEE, 2015.

[28] V. Jain and E. G. Learned-Miller. Fddb: A benchmark for face detection in unconstrained settings. UMass Amherst Technical Report, 2010. [29] O. Jesorsky, K. J. Kirchberg, and R. W. Frischholz. Robust face detection using the hausdorff distance. In Audio-and Video-based Biometric Person Authentication, pages 90–95. Springer, 2001. [30] J. Kittler, P. Huber, Z.-H. Feng, G. Hu, and W. Christmas. 3D Morphable Face Models and Their Applications. In International Conference on Articulated Motion and Deformable Objects, pages 185–206. Springer, 2016. [31] M. Koestinger, P. Wohlhart, P. M. Roth, and H. Bischof. Annotated Facial Landmarks in the Wild: A Large-scale, Realworld Database for Facial Landmark Localization. In First IEEE International Workshop on Benchmarking Facial Image Analysis Technologies, 2011. [32] V. Le, J. Brandt, Z. Lin, L. Bourdev, and T. S. Huang. Interactive facial feature localization. In European Conference on Computer Vision (ECCV), pages 679–692. Springer, 2012. [33] I. Matthews and S. Baker. Active Appearance Models Revisited. International Journal of Computer Vision, 60(2):135– 164, 2004. [34] K. Messer, J. Matas, J. Kittler, J. Luettin, and G. Maitre. XM2VTSDB: The extended M2VTS database. In International Conference on Audio and Video-based Biometric Person Authentication, volume 964, pages 965–966. Citeseer, 1999. [35] M. M. Nordstrøm, M. Larsen, J. Sierakowski, and M. B. Stegmann. The IMM Face Database - An Annotated Dataset of 240 Face Images. Technical report, Informatics and Mathematical Modelling, Technical University of Denmark, DTU, Richard Petersens Plads, Building 321, DK-2800 Kgs. Lyngby, May 2004. [36] M. Piotraschke and V. Blanz. Automated 3d face reconstruction from multiple images using quality measures. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016. [37] S. Ren, X. Cao, Y. Wei, and J. Sun. Face alignment at 3000 fps via regressing local binary features. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1685–1692, 2014. [38] S. Ren, K. He, R. Girshick, and J. Sun. Faster R-CNN: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems (NIPS), pages 91–99, 2015. [39] C. Sagonas, E. Antonakos, G. Tzimiropoulos, S. Zafeiriou, and M. Pantic. 300 faces in-the-wild challenge: Database and results. Image and Vision Computing, 47:3–18, 2016. [40] P. Sauer, T. Cootes, and C. Taylor. Accurate Regression Procedures for Active Appearance Models. In British Machine Vision Conference (BMVC), pages 1–11, 2011. [41] P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun. Overfeat: Integrated recognition, localization and detection using convolutional networks. arXiv preprint arXiv:1312.6229, 2013. [42] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.

[43] X. Song, Z.-H. Feng, G. Hu, J. Kittler, W. Christmas, and X.-J. Wu. Dictionary Integration using 3D Morphable Face Models for Pose-invariant Collaborative-representationbased Classification. arXiv preprint arXiv:1611.00284, 2016. [44] Y. Sun, X. Wang, and X. Tang. Deep Convolutional Network Cascade for Facial Point Detection. In IEEE Conference on Computer Vision and Pattern Recognition, pages 3476– 3483, 2013. [45] Y. Sun, X. Wang, and X. Tang. Deep learning face representation from predicting 10,000 classes. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2014. [46] Y. Taigman, M. Yang, M. Ranzato, and L. Wolf. Deepface: Closing the gap to human-level performance in face verification. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1701–1708, 2014. [47] G. Trigeorgis, P. Snape, M. A. Nicolaou, E. Antonakos, and S. Zafeiriou. Mnemonic Descent Method: A Recurrent Process Applied for End-To-End Face Alignment. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016. [48] A. Vedaldi and B. Fulkerson. VLFeat: An Open and Portable Library of Computer Vision Algorithms. http://www.vlfeat.org/, 2008. [49] S. Xiao, J. Feng, J. Xing, H. Lai, S. Yan, and A. Kassim. Robust facial landmark detection via recurrent attentiverefinement networks. In European Conference on Computer Vision (ECCV), 2016. [50] X. Xiong and F. De la Torre. Supervised descent method and its applications to face alignment. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 532–539, 2013. [51] X. Xiong and F. De la Torre. Global supervised descent method. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2664–2673, 2015. [52] J. Yan, Z. Lei, D. Yi, and S. Z. Li. Learn to Combine Multiple Hypotheses for Accurate Face Alignment. In International Conference of Computer Vision - Workshops, 2013. [53] S. Yang, P. Luo, C. C. Loy, and X. Tang. WIDER FACE: A Face Detection Benchmark. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. [54] S. Zafeiriou, G. Trigeorgis, G. Chrysos, J. Deng, and J. Shen. The Menpo Facial Landmark Localisation Challenge: A step closer to the solution. In IEEE Conference on Computer Vision and Pattern Recognition - Workshops (CVPRW), 2017. [55] J. Zhang, M. Kan, S. Shan, and X. Chen. Occlusion-Free Face Alignment: Deep Regression Networks Coupled With De-Corrupt AutoEncoders. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016. [56] K. Zhang, Z. Zhang, Z. Li, and Y. Qiao. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Processing Letters, 23(10):1499–1503, 2016. [57] E. Zhou, H. Fan, Z. Cao, Y. Jiang, and Q. Yin. Extensive Facial Landmark Localization with Coarse-to-Fine Convo-

lutional Network Cascade. In Iternational Conference on Computer Vision - Workshops, pages 386–391, 2013. [58] X. Zhu, Z. Lei, X. Liu, H. Shi, and S. Z. Li. Face Alignment Across Large Poses: A 3D Solution. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016. [59] X. Zhu and D. Ramanan. Face detection, pose estimation, and landmark localization in the wild. In IEEE Conference on Computer Vision and Pattern Recognition, pages 2879– 2886. IEEE, 2012. [60] X. Zhu, J. Yan, D. Yi, Z. Lei, and S. Z. Li. Discriminative 3D morphable model fitting. In IEEE International Conference and Workshops onAutomatic Face and Gesture Recognition, volume 1, pages 1–8, 2015.