• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    3D Head Pose Estimation through Facial Features and Deep Convolutional Neural Networks

    2021-12-15 12:47:50KhalilKhanJehadAliKashifAhmadAsmaGulGhulamSarwarSahibKhanQuiThanhHoaiTaTaeSunChungandMuhammadAttique
    Computers Materials&Continua 2021年2期

    Khalil Khan,Jehad Ali, Kashif Ahmad, Asma Gul, Ghulam Sarwar,Sahib Khan,Qui Thanh Hoai Ta,Tae-Sun Chung and Muhammad Attique

    1Department of Computer Science and Software Engineering, Pak-Austria Fachhochschule:Institute of Applied Sciences and Technology, Haripur-KPK, Pakistan

    2Department of Computer Engineering and Department of AI Convergence Network,Ajou University,Suwon,16499,South Korea

    3Hamad Bin Khalifa University, Doha,Qatar

    4Department of Statistics, Shaheed Benazir Bhutto Women University, Peshawar, Pakistan

    5Department of Software Engineering, University of Azad Jammu and Kashmir, Pakistan

    6Department of Electronics and Telecommunications, Politecnico di Torino, Torino, 10129,Italy

    7Institute of Research and Development, Duy Tan University, Danang, 550000, Vietnam

    8Department of Computer Engineering, Ajou University, Suwon, 16499,Korea

    9Department of Software, Sejong University, Seoul,05006, South Korea

    Abstract:Face image analysis is one among several important cues in computer vision.Over the last five decades, methods for face analysis have received immense attention due to large scale applications in various face analysis tasks.Face parsing strongly benefits various human face image analysis tasks inducing face pose estimation.In this paper we propose a 3D head pose estimation framework developed through a prior end to end deep face parsing model.We have developed an end to end face parts segmentation framework through deep convolutional neural networks (DCNNs).For training a deep face parts parsing model,we label face images for seven different classes,including eyes,brows,nose,hair,mouth, skin, and back.We extract features from gray scale images by using DCNNs.We train a classifier using the extracted features.We use the probabilistic classification method to produce gray scale images in the form of probability maps for each dense semantic class.We use a next stage of DCNNs and extract features from grayscale images created as probability maps during the segmentation phase.We assess the performance of our newly proposed model on four standard head pose datasets, including Pointing’04, Annotated Facial Landmarks in the Wild (AFLW), Boston University (BU), and ICT-3DHP, obtaining superior results as compared to previous results.

    Keywords: Face image analysis;face parsing;face pose estimation

    1 Introduction

    Face pose estimation,also known as head pose estimation is a challenging task in the field of computer vision.Head pose estimation plays an important role in many real-world applications, including gaze estimation[1],human computer interaction[2],and augmented reality[3].However,face pose estimation is still a difficult task for various reasons such as variations in facial appearance,complex and unconstrained background, and different facial expressions.Head pose estimation is particularly confronted with problems in the uncontrolled and wild conditions.Some of the applications that rely heavily on an accurate head pose estimation system are human behavior analysis, safety during driving, surveillance applications,and targeted advertisements.

    Head pose estimation is linked with gaze estimation; it is confirmed by the research conducted in the 19th century [4].The relationship between head pose estimation and gaze prediction is also confirmed by later stage in research [5].The research conducted in Langton et al.[5] suggests that gaze estimation comes from both eyes and head pose direction.Apart from eyes, this mutual relationship between various face parts and head pose is also confirmed by Huang et al.[6].The work proposed by Huang et al.[6]suggests that relationship between face parts can be exploited for several mid-level vision tasks along with head pose estimation.Our work is also impressed by this idea of mutual relationship.

    The face parsing method proposed in Khan et al.[7],segments a face image into face classes including,mouth, nose, hair, skin, back, and eyes.We also use face parts information by first developing a face segmentation framework.As can be seen these days, a shift in state-of-the-art methods from traditional machine learning algorithms towards recently introduced deep learning methods is prevalent.We also develop a DCCNs based face parsing method for seven different classes.

    Our work is inspired from Huang et al.’s[6]idea.We argue that all face analysis tasks are related and can assist each other in specific applications.The performance of the face pose prediction can be improved if a prior efficiently parsed image having information about various face features is provided as input.The same fact is also confirmed by psychology literature,for example,[8,9].In a nutshell,the performance of the face pose estimation can be improved if the information from various face features is extracted from a segmented image and given as input to the face pose estimation framework.

    Face pose estimation is already being predicted through various face parts information in literature[10,11].These methods involve landmarks localization before face pose estimation.However, the performance of the system in such cases depends on this method [12,13], which is itself another challenge.The landmarks localization algorithms are greatly affected in certain cases such as complicated facial expressions, changes in face rotation and lighting conditions, occlusions, and far field imagery conditions.All these factors make this method a rather challenging task, which ultimately drops the performance of the face pose estimation system; if it depends on it.Unlike the landmarks localization method, we introduce a face pose estimation method which does not need landmarks information but rather depends on various face parts information.

    We propose a face segmentation method based on deep learning that segments a face image into seven different classes.For building a face parsing framework, we labeled 200 face images from each database through image editing software.The deep learning-based model extracts features through Convolutional Neural Networks (CNNs) and build a Soft-Max classifier.When a testing image is provided as input to the face parsing framework, it is segmented into seven face classes.We use a probabilistic classification method and create probability maps (PMAPS) along with segmentation results.We use five different face features and extract information through CNNs to build another Soft-Max classifier.To summarize,contributions of this paper are:

    ? We propose a face parsing method that segments a face image into seven different classes.The face parsing method is based on DCNNs.

    ? We develop a new face pose estimation algorithm.The proposed face pose estimation method is based on a prior face parsing method.

    ? We conduct experiments on state-of-the-art (SOA) datasets for face pose estimation and obtained much better results compared to previous results.

    The structure of the paper is as follows:Section 2 presents related work for head pose estimation.Several datasets are reported by previous literature for head pose estimation.The datasets used in this paper are presented in Section 3.The face segmentation part is presented in Section 4, whereas the proposed head pose estimation algorithm is discussed in Section 5.The obtained results are discussed and then a comparison with SOA is shown in Section 6.We summarize the article with future directions in Section 7.

    2 Related Work

    2.1 Genetic Algorithm (GA)

    Face parsing algorithms can be categorized into local and global based parsing methods.These methods are described in the following paragraphs.

    Local methods:Local methods adopt a specific strategy of coarse-to-fine.Local based methods consider both local precision and global consistency.In local-based algorithms, separate models are trained for various face components, e.g., mouth, nose, eyes, etc.A method proposed by Luo et al.[14]trains a model that segments each face part individually.Zhou et al.[15] propose an interlinked CNNs based model after localizing face parts.The interlinked based method passes information bidirectionally,i.e., coarse and fine levels.The computational cost and memory consumption of the proposed method is large due to the bidirectional level information exchange.Another approach [16] combines the CNNs with Recurrent Neural Networks(RNNs)in two successive stages achieving SOA results on challenging.

    Global methods:Global based methods treat different face parts information globally.Accuracy of these algorithms is less, as single face parts are not targeted.These methods estimate a label for each pixel over the entire face image.Some earlier works represent the spatial relationship between face parts through different models, for example, [17] and exemplar-based model [18].The CNNs structure and loss function were processed by Liu et al.[16], which encodes the underlying layouts of the face image.This method integrates conditional random fields with CNNs, which the authors named Multi-Objective learning method.Jackson et al.[19] integrated CNNs with boundary cue to confine face regions.This method utilizes facial landmarks in the first step.Super-pixel information, Conditional Random Fields(CRFs), and CNNs are combined by Zhou et al.[20].The method proposed in Zhou et al.[20] employs fully convolutional networks, therein obtaining better performance compared to SOA.The method proposed by Wei et al.[21] regulates receptive fields in a face parsing network.To achieve good performance on real time scenario, Saito et al.[22] propose another algorithm.The computational cost of this method is much lower than other methods.

    2.2 Face Pose Estimation

    Before describing the proposed face pose estimation model, we review related work on face pose estimation algorithms in this Section of the paper.A rich literature already exists for head pose estimation; however, in this Section of the article, we will try to provide maximum information about the recently introduced algorithms for face pose estimation.

    Face pose can be classified into three categories, including yaw, roll, and pitch.The horizontal orientation is represented with yaw, vertical orientation with pitch, and the image plane by roll angle.We evaluate our proposed face pose estimation method with four large scale datasets, including Pointing’04 [23], AFLW [24], ICT-3DHPE [25], and Boston University (BU) [26] datasets.We classify the face pose estimation methods into three categories, including holistic approaches, geometric, and deep learning-based methods.These methods are described below.

    2.2.1 Holistic Methods

    In holistic approaches, the face image is considered as one-dimension vector, and certain features are extracted.Holistic methods assume a certain relationship between 2D face image properties and their 3D pose.A large number of face images are used for training purposes,and various statistical learning methods are exploited with different classification tools.The trained model then differentiates between various face poses.Some methods which used holistic approaches to address head pose estimation can be explored in[27-30].

    Holistic methods show some advantages over its competitive methods.These approaches are comparatively simple and easy to implement.These methods are fit both for high- and low-resolution images.Moreover, no negative training data is needed in the training stage.Extension of the template models is also easy and can be extended any time without doing sufficient changes in the architecture.

    Holistic methods also face some serious weaknesses.Like other methods,these algorithms also suffer from the limitation that the system must detect the face part.The system accuracy also degrades drastically with localization errors.Holistic methods become unreliable with variations in face appearance,changing in lighting conditions, and occlusions.A significant problem faced by these methods is pair wise similarity,which is the faulty assumption of the images of the same candidate in two different positions.

    2.2.2 Geometric Methods

    Geometric methods are also known as model-based methods.These methods require the localization of certain facial key points such as eyebrow,eyes,nose,the tip of the nose,lips,etc.A single feature vector is extracted from the located facial key points, and then the desired pose is predicted based on the relative position of the extracted key points.These methods are almost similar to how the human brain estimates the head pose of a face image.

    The literature reports different face features in different combinations for head pose estimation.The intraocular distance and eyes are commonly used for head pose discrimination due to their easy detection[31].In some cases, the mouth is also used, but mouth detection is comparatively difficult if the facial expressions are complicated.The tip of the nose and hair is another discriminative cue which is used for modeling of an efficient head pose estimation system.

    Geometric methods are very fast,as very few features are needed for modeling.However,if the number of features is increased,the computational cost also raises;for example,Cootes et al.[32]uses a combination of fifteen different feature points around the mouth,eyes,and nose regions.Another main advantage noted for these methods is; the extracted points are robust to rotation and translation changes.

    For different key points localization Active Shape Modeling(ASM)[33]is frequently used.However,ASM fails drastically in some critical conditions,for example,changing in lighting conditions,complicated facial expressions, and occlusions.If ASM does not perform well, the performance of the head pose estimation drops significantly.Some methods which use geometric based methods can be explored in references[34-36].

    2.2.3 Deep Learning Methods

    The performance of various visual recognition tasks has been greatly improved with deep learning architecture.Some very complex scenarios of computer vision tasks are improved with these deep learning methods.Major weaknesses of the conventional machine learning techniques are mitigated with this transition of deep learning architecture.The same is the case with face pose estimation.

    Ruiz et al.[37] propose a deep learning-based method that does not depend on prior landmarks estimation.Some hybrid models are proposed in [38-40], which address head pose estimation along with other face image analysis tasks such as gender and race classification, face recognition, and detection, etc.Similarly, Hsu et al.[41] propose a method that combines regression-based function with deep learning modules.The lastly proposed method is named QuatNet (Quaternions) by the authors.A new deep learning-based method is proposed by Lee et al.[42], which is evaluated with Pointing’04.The method proposed in Lee et al.[42] is fast and comparatively robust to certain environmental factors; therefore,presents a better choice for the datasets collected in the wild conditions.

    The performance of conventional machine learning methods is satisfactory on the datasets collected in constrained conditions.However, when these traditional machine learning methods are exposed to such unconstrained datasets, performance drops drastically.Unlike conventional machine learning methods,deep learning methods show much improvement with the challenging database.In a nutshell, some research work on face pose estimation already exists, but still, face pose estimation is an open challenge for researchers.

    3 Databases

    We evaluate our face pose estimation framework with five datasets,including Pointing’04,BU,AFLW,and ICT-3DHPE.This Section presents details about the databases used in our proposed work.

    ?Pointing’04:Images in the Pointing’04 are manually annotated.Although it is an old dataset,it is still used by researchers[43-45]due to the diversity in the images and its challenging nature.The database consists of fifteen sets of low-resolution images.Each individual set has a further two subsets having 93 face images for every subject in different orientations.The age of all participants in the dataset is between 20-40 years.Some participants in the datasets also include facial hair and some wearing glasses.The head pose of a subject is determined with the pan and tilt angle.Each participant is asked to look into 93 different markers marked on a wall in a measurement room.Each point represents a specific head pose.Due to manual labeling, the given face localization may not be accurate in some cases.The head orientation varies between -90oto +90ofor yaw pose.The step size between two consecutive poses is 15.Similarly, for pitch, the top poses are represented with positive values and bottom poses with negative values.The difference between the two poses for pitch is 30o.

    ?AFLW:These images are collected in the unconstrained condition with large variations in lighting conditions, facial expression, appearance, and some environmental factors.All images in AFLW are collected from the Internet.These images are collected in 9 different lighting conditions.The total number of face images is 13,000, whereas the number of participants is 5,749.The head pose is manually annotated in AFLW, where the yaw angle varies between ±120oand pitch and roll in the range±90o.

    ?BU Data-set:This dataset has two sequences; images exposed to changing lighting conditions and those collected in controlled and uniform lighting conditions.The database consists of both RGB and depth images.We use only RGB images in our experiments.We consider all three rotation angles, i.e., pitch, yaw, and roll.The number of participants in the database is only five.To collect ground truth data,magnetic trackers are attached to every subject’s head.

    ?ICT-3DHPE:These images are collected through Kinect sensor.Both RGB and depth images are included; however, we use only RGB for our experiments.The total number of participants in the dataset is ten, with six male and four females.The ground truth data, in this case, is also accurate as again magnetic tracker is attached to each subject’s head.

    4 Proposed Face Parsing Framework

    The face parsing module of the proposed framework is presented in this Section of the paper.We make this face parsing model for each dataset separately.Some of the datasets do not provide cropped face images;we apply a face detection method in the initial phase.As face detection is a mature research area,we use a face detection algorithm already proposed in Wei et al.[46]to each image.We re-scaled each face image to a fixed size of 227 × 227 after face detection.The proposed DCNNs based face segmentation model and its architecture is presented in Tab.1.The Fig.1 shows the proposed face parsing module.

    Table 1:Information about each CNNs layer

    Figure 1:Proposed deep CNNs face parsing framework

    Architecture:There are some parameters that greatly affect the performance of the DCNNs based model.For example, the size of the kernel used for CNNs and the pooling layer, the number of layers used for convolution, and the number of filters in every layer.In our face parsing module, we use four sets of convolutional layers (C1-C4), followed by maximum pooling layers (P1-P4) and, at the end, two fully connected layers.We fix the size of the kernel in convolutional as 5 × 5.We also fix the down sampling stride, as can be seen in Tab.1.Details about the convolutional layer, feature map, and kernel size are shown in Tab.1.Some other parameters of the proposed DCNNs are presented in Tab.2.

    Table 2:CNNs Parameters setting for training

    We use a rectified linear unit for activation function.We embed the maximum pooling layer after each convolutional layer.Our proposed DCNNs face parsing model has main three parts, i.e., convolutional layers, maximum pooling layers, and two fully connected layers.We represent the convolutional layer kernel byN*M*C.Where height and width of the kernel is represented byNandMand the channel byC.The maximum pooling layer kernel is represented byP*Q, wherePare representing height andQwidth of the kernel.The fully connected layer performs the task of classification.For optimization of the deep learning architecture and more details,Goodfellow et al.[47]can be explored further.The overview of the face parsing module is summarized in Tab.1.We train a face parsing module for each database individually.

    5 Proposed Features Based Face Pose Estimation Algorithm

    Our proposed face pose estimation model is summarized in Algorithm 1.Initially, we develop a face segmentation model through DCNNs.The face parsing model outputs the most likely class for each pixel in a face image.We created PMAPS during the segmentation phase, which we further use for face pose estimation.We investigate different combinations of these PMAPS to know which face parts help face pose differentiation.We represent these PMAPS as:PMAPSnose, PMAPSback, PMAPSeyes,PMAPSeyebrows, PMAPSskin, PMAPSmouth, and PMAPShair.Fig.1 shows some face images from Pointing’04 along with PMAPS for all five face classes, which we use in our face pose prediction model.PMAPS are grey scale images where higher intensity shows more probability of estimation for a face class and vice versa.

    Algorithm 1:Proposed face pose estimation algorithm Input: Gtraining = Mn,Tn ( ){ }m n=1 , Gtesting Where Gtraining is the training data for DCNNs model,Gtesting is the testing data,M is the training image and T i,j( )∈ 0,1,2,3,4,5,6{}is the data used as ground truth.a.Face parsing part:Step a.1:Training a face parsing DCNNs model through training data.Step a.2:Producing face parts segmentation and probability maps for each dense semantic class Step a.3:Using the DCNNs model to create PMAPS for the seven classes listed as;PMAPSnose,PMAPSback,PMAPSeyes,PMAPSeyebrows,PMAPSskin ,PMAPSmouth,and PMAPShair b.Face parsing part:Step b.1:Extracting information from the PMAPS through DCNNs from five semantic classes including nose,hair,mouth,eyes,and eyebrows.Step b.2:Training a Soft-Max classifier A by creating feature vector such that;f =PMAPSeyes +PMAPSmouth +PMAPSeyebrow+ PMAPSnose Output:estimated face pose.

    For head pose estimation, we used extracted features from PMAPS images through CNNs.After extracting the features, we train another Soft-Max classifier for each dataset.

    We manually labeled 200 face images from each dataset for seven face classes.The manually labeled images are used to build a face parsing model.For all images of every dataset, the PMAPS are generated.When a testing image is provided as input, the face parsing model creates the PMAPS for seven face classes.We conduct detailed experiments to investigate which face features can help face pose estimation.

    After detailed analyses, we decided to use PMAPS for five classes, including the eyes, nose, mouth,eyebrow, and hair.After extracting features through CNNs we concatenate five feature vectors with each other to build a single unique feature vector.We trained another Soft-Max classifier using the feature vector.To validate our model more precisely, we use ten-fold cross-validation experiments.However, we exclude those 200 images which we previously used to build our classifier.The PMAPS of a subject from Pointing’04 can be seen in Fig.2.From Fig.2, it can be noticed that changes in PMAPS are occurring as we move from one pose to another.

    Figure 2:Segmentation results and probability maps from images of Pointing’04 dataset,where pose varies from-90°to+90°with 15°difference between two adjacent poses.The images are in the order where:row 1:original testing images,row 2:segmentation results with proposed face parsing module,row 3:probability maps for hair,row 4:probability maps for mouth,row 5:probability maps for nose,row 6:probability maps for eyes,and row 7:probability maps for eye brows

    We investigate some interesting points during our experiments.We came to know that minor classes have more contribution towards face pose estimation system as compared to major classes (except hair).Hence, we use PMAPS of the four small classes (nose, eyes, brow, and mouth) and one major class hair.We ignore two major classes, skin and back.It can be seen from Fig.2, that PMAPS for minor classes highly differs from one pose to another.For example, considering the fifth row (i.e., PMAPS for the nose), in frontal face images, the nose is more exposed to the camera, and as a result, the nose is almost in the middle of the face image.As pose of the image changes from the center to left (0oto -90o) and right profile (0oto +90o), the PMAPS of the nose also moves accordingly.We use this information as a feature and encode it in a unique feature vector.

    The same difference can be noticed from Fig.2 for four other PMAPS,including eyes,mouth,eyebrows,and hair.In extreme left and right profile images in some cases, the class information is entirely missing.This also clearly shows that our feature-based face pose estimation method highly depends on accurate face parts segmentation.

    Hair class has a very complex geometry that varies from person to person.Our proposed face parsing part reports excellent labeling accuracy for hair.From the segmentation results in Fig.2, it can be observed how efficiently our face parsing module segments a hair class.The borderline for hair is detected by our face segmentation part in a much better way.

    For face pose estimation,we label 200 images from each dataset.We build a face parsing model for each database using 200 manually labeled images.We train a Soft-Max classifier for face parsing.When a test face image is given to the face parsing module,segmentation results are produced,as some of these are shown in Fig.2.We use the probabilistic classification strategy and created PMAPS for each face class.We concatenate the five face classes’ information by first extracting information through CNNs and create a unique feature vector, with which we again train a Soft-Max classifier.In experiments, we adapt 10-fold cross-validation experiments and report average results in the paper.

    6 Results and Discussion

    6.1 Experimental Setup

    We use the Intel i7 CPU for our experiments.We use 16G RAM with NVIDIA 840 M graphical processing unit.We use Tensor flow and Keras as experimental tools.We train our model for 40 Epochs and batch size 150.We keep this setting for all face parsing models developed for all four datasets.

    6.2 Face Segmentation Results

    Some remarks for face parsing results are summarized in the following paragraphs:

    Some qualitative results are shown in Fig.2.The results show that face segmentation is better for the frontal poses as compared to profile.In contrast, labeling accuracy drops as the pose moves to right or left profile, which was expected as well, as minimal information are provided for training in case of extreme profile face images.

    We also observe that as the pose moves to the right or left,labeling accuracy drops particularly for minor classes(nose,brow,eyes,and mouth).For extreme right or left profile face images,in some cases,the minor classes in some cases are completely missing.This can be seen from the images in Fig.2.In such cases,the PMAPS produced are also unclear and the segmentation part provides minimal information.

    The performance of the face parsing part also highly depends on image quality.For example, for low quality images (AFLW), comparatively poor results are reported by our proposed method.While for images from high quality datasets, such as Pointing’04 we obtained better results and also surpassed previously reported results.

    We labeled all face images through image editing software.We used the manually labeled images to build a face parsing model.We used no automatic segmentation tool in all this process.This kind of labeling strictly depends on the subjective perception of a single human involved in manual labeling.To provide accurate face labels to face images with such type of labeling is very difficult.Differentiating boundary regions of face parts in such cases is very difficult; for example, explicitly drawing a boundary region between skin and nose is not accurate.And lastly, this labeling method is very tedious and timeconsuming task.To label large number of images,sufficient time is needed.

    6.3 Face Pose Estimation

    To investigate which face features contribute significantly towards face pose estimation,we exploit the feature importance measure as reported in Pedregosa et al.[48].It is a Random Forest implementation that calculates how certain features contribute to a specific task.Fig.3 shows the importance of each face feature in face pose estimation.From Fig.3,it can be seen that the maximum contribution to face pose estimation is provided by five classes,including eyes,hair,nose,mouth,and brows.Therefore,we use PMAPS of only five features as feature descriptors and discarded the remaining two classes.

    Figure 3:Feature importance of all the seven face classes including nose,hair,mouth,eyes,eyebrow,skin,and back mouth,and nose for face pose estimation

    We evaluate our propose face pose estimation framework with two metrics,mean absolute error(MAE)and pose estimation accuracy (PEA).The MAE is a regression and PEA classification measures.MAE calculates the error between the estimated and actual pose, whereas PEA estimates how accurately a trained model estimates a pose.

    The results obtained with the proposed face pose estimation framework on Poinitng’04 and its comparison with previous results are shown in Tab.3.From Tab.3, it can be seen that better results are obtained with the proposed model as compared to previous results.We explore all combinations of facial features and conclude to use just five classes.We obtain the best results as can be noticed from Tab.3.Some of the previous techniques in Tab.3 may use different validation protocols; for example, some methods use five-fold cross-validation during their experiments, whereas we perform our experiments with ten-fold cross-validation, which is more frequently used in the literature.

    For AFLW, BU, and ICT-3DHPE results are reported only for MAE values.For a fair and exact comparison, we also report MAE values only.A complete summary of the results and then comparison with SOA is shown in Tab.3.The results reveal that we have better results for two datasets BU and ICT-3DHPE.AFLW images are collected from the internet with a very complex background.Most of these images are in wild conditions with poor resolution as well.Our reported results for AFLW are less as compared to previous results.From the segmentation results, we note that face parsing results of the face segmentation module are weak for these complex images.One possible reason for poor results for AFLW,is comparatively poor performance of the face segmentation part.

    The other two datasets,BU and ICT-3DHPE are also collected in real-world conditions.However,the quality of the images is comparatively better,and the background scenario is also not much complex.As a result, we obtain better results as compared to results reported in the literature.

    Table 3:Head pose estimation results with proposed method and its comparison with SOA

    7 Conclusion

    In the proposed work we introduced an end to end face parsing algorithm which tries to address a challenging problem of face pose estimation.We train a face parsing model through DCNNs by extracting useful information from different face parts.The face parsing model provides a class label for each pixel in a face image.We use a probabilistic classification technique and create PMAPS in the form of grey scale images for each face class.We perform a series of experiments to know which face feature helps in face pose estimation and conclude to use only five classes.We evaluate our proposed face pose estimation method with four datasets, including Pointing’04, BU, AFLW, and ICT-3DHPE obtaining better and competitive results.

    Optimization of the face parsing part is one scenario to be addressed in the future.An important point to improve the performance of the face parsing system is by applying carefully well-managed engineering methods.For example,data augmentation[2]and foveated architecture[59]are some possible options to be adapted.Secondly,sufficient information is provided by the face segmentation part to address different visual recognition problems relating to the face.We provide a simple route towards some other complicated face analysis tasks, for example, gesture recognition,face beautification, and many more.

    Funding Statement:This work was partially supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (2020-0-01592) and Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Education under Grant (2019R1F1A1058548) and Grant(2020R1G1A1013221).

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲成av人片免费观看| 亚洲人成网站在线播放欧美日韩| 欧美av亚洲av综合av国产av| 免费观看人在逋| 午夜精品久久久久久毛片777| 精品久久蜜臀av无| 欧美乱色亚洲激情| 亚洲成人精品中文字幕电影| 久久精品人人爽人人爽视色| 亚洲性夜色夜夜综合| 久久精品国产清高在天天线| 精品一品国产午夜福利视频| 欧美绝顶高潮抽搐喷水| 九色国产91popny在线| 国产aⅴ精品一区二区三区波| 一本大道久久a久久精品| 不卡一级毛片| 在线天堂中文资源库| 国产成人精品久久二区二区91| 一级,二级,三级黄色视频| 99香蕉大伊视频| 亚洲国产精品久久男人天堂| 国产精品国产高清国产av| 999久久久精品免费观看国产| 亚洲av成人一区二区三| 欧美中文综合在线视频| 国产97色在线日韩免费| 色精品久久人妻99蜜桃| 久久人妻熟女aⅴ| 国产精品九九99| 欧美激情 高清一区二区三区| 99re在线观看精品视频| 国产三级在线视频| 自线自在国产av| ponron亚洲| 国产一区二区三区视频了| 午夜福利高清视频| 黑丝袜美女国产一区| 人人妻人人澡人人看| 日日夜夜操网爽| 亚洲精品在线美女| 久久香蕉国产精品| 精品一区二区三区四区五区乱码| 亚洲av成人一区二区三| 精品国产一区二区三区四区第35| 一个人观看的视频www高清免费观看 | 亚洲五月天丁香| 欧美午夜高清在线| 桃红色精品国产亚洲av| 黄色毛片三级朝国网站| 国产av在哪里看| 成人免费观看视频高清| 日本撒尿小便嘘嘘汇集6| 国产精品电影一区二区三区| 国产av在哪里看| 91成年电影在线观看| 18禁黄网站禁片午夜丰满| 亚洲色图综合在线观看| 男人舔女人的私密视频| 18禁美女被吸乳视频| 老司机午夜福利在线观看视频| 熟妇人妻久久中文字幕3abv| 国产亚洲精品第一综合不卡| 满18在线观看网站| 日本三级黄在线观看| 国产乱人伦免费视频| 久久精品aⅴ一区二区三区四区| 国产精品 国内视频| 亚洲国产中文字幕在线视频| 日韩中文字幕欧美一区二区| 国产午夜精品久久久久久| 女警被强在线播放| 国产熟女xx| 免费看a级黄色片| 免费看美女性在线毛片视频| 亚洲国产中文字幕在线视频| 亚洲一区二区三区色噜噜| 午夜免费观看网址| 18禁黄网站禁片午夜丰满| 亚洲中文日韩欧美视频| 精品久久久久久,| 亚洲九九香蕉| 国产亚洲欧美98| 一本综合久久免费| 亚洲视频免费观看视频| 中文字幕人妻熟女乱码| 精品国产国语对白av| 亚洲少妇的诱惑av| 99国产精品一区二区蜜桃av| 我的亚洲天堂| av天堂在线播放| 久久中文字幕一级| 亚洲精品国产色婷婷电影| 亚洲欧美日韩另类电影网站| 亚洲熟妇熟女久久| e午夜精品久久久久久久| 亚洲熟女毛片儿| 亚洲一区高清亚洲精品| 成人亚洲精品av一区二区| 一级毛片高清免费大全| 日本黄色视频三级网站网址| 午夜福利一区二区在线看| 午夜亚洲福利在线播放| 久9热在线精品视频| 人妻久久中文字幕网| 国产高清videossex| 亚洲国产欧美网| 身体一侧抽搐| 在线永久观看黄色视频| 中文字幕另类日韩欧美亚洲嫩草| 国产极品粉嫩免费观看在线| 精品国产乱子伦一区二区三区| 午夜激情av网站| 黑人巨大精品欧美一区二区蜜桃| 两个人免费观看高清视频| 国产真人三级小视频在线观看| 色在线成人网| 精品人妻在线不人妻| 欧美午夜高清在线| 国产精品九九99| 日韩欧美免费精品| a在线观看视频网站| 精品熟女少妇八av免费久了| 国产单亲对白刺激| 91精品国产国语对白视频| 女人爽到高潮嗷嗷叫在线视频| 亚洲精品一卡2卡三卡4卡5卡| 久久婷婷成人综合色麻豆| 国产私拍福利视频在线观看| 极品教师在线免费播放| 婷婷丁香在线五月| 亚洲电影在线观看av| 亚洲国产欧美一区二区综合| 黄片播放在线免费| 亚洲精品在线观看二区| 这个男人来自地球电影免费观看| 精品国产乱子伦一区二区三区| 人人妻,人人澡人人爽秒播| 18禁美女被吸乳视频| 校园春色视频在线观看| 老熟妇仑乱视频hdxx| 叶爱在线成人免费视频播放| 亚洲国产中文字幕在线视频| 午夜福利高清视频| 亚洲国产精品成人综合色| 色尼玛亚洲综合影院| 精品国产超薄肉色丝袜足j| 国产一区二区激情短视频| 国产麻豆成人av免费视频| 热re99久久国产66热| 91成人精品电影| 欧美在线一区亚洲| 亚洲精品av麻豆狂野| 天天躁狠狠躁夜夜躁狠狠躁| 成年人黄色毛片网站| 成人精品一区二区免费| 国产蜜桃级精品一区二区三区| 香蕉国产在线看| 人妻丰满熟妇av一区二区三区| 日韩欧美国产一区二区入口| 性色av乱码一区二区三区2| 久久久久精品国产欧美久久久| 欧美亚洲日本最大视频资源| 亚洲av熟女| 日本在线视频免费播放| 亚洲伊人色综图| 国产成人精品无人区| 最新美女视频免费是黄的| 此物有八面人人有两片| 大型av网站在线播放| 国产成人啪精品午夜网站| 亚洲中文字幕一区二区三区有码在线看 | 中国美女看黄片| 久久青草综合色| 此物有八面人人有两片| 桃红色精品国产亚洲av| 国产精品乱码一区二三区的特点 | 日本三级黄在线观看| 一边摸一边做爽爽视频免费| www.999成人在线观看| 日本撒尿小便嘘嘘汇集6| 国产精品免费视频内射| 美女高潮到喷水免费观看| 狂野欧美激情性xxxx| 亚洲国产高清在线一区二区三 | 国产高清有码在线观看视频 | 真人一进一出gif抽搐免费| 在线观看舔阴道视频| 日韩欧美一区视频在线观看| 国产麻豆69| 国产亚洲精品综合一区在线观看 | 久久国产精品男人的天堂亚洲| 这个男人来自地球电影免费观看| 好看av亚洲va欧美ⅴa在| 他把我摸到了高潮在线观看| www国产在线视频色| 中文亚洲av片在线观看爽| 欧美激情极品国产一区二区三区| 色av中文字幕| 亚洲 国产 在线| 大型av网站在线播放| 啦啦啦观看免费观看视频高清 | 久久人妻av系列| 亚洲少妇的诱惑av| 国产精品 国内视频| 老鸭窝网址在线观看| 亚洲成av片中文字幕在线观看| 欧美日韩中文字幕国产精品一区二区三区 | 久久久水蜜桃国产精品网| 99精品欧美一区二区三区四区| 在线观看免费午夜福利视频| 亚洲一区二区三区不卡视频| 亚洲片人在线观看| 真人一进一出gif抽搐免费| 国产精品野战在线观看| 日本欧美视频一区| 亚洲男人天堂网一区| 午夜老司机福利片| 99国产精品99久久久久| 国产三级在线视频| 免费看十八禁软件| 国产成人免费无遮挡视频| 精品久久久久久,| 久久精品国产亚洲av高清一级| 伦理电影免费视频| 成人av一区二区三区在线看| 欧美激情高清一区二区三区| 两人在一起打扑克的视频| 一级毛片女人18水好多| 亚洲精品中文字幕在线视频| 久久香蕉国产精品| 可以在线观看的亚洲视频| 又黄又粗又硬又大视频| 精品一品国产午夜福利视频| 国产精华一区二区三区| 麻豆一二三区av精品| 亚洲精品一区av在线观看| 欧美在线一区亚洲| 91大片在线观看| 麻豆久久精品国产亚洲av| 国内毛片毛片毛片毛片毛片| 校园春色视频在线观看| 日本黄色视频三级网站网址| 国产三级黄色录像| 国产精品美女特级片免费视频播放器 | 制服诱惑二区| 精品国产乱子伦一区二区三区| 自线自在国产av| 免费搜索国产男女视频| 国产片内射在线| 麻豆久久精品国产亚洲av| 午夜亚洲福利在线播放| 国产亚洲欧美精品永久| 精品国产超薄肉色丝袜足j| 欧美黑人欧美精品刺激| 欧美激情高清一区二区三区| 正在播放国产对白刺激| 欧美成人午夜精品| 免费av毛片视频| 亚洲精品国产精品久久久不卡| 一进一出抽搐动态| 在线十欧美十亚洲十日本专区| 亚洲激情在线av| 久久草成人影院| 精品福利观看| 婷婷六月久久综合丁香| a在线观看视频网站| 久久国产亚洲av麻豆专区| avwww免费| 嫁个100分男人电影在线观看| 国产成人啪精品午夜网站| 悠悠久久av| 国内毛片毛片毛片毛片毛片| 日本撒尿小便嘘嘘汇集6| 久热爱精品视频在线9| 午夜福利,免费看| 午夜精品久久久久久毛片777| 久久人人爽av亚洲精品天堂| 婷婷精品国产亚洲av在线| 亚洲欧美一区二区三区黑人| 亚洲成a人片在线一区二区| 一本综合久久免费| 国产又爽黄色视频| 两个人视频免费观看高清| 精品久久蜜臀av无| 高潮久久久久久久久久久不卡| 亚洲av五月六月丁香网| 国产亚洲精品第一综合不卡| 长腿黑丝高跟| 欧美色欧美亚洲另类二区 | 亚洲成人免费电影在线观看| 欧美av亚洲av综合av国产av| 日韩有码中文字幕| 纯流量卡能插随身wifi吗| 亚洲五月婷婷丁香| 精品少妇一区二区三区视频日本电影| 亚洲av熟女| 精品久久久精品久久久| 亚洲精品国产精品久久久不卡| 国产日韩一区二区三区精品不卡| 久久婷婷人人爽人人干人人爱 | 美女午夜性视频免费| 亚洲国产欧美一区二区综合| 午夜免费鲁丝| 亚洲电影在线观看av| 岛国视频午夜一区免费看| 亚洲精华国产精华精| 中文字幕最新亚洲高清| 精品国产超薄肉色丝袜足j| 国产精品久久久人人做人人爽| 后天国语完整版免费观看| 黄色毛片三级朝国网站| 国产在线观看jvid| 国产精品久久视频播放| 夜夜躁狠狠躁天天躁| 伊人久久大香线蕉亚洲五| 国产精品九九99| 三级毛片av免费| 好看av亚洲va欧美ⅴa在| 在线永久观看黄色视频| 99香蕉大伊视频| 国产伦一二天堂av在线观看| 久久狼人影院| 国产色视频综合| 精品卡一卡二卡四卡免费| 一区福利在线观看| 一区在线观看完整版| 欧美性长视频在线观看| 岛国视频午夜一区免费看| 亚洲成a人片在线一区二区| 久热爱精品视频在线9| videosex国产| 成人亚洲精品av一区二区| 天天躁狠狠躁夜夜躁狠狠躁| 大码成人一级视频| 搡老岳熟女国产| av片东京热男人的天堂| 18禁观看日本| 免费在线观看亚洲国产| 午夜影院日韩av| 国产成人一区二区三区免费视频网站| 亚洲专区中文字幕在线| x7x7x7水蜜桃| 村上凉子中文字幕在线| 久久久久国产一级毛片高清牌| 女人高潮潮喷娇喘18禁视频| 成人国语在线视频| 美女免费视频网站| 国产成人啪精品午夜网站| 一a级毛片在线观看| 欧美一级毛片孕妇| 精品国产一区二区三区四区第35| 国产精品久久电影中文字幕| 日韩欧美一区二区三区在线观看| 午夜久久久久精精品| 两个人免费观看高清视频| 午夜成年电影在线免费观看| 久久婷婷人人爽人人干人人爱 | 丰满的人妻完整版| 亚洲天堂国产精品一区在线| 国产成年人精品一区二区| 人人澡人人妻人| 亚洲成人国产一区在线观看| 精品第一国产精品| 99在线人妻在线中文字幕| 99精品在免费线老司机午夜| 91成年电影在线观看| 色播在线永久视频| 日本免费a在线| 欧美绝顶高潮抽搐喷水| 亚洲黑人精品在线| 久久久久国产精品人妻aⅴ院| 91av网站免费观看| 超碰成人久久| 两个人视频免费观看高清| 亚洲国产精品合色在线| 久久久久久久久久久久大奶| 亚洲人成网站在线播放欧美日韩| 制服诱惑二区| 亚洲 国产 在线| 精品一区二区三区四区五区乱码| 久久国产乱子伦精品免费另类| 国产亚洲欧美98| 无遮挡黄片免费观看| 精品久久蜜臀av无| 男女做爰动态图高潮gif福利片 | 国产一区在线观看成人免费| √禁漫天堂资源中文www| 少妇被粗大的猛进出69影院| 亚洲情色 制服丝袜| 日本精品一区二区三区蜜桃| 亚洲色图综合在线观看| 黑丝袜美女国产一区| 午夜福利影视在线免费观看| cao死你这个sao货| 久热爱精品视频在线9| a在线观看视频网站| 在线观看www视频免费| 后天国语完整版免费观看| 国产精品综合久久久久久久免费 | xxx96com| 国产真人三级小视频在线观看| 麻豆av在线久日| 不卡av一区二区三区| 欧美日本中文国产一区发布| 99久久综合精品五月天人人| 12—13女人毛片做爰片一| 久久精品91无色码中文字幕| 久久精品国产综合久久久| 中文字幕人妻丝袜一区二区| 欧美另类亚洲清纯唯美| 国产三级在线视频| 在线观看免费午夜福利视频| 亚洲熟女毛片儿| 亚洲一码二码三码区别大吗| 真人做人爱边吃奶动态| 久久中文看片网| 欧美中文日本在线观看视频| 大型黄色视频在线免费观看| 在线观看免费日韩欧美大片| 亚洲午夜精品一区,二区,三区| 91成年电影在线观看| 怎么达到女性高潮| 日韩欧美一区二区三区在线观看| 黄色a级毛片大全视频| 啦啦啦 在线观看视频| 国产日韩一区二区三区精品不卡| 国产精品二区激情视频| 色播在线永久视频| 大型av网站在线播放| 精品少妇一区二区三区视频日本电影| 99在线视频只有这里精品首页| 97碰自拍视频| 成人亚洲精品av一区二区| 欧美午夜高清在线| 人妻丰满熟妇av一区二区三区| 久久久久久久午夜电影| 国产伦一二天堂av在线观看| 亚洲五月天丁香| 妹子高潮喷水视频| 久久久精品国产亚洲av高清涩受| 99re在线观看精品视频| 九色国产91popny在线| 国产一区二区在线av高清观看| 每晚都被弄得嗷嗷叫到高潮| 女警被强在线播放| 国产精品久久久av美女十八| 国产精品永久免费网站| 一二三四社区在线视频社区8| 亚洲中文日韩欧美视频| 国产成人精品无人区| 国产视频一区二区在线看| 一级作爱视频免费观看| 免费看a级黄色片| 美女扒开内裤让男人捅视频| 国产精品二区激情视频| 欧美乱码精品一区二区三区| 久久亚洲真实| 亚洲一区二区三区色噜噜| 很黄的视频免费| 亚洲国产精品sss在线观看| 成人18禁在线播放| 日韩三级视频一区二区三区| 两性午夜刺激爽爽歪歪视频在线观看 | 国产精品久久久久久亚洲av鲁大| 大码成人一级视频| 88av欧美| 女人爽到高潮嗷嗷叫在线视频| 啪啪无遮挡十八禁网站| 每晚都被弄得嗷嗷叫到高潮| 少妇的丰满在线观看| 日韩欧美一区二区三区在线观看| 国产激情欧美一区二区| 一夜夜www| 国产精品亚洲一级av第二区| 免费在线观看黄色视频的| 久久久国产欧美日韩av| 丰满人妻熟妇乱又伦精品不卡| 欧美日本中文国产一区发布| 欧美绝顶高潮抽搐喷水| 久久性视频一级片| 国产三级在线视频| 国产精品一区二区在线不卡| or卡值多少钱| 亚洲欧美精品综合一区二区三区| 国产熟女午夜一区二区三区| 成人三级黄色视频| 亚洲情色 制服丝袜| 黑丝袜美女国产一区| 美女国产高潮福利片在线看| 麻豆国产av国片精品| 女人高潮潮喷娇喘18禁视频| 亚洲精品国产色婷婷电影| 国产一区二区激情短视频| 色婷婷久久久亚洲欧美| 波多野结衣一区麻豆| 无遮挡黄片免费观看| 中文字幕最新亚洲高清| 天天添夜夜摸| 国产精品1区2区在线观看.| 亚洲人成电影免费在线| 国产xxxxx性猛交| 精品一区二区三区av网在线观看| 免费看美女性在线毛片视频| 成人特级黄色片久久久久久久| 免费在线观看亚洲国产| 色综合亚洲欧美另类图片| 一区二区三区国产精品乱码| 国产精品 国内视频| 纯流量卡能插随身wifi吗| 老司机福利观看| 亚洲熟妇中文字幕五十中出| 欧美性长视频在线观看| 好男人在线观看高清免费视频 | 亚洲久久久国产精品| 午夜免费鲁丝| 欧美日韩乱码在线| 国产午夜福利久久久久久| 十分钟在线观看高清视频www| 亚洲视频免费观看视频| 亚洲国产日韩欧美精品在线观看 | 91大片在线观看| 天天添夜夜摸| 亚洲熟女毛片儿| e午夜精品久久久久久久| 1024香蕉在线观看| 搡老熟女国产l中国老女人| 亚洲精华国产精华精| 欧美+亚洲+日韩+国产| 欧美另类亚洲清纯唯美| 国产一区二区激情短视频| 老熟妇乱子伦视频在线观看| 国产欧美日韩一区二区精品| 99re在线观看精品视频| 丁香六月欧美| 中亚洲国语对白在线视频| 纯流量卡能插随身wifi吗| 精品久久久久久成人av| 国产黄a三级三级三级人| 亚洲一区中文字幕在线| 国产极品粉嫩免费观看在线| 在线观看一区二区三区| 国产激情欧美一区二区| 亚洲国产欧美一区二区综合| 国产精品影院久久| 久久久久精品国产欧美久久久| 99国产精品免费福利视频| 亚洲色图综合在线观看| 母亲3免费完整高清在线观看| 自线自在国产av| 亚洲精品国产色婷婷电影| 亚洲性夜色夜夜综合| 夜夜看夜夜爽夜夜摸| 亚洲成国产人片在线观看| 国产精品一区二区三区四区久久 | 动漫黄色视频在线观看| 欧美一级毛片孕妇| 亚洲精品国产精品久久久不卡| 国语自产精品视频在线第100页| 午夜老司机福利片| 91麻豆精品激情在线观看国产| 亚洲一区二区三区不卡视频| 成年人黄色毛片网站| 精品国产一区二区三区四区第35| 日日摸夜夜添夜夜添小说| 免费人成视频x8x8入口观看| 日韩免费av在线播放| 亚洲第一电影网av| 国内毛片毛片毛片毛片毛片| 国产成人欧美| 午夜成年电影在线免费观看| 亚洲性夜色夜夜综合| 妹子高潮喷水视频| 欧美另类亚洲清纯唯美| 国产精品爽爽va在线观看网站 | 嫩草影视91久久| 午夜久久久久精精品| 欧美日韩乱码在线| 国产精品综合久久久久久久免费 | 亚洲第一av免费看| 国产成人免费无遮挡视频| 免费看十八禁软件| 天天躁狠狠躁夜夜躁狠狠躁| 日日爽夜夜爽网站| 这个男人来自地球电影免费观看| 午夜福利18| 日日夜夜操网爽| 久久久久久久精品吃奶| 国产精品影院久久| 国产成人av激情在线播放| 欧美在线一区亚洲| av中文乱码字幕在线| 国内精品久久久久精免费| 免费搜索国产男女视频| 国产男靠女视频免费网站| 日韩大尺度精品在线看网址 | 欧美一区二区精品小视频在线| 在线播放国产精品三级| 国产伦人伦偷精品视频| 少妇粗大呻吟视频| 亚洲国产精品久久男人天堂| 久久人妻福利社区极品人妻图片| 国产一区二区三区视频了| 自拍欧美九色日韩亚洲蝌蚪91| 精品国产美女av久久久久小说| 日韩大尺度精品在线看网址 | 日日爽夜夜爽网站| 极品教师在线免费播放| 色老头精品视频在线观看| 在线观看免费日韩欧美大片| 涩涩av久久男人的天堂| 国产99久久九九免费精品| 欧美成人一区二区免费高清观看 | 国产aⅴ精品一区二区三区波| 两个人视频免费观看高清|