• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Automated Facial Expression Recognition and Age Estimation Using Deep Learning

    2022-08-23 02:19:28SyedaAmnaRizwanYazeedYasinGhadiAhmadJalalandKibumKim
    Computers Materials&Continua 2022年6期

    Syeda Amna Rizwan,Yazeed Yasin Ghadi,Ahmad Jalal and Kibum Kim

    1Department of Computer Science,Air University,Islamabad,44000,Pakistan

    2Department of Computer Science and Software Engineering,Al Ain University,Abu Dhabi,122612,UAE

    3Department of Human-Computer Interaction,Hanyang University,Ansan,15588,Korea

    Abstract: With the advancement of computer vision techniques in surveillance systems,the need for more proficient,intelligent,and sustainable facial expressions and age recognition is necessary.The main purpose of this study is to develop accurate facial expressions and an age recognition system that is capable of error-free recognition of human expression and age in both indoor and outdoor environments.The proposed system first takes an input image pre-process it and then detects faces in the entire image.After that landmarks localization helps in the formation of synthetic face mask prediction.A novel set of features are extracted and passed to a classifier for the accurate classification of expressions and age group.The proposed system is tested over two benchmark datasets,namely,the Gallagher collection person dataset and the Images of Groups dataset.The system achieved remarkable results over these benchmark datasets about recognition accuracy and computational time.The proposed system would also be applicable in different consumer application domains such as online business negotiations,consumer behavior analysis,E-learning environments,and emotion robotics.

    Keywords:Feature extraction;face expression model;local transform features and recurrent neural network(RNN)

    1 Introduction

    Recognition of human age and expressions has engaged many researchers in various fields including sustainable security [1], forensics [2], biometrics [2], and cognitive psychology.Interest in this field is spreading fast and is fuelled by scientific advances that provide a better understanding of personal identity,attitudes,and intentions based on facial expressions and age.Facial expressions have a great impact on interpersonal communication.Human emotional responses are very complex and are most directly expressed in facial expressions.In the Mehrabian oral communication effects model, it is stated that 7% intonation, 38% expressions account when the people speak, and 55%body language accounts along with the facial expressions.Over the past few decades,researchers have conducted studies for human facial expressions recognition and age estimation(FERAE)systems that use advanced sensors such as video cameras,eye trackers,thermal cameras,human vision component sensors [3–5], and stereo-cam [6,7] to intelligently recognize the human behaviours, gestures [8–10],emotions and to predict the age of an individual.Problems that arise in automatic FERAE systems are pose variations,uncontrolled lightning,complex backgrounds,partial occlusions,etc.Researchers face many challenges in attempting to overcome these problems.

    Human subjects normally present various expressions all the time in daily life.To develop a sustainable expression recognition and age estimation system, we need to determine whether age estimation is influenced by changes in facial expression, how significant the influence is, and if a solution can be developed to solve the problem caused by facial expressions.Existing works on age estimation are mostly founded on expressionless faces.Most age estimation and expression recognition systems contain mainly frontal-view,neutral expressions,although some used variations in illumination,pose,and expression.To perform a systematic study on age estimation with various expressions,we need to use databases with clear ground truth labels for both age and expression.

    The main distribution of our proposed model is as follows; First, face detection is done using the YCbCr skin color segmentation model.Second, landmark points are plotted on the face based on the connected components technique.Third,Synthetic face a mask is mapped on the face,based on landmark points localization.Fourth, features are extracted and subdivided into two categories.For age estimation, Anthropometric model, energy-based point clouds, and wrinkles are used for feature extraction.For expression recognition, HOG-based symmetry identification, energy-based point clouds,and geodesic distances between landmark points are extracted.Finally,Recurrent Neural Network(RNN)is used for the correct recognition of facial expressions and age.

    The main contributions of the proposed system are:

    · Synthetic face mask mapping increases the multi-face expressions and age recognition accuracy.

    · Our local transform features of both age and expression recognition provide far better accuracy than other state-of-the-art methods.

    · Recurrent Neural Network (RNN) classifier for the accurate age prediction and expressions recognition of individuals.

    Our proposed sustainable FERAE model is evaluated using different performance measures over two multi-face benchmark datasets,namely,the Gallagher collection person dataset and an images of groups dataset which fully validated our system’s efficacy showing that it outperforms other state-ofthe-art methods.

    This article is structured as follows:Section 2 describes related work for both facial expression and age recognition.Section 3 gives a detailed overview of the proposed model that intelligently recognizes multi facial expressions and age.In Section 4, the proposed model performance is experimentally assessed on two publicly available benchmark datasets.Lastly,in Section 5 we sum up the paper and future directions are outlined.

    2 Related Work

    Over the past few years, many researchers have done remarkable work on both single and multi-facial expressions recognition and age estimation.In this section, a comprehensive review of recent related studies of both facial expressions recognition and age estimation models are given in Section 2.1 and 2.2 respectively.

    2.1 Multi-facial Expressions Recognition Systems

    In recent years, many RGB-based facial expressions recognition systems have been proposed.In [11], the authors first detected facial features using Multi-task Cascaded Convolution Neural Network.After that,CNN and a VGG-16 model were used for the classification of facial expressions as Neutral,Positive,or Negative.The facial expression recognition accuracy on the Public dataset was 74%.In[12],the authors developed a system to recognize facial expressions in a variety of social events.Seetaface was used to detect the faces and align them.Visual facial features,i.e.,PHOG,CENTRIST,DCNN features,and VGG features using VGGFace-LSTM and DCNN-LSTM,were then extracted.The system was tested on Group Affect Database 2.0 and achieved recognition accuracy of 79.78%.In[13],a hybrid network was developed in which CNN was used to pretrain the faces and extract scene features,skeleton features,and local features.These fused features were used to predict emotions.The system was tested on a public dataset and achieved system validation and testing accuracies 80.05%and 80.61%respectively.In[14],the authors developed a mood recognition system by first capturing the images from the web cam and trained two machine learning algorithm i-e., Gradient boosting classifier and K-Nearest Neighbor (KNN).The recognition accuracies achieved are 81% and 73%respectively.

    2.2 Multi-facial Age Estimation Systems

    In recent years, different methodologies have been adopted by researchers for the estimation of age or age group.In [15], The authors developed the system to estimate the age of real-life persons.features were extracted via Local Binary Pattern(LBP)and Gabor techniques.For classification,SVM was used.The system was tested on the Images of Group dataset and achieved an accuracy of 87.7%.In [16] extracted the features using LBP and FPLBP techniques.SVM was used for accurate age group classification and achieved an accuracy of 66.6%.In [17] developed a system for automatic classification of age and gender.Features were extracted via Multi-Level Local Binary Pattern(MLLBP)whereas SVM with non-linear RBF kernel was used to classify according to correct age groups and gender.The system was tested on the Images of Group dataset and achieved an accuracy of 43.4%.In[18]proposed a system,the authors extracted the features and classified the correct age group using Convolution Neural Network(CNN).The system was tested on the OUI Adience dataset and achieved an accuracy of 62.34%.

    3 Material and Methods

    This section describes the proposed framework for facial expressions recognition and age estimation.Fig.1 shows the general architecture of the proposed system.

    3.1 Pre-processing and Face Detection

    Our sustainable FERAE system starts with the preprocessing step, which involves two steps;1)background subtraction,and 2)aligning the faces of both datasets at an angle of 180?.First,complex backgrounds are removed from the images to detect the faces more accurately.This is done using the median filtering technique using a 5×5 window to remove the noise and suppress the undesirable distortion in the images.Then,the K means clustering is used for background subtraction.Secondly,if the positions of most of the faces in both datasets are not aligned properly this can be problematic for the detection of faces in the images.Thus,we set the face alignment of both the Gallagher collection person dataset and the Images of Groups dataset using the code available on GitHub[19].

    For face detection,a skin color segmentation technique YCbCr is used.This skin color segmentation model provides remarkable results to detect faces in a scene using YCbCr color space.The skin color of each individual varies,so to get full coverage of each skin pixel the RGB images are converted to YCbCr color space to easily distinguish skin from non-skin pixels.Fig.2 shows the examples of face detection in the Images of Groups dataset.This technique is not affected by the illumination condition Y(luma)factor.Skin representation is based on two componentsCb(blue difference)andCr(red difference).The skin color model is formulated as in Eqs.(1)and(2)[2].

    Figure 2:Some results of pre-processing and face detection over the Images of the group dataset

    3.2 Landmarks Tracking

    Landmarks tracking is the primary step towards face mask mapping.The landmarks are plotted on the facial features to track the pixel positions.They will help us extract different point-based features for the accurate classification of multi-face expressions and age.This section is divided into two subsections;Section 3.2.1 explains landmarks tracking over the Gallagher benchmark dataset for multi-face expressions recognition and Section 3.2.2 describes landmarks tracking over the Images of groups dataset for multi-face age estimation.

    3.2.1 Landmarks Tracking for Multi-face Expressions Recognition

    To plot the landmarks over the Gallagher collection person dataset,the same procedure is used for marking the landmarks on eyebrows,eyes,and lips as mentioned in Section 3.3.1.For the localization of landmarks on the nose.First, the nose is detected using a cascade algorithm.The two nostril points are obtained by applying the concept of connected components inside the bounding box.Then, 3 points are obtained, one on the nose tip and two are on the nostrils.Therefore, a total of 23 landmarks are plotted on the entire face.Figs.3a and 3b show the landmark points symmetry over both benchmark datasets,respectively.

    Figure 3: Landmark points symmetry over the (a) Gallagher collection person dataset and (b) the Images of groups dataset,respectively

    3.2.2 Landmarks Tracking for Multi-face Age Estimation

    After detection of the face,35 landmarks are plotted on the face,on the eyebrows,eyes,and lips,by converting the RGB image into a binary image and detecting the facial features using blob detection.The edges of each facial feature blob are marked with landmarks by taking the central point of each edge using Eq.(3).The nose is detected using the ridge contour method and a total of seven landmark points are marked on the nose.To plot the area of the chin,jawline,and forehead,the midpoints of the face blob or bounding box edges are marked and these are calculated using Eq.(4)[2];

    wherea,b,c,ddenotes the edges length and thee1,e2,e3,ande4are the midpoints of the blob edges.

    3.3 Synthetic Face Mask Prediction

    Synthetic mask prediction is a robust technique for the accurate prediction of the multi-face age of an individual and to recognize the expressions or emotions of a person.This technique is widely used for face detection,face recognition,face aging estimation,etc.For the generation of synthetic masks on the face, we utilized the 35 landmark points for age estimation and 23 landmarks for multi-face expressions recognition.The technique used for both the masks is the same,i.e.,three-sided polygon meshes and perpendicular bisection of a triangle are applied[15].However,the synthetic mask is only generated on facial features for multi-face expression recognition using the sub-triangle formation.The main variations appear on the facial features during changes in facial expressions.Algorithm 1 describes the overall procedure of synthetic face mask prediction over the Gallagher collection person dataset for multi-face expressions recognition.

    Given a face image with 35 and 23 landmarks points over the Images of the Group dataset and the Gallagher collection person dataset images respectively, a multivariate shape model is generated using the landmarks points via polygon meshes and the perpendicular bisection of triangles for age estimation and expression recognition, the large triangles, and sub-triangle formation rule is used.The perpendicular bisections help us distinguish the changes occurring from infancy till adulthood where triangular meshes will further help to extract features for both multi facial expressions and age estimation.Figs.4a and 4b show the synthetic mask prediction over the Images of the Group dataset and the Gallagher collection person dataset,respectively.

    Figure 4:Synthetic face mask prediction over(a)the Images of Groups dataset for age estimation and over(b)the Gallagher collection person dataset for expression recognition respectively

    Algorithm 1:Multi-face expression recognition synthetic mask prediction Input: Input X=Position of 23 landmarks localization points;Output: Mesh of triangles of Y:TM(Y);//initiating feature descriptors matrix//begin 1 Calculate the pixel positions of three outer corners of a triangle(c1,c2,c3);2 for 3 TM(Y):=(c1,c2,c3);(Continued)

    Algorithm 1:Continued 4 /*Initialize TM(Y)a large triangle*/5 The sub-triangles formed inside the TM(Y)is S:=(s1,s2,s3);6 do 7 c1A bisects /c_1;8 c2B bisects /c_2;9 c3C bisects /c_3;10 end for 11 TM(Y)←S;12 return TM(Y);13 end

    3.4 Feature Descriptors

    For the estimation of age and the accurate recognition of facial expressions,we have extracted the age and expression features individually.For age group prediction, the features extraction methods include; 1) Anthropometric model, 2) Interior Angles formulation and, 3) Wrinkles detection (See Section 3.4.1).For expressions recognition,the extracted features are,1)Geodesic distance,2)Energybased point clouds,and 3)0–180?intensity(See Section 3.4.2).

    3.4.1 Feature Extraction for Age Group Classification

    The anthropometric model is the study of the human face and facial features by dimensions and sizes [20].The landmark points marked on the facial features are known by anatomical names e.g.,the lip corners are known as the left and right cheilion and are denoted by lch and rch,likewise,the inner corners of the eyebrows are known as Nasion and are denoted by n.By using this model,we have taken several distances between the facial features which are calculated using the Euclidean distance using Eq.(5)[21].

    wherep1,p2,q1 and q2are the pixel locations along x and y coordinates,respectively.Fig.5 shows the anatomical names and calculated dimensions.

    For the calculation of the interior angles,the above-mentioned face mask is used.From infancy to adulthood the shape of the face mask changes and this results in the variations of angles.We calculated the interior anglesθ1,θ2,θ3using the law of cosine in Eqs.(6),(7)and(8)[21];

    wherep, qandrare the sides of the triangles formed by the face mask.Different measurements of interior angles on two different age groups are shown in Fig.6.

    Figure 5:Anatomical names of the given dimensions

    Figure 6:Interior angle formulations over the Images of groups dataset

    With time,human skin texture changes due to environment,stress,health issues,and many other factors.This texture variation appears in the form of wrinkles,under-eye bags,sagging skin,etc.For wrinkle detection,the Canny edge detection method is used.In Fig.7,the winkles are displayed in the form of edges,i.e.,the white pixels in the binary image over the Images of groups dataset.The quantity of the edges is equal to the number of wrinkles on the face which exhibits the age of the person.These wrinkles are calculated using the Eq.(9)[22];

    whereF,LE,RE,UEandALare the white pixels,i.e.,T1,T2,T3,T4andT5are the total number of pixels on the forehead,left-eyelid,right-eyelid,under-eyes and around the lips.

    Figure 7:The results of wrinkles formation over the images of groups dataset

    3.4.2 Feature Extraction for Facial Expressions Recognition

    The geodesic distance on the surface of the face is the shortest distance between two points.To calculate the geodesic distance,Kimmel and Sethian proposed a method known as fast marching using the Eikonal equation as in Eq.(10)[23];

    The fast-marching algorithm is based on Dijkstra algorithm which computes the shortest distances between two points.In this work,we calculate geodesic distances on the surface of the face using the values of gradient only.Imgis an image having multiple landmark points.To calculate the geodesic distance between two landmark points the distance is(D=d1,d2).The geodesic distance is taken as the parametric manifold which can be represented by mappingF:R2→R3from the parameterization P to the manifold which is given as in Eq.(11);

    The metric tensorgijof the manifold is given as in Eq.(12);

    The geodesic distance is calculated as the shortest distance between the two points.We can calculate the geodesic distance on the surface of the face between 15 landmark points.The geodesic distanceδ(A,B)between two pointsAandBis calculated as in Eq.(13);

    The distance element on the manifold is given as in Eq.(14);

    where the values ofcanddare 1 and 2.We can compute the geodesic distance between 15 landmark points and select the most significant distance that helps in expression recognition.As a result, we obtained a total of 15 distances.

    Energy-based point clouds are the techniques that work on the principle of the Dijkstra algorithm.According to our best knowledge, this technique is used for the very first time for age estimation and expression recognition, simultaneously.This technique is efficient, robust, and quite simple to implement.Using this technique, a central landmark point labeledf∈Fis marked at the center of the face.Its distance is fixed to zero, i.e.,d(f)=0.After that, this value is inserted into a priority queueQ, where the priority is based on the smallest distance between the landmark points.The remaining points are marked asd(q)= ∞.In the priority queue, one point f is selected then the shortest distances between that point to other points are calculated based on the Dijkstra algorithm.Based on those distances,energy-based point clouds are displayed on the face.The alignment of these point clouds changes with variations in the distances from the central point to the other landmark points.The distances from the central point to other varying landmark points are known as optimal distances[24].Fig.8 shows the hierarchical steps for energy-based point clouds extraction.

    Figure 8:The hierarchical steps for energy-based point clouds extraction

    In the 0–180?intensity feature extraction technique,radon transform calculates the projection of an image matrix with some specific axis.The specific axis is used to predict the 2D approximation of the facial expression through different parts of the face using the intensity estimationqalong with the specific set of radial line anglesθdefined as in Eq.(15)[25];

    whereI(q,θ)is the line integral of the image intensity andf(a, b)is the distance from the origin at angleθof the line junction.All these points on a line satisfy Eq.(8)and the projection function can be rewritten as Eqs.(16)and(17)[25];

    Finally, we extracted the top 180 levels of each pixel’s intensity and combined them into a unified vector for different facial expressions.Fig.9 shows the different expression intensity levels(0–180?).

    Our ball can compare favourably20 with the king s, he said, andturned with contempt towards the gazing crowd in the street. What hethought was sufficiently21 expressed in his features and movements: Miserable beggars, who are looking in, you are nothing incomparison to me.

    Figure 9:0–180?intensity levels for different expressions over the Gallagher collection person dataset

    3.5 Long Short-Term Memory Based Recurrent Neural Network(RNN-LSTM)

    Variations in the facial features,while expressions are changing,can exhibit various positions of the facial features.For instance,in the state of sadness,an individual has drooping eyelids,the outer corners of the lips are pulled in a downward direction and very slow eye blinking occurs.In a state of happiness,fast eye blinking and movement of the cheek muscles around the eyes occur and puffiness occurs under the eyes.By comparison with the state of anger, the eyes open widely, eyebrows are drawn together and the lips are tightly closed and become narrower and thin,or the lips are opened to form a rectangle.Similarly,for the prediction of accurate age group classification,changes in facial textures and features occur.In childhood,an individual has more tight skin,no wrinkles on the face,and no under-eye puffiness, whereas, in adulthood, more wrinkles are formed around the eyes, lips,and cheeks, sagging skin and skin color varies.These feature and texture variations are extracted in the form of feature vectors and the Recurrent Neural Network(RNN) takes advantage of them for accurate classification of multi-face expressions and age.

    The feature vectors of expressions and age are fed to the RNN classifier after the features extraction and optimization stage.Our RNN uses one hidden layer along with the 210 unidirectional LSTM fully interconnected cells.The input layer is comprised of 5080 images of the Images of Groups dataset and 589 images of the Gallagher collection person dataset.The vectors size of the Images of groups dataset is 28,231×550 and for the Gallagher collection person dataset it is 931×623 Each features vector is the depiction of the participant facial expressions and age.At the output layer, a SoftMax function,which is responsible for a 1 out K classification task,is used.The SoftMax function output range lies between 0 and 1 where the sum is equal to 1 at every time step.The RNN is trained using the Adam Optimizer with a learning rate of 0.001[26].Fig.12 depicts the hierarchy of RNN for age and expression classification.Algorithm 2 defines the RNN_LSTM training for age estimation and expression recognition.

    Algorithm 2:RNN-LSTM Training Input: Classes ←{“7”,“6”};Features ←{“Age Estimation”,“Expression Recognition”};Output: A ←dataset{n}.Values;B ←dataset{Features}.Values;1 Train_Data,Test_Data,Valid_Data ←Split_Data_Train_Test(A,B,0.33,0.25);2 Size_of_Batch←4;3 RNN_LSTM←Sequential_Model({4 Embedded_Layer(Train_Data.Length,Output_Data_Length,Train_Data.Columns),5 RNN_LSTM_Layer(Output_Data_Length),6 Dense_Layer(Output_Data_Length,activation_Function=‘Sigmoid’)});7 Optimizer←Adam,Epochs←20;8 RNN_LSTM.Compile(Optimizer);9 RNN_LSTM.train(Train_Data,Epochs,Size_of_Batch,Valid_Data);

    4 Performance Evaluation

    This section gives a brief description of two datasets used for facial expressions recognition and age estimation,results of experiments conducted to evaluate the proposed FERAE system and comparison with other systems.

    4.1 Datasets Description

    The description of each dataset used in FERAE system is given in Sections 4.1.1,4.1.2 and 4.1.3.

    4.1.1 The Gallagher Collection Person Dataset for Expression Recognition

    The first dataset used for multi-face expression recognition is the Gallagher Collection Person dataset [27].The images in this dataset were shot in real life at real events of real people with real expressions.The dataset comprises 589 images with 931 faces.Each face is labeled in the image with an expression of Neutral,Happy,Sad,FERAE,Angry and Surprise.The dataset is publicly available.Some examples from this dataset are shown in Fig.10.

    Figure 10:Some examples from the Gallagher collection person dataset

    4.1.2 Images of Groups Dataset for Age Estimation

    The second dataset is the Images of Groups dataset which is used for multi-face age group classification[28].The dataset is the largest dataset comprising 5080 images containing 28,231 faces that are labeled with age and gender.The seven-age group labels of this dataset are 0–2, 3–7, 8–12,13–19, 20–36, 37–65, and 66+.This dataset is publicly available.Some examples of this dataset are shown in Fig.11.

    Figure 11:Some examples from the images of groups dataset

    4.2 Experimental Settings and Results

    All the processing and experimentation are being performed on MATLAB(R2019).The hardware system used is Intel Core i5 with 64-bit Windows-10.The system has 16 GB and 5(GHz)CPUs.To evaluate the performance of the proposed system,we used a Leave One Person Out(LOPO)[29]crossvalidation method.Experiment 1 determined the facial features detection accuracy rates over both benchmark datasets.Experiment 2 determined the multi-face expressions recognition accuracy rates as shown in the form of a confusion matrix.Experiment 3 determined the multi-face age estimation accuracy rates over the Images of groups dataset.Experiments 4 reveal comparisons in a ROC curve graph of the proposed model with another state-of-the-art models for both multi-face expression recognition and age estimation,respectively.

    4.2.1 Experiment 1:Facial Features Detection Accuracies

    In this experiment, facial features detection accuracies over the Images of Groups dataset and Gallagher collection person dataset were determined as shown Fig.12.

    4.2.2 Experiment 2:Multi-face Expressions Recognition Accuracy

    For multi-face expression recognition, the RNN model is used for the accurate classification of expression and age.The Leave One Subject Out (LOSO) cross validation technique is used for the evaluation of the proposed system.Tab.1 shows the confusion matrix of multi-face expressions recognition.

    4.2.3 Experiment 3:Multi-face Age Estimation Accuracy

    For multi-face age estimation,the RNN model was used for the accurate classification of age.The Leave One Subject Out (LOSO) cross validation technique was used for the evaluation of proposed system.Tab.2 shows the confusion matrix for multi-face age estimation.

    Figure 12:Facial features detection accuracies over both benchmark datasets

    Table 1:Confusion matrix for multi-face expressions recognition over the Gallagher person collection dataset

    Table 2: Confusion matrix for multi-face age estimation over the images of groups dataset

    4.2.4 Experiment 4: Results for Comparison of the Proposed Multi-expressions Recognition and Age Estimation Model with Other State of the Art Models.

    Figs.13a–13f, 14a–14f shows the ROC curve graph for all multi-facial expressions and age estimation.The ROC curve is the relationship between the true positive rate and the false positive rate.The true positive is basically showing the sensitivity and false positive rate is showing the 1-specificity.Both the true positive and false positive can be calculated in Eqs.(18)and(19)respectively;

    Figure 13:The ROC curve graphs for all multi-facial expressions over the Gallagher collection person dataset.The lowest and highest values in the expressions ROC curve graphs of both the true positive and false positive using RNN are;Neutral:(0.03,0.00)and(0.80,1.00),Happy:(0.02,0.02)and(0.92,0.98), Sad: (0.14, 0.027) and (0.80, 1.00), Fear: (0.10, 0.00) and (0.81, 0.98), Angry: (0.09, 0.01) and(0.77 and 1.00)and Surprise:(0.12,0.03)and(0.93,0.98)

    We have tested our multi-facial expression recognition and age estimation system (FERAE)model using the state-of the art methods i-e.,Convolution Neural Network(CNN),Recurrent Neural Network(RNN),and Deep Belief Neural Network(DBNN).Experimental Results 4 shows that the RNN along with the other salient feature descriptors of both expression and age provides better results against CNN and DBNN.

    Figure 14:The ROC curve graphs for all the age groups over the Images of groups dataset.The lowest values and highest values in the age groups ROC curve graphs of both the true positive and false positive using RNN are; 0–2: (0.00, 0.00) and (0.90, 1.00), 3–7: (0.18, 0.01) and (0.92, 1.00), 8–12:(0.00, 0.00) and (0.89, 0.99), 13–19: (0.10, 0.00) and (0.90, 1.00), 20–36: (0.00, 0.00) and (0.98 and 0.99)and 37–65:(0.09,0.01)and(0.91,0.97)

    5 Conclusion

    In this paper,a fused model of multi facial expressions recognition and age estimation is proposed.A synthetic face mask is mapped on the face formed by the localization of the landmarks points.The novel point-based and texture-based features obtained using different feature extraction techniques are passed to the RNN classifier for the classification of expressions and age groups The proposed system is tested using the Gallagher collection person dataset for expression recognition and the Images of groups dataset for age estimation.Experimental results show that our approach produced superior classification accuracies i-e.,85.5%over the Gallagher collection person dataset and 91.4%over the images of groups dataset.The proposed system applies to surveillance systems,video gaming,consumer applications, e-learning, audience analysis, and emotion robots.As for limitations, the system fails to detect the detailed facial features of persons from images that are captured too far from the cameras.In the future,we will work on the computational time complexity of the system and also evaluate our system on RGB-D datasets.

    Acknowledgement:This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF), funded by the Ministry of Education (No.2018R1D1A1A02085645).Also,this work was supported by the Korea Medical Device Development Fund grant funded by the Korean government (the Ministry of Science and ICT, the Ministry of Trade,Industry and Energy,the Ministry of Health&Welfare,the Ministry of Food and Drug Safety)(Project Number:202012D05-02).

    Funding Statement:The authors received no specific funding for this study.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    久久精品夜色国产| 亚洲av电影在线观看一区二区三区 | 一级毛片 在线播放| 欧美bdsm另类| 婷婷色麻豆天堂久久| 亚洲av欧美aⅴ国产| 国产熟女欧美一区二区| 亚洲第一区二区三区不卡| 国产白丝娇喘喷水9色精品| 黄色一级大片看看| 国产精品久久久久久久电影| 亚洲精品国产成人久久av| 久久久成人免费电影| 日韩伦理黄色片| 亚洲成人中文字幕在线播放| 熟女电影av网| 亚洲av不卡在线观看| 尾随美女入室| 美女高潮的动态| 黄色怎么调成土黄色| 69人妻影院| 亚洲色图综合在线观看| 久久午夜福利片| 亚洲国产精品成人久久小说| 久久久久精品性色| 日韩一区二区三区影片| 久久久久国产精品人妻一区二区| 亚洲国产色片| 最近的中文字幕免费完整| 国产精品国产三级国产专区5o| 日日啪夜夜爽| 极品少妇高潮喷水抽搐| 久久精品国产a三级三级三级| 久久精品国产a三级三级三级| 观看免费一级毛片| 大又大粗又爽又黄少妇毛片口| 免费av观看视频| 亚洲国产精品999| 九草在线视频观看| 日本熟妇午夜| 国产成人精品久久久久久| 国产精品无大码| 国产视频首页在线观看| a级毛色黄片| 久久久精品免费免费高清| 永久网站在线| 国产精品三级大全| 国产精品国产三级国产专区5o| 久热这里只有精品99| 精品国产一区二区三区久久久樱花 | 日韩欧美一区视频在线观看 | 久久久欧美国产精品| 国产免费一级a男人的天堂| 色吧在线观看| 一区二区三区免费毛片| 尤物成人国产欧美一区二区三区| 夜夜看夜夜爽夜夜摸| www.色视频.com| 91久久精品国产一区二区成人| 婷婷色av中文字幕| av在线亚洲专区| 汤姆久久久久久久影院中文字幕| 舔av片在线| 国产美女午夜福利| 国产伦理片在线播放av一区| freevideosex欧美| 在线观看美女被高潮喷水网站| 三级国产精品片| 国产午夜精品一二区理论片| 国产 一区精品| 偷拍熟女少妇极品色| 少妇 在线观看| 成年女人在线观看亚洲视频 | 欧美3d第一页| 大香蕉久久网| 一级毛片 在线播放| 插阴视频在线观看视频| 精品一区二区三卡| 交换朋友夫妻互换小说| 亚洲精品国产成人久久av| 久久精品夜色国产| 亚洲国产日韩一区二区| 国产精品秋霞免费鲁丝片| 卡戴珊不雅视频在线播放| 如何舔出高潮| 国产精品一区二区在线观看99| 国内揄拍国产精品人妻在线| 免费观看无遮挡的男女| 中文字幕人妻熟人妻熟丝袜美| 成人亚洲精品一区在线观看 | 爱豆传媒免费全集在线观看| 久久午夜福利片| 精品一区二区三区视频在线| 免费黄网站久久成人精品| 97在线人人人人妻| 人人妻人人爽人人添夜夜欢视频 | 国产精品女同一区二区软件| 久久午夜福利片| 在线免费十八禁| 麻豆精品久久久久久蜜桃| 亚洲av不卡在线观看| 国产色爽女视频免费观看| 午夜视频国产福利| 你懂的网址亚洲精品在线观看| 交换朋友夫妻互换小说| 各种免费的搞黄视频| 99九九线精品视频在线观看视频| 色婷婷久久久亚洲欧美| 国产在线一区二区三区精| 成年女人在线观看亚洲视频 | 人妻系列 视频| 高清视频免费观看一区二区| 国产一区二区在线观看日韩| 亚洲四区av| 成人欧美大片| 五月伊人婷婷丁香| 亚洲国产最新在线播放| 午夜免费男女啪啪视频观看| 日本与韩国留学比较| 久久久久久久精品精品| 99热这里只有是精品50| 免费高清在线观看视频在线观看| av黄色大香蕉| 你懂的网址亚洲精品在线观看| 欧美精品国产亚洲| 国产美女午夜福利| 成年版毛片免费区| 精品久久国产蜜桃| 99热这里只有精品一区| 国产免费一级a男人的天堂| 91久久精品国产一区二区三区| 性色av一级| 免费大片黄手机在线观看| 国产精品秋霞免费鲁丝片| 大片免费播放器 马上看| 新久久久久国产一级毛片| 欧美xxⅹ黑人| 色视频在线一区二区三区| 久久久久久久国产电影| 两个人的视频大全免费| 最近中文字幕高清免费大全6| 高清毛片免费看| 欧美三级亚洲精品| 成人漫画全彩无遮挡| 中文字幕亚洲精品专区| 欧美日韩综合久久久久久| 大码成人一级视频| 国产69精品久久久久777片| 大香蕉久久网| 久久久久久久午夜电影| 久久精品国产鲁丝片午夜精品| 寂寞人妻少妇视频99o| a级毛片免费高清观看在线播放| 干丝袜人妻中文字幕| 国产黄片美女视频| 国产亚洲午夜精品一区二区久久 | 色吧在线观看| 亚洲成人一二三区av| 三级男女做爰猛烈吃奶摸视频| 黑人高潮一二区| 97精品久久久久久久久久精品| 欧美+日韩+精品| 亚洲美女视频黄频| 丝袜脚勾引网站| 国产伦在线观看视频一区| 免费播放大片免费观看视频在线观看| 午夜爱爱视频在线播放| 波野结衣二区三区在线| 久久ye,这里只有精品| 国产91av在线免费观看| 亚洲欧美精品自产自拍| 听说在线观看完整版免费高清| 国产爽快片一区二区三区| 肉色欧美久久久久久久蜜桃 | 国产精品.久久久| 亚州av有码| 成年人午夜在线观看视频| 老女人水多毛片| 免费看av在线观看网站| 国产精品一二三区在线看| 日本一本二区三区精品| 国产精品一区二区在线观看99| 亚洲成人精品中文字幕电影| 亚洲国产欧美人成| 免费观看的影片在线观看| 国产91av在线免费观看| 久久鲁丝午夜福利片| 亚洲真实伦在线观看| av又黄又爽大尺度在线免费看| 亚洲高清免费不卡视频| 香蕉精品网在线| 国产精品99久久久久久久久| 国产黄色免费在线视频| 一级毛片电影观看| 人妻少妇偷人精品九色| 午夜精品一区二区三区免费看| 国产伦在线观看视频一区| 成年免费大片在线观看| 午夜精品一区二区三区免费看| 伊人久久精品亚洲午夜| 国产乱人视频| 国产男人的电影天堂91| 日韩免费高清中文字幕av| 久久精品国产亚洲av天美| 国产精品.久久久| 久久精品久久久久久噜噜老黄| 国产av国产精品国产| 美女视频免费永久观看网站| 99re6热这里在线精品视频| 久久99蜜桃精品久久| 99九九线精品视频在线观看视频| 免费在线观看成人毛片| 另类亚洲欧美激情| 国产精品福利在线免费观看| 欧美xxxx黑人xx丫x性爽| 午夜免费男女啪啪视频观看| av又黄又爽大尺度在线免费看| 成人二区视频| 中文字幕av成人在线电影| 少妇猛男粗大的猛烈进出视频 | 爱豆传媒免费全集在线观看| 国产一级毛片在线| 午夜日本视频在线| 亚洲最大成人手机在线| 三级经典国产精品| av一本久久久久| 别揉我奶头 嗯啊视频| 国产永久视频网站| 亚洲综合精品二区| 在线精品无人区一区二区三 | 色婷婷久久久亚洲欧美| 三级男女做爰猛烈吃奶摸视频| 综合色av麻豆| 日韩电影二区| 最近最新中文字幕免费大全7| a级毛色黄片| 免费观看av网站的网址| 亚洲精品影视一区二区三区av| 国产成人午夜福利电影在线观看| 日日撸夜夜添| 欧美成人午夜免费资源| 97人妻精品一区二区三区麻豆| 五月天丁香电影| 99久久九九国产精品国产免费| 少妇的逼好多水| 丝瓜视频免费看黄片| 亚洲精品视频女| 国模一区二区三区四区视频| 尾随美女入室| 男女边吃奶边做爰视频| 97在线人人人人妻| 成人国产麻豆网| 国产中年淑女户外野战色| 色网站视频免费| 听说在线观看完整版免费高清| 久久久久久久亚洲中文字幕| 日韩伦理黄色片| 久久精品夜色国产| 91久久精品电影网| 欧美成人一区二区免费高清观看| 日韩,欧美,国产一区二区三区| 九色成人免费人妻av| 内地一区二区视频在线| av国产免费在线观看| 国产人妻一区二区三区在| 人人妻人人澡人人爽人人夜夜| 久久影院123| tube8黄色片| 欧美人与善性xxx| 五月玫瑰六月丁香| 精品人妻一区二区三区麻豆| 99热国产这里只有精品6| 午夜激情福利司机影院| 欧美精品一区二区大全| 禁无遮挡网站| 国产熟女欧美一区二区| videossex国产| 人人妻人人爽人人添夜夜欢视频 | 91午夜精品亚洲一区二区三区| 日韩免费高清中文字幕av| 黄色配什么色好看| 综合色av麻豆| 欧美变态另类bdsm刘玥| 精品久久久久久久久av| 天堂网av新在线| 成人国产麻豆网| 国内精品宾馆在线| 99热这里只有是精品50| 国产精品久久久久久精品古装| 久热这里只有精品99| 99热国产这里只有精品6| 国产精品熟女久久久久浪| 免费看不卡的av| 99久久精品热视频| 精品人妻熟女av久视频| 欧美激情国产日韩精品一区| 亚洲精品日韩av片在线观看| 久久99热这里只频精品6学生| 中文字幕人妻熟人妻熟丝袜美| 亚洲av欧美aⅴ国产| 久久精品久久精品一区二区三区| 中文字幕人妻熟人妻熟丝袜美| 色网站视频免费| 国产成人精品婷婷| 韩国av在线不卡| 十八禁网站网址无遮挡 | 成人鲁丝片一二三区免费| 国产精品.久久久| 国产男人的电影天堂91| 99热6这里只有精品| 不卡视频在线观看欧美| 亚洲av.av天堂| 国产人妻一区二区三区在| 伦精品一区二区三区| 久久99热这里只有精品18| 日本黄色片子视频| 亚洲欧美精品专区久久| 国产精品国产三级国产专区5o| 欧美精品一区二区大全| 丝袜脚勾引网站| 日本黄大片高清| 18禁裸乳无遮挡免费网站照片| 色综合色国产| 日日摸夜夜添夜夜爱| 亚洲,欧美,日韩| 又粗又硬又长又爽又黄的视频| 老女人水多毛片| 少妇人妻久久综合中文| 亚洲天堂国产精品一区在线| 国产成人91sexporn| 内地一区二区视频在线| 亚洲国产av新网站| 国产乱人偷精品视频| 国内揄拍国产精品人妻在线| 久久精品国产亚洲av天美| 精品久久久久久久久亚洲| 麻豆精品久久久久久蜜桃| 亚洲精品亚洲一区二区| 黄色视频在线播放观看不卡| 在线观看av片永久免费下载| 国产精品99久久99久久久不卡 | 成人二区视频| av免费观看日本| 成年免费大片在线观看| 少妇人妻一区二区三区视频| 国产亚洲一区二区精品| 色网站视频免费| 少妇人妻久久综合中文| 亚洲欧美日韩无卡精品| 超碰av人人做人人爽久久| 欧美丝袜亚洲另类| 亚洲欧美日韩东京热| 少妇人妻久久综合中文| 黄色视频在线播放观看不卡| 九草在线视频观看| 又粗又硬又长又爽又黄的视频| 欧美 日韩 精品 国产| 欧美三级亚洲精品| 国产综合精华液| 国产 一区 欧美 日韩| 免费看a级黄色片| 一级毛片我不卡| 欧美老熟妇乱子伦牲交| 亚洲美女视频黄频| 97精品久久久久久久久久精品| 欧美最新免费一区二区三区| 国产av不卡久久| 一级爰片在线观看| 亚洲av福利一区| 国产免费又黄又爽又色| 国产av不卡久久| 日韩电影二区| 国产视频首页在线观看| 伦精品一区二区三区| 亚洲av一区综合| 91狼人影院| 青青草视频在线视频观看| 欧美人与善性xxx| 亚洲欧洲日产国产| 国产久久久一区二区三区| 色视频www国产| 日韩一区二区视频免费看| 国产黄色免费在线视频| 日韩成人av中文字幕在线观看| av线在线观看网站| 精品人妻偷拍中文字幕| 久久6这里有精品| 久久久a久久爽久久v久久| 国产午夜精品久久久久久一区二区三区| 国产亚洲精品久久久com| 日本熟妇午夜| 国产亚洲午夜精品一区二区久久 | 91午夜精品亚洲一区二区三区| 欧美日韩国产mv在线观看视频 | 国产毛片a区久久久久| 色哟哟·www| 新久久久久国产一级毛片| 国产真实伦视频高清在线观看| 一本一本综合久久| 岛国毛片在线播放| 搡老乐熟女国产| 国产又色又爽无遮挡免| 欧美一区二区亚洲| 五月开心婷婷网| 久久韩国三级中文字幕| 国产有黄有色有爽视频| 色婷婷久久久亚洲欧美| 亚洲av国产av综合av卡| 91精品国产九色| 99热国产这里只有精品6| 别揉我奶头 嗯啊视频| 我的女老师完整版在线观看| 黄色欧美视频在线观看| 久久久久久久午夜电影| 欧美人与善性xxx| 久久6这里有精品| 人人妻人人澡人人爽人人夜夜| 能在线免费看毛片的网站| av.在线天堂| 五月天丁香电影| 中文字幕av成人在线电影| 三级国产精品片| 亚洲精品日韩av片在线观看| 亚洲国产精品国产精品| 国产亚洲91精品色在线| 91精品一卡2卡3卡4卡| 久久人人爽av亚洲精品天堂 | 少妇丰满av| 久久久久国产网址| 最后的刺客免费高清国语| 亚洲欧美一区二区三区黑人 | 春色校园在线视频观看| 边亲边吃奶的免费视频| 色视频www国产| 日韩伦理黄色片| 又爽又黄a免费视频| 亚洲国产欧美在线一区| 在线播放无遮挡| av在线播放精品| 三级经典国产精品| 欧美成人一区二区免费高清观看| 午夜福利网站1000一区二区三区| 国产视频内射| 草草在线视频免费看| 亚洲天堂国产精品一区在线| 国产精品三级大全| 一个人看视频在线观看www免费| 色播亚洲综合网| av.在线天堂| 国产黄色视频一区二区在线观看| 街头女战士在线观看网站| 麻豆精品久久久久久蜜桃| 美女主播在线视频| 观看免费一级毛片| 激情五月婷婷亚洲| 大香蕉久久网| 日韩亚洲欧美综合| 久久99蜜桃精品久久| 99热这里只有精品一区| xxx大片免费视频| 免费看不卡的av| 国产av国产精品国产| 在线天堂最新版资源| 乱系列少妇在线播放| 91aial.com中文字幕在线观看| 国产黄频视频在线观看| 午夜精品国产一区二区电影 | 国产欧美日韩一区二区三区在线 | 亚洲人成网站在线观看播放| 黄色怎么调成土黄色| 国产黄片美女视频| 男人爽女人下面视频在线观看| 国产极品天堂在线| 亚洲久久久久久中文字幕| 国产成年人精品一区二区| 国产成人午夜福利电影在线观看| 最近中文字幕高清免费大全6| 欧美日韩综合久久久久久| 简卡轻食公司| 哪个播放器可以免费观看大片| 亚洲国产精品国产精品| 国国产精品蜜臀av免费| 亚洲国产日韩一区二区| 大香蕉97超碰在线| 日本与韩国留学比较| 别揉我奶头 嗯啊视频| 又粗又硬又长又爽又黄的视频| 国产伦理片在线播放av一区| 久久午夜福利片| av网站免费在线观看视频| 成人鲁丝片一二三区免费| 亚洲伊人久久精品综合| 美女被艹到高潮喷水动态| 校园人妻丝袜中文字幕| 欧美日韩综合久久久久久| 亚洲成人久久爱视频| 97人妻精品一区二区三区麻豆| 涩涩av久久男人的天堂| 久久久精品94久久精品| 色综合色国产| 成人无遮挡网站| 亚洲成色77777| 国产精品精品国产色婷婷| 大片免费播放器 马上看| 久久女婷五月综合色啪小说 | 一级黄片播放器| 国产黄频视频在线观看| 如何舔出高潮| 精品国产三级普通话版| 中文字幕人妻熟人妻熟丝袜美| 亚洲精品色激情综合| 国产探花在线观看一区二区| 男女边吃奶边做爰视频| 日韩在线高清观看一区二区三区| 少妇裸体淫交视频免费看高清| 少妇高潮的动态图| 性插视频无遮挡在线免费观看| 岛国毛片在线播放| 午夜福利视频精品| 精品酒店卫生间| 91久久精品国产一区二区成人| 国产探花极品一区二区| 黄色怎么调成土黄色| 色5月婷婷丁香| 午夜亚洲福利在线播放| 久久精品国产亚洲av涩爱| 99九九线精品视频在线观看视频| 欧美日本视频| 91久久精品国产一区二区三区| 免费av观看视频| 深爱激情五月婷婷| av在线亚洲专区| 欧美日韩视频高清一区二区三区二| 久久精品国产亚洲av涩爱| 免费观看无遮挡的男女| 国产黄片美女视频| 精品久久久精品久久久| 全区人妻精品视频| 亚洲四区av| 亚洲av欧美aⅴ国产| 国产美女午夜福利| 午夜福利在线在线| 大香蕉97超碰在线| 熟女人妻精品中文字幕| 麻豆久久精品国产亚洲av| 国产精品嫩草影院av在线观看| 亚洲av日韩在线播放| av女优亚洲男人天堂| 国产永久视频网站| 下体分泌物呈黄色| 中文字幕av成人在线电影| 一级黄片播放器| 日韩欧美精品v在线| 欧美 日韩 精品 国产| 又大又黄又爽视频免费| av黄色大香蕉| 大码成人一级视频| 午夜免费观看性视频| 中国国产av一级| 在线精品无人区一区二区三 | 舔av片在线| 欧美激情国产日韩精品一区| 久久6这里有精品| 国产老妇女一区| 精品一区二区三区视频在线| 男人舔奶头视频| 嘟嘟电影网在线观看| 久久久久精品久久久久真实原创| videos熟女内射| 高清av免费在线| 亚洲av中文字字幕乱码综合| 国产成人免费无遮挡视频| 日韩一本色道免费dvd| 尤物成人国产欧美一区二区三区| 欧美一区二区亚洲| 国产精品一区二区性色av| 成人亚洲精品av一区二区| 97在线视频观看| 人体艺术视频欧美日本| 女人被狂操c到高潮| 在线观看三级黄色| 黄色配什么色好看| 97精品久久久久久久久久精品| 七月丁香在线播放| 国产精品99久久久久久久久| 久久久久精品性色| 亚洲国产精品专区欧美| 欧美激情国产日韩精品一区| 亚洲欧美中文字幕日韩二区| 九九久久精品国产亚洲av麻豆| 尾随美女入室| 国产片特级美女逼逼视频| 一区二区av电影网| 国产毛片在线视频| 简卡轻食公司| 亚洲av.av天堂| 天天一区二区日本电影三级| 国产精品一区www在线观看| 亚洲美女视频黄频| 国产伦在线观看视频一区| av女优亚洲男人天堂| www.色视频.com| 免费看光身美女| 亚洲人与动物交配视频| 18禁裸乳无遮挡动漫免费视频 | 国产老妇伦熟女老妇高清| 亚洲怡红院男人天堂| 国产精品99久久久久久久久| 亚洲成人久久爱视频| 亚洲一级一片aⅴ在线观看| 国产毛片在线视频| 下体分泌物呈黄色| 七月丁香在线播放| 国产成人精品福利久久| 国产精品一二三区在线看|