• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    An Intelligent HealthCare Monitoring Framework for Daily Assistant Living

    2022-08-24 06:59:06YazeedYasinGhadiNidaKhalidSulimanAlsuhibanyTamaraalShloulAhmadJalalandJeongminPark
    Computers Materials&Continua 2022年8期

    Yazeed Yasin Ghadi,Nida Khalid,Suliman A.Alsuhibany,Tamara al Shloul,Ahmad Jalal and Jeongmin Park

    1Department of Computer Science and Software Engineering,Al Ain University,Al Ain,15551,UAE

    2Department of Computer Science,Air University,Islamabad,44000,Pakistan

    3Department of Computer Science,College of Computer,Qassim University,Buraydah,51452,Saudi Arabia

    4Department of Humanities and Social Science,Al Ain University,Al Ain,15551,UAE

    5Department of Computer Engineering,Korea Polytechnic University,Siheung-si,Gyeonggi-do,237,Korea

    Abstract: Human Activity Recognition (HAR) plays an important role in life care and health monitoring since it involves examining various activities of patients at homes,hospitals,or offices.Hence,the proposed system integrates Human-Human Interaction(HHI)and Human-Object Interaction(HOI) recognition to provide in-depth monitoring of the daily routine of patients.We propose a robust system comprising both RGB (red,green,blue) and depth information.In particular,humans in HHI datasets are segmented via connected components analysis and skin detection while the human and object in HOI datasets are segmented via saliency map.To track the movement of humans,we proposed orientation and thermal features.A codebook is generated using Linde-Buzo-Gray (LBG) algorithm for vector quantization.Then,the quantized vectors generated from image sequences of HOI are given to Artificial Neural Network (ANN) while the quantized vectors generated from image sequences of HHI are given to K-ary tree hashing for classification.There are two publicly available datasets used for experimentation on HHI recognition:Stony Brook University(SBU)Kinect interaction and the University of Lincoln’s(UoL)3D social activity dataset.Furthermore,two publicly available datasets are used for experimentation on HOI recognition:Nanyang Technological University(NTU)RGB-D and Sun Yat-Sen University(SYSU)3D HOI datasets.The results proved the validity of the proposed system.

    Keywords: Artificial neural network;human-human interaction;humanobject interaction;k-ary tree hashing;machine learning

    1 Introduction

    Recent years have seen an advanced use of multi-vision sensors to attain robustness and highperformance rates while tackling many of the existing challenges in visual recognition systems [1].Moreover,low-cost depth sensors such as Microsoft Kinect [2]are being used extensively ever since their introduction.In comparison with conventional visual systems,depth maps are unaffected by varying brightness and lighting conditions [3]which motivate reflection over a wide variety of applications of Human Activity Recognition (HAR).These applications include assisted living,behaviour understanding,security systems,human-robot interactions,e-health care,smart homes,and others[4].

    To monitor the daily lifecare routine of humans thoroughly,this paper proposes a system that integrates the recognition of Human-Human Interaction (HHI) and Human-Object Interaction(HOI).In the proposed system,the silhouette segmentation of red,green,blue (RGB) and depth images from HHI and HOI datasets is carried out separately.After silhouette segmentation,there is the feature extraction phase which consists of mining two unique features,namely thermal and orientation features.Both HHI and HOI descriptors are combined and processed via Linde-Buzo-Gray(LBG)algorithm for compact vector representation.In the end,K-ary tree hashing is used for the classification of HHI classes,while Artificial Neural Network(ANN)is applied for the classification of HOI classes.

    We have used two publicly available datasets for experimentation on HHI recognition:Stony Brook University(SBU)Kinect interaction and the University of Lincoln’s(UoL)3D social activity datasets.Furthermore,we have used two different publicly available datasets for experimentation on HOI recognition:Nanyang Technological University (NTU) RGB+D and Sun Yat-Sen University(SYSU)3D HOI datasets.

    The main contributions of this paper are:

    ? Developing an efficient way of segmenting human silhouettes from both RGB and depth images via connected components,skin detection,morphological operations,and saliency maps.

    ? Designing a high-performance recognition system based on the extraction of unique orientation and thermal features.

    ? Accurate classification of HHI classes via K-ary tree hashing and HOI classes via ANN.

    The rest of this paper is organized as follows:Section 2 explains and analyzes the research work relevant to the proposed system.Section 3 describes the proposed methodology of the system which involves an extensive pre-classification process.Section 4 describes the datasets used in the proposed work and proves the robustness of the system through different experiments.Section 5 concludes the paper and notes some future works.

    2 Related Work

    The related work can be divided into two subsections including some recently developed recognition systems for both HHI and HOI.

    2.1 HHI Recognition Systems

    In recent years,many RGB-D(red,green,blue,and depth)human-human interaction recognition systems have been proposed [5].Prati et al.[6]proposed a system in which multiple camera views were used to extract features from depth maps using regression learning.However,despite the use of multiple cameras,their system had restricted applicability on large areas and was not robust against occlusion.Q.Ye et al.[7]proposed a system comprising of Gaussian time-phase features using ResNet(Residual Network).A high performance rate was achieved but their system also had a high complexity rate.In [8],Ouyed et al.extracted motion features from the joints of two persons involved in an interaction.They used multinomial kernel logistic regression to evaluate HIR but the system lacked spatiotemporal context for interaction recognition.Moreover,Ince et al.[9]proposed a system based on skeletal joints movement using Haar-wavelet.In this system,some confusion was observed due to the similarities in angles and positions of various actions.Furthermore,Bibi et al.[10]proposed an HIR system with local binary patterns using multi-view cameras.A high confusion rate was observed in similar interactions.

    Yanli et al.[11]proposed an HIR system that benefits from contrastive feature distribution.The authors extracted skeleton-based features and calculated the probability distribution.In[12],Subetha et al.extracted Histogram of Oriented Gradient (HOG) and pyramidal features.Besides,a study in [13]presented an action recognition system based on way-points trajectory,geodesic distance,joints motion and 3D Cartesian-plane features.This system achieved better performance but there was a slight decrease in accuracy due to the factor of silhouette overlapping.In addition to this study,an action representation was performed with shape,spatio-temporal angular-geometric and energy-based features[14].Therefore,a high recognition rate was achieved but the performance of the system is reduced in those environments where human posture changes rapidly.Also,human postures were extracted in [15]using an unsupervised dynamic X-means clustering algorithm.Features were extracted from skeleton joints obtained via depth sensors.The system lacked in the identification of static actions.Waheed et al.[16]generated 3D human postures and obtained their heat kernels to identify key body points.Then they extracted topological and geometric features using these key points.The authors also extracted full body features using CNN[17].In[18],a time interval at which social interaction is performed was detected and spatio-temporal and social features were extracted to track human actions.Moreover,a study in[19]proposed a fusion of multiple sensors.That is,they extracted HOG and statistical features.However,this system only worked on pre-segmented activities.

    2.2 HOI Recognition Systems

    Various methodologies have been adopted by researchers for identifying human activities in the past few years[20].For example,Meng et al.[21]proposed an HOI recognition system based on interjoint and joint-object distances.However,they did not identify the object individually but considered it one of the human body joints.In[22],joint distances that were invariant to the human pose were measured for feature extraction.This system was tested with only one dataset that consists of six simple interactions.Jalal et al.[23]proposed a HAR system based on the frame differentiation technique.In this system,human joints were extracted to record the spatio-temporal movements of humans.Moreover,Yu et al.[24]proposed a discriminative orderlet,i.e.,a middle-level feature that was used for the visual representation of actions.A cascaded HOI recognition system was proposed by Zhou et al.[25].It was a multi-stage architecture in which each stage of HOI was refined and then fed to the next network.In[26],zero-shot learning was used to accurately identify a relationship between a verb and an object.Their system lacked a spatial context and was tested on a simple verb-object pair.All the methodologies mentioned above in related work are either tested on RGB data or represented by a very complex set of features which increases the time complexity of the system.

    Inspired by these approaches and mindful of their limitations,the proposed system has been designed.Because of the high accuracies achieved by systems that used depth sensors or RGB-D data,our system also takes RGB-D input.Different researchers have extracted different types of features from human silhouettes.To make our model unique,we have chosen orientation and thermal features.Moreover,these features are robust against occlusion and rapid posture changes,the two major issues faced by most researchers.Furthermore,the approaches that have used multiple features have increased time complexity.To solve this issue,we used vector quantization.It was also noted that systems tested on only one dataset or limited number of classes fail to prove their general applicability.Therefore,the proposed system has been validated on four large datasets including two HHI and two HOI datasets.

    3 Material and Methods

    This section describes the proposed framework for active monitoring of the daily life of humans.Fig.1 shows the general overview of the proposed system architecture.This architecture is explained in the following sections.

    3.1 Image Normalization and Silhouette Segmentation

    The images in HHI and HOI datasets are first filtered to enhance the image features.Then,a median filter is applied to both RGB and depth image sequences to remove noise[27]using the formula in Eq.(1):

    whereiandjbelong to a windowwa having specified neighborhood centered around the pixels[m,n]in an image.

    3.1.1 Silhouette Segmentation of Human-Human Interactions

    The silhouette segmentation of RGB and depth frames of HHIs is performed separately.At first,connected components are located in an image via 4-connected pixel analysis as given through Eq.(2):

    wherexandyare coordinates of pixelp.After labeling of connected components,a threshold limit that determines the area(height and width)of the human body is specified.Then,a bounding box is drawn on only those labeled components that are within the specified limit.As a result,all the humans in a frame are identified and enclosed in a bounding box.After identification,human skin is detected inside a bounding box via HSV(hue,saturation,value)color model[28].To improve the thresholding process,all the light color intensities (white,skin,and yellow) are converted into black color having an intensity value of0.Then threshold-based segmentation is applied to generate binary silhouettes.Silhouette segmentation of RGB images is shown in Fig.2.

    Figure 2:Silhouette segmentation of an HHI image.(a)Detected humans via connected components,(b)skin color detection of right and left human,(c)binary silhouette,and(d)RGB silhouette

    The segmentation of depth images of HHI is performed via morphological operations[29].The first step is to convert a depth image into a binary image via Otsu’s thresholding method.Then morphological operations of dilation and erosion are applied which result in retaining the contour of key objects in an image.The Canny edge detection technique is then applied to detect the edges of the two people involved in HHI.In the end,the smaller detected objects are removed from the image.

    3.1.2 Silhouette Segmentation of HOI

    Spectral residual saliency map-based segmentation technique[30]is used to segment humans and objects from RGB and depth images.The saliency map is generated by evaluating the log spectrum of images and Fourier transform is used to obtain the frequencyfof the grayscale imagesG(x).The amplitudeA(f)=abs(f)and phase spectrumP(f)=angle(f)are computed through frequency image.Then,spectral residual regions of RGB and depth images are computed from Eq.(3)as follows:

    whereL(f)is the log ofA(f)andh(f)is an averaging filter.After residual regions are computed,saliency mapSis generated from Eq.(4)as follows:

    whereF-1is inverse Fourier transform.In the end,the saliency map is converted into a binary image via the binary thresholding method.After segmenting salient regions from the background,humans and objects are detected separately in an image using K-means algorithm.This process of silhouette segmentation on a depth image is depicted in Fig.3.

    Figure 3:Silhouette segmentation of an HOI image.(a) Original depth image,(b) saliency map,(c)binary silhouette,and(d)segmented depth human and object silhouette

    3.2 Feature Extraction

    The proposed system exploits two types of features:orientation features and thermal features.The details and results of these features are described in the following subsections.

    3.2.1 Orientation Features

    After obtaining human silhouettes,fourteen human joints (head,neck,right shoulder,left shoulder,right elbow,left elbow,right hand,left hand,torso,torso base,right knee,left knee,right foot,and left foot)are identified.Eight human joints are identified by Algorithm 1(detection of keybody points in human silhouette)proposed in[14].In this algorithm,eight human joints are detected by finding the topmost,bottommost,rightmost and leftmost pixels from the boundary of a human silhouette.The rest of the six human joints are identified by taking the average of the pixel locations of already identified joints.For example,the location of the neck joint is identified by taking the mean of the location of the head and torso joint.After locating joint points,a combination of three joints is taken to form a triangular shape and as a result,fourteen triangles are formed in HOI images(See Fig.4a).In HOI silhouettes,the orientation features of objects are also extracted.Four triangles(twelve angles)are formed from the centroid to all the four extreme points(See Fig.4b).While in HHI images,two people are involved so the number of triangles is twenty-eight(fourteen for each person)as shown in Fig.4c.The angle of tangent is measured between three sides of each triangle from Eq.(5)as follows:

    whereu.vare obtained by taking a dot product of two vectorsu and vwhich are any two sides of a triangle.A total of three angles are calculated for each triangle.The first angle is calculated by taking AB asuand AC asv.The second angle is calculated by taking AB asuand BC asv.The third angle is calculated by taking BC asuand AC asvas shown in Fig.4d.

    Figure 4:Triangular shapes formed by combining human joints.(a) Single person frame of playing phone HOI in SYSU dataset,(b)triangle formation on object,(c)two-person frame of conversation HHI in UoL dataset,and(d)three angles of a triangle

    3.2.2 Thermal Features

    The movements of different body parts as the human silhouettes move from one frame to the next are captured in the form of thermal maps.The parts having greater movements during an interaction have higher heat values (yellowish colors) on these maps.On the other hand,parts of a human silhouette that show lesser movements,i.e.,they are less involved in performing an interaction,are displayed in a reddish or blackish color.A matrix of index values ranging from 0 to 8000 shows heat values in thermal maps.These index values are used to extract heat of only those parts that are involved in an HOI and are represented from Eq.(6)as follows:

    wherevis a 1D vector in which the extracted values are stored,Krepresents index values and lnRrefers to the RGB values that are extracted fromK.The thermal maps of different HOI are shown in Fig.5.

    Figure 5:Thermal maps of HOI and HHI along with the scale of thermal values

    After extracting the two types of features from both HHI and HOI datasets,they are concatenated,resulting in a matrix.

    3.3 Vector Quantization

    After extracting the two types of features from all the images of the HHI and HOI datasets,the features are added as descriptors of each interaction class,separately.However,this results in a very large feature dimension size[31].Therefore,we generate an organized feature vector by considering a codebook of size 512.

    3.4 HHI Classification via K-ary Tree Hashing

    The quantized feature vectors of HHI classes are fed to the K-ary tree hashing classifier.The optimized features are represented in the form of a graphG={gi},wherei=1....NandNrepresents the number of objects in the graph[32].The graph comprises of verticesVand undirected edgesE.Moreover there is a label functionl:V→Lto assign labels to nodes ingiwheregirepresents the whole graph withV,E,andl.A class labelyiis set for eachgibased on graph’s structure.The graph structure means the values in the nodes of the graph based on the values of feature vectors.Each feature vector of the testing class is represented in the form of a graph and then used for predicting the correct label for each class.Also,a size of MinHashes{D(r)}forRiterations and the traversal tableKis also defined.For MinHashes,random permutations {} are also generated.The process of classifying various HHIs is given in Algorithm 2 which takes the graph,the traversal table,and the size of MinHashes as input.It is divided into four sections:traversal table construction,MinHash selection,recursive leaf extension,and leaf sequence.The traversal table is constructed to find the subtree patterns in using k-ary trees.Like a binary tree in which each node has two children,each node in a k-ary tree has k children.The leaf node labels of the k-ary trees can identify the patterns hidden in the data.Then the MinHash scheme is used to classify the interactions based on the identified patterns.

    Algorithm 1:HHI Classification via K-ary tree hashing Input: gi=(V,E,l),K,{D(r)}R r=1//R specifies total iterations//Output:{x(r)}Rr=1//Traversal Table Construction//1 V ←|V|2 l(V+1)←∞3 T ←(V+1)*ones(V+1,1+K)for v=1:V do 4 Nv ←neighbour(v)//MinHash Selection//5 temp ←sort(l(Nv))6 k ←min(K,|Nv|)7 T(v)←[v,index(temp(1:k))]end for//Recursive Leaf Extension//8 z(1) ←[1:V]T 9 S(1) ←l(z(1))for i=1:R do if r>1 then 10 z(r) ←reshape(T(z(r-1),:),[1,*])(Continued)

    Algorithm 1:Continued 11 S(r) ←reshape(l(z(r)),[V,*])end if//Leaf Sequence//12 f (r) ←[h(S(r)(1,:)),...h(S(r)(V,:))]T 13 x(r) ←[min(π(r)1 (f (r))),...min(π(r)D(r)(f (r)))]T end for

    3.5 HOI Classification via Artificial Neural Networks

    The Quantized vectors of HOI are then fed to ANN for training and predicting accurate results.The final vector dimension of the SYSU 3D HOI dataset is 6054×480 while for NTU RGB+D dataset is 6054×530.There are 6054 rows that represent the feature values for both thermal and orientation features.Whereas there are 480 and 530 columns representing the number of images in the SYSU 3D HOI and NTU RGB+D datasets respectively.In the LOSO validation technique,one subset is used for testing and all the remaining subsets are used to train the system.The system is then validated by taking another subset for testing and the remaining subsets for training.In this way,the system is trained and tested with all the subjects in both datasets and avoids sampling bias.There are three layers:input layer,hidden layer,and output layer in ANN[33].These layers are interconnected to each other and weights are associated with each connection.The net input at the neuron of each layer is computed using a transfer functionTjgiven in Eq.(7)as follows:

    wherewi,jare the weights,xirepresents the inputs andbjis the added bias term.An input layer is fed with feature descriptors.After adjusting weights,adding bias,and processing through hidden layer,it predicts accurate HOI classes of both datasets.

    4 Performance Evaluation

    This section gives a brief description of the four datasets used for HHI and HOI,results of the experiments conducted to evaluate the proposed HAR system and its comparison with other systems.

    4.1 Datasets Description

    The description of each dataset used for HHI recognition and HOI recognition is given in Tab.1.Each class of HHI and HOI dataset is performed by different number of subjects as described in the dataset description table.So the proposed system is trained with the different number of subjects of varying appearances resulting in high-performance rate in the testing phase.Each subject is used for both training and testing of a system via the LOSO validation technique.Cross-validation is used to avoid sampling bias via using new image classes for testing of a system other than those used for training.

    Table 1:Datasets description for HHI and HOI recognition

    Table 1:Continued

    4.2 Experimental Settings and Results

    All the processing and experimentation are performed on MATLAB (R2018a).The hardware system used is Intel Core i5 with 64-bit Windows-10.The system has an 8 GB ram and 5 (GHz)CPU.To evaluate the performance of the proposed system,we used a Leave One Subject Out(LOSO)cross-validation method.The results section is divided into two sections:experimental results on HHI datasets and experimental results on HOI datasets.

    4.2.1 Experimental Results on HHI Datasets

    Experiment I:Recognition Accuracies

    At first,classes of SBU and UoL datasets are given to the K-ary tree hashing classifier separately.The results of classification with classes of the SBU and UoL dataset is shown in the form of confusion matrices in Tabs.2 and 3 respectively.

    Experiment II:Precision,Recall and F1 Measures

    The precision is the ratio of correct positive predictions to the total positives while the recall is the true positive rate and it is the ratio of correct positives to the total predicted positives.The F1 score is the mean of precision and recall.The precision,recall and F1 score for classes of SBU and UoL dataset are given in Tabs.4 and 5 respectively.

    Table 2:Confusion matrix showing recognition accuracies over classes of SBU dataset

    Table 3:Confusion matrix showing recognition accuracies over classes of UoL dataset

    Table 4:Precision,Recall and F1 score over classes of SBU dataset

    Table 5:Precision,Recall and F1 score over classes of UoL dataset

    Experiment III:Comparison with Other Systems

    This section compares the proposed methodology with other recent methods as shown in Fig.6.These methods have been discussed in Section 2.

    Figure 6:Comparison of mean recognition accuracy of the proposed method with other recent methods over HHI datasets

    4.2.2 Experimental Results on HOI Datasets Experiment I:Recognition Accuracies

    The results of classification with classes of SYSU and NTU dataset are shown in the form of confusion matrices in Tabs.5,6 and 7,respectively.It is observed during experimentation that the interactions,which involve similar objects like packing backpacks and wearing backpacks,are confused with each other.

    Experiment II:Precision,Recall and F1 Measures

    The precision,recall and F1 scores for classes of the SYSU and the NTU dataset are given in Tabs.8 and 9 respectively.Hence an accurate system is developed which is able to recognize each HOI with high precision.

    Table 6:Confusion matrix showing recognition accuracies over classes of SYSU 3D HOI dataset

    Table 7:Confusion matrix showing recognition accuracies over classes of NTU RGB+D dataset

    Table 8:Precision,recall and F1 score over classes of SYSU dataset

    Table 9:Precision,Recall and F1 score over classes of NTU dataset

    Experiment III:Comparison with Other Systems

    This section compares the proposed methodology over HOI datasets with other recent methods as shown in Fig.7.In[36],a RGB-D HOI system based on joint heterogeneous features based learning was proposed.Also,an RGB-D HOI system based on SIFT regression was proposed in[38].A feature map was constructed by Local Accumulative Frame Feature (LAFF).Furthermore,a study in [39]explained graph regression,whereas multi-modality learning convolutional network was proposed in[40].In[41],the skeletal joints extracted via depth sensors were represented in the form of key poses and temporal pyramids.A mobile robot platform-based HIR was performed in [42]using skeletonbased features.Moreover,the overall human interactions are divided into interactions of different body parts [43].In this work,pairwise features were extracted to track human actions.In [44],the authors introduced a semi-automatic rapid upper limb assessment(RULA)technique using Kinect v2 to evaluate the upper limb motion.

    Figure 7:Comparison of mean recognition accuracy of different recent methods over HOI datasets

    5 Discussion

    A comparison of the proposed system with other systems showed that the proposed system performed better than many other systems proposed in the recent years.Moreover,the high accuracy scores justify the need of additional depth information along with RGB information.Similar findings were also presented in[16]and[17].However,there are some limitations of the systems,such as during skeletal joints extraction,it was challenging to locate the joints of occluded body parts.In order to overcome this limitation,we have adopted the methodology of dividing the silhouette into four halves and then locating the top,bottom,left and right pixels for identifying joints in each half.Moreover,most of the interactions are performed in standing positions in the datasets used in the proposed system.Due to this reason,there is less occlusion of human body parts with objects or other body parts and the performance rate is not very much affected.

    6 Conclusion and Future Works

    This paper proposes a real-time human activity monitoring system that recognizes the daily activities of humans using multi-vision sensors.This system integrates two types of HAR systems:HHI recognition systems and HOI recognition systems.After silhouette segmentation,two unique features are extracted:thermal and orientation features.In order to validate the proposed system’s performance,three types of experiments are performed.The comparison of the proposed system with other state-of-the-art systems is also provided which clearly shows the better performance of the proposed system.In real life,the proposed system should be applicable to many applications such as assisted living,behavior understanding,security systems and human-robot interactions,e-health care and smart homes.

    We are working on integrating more types of human activity recognition and developing a system that monitors human behavior in both indoor and outdoor environments as part of our future works.

    Funding Statement:This research was supported by a grant(2021R1F1A1063634)of the Basic Science Research Program through the National Research Foundation (NRF) funded by the Ministry of Education,Republic of Korea.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲人与动物交配视频| 国产av在哪里看| 国产一区亚洲一区在线观看| 高清欧美精品videossex| 蜜桃亚洲精品一区二区三区| 搞女人的毛片| 日本午夜av视频| 26uuu在线亚洲综合色| 男人舔女人下体高潮全视频| 国产毛片a区久久久久| 一级毛片久久久久久久久女| 在线免费观看的www视频| 免费不卡的大黄色大毛片视频在线观看 | 亚洲色图av天堂| 欧美最新免费一区二区三区| 国产精品久久久久久久电影| av天堂中文字幕网| 寂寞人妻少妇视频99o| 亚洲精品亚洲一区二区| freevideosex欧美| 91av网一区二区| 97超视频在线观看视频| 综合色av麻豆| 搡老妇女老女人老熟妇| 久热久热在线精品观看| 国产成人精品福利久久| 少妇丰满av| 校园人妻丝袜中文字幕| 中文字幕免费在线视频6| 亚洲,欧美,日韩| av播播在线观看一区| 国产亚洲午夜精品一区二区久久 | 老女人水多毛片| 三级经典国产精品| 国产成人免费观看mmmm| 精品欧美国产一区二区三| 免费黄频网站在线观看国产| 免费观看av网站的网址| 国产一区亚洲一区在线观看| 国语对白做爰xxxⅹ性视频网站| 午夜精品国产一区二区电影 | 精品国产一区二区三区久久久樱花 | 免费黄频网站在线观看国产| 我的老师免费观看完整版| 久久这里只有精品中国| 国模一区二区三区四区视频| 十八禁国产超污无遮挡网站| 男的添女的下面高潮视频| 国产国拍精品亚洲av在线观看| 大香蕉97超碰在线| 精品久久久久久久人妻蜜臀av| 中文天堂在线官网| 色哟哟·www| 成人av在线播放网站| 国产日韩欧美在线精品| 欧美精品一区二区大全| 嫩草影院精品99| 免费黄色在线免费观看| 熟妇人妻久久中文字幕3abv| 久久精品国产自在天天线| 一区二区三区四区激情视频| 高清在线视频一区二区三区| 波野结衣二区三区在线| 日韩三级伦理在线观看| 亚洲人与动物交配视频| 天天一区二区日本电影三级| 美女国产视频在线观看| 在线 av 中文字幕| kizo精华| 搡女人真爽免费视频火全软件| 卡戴珊不雅视频在线播放| 午夜福利成人在线免费观看| h日本视频在线播放| 日韩不卡一区二区三区视频在线| 国产在视频线精品| 国产伦精品一区二区三区四那| 午夜福利高清视频| 国产精品女同一区二区软件| 少妇人妻精品综合一区二区| 日本熟妇午夜| 免费电影在线观看免费观看| 乱系列少妇在线播放| 成年女人在线观看亚洲视频 | 久久久久久国产a免费观看| 能在线免费观看的黄片| 80岁老熟妇乱子伦牲交| 国产黄频视频在线观看| 欧美成人一区二区免费高清观看| 国产探花在线观看一区二区| 国产三级在线视频| 欧美成人a在线观看| 久久久久久久久久久丰满| 黄色日韩在线| 少妇的逼好多水| 97精品久久久久久久久久精品| 人人妻人人看人人澡| 国产极品天堂在线| 一级毛片久久久久久久久女| 乱码一卡2卡4卡精品| 亚洲第一区二区三区不卡| 国产成人福利小说| 亚洲精品456在线播放app| 亚洲最大成人中文| 夫妻性生交免费视频一级片| 一边亲一边摸免费视频| 免费看日本二区| 国产精品99久久久久久久久| 91av网一区二区| 午夜福利视频精品| 一级毛片aaaaaa免费看小| 中文字幕av成人在线电影| 综合色av麻豆| 久久精品国产鲁丝片午夜精品| 成年版毛片免费区| 久久人人爽人人爽人人片va| 成人二区视频| 亚洲欧美一区二区三区黑人 | 亚洲国产精品成人综合色| 看免费成人av毛片| 国产探花极品一区二区| 内射极品少妇av片p| 久久热精品热| 精品国产三级普通话版| 色吧在线观看| 成人漫画全彩无遮挡| 午夜福利在线观看吧| 日本欧美国产在线视频| 国产精品一二三区在线看| 啦啦啦韩国在线观看视频| 亚洲激情五月婷婷啪啪| 嫩草影院新地址| 蜜桃久久精品国产亚洲av| 熟女电影av网| 免费看美女性在线毛片视频| 熟妇人妻不卡中文字幕| 中文天堂在线官网| 嫩草影院新地址| 精品人妻一区二区三区麻豆| 日韩精品青青久久久久久| 观看美女的网站| 丝瓜视频免费看黄片| 日本黄大片高清| 欧美另类一区| 嘟嘟电影网在线观看| 人人妻人人澡人人爽人人夜夜 | 午夜精品在线福利| 久久久欧美国产精品| 日韩欧美精品v在线| 嫩草影院入口| 国产在线男女| 高清日韩中文字幕在线| 久99久视频精品免费| 亚洲无线观看免费| 在线免费观看的www视频| 狂野欧美白嫩少妇大欣赏| 日韩,欧美,国产一区二区三区| 精品一区二区免费观看| 麻豆精品久久久久久蜜桃| 久久久久久久久久久免费av| 嫩草影院精品99| 免费黄网站久久成人精品| 免费观看在线日韩| 一夜夜www| 高清在线视频一区二区三区| 国产午夜精品久久久久久一区二区三区| 免费观看无遮挡的男女| 久久精品国产亚洲av天美| 精品亚洲乱码少妇综合久久| 色哟哟·www| 视频中文字幕在线观看| 亚洲精品成人av观看孕妇| 欧美一级a爱片免费观看看| 亚洲成人av在线免费| 久久精品夜夜夜夜夜久久蜜豆| 国产黄片视频在线免费观看| 好男人在线观看高清免费视频| 久久久久久久大尺度免费视频| 成人一区二区视频在线观看| 国产伦在线观看视频一区| 纵有疾风起免费观看全集完整版 | 国国产精品蜜臀av免费| 免费观看的影片在线观看| 直男gayav资源| 国产毛片a区久久久久| 哪个播放器可以免费观看大片| av在线天堂中文字幕| 在线免费观看不下载黄p国产| 黄色欧美视频在线观看| 国产探花极品一区二区| 亚洲国产精品sss在线观看| 我要看日韩黄色一级片| 最近的中文字幕免费完整| 看黄色毛片网站| 人妻夜夜爽99麻豆av| 一级片'在线观看视频| 亚洲最大成人手机在线| 日日摸夜夜添夜夜添av毛片| 亚洲精品色激情综合| 久久精品久久久久久久性| 在线a可以看的网站| 在线免费观看的www视频| 国产精品久久久久久av不卡| 亚洲一级一片aⅴ在线观看| 黄片wwwwww| 日韩精品有码人妻一区| or卡值多少钱| 2021少妇久久久久久久久久久| 成年免费大片在线观看| 久久久a久久爽久久v久久| 亚洲av二区三区四区| 五月伊人婷婷丁香| 亚洲av电影在线观看一区二区三区 | 汤姆久久久久久久影院中文字幕 | 97精品久久久久久久久久精品| 真实男女啪啪啪动态图| 三级男女做爰猛烈吃奶摸视频| 亚洲熟女精品中文字幕| 欧美性猛交╳xxx乱大交人| 亚洲国产色片| 国产日韩欧美在线精品| 国产精品一区二区性色av| 又粗又硬又长又爽又黄的视频| 国产免费视频播放在线视频 | 美女脱内裤让男人舔精品视频| 在线免费观看的www视频| 一二三四中文在线观看免费高清| 久久99热6这里只有精品| 少妇的逼水好多| 国产免费一级a男人的天堂| 免费av毛片视频| 午夜福利视频精品| 成人二区视频| 亚洲人成网站在线观看播放| 亚洲av日韩在线播放| 日日撸夜夜添| 天天躁日日操中文字幕| 国产乱人偷精品视频| 亚洲成色77777| 日日撸夜夜添| 成人无遮挡网站| 少妇的逼水好多| 日日撸夜夜添| 九九爱精品视频在线观看| 国产黄色免费在线视频| 国产精品久久久久久精品电影小说 | 又爽又黄无遮挡网站| 麻豆久久精品国产亚洲av| 日日干狠狠操夜夜爽| 成人午夜精彩视频在线观看| 黑人高潮一二区| 亚洲av不卡在线观看| 麻豆久久精品国产亚洲av| 国产免费视频播放在线视频 | 亚洲高清免费不卡视频| 99re6热这里在线精品视频| 日韩av免费高清视频| 精品99又大又爽又粗少妇毛片| 天天躁夜夜躁狠狠久久av| 一区二区三区乱码不卡18| 伊人久久精品亚洲午夜| 久久久精品94久久精品| 精品久久久噜噜| 久久精品国产自在天天线| 久久精品国产亚洲网站| 两个人的视频大全免费| 国产在视频线精品| 精品国产一区二区三区久久久樱花 | av天堂中文字幕网| 国产免费视频播放在线视频 | 六月丁香七月| 99久久九九国产精品国产免费| 久久久久久久久久黄片| 久久久久久久久久久丰满| 亚洲真实伦在线观看| 国产成人a∨麻豆精品| 成人美女网站在线观看视频| 亚洲欧美一区二区三区黑人 | 欧美97在线视频| 尾随美女入室| 亚洲乱码一区二区免费版| 久久久精品94久久精品| 成人性生交大片免费视频hd| 80岁老熟妇乱子伦牲交| 免费观看在线日韩| 日本猛色少妇xxxxx猛交久久| 高清欧美精品videossex| 日本一本二区三区精品| 国产精品一区二区性色av| 亚洲av免费在线观看| 老司机影院成人| 日韩视频在线欧美| 22中文网久久字幕| 亚洲最大成人av| 午夜激情福利司机影院| 亚洲国产日韩欧美精品在线观看| 少妇被粗大猛烈的视频| 水蜜桃什么品种好| av网站免费在线观看视频 | 白带黄色成豆腐渣| 最近中文字幕高清免费大全6| kizo精华| 国产精品精品国产色婷婷| 可以在线观看毛片的网站| 色综合亚洲欧美另类图片| 亚洲成人一二三区av| 永久网站在线| 日本-黄色视频高清免费观看| 夜夜看夜夜爽夜夜摸| 久久99热这里只有精品18| 卡戴珊不雅视频在线播放| 亚洲图色成人| 最近中文字幕2019免费版| 亚洲一级一片aⅴ在线观看| 狂野欧美白嫩少妇大欣赏| 国产伦理片在线播放av一区| 少妇人妻一区二区三区视频| 日本一二三区视频观看| 国产探花极品一区二区| 国产精品国产三级国产av玫瑰| 国产精品美女特级片免费视频播放器| 午夜视频国产福利| 少妇人妻一区二区三区视频| 超碰av人人做人人爽久久| 国产精品久久久久久久久免| 伦精品一区二区三区| 久久99热这里只有精品18| 欧美zozozo另类| 国产综合懂色| 日韩欧美 国产精品| 精品午夜福利在线看| 麻豆成人午夜福利视频| 中文字幕av在线有码专区| 亚洲精品,欧美精品| 校园人妻丝袜中文字幕| 丝瓜视频免费看黄片| 人体艺术视频欧美日本| xxx大片免费视频| 国产精品嫩草影院av在线观看| 久久精品国产亚洲av涩爱| 国产亚洲最大av| a级一级毛片免费在线观看| 中文精品一卡2卡3卡4更新| 国产高清三级在线| 国产精品一及| 一级爰片在线观看| av又黄又爽大尺度在线免费看| 又爽又黄a免费视频| 亚洲怡红院男人天堂| 内地一区二区视频在线| 国产伦精品一区二区三区视频9| 国产精品久久久久久久久免| 久久精品久久久久久噜噜老黄| 成人二区视频| 五月伊人婷婷丁香| 国产69精品久久久久777片| freevideosex欧美| 超碰97精品在线观看| 中文欧美无线码| 天堂俺去俺来也www色官网 | 在线 av 中文字幕| 欧美精品一区二区大全| 非洲黑人性xxxx精品又粗又长| 男的添女的下面高潮视频| 亚洲欧美中文字幕日韩二区| 亚洲精品乱码久久久久久按摩| 狂野欧美白嫩少妇大欣赏| 午夜福利高清视频| 最近2019中文字幕mv第一页| 日韩精品青青久久久久久| 非洲黑人性xxxx精品又粗又长| 神马国产精品三级电影在线观看| 亚洲一区高清亚洲精品| 哪个播放器可以免费观看大片| 欧美成人精品欧美一级黄| 婷婷色综合大香蕉| 国产91av在线免费观看| 国产成人免费观看mmmm| 午夜精品一区二区三区免费看| 一级片'在线观看视频| 啦啦啦韩国在线观看视频| 麻豆国产97在线/欧美| 成人欧美大片| 听说在线观看完整版免费高清| av在线亚洲专区| 观看美女的网站| 日韩,欧美,国产一区二区三区| 久久97久久精品| 女人久久www免费人成看片| 少妇猛男粗大的猛烈进出视频 | 九九久久精品国产亚洲av麻豆| 亚洲精品乱码久久久v下载方式| av免费观看日本| 中文字幕制服av| 成人国产麻豆网| 亚洲精品国产av蜜桃| 草草在线视频免费看| 极品少妇高潮喷水抽搐| 国产成年人精品一区二区| 国产精品麻豆人妻色哟哟久久 | 久久久久久久久久成人| 色播亚洲综合网| 网址你懂的国产日韩在线| 欧美成人一区二区免费高清观看| 一级毛片黄色毛片免费观看视频| 国产精品嫩草影院av在线观看| 男的添女的下面高潮视频| 午夜激情福利司机影院| 亚洲精品日韩av片在线观看| 伊人久久国产一区二区| 黑人高潮一二区| 亚洲丝袜综合中文字幕| 日韩强制内射视频| 人妻制服诱惑在线中文字幕| 美女内射精品一级片tv| 国内精品美女久久久久久| 国产成人aa在线观看| 亚洲真实伦在线观看| 久久草成人影院| 亚洲欧美成人综合另类久久久| 直男gayav资源| 日韩三级伦理在线观看| 少妇的逼水好多| 一级毛片久久久久久久久女| h日本视频在线播放| 精品久久久久久久末码| 亚洲国产高清在线一区二区三| 国产淫片久久久久久久久| 男人舔奶头视频| 免费看av在线观看网站| 精品久久久久久电影网| 日韩电影二区| 国产综合精华液| 免费无遮挡裸体视频| av国产久精品久网站免费入址| 波野结衣二区三区在线| 男人爽女人下面视频在线观看| 午夜日本视频在线| 成人亚洲精品av一区二区| 久久久成人免费电影| 丝袜喷水一区| 日韩强制内射视频| 少妇高潮的动态图| 午夜免费激情av| 麻豆久久精品国产亚洲av| 秋霞在线观看毛片| 久久精品久久久久久噜噜老黄| 极品少妇高潮喷水抽搐| 国产精品日韩av在线免费观看| 麻豆成人av视频| 69av精品久久久久久| 亚洲成人av在线免费| 哪个播放器可以免费观看大片| a级毛色黄片| 久久精品国产自在天天线| 国产在视频线在精品| 一边亲一边摸免费视频| 欧美激情国产日韩精品一区| 丰满少妇做爰视频| 中文字幕av成人在线电影| 男人狂女人下面高潮的视频| 草草在线视频免费看| 夜夜爽夜夜爽视频| 免费看美女性在线毛片视频| 老司机影院毛片| 亚洲av免费高清在线观看| 国产黄色免费在线视频| 九九爱精品视频在线观看| 亚洲18禁久久av| 在线播放无遮挡| 亚洲国产精品成人综合色| 成年女人看的毛片在线观看| 三级国产精品片| 看非洲黑人一级黄片| 天堂影院成人在线观看| 国产亚洲午夜精品一区二区久久 | 亚洲精品成人久久久久久| 极品少妇高潮喷水抽搐| 综合色av麻豆| 一个人看的www免费观看视频| 禁无遮挡网站| 蜜臀久久99精品久久宅男| 国产亚洲一区二区精品| 亚洲综合色惰| 午夜免费男女啪啪视频观看| 国精品久久久久久国模美| 成年女人看的毛片在线观看| 51国产日韩欧美| 久久国产乱子免费精品| 午夜久久久久精精品| 色综合站精品国产| 国产白丝娇喘喷水9色精品| 好男人在线观看高清免费视频| 国产三级在线视频| 午夜久久久久精精品| 91久久精品国产一区二区三区| 岛国毛片在线播放| av女优亚洲男人天堂| 国产 一区精品| 中文在线观看免费www的网站| 欧美另类一区| h日本视频在线播放| 国产精品一区www在线观看| av免费观看日本| 亚洲av成人精品一区久久| 成人漫画全彩无遮挡| 在线a可以看的网站| 最近2019中文字幕mv第一页| 成人国产麻豆网| 亚洲丝袜综合中文字幕| 亚洲aⅴ乱码一区二区在线播放| 身体一侧抽搐| 女人被狂操c到高潮| 亚洲一级一片aⅴ在线观看| 在线播放无遮挡| 欧美成人一区二区免费高清观看| 午夜日本视频在线| 精品亚洲乱码少妇综合久久| 男女国产视频网站| av天堂中文字幕网| 免费在线观看成人毛片| 亚洲伊人久久精品综合| 少妇的逼水好多| 美女内射精品一级片tv| 亚洲精品日韩在线中文字幕| 国产乱人偷精品视频| 国产精品爽爽va在线观看网站| av福利片在线观看| 久久精品国产亚洲网站| 国产精品一及| 欧美+日韩+精品| 2021少妇久久久久久久久久久| 日本与韩国留学比较| 国产色爽女视频免费观看| 超碰av人人做人人爽久久| 国内精品一区二区在线观看| 国产精品久久久久久精品电影| 国产精品精品国产色婷婷| 精品久久久噜噜| 在线观看一区二区三区| 亚洲成人久久爱视频| 在线观看av片永久免费下载| 乱系列少妇在线播放| 亚洲精品亚洲一区二区| 麻豆乱淫一区二区| 免费高清在线观看视频在线观看| 99热这里只有是精品在线观看| 亚洲av国产av综合av卡| 日韩一区二区三区影片| 亚洲国产精品sss在线观看| 久久久久久久国产电影| 日本wwww免费看| 最近最新中文字幕大全电影3| 亚洲av福利一区| 免费人成在线观看视频色| 狂野欧美激情性xxxx在线观看| av在线亚洲专区| 女人十人毛片免费观看3o分钟| 色网站视频免费| 男插女下体视频免费在线播放| 午夜福利在线观看免费完整高清在| av女优亚洲男人天堂| 亚洲精品日韩在线中文字幕| 建设人人有责人人尽责人人享有的 | 国产在线一区二区三区精| or卡值多少钱| 欧美三级亚洲精品| 成人鲁丝片一二三区免费| 在线免费观看的www视频| 久久这里只有精品中国| 日韩在线高清观看一区二区三区| 春色校园在线视频观看| 亚洲成人中文字幕在线播放| 亚洲丝袜综合中文字幕| 久久午夜福利片| 2021天堂中文幕一二区在线观| 国产成人精品一,二区| 欧美激情在线99| 日韩精品青青久久久久久| 精品亚洲乱码少妇综合久久| 免费大片18禁| 国内精品美女久久久久久| 国产91av在线免费观看| 噜噜噜噜噜久久久久久91| 亚洲电影在线观看av| 亚洲av不卡在线观看| 淫秽高清视频在线观看| av.在线天堂| 亚洲av电影不卡..在线观看| 国产视频首页在线观看| 精品国产露脸久久av麻豆 | 大话2 男鬼变身卡| 久久久久久久大尺度免费视频| 久久久精品94久久精品| 夜夜看夜夜爽夜夜摸| 亚洲国产成人一精品久久久| 亚洲精品国产成人久久av| 欧美日韩综合久久久久久| 麻豆乱淫一区二区| 国产亚洲91精品色在线| 欧美不卡视频在线免费观看| 人妻系列 视频| 亚洲av成人av| 国产男女超爽视频在线观看| 婷婷色av中文字幕| 97在线视频观看| 免费观看在线日韩| 熟妇人妻不卡中文字幕| 18禁在线播放成人免费| 91久久精品电影网| 天堂影院成人在线观看| av在线播放精品| 亚洲激情五月婷婷啪啪| 亚洲av成人精品一区久久| 欧美成人a在线观看| 91aial.com中文字幕在线观看|