• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Human Pose Estimation and Object Interaction for Sports Behaviour

    2022-08-24 12:56:16AyeshaArifYazeedYasinGhadiMohammedAlarfajAhmadJalalShaharyarKamalandDongSeongKim
    Computers Materials&Continua 2022年7期

    Ayesha Arif, Yazeed Yasin Ghadi, Mohammed Alarfaj, Ahmad Jalal, Shaharyar Kamaland Dong-Seong Kim

    1Department of Computer Science, Air University, Islamabad, 44000, Pakistan

    2Department of Computer Science and Software Engineering, Al Ain University, Al Ain, 15551, UAE

    3Department of Electrical Engineering, College of Engineering, King Faisal University, Al-Ahsa, Saudi Arabia

    4Department of IT Convergence Engineering, Kumoh National Institute of Technology, Gumi, Korea

    Abstract: In the new era of technology, daily human activities are becoming more challenging in terms of monitoring complex scenes and backgrounds.To understand the scenes and activities from human life logs, human-object interaction (HOI) is important in terms of visual relationship detection and human pose estimation.Activities understanding and interaction recognition between human and object along with the pose estimation and interaction modeling have been explained.Some existing algorithms and feature extraction procedures are complicated including accurate detection of rare human postures, occluded regions, and unsatisfactory detection of objects, especially small-sized objects.The existing HOI detection techniques are instancecentric (object-based) where interaction is predicted between all the pairs.Such estimation depends on appearance features and spatial information.Therefore, we propose a novel approach to demonstrate that the appearance features alone are not sufficient to predict the HOI.Furthermore, we detect the human body parts by using the Gaussian Matric Model (GMM) followed by object detection using YOLO.We predict the interaction points which directly classify the interaction and pair them with densely predicted HOI vectors by using the interaction algorithm.The interactions are linked with the human and object to predict the actions.The experiments have been performed on two benchmark HOI datasets demonstrating the proposed approach.

    Keywords: Human object interaction; human pose estimation; object detection; sports estimation; sports prediction

    1 Introduction

    In the digital era, technology is the most significant tool to ease daily human life.Artificial Intelligence (AI) is a vast field of technology used in various research developments of expert systems and computer vision.In automated systems, technology has progressed significantly in the last couple of decades towards the computerization of humans in several applications [1].Human object interaction is a vast domain and it has many complexities in artificially intelligent systems.Moreover,in a recent study, psychophysicists declared that the understanding of an image or video in a single glimpse is not easy for humans [2].Although all the social events have been categorized in different fields whereas each field consists of different circumstances.This variety of events requires different kinds of classification between human, object, scene, and background.Therefore, a lot of research has been done in the past and on humans and objects for understanding the events.To detect the human,the authors have considered the human first followed by differentiating it from the background and estimating the pose of the human.After this, they employ object detection and classification techniques.

    Event classification and human-object interaction have been used in many applications such as surveillance systems, railways platforms, airports, and seaports where detection of normal and abnormal events along with detection for real-time data is critical [3].However, there are massive challenges in the way of improving the accuracy of human-object interaction for sports and security agencies that need to identify daily activities [4], office work, gym activities, smart homes and smart hospitals, and understanding of activities in educational institutions.All human activity-based and smart systems need to understand the event and take a decision to arrange activities in a well-organized manner.

    In this article, we proposed a unique method forHOI recognition using object detection via Gaussian Matric Model (GMM) and human pose estimation (HPE).We developed a hybrid approach for pre-processing images including salientmaps, skin detection, HSV plusRGB detection, and extraction of geometrical features.We designed a systemwith the combination ofGaussian matrix model (GMM) and k-means to detect the human skeleton, draw ellipsoids of human body parts, and detect the object by a combination of k-means as well as YOLO.A combination of SK-learn and HOG was used for classification and activity recognition.We used two publically available benchmarks datasets for our model and fully validated our results against other state-of-the-art models.The proposed technique processes the data through four different aspects: unwanted components elimination from the image, extraction of hierarchal features, detection of an object based on some factors, and classifiers.In addition, the proposed methodology has been applied to publically available datasets: the PAMI’09 and the UIUC Sports datasets and obtained significant improvement in activities recognition rate over other state-of-the-art techniques.

    The rest of the paper is organized as follows.Section 2 contains relatedwork, Section 3 presents the architecture of the proposed model.Section 4 describes the performance evaluation of the proposed work, Section 5 discusses the proposed methodology.Section 6 reports related discussions.Section 6 concludes the paper and provides some future directions

    2 Related Works

    In this research article, we discuss human-object interactions using interaction algorithms over sports datasets.

    2.1 Human Pose Estimation

    Various researches have been done using different approaches to improve the labeling of scenes since it causes false recognition [5].Designed the object recognition and representation methods that compare the overall pixels to understand the status of the image.Then, they match the kernel to understand the object properly.By combining two different techniques, the MRG technique and the segmentation tree, they show the contextual relationship with respect to detection of the edges by following the connected components [6].In [7] the adopted an approach is to use depth maps for CRF modeling and system development for scene understanding using less bright images with a simple background.Classification is done on the basis of kernel features [8].By using depth images, detection and localization of objects in 3D and foreground segmentation of RGB-D images are performed by [9].

    2.2 Action Recognition via Inertial Sensor

    In [10] the extended model of a semantic manifold by combining manually local contextual relationships and semantic relationships to classify the events.Many researchers considered the human object interaction detection by using object detection and human body detection.Human and object attention maps in that approach are constructed using contextual appearance features and local encoding.Only detection of object takes place instead of identication, Use of the complex in the term of time and computationally expensive because of the additional neural network inference [11].Similarly; hierarchical segmentation proposed by [12] performed contour detection and boundarydetection techniques onRGB images.After the segmentation, a histogram of oriented gradient (HOG) is obtained along with the combination of deformable part model (DPM) for the object detection.

    3 Material and Methods

    The proposed system based on pre-processes, segmentation and objects detection as initial steps in the input images.After the detection of an object from the image, the human has been detected using salient maps and the pose of the human body has been estimated using GMM [13].After detecting the human and object, we compute the geometrical features from the segmented images including object centroid, object length, object width, and object acquired area.We take the extreme points of the object (extreme left point, extreme right point, topmost point, and bottom-most point) from the centroid of the object.Next, we apply naive Bayes to find the features from both techniques.On the other hand, from the human body, we detect SK features and full-body features from ellipsoids.Then,to symbolize the reduced features, we optimize the features and find human object co-occurrence.We optimize the features and find co-occurrence between human and object.Finally, the last step is the classification of the event.An overview of the proposed system is shown in Fig.1.

    Figure 1: The system architecture of the proposed model of Human Object Interaction (HOI)

    3.1 Preprocessing of the Data

    First of all, we need segmented noticeable regions to detect the human-object interaction.For this,we perform foreground extraction [14] and smart image resizing.We used the Salient maps method to extract the salient regions and salient object detection as shown in Fig.2.Low-Rank and LSDM models extract the saliency maps and also capture tree-structure sparsity with the norm.For efficient results, we make partitions of the images and non-overlapping patches.The input image has been divided intoNpatches {Pi}N(i= 1).

    Figure 2: Steps of human silhouette detection.(a) Result of skin tone detection (b) result of salient region detection and (c) result of smoothing filters and bounding box

    we extract theDdimension features for each patchPiand use a vectorFi∈R Dfor representation.The feature vectorF= {f1,f3,f3,...fn}∈R D×Nis extracted from the matrix representation of the input image.

    Therefore, we have designed an algorithm for the decomposition of feature matrix into(F=L+S), some redundant information (L) as well as some structured salient maps (S) in which,Ω(S) is the norm of s-sparsity-inducing.To regularize and preserve the structures, we use the relevant and latent structures and their relationships.

    3.1.1 Foreground Extraction

    To enhance the silhouette extracted by saliency maps, we perform the segmentation via skin tone detection by using the color-space transformation approach [15] which is achieved by using heuristic thresholds.We extract some skin tone regions by using the YCBCRmodel.The values of the threshold are specified as R= 0.299, G= 0.287 and B= 0.11.

    Through random thresholds of color space transformation, some enhanced regions are extracted.For the chrominance segmentation, we classify the skin and the non-skin regions into different parts precisely.

    3.1.2 Smoothing Threshold

    We applied filters and some manual thresholds on the foreground to make the resultant image accurate.We have applied a manual threshold to fill the holes and region connector for the small region connection.The threshold range of above 30 is considered as 255 and below 30 as 0.

    These are the detection steps of human silhouette from images with 2 different techniques and in Fig.2c both the techniques have been merged.

    3.2 Human Detection

    The segmentation of the human body leads to the extraction of active regions between human and object.We extract the centroid, obtain the boundary, and highlight the boundary points and eight peak points [16].

    3.2.1 Centroid

    After the detection of one human body from an image, the spatial feature detection has been performed and the centroid of the human body has been found, which is a torso point Fig.3a.This helps in detecting the human posture, leading towards finding the action.

    Figure 3: Detection of human body (a) Centroid (b) Boundary (c) Boundary points (d) peak points

    3.2.2 Boundary

    We performed the boundary extraction algorithms to find the boundary and highlight spatial points on the boundary Fig.3c.This leads to estimating the human body posture.

    3.2.3 Peak Points

    After boundary extraction, we find the peak point on the boundary Fig.3d to estimate the human posture

    3.3 Object Segmentation

    The segmentation of objects leads to active regions between humans and objects [17].We detected the objects using YOLO.Image classification along with localization on each matrix has been performed to predict the bounding boxes and probabilities for objects by extracting geometrical features.

    In object segmentation to extract the spatial features, we extract four different extreme points and four features of length, width, centroid, and area of the region.These extracted points and features are in further processing, i.e., the parameter of rectangle or square and area of a rectangle (See Fig.4).For the measuring of the distance between the x and y extreme points, we use Euclidean distance as

    Figure 4: Detection of objects by using K-mean and geometrical features extraction

    where, ||d|| represents the distance between two points by using Euclidean,AxandA′xrepresent the x-coordinates.WhileByandB′yrepresent the y-coordinates.

    3.4 Human Body Parts Detection

    The skeleton model has been used to detect the human silhouette and by using k-means clustering,circles have been drawn on the human body.GMM model has been used for the ellipse fitting on human body poses by usingKnumber of fixed ellipses for performance prediction.GMM Ellipse fitting algorithm is responsible for drawing theKnumber ofEellipses for the best coverage of the targeted region α(E).

    We use skeleton tree by implementing compact representation for all the skeleton branches, we draw all the possiblecircles by using the tangent from the central branch.The compact and the lossless representation uses the medial axis transform (MAT) for the denotation of the centroid and the radius which represents the maximum inscribed circle and their coloring.“V”and“W”are the edges of a node of the graph G = (V,W), where V is endpoint node and W is skeleton segment which is the part of Li∈S, S∈{1,..., |W|} denoting the number of edges.

    where Pij is jth bin histogram of the ith node.The term log|S| represents the overall information of the skeleton as shown in Fig.5.

    Figure 5: Estimation of skeleton for pose estimation by using skeleton model.(a) 2D image, (b) Medial Axis Transform and (c) Centroid of a circle that is tangent from the boundary

    3.5 Object Detection

    We used two novel approaches, Fuzzy C-Means and Random forest, for super-pixels and objects segmentation respectively.

    3.5.1 Clustering

    First of all, clusters are the centroids for each object of the class and then randomly initialized.Feature vectors determine the dimensions of centroids and Euclidian distance is used for assigning the cluster for each object [18].Additionally, the pixels are assigned to the clusters having the minimum distance from the centroid.These clusters are assigned to classes of objects and the object mean is calculated again to check the difference from the previous value, and the process continues until a constant value is obtained.Fig.6 represents the clustering.We used 6 classes of PAMI’09 and 8 activities of the UIUC dataset’s feature vectors divided into 6 and 8 clusters respectively, which is obtained by some primary steps.To calculate the area ? of a rectangle, we use extreme connected points and connect them to make a rectangle for the object segmentation.For the particular rectangle,the area ? is calculated as:

    Figure 6: Similar-pixels clustering for object detection

    where ? represents one side of the perimeter, (Q-P) is one side of the rectangle, and (m-n) is the other side of the rectangle.? symbolizes the area of a rectangle.

    3.5.2 Features Computation Using HOG

    We performed segmentation on the images by using edge detection and preserved the edges and used HOG to improve the accuracy.

    3.6 Features of Human Pose Using GMM

    For the extraction of features, we draw ellipses on the human body using a fixed number of ellipses [19].GMM is performed for computing parameters and best coverage for ellipses.We did this using a two-step method: drawing the ellipses using k-means and fixing the maximum numberK.

    3.6.1 Ellipses Using K-Means

    The central skeleton is the medial axis as used for the tangent drawing on the boundary and is continuously changing.A 16-bit histogram is used for each circle and the radius of each circle is computed.The circular shape is defined using MAT-based histograms.

    3.6.2 Fixing the Maximum Number K

    A GMM model is used to draw the ellipsoids on the different body poses by a fixed number K of ellipsesEof the image and to predict the performed activity accordingly [20].GMM-EM algorithm is responsible for computing the parameters for a fixed numberkof ellipsesEthat has achieved the best coverage α(E) in two steps.Ellipses are devolved using the GMM Ellipses fitting model.We wanted to calculate the parameters of ellipsesEand by fixing them untilKin Algorithm 1 for the best coverage of the region from the silhouette.

    Algorithm 1: Ellipse Fitting Algorithm (EFA)Input: EHS: Extracted Human Silhouettes Input: Binary image I.Output: Set of ellipses E with the lowest IC[S,R] = ShapeSkeleton(I)C = ShapeComplexity(S,R)CC = InitializeEllipse(S,R)K=1 IC*=∞Repeat SCC = SelectHypothesis(k,CC)E = GMM-EM(I,SCC,K)IC = ComputeIC(I,E,C)ICmin = C.log(1-0.99)+ 2.k if IC<AIC*then IC*= IC E*= E End K= K+1 Until K==12

    P∈F×Gis the probability of pixels belonging to the ellipseEiin the model.Ciis the origin ofEiandMiis a positive-definite 2×2 matrix which represents the eccentric orientation ofEi.The Gaussian amplitude is used asAi= 1, so the value of probabilityPi(p) on the boundary is the same for all ellipses.The possibility that a point belongs to an ellipseEiis not dependent on the size of the ellipse.We checked this on both datasets as we fixed the number of ellipses by fixing the value of K with the minimum range of 12.This threshold value of K is only applied to get the fixed number of ellipses as represented in Fig.7.

    Figure 7: Detection of human body parts by using skeleton model and GMM (a) Ellipsoids from the circles on same centroids by covering all possible pixels (b) Sub-regions of GMM and (c) Limit the number of ellipses k = 12

    3.7 Human Object Co-Occurrence

    We apply a 3×3 convolutional layer to make the vector V as interaction vector of size×2.At the interface, we extracted the four possible location points for the human body center based on interaction points and interaction vectors.

    During the training process, the interaction points (IP) and the related human body and the object centroids have fixed geometric structures.While at the inference stage, our generated interaction-points(IP) need to be in grouped-form with the detection of object and their results (bounding-boxes of human-body and object).The points generated by using the interaction points p, center of human h,and center of object o imply a condition on the model: h≈P + v and O≈P - v.

    In Fig.8, it illustrates the interaction points grouping.It has 3 different inputs including the human body/object bounding boxes (green and red), the interaction points (redpoint) extracted from interaction heat maps, and interaction vectors (IV) in (red arrow) at the location of the interaction points (IP).The four corners of the (green) outlines of interaction boxes (red) are obtained by the given interaction points and the un-signed interaction vectors as shown in Eqs.(9)-(11).

    Figure 8: 3×3 convolutional layer is used to examine human-object interaction

    Fig.8 represents the procedure of finding the interaction between human and object.The three inputs,namely,the human body/object bounding-box from the object detection branch,the interaction vector is from the interaction points and it is predicted by the interaction vector branch [21].The current human body/object bounding boxes and their interaction points are regarded as true positive human-object pairs.

    Here,HboxandOboxare the obtained boxes of humans and objects from the human and object detection.ibox is the interaction box and it is generated by combining the interaction points and corresponding interaction vectors.dtl,dtr,dbl, anddbrare four different vectors with lengths of corners in between the interaction boxiboxand the reference boxrbox.dτ is the threshold of vector length set for filtering the negative human-object interaction pairs.The interaction grouping schemes are presented in Algorithm 2.

    For the prediction of interaction vectors, we compared point heat mapsPfrom the ground-truth with the heat maps∧Pof all inter-action points.All of these points are with the Gaussian kernel.We used the modified local loss which is proposed in for balancing the positive and negative values.Where Np represents the number of interaction pointsIPin the concerned image.The α and β points are the hyper-parameter points for the contribution of each point.For the interaction points prediction and interaction vector mapsVprediction.We use the value of the un-signed vectorV′k= (|Vx|k, |Vy|k)at various interaction points’pandkas the ground-truth.After that, theL1loss is used for the related interaction points and here |Vp|krepresents the vector and the point which is predicted by loss l function.

    Algorithm 2: Co-Occurance Human Object Interaction(HOI)Input: Human/Object detector H,O.Interaction points and vectors P,V Human, Object and Interaction threshold Hτ, Oτ, Iτ Output: Final Human Object Interaction box If//Interaction point p makes point set P//Interaction vector v makes set of vectors V for Hbox C Hτ, Oscore C Oτ, p CP do if H scorebox>H τ, O score box>Oτ, p score>pτ then//Interaction box ibox//calculate reference box rbox and Obox if Hbox, Obox, ibox, rbox satisfy condition 2 then Sf←Hscore .Oscore .pscore// output the current HOI Score Sf End if End if End if

    Meanwhile, we used the point heat mapPand ground heat mapP′for the prediction and all of them are defined by the Gaussian kernel.

    In(Eq.(14)),Nprepresents the number of interaction points and αandβ are the parameters that control the contribution of every point.For interaction, we use the vector maps V by using the value of the interaction vector at the interaction pointIPas the ground truth.The interaction vectorVisvi= (|vx|i,|vy|i) and the lossL1is directed for all the corresponding interaction points.

    whereVpirepresents the interaction vectorsVat theIPpoint.The loss function is:

    Here,?mis a weight for all vector loss terms.Here, we simply specify?m= 0.1 for all the experiments.

    4 Experiments and Results

    This section is organized into five sub-sections.First, two benchmark datasets are described in detail.Second, results evaluation is discussed.Third, human pose estimation is discussed.Fourth,estimation of human-object interaction is explained, and fifth, our proposed work is compared with other state-of-the-art advanced deep learning techniques.

    4.1 Datasets Description

    To evaluate the performance of the proposed system, we used images based benchmark datasets,namely, the PAMI’09 sports dataset and the UIUC sports dataset contains vast range of backgrounds and verity of sports.These datasets are further divided into testing sets for experiments and testing purposes.These two datasets have been used to detect human bodies and objects and to find the interactions between humans and objects.Both datasets are further classified into different classes of different sports and activities to recognize the different outdoor and indoor activities.

    4.1.1 PAMI’09 Dataset

    This dataset contains six classes and each class has thirty train, thirty ground truth, and twenty test images.The PAMI2009 dataset contains 480 images with few annotations [22].Each class has 80 images including 30 images for training, 30 for ground truth, and 20 for testing.Each picture has been cataloged with 12 ellipsoids.

    4.1.2 UIUC Dataset

    TheUIUC sports dataset consists of eight sports activities.In each class, there are100-240images.This dataset is comprised of 2000 images, mainly of sportsmen and sportswomen [23].

    4.2 Results Evaluations

    For efficient results, the dataset has been provided to the Gaussian mixture models in batches of classes.To minimize reconstruction errors, we set the number of training samples according to crossvalidation.

    4.2.1 Experiment I: Human Pose Estimation

    The accuracy of human pose estimation has been measured using Euclidean distance from ground truth [22] of the dataset, which is explained in Eq.(15).

    where the ground truth of datasetXis the position of human body parts.D’is the threshold, which is 12, and it is used to measure the accuracy between the ground truth and our model.

    In Tab.1, columns 2 and 4 represent the distances fromthe dataset’s ground truthwhereas columns 3 and 5 show the human body part recognition accuracies over the PAMI’09 andUIUC sports datasets respectively.

    Table 1: Human body key point’s detection accuracy

    Tabs.2 and 3 represent the mean accuracy of both the datasets respectively.

    Table 2: Mean recognition accuracy of PAMI’09 sports dataset

    Table 3: Mean recognition accuracy of UIUC sports dataset

    4.2.2 Experiment II: HOI

    For human-object interaction (HOI) detection and prediction, we use Hourglass as a feature extraction method for pre-training.We randomly initialized the network for generating the interaction points and vectors.During the training of the system, we resize the input images to a resolution of 512×512.Standard data augmentation techniques have been employed and an Adam optimizer has been used for the optimization of the loss function during training.Through the testing phase, we perform the flip augmentation method to get final detections and predictions.Moreover, we use a batch with the size of 30 and a learning rate of 2.5.

    For the detection branch, we go after the previously proposed HOI estimation methods and employ the Faster R-CNN method with the ResNet-50-FPN and pre-train it on the UIUC training dataset.To acquire the bounding-boxes at the inference, we have set the score thresh-hold for the human to be greater than 0.4 and for the object, it is 0.1.When the interaction box is generated by our interaction points and vectors.The generation of interaction systems has taken about 7s.The interaction group we have has the complexity ofO(Nh No Ni), whereNh,No,Niis the number of humans, objects, and interaction points, respectively.In the testing, our grouping scheme is timeefficient and takes less than 2s (<20% of total time).

    4.2.3 Experiment III: Classification of HOI

    By following the standard evaluation and testing methods as performed in [24] to analyze our proposed approach, the results are assembled in the form of role-mean-average precision (mAProle).In role-mean-average precision (mAProle), we apply the HOI model and perform it in a way that if and only if one HOI triplet is rewarded as a true-positive when both of the bounding boxes have IoU intersection-over-union (union of interactions) greater than or equal to 0.5 with the labeled data (ground-truth) [25] and the linked interaction class is accurately classified.Firstly, we compare our proposed technique with other state-of-the-art techniques in the literature.Tab.4 represents the comparison on the PAMI’09 and UIUC dataset.The existing approaches utilize human and object features in multi-stream architecture.

    Table 4: State-of-the-art comparison (in terms of mAProle) on the PAMI’09 and UIUC datasets, our approach by combining the HOI and IG with the mAProle of 53.6

    The work of denoted in Tab.4 as DCA, introduces an interactive network to put on noninteraction suppression and reports with a mAProle of 48.3.Our technique achieves state-of-the-art performance by comparing it to existing techniques with a mAProle of 53.4.Fig.9 shows that our results are improved by comparing them (mAProle of 53.6) in first pre-training our model on PAMI’09 and UIUC datasets and then fine-tuning and pre-training the model on both datasets.

    Figure 9: Some results by using Human Object Interaction (HOI) model

    4.2.4 Experiment IV: Qualitative Analysis of our Proposed System

    Finally, after the classification and recognition of human lifelong activities those are performed in this phase.Tab.5 shows the accuracy of different classes in the form of confusion matrix of PAMI’09 dataset with 90.0% of mean accuracy.This shows the significant improvement and better results from the proposed methodology.

    Table 5: Confusion matrix table on PAMI’09 sports dataset

    After that, the classification and recognition of human activities are performed over the UIUC sports dataset set.Tab.6 shows the accuracy of different classes in the form of confusion matrix of the UIUC sports dataset with 87.71% of mean accuracy, which shows significant improvement and better results from the proposed methodology.

    Table 6: Confusion matrix table on UIUC sports dataset

    5 Conclusion

    We proposed a novel approach to estimate the HOI in images.Our approach refers to HOI estimation as a fundamental problem of research work in which we perform the pose estimation using skeleton model and GMM.After that, we detect the object by combining the features of K-means clustering and YOLO.Moreover, we generate the interaction points and interaction vectors by using key-point detection and pair those interaction points and vectors with the human and object by using the bounding boxes.HOI interaction was performed by using the HOI interaction group method.Through reference boxes and reference vectors, we estimate the interaction.Our experiments are performed on two HOI benchmark sports datasets, PAMI’09 and UIUC.Our approach outperforms state-of-the-art methods on both datasets with the accuracies of 90.0% and 87.71%, respectively.

    In the future, we will extend the interaction vector concept by using multiple vectors from the interaction point to the human body and object to improve the results of our model.We also aim to implement this model in other applications and indoor HOI datasets.

    Funding Statement:This research work was supported by Priority Research Centers Program through NRF funded by MEST (2018R1A6A1A03024003) and the Grand Information Technology Research Center support program IITP-2020-2020-0-01612 supervised by the IITP by MSIT, Korea.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    日韩视频在线欧美| 大香蕉97超碰在线| 99久久中文字幕三级久久日本| 亚洲av日韩在线播放| 亚洲人成网站在线播| 婷婷色综合www| 国产精品免费大片| 久久久久精品久久久久真实原创| 日本-黄色视频高清免费观看| 国产午夜精品一二区理论片| 亚洲精品成人av观看孕妇| 卡戴珊不雅视频在线播放| 久久久精品94久久精品| 一本—道久久a久久精品蜜桃钙片| 3wmmmm亚洲av在线观看| 日本免费在线观看一区| 另类亚洲欧美激情| 中文乱码字字幕精品一区二区三区| 亚洲欧美日韩卡通动漫| 女人精品久久久久毛片| 日本与韩国留学比较| 免费看光身美女| 亚洲成色77777| 欧美人与善性xxx| 亚洲成色77777| 男女啪啪激烈高潮av片| 久久久久久久久大av| 高清av免费在线| 久久久久久久精品精品| 久久国内精品自在自线图片| 亚洲美女搞黄在线观看| 国产精品国产av在线观看| 国产在线视频一区二区| 亚洲国产精品一区二区三区在线| 你懂的网址亚洲精品在线观看| 免费观看av网站的网址| 久久99精品国语久久久| 亚洲国产精品一区二区三区在线| 久久亚洲国产成人精品v| 久热久热在线精品观看| 欧美97在线视频| 国产精品一区二区三区四区免费观看| 这个男人来自地球电影免费观看 | 菩萨蛮人人尽说江南好唐韦庄| 黑人猛操日本美女一级片| 老司机亚洲免费影院| 久久久久久久亚洲中文字幕| 国产熟女欧美一区二区| 啦啦啦在线观看免费高清www| 嫩草影院入口| av不卡在线播放| 91在线精品国自产拍蜜月| 久久综合国产亚洲精品| 国产精品欧美亚洲77777| 搡老乐熟女国产| 国产精品欧美亚洲77777| 大香蕉97超碰在线| 国产在线一区二区三区精| 久久婷婷青草| 男的添女的下面高潮视频| 国产一区二区在线观看日韩| 中国国产av一级| 一级毛片我不卡| 中文字幕av电影在线播放| 亚洲精品日本国产第一区| 国内精品宾馆在线| 久久青草综合色| 亚洲av男天堂| 午夜av观看不卡| 国产欧美亚洲国产| 80岁老熟妇乱子伦牲交| 国产在线一区二区三区精| 少妇精品久久久久久久| 欧美最新免费一区二区三区| 国产午夜精品久久久久久一区二区三区| 欧美日韩精品成人综合77777| 国产成人精品婷婷| 亚洲色图综合在线观看| 亚洲欧美日韩另类电影网站| 亚洲第一av免费看| 国产精品一区二区三区四区免费观看| 亚洲精品日韩av片在线观看| 久久久精品94久久精品| 精品少妇内射三级| 亚洲av综合色区一区| 国产精品麻豆人妻色哟哟久久| 两个人免费观看高清视频| 国产日韩欧美在线精品| 麻豆精品久久久久久蜜桃| 人妻少妇偷人精品九色| 亚洲欧美日韩卡通动漫| av在线老鸭窝| 亚洲精品乱码久久久v下载方式| av女优亚洲男人天堂| 一二三四中文在线观看免费高清| 午夜福利在线观看免费完整高清在| 在线精品无人区一区二区三| 精品久久久噜噜| 欧美一级a爱片免费观看看| 91精品一卡2卡3卡4卡| 欧美三级亚洲精品| 久久久久精品久久久久真实原创| 丰满迷人的少妇在线观看| 国产精品久久久久成人av| 国产精品久久久久成人av| 中文字幕亚洲精品专区| 九草在线视频观看| 欧美精品一区二区免费开放| 九色成人免费人妻av| 99久久综合免费| 日韩免费高清中文字幕av| 在线观看www视频免费| 亚洲欧美色中文字幕在线| 人妻制服诱惑在线中文字幕| 看非洲黑人一级黄片| 亚洲综合色网址| 丝袜美足系列| 欧美精品亚洲一区二区| 免费少妇av软件| 国产成人精品在线电影| 午夜福利,免费看| 赤兔流量卡办理| 免费人成在线观看视频色| 亚洲精华国产精华液的使用体验| 大陆偷拍与自拍| 69精品国产乱码久久久| 欧美性感艳星| 热re99久久精品国产66热6| 成人综合一区亚洲| 欧美xxⅹ黑人| 一级毛片 在线播放| 校园人妻丝袜中文字幕| 超碰97精品在线观看| 欧美丝袜亚洲另类| 久久精品熟女亚洲av麻豆精品| a级毛片在线看网站| 99热网站在线观看| 久久久久久伊人网av| 少妇人妻 视频| 日产精品乱码卡一卡2卡三| 国产精品国产三级专区第一集| 夫妻性生交免费视频一级片| 精品一品国产午夜福利视频| 欧美一级a爱片免费观看看| 在线观看人妻少妇| 九九爱精品视频在线观看| 老女人水多毛片| 一级爰片在线观看| 亚洲人成网站在线观看播放| 黄色欧美视频在线观看| 久久久午夜欧美精品| 韩国av在线不卡| 成年人午夜在线观看视频| 啦啦啦啦在线视频资源| 日日啪夜夜爽| 日韩电影二区| 在线天堂最新版资源| 国产精品 国内视频| 中国三级夫妇交换| 国产亚洲欧美精品永久| 母亲3免费完整高清在线观看 | 亚洲精品乱码久久久久久按摩| 国产av精品麻豆| 国产精品.久久久| 国产精品.久久久| 热99久久久久精品小说推荐| 99国产精品免费福利视频| 人妻一区二区av| 人人妻人人爽人人添夜夜欢视频| 精品人妻在线不人妻| 亚洲图色成人| 人人妻人人添人人爽欧美一区卜| 国产成人91sexporn| 久久人人爽人人爽人人片va| 丰满迷人的少妇在线观看| 老司机影院成人| 妹子高潮喷水视频| 性色av一级| 久久久精品区二区三区| 少妇人妻精品综合一区二区| 99久久综合免费| 亚洲精品国产av成人精品| 亚洲精品国产av成人精品| 少妇人妻精品综合一区二区| 18禁观看日本| 精品亚洲乱码少妇综合久久| 免费黄频网站在线观看国产| 中文字幕免费在线视频6| 亚洲综合色网址| 丰满迷人的少妇在线观看| 亚洲av在线观看美女高潮| 亚洲av欧美aⅴ国产| 久久99热这里只频精品6学生| 乱码一卡2卡4卡精品| 久久热精品热| 久久影院123| 9色porny在线观看| 久久国产亚洲av麻豆专区| 国产精品国产三级国产av玫瑰| 亚洲国产欧美在线一区| 一级片'在线观看视频| 色哟哟·www| 久久久精品94久久精品| 亚洲欧洲日产国产| 欧美另类一区| 国产一区二区在线观看av| 3wmmmm亚洲av在线观看| 欧美亚洲 丝袜 人妻 在线| 男的添女的下面高潮视频| av网站免费在线观看视频| 欧美精品人与动牲交sv欧美| √禁漫天堂资源中文www| 中文乱码字字幕精品一区二区三区| 在线观看www视频免费| 高清在线视频一区二区三区| 国产爽快片一区二区三区| 女性被躁到高潮视频| 中文乱码字字幕精品一区二区三区| 一本大道久久a久久精品| 黑人欧美特级aaaaaa片| 黄色毛片三级朝国网站| 91久久精品国产一区二区成人| 好男人视频免费观看在线| 久久精品久久精品一区二区三区| 国产日韩一区二区三区精品不卡 | 久久国产亚洲av麻豆专区| 国产精品欧美亚洲77777| 日韩成人av中文字幕在线观看| 久久青草综合色| 亚洲av福利一区| 有码 亚洲区| 精品午夜福利在线看| 一级a做视频免费观看| 精品卡一卡二卡四卡免费| 欧美成人精品欧美一级黄| 免费大片黄手机在线观看| 成人手机av| 人人妻人人添人人爽欧美一区卜| 黑人巨大精品欧美一区二区蜜桃 | 婷婷色av中文字幕| 欧美人与善性xxx| 五月伊人婷婷丁香| 一二三四中文在线观看免费高清| 国产高清三级在线| 成人毛片60女人毛片免费| 久久久久视频综合| 建设人人有责人人尽责人人享有的| 热re99久久国产66热| 特大巨黑吊av在线直播| 97在线人人人人妻| 熟女人妻精品中文字幕| 国产亚洲最大av| 99精国产麻豆久久婷婷| 狂野欧美激情性xxxx在线观看| 国产成人午夜福利电影在线观看| 国产黄频视频在线观看| 国产黄片视频在线免费观看| 另类精品久久| 国产高清三级在线| 精品少妇黑人巨大在线播放| 国产欧美亚洲国产| 一级,二级,三级黄色视频| 啦啦啦在线观看免费高清www| 妹子高潮喷水视频| 国产白丝娇喘喷水9色精品| 免费不卡的大黄色大毛片视频在线观看| 美女福利国产在线| videos熟女内射| 免费观看在线日韩| 国产男女内射视频| 中文乱码字字幕精品一区二区三区| 亚洲精品456在线播放app| 国产精品麻豆人妻色哟哟久久| a级片在线免费高清观看视频| 人人妻人人澡人人爽人人夜夜| 亚洲精品自拍成人| 黄色毛片三级朝国网站| 日本91视频免费播放| 美女大奶头黄色视频| 视频在线观看一区二区三区| 国产免费福利视频在线观看| 久久女婷五月综合色啪小说| 在线精品无人区一区二区三| 国产又色又爽无遮挡免| h视频一区二区三区| 中文字幕人妻熟人妻熟丝袜美| 婷婷成人精品国产| 亚洲精品成人av观看孕妇| 下体分泌物呈黄色| 在线 av 中文字幕| 中文字幕精品免费在线观看视频 | 狠狠婷婷综合久久久久久88av| 婷婷色麻豆天堂久久| 一级二级三级毛片免费看| 97在线视频观看| 又粗又硬又长又爽又黄的视频| 99久久中文字幕三级久久日本| 成人国产麻豆网| 国产亚洲精品第一综合不卡 | 涩涩av久久男人的天堂| 女的被弄到高潮叫床怎么办| 国产日韩欧美视频二区| 久久精品夜色国产| 亚洲人与动物交配视频| 美女国产高潮福利片在线看| 亚洲四区av| 热99国产精品久久久久久7| 欧美日韩国产mv在线观看视频| 久久精品久久精品一区二区三区| 久久综合国产亚洲精品| 国产日韩一区二区三区精品不卡 | 丝袜美足系列| 18禁在线无遮挡免费观看视频| 男男h啪啪无遮挡| 免费观看无遮挡的男女| 一级二级三级毛片免费看| 中国国产av一级| 亚洲国产精品一区三区| 高清av免费在线| 国产欧美另类精品又又久久亚洲欧美| 精品久久久久久电影网| 一区二区av电影网| 亚洲欧美成人综合另类久久久| 少妇被粗大猛烈的视频| 亚洲熟女精品中文字幕| av免费观看日本| 婷婷色综合大香蕉| 五月伊人婷婷丁香| 久久精品久久久久久噜噜老黄| 最新中文字幕久久久久| 亚洲美女视频黄频| 日本vs欧美在线观看视频| 成人国产av品久久久| 日本午夜av视频| 久久久午夜欧美精品| 只有这里有精品99| 69精品国产乱码久久久| 亚洲美女视频黄频| 色吧在线观看| 日韩av免费高清视频| 黑人猛操日本美女一级片| 搡老乐熟女国产| 人人澡人人妻人| 久久97久久精品| av国产久精品久网站免费入址| 一级毛片我不卡| 又黄又爽又刺激的免费视频.| 大香蕉久久成人网| 热re99久久国产66热| 一本久久精品| 国产精品久久久久久精品电影小说| 亚洲人与动物交配视频| 欧美丝袜亚洲另类| 在线亚洲精品国产二区图片欧美 | 久久久亚洲精品成人影院| 欧美日韩成人在线一区二区| 五月伊人婷婷丁香| 亚洲av男天堂| 亚洲欧洲日产国产| 观看美女的网站| 国模一区二区三区四区视频| 久久久久网色| 中文乱码字字幕精品一区二区三区| 亚洲欧美日韩另类电影网站| 嘟嘟电影网在线观看| 亚洲精品乱码久久久v下载方式| 丁香六月天网| 日韩精品免费视频一区二区三区 | 在线观看免费高清a一片| 美女cb高潮喷水在线观看| a级毛片黄视频| 美女视频免费永久观看网站| 黑人巨大精品欧美一区二区蜜桃 | 麻豆乱淫一区二区| 九九爱精品视频在线观看| 亚洲第一av免费看| 尾随美女入室| 国产av码专区亚洲av| 人成视频在线观看免费观看| 日本av免费视频播放| 色吧在线观看| 男女高潮啪啪啪动态图| 免费大片18禁| 夜夜骑夜夜射夜夜干| 国产日韩欧美视频二区| 99精国产麻豆久久婷婷| 亚洲国产精品一区三区| 特大巨黑吊av在线直播| 黄色一级大片看看| 国产精品国产三级国产av玫瑰| 人妻系列 视频| 国产成人午夜福利电影在线观看| 欧美精品一区二区大全| a级毛片免费高清观看在线播放| 免费看av在线观看网站| 亚洲精华国产精华液的使用体验| 精品一区二区三区视频在线| 三级国产精品片| 丝袜喷水一区| 日本黄色日本黄色录像| 成年美女黄网站色视频大全免费 | 一二三四中文在线观看免费高清| 亚洲国产精品成人久久小说| 18禁动态无遮挡网站| 人妻 亚洲 视频| 亚洲人成77777在线视频| 日产精品乱码卡一卡2卡三| 99久久中文字幕三级久久日本| 青青草视频在线视频观看| 成年人午夜在线观看视频| 大香蕉久久网| 黑人猛操日本美女一级片| 免费黄网站久久成人精品| 亚洲成人av在线免费| 在线观看三级黄色| 人妻少妇偷人精品九色| 国产又色又爽无遮挡免| 成人手机av| 91精品伊人久久大香线蕉| 欧美精品高潮呻吟av久久| 只有这里有精品99| 2021少妇久久久久久久久久久| 自线自在国产av| 国产伦理片在线播放av一区| 人成视频在线观看免费观看| 久久久久久伊人网av| 黄色怎么调成土黄色| 亚洲av免费高清在线观看| 中文字幕免费在线视频6| 一区二区日韩欧美中文字幕 | 毛片一级片免费看久久久久| 三级国产精品片| 中文字幕人妻熟人妻熟丝袜美| h视频一区二区三区| 免费观看在线日韩| 精品久久久久久电影网| 亚洲精品久久午夜乱码| 99热6这里只有精品| 国产亚洲av片在线观看秒播厂| 国产精品一区二区三区四区免费观看| 汤姆久久久久久久影院中文字幕| 精品久久蜜臀av无| 插阴视频在线观看视频| 欧美成人精品欧美一级黄| 久久97久久精品| av不卡在线播放| videos熟女内射| 国产免费现黄频在线看| 日本色播在线视频| 91久久精品国产一区二区三区| 在现免费观看毛片| 国产日韩欧美亚洲二区| 不卡视频在线观看欧美| 男女啪啪激烈高潮av片| 只有这里有精品99| 夫妻午夜视频| 国产视频首页在线观看| 免费日韩欧美在线观看| 国产色爽女视频免费观看| 亚洲成人av在线免费| 一区二区av电影网| 在线观看免费高清a一片| 中文字幕制服av| 日韩,欧美,国产一区二区三区| 国产片内射在线| 亚洲精品久久成人aⅴ小说 | 91午夜精品亚洲一区二区三区| 啦啦啦视频在线资源免费观看| 日韩,欧美,国产一区二区三区| 黄片播放在线免费| 国产精品国产三级国产专区5o| 少妇被粗大的猛进出69影院 | 建设人人有责人人尽责人人享有的| 亚洲人成网站在线观看播放| 亚洲激情五月婷婷啪啪| 丰满乱子伦码专区| 国产黄色视频一区二区在线观看| 亚洲精品久久成人aⅴ小说 | 精品国产乱码久久久久久小说| 国产黄片视频在线免费观看| 欧美日韩亚洲高清精品| 大香蕉久久网| 九九久久精品国产亚洲av麻豆| 久久午夜福利片| 丁香六月天网| 大话2 男鬼变身卡| xxx大片免费视频| 丝袜脚勾引网站| 久久毛片免费看一区二区三区| 视频在线观看一区二区三区| 久久久久人妻精品一区果冻| 少妇熟女欧美另类| 欧美人与善性xxx| 欧美bdsm另类| 毛片一级片免费看久久久久| 日日啪夜夜爽| 制服丝袜香蕉在线| 国产精品欧美亚洲77777| 韩国av在线不卡| 亚洲,欧美,日韩| 26uuu在线亚洲综合色| 午夜福利视频精品| av在线老鸭窝| 久久免费观看电影| 久久影院123| 中文字幕制服av| 亚洲精品一区蜜桃| 九九在线视频观看精品| 久久人人爽人人爽人人片va| 99热6这里只有精品| 国产精品99久久久久久久久| 又大又黄又爽视频免费| 午夜福利影视在线免费观看| videos熟女内射| 亚洲精品aⅴ在线观看| 久久久久人妻精品一区果冻| 丝袜在线中文字幕| 欧美日韩综合久久久久久| 国产成人精品婷婷| 国产视频内射| 久久久久久久久久久免费av| 日韩三级伦理在线观看| 久久久久国产精品人妻一区二区| 在线精品无人区一区二区三| 国产一区二区在线观看av| 久久久久网色| 国产毛片在线视频| 亚洲av成人精品一二三区| 在线 av 中文字幕| 麻豆成人av视频| 99热网站在线观看| 亚洲人与动物交配视频| 成年人免费黄色播放视频| 国产精品国产三级国产av玫瑰| 免费黄色在线免费观看| 精品酒店卫生间| 国产成人精品婷婷| 蜜桃国产av成人99| 精品熟女少妇av免费看| av天堂久久9| 极品少妇高潮喷水抽搐| 亚洲激情五月婷婷啪啪| 一二三四中文在线观看免费高清| 国产日韩欧美在线精品| 91aial.com中文字幕在线观看| 最后的刺客免费高清国语| 免费观看在线日韩| 国产黄色视频一区二区在线观看| 免费人妻精品一区二区三区视频| 国产精品蜜桃在线观看| 最新中文字幕久久久久| 简卡轻食公司| 久久精品人人爽人人爽视色| 久久久国产一区二区| 欧美成人精品欧美一级黄| 人妻夜夜爽99麻豆av| 一区二区三区四区激情视频| 欧美精品高潮呻吟av久久| 欧美日韩视频精品一区| 五月天丁香电影| 国产在线免费精品| 久久久精品94久久精品| 亚洲第一av免费看| 有码 亚洲区| 在线精品无人区一区二区三| 高清不卡的av网站| 精品少妇黑人巨大在线播放| 一级二级三级毛片免费看| 午夜免费男女啪啪视频观看| 精品久久国产蜜桃| 一级毛片aaaaaa免费看小| 中文精品一卡2卡3卡4更新| 国产精品麻豆人妻色哟哟久久| 一级毛片 在线播放| 伊人久久国产一区二区| 久久精品久久久久久噜噜老黄| 精品人妻一区二区三区麻豆| 国产成人免费观看mmmm| 飞空精品影院首页| 婷婷色麻豆天堂久久| 一本久久精品| 少妇高潮的动态图| 美女xxoo啪啪120秒动态图| 精品视频人人做人人爽| 日韩制服骚丝袜av| 欧美亚洲日本最大视频资源| a级毛片在线看网站| 春色校园在线视频观看| 免费高清在线观看日韩| 综合色丁香网| 亚洲精品乱码久久久v下载方式| 久久久久久久久久久久大奶| 高清不卡的av网站| 免费黄频网站在线观看国产| 欧美日韩成人在线一区二区| av女优亚洲男人天堂| 插逼视频在线观看| 91精品伊人久久大香线蕉| 精品少妇黑人巨大在线播放| 国产亚洲最大av| 日韩一区二区视频免费看| 久久久精品94久久精品| 一个人看视频在线观看www免费| 最近中文字幕高清免费大全6| 久久精品人人爽人人爽视色| 中文字幕精品免费在线观看视频 | 美女脱内裤让男人舔精品视频| 一本久久精品| 日本黄大片高清| 简卡轻食公司| 这个男人来自地球电影免费观看 | 久久精品国产a三级三级三级| 国产亚洲欧美精品永久| 男人添女人高潮全过程视频| 久久久久视频综合|