• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A Visual-Based Gesture Prediction Framework Applied in Social Robots

    2022-01-26 00:36:00BixiaoWuJunpeiZhongandChenguangYang
    IEEE/CAA Journal of Automatica Sinica 2022年3期

    Bixiao Wu,Junpei Zhong,,and Chenguang Yang,

    Abstract—In daily life,people use their hands in various ways for most daily activities.There are many applications based on the position,direction,and joints of the hand,including gesture recognition,gesture prediction,robotics and so on.This paper proposes a gesture prediction system that uses hand joint coordinate features collected by the Leap Motion to predict dynamic hand gestures.The model is applied to the NAO robot to verify the effectiveness of the proposed method.First of all,in order to reduce jitter or jump generated in the process of data acquisition by the Leap Motion,the Kalman filter is applied to the original data.Then some new feature descriptors are introduced.The length feature,angle feature and angular velocity feature are extracted from the filtered data.These features are fed into the long-short time memory recurrent neural network(LSTM-RNN) with different combinations.Experimental results show that the combination of coordinate,length and angle features achieves the highest accuracy of 99.31%,and it can also run in real time.Finally,the trained model is applied to the NAO robot to play the finger-guessing game.Based on the predicted gesture,the NAO robot can respond in advance.

    I.INTRODUCTION

    CURRENTLY,computers are becoming more and more popular,and the demand for human-robot interaction is increasing.People pay more attention to research of new technologies and methods applied to human-robot interactions[1]–[3].Making human-robot interaction as natural as daily human-human interaction is the ultimate goal.Gestures have always been considered an interactive technology that can provide computers with more natural,creative and intuitive methods.Gestures have different meanings in different disciplines.In terms of interaction design,the difference between using gestures and using a mouse and keyboard,etc.,is obvious,i.e.,gestures are more acceptable to people.Gestures are comfortable and less limited by interactive devices,and they can provide more information.Compared with traditional keyboard and mouse control methods,the direct control of the computer by hand movement has the advantages of being natural and intuitive.

    Gesture recognition [4] refers to the process of recognizing the representation of dynamic or static gestures and translating them into some meaningful instructions.It is an extremely significant research direction in the area of human-robot interaction technology.The method of realizing gesture recognition can be divided into two types: visual-based [5],[6] gesture recognition and non-visual-based gesture recognition.The study of non-vision approaches began in the 1970s.Non-vision methods always take advantage of wearable devices [7] to track or estimate the orientation and position of fingers and hands.Gloves are very common devices in this field,and they contain the sensory modules with a wired interface.The advantage of gloves is that their data do not need to be preprocessed.Nevertheless,they are very expensive for virtual reality applications.They also have wires,which makes them uncomfortable to wear.With the development of technology,current research on non-visual gesture recognition is mainly focused on EMG signals[8]–[11].However,EMG signals are greatly affected by noise,which makes it is difficult to process.

    Gesture recognition is based on vision and is less intrusive and contributes to a more natural interaction.It refers to the use of cameras [12]–[16],such as Kinect [17],[18] and Leap Motion [19],[20],to capture images of gestures.Then some algorithms are used to analyze and process the acquired data to get gesture information,so that the gesture can be recognized.It is also more natural and easy to use,becoming the mainstream way of gesture recognition.However,it is also a very challenging problem.

    By using the results of gesture recognition,the subsequent gesture of performers can be predicted.This process could be called gesture prediction,and it has wider applications.In recent years,with the advent of deep learning,many deep neural networks (DNN) are applied to gesture prediction.Zhanget al.[21] used an RNN model to predict gestures from raw sEMG signals.Weiet al.[22] combined a 3D convolutional residual network and bidirectional LSTM network to recognize dynamic gesture.Kumaret al.[23] proposed a multi modal framework based on hand features captured from Kinect and Leap Motion sensors to recognize gestures,using a hidden Markov model (HMM) and bidirectional long shortterm memory model (LSTM).The LSTM [24] has become an effective model for solving some learning problems related to sequence data.Hence,inspired by the previous works,we adopt the LSTM to predict gestures in our proposed framework.

    Fig.1.Pipeline of the proposed approach.

    In the method of gesture prediction,hand key point detection is one of the most important steps.In the early stage of technological development,the former mainly used color filters to segment the hands to achieve detection.However,this type of method relies on skin color,and the detection performance is poor when the hand is in a complex scene.Therefore,the researchers proposed a detection method based on 3D hand key points.The task goal of the 3D hand key point estimation is to locate the 3D coordinates of hand joints in a frame of depth image,mostly used in virtual immersive games,interactive tasks [25],[26],and so on.Leap Motion is a kind of equipment for 3D data extraction based on vision technology.This device could extract the position of the hand joints,orientation and the speed of the fingertips movement.Recently,Leap Motion has always been used by researchers for gesture recognition and prediction.For example,some scholars use it to recognize American sign language (ASL)[27],[28],and it has a high gesture recognition accuracy.Moreover,Zenget al.[29] proposed a gesture recognition method based on deterministic learning and joint calibration of the Leap Motion.And Marinet al.[30] developed a method to combine the Leap Motion and Kinect to calculate different features of hand,and a higher accuracy was obtained.In this work,we use the data of hand key points detected by the Leap Motion to predict gestures and utilize the gesture recognition results to play the finger-guessing game.This game contains three gestures: rock,paper and scissors.The winning rules of this game are: scissors wins paper,paper wins rock,rock wins scissors.Based on these game rules,this paper proposes a method to judge gestures in advance when the player has not completed the action.

    The combination of the Leap Motion and LSTM significantly improves human-robot interaction.The Leap Motion could track each joint of the hand directly and has the ability to recognize or predict gestures.Moreover,compared with other devices,the Leap Motion has higher localization precision.On the other hand,the LSTM can solve the prediction problem well in most cases,and it is one of the important algorithms of deep learning (DL).This work combines the strengths of the LSTM and Leap Motion to predict gestures.Leap Motion captures 21 three-dimensional joint coordinates in each frame,and the LSTM network is used to train and test these features.This work has some novel contributions:

    1) A method for predicting gestures based on the LSTM is proposed.The data of gestures is collected by the Leap Motion.

    2) In order to reduce or eliminate the jitter or jump generated in the process of acquiring data by the Leap Motion,the Kalman filter is applied to solve this problem effectively.

    3) We propose a reliable feature extraction method,which extracts coordinate features,length features,angle features and angular velocity features,and combines these features to predict gestures.

    4) We apply the trained model to the NAO robot and make it play the finger-guessing game with players,which effectively verifies the real-time and accuracy of the proposed approach.

    The rest part of this paper is structured as below: in Section II,the process of processing data is given.In Section III,the experiment of this work is introduced in detail and the effectiveness is verified in this section.Finally,Section IV makes a summary.The framework of this paper is shown in Fig.1.

    II.DATA PROCESSING

    A.Leap Motion Controller

    The structure of the Leap Motion is not complicated,as shown in Fig.2.The main part of the device includes two cameras and three infrared LEDs.They tracked infrared light outside the visible spectrum,which has a wavelength of 850 nanometers.Compared with other depth cameras,such as the Kinect,the information obtained from the Leap Motion is limited (only a few key points rather than complete depth information) and it works in smaller three-dimensional areas.However,it is more accurate to use Leap Motion to acquire data.Moreover,Leap Motion provides software that can recognize some movement patterns,including swipe,tap and so on.Developers can access some functions of Leap Motion through the application programming interface (API) to create new applications.For example,they can obtain information about the position and length of the user’s hand to recognize different gestures.

    Fig.2.The structure of the Leap Motion.

    Even though the manufacturers declare an accuracy of the Leap Motion in position measurement is around 0.01 mm,[31] shows that it is about 0.2 mm for static measurements and 0.4 mm for dynamic measurements in fact.And in the coordinates of the finger joints extracted by Leap Motion,there exists jitter or even jump,which could affect the accuracy of the experimental results.In order to reduce or eliminate these phenomena,this work takes advantage of the Kalman filter to correct the predicted position of hand joints.

    B.Data Acquisition

    Each finger is marked with a name: thumb,index,middle,ring,and pinky,including four bones (except thumb).As shown in Fig.3,the phalanx of the finger includes the metacarpal,proximal phalanx,middle phalanx,and distal phalanx.Particularly,the thumb has only three phalanges,one less than the other.In the algorithm design,we set the length of the thumb metacarpal bone to 0 to guarantee that all five fingers have the same number of phalanges,which is easy to programme.In this work,the main data acquired by the Leap Motion are as follows:

    1) Number of Detected Fingers:Num∈[1,5] is the number of fingers detected by Leap Motion.

    2) Position of the Finger Joints:Pi,i=1,2,3,...,20 contains the three-dimensional position of each finger joint.The Leap Motion provides a one-to-one map between coordinates and finger joints.

    3) Palm Center:Pc(x0,y0,z0) represents three-dimensional coordinates of the center of the palm area in 3D space.

    4) Fingertips Movement Speed:Vrepresents the speed in the three-dimensional direction of each fingertip detected by the Leap Motion.

    Fig.3.Definition of endoskeleton in Leap Motion.

    C.Kalman Filter

    1) Problem Formulation:In the process of gesture changes,the fingertips have the largest range of change and can more easily jitter or jump than other joints,therefore,the Kalman filter is used to process the data from fingertips.Compared with other filters,such as the particle filter,the Luenburger observer filter,etc.,the Kalman filter has sufficient accuracy and can effectively remove Gaussian noise.In addition,its low computational complexity meets the real-time requirements of this work.Therefore,the Kalman filter is used for this work.

    Suppose that the current position of the fingertips obtained by Leap Motion isPt,and the speed isVt.The Kalman filter assumes that these two variables obey a Gaussian distribution,and each variable has a mean value of μ,and variance of σ2.For clarity,Xtdenotes the best estimate at timet,andYtdenotes the covariance matrix.The equations ofXtandYtare as follows:

    2) The Prediction Process:We need to predict the current state (timet) according to the state of the last time (timet-1).This prediction process can be described as follows:

    where Δtis the time interval,which depends on the data acquisition rate of the Leap Motion,and α is the rate of speed change.

    The matrixFtis used to represent the prediction matrix,so(5) can be represented as follows:

    and through the basic operation of covariance,Ytcan be expressed as the following equation:

    3) Refining the Estimate With Measurements:From the measured sensor data,the current state of the system can be guessed roughly.However,due to uncertainty,some states may be closer to the real state than the measurements acquired from the Leap Motion directly.In this work,covarianceRtis used to express the uncertainty (such as the sensor noise),and the mean value of the distribution is defined asZt.

    Now,there are two Gaussian distributions,one near the predicted value and the other near the measured value.Therefore,two Gaussian distributions are supposed to be multiplied to calculate the optimal solution between the predicted value and the measured value of the Leap Motion,as shown in the following equations:

    where μ0,σ0represent the mean and variance of the predicted values,respectively.μ1,σ1represent the mean and variance of the measured values,respectively.μ′,σ′represent the mean and variance of the calculated values,respectively.

    By substituting (8) into (9),we can get the following equations:

    the same parts of (10) and (11) are represented byk

    Therefore,(10) and (11) can be converted as follows:

    4) Integrate All the Equations:In this section,the equations of the Kal man filter used in this paper are integrated.

    There are two Gaussian distributions,the prediction part

    and the measurement part

    We then put them into (13) and (14) to get the following equation:

    And according to (12),the Kalman gain is as follows:

    Next,the above three formulas are simplified.On both sides of (17)and(18),we left multiply the inverse matrix ofFt.On both sides of(18),weright multiply the inverse matrix of.We then can get the following simplified equation:

    is the new optimal estimation of the data collected by the Le ap Motion,which we putalong withinto the next prediction and update the equation,and iterate continuously.Through the above steps,the data collected by the Leap Motion could be more accurate.

    D.Feature Extraction

    Now,after filtering the original data,we analyze four features acquired from the filtered data.These features are introduced in the rest of this section.

    ● Coordinate:x,y,andzcoordinates of the hand joints obtained by the Leap Motion.

    ● Length:The distance from the fingertips to center of the hand.

    ● Angle:The angle between the Proximal phalanx and Intermediate phalanx of each finger (except Thumb).

    ● Angular velocity:The rate of the joints angle change.

    1) Coordinate Feature:As shown in Fig.4(a),this feature set represents the position of the finger joints in three dimensional space.The original data take the Leap Motion as the coordinate origin as shown in Fig.5.With the movement of the hand,the obtained data could change a lot,which has a certain impact on the experimental results.For the purpose of eliminating the influence of different coordinate systems,the coordinate origin is changed to the palm center,as shown in Fig.4(a).Taking the palm of the hand as the plane,the direction from palm center to the root of the middle finger is the positive direction of they-axis.The positive direction of thex-axis is the direction perpendicular to they-axis and to the right.Through the coordinate origin,perpendicular to this plane is thez-axis.

    The positive direction of they-axis in the new coordinate system can be represented by the following vector:

    Similarly,the positive direction of thex-axis in the new coordinate system can be expressed by the following vector:

    And the positive direction of thez-axis in the new coordinate system can be expressed by the following vector:

    The coordinate representation in the new coordinate system is

    Fig.4.Four different types of features extracted from the Leap Motion.

    Fig.5.The coordinate system of the Leap Motion.

    where (,,) represents the new coordinate after coordinate conversion,i=1,2,...,20 represent the points corresponding to the finger joints.Through the above equations,we can get new coordinates with the palm center as the origin of the coordinate system.Because each three-dimensional coordinate is a array of length 3,the actual dimension of the coordinate feature is 3 × 20 = 60.

    2) Length Feature:As shown in Fig.4(b),this feature refers to the length from each fingertip to the center of the palm.The coordinates of the joints collected from the Leap Motion are used to calculate length information.It can be found that the fingertips are the most variable joints,so (31) is used to calculate the distance between the palm center and the fingertips.

    wherei=4,8,12,16,20 represent the points corresponding to the fingertips in Fig.4(b),and the dimension of length feature is 5.

    3) Angle Feature:As shown in Fig.4(c),this feature represents the angle between Proximal phalanx and Intermediate phalanx of each finger (except Thumb),and the angle extracted from the thumb is between the Intermediate phalanx and Distal phalanx.The calculation process is as follows:

    4) Angular Velocity Feature:As shown in Fig.4(d),this feature represents the rate of the joint angle change.As shown in the following equation:

    wheretis the current time,Δtis the time interval,which depends on the Leap Motion’s sampling time.The dimension of the angular velocity feature is 5.

    E.Gesture Prediction

    The method in the previous section produces four different features,and each feature represents some information related to the performed gesture.In this section,the LSTM network[24] used for gesture prediction is described in detail.The internal structure of the LSTM is shown in Fig.6,wherextdenotes the input of the LSTM network andhtdenotes the output of the LSTM network.ftis the forget gate variables,itis the input gate variables,andotis the output gate variables.The subscriptstandt-1 represent the current time and previous time.ctandare the memory cell state and the memory gate,respectively.The notation of σlstmandtanhdenote the sigmoid and hyperbolic activation functions as shown in (36) and (37).

    The relevant parameters of the LSTM can be calculated by the following equations:

    Fig.6.The internal structure of the LSTM.

    Fig.7.Collect gesture paper data by using the Leap Motion.

    where subscripts off,i,o,andcare related to the parameters of the forget gate,input gate,output gate and memory cell.The parametersWf,Wi,Wc,andWodenote the weight matrices of the corresponding subscripts.Similarly,bf,bi,bc,andbopresent the biases corresponding to subscripts of the LSTM network.The notation of * denotes the element-wise product between vectors.

    In the process of data collection,the change of gestures can be divided into three stages,as shown in Fig.7.For a clearer description,we take the process of turning rock into paper as an example to explain these three stages:

    1) The Original Stage:As shown in Figs.7(a) and 7(b),the gestures at this stage are close to the original state,that is,the gesture is similar to a rock.

    2) The Intermediate Stage:As shown in Figs.7(c) and 7(d),the gestures at this stage change significantly compared to the original stage,that is,the five fingers clearly show different degrees of openness.

    3) The Completion Stage:As shown in Figs.7(e) and 7(f),the gestures at this stage are close to the completion,that is,the gestures tend to being paper.

    Since different players perform actions at different speeds,each action contains 2–6 frames.For the purpose of uniformity,theTof LSTM is set to 4,that is,4 frames of data are input into the LSTM network for prediction.This process is shown in Fig.8.The input layer of the LSTM network is features obtained by the Leap Motion.These features are the coordinate,length,angle and angular velocity calculated from Section II-D,and their dimensions are 60,5,5,and 5,respectively.In addition,the hidden layer of the LSTM network contains 100 nodes.The output of the LSTM network is the result of gesture prediction with the dimension of 3,that is,rock,scissors and paper.With the LSTM network,we can predict the gestures accurately,and the classification results will be sent to a social robot for interaction and reaction.

    Fig.8.The process of the LSTM for predicting gestures.

    III.EXPERIMENT

    A.Experimental Setup

    In this section,the performance and efficiency of the proposed framework are tested.The experiments were carried out on a laptop with an Intel Core i5-6200U CPU.The dynamic gestures of rock,paper and scissors are collected from five different players,and each player repeats each gesture 300 times at fast,medium,and slow speeds,for a total of 4500 different data samples.The experimental results of the network trained by the four features and their combination are compared.

    B.Kalman Filter

    In Section II-C,the Kalman filter is introduced in detail.In this section,it is verified by an experiment,and the measured position is directly obtained by the Leap Motion.The Kalman filter is used to process the original coordinate data to make the processed data closer to the real value.As can be seen from Fig.9,the processed data is much smoother.

    Fig.9.Data processed by Kalman filter.

    C.Experimental Result

    According to the description in Section II-D,we extract the three-dimensional coordinates feature,length feature,angle feature,and angular velocity feature from the filtered data,and train them.Figs.10 and 11 show the accuracy of features using the classification algorithm of Section II-E.

    The three-dimensional positions of the finger joints show that the accuracy of gesture prediction is 97.93%.The length feature and the angle feature have an accuracy of 95.17% and 93.79%,respectively.The angular velocity feature has lower performance,it has an accuracy of 79.31%.It is affected by the speed of the player’s movement,so it is not fast enough to make an accurate prediction.

    The combination of multiple features could enrich the input of the neural network.In some cases,it maybe improve the performance of the prediction.As can be seen from Fig.11,the combination of coordinate features,length features and angle features achieve the highest accuracy of 99.31%,better than any of the three features alone.These results suggest that different features can represent different attributes of the hand and include complementary information.

    Fig.10.The experimental results of four features.

    Fig.11.The experimental results of the combination of four features.

    We examine whether the proposed method is able to achieve real-time gesture recognition and prediction.As shown in Figs.12 and 13,it is obvious that the method proposed in this work can predict the gesture of the fingerguessing game very well.For example,when the player’s gesture changes from rock to paper,the proposed method can predict that the player’s gesture is paper before all fingers are fully open.In addition,we also verify the prediction results of the proposed method from different angles of the Leap Motion to the hand,as shown in Fig.14.

    D.Application

    Fig.12.The prediction process of turning rock into paper.

    Fig.13.The prediction process of turning rock into scissors.

    Fig.14.Predicted results from different angles of the Leap Motion to hand.

    In order to further prove the effectiveness of the proposed method,the trained network is applied to the humanoid robot NAO,as shown in Fig.15.The NAO is an autonomous,programmable humanoid robot which is designed by Aldebaran Robotics [32].The height of the NAO is 573.2 mm and the weight of it is 4.5 kg.It has two cameras,voice recognition,voice synthesis and powered by LiPo Battery.What’s more,it consists of four microphones,two sonar emitters and receivers,two IR emitters and receivers,and three tactile sensors on the top of head.

    In this work,we mainly use the NAO robot’s left hand to play the finger-guessing game with the player.As shown in Fig.16,the NAO robot has only three fingers,and they are linked.Therefore,we first define that the full opening of the robot fingers is paper,the half opening of the robot fingers is scissors,and the clenched fingers are rock.

    Then,the trained model is applied to the NAO robot,and the experimental results are shown in Fig.17.The Leap Motion is used to predict gestures,and then the computer sends the results to the NAO robot,so that the NAO robot can win or lose the game through some simple judgments.

    Fig.17.The experimental results with the NAO robot.

    IV.CONCLUSION

    Fig.15.Experimental equipment and platform.

    Fig.16.The NAO robot and rock-paper-scissors gesture.

    In this paper,a gesture prediction framework based on the Leap Motion is proposed.In the process of data acquisition by the Leap Motion,some jumps or jitters maybe occur.Therefore,the Kalman filter is used to solve these problems.Then,based on the original coordinate features collected by the Leap Motion,we extract three new features,namely,the length feature,angle feature and angular velocity feature.The LSTM network is used to train the model for gesture prediction.In addition,the trained model is applied to the NAO robot to verify the real-time and effectiveness of the proposed method.

    久久久久久久国产电影| 大香蕉久久成人网| 丝袜喷水一区| 亚洲视频免费观看视频| 十八禁高潮呻吟视频| 亚洲欧美色中文字幕在线| 国产一区有黄有色的免费视频| www.999成人在线观看| 黄色视频,在线免费观看| 超色免费av| 一区在线观看完整版| 亚洲伊人久久精品综合| netflix在线观看网站| 一级黄色大片毛片| 精品亚洲乱码少妇综合久久| 岛国毛片在线播放| 如日韩欧美国产精品一区二区三区| 桃花免费在线播放| av国产精品久久久久影院| xxxhd国产人妻xxx| 日韩 亚洲 欧美在线| 国产在线一区二区三区精| 99精品欧美一区二区三区四区| 90打野战视频偷拍视频| 久久久国产成人免费| 亚洲av男天堂| 国产亚洲精品一区二区www | 可以免费在线观看a视频的电影网站| 国产精品一区二区免费欧美 | 青草久久国产| www.精华液| av电影中文网址| 中文字幕制服av| 宅男免费午夜| 在线观看免费日韩欧美大片| 成人18禁高潮啪啪吃奶动态图| 天天躁日日躁夜夜躁夜夜| 天堂俺去俺来也www色官网| 天天躁狠狠躁夜夜躁狠狠躁| 丝袜美腿诱惑在线| 人人妻人人澡人人看| 久久国产精品人妻蜜桃| 韩国高清视频一区二区三区| 宅男免费午夜| 亚洲熟女精品中文字幕| 国产片内射在线| 久久精品国产综合久久久| 亚洲精品国产区一区二| 欧美变态另类bdsm刘玥| 亚洲av国产av综合av卡| 青春草亚洲视频在线观看| 免费在线观看视频国产中文字幕亚洲 | 免费不卡黄色视频| 色婷婷久久久亚洲欧美| 亚洲精品国产色婷婷电影| 老司机福利观看| 精品欧美一区二区三区在线| 国产精品一区二区在线不卡| 中文字幕人妻熟女乱码| 丝瓜视频免费看黄片| 悠悠久久av| 少妇 在线观看| 亚洲五月色婷婷综合| 嫩草影视91久久| 美女扒开内裤让男人捅视频| 国产精品成人在线| 欧美成狂野欧美在线观看| 母亲3免费完整高清在线观看| 午夜精品久久久久久毛片777| 国产国语露脸激情在线看| 大香蕉久久成人网| 桃花免费在线播放| 夫妻午夜视频| 91精品国产国语对白视频| 91精品伊人久久大香线蕉| 久久精品亚洲熟妇少妇任你| 亚洲精品国产av成人精品| 黄网站色视频无遮挡免费观看| 不卡一级毛片| 女性被躁到高潮视频| 搡老岳熟女国产| 精品亚洲成国产av| 久久狼人影院| 久久av网站| 一个人免费在线观看的高清视频 | 叶爱在线成人免费视频播放| 亚洲精品中文字幕一二三四区 | 欧美日韩福利视频一区二区| 18禁裸乳无遮挡动漫免费视频| 国产区一区二久久| 丰满迷人的少妇在线观看| 99国产精品一区二区三区| 人人妻,人人澡人人爽秒播| 欧美黑人欧美精品刺激| 免费观看av网站的网址| 日韩视频在线欧美| 中文字幕制服av| 国产欧美日韩一区二区三区在线| 国产一区有黄有色的免费视频| 亚洲av美国av| 高清黄色对白视频在线免费看| 男人操女人黄网站| 最新在线观看一区二区三区| 丁香六月欧美| 动漫黄色视频在线观看| 久热这里只有精品99| 国产欧美日韩一区二区三 | 成人黄色视频免费在线看| av天堂久久9| 精品福利观看| 在线观看一区二区三区激情| 中国美女看黄片| 性高湖久久久久久久久免费观看| 国产av精品麻豆| 老司机午夜十八禁免费视频| 91精品国产国语对白视频| 久久久久久人人人人人| 久久久久久久国产电影| 国产精品久久久人人做人人爽| 在线看a的网站| 亚洲国产av影院在线观看| 夫妻午夜视频| 欧美日韩国产mv在线观看视频| 欧美国产精品va在线观看不卡| 国产精品一二三区在线看| 波多野结衣av一区二区av| 欧美午夜高清在线| 免费高清在线观看日韩| 永久免费av网站大全| 一区在线观看完整版| 人妻久久中文字幕网| 亚洲精品久久午夜乱码| 曰老女人黄片| 亚洲 国产 在线| 久久av网站| 男女边摸边吃奶| 成人国产av品久久久| 中文字幕色久视频| 在线观看舔阴道视频| 久久久久久久国产电影| 国产精品二区激情视频| 欧美日韩亚洲综合一区二区三区_| 人人妻人人澡人人看| 国产一区二区 视频在线| 欧美 亚洲 国产 日韩一| 精品视频人人做人人爽| bbb黄色大片| 一本综合久久免费| 久久久精品免费免费高清| 免费人妻精品一区二区三区视频| 精品国产国语对白av| 欧美在线黄色| 午夜精品久久久久久毛片777| 女人爽到高潮嗷嗷叫在线视频| 一区福利在线观看| 国产国语露脸激情在线看| 不卡av一区二区三区| 国产一卡二卡三卡精品| 久久这里只有精品19| a级毛片黄视频| 蜜桃在线观看..| 别揉我奶头~嗯~啊~动态视频 | 国产精品久久久久久人妻精品电影 | 丰满人妻熟妇乱又伦精品不卡| 久久久精品区二区三区| 亚洲精品中文字幕一二三四区 | 不卡一级毛片| 永久免费av网站大全| 午夜免费鲁丝| 欧美97在线视频| 18在线观看网站| 亚洲成人手机| a级片在线免费高清观看视频| 精品熟女少妇八av免费久了| 亚洲一区中文字幕在线| 亚洲精品国产精品久久久不卡| 欧美黄色淫秽网站| 国产无遮挡羞羞视频在线观看| 国产亚洲欧美在线一区二区| 丰满人妻熟妇乱又伦精品不卡| 亚洲全国av大片| 中文字幕人妻丝袜一区二区| 香蕉国产在线看| 如日韩欧美国产精品一区二区三区| 国产精品久久久av美女十八| 两性午夜刺激爽爽歪歪视频在线观看 | tocl精华| 最黄视频免费看| 久久久国产成人免费| 一级片'在线观看视频| 人妻人人澡人人爽人人| 激情视频va一区二区三区| 他把我摸到了高潮在线观看 | 美国免费a级毛片| 99九九在线精品视频| 亚洲欧美激情在线| 日本欧美视频一区| 国产精品.久久久| 国产精品成人在线| 高清av免费在线| 正在播放国产对白刺激| 日日摸夜夜添夜夜添小说| 色播在线永久视频| 黄色怎么调成土黄色| 国产亚洲精品一区二区www | 成人亚洲精品一区在线观看| 国产黄频视频在线观看| e午夜精品久久久久久久| 一级毛片精品| 五月天丁香电影| 黄网站色视频无遮挡免费观看| 久久久国产欧美日韩av| 欧美精品高潮呻吟av久久| 亚洲精品一二三| 国产成人精品久久二区二区91| 51午夜福利影视在线观看| 午夜老司机福利片| 久久久久精品国产欧美久久久 | www.自偷自拍.com| 亚洲人成77777在线视频| 欧美精品一区二区免费开放| 国产99久久九九免费精品| 在线观看www视频免费| 黄片小视频在线播放| 免费高清在线观看视频在线观看| 欧美少妇被猛烈插入视频| 9热在线视频观看99| 一区二区日韩欧美中文字幕| 日韩人妻精品一区2区三区| 亚洲avbb在线观看| 97精品久久久久久久久久精品| 亚洲精品久久成人aⅴ小说| 一级片免费观看大全| 日韩一卡2卡3卡4卡2021年| 久久九九热精品免费| 国产精品偷伦视频观看了| 亚洲第一欧美日韩一区二区三区 | 蜜桃在线观看..| 亚洲精品中文字幕一二三四区 | 国产av又大| 黄片大片在线免费观看| 中文欧美无线码| 精品少妇黑人巨大在线播放| 国产亚洲精品久久久久5区| 亚洲精华国产精华精| 狠狠狠狠99中文字幕| 老司机影院毛片| 亚洲九九香蕉| 国产97色在线日韩免费| 日韩免费高清中文字幕av| www日本在线高清视频| 成人国产一区最新在线观看| 激情视频va一区二区三区| 他把我摸到了高潮在线观看 | 人人妻人人澡人人看| 久久久久视频综合| 久久精品国产亚洲av香蕉五月 | 亚洲精品自拍成人| 夜夜夜夜夜久久久久| 国产亚洲精品一区二区www | av国产精品久久久久影院| 欧美+亚洲+日韩+国产| 成年人免费黄色播放视频| 爱豆传媒免费全集在线观看| 女人高潮潮喷娇喘18禁视频| 日韩有码中文字幕| 免费在线观看影片大全网站| 香蕉国产在线看| 黄色视频在线播放观看不卡| 国产成人精品久久二区二区91| 精品少妇内射三级| 亚洲国产日韩一区二区| 丰满迷人的少妇在线观看| 久久狼人影院| 老司机福利观看| 国产精品影院久久| 免费在线观看黄色视频的| 精品福利观看| 国产精品久久久久久人妻精品电影 | 色播在线永久视频| 久久久久精品人妻al黑| 国产精品一区二区在线不卡| 国产精品国产av在线观看| 欧美日韩成人在线一区二区| 久久久久国产一级毛片高清牌| 精品人妻熟女毛片av久久网站| 亚洲精品久久久久久婷婷小说| 久久热在线av| 手机成人av网站| 国产一卡二卡三卡精品| bbb黄色大片| 久久精品亚洲熟妇少妇任你| 1024香蕉在线观看| 视频区图区小说| 老司机深夜福利视频在线观看 | 国产精品亚洲av一区麻豆| 欧美日韩精品网址| 国产精品久久久久久人妻精品电影 | 亚洲精品国产av成人精品| 国产在线免费精品| 丁香六月欧美| www.999成人在线观看| 亚洲国产中文字幕在线视频| 亚洲伊人色综图| 最黄视频免费看| 精品国产超薄肉色丝袜足j| 国产精品久久久久久人妻精品电影 | 精品少妇内射三级| 捣出白浆h1v1| 青青草视频在线视频观看| 日韩电影二区| 国产成人啪精品午夜网站| 国产精品麻豆人妻色哟哟久久| 极品人妻少妇av视频| 成年av动漫网址| 欧美精品一区二区免费开放| 亚洲精品国产区一区二| 后天国语完整版免费观看| 国产精品国产三级国产专区5o| 国产成人精品久久二区二区91| 国产在视频线精品| 俄罗斯特黄特色一大片| 高清在线国产一区| 欧美亚洲日本最大视频资源| 国产精品久久久久久人妻精品电影 | 久久人人爽av亚洲精品天堂| 深夜精品福利| 国产成人啪精品午夜网站| 青春草视频在线免费观看| 亚洲免费av在线视频| 99国产极品粉嫩在线观看| 国产熟女午夜一区二区三区| 热re99久久精品国产66热6| 国产精品久久久久久人妻精品电影 | 久久人人爽人人片av| 母亲3免费完整高清在线观看| 不卡av一区二区三区| 中文字幕人妻丝袜制服| 夜夜骑夜夜射夜夜干| 美女脱内裤让男人舔精品视频| 丁香六月欧美| 亚洲人成77777在线视频| 男男h啪啪无遮挡| 国产一区二区激情短视频 | 亚洲欧美精品自产自拍| 亚洲国产欧美网| 久久精品人人爽人人爽视色| 黑人操中国人逼视频| 亚洲av电影在线进入| 三上悠亚av全集在线观看| 香蕉国产在线看| 亚洲熟女毛片儿| 男女边摸边吃奶| 国产精品免费视频内射| 欧美日韩成人在线一区二区| 婷婷色av中文字幕| 亚洲色图 男人天堂 中文字幕| 韩国高清视频一区二区三区| 男人操女人黄网站| 狠狠精品人妻久久久久久综合| 欧美日韩福利视频一区二区| 国产高清视频在线播放一区 | 老熟妇乱子伦视频在线观看 | 亚洲专区字幕在线| 高清黄色对白视频在线免费看| av不卡在线播放| 亚洲国产精品一区二区三区在线| 欧美国产精品va在线观看不卡| 亚洲欧美精品自产自拍| 9热在线视频观看99| 欧美精品人与动牲交sv欧美| 欧美成人午夜精品| 后天国语完整版免费观看| 欧美亚洲日本最大视频资源| 国产国语露脸激情在线看| 欧美+亚洲+日韩+国产| 日本wwww免费看| 悠悠久久av| 欧美日韩福利视频一区二区| 亚洲少妇的诱惑av| 国产精品亚洲av一区麻豆| tube8黄色片| 欧美一级毛片孕妇| 亚洲国产看品久久| 亚洲综合色网址| 欧美日韩亚洲高清精品| 人人妻人人澡人人看| 91字幕亚洲| bbb黄色大片| 国产精品影院久久| 亚洲美女黄色视频免费看| 一级毛片精品| 亚洲avbb在线观看| 精品少妇一区二区三区视频日本电影| 国产一卡二卡三卡精品| 久久性视频一级片| 欧美精品av麻豆av| 亚洲精品国产区一区二| 久久人妻熟女aⅴ| 亚洲一卡2卡3卡4卡5卡精品中文| av在线播放精品| 一级毛片女人18水好多| 亚洲 国产 在线| 国产精品.久久久| 最黄视频免费看| 精品少妇内射三级| 欧美变态另类bdsm刘玥| 美女高潮喷水抽搐中文字幕| 欧美日韩视频精品一区| 无限看片的www在线观看| 搡老岳熟女国产| 久久精品亚洲熟妇少妇任你| 国产成人系列免费观看| 国产福利在线免费观看视频| 91老司机精品| 午夜福利乱码中文字幕| 国产色视频综合| 亚洲国产精品成人久久小说| 国产免费现黄频在线看| 亚洲一码二码三码区别大吗| 天天操日日干夜夜撸| 高清在线国产一区| 成年av动漫网址| 国产亚洲欧美在线一区二区| 亚洲精品国产一区二区精华液| 国产欧美日韩一区二区三区在线| 午夜福利在线观看吧| 超碰成人久久| 俄罗斯特黄特色一大片| 真人做人爱边吃奶动态| www.999成人在线观看| www.熟女人妻精品国产| 男女午夜视频在线观看| 黄色片一级片一级黄色片| 韩国高清视频一区二区三区| 日韩人妻精品一区2区三区| 一本综合久久免费| 性高湖久久久久久久久免费观看| 色视频在线一区二区三区| 久久久精品免费免费高清| 大码成人一级视频| 亚洲精品一区蜜桃| 亚洲人成电影免费在线| 啦啦啦啦在线视频资源| a级毛片在线看网站| 老司机影院成人| 天天添夜夜摸| 亚洲色图综合在线观看| h视频一区二区三区| 一本色道久久久久久精品综合| 一级a爱视频在线免费观看| 手机成人av网站| 午夜福利在线免费观看网站| 免费在线观看日本一区| 久久毛片免费看一区二区三区| 国产区一区二久久| 免费在线观看影片大全网站| 国产亚洲精品久久久久5区| 亚洲中文日韩欧美视频| 成年人免费黄色播放视频| 两个人看的免费小视频| 亚洲视频免费观看视频| 精品国产超薄肉色丝袜足j| 中文字幕人妻丝袜制服| 黑人巨大精品欧美一区二区mp4| 国产色视频综合| 老熟女久久久| 午夜激情久久久久久久| 中文字幕人妻丝袜一区二区| 色播在线永久视频| 美女大奶头黄色视频| 国产亚洲一区二区精品| 色婷婷av一区二区三区视频| 免费黄频网站在线观看国产| 中文字幕最新亚洲高清| 色94色欧美一区二区| 日韩免费高清中文字幕av| 日韩人妻精品一区2区三区| 欧美亚洲日本最大视频资源| 国产精品亚洲av一区麻豆| 久久毛片免费看一区二区三区| 亚洲 欧美一区二区三区| 亚洲精品粉嫩美女一区| 色婷婷久久久亚洲欧美| tocl精华| 欧美精品高潮呻吟av久久| 久久午夜综合久久蜜桃| 午夜福利在线免费观看网站| 国产精品99久久99久久久不卡| 男女边摸边吃奶| 亚洲精品国产av蜜桃| 中文精品一卡2卡3卡4更新| 久久av网站| 亚洲欧美一区二区三区久久| 亚洲精品自拍成人| 国产深夜福利视频在线观看| 欧美另类亚洲清纯唯美| 日韩,欧美,国产一区二区三区| 高清欧美精品videossex| 青青草视频在线视频观看| 黄色视频,在线免费观看| 狂野欧美激情性bbbbbb| 国产极品粉嫩免费观看在线| 搡老岳熟女国产| 亚洲情色 制服丝袜| 天天添夜夜摸| 国产日韩欧美视频二区| 日本vs欧美在线观看视频| 欧美变态另类bdsm刘玥| 亚洲成人手机| 亚洲一区二区三区欧美精品| 久久人人97超碰香蕉20202| 精品欧美一区二区三区在线| 十八禁网站免费在线| 老司机深夜福利视频在线观看 | 午夜91福利影院| 电影成人av| 午夜精品国产一区二区电影| 亚洲国产中文字幕在线视频| 亚洲中文av在线| 在线 av 中文字幕| 午夜精品久久久久久毛片777| 天天影视国产精品| 人人澡人人妻人| 狠狠婷婷综合久久久久久88av| 天堂俺去俺来也www色官网| e午夜精品久久久久久久| 国产一卡二卡三卡精品| 人妻人人澡人人爽人人| 成人手机av| 国产野战对白在线观看| 日本猛色少妇xxxxx猛交久久| 亚洲精品国产精品久久久不卡| 国产老妇伦熟女老妇高清| 69精品国产乱码久久久| 欧美97在线视频| 日日夜夜操网爽| 视频区欧美日本亚洲| av免费在线观看网站| 欧美人与性动交α欧美精品济南到| 啦啦啦免费观看视频1| 脱女人内裤的视频| 欧美精品人与动牲交sv欧美| 伊人久久大香线蕉亚洲五| av又黄又爽大尺度在线免费看| 精品人妻熟女毛片av久久网站| 色婷婷久久久亚洲欧美| 在线观看免费午夜福利视频| 人妻 亚洲 视频| 国产1区2区3区精品| 大型av网站在线播放| 久久久久久久国产电影| 男女高潮啪啪啪动态图| 美女脱内裤让男人舔精品视频| 美女国产高潮福利片在线看| 男女边摸边吃奶| 水蜜桃什么品种好| 999久久久国产精品视频| 午夜精品久久久久久毛片777| 国产欧美日韩一区二区三区在线| 99久久国产精品久久久| 国产野战对白在线观看| 国产成人欧美在线观看 | 别揉我奶头~嗯~啊~动态视频 | 18在线观看网站| 免费日韩欧美在线观看| 亚洲精品国产精品久久久不卡| 国产日韩一区二区三区精品不卡| 他把我摸到了高潮在线观看 | 成人黄色视频免费在线看| 成人国产一区最新在线观看| 十分钟在线观看高清视频www| 巨乳人妻的诱惑在线观看| 久久久精品94久久精品| 99精品欧美一区二区三区四区| 精品国产乱子伦一区二区三区 | 热re99久久国产66热| 国产亚洲精品久久久久5区| 777久久人妻少妇嫩草av网站| 国产色视频综合| 日韩欧美国产一区二区入口| 亚洲国产欧美在线一区| 午夜免费观看性视频| 亚洲欧洲日产国产| 日本精品一区二区三区蜜桃| 男男h啪啪无遮挡| 满18在线观看网站| 日韩一区二区三区影片| 久久香蕉激情| 免费黄频网站在线观看国产| 国产精品 欧美亚洲| 亚洲男人天堂网一区| 热re99久久国产66热| 亚洲欧美精品自产自拍| 亚洲精品第二区| 久久性视频一级片| 亚洲国产成人一精品久久久| 免费在线观看黄色视频的| 乱人伦中国视频| 18禁黄网站禁片午夜丰满| 在线亚洲精品国产二区图片欧美| 两人在一起打扑克的视频| 国产日韩欧美在线精品| www.精华液| 欧美变态另类bdsm刘玥| 黄片播放在线免费| 日韩视频在线欧美| 制服人妻中文乱码| 人人妻人人澡人人爽人人夜夜| 精品久久久精品久久久| 久久精品国产亚洲av高清一级| 久热爱精品视频在线9| 欧美日韩福利视频一区二区| 久久久国产欧美日韩av| √禁漫天堂资源中文www|