• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Sequential Bag-of-Words model for human action classification

    2016-03-20 07:11:12HongLiuHaoTangWiXiaoZiYiGuoLuTianYuanGao

    Hong Liu*,Hao Tang,Wi Xiao,ZiYi Guo,Lu Tian,Yuan Gao

    aSchool of Electronic and Computer Engineering,Peking University,Shenzhen Graduate Shcool,518055,China

    bEngineering Lab on Intelligent Perception for Internet of Things(ELIP),Peking University,Shenzhen Graduate School,518055,China

    cKey Laboratory of Machine Perception,Peking University,100871,China

    dSchool of Software and Microelectronics,Peking University,100871,China

    eInstitute of Computer Science,Christian-Albrechts-University,24118,Germany Available online 21 October 2016

    Sequential Bag-of-Words model for human action classification

    Hong Liua,b,c,*,Hao Tanga,b,Wei Xiaoa,b,ZiYi Guod,Lu Tiana,b,Yuan Gaoe

    aSchool of Electronic and Computer Engineering,Peking University,Shenzhen Graduate Shcool,518055,China

    bEngineering Lab on Intelligent Perception for Internet of Things(ELIP),Peking University,Shenzhen Graduate School,518055,China

    cKey Laboratory of Machine Perception,Peking University,100871,China

    dSchool of Software and Microelectronics,Peking University,100871,China

    eInstitute of Computer Science,Christian-Albrechts-University,24118,Germany Available online 21 October 2016

    Recently,approaches utilizing spatial-temporal features to form Bag-of-Words(BoWs)models have achieved great success due to their simplicity and effectiveness.But they still have difficulties when distinguishing between actions with high inter-ambiguity.The main reason is that they describe actions by orderless bag of features,and ignore the spatial and temporal structure information of visual words.In order to improve classification performance,we present a novel approach called sequential Bag-of-Words.It captures temporal sequential structure by segmenting the entire action into sub-actions.Meanwhile,we pay more attention to the distinguishing parts of an action by classifying subactions separately,which is then employed to vote for the final result.Extensive experiments are conducted on challenging datasets and real scenes to evaluate our method.Concretely,we compare our results to some state-of-the-art classification approaches and confirm the advantages of our approach to distinguish similar actions.Results show that our approach is robust and outperforms most existing BoWs based classification approaches,especially on complex datasets with interactive activities,cluttered backgrounds and inter-class action ambiguities.

    Action classification;Sequential Bag-of-Words;STIP;Probalibity

    1.Introduction

    Human action classification is an important research topic in the fields of computer vision and pattern recognition.It is a growing challenging problem and well worth studying in the task of automatically analyzing and interpreting video data in computer vision and image analysis.It is of significant use in many applications,such as intelligent surveillance,contentbased video retrieval and human-computer interaction.As a hot but difficult problem in the field of computer vision,action classification has been researched for years,but remains a verychallenging task.It still has many challenges such as intraclass and inter-class variations,body occlusion,camera movement,environment changing,etc.In action recognition, the performance is frequently susceptible to interference under background changes,variable illumination,camera motion and zooming,and intra-class inhomogeneity or variability.To alleviate thesedifficult circumstance,discriminative descriptors are of vital importance.These video descriptors are invariant to the variability disturbed by the multiple noise such as distortions,occlusions and so on.In addition,they need yet to encode the spatio-temporal information effectively and robustly.One of the most difficult problems is to distinguish between actions with high inter-ambiguities.Regarding an action as a connection of sub-actions,some action classes consist of similar sub-actions,which greatly increase the diffi culty of classification(Fig.3).

    Recent work in action recognition are generally divided into four categories:

    Fig.1.Sub-sections and sub-actions.Each section is called a sub-action.The ith sub-actions of all actions compose the ith sub-section.

    Fig.2.Different levels of actions,from left to right are"sub-action","action primitive","action","activity".They are separated by their complexities.

    .Space-time shape[8,20];

    .Optical flow[23,24];

    .Trajectory[11,12,19];

    .Local STIPs[4,14].

    For instance,Blank et al.[2]treat human actions as 3D shapes induced by the sihouettes in the space-time volume. Efros et al.[5]utilize cross-correlation between optical flow descriptors.Ali et al.[1]apply the trajectories of feet,hand and body into the action recognition framework.

    Generally,the first three categories are the global representation methods which rely on the recording circumstances frequently.Meanwhile,as for the last category,Dollar et al.[4] and Leptev et al.[9]concentrate on the local representations of STIPs in the video sequences.In Dollar et al.[4]developed a detector which applied linear 2D Gaussian kernel in the spatial dimensions and 1D Gabor filter in the temporal dimension. The local regions around STIPs are depicted by the proposed descriptors.At last the video sequences are encoded as the frequency histogram of the visual-words.In[9],the Harrisdetector is extended to the space-time domain.Interest points are located using extended Harris detector.The resulting points could be regarded as spatio-temporal corners.

    Fig.3.Examples of two confusing actions:"shaking"and"pushing".Segmenting them into 5 sections,the former sections are nearly the same.This leads to large ambiguity between their BoWs histograms.But if we focus on the salient parts of the actions,the classification would be much easier.

    Although BoWs model is popular,it has an essential drawback of only focusing on the number of words but ignoring the spatial-temporal information,which results in ambiguity between different classes of actions.For example, some similar sub-actions happen at different relative period of time,such as standing up at the beginning of an action or at the end.BoWs model is incapable to distinguish them.Moreover, BoWs model handles all the visual words equally,therefore does not lay emphases on the most distinguish parts.The higher proportion of similar sub-actions between classes,the more difficult for classification procedure using original BoWs.Hence,introducing temporal series of BoWs model is of great importance and necessity.

    Dollar et al.[4]present the gradient descriptor of STIPs (Spatial-Temporal Interest Points),which obtain better performance compared with other descriptors such as optical flow descriptor,brightness descriptor in their experiments.First,the spatial-temporal cuboids were smoothed in different scales and the image gradients are calculated.Next,the obtained gradient values were connected as the gradient vectors,and the vectors were projected to the lower dimensional space through the Principal Component Analysis(PCA)and finally we got these lowerdimensionalvectorsrepresentationforfurtherapplications.

    Laptev et al.[10]apply the Histogram of Oriented Gradients(HOG)as the STIPs descriptor.By using HOG descriptor, localobjectshape and appearance were able to be characterized through the local intensity gradients.Generally, there were two steps in the feature description:

    1.The cuboids were divided into small space-time regions;

    2.The local 1D-histogram of gradient directions over the pixels for each region were accumulated.

    Furthermore,Laptev also combined the Histogram of Optical Flow(HOF)and HOG to form a new descriptor,called HOG-HOF,which outperformed each separate method.

    Scovanner et al.[21]proposed the 3-Dimensional Scale Invariant Feature Transform(3D SIFT)descriptor,which accurately captured the spatio-temporal nature of the video sequences and extended the BoW paradigm from 2D to 3D, where the third dimension was time axis.In 3D SIFT descriptor[21];the gradient magnitude and orientation were represented by m3D(x,y,t),θ(x,y,t)and φ(x,y,t).By this means, every angle had been represented by a single unique(θ,φ)and each pixel had two values which represented the direction of the gradient in three dimensions.Finally,the orientations were concatenated into a histogram for each 3D sub-region,and the sub-histograms were formed as the desired descriptor.

    Zhao et al.[16,22]developed the Local Binary Pattern on Three Orthogonal Planes(LBP-TOP).Delightedly,LBP-TOP had been applied for dynamic texture description and recognition successfully and had obtained good results on facial expression analysis.The proposed algorithm extracted the LBP [15]fromthreeorthogonalXY,XT,XTplanes.Inthismanner,on one hand the spatial information were encoded in XY plane.Onthe other hand,the spatial temporal co-occurrence information were encoded in XTand YT planes.Shao et al.[13]proposed to extend the computation of LBP and computed histogramming LBP[7]named Extended LBP-TOP and Extended CSLBPTOP,respectively.The presented methods could extract more dynamic information inside the spatio-temporal cuboids and be applied on the gradient cuboids.

    However,only a small part of previous works focus on capturing visual words'temporal relationships.In some approaches[8,12,20];spatial-temporal correlations of local features were learned as neighborhoods or correlograms to capture the visual words'spatial-temporal relationships.While these approaches are still too local to capture long-term relationship between words.Other works[11,23,24]counted the co-occurrence between words,but they limited the scope within a small period of time.Recently Ryoo[17]represented an activity as an integral histogram of spatial-temporal features,efficiently modeling how feature distributions change over time,but it could not deal with ambiguities between classes which have similar sub-actions at same relative time. Glaser et al.[6]incorporated temporal context in BoWs models to capture the contextual sequence between words but it still could not focus on the distinctive parts of the actions.

    Instead of directly including time information in visual words,we take account of time dimension by segmenting the entire action into small sections.Each section is called a subaction.The ith sub-actions of all actions compose the ith subsection,as shown in Fig.1.

    ThensequentialBoWsmodelisusedforvideorepresentation, the classification is described as a series of sub-classes classi fication problems.In this way,we can apply our approach to original BoWs and spatial-improved BoWs,such as approaches in our previous works[11,24].Satkin et al.[19]extracted the most discriminative portion of a video for training according to the accuracy of a trained classifier.Inspired by this,the discriminative parts of the action is emphasised by assigning them high weights and salience values.The weight indicates the probability of a sub-action belonging to a certain class and the salience means how much a sub-section is distinguished from others.Thentheeffectofsimilarsub-actionsisminimized,sowe can focus on the difference between classes(salient parts in Fig.3).Ontheonehand,ifsimilarsub-actionshappenatdifferent relative time,they will be in different sub-section and classified separately.On the other hand,if they can not be separated,then low salience values are given to them to reduce their influence. Finally,sub-actions are classified separately and the results will be gathered together through voting.The experiments are implemented on UT-interaction dataset and a more challenging Rochester dataset.The results show our approach can achieve robust and accuracy beyond most related approaches.

    Therestofthepaperisorganizedasfollows.InSection2,we introduce the reason for using sub-actions and illustrate the framework of our approach.Section 3 and Section 4 describe the segmentation and classification approach respectively.In Section 5,we conduct experiments on UT-Interaction and Rochester datasets and compare our approach with other BoWs based approaches.Finally,conclusions are drawn in Section 6.

    2.The proposed framework

    As shown in Fig.2,human movement can be described at various levels of complexities[16].Usually,an activity refers a wholeexpressionofaseriesofactionssuchas"playtennis".An actionistheelementofactivitysuchas"running"or"jumping". It is often short and represents less motion information.As for actionprimitive,itisaveryshortperiodofactionthatcannotbe performed individually.Action primitives are components of actions and actions are components of activities as well.For example,"left leg forward"is an action primitive,whereas "running"is an action."Jumping hurdles"is an activity that contains starting,jumping and running actions.

    In general,action classification is performed at the first two levels.Many different action classes are unavoidable in sharing similar or same action primitives.This largely increases the difficulty to distinguish different classes.However, action classification at primitive level is also impractical.The reason is that there are innumerable action primitives due to the flexibility of human movement.Moreover,different action classes may be composed by different numbers of action primitives,which is not suitable to perform uniform preprocessing for all action classes.To solve this problem,we define sub-action instead of action primitive.Sub-action is also a small period of action,but it is not previously defined orfixed for each action classes.Sub-action is segmented according to a certain action automatically.All the action classes to be classified will have same number of sub-actions.This approach can not only solve the problem mentioned above but also be faster,more flexible and adaptive.

    In computer vision field,the Bag-of-Words model(BoWs model)can be applied to image classification,by treating image features as words.In document classification,a bag of words is a sparse vector of occurrence counts of words,that is, a sparse histogram over the vocabulary.In the field of computer vision,a bag of visual words is a vector of occurrence counts of a vocabulary of local image features.To represent an image using BoWs model,an image can be treated as a document.Similarly,"words"in images need to be defined too.To achieve this,three steps are usually included:

    1.Feature detection;

    2.Feature description;

    3.Codebook generation.

    A definition of the BoWs model can be the"histogram representation based on various features".

    As shown in Fig.4,first local features are extracted from action videos.The videos are then treated as volumes of visual words.To extract sub-actions,we chop the volumes into small clips according to the intensity of actions.Distances between clips are accumulated.Then segment the accumulated distance into equal parts,i.e.,the sub-actions.Sequential histograms are created for each action respectively to describe the action. Before classification,similarity between sub-actions in the same sub-section is figured out with training data.These similarities can be regarded as weights for voting.Meanwhilesalience value for each section is figured out according to the pre-classification accuracy.After sub-section classification,a vote scheme is conducted finally.Category of a test instance A is decided by the equation below:

    Fig.4.Framework of our approach.Firstly,videos are transformed into volumes of visual words.Secondly,a two-stage segmentation is conducted.Stage-1 chop volumes into clips according to the density of words,stage-2 divide accumulated distance between clips into equal parts.Thirdly,normal classification is done to each sub-section respectively and results are recorded as cj.Finally,the final result is decided by a voting process.

    where C denotes all the possible categories,Nsrepresents the number of sub-sections,cjdenotes A's sub-classification result in sub-section j and ωj(i,cj)is the weight of the jth sub-action belonging to class i when it is classified to sub-class cj.The last sj(cj)stands for the saliency of the jth sub-action.The instancewillbeclassifiedtothecategorywiththehighestscore.

    3.Action segmentation

    Our approach takes advantage of local spatio-temporal features to represent actions.After extracting local features into visual words from video sequence,we conduct a two-stage segmentation to avoid acquiring inequality sub-actions caused by the scale,range,rate or other individual difference between actors.Optimal segmentation could narrow the classification and disambiguate similar sub-actions occur at different relative time.We briefly introduce thevisual words extraction approach in Subsection 3.1,and in Subsection 3.2,our segmentation approach to acquire sub-actions is presented.

    3.1.Visual words extraction

    As shown in Fig.5,local feature based representation is widely used in human action analysis for its resistant to clutters and noise.Firstly,Spatial-Temporal Interest Points(STIPs)are detected,small cuboid is extracted around each interest point. Then descriptors are generated to represent local information. There are many different local detectors and descriptors being proposed.Since the average length of sub-action is short,our approach avoid using those detecting approaches which are too sparse.Dense detector[18,25]is among good choices.Specifically,the cuboid detector proposed by Dollar et al.[4]and the scale-invariant detector proposed by Willems et al.[25].Dollar's descriptor is used in our experiment.

    Fig.5.Extraction of visual words.

    Typically,Dollar[4]proposed an alternative spatiotemporalfeature detectorforaction recognition.Thedetector is based on a set of separable linear filters which regards the spatial and temporal dimensions in different ways. In[4],to extract the STIPs from video sequence I,the response function is given by:

    where g(x,y,σ)is the 2D Gaussian smoothing kernal,applied along the spatial dimensions,and heven,hoddare a quadrature pair of 1D Gabor filters applied temporally which are represented as:

    In Equation(3),ω represents the underline frequency of cosine in the Gabor filter.When ω=4/τ,the number of parameters in the response function R in Equation(2)is reduced to two:σ,τ,corresponding roughly to the spatial and temporal scale of the detector.

    Fig.6.The key images of"answer phone"acted by two actors.(a),(b),(c),(d),(e),(f)correspond to six segmenting points A,B,C,D,E,F in Fig.7.(a)and(f)are the start and end of the video.In the first section,actions change from(a)to(b),both actors'hands go forward to get the phone.In the second section(b)to(c),the phones are taken closer.In the third section(c)to(d),actors open the clamshell phones.In the forth section(d)to(e),actors raise the phones.In the final section(e) to(f),actors listen to the phones.Both final sections are longer because actors move little.

    Then k-means clustering algorithm is used to build a visual word dictionary and each feature descriptor is assigned to the closest word in the vocabulary.Finally,as shown in Fig.5,a video with N frames is described as a sequence of frames with visual words:

    where,

    wnirepresents the nth visual word in the ith frame.

    3.2.Sub-action segmentation

    To segment actions efficiently,we should achieve two goals.First,ensure the sub-actions in the same sub-section of the same action class are of the same type,ignoring the speed differences between actors.Second,all sub-actions should capture enough motion information for classification.The two-stage segmentation can reach the above goals nicely, which is detailed below.

    3.2.1.Chop clips

    In first segmentation stage,the entire video is chopped into normalized clips with approximately equal numbers of feature points.The kth clip ends up in the xkth frame,xkshould satisfy the equation below:

    where Ncis the clip number.

    Here we use point density to chop clips for further segmentation instead of using number of frames to ensure there is enough motion information in all clips.Generally,dense feature points infer strenuous action.The intensity of the action may be changed over the entire action process.Some of the action primitives maybe moderate,so there could be few interesting points in these frames and result in insufficient information.Moreover,it also balances the speed difference between different actors or actions.We compute a histogram of spatial-temporal word occurrence for each video clip,the kth clip is described as:

    3.2.2.Segment sub-actions

    In second segmentation stage,motion range is calculated to segment sub-actions.The motion range is measured with χ2distance between neighbor clips.The distance from clip i to clip j is:

    Then accumulated distance of the whole action is divided into equal parts and the motion range of sub-actions is figured out:

    where Nsrepresents the number of sub-actions.T is used as the threshold to segment clip series.In fact,the distance between clips infers not only the range of motion,but also the changing extent of the action primitives'type.Finally,the sub-actions are segmented by the equation below:

    where i,j are the beginning and ending clips of the subsection.A stable segmentation over classes is achieved by concatenating the adjacent clips together.

    A segmentation example is illustrated in Fig.6,and the corresponding accumulated distance curves are shown in Fig.7.By segmenting action like this,we can eliminate speed and range differences between instances.Therefore,for a certain class,same sub-actions could be basically segmented to the same sub-section ignoring the delicate length difference between instance.Sequential BoWs is formed for each subaction respectively for weight calculation and sequential classification.All segmentation steps are shown in Algorithm 1.The algorithm focuses on sub-action segmentation to achieve equally distribution among different sections.

    4.Sub-classification and voting

    A pre-classification using training data is conducted tofigure out the weight and salience value for each sub-action. The weight shows the sub-action's discrimination with othersub-classes,while the salience indicates the sub-action's importance within it's own class.

    Fig.7.Example of accumulated χ2distance changing over time.(a)represents the upper actor in Fig.6 and(b)represents the lower one.The accumulated distance curves are similar although they are acted by different actor and rate.B,C,D,E are four equal segmenting points.Their corresponding images are shown in Fig.6. At each point,the actions are almost executed to the same ratio.

    4.1.Weight calculation

    Similar sub-actions may occur in the same sub-section of different classes,hence the result of direct voting would be poor.The pre-classification result M(i,k)represents the percentage of sub-class i be classified to sub-class k.If the jth subaction is classified as sub-class k,the probability for this subaction belonging to class i is calculated below:

    This value can be regarded as weight to eliminate the ambiguity between sub-classes.The difference between weights shows the dissimilarity between sub-classes.The smaller the difference is,the higher the similarity shows.Similar subactions give approximately equal increase to their own categoryduringvoting,sotheclassificationresultisbarelychanged.

    4.2.Saliency calculation

    The salience value for each sub-action is figured out to differ the importance of each sub-action within the class.In some sub-section,sub-actions are at low similarity so the classification accuracy would be high.We assign high salience scores to sub-actions in such sub-sections.While for some other sub-sections with large ambiguity,classification may be hard.To reduce the effect of these sub-sections,low salience scores are assigned.For a single category,we can figure out its classification accuracy in each sub-section via training sets. But at the testing time,we can not ensure the category for an action,so an average value is used as below:

    Fig.8.Examples in datasets,from top to down are UT-Interaction scene 1,UT-Interaction scene 2 and Rochester.

    Fig.9 shows an example of calculated salience maps for UT-Interaction scene-1,scene-2 and the difference between them.To each column,the salience value is proportional to the sub-action's distinctiveness.For example the first column (shake hands),the shaking parts of the action are more salient. But to the fourth column(point),the whole process is quite different from other actions,so its salience is evenly distributed.Although the environment of the two scenes is different, their salience maps are quite similar.The salience ranges from 0 to 1,and the average difference between them is 0.1777. Except the red grid on the right top of(c)caused by overact in scene-2,most difference is less than 0.2.This means our segmentation and salience calculation approach is robust to the inner-class variation and environmental change.

    Finally,category of a test instance is decided by the multiplication of those two scores Equation(11),(12)calculated above.By combining these two weights,the effect of ambiguous sub-actions in the same sub-section is weakened, the salient points at temporal dimension are stressed.We can focus the classification on the distinguishing action parts rather than being confused by those similar parts.

    5.Experiments and discussions

    In this section,the proposed classification approach is implemented and evaluated on two challenging datasets,UTInteraction[18]and Rochester[14].Fig.8 shows some actions from these two datasets.We compare our results to state-of-the-art classification approaches and confirm the advantages of our approach to distinguish similar actions.

    The segmented version of the UT-Interaction dataset contains videos of six types of human actions[3].All besides "pointing"are interactive activities and are performed by two actors.There are two sets in the dataset performed in different environment(Fig.8(a)and(b)).One is relatively simple, which is taken in a parking lot.The other is more complex with moving trees in the background.Each activity is repeated for 10 times per set by different actors,so there are totally 120 videos.The videos are taken with camera jitters and/or passerby and have been tested by several state-of-the-art approaches[11,17].Rochester dataset contains 150 videos of 10 types of actions.Each category is performed by five actors, repeated for three times in the same scenario.The inter-class ambiguities of these two datasets are both large.

    We use the cuboid feature detector[4]as local STIPs detector for its simplicity,fastness and generality[22].A 100-dimensional Dollar's gradient descriptor[4]and a 640-dimensional 3D-SIFT descriptor[21]are used respectively to describe the STIP-centered cuboids.Other feature detectors and descriptors can also be used to acquire features.For both datasets,the cuboid size is w=h=1 pixels,τ=2 frames, threshold is 0.0002 when using gradient descriptor.When using 3D-SIFT,w=h=2 pixels,τ=3 frames,the threshold is 0.0001.After extracting features from videos,k-means clustering is used to transform them into visual words.For UT-Interaction dataset,the cluster numbers in two sceneriesare 90 and 140.The cluster number in Rochester is 900. Classification is conducted using 1-NN since the performance of SVM and 1-NN is quite close[22]while 1-NN is faster.

    Fig.9.Salience maps calculated on UT-Interaction scene-1,scene-2 and the difference between them.Each grid represents a sub-action's salience.The horizontal axis represents different classes,from left to right are"shake hands","hug","kick","point","punch",and"push".The vertical axis represents sub-actions align by time.

    Fig.10.Confusion matrices for scene-1 of UT-Interaction(upper)and Rochester(lower)are shown in row one and row two.From left to right,BoWs,sequential BoWs with 3D-SIFT descriptor and sequential BoWs with gradient descriptor.K is the cluster number.

    During the segmentation,we use χ2distance to measure motion range.Theclip numberNcandthesub-actionnumber Nsare decided by iteration tests.Ncis a bit smaller than the frame number of the shortest video,Nsis associated to the action's complexity.WesetNc=140,Ns=90forUT-Interactionscnene-1,Nc=120,Ns=90forUT-Interactionscnene-2,andNc=150, Ns=20 for Rochester.Although the average video length of Rochester is longer than that of UT-Interaction,the actions are slow and some moments are even motionless,so the sub-actionnumber of Rochester is smaller.Pre-classification is done in the training set to figure outtheweight and salience.The leave-onesequence-out cross validation setting is used in all experiments. All confusion matrices are the average of 10 runs since k-means clustering is randomly initialized.

    Confusion matrices of UT-Interaction scene-1 and Rochester are shown in Fig.10.In each column,our sequential BoWs approach using different descriptors is compared with Dollar's original BoWs.The result of our approach is much better than originalBoWs.OnUT-Interactionscene-1,errorsamong"kick", "punch"and"push"are most obvious in Fig.10(a).These three actions share a lot of same sub-actions.Moreover their unique sub-actions(stick out one's arms/leg to conduct a hit/push)are notonly short but also aliketo each other.Original BoWs model does not have the ability to capture their differences.Our approachcanhighlighttheirdifferenceandsolvethisproblemin an extent.On Rochester dataset,the results of"lookup in phone book","peel banana"and"use silverware"are unsatisfactory. Our approach can decrease those errors since temporal structure and salient parts are especially considered.

    Table 1 compares the classification accuracy of our approach with state-of-the-arts on UT-Interaction and Rochester.Cluster number K is given to indicate the complexity of those approaches as there is no unified comparison approach."K"is proportional to the algorithm's dimensionality reduction ability.In our approach,Gradient descriptor[4]shows better results on UT-Interaction scene-2 and Rochester than 3D-SIFT because it extracts more feature points to describe the sub-actions more sufficiently.On UTInteraction scene-1,our result is comparable to[11,17];but manifest a faster computational speed.Noting that[11]focuses on the spatial structure between visual words,hence it can be combined with our approach.Approach in[17]aims at action prediction,and it conducts a iterate segment matching between classes and is inefficient to action classification.On both UT-Interaction scene-2 and Rochester datasets,our approach shows the best performance.

    6.Conclusions

    The typical Bag-of-Words(BoWs)model ignore the spatial and temporal structure information of visual words,which brings ambiguities among similar actions.In order to deal with it,in this paper,we present a novel approach called sequential BoWs for efficient human action classification.It reduces the ambiguity between classes sharing similar sub-actions andcaptures the action's temporal sequential structure by segmenting the action into pieces.

    Table 1Comparing proposed approach with state-of-the-arts.

    Each sub-action has a tiny movement within a narrow range of action.Then the sequential BoWs are created,in which each sub-action is assigned with a certain weight and salience to highlight the distinguishing sections.It is noted that the weight and salience are figured out in advance according to the sub-actions discrimination evaluated by training data.Salient pieces are stressed by assigning them higher weight and salience.Finally,those sub-actions are used for classification respectively,and voting for united result.Since our approach does not operate on the descriptors directly,it can be combined with many mid-level descriptors.In experiments,the proposed approach is compared with the state-of-the-arts on two challenging datasets,UT-interaction and Rochester datasets.The results demonstrate its higher robustness and accuracy overmoststate-of-the-artclassification approaches especially on complex datasets with cluttered backgrounds and inter-class action ambiguities.

    Acknowledgments

    TheauthorsacknowledgethefinancialsupportbytheNational NaturalScienceFoundationofChina(NSFC,No.61340046),the National High Technology Research and Development Program ofChina(863 Program,No.2006AA04Z247),the Scientific and Technical Innovation Commission of Shenzhen Municipality (No.JCYJ20120614152234873,JCYJ20130331144716089) and the Specialized Research Fund for the Doctoral Program of Higher Education(SRFDP,No.20130001110011).The authors wish to thank the anonymous reviewers for their helpful comments and suggestions.

    [1]S.Ali,A.Basharat,M.Shah,Chaotic invariants for human action recognition,in:IEEE International Conference on Computer Vision (ICCV),2007,pp.1-8.

    [2]M.Blank,L.Gorelick,E.Shechtman,M.Irani,R.Basri,Actions as space-time shapes,in:IEEE International Conference on Computer Vision(ICCV)vol.2,2005,pp.1395-1402.

    [3]J.M.Chaquet,E.J.Carmona,A.Fern andez-Caballero,Elsevier Comput. Vis.Image Underst.(CVIU)117(6)(2013)633-659.

    [4]P.Doll ar,V.Rabaud,G.Cottrell,S.Belongie,Behavior recognition via sparse spatio-temporal features,in:IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance,2005,pp.65-72.

    [5]A.Efros,A.C.Berg,G.Mori,J.Malik,et al.,Recognizing action at a distance,in:IEEE International Conference on Computer Vision(ICCV), 2003,pp.726-733.

    [6]T.Glaser,L.Zelnik-Manor,Incorporating temporal context in bag-ofwords models,in:IEEE International Conference on Computer Vision Workshops(ICCVW),2011,pp.1562-1569.

    [7]M.Heikkil a,M.Pietik ainen,C.Schmid,Springer Comput.Vis.Graph. Image Process.(2006)58-69.

    [8]A.Kovashka,K.Grauman,Learning a hierarchy of discriminative spacetime neighborhood features for human action recognition,in:IEEE Conference on Computer Vision and Pattern Recognition(CVPR),2010, pp.2046-2053.

    [9]I.Laptev,Springer Int.J.Comput.Vis.(IJCV)64(2-3)(2005) 107-123.

    [10]I.Laptev,M.Marsza?ek,C.Schmid,B.Rozenfeld,Learning realistic human actions from movies,in:IEEE Conference on Computer Vision and Pattern Recognition(CVPR),2008,pp.1-8.

    [11]H.Liu,M.Liu,Q.Sun,Learning directional co-occurrence for human action classification,in:IEEE International Conference on Acoustics, Speech and Signal Processing(ICASSP),2014,pp.1235-1239.

    [12]J.Liu,M.Shah,Learning human actions via information maximization, in:IEEE Conference on Computer Vision and Pattern Recognition (CVPR),2008,pp.1-8.

    [13]R.Mattivi,L.Shao,Springer Comput.Anal.Images Patterns(CAIP) (2009)740-747.

    [14]R.Messing,C.Pal,H.Kautz,Activity recognition using the velocity histories of tracked keypoints,in:IEEE International Conference on Computer Vision(ICCV),2009,pp.104-111.

    [15]T.Ojala,M.Pietik ainen,T.M aenp a a,IEEE Trans.Pattern Anal.Mach. Intell.(TPAMI)24(7)(2002)971-987.

    [16]R.Poppe,Elsevier Image Vis.Comput.(IVC)28(6)(2010)976-990.

    [17]M.Ryoo,Human activity prediction:early recognition of ongoing activities from streaming videos,in:IEEE International Conference on Computer Vision(ICCV),2011,pp.1036-1043.

    [18]M.Ryoo,C.-C.Chen,J.Aggarwal,A.Roy-Chowdhury,An overview of contest on semantic description of human activities(sdha),in:Springer Recognizing Patterns in Signals,Speech,Images and Videos,2010,pp. 270-285.

    [19]S.Satkin,M.Hebert,Modelingthetemporalextentofactions,in:Springer European Conference on Computer Vision(ECCV),2010,pp.536-548.

    [20]S.Savarese,A.DelPozo,J.C.Niebles,L.Fei-Fei,Spatial-temporal correlatons for unsupervised action classification,in:IEEE Workshop on Motion and Video Computing(WMVC),2008,pp.1-8.

    [21]P.Scovanner,S.Ali,M.Shah,A 3-dimensional sift descriptor and its application to action recognition,in:ACM International Conference on Multimedia(ACM MM),2007,pp.357-360.

    [22]L.Shao,R.Mattivi,Feature detector and descriptor evaluation in human action recognition,in:ACM International Conference on Image and Video Retrieval,2010,pp.477-484.

    [23]Q.Sun,H.Liu,Action disambiguation analysis using normalized googlelike distance correlogram,in:Springer Asian Conference on Computer Vision(ACCV),2013,pp.425-437.

    [24]Q.Sun,H.Liu,Learning spatio-temporal co-occurrence correlograms for efficient human action classification,in:IEEE International Conference on Image Processing(ICIP),2013,pp.3220-3224.

    [25]G.Willems,T.Tuytelaars,L.Van Gool,An efficient dense and scaleinvariant spatio-temporal interest point detector,in:Springer European Conference on Computer Vision(ECCV),2008,pp.650-663.

    Prof.HongLiureceivedthePh.D.degreeinmechanical electronics and automation in 1996,and serves as a Full Professor in the School of EE&CS,Peking University (PKU),China.Prof.Liu has been selected as Chinese InnovationLeadingTalentsupportedby"NationalHighlevel Talents Special Support Plan"since 2013.He is also the Director of Open Lab on Human Robot Interaction,PKU,hisresearchfields include computervision and robotics,image processing,and pattern recognition. Dr.Liu has published more than 150 papers and gained Chinese National Aero-space Award,Wu Wenjun Award on Artificial Intelligence,Excellence Teaching Award,and Candidates of TopTenOutstandingProfessorsinPKU.HeisanIEEEmember,vicepresidentof ChineseAssociationforArtificialIntelligent(CAAI),andvicechairofIntelligent RoboticsSociety ofCAAI.He has servedas keynote speakers,co-chairs,session chairs,or PC members of many important international conferences,such as IEEE/RSJ IROS,IEEE ROBIO,IEEE SMC and IIHMSP,recently also serves as reviewers for many international journals such as Pattern Recognition,IEEE Trans.on Signal Processing,and IEEE Trans.on PAMI.

    Hao Tangreceived the B.E.degree in electronics and information engineering in 2013 and is working toward the Master degree at the School of Electronics and Computer Engineering,Peking University,China. His current research interests are image classification, hand gesture recognition,gender recognition,image retrieval,action recognition and deep learning.He has published several articles in ACM Multimedia Conference(MM),IEEE International Conference on Image Processing(ICIP)and International Joint Conference on Artificial Intelligence(IJCAI).

    Wei Xiaoreceived his Ph.D in computer science and technology in Tsinghua University,China.He is currently a postdoctoral researcher on Human-Robot Interaction(HRI)in Peking University,China.His research interests include computer vision and HRI. He has published several articles in IEEE International Conference on Multi-sensor Fusion and Integration for Intelligent Systems and ACM Multimedia Conference (MM).

    Ziyi Guoreceived her B.E.degree in digital media technology in 2014,and is working toward the Master Degree in the School of software and microelectronics,Peking University,China.Her research interests include interactive media technologies, interaction design and human action recognition.

    Lu Tianreceived her master degree in computer science and technology in Human-Robot Interaction (HRI)Lab,Peking University Shenzhen Graduate School,China.Her research interest is mainly about human action recognition.She has published articles in IEEE International Conference on Image Processing(ICIP).

    Yuan Gaoreceived the B.E.degree in intelligent science and technology from Xidian University in 2012.Then he obtained the M.S.degree in computer applied technology from Peking University in 2015. Currently,he is working toward the Doctor degree in Christian-Albrechts-University of Kiel,Germany.His research interests include object detection,3D reconstruction,facial expression and gender recognition.He has published articles in IEEE International Conference on Image Processing(ICIP).

    *Corresponding author.School of Electronic and Computer Engineering, Peking University,Shenzhen Graduate Shcool,518055,China.

    E-mail addresses:hongliu@pku.edu.cn(H.Liu),haotang@sz.pku.edu.cn (H.Tang),xiaoweipkusz@pkusz.edu.cn(W.Xiao),1401210562@pku.edu.cn (Z.Guo),lutian@sz.pku.edu.cn(L.Tian),yuan.gao@stu.uni-kiel.de(Y.Gao). Peer review under responsibility of Chongqing University of Technology.

    http://dx.doi.org/10.1016/j.trit.2016.10.001

    2468-2322/Copyright?2016,Chongqing University of Technology.Production and hosting by Elsevier B.V.This is an open access article under the CC BY-NCND license(http://creativecommons.org/licenses/by-nc-nd/4.0/).

    Copyright?2016,Chongqing University of Technology.Production and hosting by Elsevier B.V.This is an open access article under the CC BY-NC-ND license(http://creativecommons.org/licenses/by-nc-nd/4.0/).

    欧美激情极品国产一区二区三区 | 精品人妻熟女av久视频| 亚洲欧洲日产国产| 日韩强制内射视频| 国产黄片视频在线免费观看| 搡老乐熟女国产| 亚洲情色 制服丝袜| 精品一区二区免费观看| 少妇丰满av| 人人妻人人澡人人看| 日韩大片免费观看网站| 精品久久久久久电影网| 精品一区二区免费观看| 自拍欧美九色日韩亚洲蝌蚪91 | 国产精品一区www在线观看| 亚洲人与动物交配视频| 国产 精品1| 2021少妇久久久久久久久久久| 最近2019中文字幕mv第一页| 亚洲成人一二三区av| 中文字幕亚洲精品专区| 十八禁高潮呻吟视频 | 欧美精品一区二区免费开放| 69精品国产乱码久久久| 精品久久久久久久久亚洲| 久久久久久久久大av| 亚洲综合色惰| 精品久久久久久电影网| 色哟哟·www| 在线观看国产h片| 天堂8中文在线网| 色视频在线一区二区三区| 国产免费视频播放在线视频| 精品少妇久久久久久888优播| 一二三四中文在线观看免费高清| 99九九线精品视频在线观看视频| 日本爱情动作片www.在线观看| 天美传媒精品一区二区| 久久精品久久精品一区二区三区| 六月丁香七月| 久久午夜福利片| 国产成人a∨麻豆精品| 我要看黄色一级片免费的| 简卡轻食公司| 男人爽女人下面视频在线观看| a级毛色黄片| 人妻系列 视频| 亚洲精品日本国产第一区| 国产精品99久久99久久久不卡 | 成人国产麻豆网| 色吧在线观看| 欧美国产精品一级二级三级 | 在线播放无遮挡| 伦理电影大哥的女人| 日本黄色日本黄色录像| 又黄又爽又刺激的免费视频.| 夜夜爽夜夜爽视频| 国产毛片在线视频| 欧美日韩综合久久久久久| 三级国产精品欧美在线观看| 22中文网久久字幕| 国产一区二区三区av在线| 菩萨蛮人人尽说江南好唐韦庄| 色吧在线观看| 一级毛片电影观看| 美女福利国产在线| 久久国内精品自在自线图片| 亚洲精品成人av观看孕妇| 国产精品一二三区在线看| 特大巨黑吊av在线直播| 久久精品国产亚洲av天美| 老女人水多毛片| 久久婷婷青草| 久久亚洲国产成人精品v| 亚洲经典国产精华液单| 久久人妻熟女aⅴ| 午夜日本视频在线| 日本午夜av视频| 精品国产乱码久久久久久小说| 99热网站在线观看| 久久久久视频综合| 大片电影免费在线观看免费| 国产精品三级大全| 一级二级三级毛片免费看| 久久国产亚洲av麻豆专区| 国内精品宾馆在线| 人妻系列 视频| 自线自在国产av| 大又大粗又爽又黄少妇毛片口| 欧美+日韩+精品| 国产av精品麻豆| 欧美精品一区二区免费开放| 精品久久久久久电影网| 国产视频首页在线观看| 久久青草综合色| 日日摸夜夜添夜夜爱| 国产美女午夜福利| 久久久久久久久久久丰满| 亚洲三级黄色毛片| 亚洲国产精品999| 99视频精品全部免费 在线| 一级,二级,三级黄色视频| 99久久中文字幕三级久久日本| 亚洲av男天堂| 国国产精品蜜臀av免费| av国产精品久久久久影院| 最近手机中文字幕大全| 一级黄片播放器| 国产日韩欧美视频二区| 亚洲精品乱码久久久v下载方式| 欧美97在线视频| 亚洲欧美成人精品一区二区| 亚洲性久久影院| 久久久国产精品麻豆| 在线观看av片永久免费下载| 午夜av观看不卡| www.色视频.com| 国产国拍精品亚洲av在线观看| 校园人妻丝袜中文字幕| 99久久精品国产国产毛片| 三级国产精品片| 内地一区二区视频在线| 青春草视频在线免费观看| 国产一区二区三区综合在线观看 | 亚洲国产精品成人久久小说| 超碰97精品在线观看| 久久久久久久亚洲中文字幕| 精品视频人人做人人爽| av有码第一页| 1024视频免费在线观看| 日本一区二区免费在线视频| 日本猛色少妇xxxxx猛交久久| 青青草视频在线视频观看| 国产成人精品久久二区二区免费| 国产老妇伦熟女老妇高清| 秋霞在线观看毛片| 999久久久国产精品视频| 精品一品国产午夜福利视频| 人妻一区二区av| 91av网站免费观看| 欧美日韩视频精品一区| 老司机福利观看| 国产99久久九九免费精品| 搡老熟女国产l中国老女人| 午夜激情av网站| 亚洲欧美色中文字幕在线| 捣出白浆h1v1| 最近最新免费中文字幕在线| 国产97色在线日韩免费| 国产精品一区二区免费欧美 | 老汉色av国产亚洲站长工具| 满18在线观看网站| 国产一区二区三区av在线| 成年女人毛片免费观看观看9 | videos熟女内射| 欧美日本中文国产一区发布| 亚洲av欧美aⅴ国产| 日本撒尿小便嘘嘘汇集6| www.熟女人妻精品国产| 真人做人爱边吃奶动态| 又紧又爽又黄一区二区| 菩萨蛮人人尽说江南好唐韦庄| 国产亚洲欧美精品永久| 一级毛片精品| 欧美日韩精品网址| 欧美av亚洲av综合av国产av| 亚洲精品粉嫩美女一区| 天天操日日干夜夜撸| 99热网站在线观看| 国产精品一区二区在线不卡| 色播在线永久视频| 欧美亚洲日本最大视频资源| 午夜福利一区二区在线看| 免费观看av网站的网址| 别揉我奶头~嗯~啊~动态视频 | 国产免费av片在线观看野外av| 欧美激情高清一区二区三区| 波多野结衣一区麻豆| 夫妻午夜视频| 欧美日韩国产mv在线观看视频| 午夜两性在线视频| 国产精品久久久av美女十八| videosex国产| 69精品国产乱码久久久| svipshipincom国产片| 国产精品久久久久久精品电影小说| 亚洲精品国产区一区二| 在线观看人妻少妇| 国产在线视频一区二区| 国产精品 国内视频| 中文字幕色久视频| 中文字幕高清在线视频| 精品国产超薄肉色丝袜足j| 80岁老熟妇乱子伦牲交| 日韩视频一区二区在线观看| 亚洲美女黄色视频免费看| 久久午夜综合久久蜜桃| 美国免费a级毛片| 黄色 视频免费看| h视频一区二区三区| 国产男女超爽视频在线观看| 青草久久国产| 久久九九热精品免费| 国产精品国产av在线观看| 日韩有码中文字幕| 国产成人精品在线电影| 久久精品aⅴ一区二区三区四区| 欧美国产精品va在线观看不卡| 亚洲av电影在线观看一区二区三区| 女性被躁到高潮视频| 国产一级毛片在线| 亚洲精品一二三| a级毛片在线看网站| 亚洲第一欧美日韩一区二区三区 | 波多野结衣一区麻豆| 法律面前人人平等表现在哪些方面 | 亚洲av国产av综合av卡| 国产男女内射视频| 亚洲精品国产av成人精品| 国产一级毛片在线| 日韩,欧美,国产一区二区三区| 在线观看免费日韩欧美大片| 免费av中文字幕在线| 一本综合久久免费| 一个人免费看片子| 国产一区二区三区在线臀色熟女 | 国产一区二区 视频在线| 欧美精品亚洲一区二区| 婷婷成人精品国产| 国产精品久久久人人做人人爽| 国产成人精品久久二区二区91| 狂野欧美激情性xxxx| 久久精品国产综合久久久| 黑人巨大精品欧美一区二区蜜桃| 99re6热这里在线精品视频| 欧美日韩成人在线一区二区| 久久久久精品国产欧美久久久 | 一区二区日韩欧美中文字幕| 男女之事视频高清在线观看| 一本—道久久a久久精品蜜桃钙片| 久久精品国产综合久久久| 成年av动漫网址| 亚洲成人国产一区在线观看| 丝袜人妻中文字幕| 久久影院123| 国产精品免费视频内射| 精品人妻一区二区三区麻豆| 成人亚洲精品一区在线观看| 一进一出抽搐动态| 真人做人爱边吃奶动态| 老熟妇仑乱视频hdxx| 亚洲国产日韩一区二区| 久久免费观看电影| 国产激情久久老熟女| 美女国产高潮福利片在线看| 亚洲第一av免费看| 纵有疾风起免费观看全集完整版| 成年女人毛片免费观看观看9 | 美女中出高潮动态图| 99国产精品一区二区蜜桃av | 精品福利永久在线观看| 精品一区在线观看国产| 精品国产超薄肉色丝袜足j| 欧美黑人欧美精品刺激| 亚洲av国产av综合av卡| 制服人妻中文乱码| 91九色精品人成在线观看| 十八禁高潮呻吟视频| 91大片在线观看| 老司机影院成人| 日韩欧美一区二区三区在线观看 | 中文字幕高清在线视频| 男人舔女人的私密视频| 99国产精品免费福利视频| 在线观看免费午夜福利视频| 99久久综合免费| 国产日韩欧美亚洲二区| 欧美激情 高清一区二区三区| 免费观看av网站的网址| 麻豆国产av国片精品| 欧美黄色淫秽网站| 欧美日韩国产mv在线观看视频| 欧美黑人精品巨大| 啪啪无遮挡十八禁网站| 高潮久久久久久久久久久不卡| 纯流量卡能插随身wifi吗| 精品人妻熟女毛片av久久网站| 免费在线观看黄色视频的| 美国免费a级毛片| 国产xxxxx性猛交| 中文欧美无线码| 啦啦啦中文免费视频观看日本| 乱人伦中国视频| 日韩,欧美,国产一区二区三区| 精品熟女少妇八av免费久了| 中国国产av一级| 99久久人妻综合| 黄频高清免费视频| 法律面前人人平等表现在哪些方面 | videos熟女内射| 亚洲av日韩在线播放| √禁漫天堂资源中文www| 妹子高潮喷水视频| 成年人免费黄色播放视频| 菩萨蛮人人尽说江南好唐韦庄| 99国产精品99久久久久| 精品国产超薄肉色丝袜足j| 久久女婷五月综合色啪小说| 两人在一起打扑克的视频| 欧美日韩av久久| 亚洲,欧美精品.| 精品国产乱码久久久久久小说| 91精品伊人久久大香线蕉| 久久久国产欧美日韩av| 久久人妻熟女aⅴ| 伦理电影免费视频| 国产精品久久久久久精品古装| 欧美日韩中文字幕国产精品一区二区三区 | av一本久久久久| 欧美亚洲日本最大视频资源| 香蕉丝袜av| 正在播放国产对白刺激| 婷婷成人精品国产| 丝袜美腿诱惑在线| 成人三级做爰电影| 动漫黄色视频在线观看| 国产精品一区二区在线观看99| 99久久国产精品久久久| 免费久久久久久久精品成人欧美视频| 久久久久网色| 1024香蕉在线观看| 一区在线观看完整版| 国产成人免费无遮挡视频| 亚洲少妇的诱惑av| 免费女性裸体啪啪无遮挡网站| 夫妻午夜视频| 午夜两性在线视频| 亚洲av日韩精品久久久久久密| 亚洲第一欧美日韩一区二区三区 | 国产av国产精品国产| 亚洲第一av免费看| 香蕉国产在线看| 女人爽到高潮嗷嗷叫在线视频| 久久精品aⅴ一区二区三区四区| 中文字幕人妻熟女乱码| 99九九在线精品视频| 蜜桃在线观看..| 日韩有码中文字幕| 欧美性长视频在线观看| 免费在线观看完整版高清| 日韩一卡2卡3卡4卡2021年| a 毛片基地| 日韩三级视频一区二区三区| 久久久久视频综合| 建设人人有责人人尽责人人享有的| 国产欧美亚洲国产| 91麻豆av在线| 999精品在线视频| 免费观看人在逋| 人人澡人人妻人| 亚洲国产毛片av蜜桃av| 18在线观看网站| 99国产精品99久久久久| 日韩精品免费视频一区二区三区| 这个男人来自地球电影免费观看| 久久国产亚洲av麻豆专区| 国产又爽黄色视频| 香蕉国产在线看| 成年人午夜在线观看视频| 亚洲第一青青草原| 日本vs欧美在线观看视频| 99热全是精品| 最近中文字幕2019免费版| www.熟女人妻精品国产| 999久久久国产精品视频| 国产精品久久久久久人妻精品电影 | av网站免费在线观看视频| 亚洲 欧美一区二区三区| 日韩视频在线欧美| 免费在线观看黄色视频的| a级毛片黄视频| 十八禁网站网址无遮挡| 大片电影免费在线观看免费| 纯流量卡能插随身wifi吗| 啪啪无遮挡十八禁网站| 美女福利国产在线| 久久狼人影院| 在线观看免费午夜福利视频| 男人爽女人下面视频在线观看| 欧美日本中文国产一区发布| 婷婷丁香在线五月| 亚洲国产欧美日韩在线播放| 免费看十八禁软件| 在线观看舔阴道视频| 1024香蕉在线观看| 欧美黄色淫秽网站| 一边摸一边做爽爽视频免费| 国产成人精品在线电影| www.熟女人妻精品国产| 午夜福利视频在线观看免费| 久久久久国内视频| 精品卡一卡二卡四卡免费| 国产欧美日韩一区二区三 | 一级毛片精品| 国产日韩欧美在线精品| 精品国产乱码久久久久久小说| 一级a爱视频在线免费观看| 国产精品自产拍在线观看55亚洲 | 欧美激情高清一区二区三区| 啦啦啦中文免费视频观看日本| 成年动漫av网址| 一本—道久久a久久精品蜜桃钙片| 老汉色∧v一级毛片| 国产一区二区三区在线臀色熟女 | 国产黄色免费在线视频| 又黄又粗又硬又大视频| 五月开心婷婷网| 国产人伦9x9x在线观看| av在线app专区| 国产熟女午夜一区二区三区| 精品福利观看| 大香蕉久久网| 高清在线国产一区| 极品少妇高潮喷水抽搐| 午夜精品久久久久久毛片777| 亚洲精品成人av观看孕妇| 伊人亚洲综合成人网| 国产免费一区二区三区四区乱码| 美女高潮喷水抽搐中文字幕| 777米奇影视久久| 天天躁夜夜躁狠狠躁躁| 亚洲精品久久久久久婷婷小说| 国产成+人综合+亚洲专区| av视频免费观看在线观看| 一本色道久久久久久精品综合| 久久久久久人人人人人| 一级,二级,三级黄色视频| 亚洲va日本ⅴa欧美va伊人久久 | 丝袜人妻中文字幕| 无限看片的www在线观看| 大香蕉久久网| 成人18禁高潮啪啪吃奶动态图| 午夜影院在线不卡| 亚洲国产av影院在线观看| 超碰97精品在线观看| 视频在线观看一区二区三区| 五月天丁香电影| 久久久久精品人妻al黑| 搡老乐熟女国产| h视频一区二区三区| 国产成人欧美在线观看 | 欧美另类一区| 无遮挡黄片免费观看| 亚洲成国产人片在线观看| 日韩电影二区| 久久精品久久久久久噜噜老黄| 啦啦啦啦在线视频资源| 交换朋友夫妻互换小说| 一本久久精品| 女人久久www免费人成看片| 人人澡人人妻人| 久久久精品免费免费高清| 午夜激情久久久久久久| 99久久人妻综合| 国产熟女午夜一区二区三区| 精品少妇一区二区三区视频日本电影| 成年美女黄网站色视频大全免费| 99久久综合免费| 老汉色av国产亚洲站长工具| 亚洲精品国产色婷婷电影| 搡老岳熟女国产| 色婷婷av一区二区三区视频| 妹子高潮喷水视频| 久久久久国内视频| 水蜜桃什么品种好| 日韩欧美免费精品| 亚洲av片天天在线观看| 亚洲久久久国产精品| 久久久久精品国产欧美久久久 | 精品福利永久在线观看| 国产亚洲精品一区二区www | 亚洲精品一卡2卡三卡4卡5卡 | 男女国产视频网站| 免费少妇av软件| 一本色道久久久久久精品综合| 嫩草影视91久久| 亚洲欧美成人综合另类久久久| 久久久久久久大尺度免费视频| 嫁个100分男人电影在线观看| 五月天丁香电影| 亚洲精品中文字幕一二三四区 | 少妇 在线观看| 人成视频在线观看免费观看| 在线 av 中文字幕| 91av网站免费观看| 丰满少妇做爰视频| 国产精品一区二区在线观看99| 日日摸夜夜添夜夜添小说| 日韩三级视频一区二区三区| 我的亚洲天堂| 深夜精品福利| 成人影院久久| 国产不卡av网站在线观看| 黑人猛操日本美女一级片| 精品一品国产午夜福利视频| 18禁裸乳无遮挡动漫免费视频| 黑人巨大精品欧美一区二区蜜桃| 精品熟女少妇八av免费久了| 丝袜在线中文字幕| 久久久国产成人免费| 蜜桃国产av成人99| 如日韩欧美国产精品一区二区三区| 久久久久国产精品人妻一区二区| 伊人亚洲综合成人网| 亚洲一区中文字幕在线| 免费观看人在逋| 又黄又粗又硬又大视频| 狠狠婷婷综合久久久久久88av| 两性午夜刺激爽爽歪歪视频在线观看 | 黄网站色视频无遮挡免费观看| 超色免费av| 亚洲熟女毛片儿| 欧美亚洲 丝袜 人妻 在线| 一级黄色大片毛片| 亚洲激情五月婷婷啪啪| 黑丝袜美女国产一区| 久久99热这里只频精品6学生| 久久久水蜜桃国产精品网| 午夜福利视频精品| 99国产综合亚洲精品| 中文字幕精品免费在线观看视频| 狠狠狠狠99中文字幕| 国产老妇伦熟女老妇高清| 亚洲精华国产精华精| 国产日韩欧美视频二区| 欧美精品一区二区免费开放| 动漫黄色视频在线观看| 亚洲一卡2卡3卡4卡5卡精品中文| 好男人电影高清在线观看| 老司机午夜十八禁免费视频| 新久久久久国产一级毛片| 性少妇av在线| 无遮挡黄片免费观看| 少妇的丰满在线观看| 美女午夜性视频免费| 色综合欧美亚洲国产小说| 日韩 欧美 亚洲 中文字幕| 亚洲精品国产一区二区精华液| 美女午夜性视频免费| 欧美国产精品一级二级三级| 亚洲视频免费观看视频| 久久综合国产亚洲精品| www.999成人在线观看| 国产成人av激情在线播放| 国产精品二区激情视频| 啦啦啦在线免费观看视频4| 国产亚洲欧美在线一区二区| netflix在线观看网站| 大陆偷拍与自拍| 搡老熟女国产l中国老女人| 久久中文字幕一级| 欧美日韩国产mv在线观看视频| 叶爱在线成人免费视频播放| 亚洲精品国产av成人精品| 亚洲欧美一区二区三区黑人| 手机成人av网站| 十八禁人妻一区二区| 妹子高潮喷水视频| 中文字幕人妻熟女乱码| 亚洲专区中文字幕在线| 国产一区二区在线观看av| 天天躁狠狠躁夜夜躁狠狠躁| 大码成人一级视频| 久久狼人影院| 搡老乐熟女国产| 黄色视频不卡| 搡老熟女国产l中国老女人| 纵有疾风起免费观看全集完整版| 97精品久久久久久久久久精品| 嫁个100分男人电影在线观看| 亚洲精品国产区一区二| 亚洲人成电影免费在线| 国产深夜福利视频在线观看| 亚洲免费av在线视频| 国产xxxxx性猛交| 欧美大码av| 亚洲精品粉嫩美女一区| 久久精品国产综合久久久| 久久99一区二区三区| 精品少妇一区二区三区视频日本电影| 国产欧美日韩一区二区精品| 乱人伦中国视频| 可以免费在线观看a视频的电影网站| av欧美777| 久久综合国产亚洲精品| 90打野战视频偷拍视频| 超碰成人久久| 亚洲一区二区三区欧美精品| 欧美日韩黄片免| 国产欧美日韩综合在线一区二区| 在线观看一区二区三区激情| 国产高清国产精品国产三级| 在线天堂中文资源库| 女人爽到高潮嗷嗷叫在线视频| 伊人亚洲综合成人网| 超碰97精品在线观看| av片东京热男人的天堂| av网站免费在线观看视频| 99久久精品国产亚洲精品| 久久精品aⅴ一区二区三区四区| 久久青草综合色| 久久中文字幕一级| 一级黄色大片毛片| 在线观看免费午夜福利视频| 亚洲欧美色中文字幕在线|