• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    DTHN:Dual-Transformer Head End-to-End Person Search Network

    2023-12-12 15:49:42ChengFengDezhiHanandChongqingChen
    Computers Materials&Continua 2023年10期

    Cheng Feng,Dezhi Han and Chongqing Chen

    School of Information Engineering,Shanghai Maritime University,Shanghai,201306,China

    ABSTRACT Person search mainly consists of two submissions,namely Person Detection and Person Re-identification (re-ID).Existing approaches are primarily based on Faster R-CNN and Convolutional Neural Network(CNN)(e.g.,ResNet).While these structures may detect high-quality bounding boxes,they seem to degrade the performance of re-ID.To address this issue,this paper proposes a Dual-Transformer Head Network(DTHN)for end-to-end person search,which contains two independent Transformer heads,a box head for detecting the bounding box and extracting efficient bounding box feature,and a re-ID head for capturing high-quality re-ID features for the re-ID task.Specifically,after the image goes through the ResNet backbone network to extract features,the Region Proposal Network(RPN)proposes possible bounding boxes.The box head then extracts more efficient features within these bounding boxes for detection.Following this,the re-ID head computes the occluded attention of the features in these bounding boxes and distinguishes them from other persons or backgrounds.Extensive experiments on two widely used benchmark datasets,CUHK-SYSU and PRW,achieve state-of-the-art performance levels,94.9 mAP and 95.3 top-1 scores on the CUHK-SYSU dataset,and 51.6 mAP and 87.6 top-1 scores on the PRW dataset,which demonstrates the advantages of this paper’s approach.The efficiency comparison also shows our method is highly efficient in both time and space.

    KEYWORDS Transformer;occluded attention;end-to-end person search;person detection;person re-ID;Dual-Transformer Head

    1 Introduction

    Person search aims to localize a specific target person from the gallery set,which means it contains two submissions,Person Detection,and Person re-ID.Depending on these two different submissions,existing work can be divided into two-step and end-to-end methods.Two-step methods[1–6]treat them separately by conducting re-ID[7–10]on cropped person patches found by a standalone person box detector.They trade time and resource consumption for better performance,as shown in Fig.1a.

    By comparison,in a multi-task framework,end-to-end methods [11–17] effectively tackle both detection and re-ID simultaneously,as seen in Fig.1b.These approaches commonly utilize a person detector (e.g.,Faster R-CNN [18],RetinaNet [19],or FCOS [20]) for detection and then feed the feature into re-ID branches.To address the issue caused by the parallel structure of Faster R-CNN,Li et al.[12]proposed SeqNet to perform detection and re-ID sequentially for extracting high-quality features and achieving superior re-ID performance.Yu[17]introduced COAT to solve the imbalance between detection and re-ID by learning pose/scale-invariant features in a coarse-to-fine manner and achieving improved performance.However,end-to-end methods still suffer from several challenges:

    ■Handing occlusions with background objects or partial appearance poses a significant challenge.The detection and correct re-ID of persons become more challenging when they are obscured by objects or positioned at the edges of the captured image.While current models may perform well in person search,they are prone to failure in complex occlusion situations.

    ■The significant scale of pose variations makes it complicated to re-ID.Since current models mainly utilize CNN to extract re-ID features,they tend to suffer from the scale of pose variations due to inconsistent perceptual fields,which degrades the re-ID performance.

    ■Efficient re-ID feature extraction remains a thorny problem.Existing methods either re-ID first or detection first,but still leave the unsolved issue of how to efficiently extract the re-ID feature for better performance.

    Figure 1:Classification and comparison of two person search network

    For such cases,we propose a Dual-Transformer Head End-to-End Person Search Network(DTHN)method to address the above limitations.First,inspired by SeqNet,an additional Faster RCNN head is used as an enhanced RPN to provide high-quality bounding boxes.Then a Transformerbased box head is utilized to efficiently extract box features to perform high-accuracy detection.Next,a Transformer-based re-ID head is employed to efficiently obtain the re-ID representation from the bounding boxes.Moreover,we randomly mix up partial tokens of instances in a mini-batch to learn the cross-attention.Compared to previous works that have difficulty dealing with the balance issue between detection and re-ID,DTHN can achieve high detection accuracy without degrading re-ID performance.

    The main contributions of this paper are as follows:

    ■we propose a Dual-Transformer Head End-to-End Person Search Network,refining the box and re-ID feature extraction problem previous end-to-end frameworks were limited.The performance is improved by designing a Dual-Transformer Head structure containing two independent Transformer heads for handling high-quality bounding box feature extraction and high-quality re-ID feature extraction,respectively.

    ■we improve the end-to-end person search efficiency by using a Dual-Transformer Head instead of traditional CNN,reducing the number of parameters and remain a comparable accuracy.By employing the occlusion attention mechanism,the network can learn person features under occlusion,which substantially improves the performance of the re-ID in small-scale person and occlusion situations.

    ■we validate the effectiveness of our approach by achieving state-of-the-art performance on two widely used datasets,CUHK-SYSU and PRW.94.9 mAP and 95.3 top-1 scores were achieved on the CUHK-SYSU dataset,and 51.6 mAP and 87.6 top-1 scores were achieved on the PRW dataset.

    The remainder of this paper is organized as follows: Section 2 presents the research related to this work in recent years;Section 3 reviews the relative preparatory knowledge and presents the proposed DTHN design in detail;Section 4 presents some relevant experimental setups and verifies the effectiveness of the proposed method through experiments;Section 5 summarizes this work and provides an outlook for feature work.

    2 Related Work

    2.1 Person Search

    Person search has received increasing attention since the release of CUHK-SYSU and PRW,two large-scale datasets.This development marked a shift in researchers’approach to person search,as they began viewing it as a holistic task instead of treating it separately.The early solutions were two-step methods,using a person detector or manually constructing the person box,then constructing a person re-ID model to search for targets in the gallery.With high performance comes high time and resource consumption,two-step methods tend to consume more computational resources and time to perform at the same level as end-to-end methods.End-to-end person search has attracted extensive interest due to the integrity of solving two submissions together.Li et al.[12]shared the stem representations of person detection and re-ID,solving two submissions sequentially.Yan[14]proposed the first anchorfree person search method to address the misalignment problem at different levels.Furthermore,Yu[17]presented a three-cascade framework for progressively balancing person detection and re-ID.

    2.2 Vision Transformer

    Transformer [21] was initially designed to solve problems in natural language processing.Since the release of Vision Transformer (ViT) [22],it has become popular in computer vision (CV) [23–26].This pure Transformer backbone achieves state-of-the-art performance on many CV problems and has been shown to extract multi-scale features that traditional CNNs struggle with.The re-ID process heavily relies on fine-grained features,making it a promising technology in this field.Several efforts have been made to explore the application of ViT in person re-ID.Li et al.[27]proposed the part-aware Transformer to perform occluded person re-ID through diverse part discovery.Yu [17]performed the person search with multi-scale convolutional Transformers,learning discriminative re-ID features and distinguishing people from the background in a cascade pipeline.Our paper proposes a Dual-Transformer Head for the end-to-end person search network to efficiently extract high-quality bounding boxes feature and re-ID feature.

    2.3 Attention Mechanism

    The attention mechanism plays a crucial role in the operation and function of the whole Transformer.After the proposal of ViT,numerous variants of ViT have tried to bring different features to the Transformer by changing the attention mechanism.Among them,in the target detection task,using a combination of artificial token transformations has become a mainstream approach to solve the detection of occluded targets.Based on this,Yu [17] proposed an occlusion attention module in which both positive and negative samples in the same mini-batch are randomly partially swapped to simulate the encountered background occlusion of a person,achieving good performance.This is also mainly the attention mechanism used in this paper.

    To give the reader further insight into the work in this paper,Table 1 provides a summary of the related work and the work in this paper.

    Table 1:A summary of related person search works and our work

    3 Methods

    As previously mentioned,existing end-to-end person search works still struggle with the conflict of person detection and person re-ID.Prior studies have indicated that,despite a potential decrease in detection precision,the precision of re-ID can be maintained or even improved through serialization.However,achieving a high-level detection precision results in accurate bounding box features,which are beneficial for re-ID.Thus,we propose the Dual-Transformer Head Person Search Network(DTHN)manage to get both high-quality detection and refined re-ID accuracy.

    3.1 End-to-End Person Search Network

    As shown in Fig.2,our network is based on the Faster R-CNN object detector backbone with Region Proposal Network.We start by pre-processing the image to be searched for which will be converted to a size of 800 ?1500 as a standard input.We then use the ResNet-50 [28] backbone to extract the 1024-dim backbone feature in a size of 1024 ?58 ?76,then fed it into the RPN to obtain the region proposals.During training,RoI-Align is performed using the proposals generated by RPN to obtain the features of the region of interest for bounding box search,but RoI-Align is performed using a Ground-truth bounding box during the re-ID phase.Note that instead of using ResNet-50 stage 5 (res5) as our box head,we utilize a Transformer to extract high-quality box features and get high detection accuracy,and use the predictor head of Faster R-CNN to obtain high-confidence detection boxes.The RoI-Align operation is applied to pool ah?wregion as our region of interest,we use it as the stem featureF∈Rh?w?c.Note that F has the height of h and the width ofw,andcdenotes the number of channels.We set the intersection-over-union (IoU) thresholds at 0.5 in the training phase to distinguish positive and negative samples,and 0.8 IoU in the testing phase to get high-confidence bounding boxes.Then a Transformer re-ID head is utilized to extract distinguish features from theF.In each Transformer head,we learn the feature supervised by two lossesLreg1andLreg2.WhereNpdenotes the number of positive samples,ridenotes the calculated regression ofi-th positive samples,Δidenotes the corresponding ground truth regression,andLlocdenotes the Smooth-L1-Loss.The expressions forLreg1andLreg2are identical,as shown in the equation forLregbelow.

    Figure 2:Structural framework of the DTHN,the dotted line means only happens in the testing phase

    In addition,we also calculate the classification lossLcls1,andLcls2after two transformer heads.WhereNdenotes the number of samples,pidenotes the predicted classification probability ofi-th sample,andcidenotes the ground truth label.

    Note thatLcls2and the re-ID lossLreidare two different losses calculated by the Norm-Aware Embedding(NAE)Lnae(.),wherefdenotes the extracted 256-dim features.

    3.2 Occluded Attention

    The attention mechanism plays a crucial role in the Transformer.In our application,where we aim to extract high-quality bounding boxes and re-ID features,we must address the issue of occlusion.To this end,we use occluded attention in the DTH to prompt the model to learn the occlusion feature and address it in real applications,as shown in Fig.3.Equations should be flushed to the left of the column.First,we build the token bankwherepdenotes the number of box proposals,andxidenotes the token in one mini-batch.We exchange part of the tokens with another token from the token bank according to the index,using Token-Mix-Up(TMU)function,wherexiandxjdenote the token to be handled,Rdenotes the random value generated by the system,Tdenotes the exchange threshold.

    Figure 3:The occluded attention mechanism in DTHN

    After random swapping,we transform the tokenized features into three matrices through three fully connected(FC)layers:query matrixQ,key matrixKand value matrixV,and then we compute the multi-head self-attention(MSA)as follows,where ?cdenotes the channel scale of the token,it equals,nis the number of slices during tokenization,mdenotes the number of heads MSA has:

    After MSA,we perform Feed Forward Network(FFN)to output features for feature regression,classification,and re-ID.

    3.3 Dual-Transformer Head

    The Dual-Transformer Head(DTH)consists of two individual Transformer heads designed for detection and re-ID.Although working in different parts of the network,the detection and re-ID heads share the same mechanism.The Transformer box head takes box proposals as input and generates processed features as output.In contrast,the Transformer re-ID head takes ground truth as input during the training phase but proposals during the testing phase.Therefore,we hypothesize that the quality of detection can positively impact the re-ID performance.To provide a visual representation,the structure of the DTH is visualized in Fig.4.

    Figure 4:The structure of DTH and how it works

    First,the pooled stem featureF∈Rh?w?cis fed into the Transformer box head and obtains the proposal feature,which is fed into Faster R-CNN to calculate the proposal regression and proposal classification.After that,Fis re-fed into the Transformer re-ID head and obtains box feature,which is fed into the bounding box regressor and Norm-Aware Embedding to calculate the box regression and box classification.The loss function of NAE to calculate the box classificationLcls2is shown in equation below:

    wherey∈{0,1} denotes that the box is a person or background.norm r∈[0,∞).σdenotes the sigmoid activation function,within which is a batch normalization layer.The OIM loss is calculated using the features processed by NAE.OIM only consider the labeled and unlabeled identities,while leave the other proposals untouched.OIM has two auxiliary structures,Look-Up Table(LUT)to store all feature vectors with tagged identities and Circular Queue(CQ)to store untagged identities detected in the recent mini-batch.Based on these two structures,the probability ofxbeing recognized as the identity with class-idiand thei-th unlabeled identity by two Softmax function.OIM loss is calculated as equation below as our re-ID loss.

    wheredenotes the i-th column of the LUT,denotes the i-th column of the CQ,τdenotes softer probability distribution,Exdenotes the expectation,ptdenotes the probability of being judged ast.

    We take the Transformer re-ID head as an example to demonstrate the process.After the feature has been pooled intoF∈Rh?w?c,Fwill go through the tokenization.We splitFtonslices channelwise getting∈Rh?w?c?.We utilize series convolutional layers to generate tokens based ongetting∈R?h??w??c.By flattening ?Finto tokenx∈R?h?w??c.After finishing TMU,go through the MSA and FFN mentioned above transforming each token to enhance its representation ability.The enhanced feature will be projected into the same size it gets in,Then we concatenate the features of the n scales of transformers to the original sizeh?w?c.There is a residual connection outside each transformer.After the global average pooling (GAP) layer,the feature Transformer outputs will be pooled and delivered to different loss functions according to the type of Transformer head.The internal structure of the Transformer head is shown in Fig.5.

    4 Experiment

    All training processes are conducted in PyTorch with one NVIDIA A40 GPU,while testing processes are conducted with one NVIDIA 3070Ti GPU.The origin image will go through the ResNet-50 stage 4 and be resized to 900 ?1500 as the input.The source code and implementation details can be found in https://github.com/FitzCoulson/DTHN/tree/master.

    Figure 5:The internal structure of Transformer head

    4.1 Datasets and Metrics

    We conduct our experiments on two wildly used datasets.The CUHK-SYSU dataset[13]contains images from 18184 scenes with 8432 identities and 96143 bounding boxes.The default gallery contains 2900 testing identities in 6978 images with a default size of 100.While the PRW dataset [6] collects 11816 video frames from 6 cameras with 5704 frames and 482 identities,dividing into a training set with 5705 frames and 482 identities and a testing set with 2057 query persons in 6112 frames.

    We evaluate our model following the standard evaluation metrics.According to the Cumulative Matching Characteristic (CMC),the detection box will only be considered correct when the IoU is more than 0.5.So,we use Recall and Average Precision (AP) as the performance metric for person detection.While the person re-ID uses the mean Average Precision(mAP)and top-1 scores.All the metrics the higher the better.

    whereRnandPnseparately denote the recall and precision of then-th confidence threshold,Cdenotes the number of all classifications.The top-1 score denotes the result with the highest accuracy under the classification.

    4.2 Implementation Detail

    We take ResNet-50 pre-trained on the ImageNet as the backbone.The batch size is set to 5 during training and 1 during testing.The size of theFwill be set to 14 ?14 ?1024.The number of heads m in MSA is set to 8.The loss weightλ1is set to 10,and others are set to 1.We use the SGD optimizer with a momentum of 0.9 to train 20 epochs.The initial learning rate will warm up to 0.003 during the first epoch and decrease by 10 after the 16th epoch.The CQ size of OIM is set to 5000 for CUHK-SYSU and 500 for PRW.The IoU threshold is set to 0.4 in the testing phase.

    4.3 Ablation Study

    We conducted several experiments on the PRW dataset to analyze our proposed method.As shown in Table 2,we test several combinations of different box heads and re-ID heads and evaluate their performance on the PRW dataset.

    We set the default box head and re-ID head as ResNet-50(stage 5)and conduct one experiment,follow by two experiments by setting the box head or the re-ID head to the corresponding Transformer head,respectively,and finally set both the box head and the re-ID head to the Transformer head for one experiment.As we can see from Table 2,when using ResNet-50 (stage 5) as the box head and the re-ID head,both detection and re-ID are at a moderate level.However,when we change the box head to Transformer,the detection accuracy does not improve,while the re-ID accuracy is also slightly reduced,so Transformer cannot play a good effect only for the box head.When we maintain the box head as ResNet-50(stage 5),and replace the re-ID head with Transformer,the re-ID accuracy increases significantly,which shows that Transformer can maximize information extracted from the feature for re-ID.Finally,we replace both the box head and re-ID head with Transformer,while the detection accuracy is slightly reduced,the re-ID accuracy is significantly improved with the support of the DTH.As can be seen,although the Transformer box head reduces the detection accuracy,it efficiently extracts the valid information and improves the overall re-ID performance with the Transformer re-ID head.The Transformer re-ID head undoubtedly enhances the re-ID performance in various occlusion scenarios,and significantly increases the overall re-ID performance.

    Therefore,we believe that our design of the DTHN can fully extract both the box features and the unique features of the person for efficient re-ID.

    4.4 Comparison with State-of-the-Art Models

    We compare our DTHN with state-of-the-art methods on CUHK-SYSU and PRW,including two-step and end-to-end methods.The results are shown in Table 3.

    Table 3:Comparison with SOTA models

    Context Bipartite Graph Matching(CBGM)is a algorithm used in test phase to integrate context information into the matching process.It compares the two most similar targets and use K-M algorithm to the optimal matching with largest weight.

    The results of using CBGM are shown in Table 4.

    Table 4:Comparison with SOTA models using CBGM

    The graphical representations of each dataset’s results are shown in Figs.6 and 7.The horizontal axis is mAP and the vertical axis is top-1.

    Figure 6:Comparison with SOTA end-to-end models in CUHK-SYSU

    4.4.1 Result on CUHK-SYSU

    As shown in the table,we achieved the same 93.9 mAP and a comparable 94.3 top-1 scores compared to the state-of-the-art two-step method TCTS.Compared with the recent end-to-end works,our mAP outperforms the AlignPS,SeqNet,and AGWF,and our top-1 score outperforms the AlignPS and AGWF.Additionally,by using the post-processing operation CBGM,both mAP and top-1 scores of our method improved to 94.9 and 95.3,achieving the best mAP in all methods with a highly competitive top-1 scores.

    4.4.2 Result on PRW

    PRW dataset is well known as more challenging.We achieved 50.7 mAP and 85.1 top-1 scores.Our mAP outperforms all the two-step methods.Among the end-to-end methods,our mAP and top-1 score outperform AlignPS and SeqNet,while remaining a 2.5 gap with AGWT and COAT.Due to the structural advantage of COAT,it remains state-of-the-art status on the PRW dataset,but the DTHN proposed in this paper still achieves respectable results with a smaller number of parameters and computational effort.However,by applying CBGM as a post-processing operation,we obtain a slight gain of 0.9 mAP and a significant gain of 2.5 for the top-1 score,further improving the performance of our method and reducing the gap with COAT.This means that our proposed DTHN is effective in handling the challenging PRW dataset.

    Figure 7:Comparison with SOTA end-to-end models in PRW

    4.4.3 Efficiency Comparison

    We compare our efficiency with two end-to-end networks SeqNet and COAT.All experiments are conducted on the RTX 3070Ti GPU on the PRW dataset.As shown in Table 5,we include the number of parameters,the multiply-accumulate operations(MACs),and the running speed in frames per second(FPS)in the comparison.

    Table 5:Efficiency comparison

    Compared with SeqNet and COAT,we significantly reduce the number of parameters and remain the equivalent MACs,achieving a comparable accuracy.In terms of FPS,SeqNet has the highest 9.43 because it does not need to compute attention,and we have a slight advantage in running speed compared to COAT with also computes attention.In summary,our model can run efficiently while having a good performance.

    4.5 Visualization Analysis

    To show the recognition accuracy of DTHN in different scenes,several scenes are selected as demonstrations as shown in Fig.8.The green bounding box indicates the detection results that are higher than 0.5 similarity.

    Person search is difficult for several reasons,such as camera distance,occlusion,resolution,complex background,and lighting environments.DTHN can extract the features of the target well,thanks to the inclusion of DTH structure.The visualization demonstrates the model’s ability to make sound judgments despite a variety of difficult situations,proving the model’s effectiveness.

    The network takes the query picture as the target and search the person in the gallery.In case(1),the target is a dancing girl on the dance floor.Despite the dim lighting and the fact that dance movements may make the target difficult to recognize,the model is still able to find the target among the many dancers in the scene.In case (2),the target is a young man with a suitcase which covered his lower half body.Despite the lack of information about the lower half,the model can still target in multi-crowd scenarios based on existing information,even with the target’s back toward the camera.In case(3),the target is a male with his back to the camera.In the absence of front side information,the model does a good job of identifying the target based on other information such as clothing.In the same back scene with target undressing,the model is still able to correctly recognize the target.

    5 Conclusion and Outlook

    After noticing the challenges of occlusion and efficiency in end-to-end person search,we propose a DTHN to address the problems.We use two Transformer heads to deal with box detection and re-ID tasks separately,handling high-quality bounding box feature extraction and high-quality re-ID feature extraction.DTHN outperforms existing methods in the CUHK-SYSU dataset and achieves competitive results in the PRW dataset,which demonstrates the method’s superior structural design and effectiveness.

    Although our method is slightly slower than traditional CNN methods due to the scale dot production used by the attention mechanism in the Transformer,which consumes more computational resources.However,thanks to the small size of the Transformer,we have cut down the number of parameters compared to traditional CNNs,which gives us hope for deployment on terminal devices.Despite the good results,we believe that there is still room for improvement in our approach,either in terms of better and more convenient attention computation methods or in terms of adaptive attention mechanisms.Eventually,we may be able to create a pure Transformer model,using different attention heads on a single Transformer to accomplish different tasks.This is the main focus of our team afterward.We believe that the deployment of person search on terminal devices is just around the corner.

    Acknowledgement:Thank you to laboratory colleagues for their support of this paper.

    Funding Statement:This research is supported by the Natural Science Foundation of Shanghai under Grant 21ZR1426500,and the National Natural Science Foundation of China under Grant 61873160.

    Author Contributions:The authors confirm their contribution to the paper as follows:study conception and design: Cheng Feng;data collection: Cheng Feng;analysis and interpretation of results: Cheng Feng;draft manuscript preparation:Cheng Feng,Dezhi Han,Chongqing Chen.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data that support the findings of this study are available upon request from the corresponding author,Cheng Feng,upon reasonable request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    欧美高清成人免费视频www| 1000部很黄的大片| 亚洲怡红院男人天堂| 亚洲精品国产成人久久av| 亚洲婷婷狠狠爱综合网| 久久久久久久大尺度免费视频| 国产在线男女| 麻豆久久精品国产亚洲av| 九九久久精品国产亚洲av麻豆| 午夜亚洲福利在线播放| 免费黄色在线免费观看| 国产伦精品一区二区三区四那| 欧美日韩综合久久久久久| 亚洲国产精品sss在线观看| 日韩欧美三级三区| 一区二区三区免费毛片| 极品少妇高潮喷水抽搐| 国产精品1区2区在线观看.| 91aial.com中文字幕在线观看| 久久久久久久午夜电影| 久99久视频精品免费| 中文在线观看免费www的网站| 免费大片黄手机在线观看| 婷婷色麻豆天堂久久| 国产成人精品婷婷| 亚洲国产精品成人久久小说| 国产淫语在线视频| av网站免费在线观看视频 | 色视频www国产| 欧美97在线视频| 一个人看的www免费观看视频| 亚洲av国产av综合av卡| 国产黄a三级三级三级人| 国语对白做爰xxxⅹ性视频网站| 欧美极品一区二区三区四区| 亚洲欧美日韩东京热| 午夜福利在线在线| 91久久精品国产一区二区三区| 亚洲av免费高清在线观看| 日本三级黄在线观看| 99re6热这里在线精品视频| 中文字幕av成人在线电影| 亚洲av免费在线观看| 久久6这里有精品| 欧美3d第一页| 国产成人freesex在线| 91久久精品电影网| 精品一区在线观看国产| 人妻少妇偷人精品九色| 亚洲在线观看片| 亚洲在久久综合| 日本-黄色视频高清免费观看| 国产一级毛片七仙女欲春2| 永久免费av网站大全| 国产精品一区二区性色av| 国产精品一区二区性色av| 国产成人精品久久久久久| 精品一区二区三区人妻视频| 黄色日韩在线| 波野结衣二区三区在线| 国产精品一区二区在线观看99 | 国产一区二区在线观看日韩| 精品人妻视频免费看| 日日撸夜夜添| 亚洲精品日本国产第一区| 国产大屁股一区二区在线视频| 亚洲成色77777| eeuss影院久久| 啦啦啦中文免费视频观看日本| 韩国高清视频一区二区三区| www.av在线官网国产| 97人妻精品一区二区三区麻豆| 99九九线精品视频在线观看视频| 女的被弄到高潮叫床怎么办| 国产黄片美女视频| 丰满人妻一区二区三区视频av| 天天躁日日操中文字幕| 亚洲美女搞黄在线观看| 夜夜看夜夜爽夜夜摸| 日韩大片免费观看网站| 成人性生交大片免费视频hd| 18+在线观看网站| 国产精品综合久久久久久久免费| 床上黄色一级片| 国产老妇伦熟女老妇高清| 99热这里只有是精品50| 寂寞人妻少妇视频99o| 在线观看人妻少妇| 我的老师免费观看完整版| 亚州av有码| 免费观看的影片在线观看| 丝袜喷水一区| 午夜福利网站1000一区二区三区| 一级片'在线观看视频| 2021少妇久久久久久久久久久| 国产综合懂色| h日本视频在线播放| 成人国产麻豆网| 少妇猛男粗大的猛烈进出视频 | 亚洲在久久综合| 国产三级在线视频| 91精品伊人久久大香线蕉| 18禁动态无遮挡网站| 日韩,欧美,国产一区二区三区| 国产真实伦视频高清在线观看| 偷拍熟女少妇极品色| 99热6这里只有精品| 日日摸夜夜添夜夜添av毛片| 男人舔女人下体高潮全视频| 久久国内精品自在自线图片| 少妇的逼水好多| 亚洲精品视频女| 日韩强制内射视频| 亚洲美女搞黄在线观看| 精品一区二区三区人妻视频| 亚洲最大成人av| 建设人人有责人人尽责人人享有的 | 久久99蜜桃精品久久| 欧美日韩视频高清一区二区三区二| 亚洲一区高清亚洲精品| 熟女电影av网| 色综合站精品国产| 久久久久久久国产电影| 国内精品美女久久久久久| 欧美精品国产亚洲| 三级毛片av免费| 欧美日本视频| 日韩精品有码人妻一区| 久久久色成人| 色5月婷婷丁香| 精品人妻视频免费看| 69av精品久久久久久| 直男gayav资源| 在线天堂最新版资源| 国产一区亚洲一区在线观看| 非洲黑人性xxxx精品又粗又长| 免费av观看视频| 国产精品无大码| 欧美zozozo另类| 中文字幕免费在线视频6| 亚洲av二区三区四区| 久久久久久久久久黄片| 国产精品熟女久久久久浪| 可以在线观看毛片的网站| 日韩精品青青久久久久久| 欧美性感艳星| 久久久久久久国产电影| 永久免费av网站大全| 搡老乐熟女国产| 极品教师在线视频| 麻豆久久精品国产亚洲av| 听说在线观看完整版免费高清| 免费看av在线观看网站| 九九在线视频观看精品| 国产亚洲最大av| 又爽又黄无遮挡网站| 国产精品一区二区三区四区免费观看| 99久久精品热视频| 99热全是精品| 人人妻人人澡人人爽人人夜夜 | 欧美 日韩 精品 国产| 欧美最新免费一区二区三区| 久久久久久国产a免费观看| 成人鲁丝片一二三区免费| 色5月婷婷丁香| 日韩精品有码人妻一区| 欧美xxxx性猛交bbbb| 伦精品一区二区三区| 亚洲欧洲国产日韩| 大香蕉97超碰在线| 麻豆成人av视频| 日韩精品有码人妻一区| 国精品久久久久久国模美| 日本黄色片子视频| 国产探花极品一区二区| 亚洲国产欧美在线一区| 女人被狂操c到高潮| 亚洲在线自拍视频| 国产爱豆传媒在线观看| 我要看日韩黄色一级片| 久久午夜福利片| 国产69精品久久久久777片| 亚洲精品成人av观看孕妇| 乱码一卡2卡4卡精品| 国产成人午夜福利电影在线观看| 国产精品一区二区三区四区免费观看| 男的添女的下面高潮视频| 亚洲综合色惰| 国产黄色小视频在线观看| 2021少妇久久久久久久久久久| 免费看av在线观看网站| 联通29元200g的流量卡| 精品国产一区二区三区久久久樱花 | 国产成人91sexporn| 丝袜美腿在线中文| 99久久精品一区二区三区| 亚洲欧美精品专区久久| 男插女下体视频免费在线播放| 亚洲av中文字字幕乱码综合| 久久精品国产自在天天线| 久久久久九九精品影院| 国产伦理片在线播放av一区| 亚洲怡红院男人天堂| 精品人妻熟女av久视频| 国产成人精品福利久久| 久久精品综合一区二区三区| 国产91av在线免费观看| 成人毛片60女人毛片免费| av在线天堂中文字幕| 别揉我奶头 嗯啊视频| 永久免费av网站大全| 深爱激情五月婷婷| 久久久久久久亚洲中文字幕| 亚洲精品,欧美精品| 亚洲婷婷狠狠爱综合网| 亚洲性久久影院| 菩萨蛮人人尽说江南好唐韦庄| 国产精品一区www在线观看| 国产探花极品一区二区| 日韩三级伦理在线观看| 狠狠精品人妻久久久久久综合| 欧美xxⅹ黑人| 午夜福利视频精品| 亚洲国产欧美人成| 免费看av在线观看网站| av天堂中文字幕网| 亚洲国产精品专区欧美| 欧美性感艳星| 久久精品国产亚洲网站| 免费少妇av软件| 三级男女做爰猛烈吃奶摸视频| 久久99热这里只有精品18| 亚洲精品影视一区二区三区av| 久久草成人影院| 看黄色毛片网站| 欧美xxⅹ黑人| 日韩中字成人| 亚洲av一区综合| 极品教师在线视频| 亚洲欧美成人精品一区二区| 成人亚洲精品一区在线观看 | 久久国内精品自在自线图片| 精品久久久久久久久久久久久| 少妇高潮的动态图| 亚洲欧美精品专区久久| 日本免费a在线| 99热这里只有是精品在线观看| 日韩一区二区视频免费看| 精品国产露脸久久av麻豆 | 色播亚洲综合网| 久久精品久久久久久久性| 欧美日韩视频高清一区二区三区二| 欧美3d第一页| 极品少妇高潮喷水抽搐| 熟女电影av网| 亚洲av成人av| 亚洲人与动物交配视频| 精品国内亚洲2022精品成人| 欧美日韩综合久久久久久| 国产成人91sexporn| 国产片特级美女逼逼视频| 又黄又爽又刺激的免费视频.| 精品国产露脸久久av麻豆 | www.色视频.com| 日日干狠狠操夜夜爽| 欧美 日韩 精品 国产| 黄色配什么色好看| 国产不卡一卡二| 日韩av在线大香蕉| 18禁动态无遮挡网站| 岛国毛片在线播放| 亚洲欧美日韩无卡精品| 亚洲av电影在线观看一区二区三区 | 久久久国产一区二区| 国产精品嫩草影院av在线观看| 免费少妇av软件| 男人舔奶头视频| 久久久久久国产a免费观看| 狂野欧美白嫩少妇大欣赏| 免费看不卡的av| 精品人妻偷拍中文字幕| 丰满乱子伦码专区| 亚洲精品成人av观看孕妇| av在线老鸭窝| 丝瓜视频免费看黄片| 久久久久免费精品人妻一区二区| www.av在线官网国产| 免费电影在线观看免费观看| 一级毛片电影观看| 免费不卡的大黄色大毛片视频在线观看 | 久久99热这里只有精品18| 欧美性感艳星| or卡值多少钱| 国精品久久久久久国模美| 美女内射精品一级片tv| 亚洲aⅴ乱码一区二区在线播放| 国产黄片视频在线免费观看| 一级毛片我不卡| 免费看日本二区| 久久久久性生活片| 亚洲一区高清亚洲精品| 成人性生交大片免费视频hd| 亚洲四区av| 美女内射精品一级片tv| 91久久精品国产一区二区成人| 一区二区三区免费毛片| 六月丁香七月| 国产亚洲精品久久久com| 男人爽女人下面视频在线观看| 久久久色成人| 久久人人爽人人片av| 欧美丝袜亚洲另类| 亚洲av不卡在线观看| 成人欧美大片| 成人亚洲欧美一区二区av| 99re6热这里在线精品视频| 三级男女做爰猛烈吃奶摸视频| 免费高清在线观看视频在线观看| av国产免费在线观看| 日韩大片免费观看网站| 国产精品嫩草影院av在线观看| 看免费成人av毛片| 搡老乐熟女国产| 国产色婷婷99| 国产精品久久久久久精品电影| 天堂俺去俺来也www色官网 | 久久99热这里只频精品6学生| 午夜精品一区二区三区免费看| 国产一区二区在线观看日韩| 韩国av在线不卡| 天堂中文最新版在线下载 | 午夜福利视频1000在线观看| 夫妻性生交免费视频一级片| 一二三四中文在线观看免费高清| 日本熟妇午夜| 国产综合精华液| 日韩视频在线欧美| av又黄又爽大尺度在线免费看| 亚洲人成网站在线播| 伦理电影大哥的女人| 国产精品久久久久久久电影| 成人鲁丝片一二三区免费| 又黄又爽又刺激的免费视频.| 黄色一级大片看看| 亚洲精品日韩av片在线观看| 国产黄a三级三级三级人| 91久久精品国产一区二区三区| 国产精品女同一区二区软件| 丰满人妻一区二区三区视频av| 激情五月婷婷亚洲| 少妇裸体淫交视频免费看高清| 日本一二三区视频观看| 婷婷色av中文字幕| 久久亚洲国产成人精品v| 精品一区二区三区人妻视频| 欧美日韩精品成人综合77777| 亚洲欧美日韩无卡精品| 欧美高清成人免费视频www| 久久久久久久久久成人| 欧美不卡视频在线免费观看| 成人毛片a级毛片在线播放| 国产精品人妻久久久影院| 精品一区二区三区视频在线| 免费黄频网站在线观看国产| 插逼视频在线观看| 一个人免费在线观看电影| 成人欧美大片| 成人特级av手机在线观看| 国产免费又黄又爽又色| 免费电影在线观看免费观看| 真实男女啪啪啪动态图| 亚洲精品色激情综合| 伊人久久国产一区二区| 精品久久久久久电影网| 久久精品国产鲁丝片午夜精品| 久久精品夜夜夜夜夜久久蜜豆| 久久草成人影院| 汤姆久久久久久久影院中文字幕 | 日韩欧美一区视频在线观看 | 日韩欧美 国产精品| 精品午夜福利在线看| 午夜精品在线福利| 观看免费一级毛片| 国产精品久久久久久久电影| 久久久久久久久久黄片| 男人爽女人下面视频在线观看| 久久久精品免费免费高清| 日日啪夜夜撸| 大片免费播放器 马上看| 免费黄频网站在线观看国产| 国产爱豆传媒在线观看| 一级毛片电影观看| 日本黄色片子视频| 色5月婷婷丁香| 精品国产三级普通话版| 18禁在线无遮挡免费观看视频| 欧美高清性xxxxhd video| 国产男女超爽视频在线观看| av又黄又爽大尺度在线免费看| 国模一区二区三区四区视频| 精品人妻一区二区三区麻豆| 三级国产精品欧美在线观看| 久久久精品欧美日韩精品| 亚洲久久久久久中文字幕| 全区人妻精品视频| 女人十人毛片免费观看3o分钟| 日韩,欧美,国产一区二区三区| 在线a可以看的网站| 嫩草影院入口| 夫妻性生交免费视频一级片| 亚洲综合精品二区| 午夜福利在线观看免费完整高清在| 亚洲精品国产av成人精品| 国产精品久久久久久久电影| 国产片特级美女逼逼视频| 午夜福利高清视频| 亚洲精品第二区| 亚洲精品日韩在线中文字幕| 欧美丝袜亚洲另类| 国产黄片美女视频| 国产精品av视频在线免费观看| 国产有黄有色有爽视频| 免费av毛片视频| 欧美成人a在线观看| 大又大粗又爽又黄少妇毛片口| 最新中文字幕久久久久| 熟女人妻精品中文字幕| 日日摸夜夜添夜夜添av毛片| 国内揄拍国产精品人妻在线| 一级毛片我不卡| 青春草视频在线免费观看| 欧美最新免费一区二区三区| 国产欧美另类精品又又久久亚洲欧美| 简卡轻食公司| 简卡轻食公司| 欧美日韩视频高清一区二区三区二| 夜夜爽夜夜爽视频| 精华霜和精华液先用哪个| 男人爽女人下面视频在线观看| 欧美日韩综合久久久久久| 最近视频中文字幕2019在线8| 99久久人妻综合| 亚洲熟女精品中文字幕| 亚洲精品第二区| 国产精品综合久久久久久久免费| 日本一二三区视频观看| av福利片在线观看| 日日干狠狠操夜夜爽| 亚洲精华国产精华液的使用体验| 日日撸夜夜添| 内地一区二区视频在线| av免费观看日本| 少妇人妻精品综合一区二区| 欧美区成人在线视频| 日本wwww免费看| 在线免费观看的www视频| 久久久久网色| 欧美+日韩+精品| 国内揄拍国产精品人妻在线| 国产麻豆成人av免费视频| 国产成人一区二区在线| 日本wwww免费看| 国产老妇伦熟女老妇高清| 日日摸夜夜添夜夜添av毛片| 国产不卡一卡二| 国产亚洲av嫩草精品影院| 中文字幕久久专区| 欧美日韩精品成人综合77777| 欧美xxxx黑人xx丫x性爽| 啦啦啦中文免费视频观看日本| 久久久久久久亚洲中文字幕| 1000部很黄的大片| 午夜福利网站1000一区二区三区| 国产精品无大码| 国产片特级美女逼逼视频| 色5月婷婷丁香| 亚洲内射少妇av| 丝瓜视频免费看黄片| 国产精品福利在线免费观看| 亚洲婷婷狠狠爱综合网| 中文资源天堂在线| 我的女老师完整版在线观看| 嘟嘟电影网在线观看| 亚洲欧美精品自产自拍| 天天躁日日操中文字幕| 精品一区二区三卡| 国产一区亚洲一区在线观看| 久久久久久久大尺度免费视频| 久久久亚洲精品成人影院| videossex国产| av天堂中文字幕网| 最近中文字幕高清免费大全6| 尾随美女入室| 欧美性感艳星| 欧美区成人在线视频| 国产成人aa在线观看| 免费观看在线日韩| 韩国高清视频一区二区三区| 2021天堂中文幕一二区在线观| 丝瓜视频免费看黄片| 少妇的逼好多水| 中文字幕久久专区| 久久99热这里只频精品6学生| .国产精品久久| 一区二区三区乱码不卡18| 久久久亚洲精品成人影院| 日韩 亚洲 欧美在线| 国产亚洲5aaaaa淫片| 免费黄网站久久成人精品| 国内揄拍国产精品人妻在线| 一级毛片aaaaaa免费看小| 一级毛片久久久久久久久女| 成人高潮视频无遮挡免费网站| 性色avwww在线观看| 99视频精品全部免费 在线| 成人av在线播放网站| 黄片无遮挡物在线观看| 2021少妇久久久久久久久久久| 国产一级毛片七仙女欲春2| 亚洲欧美中文字幕日韩二区| 看十八女毛片水多多多| 免费看美女性在线毛片视频| 卡戴珊不雅视频在线播放| 国产色爽女视频免费观看| 国产成人免费观看mmmm| 极品教师在线视频| 亚洲乱码一区二区免费版| 少妇熟女欧美另类| 人人妻人人澡欧美一区二区| 亚洲欧美中文字幕日韩二区| 两个人的视频大全免费| 精品欧美国产一区二区三| 欧美激情久久久久久爽电影| 成人午夜高清在线视频| 亚洲国产日韩欧美精品在线观看| 欧美一区二区亚洲| 欧美一级a爱片免费观看看| 国产精品av视频在线免费观看| 高清毛片免费看| 亚洲熟女精品中文字幕| 亚洲久久久久久中文字幕| www.色视频.com| 午夜激情福利司机影院| 日本与韩国留学比较| 97精品久久久久久久久久精品| 国产精品国产三级专区第一集| 非洲黑人性xxxx精品又粗又长| 午夜福利网站1000一区二区三区| 水蜜桃什么品种好| 只有这里有精品99| 午夜激情福利司机影院| 美女大奶头视频| 国产单亲对白刺激| 国产黄频视频在线观看| 你懂的网址亚洲精品在线观看| 高清av免费在线| 午夜精品一区二区三区免费看| 中文精品一卡2卡3卡4更新| 成人美女网站在线观看视频| 久久久精品94久久精品| 特级一级黄色大片| 国产黄片视频在线免费观看| 一个人看的www免费观看视频| 18禁裸乳无遮挡免费网站照片| 国产成人精品久久久久久| 国产精品熟女久久久久浪| 久久久亚洲精品成人影院| 亚洲精品456在线播放app| 久久精品国产自在天天线| 亚洲国产精品国产精品| 七月丁香在线播放| 日韩欧美一区视频在线观看 | 久久精品人妻少妇| 人妻系列 视频| 亚洲欧美精品专区久久| 国产精品国产三级国产专区5o| 精品久久久久久久久亚洲| 女人被狂操c到高潮| 日日撸夜夜添| 啦啦啦韩国在线观看视频| 黑人高潮一二区| 久久久欧美国产精品| 久久久久久久久大av| 精品久久久噜噜| 高清av免费在线| 精品国内亚洲2022精品成人| 深爱激情五月婷婷| 国产精品一区二区三区四区免费观看| 久久精品久久精品一区二区三区| 亚洲av二区三区四区| 国产高清有码在线观看视频| 国产白丝娇喘喷水9色精品| 久久久久九九精品影院| h日本视频在线播放| 国产高清国产精品国产三级 | 激情 狠狠 欧美| 国产成人一区二区在线| av免费观看日本| 久热久热在线精品观看| 最新中文字幕久久久久| 99热这里只有是精品50| 久久精品夜夜夜夜夜久久蜜豆| 蜜桃久久精品国产亚洲av| 亚洲精品色激情综合| 成人午夜精彩视频在线观看| 2018国产大陆天天弄谢| 国产伦一二天堂av在线观看| 18禁在线无遮挡免费观看视频| 欧美不卡视频在线免费观看| 一区二区三区乱码不卡18| 精品亚洲乱码少妇综合久久| 少妇丰满av| 中文字幕久久专区| 国产 亚洲一区二区三区 | 狠狠精品人妻久久久久久综合|