• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Self-supervised recalibration network for person re-identification

    2024-02-29 08:22:16ShaoqiHouZhimingWangZhihuaDongYeLiZhiguoWangGuangqiangYinXinzhongWang
    Defence Technology 2024年1期

    Shaoqi Hou , Zhiming Wang , Zhihua Dong , Ye Li , Zhiguo Wang ,*,Guangqiang Yin ,**, Xinzhong Wang

    a Shenzhen Institute of Information Technology, Shenzhen, 518172, China

    b University of Electronic Science and Technology of China, Chengdu, 611731, China

    c Kash Institute of Electronics and Information Industry, Kashi, 844099, China

    Keywords: Person re-identification Attention mechanism Global information Local information Adaptive weighted fusion

    ABSTRACT The attention mechanism can extract salient features in images,which has been proved to be effective in improving the performance of person re-identification (Re-ID).However, most of the existing attention modules have the following two shortcomings:On the one hand,they mostly use global average pooling to generate context descriptors, without highlighting the guiding role of salient information on descriptor generation, resulting in insufficient ability of the final generated attention mask representation; On the other hand, the design of most attention modules is complicated, which greatly increases the computational cost of the model.To solve these problems, this paper proposes an attention module called self-supervised recalibration (SR) block, which introduces both global and local information through adaptive weighted fusion to generate a more refined attention mask.In particular, a special"Squeeze-Excitation" (SE) unit is designed in the SR block to further process the generated intermediate masks, both for nonlinearizations of the features and for constraint of the resulting computation by controlling the number of channels.Furthermore, we combine the most commonly used ResNet-50 to construct the instantiation model of the SR block,and verify its effectiveness on multiple Re-ID datasets,especially the mean Average Precision(mAP)on the Occluded-Duke dataset exceeds the state-of-the-art(SOTA) algorithm by 4.49%.

    1.Introduction

    Target recognition is an indispensable key technology for the construction of smart city and national defense.As we all know,pedestrians are the main bodies of various activities,so the accurate identification of pedestrians has always been a hot direction in the field of target recognition.The term concept of "Person Reidentification (Re-ID)" was coined in 2005 to address the problem of matching pedestrians across different scenes.In simple terms,it is to match pedestrians appearing in a certain camera with the same pedestrians appearing in other cameras.The emergence of Re-ID technology makes up for the defect that pedestrian identity is difficult to determine due to the lack of face information in surveillance scenarios, which has important application value in the fields of target precision attack,evidence collection of the forensic department, the operation of unmanned supermarket and intelligent security.However,due to the different parameters of different cameras and the variable monitoring environment, pedestrian images are vulnerable to factors such as illumination, viewing angle,blur,and occlusion,which makes the customization of Re-ID technology face great challenges at the beginning.

    With the rapid development of deep learning, since 2016,scholars have begun to apply this technology to the large-scale dataset of Re-ID and have made great progress [1].It is well known that network engineering has been one of the most important deep learning techniques, as well-designed networks ensure significant performance gains in various applications including Re-ID.From AlexNet [2] to residual-type architecture[3,4],neural networks have become deeper and deeper,which can provide rich representation.Following a similar line of thought,GoogLeNet [5] showed that width is another important factor in improving model performance.Zagoruyko et al.[6] proposed a wide residual network based on the ResNet architecture and proved in the CIFAR benchmark that a 28-layer ResNet with increased width can outperform an extremely deep ResNet with 1001 layers.

    In addition to the above means, some studies have shown that by integrating the learning mechanism into the network, it is helpful to capture the correlation between features, and the representation generated by convolutional neural networks can be strengthened.Based on this, many scholars propose that network design can learn from human visual function, that is, learning to choose "key things" for attention, and selectively ignore those things that are not interested in Ref.[7].Inspired by this,scholars at home and abroad began to introduce "attention mechanism" into network training, and it has highlighted its effectiveness in many tasks, including sequence learning [8,9], image localization and understanding [10,11], image captioning [12,13] and lip reading[14].A central theme of deep model research is to find more powerful feature representation,and the attention mechanism can capture the attributes and features that are more discriminative for a given task in the image through weight allocation, and achieve the purpose of suppressing the interference information.For example, if the gender of pedestrians is classified, the focus of model learning should be the target "whether there is long hair","whether there is a skirt" and so on.

    The commonly used attention mechanisms can be divided into two categories: spatial attention and channel attention.Spatial attention aims to focus on two-dimensional space and extract salient features of pedestrians and other targets by finding the relevance of image context.Channel is dedicated to building the dependence relationship between channel features and improving the information weight of key channel features.The originator of channel attention is SENet[15]proposed by Momenta company in 2017, which learns to use global information to selectively emphasize the importance of channel information and suppress less useful features.Based on SENet, Woo et al.[16] proposed an attention module CBAM based on convolution block, which not only adds spatial attention mechanism, but also introduces max pooling as a supplement to SENet’s use of average pooling to generate attention masks.Since then, on the basis of SENet and CBAM,a large number of classical researches on channel attention and spatial attention have appeared in the field of computer vision[17-19], and led to the application of attention mechanism in the field of Re-ID[20-22].However,the existing channel attention and spatial attention modules basically have the following shortcomings:

    (1) Most attention modules choose global average pooling to generate weight descriptors for each feature map’s channel or spatial location.Although the integrity of global information is preserved,it does not highlight the guiding role of saliency information on descriptors.

    (2) Even though the fusion of max pooling and average pooling is adopted, the fusion method is very violent, either direct Add or Cat, which does not give full play to the supervision role of network learning on the generation of attention masks,ignoring the difference in the importance of different descriptor information to different tasks.

    (3) The design of most attention modules is too complex,which greatly increases the computational cost of the model.

    In view of the above problems, our experiments find that explicitly modeling the interdependence between channel features of feature map can significantly improve the expression ability of network,and this effect is more obvious than that of spatial feature.On this basis,we start from the channel relationship and adopts the method of "local and global information fusion" to design a selfsupervised recalibration (SR) block (see Fig.1), which can supervise the generation of the optimal attention mask through its own training process and thus self-adaptively recalibrate the feature response of the channel direction of the feature map through the generated mask.In the recalibrated feature information, the proportion of the invariant feature under the pedestrian cross-scene in the whole recognition information is enhanced,thus improving the representation quality of the network generation.

    On the one hand, the attention mask of the SR block also incorporates local information descriptors in addition to the global information descriptors.In particular, different from the conventional approach of direct fusion, SR block allows the network to adaptively weight two different descriptors through training, so that the model can be task-oriented to achieve the purpose of enhanced learning.On the other hand,a small"Squeeze-Excitation"(SE) unit is designed in the SR block, which can transfer features through the "Squeeze-Excitation" operation.It can not only perform nonlinear processing on features, but also control the number of channels in the feature map, thus significantly constraining the computational cost of the model.

    Fig.1.Overall structure of SR block.

    In summary,the main contributions of our paper are as follows:

    (1) Propose a simple and effective channel attention module named SR block, which can effectively improve the expressive ability of CNN;

    (2) The effectiveness of the SR block is verified through a large number of ablation experiments;

    (3) By inserting the SR block of this paper into ResNet-50,we get the instantiated model SR-ResNet-50, and verify that it has achieved significant performance improvements on multiple Re-ID datasets.

    The rest of the paper is organized as follows: Section 2 introduces some related work;Section 3 describes the proposed selfsupervised recalibration network (SRNet), which involves the design details of the SR block;Section 4 illustrates the datasets and evaluation indicators adopted in this paper, and presents various experimental designs and analysis of results;Finally,Section 5 gives the conclusion of this paper.

    2.Related work

    In recent years, many new methods combining depth framework and attention mechanism have emerged in Re-ID task, and they have achieved excellent performance on the Re-ID dataset.For example,Jiang et al.[23]proposed a cross-modal multi-granularity attention network (CMGN) in 2020, which can learn the common features of different modes and map them to the same feature space, thus solving the problem of cross-modal pedestrian matching under infrared and visible light.In the same year, Zhu et al.[24] proposed a curriculum enhanced supervised attention network (CE-SAN) to solve the problem of pedestrian misalignment.The CE-SAN aims to train the attention module under supervision, it links local and global branches together to better emphasize key information and exploit discriminative features.In 2021, Ning et al.[25] proposed a joint weak saliency mechanism and attention-aware model (JWSAA) for the background bias caused by significance detection bias, which makes the model gradually focus on all valuable regions of pedestrian features by weakening the high-response features.Then, Zhang et al.[26]proposed a hybrid-attention guided network with multiple resolution features to ensure the spatial alignment of feature maps at different stages during feature fusion.In addition,Sun et al.(2021)[27]cascaded spatial-level attention block(SAB)and channel-level attention block(CAB)to form a multi-level-attention block(MAB),which is used to extract the robust features of pedestrians across scenes.In the latest work in 2022, Tan et al.[28] proposed a multihead self-attention network (MHSA-Net).MHSA-Net can adaptively obtain key local information from pedestrian images and eliminate unimportant noise information, significantly improving the performance of Re-ID under occlusion conditions.In addition,in order to take into account model performance and computational complexity, Chen et al.[29] proposed an attention aware feature learning framework, which can carry out supervised training without changing the structure of the Re-ID model.

    3.Self-supervised recalibration network

    This section contains four parts:First,subsection 3.1 introduces the overall architecture and design ideas of SR block; Then, in subsections 3.2 and 3.3, two key design schemes in SR block are elaborated respectively; Finally, subsection 3.4 details the instantiation model for SR block.

    3.1.Self-supervised recalibration block

    As shown in Fig.1,the SR block is a computing unit.The specific process of feature map recalibration is as follows:

    Step 1.Given an intermediate feature mapF∈RC×H×Was input.In order to take into account the integrity and specificity of the feature map information of each channel,the SR block uses average pooling and max pooling,respectively,to aggregate the spatial information of the input feature map,and generate two different spatial context descriptorsFavgtheFmax.

    whereu(i,j) represents the pixel value at the feature map coordinate(i,j).

    Step 2."Squeeze-Excitation" operation is carried out for the two feature maps respectively (see Section 3.2 for details) to further process the features and improve the nonlinearities of the model while controlling the computational load.

    In the equation,twoFsq-ex(·)have the same operation process but do not share parameters.

    Step 3.The two intermediate masksandare multiplied by weight factorswC×1and (1C×1-wC×1) for adaptive weighted fusion (see subsection 3.3 for details), and the fused mask passes through a composite convolutional layer (see Fig.2) further processes the feature details to obtain the hybrid maskM.

    In the equation,·is the multiplication by channel;⊕represents the channel-concat operation;-represents the subtraction by channel.In the actual training process, the weight factors are broadcast(replicated) in the spatial dimension, so that the program can perform element-by-element operations.

    Step 4.The hybrid maskMuses the same form of "Squeeze-Excitation"operation as the second step to perform fine processing of features to obtain the final maskM'.

    Fig.2.Structure of composite convolutional layer in SR block.

    Step 5.The channel attention maskM' is multiplied channel-bychannel with the input feature map to generate the recalibrated feature map.

    where ?represents element-wise multiplication.Similarly, in the actual training process, the attention maskM' will first be broadcast (replicated) correspondingly along the space dimension(H×W) of feature map, changing from dimensionRC×1×1toRC×H×W.

    The above is the whole process of the input feature map by the SR block.It can be seen that the SR block processes the features in a more detailed manner.In particular, the generation of the weighting factor and the final mask of the SR block completely depends on the adaptive learning of the network.During the training process,the model supervises the selection of parameters,so that the model can independently decide the weight allocation of the average pooling mask and the max pooling mask.To be clear,the design details in each processing step of the SR block in this paper are as follows.

    (1) In Step 1, based on the generality of average pooling, this paper introduces max pooling as a supplement to global spatial information, which aims to use the important clues about unique object features collected by max pooling to infer more refined channel attention;

    (2) The use of weight factorswC×1and (1C×1-wC×1) in Step 2 allows the model to have more autonomy in the selection of mask descriptors.On the basis of empirical evidence (see the ablation experiments in subsection 4.4), we find that letting each corresponding channel learn a mutual exclusion mechanism ofand1-(i=1,2,3,…,C)can make the generated hybrid masks more specific and representative;

    (3) TheFsq-ex(·)operation in Steps 2 and 4 will further enhance the feature richness of the intermediate mask, and its"Squeeze-Excitation" operation can effectively control the amount of computation;

    (4) The fusion method of the two weighted masks in Step 3 is channel-concat rather than element-add to avoid feature confusion caused by direct addition, and it’s also proved in this paper that channel-concat is a special form of elementadd, which can be converted to each other under certain conditions (see subsection 4.4 for the corresponding proof);

    (5) In Step 3,the operation details of the composite convolution layer are shown in Fig.2.In addition to processing features,it also has the function of channel compression, which is responsible for halving the channels of fused mask.In particular,we place the batch normalization(BN)layer after the activation function ReLU, because the non-negative response of ReLU will cause the weight layer to be updated in a suboptimal way, which is also verified in the ablation experiments.

    3.2."Squeeze-Excitation" operation

    Fig.3.Schematic diagram of Fsq-ex(·).

    "Squeeze-Excitation"operationFsq-ex(·)is a design used in this paper to trade off the accuracy and computational complexity of the model.As shown in Fig.3,it can control the channel dimension by adjusting the "scaling factor"r(rdefaults to 16, which will be discussed in the ablation experiment), thereby increasing or decreasing the capacity of the model.

    Fsq-ex(·)consists of two series operations ofFsq(·)(squeezing)andFex(·) (exciting).Fsq(·) is responsible for compressing the number of channels of the attention descriptor to 1/rof the original.AfterFex(·),the channels is restored.That is,given the descriptorsFavgandFmax, the process of calculating the intermediate masksandis as follows:

    Given a hybrid mask M,the process of calculating the final mask is

    The meanings of each symbol in the equation are the same as those defined in subsection 3.1.

    Through the "Squeeze-Excitation" operation, the channel dependencies of the attention mask are fully interacted, which also greatly reduces the computation of subsequent filters.It should be noted that although the same form of operation has been performed,the parameters of the threeFsq-ex(·)functions in Fig.1 are not shared, and the details are shown in Fig.4.

    Fig.4.Detailed structure of Fsq-ex(·): (a) (·); (b) (·); (c)(·).

    As can be seen from Fig.4 above,except for(·),the basic unit of each "Squeeze-Excitation" operation is the composite convolution layer as shown in Fig.2, and the Conv1d in each composite convolutional layer is a one-dimensional convolution of size 1 × 1 (essentially equivalent to the fully connected layer).Among them, the joint use of the filter Conv1d and the activation function ReLU enables the model to fully learn the nonlinear interaction between the channels of feature map,and the BN layer is responsible for constraining the data distribution while accelerating the convergence of the SR block.In(·), we use the sigmoid function to map the value of the attention mask to the(0,1)interval.

    In short,the detailed calculation process of the three"Squeeze-Excitation" operations in the SR block is

    whereRdenotes the activation function ReLU; σ denotes the Sigmoid function; * denotes matrix multiplication;Wsq∈RC/r×C,Wex∈RC/r×C; Other parameters are the same as those defined in Section 3.1.

    3.3.Adaptive weighted fusion

    In order to make each channel attention mask take into account both global semantic information and local semantic information,we specially adopt a weighted method to adaptively fuse two different intermediate masksand(see Fig.1).As shown in Fig.5,supposewherei= 1,2,…,C.Then the calculation equation of the fusion mask is

    In the equation,pi,qi∈(0,1) andpi+qi= 1, the mutual exclusivity of the two allows the hybrid mask to autonomously learn the importance of global information and local information according to the Re-ID task,and then determine the weight vectors P andQ.

    In the actual training process of the model, since the network can only learn real parameters, this paper cleverly introduces the softmax function to constrain the parameters learned by the network to the interval (0,1).Which is

    At this point,the network learning of the weight vectorsPandQis transformed into the learning of two real vectorsXavgandXmaxrespectively.

    3.4.Instantiated model

    The SR block can be flexibly inserted into standard backbones such as VGGNet and ResNet,or serve as an efficient replacement for other plug-and-play modules.For a fair comparison, we integrate the SR block into the most widely used ResNet-50, and conduct experiments with this as the research object.ResNet-50 is the most classic backbone that gives consideration to both "depth" and"accuracy" of the network, it’s composed of multiple residual blocks stacked according to certain rules.As the basic unit of ResNet-50,the residual block contains two parts:"residual branch"and "identity mapping branch".The authors He et al.[3] believe that if the optimization objective function approximates an identity map, it will be easier to learn to find the disturbance (i.e., the residual)to the identity map than to relearn a new mapping function.Therefore, adding "identity mapping branch" can effectively alleviate the network degradation problem caused by the increase of depth.

    As shown in Table 1, we insert the SR block after the nonlinear operation and before the summation of the "identity map branch"and call it the SR-ResNet (see Fig.6).Similarly, other variants that integrate SR blocks with InceptionNet[30],ResNeXt[4],MobileNet[31], and ShuffleNet [32] can be constructed by similar schemes above.For the concrete example of SRNet,Fig.7 shows the detailed description of ResNet-50 and SR-ResNet-50.It should be noted that the experiment in this paper will be carried out around the instantiation model SR-ResNet-50.

    4.Experimental results and analysis

    This section contains four parts: Subsection 4.1 introduces the datasets and evaluation indicators used in the experiments; Subsection 4.2 describes the loss function and some specific parameter settings;In subsection 4.3,we focus on testing the performance of the model before and after instantiation; Then, in subsection 4.4,several groups of ablation experiments were set up to verify the effectiveness of each design scheme of SR block; Subsection 4.5 shows the comparison between the proposed algorithm and the advanced algorithm, highlighting the superiority of the overall model; Finally, in subsection 4.6, some inference results are visualized to facilitate intuitive understanding of the effect of the model before and after improvement.

    Fig.5.Schematic representation of adaptive weighted fusion in SR blocks.

    Fig.6.Integration scheme of SR block and ResBlock: (a) ResBlock; (b) SR-ResBlock.

    Table 1Detailed configuration of ResNet-50 and SR-ResNet-50.

    Fig.7.Image display of pedestrians in Market1501.

    4.1.Datasets and evaluation indicators

    4.1.1.Datasets

    The datasets used in the experiments in this paper include two regular datasets Market1501[33]and DukeMTMC-reID[34],and an occlusion dataset Occluded-Duke [35].

    Market1501 is an authoritative dataset collected on the campus of Tsinghua University in summer,constructed and made public in 2015, as shown in Fig.7 for a sample of pedestrian images.Market1501 contains a total of 1501 pedestrian IDs and 32,668 detection frames captured by 6 cameras (including 5 high-definition cameras and 1 low-definition camera).Each pedestrian is captured by at least 2 cameras and may have multiple images under one camera.Among them, the training set has 751 pedestrian IDs,including 12,936 images; the query set has 750 pedestrian IDs,including 3368 images;the gallery set has 19,732 images with 750 pedestrian IDs,and the IDs of the gallery set is the same as that of the query set.

    DukeMTMC is a large-scale multi-target multi-camera pedestrian tracking dataset constructed in 2016.It provides a large HD video dataset recorded by 8 simultaneous cameras,with more than 7000 single-camera trajectories and more than 2700 individual pedestrians, as shown in Fig.8 for a sample pedestrian images.DukeMTMC-reID is a Re-ID subset of DukeMTMC and provides human-annotated bounding boxes containing 16,522 training images for 702 pedestrian IDs,2228 query images from an additional 702 pedestrian IDs, and a gallery containing 1110 pedestrian IDs and 17,661 images.

    In order to fully demonstrate the superior performance and transferability of the proposed algorithm,we introduce the difficult dataset Occluded-Duke, which is the largest dataset for occluded Re-ID so far.Occluded-Duke is constructed from DukeMTMC-reID by leaving occluded images and filtering out some overlapping images, and all its query images and 10% of gallery images are occluded pedestrian images, as shown in Fig.9 for a sample of pedestrian images.Specifically, the training set of Occluded-Duke contains 15,618 images, covering 702 pedestrian IDs; The test set contains a total of 1110 pedestrian IDs,including 2210 query images and 17,661 gallery images.In particular,the directory structure and naming of the Occluded-Duke dataset is the same as DukeMTMCreID.

    4.1.2.Evaluationindicators

    The evaluation of all Re-ID algorithms in this paper includes two general evaluation indicators:the Cumulative Match Characteristic(CMC) of Rank-1 and the mean Average Precision(mAP).

    Rank-1 represents the probability that the image ranked first in gallery search results is the target, also known as the first hit rate;Similarly, Rank-k represents the probability that the topkgallery images retrieved from each query image contain the correct target.In terms of practical tasks, the most meaningful metric is Rank-1.

    mAP is a comprehensive evaluation, taking into account both Precision and Recall.Precision refers to the proportion of true examples among all retrieved samples that are considered to be correct targets; However, Recall refers to the proportion of retrieved as true examples among samples that are actually correct targets.The two are used as vertical and horizontal coordinates respectively to form aP-R curve, and the area under the curve is called Average Precision (AP), which is the integral sum of the precision at different recall points,as shown in Eq.(11).The larger the value of AP, the higher the average accuracy of the algorithm.

    In general,the query image set contains multiple samples with different IDs to be retrieved.AP after arithmetic average based on the number of samples is called mAP, namely

    wheremrepresents the total number of samples in the query set;iis theith sample to be retrieved in the query set.

    In addition to the accuracy indicators Rank-1 and mAP,we also add the computational load indicator GFLOPs (Giga Floating-point Operations) for comprehensive evaluation of the model.

    FLOPs represents the number of floating-point operations(that is,one GFLOPs is equal to 109floating-point operations),which is a key indicator for measuring the time complexity of the model,and it comes from the paper [36] in ICLR 2017.In the convolutional neural network, the computational load of the model is mainly concentrated in the convolutional layer and the fully connected layer.Specifically, the theoretical calculation of the FLOPs for the convolutional layer and the fully connected layer is shown in Eqs.(13) and (14), respectively.

    In the equation,kwandkhrepresent the width and height of the convolution kernel respectively;kw×kh×Cinrepesents the amount of multiplication;kw×kh×Cin-1 repesents the amount of addition; +1 represents the offset;w×h×Coutrepresents the number of elements in the output feature map;wandhrepresent the width and height of the output feature map, respectively.

    Fig.8.Image display of pedestrians in DukeMTMC-reID.

    Fig.9.Image display of pedestrians in Occluded-Duke.

    whereninrepresents the number of input features (that is, the number of neurons);at the same time,ninrepresents the amount of multiplication operations,nin-1 represents the amount of addition operations;+1 represents the bias;noutrepresents the number of output features.

    4.2.Experimental settings

    4.2.1.Lossfunction

    In this paper, Triplet Loss [37] is selected as the metric loss function (also known as distance loss function) while the Softmax Loss is selected as the classification loss function during training.In the metric loss function, a target sample, a positive sample and a negative sample are given.Among them, the target sample has the same label as the positive sample and a different label from the other negative sample.The Triplet Loss is calculated by taking the three samples into the following equation:

    wheredpis the distance between the target sample and the positive sample;dnis the distance between the target sample and the negative sample; α is the threshold of Triplet Loss.The goal of Triplet Loss is to pulldpcloser anddnfarther through training optimization, as shown in Fig.10.

    Fig.10.Schematic diagram of Triplet Loss.

    Softmax Loss is a loss function combined with Softmax function and Cross-entropy Loss.The function of the Softmax function is to map the output values of multiple neurons calculated by the network to the interval (0,1), thereby giving the predicted probability that the target sample belongs to each category.The Crossentropy Loss is a way to measure the difference between the predicted value of the model and the actual value based on the Softmax function,and the two are widely used in various classification tasks.

    Supposexrepresents the target sample andmrepresents the number of categories.Then the calculation formula of the Softmax function is

    The equation for calculating Cross-entropy Loss is

    In the equation,p(xj) andq(xj) represent the predicted probability and true probability that the target samplexbelongs to categoryj, respectively;Lcrossis the Softmax Loss value of the samplex.

    If there arensamples in a batch, the calculation equation of Softmax Loss is

    4.2.2.Parametersettings

    In order to ensure the consistency of the experimental results,the experimental process in this paper is carried out in the same software and hardware environment.The computing platform is a linux system of version 5.4.0, using a single GTX 1080Ti graphics card(memory is 11G),Pytorch version is 1.6.0,CUDA version is 10.1,and Python version is 3.7.7; During the training process, the training set is uniformly scale to 256×128 size,and perform data enhancement operations such as horizontal flip(probability is 0.5),pixel filling(padding is 10),and random erasure(probability is 0.5);The training batch is set to 64, including 16 randomly selected Pedestrian IDs,4 images are randomly sampled for each ID;The loss functions used for training are described in subsection 4.2.1,where the classification loss function is Softmax Loss(regularized by label smoothing with eps= 0.1),the metric loss function is Triplet Loss,and the Cosine function is used for distance calculation; The training process uses the adam optimizer to update the modelparameters, momentum is set to 0.9, and weight decay is set to 0.0005;The learning rate adopts a multi-step learning strategy,and the initial learning rate is set to 0.00035.When the iteration reaches 40th and 90th epochs,the learning rate is reduced to 0.1 times of the original one in turn,and the training process is stopped after 120 epochs; For testing process, the test set is also scaled to 256 × 128.

    Table 2Accuracy of SR-ResNet-50 and ResNet-50 on three different datasets.

    4.3.Performance of instantiated model

    We first test the instantiated model SR-ResNet-50 on three Re-ID authoritative datasets.As shown in Table 2, compared with ResNet-50,SR-ResNet-50 has significantly improved accuracy on all three datasets.On Market1501 dataset, SR-ResNet-50 is 1.55%higher on Rank-1 and 2.19%higher on mAP than ResNet-50;On the DukeMTMC-reID dataset, the accuracy of SR-ResNet-50 is also significant,with the Rank-1 reaching 88.51%and the mAP reaching 77.91%, which are 2.15% and 1.45% higher than the 86.36% and 76.46% of the ResNet-50 respectively.In the occlusion scene, SRResNet-50 algorithm shows strong robustness, its Rank-1 and mAP are increased from 49.73% to 43.33%-55.70% and 48.29%,which are increased by 5.97%and 4.96%,respectively.The necessity of explicitly modeling the interdependence between channel features is fully demonstrated.

    In addition to the accuracy improvement, the paper also measured the average training speed and average inference speed of SR-ResNet-50 on the above three different datasets to be as high as 206 FPS (i.e.Frames Per Second) and 562 FPS, respectively.Of course,in order to further prove the advantages of the proposed SR block, we conduct relevant experiments in terms of model computational complexity.Specifically, we selected four highly cited attention mechanisms such as CBAM [16], EPSA [38], GAM[39] and SK [40] to replace our SR block, and tested their performance on the Market1501 dataset.Firstly, as can be seen from experiments 1-5 in Table 3, the computational cost of other attention mechanisms except CBAM-ResNet-50 is much higher than SR-ResNet-50.In particular, the GFLOPs (taking a single color pedestrian image of 256×128 pixels as input)of SK-ResNet-50 and inference speed of GAM-ResNet-50 reach 118.202 and 8.740 ms/f(millisecond/frame),respectively.Then,experiments 5 and 6 show that SR-ResNet-50, compared with the pure ResNet-50, has achieved a good trade-off between improved accuracy and increased computational cost (only increases by 0.052 and 0.497 ms/f on GFLOPs and Inference speed, respectively).

    4.4.Ablation experiments

    In order to further prove the influence of using different configurations on the SR block, we conduct a series of ablation experiments on the instantiated model SR-ResNet-50 to explore the optimal design of the SR block in seven aspects,including effect of max pooling, scaling factor, weighting method, fusion method, order of components in the composite convolutional layer, different stages and integration strategy.It should be emphasized that, to accommodate the control-variable method, all ablation experiments only show the test results on the single dataset Market1501(other datasets are also found to have semantic homology in this paper), and all experiments adopt the same experimental settings as described in subsection 4.2.In particular, this paper supports a practical exploration of the various ablation strategies proposed for different computer vision tasks, which means that the design techniques of SR-ResNet-50 model (including SR block) may vary depending on the task.

    4.4.1.Effectofmaxpooling

    First, we conduct experiments to verify the effect of max pooling.As shown in Table 4, after deleting the max pooling branch in the SR block, the Rank-1 and mAP of SR-ResNet-50 decreased by 0.54% and 1.31% respectively despite saving the computing cost of 0.045 GFLOPs.It can be analyzed that the specificity cues introduced by the max pooling can help the SR to infer finer attention mask, which is an indispensable supplement to global spatial information.

    4.4.2.Scalingfactor

    The existence of "Squeeze-Excitation" operation makes the capacity and computational cost of SR block easier to adjust.As can be seen from Table 5, with the increase of the scaling factorr, all indicators have significantly different changes.The value of accuracy indicator Rank-1 is robust in a certain range (fluctuating around 95.00%), while the other accuracy indicator mAP shows a trend of increasing first and then decreasing.Different from Rank-1 and mAP, the computational complexity of the model gradually decreases with the increase ofr.Whenr= 16, the mAP reaches the maximum value of 88.30%, which is 0.27% higher than that when the channel is not scaled (r= 1), and the computation load is also reduced from 4.219 GFLOPs to 4.105 GFLOPs (reduced by about 2.7%).At this time, Rank-1 also approaches the maximum value of 95.46% with only 0.03% accuracy difference.

    Table 3Performance comparison of different attention mechanisms on the Market1501 dataset.

    Table 4Performance comparison of SR-ResNet-50 before and after introducing the max pooling.

    Table 5Effect of different r on model performance.

    Comprehensive analysis shows that the reduction of model complexity will not monotonically improve the performance,and a good balance between accuracy and computational complexity can be achieved whenr= 16.

    4.4.3.Weightingmethods

    In subsection 3.3,the weighting method adopted by SR block is introduced in detail,and in order to prove its rationality,all variants of the weighting method are compared experimentally.As shown in Fig.11,from the form,the weighting factor has two kinds:vector(with the same dimension as mask)and coefficient.For each form,the value of each element in the weighting factorsPandQhas three relations:pi,qi∈R;pi,qi∈(0,1);pi,qi∈(0,1)andpi+qi=1.In this paper, pairwise combination of the form of weighting factors and value relations is carried out to obtain a total of six combinations, and the six combinations are verified one by one.

    It can be clearly seen from Table 6 that the third group of experiments has the best accuracy,and the fifth group of experiments has the worst accuracy, but the computational complexity of the model remains unchanged.That is, compared with "P,Qare coefficients,pi,qi∈(0,1)", when "P,Qare vectors,pi,qi∈(0,1) andpi+qi= 1", SR-ResNet-50 achieves 0.72% and 0.51% improvement in Rank-1 and mAP, respectively, on the Market1501 dataset.In particular, comparing experiments 2 and 3, it can be seen that when"pi+qi=1",the Rank-1 and mAP of the model are increased from 94.92%to 88.10%-95.43%and 88.30%,respectively,indicating that there is indeed a"mutual exclusion"relationship between the max pooling mask and the average pooling mask in the Re-ID task,and this mutually exclusive relationship can be adjusted autonomously through network training;In addition,experiments 3 and 6 show that the performance indicators of SR-ResNet-50 when"P,Q are vectors"are better than those of SR-ResNet-50 when"P,Q are coefficients",which fully proves that the above"mutual exclusion"relationship is different in different channels.

    Fig.11.Schematic diagram of the weighting method of SR blocks: (a) The weighting factors are vectors; (b) The weighting factors are coefficients.

    Table 6Effect of different weighting methods on model performance.

    4.4.4.Fusionmethods

    Element-add and channel-concat are two commonly used feature fusion methods in deep learning.The method of Add is equivalent to the assumption that the feature maps of the corresponding channels of two input features are semantically similar,which is widely used in many models, especially in the ResNet series.When the relationship between the two-branch input features is not clear, Cat can splice the two inputs along the channel dimension,so as to achieve the purpose of allowing the network to autonomically learn the two-branch feature fusion method.InceptionNet [5] and DenseNet [41] are typical representatives of this fusion method.

    However, in practical application, the two fusion methods can be transformed into each other theoretically through certain processing [42].As shown in Fig.12, two input branches A and B are fused by Cat and then processed by convolution layer, which is equivalent to A and B being processed by convolutional layer and then performing add operation.Mathematically, it can be expressed as

    where * denotes matrix multiplication, which is the convolution operation; Cat(·,·) represents the merging of channel directions;Wrepresents convolutional layer,brepresents bias; ⊕ indicates that the elements are added.

    From Eq.(19), givenWandb, it’s easy to find the equivalent combination ofW1andW2,b1andb2, that is, two independent convolutional layers.It should be noted that, except for changing the fusion method, the structures of other network layers in the experiment remain the same.It can be seen from Table 7 that compared with the fusion method of element-add,the accuracy of Rank-1 and mAP is increased from 94.98% to 88.10%-95.43% and 88.30%,respectively,with a slight increase in the computation load.This coincides with the above analysis, indicating that channel merging is more robust to the fusion of two-branch input features with unknown relationships.

    4.4.5.Componentorder

    In Fig.2 and subsection 3.2,we elaborate several times that the component order of composite convolutional layer in the SR block is set as Conv1d-ReLU-BN(C-R-B).Scholars have different opinionson whether the BN layer should be placed in front of or behind the nonlinear activation layer ReLU:Ioffe[43],the original author of the BN layer technology, believes that it should be placed before the nonlinear layer, because the role of the BN layer is to smooth the input distribution of the hidden layer,so as to achieves the purpose of facilitating gradient propagation; However, Chen et al.[44]believe that BN layer should not be placed before ReLU,because the non-negative response of ReLU will make the weight layer updated in a suboptimal way.

    Table 7Effect of different fusion methods on model performance.

    We argue that the BN layer should be placed before or after the ReLU is related to the specific task and network structure.On the one hand, where BN layers are placed is dependent on the nonlinear activation function used (e.g., the difference between ReLU and Sigmoid); On the other hand, since most convolutional neural networks use a cyclic superposition of C,B,and R layers(e.g.-C-B-R-C-B-R-),the two ways should be equivalent to some extent.Therefore, the pros and cons should also be subject to the specific experimental conditions.

    As shown in Table 8 (R-B-C and R-C-B combinations are not discussed here because the nonlinear layer cannot directly process features), compared with other combinations, the composite convolutional layer in the order of C-R-B has the best performance.According to the experimental results of the first and second groups,no matter the index Rank-1 or mAP,for the re-ID task using ReLU activation function, the model with BN layer after the nonlinear layer has better accuracy.The combination effect of the third group B-R-C was the worst,with a difference of 1.01%in Rank-1 and 1.58% in mAP compared with the third group.The analysis shows that using BN layer to process the feature map directly will seriously destroy the correlation of the original information, and the subsequent non-negative response of ReLU function will cause further loss of the feature map information.

    Table 8Effect of different orders of C/B/R in composite convolutional layers on model performance.

    Fig.12.Equivalent schematic diagram of the fusion method of Cat and Add.

    Fig.13.Network structure of SR-ResNet-50.

    Table 9Effect of different stages of SR block insertion on model performance.

    4.4.6.Differentstages

    In order to explore the performance of SR blocks on different stages of the ResNet-50, we insert the SR blocks into Stage_1--Stage_4 of the ResNet-50 in turn, and obtain five different models SR-ResNet-50_Stage1,SR-ResNet-50_Stage2,SR-ResNet-50_Stage3,SR-ResNet-50_Stage4 and SR-ResNet-50 (indicating that SR blocks are inserted in all four stages, as shown in Fig.13), respectively.

    As can be seen from Table 9,the performance of the model first increases and then decreases as the SR block is inserted deeper into the network.According to experiments 2-5,SR-ResNet-50_Stage2 has the best effect in four different stages,reaching 95.28%of Rank-1 and 88.01% of mAP; Compared with SR-ResNet-50_Stage2, the performance of model SR-ResNet-50_Stage4 is significantly decreased,with only 94.33%of Rank-1 and 86.51%of mAP,and the Rank-1 and mAP are decreased by 0.95% and 1.5%, respectively.At the same time, the computational complexity increases by 0.029 GFLOPs,which corresponds to the phenomenon that the deeper the network is, the more channels there are.

    Looking across the first five sets of experiments,SR blocks lead to performance gains when introduced at each stage of the model.Moreover,from the experiment 6,it is known that the performancegains generated by the SR blocks at different stages are complementary, in the sense that they can be effectively combined to further improve the performance of the network.

    Table 10Effect of different integration strategies of SR blocks on model performance.

    4.4.7.Integrationstrategies

    Finally, in addition to the proposed SR block, four variants are considered in this paper,as shown in Fig.14:(c)SR-Pre,where the SR block is moved before the residual unit; (d) SR-Post, where the SR block is moved after the summation with the identity branch(after ReLU); (e) SR-Identity, where the SR block is placed on the identity connection,parallel to the residual unit;(f)SR-Parallel,The SR block is a separate branch,which is parallel to both the identity connection and the residual unit.

    Table 10 shows that, except for the standard SRBlock, other integration strategies of the SR block lead to the degradation of the model performance.Compared with the standard SRBlock, SRPostblock has the most significant performance degradation, with the Rank-1 and mAP are only 89.01% and 72.00%, respectively,which are 6.42%and 16.3%lower than the standard SRBlock on the same indicator.For SR-Parallelblock,Rank-1 and mAP also decrease significantly,to 92.52%and 80.66%,respectively,compared with the standard SRBlock.In addition,SR-ParallelBlock is much higher than the standard SRBlock in terms of computation complexity, which increases from 4.105 GFLOPs to 4.542 GFLOPs, with an increase of 10.6%.Taken together,standard SRblock is the best way to integrate SR block in the Re-ID task.

    Fig.14.Schematic diagram of the integration strategy for SR blocks: (a) ResBlock; (b) SRBlock; (c) SR-PreBlock; (d) SR-PostBlock; (e) SR-IdentityBlock; (f) SR-ParallelBlock.

    4.5.Algorithm comparison experiments

    In order to prove the superiority of the proposed algorithm,we selected a series of advanced algorithms for Re-ID task from 2018 to 2022 for comparison,including two literatures in 2018 such as PAN[45]and PCB[46],and three literatures in 2019 such as AANet[47],CAMA[48]and OSNet[49].REDA[50]and CAP[51]were published in 2020, CBN [52], GPS [53] and CAL [54] were published in 2021,and MHSA-Net[28]and DAAF-BOT[29]were published in 2022.So that adds up to 12 different algorithms,the selection principles are as follows:

    (1) Their structures are all based on convolutional neural networks;

    (2) They are representative algorithms in different Re-ID genres;

    (3) Experiments were carried out on Market1501 and DukeMTMC-reID datasets at the same time, and evaluation indicators Rank-1 and mAP were used.

    As can be seen from Table 11, for the Market1501 and DukeMTMC-reID datasets,the SR-ResNet-50 proposed in this paper outperforms most classical algorithms in performance.Compared with the PAN algorithm, the performance of SR-ResNet-50 is significantly improved on the two datasets.The Rank-1 and mAP on Market1501 exceed 13.23%and 24.9%,respectively,and the Rank-1 and mAP on DukeMTMC-reID also exceed 16.92% and 26.40%respectively;For the MHSA-Net algorithm in 2022,the mAP of SRResNet-50 in this paper still has 4.30%and 4.81%advantages on the two datasets respectively;Of course,for the state-of-the-art(SOTA)CAL, SR-ResNet-50 still has a slight gap of 0.07% in Rank-1 of Market1501.However,it should be noted that the SR-ResNet-50 in this paper does not make other improvements except backbone,nor does it add any fancy training techniques(such as Re-rank,etc.).In this paper,Re-rank post-processing technology was added on the basis of SR-ResNet-50, and experiments show that it can easily surpass all the algorithms listed above, but it is meaningless to do so.From this level, the performance of SR-ResNet-50 is very competitive.

    Next, we explore the generalization ability of SR-ResNet-50 on the occlusion dataset Occluded-Duke.It can be clearly seen from Table 12 that SR-ResNet-50 far exceeds other algorithms whetherit’s the indicator Rank-1 or mAP; For the most advanced HOReID,SR-ResNet-50 still outperforms by 0.6% and 4.49% on Rank-1 and mAP,respectively.The above fully proves the robustness of the SRResNet-50 in the occlusion scene, indicating that feature recalification is also beneficial to model the saliency information of pedestrians.

    Table 11Accuracy comparison of SR-ResNet-50 and advanced algorithms on the regular datasets.

    Table 12Accuracy comparison of SR-ResNet-50 and advanced algorithms on the Occluded-Duke dataset.

    Table 13Accuracy comparison of SR-ResNet-50 and advanced algorithms on the MSMT17 dataset.

    The results are shown in Table 13.Firstly,compare experiments 7 and 8,we can find that the Rank-1/mAP of ResNet-50 embedded with SR block is improved by 5.47%/3.75% compared with the original one,which proves that the improvement effect of SR block on model performance is still effective on MSMT17 dataset.Next,we select several advanced algorithms tested on the MSMT17 from 2019 to 2022 for comparison.From experiments 1-6 and 8, it can be seen that the our SR-ResNet-50 has certain advantages in accuracy.However,for the most advanced PNL algorithm,SR-ResNet-50 still has a large gap.Of course, this also provides us with an inspiration for the next step, which is that the performance of models that only embed simple enhancement modules can be further improved by adopting multi-aspect improvements or employing multiple training skills.

    4.6.Visualization results

    In order to directly reflect the actual retrieval performance of the model, we randomly select 3 query images for testing, and select the Top10 matching results according to the similarity from high to low.As shown in Fig.15,the first row of matching results for each query comes from the ResNet-50,while the second row comes from the proposed SR-ResNet-50 in this paper.As we can clearly see, a "yellow box" indicates a correctly matched target, while a"blue box" indicates a wrong match.Specifically, each retrieved instance is labeled as "201.243/false/cam0", where "201.243" indicates the similarity between the matched image and the query image, and "false" indicates that they are not the same ID (i.e., a wrong match)."cam0" means that the matched image originates from camera"0".

    Fig.15.Display of some matching results between SR-ResNet-50 and ResNet-50.

    From the presentation in Fig.15,we can find that the number of correct targets and ranking of SR-ResNet-50 in this paper are basically better than that of ResNet-50(baseline),and it has certain adaptability to complex environments such as blur, occlusion, and background interference.

    5.Conclusions

    In this paper, we propose a self-supervised recalibration network with plug-and-play SR block as the core, aiming at"improve performance + control computation load" for the Re-ID task.On the one hand, the SR block improves the representation ability of the model by explicitly modeling the interdependence between channel features, different from other channel attention modules,the channel recalibration factor of the SR block fuses both the global average information and local saliency information,and can autonomously recalibrate the feature response of the channel direction by means of channel weighted;On the other hand,the SR block regulates the channel dimension of the feature map through the designed SE unit, which effectively constrains the computational cost generated by the model.In particular, we integrate the SR block into ResBlock to build a new SR-ResNet-50 and take this as the research object.A series of ablation experiments and comparison experiments show that SR-ResNet-50 has better performance advantages on both regular dataset and occlusion dataset (especially the mAP exceeds the SOTA algorithm by 4.49% on the Occluded-Duke dataset), which plays a certain reference value for other computer vision tasks.

    Declaration of competing interest

    The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    Acknowledgements

    This work was supported in part by the Natural Science Foundation of Xinjiang Uygur Autonomous Region (Grant No.2022D01B186 and No.2022D01B05).

    国产黄a三级三级三级人| 一本一本综合久久| 99热只有精品国产| 成人国产一区最新在线观看| 搞女人的毛片| 熟女少妇亚洲综合色aaa.| 麻豆av在线久日| 国产午夜精品久久久久久| 婷婷精品国产亚洲av在线| 欧美激情久久久久久爽电影| 国产人伦9x9x在线观看| 亚洲人成电影免费在线| 丝袜人妻中文字幕| 国产一区在线观看成人免费| 欧美最黄视频在线播放免费| 欧美大码av| 亚洲七黄色美女视频| 日本免费一区二区三区高清不卡| 久久久久国产一级毛片高清牌| 国产精品久久久av美女十八| 久久国产精品影院| 欧美日韩精品网址| 人妻丰满熟妇av一区二区三区| 国产成人福利小说| 久久中文看片网| 国产不卡一卡二| 黑人巨大精品欧美一区二区mp4| www.自偷自拍.com| 亚洲欧美精品综合久久99| 又大又爽又粗| 男插女下体视频免费在线播放| 亚洲熟妇熟女久久| 亚洲国产日韩欧美精品在线观看 | 极品教师在线免费播放| 国产精品久久久久久亚洲av鲁大| 少妇熟女aⅴ在线视频| 一区二区三区国产精品乱码| 精品久久久久久成人av| 久99久视频精品免费| 亚洲五月天丁香| 欧美zozozo另类| 国产精品一区二区精品视频观看| www日本在线高清视频| 非洲黑人性xxxx精品又粗又长| 成人性生交大片免费视频hd| 免费看美女性在线毛片视频| 男女下面进入的视频免费午夜| 两个人的视频大全免费| 久久久久国产精品人妻aⅴ院| 亚洲中文日韩欧美视频| 国产精品一区二区免费欧美| 亚洲av电影不卡..在线观看| 欧美日韩福利视频一区二区| 黄色视频,在线免费观看| a级毛片a级免费在线| 国内少妇人妻偷人精品xxx网站 | 欧美在线黄色| 老汉色∧v一级毛片| 色老头精品视频在线观看| 国产精品一区二区免费欧美| 亚洲一区高清亚洲精品| 一本久久中文字幕| 亚洲性夜色夜夜综合| 一区福利在线观看| 成人特级黄色片久久久久久久| 免费在线观看日本一区| 午夜福利成人在线免费观看| 可以在线观看的亚洲视频| 亚洲九九香蕉| 中文字幕人成人乱码亚洲影| 男人舔奶头视频| 夜夜看夜夜爽夜夜摸| 欧美最黄视频在线播放免费| 99热这里只有精品一区 | 久久草成人影院| 欧美zozozo另类| 搡老岳熟女国产| 精品久久久久久,| 久久亚洲精品不卡| 久久久成人免费电影| 色在线成人网| 婷婷精品国产亚洲av| 成人高潮视频无遮挡免费网站| 中文资源天堂在线| 亚洲无线观看免费| 一个人看视频在线观看www免费 | 欧美3d第一页| 欧美乱妇无乱码| 波多野结衣高清作品| 中文字幕人妻丝袜一区二区| 国产一区二区在线av高清观看| 欧美黄色片欧美黄色片| www日本黄色视频网| 无遮挡黄片免费观看| 免费在线观看影片大全网站| 白带黄色成豆腐渣| 在线观看免费视频日本深夜| 午夜福利在线观看免费完整高清在 | 午夜影院日韩av| 国产一级毛片七仙女欲春2| 三级国产精品欧美在线观看 | 波多野结衣高清无吗| 麻豆成人av在线观看| 曰老女人黄片| 亚洲自偷自拍图片 自拍| 欧美日韩国产亚洲二区| 看片在线看免费视频| 亚洲国产欧美网| 国产精品 欧美亚洲| 国产综合懂色| 99热这里只有精品一区 | 淫妇啪啪啪对白视频| 婷婷丁香在线五月| 夜夜夜夜夜久久久久| 在线观看日韩欧美| 给我免费播放毛片高清在线观看| 一个人观看的视频www高清免费观看 | 精华霜和精华液先用哪个| www.自偷自拍.com| 日本免费一区二区三区高清不卡| 久久精品91蜜桃| 在线观看免费视频日本深夜| 成人国产综合亚洲| 熟妇人妻久久中文字幕3abv| 麻豆一二三区av精品| 精品一区二区三区av网在线观看| 制服人妻中文乱码| 亚洲欧美激情综合另类| 亚洲第一欧美日韩一区二区三区| 欧美中文综合在线视频| 精品久久久久久久久久免费视频| 国产久久久一区二区三区| 成人亚洲精品av一区二区| 欧美日韩福利视频一区二区| 欧美中文日本在线观看视频| 国产人伦9x9x在线观看| 精品国内亚洲2022精品成人| 美女被艹到高潮喷水动态| 91九色精品人成在线观看| 成人国产综合亚洲| 一区二区三区高清视频在线| 女人高潮潮喷娇喘18禁视频| 人人妻人人澡欧美一区二区| 成人高潮视频无遮挡免费网站| 三级男女做爰猛烈吃奶摸视频| 久久久久国产精品人妻aⅴ院| 精品国产亚洲在线| 亚洲精品一区av在线观看| 少妇人妻一区二区三区视频| 波多野结衣高清无吗| 日本五十路高清| 国产精品久久久av美女十八| 午夜福利在线在线| 日韩成人在线观看一区二区三区| 99久久精品热视频| 哪里可以看免费的av片| or卡值多少钱| 亚洲aⅴ乱码一区二区在线播放| 国产亚洲av嫩草精品影院| 在线观看日韩欧美| 亚洲精品456在线播放app | 日本一二三区视频观看| 一个人免费在线观看的高清视频| 九九久久精品国产亚洲av麻豆 | 国产黄色小视频在线观看| 欧美日韩黄片免| 激情在线观看视频在线高清| 精品99又大又爽又粗少妇毛片 | 国产精品永久免费网站| 啪啪无遮挡十八禁网站| 免费看美女性在线毛片视频| 国产成年人精品一区二区| 亚洲熟妇中文字幕五十中出| АⅤ资源中文在线天堂| 亚洲精品国产精品久久久不卡| 国产精品日韩av在线免费观看| 久久久久久国产a免费观看| 淫秽高清视频在线观看| 90打野战视频偷拍视频| 中文字幕熟女人妻在线| 一个人免费在线观看电影 | 成人精品一区二区免费| 亚洲成av人片在线播放无| 久久精品亚洲精品国产色婷小说| 亚洲专区字幕在线| 两性午夜刺激爽爽歪歪视频在线观看| 动漫黄色视频在线观看| 免费av不卡在线播放| 手机成人av网站| 国产亚洲av高清不卡| 人妻夜夜爽99麻豆av| 午夜免费成人在线视频| 一a级毛片在线观看| 中文字幕人妻丝袜一区二区| 18禁黄网站禁片免费观看直播| 国产又色又爽无遮挡免费看| 精品国产三级普通话版| 亚洲18禁久久av| 欧美zozozo另类| 国产成人aa在线观看| 欧美3d第一页| 亚洲天堂国产精品一区在线| 国产免费男女视频| 悠悠久久av| 久久精品夜夜夜夜夜久久蜜豆| 亚洲18禁久久av| 丰满的人妻完整版| 国产高清激情床上av| 一进一出好大好爽视频| 国产精品精品国产色婷婷| 国产精品 国内视频| 国产成人一区二区三区免费视频网站| 性欧美人与动物交配| 欧洲精品卡2卡3卡4卡5卡区| 国产在线精品亚洲第一网站| 精品久久久久久成人av| 亚洲人成网站高清观看| 国产精品一区二区三区四区免费观看 | 色噜噜av男人的天堂激情| 黄片小视频在线播放| 俄罗斯特黄特色一大片| 国产精品美女特级片免费视频播放器 | 国产 一区 欧美 日韩| 成人精品一区二区免费| 成人国产综合亚洲| 三级国产精品欧美在线观看 | 午夜激情福利司机影院| 久久久精品大字幕| 国产伦在线观看视频一区| 亚洲精品一卡2卡三卡4卡5卡| 国产乱人视频| 国产一区二区在线观看日韩 | 国产精品亚洲一级av第二区| 午夜福利在线在线| 香蕉av资源在线| 90打野战视频偷拍视频| 欧美xxxx黑人xx丫x性爽| 在线观看日韩欧美| 真实男女啪啪啪动态图| 国产亚洲精品av在线| 激情在线观看视频在线高清| 国产一级毛片七仙女欲春2| 99国产极品粉嫩在线观看| 久久亚洲精品不卡| 日韩三级视频一区二区三区| 俄罗斯特黄特色一大片| 中国美女看黄片| 欧美日韩一级在线毛片| 久久草成人影院| 怎么达到女性高潮| 免费在线观看视频国产中文字幕亚洲| 国产亚洲av高清不卡| 波多野结衣高清无吗| 久久婷婷人人爽人人干人人爱| 一二三四社区在线视频社区8| 欧美日韩乱码在线| 毛片女人毛片| 少妇的丰满在线观看| 亚洲人成网站高清观看| 国产成+人综合+亚洲专区| 全区人妻精品视频| 国产野战对白在线观看| 男女做爰动态图高潮gif福利片| 国产三级黄色录像| 啪啪无遮挡十八禁网站| 久久久久性生活片| 观看免费一级毛片| 国产伦精品一区二区三区四那| 成人特级黄色片久久久久久久| 久久这里只有精品中国| www.999成人在线观看| 性色av乱码一区二区三区2| 久9热在线精品视频| 99久久综合精品五月天人人| 一本综合久久免费| 日韩欧美国产一区二区入口| 免费高清视频大片| 天天躁日日操中文字幕| 欧美成人一区二区免费高清观看 | 国产精品久久久久久久电影 | 日韩精品青青久久久久久| 国产免费男女视频| 精品国产亚洲在线| 日韩欧美 国产精品| 午夜免费观看网址| 久久天躁狠狠躁夜夜2o2o| 欧美中文综合在线视频| 欧美绝顶高潮抽搐喷水| 每晚都被弄得嗷嗷叫到高潮| 一进一出好大好爽视频| 国产高清有码在线观看视频| 欧美日韩福利视频一区二区| 高清在线国产一区| av福利片在线观看| 最新中文字幕久久久久 | 青草久久国产| 国产精品一区二区免费欧美| 亚洲国产高清在线一区二区三| 亚洲成人中文字幕在线播放| 久久久久久大精品| 国产欧美日韩一区二区精品| 国产人伦9x9x在线观看| 久久精品亚洲精品国产色婷小说| www.自偷自拍.com| 国产精品精品国产色婷婷| 欧美色欧美亚洲另类二区| 一级毛片高清免费大全| 身体一侧抽搐| 久久久国产欧美日韩av| 香蕉av资源在线| 最近最新中文字幕大全免费视频| 老司机午夜福利在线观看视频| 亚洲自偷自拍图片 自拍| 1000部很黄的大片| 亚洲激情在线av| 91久久精品国产一区二区成人 | 亚洲熟妇中文字幕五十中出| 国产精品 国内视频| 91麻豆av在线| 国产精品亚洲美女久久久| 不卡一级毛片| 中文字幕人妻丝袜一区二区| 嫩草影院入口| 亚洲一区高清亚洲精品| 国产伦精品一区二区三区四那| 午夜两性在线视频| 欧美在线黄色| 久久中文字幕人妻熟女| 99久久国产精品久久久| 午夜精品一区二区三区免费看| 精品免费久久久久久久清纯| 欧美zozozo另类| 免费看日本二区| 757午夜福利合集在线观看| 欧美在线黄色| a级毛片在线看网站| x7x7x7水蜜桃| 精品久久久久久久人妻蜜臀av| 1000部很黄的大片| 久久精品国产亚洲av香蕉五月| 欧美性猛交╳xxx乱大交人| 国产精品亚洲av一区麻豆| 99国产精品99久久久久| 国产爱豆传媒在线观看| 哪里可以看免费的av片| 亚洲成人久久爱视频| 亚洲成av人片在线播放无| 欧美日本亚洲视频在线播放| 国产黄色小视频在线观看| 中出人妻视频一区二区| 亚洲成av人片在线播放无| 老汉色∧v一级毛片| 黄色女人牲交| 高清毛片免费观看视频网站| 最新美女视频免费是黄的| 黄色 视频免费看| 亚洲无线在线观看| av片东京热男人的天堂| 国产1区2区3区精品| 97碰自拍视频| 亚洲九九香蕉| 亚洲欧美激情综合另类| 午夜久久久久精精品| 三级国产精品欧美在线观看 | 国产v大片淫在线免费观看| 特级一级黄色大片| 一区福利在线观看| 99riav亚洲国产免费| 成人18禁在线播放| 在线观看免费视频日本深夜| 黄片小视频在线播放| 精品欧美国产一区二区三| 99久久精品一区二区三区| 91字幕亚洲| 免费观看人在逋| 国产久久久一区二区三区| 亚洲国产高清在线一区二区三| 免费高清视频大片| 久9热在线精品视频| 午夜成年电影在线免费观看| 天天一区二区日本电影三级| 一a级毛片在线观看| 国产精华一区二区三区| 天堂影院成人在线观看| 看黄色毛片网站| 亚洲av成人不卡在线观看播放网| 久久婷婷人人爽人人干人人爱| 亚洲五月天丁香| 国产高清三级在线| 国产一级毛片七仙女欲春2| 国产三级黄色录像| 97超级碰碰碰精品色视频在线观看| 搞女人的毛片| 精品国产三级普通话版| 亚洲五月婷婷丁香| 最新美女视频免费是黄的| 成人av一区二区三区在线看| 亚洲性夜色夜夜综合| 老司机午夜福利在线观看视频| 婷婷亚洲欧美| 久久久成人免费电影| 高清毛片免费观看视频网站| 精品一区二区三区av网在线观看| 1000部很黄的大片| 九九热线精品视视频播放| 国产亚洲欧美98| 亚洲欧美日韩东京热| 国产精品av视频在线免费观看| 丝袜人妻中文字幕| 欧美一级a爱片免费观看看| 国产爱豆传媒在线观看| 国产精品av久久久久免费| e午夜精品久久久久久久| 亚洲va日本ⅴa欧美va伊人久久| 一进一出好大好爽视频| 无人区码免费观看不卡| 欧美不卡视频在线免费观看| АⅤ资源中文在线天堂| 国产麻豆成人av免费视频| 欧美极品一区二区三区四区| 黑人巨大精品欧美一区二区mp4| 欧美午夜高清在线| 亚洲色图 男人天堂 中文字幕| 色综合欧美亚洲国产小说| 老熟妇乱子伦视频在线观看| 久久精品国产清高在天天线| 亚洲精品色激情综合| 亚洲色图 男人天堂 中文字幕| 熟妇人妻久久中文字幕3abv| 99热只有精品国产| 欧美丝袜亚洲另类 | 精品久久久久久久久久久久久| 免费在线观看影片大全网站| 成人特级av手机在线观看| 亚洲欧美精品综合久久99| 日本成人三级电影网站| 天堂√8在线中文| 成人三级做爰电影| av在线蜜桃| 国产黄色小视频在线观看| 久久久国产成人精品二区| av福利片在线观看| xxx96com| av片东京热男人的天堂| 欧美激情在线99| 国产成人欧美在线观看| 夜夜躁狠狠躁天天躁| 人妻夜夜爽99麻豆av| 亚洲狠狠婷婷综合久久图片| 国内久久婷婷六月综合欲色啪| 18禁观看日本| 久久久久久久精品吃奶| 欧美3d第一页| 亚洲精品一卡2卡三卡4卡5卡| 一a级毛片在线观看| 日日干狠狠操夜夜爽| 日本在线视频免费播放| 日本在线视频免费播放| 久久草成人影院| 国产aⅴ精品一区二区三区波| 少妇的逼水好多| 全区人妻精品视频| 欧洲精品卡2卡3卡4卡5卡区| 午夜精品一区二区三区免费看| 国产乱人伦免费视频| 国产高清激情床上av| 99精品久久久久人妻精品| 香蕉久久夜色| 在线观看一区二区三区| 久久九九热精品免费| 免费观看精品视频网站| 伊人久久大香线蕉亚洲五| 亚洲国产欧美一区二区综合| 亚洲 欧美 日韩 在线 免费| 国产成人aa在线观看| 日韩人妻高清精品专区| 免费在线观看亚洲国产| 国产不卡一卡二| 级片在线观看| 国产精品日韩av在线免费观看| 国产探花在线观看一区二区| 午夜福利免费观看在线| www日本在线高清视频| 久久精品91无色码中文字幕| 我的老师免费观看完整版| 午夜激情欧美在线| 中文字幕人成人乱码亚洲影| 午夜久久久久精精品| 九色成人免费人妻av| 亚洲av片天天在线观看| av视频在线观看入口| 午夜两性在线视频| 国内精品久久久久久久电影| 午夜福利欧美成人| 国产精品一区二区免费欧美| 中文字幕熟女人妻在线| 欧美不卡视频在线免费观看| 精品欧美国产一区二区三| 99国产极品粉嫩在线观看| 特大巨黑吊av在线直播| 婷婷六月久久综合丁香| 99热精品在线国产| 国产三级黄色录像| 超碰成人久久| 桃红色精品国产亚洲av| 久久欧美精品欧美久久欧美| 亚洲avbb在线观看| АⅤ资源中文在线天堂| 国产伦一二天堂av在线观看| 午夜福利在线观看吧| 每晚都被弄得嗷嗷叫到高潮| 亚洲第一欧美日韩一区二区三区| 国产私拍福利视频在线观看| 国产成人精品久久二区二区免费| 欧美zozozo另类| 日本黄大片高清| 日本黄色片子视频| 国产精品一及| 看黄色毛片网站| cao死你这个sao货| 国产高潮美女av| 久久精品国产99精品国产亚洲性色| 久久久精品大字幕| 国产精品自产拍在线观看55亚洲| 日韩精品中文字幕看吧| tocl精华| 成人三级黄色视频| 两个人看的免费小视频| 香蕉丝袜av| 欧美日韩综合久久久久久 | 天堂动漫精品| 国产伦在线观看视频一区| 亚洲精华国产精华精| 999精品在线视频| 国产97色在线日韩免费| 国产91精品成人一区二区三区| 久久久久久久久中文| 91在线精品国自产拍蜜月 | 国产高清视频在线观看网站| 91在线观看av| 日韩高清综合在线| 99国产极品粉嫩在线观看| 99久久成人亚洲精品观看| 国产一区二区激情短视频| 日韩国内少妇激情av| 久久性视频一级片| 女人高潮潮喷娇喘18禁视频| 级片在线观看| 偷拍熟女少妇极品色| 亚洲人成网站在线播放欧美日韩| 精品一区二区三区四区五区乱码| 久久精品国产亚洲av香蕉五月| 少妇的逼水好多| 欧美xxxx黑人xx丫x性爽| 国产三级中文精品| 久久伊人香网站| 久久久久久久午夜电影| 网址你懂的国产日韩在线| 欧美日韩福利视频一区二区| 午夜福利欧美成人| 精品久久久久久久人妻蜜臀av| 一个人观看的视频www高清免费观看 | 日韩精品青青久久久久久| 亚洲精品乱码久久久v下载方式 | av片东京热男人的天堂| 欧美黄色片欧美黄色片| 天天一区二区日本电影三级| tocl精华| 精品国产三级普通话版| 色综合婷婷激情| 午夜免费激情av| 女人高潮潮喷娇喘18禁视频| 久久久久久久久免费视频了| 在线观看美女被高潮喷水网站 | 啦啦啦观看免费观看视频高清| 欧美日韩乱码在线| 每晚都被弄得嗷嗷叫到高潮| 法律面前人人平等表现在哪些方面| 亚洲国产欧美人成| 又紧又爽又黄一区二区| 国产69精品久久久久777片 | 狂野欧美白嫩少妇大欣赏| 在线a可以看的网站| а√天堂www在线а√下载| 怎么达到女性高潮| 男女午夜视频在线观看| 亚洲国产色片| 国产高潮美女av| 国产真实乱freesex| 999久久久精品免费观看国产| 毛片女人毛片| 12—13女人毛片做爰片一| 99在线视频只有这里精品首页| 日本免费一区二区三区高清不卡| 美女 人体艺术 gogo| 2021天堂中文幕一二区在线观| 国产单亲对白刺激| 日韩成人在线观看一区二区三区| 美女扒开内裤让男人捅视频| 国产精品香港三级国产av潘金莲| 成人18禁在线播放| 1024香蕉在线观看| 久久伊人香网站| 亚洲熟妇中文字幕五十中出| 成人欧美大片| 亚洲av日韩精品久久久久久密| 欧美乱码精品一区二区三区| 久久精品综合一区二区三区| 91九色精品人成在线观看| 国产精品女同一区二区软件 | 欧美日韩国产亚洲二区| av在线天堂中文字幕| 99热这里只有是精品50| 亚洲av免费在线观看| 最近最新中文字幕大全电影3|