• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Few-shot object detection based on positive-sample improvement

    2023-11-11 04:07:06YanOuyangXinqingWangRuizheHuHonghuiXu
    Defence Technology 2023年10期

    Yan Ouyang, Xin-qing Wang, Rui-zhe Hu, Hong-hui Xu

    Department of Mechanical Engineering, College of Field Engineering, Army Engineering University of PLA, Nanjing, 210007, China

    Keywords:Few-shot learning Object detection Sample augmentation Attention mechanism

    ABSTRACT Traditional object detectors based on deep learning rely on plenty of labeled samples, which are expensive to obtain.Few-shot object detection(FSOD)attempts to solve this problem,learning detection objects from a few labeled samples, but the performance is often unsatisfactory due to the scarcity of samples.We believe that the main reasons that restrict the performance of few-shot detectors are: (1)the positive samples is scarce,and(2)the quality of positive samples is low.Therefore,we put forward a novel few-shot object detector based on YOLOv4, starting from both improving the quantity and quality of positive samples.First, we design a hybrid multivariate positive sample augmentation (HMPSA)module to amplify the quantity of positive samples and increase positive sample diversity while suppressing negative samples.Then,we design a selective non-local fusion attention(SNFA)module to help the detector better learn the target features and improve the feature quality of positive samples.Finally,we optimize the loss function to make it more suitable for the task of FSOD.Experimental results on PASCAL VOC and MS COCO demonstrate that our designed few-shot object detector has competitive performance with other state-of-the-art detectors.

    1.Introduction

    Object detection is one of the core tasks of computer vision,and the main task is to classify and locate objects in images.For the past few years, thanks to the extensive research and improvement of deep learning methods, the task of object detection has achieved unprecedented breakthroughs.Various deep learning based object detection methods[1-5]have been proposed one after another and gradually replaced traditional methods to become the mainstream in this field.However, the prosperity of these deep learning based methods in object detection tasks largely relies on plenty of annotated training data.A good deal of training data mainly comes from manual annotation,which requires plenty of time and human resources, and the cost is high [6].Moreover, in some special application scenarios, such as the detection of rare species or military targets,even a large amount of image data cannot be obtained.When the detector cannot obtain enough annotated training data,the detection model tends to be severely overfitted, resulting in degraded model performance and poor generalization ability.By comparison, humans can quickly learn new objects from a few examples [7]: for example, a human child can learn to recognize new objects that have never been seen in reality(such as a giraffe)from just one or a few pictures.Inspired by the human learning perspective [8], the concept of few-shot learning [9,10] was proposed and received extensive attention.

    Affected by complex environments (such as lighting, occlusion,etc.), objects in the real world have different shapes, and few-shot learning is a challenging task[11].In recent years,although various methods on few-shot learning[12-19]have emerged and achieved good results, most of them focus on solving image classification problems, and there are few methods involving few-shot object detection(FSOD).Object detection not merely requires to complete the task of category prediction, but also requires to accurately locate the object, which is harder than the simple image classification task.At present,for the exploration of FSOD methods,most of them inherit the idea of few-shot classification [7,11,20,21].Different from previous methods, for the problem of FSOD, we believe that from the perspective of samples, improving the performance of the detector by improving the quantity and quality of positive samples is a new idea worth studying.

    Our paper puts forward a new general FSOD model by improving positive samples.The main contributions of our paper are displayed as follows:

    (1) We design a hybrid multivariate positive sample augmentation module according to the actual situation, which can effectively reduce the risk of overfitting and improve the generalization performance of the model by expanding the annotated sample capacity and improving the sample distribution in a targeted manner.

    (2) We design a selective non-local fusion attention (SNFA)module to enhance the target feature expression and suppress other interfering features, thereby improving the feature extraction quality of samples for better detection performance.

    (3) We optimize the loss function of the detector.By introducing an orthogonal loss and modifying the original classification loss term, the model is more suitable for the task of FSOD.

    (4) Experimental results on PASCAL VOC and MS COCO data sets show that the FSOD model designed by us has good detection performance, and is superior to other FSOD models in the case of scarce samples, which can better adapt to the FSOD task.

    The rest of this paper is organized as follows:Section 2 reviews related work on FSOD.Section 3 details our raised model.Section 4 introduces the implementation details and analyzes the experimental results.Finally,Section 5 concludes the work of this article.

    2.Related work

    2.1.General object detection

    Object detection has always been a research hotspot in the field of computer vision.Especially with the rise of deep learning [22],the object detection method based on CNN has attracted the attention and research of many scholars.The current mainstream object detection models can be split into two types: two-stage detectors based on proposals and one-stage detectors without proposals.The R-CNN series [1-3] is representative of two-stage detectors, which produce several ROI proposals that may contain objects in the first stage,and then use bounding box regression and classifier to detect objects on the proposed area in the second stage.While one-stage detectors, such as the YOLO series [4,23-25] and SSD [5], do not need to generate ROI proposals and can directly classify and locate objects.Generally speaking, one-stage object detectors are much faster than two-stage object detectors,but have relatively poor detection accuracy, which is determined by their structure.Although these detectors can achieve excellent detection performance, they are all based on plenty of annotated training samples,and it is difficult to learn to detect novel classes through a few samples.

    2.2.Few-shot learning

    The task of few-shot learning(FSL)is to learn novel classes from a small amout of training examples.So far,a great deal of FSL work has investigated how to solve this novel classes generalization problem with scarce samples.At present,the main FSL methods can be split into three types: transfer-learning based, metric-learning based and meta-learning based.Transfer learning is to solve new problems by exploiting previously learned knowledge [26].Early transfer-learning based FSL methods [10,27,28] used a Bayesian framework to transfer prior knowledge of previously learned categories to novel categories.Hariharan et al.[29] proposed a fewshot transfer learning method combined with deep neural networks.Blaes et al.[30] proposed a learning process based on a global feature layer (Global Prototype Learning, GPL) for few-shot learning.Bauer et al.[31] proposed DLPM to transfer representative knowledge and classifier weights to FSL.Metric learning achieves classification by learning the distribution of distances between samples.Triplet network [32] induces similarity with respect to triplet images through a deep ranking neural network.Vinyals et al.[15]proposed a matching network based on memory and attention mechanism, which can identify unlabeled data using the matching metric of the data in the attention embedding.The prototype network [12] learns a metric space for classification via calculating the distance to the prototype representation of each class.Meta-learning elevates learning from the data to the task level, and can generalize well to new tasks by training in the labeled task space.Ravi et al.[13]Proposed a metalearner based on LSTM for few-shot learning.Finn et al.[16] proposed a model-agnostic meta-learning algorithm (MAML) to reduce training steps by learning better initial parameters.Li et al.[18].proposed a stochastic gradient descent (Meta-SGD) based meta-learner,which overcame the difficulty of Meta-LSTM training.Lee et al.[33]analyzed the best conditions’implicit differentiation for the convex problem of linear classifiers,allowing the network to embed high-dimensional features for fast generalization.However,most of these works are for few-shot classification, which are unable to use directly for FSOD.

    2.3.Few-shot object detection

    For the past few years, some FSOD methods have emerged one after another,but most of the existing methods are inherited from the few-shot classification task.Chen et al.[20] combined the characteristics of Faster R-CNN [3] and SSD [5] on classification tasks and bounding box regression tasks, and put forward a lowshot transfer detector (LSTD) based on knowledge transfer regularization and background suppression regularization, which utilizes knowledge of source and target domains respectively to strengthen the model's generalization ability to few-shot data.But it relies heavily on source domain knowledge and is less robust.RepMet[21] introduces prototype clustering, constructs prototype space to extract image embedding features, and uses Euclidean distance to compute the distance between embedding features and support set class representation.But this network ignores the object location and the performance is not outstanding.Kang et al.[7]used a meta-weighted model to fuse the support set features to the query set features in a channel multiplication manner, which requires an additional mask branch and increases the computational complexity.Fan et al.[11] combined an attention-based region proposal network (RPN) with a strategy of comparative training,using the idea of a matching network[15]to calculate the similarity between query instances and support instances.P′erez-Rúa et al.[34]proposed a meta-learning method for central point prediction by referring to the structure and ideas of CenterNet[35].This model can realize incremental learning, that is, no need to access base class data after adding new classes.Hu et al.[36]designed a Gabor convolutional kernel-based method for few-shot object detection,which achieved better accuracy and recall on few-shot datasets.MPSR [37] and SAR [38] studied the multi-scale problem in FSOD,and designed a feature pyramid module to generate multi-scale features and refine them at different scales.Zhang et al.[39]combined the recently popular Transformer [40] with metalearning, and proposed an image-level meta-learning FSOD model.Different from the above work, we try to focus on a few number of given positive samples instead of relying on the relation with the outside world, which could be labile.

    3.Proposed method

    In order to better expand the discussion of this paper,we restate the protocols of the FSOD task as follows:The classes in all training data are separated into base classes and novel classes,and the two classes of data do not overlap each other.The base class includes plenty of labeled training data, while the novel class has only a small amount of labeled training data available.The training of FSOD is generally divided into two phases.In training phase 1, a large amount of data containing only the base class is used for basic training of the detection network.Then, in training phase 2, the detection network is fine-tuned by using a small amount of balanced training data containing the novel and base classes.The final trained detection network can be used to detect novel classes.The flow chart of FSOD is displayed in Fig.1.

    Detecting objects of novel classes as accurately as possible by using a few number of samples is a challenging task.In this paper,we design a FSOD model based on YOLOv4[25]for addressing this challenge.The overall network structure of our detection model is displayed in Fig.2.Overall,our model is divided into two parts:the basic detection branch and the reinforcement branch.To satisfy the balance between detection accuracy and detection speed,we adopt the one-stage detector YOLOv4 as the baseline.The basic detection branch is mainly implemented through the network structure of YOLOv4.The reinforcement branch is mainly implemented through a parallel Siamese network structure that shares weights with the basic branch.The advantage of using Siamese network structure is that it can maintain the continuity of training and inference.As shown in Fig.2, in the inference process, we only need to use the basic detection branch at the bottom, while the reinforcement branch at the top can be removed.Specifically, we first preprocess the novel class images using our designed HMPSA module to increase the diversity and difference of the novel class samples.The augmented sample images are then sent to the reinforcement branch for feature extraction and prediction.During feature extraction,we optimize the feature representation of the target by using a SNFA module,so that the network can extract better novel class target features.As shown at the bottom of Fig.2, our SNFA includes two parallel branches,namely the spatial attention branch and the channel attention branch.The attention features generated by the two branches are sent to 1×1 convolution layer to reshape the channel size after concatenate fusion.Finally, the result obtained by the reinforcement branch is sent to the optimized loss function together with the result of the basic branch.

    Following the basic setup of transfer learning-based FSOD models, our training process is separated into two stages: in the first stage, only the base class is trained on the basic detection branch, following the training process of YOLOv4, while the reinforcement branch is frozen.In the second stage,the network is finetuned, except for the detection head used for bounding box regression and classification, other weight parameters are frozen,that is, the second stage is initialized with the weight parameters obtained from the first stage training to perform the transfer of base class knowledge.Fine-tuning is only for the last layer, using a balanced dataset containing all novel class samples and an approximate number of base class samples as training data,avoiding problems of overfitting and poor generalization performance.During inference,the reinforcement branch is removed,and the data of novel classes without labels get detection results through the basic detection branch.

    3.1.Basic detection model

    Fig.1.Flow chart of few-shot object detection.

    Fig.2.Network structure of our proposed YOLOv4-based few-shot object detection model.SAM: spatial attention module, CAM: channel attention module.

    We chose YOLO v4 as the base detector in order to balance detection speed and accuracy.As shown at the bottom of Fig.2,YOLOv4 consists of three components: backbone, neck and detection head.The backbone adopts CSPDarknet for feature extraction.In our detection model,we use SNFA to optimize the backbone for better feature extraction performance.The neck uses a modified path aggregation network (Modified-PAN) for feature fusion.The top-down and bottom-up feature fusion mechanism can augment the semantic information of features and obtain more accurate location information.After fusion, the three-layer features with different resolutions are sent to the detection head for classification and localization.

    3.2.Hybrid multivariate positive sample augmentation

    In the task of FSOD, the number of labeled positive samples is very rare, which cannot be fully trained as in traditional object detection, which is the key to what plagues few-shot detector performance.The scarcity of samples makes it impossible for the network model to learn comprehensive target information.When the target environment changes greatly (such as scale changes,illumination changes, and occlusions), the detector often has false detections or even missed detections.For this problem, sample augmentation methods can provide help.However, the object processed by the traditional sample augmentation method is the entire image,and a large amount of background information will be introduced when the positive sample is augmented.There are often a lot of interference information (such as negative samples irrelevant to the target) in these backgrounds.Adding more negative samples while increasing the number of positive samples will lead to more imbalance between positive and negative samples, which is not wise, especially in few-shot detection tasks.

    In response to the above problems, we design a hybrid multivariate positive sample augmentation(HMPSA)module to amplify the number of our positive samples while suppressing the interference of negative samples.Without increasing the cost of additional labeling, our HMPSA module starts from the data domain,increases the capacity of labeled positive samples in a targeted manner,and enriches the training samples,thereby improving the performance of the detection model.

    As displayed in Fig.3, our HMPSA module is mainly separated into two major steps: background sparsity and sample augmentation.Firstly,in order to suppress the negative samples’interference,we sparse the background of the image and extract the positive sample region from the image.The positive sample region takes the ground truth box as the main body and expanded outward by 20 pixels to simulate common scenes.The rest of the image is used as the background and its pixel value is set to 0.Then,to increase the diversity of positive samples, we perform hybrid multivariate augmentation on the positive sample regions.In order to fully simulate the environment of the sample and enhance the robustness of the detection model, we employ five augmentation submodules.Specifically, for different lighting environments (such as cloudy days,sunny days,dusk,etc.)where the target is located,we use HSV color space augmentation to augment the sample data.We use rotation augmentation to augment sample data for different geographic environments where the target is located.For different scales and positions of the target, we use scaling and translation augmentation to increase the sample data.To simulate blurring during image acquisition, we augment the sample data using blur augmentation.In order to simulate various occlusion situations in the actual environment, we adopt object-aware random erasure augmentation [41] to increase the sample size.Fig.3 shows the workflow of our designed positive sample augmentation module.First, the annotated positive sample regions first implement HSV color space augmentation,rotation augmentation,and scaling and translation augmentation respectively.Then, the data augmented by these sub-modules is augmented a second time, which is subjected to blur augmentation and object-aware random erasure augmentation [41].Finally, the augmented data of the first three sub-modules and the augmented data of the last two sub-modules are output together.The specific augmentation operations of each sub-module are as follows:

    (1) HSV color space augmentation:The representation space of the original image is RGB color space, we first convert the original image from RGB space to HSV space

    where (h,s,v) represent the three components of Hue, Saturation and Value in HSV space, respectively.(r,g,b) represent the Red,Green and Blue components of the RGB space,respectively.max =max(r,g,b),min = min(r,g,b).

    Then, we add different degrees of random disturbance to the three channels of (h,s,v) for tone transformation, so as to expand the samples.Specifically,we divide the random disturbance weight W into four ranges(w1,w2,w3,w4)to simulate the transformation of four different lighting environments.Where w1?[ - 1, - 0.5),w2?[ - 0.5,0),w3?[0,0.5),w4?[0.5,1].The HSV channel components (hg,sg,vg) of the sample after increasing the disturbance are

    where λh,λs,λvare the disturbance coefficients corresponding to the three channels (h,s,v) respectively.The random disturbance weight W is randomly selected within the defined four sub-ranges,and each sub-weight corresponds to an augmented sample.

    (2) Rotation augmentation: In different geographical environments,the poses of the objects in the image are different and diverse.To meet the pose diversity requirement of the sample object, we design 6 poses through rotation augmentation.They are 18°, 45°, 72°clockwise rotation and 18°, 45°, 72°counterclockwise rotation, respectively.

    Fig.3.Hybrid multivariate positive sample augmentation module: (a) background sparsity; (b) HSV color space augmentation; (c) rotation augmentation; (d) scaling and translation augmentation; (e) blur augmentation; (f) object-aware random erasure augmentation.The red box represents the ground truth box, and the blue box represents the expanded positive sample region.

    (3) Scaling and translation augmentation: For the diversity of scales and locations of objects, we augment the samples by scaling and translation augmentation.First, multi-scale scaling is performed on the input image, and the scope of scaling is limited to: the size of the positive sample area is smaller than the original input image, and n samples of different scales are obtained.Then,we randomly crop these n samples to the same size as the original input image,and the cropped region does not contain the object.Each sample is randomly cropped for m times, so that a total of n× m samples with different scales and location distribution can be obtained.

    (4) Blur augmentation: We simulate the blurring of objects in image samples in real environments by blurring the images.We use a Gaussian distribution to calculate the value of each pixel in the region, with a blur radius of 5 and a variance of 1.5.

    (5) Object-aware random erasure augmentation [41]: We select a random erase area for the object bounding box to achieve sample amplification for occlusion situations, and the specific operation is borrowed from Ref.[41].

    3.3.Selective non-local fusion attention

    The traditional object detection method learns the feature distribution of the object through the training of abundant positive samples, and distinguishes different classes according to the feature distribution law.For the task of FSOD,positive samples are scarce, and it is difficult to grasp the object features through traditional training methods.In response to this problem, we consider to enhance the feature expression of the object by designing a selective non-local fusion attention (SNFA) module,suppress other interfering features, start with the feature quality,and improve the performance of the object detector by learning high-quality positive sample features.Compared with the traditional attention mechanism,the SNFA we designed starts from both local feature saliency and global feature correlation, which can allocate attention more reasonably,highlight the key features of the object, and exclude the interference of background feature at the same time.

    The SNFA we designed contains two sub-modules, a spatial attention sub-module and a channel attention sub-module, which are used to tell the neural network to focus on“where”and“what”[42].Fig.4 visualizes the workflow of the two attention submodules.

    where msijrepresents the influence of the j th position on the i th position.At the same time,we create a novel feature C?RC×H×Wby feeding the feature X into the convolutional layer.Likewise, we reshape C into C?RC×M.Then perform matrix multiplication on C and the transpose of Msto get the correlation feature D?RC×M,and then reshape D into D?RC×H×W.In the branch at the bottom, we first perform the maximum pool and average pool operations [42]on the input feature X?RC×H×Walong the channel axis, concatenate the obtained results and send them to the 7×7 convolution layer and sigmoid to obtain the spatial importance weight Ws?R1×H×W.Then perform element-wise multiplication of Wsand the spatial pixels of X to obtain the importance feature E?RC×H×W

    where Θ represents element-wise multiplication,σ represents the sigmoid function, and c7×7represents the 7×7 convolution operation.Finally,we weighted the features of the two branches to get the final output feature Ys?RC×H×W:

    Fig.4.Schematic diagram of selective non-local fusion attention.Two attention sub-modules are included: (a) spatial attention module; (b) channel attention module.

    where α and 1-α are the scale coefficients of D and E,respectively,α is initialized to 0.5, and the weight gradually updated through learning.

    (2) Channel attention module:Likewise,our channel attention module also contains two branches.As shown in Fig.4(b),in the top branch,for the input feature X?RC×H×W,we reshape X into X?RC×M,then perform matrix multiplication on X and the transpose of X,and obtain the channel correlation matrix Mc?RC×Cthrough a softmax layer

    where mcijrepresents the influence of the j th channel on the i th channel.At the same time,we conduct matrix multiplication on the transpose of Mcand X to obtain the correlation feature F?RC×M,and then reshape F into F?RC×H×W.In the bottom branch,we first perform max-pooling and average-pooling operations [42] on the input features X?RC×H×Walong the spatial axis, and send the obtained results into two fully connected layers.Perform elementwise summation on the output of the fully connected layer and send it to sigmoid to get the channel importance weight Wc?RC×1×1.Then perform element-wise multiplication of Wcwith the channel of the input feature X to obtain the importance feature G?RC×H×W.

    where fc1and fc2represent two fully connected layers,respectively,and ρ represents the ReLU function.Finally,the features of the two branches are also weighted and fused to get the final output feature Yc?RC×H×W:

    where β and 1-β are the scale coefficients of F and G,respectively,β is initialized to 0.5, and the weight gradually updated through learning.

    3.4.Optimized loss function

    The loss function is the objective function of the detection network learning and updating.A good loss function can often effectively boost the performance of model.In this paper, the training of the network is separated into two processes: the pretraining process based on the base class and the fine-tuning process based on the novel class(with the appropriate base class added to balance the positive and negative samples).The pre-training process is no different from traditional object detection.In this stage, our loss function follows the original design of yolov4.However, in the process of fine-tuning, due to the scarcity of annotated samples, the detection head of the network is difficult to learn reliable feature representation like general detection task.In this case,it is easy for the classifier to confuse the novel class target with other similar base classes,resulting in a large number of false detections and hurting the performance of the model.The research in Ref.[45] shows that in the FSOD task, the bounding box regression task can achieve good results after being fully pretrained on the base class data and then fine-tuned on the novel class.Likewise, the confidence loss measures the reliability of the predicted bounding box, independent of the number of positive samples.Therefore, for the fine-tuning process for few-shot detection, we only optimize the classification loss on the original basis.Specific improvements are described below.

    First, for the characteristics of FSOD tasks, we introduce an orthogonal loss Lo[46].As shown in Fig.5, by applying orthogonality to the feature space before classification, the features of different classes are separated and the features of the same class are aggregated, thereby ensuring that the similarity between features can be well measured under the condition of few samples.Let the input and output pairs of the network be {xi,yi}, xiis the input image,yiis the class label,and the feature before the classification layer is F = {f1,f2,……,fn}, The orthogonal loss Lois defined as follows:

    where φ is the hyperparameter that controls the weight, |?|represents the absolute value operation, and ‖?‖2represents the ?2norm operation.In the orthogonal loss Lo, the clustering between samples of the same class is promoted by constraining 1-p to approach 0, and the separation of samples of different classes is promoted by constraining q to approach 0.

    Then, we improve the original classification loss and suppress the confusing targets of other classes by highlighting the weight of similar classes.The modified classification loss is expressed as follows:

    where i is the grid cell,which is divided into K×K in total.j is the anchor,and a grid cell has a total of S anchors.Iobjijindicates whether the corresponding anchor contains the object, and ^pi(c) and pi(c)indicate the truth value and predicted value of the class, respectively.The value of ^pi(c) is 0 or 1.Different from the original classification loss function, we add a weight term (1 +pi(c))μto the second term to suppress hard negative samples, where μ is a hyperparameter greater than 1 to control the degree of suppression according to the magnitude of different predicted values.When^pi(c) =0,it means that the sample is a negative sample.The larger pi(c)is,the more likely it is to be wrongly determined as a positive sample,that is,it is easier to become a hard negative sample that is difficult to identify by the network.The weight item we add can adaptively increase the loss proportion of hard samples according to the value of pi(c)and effectively suppress hard negative samples.To sum up, our optimized classification loss is

    Fig.5.Schematic diagram of the action process of the orthogonal loss.The same shapes and colors in the figure represent the same class(such as orange stars),and different colors and shapes belong to different classes (such as green triangles and blue squares).

    where θ is the hyperparameter controlling the orthogonal loss ratio.

    4.Experiments

    In this section, we evaluate the performance of our FSOD detector through experiments on two general object detection benchmarks.The main contents include: dataset setup, experimental details, experimental results and ablation experiments.

    4.1.Dataset setup

    We mainly conduct experiments on two widely used object detection benchmarks,PASCAL VOC 2007[47]+2012[48]and MS COCO [49].When experimenting on the PASCAL VOC, following convention[3,23],we use the training set of VOC2007 +2012 and the validation set of VOC2012 to train the detector,and use the test set of VOC2007 for evaluation.PASCAL VOC contains a total of 20 classes, following the settings of Refs.[7,37,49], we perform three different partitions for these 20 classes, each of which contains 15 base classes and 5 novel classes.The three division methods are as follows (the first five are novel classes, and rest represents all the remaining classes as base classes):

    ◆Split1:{bird,cow,bus,motorbike,sofa/rest}

    ◆Split2:{horse,cow,bottle,aeroplane,sofa/rest}

    ◆Split3:{cat,sheep,boat,motorbike,sofa/rest}

    In the first stage of training, all training samples are from the base class.In the second stage of training, only the novel class samples containing K labeled objects and the base class samples containing K labeled objects are used,and K in the experiment is set to 1, 2, 3, 5,10.MS COCO is a larger-scale object detection dataset consisting of 80 classes, with a total of 80 k training images, 40 k validation images and 20 k testing images.For experiments on the MS COCO, following the setting of Ref.[50], we use all training set images and 35 k validation set images to train the detector and use the remaining 5 k validation set images for evaluation.At the same time, the same 20 classes as PASCAL VOC are considered as novel classes, and the rest 60 classes are treated as base classes.In the fine-tuning stage, we set K to 10 and 30.For the sake of better evaluation about the cross-domain generalization ability of the model [51,52], we use 60 classes different from Pascal VOC in MS COCO as the base class training model,and detect 20 novel classes on Pascal VOC.For evaluation metrics, we use AP50to evaluate model performance on PASCAL VOC, using AP50∶95, AP50, and AP75to evaluate model performance on MS COCO.

    4.2.Implementation details

    In all training stages, we use SGD as the optimizer with momentum of 0.9, mini-batch size set to 16, and L2weight decay of 0.0005.In the first stage of training,the batch size is set to 64,the basic detection branch trains 160 epochs in total, and the initial learning rate is set to 10-4.and the strategy of gradual learning rate decay is used:the learning rate becomes 10-3at 10 epochs,10-4at 80 epochs, and 10-5at 120 epochs.For the second-stage finetuning,considering that the CutMix[53]and Mosaic[25]strategies on YOLOv4 will damage a few number of novel samples,we turn off the original data augmentation strategy of YOLOv4 and only use our HMPSA module processes samples.Meanwhile, only the last fully connected layer of the detection head is trained, other layers are frozen.The reinforcement branch's parameters are consistent with the basic detection branch.The model is trained for a total of 500 epochs,and the learning rate is set to 10-3.Other hyperparameter settings are displayed in Table 1.

    4.3.Comparison with the state-of-the-art

    In this section, we mainly verify the performance of our raised few-shot object detector through three groups of experiments.The three groups of experiments are: experiments on PASCAL VOC,experiments on MS COCO and cross-domain experiments from MS COCO to PASCAL VOC.Specifically, we compare our detector with several state-of-the-art few-shot object detectors: Meta R-CNN[50], TFA [54], MPSR [37], Retentive-RCNN [55], DCNet [56],LSTD(YOLO)-full[20]and Meta YOLO[7].At the same time,we also use the unmodified YOLOv4 as the few-shot detector for comparison.The training strategy of YOLOv4 is also separated into two phrases.In the first phrase, the same training strategy as our detector is adopted.In the second phrase, the network is fine-tuned by using the novel class data until the network is fully converged.We refer to this method as YOLOv4-tf-full for short.

    4.3.1.Results of PASCAL VOC

    Table 2 reveals the comparison of novel class detection performance between our model and other detectors on PASCAL VOC.First, the experimental results indicate that our proposed optimization strategy for FSOD can indeed effectively boost the detection performance compared with the YOLOv4 baseline detector.Especially in the case of extremely scarce samples(K=1,2,3),the gain amplitude is particularly obvious.In split 1, when K = 1, the performance improves by as much as 24.9%.When the sample increases (K = 5, 10), the performance of the YOLOv4 baseline is greatly improved, but our method still surpasses the baseline.Compared to the baseline, our method achieves an average performance improvement of 5.2% and 5.8% at K = 5 and K = 10,respectively.

    Furthermore,our model also displays outstanding performance compared to other FSOD detectors.For the LSTD and Meta YOLO models whose baselines are also YOLO series, our model has a significant performance advantage.On the one hand, this benefits from the improved baseline performance,but on the other hand,it relies more on our proposed model improvements for FSOD.Our model is also competitive compared to the models designed based on Faster R-CNN.As revealed in Table 2, our model can achieve optimal or sub-optimal performance when samples are scarce(K = 1, 2, 3).Even though the sample is gradually increased, our model is at a relatively advanced level.

    Table 1Hyperparameter settings.

    Table 2Performance(mAP50)comparison of various FSOD methods on PASCAL VOC.The novel class detection performance for three different partitions is reported separately.“FRCN R-101” is an abbreviation for Faster R-CNN ResNet-101, “-" indicates that no results were reported in the original text, and the best performance is shown in bold.

    4.3.2.Results of MS COCO

    Compared to PASCAL VOC, MS COCO is a more challenging detection benchmark with more categories and more complex backgrounds.Table 3 reports our detection performance evaluation results for 20 novel classes in the 10-shot and 30-shot cases.Likewise, our model has a clear performance advantage over the YOLOv4 baseline detector and other state-of-the-art FSOD models.Whether it is 10-shot or 30-shot, our model achieves leading performance.This also strongly proves the availability of our raised improvement strategy for improving FSOD.

    4.3.3.Results of MS COCO to PASCAL VOC

    To evaluate the cross-domain generalization ability of our model,we combine PASCAL VOC and MS COCO for testing.We train the model with 10 novel class annotated objects on MS COCO and evaluate on the VOC2007 test set.We adopt Meta R-CNN, MPSR,LSTD(YOLO)-full,Meta YOLO and YOLOv4+ft-full as the comparison benchmarks for cross-domain performance.Fig.6 displays the experimental results.From Fig.6 we can clearly find that our method achieves the best performance.At the same time, we also find that the cross-domain novel class detection performance becomes worse compared to the intra-domain experimental results on PASCAL VOC.On the one hand, this is related to the increase in the number of novel classes for evaluation (from 5 to 20), on the other hand, it also shows that there is a performance degradation problem in cross-domain detection, which needs further research to solve.

    4.4.Ablation studies

    Fig.6.10-shot cross-domain detection performance(mAP50)comparison for 20 novel classes under COCO to PASCAL.

    To evaluate the different factors’influence on the results in our model, we performed an ablation study on split 1 of PASCAL VOC,and Table 4 reveals the results of our ablation experiments.HMPSA,SNFA and OCL (including OL and IMP) represent our proposed improved components, namely a hybrid multivariate positive sample augmentation module, a selective non-local fusion attention module and an optimized classification loss, respectively.We combine these components in various ways to observe their specific contributions to the results.

    4.4.1.Hybrid multivariate positive sample augmentation module

    The first two rows of Table 4 indicate that the HMPSA module dramatically boosts the model's performance.In different settings of K = 1, 2, 3, 5,10, mAP is improved by 9.1%, 7.6%, 8.3%, 4.0% and 3.4%, respectively.Moreover, we find that it helps performancemore when the samples are scarce,which is exactly what is needed in few-shot detection.Meanwhile,HMPSA contributes more to the performance improvement than the other two factors.HMPSA optimizes the distribution of novel class positive samples,increases samples' diversity, and enables the model to better learn sample features in different environments when given a few positive samples, which fundamentally effectively avoids the overfitting and poor generalization caused by scarcity.At the same time,HMPSA suppressed the negative samples' interference, which further improves the model performance.

    Table 3Comparison of the performance of various FSOD methods on the MS COCO (using AP50:95,AP50 and AP75, respectively).The performance under 10-/30-shot is reported respectively.“FRCN R-101”is an abbreviation for Faster R-CNN ResNet-101,“-"indicates that no results were reported in the original text,and the best performance is shown in bold.

    Furthermore, for the sake of further demonstration about the superiority of our raised sample augmentation method,we make a comparison between our HMPSA and several other sample augmentation methods on the same benchmark.These methods include: HMPSA-removed benchmarks (which we denote by ‘*‘),SINP [57], SNIPER [58], and Object Pyramid [37].SINP and its improved version SNIPER enhance specific areas rather than all pixels.The Object Pyramid expands the scale of positive samples by sampling.We did a comparative experiment with K = 1, 5, and 10 on split 1 of PASCAL VOC,and the experimental results are revealed in Fig.7.It can be found from Fig.7 that our method is superior to several other methods.In SINP and SNIPER,the number of negative samples is also amplified, so the performance is even lower than the baseline.Although the Object Pyramid eliminates the interference of negative samples and is helpful for the detection of few samples, it only expands the samples in terms of scale and cannot meet the actual detection needs.On the contrary, our method can not only suppress the interference of negative samples, but also amplify positive samples according to the actual situation and increase the diversity of samples.

    Fig.7.Comparison of the performance of our HMPSA with several methods on the PASCAL VOC(taking split 1 as an example).K=1,5 and 10 are compared respectively.*: Baseline +SNFA +OCL.

    4.4.2.Selective non-local fusion attention module

    Comparing the data in the first and third rows of Table 4,we can see that SNFA also help a lot in improving the performance of the model.When K = 1, 2, 3, 5, and 10, mAP increased by 6.8%, 2.5%,3.4%, 1.4% and 1.5%, respectively.This verifies the effectiveness of SNFA for improving detection performance.For the sake of more intuitive demonstration about the help of our raised SNFA for performance improvement, we visualize the attention of some novel classes on two datasets respectively.Fig.8 shows the attention comparison results: Baseline vs.SNFA.The top image of each dataset is the baseline attention result with the SNFA module removed, and the bottom image is the attention result with the SNFA module added.As can be seen from Fig.8,our SNFA is able to provide better attention regions for novel class detection,i.e.better express novel class features.For a simple single object existing in the image,the attention of our SNFA and the baseline is not much different, but the SNFA can pay more attention to the object itself rather than the environment,as shown in Fig.8(a),Fig.8(c),Fig.8(f)and Fig.8(g).Our SNFA is also able to provide more attention when the objects are occluded, such as the bird in Fig.8(b).When there are other significant interfering objects in the image, such as the motorbike in Fig.8(d)and the dog in Fig.8(i),SNFA can reasonably allocate attention and avoid confusing the detector.For the case where there are multiple objects and small objects in the image,the SNFA focus area is also more reasonable, such as horse in Fig.8(e)and boat in Fig.8(j).Overall,our proposed SNFA is able to optimize the feature representation of positive samples,making it easier for the network to learn purely novel class features,thereby improving detection accuracy.

    4.4.3.Optimized loss function

    We split the optimized loss function into an orthogonal loss and an improved classification loss and evaluate them separately.The fourth and fifth rows of Table 4 reveal the evaluation results.Under the settings of K=1,2,3,5,10,the mAP is improved by 4.7%,1.7%,1.9%,0.9%and 0.9%for the orthogonal loss part,respectively.For the improved classification loss part, mAP is improved by 3.8%, 1.4%,1.5%, 0.7% and 0.6%, respectively.In general, for the case of scarce samples(K=1,2,3),the optimized loss function can bring at least 3% performance improvement, which can further boost the fewshot detector's performance.

    In addition, we also conduct experimental analysis of three hyperparameters in the process of optimizing the loss function.Table 5 displays the experimental results.Our experiments are conducted on split 2 of PASCAL VOC and report results for K=1,5,10 settings.The first row of Table 5 is the best hyperparameters that we adjusted through repeated experiments.First, we fix the hyperparameter μ in the improved classification loss part and modify the two hyperparameters φ and θ related to the orthogonal loss.When θ is held constant and φ is changed, we find that the model performance has a trend of increasing first and then decreasing as the value of φ increases.This shows that in the orthogonal loss,although high weights that promote the separation of different classes of samples are more suitable for few-shot detection, the weights cannot be increased excessively, otherwise the model performance will be compromised.Similarly, this trend also exists on θ, according to the data, higher orthogonal loss weights can lead to better performance for the model compared to the improved classification loss term.Furthermore, we found that changing θ has a larger impact on performance(more performance fluctuations) than changing φ.Then, we fix φ and θ, change μ and see how it affects the results.We found that the trend of rising first and then falling also exists, which indicates that although increasing the weight of hard negative samples can improve the model's performance,we should not give excessive concern to the hard negative samples.

    Fig.8.Visualization results of attention on some novel classes on two datasets:Baseline vs.SNFA:(a)Bus;(b)Bird;(c)Cow;(d)Motorbike;(e)Horse; (f)Train;(g)Aeroplane; (h)Cow; (i) Dog; (j) Boat.

    Table 5Comparison of the performance impact of different hyperparameter settings in the optimized classification loss function (taking VOC split 2 as an example).

    Finally,we combine the three components in pairs and observe their impact on performance.The sixth,seventh and eighth rows in Table 4 show the experimental results.It can be revealed from the obtained data that the superposition of each component will not harm each other, but can achieve better performance after superposition.

    4.5.Visualization and analysis of test results

    We present visualizations of our detection results for novel classes on both datasets in Fig.9.We separately picked some typical success and failure cases to help analyze our model's performance.From the success stories in the figure, we can see that our model can obtain good performance when the object is relatively simple and single.When the object is located in a complex environment,such as occlusion,small object scale,or other significant interfering objects in the scene, our model can still show excellent performance.But even so, our model still suffers from some false detection failures, as shown in the rightmost column in Fig.9.It can be found from the figure that these failure cases are mainly caused by the lack of key features of the object or being too similar to other classes, which shows that distinguishing confusing objects under the circumstance of scarce samples is a worthy research direction.

    5.Conclusions

    This article raises a new FSOD model on the benchmark of YOLOv4.Our model mainly starts from improving the quantity and quality of samples,which can effectively boost the performance of FSOD tasks.Specifically, we first put forward a HMPSA module,which can amplify the number of positive samples for different actual situations and increase the diversity of samples.Second,we design a SNFA module, which can achieve better feature representation and enable the network to learn higher-quality positive sample features.Finally,for the specific task of few-shot detection,we optimize the loss function of the baseline detector to better serve the task of few-shot detection.Experiments on two benchmark data sets indicate that our model is better than other FSOD methods when samples are scarce,and can better adapt to the task of FSOD.

    Fig.9.Visualization of detection results of our model for novel classes on PASCAL VOC and MS COCO.Including success cases(green boxes)and failure cases(red boxes),gray boxes are detected irrelevant classes.

    Declaration of competing interest

    The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    Acknowledgements

    The authors would like to acknowledge the China National Key Research and Development Program(Grant No.2016YFC0802904),National Natural Science Foundation of China(Grant No.61671470)and 62nd batch of funded projects of China Postdoctoral Science Foundation (Grant No.2017M623423) to provide fund for conducting experiments.

    精品99又大又爽又粗少妇毛片| 亚洲精品乱码久久久久久按摩| 国产精品福利在线免费观看| 国产精品久久久久久精品电影| 激情五月婷婷亚洲| 欧美激情在线99| 少妇裸体淫交视频免费看高清| 大香蕉97超碰在线| 香蕉精品网在线| 久热这里只有精品99| 三级经典国产精品| 久久99精品国语久久久| 国产高清国产精品国产三级 | 美女主播在线视频| 啦啦啦在线观看免费高清www| 久久精品久久精品一区二区三区| 免费在线观看成人毛片| 国产色爽女视频免费观看| 午夜免费男女啪啪视频观看| 国产女主播在线喷水免费视频网站| 中文欧美无线码| 久久精品人妻少妇| 亚洲在久久综合| 久久ye,这里只有精品| 国产高潮美女av| 国产欧美日韩精品一区二区| 一边亲一边摸免费视频| 波野结衣二区三区在线| 日日摸夜夜添夜夜添av毛片| 国产 一区 欧美 日韩| 午夜精品一区二区三区免费看| 高清在线视频一区二区三区| 亚洲成人久久爱视频| 热re99久久精品国产66热6| 女人被狂操c到高潮| 熟妇人妻不卡中文字幕| 久久久a久久爽久久v久久| 新久久久久国产一级毛片| 亚洲av在线观看美女高潮| 成年av动漫网址| 欧美成人精品欧美一级黄| 香蕉精品网在线| 乱码一卡2卡4卡精品| 成人二区视频| 国产精品一区www在线观看| 午夜激情福利司机影院| 久久99热6这里只有精品| 99热这里只有是精品在线观看| 欧美xxxx性猛交bbbb| 国产一级毛片在线| 日日啪夜夜爽| 成年av动漫网址| 亚洲人与动物交配视频| 五月天丁香电影| 亚洲欧美日韩卡通动漫| 18禁在线无遮挡免费观看视频| 好男人视频免费观看在线| 18禁在线无遮挡免费观看视频| 最近手机中文字幕大全| 成年免费大片在线观看| 成人亚洲精品av一区二区| 欧美 日韩 精品 国产| 观看免费一级毛片| 人妻 亚洲 视频| 又爽又黄a免费视频| 久久精品久久久久久噜噜老黄| 各种免费的搞黄视频| 18禁裸乳无遮挡动漫免费视频 | 亚洲av男天堂| av又黄又爽大尺度在线免费看| 国产一区二区在线观看日韩| 草草在线视频免费看| 一级毛片 在线播放| 亚洲av.av天堂| 91久久精品国产一区二区三区| 亚洲精品久久午夜乱码| 亚洲国产欧美人成| 少妇的逼好多水| 国产男人的电影天堂91| 高清av免费在线| 欧美另类一区| 校园人妻丝袜中文字幕| 国产精品久久久久久av不卡| 免费大片18禁| 天美传媒精品一区二区| 欧美zozozo另类| 男女边吃奶边做爰视频| 免费不卡的大黄色大毛片视频在线观看| 搡女人真爽免费视频火全软件| 日本午夜av视频| 亚洲精品日韩在线中文字幕| 国产女主播在线喷水免费视频网站| 欧美最新免费一区二区三区| 国产伦精品一区二区三区四那| 亚洲美女搞黄在线观看| 直男gayav资源| 国产成人freesex在线| 亚洲真实伦在线观看| 成人漫画全彩无遮挡| 秋霞在线观看毛片| 另类亚洲欧美激情| 亚洲真实伦在线观看| av一本久久久久| 丝袜脚勾引网站| 啦啦啦啦在线视频资源| 亚洲性久久影院| 婷婷色麻豆天堂久久| 精品久久久久久久久av| 国产片特级美女逼逼视频| 亚洲成人一二三区av| 18+在线观看网站| 国产乱人视频| 国产成人福利小说| 免费在线观看成人毛片| 欧美激情国产日韩精品一区| 日本午夜av视频| 亚洲精品久久午夜乱码| 亚洲图色成人| 国产老妇女一区| 美女主播在线视频| 日本爱情动作片www.在线观看| 色视频www国产| 熟女av电影| 亚洲精品自拍成人| 男女国产视频网站| 免费看av在线观看网站| 国产精品.久久久| 精品国产乱码久久久久久小说| 日本av手机在线免费观看| 国产精品一二三区在线看| 看黄色毛片网站| 纵有疾风起免费观看全集完整版| 国产亚洲午夜精品一区二区久久 | 国产黄片美女视频| 久久久久久久久大av| 精品久久久久久久末码| 国产综合懂色| 麻豆久久精品国产亚洲av| 看非洲黑人一级黄片| 最近中文字幕高清免费大全6| 国产精品一二三区在线看| 97超碰精品成人国产| 天天躁夜夜躁狠狠久久av| 亚洲内射少妇av| 国产片特级美女逼逼视频| 亚洲三级黄色毛片| 寂寞人妻少妇视频99o| 免费电影在线观看免费观看| 久久久精品欧美日韩精品| av卡一久久| 亚洲va在线va天堂va国产| 春色校园在线视频观看| 美女视频免费永久观看网站| 久久人人爽av亚洲精品天堂 | 日韩一本色道免费dvd| 岛国毛片在线播放| 亚洲精品456在线播放app| 一本一本综合久久| 下体分泌物呈黄色| 国产亚洲av片在线观看秒播厂| 免费看不卡的av| 三级经典国产精品| 在线观看一区二区三区激情| 老女人水多毛片| 欧美变态另类bdsm刘玥| 国产黄片美女视频| 高清午夜精品一区二区三区| 亚洲一区二区三区欧美精品 | 久久热精品热| 亚洲一区二区三区欧美精品 | 精品久久国产蜜桃| 亚洲国产av新网站| 色视频在线一区二区三区| 国产精品嫩草影院av在线观看| 亚洲人成网站在线观看播放| 日本wwww免费看| 亚洲欧美中文字幕日韩二区| 国产毛片a区久久久久| 各种免费的搞黄视频| 麻豆精品久久久久久蜜桃| 亚洲性久久影院| 3wmmmm亚洲av在线观看| 亚洲国产最新在线播放| 欧美97在线视频| 纵有疾风起免费观看全集完整版| 久久ye,这里只有精品| 国产精品嫩草影院av在线观看| 在线观看一区二区三区激情| .国产精品久久| 精品久久久久久电影网| 老司机影院成人| 免费黄网站久久成人精品| 夫妻性生交免费视频一级片| 在线观看三级黄色| 水蜜桃什么品种好| 亚洲,欧美,日韩| 2021天堂中文幕一二区在线观| 午夜亚洲福利在线播放| 成人二区视频| 看十八女毛片水多多多| 国产伦精品一区二区三区视频9| 视频区图区小说| 精品一区二区三卡| 精品久久久噜噜| 另类亚洲欧美激情| 成人黄色视频免费在线看| 啦啦啦中文免费视频观看日本| 又爽又黄无遮挡网站| 黄色配什么色好看| 高清日韩中文字幕在线| 2021少妇久久久久久久久久久| 日韩电影二区| 日日啪夜夜撸| 国产成人a∨麻豆精品| 国产在线一区二区三区精| 成年女人看的毛片在线观看| 有码 亚洲区| 秋霞在线观看毛片| av福利片在线观看| 日韩制服骚丝袜av| 亚洲av免费高清在线观看| 欧美97在线视频| 久久久亚洲精品成人影院| 久久ye,这里只有精品| 建设人人有责人人尽责人人享有的 | 不卡视频在线观看欧美| 亚洲伊人久久精品综合| 国产在视频线精品| 免费人成在线观看视频色| 亚洲精品乱码久久久久久按摩| 国国产精品蜜臀av免费| 黄色视频在线播放观看不卡| 日韩在线高清观看一区二区三区| 亚洲怡红院男人天堂| 丰满少妇做爰视频| 秋霞伦理黄片| 久久国产乱子免费精品| 男男h啪啪无遮挡| 在线观看人妻少妇| 午夜日本视频在线| 69人妻影院| 男女那种视频在线观看| 成人毛片a级毛片在线播放| 男女边吃奶边做爰视频| 亚洲精品456在线播放app| av天堂中文字幕网| 亚洲国产av新网站| 丝袜脚勾引网站| 国产伦在线观看视频一区| 亚洲精品成人久久久久久| 性色avwww在线观看| 精品人妻一区二区三区麻豆| 永久免费av网站大全| 国产高清国产精品国产三级 | 插阴视频在线观看视频| 80岁老熟妇乱子伦牲交| av免费观看日本| 久久久久九九精品影院| 日韩欧美一区视频在线观看 | 99热全是精品| 国产欧美亚洲国产| 欧美精品国产亚洲| 中文字幕免费在线视频6| 18+在线观看网站| 免费大片黄手机在线观看| 成人午夜精彩视频在线观看| 一级毛片aaaaaa免费看小| 男人添女人高潮全过程视频| av在线老鸭窝| tube8黄色片| 夫妻性生交免费视频一级片| 精品一区在线观看国产| 成人美女网站在线观看视频| 深夜a级毛片| 中文乱码字字幕精品一区二区三区| 青春草视频在线免费观看| av国产免费在线观看| 日本一二三区视频观看| 国产精品一区二区在线观看99| 国产精品人妻久久久影院| 欧美 日韩 精品 国产| 久久久久久九九精品二区国产| 嫩草影院入口| 一级毛片黄色毛片免费观看视频| 国产欧美另类精品又又久久亚洲欧美| 国产久久久一区二区三区| 国产极品天堂在线| 国产 精品1| 国产老妇女一区| 少妇裸体淫交视频免费看高清| 大又大粗又爽又黄少妇毛片口| 精品国产三级普通话版| av女优亚洲男人天堂| 白带黄色成豆腐渣| 深爱激情五月婷婷| 国产又色又爽无遮挡免| 国产成人福利小说| 一级毛片aaaaaa免费看小| 色吧在线观看| 久久久久久久久久久丰满| 久久99蜜桃精品久久| 国产片特级美女逼逼视频| 免费av不卡在线播放| 免费大片18禁| 免费av观看视频| 自拍欧美九色日韩亚洲蝌蚪91 | 中文字幕制服av| 午夜日本视频在线| 国产精品一区www在线观看| 国产高清不卡午夜福利| 男男h啪啪无遮挡| 日韩在线高清观看一区二区三区| 一级a做视频免费观看| 新久久久久国产一级毛片| 亚洲成人精品中文字幕电影| 少妇人妻精品综合一区二区| 午夜福利高清视频| 国产精品99久久99久久久不卡 | 99久久精品国产国产毛片| 一个人看的www免费观看视频| 少妇裸体淫交视频免费看高清| 另类亚洲欧美激情| 亚洲一级一片aⅴ在线观看| 在线亚洲精品国产二区图片欧美 | 午夜精品国产一区二区电影 | 人妻少妇偷人精品九色| 国产精品一区二区三区四区免费观看| 国产成年人精品一区二区| 特大巨黑吊av在线直播| 亚洲精品自拍成人| 91精品伊人久久大香线蕉| 国产成人午夜福利电影在线观看| 久久久久久国产a免费观看| tube8黄色片| 亚洲av男天堂| 极品少妇高潮喷水抽搐| 日韩制服骚丝袜av| 别揉我奶头 嗯啊视频| 亚洲熟女精品中文字幕| 成人二区视频| 亚洲成人中文字幕在线播放| 国产精品伦人一区二区| 国产真实伦视频高清在线观看| 最近的中文字幕免费完整| 三级国产精品欧美在线观看| 18禁在线无遮挡免费观看视频| 日韩国内少妇激情av| 狂野欧美激情性xxxx在线观看| 国产探花在线观看一区二区| 久久人人爽人人爽人人片va| 欧美潮喷喷水| 色哟哟·www| 国产日韩欧美在线精品| 五月玫瑰六月丁香| 欧美一级a爱片免费观看看| 我的女老师完整版在线观看| 又爽又黄a免费视频| 99热国产这里只有精品6| 干丝袜人妻中文字幕| 汤姆久久久久久久影院中文字幕| 秋霞在线观看毛片| 国产极品天堂在线| 免费看av在线观看网站| 少妇人妻一区二区三区视频| 亚洲成人一二三区av| av在线观看视频网站免费| 国产黄色视频一区二区在线观看| 久久久亚洲精品成人影院| 成年人午夜在线观看视频| 午夜福利在线观看免费完整高清在| 丰满少妇做爰视频| 熟女电影av网| 亚洲四区av| 九九在线视频观看精品| 亚洲欧美清纯卡通| 国产午夜福利久久久久久| 18禁裸乳无遮挡免费网站照片| 波野结衣二区三区在线| 婷婷色综合www| av天堂中文字幕网| 王馨瑶露胸无遮挡在线观看| 在线a可以看的网站| 一个人观看的视频www高清免费观看| 日韩,欧美,国产一区二区三区| 亚洲av一区综合| 成人午夜精彩视频在线观看| 欧美性猛交╳xxx乱大交人| av在线亚洲专区| 欧美性猛交╳xxx乱大交人| 99久久精品一区二区三区| 日韩av免费高清视频| 久久久久久久亚洲中文字幕| 午夜激情久久久久久久| 午夜免费观看性视频| 日本猛色少妇xxxxx猛交久久| 免费黄频网站在线观看国产| 2018国产大陆天天弄谢| 午夜视频国产福利| 国产精品精品国产色婷婷| 97超碰精品成人国产| 亚洲内射少妇av| 欧美97在线视频| 又粗又硬又长又爽又黄的视频| 夜夜爽夜夜爽视频| 国国产精品蜜臀av免费| 97超碰精品成人国产| 青青草视频在线视频观看| 高清视频免费观看一区二区| 色播亚洲综合网| 亚洲色图av天堂| 毛片一级片免费看久久久久| 国产老妇女一区| 国产亚洲av嫩草精品影院| 亚洲第一区二区三区不卡| 亚洲最大成人手机在线| 日韩欧美 国产精品| 国产精品久久久久久精品电影小说 | 亚洲av中文字字幕乱码综合| 久久人人爽av亚洲精品天堂 | eeuss影院久久| 丝瓜视频免费看黄片| 精品久久久久久电影网| 男女那种视频在线观看| 久久久久久久午夜电影| 色5月婷婷丁香| 午夜免费鲁丝| 三级经典国产精品| 春色校园在线视频观看| 亚洲精品aⅴ在线观看| 国产中年淑女户外野战色| 久久99精品国语久久久| 国产美女午夜福利| 亚洲欧美精品专区久久| 狠狠精品人妻久久久久久综合| 亚洲av成人精品一二三区| 免费黄频网站在线观看国产| 97在线视频观看| 亚洲精品一二三| 在线观看美女被高潮喷水网站| 一个人看视频在线观看www免费| 久久6这里有精品| 波野结衣二区三区在线| 亚洲国产色片| 免费看a级黄色片| 国产精品伦人一区二区| 99热网站在线观看| 人妻 亚洲 视频| 欧美性感艳星| 老司机影院毛片| 毛片女人毛片| 久久热精品热| 特大巨黑吊av在线直播| 亚洲精品乱久久久久久| 国产成人午夜福利电影在线观看| 亚洲aⅴ乱码一区二区在线播放| 国产av码专区亚洲av| 五月天丁香电影| 亚洲精品乱码久久久久久按摩| 日本黄色片子视频| 国产精品av视频在线免费观看| 观看免费一级毛片| 搡老乐熟女国产| 国产一区二区三区综合在线观看 | 国产精品一区二区在线观看99| 大香蕉97超碰在线| 国产黄色免费在线视频| 免费黄网站久久成人精品| 免费电影在线观看免费观看| 日韩欧美一区视频在线观看 | 久久精品综合一区二区三区| 成人高潮视频无遮挡免费网站| 性色avwww在线观看| 午夜福利网站1000一区二区三区| 99久久精品国产国产毛片| 九九在线视频观看精品| 久久精品夜色国产| 久久久久精品久久久久真实原创| 国产高清三级在线| 成人漫画全彩无遮挡| 99久久九九国产精品国产免费| 丰满少妇做爰视频| 亚洲一区二区三区欧美精品 | 在线观看三级黄色| 欧美老熟妇乱子伦牲交| 精品久久久久久久末码| 亚洲精品aⅴ在线观看| 热re99久久精品国产66热6| 精品一区二区免费观看| 一级毛片aaaaaa免费看小| 午夜免费鲁丝| 成年女人看的毛片在线观看| 久久人人爽人人片av| 看非洲黑人一级黄片| 免费少妇av软件| 噜噜噜噜噜久久久久久91| 国产毛片a区久久久久| 男人和女人高潮做爰伦理| 亚洲精品久久久久久婷婷小说| 亚洲av不卡在线观看| 中国三级夫妇交换| 免费av观看视频| 六月丁香七月| av网站免费在线观看视频| 如何舔出高潮| 黑人高潮一二区| 97超碰精品成人国产| 大片免费播放器 马上看| 永久免费av网站大全| 干丝袜人妻中文字幕| 日韩大片免费观看网站| 日日摸夜夜添夜夜添av毛片| 亚洲婷婷狠狠爱综合网| 亚洲精品视频女| 午夜精品国产一区二区电影 | 中文在线观看免费www的网站| 男女边吃奶边做爰视频| 网址你懂的国产日韩在线| 五月天丁香电影| 亚洲精品日韩av片在线观看| 精品久久久噜噜| 久久人人爽av亚洲精品天堂 | 又粗又硬又长又爽又黄的视频| 搞女人的毛片| 在现免费观看毛片| 欧美日韩精品成人综合77777| 精品久久久噜噜| 嫩草影院新地址| av.在线天堂| 麻豆精品久久久久久蜜桃| 国产精品秋霞免费鲁丝片| 小蜜桃在线观看免费完整版高清| 亚洲婷婷狠狠爱综合网| 国产精品国产三级专区第一集| 亚洲精品乱码久久久v下载方式| 97在线视频观看| 黄色日韩在线| 97精品久久久久久久久久精品| 又大又黄又爽视频免费| 国产在线男女| 丰满少妇做爰视频| 2021天堂中文幕一二区在线观| 高清在线视频一区二区三区| 国产亚洲最大av| 大陆偷拍与自拍| 午夜激情福利司机影院| 七月丁香在线播放| 中国美白少妇内射xxxbb| 婷婷色综合www| 日韩欧美精品v在线| 国产精品久久久av美女十八| 久久精品久久精品一区二区三区| 在线看a的网站| 老司机靠b影院| 777米奇影视久久| 黄色视频不卡| 一级a爱视频在线免费观看| 亚洲av成人精品一二三区| 久久久欧美国产精品| 亚洲国产欧美网| av在线观看视频网站免费| 男女午夜视频在线观看| 综合色丁香网| 人成视频在线观看免费观看| 亚洲av成人不卡在线观看播放网 | 啦啦啦在线观看免费高清www| 老汉色av国产亚洲站长工具| 9热在线视频观看99| 一区二区av电影网| 亚洲精华国产精华液的使用体验| 咕卡用的链子| 国产精品偷伦视频观看了| 丝袜喷水一区| 观看美女的网站| 国产免费福利视频在线观看| 国产精品久久久久久久久免| 国产亚洲午夜精品一区二区久久| 亚洲av日韩精品久久久久久密 | 九九爱精品视频在线观看| 久久ye,这里只有精品| 免费观看av网站的网址| 国产成人欧美| 2018国产大陆天天弄谢| 夫妻性生交免费视频一级片| 久久久久久免费高清国产稀缺| 久久久国产一区二区| 伦理电影大哥的女人| 一边摸一边做爽爽视频免费| 国产一区亚洲一区在线观看| 黄色视频不卡| 久久人人爽人人片av| av又黄又爽大尺度在线免费看| 精品酒店卫生间| 麻豆乱淫一区二区| 亚洲av福利一区| 90打野战视频偷拍视频| 国产精品国产三级国产专区5o| 一本—道久久a久久精品蜜桃钙片| a级片在线免费高清观看视频| 婷婷色综合www| 99热国产这里只有精品6| 18禁国产床啪视频网站| 国产精品嫩草影院av在线观看| 日本一区二区免费在线视频| 午夜免费男女啪啪视频观看| 国产在视频线精品| 一区在线观看完整版| 在线观看免费视频网站a站| 亚洲国产欧美网| 欧美日韩视频精品一区| 日韩精品免费视频一区二区三区| 欧美日韩国产mv在线观看视频| 街头女战士在线观看网站| 午夜福利乱码中文字幕| 亚洲欧美清纯卡通| 日韩精品有码人妻一区|