• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    CAW-YOLO:Cross-Layer Fusion and Weighted Receptive Field-Based YOLO for Small Object Detection in Remote Sensing

    2024-03-23 08:17:26WeiyaShiShaowenZhangandShiqiangZhang

    Weiya Shi ,Shaowen Zhang and Shiqiang Zhang

    1College of Artificial Intelligence and Big Data,Henan University of Technology,Zhengzhou,450001,China

    2College of Information Science and Engineering,Henan University of Technology,Zhengzhou,450001,China

    ABSTRACT In recent years,there has been extensive research on object detection methods applied to optical remote sensing images utilizing convolutional neural networks.Despite these efforts,the detection of small objects in remote sensing remains a formidable challenge.The deep network structure will bring about the loss of object features,resulting in the loss of object features and the near elimination of some subtle features associated with small objects in deep layers.Additionally,the features of small objects are susceptible to interference from background features contained within the image,leading to a decline in detection accuracy.Moreover,the sensitivity of small objects to the bounding box perturbation further increases the detection difficulty.In this paper,we introduce a novel approach,Cross-Layer Fusion and Weighted Receptive Field-based YOLO (CAW-YOLO),specifically designed for small object detection in remote sensing.To address feature loss in deep layers,we have devised a cross-layer attention fusion module.Background noise is effectively filtered through the incorporation of Bi-Level Routing Attention(BRA).To enhance the model’s capacity to perceive multi-scale objects,particularly small-scale objects,we introduce a weighted multi-receptive field atrous spatial pyramid pooling module.Furthermore,we mitigate the sensitivity arising from bounding box perturbation by incorporating the joint Normalized Wasserstein Distance(NWD)and Efficient Intersection over Union(EIoU)losses.The efficacy of the proposed model in detecting small objects in remote sensing has been validated through experiments conducted on three publicly available datasets.The experimental results unequivocally demonstrate the model’s pronounced advantages in small object detection for remote sensing,surpassing the performance of current mainstream models.

    KEYWORDS Small object detection;attention mechanism;cross-layer fusion;discrete cosine transform

    1 Introduction

    With the advancement of aerospace remote sensing technology,the tasks such as classification,segmentation,detection,and tracking of remote sensing images have emerged as focal points in the modern field of image processing.In recent years,deep learning-based object detection algorithms have achieved notable breakthroughs in various domains [1].Despite the success of these object detection algorithms[2–5]in detecting medium and large objects in remote sensing object detection tasks,challenges persist when it comes to small objects characterized by a limited number of pixel values.Even the current more advanced object detection algorithms[6–9]cannot completely overcome this challenge.The definition of small objects varies in different scenarios,and there are fundamental approaches for classification:1)Based on relative scale,object instances whose relative area,i.e.,the ratio of the area of the bounding box to the area of the image,has a median between 0.08%and 0.58%for all object instances in the same category are defined as small objects;2)Based on absolute scale,i.e.,from the perspective of the absolute pixel size of the object,the most common definition comes from the MSCOCO dataset[10],a generalized dataset in the field of object detection,which defines a small object as an object with a resolution of less than 32 pixels by 32 pixels.

    Presently,there are two types of object detection models based on deep learning: single-stage detection models and two-stage detection models.Unfortunately,both detection models exhibit notable shortcomings in the performance of small object detection compared to other sizes of objects[11].Persistent challenges in small object detection include the following: Firstly,the risk of losing object information is pronounced.Common contemporary object detectors are typically comprised of a backbone and a detection head,with the latter making decisions based on the representation output by the former.While these models have proven effective and yielded unprecedented success,general feature extraction networks often employ down-sampling operations to filter out noise and reduce the spatial resolution of the feature maps.While this loss has minimal impact on the model’s performance for medium and large object detection,it proves fatal for small objects.The detection head struggles to provide accurate predictions in highly structured representations when the subtle features of small objects are nearly eliminated from the deep features.Secondly,noise adversely affects feature representation,and discriminative features are crucial for both classification and localization tasks.Small objects typically exhibit low resolution and poor feature quality,posing challenges in learning discriminatively from the distorted structures.Simultaneously,regional features of small objects are susceptible to background and other environmental factors,introducing additional noise into the deep learning process.Thirdly,small objects exhibit low tolerance for perturbation of bounding frames.In most object detection algorithms,localization—one of the primary tasks—is formulated as a regression problem.The localization branch is designed to output bounding box offsets,and the Intersection over Union(IoU)metric is commonly used for accuracy evaluation.However,even a slight deviation significantly lowers the IoU,exacerbating the learning difficulty of the regression branch.Numerous scholars have proposed solutions to these challenges in small object detection.For example,Bai et al.[12]proposed a feature-enhanced pyramid and shallow feature reconstruction network (FEPS-Net) based on feature enhancement to address the difficulty of detecting densely distributed small ships in synthetic aperture radar(SAR)ship images.FEPS-Net proposes a feature enhancement pyramid that includes a spatial enhancement module for enhancing spatial location information and suppressing background noise,and a feature alignment module for solving the problem of feature misalignment during feature fusion;in addition,for the problem of detecting small ships in SAR ship images,a shallow feature reconstruction module is designed to extract the semantic information of small ships.Huang et al.[13]proposed a cross-scale feature fusion pyramid network for the problem of fewer small object features in the deep network,in which the Cross-Scale Fusion Module(CSFM)is introduced to perform multi-scale fusion in order to ensure that there are enough small object features in the output features,and in this way to solve the problem of difficult detection of small objects in remote sensing images.While these existing methods have improved detection accuracy to some extent,they still face limitations:1)Most detection models targeting small objects adopt only the multi-scale fusion method,neglecting the cross-layer fusion strategy and the determination of the fusion factor;2) The Spatial Pyramid Pooling (SPP) module in many models does not incorporate smaller receptive fields and self-learning weighted ideas,hindering small object detection;3) The optimization of the frame regression loss calculation method is overlooked in most models,resulting in poor locating ability for small objects.

    To tackle the challenges associated with detecting small objects,we propose the Cross-Layer Fusion and Weighted Receptive Field-based YOLO [14] (CAW-YOLO) for small object detection in remote sensing.This approach utilizes the YOLOX-S object detection model as the baseline model.Firstly,Bi-Level Routing Attention (BRA) [15] is introduced into the backbone network to enhance feature extraction while effectively filtering out noise.Secondly,we propose MSCACBAM,an enhancement of the Convolutional Block Attention Module (CBAM) [16],addressing the limitations of CBAM.Thirdly,we present a weighted multi-receptive field atrous spatial pyramid pooling module,leveraging the concept of receptive field weighting to augment the detection capability for small objects.Fourthly,a cross-layer attention fusion module is constructed using a multi-scale fusion scheme,which fuses shallow and deep features,so that the information of small objects is retained to the maximum extent in the fused feature map.Finally,we address the low tolerance of small objects to bounding box perturbations by incorporating the joint Normalized Wasserstein Distance(NWD) [17] and Efficient Intersection over Union (EIoU) [18] losses as the model’s box regression loss function.The following summarizes the contributions of the paper:

    1.Utilizing YOLOX-S [19] as the baseline model,we propose CAW-YOLO for remote sensing small objects,which surpass current mainstream object detection models used for multiple small object datasets in remote sensing.

    2.A cross-layer attention fusion module is proposed which is to preserve the information of small objects in the fused feature map to the greatest extent and reduce the redundancy in the feature fusion process through the attention mechanism.It is worth mentioning that we propose an improved CBAM based on CBAM in this module.

    3.Based on a weighting mechanism,a multi-receptive field atrous spatial pyramid pooling module is proposed so that the model can dynamically adjust the importance of different sizes of receptive field feature maps according to different inputs.

    4.We adopt a joint box regression loss function to address the low tolerance of small objects to bounding box disturbances while maintaining a high convergence speed in the model.

    The remainder of the article is organized as follows:Section 2 reviews some related work.Section 3 describes the materials and methods of our work.Section 4 presents our experimental results and discusses their significance.Finally,Section 5 provides a concise summary of our findings and suggests directions for future research.

    2 Related Work

    2.1 Attention Mechanism

    In recent years,the attention mechanism has gained widespread use across various task models,such as language and vision,due to its exceptional performance.Attention systems operate by selectively weighting input variables to determine output variables.In visual models,attention systems suppress irrelevant information by assigning weights to features in the object area.SENet[20]obtains the weights corresponding to each channel through compression and assigns the corresponding weights to each channel;CBAM calculates the attention weights in both spatial and channel dimensions to better focus on the features of interest;CA(Coordinate Attention)[21]is essentially spatial attention by performing average pooling horizontally and vertically,and assigns different weight coefficients to different locations in space after embedding the location information in the channel attention.

    2.2 Feature Fusion

    Multi-scale feature fusion forms the foundational structure for addressing small object detection problems.Deeper Convolutional Neural Network(CNN)structures generate multi-layer feature maps with diverse spatial resolutions.Lower layers contain finer-grained and more localized features,while higher layers offer richer semantic information.The challenge arises because features of small objects may vanish in deeper layers due to down-sampling.Therefore,feature fusion integrates features at different levels or branches,enhancing the representation of small object features.SSD[22]first attempts to use pyramid features for object detection.Feature Pyramid Networks(FPN)[23]uses lateral connectivity and top-down feature fusion to construct a feature pyramid with stronger semantics.In addition,Path Aggregation Network(PAN)[24]adds bottom-up enhancement based on FPN to improve the information flow.EfficientDet[25]repeats the bidirectional path multiple times to facilitate higher-level feature fusion.In addition to manually designing the network,NAS-FPN[26]attempts to find stronger feature pyramid structures with the help of neural structure search.

    2.3 Current Mainstream Small Object Detection Model

    Presently,numerous mainstream small object detection models exist.Wang et al.[27] proposed M-CenterNet,which was an improvement on CenterNet.Instead of predicting a single central point,M-CenterNet predicts four central points,enhancing the model’s accuracy in pinpointing small objects and improving detection accuracy.Li et al.[28] proposed TridentNet,emphasizing the impact of receptive fields on small objects.TridentNet suggested that smaller receptive fields are more suitable for detecting these objects and employs a multi-scale detection framework to balance them with objects of other sizes.Most current object detection models use rectangular bounding boxes to represent objects across various recognition stages like anchors,proposals,and final predictions.While convenient,these boxes only offer initial object positioning,resulting in rough feature extraction.To address this,Yang et al.[29] proposed RepPoints.RepPoints uses a set of points for object localization,employing adaptive to position these points accurately.This approach limits the spatial scope of the object,highlighting semantically important areas and providing a more refined positioning method.Hong et al.[30]proposed SSPNet(Scale Selection Pyramid Network)to enhance the representation of small objects in FPN by employing a scale enhancement module and scale selection module.The former ensures that the model focuses on objects of different scales throughout each layer to prevent the small objects from being overwhelmed by the large number of backgrounds,while the latter facilitates feature sharing between adjacent layers in FPN,ensuring gradient consistency.Lu et al.[31]proposed AF-SSD(Attention and Feature-fusion SSD),which enhances the detection ability for small objects by fusing shallow and deep feature information through a multi-layer feature fusion structure.

    3 Methodology

    In this section,we elucidate the proposed method and delve into the implementation details of each module.

    3.1 General Structure of the Model

    We use YOLOX-S as the baseline model and propose an improved model,named CAW-YOLO,for small objects in remote sensing.The overall structure of the model is shown in Fig.1,illustrating the data flow with arrows.The whole model is comprised of the Backbone,Neck and Head.The Backbone plays the role of extracting image features.Considering that small object features are easily affected by background noise in the process of feature extraction,CAW-YOLO adds Bi-Level Routing Attention(BRA) to the last layer of Backbone,so that the model can filter background noise in the process of feature extraction.The BRA module is based on sparse sampling rather than down-sampling,it can retain the fine-grained detail information on the one hand,and achieve the purpose of saving the computation quantity on the other hand as well.The retention of fine-grained features in the network can improve the model’s ability to discriminate small objects.In order to enhance the model’s ability of multi-scale perception,especially for small-scale objects,we propose WMFASPP instead of SPP[32]in YOLOX-S,which enhances the model’s multi-scale perception ability through dynamic weighting.In the Neck,we adopt the cross-layer fusion method,fusing feature maps d2,d3 and d4 through the Cross-layer Attention Fusion Module.Finally,the detection results are processed in the Head.To be clear here,although BRA has a good ability to optimize the model,it also has a negative impact,that is,excessive use of BRA in the model will introduce too many parameters,affecting the detection speed of the model.For an object detection model for industrial applications,its real-time performance must be ensured,as a result,only one BRA is used in order to avoid introducing too many parameters.

    Figure 1:Overall architecture of CAW-YOLO

    3.2 Cross-Layer Attention Fusion Module

    In the task of object detection,the shallow features of the network have a smaller receptive field,and the extracted features are more similar to the input,including more fine-grained features such as color,texture,corners and edges,which are more conducive to the network’s localization of small objects.While the deep features of the network,after multi-layer convolution operations,contain more abstract semantic information,in which the detailed features of small objects are almost eliminated,which is the reason why many object detection models have poor performance for small object detection.Addressing this,the improved model performs cross-layer feature fusion in YOLOXS’s Neck.We propose a Cross-layer Attention Fusion Module(CAFM)to fuse shallow features with deep features,and use the attention mechanism to reduce the redundancy of feature fusion.

    3.2.1 MSCA-CBAM

    Before introducing the entire CAFM in detail,it is necessary to introduce the attention module used in CAFM:MSCA-CBAM in detail.The main function of MSCA-CBAM in CAFM is to filter useless features and reduce feature information redundancy during feature fusion.The structure of CBAM will be introduced below,and its defects will be analyzed.Finally,how MSCA-CBAM is improved from CBAM will be introduced.

    CBAM is comprised of a Channel Attention Module (CAM) and a Spatial Attention Module (SAM),which are cascaded together.The CAM utilizes max-pooling and average-pooling to respectively compress the input featuresF∈RC×H×W,aiming to derive channel information from the input features.The resulting two generated channel information features∈RC×1×1undergo processing through a multilayer perceptual machine—a fully connected three-layer structure.Finally,the obtained features are aggregated and subjected to a sigmoid function to obtain the channel attention weightsMc∈RC×1×1,which are computed as Eq.(1):

    Here,σrepresents the sigmoid activation function,andW0andW1represents the weights in the fully connected three-layer network.

    The feature mappingF1∈RC×H×Wincorporating the channel attention is derived by producing the channel attention weightsMcwith the input featuresF.Subsequently,SAM processesF1as input,performing maximum pooling and average pooling on the channel dimensions to obtain two spatial information mappings∈R1×H×W,which are then concatenated.Then the spatial attention weightsMs∈R1×H×Ware obtained after a 7×7 convolution operation and sigmoid function processing,and the computation is as Eq.(2):

    Here,σrepresents the sigmoid activation function,andf7×7is a 7×7 convolution operation.The CBAM-processed feature map is obtained by performing the inner product operation onMsandF1.The CBAM-processed feature map is then used as the basis for the feature map.

    While CBAM proves efficient as an attention module,it has still some drawbacks.Firstly,the use of global average-pooling in the CAM for feature compression can lead to some amount of missing feature information.Secondly,it demands more computing resources,increasing computational complexity.The work[33]proposed the Multi-Spectral Channel Attention Module(MSCAM),which proves that the global average-pooling used by CAM in calculating channel attention is essentially the zero-frequency pooling of the Discrete Cosine Transform(DCT)from the frequency perspective.However,this obviously ignores other useful frequency components.Although maximum pooling is also utilized in CAM to supplement the frequency information,this is still not an optimal solution.Based on the work[33],the CBAM is improved by utilizing 2D DCT instead of global average-pooling in CAM,which introduces frequency information into the attention process,includes more useful frequency information during feature compression.This avoids missing feature information attributed to global average pooling,allowing the network to better filter noise during feature extraction.The improved CBAM is referred to as MSCA-CBAM,and its structure is shown in Fig.2.

    Figure 2:Structure of MSCA-CBAM

    Instead of global average pooling,2D DCT is utilized as follows:the feature map is divided into multiple parts according to the number of channels,and each undergoes 2D DCT calculation using different frequency components,including the zero-frequency component (i.e.,global average pooling),thereby compressing more frequency information.As shown in Fig.2,F0,F1,...,Fn-1denotes the partitions into which the input featureF∈RC×H×Wis divided,whereFi∈,i∈{0,1,...,n-1},,andCshould be divided byn.For each partition,the corresponding 2D DCT frequency component is assigned,and the 2D DCT transform is performed.Finally,the compressed components are concatenated to obtain the 2D DCT compressed feature mapF2DDCT∈RC×1×1.The calculation process is as follows.In addition,the 16 frequency components with the best performance based on the heuristic two-step criterion from the work[28]are used our works.

    Here,Eq.(4) is the basis function of 2D DCT.In Eq.(3),uiandviare the preset 2D frequency component weights corresponding toFi,andFreqi∈denotes the results of 2D DCT compression for each channel partition.

    3.2.2 The Structure of the Cross-Layer Attention Fusion Module

    The implementation process of the cross-layer attention fusion module will be described in detail below,including the selection of required fusion features and the selection of fusion methods.

    As shown in Fig.3,the output features of layers d2,d3,and d4 in the Backbone are utilized for cross-layer feature fusion.This involves combining shallow features with deep features using the three output features p1,p2,and p3 of the PAN in the Neck.In contrast to the conventional cross-layer fusion,this approach directly fuses Backbone features with PAN output features,avoiding additional processing by the FPN and the PAN.Compared with the latter,by directly fusing with the output features of PAN,the number of parameters of the model can be minimally increased while fusing the deep and shallow features to avoid overfitting the network;secondly,in the backbone part of the network,since the features are only less processed,there are more fine-grained features of the retained small objects,which will inevitably result in the loss of the fine-grained features of the small objects if they are fused in the FPN and the PAN and undergo too much processing.

    Figure 3:Cross-layer fusion network structure

    The structure of CAFM is shown in Fig.4a.Here,X∈RC×H×Wrepresents shallow features output from backbone,andY∈represents deep features output from the PAN.Inspired by the prior work[34],down-sampling utilizes the SPD(Space-to-depth)layer,as shown in Fig.4b.The SPD layer slices the feature mapXby the separated feature points to generate four sub-feature maps,where the size of each sub-feature map is,and finally performs a stitching operation to generate the down-sampled feature mapX′∈.In contrast to direct utilize of down-sampling methods such as pooling,the SPD layer converts the fine-grained features of the space into channel depth features and does not directly erase the fine-grained features.After the SPD layer,X′is processed utilizing a non-spanning(step size of 1 and kernel size of 1)dynamic convolution[35].The structure of dynamic convolution is shown in Fig.4c.It firstly performs attention computation for the input feature maps,generates n attention weights,and then performs linear summation on the n convolution kernel parameters.Finally,the summated convolution kernel is used as the convolution kernel of dynamic convolution for the operation.Compared to traditional convolution,dynamic convolution can change with the input.After using the SPD layer,dynamic convolution can more effectively extract small object features with fewer operations.In addition,utilizing non-stride methods is more helpful for the network to retain the fine-grained information of small objects,as opposed to s strides greater than 1,which results in an undifferentiated loss of feature information.Finally,CAFM utilize the MSCA-CBAM to compute the attention of the two feature maps,reducing the redundancy of feature information during the feature fusion process.

    Figure 4:Structure of CAFM.(a)Overall structure of CAFM;(b)SPD layer;(c)Overall structure of dynamic convolution

    3.3 Weighted Multi-Receptive Field WMFASPP

    Spatial Pyramid Pooling(SPP)is utilized in the Backbone of YOLOX-S,which aims at extracting spatial feature information at different scales of the receptive field,and improving the robustness of the model to spatial layout and object variability.Chen et al.[36]proposed Atrous Spatial Pyramid Pooling(ASPP).ASPP fundamentally differs from SPP by eschewing pooling operations and employing a series of dilation convolutions with varying dilation rates to create a pooling pyramid.This approach offers distinct advantages:compared to pooling operations,downsampling with dilated convolutions acquires feature maps with diverse receptive fields while preserving fine-grained image features crucial for small object detection.Based on the above theories,we propose the Weighted Multi-receptive Field Atrous Spatial Pyramid Pooling(WMFASPP),Its structure is shown in Fig.4.

    WMFASPP constructs feature maps with different receptive fields through ordinary convolution and dilated convolution.The dilation rates of dilated convolutions are set at 2,3,4,and 5,corresponding to receptive field sizes of 1 × 1,3 × 3,7 × 7,9 × 9,and 11 × 11,respectively.The smaller receptive fields ensure the network captures features of small objects,while the multi-scale receptive fields empower the network to gather information across various scales.Leveraging dilated convolution ensures the retention of fine-grained features during downsampling,preserving detailed image information and enhancing small-scale object detection.To adaptively learn the importance of different receptive field feature maps,a weighted fusion concept is employed.Different weights are multiplied with distinct receptive field feature maps before the splicing operation.Moreover,regarding the generation of weights,instead of taking the traditional way of generating compressed feature maps using global average pooling,the 2D DCT mentioned in the previous section is utilized.As shown in Fig.5,the input feature map of WMFASPP undergoes 2D DCT compression,followed by feature extraction via a fully connected two-layer structure.Finally,the Sigmoid function activate to yield the 6 feature map weightsW∈R6×1×1.These six weights are multiplied with the feature maps of different receptive fields before the subsequent step.This adaptive weighting enhances the model’s multi-scale detection capabilities,assigning greater importance to small receptive field feature maps when increased attention to small objects is warranted,and vice versa for larger objects.The selection of frequency components in the 2D DCT transformation aligns with the approach detailed in the previous section.

    Figure 5:Overall structure of WMFASPP

    3.4 Joint Box Regression Loss of NWD and EIoU

    To mitigate the susceptibility of small objects to bounding box perturbations,Wang et al.[21]proposed a novel metric utilizing the Wasserstein distance to assess small objects,which better reflects the distance between distributions even if there is no overlap between them.Compared with IoU,this metric has better properties in measuring the similarity between small objects.The Normalized Wasserstein Distance is a distance metric that measures the distance between probability distributions using the Wasserstein distance derived from optimal transportation theory.When comparing two 2D Gaussian distributions represented byμ1=N(m1,Σ1)andμ2=N(m2,Σ2),the distance between them can be quantified using the second-order Wasserstein distance,as shown in Eq.(6).

    The equation can be simplified as:

    Here,‖·‖F(xiàn)represents the Frobenius norm.Furthermore,since modeling is performed for two bounding boxes A=(cxa,cya,wa,ha)and B=(cxb,cyb,wb,hb),the above equation can be simplified to:

    Normalizing it results in NWD:

    Here,C is the usual size of the dataset instance.

    Recognizing that relying solely on NWD as a distance metric for object and prediction boxes might impede model convergence,both EIoU and NWD are employed as the distance metrics in this paper.Their collective loss is the joint loss of EIoU and NWD,computed as:

    Here,EIoU is calculated as:

    Here,crepresents the diagonal length of the minimum outer rectangle of the prediction box and the real box.CwandChdenote the width and length of the minimum outer rectangle,andbandbgtdenote the centers of the prediction box and the real box,respectively.ρdenotes the Euclidean distance,w,wgt,handhgtrepresent the width and length of the prediction box and the real box,respectively.

    4 Experimental Details

    4.1 Datasets

    Currently,there are many datasets used in the field of remote sensing object detection.However,in most of datasets,the number of large objects is much larger than that of small objects;at the same time,some small object annotations have poor quality and may contain errors.In recent years,numerous datasets proposed for remote sensing small objects,such as DOTA [37],NWPUVHR-10 [38],UCAS_AOD [39],DIOR [40],RSOD [41],and AI-TOD [27].Considering the number of categories and the richness of small objects in the dataset,we choose to utilize AI-TOD,DIOR,and RSOD as the experimental dataset.

    DIOR is a large-scale dataset for optical remote sensing detection containing such as high-speed service areas,high-speed toll booths,airplanes,airports,baseball stadiums,basketball courts,bridges,chimneys,dams,golf courses,athletic stadiums,harbors,overpasses,boats,stadiums,storage tanks,tennis courts,train stations,vehicles,and windmills in total 20 categories.The dataset has a total of 23463 images and 192472 instances,out of which 5862 will be used as a train set,5863 as a validation set,and 11738 as a test set.RSOD is an open remote sensing dataset including 4 categories of airplanes,fuel tanks,playgrounds,and overpasses.The dataset has a total of 976 images and 6950 instances,which are randomly divided into a train set and a test set according to the ratio of 8:2.AI-TOD is a newer remote sensing object detection dataset,which mainly focuses on the small objects,including airplane bridge storage-tank ship swimming-pool vehicle person AI-TOD is a newer remote sensing object detection dataset,which mainly focuses on small objects,including airplanes,bridges,storagetanks,ships,swimming-pools,vehicles,persons,and wind-mills,with 8 categories and a total of 28,036 images,of which 14,018 are used as a train set,and 14,018 as a test set (such as PASCAL VOC(156.6 pixels),MS COCO(99.5 pixels)or DOTA(55.3 pixels)).The details of the three datasets are summarized in Table 1.

    Table 1:Dataset overview

    In order to prove the reliability of the dataset used for the experiment,the aspect ratio of the object boxes in the dataset was counted as shown in Fig.6.There are a large number of small objects in all three data mentioned in the experiment,which meets the experimental requirements of remote sensing small objects.

    Figure 6:Aspect ratio of the object box in the datasets.(a)aspect ratio of the object box in AI-TOD;(b)aspect ratio of the object box in DIOR;(c)aspect ratio of the object box in RSOD

    4.2 Experimental Details

    In our experiment,the hardware configuration for model training and performance evaluation is Intel Core i7-12700KF(3.60 GHz),with 32 GB of RAM,GPU model NVIDIA RTX A4000,and 16 GB of video memory;the software environment is Windows 10,Python 3.8,PyTorch 1.9,and CUDA framework for parallel acceleration.Parallel accelerated computing,CUDA version 11.0.In order to have a fair comparison of the model performance,the Adam optimizer is used uniformly in the experiments,the momentum size is set to 0.937,the learning rate adopts the cosine annealing algorithm,the initial learning rate is 1e-3,the minimum learning rate is 0.01 times of the initial learning rate,the batch size and the epoch is set to 16 and 300,respectively.

    4.3 Evaluation Metric

    In the object detection task,the commonly used evaluation metrics are Average Precision(AP),detection speed FPS (Frames Per Second) and the amount of model parameters (Params).The AP is a comprehensive metric in the task of object detection.For the detection of multi-category objects,the AP usually denotes mean average precision(mAP)which is obtained by the average of different category APs.But,in order to specifically show the detection effect of the proposed model for small objects.we adopts the same evaluation metrics as in the work[27]in the comparison experiments for the AI-TOD dataset.This is a more stringent COCO metric,which includes AP,AP0.5,AP0.75,APvt,APt,APs,APmand APl.AP denotes the AP with IoU=0.50:0.05:0.95(the average of the values on the IOU thresholds),and AP0.5denotes the AP with IOU=0.50(equally to mAP in PASCAL VOC metric,is the metric that can best represent the comprehensive performance of the detection model).AP0.75denotes APs with IOU=0.75(a very strict metric).APsdenotes AP for small objects with areas less than 322and greater than 162.APmdenotes AP for medium-sized objects with areas between 322and 962.APldenotes AP for large objects with areas greater than 962.In addition,APvtand APtare very tiny and tiny objects,corresponding to areas ranging from greater than 22less than 82and greater than 82less than 162,respectively.

    4.4 Experimental Results and Analysis

    In this section,the effectiveness of the proposed method is evaluated and analyzed,utilizing YOLOX-S as the baseline network.It is worth mentioning that all three datasets were utilized in a set of comparative experiments in this chapter,and the AI-TOD dataset was utilized in the subsequent improved module comparisons

    4.4.1 Comparative Experiments

    In order to prove the superiority of the proposed model over the current mainstream object detection models and its generalizability among different datasets,comparisons are made with multiple models on the AI-TOD dataset,the RSOD dataset and the DIOR dataset,respectively,and the comparison results are shown in Tables 2 and 3.

    The results of the comparison experiments for the AI-TOD dataset are shown in Table 2.Even the current state-of-the-art detection model’s APvttends to be close to 0 (which indicates that it is difficult to detect tiny objects with area sizes of 22~82),while the APvtvalue of the proposed model in this paper reaches 6.3%.Compared with YOLOv8n,YOLOv6-M,M-CenterNet,YOLOX-S,DETR,and Faster R-CNN,the APvtvalues of the model in this paper are 4.3%,3.7%,0.2%,3.4%,6.3%,and 6.3% higher,respectively.In addition,the AP0.5and APtvalues of the proposed model are the highest,which indicates that the proposed model has the best detection performance for small objects compared to the other models.It is worth mentioning that many metrics of the model in this paper have reached the highest in Table 2,but the values of AP and APmare not the highest.Because,for objects of different scales,it is a problem to keep the AP of small objects rising while the AP of large objects is also rising.And,the model proposed in this paper,on the basis of YOLOX-S,strengthens the attention to small objects,it is inevitable that the model’s attention to medium and large objects will decline.This is also the reason why the APmof the model proposed in this paper is smaller than that of YOLOX-S,and also causes the AP metric of the model proposed in this paper to be slightly inferior to that of YOLOX-S.However,from a global perspective,the APvt,APtand APsmetrics of the model proposed in this paper have been greatly improved compared with YOLOX-S,while the AP metric has only decreased by 0.7%.It is very cost-effective to reduce the AP metric or exchange for a large improvement in the accuracy of small objects.In the comparison experiments on the RSOD and DIOR datasets,the results are shown in Table 3,the proposed model outperforms the current mainstream object detection models in the comparison of each metric.The APsreach 13.9% and AP0.5reaches 73.98% on the DIOR dataset,while the APsreach 44.8% and AP0.5reaches 97.50% on the RSOD dataset.

    Table 2:Evaluation metrics of different methods based on AI-TOD dataset

    Table 3:Evaluation metrics of different methods based on DIOR dataset and RSOD dataset

    In order to explore the difference between the proposed detection model in terms of parameter amount and detection speed with respect to the current mainstream object detection model,parameter amount and detection speed are compared between the current mainstream detection models and the proposed model,and the comparison results are shown in Table 3.The parameter amount of the detection model proposed in this paper is only 13.22 M,and the detection speed can reach 90.5 frame·s-1.The parameter amount is only 4.29 M higher than the baseline model YOLOX-S,and the detection speed is up to 90.5 frame·s-1to meet the requirement of real-time detection.In addition,compared with the other models in the table,the parameter amount of the proposed model is only larger than that of YOLOX-S,YOLOv8-S and YOLOv5-S,and the number of parameters of the rest of the models is larger than that of the model in this paper.Finally,it should be noted that with the increase of modules,the detection speed and parameter number of the model will be affected to a certain extent,but at the cost of a small number of parameters and detection speed,improving the detection accuracy of small objects is cost-effective,which is not inferior to the current mainstream detectors(YOLOv8,etc.).

    In order to directly reflect the detection ability of the improved model on remote sensing small objects,the detection effect is compared using the BASELINE,the proposed model,YOLOv8-S and YOLOv7,and the detection effect is shown in Fig.7.Compared with the other three current mainstream models in the detection of remote sensing small objects in the process of a large number of leakage detection,the improved model has greatly improved this situation.This proves that the model structure proposed in this paper plays an important role in the detection process of small objects.

    4.4.2 Ablation Experiment

    In order to verify the effectiveness of each improvement method proposed in this paper,a set of ablation experiments were designed on the AI-TOD dataset,and the results of the experiments are shown in Table 4.A“√”indicates that the method was added.All the methods proposed in this paper improve the model AP0.5by a total of 1.2%,APtby 2.1% and APvtby 3.4%.It illustrates that the model’s detection ability for remote sensing objects,especially small objects,can be significantly improved when all the improved methods are utilized.In addition,after utilizing all the methods,the number of parameters only increases by 4.29 M compared with the original model,and the FPS only decreases by 27.4 frame·s-1,which only uses less overhead in exchange for a larger improvement.

    Table 4:Evaluation metrics of ablation experiment

    From the overall point of view,the improved model in this paper has some advantages in improving both AP0.5and APs.At the same time,due to the addition of each improvement method,the complexity of the model will inevitably be enhanced,and it will be slightly inferior to the original model in terms of the number of parameters and detection speed.However,the small increase in the number of parameters and detection speed does not affect the lightweight and real-time performance of the model,and it is very cost-effective to exchange a small amount of overhead for the improvement of detection accuracy.

    Figure 7:The detection effect is compared with the BASELINE,the proposed model,YOLOv8-S,and YOLOv7.(a)The detection effect of the proposed model;(b)The detection effect of YOLOv8-S;(c)The detection effect of YOLOv7;(d)The detection effect of baseline

    4.4.3 Experimental Analysis of MSCA-CBAM

    To further validate the effectiveness of MSCA-CBAM,a set of side-by-side comparison experiments of attention modules are done in this paper,replacing MSCA-CBAM with CBAM,CA,SK and MSCAM attention modules respectively on the basis of the model proposed in this paper on the AI-TOD dataset,and the experimental results are shown in Table 5.As can be seen from Table 5,the different attention modules all bring different changes in performance,while the model utilizing MSCA-CBAM achieves the highest in terms of detection accuracy.In terms of the number of model parameters,the addition of MSCA-CBAM only results in a model parameter of 13.22 M,which is only 0.01 M higher than that of CA and lower than that of CBAM,which is 14.77 M.This is due to the fact that pre-determined frequency components are utilized in the 2D DCT transforms,so no additional parameter number is introduced when using multi-band components for feature compression.

    Table 5:Comparison of the effects of different attention models

    In order to increase the interpretability of the performance of each attention model,the effect of each attention model is visualized and analyzed on a small object remote sensing image using a heat map,and the visualization results are shown in Fig.8.Among them,MSCA-CBAM achieves the best both in terms of heat map response strength and response tightness.

    4.4.4 Experimental Analysis of WMFASPPF

    In order to explore the superiority of WMFASPP,a set of comparative experiments were conducted utilizing the models proposed in this paper with SPP,ASPP and WMFASPP,respectively,and the results are shown in Table 6.Among them,the model with WMFASPP achieves the best detection accuracy,with an improvement of 2.5% and 2.3% in AP0.5,2.8% and 3.1% in APt,and 2.8% and 3.2% in APvt,relative to the use of SPP and the use of ASPP,respectively.Therefore,by constructing additional small receptive fields and adaptive weighting in the spatial pyramid pooling module,it can strengthen the model’s multi-scale perception ability and adaptive scale reinforcement ability,which can effectively improve the model’s accuracy for small object detection.In addition,WMFASPP has a smaller number of parameters than ASPP,for several reasons: Firstly,the 2D DCT utilized in WMFASPP does not add additional parameters.Secondly,more 1×1 convolution is utilized in ASPP for feature extraction,which is the main reason for the large number of ASPP parameters.Thirdly,WMFASPP initially reduces the channel by more than ASPP,which results in a lower number of convolution module parameters in the middle.

    Table 6:Comparison of the effects of different spatial pyramid pooling models

    4.4.5 Experimental Analysis Box Regression Loss Function

    In order to investigate the effectiveness of the joint loss of EIoU and NWD,a set of cross-sectional comparison experiments of the box regression loss function are designed in this paper,utilizing the IoU loss,EIoU loss,NWD loss and the joint box regression loss in the model proposed in this paper,respectively,and the experimental results are shown in Table 7.Compared to the model utilizing IoU loss alone as the box regression loss,higher accuracy is obtained by utilizing the joint box regression loss(0.7% improvement in AP0.5,0.6% improvement in APt,and 0.3 improvement in APvt).

    Table 7:Comparison of the effects of different box regression loss function

    Figure 8:Heat map visualization results.(a)Original image;(b)Heat map of MSCA-CBAM;(c)Heat map of CBAM;(d)Heat map of CA;(e)Heat map result of SK;(f)Heat map of MSCAM

    4.4.6 Comparison of Cross-Layer Fusion Feature Maps

    In order to demonstrate the effectiveness of CAFM proposed in this paper,feature maps will be visualized utilizing the output from the bottom layer of the cross-layer fusion model and without utilizing the cross-layer fusion model,and the results of the visualization are shown in Fig.9(Fig.8a is the input image).Compared with the model without cross-layer fusion method,the object features in the detected images can be clearly observed in the output feature map of the model utilizing the CAFM proposed in this paper,and the feature response is stronger in comparison.

    5 Conclusion

    Aiming at the problem that it is difficult to detect small objects in remote sensing images,this paper proposes CAW-YOLO for small objects in remote sensing by utilizing the YOLOX-S object detection model as the baseline model.It firstly incorporates BRA into the feature extraction stage of the model so that the model can filter out the background noise interfering with the detection of small objects in the feature extraction stage.A weighted multi-receptive field atrous spatial pyramid pooling module is then constructed by increasing the feature maps of small receptive field and adaptive weights to enhance the multi-scale,especially small-scale,perception capability of the module.Furthermore,a cross-layer attention fusion module is proposed,so that the model can maximize the retention of small object features,and an improved CBAM attention module is also proposed in which the CBAM attention module is improved by a 2D DCT transform,which introduces frequency information into the attention so that it can optimize the performance of CBAM attention;finally,the joint box regression loss of NWD and EIoU is utilized to cope with the sensitivity of the small object to the slight offset of the object frame.The effectiveness and generalizability of the improved model are demonstrated on three remote sensing datasets,AI-TOD,DIOR and RSOD,and on the AI-TOD dataset,which is the most difficult to detect,the APvtreached 6.3%,a 3.4% improvement over the baseline.However,due to the inclusion of the improved module,it is bound to increase a certain number of parameters and detection speed,and the future model needs to be researched in the direction of high accuracy and lightweight.

    Acknowledgement:The authors wish to express their appreciation to the reviewers for their helpful suggestions which greatly improved the presentation of this paper.We are also very grateful to the editors for their patient and friendly reminders,as well as for their efforts in editing the manuscript.

    Funding Statement:This work was supported in part by the National Natural Science Foundation of China under Grant 62006071,and part by the Science and Technology Research Project of Henan Province under Grant 232103810086.

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design:Weiya Shi;experimental data collection,analysis and interpretation of results,and draft manuscript preparation: Shaowen Zhang;review of paper content: Shiqiang Zhang.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:Publicly available datasets were analyzed in this study.AI-TOD data set download address is:https://github.com/jwwangchn/AI-TOD;The current address of the ROSD dataset is: https://github.com/RSIA-LIESMARS-WHU/RSOD-Dataset-;DIOR download address is:http://www.escience.cn/people/gongcheng/DIOR.html.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产黄色小视频在线观看| 2021天堂中文幕一二区在线观| 久久久精品大字幕| 蜜桃久久精品国产亚洲av| 亚洲美女搞黄在线观看| 乱码一卡2卡4卡精品| 天堂网av新在线| 国产成人午夜福利电影在线观看| 亚洲无线观看免费| 波多野结衣巨乳人妻| 免费一级毛片在线播放高清视频| 亚洲第一区二区三区不卡| 久久久久精品久久久久真实原创| 狠狠狠狠99中文字幕| 久久精品国产亚洲网站| 亚洲av中文字字幕乱码综合| 成人毛片a级毛片在线播放| 免费av毛片视频| 久久精品久久精品一区二区三区| 极品教师在线视频| 国产精品99久久久久久久久| 国产亚洲5aaaaa淫片| 国产一区二区亚洲精品在线观看| 秋霞在线观看毛片| 日韩在线高清观看一区二区三区| 在线观看av片永久免费下载| av国产免费在线观看| 久久国产乱子免费精品| 午夜免费激情av| 亚洲欧美清纯卡通| 欧美丝袜亚洲另类| 精品99又大又爽又粗少妇毛片| 亚洲在线自拍视频| 国产探花在线观看一区二区| 一个人看视频在线观看www免费| www.av在线官网国产| 国产高清不卡午夜福利| 久久国产乱子免费精品| 床上黄色一级片| 69av精品久久久久久| 精品无人区乱码1区二区| 中文字幕制服av| 成人美女网站在线观看视频| 91狼人影院| 在线a可以看的网站| 永久免费av网站大全| 亚洲欧美日韩无卡精品| 日日撸夜夜添| 乱人视频在线观看| 日韩av不卡免费在线播放| 黄片wwwwww| 国产高清视频在线观看网站| 天美传媒精品一区二区| 国产亚洲av嫩草精品影院| 麻豆精品久久久久久蜜桃| 国产在线一区二区三区精 | 欧美日韩国产亚洲二区| 人妻夜夜爽99麻豆av| 亚洲成人中文字幕在线播放| 亚洲伊人久久精品综合 | 国产精品久久视频播放| 国产免费视频播放在线视频 | 久久久久久大精品| 亚洲av熟女| 亚洲av免费高清在线观看| 免费在线观看成人毛片| 99热这里只有精品一区| 午夜精品一区二区三区免费看| 美女被艹到高潮喷水动态| 国产探花极品一区二区| 免费av不卡在线播放| av天堂中文字幕网| 久久99精品国语久久久| 男女国产视频网站| 久久草成人影院| 两性午夜刺激爽爽歪歪视频在线观看| 亚洲av成人精品一二三区| av国产久精品久网站免费入址| 99久久人妻综合| 黄色日韩在线| 免费电影在线观看免费观看| 亚洲欧美中文字幕日韩二区| 精品99又大又爽又粗少妇毛片| 久久久久久久久久黄片| 人妻制服诱惑在线中文字幕| 免费看光身美女| 免费无遮挡裸体视频| 国产免费一级a男人的天堂| 日产精品乱码卡一卡2卡三| 欧美日韩在线观看h| 最近最新中文字幕免费大全7| 成年女人永久免费观看视频| 91在线精品国自产拍蜜月| 亚洲av免费高清在线观看| 欧美丝袜亚洲另类| 又爽又黄a免费视频| 国产午夜精品一二区理论片| 国产黄片美女视频| 日韩一本色道免费dvd| av免费观看日本| 九色成人免费人妻av| 久久精品夜色国产| 网址你懂的国产日韩在线| 亚洲国产精品专区欧美| 最近手机中文字幕大全| 久久亚洲国产成人精品v| 亚洲欧美精品专区久久| 亚洲av熟女| av黄色大香蕉| 中文字幕av成人在线电影| 99九九线精品视频在线观看视频| 高清在线视频一区二区三区 | 97在线视频观看| 一级黄片播放器| av在线亚洲专区| 日韩中字成人| 在现免费观看毛片| 高清视频免费观看一区二区 | 丝袜美腿在线中文| 国产探花在线观看一区二区| 精华霜和精华液先用哪个| 六月丁香七月| 人人妻人人看人人澡| 久久国产乱子免费精品| 男人和女人高潮做爰伦理| 欧美变态另类bdsm刘玥| 欧美日韩在线观看h| av卡一久久| 一级爰片在线观看| 男女啪啪激烈高潮av片| 欧美zozozo另类| 国产亚洲av片在线观看秒播厂 | 国产午夜精品久久久久久一区二区三区| 一个人免费在线观看电影| 精品久久久久久成人av| 国内少妇人妻偷人精品xxx网站| 国产精品爽爽va在线观看网站| 久久草成人影院| 色网站视频免费| 亚洲av福利一区| 非洲黑人性xxxx精品又粗又长| 天天一区二区日本电影三级| 亚洲精品色激情综合| 亚洲人与动物交配视频| 少妇熟女aⅴ在线视频| 九九久久精品国产亚洲av麻豆| 一本久久精品| 国产又色又爽无遮挡免| 在线观看av片永久免费下载| 99在线人妻在线中文字幕| 美女内射精品一级片tv| 麻豆成人午夜福利视频| 中文字幕人妻熟人妻熟丝袜美| 欧美激情久久久久久爽电影| 中文字幕av成人在线电影| 美女黄网站色视频| 欧美性猛交╳xxx乱大交人| 99久久精品国产国产毛片| 国产精品伦人一区二区| 日韩在线高清观看一区二区三区| 国产精品国产高清国产av| 久久人人爽人人片av| 又爽又黄无遮挡网站| av黄色大香蕉| 少妇熟女aⅴ在线视频| 在线观看66精品国产| 亚洲精品aⅴ在线观看| 九九久久精品国产亚洲av麻豆| 一级毛片久久久久久久久女| 男人和女人高潮做爰伦理| 高清在线视频一区二区三区 | 级片在线观看| 日韩成人伦理影院| 国产精品综合久久久久久久免费| 99久久中文字幕三级久久日本| 免费观看性生交大片5| 久久99热6这里只有精品| 看黄色毛片网站| 美女脱内裤让男人舔精品视频| 青春草国产在线视频| 在线播放国产精品三级| 国产精品.久久久| 成人毛片60女人毛片免费| 色噜噜av男人的天堂激情| 亚州av有码| 一级毛片久久久久久久久女| 久久精品夜色国产| 黄色欧美视频在线观看| 亚洲成人av在线免费| 精品人妻视频免费看| 97在线视频观看| 老司机影院成人| 最近最新中文字幕大全电影3| 女人久久www免费人成看片 | 国产亚洲av嫩草精品影院| 亚洲精品一区蜜桃| 99久久九九国产精品国产免费| 国产一区二区在线观看日韩| 热99re8久久精品国产| 99久久九九国产精品国产免费| 亚洲av成人av| 亚洲美女搞黄在线观看| 午夜日本视频在线| 亚洲国产精品成人综合色| 久久精品久久久久久久性| 中文资源天堂在线| 亚洲av二区三区四区| 中文字幕亚洲精品专区| 看片在线看免费视频| 日韩av不卡免费在线播放| 激情 狠狠 欧美| 热99re8久久精品国产| 欧美日本视频| 自拍偷自拍亚洲精品老妇| 日本猛色少妇xxxxx猛交久久| 国产黄色视频一区二区在线观看 | 一边亲一边摸免费视频| 国产精品熟女久久久久浪| 国产成人精品一,二区| 身体一侧抽搐| av又黄又爽大尺度在线免费看 | av线在线观看网站| .国产精品久久| 搡女人真爽免费视频火全软件| 嫩草影院入口| 亚洲精品乱码久久久v下载方式| 国产毛片a区久久久久| 精品久久久久久久久亚洲| 国产精品1区2区在线观看.| 热99在线观看视频| 国产精品久久视频播放| 精品人妻一区二区三区麻豆| 精品国产露脸久久av麻豆 | 乱人视频在线观看| 亚洲国产成人一精品久久久| 久久久久久九九精品二区国产| 狠狠狠狠99中文字幕| 欧美潮喷喷水| 在线免费观看的www视频| 成人二区视频| 汤姆久久久久久久影院中文字幕 | 中文字幕精品亚洲无线码一区| 免费无遮挡裸体视频| 日韩强制内射视频| 可以在线观看毛片的网站| 久久婷婷人人爽人人干人人爱| 国产乱来视频区| 高清午夜精品一区二区三区| 一个人观看的视频www高清免费观看| 中文字幕av在线有码专区| 久久99热这里只有精品18| 精品无人区乱码1区二区| 你懂的网址亚洲精品在线观看 | 国产午夜福利久久久久久| 少妇裸体淫交视频免费看高清| 国产色婷婷99| 亚洲av男天堂| 蜜桃久久精品国产亚洲av| 欧美又色又爽又黄视频| 国产片特级美女逼逼视频| 亚洲成色77777| 中国国产av一级| 国产 一区 欧美 日韩| 亚洲av中文av极速乱| 国产精华一区二区三区| 国产老妇女一区| 日本爱情动作片www.在线观看| 国产探花在线观看一区二区| 九九久久精品国产亚洲av麻豆| 国产精品国产三级国产专区5o | 中文字幕精品亚洲无线码一区| 亚洲中文字幕日韩| 最后的刺客免费高清国语| 国产精品99久久久久久久久| 亚洲精品色激情综合| 免费黄色在线免费观看| 国产一区有黄有色的免费视频 | av播播在线观看一区| 在线观看美女被高潮喷水网站| 午夜精品一区二区三区免费看| 在线观看一区二区三区| 久热久热在线精品观看| 亚洲乱码一区二区免费版| 国产人妻一区二区三区在| 插逼视频在线观看| 亚洲高清免费不卡视频| 国产精品国产高清国产av| 国产精品精品国产色婷婷| 成人一区二区视频在线观看| 爱豆传媒免费全集在线观看| 级片在线观看| 婷婷六月久久综合丁香| 亚洲精品乱码久久久v下载方式| 国产一区二区三区av在线| 国产黄色视频一区二区在线观看 | 国产精品熟女久久久久浪| 成年版毛片免费区| 卡戴珊不雅视频在线播放| 国产精品人妻久久久久久| 99在线人妻在线中文字幕| 亚洲天堂国产精品一区在线| 国产成人91sexporn| 亚洲欧美精品综合久久99| 日日啪夜夜撸| 最近2019中文字幕mv第一页| 午夜福利网站1000一区二区三区| 一本一本综合久久| 小说图片视频综合网站| 国产精品久久视频播放| 在线观看66精品国产| 日日摸夜夜添夜夜爱| 亚洲,欧美,日韩| 久久精品综合一区二区三区| 亚洲在线观看片| 久久久久久久久大av| 内地一区二区视频在线| 少妇丰满av| 又爽又黄无遮挡网站| 欧美又色又爽又黄视频| 麻豆乱淫一区二区| 欧美日本亚洲视频在线播放| 成人鲁丝片一二三区免费| 久久精品人妻少妇| 少妇人妻精品综合一区二区| 亚洲av中文av极速乱| 水蜜桃什么品种好| 国产精品综合久久久久久久免费| 久99久视频精品免费| 亚洲第一区二区三区不卡| 成人无遮挡网站| 国产精品1区2区在线观看.| 欧美激情在线99| 国产午夜精品一二区理论片| 久久久精品94久久精品| 人人妻人人看人人澡| 亚洲国产欧美人成| 18+在线观看网站| 久久久精品94久久精品| kizo精华| 免费观看人在逋| 亚洲国产日韩欧美精品在线观看| 噜噜噜噜噜久久久久久91| 日韩一区二区视频免费看| 亚洲国产欧洲综合997久久,| 可以在线观看毛片的网站| 国产亚洲91精品色在线| 成人欧美大片| 欧美一区二区国产精品久久精品| 亚洲av中文字字幕乱码综合| or卡值多少钱| 麻豆成人av视频| 舔av片在线| 99久久精品一区二区三区| 国产不卡一卡二| 欧美xxxx性猛交bbbb| 九九在线视频观看精品| 一二三四中文在线观看免费高清| 亚洲人成网站在线观看播放| 久久精品国产自在天天线| 国产成人aa在线观看| 精品酒店卫生间| 淫秽高清视频在线观看| 在线播放无遮挡| 淫秽高清视频在线观看| 欧美日本视频| 久久99热这里只频精品6学生 | 国产成人福利小说| a级一级毛片免费在线观看| 长腿黑丝高跟| 秋霞伦理黄片| 亚洲熟妇中文字幕五十中出| 蜜桃亚洲精品一区二区三区| 男人狂女人下面高潮的视频| 日韩在线高清观看一区二区三区| 少妇丰满av| 日韩一区二区三区影片| 欧美人与善性xxx| 亚州av有码| 国产精品蜜桃在线观看| 国产精品久久久久久久电影| 九色成人免费人妻av| 如何舔出高潮| 国产免费福利视频在线观看| 日韩成人av中文字幕在线观看| 久久精品国产鲁丝片午夜精品| 日韩人妻高清精品专区| 国模一区二区三区四区视频| 91精品一卡2卡3卡4卡| 免费不卡的大黄色大毛片视频在线观看 | 欧美区成人在线视频| 国产视频首页在线观看| 国产在视频线在精品| 蜜臀久久99精品久久宅男| av国产久精品久网站免费入址| av在线亚洲专区| 国产一区亚洲一区在线观看| 欧美另类亚洲清纯唯美| 狂野欧美白嫩少妇大欣赏| 精品人妻偷拍中文字幕| 在线免费观看不下载黄p国产| 在线观看av片永久免费下载| a级毛色黄片| 亚洲av中文av极速乱| 韩国av在线不卡| av播播在线观看一区| 如何舔出高潮| 美女国产视频在线观看| 天堂网av新在线| 国产一区二区在线av高清观看| 波多野结衣巨乳人妻| 欧美性猛交╳xxx乱大交人| 搞女人的毛片| 欧美成人a在线观看| 在线观看66精品国产| 99久久精品国产国产毛片| 两个人的视频大全免费| 麻豆精品久久久久久蜜桃| 国产一区二区三区av在线| 亚洲国产成人一精品久久久| 亚洲国产精品专区欧美| 99热精品在线国产| 美女xxoo啪啪120秒动态图| 天天一区二区日本电影三级| 亚洲国产最新在线播放| 日韩欧美 国产精品| 亚洲欧美一区二区三区国产| 日韩欧美精品v在线| 中文字幕免费在线视频6| 国产成人91sexporn| 国产日韩欧美在线精品| 欧美一区二区国产精品久久精品| 国产午夜精品论理片| 亚洲av熟女| 三级男女做爰猛烈吃奶摸视频| 毛片女人毛片| 久久久午夜欧美精品| 超碰97精品在线观看| 少妇丰满av| 国产大屁股一区二区在线视频| 国产成人精品婷婷| 国产又色又爽无遮挡免| 日日摸夜夜添夜夜爱| 美女脱内裤让男人舔精品视频| 欧美色视频一区免费| 22中文网久久字幕| 91精品国产九色| 菩萨蛮人人尽说江南好唐韦庄 | eeuss影院久久| 久久久久久久久久久丰满| 日韩大片免费观看网站 | 久久婷婷人人爽人人干人人爱| 毛片女人毛片| 99九九线精品视频在线观看视频| 久久精品国产鲁丝片午夜精品| 五月伊人婷婷丁香| 韩国av在线不卡| 人妻夜夜爽99麻豆av| 少妇猛男粗大的猛烈进出视频 | 九九热线精品视视频播放| 国产大屁股一区二区在线视频| 26uuu在线亚洲综合色| 麻豆精品久久久久久蜜桃| 日韩欧美 国产精品| 69人妻影院| 精品少妇黑人巨大在线播放 | a级一级毛片免费在线观看| 国产一区亚洲一区在线观看| 国产av一区在线观看免费| 男人舔女人下体高潮全视频| 成人二区视频| 国产真实伦视频高清在线观看| 精品一区二区三区视频在线| 精品熟女少妇av免费看| 久久午夜福利片| 黄色一级大片看看| 国内精品宾馆在线| 只有这里有精品99| 欧美高清性xxxxhd video| 亚洲18禁久久av| 亚洲国产日韩欧美精品在线观看| av卡一久久| 成人毛片60女人毛片免费| 桃色一区二区三区在线观看| 国产伦精品一区二区三区视频9| 夜夜爽夜夜爽视频| 欧美性感艳星| 精品国产露脸久久av麻豆 | 国产单亲对白刺激| 美女黄网站色视频| 日韩成人伦理影院| 日韩欧美精品免费久久| 在线观看av片永久免费下载| 午夜精品国产一区二区电影 | 国产午夜福利久久久久久| 韩国av在线不卡| 亚洲欧美成人综合另类久久久 | 亚洲美女视频黄频| 成年av动漫网址| 国产午夜精品久久久久久一区二区三区| 亚洲人与动物交配视频| 身体一侧抽搐| 免费观看的影片在线观看| 你懂的网址亚洲精品在线观看 | 3wmmmm亚洲av在线观看| 精品久久久噜噜| 听说在线观看完整版免费高清| 波野结衣二区三区在线| 欧美xxxx性猛交bbbb| 在线观看一区二区三区| 国产 一区 欧美 日韩| 亚州av有码| 国产69精品久久久久777片| 日产精品乱码卡一卡2卡三| 老司机福利观看| 亚洲熟妇中文字幕五十中出| 伊人久久精品亚洲午夜| 一本一本综合久久| 成人三级黄色视频| 午夜精品一区二区三区免费看| 免费观看人在逋| 日本免费在线观看一区| 啦啦啦啦在线视频资源| 黑人高潮一二区| av在线亚洲专区| 午夜免费激情av| 免费看a级黄色片| 中文字幕人妻熟人妻熟丝袜美| 国产亚洲最大av| av天堂中文字幕网| 免费在线观看成人毛片| 高清日韩中文字幕在线| 欧美日韩国产亚洲二区| 美女cb高潮喷水在线观看| av在线观看视频网站免费| 中文字幕人妻熟人妻熟丝袜美| 嫩草影院精品99| 精品国产露脸久久av麻豆 | 亚洲图色成人| 一级毛片电影观看 | 国产精品精品国产色婷婷| 欧美日韩综合久久久久久| 热99re8久久精品国产| 99久久精品热视频| 国产在线男女| av卡一久久| 亚洲av中文av极速乱| 精品久久久久久久久久久久久| 我要看日韩黄色一级片| 夜夜看夜夜爽夜夜摸| 免费看av在线观看网站| 亚洲自偷自拍三级| 看免费成人av毛片| 能在线免费观看的黄片| 亚洲av成人精品一二三区| 看黄色毛片网站| 男女视频在线观看网站免费| 三级男女做爰猛烈吃奶摸视频| 国产在视频线精品| av.在线天堂| 麻豆成人午夜福利视频| 久久精品人妻少妇| 色视频www国产| 国产 一区精品| 亚洲不卡免费看| 久久精品国产自在天天线| 一卡2卡三卡四卡精品乱码亚洲| 亚洲欧洲国产日韩| 亚洲国产精品国产精品| 爱豆传媒免费全集在线观看| 国产激情偷乱视频一区二区| 综合色av麻豆| 国产伦在线观看视频一区| av在线播放精品| 色网站视频免费| 老司机影院成人| 卡戴珊不雅视频在线播放| 国产精品1区2区在线观看.| 国产亚洲av嫩草精品影院| 身体一侧抽搐| 国产美女午夜福利| 91aial.com中文字幕在线观看| 亚洲久久久久久中文字幕| 日韩欧美三级三区| av专区在线播放| 偷拍熟女少妇极品色| 国产白丝娇喘喷水9色精品| 男插女下体视频免费在线播放| 中文天堂在线官网| 欧美成人精品欧美一级黄| 久久久久久国产a免费观看| 亚洲经典国产精华液单| 日本黄色片子视频| 高清av免费在线| 国产高清不卡午夜福利| 婷婷色av中文字幕| 国产黄a三级三级三级人| 久久久久久久久久久免费av| 免费在线观看成人毛片| 五月玫瑰六月丁香| 亚洲色图av天堂| 中文乱码字字幕精品一区二区三区 | 黄片无遮挡物在线观看| 国产精品电影一区二区三区| videossex国产| 国产高潮美女av| 干丝袜人妻中文字幕| 亚洲欧洲国产日韩| 十八禁国产超污无遮挡网站| 变态另类丝袜制服| 少妇熟女aⅴ在线视频| 在线免费观看不下载黄p国产| 亚洲精品乱久久久久久| 白带黄色成豆腐渣| 久久精品久久久久久久性| 麻豆国产97在线/欧美|