• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    An Intelligent Detection Method for Optical Remote Sensing Images Based on Improved YOLOv7

    2024-01-12 03:45:56ChaoDongandXiangkuiJiang
    Computers Materials&Continua 2023年12期

    Chao Dong and Xiangkui Jiang

    School of Automation,Xi’an University of Posts and Telecommunications,Xi’an,710121,China

    ABSTRACT To address the issue of imbalanced detection performance and detection speed in current mainstream object detection algorithms for optical remote sensing images,this paper proposes a multi-scale object detection model for remote sensing images on complex backgrounds,called DI-YOLO,based on You Only Look Once v7-tiny(YOLOv7-tiny).Firstly,to enhance the model’s ability to capture irregular-shaped objects and deformation features,as well as to extract high-level semantic information,deformable convolutions are used to replace standard convolutions in the original model.Secondly,a Content Coordination Attention Feature Pyramid Network(CCAFPN)structure is designed to replace the Neck part of the original model,which can further perceive relationships between different pixels,reduce feature loss in remote sensing images,and improve the overall model’s ability to detect multi-scale objects.Thirdly,an Implicitly Efficient Decoupled Head (IEDH) is proposed to increase the model’s flexibility,making it more adaptable to complex detection tasks in various scenarios.Finally,the Smoothed Intersection over Union (SIoU) loss function replaces the Complete Intersection over Union (CIoU)loss function in the original model,resulting in more accurate prediction of bounding boxes and continuous model optimization.Experimental results on the High-Resolution Remote Sensing Detection (HRRSD) dataset demonstrate that the proposed DI-YOLO model outperforms mainstream target detection algorithms in terms of mean Average Precision(mAP)for optical remote sensing image detection.Furthermore,it achieves Frames Per Second(FPS)of 138.9,meeting fast and accurate detection requirements.

    KEYWORDS Object detection;optical remote sensing images;YOLOv7-tiny;real-time detection

    1 Introduction

    The detection of objects in optical remote sensing images has become a hot research topic in the field of remote sensing,which is widely used in agriculture,the ocean,the military,and other fields[1–3].However,optical remote sensing images are characterized by a wide image field,large differences in the scales of objects in the image,and intricate backgrounds[4],which leads to a serious impact on detection performance.As artificial intelligence technology continues to develop,it is necessary to design an accurate and fast method to meet actual detection needs.

    Traditional image processing algorithms such as Haar-like Feature Classifier (Haar-like) [5],Histogram of Oriented Gradients(HOG)[6],and Deformable Part-based Model(DPM)[7]are used in the field of object detection.Yao et al.[8]proposed to first use the AdaBoost classifier trained on Haar-like features for coarse detection of ships,and then Line Segment Detector (LSD) to achieve fine detection,which achieves a low false detection rate.Yu et al.[9] proposed a method for vehicle detection.Firstly,the optical remote sensing image is converted into hyperspectral form by using the image conversion technique,and then the AdaBoost algorithm and Haar-like verify whether the candidate vehicle is true or not,to accomplish the detection task.Qi et al.[10]proposed the S-HOG method based on HOG,which solves the problem of HOG being sensitive to direction and unable to detect low-resolution small objects,and the proposed S-HOG method has strong robustness in complex scenes and scenes with large ship transformations.Luo et al.[11]proposed a method based on the object proposal technique using HOG-SVM(Support Vector Machine)classifiers to detect,which has good detection performance while reducing the amount of computation.Qiu et al.[12]proposed an improved model Partial Configuration Model(PCM),which solves the problem of DPM in detecting occluded objects.The problem of performance degradation in detecting remote sensing images with occlusions.Although traditional object detection methods have achieved some success in the field of computer vision,traditional object detection algorithms suffer from the problems of difficulty in extracting image features in complex scenes,high model complexity,and limited generalization ability.

    Convolutional Neural Networks (CNN) have introduced a series of object detection methods that have achieved significant detection performance.Arikumar et al.[13] used a convolutional neural network to further process the point cloud data from 3D LiDAR sensors and camera sensors to improve the accuracy of the model in recognizing objects,which significantly improves the environment-sensing ability of self-driving vehicles.With convolutional neural networks,features of an image are automatically extracted using convolutional kernels,in contrast to traditional image processing algorithms.There are two main types of object detection methods currently available:region extraction-based (two-stage) and regression-based (single-stage).It is common for most of these two-stage object detection algorithms to make use of the Region-CNN(R-CNN)[14–16]family,and the single-stage object detection algorithms include the You Only Look Once(YOLO)family of algorithms[17–24]and the Single Shot MultiBox Detector(SSD)family of algorithms[25,26].

    Based on YOLOv7-tiny,this paper designs a detection model that can effectively identify remote sensing images under complex backgrounds.Specifically,to effectively capture the feature information of multi-scale objects,this paper reconstructs the backbone network to improve the perception ability of objects of different sizes.To enable the network to learn the different information between different features,this paper designs the Content Coordination Attention Feature Pyramid Network (CCAFPN) structure to improve the feature representation ability.To provide more accurate prediction information,an Implicit Efficient decoupling head (IEDH) is designed to provide more detailed texture content and edge information for classification tasks and regression tasks.Finally,the Smoothed Intersection over Union(SIoU)loss function is used to accelerate the convergence speed of the network.

    The contributions of this paper are summarized as follows:

    (1) To enhance the learning ability of the model for multi-scale objects,the backbone network structure is improved.

    (2) To obtain rich channel and spatial information,the CCA-FPN structure is proposed to improve the ability to capture global information.

    (3) To improve the overall recognition accuracy of the model,this paper proposes an IEDH structure to provide the required feature information for regression tasks and classification tasks,respectively.

    (4) The use of the SIoU loss function aims to alleviate the discrepancy between the predicted bounding box and the ground truth,thereby enhancing the detection performance of the model.

    2 Related Works

    Recent years have seen the application of deep learning to optical remote sensing images and remarkable results.Detecting multiple objects accurately in optical remote sensing images,for instance,is crucial in practical applications.For the purpose of resolving this issue,Zhang et al.[27]proposed a novel context-aware detection network(CAD-Net)based on Faster Region-based CNN(Faster RCNN),which utilizes an attention mechanism to regulate the feature map while combining the feature maps.Based on Mask R-CNN,Fu et al.[28] developed an enhanced algorithm that enables rapid assessment of the range and scale of post-earthquake landslides,contributing to post-earthquake rescue and disaster assessment efforts.Zhang et al.[29] presented an enhanced algorithm based on R-CNN,reducing false detection rates and improving overall detection results for ship detection,enabling more accurate detection of vessels in coastal and inland rivers.The research presented by Zhang et al.[30]addressed the challenges involved in extracting diverse and complex buildings from optical remote sensing images.These challenges included an insufficient ability to detect edges and partial feature extraction.Using the Mask R-CNN framework,the authors presented a method for combining edge features to enhance building identification efficiency.

    Despite the advantages of two-stage object detection algorithms with respect to accuracy and performance,these algorithms require complex model structures and extensive parameter computations,which make it difficult to maintain a balance between detection speed and accuracy.Furthermore,two-stage algorithms may lose spatial information about local objects within the entire image.As a result,researchers have started researching single-stage algorithms that balance speed and accuracy.According to Lu et al.[31],attention mechanisms and multiscale feature fusion modules based on SSD were introduced to address the challenges posed by small objects,complex backgrounds,and scale variations.The images can be successfully detected using this method for small objects,but the overall detection performance is inadequate.A YOLOrs network model[32]was proposed based on ResNet[33]and YOLOv3.As a solution,YOLOrs has been designed for real-time object detection,and it is capable of detecting objects at multiple scales.Although this model can meet real-time requirements,it has certain limitations in detection performance.

    In conclusion,applying deep learning methods to the field of optical remote sensing image detection is feasible,but there are still some challenges that need to be addressed.For example,remote sensing image datasets contain targets of various sizes and scales,which poses a challenge for models to adapt to these variations in target sizes.Additionally,some targets in the dataset may have indistinct features in the images,making it difficult for traditional feature extraction methods to accurately detect and identify them.This underscores the need for models with enhanced feature learning and representation capabilities.Furthermore,these datasets may include densely distributed targets,resulting in multiple targets overlapping each other,making their separation a challenging task.This situation increases the complexity of object detection,as models must be capable of distinguishing and precisely locating overlapping targets.Consequently,these challenges present difficulties for most object detection algorithms in achieving a balance between detection speed and performance.Therefore,there is a pressing need to develop more efficient and accurate object detection algorithms to overcome these obstacles,ultimately leading to improved results in practical applications.

    The current YOLO family of algorithms includes YOLOv3-tiny,YOLOv4-tiny,and YOLOv7-tiny.These“Tiny”versions are lightweight versions of the original YOLO models designed to reduce the number of parameters and the computational complexity of the models while maintaining high detection performance for embedded devices and resource-limited environments,where YOLOv7-tiny can achieve faster detection speeds with the same YOLOv7-tiny can achieve faster detection speeds with the same computational resources.To solve the problem of mismatch between detection speed and detection performance,this paper proposes a DI-YOLO object detection model based on YOLOv7-tiny,which realizes fast and accurate detection of optical remote sensing images.In this paper,the effectiveness of this network model is verified on the HRRSD-Dataset: High-Resolution Remote Sensing Detection(HRRSD)[34]dataset and NWPU VHR-10[35]dataset.

    3 Method

    3.1 Overall Structure of DI-YOLO

    In order to ensure that the model has a high detection accuracy while guaranteeing that the model has a fast detection speed,this paper proposes an improved model DI-YOLO based on YOLOv7-tiny,whose overall structure is shown in Fig.1.First,to better deal with the presence of irregular shapes and deformed features in the image,this paper replaces the standard convolution in the backbone network with deformable convolution[36],which enhances the extraction ability of the backbone network for multi-scale target features.Secondly,to obtain detailed information in the high-resolution feature map,this paper designs the Content Coordination Attention Feature Pyramid Network (CCA-FPN) structure for feature fusion to enhance the neck network’s ability to perceive spatial relationships.Then,focusing on the texture content and edge information of the target to improve model performance,this paper designs an Implicit Efficient Decoupling Head(IEDH)as the final network output.Finally,to improve the quality of bounding box matching by considering the degree of overlap between the target shape and the bounding box,this paper optimizes the prediction results using the Smoothed Intersection over Union(SIoU)loss function.

    3.2 Deformable Convolutional

    With deformable convolution,the sampling position of the convolution kernel can be adjusted as opposed to standard convolution.The comparison of deformable convolution with standard convolution is shown in Fig.2.

    Deformable convolution and pooling operations are both two-dimensional operations performed within the same channel.The difference between deformable convolution and standard convolution lies in deformable convolution’s ability to dynamically adjust the sampling positions of the convolution kernels using learned offsets.This allows it to adapt to object shape variations.This capability enables deformable convolution to sample input features at more precise locations.

    The operation of standard convolution is mainly divided into sampling and weighting operations on the input feature map through a regular gridR,whereRdefines the perceptual field and dilation.As shown in Eq.(1).

    Figure 1:DI-YOLO overall structure diagram

    Figure 2:Comparison of deformable convolution and traditional standard convolution.(a) shows a regular sampling grid(green dots);(b)shows demonstrates the sampled positions(deep blue dots)of deformable convolution with increased offsets(light blue arrows);(c)and(d)show depict special cases of(b),including scaling,aspect ratios(anisotropy),and rotation of the image

    For each positionP0on the output feature map,computed using Eq.(2),wherePnenumerates the positions listed inR,wrepresents the weights of the convolutional kernel,andxdenotes the input feature map.

    However,the operation of deformable convolution is different as it incorporates an offsetΔPnon top of the regular sampling positions.The offsetΔPnis generated by a deformable offset generation module,which learns to extract information from the input feature map that is relevant to variations in the object shape.This process is described by the Eq.(3).

    Now,the sampling positions have become irregular due to the introduction of the offsetΔPn.Since the offsetΔPnis typically a decimal value that does not correspond directly to actual points in the feature map,it requires implementation using bilinear interpolation.The transformation formula for this process is represented as Eq.(4).

    wherex(q) represents the value at the integer positions of the feature map,andx(p)=x(P0+Pn+ΔPn) represents the value at all decimal positions after adding the offset.The functiong(a,b) is defined asg(a,b)=max(0,1-|a-b|).

    In summary compared to standard convolution deformable convolution can improve the network model’s ability to model multi-scale object deformation in optical remote sensing images,enabling models to adapt more effectively to objects of different sizes and shapes.

    3.3 CCA-FPN Structure

    This paper presents a feature fusion network CCA-FPN structure as a means of improving the model’s ability to represent image features.Among them,Content-Aware ReAssembly of Features(CARAFE)[37]has a larger sensory field,which can utilize the feature information more efficiently despite the introduction of a small number of parameters and computation.Coordinate Attention(CA) [38] module,which further improves the model’s ability to acquire desired object features in complex environments.Fig.3 shows its structure.

    3.3.1 Content-Aware ReAssembly of Features Lightweight Upsampling Operator Module

    The up-sampling operation in YOLOv7-tiny uses the nearest neighbor interpolation method.Although the method is fast in computation,has a small number of parameters,and is simple to implement,the method only considers the pixel value closest to the object pixel without considering the relationship between the neighboring pixel values,which image may lose some of its features due to this process.Therefore,this paper proposes to replace the original up-sampling module of YOLOv7-tiny with a lightweight up-sampling operator CARAFE,which views the feature up-sampling process as a process of feature reorganization.

    Fig.4 illustrates that CARAFE consists of two main modules:the upsampling kernel prediction moduleψand the content-aware reassembly moduleφ.

    Figure 3:CCA-FPN structure

    Figure 4:Content-Aware ReAssembly of Features Module structure diagram

    Upsampling kernel prediction moduleψpredicts an internal kernelWl′for each positionl′based on the neighboring positions ofXl,as shown in Eq.(5).

    Content-aware reassembly moduleφrecombines the neighboring positions ofXlwith the predicted internal kernelWl′,as shown in Eq.(6).

    The CARAFE algorithm employs the feature recombination module to perform upsampling on a feature input map that has a shape ofC×H×W,assuming an upsampling factor ofα.Based on the prediction of upsampling kernels,CARAFE then performs the upsampling by utilizing upsampling kernel prediction module.As a result,C×αH×αWis the shape of the output feature map.

    In summary,the CARAFE module possesses a substantial receptive field and accomplishes lightweight operations through the integration of upsampling kernels that exploit the semantic information within the feature map.

    3.3.2 Coordinate Attention Module

    Hybrid attention mechanisms,which consider multiple types of information simultaneously,enable more comprehensive feature extraction and improve model performance.Fig.5 illustrates the basic structure of the CA model.The CA model incorporates not only channel information but also position information related to directions.Due to its lightweight and flexible design,CA can be easily integrated into various deep learning models without introducing excessive computational burden,thereby enhancing model efficiency and speed while maintaining performance.

    Figure 5:Coordinate attention module structural diagram

    The combination of information embedding and attention generation makes this attention mechanism capable of encoding long-term dependencies and channel relationships.The following describes these two steps.

    First is the coordinate information embedding.Given an input feature mapF∈RC?H?W,traditional global average pooling is divided into two steps as shown in Eq.(7).It involves two onedimensional average pooling operations using pooling kernels(H,1)and(1,W)in different directions of the feature map.This results in two embedded feature maps.A detailed description is provided below:

    Along the horizontal direction,specifically for the output at height h in channelc,we obtain a feature vector,which can be represented by Eq.(8).

    Along the vertical direction,specifically for the output at widthwin channelc,we obtain a feature vector,which can be represented by Eq.(9).

    A pair of direction-aware feature maps is derived by aggregating the two transformations mentioned above along the two spatial directions,(h)and(w),as shown in Fig.6.It is possible to effectively capture spatial dependencies and maintain spatial sensitivity by aggregating transformed feature maps both horizontally and vertically.This helps the network focus on relevant regions and enhances its ability to localize objects accurately.

    Figure 6:Direction-aware feature map

    Next is the process of attention generation,known as coordinate attention generation.According to the two feature maps obtained,(h)and(w),they are concatenated along the spatial dimension.In the next step,they are subjected to a 1×1 convolutional transformation followed by an activation function,resulting inf∈RC/r×(H+W),as shown in Eq.(10).

    Subsequently,a split operation is performed along the spatial dimension,yielding two separate feature maps,f h∈Rc/r?H?1andf w∈Rc/r?1?w.Here,ris a reduction factor similar to the one used in Squeeze-and-Excitation networks (SENet) [39],which aims to reduce the model’s parameter count.Finally,the transform and sigmoid(σ)operations,as shown in Eqs.(11)and(12),are applied to these two feature maps,resulting in the attention vectorsghandgw.

    Finally,the attention vectorsghandgware expanded through row-wise and column-wise broadcasting operations,as illustrated in Fig.7.They are transformed to have the same dimensions as the input image and then undergo element-wise multiplication at corresponding positions.Eq.(13)shows the output of the Coordinate Attention blocky.

    Figure 7:Feature expansion diagram

    Unlike other attention mechanisms such as SENet that focus solely on constructing interdependencies among channels while neglecting spatial information,thereby exhibiting certain limitations[40],in CA attention mechanisms,both channel information and spatial information are considered.The module generates two attention maps by attending to the input tensor in both horizontal and vertical directions.According to attention maps,each element reflects whether the object of interest appears in the corresponding row and column.Through the encoding process,attention is more effectively coordinated to locate objects of interest,improving recognition performance.

    3.4 Implicitly Efficient Decoupled Head

    By combining the decoupled operations from YOLOX,YOLOv6,and the implicit knowledge learning operation introduced in YOLOR,this paper proposes a method with an IEDH.As shown in Fig.8,this decoupled head exhibits improved detection performance and faster speed.

    The construction process of the implicit efficient decoupling head is as follows: first,the input feature layer is feature-integrated by 1×1 convolution,and then the whole prediction process is divided into a classification part and a regression part.The classification part uses 3 × 3 convolution for feature extraction,and then the classification task is performed by 1 × 1 convolution to determine which category it belongs to.The regression part uses the same operation as the classification part,but after feature extraction,the regression part employs two 1×1 convolutions to obtain regression prediction results to determine whether the feature has a corresponding object.Finally,these three predictions are superimposed and integrated.

    Figure 8:Implicit efficient decoupled head structure

    3.5 Smoothed Intersection over Union Loss Function

    Complete Intersection over Union (CIoU) loss function does not incorporate the direction information contained in the ground truth,which may affect the training speed of network models and inference accuracy.Given the diverse range of sizes and shapes exhibited by objects,we propose the SIoU[41]loss function as a viable alternative to the CIoU loss function.There are four cost functions in the SIoU loss function.

    According to Eq.(14),angular loss cost is defined as follows:

    whereChis the difference in height between a real frame center point and the predicted frame’s center point as shown in Eq.(15):

    σdistance between the center points of the predicted and ground truth boxes as shown in Eq.(16):

    The angle loss is zero whenαisor 0,and if α ≤,αis minimized;otherwise,βis minimized.The computational graph is illustrated in Fig.9.

    Figure 9:Angular loss value

    whereρx,ρy,γare shown in Eq.(19),and hereCw,Chrepresent the smallest external rectangle of the real and predicted boxes,which are defined by their width and height.As shown in Fig.10.

    Figure 10:Distance loss calculation

    The shape loss cost function is defined as in Eq.(20).

    where theww,whexpressions are shown in Eq.(21).

    in Eq.(21),w,h,wgt,andhgtrepresent the width and height of the predicted box and the ground truth box,respectively.θcontrols the emphasis on the shape loss.

    According to Eq.(22),the IoU cost function is as follows:

    As a summary,Eq.(23)illustrates the SIoU loss function.

    4 Experiments

    4.1 Experimental Conditions

    4.1.1 Experimental Data Set

    To validate the reliability of the proposed model,two datasets were used:HRRSD and NWPU VHR-10.HRRSD is a dataset developed in 2017 by the Center for Optical Imagery Analysis and Learning of the Xi’an Institute of Optics and Precision Mechanics of the Chinese Academy of Sciences,specifically designed for remote sensing image object detection research.A total of 55,740 object instances are included in the HRRSD dataset,with approximately 4,000 instances per category.It covers 13 types of remote sensing land object categories,including ships,bridges,ground track fields,storage tanks,basketball courts,tennis courts,airplanes,baseball fields,harbors,vehicles,intersections,T-intersections,and parking lots.One notable feature of this dataset is the relatively balanced sample distribution among different categories,with each category containing over 4,000 samples.Moreover,the dataset includes a diverse range of detection objects,with each image containing various sizes and types of objects,and some images containing densely packed objects.Northwestern Polytechnical University,China,has released the NWPU VHR-10 dataset.With 650 images containing objects,150 background images,and 10 object classes,it is comprised of 650 images from Google Earth.

    4.1.2 Experimental Environment

    The experimental setup for this study is shown in Table 1.

    Table 1:Experimental setup

    The hyperparameter settings before model training in this paper are shown in Table 2.

    Table 2:Hyperreference settings

    4.2 Evaluation Metrics

    A metric used to evaluate network models’object detection performance is mean Average Precision (mAP).mAP represents the average of all class results and is used to evaluate the overall performance of the detection algorithm,where mAP0.5 evaluates the object detection performance at an IoU threshold of 0.5,mAP0.5:0.95 provides a more detailed evaluation by considering a range of IoU thresholds between 0.5 and 0.95 in steps of 0.05.This study utilizes Precision(P),Recall(R),Average Precision (AP),and mAP as performance evaluation metrics for DI-YOLO.The following are specific expressions:

    True Positives(TP)in Eq.(24)are the number of positive samples correctly recognized as positive samples;False Positives (FP) are false positive samples,the number of negative samples incorrectly recognized as positive samples.

    False Negatives(FN)in Eq.(25)is the number of false negatives,the number of positive samples that are incorrectly recognized as negative.

    In Eq.(26)P stands for precision and R stands for recall.

    N in Eq.(27)denotes N classifications in the dataset.

    4.3 Experimental Results

    4.3.1 Ablation Experiments

    Employing YOLOv7-tiny as the base model,we performed ablation experiments on the HRRSD dataset to substantiate the efficacy of the proposed algorithm for optical remote sensing image detection.Various improvement modules were evaluated in the DI-YOLO model to determine their performance.The results of the experiments are presented in Table 3.

    Table 3:Ablation experiments

    Based on deformable convolution characteristics,this paper replaces standard convolution with deformable convolution in the YOLOv7-tiny backbone network.As indicated in Table 3,the model’s overall performance continues to improve following the implementation of deformable convolution,while reducing the computational burden.

    In Fig.11,the gradient-weighted class activation mapping(Grad-CAM)[42]is used to visualize the heatmap of CCA-FPN.It can be clearly observed that with the addition of the CCA-FPN structure again,the high-confidence regions of the object become more prominent,thus focusing on the important elements and reducing the interference in the recognition process.

    To fully extract information from the feature map,IEDH is used as the final output of the network.Lastly,the SIoU loss function is used in this paper to accelerate network convergence.The experimental results show that although DI-YOLO increases the parameters and computation,its precision increases by 4.2%,recall by 2.3%,mAP0.5 by 2.9%,and mAP0.5:0.95 by 2.5%.According to the results above,the model maintains high recognition performance while meeting the requirements of real-time processing.

    Figure 11:Heat map of detection results using Grad-CAM before and after adding the CCA-FPN module

    In addition,Fig.12 illustrates a performance comparison graph between the DI-YOLO and YOLOv7-tiny models based on ablation results.These graphs include precision curves,recall curves,and mAP0.5 curves.Based on the comparison graphs,it is evident that the DI-YOLO model outperforms YOLOv7-tiny in each of the aforementioned evaluation metrics.

    Figure 12:Performance comparison between YOLOv7-tiny and DI-YOLO models

    Fig.13 illustrates the loss comparison graph.The graph indicates that the proposed model introduced in this paper exhibits a faster convergence speed compared to the baseline model YOLOv7-tiny.

    Figure 13:Loss comparison between YOLOv7-tiny and DI-YOLO models

    4.3.2 Comparison Experiments

    As shown in Table 4,the proposed DI-YOLO is compared with classical object detection algorithms using Resnet-50 as the backbone of Faster R-CNN,YOLOv3-tiny,YOLOv4-tiny,YOLOv5n,YOLOv7-tiny and YOLOv8-n.

    Table 4:Performance comparison of different algorithms under HRRSD dataset

    In Fig.14,it is shown that DI-YOLO has high performance,can accurately extract image features,and is suitable to detect multiscale image detection under complex backgrounds such as those encountered in optical remote sensing.

    Figure 14:The HRRSD dataset was used to compare the detection effects of DI-YOLO

    From the experimental results,it is shown that the mAP of DI-YOLO is the best,which is because the CCA-FPN structure improves the feature resolution by feature reorganization,introduces context modeling in the feature up-sampling process,which in turn can adapt to targets with different scales,orientations,and shapes,and improves the sensitivity to various types of targets in remote sensing images,which proves that the structure can be used for multiscale in the complicated context of object detection tasks in complex backgrounds.Combined with Fig.14,it can be concluded that DI-YOLO performs better in the case of complex background features or large differences between background feature scales and targets,which is due to the fact that the efficient decoupling head proposed in this paper better handles both the classification and the localization task.

    4.3.3 Generalizability Verification

    The DI-YOLO model needs to be evaluated on an image set that does not include HRRSD data to further validate its generalization ability.The HRRSD dataset contains the same categories as the NWPU VHR-10 dataset for validation purposes.The experimental samples include but are not limited to,ships,aircraft,baseball fields,and ports of various sizes in different contexts.Fig.15 illustrates the comparison of detection results between YOLOv7-tiny and DI-YOLO.Although DI-YOLO has a few misclassifications based on experimental results,the proposed DI-YOLO model shows better object detection accuracy compared to YOLOv7-tiny while maintaining real-time performance.

    Figure 15:The NWPU VHR-10 dataset was used to compare the detection effects of DI-YOLO

    5 Conclusion

    This paper presents a model for object detection,DI-YOLO,by improving on the basic framework of YOLOv7-tiny.The deformable convolution was first used to make the model optimized for feature extraction of multi-scale objects.Secondly,a CCA-FPN structure is designed to enhance the model’s capability to recognize multi-objects.Next,the IEDH is employed to further focus on the object’s texture content and edge information.Lastly,the bias effect can be reduced by using the SIoU loss function.The experiments show that compared with other mainstream algorithms,DI-YOLO can obtain better detection results while ensuring real-time performance.Although the overall detection performance of DI-YOLO performs better,its effect on detecting optical remote sensing images with severely dense targets and tiny targets performs poorly compared to mainstream models,which will be a key research focus in the future.

    Acknowledgement:The authors gratefully acknowledge the support of Shaanxi Province’s Key Research and Development Plan.

    Funding Statement:Funding for this research was provided by 511 Shaanxi Province’s Key Research and Development Plan(No.2022NY-087).

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design:Chao Dong and Xiangkui Jiang;data collection:Chao Dong;analysis and interpretation of results: Chao Dong and Xiangkui Jiang;draft manuscript preparation: Chao Dong.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The datasets presented in this study are available through:https://github.com/CrazyStoneonRoad/TGRS-HRRSD-Dataset;https://gcheng-nwpu.github.io/#Datasets.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    极品人妻少妇av视频| 久久精品国产a三级三级三级| 26uuu在线亚洲综合色| 亚洲av欧美aⅴ国产| 成人毛片60女人毛片免费| 黑丝袜美女国产一区| 久久久精品免费免费高清| 精品一区二区三卡| 丰满少妇做爰视频| 久久久久久久久免费视频了| 国产精品欧美亚洲77777| 男女边吃奶边做爰视频| av一本久久久久| 国产成人精品婷婷| 欧美日韩亚洲国产一区二区在线观看 | 亚洲熟女精品中文字幕| 在线精品无人区一区二区三| 午夜福利乱码中文字幕| 晚上一个人看的免费电影| 亚洲精品在线美女| 亚洲五月色婷婷综合| 久久久久精品性色| 日本欧美国产在线视频| 久久国内精品自在自线图片| 久久国内精品自在自线图片| 在现免费观看毛片| 男人爽女人下面视频在线观看| 色播在线永久视频| 婷婷色综合www| 亚洲国产欧美网| 久久午夜福利片| 丝袜美足系列| 亚洲婷婷狠狠爱综合网| 亚洲国产最新在线播放| 亚洲欧美清纯卡通| 9热在线视频观看99| 性色av一级| 国产乱来视频区| 免费日韩欧美在线观看| 9色porny在线观看| 亚洲精品国产色婷婷电影| 在线观看三级黄色| 熟女少妇亚洲综合色aaa.| 国产精品一区二区在线不卡| 国产精品女同一区二区软件| 中文字幕色久视频| 亚洲精品国产色婷婷电影| 午夜老司机福利剧场| 中文乱码字字幕精品一区二区三区| 国产成人aa在线观看| 蜜桃国产av成人99| 熟妇人妻不卡中文字幕| 欧美精品高潮呻吟av久久| 午夜福利视频在线观看免费| av视频免费观看在线观看| 免费在线观看完整版高清| 免费大片黄手机在线观看| 国产精品二区激情视频| 性高湖久久久久久久久免费观看| 亚洲欧美成人精品一区二区| 日本欧美视频一区| 在线观看人妻少妇| av在线app专区| 精品亚洲成国产av| 成年女人在线观看亚洲视频| 不卡视频在线观看欧美| 亚洲成色77777| 边亲边吃奶的免费视频| 1024视频免费在线观看| 亚洲av欧美aⅴ国产| 国产精品一国产av| 肉色欧美久久久久久久蜜桃| 大话2 男鬼变身卡| 欧美亚洲 丝袜 人妻 在线| 日韩一卡2卡3卡4卡2021年| 日本猛色少妇xxxxx猛交久久| 男女高潮啪啪啪动态图| 一级片'在线观看视频| 午夜激情久久久久久久| 两个人看的免费小视频| 久久人人爽av亚洲精品天堂| 黄色毛片三级朝国网站| 不卡av一区二区三区| 激情视频va一区二区三区| 自线自在国产av| 宅男免费午夜| 老司机影院毛片| 一级,二级,三级黄色视频| 欧美人与性动交α欧美精品济南到 | 人妻 亚洲 视频| 日韩欧美一区视频在线观看| av视频免费观看在线观看| 亚洲精品美女久久久久99蜜臀 | 有码 亚洲区| 人人澡人人妻人| 久久久久久久久免费视频了| 丰满乱子伦码专区| 亚洲成国产人片在线观看| 2018国产大陆天天弄谢| 欧美日韩视频高清一区二区三区二| 一级毛片我不卡| 免费黄色在线免费观看| 国产爽快片一区二区三区| 乱人伦中国视频| 老司机影院成人| 国产精品国产三级国产专区5o| 国产精品久久久久久精品古装| 欧美97在线视频| a级片在线免费高清观看视频| 免费观看在线日韩| 国产黄频视频在线观看| 亚洲国产欧美在线一区| 一区二区三区四区激情视频| 咕卡用的链子| 亚洲av综合色区一区| 久久婷婷青草| 日本黄色日本黄色录像| 满18在线观看网站| 欧美日韩精品成人综合77777| 免费久久久久久久精品成人欧美视频| 亚洲成人手机| 中文乱码字字幕精品一区二区三区| 久久久久久久大尺度免费视频| 亚洲国产看品久久| 大香蕉久久网| 成人毛片60女人毛片免费| 国产国语露脸激情在线看| 久久热在线av| 亚洲av免费高清在线观看| 久久久精品区二区三区| 日本欧美视频一区| 国产精品.久久久| 亚洲av中文av极速乱| 99久久中文字幕三级久久日本| 青春草亚洲视频在线观看| 一本久久精品| 精品国产超薄肉色丝袜足j| 国产乱人偷精品视频| 十八禁网站网址无遮挡| 91午夜精品亚洲一区二区三区| 街头女战士在线观看网站| 一边亲一边摸免费视频| 久久午夜福利片| 亚洲精品中文字幕在线视频| 久热这里只有精品99| 精品国产露脸久久av麻豆| 国产一区二区三区av在线| 亚洲av福利一区| 天美传媒精品一区二区| 午夜激情av网站| 国产精品麻豆人妻色哟哟久久| 亚洲男人天堂网一区| 高清黄色对白视频在线免费看| 永久免费av网站大全| 亚洲人成电影观看| 国产成人精品婷婷| 最近手机中文字幕大全| 亚洲成人一二三区av| 国产不卡av网站在线观看| 成人亚洲精品一区在线观看| 麻豆av在线久日| 亚洲伊人久久精品综合| 久久久久久久久免费视频了| 国产精品久久久久成人av| 一级黄片播放器| 久久精品熟女亚洲av麻豆精品| 黑丝袜美女国产一区| 男人操女人黄网站| 国产精品一区二区在线不卡| 9热在线视频观看99| 少妇精品久久久久久久| 午夜福利在线免费观看网站| 亚洲人成电影观看| 免费久久久久久久精品成人欧美视频| 丰满少妇做爰视频| 中文字幕制服av| www.精华液| 成人国产av品久久久| 天堂8中文在线网| 亚洲欧美清纯卡通| 免费黄网站久久成人精品| 亚洲欧美成人精品一区二区| 国产成人精品久久久久久| 久久99蜜桃精品久久| 可以免费在线观看a视频的电影网站 | 黄色一级大片看看| 久久免费观看电影| 人人妻人人澡人人看| 亚洲精品久久午夜乱码| 麻豆乱淫一区二区| 一区福利在线观看| 国产免费福利视频在线观看| 中国三级夫妇交换| 最近手机中文字幕大全| 亚洲欧美中文字幕日韩二区| 日本91视频免费播放| 久久久久久久久久久免费av| 波多野结衣一区麻豆| 午夜久久久在线观看| 一本大道久久a久久精品| 999精品在线视频| 蜜桃在线观看..| 精品福利永久在线观看| 中文字幕色久视频| 亚洲精品国产av成人精品| 日韩制服丝袜自拍偷拍| 亚洲精品美女久久久久99蜜臀 | 女人被躁到高潮嗷嗷叫费观| 亚洲精品美女久久久久99蜜臀 | 国产精品一二三区在线看| 久久精品熟女亚洲av麻豆精品| 国产精品一区二区在线观看99| av福利片在线| 综合色丁香网| www日本在线高清视频| 午夜精品国产一区二区电影| 黄片小视频在线播放| 91午夜精品亚洲一区二区三区| 国产精品欧美亚洲77777| 美女国产高潮福利片在线看| 少妇精品久久久久久久| 黄频高清免费视频| 一级毛片 在线播放| 美女福利国产在线| 伊人久久大香线蕉亚洲五| 女性生殖器流出的白浆| 久久久精品国产亚洲av高清涩受| 黑人欧美特级aaaaaa片| 国产成人aa在线观看| 亚洲三级黄色毛片| 欧美xxⅹ黑人| 久久久久视频综合| 亚洲欧美清纯卡通| 一级爰片在线观看| 看十八女毛片水多多多| 精品国产一区二区久久| 黄色毛片三级朝国网站| 自线自在国产av| 狠狠婷婷综合久久久久久88av| 大码成人一级视频| 99久国产av精品国产电影| 国产成人精品一,二区| 色网站视频免费| 国产精品成人在线| 成人影院久久| 国产高清国产精品国产三级| 亚洲一区中文字幕在线| 久久久欧美国产精品| 日韩一本色道免费dvd| 在线观看美女被高潮喷水网站| 熟妇人妻不卡中文字幕| 欧美激情高清一区二区三区 | 日韩制服骚丝袜av| 毛片一级片免费看久久久久| 久久精品国产亚洲av天美| 1024香蕉在线观看| 下体分泌物呈黄色| 韩国精品一区二区三区| 在线天堂中文资源库| 香蕉国产在线看| 9色porny在线观看| 欧美成人午夜精品| 欧美人与善性xxx| 超色免费av| 校园人妻丝袜中文字幕| 纯流量卡能插随身wifi吗| 99re6热这里在线精品视频| 超色免费av| 黑人欧美特级aaaaaa片| 亚洲av综合色区一区| 亚洲精品美女久久av网站| 亚洲色图 男人天堂 中文字幕| 免费av中文字幕在线| 久久人人爽人人片av| 午夜福利一区二区在线看| 伊人亚洲综合成人网| 777米奇影视久久| 五月天丁香电影| 一区二区日韩欧美中文字幕| 欧美精品一区二区大全| 精品一区二区免费观看| 一本久久精品| 天天影视国产精品| 亚洲精品aⅴ在线观看| 高清av免费在线| 最近2019中文字幕mv第一页| 18+在线观看网站| 男女啪啪激烈高潮av片| 国产乱人偷精品视频| 亚洲,欧美精品.| 69精品国产乱码久久久| 啦啦啦视频在线资源免费观看| 九九爱精品视频在线观看| 中文字幕另类日韩欧美亚洲嫩草| 777米奇影视久久| 三级国产精品片| 宅男免费午夜| 久久这里只有精品19| 老司机亚洲免费影院| 尾随美女入室| 亚洲av成人精品一二三区| 色婷婷久久久亚洲欧美| 国产日韩欧美在线精品| 亚洲av日韩在线播放| 黄片播放在线免费| 亚洲三区欧美一区| 90打野战视频偷拍视频| 日韩大片免费观看网站| 亚洲精品日本国产第一区| 久热这里只有精品99| 只有这里有精品99| 日日爽夜夜爽网站| 国产精品麻豆人妻色哟哟久久| 成人毛片60女人毛片免费| 欧美日韩视频高清一区二区三区二| 日日啪夜夜爽| a级毛片在线看网站| 一区二区三区四区激情视频| 丁香六月天网| 寂寞人妻少妇视频99o| 国产视频首页在线观看| 妹子高潮喷水视频| 性高湖久久久久久久久免费观看| 亚洲国产欧美在线一区| 天天躁日日躁夜夜躁夜夜| 国产xxxxx性猛交| 国产精品国产三级专区第一集| videossex国产| 青春草视频在线免费观看| 超色免费av| 精品卡一卡二卡四卡免费| 亚洲国产看品久久| 中文欧美无线码| 不卡av一区二区三区| 精品国产一区二区三区四区第35| 国产极品天堂在线| 午夜日韩欧美国产| 在线观看免费视频网站a站| 成人手机av| 99九九在线精品视频| 18禁国产床啪视频网站| 成人国产av品久久久| 好男人视频免费观看在线| 丝袜喷水一区| 1024视频免费在线观看| 看免费av毛片| 亚洲欧美色中文字幕在线| 亚洲人成电影观看| 精品卡一卡二卡四卡免费| 午夜福利一区二区在线看| 亚洲av成人精品一二三区| 国产成人aa在线观看| 久久鲁丝午夜福利片| 日韩一卡2卡3卡4卡2021年| 亚洲精品成人av观看孕妇| 久久国产精品大桥未久av| av一本久久久久| 男女午夜视频在线观看| 日韩人妻精品一区2区三区| 国精品久久久久久国模美| 少妇的逼水好多| 日本vs欧美在线观看视频| 国产精品国产三级国产专区5o| 黄色配什么色好看| 色哟哟·www| 久久久久久久国产电影| 伊人亚洲综合成人网| 91精品伊人久久大香线蕉| 午夜福利一区二区在线看| 99热网站在线观看| 建设人人有责人人尽责人人享有的| 黑丝袜美女国产一区| 中国国产av一级| 免费黄频网站在线观看国产| 成人手机av| 热99国产精品久久久久久7| 99国产精品免费福利视频| 午夜久久久在线观看| 97在线人人人人妻| 香蕉国产在线看| 欧美国产精品va在线观看不卡| 亚洲精品美女久久久久99蜜臀 | 欧美变态另类bdsm刘玥| 亚洲欧洲国产日韩| a 毛片基地| 看免费av毛片| 三上悠亚av全集在线观看| 国产精品 国内视频| 热re99久久国产66热| 久久精品国产亚洲av天美| 亚洲精品久久久久久婷婷小说| 99久久综合免费| 人人妻人人爽人人添夜夜欢视频| 免费不卡的大黄色大毛片视频在线观看| 亚洲精品成人av观看孕妇| 成年美女黄网站色视频大全免费| 黄色毛片三级朝国网站| 精品视频人人做人人爽| 久久精品国产自在天天线| 国产精品国产三级国产专区5o| 亚洲第一区二区三区不卡| 国产成人免费无遮挡视频| 国产又色又爽无遮挡免| 99久久中文字幕三级久久日本| 18禁动态无遮挡网站| 免费高清在线观看视频在线观看| 免费在线观看视频国产中文字幕亚洲 | 新久久久久国产一级毛片| freevideosex欧美| 国产精品久久久久成人av| 亚洲精品美女久久av网站| 黑人猛操日本美女一级片| 女人高潮潮喷娇喘18禁视频| 国产极品天堂在线| 欧美 亚洲 国产 日韩一| 欧美激情极品国产一区二区三区| 国产精品 欧美亚洲| 精品亚洲乱码少妇综合久久| 一区二区三区精品91| 99九九在线精品视频| 亚洲成av片中文字幕在线观看 | 中文字幕精品免费在线观看视频| 伦精品一区二区三区| 欧美精品av麻豆av| 国产亚洲av片在线观看秒播厂| 欧美日韩精品成人综合77777| 曰老女人黄片| 80岁老熟妇乱子伦牲交| 亚洲国产av新网站| av.在线天堂| 一本—道久久a久久精品蜜桃钙片| 超碰97精品在线观看| 大陆偷拍与自拍| 91精品国产国语对白视频| 国产伦理片在线播放av一区| 大码成人一级视频| 国产乱人偷精品视频| 午夜av观看不卡| 亚洲欧美日韩另类电影网站| 亚洲av日韩在线播放| 99久久综合免费| 日韩,欧美,国产一区二区三区| 有码 亚洲区| 国产又爽黄色视频| 性少妇av在线| 香蕉精品网在线| 亚洲av国产av综合av卡| 国产精品久久久久久精品电影小说| 免费观看在线日韩| 美国免费a级毛片| 久久精品人人爽人人爽视色| 天堂8中文在线网| 亚洲欧美成人精品一区二区| 黑人猛操日本美女一级片| 麻豆乱淫一区二区| 久久国产亚洲av麻豆专区| 国产色婷婷99| 免费高清在线观看视频在线观看| 美女大奶头黄色视频| 久久精品久久精品一区二区三区| 国产在线视频一区二区| 成人漫画全彩无遮挡| 久久毛片免费看一区二区三区| 国产精品免费大片| 精品人妻一区二区三区麻豆| 欧美日韩精品成人综合77777| 久久久久久免费高清国产稀缺| 男女啪啪激烈高潮av片| a级毛片黄视频| 国产成人精品久久二区二区91 | 欧美精品一区二区免费开放| 久久久久国产精品人妻一区二区| 欧美精品人与动牲交sv欧美| av不卡在线播放| 亚洲国产日韩一区二区| 久久人人爽av亚洲精品天堂| 欧美黄色片欧美黄色片| 我的亚洲天堂| 一区二区三区四区激情视频| 亚洲男人天堂网一区| 国产人伦9x9x在线观看 | 国产精品无大码| 亚洲国产欧美日韩在线播放| av福利片在线| 9热在线视频观看99| 亚洲av欧美aⅴ国产| 黄色视频在线播放观看不卡| 在线看a的网站| 亚洲精品国产av成人精品| 欧美日本中文国产一区发布| 青春草亚洲视频在线观看| 久久毛片免费看一区二区三区| 国产精品人妻久久久影院| 免费黄频网站在线观看国产| 久久精品久久久久久噜噜老黄| 伊人久久国产一区二区| 亚洲av.av天堂| 国产深夜福利视频在线观看| 免费观看a级毛片全部| 婷婷成人精品国产| 男人操女人黄网站| 女的被弄到高潮叫床怎么办| 亚洲欧洲国产日韩| 亚洲精品一区蜜桃| 日韩制服骚丝袜av| 精品99又大又爽又粗少妇毛片| 亚洲av成人精品一二三区| av在线app专区| 人人妻人人澡人人看| 欧美日韩亚洲高清精品| 成人黄色视频免费在线看| 亚洲欧美日韩另类电影网站| 丝袜人妻中文字幕| 黄色视频在线播放观看不卡| 国产女主播在线喷水免费视频网站| 免费在线观看视频国产中文字幕亚洲 | 亚洲男人天堂网一区| 久久精品久久精品一区二区三区| 捣出白浆h1v1| 中文字幕av电影在线播放| 成年人免费黄色播放视频| 看免费av毛片| 欧美激情极品国产一区二区三区| 中文字幕av电影在线播放| 色94色欧美一区二区| 色播在线永久视频| 熟女电影av网| 日本欧美国产在线视频| 男女下面插进去视频免费观看| 亚洲国产色片| 男的添女的下面高潮视频| 少妇人妻精品综合一区二区| 丝袜人妻中文字幕| 久久久久久免费高清国产稀缺| 五月伊人婷婷丁香| 三上悠亚av全集在线观看| 精品国产一区二区久久| 考比视频在线观看| 99九九在线精品视频| 汤姆久久久久久久影院中文字幕| 亚洲国产精品国产精品| www.熟女人妻精品国产| 欧美最新免费一区二区三区| 一边摸一边做爽爽视频免费| 菩萨蛮人人尽说江南好唐韦庄| 国产av一区二区精品久久| 99久久综合免费| 亚洲精品日韩在线中文字幕| 国产在线一区二区三区精| 成人影院久久| 欧美日韩一区二区视频在线观看视频在线| 亚洲三区欧美一区| 可以免费在线观看a视频的电影网站 | 九草在线视频观看| 日韩 亚洲 欧美在线| 久久这里只有精品19| 18+在线观看网站| 美国免费a级毛片| 两性夫妻黄色片| 欧美av亚洲av综合av国产av | 亚洲精品国产av蜜桃| 国产精品 欧美亚洲| 国产白丝娇喘喷水9色精品| 在线免费观看不下载黄p国产| 日韩制服骚丝袜av| 亚洲国产精品999| 99久久精品国产国产毛片| 免费观看a级毛片全部| 成年动漫av网址| 91精品伊人久久大香线蕉| 国产黄色免费在线视频| 女人高潮潮喷娇喘18禁视频| 人人澡人人妻人| 国产福利在线免费观看视频| 18禁动态无遮挡网站| 国产男人的电影天堂91| 欧美黄色片欧美黄色片| 久久午夜福利片| 天天操日日干夜夜撸| 黄网站色视频无遮挡免费观看| 成人毛片a级毛片在线播放| 欧美人与性动交α欧美软件| 中文字幕人妻丝袜制服| 日韩 亚洲 欧美在线| 久久精品国产自在天天线| 久久国产精品男人的天堂亚洲| 欧美少妇被猛烈插入视频| 免费在线观看完整版高清| 久久久久久久精品精品| 性高湖久久久久久久久免费观看| 久久精品久久精品一区二区三区| 久久久久精品人妻al黑| 9191精品国产免费久久| 亚洲欧美精品自产自拍| 黄色视频在线播放观看不卡| 亚洲精品,欧美精品| 亚洲精品国产av蜜桃| 国产精品人妻久久久影院| 免费看不卡的av| 久久青草综合色| 欧美激情高清一区二区三区 | 国产成人欧美| 一级毛片 在线播放| 欧美变态另类bdsm刘玥| 亚洲情色 制服丝袜| 五月开心婷婷网| 国语对白做爰xxxⅹ性视频网站| 欧美精品人与动牲交sv欧美| 亚洲av综合色区一区| 国产成人91sexporn| 久久精品aⅴ一区二区三区四区 | 国产爽快片一区二区三区| 国产一区亚洲一区在线观看| 欧美激情 高清一区二区三区| 男男h啪啪无遮挡|