• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Swin-PAFF:A SAR Ship Detection Network with Contextual Cross-Information Fusion

    2023-12-15 03:56:38YujunZhangDezhiHanandPengChen
    Computers Materials&Continua 2023年11期

    Yujun Zhang,Dezhi Han and Peng Chen

    School of Information Engineering,Shanghai Maritime University,Shanghai,201306,China

    ABSTRACT Synthetic Aperture Radar (SAR) image target detection has widespread applications in both military and civil domains.However,SAR images pose challenges due to strong scattering,indistinct edge contours,multi-scale representation,sparsity,and severe background interference,which make the existing target detection methods in low accuracy.To address this issue,this paper proposes a multi-scale fusion framework(Swin-PAFF)for SAR target detection that utilizes the global context perception capability of the Transformer and the multi-layer feature fusion learning ability of the feature pyramid structure(FPN).Firstly,to tackle the issue of inadequate perceptual image context information in SAR target detection,we propose an end-to-end SAR target detection network with the Transformer structure as the backbone.Furthermore,we enhance the ability of the Swin Transformer to acquire contextual features and cross-information by incorporating a Swin-CC backbone network model that combines the Spatial Depthwise Pooling(SDP)module and the self-attentive mechanism.Finally,we design a cross-layer fusion neck module(PAFF)that better handles multi-scale variations and complex situations(such as sparsity,background interference,etc.).Our devised approach yields a noteworthy AP@0.5:0.95 performance of 91.3%when assessed on the HRSID dataset.The application of our proposed technique has resulted in a noteworthy advancement of 8%in the AP@0.5:0.95 scores on the HRSID dataset.

    KEYWORDS Transformer;deep learning;SAR object detection;ship detection

    1 Introduction

    Synthetic Aperture Radar(SAR)is an advanced active microwave sensor used for high-resolution remote sensing observations of the Earth [1].It has an all-weather,all-day operation capability and performs an essential role in ocean monitoring [2].SAR ship detection is a fundamental maritime task with important applications in offshore traffic management,fisheries management,and offshore emergency rescue[3,4].Therefore,more and more scholars have devoted their attention to this topic[5-16].

    Previous standard SAR ship detection methods require manual design of ship features,such as constant false alarm rate (CFAR) [1],saliency [2],superpixel [3],and transformation [4].However,these traditional methods usually have complex algorithms,limited migration capability,and require tedious manual design,which limit their application in practical scenarios.In addition,these methods generally use a limited number of ship images for theoretical analysis to define ship features.Moreover,they cannot capture the ship features of diverse sizes in various backgrounds[5].Therefore,their multiscale and multi-scene detection performance is usually poor.

    In recent years,with the development of deep learning and the rise of convolutional neural networks(CNNs),the current state-of-the-art CNN-based SAR ship detectors have addressed some of these problems to some extent[17].However,recent research has shown that the size of the actual receptive field in convolutional neural networks is much smaller than the theoretical receptive field,which means that CNNs may not be able to make full use of the contextual information of the input data.In addition,the feature capture capability of CNNs is also limited,and they cannot extract global representation features well.To enhance the global capture capability of CNNs,deeper convolutional layers can be stacked,but this results in models that are too large,with too many learning parameters to train and converge effectively,which causes a dramatic increase in computation and reducing the timeliness of the models.Therefore,there is a requirement to explore alternative model structures and techniques to address these issues.

    The Transformer structure[18]proposed by Google has been gradually applied to computer vision tasks after achieving good results in natural language processing.In 2020,Vision Transformer(ViT)[19]became the first Transformer structure adopted in computer vision,and it achieved state-of-the-art results in optical natural scene recognition,demonstrating the feasibility of Transformers in this field.Subsequently,other Transformer-based network models have emerged,such as Detection Transformer(DETR)[20]and Swin Transformer[21].

    Swin Transformer is a novel Transformer architecture used for computer vision tasks such as image classification,object detection,and segmentation [22].It employs a hierarchical attention mechanism by dividing the input image into fixed-sized patches,allowing each patch to capture global information effectively.This approach helps extract more contextual and global features.Compared to traditional Transformer networks,Swin Transformer introduces a local window mechanism,transforming global attention computation into local window-based computation,significantly reducing computational complexity.This enables Swin Transformer to operate more efficiently when dealing with large-scale images while reducing storage requirements.Therefore,the main reason for choosing Swin Transformer is its ability to extract rich contextual features while reducing computational overhead compared to other Transformer networks.

    To better handle the extracted features from Swin Transformer,this paper proposes an innovative cross-level fusion neck module called PAFF,which offers several advantages including multi-scale receptive fields,feature fusion,enhanced contextual information,and end-to-end training.PAFF constructs a feature pyramid to obtain rich multi-scale features,facilitating the detection of objects at different scales.Additionally,PAFF can fuse features from different levels,integrating information from both low-level and high-level features to improve object detection performance.It also propagates contextual information through an upsampling path,enhancing the semantic representation of features.Importantly,PAFF supports end-to-end training,enabling the entire network to learn collaboratively and enhance model performance.

    This paper focuses on optimizing the design of the backbone and neck parts of the target detection framework,using YOLOX [23] as the baseline.The integration of the Transformer model and the single-stage model in our research presents a novel approach that enhances the existing methodology for ship detection in Synthetic Aperture Radar (SAR) imagery.This innovative combination offers significant contributions to the field,bringing about advancements in the current SAR ship detection framework,and its main contributions include the following:

    1.In order to solve the existing problem of SAR target detection which lacks weak contextual information of the sensed images,we propose an end-to-end SAR target detection network with the Transformer structure as the backbone.

    2.Combining strategies such as the SDP module and the self-attentive mechanism,a Swin-CC backbone model is proposed to improve the ability of the Swin Transformer to acquire contextual features and cross information.

    3.A cross-layer fusion neck module PAFF is designed,which can better cope with multi-scale changes and complex practical situations.The module can effectively fuse features from different layers to improve the performance and capability of the model,and it also improves the model’s ability to detect and recognise objects at different scales.

    The remainder of the paper is organised as follows:Section 2 of this paper presents an expanded elucidation of the primary framework employed herein.The subsequent Section 3 delineates the core methodology employed in this study,as well as the evaluation metrics utilized to gauge the experimental outcomes.Section 4 provides comprehensive details regarding the experimental configurations,presenting the results of the conducted ablation experiments and a substantial number of comparative experiments.Finally,Section 5 serves as the conclusive section of this paper,summarizing the overall findings and presenting prospects for advancing the research.

    2 Related Work

    2.1 YOLOX

    The SAR ship detection problem poses challenges due to the unclear edge information of targets and the limitations of traditional anchor-based methods.In order to overcome these shortcomings and leverage the benefits of anchor-free detection frameworks,we draw inspiration from the latest anchor-free model called Exceeding YOLO Series in 2021(YOLOX)[23],which represents a significant improvement over the YOLO series.YOLOX introduces an explicit definition of positive sample regions by projecting the 3×3 region of the ground truth bounding box onto the center of the feature graph.It predicts four values for each target,namely the offset distance of the upper left corner and the height and width of the bounding box.To handle fuzzy samples more effectively,YOLOX adopts the simOTA algorithm for positive and negative sample matching[23].The simOTA algorithm involves calculating the matching degree for each pair,selecting the top k prediction boxes with the smallest cost in a fixed central area,and marking the grids associated with these positive samples as positive.Given the sparsity,small sample characteristics,and target scattering in SAR images,YOLOX’s characteristics of a decoupling head,a new tag allocation strategy,and an anchor-free mechanism make it an ideal baseline detector with a trade-off between accuracy and speed.Therefore,we have chosen YOLOX as the benchmark network for our research.

    2.2 Swin Transformer

    In SAR images,small target ships may lose information during downsampling.The Swin Transformer infrastructure can address this issue with its large throughput and massively parallel processing capability.To extract features from an image,firstly,input image is first divided into tokens with high-resolution properties of the image.The Swin Block employs the Shift Window approach to limit the self-focused computation to non-overlapping partial windows,enabling cross-window connections and improving efficiency[22].

    The Swin Transformer Block is an elementary module within the Swin Transformer that extracts features and transfers information in the input image.It consists of two primary components: the Swin Self-Attention(MSA)and the Swin Transformer Feed Forward(FFN).The MSA captures the relative position encoding between elements within each window by computing global contextual information.It then performs a multi-headed self-attention calculation to enable each element to aggregate information from others.This attention calculation considers the relative position relationships between elements and uses scaling parameters to balance the importance between different parts of the attention calculation.The FFN comprises two linear layers and an activation function that perform a non-linear transformation on the elements within each window.The output of the FFN calculation is added to the output of the MSA and normalized by Layer Normalization to produce the final output of the Swin Transformer Block.In the Swin Transformer,the input and output of each Swin Transformer Block consist of a set of windows,rather than individual pixels or feature vectors.This approach enables the Swin Transformer to start with locally aggregated information and gradually expand to globally aggregated information,thus capturing the structural and contextual information of the input image better.Additionally,the Swin Transformer uses a window grouping strategy that allows multiple Swin blocks to share windows,enhancing the coherence of information flow while maintaining efficiency.

    The Swin Transformer adopts the concept of Windows Multi-Head Self-Attention(W-MSA)to partition the feature map into multiple non-overlapping regions,or windows,and performs Multi-Head Self-Attention within each window.This approach reduces the computational cost compared to applying Multi-Head Self-Attention(MSA)directly to the entire global feature map as in the Vision Transformer,particularly when the shallow feature map is large.

    3 Main Methods

    In this section,Swin Transformer is selected as the basic architecture of the backbone network,based on which the Contextual Intersection Information Extractor(CSC)module and the backbone Swin-CC network based on contextual cross-information fusion are designed.In addition,to better extract multi-scale SAR target features,an adaptive spatial feature fusion feature pyramid structure with an enhancement neck(PAFF)is designed.The general framework diagram is shown in Fig.1.

    3.1 General Architecture of Swin-PAFF

    This paper uses YOLOX as the basic architecture and introduces Swin Transformer as the backbone model of the network,and designs the feature extraction module CSC to fully capture contextual cross-information.In particular,the CSC module achieves more comprehensive and finer information interaction and better contextual information capture by cross-linking different blocks in the target feature map and can help the network to better capture the relationship between the target and its surroundings,thus improving the accuracy of target detection.The module has higher computational efficiency and better interpretability,can optimizes the utilization of contextual and cross-position information,can extract richer feature information,and improves the multi-scale SAR target multi-characterization and description capability.

    Figure 1:General framework of Swin-PAFF

    In this paper,a novel enhanced neck module called PAFF (Cross-Layer Augmented Features and Adaptive Spatial Fusion)is proposed.The module incorporates a feature pyramid network[24]to facilitate the transfer of robust semantic features from the top to the bottom layers.Employing top-down and bottom-up feature pyramid networks,it enhances multi-scale connectivity operations.Furthermore,the proposed module leverages an attention mechanism to generate a weight map of equal dimensions from various layer parameters.Each pixel in the weight map corresponds to a specific coordinate,which is then multiplied element-wise with the corresponding element in the feature map to generate a new feature map[25].This approach effectively captures spatial relationships within the feature map,leading to improved model accuracy.Moreover,unlike traditional attention mechanisms,explicit calculation of attention weights is unnecessary,thereby reducing computational costs.Finally,the spatial adaptive fusion module facilitates better adaptation of the model to target scale variations and positional changes,resulting in enhanced feature fusion.

    Swin-PAFF consists of three main components: Swin-CC,PAFF,and YOLOX-Head.First,features are extracted from the input image using the designed Swin-CC backbone network,and the obtained feature maps are fused at multiple scales using the PAFF module.The YOLOX-Head module receives the multi-scale feature maps and performs multi-scale feature fusion.

    3.2 Swin-CC Backbone Network Based on Contextual and Cross-Information Fusion

    For the characteristics of SAR targets such as strong scattering,sparse and multi-scale,a target detection backbone network based on context and cross information fusion,Swin-CC,is designed by combining the advantages of Transformer,which can extract richer contextual features and cross information and improve multi-scale SAR target characterization.First,this paper uses the Swin Transformer,a current state-of-the-art model in the field of target detection and instance segmentation,as the base backbone network.Next,inspired by CCNet [26] and CSP_Darknet [27],this paper extends the field of perception,depth,and resolution by introducing the CSC attention mechanism module in the Patchembed section,which enhances the feature-aware domain,strengthens the comparison between different windows on the feature map,and performs the ability to fuse contextual and cross-information.

    3.2.1 Context-Based and Cross-Information Fusion CSC Module

    In this study,we integrated the CSC attention mechanism module into the Patchembed module and restructured it.The structure of the network for extracting the CSC attention mechanism is depicted in Fig.2.By incorporating the CSC module into the Patchembed module,we expanded the field of perception and the depth of the network.As a result,we improved the resolution of the network and its overall performance.

    Figure 2:CSC module network structure

    The figure presented in this paper,Fig.3 illustrates that the CSC feature extraction module comprises two components: the CrissCross Attention (CCA) [26] mechanism module and the SDP module.The CrissCross Attention is a multi-headed attention mechanism that is designed to improve a model’s ability to model spatial and channel relationships between features.It operates by mapping input features into multiple query,key,and value vectors and then computing an attention score by taking the inner product of queries and keys to assign a weight to each location and channel.Then,the weighted value vectors are summed to produce a weighted feature representation.The advantage of CrissCross Attention is that it can effectively capture spatial relationships and crosschannel dependencies between features,thus improving the model performance while consuming less GPU memory.

    We have innovatively introduced the CSC operation to address the limitation of the CCA module,which can only capture contextual information both horizontally and vertically.The CSC operation first fuses the feature information extracted from the previous CCA module with the SDP operation information features,reducing the number of computed parameters,before adding another layer of CCA modules.In the first CCA attention module,the feature map H extracted from each patch’s feature picture serves as the input,producing an output feature map H′of the same shape.In the second CCA attention module,the input is H′that has undergone feature fusion in the SDP module,producing an output feature map H′′.Fig.3 illustrates this process.By fusing features from two CCA modules and one SDP module,we obtain full image contextual information from all pixel points and generate new features that are contextually dense and rich in information.

    Figure 3:Example diagram after two CCA propagations

    Overall,our CSC module compensates for the fact that cross-information cannot be obtained from all pixels with dense contextual information.Compared to the CCA module,the CSC module does not bring in additional parameters and can obtain improved performance with lower computational effort.So we introduce the CSC module to enhance the network’s ability to extract contextual features and fuse cross-information while reducing the computational effort.

    3.2.2 Multi-Scale Feature Extraction

    The SDP method is a novel approach to enhancing the receptive field of neurons and achieving multi-scale feature extraction by utilizing maximum pooling and depth separable convolution(DWconv)techniques.Through the integration of convolution kernels of varying sizes across different layers,the neural network is capable of capturing a broader range of features and improving its receptive field,resulting in improved target detection.Furthermore,the employment of multi-layer features helps to mitigate the impact of environmental factors and enhance the background data at varying scales,thus facilitating the detection of small targets.

    3.3 Feature Pyramid and Spatial Adaptive Fusion Module PAFF

    The architecture of the feature pyramid and spatial adaptive fusion module(PAFF)proposed in this paper consists of two main parts,PAFPN and ASFF[28],and its structural diagram is shown in Fig.4.

    PAFPN is a feature extraction and integration network that can detect targets of various sizes at different scales.It improves on the Feature Pyramid Network(FPN)by not only upsampling and fusing features between each level but also combining features between levels to enhance sampling and fusion.The path aggregation module in PAFPN captures the importance between different feature layers and fuses their features with the corresponding up/down-sampled features into a unified feature pyramid.This approach allows PAFPN to better handle targets of different sizes and improve detection performance for both large and small targets.

    Research on attention mechanisms has revealed that inter-channel attention significantly improves model performance.However,it often neglects inter-pixel location information.Therefore,this paper incorporates Coordinate Attention(CA)[29]operations into the feature map before up/downsampling by PAFPN.As shown in Fig.5,CA encodes the relationship between channel and remote location information.The overall structure consists of two steps: Coordinate Information Embedding and Coordinate Attention Generation.

    Figure 4:Structure of the PAFF model

    Figure 5:CA model structure

    Modeling the relationships between channels with ordinary convolution is difficult.Explicitly constructing interdependencies between channels can enhance the model sensitivity to informative channels,thereby facilitating the final classification decision.Furthermore,global average pooling can help the model capture global information,which is not available through convolution.The compression step for the cth channel given an input X can be expressed as shown in Eq.(1).

    Feature pyramids are a commonly used technique in object detection but suffer from inconsistency between features at different scales,particularly for first-level detectors.To address this issue,we propose the Attentional Spatial Fusion Network (ASFF) feature fusion strategy to improve target detection performance within the feature pyramid in single-level detectors.The ASFF network module filters and fuses feature maps from different levels to retain valuable information,adaptively blending features from different levels at each spatial location.From Fig.4,we show that fusion vectors are a weighted combination of vectors from the first three feature maps,with network-learned coefficients shared across all channels.We set the feature maps to different levels(l ∈0,1,2)based on the input mapping’s dimensionality,with the corresponding feature maps referred to as Xl.In this paper,we set level l(l ∈0,1,2)to 192,384,and 768,respectively,depending on the input size and the characteristics of the SAR ship target.The output of the ASFF module is defined as follows:

    wherexn→ldenotes the feature vector after adjusting the features at level n to level l.α,β and γ are the learning rates at level n,level n+1 and level n+2,respectively,defined as follows:

    3.4 Loss Function

    Swin-PAFF is derived from the YOLOX model,with improvements and optimizations made to the original model.As a result,the loss function used in Swin-PAFF is similar to that of the YOLOX model.Below are the formulas for the operations ofLcls,Lreg,Lobj,and Loss,respectively:

    To calculate theLclsclassification loss,we utilize the Varifocal loss.This loss function allows for more flexible control of the model’s attention on samples of different difficulty levels by adjusting the α and β parameters accordingly.The formula for the classification loss is as follows:

    For theLobjloss function,we use the binary cross-entropy loss function(BCELoss).It is calculated as follows:

    For the final 15 rounds,we disabled the data augmentation mode and employed the SmoothL1Loss function on the unencoded predictions of the positive sample bounding boxes and their corresponding ground truth bounding boxes.The calculation of the SmoothL1Loss is presented in Eq.(12).

    4 Experimental Results

    The present section presents the experimental evaluation of our proposed detection method.The HRSID dataset is utilized as the experimental data,and the model’s effectiveness is validated using data from the SSDD [29] and a subset of the HRSID [30] dataset.Subsequently,we assess the effectiveness of the Swin-CC attention mechanism module and the PAFF module on the model by performing ablation experiments,as described in the following section.Finally,we compare our approach with other existing methods to validate its effectiveness.The experimental environment includes the MMdetection[31]based framework,NVIDIA RTX3090 GPU with 24 GB video memory,and Ubuntu 18.04 operating system.The AdamW optimizer is used with a batch size of 8 and the model epoch set to 100.

    4.1 Calculation Methods of Evaluation Metrics

    This paper utilizes the COCO evaluation metric,which employs average precision (AP) as the primary metric.AP is calculated by determining the mean precision at ten intersections with an IoU threshold that ranges from 0.50 to 0.95 in intervals of 0.05.AP50refers to the AP with an IoU threshold of 0.50,while AP75represents the AP with an IoU threshold of 0.75.In addition,APSand APMdenote the AP for small and medium ships,respectively.The evaluation metric uses true positives(TP),false positives(FP),and false negatives(FN)to determine the number of samples in each category.Recall(R),precision(P),and mean average precision(MAP)are defined as follows:

    4.2 Ablation Experiments

    This paper presents a series of ablation experiments to verify the efficacy of the Swin-CC and PAFF modules proposed in this study.The experiments use the HRSID dataset published by Wei et al.[30]in 2020,which comprises 5604 images of SAR ships with resolutions ranging from 0.1 to 3 m and featuring HH,HV,and VV polarisations.

    Our selection of YOLOX as the current baseline is supported by the data presented in Table 1.We conducted training on the HRSID dataset using several widely used target detection models,including YOLOF [32],YOLOv3 [33],YOLOv7 [34],YOLOv8 [35],and YOLOX [23].Upon comparing their performance,we observed that YOLOX achieved a slightly lower AP50value than YOLOv8,but with a smaller number of parameters.This finding indicates that YOLOX can deliver satisfactory results while requiring relatively fewer computational resources.As a result,we have chosen YOLOX as our benchmark model for further analysis.

    Table 1: Comparison with the state-of-the-art detectors on HRSID dataset

    The benchmark model is augmented with the Swin-CC and PAFF modules,and their impact on the model’s performance is evaluated.The experimental results presented in Table 2 indicate that the Swin-CC module significantly improves detection accuracy.That is because the Swin Transformer’s remote modeling capability and the CSC module’s powerful feature extraction capability enable the model to capture target object features more effectively.Furthermore,the PAFF module is used as a feature enhancement extraction tool,and the results demonstrate that incorporating the PAFF module improves the model’s detection accuracy by 1%.Most notably,the combined use of the Swin-CC and PAFF modules leads to an exponential increase of approximately 8% in feature extraction performance compared to the baseline model.These experimental results establish the proposed method’s scalability,applicability,and efficacy in enhancing detection performance.

    She had run to her little room and had quickly taken off her dress, made her face and hands black, put on the fur cloak, and was once more the Many-furred Creature

    Table 2: Ablation experiments

    Table 2 illustrates that the Swin-CC module,serving as the backbone network module,has superior feature extraction capabilities due to its combination of global-local and spatial-location characteristics.Furthermore,the use of the Swin-CC module improves the detection performance of small,medium,and large ships compared to the baseline YOLOX.Specifically,the APSand APMfor small and medium-sized vessels increased by 1.1% and 1.3%,respectively,while AP,AP50,and AP75increased by 1.6%,0.7%,and 2.6%,respectively.In addition,the PAFF module improves inter-feature fusion and adaptive spatial fusion compared to the baseline YOLOX,resulting in improvements of 0.9%,1.3%,and 1.2%for AP,AP50,and AP75,and 0.7%and 1.1%for APSand APM.When combined,the complete Swin-PAFF yields significant improvements of 7.9%,7.8%,9.3%,6.6%,and 14.4% in each metric compared to the baseline.These findings suggest that both the Swin-CC module and the PAFF module of Swin-PAFF improve model detection performance and their combination results in a more robust model.

    4.3 Comparative Experiments

    To substantiate the effectiveness of the proposed approach,this paper conducts a comparative analysis with state-of-the-art SAR ship detection methods,utilizing HRSID dataset as presented in Table 3.The Swin-PAFF model is a hybrid model combining a single-stage model and a transformer model.Hence,we selected approaches primarily based on single-stage and transformer models for comparison,including YOLOF [32],YOLOv3 [33],YOLOv7 [34],YOLOv8 [35],RetinaNet [36],SSD[37],FCOS[38],TOOD[39],VFNet[40],PVT[41],and deformable DETR[42].Furthermore,we evaluated the performance of our model against classical one-stage models using the LS-SSDDv1.0 dataset,with the results reported in Table 4.Additionally,to assess the model’s generalization capability,we directly applied the weights of the aforementioned models to predict the accuracy of the SSDD dataset.We conducted predictions for the test set,training set,nearshore hull,and offshore hull,and the outcomes are presented in Tables 5-8.The Swin-PAFF model exhibited promising results across all scenarios,highlighting its robustness and superior performance.

    Table 3: Results of ten network models with the HRSID dataset

    Table 4: Results of ten network models with the LS-SSDD-v1.0 dataset

    Table 5: Results of ten network models under the train dataset in SSDD

    Table 6: Results of ten network models under val dataset in SSDD

    Table 7: Results of the ten network models under the inshore dataset in SSDD

    Table 8: Results of the ten network models with the offshore dataset in SSDD

    Tables 4 and 5 depict that the proposed Swin-PAFF approach outperforms other models in detecting ships on the HRSID dataset and the LS-SSDD-v1.0 dataset.Additionally,Tables 3-6 exhibit that the Swin-PAFF framework not only achieves outstanding results on the HRSID dataset but also showcases strong generalization capabilities on the SSDD dataset,especially in complex nearshore scenes.The Swin-CC’s ability to extract contextual cross-information and the PAFF enhancement neck’s spatially adaptive fusion capability are the two main factors behind this remarkable performance,which improves the feature extraction performance of the Swin-PAFF framework and enhances the network’s robustness.

    The analysis presented in Table 4 reveals that our model exhibits a higher number of parameters in comparison to other YOLO models.While the fusion of YOLO and Swin Transformer for static target detection is a novel approach,the significant increase in the number of parameters after incorporating the Swin Transformer raises concerns regarding the practicality of real-time detection and tracking tasks.It is imperative to acknowledge this issue as a crucial area that requires further development in the future.

    Moreover,Fig.6 presents that the Swin-PAFF model possesses powerful generalization capabilities and surpasses other models in detecting near-shore ships without relying on the SSDD dataset for training.Furthermore,Fig.7 demonstrates that the Swin-PAFF model exhibits exceptional performance in detecting small nearshore objects on the HRSID dataset,further highlighting its exceptional generalization capabilities.

    Figure 6:Detection nearshore in the SSDD dataset.(a)Green markers indicate true boxes.(b)Orange markers indicate targets with false detections(i.e.,false positives,FP).(c)Blue markers indicate missed detections(i.e.,false negatives,FN)

    Figure 7: HRSID dataset detection Swin-PAFF module (a) Green markers indicate true boxes.(b)Orange markers indicate targets with false detections(i.e.,false positives,FP).(c)Blue markers indicate missed detections(i.e.,false negatives,FN)

    5 Conclusion

    In this paper,we present Swin-PAFF,a Transformer network designed for SAR ship detection,which incorporates contextual cross-information fusion and adaptive spatial feature fusion.Swin-PAFF builds on the YOLOX,an advanced single-stage target detection algorithm,and addresses challenges related to complex backgrounds,large scale differences,and dense distribution of small targets in SAR ship detection.

    In order to enrich the extraction of comprehensive contextual intersection information,we have developed a dedicated feature extraction module,referred to as the Contextual Intersection Information Extractor (CSC),and seamlessly integrated it with the Swin Transformer architecture.Furthermore,we propose a novel technique,named Adaptive Spatial Feature Fusion (PAFF),for enhancing the feature pyramid structure,thereby bolstering the efficacy of feature extraction.Our methodology was rigorously evaluated on the HRSID dataset,yielding an accuracy improvement of approximately 8%.Additionally,Swin-PAFF exhibits exceptional performance as a single-stage network,thereby underscoring its robustness and practical utility.Nonetheless,the incorporation of the Transformer architecture significantly escalates the computational requirements of our model,necessitating the exploration of strategies to alleviate this computational burden.

    Although Swin-PAFF combines Transformer and CNN detection methods,we intend to explore ways to reduce the computational complexity associated with the Transformer model by integrating it with blockchain in future research [43,44].In addition,we plan to explore ways to make the Transformer model lighter.

    Acknowledgement:None.

    Funding Statement:This research is supported by the National Key Research and Development Program of China under Grant 2021YFC2801001,the Natural Science Foundation of Shanghai under Grant 21ZR1426500,and the 2022 Graduate Top Innovative Talents Training Program at Shanghai Maritime University under Grant 2022YBR004.

    Author Contributions:Conceptualization:Y.Z.;methodology:Y.Z.,D.H.;software:Y.Z.,P.C.;validation:Y.Z.,D.H.and P.C.;formal analysis:Y.Z.,P.C.;investigation:Y.Z.;resources:Y.Z.;data curation:Y.Z.;writing and original draft preparation:Y.Z.,D.H.and P.C.;writing—review and editing:Y.Z.,D.H.and P.C.;visualization: Y.Z.;supervision: D.H.,P.C.;project administration: D.H.;funding acquisition:D.H.,P.C.All authors have read and agreed to the published version of the manuscript.

    Availability of Data and Materials:The data presented in this study are available on request from the corresponding author.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    午夜精品久久久久久毛片777| 在线观看舔阴道视频| 一级黄色大片毛片| 久热这里只有精品99| 免费观看人在逋| 欧美日韩一级在线毛片| 99精品在免费线老司机午夜| 侵犯人妻中文字幕一二三四区| 亚洲第一av免费看| 满18在线观看网站| 一卡2卡三卡四卡精品乱码亚洲| 日日摸夜夜添夜夜添小说| 色婷婷久久久亚洲欧美| 在线观看一区二区三区| a级毛片在线看网站| 老司机午夜福利在线观看视频| 亚洲 欧美一区二区三区| 亚洲国产高清在线一区二区三 | 操出白浆在线播放| 怎么达到女性高潮| 99久久久亚洲精品蜜臀av| 国产私拍福利视频在线观看| 桃红色精品国产亚洲av| 亚洲欧洲精品一区二区精品久久久| 岛国在线观看网站| 久久狼人影院| 最近最新免费中文字幕在线| 精品久久久久久久久久久久久 | 久久中文字幕一级| 久久这里只有精品19| 99精品在免费线老司机午夜| 国产精品久久久久久亚洲av鲁大| 99国产综合亚洲精品| 亚洲精华国产精华精| 国产亚洲欧美精品永久| 亚洲av五月六月丁香网| 男女下面进入的视频免费午夜 | 欧美日本视频| 亚洲国产精品sss在线观看| 久久久精品欧美日韩精品| 日韩有码中文字幕| 国产亚洲精品av在线| 亚洲 国产 在线| 19禁男女啪啪无遮挡网站| 后天国语完整版免费观看| 久久国产精品人妻蜜桃| 国产精品亚洲一级av第二区| 一进一出好大好爽视频| 久久性视频一级片| 精品国内亚洲2022精品成人| 此物有八面人人有两片| 免费高清在线观看日韩| 国产亚洲欧美98| 9191精品国产免费久久| 村上凉子中文字幕在线| av在线天堂中文字幕| 非洲黑人性xxxx精品又粗又长| 日日摸夜夜添夜夜添小说| 国产爱豆传媒在线观看 | 亚洲人成电影免费在线| 中文资源天堂在线| 亚洲国产看品久久| 一本久久中文字幕| 久久国产精品人妻蜜桃| 12—13女人毛片做爰片一| 亚洲精品美女久久久久99蜜臀| 黄片大片在线免费观看| 亚洲 欧美一区二区三区| 一进一出抽搐gif免费好疼| 长腿黑丝高跟| av福利片在线| 精品国产国语对白av| 最好的美女福利视频网| 亚洲国产精品成人综合色| 亚洲国产精品久久男人天堂| 久久久久久免费高清国产稀缺| 亚洲男人天堂网一区| 久久欧美精品欧美久久欧美| 琪琪午夜伦伦电影理论片6080| 无人区码免费观看不卡| 亚洲国产欧美网| 亚洲av成人不卡在线观看播放网| 黄色视频,在线免费观看| e午夜精品久久久久久久| 欧美乱码精品一区二区三区| 亚洲国产精品999在线| 69人妻影院| 久久99热6这里只有精品| 真实男女啪啪啪动态图| 国产精品一区二区免费欧美| 性色avwww在线观看| 麻豆乱淫一区二区| av在线播放精品| 女人被狂操c到高潮| 精品国内亚洲2022精品成人| 国产伦一二天堂av在线观看| 国产老妇女一区| 男人舔奶头视频| 18禁黄网站禁片免费观看直播| 香蕉av资源在线| 美女xxoo啪啪120秒动态图| 嫩草影院入口| 国产麻豆成人av免费视频| 亚洲一区高清亚洲精品| 亚洲天堂国产精品一区在线| 看黄色毛片网站| av福利片在线观看| 男女啪啪激烈高潮av片| 少妇人妻一区二区三区视频| 99久久精品热视频| 3wmmmm亚洲av在线观看| 日韩欧美一区二区三区在线观看| 综合色av麻豆| av卡一久久| 亚洲精品久久国产高清桃花| 亚洲国产色片| 欧美丝袜亚洲另类| 国产精品久久电影中文字幕| 看片在线看免费视频| 国产高清不卡午夜福利| 亚洲精品亚洲一区二区| 热99在线观看视频| 国国产精品蜜臀av免费| av卡一久久| 韩国av在线不卡| 十八禁国产超污无遮挡网站| 国产伦一二天堂av在线观看| 少妇熟女aⅴ在线视频| 亚洲欧美精品自产自拍| av国产免费在线观看| 亚洲综合色惰| 舔av片在线| 亚洲欧美清纯卡通| 亚洲国产精品国产精品| 亚洲无线观看免费| 亚洲成人久久性| 国产高清激情床上av| 成人漫画全彩无遮挡| av女优亚洲男人天堂| 一卡2卡三卡四卡精品乱码亚洲| 免费一级毛片在线播放高清视频| 少妇高潮的动态图| 22中文网久久字幕| 我的女老师完整版在线观看| 我要搜黄色片| 久99久视频精品免费| 女的被弄到高潮叫床怎么办| 真实男女啪啪啪动态图| 亚洲av五月六月丁香网| 最好的美女福利视频网| 日韩三级伦理在线观看| 有码 亚洲区| 色播亚洲综合网| 国产中年淑女户外野战色| 丝袜喷水一区| 成人欧美大片| 亚洲国产高清在线一区二区三| 男女下面进入的视频免费午夜| 天堂√8在线中文| 天天躁日日操中文字幕| 中文字幕精品亚洲无线码一区| 男女那种视频在线观看| 成人高潮视频无遮挡免费网站| 嫩草影视91久久| 亚洲精品色激情综合| 国模一区二区三区四区视频| 校园春色视频在线观看| 搞女人的毛片| 亚洲精品在线观看二区| 国产国拍精品亚洲av在线观看| 国产成人福利小说| 床上黄色一级片| 久久久久九九精品影院| 亚洲一区高清亚洲精品| 如何舔出高潮| 欧美最黄视频在线播放免费| 看片在线看免费视频| 成年女人看的毛片在线观看| 久久亚洲精品不卡| 亚洲av一区综合| 三级毛片av免费| 亚洲三级黄色毛片| 亚洲精品一卡2卡三卡4卡5卡| 少妇被粗大猛烈的视频| 我要看日韩黄色一级片| 欧美xxxx性猛交bbbb| 日本一本二区三区精品| 亚洲最大成人中文| 日本与韩国留学比较| 久久精品综合一区二区三区| 亚洲四区av| 久久精品国产鲁丝片午夜精品| 国内精品一区二区在线观看| 国产精品久久电影中文字幕| 国内久久婷婷六月综合欲色啪| 国内精品一区二区在线观看| 一级a爱片免费观看的视频| 国产精品日韩av在线免费观看| 俺也久久电影网| 国产乱人偷精品视频| 天天一区二区日本电影三级| 日本成人三级电影网站| 少妇丰满av| 国产高清有码在线观看视频| 美女高潮的动态| 亚洲欧美成人精品一区二区| 日韩欧美 国产精品| 99久久中文字幕三级久久日本| 成人漫画全彩无遮挡| 欧美xxxx性猛交bbbb| 精品久久久久久久人妻蜜臀av| 精品一区二区三区人妻视频| 亚洲综合色惰| 亚洲高清免费不卡视频| 内地一区二区视频在线| 人妻制服诱惑在线中文字幕| 欧美日韩在线观看h| 日韩欧美免费精品| 99热全是精品| 日韩一区二区视频免费看| 老熟妇仑乱视频hdxx| 亚洲国产欧洲综合997久久,| 一本一本综合久久| 国产高清三级在线| 国产极品精品免费视频能看的| 深夜精品福利| 在线观看午夜福利视频| 99久久成人亚洲精品观看| 两性午夜刺激爽爽歪歪视频在线观看| 人人妻,人人澡人人爽秒播| 亚洲精品粉嫩美女一区| 在线免费十八禁| 国产探花在线观看一区二区| 国产熟女欧美一区二区| 国内揄拍国产精品人妻在线| 中文字幕熟女人妻在线| 精品一区二区三区视频在线| 最近最新中文字幕大全电影3| 天堂网av新在线| 中文亚洲av片在线观看爽| 国产精品av视频在线免费观看| 91久久精品国产一区二区成人| 晚上一个人看的免费电影| 尾随美女入室| 一级a爱片免费观看的视频| 午夜亚洲福利在线播放| 99热6这里只有精品| 成人精品一区二区免费| 12—13女人毛片做爰片一| 国产av不卡久久| 在现免费观看毛片| 久久久a久久爽久久v久久| 亚洲电影在线观看av| 美女黄网站色视频| 卡戴珊不雅视频在线播放| 精品日产1卡2卡| 亚洲国产日韩欧美精品在线观看| 亚洲精品亚洲一区二区| 嫩草影院精品99| 久久久久性生活片| 亚洲无线观看免费| av中文乱码字幕在线| 黄色视频,在线免费观看| 午夜激情欧美在线| 人人妻人人看人人澡| 嫩草影视91久久| 露出奶头的视频| 日本撒尿小便嘘嘘汇集6| 日韩在线高清观看一区二区三区| 人人妻人人澡欧美一区二区| 欧美性猛交╳xxx乱大交人| 干丝袜人妻中文字幕| 两性午夜刺激爽爽歪歪视频在线观看| 国产伦精品一区二区三区视频9| 欧美成人a在线观看| 亚洲国产精品久久男人天堂| 久久精品91蜜桃| 变态另类丝袜制服| 99热全是精品| 免费看光身美女| 一a级毛片在线观看| 毛片一级片免费看久久久久| 欧美极品一区二区三区四区| 亚洲av五月六月丁香网| 亚洲av.av天堂| 一本一本综合久久| 精品国内亚洲2022精品成人| 国产免费一级a男人的天堂| 婷婷六月久久综合丁香| 国产一区二区在线av高清观看| 在线免费观看不下载黄p国产| 一级av片app| 国产在线男女| 亚洲乱码一区二区免费版| 六月丁香七月| 亚洲三级黄色毛片| 在线a可以看的网站| 亚州av有码| 在线免费十八禁| 尾随美女入室| 热99在线观看视频| 亚洲四区av| 亚洲精华国产精华液的使用体验 | 国产精品久久视频播放| 在线播放国产精品三级| 在线天堂最新版资源| 女生性感内裤真人,穿戴方法视频| 性欧美人与动物交配| 亚洲av二区三区四区| 99热只有精品国产| 韩国av在线不卡| 大香蕉久久网| 国产高清视频在线观看网站| 亚洲一区高清亚洲精品| 三级男女做爰猛烈吃奶摸视频| 亚洲国产欧美人成| 国产精品久久久久久久久免| 免费av不卡在线播放| 哪里可以看免费的av片| 又爽又黄无遮挡网站| 亚洲av五月六月丁香网| 亚洲熟妇中文字幕五十中出| aaaaa片日本免费| 成人永久免费在线观看视频| 婷婷精品国产亚洲av在线| 91久久精品国产一区二区成人| 亚洲在线自拍视频| av天堂在线播放| 日本免费a在线| 国产av一区在线观看免费| 永久网站在线| а√天堂www在线а√下载| 校园春色视频在线观看| 99久久久亚洲精品蜜臀av| 男女之事视频高清在线观看| 精品99又大又爽又粗少妇毛片| 身体一侧抽搐| 校园人妻丝袜中文字幕| 一本精品99久久精品77| 成人毛片a级毛片在线播放| 精品免费久久久久久久清纯| 天堂影院成人在线观看| 亚洲av中文av极速乱| а√天堂www在线а√下载| 欧美另类亚洲清纯唯美| 国产精品一区二区性色av| 综合色丁香网| 亚洲天堂国产精品一区在线| 国产精品99久久久久久久久| 久久婷婷人人爽人人干人人爱| 能在线免费观看的黄片| 国产伦一二天堂av在线观看| 亚洲一级一片aⅴ在线观看| 一进一出抽搐动态| 国产免费男女视频| 久久精品国产清高在天天线| 国产麻豆成人av免费视频| 日韩国内少妇激情av| 国产精品一区二区三区四区久久| 最近的中文字幕免费完整| 欧美+日韩+精品| 日日撸夜夜添| 韩国av在线不卡| 亚洲专区国产一区二区| 亚洲熟妇熟女久久| 国产精品一二三区在线看| 在线a可以看的网站| 中国国产av一级| 久久人妻av系列| 午夜影院日韩av| 毛片女人毛片| 搡老熟女国产l中国老女人| 久久99热这里只有精品18| 国产成人aa在线观看| 亚洲成人av在线免费| 俄罗斯特黄特色一大片| 最近最新中文字幕大全电影3| 美女免费视频网站| 亚洲av免费高清在线观看| 日韩精品中文字幕看吧| 国产一区二区激情短视频| 欧美高清性xxxxhd video| 天天躁夜夜躁狠狠久久av| 最后的刺客免费高清国语| 欧美激情国产日韩精品一区| 97超视频在线观看视频| 99久国产av精品| 麻豆久久精品国产亚洲av| 一区二区三区四区激情视频 | 日韩在线高清观看一区二区三区| 草草在线视频免费看| 亚洲人与动物交配视频| 免费无遮挡裸体视频| 国产亚洲91精品色在线| 晚上一个人看的免费电影| 寂寞人妻少妇视频99o| 女人十人毛片免费观看3o分钟| 久久久久久久亚洲中文字幕| 久久人人精品亚洲av| 插阴视频在线观看视频| 色5月婷婷丁香| 免费不卡的大黄色大毛片视频在线观看 | 不卡一级毛片| av在线播放精品| 男女那种视频在线观看| 日本色播在线视频| 国产亚洲av嫩草精品影院| 国产伦在线观看视频一区| 99在线人妻在线中文字幕| 国产av一区在线观看免费| 人人妻人人澡欧美一区二区| 久久九九热精品免费| 欧美绝顶高潮抽搐喷水| 免费高清视频大片| 欧洲精品卡2卡3卡4卡5卡区| 此物有八面人人有两片| 国内精品宾馆在线| 色综合色国产| 国产精品日韩av在线免费观看| 又黄又爽又刺激的免费视频.| 亚洲成人中文字幕在线播放| 国产黄色视频一区二区在线观看 | 蜜桃亚洲精品一区二区三区| 国产爱豆传媒在线观看| 久久午夜福利片| 欧美+日韩+精品| 久久久久久久午夜电影| 亚洲国产精品久久男人天堂| av天堂在线播放| 高清毛片免费观看视频网站| 欧美激情久久久久久爽电影| 色哟哟·www| 久久久久久久久久久丰满| 国产午夜精品久久久久久一区二区三区 | 久久6这里有精品| 国产精品不卡视频一区二区| 国产av不卡久久| 中文字幕人妻熟人妻熟丝袜美| 日韩欧美精品v在线| 成人性生交大片免费视频hd| 看黄色毛片网站| 国产不卡一卡二| 麻豆乱淫一区二区| 亚洲性久久影院| 国产一区二区在线观看日韩| 简卡轻食公司| 成人特级黄色片久久久久久久| 嫩草影院新地址| 在线播放国产精品三级| 你懂的网址亚洲精品在线观看 | 久久99热6这里只有精品| 免费一级毛片在线播放高清视频| 精品久久久久久久人妻蜜臀av| 国产精品国产高清国产av| 人人妻人人澡人人爽人人夜夜 | av天堂在线播放| 熟女电影av网| 十八禁网站免费在线| 韩国av在线不卡| 日日啪夜夜撸| 男插女下体视频免费在线播放| 人妻制服诱惑在线中文字幕| 欧美日韩综合久久久久久| 免费观看人在逋| 亚洲中文日韩欧美视频| 天堂√8在线中文| 人人妻人人看人人澡| 亚洲国产精品久久男人天堂| 12—13女人毛片做爰片一| 一边摸一边抽搐一进一小说| 亚洲av五月六月丁香网| 天美传媒精品一区二区| 中文字幕av在线有码专区| 久久人人爽人人爽人人片va| 国内久久婷婷六月综合欲色啪| 精品人妻视频免费看| 一级黄片播放器| 日本黄色视频三级网站网址| 欧美在线一区亚洲| 91在线观看av| 一级毛片电影观看 | 精品一区二区三区人妻视频| 亚洲国产精品成人综合色| 国产免费男女视频| 国产精品爽爽va在线观看网站| 青春草视频在线免费观看| 亚洲av.av天堂| 九九爱精品视频在线观看| 国产真实伦视频高清在线观看| 国产中年淑女户外野战色| 亚洲中文字幕一区二区三区有码在线看| 欧美xxxx黑人xx丫x性爽| 伊人久久精品亚洲午夜| 听说在线观看完整版免费高清| 看黄色毛片网站| 国产av不卡久久| 久久人妻av系列| 久久婷婷人人爽人人干人人爱| 一卡2卡三卡四卡精品乱码亚洲| 精品一区二区三区视频在线观看免费| 亚洲婷婷狠狠爱综合网| 国产亚洲精品av在线| 欧美性猛交黑人性爽| 成年av动漫网址| 18禁黄网站禁片免费观看直播| 美女 人体艺术 gogo| 97碰自拍视频| 久99久视频精品免费| 免费在线观看影片大全网站| 免费观看的影片在线观看| 小说图片视频综合网站| 久久午夜福利片| 欧美日韩综合久久久久久| 一区二区三区免费毛片| 三级经典国产精品| 亚洲aⅴ乱码一区二区在线播放| av在线亚洲专区| 舔av片在线| 中国美白少妇内射xxxbb| 亚洲最大成人av| 亚洲中文日韩欧美视频| 九九热线精品视视频播放| 国产精品人妻久久久久久| 色综合站精品国产| 精品久久久噜噜| 国产欧美日韩精品亚洲av| 联通29元200g的流量卡| 亚洲熟妇熟女久久| 天堂网av新在线| 亚洲精品色激情综合| 国产大屁股一区二区在线视频| 我的女老师完整版在线观看| 亚洲美女搞黄在线观看 | 国产视频内射| 亚洲美女黄片视频| 男人狂女人下面高潮的视频| 村上凉子中文字幕在线| 久久久久性生活片| 日本五十路高清| 国产亚洲欧美98| 少妇人妻一区二区三区视频| 亚洲av不卡在线观看| 国产 一区精品| 精品午夜福利视频在线观看一区| 久久精品久久久久久噜噜老黄 | 男女做爰动态图高潮gif福利片| 亚洲国产精品国产精品| 国产精品av视频在线免费观看| 精品午夜福利视频在线观看一区| 少妇被粗大猛烈的视频| 成年av动漫网址| 久久这里只有精品中国| 又爽又黄a免费视频| 在线免费观看的www视频| 丝袜喷水一区| 国产av不卡久久| 亚洲三级黄色毛片| 久久精品综合一区二区三区| 国产在线男女| 我要看日韩黄色一级片| 伦精品一区二区三区| 国产黄色小视频在线观看| 日韩国内少妇激情av| 久久6这里有精品| 亚洲经典国产精华液单| 国模一区二区三区四区视频| 成人特级黄色片久久久久久久| 老司机影院成人| 日韩欧美 国产精品| 99在线视频只有这里精品首页| 亚洲,欧美,日韩| 精品免费久久久久久久清纯| 国产伦在线观看视频一区| 日日干狠狠操夜夜爽| 亚洲色图av天堂| 午夜老司机福利剧场| 小说图片视频综合网站| 日本一二三区视频观看| 亚洲中文字幕日韩| 老司机影院成人| 此物有八面人人有两片| 狂野欧美白嫩少妇大欣赏| 村上凉子中文字幕在线| 成人亚洲精品av一区二区| 久久精品夜夜夜夜夜久久蜜豆| 国产又黄又爽又无遮挡在线| 亚洲国产欧洲综合997久久,| 日日摸夜夜添夜夜添小说| 我要看日韩黄色一级片| 日韩国内少妇激情av| 一边摸一边抽搐一进一小说| 免费观看在线日韩| 成人av在线播放网站| 人人妻人人澡欧美一区二区| 蜜桃亚洲精品一区二区三区| 久久亚洲精品不卡| 日本一本二区三区精品| 国产成人福利小说| 亚洲无线观看免费| 卡戴珊不雅视频在线播放| 亚洲国产高清在线一区二区三| 男人和女人高潮做爰伦理| 国产三级中文精品| 亚洲成a人片在线一区二区| 97人妻精品一区二区三区麻豆| 日本黄色片子视频| 久久热精品热| 国产精品伦人一区二区| 久久这里只有精品中国| 天堂网av新在线| 嫩草影院精品99| 亚洲精品一卡2卡三卡4卡5卡| 亚洲av免费高清在线观看| 国产黄色视频一区二区在线观看 | 色综合亚洲欧美另类图片|