• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Fine-Grained Classification of Remote Sensing Ship Images Based on Improved VAN

    2023-12-15 03:57:20GuoqingZhouLiangHuangandQiaoSun
    Computers Materials&Continua 2023年11期

    Guoqing Zhou,Liang Huang and Qiao Sun

    College of Electronic Engineering,Naval University of Engineering,Wuhan,430033,China

    ABSTRACT The remote sensing ships’fine-grained classification technology makes it possible to identify certain ship types in remote sensing images,and it has broad application prospects in civil and military fields.However,the current model does not examine the properties of ship targets in remote sensing images with mixed multi-granularity features and a complicated backdrop.There is still an opportunity for future enhancement of the classification impact.To solve the challenges brought by the above characteristics,this paper proposes a Metaformer and Residual fusion network based on Visual Attention Network(VAN-MR)for fine-grained classification tasks.For the complex background of remote sensing images,the VAN-MR model adopts the parallel structure of large kernel attention and spatial attention to enhance the model’s feature extraction ability of interest targets and improve the classification performance of remote sensing ship targets.For the problem of multi-grained feature mixing in remote sensing images,the VAN-MR model uses a Metaformer structure and a parallel network of residual modules to extract ship features.The parallel network has different depths,considering both high-level and lowlevel semantic information.The model achieves better classification performance in remote sensing ship images with multi-granularity mixing.Finally,the model achieves 88.73%and 94.56%accuracy on the public fine-grained ship collection-23(FGSC-23)and FGSCR-42 datasets,respectively,while the parameter size is only 53.47 M,the floating point operations is 9.9 G.The experimental results show that the classification effect of VAN-MR is superior to that of traditional CNNs model and visual model with Transformer structure under the same parameter quantity.

    KEYWORDS Fine-grained classification;metaformer;remote sensing;residual;ship image

    1 Introduction

    Remote sensing ship classification technology is widely used in maritime trade,military security,and other fields.The classification of remote sensing ship targets can be divided into coarse-grained classification and fine-grained classification.The coarse-grained classification takes the ship target as the fundamental category to participate in the overall classification.The coarse-grained classification focuses on the ship subclasses that fall under the same primary category,like Ford-class aircraft carriers,Arleigh Burke-class destroyers.Therefore,the fine-grained classification task requires that the model can extract more abundant and fine-grained features.

    Detailed analysis of the remote sensing ship image data shows it has two typical characteristics:mixed multi-grain size features and complicated backgrounds.First,the mixed multi-grain size features are caused by the minor intra-class differences and the enormous intra-class differences.There are tiny differences in the same type of ships.Take the destroyer,for example,Arleigh Burke-class destroyers,Autumn Moon class destroyers,and others.They have remarkable similarities in hull appearance,superstructure,and weapon configuration.However,fine-grained classification requires accurate classification of specific models,which is more complicated.Second,the background in remote sensing ship images is complex.For example,port ship targets can be mixed with complex shore structures or other ships and targets,interfering with the model classification results.Therefore,more attention should be given to the complex background and mixed granularity in ship target classification tasks to achieve good generality on natural ship image data.

    For the problem of complex backgrounds and diverse information granularity in ship images,this paper proposed a Metaformer [1] and Residual fusion network based on Visual Attention Network(VAN-MR) to solve these problems.Unlike Transformer [2],which uses block stacking to form a backbone network,VAN-MR proposes the structure of the network,which fused the Metaformer and residual structure shown in Fig.1a.With the fused novel structure,VAN-MR can retain adequate underlying semantic information while acquiring high-level semantic information.Using this structure resolves the issue with multi-granularity hybrid scale classification.In addition,unlike traditional CNN networks that focus on picture information indiscriminately,VAN-MR adopts an attention structure with large kernel attention and spatial attention in parallel to enhance the model’s attention to the region of interest.The architecture enhances the model’s capacity to extract features from complicated backgrounds.

    Figure 1:Metaformer and residuals structure.The backbone of the classical visual Transformer and visual Multilayer Perceptron(MLP)consists of the block shown in(b),on which the parallel residual feature extraction module(c)is designed in this paper,which together forms a new backbone for the remote sensing ship image classification(a)

    The main contributions of this paper are as follows:

    1.This paper analyzes two challenges faced by the fine-grained classification task of ship images:complex background and multi-granularity feature confusion.Meanwhile,the paper proposes a solution based on a new backbone network to solve the challenges traditional image classification networks face in the above feature problems.

    2.For remote sensing ship targets under complex backgrounds,this paper adopts the method of large kernel attention and space attention in parallel.In order to solve the problem of multi-granularity feature confounding in remote sensing ship images,a structure of network based on the fusion of Metaformer structure and residual structure is proposed.The new structure of the network has better image understanding ability and performs well in fine-grained classification tasks of remote sensing ships.

    3.Experiments have been carried out on the model proposed in this paper,which thoroughly verifies the model’s effectiveness.Compared with other methods,the framework has shown the most advanced performance on the publicly available remote sensing ship classification datasets FGSC-23[3]and FGSCR-42[4].

    This paper is structured as follows:Section 1 presents the definition of fine-grained classification tasks for remote sensing ships and the research direction of this paper.Section 2 introduces the background and motivation of this study.Section 3 presents the technical development related to the research in this paper.Section 4 introduces the overall structure of the model and details of improvements.In Section 5,experiments and analyses are presented.Ablation and comparison experiments are designed to verify the effect of the model.Section 6 discusses the content of adversarial attacks,and Section 7 summarizes the work of this paper.

    2 Background

    There are many excellent methods in the field of remote sensing ship image classification,but there are also many problems that need to be solved.This paper hopes to provide a solution by designing a new backbone network.This section first analyzes the effectiveness threats faced by research,then introduces the application and existing problems of CNN and Transformer in ship image classification technology,and finally puts forward the research motivation of this paper.

    2.1 Threat to a Validity

    As a threat to external validity,we note that our method is specific to the Python language and has not been tried to run in other language environments.In addition,the accuracy of labels can also affect the classification results.Although experienced experts label the data sets,errors inevitably occur.As a threat to internal validity,we note that the model may generate noise data in the data preprocessing stage.Secondly,the hardware device states that the model training depends on is fluctuating,resulting in subtle differences in the weight data generated during the training process,which affects the effect of model training.As a threat to conclusion validity,we note that this paper only selects four evaluation indicators commonly used in image classification tasks,reducing the randomness of the evaluation results.

    2.2 Background and Motivation

    The image classification technology based on deep learning is different from traditional methods[5,6],Contrary to traditional methods,it automatically extracts shallow and deep information using convolution operations without requiring artificial feature extraction criteria.In computer vision,the method based on CNN first achieves excellent performance beyond the traditional method.Alex-Net [7] adapted a convolutional neural network (CNN) to deep and comprehensive models in 2012.Subsequently,convolutional neural networks like VGG [8],Google-Net [9],ResNet [10],and others were proposed,gradually boosting their complexity and efficacy.Liu et al.[11]improved the inceptionV3 [12] network,and the network achieved better results in classifying obscured ship targets.Zhao et al.[13]proposed a feature balancing strategy and,at the same time,using an iterative interaction mechanism that could effectively enhance the model classification effect generation by generation.Zhang et al.[14]proposed the AMEFRN model and designed an attribute feature learning branch supervised by attribute information to enhance the model learning ability.Song et al.[15]used methods such as attention cropping and attention classification module to enhance the classification ability of the model.

    Although the rapid development of CNN series models has made CNN effectively applied in the field of fine-grained classification of remote sensing ships,most CNN models are designed for natural pictures,and their effects in remote sensing image classification tasks are limited due to the complex background information in remote sensing images.In addition,in fine-grained classification tasks,traditional CNN models use large depth feature extraction networks to extract features,which is easy to lose the target spatial location information,resulting in the loss of coarse-grained features and affecting the classification effect of the model.Therefore,the above series of CNN-based methods have room for improvement in accuracy and complexity.The attention mechanism adopted in this paper can help CNN model to extract more fine-grained features and enhance model performance.

    In recent years,the image classification network based on Transformer[2]structure has achieved better results than basic CNN-based methods.In 2020,Google proposed VIT [16] network and successfully applied the Transformer structure in computer vision.Swin-Transformer [17] further improves the performance of the Transformer structure in the CV field.Xu et al.[18] integrated the advantages of ResNet and VIT models through knowledge distillation (KD) to improve the performance of VIT in remote sensing image scene classification.Huang et al.[19] applied Swin to the fine-grained classification task of remote sensing ships for the first time and designed a feature extraction backbone network integrating CNN and Transformer structure.The network can effectively extract ship targets’coarse-grained and fine-grained features and perform well in the available remote sensing ship classification dataset.

    Although the classification network based on the Transformer structure has achieved a classification accuracy surpassing that of the CNN network,its structural complexity results in significant computational overhead,leading to decreased model inference speed in ship scenes.Furthermore,the existing Transformer-based methods fail to adequately capture spatial location information.Given the challenge posed by coarse and fine granularity in ship images,there is scope for enhancing the performance of Transformer-based classification networks.The adoption of a parallel residual structure in this study enables the incorporation of crucial semantic information such as spatial location,thereby improving model performance in fine-grained classification tasks involving remote sensing ship images.

    Since the success of the Transformer structure in the field of computer vision[20],many scholars have explored the reasons why Transformer can be effective.Many believe the attention-based Token Mixer contributes significantly to the model’s performance [21].Therefore,some scholars have proposed improving the attention module to enhance the model’s performance,providing many ideas for improving the attention mechanism[22,23].However,Lee-Thorp et al.[24]found that the model could achieve similar results even when using the Fourier transform instead of a complex attention network.Further,Yu et al.[1] argued that VIT and other improved models are effective because of the use of a generic structure Metaformer.The authors proposed the PoolFormer [1] network based on a generic architecture using a simple pooling operation instead of an attention network,and experimental results showed that the new model achieved results beyond Swin.The comparison of the different classification networks is shown in Table 1.

    Table 1: Comparison of classification models based on different architectures

    With fewer parameters and a straightforward structure,the pure CNN network built on Metaformer architecture has the potential to outperform the Transformer network.It has been successful in several tasks in the vision field.However,it is rarely applied to the remote sensing field.We hope to apply the Metaformer structure to the fine-grained classification task of remote sensing ships and,at the same time,propose an improvement scheme to address the problems of insufficient feature extraction ability and failure to consider the multi-granularity feature mixing in the traditional network,to adapt to the fine-grained classification task of remote sensing ships.

    3 Literature Review/Related Work

    In the field of computer vision,many scholars have done a lot of research work,which has accelerated the development of image classification,object detection and other technologies.In these works,deep learning method and remote sensing ship image classification technology have a lot of relevance to this research.

    3.1 Deep Learning Methods

    Deep learning is a deep machine learning model.Deep learning methods bring significant progress to computer vision and machine learning.At present,deep learning has been applied in many fields.Chakravarthy et al.[25] applied the decision tree to web mining to find the web patterns that users are interested in more accurately.Arunkumar et al.[26] adopted support vector machine and other machine learning methods to improve the accuracy of emotional evaluation of medical videos.Devi et al.[27] used stacked convolutional neural network to analyze the authors of WHATSAPP accurately.Raja et al.[28]proposed a new conditional generation network C-GAN,which improved the recognition accuracy of autism spectrum disorders.

    In recent years,pure CNN networks based on Metaformer architecture have achieved rapid development in computer vision.They have outperformed both traditional CNN-based networks and Transformer-based networks on public datasets.Touvron et al.proposed a ResMLP[29]model based on Metaformer architecture,which did not adopt any self-attention module,but only contained a linear layer and GeLU [30].It has achieved competitive results on the public data set ImageNet-1k.Guo et al.proposed a VAN[31]network based on the Meta-Former architecture,using a large kernel attention mechanism LKA instead of a complex self-attentive mechanism,which significantly reduces the number of parameters of the model,and experiments show that the VAN network achieves beyond Swin in down-stream tasks such as classification and detection in vision domain.Guo et al.proposed the SegNeXt[32]network for segmentation task,which achieved good performance on ADE20K[33],Cityscapes[34],and COCO-Stuff[35]datasets by using multi-branch large kernel attention instead of self-attention networks.

    3.2 Fine-Grained Ship Classification

    There are two essential datasets in the fine-grained classification of remote sensing ships.The first is FGSC-23[3],proposed by Yao et al.,and the second is FGSCR-42[4],proposed by Di et al.These publicly available fine-grained classification datasets of remote sensing ships provide data support for the research work of other scholars.

    Facing the challenges of fine-grained classification,many scholars have achieved better application results by improving classification networks.Chen et al.[36] used gradient-weighted class activation feature maps,which can effectively locate regions in the pictures where ship features are concentrated.In addition,they mirrored self-calibrated convolution in the network and expanded the model horizon by collecting richer contextual information in this way.Then,to enhance the model’s capacity to extract global features,Zhang et al.[37] replaced some branches of the BCNN model based on the BCNN with the inception module and added a new activation function,AM-Softmax,to minimize inter-class differences while increasing intra class similarities.A memory network with an adaptive selection learning strategy was proposed by Xu et al.[38]after they thoroughly analyzed the fine-grained classification problem under the sample imbalance condition.This memory network can extract compelling features from the imbalanced ship dataset and improve classification performance.Chen et al.[39] proposed P2Net based on contrast learning,which uses a two-branch network and an image aggregation module to achieve good fine-grained classification performance.Li et al.[40]proposed a foreground aware Feature Pyramid Network(FPN)network for the small sample problem.They proposed two methods for computing foreground weights to effectively solve the fine-grained classification problem of ships in the small sample case.

    The above model proposes solutions to the problems in remote sensing image classification tasks.In this paper,based on the above models,this paper continue exploring the fine-grained classification problem in the case of complex backgrounds and multi-granularity feature mixing and propose a new feature extraction network to solve such problems better.

    4 Proposed Work

    In the fine-grained classification task,the model have to pay attention to the target’s minute features and overall features at the coarse-grained level.Moreover,the remote sensing ship images have mixed multi-grain size features and complex backgrounds.The traditional backbone network improves the model’s ability to extract high-level semantic information by deepening the network.However,it needs to gain the underlying semantic information,leading to poor results in the case of complex background of remote sensing images.Based on the VAN model,the paper proposes the framework design of the parallelized residual network and improved attention.The residual structure can retain the underlying semantic information of the image,and the improved attention structure can better focus on and extract fine-grained features.The overall structure of the model is shown in Fig.2.

    Figure 2:VAN-MR model architecture

    The backbone network of the model is composed of two parallel feature extraction branches,namely the Metaformer branch and the residual branch.Metaformer branch learns from the classic architecture of the VAN network and is repeatedly stacked by the same blocks.The 2D convolution method is adopted in the patch embedding to downsample the feature map.The four patch embedding convolution steps are 4,2,2,2,and the size of the particular map after downsampling is H/4×W/4,H/8×W/8,H/16×W/16.H/32×W/32,the subsampled feature map is input into the corresponding stage,and the blocks contained in each stage are {3,5,27,3}.The output feature map size of each stage is H/4×W/4,H/8×W/8,H/16×W/16,and H/32×W/32,respectively,which is the same as the input.The residual branch adopts residual connection mode,which is composed of a convolution layer,normalization layer,and activation function.The input of the first residual structure is the output feature map of stage 1,and the input of other residual structures is obtained by adding the output feature map of the last residual structure and stage.After each residual,the size of the feature map is halved,which is consistent with the result generated by the patch embedding structure.At the end of the feature extraction network,the feature graphs generated by the two branches are added and output to the classification network to get the final result.

    4.1 Metaformer Branch

    The Metaformer branch of the model consists of a patch embedding and a block with the same structure.Patch embedding performs a downsampling operation on the feature map and outputs it to the block.The block structure is shown in Fig.1b,where the input features pass through two residual serial blocks.The first residual block consists of Batch Normalization(BN)layer,an attention mechanism,and the second residual block consists of a BN layer and channel MLP.

    The VAN model uses the large kernel attention (LKA) instead of the self-attention module.In neural networks,a large convolution kernel will have a larger effective receptive field,which is conducive to obtaining remote global information.Still,a large convolution kernel will increase the computational load,and a small convolution kernel will help get local information,but the effective receptive field needs to be more significant.Although self-attention can capture remote dependencies,the number of parameters is vast.LKA divided large kernel convolution into deep convolution(DW-Conv),profound dilation(DW-D-Conv)convolution,and Pointwise Conv,and comprehensively considered local context information and remote relations while maintaining a small number of parameters.As a result,the VAN network is thriving in natural image classification tasks.However,since remote sensing ship images have characteristics such as complex backgrounds (described in Section 2.2).Therefore,the attention module needs to locate the target of interest more accurately in the feature space for the classification model to accomplish the fine-grained classification task of remote sensing ships.

    Inspired by the attention network of the Convolutional Block Attention Module(CBAM)[41],the model added a parallel spatial attention module to the LKA attention network to further enhance the feature extraction capability of the attention network.The overall structure of the attention network is shown in Fig.3a.

    Figure 3:Attention structure

    The structure of the large kernel attention is shown in Fig.3b.A large kernel convolution can be divided into three components: a spatial local convolution (depth-wise convolution),a spatial longrange convolution(depth-wise dilation convolution),and a channel convolution(1×1 convolution),as shown in Eq.(1):

    wherexrepresents the input feature graph,F′represents the intermediate state feature graph,F represents the output feature graph,and ?represents the product of elements.LKA module adopts deep convolution size of 5×5,deep dilation convolution size of 7×7,dilation parameter of 3,and dot product kernel size of 1×1.After decomposition,the three kinds of convolution can obtain the same convolution field of view as 21×21 convolution kernel,effectively obtaining remote information,and the number of parameters of convolution kernel is much smaller than that of 21×21 convolution kernel.

    In the SAM part,the feature maps are respectively subjected to maximum pooling and average pooling,then the two feature maps are fused,and finally,the spatial attention features are obtained through convolution operation.The expression is shown in Eq.(2):

    wherexrepresents the input feature graph,F′represents the intermediate state feature graph,and F represents the output feature graph.By referring to the SAM module in the CBAM network,VANMR designed the parallel attention network of LKA and SAM.The structure of the network is shown in Fig.3a.After the activation function,the feature maps are respectively passed through the LKA module and SAM module,and then the resulting two feature maps are combined and output.Finally,the feature maps are added to the original image through the convolutional layer to form a complete attention module.

    The following pseudocode visually demonstrates the workflow of the attention module.

    In the pseudo-code above,embed_dims is the input dimension,which in the Metaformer structure is the same as the output dimension,so the output dimension is also represented as embed_dims.As can be seen from the pseudo-code,the input feature graph passes through two attention networks,respectively,and finally,the output feature graph is fused to achieve the parallel attention mechanism.

    4.2 Residual Branch

    In the fine-grained classification task of remote sensing ship images,the classification model will be interfered with by the mixed information of multi-grained information,resulting in poor classification performance.The classification model needs to focus on both coarse-grained features and finegrained features of the target,which is the key to accomplishing the fine-grained classification task of remote sensing ships.Therefore,paper proposed a residual structure in parallel with Metaformer.The structure has the ability to extract high-level semantic information,and at the same time,it retains sufficient underlying semantic information for the model.Therefore,it helps the model to achieve complementary underlying and deep semantic information and enhances the image understanding capability of the model.

    The residual branch consists of 3 residual modules in series.The internal structure of the three residual modules is the same.Each residual network receives the fusion feature map generated by the previous residual network and the previous stage as input.In the residual module,the feature map first passes through a 2×2 convolution layer,which downsamples the feature map to 1/2 of its original size,maintaining the same feature map size as the sibling stage,and this operation ensures that the feature maps can be fused in subsequent operations.The featured graph after subsampling will pass through a residual connection,and the connection structure is shown in Fig.1c.The residual connection is composed of three BN layers and two 3×3 convolution layers alternately.Among them,the step size of the 3×3 convolution operation is 1,and the padding is 1.Therefore,the convolution layer will not change the size of the feature graph.Compared with a large number of block structures and convolution operations in the stage part,two 3×3 convolution structures are adopted for residual connection,which avoids the destruction of underlying semantic information based on ensuring the model’s effective extraction of target features.The BN layer can speed up model training and convergence without increasing the number of parameters and excessive memory usage.The operation process of the residual error module can be expressed by the formula as follows:

    wheref(x) represents the output result of the feature graph obtained through the first 2×2 convolution layer,G(x) represents the output result of the feature graph obtained through the internal calculation of the residual connection,andYrepresents the output result of the residual module.Inside the residual connection,the feature graph first passes through the BN layer,then through the first 3×3 convolution layer and the second BN layer,then through the second 3×3 convolution layer and the third BN layer,and finally,the outputG(x) andf(x)are added together to getY,which is the final result of the residual module.Ywill be added with the operation result of the same stage as the input of the next residual module and continue to participate in the feature extraction process.

    The following pseudocode visually illustrates the workflow of parallelizing residual branches.

    In the above pseudo-code,the input dimension differs from the output dimension,and the dimension adjustment is realized through the convolution operation before the residual join.In the residual join,the size of the feature graph will not be changed.The output of the residual connection will be sent to the next stage of the network.

    5 Experiment and Analysis

    In this paper,the effectiveness of the method was verified in a large number of experiments,including sensitivity analysis,comparison test and ablation experiment,and the experimental results were analyzed.

    5.1 Datasets

    To verify the model’s validity in fine-grained tasks of remote sensing ships,paper conducted experiments on typical publicly available fine-grained classification datasets of remote sensing ships,namely FGSC-23 and FGSCR-42.

    FGSC-23 is a high-resolution remote sensing ship image dataset derived from publicly available remote sensing images of Google Earth and GF-2 satellite surface scenes,containing 23 categories of more than 4,000 ship targets.Compared with other ship datasets,FGSC-23 is characterized by various targets,diverse backgrounds,and high precision compared with other ship datasets.

    FGSCR-42 consists of 9,320 remote sensing images with different spatial resolutions,including 42 models from 10 major ship categories.The image size ranges from 50×50 to 1500×1500 pixels.Fig.4 shows some pictures of the FGSC-23 dataset,and Fig.5 shows some of the FGSCR-42 dataset.

    Figure 4:Image of FGSC-23 dataset

    As shown in Figs.4a,4c,and 4f,there are significant differences in the appearance of different types of ships,such as aircraft carriers,cargo ships,and cruisers.The classification model can complete the task according to the global coarse-grained features.However,some different types of ships will also show similar shapes from a specific perspective,as shown in Figs.4b,4d,and 4f.In addition,different subclasses belonging to the same ship class have very similar shape features and superstructures.As shown in Figs.5a and 5d belong to the same aircraft carrier but belong to different subcategories.They have similar appearance and superstructure.In the case of mixed multi-granularity features,the model must pay attention to coarse and fine-grained features to complete the fine-grained classification task.Observing Figs.4 and 5,we can also find that remote sensing ship targets usually contain complex background information.As shown in Figs.4b,4c,and 4e,the background of the ship image in the dock contains a variety of objects.In Fig.4f,Figs.5c,and 5f,the background of the port ship image contains many shore targets.In remote sensing ship images with a complex background,the classification model must accurately locate the target of interest.

    Figure 5:Image of FGSCR-42 dataset

    5.2 Parameter Setting

    The experiment was completed using a 64-bit computer with an NVIDIA-2080 GPU equipped with the Ubuntu20.04 system.In this paper,all the experiments adopted the MMClassification framework,and the model training lasted 100 rounds.batch_size is 8,and the data enhancement methods are BatchMixup,BatchCutMix,RandomResizedCrop,and RandomFlip.

    5.3 Evaluation Index

    In this study,top-1 accuracy,top-5 accuracy,Precision,Recall,and F1-score were used as evaluation values to measure the model’s performance in the data set.Top-1 accuracy refers to the ratio of the number of correct labels included in the first classification probability of all test pictures to the number of total test pictures;top-5 accuracy refers to the ratio of the number of correct labels included in the first five classification probabilities of all test pictures to the number of total test pictures.The calculation formula of the evaluation index is as follows:

    In the above formula,TP represents the number of real targets marked as positive samples by the classification model,FP represents the number of false targets marked as positive samples by the classification model,and FN represents the number of real targets marked as negative samples by the classification model.

    5.4 Results and Analysis

    In order to verify the validity of the model in the fine-grained classification task of remote sensing ships,the paper designed a comparison experiment and an ablation experiment,all of which adopted the same parameter setting and data enhancement methods.In the comparison experiment,We used the pre-improved classification network VAN as the baseline,the classical CNN classification model and Swin-Transformer are selected as comparison models.In the ablation experiment,the paper respectively verify the effectiveness of the improved attention module and parallel residual structure.By comparison of classification accuracy and class activation maps,the effectiveness of the proposed model in the fine-grained classification task of remote sensing ships is demonstrated.

    The variation curves of top-1 and top-5 accuracy during model training in this paper are shown in Fig.6.As seen from the figure,the accuracy increased rapidly in the early stage of training.After training rounds,top-5 accuracy reached the maximum value near 98%and then changed steadily.After 40 rounds of training,top-1 accuracy reached the maximum value near 88%.The final top-1 accuracy reached 88.73%.

    Figure 6:Variation curves of top-1 and top-5 accuracy

    5.4.1 Sensitivity Analysis

    The model can perform better classification tasks in this experiment by adjusting batch size,iteration number,and learning rate attenuation strategy.The experimental results of parameter sensitivity with 50 epochs are shown in Table 2,and the evaluation standard is top-1 accuracy(%).

    Table 2: Experimental results of parameter adjustment during 50 epochs

    The experimental results of parameter sensitivity with 100 epochs are shown in Table 3.

    Table 3: Experimental results of parameter adjustment during 100 epochs

    By comparing the data in Tables 2 and 3,it can be seen that the model has achieved better classification performance in 50 rounds of training,which can also be seen from the trend of accuracy change shown in Fig.5.In addition,Tables 2 and 3 reflect that different batches only impact the model’s progress a little.When batches 2,4,8,and 16 are selected,respectively,the difference in accuracy brought by them is no more than 1.27%,and the best training batch is 8.In addition,different learning rate attenuation strategies significantly impact the model’s training effect.In this experiment,four commonly used attenuation strategies are selected:CosineAnnealingLR,StepLR,LinearLR,and MultiStepLR.The cosine annealing strategy is the widely used learning rate attenuation strategy.This experiment also shows the best training effect.Under the same iteration number and batch conditions,the classification accuracy of the trained model is 0.49%-4.98%higher than that of the model using other attenuation strategies.

    5.4.2 Comparison Experiment

    First,as a pure CNN model,the paper designed a comparison experiment with traditional CNN models such as the VGG,RegNet[42],and ConvNeXt[43].Then,as an efficient classification backbone network,paper design a comparison experiment with the Swin-Transformer network.The experiment also includes a performance comparison with the latest classification backbone networks,specifically RIFormer[44],Rev-vit[45]and EVA-02[46].In addition,the paper shows the comparison results of the number of parameters between the VAN-MR model and the above models,which is convenient for showing the degree of resource consumption of the model.Table 4 shows the top-1 accuracy,top-5 accuracy,and parameter quantity obtained by various networks in fine-grained classification tasks.

    Table 4: Performance of the models on FGSC-23 and FGSCR-42-1

    As can be seen from Table 4,in the FGSC-23 dataset,the accuracy of the VGG network top-1 was 80%,the accuracy of ConvNeXt was 83.15%,and the classification effect of the Swin-small network exceeded that of the traditional CNN network,reaching 86.92%.The top-1 accuracy of Swin-base,more extensive structure network of the same type,is 87.21%.The accuracy of the top-1 model in this paper is 88.73%.Transformer has achieved a good effect compared to the traditional CNN network,and the ConvNeXt network,with similar structure and training skills,has also achieved high accuracy.Compared with the latest backbone network,the proposed model also maintains advantages.The classification accuracy of EVA-02 and Rev-vit is close to that of the proposed model.Still,the number of parameters and time complexity are higher.In contrast,the time complexity of RIFormer is lower,but the classification accuracy is slightly lower than that of the proposed model.In addition,we can find that compared with the benchmark model VAN,the time complexity of the model in this paper has a slight increase.Still,compared with other models,the time complexity of VAN-MR remains at a low level,and at the same time,the classification accuracy of VAN-MR is much greater than that of the benchmark model,surpassing other models.Compared with the above model,the new backbone network,which provides more comprehensive granular information to the model,achieved the best results.

    Compared with other classical networks,it can be found that the top-1 accuracy of the VAN-MR network is 4.49%higher than that of VAN,but the number of parameters only increases by 6.19 M.This is because the parallel residual network provides adequate underlying semantic information,while the number of parameters increased by the simple residual structure is very limited.Compared with Swin-small,the number of VAN-MR parameters is similar,but top-1 accuracy is improved by 2.37%.Swin-base is much closer to VAN-MR in terms of accuracy,with a difference of only 1.52%,but the number of parameters in Swin-base is already much more extensive than VAN-MR.Swinbase uses a deeper network,meaning more blocks and self-attention modules.These structures bring additional parameters.However,the parallel structure of large kernel attention and spatial attention adopted by VAN-MR has only a small number of parameters.However,it provides a compelling image understanding ability,enabling VAN-MR to focus better on the target area of interest,thus improving model performance.

    FGSCR-42 contains more samples than FGSC-23,so the same model will perform better on the FGSCR-42 dataset.In Table 4,we can see that VAN-MR also achieves sota effect.

    Table 5 shows the performance of the proposed model and other models in Precision,Recall and F1-score.As shown in Table 5,the Precision and Recall of VAN-MR have achieved the best performance on the two data.Compared with the benchmark model VAN,VAN-MR has made significant progress.Compared with SWIN series classification networks,the proposed model also has advantages.The network that most closely matches the F1-score of VAN-MR is the swin-base,but VAN-MR has only 62%of the parameters of the swin-base.Through observation and comparison of experimental results,we can find that VAN-MR has more robust feature extraction and feature fusion capabilities,which is more suitable for fine-grained classification tasks of remote sensing ships and has achieved good performance on fine-grained classification data sets of remote sensing ships with complex background.

    Table 5: Performance of the models on FGSC-23 and FGSCR-42-2

    Fig.7 shows the confusion matrix based on the classification results of the four models on the FGSC-23 dataset.The confusion matrix is mainly used for the classification performance of the visualization algorithm,where the horizontal axis is the predicted result,and the vertical axis is the natural result.The darkest regions in the confusion matrix are clustered along the diagonal,indicating successful classification.In the figure,the X-axis of the confusion matrix represents the prediction category,the Y-axis represents the actual category,and the diagonal numbers represent each category’s prediction accuracy.As can be seen from the confusion matrix,Swin and VAN networks have similar classification effects,both of which are better than VGG networks.For categories 12 and 19,Swin has a classification accuracy of 84% and 73%,respectively,while VAN has a classification accuracy of 81% and 55% for the same target.The classification accuracy of the proposed algorithm for the same target is 87% and 92%,respectively,significantly improved compared with the VAN network and exceeds Swin.

    As shown in Fig.7d,the classification accuracy of categories 15,20,and 22 reached 100%,significantly improving the classification effect compared with VANs.Categories 15 and 22 are car carriers and barge,respectively.They appear similar to other ships in remote sensing images,but their superstructures are very different,and the models must fully extract fine-grained features.Category 20 is a passenger ship with many shore-based targets in its background.In this case,the classification model must focus more on the passenger ship and suppress background interference.VAN-MR used the parallel spatial attention and residual method to strengthen the model’s ability to extract multigranularity features and focus on targets of interest,which improved the classification effect.At the same time,we found that the model’s classification accuracy could have been higher for categories 5,13,and 17,which were amphibious transport dock,auxiliary ship,and bulk carrier,respectively.The remote sensing images of the three kinds of ships mentioned above contained pictures of subclasses with very different structures belonging to the same big category(amphibious transport dock/auxiliary ship/bulk carrier).This is a problem for fine-grained classification models,which mistakenly think they belong to different categories,resulting in classification failures.

    By comparing the confusion matrix of VAN and VAN-MR,it can be found that among the 23 types of targets,the detection accuracy of 13 types of targets has improved,demonstrating the excellent classification ability of the proposed model in complex background images.

    Figure 7:Confusion matrices of different models on FGSC-23 data set(a)VGG-16,(b)Swin-small,(c)VAN,(d)VAN-MR

    5.4.3 Comparison Experiment

    In order to understand the different effects of different model structures on the final classification effect,the paper designed an ablation experiment.Table 6 shows the performance of different model structures on the FGSC-23 dataset.

    Table 6: Model accuracy of different configurations

    As shown in Table 6,the top-1 accuracy of the VAN network on the FGSC-23 dataset is 84.24%,and when the parallel SAM module is added alone,the top-1 accuracy is 87.23%,an increase of 2.99%.At the same time,the parameter number of the model is minimal,only 0.01 M,and the computational complexity is only slightly increased.When a residual module is added alone,the top-1 accuracy of the network is 86.06%,an increase of 1.82%.Because the residual convolutional layer is too large,about 9.21 M parameters are introduced,and the time complexity is increased by 0.92 G.When the residual module is added alone,the top-1 accuracy of the network is 86.06%,up by 1.82%.When two improvement schemes were added,the model effect reached the highest value,88.73%.Compared with the baseline model,the classification accuracy of VAN-MR is improved by 4.49%.

    In order to better understand the contribution of different model structures to model classification performance,the experiment shows the class activation maps of models with different structures in the same picture.The class activation maps are used to show which parts of the input image play a role in the classification and judgment of the model,which reflects the model’s ability to understand the image features.Generally speaking,the brightly colored part of the class activation maps is the primary source of features extracted from the network.The class activation maps are shown in Fig.8.As can be seen from the first line in the figure,feature extraction of remote sensing ship targets by the VAN network focuses on fine-grained local features.In ship images with a complex background,the location of feature extraction needs to be more accurate,as shown in the class activation maps of the frigate.The class activation maps with the SAM attention network added are shown in the second line.SAM provides more comprehensive spatial information,and the model can pay attention to more local features but still needs to provide a broad view.The third line shows the class activation maps effect after the residual module is only added.We can see that the residual module retains sufficient underlying semantic information for the model,and the model pays attention to all target information.The target area has the deepest color but contains more background information.The class activation maps of the VAN-MR model are shown in the fourth line.VAN-MR can cover all features of the ship target.Meanwhile,due to the better image understanding ability provided by the attention network,VAN-MR can better suppress the background information.The addition of the SAM and residual modules,as indicated by the class activation maps,improved the model’s ability to extract features,reflected in the classification accuracy,as shown in the table.Therefore,VANMR has strong classification ability in fine-grained data sets of remote sensing ships with complex backgrounds.

    5.4.4 Statistical Tests

    To check the performance changes of the comparison method under different training amounts and the performance fluctuations of the model in multiple experiments,we designed five rounds of experiments for each case,and the accuracy changes obtained in the experiments are shown in Table 7.What’s more,the number of training samples and test samples of FGSC-23 is 8:2,but we split the FGSC-23 by 7:3,6:4,5:5,4:6,3:7 for training and validation.The results are as follow:

    Table 7: Performance comparison with different slitting ratios of dataset

    Figure 8:Class activation maps of different configuration models

    We can discover from the table that in Dataset-7:3,while we have 5 runs for our model,the variance of the results is controlled to within 0.1%.Its accuracy is still very competitive compared with other state-of-the-art models,proving that the model in this paper has high stability.In addition,we designed comparative experiments with different data segmentation ratios.As can be seen from the table,the smaller the ratio of training data to test data,the lower the classification accuracy of the model,but in each case of dataset distribution,compared with other models,the model in this paper maintains the advantage of accuracy.

    6 Discussion

    This paper proposes a fine-grained classification model for remote sensing ship images.Like most image classification models,they all rely on the quality of training data sets.Accurate labeling and explicit image content are the keys to successful model training.Although most deep neural networks can extract the target features of the input images,the input-output mapping learned by the network is discontinuous,and introducing subtle perturbations in the images may lead to model classification errors.Therefore,training the model against attack and attack and defense is significant.The method in this paper has achieved specific results on the open remote sensing ship image classification data set.Still,the premise of the work is that no counterattack has been accepted.How to deal with the attack and strengthen the anti-jamming ability of the model is an essential direction of future research.

    7 Conclusion and Future Scope

    Compared with the natural image classification task,the remote sensing ship fine-grained classification task has two unique characteristics: complex background and mixed multi-grained features.The general CNN network and Transformer structure are used in vision tasks but lack the attention for these unique characteristics.Therefore,the paper proposed a classified backbone network VAN-MR based on the fusion of the Metaformer and residual structure.Specifically,the paper designed a parallel structure of a large core and spatial attention to enhance the feature extraction capability under complex backgrounds.This parallel structure can effectively suppress the interference of complex backgrounds.In addition,the model adopts a concurrent residual structure to extract high-level semantic information and low-level semantic information.This contemporary residual structure retains the mixed multi-granularity characteristics through top-down convolution operation.In general,VAN-MR achieved 88.71%and 94.56%accuracy on FGSC-23 and FGSCR-42 data sets,respectively,while retaining only 50.47 M parameters,providing a solution to the problem of complex background and mixed granularity information in remote sensing ship images.

    The model in this paper has achieved specific results on the public remote sensing ship image classification data set.Still,we found that the background suppression ability of the model will be slightly decreased when only residual structure is introduced.We will explore other CNN structures to pursue better feature extraction and background suppression ability.In the next step,we will further the model’s lightweight transformation to better meet the needs of platform deployment.On the other hand,maintaining a good classification ability after introducing perturbations is also a field that will be explored further.

    Acknowledgement:The authors extend their appreciation to the anonymous reviewers for their constructive comments and suggestions.

    Funding Statement:The authors received no specific funding for this study.

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design: Guoqing Zhou,Liang Huang;data collection: Guoqing Zhou,Qiao Sun;analysis and interpretation of results:Guoqing Zhou,Liang Huang;draft manuscript preparation:Guoqing Zhou,Qiao Sun.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:Not applicable.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    赤兔流量卡办理| 欧美bdsm另类| 嫩草影院新地址| 下体分泌物呈黄色| 国产成人精品一,二区| 女性生殖器流出的白浆| 日本wwww免费看| 美女xxoo啪啪120秒动态图| 热99国产精品久久久久久7| av在线app专区| 亚洲久久久国产精品| 纵有疾风起免费观看全集完整版| 最近最新中文字幕免费大全7| 美女福利国产在线 | 亚洲丝袜综合中文字幕| 国产爱豆传媒在线观看| 国产亚洲最大av| 婷婷色综合www| 丝瓜视频免费看黄片| 国产伦精品一区二区三区四那| 男人添女人高潮全过程视频| 日本vs欧美在线观看视频 | 男女国产视频网站| 亚洲综合精品二区| 熟女人妻精品中文字幕| 在线免费十八禁| 熟妇人妻不卡中文字幕| 国产91av在线免费观看| 日韩电影二区| 亚洲成人av在线免费| 男女啪啪激烈高潮av片| 久久久久久伊人网av| 成人特级av手机在线观看| 偷拍熟女少妇极品色| 91精品伊人久久大香线蕉| 亚洲欧美中文字幕日韩二区| 蜜桃亚洲精品一区二区三区| 在线亚洲精品国产二区图片欧美 | 观看美女的网站| 毛片一级片免费看久久久久| 2021少妇久久久久久久久久久| 国产中年淑女户外野战色| 在线亚洲精品国产二区图片欧美 | 国产成人a∨麻豆精品| 亚洲精品日本国产第一区| 蜜桃亚洲精品一区二区三区| 男人添女人高潮全过程视频| 岛国毛片在线播放| 人人妻人人澡人人爽人人夜夜| 日韩中字成人| videossex国产| 亚洲精品乱码久久久v下载方式| 人妻制服诱惑在线中文字幕| 国内揄拍国产精品人妻在线| videossex国产| av在线app专区| 99视频精品全部免费 在线| 免费观看a级毛片全部| 大码成人一级视频| 99视频精品全部免费 在线| 久久毛片免费看一区二区三区| kizo精华| 网址你懂的国产日韩在线| 老司机影院毛片| 日韩人妻高清精品专区| 亚洲av国产av综合av卡| 日韩人妻高清精品专区| 中文字幕制服av| 免费黄频网站在线观看国产| 80岁老熟妇乱子伦牲交| 精品酒店卫生间| 人妻制服诱惑在线中文字幕| av在线观看视频网站免费| 激情五月婷婷亚洲| 国产精品一区www在线观看| 亚洲精品乱码久久久v下载方式| 国产在线视频一区二区| freevideosex欧美| 久久人人爽人人爽人人片va| 久久99热这里只频精品6学生| 国产成人免费无遮挡视频| 久久人人爽人人爽人人片va| 国国产精品蜜臀av免费| 日本午夜av视频| 国产伦精品一区二区三区视频9| 亚洲伊人久久精品综合| 亚洲伊人久久精品综合| 亚洲自偷自拍三级| av一本久久久久| 一级毛片我不卡| 韩国高清视频一区二区三区| 观看免费一级毛片| 免费看光身美女| 高清黄色对白视频在线免费看 | 在线观看一区二区三区激情| 欧美变态另类bdsm刘玥| 国产精品不卡视频一区二区| 亚洲人成网站在线观看播放| 午夜免费男女啪啪视频观看| 国产av精品麻豆| 久久热精品热| 久久99热这里只有精品18| 激情五月婷婷亚洲| 亚洲精品国产av成人精品| 男人狂女人下面高潮的视频| 精品视频人人做人人爽| 成人亚洲精品一区在线观看 | 精品少妇久久久久久888优播| 中文乱码字字幕精品一区二区三区| 亚洲人成网站在线播| 十八禁网站网址无遮挡 | 偷拍熟女少妇极品色| 蜜桃亚洲精品一区二区三区| 亚洲国产精品999| 久久精品人妻少妇| 爱豆传媒免费全集在线观看| 一二三四中文在线观看免费高清| 亚洲欧洲日产国产| 18禁在线播放成人免费| 欧美少妇被猛烈插入视频| 美女国产视频在线观看| 亚洲国产欧美在线一区| 青春草亚洲视频在线观看| 亚洲av电影在线观看一区二区三区| 六月丁香七月| 大香蕉97超碰在线| 国产精品久久久久成人av| 国产淫片久久久久久久久| 天天躁日日操中文字幕| 国产亚洲精品久久久com| 国产淫片久久久久久久久| 午夜日本视频在线| 男女免费视频国产| 久久热精品热| 亚洲熟女精品中文字幕| 97在线视频观看| 久久国产亚洲av麻豆专区| 国产av国产精品国产| 欧美成人a在线观看| 精品久久久久久久末码| 国产免费一区二区三区四区乱码| 亚洲国产精品成人久久小说| 日韩 亚洲 欧美在线| 美女高潮的动态| 免费看日本二区| 久久精品久久久久久久性| 国产v大片淫在线免费观看| 亚洲av成人精品一区久久| 免费黄频网站在线观看国产| 18+在线观看网站| 国产精品人妻久久久久久| 免费av不卡在线播放| 自拍偷自拍亚洲精品老妇| 欧美精品国产亚洲| 最近中文字幕高清免费大全6| 国产精品久久久久久精品电影小说 | 一级片'在线观看视频| 欧美xxxx黑人xx丫x性爽| 国产亚洲精品久久久com| 三级国产精品片| 韩国高清视频一区二区三区| 最近最新中文字幕大全电影3| 亚洲色图av天堂| 91aial.com中文字幕在线观看| 亚洲精品色激情综合| 天天躁日日操中文字幕| 国产爽快片一区二区三区| 国产毛片在线视频| 亚洲精品乱码久久久久久按摩| 777米奇影视久久| 最后的刺客免费高清国语| 欧美高清性xxxxhd video| 99re6热这里在线精品视频| 精品人妻偷拍中文字幕| 777米奇影视久久| 日韩亚洲欧美综合| 亚洲第一av免费看| 高清午夜精品一区二区三区| 国产成人91sexporn| 亚洲高清免费不卡视频| 亚洲精品乱码久久久v下载方式| 性高湖久久久久久久久免费观看| 久久久精品94久久精品| 久久久久久久大尺度免费视频| 伊人久久精品亚洲午夜| 国产男人的电影天堂91| a 毛片基地| 人妻制服诱惑在线中文字幕| 少妇的逼好多水| 日韩欧美精品免费久久| 又黄又爽又刺激的免费视频.| 纵有疾风起免费观看全集完整版| 麻豆国产97在线/欧美| 国产精品一区www在线观看| 日韩电影二区| 国产一区二区三区av在线| 在线观看免费高清a一片| 国产成人精品一,二区| 国产一区二区三区av在线| 日本一二三区视频观看| 亚洲国产精品999| 亚洲av男天堂| 免费看av在线观看网站| 边亲边吃奶的免费视频| 性色avwww在线观看| 国产在线男女| 最近最新中文字幕免费大全7| 三级国产精品片| 欧美区成人在线视频| 在线观看国产h片| 国产一级毛片在线| 亚洲av国产av综合av卡| 色婷婷久久久亚洲欧美| 看十八女毛片水多多多| 少妇裸体淫交视频免费看高清| 又黄又爽又刺激的免费视频.| 欧美日韩在线观看h| 亚洲四区av| 伦精品一区二区三区| 十分钟在线观看高清视频www | 欧美极品一区二区三区四区| 亚洲激情五月婷婷啪啪| 色婷婷久久久亚洲欧美| 黑丝袜美女国产一区| 国产精品一区二区在线不卡| 欧美97在线视频| 精品久久久久久久久亚洲| 亚洲av免费高清在线观看| 亚洲国产日韩一区二区| 亚洲精品国产成人久久av| 99久久人妻综合| 亚洲av中文av极速乱| 六月丁香七月| 成人高潮视频无遮挡免费网站| 国产男女内射视频| 久久精品国产a三级三级三级| 亚洲精品乱码久久久久久按摩| 欧美成人一区二区免费高清观看| www.av在线官网国产| 2018国产大陆天天弄谢| 高清在线视频一区二区三区| 精华霜和精华液先用哪个| 亚洲欧美中文字幕日韩二区| 亚洲电影在线观看av| 久久综合国产亚洲精品| 久久毛片免费看一区二区三区| 高清日韩中文字幕在线| 欧美成人午夜免费资源| av国产免费在线观看| 国产老妇伦熟女老妇高清| 久久久久国产网址| 国产综合精华液| 国产免费又黄又爽又色| 中国三级夫妇交换| 亚洲精品乱码久久久v下载方式| 午夜福利在线在线| 欧美精品亚洲一区二区| 高清黄色对白视频在线免费看 | 亚洲国产精品国产精品| 自拍偷自拍亚洲精品老妇| 国内精品宾馆在线| 亚洲色图av天堂| 亚洲av电影在线观看一区二区三区| 成人美女网站在线观看视频| 色吧在线观看| 亚洲国产精品国产精品| 日韩精品有码人妻一区| 韩国高清视频一区二区三区| 亚洲一区二区三区欧美精品| 国精品久久久久久国模美| 日韩视频在线欧美| 爱豆传媒免费全集在线观看| 极品少妇高潮喷水抽搐| 亚洲怡红院男人天堂| 人妻系列 视频| 午夜福利在线观看免费完整高清在| 91在线精品国自产拍蜜月| 欧美变态另类bdsm刘玥| 有码 亚洲区| 老女人水多毛片| 欧美bdsm另类| 亚洲成人一二三区av| 边亲边吃奶的免费视频| 五月天丁香电影| 亚洲成人av在线免费| 久久精品久久精品一区二区三区| 国产精品99久久久久久久久| 美女xxoo啪啪120秒动态图| 男女无遮挡免费网站观看| 久久久久久久久大av| 亚洲成人一二三区av| 久久热精品热| 熟女人妻精品中文字幕| 欧美精品人与动牲交sv欧美| 国产淫片久久久久久久久| 99视频精品全部免费 在线| 欧美zozozo另类| 欧美日本视频| 精品一区在线观看国产| 小蜜桃在线观看免费完整版高清| 欧美精品国产亚洲| 国产成人a区在线观看| 九九久久精品国产亚洲av麻豆| 欧美丝袜亚洲另类| 精品人妻偷拍中文字幕| 黄色视频在线播放观看不卡| 1000部很黄的大片| 久久久色成人| 久久人妻熟女aⅴ| 欧美成人一区二区免费高清观看| 免费av不卡在线播放| 亚洲色图综合在线观看| 内射极品少妇av片p| 久久久精品94久久精品| 另类亚洲欧美激情| 国产视频首页在线观看| 少妇的逼水好多| xxx大片免费视频| 中文在线观看免费www的网站| 国内揄拍国产精品人妻在线| 高清毛片免费看| 深夜a级毛片| 99热这里只有是精品在线观看| 精品人妻偷拍中文字幕| 亚洲欧美一区二区三区国产| 日韩av在线免费看完整版不卡| 久久久久网色| 91久久精品电影网| 亚洲国产色片| 肉色欧美久久久久久久蜜桃| 免费观看的影片在线观看| 久久久久久久久久人人人人人人| 国产精品久久久久久av不卡| .国产精品久久| 卡戴珊不雅视频在线播放| 国产伦在线观看视频一区| 丰满乱子伦码专区| 蜜桃在线观看..| 欧美国产精品一级二级三级 | 成人毛片a级毛片在线播放| 国产在视频线精品| 国产精品久久久久久久电影| 91精品伊人久久大香线蕉| 看免费成人av毛片| 久久99热这里只频精品6学生| 尾随美女入室| 久久久久久久精品精品| 九草在线视频观看| av卡一久久| 国产毛片在线视频| 国产黄片视频在线免费观看| 国产精品一区二区三区四区免费观看| 免费观看无遮挡的男女| 性高湖久久久久久久久免费观看| 亚洲婷婷狠狠爱综合网| 高清午夜精品一区二区三区| 国产精品熟女久久久久浪| 国产精品久久久久久久电影| 久久久色成人| 国内少妇人妻偷人精品xxx网站| 成年人午夜在线观看视频| 三级国产精品欧美在线观看| 国产片特级美女逼逼视频| 国产精品一区二区性色av| 国产亚洲精品久久久com| 亚洲电影在线观看av| 中文字幕精品免费在线观看视频 | 午夜激情福利司机影院| 自拍偷自拍亚洲精品老妇| 建设人人有责人人尽责人人享有的 | 少妇的逼好多水| 男女无遮挡免费网站观看| 久久久国产一区二区| 国产淫语在线视频| 欧美另类一区| 日韩成人伦理影院| 人人妻人人添人人爽欧美一区卜 | 一级a做视频免费观看| 成人一区二区视频在线观看| 狂野欧美激情性xxxx在线观看| 午夜免费男女啪啪视频观看| 免费大片18禁| 看免费成人av毛片| 欧美xxxx黑人xx丫x性爽| 黑丝袜美女国产一区| 丰满少妇做爰视频| 天堂中文最新版在线下载| 在现免费观看毛片| 日韩av不卡免费在线播放| 午夜激情福利司机影院| 欧美精品一区二区免费开放| 午夜精品国产一区二区电影| 日本wwww免费看| 新久久久久国产一级毛片| 久久久久久久久大av| 观看免费一级毛片| 日本猛色少妇xxxxx猛交久久| 日韩欧美一区视频在线观看 | 精品久久久久久久久亚洲| 欧美日韩视频精品一区| 日本欧美国产在线视频| 亚洲精品一区蜜桃| 最黄视频免费看| 色视频www国产| 日本黄大片高清| 国产免费一区二区三区四区乱码| 精品一区二区三卡| 久久99热这里只有精品18| 亚洲人成网站在线播| av免费观看日本| 国产一级毛片在线| 国产真实伦视频高清在线观看| 男人爽女人下面视频在线观看| 在线天堂最新版资源| 嫩草影院入口| 一个人看的www免费观看视频| 啦啦啦中文免费视频观看日本| 国产精品国产av在线观看| 精品午夜福利在线看| 精品人妻一区二区三区麻豆| 99热国产这里只有精品6| 久久人人爽人人片av| 99精国产麻豆久久婷婷| av国产精品久久久久影院| 亚洲无线观看免费| 精品久久久久久久久av| 一个人免费看片子| 欧美日韩精品成人综合77777| 秋霞在线观看毛片| 日韩欧美 国产精品| 欧美xxxx黑人xx丫x性爽| 国产在线视频一区二区| 丰满乱子伦码专区| 少妇精品久久久久久久| 又大又黄又爽视频免费| 最近2019中文字幕mv第一页| 国产精品久久久久久久电影| 免费观看无遮挡的男女| 久久av网站| 久久国内精品自在自线图片| 一级a做视频免费观看| 日日啪夜夜爽| 免费播放大片免费观看视频在线观看| 人妻系列 视频| av网站免费在线观看视频| 国产在视频线精品| 国产精品99久久久久久久久| 精品久久久久久久久av| 久久精品国产a三级三级三级| 观看av在线不卡| 亚洲欧美一区二区三区国产| 岛国毛片在线播放| 久久影院123| 久久久a久久爽久久v久久| av免费观看日本| 麻豆精品久久久久久蜜桃| 大话2 男鬼变身卡| 能在线免费看毛片的网站| 在线播放无遮挡| 成人国产av品久久久| 久久久久精品性色| 黄色日韩在线| 久久国内精品自在自线图片| 久久久亚洲精品成人影院| 午夜激情福利司机影院| 久久精品久久久久久噜噜老黄| 国产日韩欧美亚洲二区| 国产精品三级大全| 黄色配什么色好看| 五月天丁香电影| 只有这里有精品99| 97在线人人人人妻| 欧美另类一区| 亚洲色图av天堂| 久久久色成人| 深夜a级毛片| 极品少妇高潮喷水抽搐| a级毛片免费高清观看在线播放| 亚洲,欧美,日韩| 王馨瑶露胸无遮挡在线观看| 精品人妻偷拍中文字幕| 国产欧美日韩一区二区三区在线 | 欧美最新免费一区二区三区| 丰满迷人的少妇在线观看| 99re6热这里在线精品视频| 国精品久久久久久国模美| 国产白丝娇喘喷水9色精品| 王馨瑶露胸无遮挡在线观看| 精品国产一区二区三区久久久樱花 | 性色avwww在线观看| 寂寞人妻少妇视频99o| 国产淫片久久久久久久久| 交换朋友夫妻互换小说| 午夜激情福利司机影院| 2022亚洲国产成人精品| 麻豆精品久久久久久蜜桃| 久久99蜜桃精品久久| 亚洲精品456在线播放app| 欧美日韩视频精品一区| 91精品国产九色| 精品久久久久久久久亚洲| 18禁动态无遮挡网站| 国产黄频视频在线观看| 天堂俺去俺来也www色官网| 蜜臀久久99精品久久宅男| 国产中年淑女户外野战色| 亚洲一区二区三区欧美精品| 最近2019中文字幕mv第一页| 国产爱豆传媒在线观看| 亚洲欧美一区二区三区国产| 国产永久视频网站| 最近2019中文字幕mv第一页| 91午夜精品亚洲一区二区三区| 51国产日韩欧美| 亚洲精品自拍成人| 最近2019中文字幕mv第一页| 男女啪啪激烈高潮av片| 国产中年淑女户外野战色| 自拍偷自拍亚洲精品老妇| 美女高潮的动态| 免费观看a级毛片全部| 18禁动态无遮挡网站| 久久精品久久久久久噜噜老黄| 狠狠精品人妻久久久久久综合| 3wmmmm亚洲av在线观看| 欧美高清性xxxxhd video| 久久久久久伊人网av| 在线观看人妻少妇| 亚洲av福利一区| 精品久久久噜噜| 国产一区亚洲一区在线观看| 又爽又黄a免费视频| 丰满人妻一区二区三区视频av| 一区二区三区乱码不卡18| 日本黄大片高清| 亚洲欧美日韩另类电影网站 | 毛片一级片免费看久久久久| 你懂的网址亚洲精品在线观看| 麻豆成人午夜福利视频| 日韩av不卡免费在线播放| 国产精品一二三区在线看| 成人黄色视频免费在线看| 久久久久久久大尺度免费视频| 日韩制服骚丝袜av| 国产成人精品一,二区| 日本猛色少妇xxxxx猛交久久| a级一级毛片免费在线观看| 多毛熟女@视频| 另类亚洲欧美激情| 美女xxoo啪啪120秒动态图| 成人高潮视频无遮挡免费网站| 久久韩国三级中文字幕| 国产精品.久久久| 一级毛片我不卡| 在线看a的网站| 乱系列少妇在线播放| 亚洲精品久久久久久婷婷小说| 干丝袜人妻中文字幕| 最黄视频免费看| 久久精品国产a三级三级三级| 亚洲av男天堂| 国产免费又黄又爽又色| 久久久精品94久久精品| 国产成人一区二区在线| 欧美zozozo另类| 美女脱内裤让男人舔精品视频| 中文字幕精品免费在线观看视频 | 国产男女内射视频| 欧美高清成人免费视频www| 一级毛片我不卡| 欧美精品一区二区免费开放| 精品一区二区三卡| 久久韩国三级中文字幕| 2022亚洲国产成人精品| 国产大屁股一区二区在线视频| 久久久久人妻精品一区果冻| 欧美日韩一区二区视频在线观看视频在线| 男人狂女人下面高潮的视频| 亚洲自偷自拍三级| 色综合色国产| 日韩一区二区三区影片| 丰满人妻一区二区三区视频av| 伊人久久精品亚洲午夜| 国产美女午夜福利| 美女国产视频在线观看| 精品少妇黑人巨大在线播放| 老司机影院成人| 搡女人真爽免费视频火全软件| 国产成人午夜福利电影在线观看| 日韩免费高清中文字幕av| 日本午夜av视频| 成人午夜精彩视频在线观看| 亚洲va在线va天堂va国产| 国产一区二区三区av在线| 久久婷婷青草| 一区二区三区乱码不卡18| 免费观看a级毛片全部| 国产av一区二区精品久久 | 成人免费观看视频高清| 麻豆精品久久久久久蜜桃| 国产成人精品婷婷| 夫妻性生交免费视频一级片| 美女高潮的动态| 麻豆国产97在线/欧美| 九九在线视频观看精品| 国产精品蜜桃在线观看| 国产亚洲午夜精品一区二区久久| 美女脱内裤让男人舔精品视频| 国产v大片淫在线免费观看| 欧美少妇被猛烈插入视频| av.在线天堂| 九九爱精品视频在线观看| 国产精品女同一区二区软件| 日韩欧美 国产精品| 乱码一卡2卡4卡精品|